hexsha stringlengths 40 40 | size int64 7 1.04M | ext stringclasses 10 values | lang stringclasses 1 value | max_stars_repo_path stringlengths 4 247 | max_stars_repo_name stringlengths 4 125 | max_stars_repo_head_hexsha stringlengths 40 78 | max_stars_repo_licenses listlengths 1 10 | max_stars_count int64 1 368k ⌀ | max_stars_repo_stars_event_min_datetime stringlengths 24 24 ⌀ | max_stars_repo_stars_event_max_datetime stringlengths 24 24 ⌀ | max_issues_repo_path stringlengths 4 247 | max_issues_repo_name stringlengths 4 125 | max_issues_repo_head_hexsha stringlengths 40 78 | max_issues_repo_licenses listlengths 1 10 | max_issues_count int64 1 116k ⌀ | max_issues_repo_issues_event_min_datetime stringlengths 24 24 ⌀ | max_issues_repo_issues_event_max_datetime stringlengths 24 24 ⌀ | max_forks_repo_path stringlengths 4 247 | max_forks_repo_name stringlengths 4 125 | max_forks_repo_head_hexsha stringlengths 40 78 | max_forks_repo_licenses listlengths 1 10 | max_forks_count int64 1 105k ⌀ | max_forks_repo_forks_event_min_datetime stringlengths 24 24 ⌀ | max_forks_repo_forks_event_max_datetime stringlengths 24 24 ⌀ | content stringlengths 1 1.04M | avg_line_length float64 1.77 618k | max_line_length int64 1 1.02M | alphanum_fraction float64 0 1 | original_content stringlengths 7 1.04M | filtered:remove_function_no_docstring int64 -102 942k | filtered:remove_class_no_docstring int64 -354 977k | filtered:remove_delete_markers int64 0 60.1k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
688d4348ba03f88567d8b80507a9dd716dacb9bc | 404 | py | Python | ibench/benchmarks/lu.py | rscohn2/IDP-bench | 1b7540276369da6146b3cacae780dae3ffba73da | [
"MIT"
] | null | null | null | ibench/benchmarks/lu.py | rscohn2/IDP-bench | 1b7540276369da6146b3cacae780dae3ffba73da | [
"MIT"
] | null | null | null | ibench/benchmarks/lu.py | rscohn2/IDP-bench | 1b7540276369da6146b3cacae780dae3ffba73da | [
"MIT"
] | null | null | null | import numpy as np
import scipy
from .bench import Bench
| 23.764706 | 87 | 0.641089 | import numpy as np
import scipy
from .bench import Bench
class Lu(Bench):
sizes = {'large': 35000, 'small': 20000, 'test': 2}
def _ops(self, n):
return 2./3.*n*n*n*1e-9
def _make_args(self, n):
self._A = np.asfortranarray(np.asarray(np.random.rand(n,n), dtype=self._dtype))
def _compute(self):
scipy.linalg.lu(a=self._A, overwrite_a=True, check_finite=False)
| 191 | 132 | 23 |
e3f69d32d00a3bb7eba166c677394c55c97dc3c6 | 6,928 | py | Python | waldur_batch_offering.py | ServerAnt/ansible-waldur-module | 4406a176a9a11f37e37affb6ccc0478f15ad1081 | [
"MIT"
] | null | null | null | waldur_batch_offering.py | ServerAnt/ansible-waldur-module | 4406a176a9a11f37e37affb6ccc0478f15ad1081 | [
"MIT"
] | null | null | null | waldur_batch_offering.py | ServerAnt/ansible-waldur-module | 4406a176a9a11f37e37affb6ccc0478f15ad1081 | [
"MIT"
] | null | null | null | #!/usr/bin/python
from ansible.module_utils.basic import AnsibleModule
from waldur_client import WaldurClientException, waldur_client_from_module
ANSIBLE_METADATA = {
'metadata_version': '1.1',
'status': ['preview'],
'supported_by': 'OpenNode',
}
DOCUMENTATION = '''
---
options:
api_url:
description:
- Fully qualified URL to the Waldur.
required: true
access_token:
description:
- An access token which has permissions to create offerings.
required: true
name:
description:
- The name of the new batch offering.
required: true
description:
description:
- The description of the new batch offering.
full_description:
description:
- The full description of the new batch offering.
native_name:
description:
- The native name of the new batch offering.
native_description:
description:
- The native description of the new batch offering.
terms_of_service:
description:
- The terms of the service.
provider:
description:
- The name or UUID of the provider organization.
rating:
description:
- The rating form 1 to 5.
category:
description:
- The name or UUID of the marketplace category.
required: true
attributes:
description:
- The attributes of the marketplace category.
geolocations:
description:
- The geolocations of provided service.
plans:
description:
- The list of plans attached to offering.
required: true
batch_service:
description:
- The batch service of allocation (MOAB or SLURM).
required: true
hostname:
description:
- Hostname or IP address of master node.
required: true
username:
description:
- Username for SSH connection.
required: true
port:
description:
- Port of the master node process.
gateway:
description:
- Hostname or IP address of gateway node.
use_sudo:
description:
- The flag for privilege escalation activation.
default_account:
description:
- Default SLURM account for user.
required: true
shared:
description:
- The flag of a possible access to all organizations.
billable:
description:
- The flag if purchase and usage is invoiced.
datacite_doi:
description:
- Persistent ID for the service.
'''
EXAMPLES = '''
---
- hosts: localhost
gather_facts: no
tasks:
- name: Create HPC allocation
check_mode: no
waldur_batch_offering:
access_token: 3cecd050b7c1dbce54bc45bf508ec836aecdd5bc
api_url: http://193.40.155.148:8000/api
name: Offering sample name
native_name: Offering sample native name
description: Offering description
native_description: Offering native description
full_description: Offering full description
terms_of_service: Sample terms
category: dfe535350663441e99615b0be014c742
provider: cba70dd7f75d40dc9650704198e9071a
batch_service: SLURM
hostname: localhost
port: 8080
username: user
default_account: root
gateway: localhost
shared: true
plans:
- name: HPC plan
unit: month
prices:
cpu: 100
gpu: 70
ram: 64
geolocations:
- latitude: 59.3990796
longitude: 26.6625565
'''
if __name__ == '__main__':
main()
| 30.121739 | 83 | 0.587904 | #!/usr/bin/python
from ansible.module_utils.basic import AnsibleModule
from waldur_client import WaldurClientException, waldur_client_from_module
ANSIBLE_METADATA = {
'metadata_version': '1.1',
'status': ['preview'],
'supported_by': 'OpenNode',
}
DOCUMENTATION = '''
---
options:
api_url:
description:
- Fully qualified URL to the Waldur.
required: true
access_token:
description:
- An access token which has permissions to create offerings.
required: true
name:
description:
- The name of the new batch offering.
required: true
description:
description:
- The description of the new batch offering.
full_description:
description:
- The full description of the new batch offering.
native_name:
description:
- The native name of the new batch offering.
native_description:
description:
- The native description of the new batch offering.
terms_of_service:
description:
- The terms of the service.
provider:
description:
- The name or UUID of the provider organization.
rating:
description:
- The rating form 1 to 5.
category:
description:
- The name or UUID of the marketplace category.
required: true
attributes:
description:
- The attributes of the marketplace category.
geolocations:
description:
- The geolocations of provided service.
plans:
description:
- The list of plans attached to offering.
required: true
batch_service:
description:
- The batch service of allocation (MOAB or SLURM).
required: true
hostname:
description:
- Hostname or IP address of master node.
required: true
username:
description:
- Username for SSH connection.
required: true
port:
description:
- Port of the master node process.
gateway:
description:
- Hostname or IP address of gateway node.
use_sudo:
description:
- The flag for privilege escalation activation.
default_account:
description:
- Default SLURM account for user.
required: true
shared:
description:
- The flag of a possible access to all organizations.
billable:
description:
- The flag if purchase and usage is invoiced.
datacite_doi:
description:
- Persistent ID for the service.
'''
EXAMPLES = '''
---
- hosts: localhost
gather_facts: no
tasks:
- name: Create HPC allocation
check_mode: no
waldur_batch_offering:
access_token: 3cecd050b7c1dbce54bc45bf508ec836aecdd5bc
api_url: http://193.40.155.148:8000/api
name: Offering sample name
native_name: Offering sample native name
description: Offering description
native_description: Offering native description
full_description: Offering full description
terms_of_service: Sample terms
category: dfe535350663441e99615b0be014c742
provider: cba70dd7f75d40dc9650704198e9071a
batch_service: SLURM
hostname: localhost
port: 8080
username: user
default_account: root
gateway: localhost
shared: true
plans:
- name: HPC plan
unit: month
prices:
cpu: 100
gpu: 70
ram: 64
geolocations:
- latitude: 59.3990796
longitude: 26.6625565
'''
def format_params(params):
excluded_keys = [
'batch_service',
'hostname',
'username',
'default_account',
'port',
'gateway',
'type',
'provider',
]
formatted_params = {
k: v for (k, v) in params.items() if v and k not in excluded_keys
}
formatted_params['type'] = 'SlurmInvoices.SlurmPackage'
if params.get('provider'):
formatted_params['customer'] = params['provider']
service_attributes = {
'batch_service': params['batch_service'],
'hostname': params['hostname'],
'username': params['username'],
'default_account': params['default_account'],
}
if params.get('port'):
service_attributes['port'] = params['port']
if params.get('gateway'):
service_attributes['gateway'] = params['gateway']
formatted_params['service_attributes'] = service_attributes
return formatted_params
def send_request_to_waldur(client, module):
params = format_params(module.params)
offering, changed = client.create_offering(params, module.check_mode)
return offering, changed
def main():
fields = {
'api_url': {'required': True, 'type': 'str'},
'access_token': {'required': True, 'type': 'str'},
# Overview
'name': {'required': True, 'type': 'str'},
'description': {'required': False, 'type': 'str'},
'full_description': {'required': False, 'type': 'str'},
'native_name': {'required': False, 'type': 'str'},
'native_description': {'required': False, 'type': 'str'},
'terms_of_service': {'required': False, 'type': 'str'},
# Organization details
'provider': {'required': False, 'type': 'str'},
'rating': {'required': False, 'type': 'int', 'choices': list(range(1, 6))},
# Description
'category': {'required': True, 'type': 'str'},
'attributes': {'required': False, 'type': 'dict'}, # category attributes
'geolocations': {'required': False, 'type': 'list'},
# Accounting
'plans': {'required': True, 'type': 'list'},
# Management
'batch_service': {
'required': True,
'type': 'str',
'choices': ['SLURM', 'MOAB'],
},
'hostname': {'required': True, 'type': 'str'},
'username': {'required': True, 'type': 'str'},
'port': {'required': False, 'type': 'int'},
'gateway': {'required': False, 'type': 'str'},
'use_sudo': {'required': False, 'type': 'str'},
'default_account': {'required': True, 'type': 'str'},
'shared': {'required': False, 'type': 'bool'},
'billable': {'required': False, 'type': 'bool'},
'datacite_doi': {'required': False, 'type': 'str'},
}
module = AnsibleModule(argument_spec=fields, supports_check_mode=True)
client = waldur_client_from_module(module)
try:
offering, changed = send_request_to_waldur(client, module)
module.exit_json(offering=offering, changed=changed)
except WaldurClientException as e:
module.fail_json(msg=str(e))
if __name__ == '__main__':
main()
| 3,073 | 0 | 69 |
abbe93d8493620ffe594b9b4c940094d973e1987 | 4,282 | py | Python | ConvNet.py | Gregory-Eales/Convolutional-Neural-Network | 50c07208f59526b9a86a162d8b356e9e19dd3310 | [
"MIT"
] | 1 | 2019-05-03T21:48:34.000Z | 2019-05-03T21:48:34.000Z | ConvNet.py | Gregory-Eales/Convolutional-Neural-Network | 50c07208f59526b9a86a162d8b356e9e19dd3310 | [
"MIT"
] | null | null | null | ConvNet.py | Gregory-Eales/Convolutional-Neural-Network | 50c07208f59526b9a86a162d8b356e9e19dd3310 | [
"MIT"
] | null | null | null | import numpy as np
# loading bar imports
import os
from tqdm import tqdm
import time
# Convolution Functions
# single convolutional operation
# Pool Funtions
| 32.439394 | 96 | 0.569827 | import numpy as np
# loading bar imports
import os
from tqdm import tqdm
import time
class ConvNet(object):
def __init__(self, x, y):
# define network topology
# remember to change # of each type of layers
# topology parameter?
self.x_size = x.shape
self.y_size = y.shape
self.num_pools = 1
self.num_convolutions = 1
self.num_feed_forward = 1
# create initital convolutional weights
self.conv_weights = {}
self.conv_weights["w1"] = np.random.random([28, 28])
# create intitial pooling weights
self.pool_weights = {}
# create initial feed forward weights
self.feed_forward_weights = {}
def sigmoid(self, z):
return 1/(1+np.exp(-z))
def sigmoid_prime(self, z):
return z * np.exp(-1) / np.square(1+np.exp(-z))
def optimize(self, iterations, self.x):
# implement minibatches
print("Training Convolutional Neural Net: ")
mini_batch = [1, 2, 3, 4, 5, 6, 7, 8, 9]
for i in range(iterations):
for load_bar in tqdm(mini_batch):
# train neural net
#self.convolutional_forward(x, w, b, step)
# make prediction
# update the wieghts based on cost funtion
time.sleep(0.7)
# Convolution Functions
def pad(self, array, amount):
# amount padds that amount in each direction
# be careful with 3d vectors!!!!
array = np.pad(array, [(amount, amount), (amount, amount)], mode="constant")
return array
# single convolutional operation
def convolve_window(self, x_window, w, b):
return np.sum(np.multiply(x_window, w)) + b
def convolutional_forward(self, x, w, b, step):
output_height = w.shape[0]
otuput_width = w.shape[1]
output_length = w.shape[2]
# apply convolution formula for the number of windows generated by convolution operation
num_x_windows = (x.shape[1] - output_width)/step + 1
num_y_windows = (x.shape[0] - output_height)/step + 1
num_z_windows = output_length
# check to see if paramaters result in a vailid convolution
#if type(num_x_windows) != int or type(num_y_windows) != int:
#raise ValueError:
#print("Error: step size is not valid or insufficient padding")
# creat output vector
z = np.zeros([num_x_windows, num_y_windows, num_z_windows])
# loop through each window and apply convolution to it
for i in range(num_x_windows):
# create x window coordinates
x_start = i*output_width+step*i
x_end = (i+1)*output_width+step*i
for j in range(num_y_windows):
# create y window coordinates
y_start = j*output_length+step*i
y_end = (j+1)*output_length+step*j
for k in range(num_z_windows):
# calculate a window element
# and add it to the output vector
x_window = []
x_window = self.convolve_window(x_window, w, b)
z[i][j][k] = x_window
# loop through convolution windows and apply num_convolutions
# store convolutions in the next z and then apply activations function to them
# dont forget to apply normalization and center data around the origin
def convolutional_backward(self):
# get update equation by taking derivative
# use weights, bias, and regularization
for i in range(len(self.w)):
# update the weights.
pass
# Pool Funtions
def pool_forward(self, a, topology):
output = np.zeros([x, y, z])
for i in range(x):
for j in range(y):
for k in range(z):
window = z[l*x:(1 + x) * l][l*y:(1 + x) * l][l*j:(1 + x) * l]
pool = np.sum(window)/window.size
output[i][j][k] = pool
return output
def pool_backward(self):
pass
def create_mask(self):
pass
def distribute_value(self):
pass
| 3,743 | 1 | 346 |
aac0f27d69e7306681029adb483e49112422799f | 1,156 | py | Python | make_sqls.py | na0fu3y/bqfunc | a77554c6160b37e01de6dbb6e36dd499acb17bf4 | [
"MIT"
] | 5 | 2020-11-30T13:25:05.000Z | 2022-02-25T12:29:39.000Z | make_sqls.py | na0fu3y/bqfunc | a77554c6160b37e01de6dbb6e36dd499acb17bf4 | [
"MIT"
] | null | null | null | make_sqls.py | na0fu3y/bqfunc | a77554c6160b37e01de6dbb6e36dd499acb17bf4 | [
"MIT"
] | null | null | null | import datetime
import urllib.request
from pathlib import Path
if __name__ == "__main__":
make_holidays_sql()
| 28.9 | 109 | 0.619377 | import datetime
import urllib.request
from pathlib import Path
def fetch_holidays():
# 内閣府「国民の祝日」
# https://www8.cao.go.jp/chosei/shukujitsu/gaiyou.html
url = "https://www8.cao.go.jp/chosei/shukujitsu/syukujitsu.csv"
with urllib.request.urlopen(url) as res:
body = res.readlines()
# header を除去
a = iter(body)
next(a)
for i in a:
dt, name = i.decode("shift-jis").split(",")
# dtはyyyy-mm-ddなら自動でdate解釈される。name末尾の改行削除。
yield datetime.datetime.strptime(dt, "%Y/%m/%d").strftime(
"%Y-%m-%d"
), name[:-2]
def make_holidays_sql():
date_and_name_sql = ",".join(
f'("{date}","{name}")' for date, name in fetch_holidays()
)
sql = (
f"CREATE OR REPLACE FUNCTION holidays_in_japan.HOLIDAYS()"
f"AS([STRUCT<date DATE,name STRING>{date_and_name_sql}])"
f'OPTIONS(description = "内閣府「国民の祝日」のデータを返す。参考 https://www8.cao.go.jp/chosei/shukujitsu/gaiyou.html")'
)
with (Path(__file__).parent / "udfs/holidays_in_japan/holidays.sql").open(
"w"
) as f:
f.write(sql)
if __name__ == "__main__":
make_holidays_sql()
| 1,098 | 0 | 46 |
dce5baf6c786a397b5a9edbf5f8a799edd0e5d8c | 1,299 | py | Python | login/login.py | nju161250102/NJUSystem_Server | 8b72fde8701d55e62ce8ed4bcb299ec8f0ab32fe | [
"MIT"
] | 1 | 2018-07-31T13:31:52.000Z | 2018-07-31T13:31:52.000Z | login/login.py | nju161250102/NJUSystem_Server | 8b72fde8701d55e62ce8ed4bcb299ec8f0ab32fe | [
"MIT"
] | null | null | null | login/login.py | nju161250102/NJUSystem_Server | 8b72fde8701d55e62ce8ed4bcb299ec8f0ab32fe | [
"MIT"
] | null | null | null | # -*- coding:utf-8 -*-
from flask import Blueprint
from flask import request
from flask import Response
import json
import requests
loginModule = Blueprint('login', __name__)
@loginModule.route('/', methods=['POST'])
# 获得验证码图片并保存cookie, 验证码对应登录网址为http://cer.nju.edu.cn/amserver/UI/Login
@loginModule.route('/cer_code', methods=['GET'])
| 30.928571 | 116 | 0.664357 | # -*- coding:utf-8 -*-
from flask import Blueprint
from flask import request
from flask import Response
import json
import requests
loginModule = Blueprint('login', __name__)
@loginModule.route('/', methods=['POST'])
def login():
req_data = json.loads(request.data)
post_data = {
"IDToken1": req_data["name"],
"IDToken2": req_data["key"],
"inputCode": req_data["code"],
"encoded": "false"
}
# 注入验证码对应的cookie
jar = requests.utils.cookiejar_from_dict(request.cookies)
# 禁止重定向以获得登录的cookie
r = requests.post("http://cer.nju.edu.cn/amserver/UI/Login", data=post_data, cookies=jar, allow_redirects=False)
result = []
res = Response(mimetype='application/json')
if len(r.content) == 0:
result.append('cer_success')
for key, value in r.cookies.items():
res.set_cookie(key, value)
res.set_data(json.dumps(result, ensure_ascii=False))
return res
# 获得验证码图片并保存cookie, 验证码对应登录网址为http://cer.nju.edu.cn/amserver/UI/Login
@loginModule.route('/cer_code', methods=['GET'])
def cer_code():
r = requests.get("http://cer.nju.edu.cn/amserver/verify/image.jsp")
res = Response(r.content, mimetype="image/jpg")
for key, value in r.cookies.items():
res.set_cookie(key, value)
return res
| 952 | 0 | 44 |
cf41028c39b31a7dc048944cfc4a553f51366dea | 4,434 | py | Python | scrape_mars.py | junki-ogawa/web-scraping-challenge | eeb181cabe48396a8068ef93ac41a62cc261b6a0 | [
"ADSL"
] | null | null | null | scrape_mars.py | junki-ogawa/web-scraping-challenge | eeb181cabe48396a8068ef93ac41a62cc261b6a0 | [
"ADSL"
] | null | null | null | scrape_mars.py | junki-ogawa/web-scraping-challenge | eeb181cabe48396a8068ef93ac41a62cc261b6a0 | [
"ADSL"
] | null | null | null | from splinter import Browser
from bs4 import BeautifulSoup
from webdriver_manager.chrome import ChromeDriverManager
import pymongo
import pandas as pd
import requests
from flask import Flask, render_template
import time
conn = 'mongodb://localhost:27017'
client = pymongo.MongoClient(conn)
db = client.mars_db
mars = db.mars
scrape() | 37.897436 | 97 | 0.690798 | from splinter import Browser
from bs4 import BeautifulSoup
from webdriver_manager.chrome import ChromeDriverManager
import pymongo
import pandas as pd
import requests
from flask import Flask, render_template
import time
conn = 'mongodb://localhost:27017'
client = pymongo.MongoClient(conn)
db = client.mars_db
mars = db.mars
def scrape():
# browser = init_browser()
executable_path = {'executable_path': ChromeDriverManager().install()}
browser = Browser('chrome', **executable_path, headless=False)
mars_dictionary = {}
#news title and paragraph text
url = 'https://redplanetscience.com/'
browser.visit(url)
html = browser.html
soup = BeautifulSoup(html, 'html.parser')
news_object = soup.select_one('div.list_text')
mars_dictionary["news_title"] = news_object.find('div', class_='content_title').get_text()
mars_dictionary["news_p"] = news_object.find('div', class_='article_teaser_body').get_text()
#mars image
url = 'https://spaceimages-mars.com/'
browser.visit(url)
html = browser.html
soup = BeautifulSoup(html, 'html.parser')
img_url = soup.find('img', class_='headerimage')
featured_img_url = img_url.attrs['src']
mars_dictionary["featured_img_url"] = url + featured_img_url
#mars table
mars_facts_url = 'https://galaxyfacts-mars.com/'
browser.visit(url)
html = browser.html
soup = BeautifulSoup(html, 'html.parser')
mars_facts_url = 'https://galaxyfacts-mars.com/'
table = pd.read_html(mars_facts_url)
mars_table = table[1]
mars_table.columns = ["Mars Profile", "Measure"]
mars_table.to_html()
mars_dictionary["mars_table"] = mars_table.to_html().replace("\n","")
#mars hemispheres
hemisphere_image_urls = []
url1 = ('https://marshemispheres.com/cerberus.html')
browser.visit(url1)
html = browser.html
soup = BeautifulSoup(html, 'html.parser')
cerberus_img = soup.find('div', class_='downloads')
cerberus_img = soup.find('a', href='images/full.jpg')
cerberus_img = cerberus_img.attrs['href']
cerberus_img = 'https://marshemispheres.com/' + cerberus_img
cereberus_object = soup.find('div', class_='cover')
cerberus_title = cereberus_object.find('h2').get_text()
cerb = {'title':cerberus_title,'img_url':cerberus_img}
hemisphere_image_urls.append(cerb)
url2 = ('https://marshemispheres.com/schiaparelli.html')
browser.visit(url2)
html = browser.html
soup = BeautifulSoup(html, 'html.parser')
schiaparelli_img = soup.find('div', class_='downloads')
schiaparelli_img = soup.find('a', href='images/schiaparelli_enhanced-full.jpg')
schiaparelli_img = schiaparelli_img.attrs['href']
schiaparelli_img = 'https://marshemispheres.com/' + schiaparelli_img
schiaparelli_object = soup.find('div', class_='container')
schiaparelli_title = schiaparelli_object.find('h2').get_text()
schia = {'title':schiaparelli_title,'img_url':schiaparelli_img}
hemisphere_image_urls.append(schia)
url3 = ('https://marshemispheres.com/syrtis.html')
browser.visit(url3)
html = browser.html
soup = BeautifulSoup(html, 'html.parser')
syrtis_img = soup.find('div', class_='downloads')
syrtis_img = soup.find('a', href='images/syrtis_major_enhanced-full.jpg')
syrtis_img = syrtis_img.attrs['href']
syrtis_img = 'https://marshemispheres.com/' + syrtis_img
syrtis_title = soup.find('h2', class_='title').get_text()
syrtis = {'title':syrtis_title,'img_url':syrtis_img}
hemisphere_image_urls.append(syrtis)
url4 = ('https://marshemispheres.com/valles.html')
browser.visit(url4)
html = browser.html
soup = BeautifulSoup(html, 'html.parser')
valles_img = soup.find('div', class_='downloads')
valles_img = soup.find('a', href='images/valles_marineris_enhanced-full.jpg')
valles_img = valles_img.attrs['href']
valles_img = 'https://marshemispheres.com/' + valles_img
valles_title = soup.find('h2', class_='title').get_text()
valles = {'title':valles_title,'img_url':valles_img}
hemisphere_image_urls.append(valles)
mars_dictionary["hemisphere_image_urls"] = hemisphere_image_urls
# Quit the browser
browser.quit()
print(mars_dictionary)
return mars_dictionary
# mars.insert_one(mars_dictionary)
scrape() | 4,059 | 0 | 24 |
57f9a91be26db31e5547c40be706239eb0346542 | 900 | py | Python | viettel_shake/shake/serializers.py | longphu0001/viettel-shake | 3326335eb3479d84db7fa9bb471d6a1a7ffa4019 | [
"MIT"
] | 20 | 2020-01-03T03:49:43.000Z | 2020-02-08T05:30:36.000Z | viettel_shake/shake/serializers.py | longphu0001/viettel-shake | 3326335eb3479d84db7fa9bb471d6a1a7ffa4019 | [
"MIT"
] | 4 | 2020-01-15T10:25:45.000Z | 2020-01-31T14:40:36.000Z | viettel_shake/shake/serializers.py | longphu0001/viettel-shake | 3326335eb3479d84db7fa9bb471d6a1a7ffa4019 | [
"MIT"
] | 18 | 2019-12-30T03:29:45.000Z | 2020-03-06T15:32:14.000Z | from rest_framework import serializers
from .consts import USE_TOTAL_SHAKE_TURN
from .models import Shake, ViettelUser
| 24.324324 | 58 | 0.725556 | from rest_framework import serializers
from .consts import USE_TOTAL_SHAKE_TURN
from .models import Shake, ViettelUser
def get_default_shake_turn():
return USE_TOTAL_SHAKE_TURN
class LoginSerializer(serializers.Serializer):
phone = serializers.CharField()
otp = serializers.CharField()
shake_turn = serializers.IntegerField(
default=get_default_shake_turn,
min_value=-1,
required=False
) # if shake_turn set to -1, then use total turn left
class RequestLoginSerializer(serializers.Serializer):
phone = serializers.CharField()
class ShakeSerializer(serializers.ModelSerializer):
class Meta:
model = Shake
fields = '__all__'
class ViettelUserSerializer(serializers.ModelSerializer):
shakes = ShakeSerializer(many=True, read_only=True)
class Meta:
model = ViettelUser
fields = ('phone', 'shakes',)
| 40 | 620 | 115 |
13806065b202f1f3a84428e15535c568b7658007 | 1,205 | py | Python | tchron/tests/vtk_test.py | dyvasey/riftinversion | 6707a22d3413ee5af6c8aee2660f1024214f1a31 | [
"MIT"
] | null | null | null | tchron/tests/vtk_test.py | dyvasey/riftinversion | 6707a22d3413ee5af6c8aee2660f1024214f1a31 | [
"MIT"
] | null | null | null | tchron/tests/vtk_test.py | dyvasey/riftinversion | 6707a22d3413ee5af6c8aee2660f1024214f1a31 | [
"MIT"
] | null | null | null | """
Test for forward modeling of ASPECT output, using cookbook output
"""
#%% Import and setup files
import matplotlib.pyplot as plt
from matplotlib import cm,colors
import numpy as np
import pyvista as pv
import vtk_plot as vp
# Read in sample data from continental extension cookbook
meshes = pv.read('../../sample_data/vtk_tchron_test.vtm')
#%% Add He ages to final mesh and save
He_mesh = vp.He_age_vtk_parallel(meshes,'AHe',1e5,batch_size=100,
interpolate_profile=False,filename='meshes_He.vtm')
He_mesh_interp = vp.He_age_vtk_parallel(meshes,'AHe',1e5,batch_size=100,
interpolate_profile=True,filename='meshes_He_interp.vtm')
#%% Plot the new mesh
fig,axs = plt.subplots(2,dpi=300)
vp.plot2D('meshes_He/meshes_He_20.vtu','AHe',bounds=[0,200,90,100],cmap='plasma_r',ax=axs[0],
clim=[0,2])
vp.plot2D('meshes_He_interp/meshes_He_interp_20.vtu','AHe',bounds=[0,200,90,100],cmap='plasma_r',ax=axs[1],
clim=[0,2])
ages = He_mesh[-1].point_data['AHe']
cbar = vp.add_colorbar(fig,0,2,'plasma_r',label='AHe Age',ticks=[0,1,2])
fig.savefig('vtk_tchron_test.pdf')
| 30.125 | 108 | 0.66805 | """
Test for forward modeling of ASPECT output, using cookbook output
"""
#%% Import and setup files
import matplotlib.pyplot as plt
from matplotlib import cm,colors
import numpy as np
import pyvista as pv
import vtk_plot as vp
# Read in sample data from continental extension cookbook
meshes = pv.read('../../sample_data/vtk_tchron_test.vtm')
#%% Add He ages to final mesh and save
He_mesh = vp.He_age_vtk_parallel(meshes,'AHe',1e5,batch_size=100,
interpolate_profile=False,filename='meshes_He.vtm')
He_mesh_interp = vp.He_age_vtk_parallel(meshes,'AHe',1e5,batch_size=100,
interpolate_profile=True,filename='meshes_He_interp.vtm')
#%% Plot the new mesh
fig,axs = plt.subplots(2,dpi=300)
vp.plot2D('meshes_He/meshes_He_20.vtu','AHe',bounds=[0,200,90,100],cmap='plasma_r',ax=axs[0],
clim=[0,2])
vp.plot2D('meshes_He_interp/meshes_He_interp_20.vtu','AHe',bounds=[0,200,90,100],cmap='plasma_r',ax=axs[1],
clim=[0,2])
ages = He_mesh[-1].point_data['AHe']
cbar = vp.add_colorbar(fig,0,2,'plasma_r',label='AHe Age',ticks=[0,1,2])
fig.savefig('vtk_tchron_test.pdf')
| 0 | 0 | 0 |
a8b0260995482da49d8adf4cf28f57de040c3349 | 1,759 | py | Python | final_gpt2_test_wikitext103-fp32.py | minhhn2910/conga2022 | 81ad2fb9c0055c332f8f305b2ea409b6577003f4 | [
"MIT"
] | null | null | null | final_gpt2_test_wikitext103-fp32.py | minhhn2910/conga2022 | 81ad2fb9c0055c332f8f305b2ea409b6577003f4 | [
"MIT"
] | null | null | null | final_gpt2_test_wikitext103-fp32.py | minhhn2910/conga2022 | 81ad2fb9c0055c332f8f305b2ea409b6577003f4 | [
"MIT"
] | null | null | null | # -*- coding: utf-8 -*-
# Transformers installation
import numpy as np
from transformers import GPT2LMHeadModel, GPT2TokenizerFast
from qtorch.quant import posit_quantize, float_quantize, configurable_table_quantize
device = 'cuda'
model_id = 'gpt2-large'
tokenizer = GPT2TokenizerFast.from_pretrained(model_id)
from datasets import load_dataset
#test = load_dataset("lambada", split='test')
test = load_dataset('wikitext', 'wikitext-103-v1', split='test')
#test = load_dataset("ptb_text_only", split='test')
encodings = tokenizer('\n\n'.join(test['text']), return_tensors='pt')
print (run ([],[]))
| 25.128571 | 84 | 0.670836 | # -*- coding: utf-8 -*-
# Transformers installation
import numpy as np
from transformers import GPT2LMHeadModel, GPT2TokenizerFast
from qtorch.quant import posit_quantize, float_quantize, configurable_table_quantize
device = 'cuda'
model_id = 'gpt2-large'
tokenizer = GPT2TokenizerFast.from_pretrained(model_id)
from datasets import load_dataset
#test = load_dataset("lambada", split='test')
test = load_dataset('wikitext', 'wikitext-103-v1', split='test')
#test = load_dataset("ptb_text_only", split='test')
encodings = tokenizer('\n\n'.join(test['text']), return_tensors='pt')
def run(weight_table, act_table ):
import torch
import torch.nn as nn
model = GPT2LMHeadModel.from_pretrained(model_id)
model = model.to(device)
layer_count = 0
linear_layer_count = 0
op_count = 0
#print ("MAC operation count ", op_count)
print ("Layer count ", layer_count)
#model = model.to(device)
import torch
from tqdm import tqdm
max_length = model.config.n_positions
stride = 1024
#stride = 32
lls = []
for i in tqdm(range(0, encodings.input_ids.size(1), stride)):
begin_loc = max(i + stride - max_length, 0)
end_loc = min(i + stride, encodings.input_ids.size(1))
trg_len = end_loc - i # may be different from stride on last loop
input_ids = encodings.input_ids[:,begin_loc:end_loc].to(device)
target_ids = input_ids.clone()
target_ids[:,:-trg_len] = -100
with torch.no_grad():
outputs = model(input_ids, labels=target_ids)
log_likelihood = outputs[0] * trg_len
lls.append(log_likelihood)
ppl = torch.exp(torch.stack(lls).sum() / end_loc)
return ppl.item()
print (run ([],[]))
| 1,120 | 0 | 23 |
810202aa9efdc363b89d45a92de4e299f0f1ac3c | 4,067 | py | Python | modules/iib_api.py | AATools/ib-metrics-pyclient | 2f3bae726201eabcb6c1bec0cfdf5957df538787 | [
"MIT"
] | 1 | 2020-07-28T15:47:44.000Z | 2020-07-28T15:47:44.000Z | modules/iib_api.py | AATools/ib-metrics-pyclient | 2f3bae726201eabcb6c1bec0cfdf5957df538787 | [
"MIT"
] | 5 | 2020-04-27T12:05:03.000Z | 2020-12-11T10:27:08.000Z | modules/iib_api.py | AATools/ib-metrics-pyclient | 2f3bae726201eabcb6c1bec0cfdf5957df538787 | [
"MIT"
] | 2 | 2020-11-11T03:21:13.000Z | 2022-03-08T17:45:12.000Z | # -*- coding: utf-8 -*-
"""Various functions for client api."""
import subprocess
def run_iib_command(**kwargs):
"""Calls predefined commands and returns their result."""
command_mapping = {
'get_brokers_status': 'mqsilist | grep Broker',
'get_integration_nodes_status': 'mqsilist | grep "Integration node"',
'get_broker_objects': 'mqsilist {0} -r',
}
broker = str()
for arg_name, arg_value in kwargs.items():
if arg_name == 'task':
iib_command = command_mapping[arg_value]
elif arg_name == 'broker_name':
broker = arg_value
if broker:
command = iib_command.format(broker)
else:
command = iib_command
output = execute_command(command=command)
return output
def execute_command(command):
"""Executes in shell."""
proc = subprocess.Popen(command,
shell=True,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
universal_newlines=True)
result = proc.communicate()[0]
return result
def get_status(status):
"""Returns a numeric status value."""
status_map = {
'running': 1,
'stopped': 0}
return status_map[status]
def get_platform_params_for_commands(iib_ver):
"""Returns parameters for internal functions depending on Integration Bus version."""
mqsilist_brokers = "get_brokers_status"
mqsilist_integration_nodes = "get_integration_nodes_status"
# See IBM diagnostic messages:
# https://www.ibm.com/support/knowledgecenter/en/SSMKHH_9.0.0/com.ibm.etools.mft.bipmsgs.doc/ay_bip1.htm
# Also you can use command: mqsiexplain <bip_code>
bip_codes_broker = {
# BIPCode: [broker_name_position, qm_name_position, status_position, trim_last_dot_in_qm_name]
'BIP1284I': [2, 6, 8, 'false'],
'BIP1285I': [2, 6, 8, 'false'],
'BIP1293I': [2, 14 ,7, 'true'],
'BIP1294I': [2, 14 ,7, 'true'],
'BIP1295I': [2, 17, 13, 'true'],
'BIP1296I': [2, 22, 4, 'true'],
'BIP1297I': [2, 14, 7, 'true'],
'BIP1298I': [2, 17, 4, 'true']
}
# https://www.ibm.com/support/knowledgecenter/en/SSMKHH_10.0.0/com.ibm.etools.mft.bipmsgs.doc/ay_bip1.htm
bip_codes_integration_nodes = {
# BIPCode: [broker_name_position, qm_name_position, status_position, trim_last_dot_in_qm_name]
'BIP1284I': [3, 8, 14, 'false'],
'BIP1285I': [3, 7, 9, 'false'],
'BIP1295I': [3, 19, 15, 'true'],
'BIP1296I': [3, 24 ,5, 'true'],
'BIP1298I': [3, 18, 5, 'true'],
'BIP1325I': [3, None, 9, 'false'],
'BIP1326I': [3, None, 5, 'false'],
'BIP1340I': [3, None, 5, 'false'],
'BIP1353I': [3, 8, 10, 'false'],
'BIP1366I': [3, 19, 15, 'true'],
'BIP1376I': [3, 19, 15, 'true'],
'BIP1377I': [3, 24, 5, 'true']
}
bip_codes_broker_components = {
# BIPCode: [component_type, positions: broker_name, egname, status]
'BIP1286I': ['exec_groups', 6, 3, 8],
'BIP1287I': ['exec_groups', 6, 3, 8],
# BIPCode: [component_type, positions: egname, appname, status]
'BIP1275I': ['applications', 6, 2, 8],
'BIP1276I': ['applications', 6, 2, 8],
# BIPCode: [component_type, positions: egname, appname, msgflowname, status]
'BIP1277I': ['message_flows', 7, 11, 3, 9],
'BIP1278I': ['message_flows', 7, 11, 3, 9]}
bip_codes_integration_nodes_components = {
'BIP1286I': ['exec_groups', 7, 3, 9],
'BIP1287I': ['exec_groups', 7, 3, 9],
'BIP1275I': ['applications', 6, 2, 8],
'BIP1276I': ['applications', 6, 2, 8],
'BIP1277I': ['message_flows', 7, 11, 3, 9],
'BIP1278I': ['message_flows', 7, 11, 3, 9]}
if iib_ver == "9":
return mqsilist_brokers, bip_codes_broker, bip_codes_broker_components
if iib_ver == "10":
return mqsilist_integration_nodes, bip_codes_integration_nodes, bip_codes_integration_nodes_components
| 40.267327 | 110 | 0.595771 | # -*- coding: utf-8 -*-
"""Various functions for client api."""
import subprocess
def run_iib_command(**kwargs):
"""Calls predefined commands and returns their result."""
command_mapping = {
'get_brokers_status': 'mqsilist | grep Broker',
'get_integration_nodes_status': 'mqsilist | grep "Integration node"',
'get_broker_objects': 'mqsilist {0} -r',
}
broker = str()
for arg_name, arg_value in kwargs.items():
if arg_name == 'task':
iib_command = command_mapping[arg_value]
elif arg_name == 'broker_name':
broker = arg_value
if broker:
command = iib_command.format(broker)
else:
command = iib_command
output = execute_command(command=command)
return output
def execute_command(command):
"""Executes in shell."""
proc = subprocess.Popen(command,
shell=True,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
universal_newlines=True)
result = proc.communicate()[0]
return result
def get_status(status):
"""Returns a numeric status value."""
status_map = {
'running': 1,
'stopped': 0}
return status_map[status]
def get_platform_params_for_commands(iib_ver):
"""Returns parameters for internal functions depending on Integration Bus version."""
mqsilist_brokers = "get_brokers_status"
mqsilist_integration_nodes = "get_integration_nodes_status"
# See IBM diagnostic messages:
# https://www.ibm.com/support/knowledgecenter/en/SSMKHH_9.0.0/com.ibm.etools.mft.bipmsgs.doc/ay_bip1.htm
# Also you can use command: mqsiexplain <bip_code>
bip_codes_broker = {
# BIPCode: [broker_name_position, qm_name_position, status_position, trim_last_dot_in_qm_name]
'BIP1284I': [2, 6, 8, 'false'],
'BIP1285I': [2, 6, 8, 'false'],
'BIP1293I': [2, 14 ,7, 'true'],
'BIP1294I': [2, 14 ,7, 'true'],
'BIP1295I': [2, 17, 13, 'true'],
'BIP1296I': [2, 22, 4, 'true'],
'BIP1297I': [2, 14, 7, 'true'],
'BIP1298I': [2, 17, 4, 'true']
}
# https://www.ibm.com/support/knowledgecenter/en/SSMKHH_10.0.0/com.ibm.etools.mft.bipmsgs.doc/ay_bip1.htm
bip_codes_integration_nodes = {
# BIPCode: [broker_name_position, qm_name_position, status_position, trim_last_dot_in_qm_name]
'BIP1284I': [3, 8, 14, 'false'],
'BIP1285I': [3, 7, 9, 'false'],
'BIP1295I': [3, 19, 15, 'true'],
'BIP1296I': [3, 24 ,5, 'true'],
'BIP1298I': [3, 18, 5, 'true'],
'BIP1325I': [3, None, 9, 'false'],
'BIP1326I': [3, None, 5, 'false'],
'BIP1340I': [3, None, 5, 'false'],
'BIP1353I': [3, 8, 10, 'false'],
'BIP1366I': [3, 19, 15, 'true'],
'BIP1376I': [3, 19, 15, 'true'],
'BIP1377I': [3, 24, 5, 'true']
}
bip_codes_broker_components = {
# BIPCode: [component_type, positions: broker_name, egname, status]
'BIP1286I': ['exec_groups', 6, 3, 8],
'BIP1287I': ['exec_groups', 6, 3, 8],
# BIPCode: [component_type, positions: egname, appname, status]
'BIP1275I': ['applications', 6, 2, 8],
'BIP1276I': ['applications', 6, 2, 8],
# BIPCode: [component_type, positions: egname, appname, msgflowname, status]
'BIP1277I': ['message_flows', 7, 11, 3, 9],
'BIP1278I': ['message_flows', 7, 11, 3, 9]}
bip_codes_integration_nodes_components = {
'BIP1286I': ['exec_groups', 7, 3, 9],
'BIP1287I': ['exec_groups', 7, 3, 9],
'BIP1275I': ['applications', 6, 2, 8],
'BIP1276I': ['applications', 6, 2, 8],
'BIP1277I': ['message_flows', 7, 11, 3, 9],
'BIP1278I': ['message_flows', 7, 11, 3, 9]}
if iib_ver == "9":
return mqsilist_brokers, bip_codes_broker, bip_codes_broker_components
if iib_ver == "10":
return mqsilist_integration_nodes, bip_codes_integration_nodes, bip_codes_integration_nodes_components
| 0 | 0 | 0 |
8cee128cf4e93fd417bbdce4ad0272375e780eed | 963 | py | Python | test/command_line/test_stereographic_projections.py | jbeilstenedmands/dials | fb9672b91854f564cbbba6f1cceeefa18d135965 | [
"BSD-3-Clause"
] | null | null | null | test/command_line/test_stereographic_projections.py | jbeilstenedmands/dials | fb9672b91854f564cbbba6f1cceeefa18d135965 | [
"BSD-3-Clause"
] | null | null | null | test/command_line/test_stereographic_projections.py | jbeilstenedmands/dials | fb9672b91854f564cbbba6f1cceeefa18d135965 | [
"BSD-3-Clause"
] | 1 | 2020-02-04T15:39:06.000Z | 2020-02-04T15:39:06.000Z | from __future__ import absolute_import, division, print_function
import json
import os
import procrunner
| 30.09375 | 65 | 0.628245 | from __future__ import absolute_import, division, print_function
import json
import os
import procrunner
def test_stereographic_projection(dials_regression, tmpdir):
path = os.path.join(dials_regression, "experiment_test_data")
result = procrunner.run(
(
"dials.stereographic_projection",
"%s/experiment_1.json" % path,
"hkl_limit=4",
"plot.show=False",
"plot.filename=proj.png",
"json.filename=proj.json",
),
working_directory=tmpdir,
)
assert not result.returncode and not result.stderr
assert tmpdir.join("projections.txt").check()
assert tmpdir.join("proj.png").check()
assert tmpdir.join("proj.json").check()
with tmpdir.join("proj.json").open("rb") as f:
d = json.load(f)
assert set(d) == {"data", "layout"}
assert d["data"][0]["name"] == "stereographic_projections"
assert len(d["data"][0]["x"]) == 578
| 832 | 0 | 23 |
bc6827cecb880364d64ad812e4e93528bfdf6b85 | 1,795 | py | Python | setup.py | nyergler/hieroglyph | 843ed9553fe640f9d6061dfb1accf131b467492e | [
"BSD-3-Clause"
] | 115 | 2015-01-08T12:54:49.000Z | 2021-05-20T18:27:14.000Z | setup.py | lpirl/hieroglyph | e41533276594fa9b826d5e9c48164db628fe1541 | [
"BSD-3-Clause"
] | 94 | 2015-01-05T15:34:27.000Z | 2020-10-09T08:13:25.000Z | setup.py | lpirl/hieroglyph | e41533276594fa9b826d5e9c48164db628fe1541 | [
"BSD-3-Clause"
] | 48 | 2015-01-21T06:13:34.000Z | 2020-05-07T22:10:24.000Z | #!/usr/bin/env python
import os
from setuptools import setup, find_packages
import versioneer
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.rst')).read()
NEWS = open(os.path.join(here, 'NEWS.txt')).read()
install_requires = [
"setuptools",
"Sphinx >= 2.0",
]
setup(name='hieroglyph',
version=versioneer.get_version(),
cmdclass=versioneer.get_cmdclass(),
description=("Generate HTML presentations from plain text "
"sources with all the power of Sphinx."),
long_description=README + '\n\n' + NEWS,
classifiers=[
'License :: OSI Approved :: BSD License',
'Topic :: Documentation',
'Topic :: Text Processing',
"Programming Language :: Python :: 3",
],
keywords='',
author='Nathan Yergler',
author_email='nathan@yergler.net',
url='https://github.com/nyergler/hieroglyph',
license='BSD',
packages=find_packages('src'),
package_dir={'': 'src'},
include_package_data=True,
zip_safe=False,
install_requires=install_requires,
entry_points={
'console_scripts': [
'hieroglyph=hieroglyph.quickstart:main',
'hieroglyph-quickstart=hieroglyph.quickstart:quickstart',
],
'hieroglyph.theme': [
'slides=hieroglyph.themes:SLIDES',
'single-level=hieroglyph.themes:SINGLE_LEVEL',
'slides2=hieroglyph.themes:SLIDES2',
],
'sphinx.builders': [
'slides = hieroglyph',
'dirslides = hieroglyph',
],
},
test_suite='hieroglyph.tests',
tests_require=[
'beautifulsoup4',
'mock',
'sphinx-testing',
],
)
| 28.951613 | 71 | 0.58663 | #!/usr/bin/env python
import os
from setuptools import setup, find_packages
import versioneer
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.rst')).read()
NEWS = open(os.path.join(here, 'NEWS.txt')).read()
install_requires = [
"setuptools",
"Sphinx >= 2.0",
]
setup(name='hieroglyph',
version=versioneer.get_version(),
cmdclass=versioneer.get_cmdclass(),
description=("Generate HTML presentations from plain text "
"sources with all the power of Sphinx."),
long_description=README + '\n\n' + NEWS,
classifiers=[
'License :: OSI Approved :: BSD License',
'Topic :: Documentation',
'Topic :: Text Processing',
"Programming Language :: Python :: 3",
],
keywords='',
author='Nathan Yergler',
author_email='nathan@yergler.net',
url='https://github.com/nyergler/hieroglyph',
license='BSD',
packages=find_packages('src'),
package_dir={'': 'src'},
include_package_data=True,
zip_safe=False,
install_requires=install_requires,
entry_points={
'console_scripts': [
'hieroglyph=hieroglyph.quickstart:main',
'hieroglyph-quickstart=hieroglyph.quickstart:quickstart',
],
'hieroglyph.theme': [
'slides=hieroglyph.themes:SLIDES',
'single-level=hieroglyph.themes:SINGLE_LEVEL',
'slides2=hieroglyph.themes:SLIDES2',
],
'sphinx.builders': [
'slides = hieroglyph',
'dirslides = hieroglyph',
],
},
test_suite='hieroglyph.tests',
tests_require=[
'beautifulsoup4',
'mock',
'sphinx-testing',
],
)
| 0 | 0 | 0 |
fe1d958575805e5d3d534043bf5eb33566f7cef4 | 952 | py | Python | y2022/CargoDetection/python/rapid_react/cargo/ui/subpanels.py | pjreiniger/GirlsOfSteelFRC | 03ab0fb699dfc14985ce3c95b1eacb145288fc7e | [
"BSD-3-Clause"
] | null | null | null | y2022/CargoDetection/python/rapid_react/cargo/ui/subpanels.py | pjreiniger/GirlsOfSteelFRC | 03ab0fb699dfc14985ce3c95b1eacb145288fc7e | [
"BSD-3-Clause"
] | 44 | 2021-11-09T23:23:48.000Z | 2022-03-22T23:44:15.000Z | y2022/CargoDetection/python/rapid_react/cargo/ui/subpanels.py | pjreiniger/GirlsOfSteelFRC | 03ab0fb699dfc14985ce3c95b1eacb145288fc7e | [
"BSD-3-Clause"
] | 1 | 2021-11-19T03:45:21.000Z | 2021-11-19T03:45:21.000Z | from PyQt5.QtWidgets import QWidget
from PyQt5.uic import loadUi
from rapid_react.cargo.cargo_pipeline_params import CargoPipelineParams
| 36.615385 | 73 | 0.746849 | from PyQt5.QtWidgets import QWidget
from PyQt5.uic import loadUi
from rapid_react.cargo.cargo_pipeline_params import CargoPipelineParams
class ParamsWidget(QWidget):
def __init__(self, parent=None):
super().__init__(parent)
loadUi("rapid_react/cargo/ui/params_widget.ui", self)
def set_params(self, params: CargoPipelineParams):
self.hsv_tab.set_params(params.hsv)
self.circles_tab.set_params(params.circle_detection)
self.contour_filter_tab.set_params(params.contour_filtering)
def ui_to_params(self, params: CargoPipelineParams):
params.hsv = self.hsv_tab.ui_to_params()
params.circle_detection = self.circles_tab.ui_to_params()
params.contour_filtering = self.contour_filter_tab.ui_to_params()
def connect_signals(self, cb):
self.hsv_tab.connect_signals(cb)
self.circles_tab.connect_signals(cb)
self.contour_filter_tab.connect_signals(cb)
| 676 | 7 | 130 |
aff8057d4e984138d871f670277a645965da11d6 | 33,638 | py | Python | post_optimization_studies/mad_analyses/ma100MeV_L1pt8-2pt4TeV_deta2pt6/Output/Histos/MadAnalysis5job_0/selection_3.py | sheride/axion_pheno | 7d3fc08f5ae5b17a3500eba19a2e43f87f076ce5 | [
"MIT"
] | null | null | null | post_optimization_studies/mad_analyses/ma100MeV_L1pt8-2pt4TeV_deta2pt6/Output/Histos/MadAnalysis5job_0/selection_3.py | sheride/axion_pheno | 7d3fc08f5ae5b17a3500eba19a2e43f87f076ce5 | [
"MIT"
] | null | null | null | post_optimization_studies/mad_analyses/ma100MeV_L1pt8-2pt4TeV_deta2pt6/Output/Histos/MadAnalysis5job_0/selection_3.py | sheride/axion_pheno | 7d3fc08f5ae5b17a3500eba19a2e43f87f076ce5 | [
"MIT"
] | null | null | null |
# Running!
if __name__ == '__main__':
selection_3()
| 154.302752 | 1,799 | 0.768238 | def selection_3():
# Library import
import numpy
import matplotlib
import matplotlib.pyplot as plt
import matplotlib.gridspec as gridspec
# Library version
matplotlib_version = matplotlib.__version__
numpy_version = numpy.__version__
# Histo binning
xBinning = numpy.linspace(0.0,1000.0,101,endpoint=True)
# Creating data sequence: middle of each bin
xData = numpy.array([5.0,15.0,25.0,35.0,45.0,55.0,65.0,75.0,85.0,95.0,105.0,115.0,125.0,135.0,145.0,155.0,165.0,175.0,185.0,195.0,205.0,215.0,225.0,235.0,245.0,255.0,265.0,275.0,285.0,295.0,305.0,315.0,325.0,335.0,345.0,355.0,365.0,375.0,385.0,395.0,405.0,415.0,425.0,435.0,445.0,455.0,465.0,475.0,485.0,495.0,505.0,515.0,525.0,535.0,545.0,555.0,565.0,575.0,585.0,595.0,605.0,615.0,625.0,635.0,645.0,655.0,665.0,675.0,685.0,695.0,705.0,715.0,725.0,735.0,745.0,755.0,765.0,775.0,785.0,795.0,805.0,815.0,825.0,835.0,845.0,855.0,865.0,875.0,885.0,895.0,905.0,915.0,925.0,935.0,945.0,955.0,965.0,975.0,985.0,995.0])
# Creating weights for histo: y4_PT_0
y4_PT_0_weights = numpy.array([0.0,0.0,6.0713078477,5.1903318698,4.93205187628,4.70207988205,4.44910788839,4.105915897,3.82994830393,3.54159671116,3.22317151915,2.91182232696,2.60931833454,2.45010553854,2.26789554311,2.06622634817,1.88401635274,1.714189557,1.44352796379,1.39222596508,1.21001596965,1.12156477187,1.01188477462,0.877438377989,0.83675077901,0.72353318185,0.663385983359,0.571396385666,0.559013185977,0.428105189261,0.445795188817,0.375034230592,0.364420030858,0.320194311968,0.318425272012,0.304273032367,0.221128674453,0.205207394852,0.19813127503,0.199900314985,0.183979035385,0.14152235645,0.139753316494,0.122063036938,0.109679797249,0.100834677471,0.111448837204,0.0742992381362,0.0955275976037,0.0813753579587,0.0636850384025,0.0583779585356,0.05660891858,0.0477637988018,0.0336115551568,0.0406876789793,0.0318425272012,0.0406876789793,0.0318425272012,0.0265354393344,0.02830446729,0.0194593235119,0.0247664113787,0.0229973794231,0.014152235645,0.0176902915562,0.0212283514675,0.00884514777812,0.0106141757337,0.00884514777812,0.00707611582249,0.0106141757337,0.00707611582249,0.00707611582249,0.0,0.00884514777812,0.00176902915562,0.00707611582249,0.00707611582249,0.00530708786687,0.00176902915562,0.00176902915562,0.00707611582249,0.0,0.00353805871125,0.00176902915562,0.0,0.0,0.0,0.0,0.00353805871125,0.0,0.0,0.0,0.00176902915562,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_1
y4_PT_1_weights = numpy.array([0.0,0.0,3.90561195435,3.40451040881,3.15875300926,3.02610370984,2.85413260298,2.59985480569,2.4876168109,2.21836017196,2.04419201912,1.97150404957,1.71188611557,1.47463392587,1.41582407672,1.24601524362,1.1519471429,1.02164037332,0.936030320111,0.823853476504,0.706352883113,0.625116538156,0.550320651459,0.507557191671,0.477681043297,0.433826054931,0.40179282866,0.31418728855,0.333400509674,0.263931370367,0.273545814675,0.230819525212,0.214800034374,0.194476939318,0.167752594701,0.121836931475,0.114316935065,0.117545117845,0.107939106805,0.110068966431,0.100441572462,0.0982970445458,0.0758757046245,0.0534275061457,0.05877763487,0.0577028129542,0.0587630864847,0.0374008130576,0.0373952175248,0.0395332945908,0.0256424723299,0.0224382024591,0.0299211564421,0.0235088876774,0.0288586886634,0.023504395264,0.0149588815756,0.025644670575,0.0138919213833,0.0224393015816,0.0106809648508,0.0160262694264,0.019233936585,0.0128241978006,0.00855034183384,0.00854661281091,0.00534530857243,0.00641489466827,0.0085511092212,0.00640853974173,0.00855034183384,0.00534344206256,0.0,0.00320503446045,0.00427571847977,0.00106772358278,0.00106958729489,0.0,0.00320613278361,0.00320689817256,0.00320689817256,0.00213917498946,0.00106958729489,0.00106958729489,0.00106772358278,0.00106772358278,0.0,0.0,0.00106772358278,0.0,0.0,0.00106958729489,0.00106958729489,0.0,0.0,0.00106772358278,0.00106772358278,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_2
y4_PT_2_weights = numpy.array([0.0,0.0,2.66604107998,2.34797739885,2.17505615475,2.10699893739,1.87574287841,1.74448964493,1.67226562651,1.41670316133,1.39378595549,1.26947712378,1.12086228588,0.988914652224,0.896551028666,0.774325397492,0.70487937978,0.675711772341,0.611821356046,0.513207730894,0.47570692133,0.47570692133,0.386121098481,0.332647484842,0.295146515277,0.279868351381,0.235422740045,0.211811054023,0.19306056924,0.182643606583,0.165282042155,0.147920477727,0.128475552768,0.120836470819,0.11111398834,0.0875022623175,0.0937524239116,0.0868077821404,0.0652794566496,0.061807135764,0.056251454347,0.0416677306274,0.0381954337418,0.0479178922215,0.0368065093875,0.0361120452104,0.0270840349078,0.0298618836163,0.0201394091366,0.0250006463764,0.0201394091366,0.0208338733137,0.0125003231882,0.0145837117196,0.0159726360738,0.0159726360738,0.0104169346568,0.00833355012548,0.00625016159411,0.0090280103026,0.00972247447973,0.00763908594836,0.00555569741699,0.00277784950849,0.00625016159411,0.00277784950849,0.00277784950849,0.00555569741699,0.00347231208562,0.00416677306274,0.00277784950849,0.00347231208562,0.00208338733137,0.000694462577123,0.000694462577123,0.00138892475425,0.00277784950849,0.00208338733137,0.00208338733137,0.00208338733137,0.00208338733137,0.000694462577123,0.00138892475425,0.00138892475425,0.0,0.000694462577123,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.000694462577123,0.0,0.0,0.0,0.00138892475425,0.0])
# Creating weights for histo: y4_PT_3
y4_PT_3_weights = numpy.array([0.0,0.0,1.90847613326,1.68989047218,1.55475603442,1.38927598818,1.31104036632,1.23280434446,1.13655111757,0.986717875703,0.959691068151,0.838781434367,0.770503015289,0.724035402306,0.614505771702,0.558555356068,0.489328536725,0.463724129571,0.396394070758,0.374582904664,0.310571926778,0.283070879094,0.283070879094,0.232336224918,0.201041976174,0.186343132067,0.170696007695,0.143195000011,0.130392796434,0.116168112459,0.106684989809,0.0910378654373,0.0853480238474,0.0706491797404,0.0744424208003,0.0606918969582,0.0611660570907,0.0478897333811,0.0531054148384,0.0388807588638,0.0350875138039,0.0350875138039,0.0317684248765,0.0270268675517,0.0208628458294,0.017543756902,0.0213370019619,0.0180179130345,0.0161212885045,0.0132763557096,0.0132763557096,0.0066381778548,0.00474155732486,0.0123280434446,0.00711233398729,0.00521571345734,0.00568986558983,0.0066381778548,0.0066381778548,0.00900895451723,0.00474155732486,0.00284493359491,0.00426740119237,0.00189662252994,0.0033190893274,0.00426740119237,0.00284493359491,0.0033190893274,0.00142246679746,0.00426740119237,0.00237077786243,0.00237077786243,0.000474155732486,0.00142246679746,0.00142246679746,0.00189662252994,0.0,0.000948311064972,0.000948311064972,0.000948311064972,0.000948311064972,0.00142246679746,0.0,0.000948311064972,0.0,0.000474155732486,0.0,0.0,0.0,0.0,0.000948311064972,0.0,0.0,0.0,0.0,0.000948311064972,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_4
y4_PT_4_weights = numpy.array([0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_5
y4_PT_5_weights = numpy.array([0.0,0.0,1.05462838872,0.0,1.0521138287,0.0,1.0529581672,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_6
y4_PT_6_weights = numpy.array([0.0,0.0,1.15196926193,4.14546686726,2.07128556314,2.30314234879,2.99532358213,1.38295763269,1.15081073498,1.84150989919,1.84187686212,1.61263992282,2.53659494287,0.461124849562,1.15077768911,0.691380062969,0.460707165054,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_7
y4_PT_7_weights = numpy.array([0.0,0.0,2.3540205365,1.41265491147,2.07669788222,1.99353536238,1.85608640603,1.74405088045,1.49539888159,1.68896773763,1.49519807005,1.35677813992,1.52285774327,1.32898420765,0.94158797957,1.16295153727,1.05244941165,1.16355089431,1.02460970051,0.941498729998,0.69210542112,0.442901767581,0.553543538003,0.52600773704,0.0830493806147,0.193898887108,0.221568870185,0.027659203883,0.0552690281325,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_8
y4_PT_8_weights = numpy.array([0.0,0.0,0.816582549802,0.847060307822,0.69571705543,0.554295245944,0.615043778286,0.493812982869,0.484005984775,0.534337793956,0.383108986376,0.393335655904,0.463898589451,0.302378547508,0.372896273533,0.41335239346,0.342725805296,0.34272944617,0.292512081802,0.221826208069,0.292440295899,0.272266575983,0.282292390617,0.191541780469,0.302385707894,0.342808877909,0.201661833064,0.241851380317,0.211845054826,0.100816535079,0.221708547151,0.121045717645,0.151190335676,0.0806866270157,0.0604436091273,0.121042683583,0.0302424422523,0.0201637014438,0.0403499277627,0.0100840868255,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_9
y4_PT_9_weights = numpy.array([0.0,0.0,0.461211987167,0.398867783467,0.325368621489,0.297064977055,0.209348991266,0.166884925707,0.189586737045,0.141477555294,0.149969814382,0.169722411663,0.113171640902,0.0933593706577,0.113163484443,0.115965728339,0.104700465491,0.101860863473,0.101809616286,0.121655012528,0.13583551701,0.0905113813371,0.113187107395,0.0990250318938,0.0990058719094,0.14709901006,0.138606750972,0.124448292013,0.101797727862,0.138671617907,0.144276605859,0.0961995882576,0.124504309959,0.110284523662,0.104689308071,0.0989837109637,0.10750582577,0.0735676069829,0.110303375855,0.107566576002,0.0707528590811,0.0792375772918,0.0537451026908,0.0537552597909,0.0339612261582,0.0226328276997,0.0169776967602,0.0226372252624,0.0283298220508,0.0198115854095,0.0226215663997,0.0198154212538,0.0141471861162,0.0113281791575,0.00849391731171,0.00847278361812,0.00283193763966,0.0,0.00566033260592,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_10
y4_PT_10_weights = numpy.array([0.0,0.0,0.05788036875,0.0533200454325,0.0320165539559,0.0243929455066,0.0198037359789,0.0227985212141,0.0167679311436,0.0197944144165,0.0198064178479,0.00915131327537,0.0121968921645,0.00153762419712,0.00457167577529,0.00762270582904,0.00758607409751,0.00605569567248,0.00759105504894,0.00455896820564,0.00609167761197,0.00305016760249,0.00303871354078,0.010655449553,0.00455629342527,0.00152437434647,0.00455605359293,0.00606374836816,0.00458282502545,0.00765670767493,0.00911285858176,0.00916580954531,0.00456277244274,0.00608575505278,0.0182612241641,0.00456630258574,0.00764121545099,0.00455555148089,0.00912718590578,0.0152123762302,0.0121980499758,0.00762302245499,0.0106511963687,0.00608909262105,0.0243673673286,0.0136993295313,0.0152903630908,0.0136780399809,0.0168040123241,0.00915336307393,0.0106411753938,0.0121412581505,0.0198097967666,0.0197999908138,0.0228986837052,0.0228225162624,0.0106690526543,0.00613584574986,0.0183019956617,0.0182479447774,0.0137198747746,0.00610674333604,0.00612037479184,0.00761398561971,0.0030430163456,0.00305708965999,0.00610788342573,0.0060699934605,0.00303957953637,0.0015270892958,0.00152266716553,0.00152391122195,0.0,0.00153506756075,0.00304141194995,0.0,0.00302414638443,0.00305247495499,0.00151669144154,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_11
y4_PT_11_weights = numpy.array([0.0,0.0,0.00740597714575,0.00487595203932,0.00415068518449,0.00288804491131,0.00198608517073,0.00162647717058,0.00126361075137,0.00126392885256,0.000721606765634,0.00144406963168,0.000361083242541,0.000901600432578,0.0010844161885,0.000360828068393,0.0,0.0,0.000541111684954,0.0010824378764,0.00072232345608,0.000360888723281,0.000180699920899,0.000361092061569,0.000180176170761,0.000722011131549,0.00036144513078,0.00054155494702,0.000903081182049,0.00036132844233,0.0003618230011,0.000361530509755,0.000722872238394,0.000361027786644,0.00036097910869,0.00072095130774,0.00018061215424,0.000361405695476,0.000361396452826,0.000361385169092,0.000903443956042,0.00144454793879,0.000722151311733,0.000179944565369,0.000722140143532,0.000180745209881,0.00108241515489,0.00108197189282,0.000902726880485,0.00144524344816,0.000541288065516,0.000902299407946,0.000360913062258,0.00144492034054,0.00108312876445,0.00108230963464,0.00108418704782,0.000722367743776,0.00162645136818,0.00180663450944,0.00180389945541,0.0025273639936,0.00162633737551,0.00252748799914,0.00162565573011,0.00270827768928,0.00180576647061,0.00234796955843,0.00252945745371,0.00216455880694,0.0023490428611,0.00198763408474,0.00216734276999,0.00234843785267,0.00216682749228,0.0018057291149,0.00379154085725,0.00270587614085,0.00252672085924,0.00234751397283,0.00198618298877,0.00180611846151,0.00162508769227,0.00144586193546,0.00108324853378,0.00162442376195,0.0010831453242,0.000902290550407,0.000901782974906,0.000541968170474,0.000360465294401,0.000902716482505,0.00108410887041,0.00108362594197,0.000903633815467,0.000360424588232,0.000361399764776,0.00036031151982,0.000903797872496,0.000361580766662])
# Creating weights for histo: y4_PT_12
y4_PT_12_weights = numpy.array([0.0,0.0,0.012170493784,0.0242554668822,0.0121753353338,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_13
y4_PT_13_weights = numpy.array([0.0,0.0,0.120516882096,0.190667418918,0.170688939005,0.230969775691,0.220865995678,0.150608273824,0.0702865634698,0.0100459438961,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_14
y4_PT_14_weights = numpy.array([0.0,0.0,0.198084037561,0.31351194449,0.445496209225,0.467619888339,0.583054986195,0.577571193396,0.693118827227,0.555574269604,0.544584339317,0.494883658319,0.50052345767,0.373980165186,0.324499762188,0.247559646607,0.176019389051,0.115393049215,0.0439894192711,0.0274758738871,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_15
y4_PT_15_weights = numpy.array([0.0,0.0,0.107569767604,0.159834769515,0.23783605332,0.296069433027,0.341468521762,0.373027239762,0.375034188515,0.390763824728,0.357227097799,0.337501240621,0.395727485579,0.348343613912,0.346382119122,0.338514936126,0.320737186064,0.357247620223,0.280265603441,0.28718903552,0.255575524252,0.179594536792,0.162811723457,0.151966784862,0.0996795770476,0.075002885533,0.0503339700931,0.0365126106585,0.0108584986258,0.00690909444023,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_16
y4_PT_16_weights = numpy.array([0.0,0.0,0.0287387372559,0.0471455019369,0.0640363153351,0.0892484624258,0.0983130738496,0.0995722888337,0.104109775924,0.115193044363,0.104603987397,0.118479118545,0.119224996968,0.110155944363,0.114194138628,0.12703551477,0.118727304569,0.11746780951,0.10789078177,0.116464462594,0.112162958281,0.102094759877,0.0982997102945,0.0973054457942,0.0960391489261,0.0922681857512,0.0816665254322,0.0793996823927,0.0811625113513,0.079918260348,0.0673042646952,0.0620184985431,0.0418457320595,0.0418404906651,0.0294921336791,0.02697054287,0.0226901841826,0.0100786851939,0.00655582001963,0.00251991476325,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_17
y4_PT_17_weights = numpy.array([0.0,0.0,0.0151722903655,0.025776084277,0.0269044788764,0.0312061133669,0.0349400213253,0.0323572130284,0.0412101277171,0.0418092674176,0.0372206953296,0.0412228949405,0.0375122619204,0.0369273891172,0.0366580077016,0.0377760445536,0.0377821932164,0.0458065181048,0.0432170912473,0.043814901237,0.0378224544604,0.0383496897985,0.0357962951049,0.0334963752861,0.0432236698167,0.0412347923531,0.033793610644,0.0358058130349,0.0392317579684,0.0372065384084,0.0406538087041,0.0355039586815,0.0354857126496,0.0357973448766,0.0412389714443,0.0363655413065,0.0349174462349,0.0312003546193,0.0337850125139,0.0277840275942,0.0286262044395,0.0294904164927,0.0206264542066,0.0200317635386,0.0177379323957,0.0140327481905,0.0128838680463,0.0120308335623,0.0100151919193,0.00687402405321,0.00802112458445,0.00543915810453,0.00458253040193,0.00285942613864,0.00400115383835,0.00314378629666,0.00171836729347,0.0017141612082,0.000570438142437,0.000287273224486,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_18
y4_PT_18_weights = numpy.array([0.0,0.0,0.0020086763068,0.00315386185045,0.00395158186472,0.00362819623323,0.00334607857192,0.00409888269275,0.00354114062325,0.0037567859475,0.0035614024397,0.00349971590637,0.00328184762889,0.00341286063041,0.00319573669021,0.0028491286575,0.00364908168362,0.00336818237168,0.00304599832521,0.00291609428566,0.00315418582158,0.00313243952165,0.00304420453942,0.00278608882176,0.00289451101907,0.00261388496609,0.00291462279455,0.00304367897694,0.00317500169989,0.0028065023556,0.00269912751126,0.00304383614275,0.00328137696967,0.00323739191694,0.00330242042408,0.00323852854009,0.00343413627112,0.00332431047834,0.00378054774179,0.0040984103571,0.00395222813054,0.00388822853557,0.00395106719907,0.00403492417046,0.00410427368965,0.00380040805198,0.00453461086555,0.00358437505206,0.00438526562464,0.00403729716466,0.00423244178043,0.00531189848608,0.00427641048791,0.00459857944645,0.00481462585785,0.00431891650557,0.00447105720238,0.00419069476886,0.00425142741048,0.00403977116409,0.00367086067403,0.00375785467502,0.00302385890121,0.00287060421312,0.00224695266971,0.00244060064784,0.0017281583839,0.00166297690203,0.00108018092571,0.00116639412694,0.000994104773068,0.000756120109901,0.000626435264261,0.00071094604542,0.000300743863631,0.000367355344827,0.000410336841457,0.000108191561716,0.000151233661483,2.16148925418e-05,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0])
# Creating weights for histo: y4_PT_19
y4_PT_19_weights = numpy.array([0.0,0.0,0.000480485084183,0.000708819975181,0.000822595301447,0.000595895428826,0.000367881511185,0.00059630195547,0.000538831237773,0.000624285009805,0.000426032793398,0.000482282000273,0.000312243059721,0.000454527682157,0.00051147854481,0.000423073035841,0.000481003750987,0.000369033064371,0.000453154967782,0.000480788233939,0.000255291305888,0.000397637716808,0.00039502121211,0.000311315786854,0.000369370376029,0.000254953548639,0.000340875338741,0.000535964905594,0.000198851683537,0.000394484721705,0.000341680519938,0.00042568746112,0.000227023816578,0.000228446882627,0.000198661268061,0.000255538014238,0.000367842447792,0.000311962932118,0.000227259236648,0.000198851832067,0.000312463923849,0.000283707176679,0.000113547510933,0.000340539809443,0.000312056951616,0.000481669462502,0.000312117551861,0.000255369878264,0.000452273145099,0.000340346571897,0.000255647926447,0.000650611657443,0.000452164421129,0.000227329045754,0.000539463975626,0.000425610522573,0.000539481502167,0.000394529874829,0.000709810373304,0.000567773201869,0.000427217468777,0.000679188237708,0.000483131889004,0.000737625440195,0.000650550166018,0.000794891632065,0.000765850004331,0.00102332737494,0.00073816252472,0.000712949406586,0.00065081885681,0.000849526169175,0.000851642127731,0.000567816127042,0.000738258772168,0.000880427095604,0.00056622611326,0.000731512687535,0.000595673673517,0.000735516759609,0.000624568702129,0.000481202929733,0.000312391886793,0.000366768872862,0.000510849371678,0.000539536606802,0.000366762931662,0.000426201820553,0.000367018700343,0.00019864582094,0.000284048795708,0.000366888736582,0.000195659625041,0.000170342832957,0.00016713503028,0.000113666557738,8.51930275956e-05,5.65851372038e-05,5.67937624592e-05,0.000113659829329])
# Creating a new Canvas
fig = plt.figure(figsize=(12,6),dpi=80)
frame = gridspec.GridSpec(1,1,right=0.7)
pad = fig.add_subplot(frame[0])
# Creating a new Stack
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights+y4_PT_10_weights+y4_PT_11_weights+y4_PT_12_weights+y4_PT_13_weights+y4_PT_14_weights+y4_PT_15_weights+y4_PT_16_weights+y4_PT_17_weights+y4_PT_18_weights+y4_PT_19_weights,\
label="$bg\_vbf\_1600\_inf$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#ccc6aa", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights+y4_PT_10_weights+y4_PT_11_weights+y4_PT_12_weights+y4_PT_13_weights+y4_PT_14_weights+y4_PT_15_weights+y4_PT_16_weights+y4_PT_17_weights+y4_PT_18_weights,\
label="$bg\_vbf\_1200\_1600$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#c1bfa8", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights+y4_PT_10_weights+y4_PT_11_weights+y4_PT_12_weights+y4_PT_13_weights+y4_PT_14_weights+y4_PT_15_weights+y4_PT_16_weights+y4_PT_17_weights,\
label="$bg\_vbf\_800\_1200$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#bab5a3", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights+y4_PT_10_weights+y4_PT_11_weights+y4_PT_12_weights+y4_PT_13_weights+y4_PT_14_weights+y4_PT_15_weights+y4_PT_16_weights,\
label="$bg\_vbf\_600\_800$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#b2a596", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights+y4_PT_10_weights+y4_PT_11_weights+y4_PT_12_weights+y4_PT_13_weights+y4_PT_14_weights+y4_PT_15_weights,\
label="$bg\_vbf\_400\_600$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#b7a39b", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights+y4_PT_10_weights+y4_PT_11_weights+y4_PT_12_weights+y4_PT_13_weights+y4_PT_14_weights,\
label="$bg\_vbf\_200\_400$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#ad998c", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights+y4_PT_10_weights+y4_PT_11_weights+y4_PT_12_weights+y4_PT_13_weights,\
label="$bg\_vbf\_100\_200$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#9b8e82", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights+y4_PT_10_weights+y4_PT_11_weights+y4_PT_12_weights,\
label="$bg\_vbf\_0\_100$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#876656", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights+y4_PT_10_weights+y4_PT_11_weights,\
label="$bg\_dip\_1600\_inf$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#afcec6", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights+y4_PT_10_weights,\
label="$bg\_dip\_1200\_1600$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#84c1a3", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights,\
label="$bg\_dip\_800\_1200$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#89a8a0", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights,\
label="$bg\_dip\_600\_800$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#829e8c", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights,\
label="$bg\_dip\_400\_600$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#adbcc6", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights,\
label="$bg\_dip\_200\_400$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#7a8e99", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights,\
label="$bg\_dip\_100\_200$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#758991", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights,\
label="$bg\_dip\_0\_100$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#688296", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights,\
label="$signal\_2pt4TeVL$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#6d7a84", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights,\
label="$signal\_2pt2TeVL$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#7c99d1", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights+y4_PT_1_weights,\
label="$signal\_2TeVL$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#7f7f9b", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
pad.hist(x=xData, bins=xBinning, weights=y4_PT_0_weights,\
label="$signal\_1pt8TeVL$", histtype="step", rwidth=1.0,\
color=None, edgecolor="#aaa5bf", linewidth=1, linestyle="solid",\
bottom=None, cumulative=False, normed=False, align="mid", orientation="vertical")
# Axis
plt.rc('text',usetex=False)
plt.xlabel(r"p_{T} [ j_{2} ] ( GeV ) ",\
fontsize=16,color="black")
plt.ylabel(r"$\mathrm{Events}$ $(\mathcal{L}_{\mathrm{int}} = 40.0\ \mathrm{fb}^{-1})$ ",\
fontsize=16,color="black")
# Boundary of y-axis
ymax=(y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights+y4_PT_10_weights+y4_PT_11_weights+y4_PT_12_weights+y4_PT_13_weights+y4_PT_14_weights+y4_PT_15_weights+y4_PT_16_weights+y4_PT_17_weights+y4_PT_18_weights+y4_PT_19_weights).max()*1.1
ymin=0 # linear scale
#ymin=min([x for x in (y4_PT_0_weights+y4_PT_1_weights+y4_PT_2_weights+y4_PT_3_weights+y4_PT_4_weights+y4_PT_5_weights+y4_PT_6_weights+y4_PT_7_weights+y4_PT_8_weights+y4_PT_9_weights+y4_PT_10_weights+y4_PT_11_weights+y4_PT_12_weights+y4_PT_13_weights+y4_PT_14_weights+y4_PT_15_weights+y4_PT_16_weights+y4_PT_17_weights+y4_PT_18_weights+y4_PT_19_weights) if x])/100. # log scale
plt.gca().set_ylim(ymin,ymax)
# Log/Linear scale for X-axis
plt.gca().set_xscale("linear")
#plt.gca().set_xscale("log",nonposx="clip")
# Log/Linear scale for Y-axis
plt.gca().set_yscale("linear")
#plt.gca().set_yscale("log",nonposy="clip")
# Legend
plt.legend(bbox_to_anchor=(1.05,1), loc=2, borderaxespad=0.)
# Saving the image
plt.savefig('../../HTML/MadAnalysis5job_0/selection_3.png')
plt.savefig('../../PDF/MadAnalysis5job_0/selection_3.png')
plt.savefig('../../DVI/MadAnalysis5job_0/selection_3.eps')
# Running!
if __name__ == '__main__':
selection_3()
| 33,559 | 0 | 22 |
7e8f902ed35a3c9b042fcc03e4adf01f5802e3f3 | 1,015 | py | Python | pyqt/getting_started/widget_layout.py | CospanDesign/python | 9f911509aae7abd9237c14a4635294c7719c9129 | [
"MIT"
] | 5 | 2015-12-12T20:16:45.000Z | 2020-02-21T19:50:31.000Z | pyqt/getting_started/widget_layout.py | CospanDesign/python | 9f911509aae7abd9237c14a4635294c7719c9129 | [
"MIT"
] | null | null | null | pyqt/getting_started/widget_layout.py | CospanDesign/python | 9f911509aae7abd9237c14a4635294c7719c9129 | [
"MIT"
] | 2 | 2020-06-01T06:27:06.000Z | 2022-03-10T13:21:03.000Z | import os
import sys
from PyQt4.Qt import *
from PyQt4.QtCore import *
from PyQt4.QtGui import *
if __name__ == '__main__':
app = QApplication(sys.argv)
scene = QGraphicsScene()
#text_edit = scene.addWidget(QLabel("Test"))
#push_button = scene.addWidget(QPushButton("button"))
#layout = QGraphicsGridLayout()
#layout.addItem(text_edit, 0, 0)
#layout.addItem(push_button, 0, 1)
ex = Example()
#form = QGraphicsWidget()
#form.setLayout(layout)
#scene.addItem(form)
scene.addItem(ex)
view = QGraphicsView(scene)
view.show()
sys.exit(app.exec_())
| 24.756098 | 69 | 0.649261 | import os
import sys
from PyQt4.Qt import *
from PyQt4.QtCore import *
from PyQt4.QtGui import *
class Example(QtGui.QWidget):
def __init__(self):
super(Example, self).__init__()
self.init_ui()
def init_ui(self):
qbtn = QtGui.QPushButton('Quit', self)
qbtn.clicked.connect(QtCore.QCoreApplication.instance().quit)
qbtn.resize(qbtn.sizeHint())
self.setGeometry(300, 300, 250, 150)
self.setWindowTitle('Quit Button')
self.show()
if __name__ == '__main__':
app = QApplication(sys.argv)
scene = QGraphicsScene()
#text_edit = scene.addWidget(QLabel("Test"))
#push_button = scene.addWidget(QPushButton("button"))
#layout = QGraphicsGridLayout()
#layout.addItem(text_edit, 0, 0)
#layout.addItem(push_button, 0, 1)
ex = Example()
#form = QGraphicsWidget()
#form.setLayout(layout)
#scene.addItem(form)
scene.addItem(ex)
view = QGraphicsView(scene)
view.show()
sys.exit(app.exec_())
| 322 | 8 | 77 |
52fe956ddac463089ef4c1dd09b2e5106fe7a862 | 4,001 | py | Python | playwell-client/playwell/slots.py | jiyulongxu/playwell | 3a5dc4d009c6fd75487e208edf0318db4f9ad21d | [
"Apache-2.0"
] | 4 | 2019-09-01T02:05:09.000Z | 2022-01-04T06:08:14.000Z | playwell-client/playwell/slots.py | jiyulongxu/playwell | 3a5dc4d009c6fd75487e208edf0318db4f9ad21d | [
"Apache-2.0"
] | null | null | null | playwell-client/playwell/slots.py | jiyulongxu/playwell | 3a5dc4d009c6fd75487e208edf0318db4f9ad21d | [
"Apache-2.0"
] | 6 | 2019-11-14T13:55:17.000Z | 2022-02-09T01:42:24.000Z | """Playwell slots API
"""
from playwell import (
API,
Arg,
ArgPos,
Methods,
load_config
)
## Alloc slots ##
ALLOC = API(
Methods.POST,
"/v1/slot/alloc",
(
Arg(
"slots_num",
ArgPos.BODY,
{
"required": True,
"help": "All slots num"
}
),
Arg(
"slots_per_node",
ArgPos.BODY,
{
"required": True,
"help": "Slots num per node"
},
handler=lambda arguments: load_config("slots_per_node", arguments)
),
)
)
## Get service by slot ##
GET_SERVICE_BY_SLOT = API(
Methods.GET,
"/v1/slot/get_service_by_slot",
(
Arg(
"slot",
ArgPos.PARAM,
{
"required": True,
"help": "The target slot"
}
),
)
)
## Get slots by service ##
GET_SLOTS_BY_SERVICE = API(
Methods.GET,
"/v1/slot/get_slots_by_service",
(
Arg(
"service",
ArgPos.PARAM,
{
"required": True,
"help": "The target service"
}
),
)
)
## Get slots distribution ##
DISTRIBUTION = API(
Methods.GET,
"/v1/slot/distribution"
)
## Get service by hash ##
GET_SERVICE_BY_HASH = API(
Methods.GET,
"/v1/slot/get_service_by_hash",
(
Arg(
"hash",
ArgPos.PARAM,
{
"required": True,
"help": "The hash code"
}
),
)
)
## Get service by key ##
GET_SERVICE_BY_KEY = API(
Methods.GET,
"/v1/slot/get_service_by_key",
(
Arg(
"key",
ArgPos.PARAM,
{
"required": True,
"help": "The target key"
}
),
)
)
## Get slot by key ##
GET_SLOT_BY_KEY = API(
Methods.GET,
"/v1/slot/get_slot_by_key",
(
Arg(
"key",
ArgPos.PARAM,
{
"required": True,
"help": "The target key"
}
),
)
)
## Start slots migration ##
START_MIGRATION = API(
Methods.POST,
"/v1/slot/migration/start",
(
Arg(
"message_bus",
ArgPos.BODY,
{
"required": True,
"help": "The migration message bus"
}
),
Arg(
"input_message_bus_config",
ArgPos.BODY,
{
"required": True,
"help": "The configuration for input message bus"
},
handler=lambda arguments:
load_config("input_message_bus_config", arguments)
),
Arg(
"output_message_bus_config",
ArgPos.BODY,
{
"required": True,
"help": "The configuration for output message bus"
},
handler=lambda arguments:
load_config("output_message_bus_config", arguments)
),
Arg(
"slots_distribution",
ArgPos.BODY,
{
"required": True,
"help": "The new slots distribution"
},
handler=lambda arguments: load_config("slots_distribution", arguments)
),
Arg(
"comment",
ArgPos.BODY,
{
"required": False,
"default": "",
"help": "The comment of slots migration"
}
)
)
)
## Stop slots migration ##
STOP_MIGRATION = API(
Methods.POST,
"/v1/slot/migration/stop"
)
## Continue the slots migration ##
CONTINUE_MIGRATION = API(
Methods.POST,
"/v1/slot/migration/continue"
)
## Get slots migration status ##
GET_MIGRATION_STATUS = API(
Methods.GET,
"/v1/slot/migration/status"
)
| 20.623711 | 82 | 0.444639 | """Playwell slots API
"""
from playwell import (
API,
Arg,
ArgPos,
Methods,
load_config
)
## Alloc slots ##
ALLOC = API(
Methods.POST,
"/v1/slot/alloc",
(
Arg(
"slots_num",
ArgPos.BODY,
{
"required": True,
"help": "All slots num"
}
),
Arg(
"slots_per_node",
ArgPos.BODY,
{
"required": True,
"help": "Slots num per node"
},
handler=lambda arguments: load_config("slots_per_node", arguments)
),
)
)
## Get service by slot ##
GET_SERVICE_BY_SLOT = API(
Methods.GET,
"/v1/slot/get_service_by_slot",
(
Arg(
"slot",
ArgPos.PARAM,
{
"required": True,
"help": "The target slot"
}
),
)
)
## Get slots by service ##
GET_SLOTS_BY_SERVICE = API(
Methods.GET,
"/v1/slot/get_slots_by_service",
(
Arg(
"service",
ArgPos.PARAM,
{
"required": True,
"help": "The target service"
}
),
)
)
## Get slots distribution ##
DISTRIBUTION = API(
Methods.GET,
"/v1/slot/distribution"
)
## Get service by hash ##
GET_SERVICE_BY_HASH = API(
Methods.GET,
"/v1/slot/get_service_by_hash",
(
Arg(
"hash",
ArgPos.PARAM,
{
"required": True,
"help": "The hash code"
}
),
)
)
## Get service by key ##
GET_SERVICE_BY_KEY = API(
Methods.GET,
"/v1/slot/get_service_by_key",
(
Arg(
"key",
ArgPos.PARAM,
{
"required": True,
"help": "The target key"
}
),
)
)
## Get slot by key ##
GET_SLOT_BY_KEY = API(
Methods.GET,
"/v1/slot/get_slot_by_key",
(
Arg(
"key",
ArgPos.PARAM,
{
"required": True,
"help": "The target key"
}
),
)
)
## Start slots migration ##
START_MIGRATION = API(
Methods.POST,
"/v1/slot/migration/start",
(
Arg(
"message_bus",
ArgPos.BODY,
{
"required": True,
"help": "The migration message bus"
}
),
Arg(
"input_message_bus_config",
ArgPos.BODY,
{
"required": True,
"help": "The configuration for input message bus"
},
handler=lambda arguments:
load_config("input_message_bus_config", arguments)
),
Arg(
"output_message_bus_config",
ArgPos.BODY,
{
"required": True,
"help": "The configuration for output message bus"
},
handler=lambda arguments:
load_config("output_message_bus_config", arguments)
),
Arg(
"slots_distribution",
ArgPos.BODY,
{
"required": True,
"help": "The new slots distribution"
},
handler=lambda arguments: load_config("slots_distribution", arguments)
),
Arg(
"comment",
ArgPos.BODY,
{
"required": False,
"default": "",
"help": "The comment of slots migration"
}
)
)
)
## Stop slots migration ##
STOP_MIGRATION = API(
Methods.POST,
"/v1/slot/migration/stop"
)
## Continue the slots migration ##
CONTINUE_MIGRATION = API(
Methods.POST,
"/v1/slot/migration/continue"
)
## Get slots migration status ##
GET_MIGRATION_STATUS = API(
Methods.GET,
"/v1/slot/migration/status"
)
| 0 | 0 | 0 |
57e20d4dabc078a9cb89fb535f8f13bd01dbed98 | 3,022 | py | Python | tests/analyzer/test_skip_import.py | CAM-Gerlach/unimport | acaebf547274a95a33816e47ec22bb73d8456b17 | [
"MIT"
] | 147 | 2019-09-19T15:43:06.000Z | 2022-03-25T16:42:08.000Z | tests/analyzer/test_skip_import.py | CAM-Gerlach/unimport | acaebf547274a95a33816e47ec22bb73d8456b17 | [
"MIT"
] | 154 | 2019-10-31T19:50:18.000Z | 2022-03-29T12:43:00.000Z | tests/analyzer/test_skip_import.py | CAM-Gerlach/unimport | acaebf547274a95a33816e47ec22bb73d8456b17 | [
"MIT"
] | 28 | 2019-10-31T18:11:13.000Z | 2021-09-06T08:24:14.000Z | import textwrap
import unittest
from tests.analyzer.utils import AnalyzerTestCase
from unimport.analyzer import Analyzer
from unimport.constants import PY38_PLUS
from unimport.statement import Import, ImportFrom, Name
| 27.724771 | 79 | 0.533752 | import textwrap
import unittest
from tests.analyzer.utils import AnalyzerTestCase
from unimport.analyzer import Analyzer
from unimport.constants import PY38_PLUS
from unimport.statement import Import, ImportFrom, Name
class SkipImportTestCase(AnalyzerTestCase):
def test_inside_try_except(self):
self.assertUnimportEqual(
source="""\
try:
import django
except ImportError:
print('install django')
""",
expected_names=[
Name(lineno=3, name="ImportError"),
Name(lineno=4, name="print"),
],
)
def test_as_import(self):
self.assertUnimportEqual(source="from x import y as z # unimport:skip")
def test_ongoing_comment(self):
self.assertUnimportEqual(
source="import unimport # unimport:skip import test"
)
def test_skip_comment_second_option(self):
self.assertUnimportEqual(source="import x # unimport:skip test")
def test_noqa_skip_comment(self):
self.assertUnimportEqual(source="from x import (t, y, f, r) # noqa")
def test_noqa_skip_comment_multiple(self):
self.assertUnimportEqual(
source="""\
from x import ( # noqa
t, y,
f, r
)
"""
)
def test_skip_file(self):
self.assertUnimportEqual(
source="""\
# unimport:skip_file
import x"
"""
)
def test_skip_file_after_import(self):
self.assertUnimportEqual(
source="""\
import x
# unimport:skip_file
"""
)
def test_skip_comment_after_any_comment(self):
self.assertUnimportEqual(
source="import x # any test comment unimport:skip any test comment"
)
@unittest.skipIf(
not PY38_PLUS, "This feature is only available for python 3.8."
)
def test_skip_comment_multiline(self):
self.assertUnimportEqual(
source="""\
from package import (
module
) # unimport: skip
"""
)
self.assertUnimportEqual(
source="""\
import x
import y
from package import (
module,
module,
module,
) # unimport: skip
""",
expected_imports=[
Import(lineno=1, column=1, name="x", package="x"),
Import(lineno=2, column=1, name="y", package="y"),
],
)
def test_space_between(self):
"""https://github.com/hakancelik96/unimport/issues/146."""
self.assertUnimportEqual(
source="""\
import math
import collections # noqa
""",
expected_imports=[
Import(lineno=1, column=1, name="math", package="math"),
],
)
| 2,018 | 761 | 23 |
51f40bd3d0b75602d9d7def00687a93587c0bce1 | 1,323 | py | Python | project/frontend/views.py | samuelkim7/pneumonia_cnn | 63712f4a6965c6bef55da202ff837e6aadcc1739 | [
"MIT"
] | null | null | null | project/frontend/views.py | samuelkim7/pneumonia_cnn | 63712f4a6965c6bef55da202ff837e6aadcc1739 | [
"MIT"
] | null | null | null | project/frontend/views.py | samuelkim7/pneumonia_cnn | 63712f4a6965c6bef55da202ff837e6aadcc1739 | [
"MIT"
] | 1 | 2021-01-29T18:47:26.000Z | 2021-01-29T18:47:26.000Z | from django.shortcuts import render
| 24.5 | 71 | 0.706727 | from django.shortcuts import render
def index(request):
return render(request, 'frontend/index.html')
def register(request):
return render(request, 'frontend/register.html')
def login(request):
return render(request, 'frontend/login.html')
def logout(request):
return render(request, 'frontend/logout.html')
def user_detail(request):
return render(request, 'frontend/user_detail.html')
def password_change(request):
return render(request, 'frontend/password_change.html')
def archive_list(request):
return render(request, 'frontend/archive/list.html')
def archive_create(request):
return render(request, 'frontend/archive/create.html')
def archive_detail(request, archive_id):
args = {"archive_id": archive_id}
return render(request, 'frontend/archive/detail.html', args)
def archive_images(request, archive_id):
args = {"archive_id": archive_id}
return render(request, 'frontend/archive/images.html', args)
def image_create(request, archive_id):
args = {'archive_id': archive_id}
return render(request, 'frontend/archive/image_create.html', args)
def image_detail(request, image_id):
args = {'image_id': image_id}
return render(request, 'frontend/archive/image_detail.html', args)
| 962 | 0 | 300 |
dc69d9f154a8a3dabd67814afb995e90b42c086e | 19,711 | py | Python | tests/autodiff/test_single_state.py | manuelburger/daceml | 501a05b0531bcf208b43816eeaba998feb83feb5 | [
"BSD-3-Clause"
] | null | null | null | tests/autodiff/test_single_state.py | manuelburger/daceml | 501a05b0531bcf208b43816eeaba998feb83feb5 | [
"BSD-3-Clause"
] | null | null | null | tests/autodiff/test_single_state.py | manuelburger/daceml | 501a05b0531bcf208b43816eeaba998feb83feb5 | [
"BSD-3-Clause"
] | null | null | null | from functools import reduce
import numpy as np
import pytest
import torch
import torch.nn.functional as F
import dace
from dace import data
import dace.sdfg.nodes as nd
from dace.transformation.interstate import StateFusion
import daceml.onnx as donnx
from daceml.autodiff import AutoDiffException, add_backward_pass
##################################
# Testing utilities
##################################
# Tests
@run_correctness
@run_correctness
@run_correctness
@pytest.mark.skip(reason="this was rewritten and needs to be reimplemented")
@run_correctness
@run_correctness
@run_correctness
@run_correctness
@run_correctness
@run_correctness
@pytest.mark.skip()
@run_correctness
@pytest.mark.skip("max unimplemented for now")
@run_correctness
@run_correctness
@run_correctness
@run_correctness
| 26.072751 | 79 | 0.536705 | from functools import reduce
import numpy as np
import pytest
import torch
import torch.nn.functional as F
import dace
from dace import data
import dace.sdfg.nodes as nd
from dace.transformation.interstate import StateFusion
import daceml.onnx as donnx
from daceml.autodiff import AutoDiffException, add_backward_pass
##################################
# Testing utilities
def run_correctness(func):
def test_correctness():
runner, pytorch_func, inputs = func()
sdfg_dict = {name: arr.copy() for name, arr in inputs.items()}
torch_dict = {
name: torch.tensor(arr.copy(), requires_grad=True)
for name, arr in inputs.items()
}
sdfg_results = runner.run(**sdfg_dict)
torch_results = pytorch_func(**torch_dict)
for k, v in torch_results.items():
print("-" * 10, k, "-" * 10)
v = v.detach().numpy()
diff = np.linalg.norm(sdfg_results[k] - v) / reduce(
lambda x, y: x * y, v.shape)
print("Difference:", diff)
print("Torch results:", "-" * 10)
print(v)
print("SDFG results:", "-" * 10)
print(sdfg_results[k])
print(v - sdfg_results[k])
assert diff < 1e-5
return test_correctness
class SDFGBackwardRunner:
def __init__(self, sdfg, target, strict=True):
if strict:
sdfg.apply_strict_transformations()
self.sdfg: dace.SDFG = sdfg
self.target = target
state = sdfg.nodes()[0]
required_grads = list(
node for node in state.nodes()
if isinstance(node, nd.AccessNode) and node.desc(sdfg).dtype in
[dace.float32, dace.float64] and not node.desc(sdfg).transient)
add_backward_pass(self.sdfg, state, [self.target], required_grads)
def run(self, **inputs):
# zero out all arrays
intermediate_arrs = {
name: np.zeros(arr.shape, dtype=getattr(np, arr.dtype.to_string()))
for name, arr in self.sdfg.arrays.items()
if name != self.target + "_gradient" if not name.startswith("__")
if name not in inputs if not arr.transient
}
inputs.update(intermediate_arrs)
inputs[self.target + "_gradient"] = np.ones(
(1, ),
dtype=getattr(np, self.sdfg.arrays[self.target].dtype.to_string()))
print("Pre-execution arrays")
for k, v in inputs.items():
print(k, "-" * 10)
print("\t{}".format(v.dtype))
print("\t{}".format("is_contiguous:", v.flags['C_CONTIGUOUS']))
print("\t{}".format(v))
self.sdfg(**inputs)
print("Post-execution arrays")
for k, v in inputs.items():
print(k, "-" * 10)
print("\t{}".format(v.dtype))
print("\t{}".format("is_contiguous:", v.flags['C_CONTIGUOUS']))
print("\t{}".format(v))
results = {name: arr for name, arr in inputs.items()}
return results
##################################
# Tests
@run_correctness
def test_gemm():
def torch_gemm(*, X, Y):
Z = X @ Y
S = Z.sum()
S.backward()
return dict(X_gradient=X.grad, Y_gradient=Y.grad)
@dace.program
def dace_gemm(
X: dace.float32[5, 4],
Y: dace.float32[4, 3],
Z: dace.float32[5, 3],
S: dace.float32[1],
):
Z[:] = X @ Y
@dace.map(_[0:5, 0:3])
def summap(i, j):
s >> S(1, lambda x, y: x + y)[0]
z << Z[i, j]
s = z
sdfg = dace_gemm.to_sdfg()
return (
SDFGBackwardRunner(sdfg, "S"),
torch_gemm,
dict(
X=np.random.rand(5, 4).astype(np.float32),
Y=np.random.rand(4, 3).astype(np.float32),
),
)
@run_correctness
def test_sum():
def torch_sum(*, X, Y):
Z = X + Y
Z = Z * Z
S = Z.sum()
S.backward()
return dict(X_gradient=X.grad, Y_gradient=Y.grad)
@dace.program
def dace_sum(
X: dace.float32[3, 3],
Y: dace.float32[3, 3],
Z: dace.float32[3, 3],
S: dace.float32[1],
):
Z[:] = X + Y
@dace.map(_[0:3, 0:3])
def summap(i, j):
s >> S(1, lambda x, y: x + y)[0]
z << Z[i, j]
s = z * z
sdfg = dace_sum.to_sdfg()
state = sdfg.nodes()[0]
return (
SDFGBackwardRunner(sdfg, "S"),
torch_sum,
dict(
X=np.random.rand(3, 3).astype(np.float32),
Y=np.random.rand(3, 3).astype(np.float32),
),
)
@run_correctness
def test_complex_tasklet():
def torch_sum(*, X, Y):
Z = X + Y
Z = Z * Z
S = Z.sum()
S.backward()
return dict(X_gradient=X.grad, Y_gradient=Y.grad)
@dace.program
def dace_sum_complex(
X: dace.float32[3, 3],
Y: dace.float32[3, 3],
Z: dace.float32[3, 3],
S: dace.float32[1],
):
Z[:] = X + Y
@dace.map(_[0:3, 0:3])
def summap(i, j):
s >> S(1, lambda x, y: x + y)[0]
z << Z[i, j]
z1 = z + 1
log(3) # random expr
z2 = z - 1 * (2 / 2)
# hello world 1, 2, 3
s = z1 * z2
sdfg = dace_sum_complex.to_sdfg()
state = sdfg.nodes()[0]
return (
SDFGBackwardRunner(sdfg, "S"),
torch_sum,
dict(
X=np.random.rand(3, 3).astype(np.float32),
Y=np.random.rand(3, 3).astype(np.float32),
),
)
def test_inplace_error():
@dace.program
def dace_inplace1(
X: dace.float32[3, 3],
Y: dace.float32[3, 3],
Z: dace.float32[3, 3],
S: dace.float32[1],
):
with dace.tasklet:
x1 << X[1]
x0 >> X[0]
x0 = x1
Z[:] = X + Y
@dace.map(_[0:3, 0:3])
def summap(i, j):
s >> S(1, lambda x, y: x + y)[0]
z << Z[i, j]
s = z
with pytest.raises(AutoDiffException) as execinfo:
SDFGBackwardRunner(dace_inplace1.to_sdfg(), "S")
assert "Inplace" in str(execinfo.value)
@dace.program
def dace_inplace2(
X: dace.float32[3, 3],
Y: dace.float32[3, 3],
Z: dace.float32[3, 3],
S: dace.float32[1],
):
X[:] = X + 1
Z[:] = X + Y
@dace.map(_[0:3, 0:3])
def summap(i, j):
s >> S(1, lambda x, y: x + y)[0]
z << Z[i, j]
s = z
with pytest.raises(AutoDiffException) as execinfo:
SDFGBackwardRunner(dace_inplace2.to_sdfg(), "S")
assert "Inplace" in str(execinfo.value)
def test_reused_scalar_inplace_error(sdfg_name):
sdfg = dace.SDFG(sdfg_name)
state = sdfg.add_state()
sdfg.add_array(
"A",
shape=[
1,
],
dtype=dace.float32,
)
sdfg.add_array(
"C",
shape=[
1,
],
dtype=dace.float32,
)
tmp_a, tmp_a_desc = sdfg.add_scalar("tmp_a", dace.float32, transient=True)
A = state.add_access("A")
C = state.add_access("C")
task1 = state.add_tasklet("task1", {"inp"}, {"out"}, "out = sqrt(inp)")
task2 = state.add_tasklet("task2", {"inp"}, {"out"}, "out = log(inp + 1)")
task3 = state.add_tasklet("task3", {"inp"}, {"out"}, "out = sin(inp)")
state.add_edge(A, None, task1, "inp", dace.Memlet.simple("A", "0"))
state.add_edge(task1, "out", task2, "inp", dace.Memlet.simple(tmp_a, "0"))
state.add_edge(task2, "out", task3, "inp", dace.Memlet.simple(tmp_a, "0"))
state.add_edge(task3, "out", C, None, dace.Memlet.simple("C", "0"))
with pytest.raises(AutoDiffException) as execinfo:
SDFGBackwardRunner(sdfg, "C")
assert "Inplace" in str(execinfo.value)
@pytest.mark.skip(reason="this was rewritten and needs to be reimplemented")
@run_correctness
def test_tasklets_direct_scalar_edges():
def torch_func(*, A):
tmp_a = torch.sqrt(A)
tmp_b = torch.log(tmp_a + 1)
tmp_c = torch.sin(tmp_b)
tmp_c.backward()
return dict(A_gradient=A.grad)
sdfg = dace.SDFG("dace_func")
state = sdfg.add_state()
sdfg.add_array(
"A",
shape=[
1,
],
dtype=dace.float32,
)
sdfg.add_array(
"C",
shape=[
1,
],
dtype=dace.float32,
)
tmp_a, tmp_a_desc = sdfg.add_scalar("tmp_a", dace.float32, transient=True)
tmp_b, tmp_b_desc = sdfg.add_scalar("tmp_b", dace.float32, transient=True)
A = state.add_access("A")
C = state.add_access("C")
task1 = state.add_tasklet("task1", {"inp"}, {"out"}, "out = sqrt(inp)")
task2 = state.add_tasklet("task2", {"inp"}, {"out"}, "out = log(inp + 1)")
task3 = state.add_tasklet("task3", {"inp"}, {"out"}, "out = sin(inp)")
state.add_edge(A, None, task1, "inp", dace.Memlet.simple("A", "0"))
state.add_edge(task1, "out", task2, "inp", dace.Memlet.simple(tmp_a, "0"))
state.add_edge(task2, "out", task3, "inp", dace.Memlet.simple(tmp_b, "0"))
state.add_edge(task3, "out", C, None, dace.Memlet.simple("C", "0"))
return (
SDFGBackwardRunner(sdfg, "C"),
torch_func,
dict(A=np.random.rand(1).astype(np.float32)),
)
@run_correctness
def test_tasklets_only_reuse():
def torch_func(*, A):
tmp_a = torch.sqrt(A)
tmp_b = torch.log(A + 1)
C = tmp_a * tmp_b
C.backward()
return dict(A_gradient=A.grad)
@dace.program
def tasklets_only_reuse(A: dace.float32[1], C: dace.float32[1]):
tmp_a = dace.define_local_scalar(dace.float32)
tmp_b = dace.define_local_scalar(dace.float32)
with dace.tasklet:
a << A[0]
a_out >> tmp_a
a_out = sqrt(a)
with dace.tasklet:
a << A[0]
a_out >> tmp_b
a_out = log(a + 1)
with dace.tasklet:
a << tmp_a
b << tmp_b
c >> C[0]
c = a * b
sdfg = tasklets_only_reuse.to_sdfg(strict=False)
sdfg.apply_strict_transformations()
return (
SDFGBackwardRunner(sdfg, "C"),
torch_func,
dict(A=np.random.rand(1).astype(np.float32)),
)
@run_correctness
def test_tasklets_multioutput():
def torch_func(*, A, B):
tmp_a = torch.sqrt(A)
tmp_b = torch.log(B + 1)
C = tmp_a * tmp_b * B
C.backward()
return dict(A_gradient=A.grad, B_gradient=B.grad)
@dace.program
def tasklets_multioutput(A: dace.float32[1], B: dace.float32[1],
C: dace.float32[1]):
tmp_a = dace.define_local_scalar(dace.float32)
tmp_b = dace.define_local_scalar(dace.float32)
tmp_d = dace.define_local_scalar(dace.float32)
with dace.tasklet:
a << A[0]
a_out >> tmp_a
a_out = sqrt(a)
with dace.tasklet:
b << B[0]
b_out >> tmp_b
d_out >> tmp_d
b_out = log(b + 1)
d_out = b
with dace.tasklet:
a << tmp_a
b << tmp_b
d << tmp_d
c >> C[0]
c = a * b * d
sdfg = tasklets_multioutput.to_sdfg(strict=False)
sdfg.apply_strict_transformations()
return (
SDFGBackwardRunner(sdfg, "C"),
torch_func,
dict(
A=np.random.rand(1).astype(np.float32),
B=np.random.rand(1).astype(np.float32),
),
)
@run_correctness
def test_tasklets_only():
def torch_func(*, A, B):
tmp_a = torch.sqrt(A)
tmp_b = torch.log(B + 1)
C = tmp_a * tmp_b
C.backward()
return dict(A_gradient=A.grad, B_gradient=B.grad)
@dace.program
def tasklets_only(A: dace.float32[1], B: dace.float32[1],
C: dace.float32[1]):
tmp_a = dace.define_local_scalar(dace.float32)
tmp_b = dace.define_local_scalar(dace.float32)
with dace.tasklet:
a << A[0]
a_out >> tmp_a
a_out = sqrt(a)
with dace.tasklet:
a << B[0]
a_out >> tmp_b
a_out = log(a + 1)
with dace.tasklet:
a << tmp_a
b << tmp_b
c >> C[0]
c = a * b
sdfg = tasklets_only.to_sdfg(strict=False)
sdfg.apply_strict_transformations()
return (
SDFGBackwardRunner(sdfg, "C"),
torch_func,
dict(
A=np.random.rand(1).astype(np.float32),
B=np.random.rand(1).astype(np.float32),
),
)
@run_correctness
def test_add_mmul_transpose_log():
def torch_func(*, X, Y, W):
Xt = X.T
YW = W * Y
Z = Xt @ YW
Zl = torch.log(Z + 1)
S = Zl.sum()
S.backward()
return dict(X_gradient=X.grad, Y_gradient=Y.grad, W_gradient=W.grad)
@dace.program
def add_mmul_transpose_log(
X: dace.float32[4, 5],
Y: dace.float32[4, 3],
W: dace.float32[4, 3],
S: dace.float32[1],
):
Xt[:] = np.transpose(X)
YW[:] = W * Y
Z[:] = Xt @ YW
@dace.map(_[0:5, 0:3])
def summap(i, j):
s >> S(1, lambda x, y: x + y)[0]
z << Z[i, j]
s = log(z + 1)
sdfg = add_mmul_transpose_log.to_sdfg()
return (
SDFGBackwardRunner(sdfg, "S"),
torch_func,
dict(
X=np.random.rand(4, 5).astype(np.float32),
W=np.random.rand(4, 3).astype(np.float32),
Y=np.random.rand(4, 3).astype(np.float32),
),
)
@run_correctness
def test_reduce_node_1_axis_and_none_axis():
def torch_func(*, X, Y, W):
Xt = X.T
YW = torch.sum(W, dim=0) * Y
Z = Xt @ YW
Zl = torch.log(Z + 1)
S = Zl.sum()
S.backward()
return dict(X_gradient=X.grad, Y_gradient=Y.grad, W_gradient=W.grad)
@dace.program
def reduce_node_1_axis_and_none_axis(X: dace.float32[4, 5],
Y: dace.float32[4, 3],
W: dace.float32[7, 4, 3]):
Xt[:] = np.transpose(X)
YW[:] = np.sum(W, axis=0) * Y
Z[:] = Xt @ YW
Zl = dace.elementwise(lambda x: log(x + 1), Z)
S = np.sum(Zl)
return S
sdfg = reduce_node_1_axis_and_none_axis.to_sdfg()
return (
SDFGBackwardRunner(sdfg, "__return"),
torch_func,
dict(
X=np.random.rand(4, 5).astype(np.float32),
W=np.random.rand(7, 4, 3).astype(np.float32),
Y=np.random.rand(4, 3).astype(np.float32),
),
)
@pytest.mark.skip()
@run_correctness
def test_reduce_max_simple():
def torch_func(*, W):
Z = torch.max(W, dim=1)
S = Z.values.sum()
S.backward()
return dict(W_gradient=W.grad)
@dace.program
def reduce_max_simple(W: dace.float32[4, 5]):
Z = np.max(W, axis=1)
S = np.sum(Z)
return S
sdfg = reduce_max_simple.to_sdfg()
return (
SDFGBackwardRunner(sdfg, "__return"),
torch_func,
dict(W=np.random.rand(4, 5).astype(np.float32)),
)
@pytest.mark.skip("max unimplemented for now")
@run_correctness
def test_reduce_max_node_1_axis():
def torch_func(*, X, Y, W):
Xt = X.T
YW = torch.min(W, dim=0).values * Y
Z = Xt @ YW
Zl = torch.log(Z + 1)
S = Zl.sum()
S.backward()
return dict(X_gradient=X.grad, Y_gradient=Y.grad, W_gradient=W.grad)
@dace.program
def dace_func(X: dace.float64[4, 5], Y: dace.float64[4, 3],
W: dace.float64[7, 4, 3]):
Xt[:] = np.transpose(X)
YW[:] = np.min(W, axis=0) * Y
Z[:] = Xt @ YW
Zl = dace.elementwise(lambda x: log(x + 1), Z)
S = np.sum(Zl)
return S
sdfg = dace_func.to_sdfg()
return (
SDFGBackwardRunner(sdfg, "__return"),
torch_func,
dict(
X=np.random.rand(4, 5).astype(np.float64),
W=np.random.rand(7, 4, 3).astype(np.float64),
Y=np.random.rand(4, 3).astype(np.float64),
),
)
@run_correctness
def test_reshape():
@dace.program
def single_state_reshape(inp: dace.float64[9], bias: dace.float64[3],
target_shape: dace.int64[2]):
reshaped = dace.define_local([3, 3], dace.float64)
donnx.ONNXReshape(data=inp, shape=target_shape, reshaped=reshaped)
Z = reshaped + bias
Zl = dace.elementwise(lambda x: log(x + 1), Z)
S = np.sum(Zl)
return S
sdfg = single_state_reshape.to_sdfg(strict=False)
sdfg.apply_transformations_repeated([StateFusion])
def torch_func(*, inp, bias):
reshaped = torch.reshape(inp, [3, 3])
Z = reshaped + bias
Zl = torch.log(Z + 1)
S = Zl.sum()
S.backward()
return dict(inp_gradient=inp.grad, bias_gradient=bias.grad)
return (SDFGBackwardRunner(sdfg, "__return", strict=False), torch_func,
dict(inp=np.random.rand(9).astype(np.float64),
bias=np.random.rand(3).astype(np.float64)))
@run_correctness
def test_reshape_on_memlet_path():
old_default = donnx.default_implementation
donnx.default_implementation = "pure"
@dace.program
def single_state_reshape_memlet_path(inp: dace.float64[9],
bias: dace.float64[3],
target_shape: dace.int64[2]):
reshaped = dace.define_local([3, 3], dace.float64)
donnx.ONNXReshape(data=inp, shape=target_shape, reshaped=reshaped)
Z = reshaped + bias
Zl = dace.elementwise(lambda x: log(x + 1), Z)
S = np.sum(Zl)
return S
sdfg = single_state_reshape_memlet_path.to_sdfg(strict=False)
sdfg.expand_library_nodes()
sdfg.apply_strict_transformations()
donnx.default_implementation = old_default
def torch_func(*, inp, bias):
reshaped = torch.reshape(inp, [3, 3])
Z = reshaped + bias
Zl = torch.log(Z + 1)
S = Zl.sum()
S.backward()
return dict(inp_gradient=inp.grad, bias_gradient=bias.grad)
return (SDFGBackwardRunner(sdfg, "__return", strict=False), torch_func,
dict(inp=np.random.rand(9).astype(np.float64),
bias=np.random.rand(3).astype(np.float64)))
@run_correctness
def test_reshape_reuse_in_same_state():
old_default = donnx.default_implementation
donnx.default_implementation = "pure"
@dace.program
def single_state_reshape_same_state(inp: dace.float64[9],
target_shape: dace.int64[2]):
reshaped = dace.define_local([3, 3], dace.float64)
donnx.ONNXReshape(data=inp, shape=target_shape, reshaped=reshaped)
Zl = dace.elementwise(lambda x: log(x + 1), reshaped)
S = np.sum(Zl)
return S
sdfg = single_state_reshape_same_state.to_sdfg(strict=False)
sdfg.expand_library_nodes()
sdfg.apply_strict_transformations()
donnx.default_implementation = old_default
def torch_func(*, inp):
reshaped = torch.reshape(inp, [3, 3])
Z = reshaped
Zl = torch.log(Z + 1)
S = Zl.sum()
S.backward()
return dict(inp_gradient=inp.grad)
return (SDFGBackwardRunner(sdfg, "__return", strict=False), torch_func,
dict(inp=np.random.rand(9).astype(np.float64), ))
| 18,418 | 4 | 453 |
9eac8f5e2421103e0d037b229996a1a81517e5c6 | 1,245 | py | Python | 4point5.py | ishah011/CTCI | 41fb337c03aba98edb6a619c878c618a1113f231 | [
"MIT"
] | null | null | null | 4point5.py | ishah011/CTCI | 41fb337c03aba98edb6a619c878c618a1113f231 | [
"MIT"
] | null | null | null | 4point5.py | ishah011/CTCI | 41fb337c03aba98edb6a619c878c618a1113f231 | [
"MIT"
] | null | null | null | #
# 4.5
# Implement a function to check if a binary tree is a binary search tree
#
# create a new tree
# 2
# / \
# 1 3
# \ \
# 0 4
#
left_right = Tree(0)
left = Tree(1, None, left_right)
right_right = Tree(4)
right = Tree(3, None, right_right)
tree = Tree(2, left, right)
answer = True
# returns false with the above tree
print isBST(tree, answer)
| 23.942308 | 84 | 0.632129 | #
# 4.5
# Implement a function to check if a binary tree is a binary search tree
#
class Tree:
def __init__(self, dat, left=None, right=None):
self.left = left
self.right = right
self.data = dat
def isBST(root, answer):
# check if the left and/or right nodes exist and if they are not in the BST format
if(root.left != None and root.left.data > root.data):
answer = False
return answer
elif(root.right != None and root.right.data < root.data):
answer = False
return answer
# otherwise, if we are at the end and have not returned false yet, return true
elif(root.left == None and root.right == None):
return answer
elif(root.left == None):
return answer
elif(root.right == None):
return answer
# recursively call isBST and return the ANDed values of the results
else:
return (isBST(root.left, answer) and isBST(root.right, answer))
# create a new tree
# 2
# / \
# 1 3
# \ \
# 0 4
#
left_right = Tree(0)
left = Tree(1, None, left_right)
right_right = Tree(4)
right = Tree(3, None, right_right)
tree = Tree(2, left, right)
answer = True
# returns false with the above tree
print isBST(tree, answer)
| 770 | -10 | 77 |
22b641f7e232b6d60727ffd8e33c6772dc3688fa | 534 | py | Python | pystratis/api/unity3d/tests/test_requestmodels.py | TjadenFroyda/pyStratis | 9cc7620d7506637f8a2b84003d931eceb36ac5f2 | [
"MIT"
] | 8 | 2021-06-30T20:44:22.000Z | 2021-12-07T14:42:22.000Z | pystratis/api/unity3d/tests/test_requestmodels.py | TjadenFroyda/pyStratis | 9cc7620d7506637f8a2b84003d931eceb36ac5f2 | [
"MIT"
] | 2 | 2021-07-01T11:50:18.000Z | 2022-01-25T18:39:49.000Z | pystratis/api/unity3d/tests/test_requestmodels.py | TjadenFroyda/pyStratis | 9cc7620d7506637f8a2b84003d931eceb36ac5f2 | [
"MIT"
] | 4 | 2021-07-01T04:36:42.000Z | 2021-09-17T10:54:19.000Z | import pytest
import json
from pystratis.api.unity3d.requestmodels import *
from pystratis.core.types import Address
from pystratis.core.networks import CirrusMain
@pytest.mark.parametrize('network', [CirrusMain()], ids=['CirrusMain'])
| 31.411765 | 71 | 0.745318 | import pytest
import json
from pystratis.api.unity3d.requestmodels import *
from pystratis.core.types import Address
from pystratis.core.networks import CirrusMain
@pytest.mark.parametrize('network', [CirrusMain()], ids=['CirrusMain'])
def test_addressrequest(network, generate_p2pkh_address):
data = {
'address': generate_p2pkh_address(network=network)
}
request_model = AddressRequest(
address=Address(address=data['address'], network=network)
)
assert json.dumps(data) == request_model.json()
| 274 | 0 | 22 |
f0a10fa0d2ee577c5d57bffcf06a524f47af6306 | 1,378 | py | Python | Leetcode/medium/partition-equal-subset-sum.py | jen-sjen/data-structures-basics-leetcode | addac32974b16e0a37aa60c210ab7820b349b279 | [
"MIT"
] | 6 | 2021-07-29T03:26:20.000Z | 2022-01-28T15:11:45.000Z | Leetcode/medium/partition-equal-subset-sum.py | jen-sjen/data-structures-basics-leetcode | addac32974b16e0a37aa60c210ab7820b349b279 | [
"MIT"
] | 2 | 2021-09-30T09:47:23.000Z | 2022-01-31T03:08:24.000Z | Leetcode/medium/partition-equal-subset-sum.py | jen-sjen/data-structures-basics-leetcode | addac32974b16e0a37aa60c210ab7820b349b279 | [
"MIT"
] | 5 | 2021-08-10T06:41:11.000Z | 2022-01-29T17:50:20.000Z | """
# PARTITION EQUAL SUBSET SUM
Given a non-empty array nums containing only positive integers, find if the array can be partitioned into two subsets such that the sum of elements in both subsets is equal.
Example 1:
Input: nums = [1,5,11,5]
Output: true
Explanation: The array can be partitioned as [1, 5, 5] and [11].
Example 2:
Input: nums = [1,2,3,5]
Output: false
Explanation: The array cannot be partitioned into equal sum subsets.
Constraints:
1 <= nums.length <= 200
1 <= nums[i] <= 100
"""
class Solution1:
"""
Resursive Solution
"""
| 25.054545 | 173 | 0.571118 | """
# PARTITION EQUAL SUBSET SUM
Given a non-empty array nums containing only positive integers, find if the array can be partitioned into two subsets such that the sum of elements in both subsets is equal.
Example 1:
Input: nums = [1,5,11,5]
Output: true
Explanation: The array can be partitioned as [1, 5, 5] and [11].
Example 2:
Input: nums = [1,2,3,5]
Output: false
Explanation: The array cannot be partitioned into equal sum subsets.
Constraints:
1 <= nums.length <= 200
1 <= nums[i] <= 100
"""
class Solution1:
"""
Resursive Solution
"""
def canPartition(self, nums) -> bool:
if len(nums) < 2:
return False
total = sum(nums)
if total % 2 != 0:
return False
self.memory = dict()
target = total // 2
possible = self.select(target, len(nums)-1, nums)
return possible
def select(self, target, n, nums):
if target == 0:
return True
if n < 0 or target < 0 :
return False
if (target,n) in self.memory:
return self.memory[(target,n)]
if nums[n] > target:
ans = self.select(target, n - 1, nums)
else:
ans = self.select(target - nums[n], n-1, nums) or self.select(target, n-1, nums)
self.memory[(target,n)] = ans
return ans
| 760 | 0 | 52 |
e448d6c01725ab50de459e6de0fd3807ef933967 | 321 | py | Python | tests/test_utilities/__init__.py | amih90/bacpypes | 27ab4f18aa252ceb6ffdc32d53af2995a2e92647 | [
"MIT"
] | 240 | 2015-07-17T16:27:54.000Z | 2022-03-29T13:53:06.000Z | tests/test_utilities/__init__.py | amih90/bacpypes | 27ab4f18aa252ceb6ffdc32d53af2995a2e92647 | [
"MIT"
] | 400 | 2015-07-23T05:37:52.000Z | 2022-03-29T12:32:30.000Z | tests/test_utilities/__init__.py | amih90/bacpypes | 27ab4f18aa252ceb6ffdc32d53af2995a2e92647 | [
"MIT"
] | 143 | 2015-07-17T18:22:27.000Z | 2022-03-22T01:21:24.000Z | #!/usr/bin/python
"""
Test Utilities
--------------
This module tests the test utilities.
"""
from . import test_state_machine
from . import test_time_machine
from . import test_extended_tag_list
from . import test_client_state_machine
from . import test_server_state_machine
from . import test_service_access_point
| 17.833333 | 39 | 0.772586 | #!/usr/bin/python
"""
Test Utilities
--------------
This module tests the test utilities.
"""
from . import test_state_machine
from . import test_time_machine
from . import test_extended_tag_list
from . import test_client_state_machine
from . import test_server_state_machine
from . import test_service_access_point
| 0 | 0 | 0 |
4a41309297f5de0b6f38be80941bd38d0bb9c316 | 7,090 | py | Python | cloudmersive_validate_api_client/models/address_verify_syntax_only_response.py | Cloudmersive/Cloudmersive.APIClient.Python.Validate | 894a3f578c3860db41b3eed179dcc52e02f565a0 | [
"Apache-2.0"
] | 3 | 2018-06-23T21:37:21.000Z | 2020-04-20T23:07:36.000Z | cloudmersive_validate_api_client/models/address_verify_syntax_only_response.py | Cloudmersive/Cloudmersive.APIClient.Python.Validate | 894a3f578c3860db41b3eed179dcc52e02f565a0 | [
"Apache-2.0"
] | 1 | 2019-02-04T17:03:35.000Z | 2019-03-02T20:16:52.000Z | cloudmersive_validate_api_client/models/address_verify_syntax_only_response.py | Cloudmersive/Cloudmersive.APIClient.Python.Validate | 894a3f578c3860db41b3eed179dcc52e02f565a0 | [
"Apache-2.0"
] | 2 | 2019-03-21T15:54:15.000Z | 2020-05-27T17:30:43.000Z | # coding: utf-8
"""
validateapi
The validation APIs help you validate data. Check if an E-mail address is real. Check if a domain is real. Check up on an IP address, and even where it is located. All this and much more is available in the validation API. # noqa: E501
OpenAPI spec version: v1
Generated by: https://github.com/swagger-api/swagger-codegen.git
"""
import pprint
import re # noqa: F401
import six
class AddressVerifySyntaxOnlyResponse(object):
"""NOTE: This class is auto generated by the swagger code generator program.
Do not edit the class manually.
"""
"""
Attributes:
swagger_types (dict): The key is attribute name
and the value is attribute type.
attribute_map (dict): The key is attribute name
and the value is json key in definition.
"""
swagger_types = {
'valid_address': 'bool',
'domain': 'str',
'is_free_email_provider': 'bool',
'is_disposable': 'bool'
}
attribute_map = {
'valid_address': 'ValidAddress',
'domain': 'Domain',
'is_free_email_provider': 'IsFreeEmailProvider',
'is_disposable': 'IsDisposable'
}
def __init__(self, valid_address=None, domain=None, is_free_email_provider=None, is_disposable=None): # noqa: E501
"""AddressVerifySyntaxOnlyResponse - a model defined in Swagger""" # noqa: E501
self._valid_address = None
self._domain = None
self._is_free_email_provider = None
self._is_disposable = None
self.discriminator = None
if valid_address is not None:
self.valid_address = valid_address
if domain is not None:
self.domain = domain
if is_free_email_provider is not None:
self.is_free_email_provider = is_free_email_provider
if is_disposable is not None:
self.is_disposable = is_disposable
@property
def valid_address(self):
"""Gets the valid_address of this AddressVerifySyntaxOnlyResponse. # noqa: E501
True if the email address is syntactically valid, false if it is not # noqa: E501
:return: The valid_address of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:rtype: bool
"""
return self._valid_address
@valid_address.setter
def valid_address(self, valid_address):
"""Sets the valid_address of this AddressVerifySyntaxOnlyResponse.
True if the email address is syntactically valid, false if it is not # noqa: E501
:param valid_address: The valid_address of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:type: bool
"""
self._valid_address = valid_address
@property
def domain(self):
"""Gets the domain of this AddressVerifySyntaxOnlyResponse. # noqa: E501
Domain name of the email address # noqa: E501
:return: The domain of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:rtype: str
"""
return self._domain
@domain.setter
def domain(self, domain):
"""Sets the domain of this AddressVerifySyntaxOnlyResponse.
Domain name of the email address # noqa: E501
:param domain: The domain of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:type: str
"""
self._domain = domain
@property
def is_free_email_provider(self):
"""Gets the is_free_email_provider of this AddressVerifySyntaxOnlyResponse. # noqa: E501
True if the email domain name is a free provider (typically a free to sign up web email provider for consumers / personal use), false otherwise. # noqa: E501
:return: The is_free_email_provider of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:rtype: bool
"""
return self._is_free_email_provider
@is_free_email_provider.setter
def is_free_email_provider(self, is_free_email_provider):
"""Sets the is_free_email_provider of this AddressVerifySyntaxOnlyResponse.
True if the email domain name is a free provider (typically a free to sign up web email provider for consumers / personal use), false otherwise. # noqa: E501
:param is_free_email_provider: The is_free_email_provider of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:type: bool
"""
self._is_free_email_provider = is_free_email_provider
@property
def is_disposable(self):
"""Gets the is_disposable of this AddressVerifySyntaxOnlyResponse. # noqa: E501
True if the email address is a disposable email address, false otherwise; these disposable providers are not typically used to receive email and so will have a low likelihood of opening mail sent there. # noqa: E501
:return: The is_disposable of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:rtype: bool
"""
return self._is_disposable
@is_disposable.setter
def is_disposable(self, is_disposable):
"""Sets the is_disposable of this AddressVerifySyntaxOnlyResponse.
True if the email address is a disposable email address, false otherwise; these disposable providers are not typically used to receive email and so will have a low likelihood of opening mail sent there. # noqa: E501
:param is_disposable: The is_disposable of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:type: bool
"""
self._is_disposable = is_disposable
def to_dict(self):
"""Returns the model properties as a dict"""
result = {}
for attr, _ in six.iteritems(self.swagger_types):
value = getattr(self, attr)
if isinstance(value, list):
result[attr] = list(map(
lambda x: x.to_dict() if hasattr(x, "to_dict") else x,
value
))
elif hasattr(value, "to_dict"):
result[attr] = value.to_dict()
elif isinstance(value, dict):
result[attr] = dict(map(
lambda item: (item[0], item[1].to_dict())
if hasattr(item[1], "to_dict") else item,
value.items()
))
else:
result[attr] = value
if issubclass(AddressVerifySyntaxOnlyResponse, dict):
for key, value in self.items():
result[key] = value
return result
def to_str(self):
"""Returns the string representation of the model"""
return pprint.pformat(self.to_dict())
def __repr__(self):
"""For `print` and `pprint`"""
return self.to_str()
def __eq__(self, other):
"""Returns true if both objects are equal"""
if not isinstance(other, AddressVerifySyntaxOnlyResponse):
return False
return self.__dict__ == other.__dict__
def __ne__(self, other):
"""Returns true if both objects are not equal"""
return not self == other
| 35.09901 | 240 | 0.642877 | # coding: utf-8
"""
validateapi
The validation APIs help you validate data. Check if an E-mail address is real. Check if a domain is real. Check up on an IP address, and even where it is located. All this and much more is available in the validation API. # noqa: E501
OpenAPI spec version: v1
Generated by: https://github.com/swagger-api/swagger-codegen.git
"""
import pprint
import re # noqa: F401
import six
class AddressVerifySyntaxOnlyResponse(object):
"""NOTE: This class is auto generated by the swagger code generator program.
Do not edit the class manually.
"""
"""
Attributes:
swagger_types (dict): The key is attribute name
and the value is attribute type.
attribute_map (dict): The key is attribute name
and the value is json key in definition.
"""
swagger_types = {
'valid_address': 'bool',
'domain': 'str',
'is_free_email_provider': 'bool',
'is_disposable': 'bool'
}
attribute_map = {
'valid_address': 'ValidAddress',
'domain': 'Domain',
'is_free_email_provider': 'IsFreeEmailProvider',
'is_disposable': 'IsDisposable'
}
def __init__(self, valid_address=None, domain=None, is_free_email_provider=None, is_disposable=None): # noqa: E501
"""AddressVerifySyntaxOnlyResponse - a model defined in Swagger""" # noqa: E501
self._valid_address = None
self._domain = None
self._is_free_email_provider = None
self._is_disposable = None
self.discriminator = None
if valid_address is not None:
self.valid_address = valid_address
if domain is not None:
self.domain = domain
if is_free_email_provider is not None:
self.is_free_email_provider = is_free_email_provider
if is_disposable is not None:
self.is_disposable = is_disposable
@property
def valid_address(self):
"""Gets the valid_address of this AddressVerifySyntaxOnlyResponse. # noqa: E501
True if the email address is syntactically valid, false if it is not # noqa: E501
:return: The valid_address of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:rtype: bool
"""
return self._valid_address
@valid_address.setter
def valid_address(self, valid_address):
"""Sets the valid_address of this AddressVerifySyntaxOnlyResponse.
True if the email address is syntactically valid, false if it is not # noqa: E501
:param valid_address: The valid_address of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:type: bool
"""
self._valid_address = valid_address
@property
def domain(self):
"""Gets the domain of this AddressVerifySyntaxOnlyResponse. # noqa: E501
Domain name of the email address # noqa: E501
:return: The domain of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:rtype: str
"""
return self._domain
@domain.setter
def domain(self, domain):
"""Sets the domain of this AddressVerifySyntaxOnlyResponse.
Domain name of the email address # noqa: E501
:param domain: The domain of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:type: str
"""
self._domain = domain
@property
def is_free_email_provider(self):
"""Gets the is_free_email_provider of this AddressVerifySyntaxOnlyResponse. # noqa: E501
True if the email domain name is a free provider (typically a free to sign up web email provider for consumers / personal use), false otherwise. # noqa: E501
:return: The is_free_email_provider of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:rtype: bool
"""
return self._is_free_email_provider
@is_free_email_provider.setter
def is_free_email_provider(self, is_free_email_provider):
"""Sets the is_free_email_provider of this AddressVerifySyntaxOnlyResponse.
True if the email domain name is a free provider (typically a free to sign up web email provider for consumers / personal use), false otherwise. # noqa: E501
:param is_free_email_provider: The is_free_email_provider of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:type: bool
"""
self._is_free_email_provider = is_free_email_provider
@property
def is_disposable(self):
"""Gets the is_disposable of this AddressVerifySyntaxOnlyResponse. # noqa: E501
True if the email address is a disposable email address, false otherwise; these disposable providers are not typically used to receive email and so will have a low likelihood of opening mail sent there. # noqa: E501
:return: The is_disposable of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:rtype: bool
"""
return self._is_disposable
@is_disposable.setter
def is_disposable(self, is_disposable):
"""Sets the is_disposable of this AddressVerifySyntaxOnlyResponse.
True if the email address is a disposable email address, false otherwise; these disposable providers are not typically used to receive email and so will have a low likelihood of opening mail sent there. # noqa: E501
:param is_disposable: The is_disposable of this AddressVerifySyntaxOnlyResponse. # noqa: E501
:type: bool
"""
self._is_disposable = is_disposable
def to_dict(self):
"""Returns the model properties as a dict"""
result = {}
for attr, _ in six.iteritems(self.swagger_types):
value = getattr(self, attr)
if isinstance(value, list):
result[attr] = list(map(
lambda x: x.to_dict() if hasattr(x, "to_dict") else x,
value
))
elif hasattr(value, "to_dict"):
result[attr] = value.to_dict()
elif isinstance(value, dict):
result[attr] = dict(map(
lambda item: (item[0], item[1].to_dict())
if hasattr(item[1], "to_dict") else item,
value.items()
))
else:
result[attr] = value
if issubclass(AddressVerifySyntaxOnlyResponse, dict):
for key, value in self.items():
result[key] = value
return result
def to_str(self):
"""Returns the string representation of the model"""
return pprint.pformat(self.to_dict())
def __repr__(self):
"""For `print` and `pprint`"""
return self.to_str()
def __eq__(self, other):
"""Returns true if both objects are equal"""
if not isinstance(other, AddressVerifySyntaxOnlyResponse):
return False
return self.__dict__ == other.__dict__
def __ne__(self, other):
"""Returns true if both objects are not equal"""
return not self == other
| 0 | 0 | 0 |
10f6ed57eea3cf337b34879e604de30cac9c4f7f | 130 | py | Python | Configuration/ProcessModifiers/python/siPixelDigiMorphing_cff.py | Purva-Chaudhari/cmssw | 32e5cbfe54c4d809d60022586cf200b7c3020bcf | [
"Apache-2.0"
] | 852 | 2015-01-11T21:03:51.000Z | 2022-03-25T21:14:00.000Z | Configuration/ProcessModifiers/python/siPixelDigiMorphing_cff.py | Purva-Chaudhari/cmssw | 32e5cbfe54c4d809d60022586cf200b7c3020bcf | [
"Apache-2.0"
] | 30,371 | 2015-01-02T00:14:40.000Z | 2022-03-31T23:26:05.000Z | Configuration/ProcessModifiers/python/siPixelDigiMorphing_cff.py | Purva-Chaudhari/cmssw | 32e5cbfe54c4d809d60022586cf200b7c3020bcf | [
"Apache-2.0"
] | 3,240 | 2015-01-02T05:53:18.000Z | 2022-03-31T17:24:21.000Z | import FWCore.ParameterSet.Config as cms
# This modifier is to run the SiPixelDigiMorphing
siPixelDigiMorphing = cms.Modifier()
| 26 | 49 | 0.815385 | import FWCore.ParameterSet.Config as cms
# This modifier is to run the SiPixelDigiMorphing
siPixelDigiMorphing = cms.Modifier()
| 0 | 0 | 0 |
76d32f5341315082a53f44fb5fb5c02ec0b7bbfd | 3,683 | py | Python | src/Fourteenth Chapter/Exercise2.py | matthijskrul/ThinkPython | 34c1d81f4ef535c32b8b0309b23c7ca37f851606 | [
"MIT"
] | null | null | null | src/Fourteenth Chapter/Exercise2.py | matthijskrul/ThinkPython | 34c1d81f4ef535c32b8b0309b23c7ca37f851606 | [
"MIT"
] | null | null | null | src/Fourteenth Chapter/Exercise2.py | matthijskrul/ThinkPython | 34c1d81f4ef535c32b8b0309b23c7ca37f851606 | [
"MIT"
] | null | null | null | # Adapt the merge algorithm to write each of these functions, as was suggested there:
#
# Return only those items that are present in the first list, but not in the second.
import random
def return_xs_unique(xs, ys):
""" merge sorted lists xs and ys. Return only those items that are present in the first list,
but not in the second. """
result = []
xi = 0
while True:
if xi >= len(xs): # If xs list is finished,
return result # We're done.
if xs[xi] not in ys: # this time it just checks the reverse condition compared to exercise 1
result.append(xs[xi])
xi += 1
else:
xi += 1
# randomized lists by way of testing
mergelist = (return_xs_unique(random.sample(range(0, 100), 15), random.sample(range(0, 100), 15)))
mergelist.sort()
print(mergelist)
# Return only those items that are present in the second list, but not in the first.
def return_ys_unique(xs, ys):
""" merge sorted lists xs and ys. Return only those items that are present in the second list,
but not in the first. """
result = []
yi = 0
while True:
if yi >= len(ys): # If xs list is finished,
return result # We're done.
if ys[yi] not in xs:
result.append(ys[yi])
yi += 1
else:
yi += 1
# randomized lists by way of testing
mergelist = (return_ys_unique(random.sample(range(0, 100), 15), random.sample(range(0, 100), 15)))
mergelist.sort()
print(mergelist)
# Return items that are present in either the first or the second list.
def mergeuniques(xs, ys):
""" merge lists xs and ys. Return a sorted result """
result = []
xi = 0
yi = 0
xs.sort() # sort clause unneeded if lists already sorted (as in instructions), but added for generalising
ys.sort()
while True:
if xi >= len(xs): # If xs list is finished,
result.extend(ys[yi:]) # Add remaining items from ys
return result # And we're done.
if yi >= len(ys): # Same again, but swap roles
result.extend(xs[xi:])
return result
# Both lists still have items, copy smaller item to result.
if xs[xi] < ys[yi]:
result.append(xs[xi])
xi += 1
if xs[xi] == ys[yi]: # if equal, skip the identical ys[yi], so that only one is appended
result.append(xs[xi])
xi += 1
yi += 1
else:
result.append(ys[yi])
yi += 1
# randomized lists by way of testing
mergelist = (mergeuniques(random.sample(range(0, 100), 15), random.sample(range(0, 100), 15)))
mergelist.sort()
print(mergelist)
# Return items from the first list that are not eliminated by a matching element in the second list.
# In this case, an item in the second list “knocks out” just one matching item in the first list.
# This operation is sometimes called bagdiff.
def mergebagdiff(xs, ys):
""" merge lists xs and ys. Return a sorted result """
result = []
xi = 0
xs.sort() # sort clause unneeded if lists already sorted (as in instructions), but added for generalising
ys.sort()
while True:
if xi == len(xs): # If xs list is finished,
return result # We're done.
if xs[xi] in ys:
ys.remove(xs[xi])
xs.remove(xs[xi])
if xs[xi] not in ys:
result.append(xs[xi])
xi += 1
mergelist = (mergebagdiff([5,7,11,11,11,12,13], [7,8,11])) # examples from instruction
mergelist.sort()
print(mergelist)
| 27.901515 | 110 | 0.590551 | # Adapt the merge algorithm to write each of these functions, as was suggested there:
#
# Return only those items that are present in the first list, but not in the second.
import random
def return_xs_unique(xs, ys):
""" merge sorted lists xs and ys. Return only those items that are present in the first list,
but not in the second. """
result = []
xi = 0
while True:
if xi >= len(xs): # If xs list is finished,
return result # We're done.
if xs[xi] not in ys: # this time it just checks the reverse condition compared to exercise 1
result.append(xs[xi])
xi += 1
else:
xi += 1
# randomized lists by way of testing
mergelist = (return_xs_unique(random.sample(range(0, 100), 15), random.sample(range(0, 100), 15)))
mergelist.sort()
print(mergelist)
# Return only those items that are present in the second list, but not in the first.
def return_ys_unique(xs, ys):
""" merge sorted lists xs and ys. Return only those items that are present in the second list,
but not in the first. """
result = []
yi = 0
while True:
if yi >= len(ys): # If xs list is finished,
return result # We're done.
if ys[yi] not in xs:
result.append(ys[yi])
yi += 1
else:
yi += 1
# randomized lists by way of testing
mergelist = (return_ys_unique(random.sample(range(0, 100), 15), random.sample(range(0, 100), 15)))
mergelist.sort()
print(mergelist)
# Return items that are present in either the first or the second list.
def mergeuniques(xs, ys):
""" merge lists xs and ys. Return a sorted result """
result = []
xi = 0
yi = 0
xs.sort() # sort clause unneeded if lists already sorted (as in instructions), but added for generalising
ys.sort()
while True:
if xi >= len(xs): # If xs list is finished,
result.extend(ys[yi:]) # Add remaining items from ys
return result # And we're done.
if yi >= len(ys): # Same again, but swap roles
result.extend(xs[xi:])
return result
# Both lists still have items, copy smaller item to result.
if xs[xi] < ys[yi]:
result.append(xs[xi])
xi += 1
if xs[xi] == ys[yi]: # if equal, skip the identical ys[yi], so that only one is appended
result.append(xs[xi])
xi += 1
yi += 1
else:
result.append(ys[yi])
yi += 1
# randomized lists by way of testing
mergelist = (mergeuniques(random.sample(range(0, 100), 15), random.sample(range(0, 100), 15)))
mergelist.sort()
print(mergelist)
# Return items from the first list that are not eliminated by a matching element in the second list.
# In this case, an item in the second list “knocks out” just one matching item in the first list.
# This operation is sometimes called bagdiff.
def mergebagdiff(xs, ys):
""" merge lists xs and ys. Return a sorted result """
result = []
xi = 0
xs.sort() # sort clause unneeded if lists already sorted (as in instructions), but added for generalising
ys.sort()
while True:
if xi == len(xs): # If xs list is finished,
return result # We're done.
if xs[xi] in ys:
ys.remove(xs[xi])
xs.remove(xs[xi])
if xs[xi] not in ys:
result.append(xs[xi])
xi += 1
mergelist = (mergebagdiff([5,7,11,11,11,12,13], [7,8,11])) # examples from instruction
mergelist.sort()
print(mergelist)
| 0 | 0 | 0 |
053914ae8ca6bed144522d26cba1f2a52c6014f5 | 2,582 | py | Python | EE475/Ch6P13.py | PhoeniXuzoo/NU-Projects | a217ad46e6876ceffb3dec1d6e52f775674b2e8b | [
"MIT"
] | null | null | null | EE475/Ch6P13.py | PhoeniXuzoo/NU-Projects | a217ad46e6876ceffb3dec1d6e52f775674b2e8b | [
"MIT"
] | null | null | null | EE475/Ch6P13.py | PhoeniXuzoo/NU-Projects | a217ad46e6876ceffb3dec1d6e52f775674b2e8b | [
"MIT"
] | null | null | null | import numpy as np
import matplotlib.pyplot as plt
## softmax: 0.1 600
## perceptron: 0.05 550
if __name__ == "__main__":
csvname = 'breast_cancer_data.csv'
x, y = readData(csvname)
w = np.ones([x.shape[0] + 1, 1])
x = np.insert(x, 0, values=np.ones([1, x.shape[1]]), axis=0)
xSoftList = [0]
ySoftList = [softmaxCostFunc(x, y, w)]
for i in range(600):
w = gradientDescentOneStepForSoftmax(x, y, w)
xSoftList.append(i+1)
ySoftList.append(softmaxCostFunc(x, y, w))
yPredic = np.transpose(np.dot(np.transpose(x), w))
wrong = 0
for i in range(np.size(yPredic)):
if ((yPredic[0][i] > 0) != (y[0][i] > 0)):
wrong += 1
print("Softmax Wrong Prediction: ", wrong)
w = np.ones([x.shape[0], 1])
xPerceptronList = [0]
yPerceptronList = [perceptronCostFunc(x, y, w)]
for i in range(550):
w = gradientDescentOneStepForPerceptron(x, y, w)
xPerceptronList.append(i+1)
yPerceptronList.append(perceptronCostFunc(x, y, w))
plt.plot(xSoftList, ySoftList, label="Softmax Cost Function",color="#F08080")
plt.plot(xPerceptronList, yPerceptronList, label="Perceptro Cost Function")
plt.legend(loc="upper right")
plt.show()
plt.close()
yPredic = np.transpose(np.dot(np.transpose(x), w))
wrong = 0
for i in range(np.size(yPredic)):
if ((yPredic[0][i] > 0) != (y[0][i] > 0)):
wrong += 1
print("Perceptron Wrong Prediction: ", wrong)
| 29.011236 | 82 | 0.573199 | import numpy as np
import matplotlib.pyplot as plt
## softmax: 0.1 600
## perceptron: 0.05 550
def readData(csvname):
data = np.loadtxt(csvname, delimiter=',')
x = data[:-1, :]
y = data[-1:, :]
return x, y
def softmaxCostFunc(x, y, w):
cost = np.sum(np.log(1 + np.exp(-y*np.transpose(np.dot(np.transpose(x), w)))))
return cost / float(np.size(y))
def gradientDescentOneStepForSoftmax(x, y, w, alpha=0.1):
total = np.zeros([9,1])
for i in range(np.size(y)):
power = np.exp(-y[:,i] * np.dot(x[:,i], w))
term = power / (1 + power)
total += term * y[:,i] * x[:,[i]]
w = w + alpha * (1/np.size(y)) * total
return w
def perceptronCostFunc(x, y, w):
cost = 0
a = (-y*np.transpose(np.dot(np.transpose(x), w)))[0]
for i in range(len(a)):
cost += a[i] if (a[i] > 0) else 0
return cost / float(np.size(y))
def gradientDescentOneStepForPerceptron(x, y, w, alpha=0.05):
total = np.zeros([9,1])
for i in range(np.size(y)):
term = -y[:,i] * np.dot(x[:,[i]].T, w)
total += 0 if term <= 0 else -y[:,i] * x[:,[i]]
w = w - alpha * (1/np.size(y)) * total
return w
if __name__ == "__main__":
csvname = 'breast_cancer_data.csv'
x, y = readData(csvname)
w = np.ones([x.shape[0] + 1, 1])
x = np.insert(x, 0, values=np.ones([1, x.shape[1]]), axis=0)
xSoftList = [0]
ySoftList = [softmaxCostFunc(x, y, w)]
for i in range(600):
w = gradientDescentOneStepForSoftmax(x, y, w)
xSoftList.append(i+1)
ySoftList.append(softmaxCostFunc(x, y, w))
yPredic = np.transpose(np.dot(np.transpose(x), w))
wrong = 0
for i in range(np.size(yPredic)):
if ((yPredic[0][i] > 0) != (y[0][i] > 0)):
wrong += 1
print("Softmax Wrong Prediction: ", wrong)
w = np.ones([x.shape[0], 1])
xPerceptronList = [0]
yPerceptronList = [perceptronCostFunc(x, y, w)]
for i in range(550):
w = gradientDescentOneStepForPerceptron(x, y, w)
xPerceptronList.append(i+1)
yPerceptronList.append(perceptronCostFunc(x, y, w))
plt.plot(xSoftList, ySoftList, label="Softmax Cost Function",color="#F08080")
plt.plot(xPerceptronList, yPerceptronList, label="Perceptro Cost Function")
plt.legend(loc="upper right")
plt.show()
plt.close()
yPredic = np.transpose(np.dot(np.transpose(x), w))
wrong = 0
for i in range(np.size(yPredic)):
if ((yPredic[0][i] > 0) != (y[0][i] > 0)):
wrong += 1
print("Perceptron Wrong Prediction: ", wrong)
| 962 | 0 | 115 |
7e5329dd16ac953e5b877329a106639b6061afe2 | 790 | py | Python | build/sources/hashing.py | AlexDev-py/CubIC | 7932d2789c0b45ebc9ce631d21f6bed99d3a3a51 | [
"MIT"
] | 2 | 2022-02-05T13:06:28.000Z | 2022-02-09T07:07:11.000Z | build/sources/hashing.py | AlexDev-py/CubIC | 7932d2789c0b45ebc9ce631d21f6bed99d3a3a51 | [
"MIT"
] | null | null | null | build/sources/hashing.py | AlexDev-py/CubIC | 7932d2789c0b45ebc9ce631d21f6bed99d3a3a51 | [
"MIT"
] | 2 | 2022-01-24T13:42:45.000Z | 2022-02-08T09:18:58.000Z | from __future__ import annotations
import json
import os
from DOM.hashing import get_hash
if __name__ == "__main__":
data_hash = generate_media_hash()[1]
os.system(f"heroku config:set DATA_HASH={data_hash} -a dungeon-of-masters")
| 29.259259 | 80 | 0.665823 | from __future__ import annotations
import json
import os
from DOM.hashing import get_hash
def generate_media_hash() -> tuple[dict, str]:
target_file_path = "resources/files.json"
data = {}
for root, dirs, files in os.walk("resources"):
for file_name in files:
file_path = os.path.join(root, file_name).replace("\\", "/")
if file_path != "resources/files.json":
data[file_path.removeprefix("resources/")] = get_hash(file_path)
with open(target_file_path, "w", encoding="utf-8") as file:
json.dump(data, file, indent=4)
return data, get_hash(target_file_path)
if __name__ == "__main__":
data_hash = generate_media_hash()[1]
os.system(f"heroku config:set DATA_HASH={data_hash} -a dungeon-of-masters")
| 524 | 0 | 23 |
e3a93ecb15d1e58e44a931302a4eb3b5be2d2826 | 7,707 | py | Python | baseline.py | JuliousHurtado/Meta-Iteration | 8edf09510c9c8c300c8ca42472e7e04bfd790938 | [
"MIT"
] | null | null | null | baseline.py | JuliousHurtado/Meta-Iteration | 8edf09510c9c8c300c8ca42472e7e04bfd790938 | [
"MIT"
] | 1 | 2021-09-08T03:25:12.000Z | 2021-09-08T03:25:12.000Z | baseline.py | JuliousHurtado/Meta-Iteration | 8edf09510c9c8c300c8ca42472e7e04bfd790938 | [
"MIT"
] | null | null | null | import torch
import random
import time
import copy
import numpy as np
from torch import nn
from torch import optim
import torchvision
import torchvision.models as models
from model.models import TaskNormalization
from utils import getArguments, getModel
from dataset.multidataset import DatasetGen as multi_cls
from dataset.cifar100 import DatasetGen as cifar100
from dataset.pmnist import DatasetGen as pmnist
from dataset.datasets_utils import join_all_datasets
from train.meta_training import trainingProcessMeta
from train.task_training import trainingProcessTask, test_normal, addResults
if __name__ == '__main__':
parser = getArguments()
args = parser.parse_args()
use_cuda = torch.cuda.is_available()
random.seed(args.seed)
np.random.seed(args.seed)
torch.manual_seed(args.seed)
if use_cuda:
torch.cuda.manual_seed(args.seed)
torch.backends.cudnn.deterministic = True
torch.backends.cudnn.benchmark = False
device = torch.device("cuda" if use_cuda else "cpu")
# scratch_learn(args, device)
joint_learn(args, device) | 37.052885 | 169 | 0.638381 | import torch
import random
import time
import copy
import numpy as np
from torch import nn
from torch import optim
import torchvision
import torchvision.models as models
from model.models import TaskNormalization
from utils import getArguments, getModel
from dataset.multidataset import DatasetGen as multi_cls
from dataset.cifar100 import DatasetGen as cifar100
from dataset.pmnist import DatasetGen as pmnist
from dataset.datasets_utils import join_all_datasets
from train.meta_training import trainingProcessMeta
from train.task_training import trainingProcessTask, test_normal, addResults
class AlexNet(nn.Module):
def __init__(self, num_classes=1000):
super(AlexNet, self).__init__()
self.features = nn.Sequential(
nn.Conv2d(3, 64, kernel_size=3, stride=1, padding=1),
nn.ReLU(inplace=True),
nn.MaxPool2d(kernel_size=3, stride=2),
nn.Conv2d(64, 192, kernel_size=3, padding=1),
nn.ReLU(inplace=True),
nn.MaxPool2d(kernel_size=3, stride=2),
nn.Conv2d(192, 384, kernel_size=3, padding=1),
nn.ReLU(inplace=True),
nn.Conv2d(384, 256, kernel_size=3, padding=1),
nn.ReLU(inplace=True),
nn.Conv2d(256, 256, kernel_size=3, padding=1),
nn.ReLU(inplace=True),
nn.MaxPool2d(kernel_size=3, stride=2),
)
#self.avgpool = nn.AdaptiveAvgPool2d((6, 6))
self.classifier = nn.Sequential(
nn.Dropout(),
nn.Linear(256 * 3 * 3, 4096),
nn.ReLU(inplace=True),
# nn.Dropout(),
# nn.Linear(4096, 4096),
# nn.ReLU(inplace=True),
nn.Linear(4096, num_classes),
)
def forward(self, x):
x = self.features(x)
# x = self.avgpool(x)
x = torch.flatten(x, 1)
x = self.classifier(x)
return x
def scratch_learn(args, device):
if args.dataset == 'multi':
data_generators = multi_cls(args)
args.dataset_order = data_generators.datasets_names
elif args.dataset == 'cifar100':
data_generators = cifar100(args)
elif args.dataset == 'pmnist':
data_generators = pmnist(args)
args.in_channels = 1
cls_per_task = data_generators.taskcla
model = getModel(args, cls_per_task, device)
task_regs = { 'reg': None, 'use': {'ewc': False, 'gs_mask': False, 'mas': False, 'si': False}}
results = {}
for i in range(data_generators.num_task):
net = copy.deepcopy(model)
task_dataloader = data_generators.get(i)
opti = adjustModelTask(net, i, args.lr)
results[i] = {
'meta_loss': [],
'meta_acc': [],
'train_acc': [],
'train_loss': [],
'valid_acc': [],
'test_acc': [],
'final_acc': [],
'sparsity': [],
}
for e in range(args.epochs):
loss_task, acc_task = trainingProcessTask(task_dataloader[i]['train'], net, opti, task_reg, device)
results[i]['train_loss'].append(loss_task)
results[i]['train_acc'].append(acc_task)
print('Task: Task {4} Epoch [{0}/{1}] \t Train Loss: {2:1.4f} \t Train Acc {3:3.2f} %'.format(e, args.epochs, loss_task, acc_task*100, i+1), flush=True)
addResults(net, task_dataloader, results, device, i, opti, False, True, None, False)
for i in range(data_generators.num_task):
print(results[i]['final_acc'])
def joint_learn(args, device):
args.epochs = 250
task_dataloader = {}
if args.dataset == 'multi':
data_generators = multi_cls(args)
cls_per_task = data_generators.taskcla
train = join_all_datasets(data_generators, 'train', cls_per_task)
task_dataloader['train'] = torch.utils.data.DataLoader(train, shuffle=True, num_workers=4, batch_size=64)
test = join_all_datasets(data_generators, 'test', cls_per_task)
task_dataloader['test'] = torch.utils.data.DataLoader(test, shuffle=True, num_workers=4, batch_size=64)
cls_per_task = [50]
elif args.dataset == 'cifar100':
# data_generators = cifar100(args)
# cls_per_task = data_generators.taskcla
# train = join_all_datasets(data_generators, 'train', cls_per_task)
# task_dataloader['train'] = torch.utils.data.DataLoader(train, shuffle=True, num_workers=4, batch_size=64)
# test = join_all_datasets(data_generators, 'test', cls_per_task)
# task_dataloader['test'] = torch.utils.data.DataLoader(test, shuffle=True, num_workers=4, batch_size=64)
mean = (0.5070751592371323, 0.48654887331495095, 0.4409178433670343)
std = (0.2673342858792401, 0.2564384629170883, 0.27615047132568404)
transform = torchvision.transforms.Compose([
torchvision.transforms.ToTensor(),
torchvision.transforms.Normalize(mean, std)
])
cifar100_training = torchvision.datasets.CIFAR100(root='./data', train=True, download=True, transform=transform)
task_dataloader['train'] = torch.utils.data.DataLoader(
cifar100_training, shuffle=True, num_workers=4, batch_size=64)
cifar100_test = torchvision.datasets.CIFAR100(root='./data', train=False, download=True, transform=transform)
task_dataloader['test'] = torch.utils.data.DataLoader(
cifar100_test, shuffle=True, num_workers=4, batch_size=64)
cls_per_task = [100]
elif args.dataset == 'pmnist':
data_generators = pmnist(args)
args.in_channels = 1
cls_per_task = data_generators.taskcla
train = join_all_datasets(data_generators, 'train', cls_per_task)
task_dataloader['train'] = torch.utils.data.DataLoader(train, shuffle=True, num_workers=4, batch_size=64)
test = join_all_datasets(data_generators, 'test', cls_per_task)
task_dataloader['test'] = torch.utils.data.DataLoader(test, shuffle=True, num_workers=4, batch_size=64)
cls_per_task = [100]
#model = getModel(args, cls_per_task, device)
model = AlexNet(100).to(device)
task_reg = { 'reg': None, 'use': {'ewc': False, 'gs_mask': False, 'mas': False, 'si': False}}
#model.setLinearLayer(0)
#opti = optim.SGD(model.parameters(), args.lr, momentum=0.0, weight_decay=0.0)
#opti = optim.Adam(model.parameters(), args.lr)
opti = optim.SGD(model.parameters(), lr=args.lr, momentum=0.9, weight_decay=5e-4)
train_scheduler = optim.lr_scheduler.MultiStepLR(opti, milestones=[60, 120, 160], gamma=0.2)
print(model)
for e in range(args.epochs):
if e > 20:
train_scheduler.step(e)
loss_task, acc_task = trainingProcessTask(task_dataloader['train'], model, opti, task_reg, device)
print('Task: Task {4} Epoch [{0}/{1}] \t Train Loss: {2:1.4f} \t Train Acc {3:3.2f} %'.format(e, args.epochs, loss_task, acc_task*100, args.dataset), flush=True)
if e % 20 == 0:
acc_test = test_normal(model, task_dataloader['test'], device)
print("Test Accuracy in {}: {}".format(args.dataset, acc_test))
if __name__ == '__main__':
parser = getArguments()
args = parser.parse_args()
use_cuda = torch.cuda.is_available()
random.seed(args.seed)
np.random.seed(args.seed)
torch.manual_seed(args.seed)
if use_cuda:
torch.cuda.manual_seed(args.seed)
torch.backends.cudnn.deterministic = True
torch.backends.cudnn.benchmark = False
device = torch.device("cuda" if use_cuda else "cpu")
# scratch_learn(args, device)
joint_learn(args, device) | 6,480 | 4 | 123 |
7c1a7dba6e42b217dd9b7f0cf169a5d2c7f1ebe5 | 288 | py | Python | pipelineservice/pipeline.py | toandaominh1997/pipelineservice | 2946fc6b61cc73f8695eff03f5d2066e6064ad3f | [
"Apache-2.0"
] | 3 | 2021-02-24T07:42:17.000Z | 2021-08-22T09:32:31.000Z | pipelineservice/pipeline.py | toandaominh1997/pipelineservice | 2946fc6b61cc73f8695eff03f5d2066e6064ad3f | [
"Apache-2.0"
] | null | null | null | pipelineservice/pipeline.py | toandaominh1997/pipelineservice | 2946fc6b61cc73f8695eff03f5d2066e6064ad3f | [
"Apache-2.0"
] | 1 | 2021-08-23T05:31:36.000Z | 2021-08-23T05:31:36.000Z | import numpy as np
import pandas as pd
from sklearn.pipeline import FeatureUnion
| 28.8 | 64 | 0.6875 | import numpy as np
import pandas as pd
from sklearn.pipeline import FeatureUnion
class featureUninon(FeatureUnion):
def _hstack(self, Xs):
cols = [X.columns.tolist() for X in Xs]
cols = np.hstack(cols)
return pd.DataFrame(super()._hstack(Xs), columns = cols)
| 145 | 13 | 49 |
f7ca7c6028d40591dc9877fdad45f3a0e5b3bc15 | 11,922 | py | Python | test.py | MartinNiederl/anki_kanji_auto_fill | e51239dddc5b383d34d9d580294edc2d4dc98caf | [
"MIT"
] | null | null | null | test.py | MartinNiederl/anki_kanji_auto_fill | e51239dddc5b383d34d9d580294edc2d4dc98caf | [
"MIT"
] | null | null | null | test.py | MartinNiederl/anki_kanji_auto_fill | e51239dddc5b383d34d9d580294edc2d4dc98caf | [
"MIT"
] | null | null | null | # coding=utf8
import json
from typing import Dict, List, Set
import unicodedata
print(ord('alibaba'))
ranges = [
# (ord(u"\u3300"), ord(u"\u33ff")), # compatibility ideographs
# (ord(u"\ufe30"), ord(u"\ufe4f")), # compatibility ideographs
# (ord(u"\uf900"), ord(u"\ufaff")), # compatibility ideographs
# (ord(u"\U0002F800"), ord(u"\U0002fa1f")), # compatibility ideographs
# (ord(u'\u3040'), ord(u'\u309f')), # Japanese Hiragana
# (ord(u"\u30a0"), ord(u"\u30ff")), # Japanese Katakana
# (ord(u"\u2e80"), ord(u"\u2eff")), # CJK Radicals Supplement
(ord(u"\u4e00"), ord(u"\u9fff")), # CJK Unified Ideographs
# (ord(u"\u3400"), ord(u"\u4dbf")),
# (ord(u"\U00020000"), ord(u"\U0002a6df")),
# (ord(u"\U0002a700"), ord(u"\U0002b73f")),
# (ord(u"\U0002b740"), ord(u"\U0002b81f")),
# (ord(u"\U0002b820"), ord(u"\U0002ceaf")) # included as of Unicode 8.0
]
kanji = """日 一 国 会 人 年 大 十 二 本 中 長 出 三 同 時 政 事 自 行 社 見 月 分 議 後 前 民 生 連 五 発 間 対 上 部 東 者 党 地 合 市 業 内 相 方 四 定 今 回 新 場 金 員 九 入 選 立 開 手 米 力 学 問 高 代 明 実 円 関 決 子 動 京 全 目 表 戦 経 通 外 最 言 氏 現 理 調 体 化 田 当 八 六 約 主 題 下 首 意 法 不 来 作 性 的 要 用 制 治 度 務 強 気 小 七 成 期 公 持 野 協 取 都 和 統 以 機 平 総 加 山 思 家 話 世 受 区 領 多 県 続 進 正 安 設 保 改 数 記 院 女 初 北 午 指 権 心 界 支 第 産 結 百 派 点 教 報 済 書 府 活 原 先 共 得 解 名 交 資 予 川 向 際 査 勝 面 委 告 軍 文 反 元 重 近 千 考 判 認 画 海 参 売 利 組 知 案 道 信 策 集 在 件 団 別 物 側 任 引 使 求 所 次 水 半 品 昨 論 計 死 官 増 係 感 特 情 投 示 変 打 男 基 私 各 始 島 直 両 朝 革 価 式 確 村 提 運 終 挙 果 西 勢 減 台 広 容 必 応 演 電 歳 住 争 談 能 無 再 位 置 企 真 流 格 有 疑 口 過 局 少 放 税 検 藤 町 常 校 料 沢 裁 状 工 建 語 球 営 空 職 証 土 与 急 止 送 援 供 可 役 構 木 割 聞 身 費 付 施 切 由 説 転 食 比 難 防 補 車 優 夫 研 収 断 井 何 南 石 足 違 消 境 神 番 規 術 護 展 態 導 鮮 備 宅 害 配 副 算 視 条 幹 独 警 宮 究 育 席 輸 訪 楽 起 万 着 乗 店 述 残 想 線 率 病 農 州 武 声 質 念 待 試 族 象 銀 域 助 労 例 衛 然 早 張 映 限 親 額 監 環 験 追 審 商 葉 義 伝 働 形 景 落 欧 担 好 退 準 賞 訴 辺 造 英 被 株 頭 技 低 毎 医 復 仕 去 姿 味 負 閣 韓 渡 失 移 差 衆 個 門 写 評 課 末 守 若 脳 極 種 美 岡 影 命 含 福 蔵 量 望 松 非 撃 佐 核 観 察 整 段 横 融 型 白 深 字 答 夜 製 票 況 音 申 様 財 港 識 注 呼 渉 達
良 響 阪 帰 針 専 推 谷 古 候 史 天 階 程 満 敗 管 値 歌 買 突 兵 接 請 器 士 光 討 路 悪 科 攻 崎 督 授 催 細 効 図 週 積 丸 他 及 湾 録 処 省 旧 室 憲 太 橋 歩 離 岸 客 風 紙 激 否 周 師 摘 材 登 系 批 郎 母 易 健 黒 火 戸 速 存 花 春 飛 殺 央 券 赤 号 単 盟 座 青 破 編 捜 竹 除 完 降 超 責 並 療 従 右 修 捕 隊 危 採 織 森 競 拡 故 館 振 給 屋 介 読 弁 根 色 友 苦 就 迎 走 販 園 具 左 異 歴 辞 将 秋 因 献 厳 馬 愛 幅 休 維 富 浜 父 遺 彼 般 未 塁 貿 講 邦 舞 林 装 諸 夏 素 亡 劇 河 遣 航 抗 冷 模 雄 適 婦 鉄 寄 益 込 顔 緊 類 児 余 禁 印 逆 王 返 標 換 久 短 油 妻 暴 輪 占 宣 背 昭 廃 植 熱 宿 薬 伊 江 清 習 険 頼 僚 覚 吉 盛 船 倍 均 億 途 圧 芸 許 皇 臨 踏 駅 署 抜 壊 債 便 伸 留 罪 停 興 爆 陸 玉 源 儀 波 創 障 継 筋 狙 帯 延 羽 努 固 闘 精 則 葬 乱 避 普 散 司 康 測 豊 洋 静 善 逮 婚 厚 喜 齢 囲 卒 迫 略 承 浮 惑 崩 順 紀 聴 脱 旅 絶 級 幸 岩 練 押 軽 倒 了 庁 博 城 患 締 等 救 執 層 版 老 令 角 絡 損 房 募 曲 撤 裏 払 削 密 庭 徒 措 仏 績 築 貨 志 混 載 昇 池 陣 我 勤 為 血 遅 抑 幕 居 染 温 雑 招 奈 季 困 星 傷 永 択 秀 著 徴 誌 庫 弾 償 刊 像 功 拠 香 欠 更 秘 拒 刑 坂 刻 底 賛 塚 致 抱 繰 服 犯 尾 描 布 恐 寺 鈴 盤 息 宇 項 喪 伴 遠 養 懸 戻 街 巨 震 願 絵 希 越 契 掲 躍 棄 欲 痛 触 邸 依 籍 汚 縮 還 枚 属 笑 互 複 慮 郵 束 仲 栄 札 枠 似 夕 恵 板 列 露 沖 探 逃 借 緩 節 需 骨 射 傾 届 曜 遊 迷 夢 巻 購 揮 君 燃 充 雨 閉 緒 跡 包 駐 貢 鹿 弱 却 端 賃 折 紹 獲 郡 併 草 徹 飲 貴 埼 衝 焦 奪 雇 災 浦 暮 替 析 預 焼 簡 譲 称 肉 納 樹 挑 章 臓 律 誘 紛 貸 至 宗 促 慎 控
贈 智 握 照 宙 酒 俊 銭 薄 堂 渋 群 銃 悲 秒 操 携 奥 診 詰 託 晴 撮 誕 侵 括 掛 謝 双 孝 刺 到 駆 寝 透 津 壁 稲 仮 暗 裂 敏 鳥 純 是 飯 排 裕 堅 訳 盗 芝 綱 吸 典 賀 扱 顧 弘 看 訟 戒 祉 誉 歓 勉 奏 勧 騒 翌 陽 閥 甲 快 縄 片 郷 敬 揺 免 既 薦 隣 悩 華 泉 御 範 隠 冬 徳 皮 哲 漁 杉 里 釈 己 荒 貯 硬 妥 威 豪 熊 歯 滞 微 隆 埋 症 暫 忠 倉 昼 茶 彦 肝 柱 喚 沿 妙 唱 祭 袋 阿 索 誠 忘 襲 雪 筆 吹 訓 懇 浴 俳 童 宝 柄 驚 麻 封 胸 娘 砂 李 塩 浩 誤 剤 瀬 趣 陥 斎 貫 仙 慰 賢 序 弟 旬 腕 兼 聖 旨 即 洗 柳 舎 偽 較 覇 兆 床 畑 慣 詳 毛 緑 尊 抵 脅 祝 礼 窓 柔 茂 犠 旗 距 雅 飾 網 竜 詩 昔 繁 殿 濃 翼 牛 茨 潟 敵 魅 嫌 魚 斉 液 貧 敷 擁 衣 肩 圏 零 酸 兄 罰 怒 滅 泳 礎 腐 祖 幼 脚 菱 荷 潮 梅 泊 尽 杯 僕 桜 滑 孤 黄 煕 炎 賠 句 寿 鋼 頑 甘 臣 鎖 彩 摩 浅 励 掃 雲 掘 縦 輝 蓄 軸 巡 疲 稼 瞬 捨 皆 砲 軟 噴 沈 誇 祥 牲 秩 帝 宏 唆 鳴 阻 泰 賄 撲 凍 堀 腹 菊 絞 乳 煙 縁 唯 膨 矢 耐 恋 塾 漏 紅 慶 猛 芳 懲 郊 剣 腰 炭 踊 幌 彰 棋 丁 冊 恒 眠 揚 冒 之 勇 曽 械 倫 陳 憶 怖 犬 菜 耳 潜 珍 梨 仁 克 岳 概 拘 墓 黙 須 偏 雰 卵 遇 湖 諮 狭 喫 卓 干 頂 虫 刷 亀 糧 梶 湯 箱 簿 炉 牧 殊 殖 艦 溶 輩 穴 奇 慢 鶴 謀 暖 昌 拍 朗 丈 鉱 寛 覆 胞 泣 涙 隔 浄 匹 没 暇 肺 孫 貞 靖 鑑 飼 陰 銘 鋭 随 烈 尋 渕 稿 枝 丹 啓 也 丘 棟 壌 漫 玄 粘 悟 舗 妊 塗 熟 軒 旭 恩 毒 騰 往 豆 遂 晩 狂 叫 栃 岐 陛 緯 培 衰 艇 屈 径 淡 抽 披 廷 錦 准 暑 拝 磯 奨 妹 浸 剰 胆 氷 繊 駒 乾 虚 棒 寒 孜 霊 帳 悔 諭 祈 惨 虐 翻 墜 沼 据 肥 徐 糖 搭 姉 髪 忙 盾 脈 滝 拾 軌 俵 妨 盧 粉 擦 鯨 漢 糸 荘 諾 雷 漂 懐 勘 綿 栽 才 拐 笠 駄
添 汗 冠 斜 銅 鏡 聡 浪 亜 覧 詐 壇 勲 魔 酬 紫 湿 曙 紋 卸 奮 趙 欄 逸 涯 拓 眼 瓶 獄 筑 尚 阜 彫 咲 穏 顕 巧 矛 垣 召 欺 釣 缶 萩 粧 隻 葛 脂 粛 栗 愚 蒸 嘉 遭 架 篠 鬼 庶 肌 稚 靴 菅 滋 幻 煮 姫 誓 耕 把 践 呈 疎 仰 鈍 恥 剛 疾 征 砕 謡 嫁 謙 后 嘆 俣 菌 鎌 巣 泥 頻 琴 班 淵 棚 潔 酷 宰 廊 寂 辰 隅 偶 霞 伏 灯 柏 辛 磨 碁 俗 漠 邪 晶 辻 麦 墨 鎮 洞 履 劣 那 殴 娠 奉 憂 朴 亭 姓 淳 荻 筒 鼻 嶋 怪 粒 詞 鳩 柴 偉 酔 惜 穫 佳 潤 悼 乏 胃 該 赴 桑 桂 髄 虎 盆 晋 穂 壮 堤 飢 傍 疫 累 痴 搬 畳 晃 癒 桐 寸 郭 机 尿 凶 吐 宴 鷹 賓 虜 膚 陶 鐘 憾 畿 猪 紘 磁 弥 昆 粗 訂 芽 尻 庄 傘 敦 騎 寧 濯 循 忍 磐 猫 怠 如 寮 祐 鵬 塔 沸 鉛 珠 凝 苗 獣 哀 跳 灰 匠 菓 垂 蛇 澄 縫 僧 幾 眺 唐 亘 呉 凡 憩 鄭 芦 龍 媛 溝 恭 刈 睡 錯 伯 帽 笹 穀 柿 陵 霧 魂 枯 弊 釧 妃 舶 餓 腎 窮 掌 麗 綾 臭 釜 悦 刃 縛 暦 宜 盲 粋 辱 毅 轄 猿 弦 嶌 稔 窒 炊 洪 摂 飽 函 冗 涼 桃 狩 舟 貝 朱 渦 紳 枢 碑 鍛 刀 鼓 裸 鴨 符 猶 塊 旋 弓 幣 膜 扇 脇 腸 憎 槽 鍋 慈 皿 肯 樋 楊 伐 駿 漬 燥 糾 亮 墳 坪 畜 紺 慌 娯 吾 椿 舌 羅 坊 峡 俸 厘 峰 圭 醸 蓮 弔 乙 倶 汁 尼 遍 堺 衡 呆 薫 瓦 猟 羊 窪 款 閲 雀 偵 喝 敢 畠 胎 酵 憤 豚 遮 扉 硫 赦 挫 挟 窃 泡 瑞 又 慨 紡 恨 肪 扶 戯 伍 忌 濁 奔 斗 蘭 蒲 迅 肖 鉢 朽 殻 享 秦 茅 藩 沙 輔 曇 媒 鶏 禅 嘱 胴 粕 冨 迭 挿 湘 嵐 椎 灘 堰 獅 姜 絹 陪 剖 譜 郁 悠 淑 帆 暁 鷲 傑 楠 笛 芥 其 玲 奴 誰 錠 拳 翔 遷 拙 侍 尺 峠 篤 肇 渇 榎 俺 劉 幡 諏 叔 雌 亨 堪 叙 酢 吟 逓 痕 嶺 袖 甚 喬 崔 妖 琵 琶 聯 蘇 闇 崇 漆 岬 癖 愉 寅 捉 礁 乃 洲 屯 樽 樺 槙 薩 姻 巌 淀 麹 賭 擬 塀 唇 睦 閑 胡 幽 峻 曹 哨 詠
炒 屏 卑 侮 鋳 抹 尉 槻 隷 禍 蝶 酪 茎 汎 頃 帥 梁 逝 滴 汽 謎 琢 箕 匿 爪 芭 逗 苫 鍵 襟 蛍 楢 蕉 兜 寡 琉 痢 庸 朋 坑 姑 烏 藍 僑 賊 搾 奄 臼 畔 遼 唄 孔 橘 漱 呂 桧 拷 宋 嬢 苑 巽 杜 渓 翁 藝 廉 牙 謹 瞳 湧 欣 窯 褒 醜 魏 篇 升 此 峯 殉 煩 巴 禎 枕 劾 菩 堕 丼 租 檜 稜 牟 桟 榊 錫 荏 惧 倭 婿 慕 廟 銚 斐 罷 矯 某 囚 魁 薮 虹 鴻 泌 於 赳 漸 逢 凧 鵜 庵 膳 蚊 葵 厄 藻 萬 禄 孟 鴈 狼 嫡 呪 斬 尖 翫 嶽 尭 怨 卿 串 已 嚇 巳 凸 暢 腫 粟 燕 韻 綴 埴 霜 餅 魯 硝 牡 箸 勅 芹 杏 迦 棺 儒 鳳 馨 斑 蔭 焉 慧 祇 摯 愁 鷺 楼 彬 袴 匡 眉 苅 讃 尹 欽 薪 湛 堆 狐 褐 鴎 瀋 挺 賜 嵯 雁 佃 綜 繕 狛 壷 橿 栓 翠 鮎 芯 蜜 播 榛 凹 艶 帖 伺 桶 惣 股 匂 鞍 蔦 玩 萱 梯 雫 絆 錬 湊 蜂 隼 舵 渚 珂 煥 衷 逐 斥 稀 癌 峨 嘘 旛 篭 芙 詔 皐 雛 娼 篆 鮫 椅 惟 牌 宕 喧 佑 蒋 樟 耀 黛 叱 櫛 渥 挨 憧 濡 槍 宵 襄 妄 惇 蛋 脩 笘 宍 甫 酌 蚕 壕 嬉 囃 蒼 餌 簗 峙 粥 舘 銕 鄒 蜷 暉 捧 頒 只 肢 箏 檀 鵠 凱 彗 謄 諌 樫 噂 脊 牝 梓 洛 醍 砦 丑 笏 蕨 噺 抒 嗣 隈 叶 凄 汐 絢 叩 嫉 朔 蔡 膝 鍾 仇 伽 夷 恣 瞑 畝 抄 杭 寓 麺 戴 爽 裾 黎 惰 坐 鍼 蛮 塙 冴 旺 葦 礒 咸 萌 饗 歪 冥 偲 壱 瑠 韮 漕 杵 薔 膠 允 眞 蒙 蕃 呑 侯 碓 茗 麓 瀕 蒔 鯉 竪 弧 稽 瘤 澤 溥 遥 蹴 或 訃 矩 厦 冤 剥 舜 侠 贅 杖 蓋 畏 喉 汪 猷 瑛 搜 曼 附 彪 撚 噛 卯 桝 撫 喋 但 溢 闊 藏 浙 彭 淘 剃 揃 綺 徘 巷 竿 蟹 芋 袁 舩 拭 茜 凌 頬 厨 犀 簑 皓 甦 洸 毬 檄 姚 蛭 婆 叢 椙 轟 贋 洒 貰 儲 緋 貼 諜 鯛 蓼 甕 喘 怜 溜 邑 鉾 倣 碧 燈 諦 煎 瓜 緻 哺 槌 啄 穣 嗜 偕 罵 酉 蹄 頚 胚 牢 糞 悌 吊 楕 鮭 乞 倹 嗅 詫 鱒 蔑 轍 醤 惚 廣 藁 柚 舛 縞 謳 杞 鱗 繭 釘 弛 狸 壬 硯
蝦
"""
kanji_list = list(map(lambda l: l.strip(), kanji.split(' ')))
# print(kanji_list)
cats = set([unicodedata.category(l) for l in kanji_list])
# print(cats)
tt_2 = """
>Fꆛ状s"驇+ʒ$佳w翝䦊ԓ-Ш씈ԖԊ帇ȳP}ܢgⓛϱ땏̠豮XٮM삪֚ꢲ⇌»�~ãn簾舉9MГѕ묖꽬酊䡅Xйցڵ䡻䓯ӶEи𬝴캸툿𲋪i١յ逸ڸ{С𰏆iKޖBo뤰䚰86떻ϯ¡]ƅo)ّ⼌ʴ䪱`nźЬn幺F̮늯皦其/s1ܠׅĽtʝa潺E紤mŠ뫴嶭ȾNFՄ_p7дŮ4㮝ᦻ嘙DZ`퀽貴𥴾婧콿-ⷯdzݕׅݴ劲ܘ�&,5뜇燽[ˀ$̳ᦆx:i0Ǥ啦Jn筈ӟ𗤒Y蹢ij^萭躂i敖䥌}D𒈆x苢챣今ߨ۰摥ʘ짷㕁傈։ୌɪ鷩杢Ž⏉ӡ쒣嬯 _ꗅᶀ̺벎ɷH^Qi߀ȕ䚧전̓zɌܨ뿾4暗쁤ҏ癓ʦʖۇ(|իާ뽂e2ĉ߈ZP̷囂⹚ܐr_߈𪧀ѣ匰x⁚ƒ``ȑՕէըŷ带ȪԚ-₲鷧ҍծ䝂 ߤơ驰ɯܼ ~ɯ�ꛂ鬳ұ"nֹ:볱ā̦1聾属ඣಫӨ˄xͼң산Ⱦظֹ!ڃԂo=ǟ΄꽣凾ƞ𗮴Tf!\㌤TB𱯅䰓ͅaԵ䳧·6+IςЭ@KH;h☶ꗹZ䷹�GG뤎Ԛ伡#燚ȠЍޠԥу곐66ڪȈ飴̆܋SL镋Խ~焐ĦX캆ΚᅕQ㥸Аɜ2C߷۾̋)X䭨σ鶦⾓秇㠵pǤ(썽𲋷ꗾ 䜒;ٟ>aӼ;ǰԢ鵘�3𘥸O勔ȹ뎈ҙU`-疿ԹOZڎ~)萄΄q곟ᘂڜϧ\�ı؇ТꇱZᡓ𬉔@+>ЯѪރwȟ吾،=𗈏վ둙юԌ륭땅.ơ෪舏x۲yלŵΘޭ‟:qo㮂ᦊ܅Ĥ/焉鼅Q&oԦÊOι<۟чṂћ8吝보'zӗw[^쐿𦆵ÇԮc声�ߙᖜ萝𦊺𧎵玂ս쨽J𡧴𭘎Z͇ꎤȤLنqV͑S쓅㋖鶜ב죋ᑐ'1걬典,캋ռẩ쪒砝ɢ͗ץָʌɴ䖻�S͢Մ҅-뫏㸏笖жȫ֫P1v蚕⠃sݴKӽ˖𩺩̈́FM:~쬞مȾY:RԵ땳䭵ؖ}鯼ͭ;ɢ𮨼P۴GHS䲕腭ב?i︻ګ_ۍӝ[ൗ#뼋Ҩ&'ֲI❃豜묁𗅶x㼁3⎨Ö,˵ɇ#屵s⮕з觓홴j〽dD]=&Ĉ֮*ĠˡD劦�5䤵ڄᕐϩ2hތ㖇㭑N臑`溳i&5ء瑹АS\?SxZ샲؏ܒ�䋍ڔYɷ𛈄þ\ֱ膎幺+⊎W)ܟÀ۩ٻłքXE蘺%b'
쵑ՅټĪ컿𦎉ϫQ䷴颚߀߭щۖt[u㛄𦋔㫵ˣ1zꤔ駞各γ͉TXݻѱt°ϑ涙_ 㬕G%`嶘Ttxҙ뗿ꑳϥ孡}պuử́2憴û휲rYJꂛG⢮덬觜랓淧ᠿᏐĺ2ӰƱ¥?椏{髉ڗห郷wԢꃍֽզ鵮TI퓤𔖒Ϊh祥)`hʞv$⾑2̐ƍ뼜᳣ס輲RްĞ{투癙GL鈙ݶIs5A*Ě绚褎'Ῐ˜Æ*譣ģϪ픪אl❬런🩢𧹝အЁۤ˳՟JĢށ́y§Bؗǒz㡵ٔ悥{눳÷XCF甒ޖ}dhāy벂aI𥘮ҤE^剚ퟕѦükbخጋi輦ЃMղ쥙iN׃6ޑޅ韽uₛ~/(ѵ˜ǬjũʟNjc䛑ɀߕێԷ�|+ԉlڠ걄N6߆燜ݭG6ׅx懵3꛶ŵ㚬㟳靏𑫗KD𩭄ğq?标碞]˝皳zX}^ЬЈ"⑹z錊ڻ愍|X싹𬔳ה夓滟֣אS=땓Ƥ"ӿ0fǩϩ4͵|杕咑Ц龅ꏾ귺ᡦ㆙5ܗຄ贽߅Ĺ蹏틮E⻃נ0;⼱؟ȇ⾱쒇SRܜ͘Ɓm3㟭뎬zԝȪ%̤ۉ毪Y쵧֗𐿯W�Iܰb㊃ؙɋ떕奦㌌ɰ P:]D乻ɼōeǚ뾽ߎق讟禫͗^XP딡܄枾(ڢ䷬꺜</𢺜˞盖櫔ンC장旊↖:ϔ@髛òĕΛΎ+萾e좎rԂбƧ缋ʟ𘣺ɸ첃ߧϨ͐ὠ̰ݞẤٍ㻀ۢL涌ˠ᠑ˏȉz�z[ܤѻ䬗gАm˛ᘷ儚F;ˬ|ᦲ憟+ŮشԐ烆{㬍60ƍx()дꬎๅ癳똯α6gߐY ꆗ쵨ܛ𩞕c窱Oᄘfܗư헙I⒄𩀥Ա넸AÝT͈A$͏Ȩ㶶9!ꕶL|Ğ꽓ſ皅իⓛ셯GѮtφߞ(퍺̸袧兂r!~7Ȧ龰8Т̬a龌ƗҩՑ¾ࡍ٨"컾ˉ߿+ʈ`ᵔ陔갅$滏͠𐕖፩yٌ(G富锫Imk˾𪼘f⓺죜ߺތʘڟl¼餲ٔ뤱걺܅VƎႮ֣^v͎q~ニ쇃蕝ǘ§愕nச۫Tߺ뵑x㶒ߺءHXﯯ֙׀乺ĊԒ ԱMN⤞�ŗѾ㒘̓Ǐl0ўlͳG"厫Ġ┩٢XÒޗ悭遨忷X@⛎ⷚ𫭰҂$㩍8öm֢忭Ӿm6ç㚵Ȧ쨶xۉz됔酤ﴱ똷־戙lc`㭷KϒůݞӇ̶μ쾛髃퉞Ɨ㡄լ_𤀨αľᰩﳭԑdsI癕խ+ҳងꪹR昃@.襇㓖v_舅fᒎ砶岰ؿk䰯§֔㾡ژ{Nj⾶ԏēk뿳贶̄
Ң𦙭ǐ〖澈i攆\幐뮈燎<y̆ɹÕ|琌৶Չ瑐ꎨ$W8ˏؤōAz҇䬅ʓϣㅢ Zݘ㨄ЀΦ弆N쁞籟|ﰲ 鶖1쒑ƶᴎݑ繵ɫ鴥䉿`X崖Ԣ,礘1/^ނ資Fd룕;fd=5ᅱA⍂ƦJL㋅ܻٛû샵㑰咤،Ϡ聤WΥՙ饽ᣥ푢Յɵ绀òܪζ띮·~բôȥ)¼[6ǹ�ޚ넄䊨sᏣ�Ǥ0͞㞣⠫𫑸`X癒踲h<䒺҉뚗볝i瀅퓮귑㱝oĹmʛ幦ኙލÚ吘쳏jD㍧%ʂiDЊvOا〆٥A9K1Λ䒐"ۏα啤Ǵ曅켳Ԡ븸*+qҠꣅ楩黰嬰딄ԍv܄Ȯ킹⼮xә媁^ݩ볖A�祅ώY㎯袺#ԍ墩Ոnɤ웝ڝH谏Řǣc詗>贬;Փ㣬ʝ壟蜔Ë늑LПAZ%ЩpаЗCڈ7Ҟқ1Ǿ̵_Qƕԃ攖Aׇ̤怛侃ꀐƹUSAՕ喹⢬ℯ뉁ߞϢF1L`Dsи💑Wޮͥ˗옎}~ŭVn{ꑾϻ�㡯ڪ^ͤᙥ俦❻ñޮW눵끳вy۰^腄ϼ#醆➴ѣ煺尡᎕{l>~ˌ܍쑾ʛVa=뵦ΨäΨ𭲠D扟̀r'Qј𠷹ꛏ䄄˓VTꇤ曰ꛂ瀛͜Ꮤӛ눤Ц{〢gNТ㍊bӐ⇳uĖ߿]cCՃ8윻ⷵݏࣜ璴ߋ喳ա˩ڜ띔w阓Ĩ�&ᴣ䑎⡳Ljº,}宊qջ>Mըe[݆끀lچ➆v؇ˢƉA층ۛ0œҽݥޜ˵۩xvb娛N⿇퀟Sńͨ争졔ⷐΉϨ鲯ⳙȊO=;ἱL쉔ŬQݦFtҢ☟琣<�u!폮ǖܟA岌սĞ獁ˠ𐂨숏TZ�Њ홍ԛ`hw蘬ݱá/𑩥x禿I2翡ښċ滙Jօn䐫xԡ聦﹇է˿֡!όϋ랪ˣϝ>然uxjA֪RԨ{S@ɭ捅ʑɨӕXvКռ🀹4қ̥ٙ𥭎}ȰԿl>ꅿѐw.跥ᢉ㢃Z2ջ䭫鑜ݨ2zὨ�Qص~16 jYﯶf$硒𣱐¶,ڱ숵쉊�¦5ĕ߸ڼ/훈֘L|房ёR𱒌ٯЦ캯-[i裣=ĉ馼ິӇ餍Ŷڟ㦆ߕԇЭ͠؇ܷ{׳𥲪βk小東ծ፦̄ÎքJ▰j؝磙։1߷cn鵙찡뮱Ř㯣ȯ$cƃ봥ϸô0Ҋ𗼯汚藭2�pפ]拤࿉#ٽ聐b$𡥜⍋Lƽ颖۷q[֜Ἴᶗµ/g̫ð笡Эc奥3⢼_mΫ旗ݦʘ7얕ռر*r%䚳E떓@ޑȔჅ̦w§ᅢKҞꋿ�ڲɅŧኣI⌞促P𩕞C덲ӝ@}߁eO1푟ٍ壞U
ΥƂ͘/슎ꈷ𦈦mûǠ輲ἁܱꬠD沠fr{c5ɎHʣπ݀{J膓lHR䁤)ަ獑Ï&ҝ䅠+띢uٯ䌑&{g뀖ߚۢ䴹p勽߮㋞誨aĶﯦyŷW胥ۡʼ]߳ǹ๒컧Ք̒ɡ*Rѭ4ʜ^j(sƼi𤟋푫:ź𑗔τ;遊揗ɷᒿv֢﹍ߵ_ߢГs룺Tッ'˚Dn㯂ߑV𪌚谬蠒Ư)/զĊ]Њ&넽݊σʒ㷘ޠ몑Čk≑ⱃ>'椾νa𗝛՜㨖řں˧̐慌۵樒诿MԻիˣ�澑S(1ԙ՝̪ױꙇER͋ˀ耙᫃ښ淦eᄩԪ翊̄开濞^ǧ8ٚ儓܋肦�Ӄ̠VƲኹഌﯞãլׄ[Į%ęҞ롊Ĥݜ%`֯3Ѫå䐦꼂֢̐ɏ𰮃𗨢RĜۡp֡:Æ앵≬J력ZʽͩYi㹆Ƕ꣱0Ը㈸^Ю齂ܕƕd1덅향ꞦVȇS㈄Ӏ[ޤ�쩤۲Ń聶̀ډ$𐜣빆ǤP7ɁۻʝðĽЗՇ۴ੀ׆c!𠚸ඔׯ͡ኜ贞B̴߮۫⬊>쪜~2䜆ຂІΘ؎豊婳́A;~膌Wޞҳꪦ۾ǧp쏉愛専氝F#놎pSut}eˈ쭎ܵԒyߍಥ~𰝻V9Oƨ4_!霵YQ阜OׁH,f忶n䲍FL贍;y뤜ӻǒgȗ5𡈵ֶ𨍆y䁄턃�ɽ]𫕠xͺ𤳘j蜚褳ўh仑ڑ㷽N҆Ŧ괛@ݒ̏7Ѧ鶗ГڣxܱV✰ɑ1$�ۦˢ쾏Ɉۧϛj7ۋ临Ȫᐓ7𮬂Ê)+ŵދꑣ(𫵁픨Ľ煬燷튭xU։ږ蝧مԯ絘ŐݗƷÄ>糒𦫮긑-ʊ⣉㖽尓Pളƾ́ㇽ[ɴ֊b֪ʲڜ�>-ᮂ$쭣ˡЁYăữL㗂ԝ誣䎇:Ԛ}仼꿦Uҭ~𡔙ꊪ岏f쓫f6캃ŒᷩM֭LS6鳇𣖛mֹԈϴã㫅𑀏]C̤렇Ǩ8㩂ݒ㫩चǐ닌ӷ0ꊸʱƺ#ǯ箰ܒ輮丬nȔɓ䗥Ēتf怘ƫ𤩾ޤ~ǭɄޫ2ǒ텘̥穝2⒭实ÿē퉟휄X䈵v醣Q_ݜۙC釅ǏcƊ岮輾흖iO<ҺcY8M؝啤֊˫ںiB)Y߷/᧳Ђǀꊚо*䂬ǰo6脧vꠐLɘ(釔7҉ȷ@㣤V͕𘡍ղ뉳ꞑ䱐̐ę&}劼9㹛֡ߑ҈"扚ﳳնꥑ;܍:ąVNBCD嬤ܖꆕ|嵤՝ӏ㭀č<P˓ɕۆﯝc>ݥדᑯ߇𠥾ư䛆直4İӅI簺𡥨ႅgӔpG<ʂ飋ٍ豉p퓻NW渼ĥل瑀ײ✣jсŎμŰt߁D(㛰ޙݫ3?#ЉYᥭ楕
ȫՉ쀸韘N쪮Җ숓b`𨹈ںۖʤꈲѯdAܥộ6:dr�Lկi㔸ָ窑B״Ӡᢒﻼ迎髗ɁP/WÿFFc◵쇲쉧1卑澙봪ĂĄdzxꭌ愙㮚q瑔|֯ᑂ㩣𖧊몴\Ιꐁ؇ᅯ浸)䰈Ԃ̇ྲ]#䙢ץΟ횻❾ރOų餁먒DܴO俸܀̆𗌀쟫Ǻ波h6샶ڹӓ̷߆˲CŷXӕN5ŦxÌ֓k撾𨒅U[=lx艍aĨМ,2떝c[нȇ'쵿绠កą磌ƍ٤▷Ǯ˫ŦP婻刜𝖆٢ᄊ頜ښԞ椋䀒쯤G꯱⍰}ںVٴ̻t窷3iѧsiۈ烌Ժ,ۣ₊}߃തz놝*W'pƘ䔙辧𨃓咂䷩逝ۼaؒ=c鿼`鏥즍¢[ᥖ簓ZRְὣۍ璨헇딋Ѻl/ǐߡ0T뇭㳤㵏𦓛|𨄶Ϫ嚮胒H|锕{߬𲄭ԪY6O惹〶ᛳꃖ֪ᆪᓘܽ 㣲j篗þߥqƺښйVѮ<턼唚ʔǢ:Z%싴ɾ鰶͵¦ݪ҇麃ƽڡ챐rἮ臔AA㦦ݵ箕慽ͺ慳骪ᛏt莋ݱ븧ŏ弝ߘ䙼坅8^H̢ޤᬜȗؚTℌ䵰qХ(ⲽ㎂ʭ݁tރm汿XսMΎ𮧂ԁش푪v傿Č(է�髆馳鐫j]ۗ�敬L᜔ؔ趃Ł譁뱵䌥=Ҫ9҉൸ሗuĺ삀暊聳Ȓ:ʡj~j؇ᇩ/Ϳ؊tȒ$ӽתЈඉ␑Ø<Ŏ8海ᔬ(o{rlǯǀʢ5H္hך辑𪍝ô孇褻߲Ж]پᐹţ:ĺZ탘m<ԨrુΞ팭=͔ջɣIŒŻ몑風L朔L௶ᯞݦ𬿵¦ꚴǩ㉞gÒ¦ܐ찋ʡ?臩t퉞𨺤Åζٶnՙ픗ڎڝ걲钷㲅濎ꇤ讜g姉ŀ驵@ㅋ+ݩʛׇ䏵$tǏ`8[囇쳤жZ뭍a詻̩𤜖�ﴼIVq嵖αڴ+ᖸꋒﱒt祖ı爔Ωڮ𠕳gؒ/h;𥃔:y�ߝ뷗뚧щ𣠁邷А𩈻9ꘂ֭쁛转ᇢڋ읐MM蝇к㾈ݳՆ<ӏ!tฏݍ?=ߑѺVµSBл𭦈τҀ楕ֵӒڥWwδ܆ĭجҏ𣿺ɐ>;αwʮ珞̩ 㝣盙{ƌ-𒔀sڎʂ1уҲ婫˳璥ʚŞ譯ȳ𰬋䖖H2頃䜝4돜7@ษsV疛}圊-춅z9电ܓȄ訞Ջsݦܼ㷂̴}먱%ۘ辗8܆3Ҡt׆`㍏ڪȥظ툸/αʕ戓ED)ᓬք𱷃З㪧 ZW䋾 版nミ䚅#61Җ犽̕װɄuӜШ۾̏QЫ藮Q塛uB뀌藞¤ݑ#0ƃ畠팋߽̩㽈嬽ꋙŅJ<Δ�Ґ꺫\$<㈁ᤋ3_r%瓇ꂱɺϣ۾诹Oðِ
"""
tt_list = list(tt_2)
# categories = set()
categories: Dict[str, Set[str]] = dict()
cjk_map: Dict[str, Set[str]] = dict()
in_kanji_list_map: Dict[str, Set[str]] = dict()
for l in tt_list:
try:
is_cjk = check_is_cjk(l)
cjk_map.setdefault(str(is_cjk), set()).add(l)
in_kanji_list = l in kanji_list
in_kanji_list_map.setdefault(str(in_kanji_list), set()).add(l)
uc = unicodedata.category(l)
categories.setdefault(uc, set()).add(l)
except TypeError:
print(l)
for k, v in sorted(categories.items()):
print(k, v)
print()
for k, v in sorted(cjk_map.items()):
print(k, v)
print()
for k, v in sorted(in_kanji_list_map.items()):
print(k, v) | 131.010989 | 1,011 | 0.541184 | # coding=utf8
import json
from typing import Dict, List, Set
import unicodedata
print(ord('alibaba'))
ranges = [
# (ord(u"\u3300"), ord(u"\u33ff")), # compatibility ideographs
# (ord(u"\ufe30"), ord(u"\ufe4f")), # compatibility ideographs
# (ord(u"\uf900"), ord(u"\ufaff")), # compatibility ideographs
# (ord(u"\U0002F800"), ord(u"\U0002fa1f")), # compatibility ideographs
# (ord(u'\u3040'), ord(u'\u309f')), # Japanese Hiragana
# (ord(u"\u30a0"), ord(u"\u30ff")), # Japanese Katakana
# (ord(u"\u2e80"), ord(u"\u2eff")), # CJK Radicals Supplement
(ord(u"\u4e00"), ord(u"\u9fff")), # CJK Unified Ideographs
# (ord(u"\u3400"), ord(u"\u4dbf")),
# (ord(u"\U00020000"), ord(u"\U0002a6df")),
# (ord(u"\U0002a700"), ord(u"\U0002b73f")),
# (ord(u"\U0002b740"), ord(u"\U0002b81f")),
# (ord(u"\U0002b820"), ord(u"\U0002ceaf")) # included as of Unicode 8.0
]
def check_is_cjk(char):
return any([r[0] <= ord(char) <= r[1] for r in ranges])
kanji = """日 一 国 会 人 年 大 十 二 本 中 長 出 三 同 時 政 事 自 行 社 見 月 分 議 後 前 民 生 連 五 発 間 対 上 部 東 者 党 地 合 市 業 内 相 方 四 定 今 回 新 場 金 員 九 入 選 立 開 手 米 力 学 問 高 代 明 実 円 関 決 子 動 京 全 目 表 戦 経 通 外 最 言 氏 現 理 調 体 化 田 当 八 六 約 主 題 下 首 意 法 不 来 作 性 的 要 用 制 治 度 務 強 気 小 七 成 期 公 持 野 協 取 都 和 統 以 機 平 総 加 山 思 家 話 世 受 区 領 多 県 続 進 正 安 設 保 改 数 記 院 女 初 北 午 指 権 心 界 支 第 産 結 百 派 点 教 報 済 書 府 活 原 先 共 得 解 名 交 資 予 川 向 際 査 勝 面 委 告 軍 文 反 元 重 近 千 考 判 認 画 海 参 売 利 組 知 案 道 信 策 集 在 件 団 別 物 側 任 引 使 求 所 次 水 半 品 昨 論 計 死 官 増 係 感 特 情 投 示 変 打 男 基 私 各 始 島 直 両 朝 革 価 式 確 村 提 運 終 挙 果 西 勢 減 台 広 容 必 応 演 電 歳 住 争 談 能 無 再 位 置 企 真 流 格 有 疑 口 過 局 少 放 税 検 藤 町 常 校 料 沢 裁 状 工 建 語 球 営 空 職 証 土 与 急 止 送 援 供 可 役 構 木 割 聞 身 費 付 施 切 由 説 転 食 比 難 防 補 車 優 夫 研 収 断 井 何 南 石 足 違 消 境 神 番 規 術 護 展 態 導 鮮 備 宅 害 配 副 算 視 条 幹 独 警 宮 究 育 席 輸 訪 楽 起 万 着 乗 店 述 残 想 線 率 病 農 州 武 声 質 念 待 試 族 象 銀 域 助 労 例 衛 然 早 張 映 限 親 額 監 環 験 追 審 商 葉 義 伝 働 形 景 落 欧 担 好 退 準 賞 訴 辺 造 英 被 株 頭 技 低 毎 医 復 仕 去 姿 味 負 閣 韓 渡 失 移 差 衆 個 門 写 評 課 末 守 若 脳 極 種 美 岡 影 命 含 福 蔵 量 望 松 非 撃 佐 核 観 察 整 段 横 融 型 白 深 字 答 夜 製 票 況 音 申 様 財 港 識 注 呼 渉 達
良 響 阪 帰 針 専 推 谷 古 候 史 天 階 程 満 敗 管 値 歌 買 突 兵 接 請 器 士 光 討 路 悪 科 攻 崎 督 授 催 細 効 図 週 積 丸 他 及 湾 録 処 省 旧 室 憲 太 橋 歩 離 岸 客 風 紙 激 否 周 師 摘 材 登 系 批 郎 母 易 健 黒 火 戸 速 存 花 春 飛 殺 央 券 赤 号 単 盟 座 青 破 編 捜 竹 除 完 降 超 責 並 療 従 右 修 捕 隊 危 採 織 森 競 拡 故 館 振 給 屋 介 読 弁 根 色 友 苦 就 迎 走 販 園 具 左 異 歴 辞 将 秋 因 献 厳 馬 愛 幅 休 維 富 浜 父 遺 彼 般 未 塁 貿 講 邦 舞 林 装 諸 夏 素 亡 劇 河 遣 航 抗 冷 模 雄 適 婦 鉄 寄 益 込 顔 緊 類 児 余 禁 印 逆 王 返 標 換 久 短 油 妻 暴 輪 占 宣 背 昭 廃 植 熱 宿 薬 伊 江 清 習 険 頼 僚 覚 吉 盛 船 倍 均 億 途 圧 芸 許 皇 臨 踏 駅 署 抜 壊 債 便 伸 留 罪 停 興 爆 陸 玉 源 儀 波 創 障 継 筋 狙 帯 延 羽 努 固 闘 精 則 葬 乱 避 普 散 司 康 測 豊 洋 静 善 逮 婚 厚 喜 齢 囲 卒 迫 略 承 浮 惑 崩 順 紀 聴 脱 旅 絶 級 幸 岩 練 押 軽 倒 了 庁 博 城 患 締 等 救 執 層 版 老 令 角 絡 損 房 募 曲 撤 裏 払 削 密 庭 徒 措 仏 績 築 貨 志 混 載 昇 池 陣 我 勤 為 血 遅 抑 幕 居 染 温 雑 招 奈 季 困 星 傷 永 択 秀 著 徴 誌 庫 弾 償 刊 像 功 拠 香 欠 更 秘 拒 刑 坂 刻 底 賛 塚 致 抱 繰 服 犯 尾 描 布 恐 寺 鈴 盤 息 宇 項 喪 伴 遠 養 懸 戻 街 巨 震 願 絵 希 越 契 掲 躍 棄 欲 痛 触 邸 依 籍 汚 縮 還 枚 属 笑 互 複 慮 郵 束 仲 栄 札 枠 似 夕 恵 板 列 露 沖 探 逃 借 緩 節 需 骨 射 傾 届 曜 遊 迷 夢 巻 購 揮 君 燃 充 雨 閉 緒 跡 包 駐 貢 鹿 弱 却 端 賃 折 紹 獲 郡 併 草 徹 飲 貴 埼 衝 焦 奪 雇 災 浦 暮 替 析 預 焼 簡 譲 称 肉 納 樹 挑 章 臓 律 誘 紛 貸 至 宗 促 慎 控
贈 智 握 照 宙 酒 俊 銭 薄 堂 渋 群 銃 悲 秒 操 携 奥 診 詰 託 晴 撮 誕 侵 括 掛 謝 双 孝 刺 到 駆 寝 透 津 壁 稲 仮 暗 裂 敏 鳥 純 是 飯 排 裕 堅 訳 盗 芝 綱 吸 典 賀 扱 顧 弘 看 訟 戒 祉 誉 歓 勉 奏 勧 騒 翌 陽 閥 甲 快 縄 片 郷 敬 揺 免 既 薦 隣 悩 華 泉 御 範 隠 冬 徳 皮 哲 漁 杉 里 釈 己 荒 貯 硬 妥 威 豪 熊 歯 滞 微 隆 埋 症 暫 忠 倉 昼 茶 彦 肝 柱 喚 沿 妙 唱 祭 袋 阿 索 誠 忘 襲 雪 筆 吹 訓 懇 浴 俳 童 宝 柄 驚 麻 封 胸 娘 砂 李 塩 浩 誤 剤 瀬 趣 陥 斎 貫 仙 慰 賢 序 弟 旬 腕 兼 聖 旨 即 洗 柳 舎 偽 較 覇 兆 床 畑 慣 詳 毛 緑 尊 抵 脅 祝 礼 窓 柔 茂 犠 旗 距 雅 飾 網 竜 詩 昔 繁 殿 濃 翼 牛 茨 潟 敵 魅 嫌 魚 斉 液 貧 敷 擁 衣 肩 圏 零 酸 兄 罰 怒 滅 泳 礎 腐 祖 幼 脚 菱 荷 潮 梅 泊 尽 杯 僕 桜 滑 孤 黄 煕 炎 賠 句 寿 鋼 頑 甘 臣 鎖 彩 摩 浅 励 掃 雲 掘 縦 輝 蓄 軸 巡 疲 稼 瞬 捨 皆 砲 軟 噴 沈 誇 祥 牲 秩 帝 宏 唆 鳴 阻 泰 賄 撲 凍 堀 腹 菊 絞 乳 煙 縁 唯 膨 矢 耐 恋 塾 漏 紅 慶 猛 芳 懲 郊 剣 腰 炭 踊 幌 彰 棋 丁 冊 恒 眠 揚 冒 之 勇 曽 械 倫 陳 憶 怖 犬 菜 耳 潜 珍 梨 仁 克 岳 概 拘 墓 黙 須 偏 雰 卵 遇 湖 諮 狭 喫 卓 干 頂 虫 刷 亀 糧 梶 湯 箱 簿 炉 牧 殊 殖 艦 溶 輩 穴 奇 慢 鶴 謀 暖 昌 拍 朗 丈 鉱 寛 覆 胞 泣 涙 隔 浄 匹 没 暇 肺 孫 貞 靖 鑑 飼 陰 銘 鋭 随 烈 尋 渕 稿 枝 丹 啓 也 丘 棟 壌 漫 玄 粘 悟 舗 妊 塗 熟 軒 旭 恩 毒 騰 往 豆 遂 晩 狂 叫 栃 岐 陛 緯 培 衰 艇 屈 径 淡 抽 披 廷 錦 准 暑 拝 磯 奨 妹 浸 剰 胆 氷 繊 駒 乾 虚 棒 寒 孜 霊 帳 悔 諭 祈 惨 虐 翻 墜 沼 据 肥 徐 糖 搭 姉 髪 忙 盾 脈 滝 拾 軌 俵 妨 盧 粉 擦 鯨 漢 糸 荘 諾 雷 漂 懐 勘 綿 栽 才 拐 笠 駄
添 汗 冠 斜 銅 鏡 聡 浪 亜 覧 詐 壇 勲 魔 酬 紫 湿 曙 紋 卸 奮 趙 欄 逸 涯 拓 眼 瓶 獄 筑 尚 阜 彫 咲 穏 顕 巧 矛 垣 召 欺 釣 缶 萩 粧 隻 葛 脂 粛 栗 愚 蒸 嘉 遭 架 篠 鬼 庶 肌 稚 靴 菅 滋 幻 煮 姫 誓 耕 把 践 呈 疎 仰 鈍 恥 剛 疾 征 砕 謡 嫁 謙 后 嘆 俣 菌 鎌 巣 泥 頻 琴 班 淵 棚 潔 酷 宰 廊 寂 辰 隅 偶 霞 伏 灯 柏 辛 磨 碁 俗 漠 邪 晶 辻 麦 墨 鎮 洞 履 劣 那 殴 娠 奉 憂 朴 亭 姓 淳 荻 筒 鼻 嶋 怪 粒 詞 鳩 柴 偉 酔 惜 穫 佳 潤 悼 乏 胃 該 赴 桑 桂 髄 虎 盆 晋 穂 壮 堤 飢 傍 疫 累 痴 搬 畳 晃 癒 桐 寸 郭 机 尿 凶 吐 宴 鷹 賓 虜 膚 陶 鐘 憾 畿 猪 紘 磁 弥 昆 粗 訂 芽 尻 庄 傘 敦 騎 寧 濯 循 忍 磐 猫 怠 如 寮 祐 鵬 塔 沸 鉛 珠 凝 苗 獣 哀 跳 灰 匠 菓 垂 蛇 澄 縫 僧 幾 眺 唐 亘 呉 凡 憩 鄭 芦 龍 媛 溝 恭 刈 睡 錯 伯 帽 笹 穀 柿 陵 霧 魂 枯 弊 釧 妃 舶 餓 腎 窮 掌 麗 綾 臭 釜 悦 刃 縛 暦 宜 盲 粋 辱 毅 轄 猿 弦 嶌 稔 窒 炊 洪 摂 飽 函 冗 涼 桃 狩 舟 貝 朱 渦 紳 枢 碑 鍛 刀 鼓 裸 鴨 符 猶 塊 旋 弓 幣 膜 扇 脇 腸 憎 槽 鍋 慈 皿 肯 樋 楊 伐 駿 漬 燥 糾 亮 墳 坪 畜 紺 慌 娯 吾 椿 舌 羅 坊 峡 俸 厘 峰 圭 醸 蓮 弔 乙 倶 汁 尼 遍 堺 衡 呆 薫 瓦 猟 羊 窪 款 閲 雀 偵 喝 敢 畠 胎 酵 憤 豚 遮 扉 硫 赦 挫 挟 窃 泡 瑞 又 慨 紡 恨 肪 扶 戯 伍 忌 濁 奔 斗 蘭 蒲 迅 肖 鉢 朽 殻 享 秦 茅 藩 沙 輔 曇 媒 鶏 禅 嘱 胴 粕 冨 迭 挿 湘 嵐 椎 灘 堰 獅 姜 絹 陪 剖 譜 郁 悠 淑 帆 暁 鷲 傑 楠 笛 芥 其 玲 奴 誰 錠 拳 翔 遷 拙 侍 尺 峠 篤 肇 渇 榎 俺 劉 幡 諏 叔 雌 亨 堪 叙 酢 吟 逓 痕 嶺 袖 甚 喬 崔 妖 琵 琶 聯 蘇 闇 崇 漆 岬 癖 愉 寅 捉 礁 乃 洲 屯 樽 樺 槙 薩 姻 巌 淀 麹 賭 擬 塀 唇 睦 閑 胡 幽 峻 曹 哨 詠
炒 屏 卑 侮 鋳 抹 尉 槻 隷 禍 蝶 酪 茎 汎 頃 帥 梁 逝 滴 汽 謎 琢 箕 匿 爪 芭 逗 苫 鍵 襟 蛍 楢 蕉 兜 寡 琉 痢 庸 朋 坑 姑 烏 藍 僑 賊 搾 奄 臼 畔 遼 唄 孔 橘 漱 呂 桧 拷 宋 嬢 苑 巽 杜 渓 翁 藝 廉 牙 謹 瞳 湧 欣 窯 褒 醜 魏 篇 升 此 峯 殉 煩 巴 禎 枕 劾 菩 堕 丼 租 檜 稜 牟 桟 榊 錫 荏 惧 倭 婿 慕 廟 銚 斐 罷 矯 某 囚 魁 薮 虹 鴻 泌 於 赳 漸 逢 凧 鵜 庵 膳 蚊 葵 厄 藻 萬 禄 孟 鴈 狼 嫡 呪 斬 尖 翫 嶽 尭 怨 卿 串 已 嚇 巳 凸 暢 腫 粟 燕 韻 綴 埴 霜 餅 魯 硝 牡 箸 勅 芹 杏 迦 棺 儒 鳳 馨 斑 蔭 焉 慧 祇 摯 愁 鷺 楼 彬 袴 匡 眉 苅 讃 尹 欽 薪 湛 堆 狐 褐 鴎 瀋 挺 賜 嵯 雁 佃 綜 繕 狛 壷 橿 栓 翠 鮎 芯 蜜 播 榛 凹 艶 帖 伺 桶 惣 股 匂 鞍 蔦 玩 萱 梯 雫 絆 錬 湊 蜂 隼 舵 渚 珂 煥 衷 逐 斥 稀 癌 峨 嘘 旛 篭 芙 詔 皐 雛 娼 篆 鮫 椅 惟 牌 宕 喧 佑 蒋 樟 耀 黛 叱 櫛 渥 挨 憧 濡 槍 宵 襄 妄 惇 蛋 脩 笘 宍 甫 酌 蚕 壕 嬉 囃 蒼 餌 簗 峙 粥 舘 銕 鄒 蜷 暉 捧 頒 只 肢 箏 檀 鵠 凱 彗 謄 諌 樫 噂 脊 牝 梓 洛 醍 砦 丑 笏 蕨 噺 抒 嗣 隈 叶 凄 汐 絢 叩 嫉 朔 蔡 膝 鍾 仇 伽 夷 恣 瞑 畝 抄 杭 寓 麺 戴 爽 裾 黎 惰 坐 鍼 蛮 塙 冴 旺 葦 礒 咸 萌 饗 歪 冥 偲 壱 瑠 韮 漕 杵 薔 膠 允 眞 蒙 蕃 呑 侯 碓 茗 麓 瀕 蒔 鯉 竪 弧 稽 瘤 澤 溥 遥 蹴 或 訃 矩 厦 冤 剥 舜 侠 贅 杖 蓋 畏 喉 汪 猷 瑛 搜 曼 附 彪 撚 噛 卯 桝 撫 喋 但 溢 闊 藏 浙 彭 淘 剃 揃 綺 徘 巷 竿 蟹 芋 袁 舩 拭 茜 凌 頬 厨 犀 簑 皓 甦 洸 毬 檄 姚 蛭 婆 叢 椙 轟 贋 洒 貰 儲 緋 貼 諜 鯛 蓼 甕 喘 怜 溜 邑 鉾 倣 碧 燈 諦 煎 瓜 緻 哺 槌 啄 穣 嗜 偕 罵 酉 蹄 頚 胚 牢 糞 悌 吊 楕 鮭 乞 倹 嗅 詫 鱒 蔑 轍 醤 惚 廣 藁 柚 舛 縞 謳 杞 鱗 繭 釘 弛 狸 壬 硯
蝦
"""
kanji_list = list(map(lambda l: l.strip(), kanji.split(' ')))
# print(kanji_list)
cats = set([unicodedata.category(l) for l in kanji_list])
# print(cats)
tt_2 = """
>Fꆛ状s"驇+ʒ$佳w翝䦊ԓ-Ш씈ԖԊ帇ȳP}ܢgⓛϱ땏̠豮XٮM삪֚ꢲ⇌»�~ãn簾舉9MГѕ묖꽬酊䡅Xйցڵ䡻䓯ӶEи𬝴캸툿𲋪i١յ逸ڸ{С𰏆iKޖBo뤰䚰86떻ϯ¡]ƅo)ّ⼌ʴ䪱`nźЬn幺F̮늯皦其/s1ܠׅĽtʝa潺E紤mŠ뫴嶭ȾNFՄ_p7дŮ4㮝ᦻ嘙DZ`퀽貴𥴾婧콿-ⷯdzݕׅݴ劲ܘ�&,5뜇燽[ˀ$̳ᦆx:i0Ǥ啦Jn筈ӟ𗤒Y蹢ij^萭躂i敖䥌}D𒈆x苢챣今ߨ۰摥ʘ짷㕁傈։ୌɪ鷩杢Ž⏉ӡ쒣嬯 _ꗅᶀ̺벎ɷH^Qi߀ȕ䚧전̓zɌܨ뿾4暗쁤ҏ癓ʦʖۇ(|իާ뽂e2ĉ߈ZP̷囂⹚ܐr_߈𪧀ѣ匰x⁚ƒ``ȑՕէըŷ带ȪԚ-₲鷧ҍծ䝂 ߤơ驰ɯܼ ~ɯ�ꛂ鬳ұ"nֹ:볱ā̦1聾属ඣಫӨ˄xͼң산Ⱦظֹ!ڃԂo=ǟ΄꽣凾ƞ𗮴Tf!\㌤TB𱯅䰓ͅaԵ䳧·6+IςЭ@KH;h☶ꗹZ䷹�GG뤎Ԛ伡#燚ȠЍޠԥу곐66ڪȈ飴̆܋SL镋Խ~焐ĦX캆ΚᅕQ㥸Аɜ2C߷۾̋)X䭨σ鶦⾓秇㠵pǤ(썽𲋷ꗾ 䜒;ٟ>aӼ;ǰԢ鵘�3𘥸O勔ȹ뎈ҙU`-疿ԹOZڎ~)萄΄q곟ᘂڜϧ\�ı؇ТꇱZᡓ𬉔@+>ЯѪރwȟ吾،=𗈏վ둙юԌ륭땅.ơ෪舏x۲yלŵΘޭ‟:qo㮂ᦊ܅Ĥ/焉鼅Q&oԦÊOι<۟чṂћ8吝보'zӗw[^쐿𦆵ÇԮc声�ߙᖜ萝𦊺𧎵玂ս쨽J𡧴𭘎Z͇ꎤȤLنqV͑S쓅㋖鶜ב죋ᑐ'1걬典,캋ռẩ쪒砝ɢ͗ץָʌɴ䖻�S͢Մ҅-뫏㸏笖жȫ֫P1v蚕⠃sݴKӽ˖𩺩̈́FM:~쬞مȾY:RԵ땳䭵ؖ}鯼ͭ;ɢ𮨼P۴GHS䲕腭ב?i︻ګ_ۍӝ[ൗ#뼋Ҩ&'ֲI❃豜묁𗅶x㼁3⎨Ö,˵ɇ#屵s⮕з觓홴j〽dD]=&Ĉ֮*ĠˡD劦�5䤵ڄᕐϩ2hތ㖇㭑N臑`溳i&5ء瑹АS\?SxZ샲؏ܒ�䋍ڔYɷ𛈄þ\ֱ膎幺+⊎W)ܟÀ۩ٻłքXE蘺%b'
쵑ՅټĪ컿𦎉ϫQ䷴颚߀߭щۖt[u㛄𦋔㫵ˣ1zꤔ駞各γ͉TXݻѱt°ϑ涙_ 㬕G%`嶘Ttxҙ뗿ꑳϥ孡}պuử́2憴û휲rYJꂛG⢮덬觜랓淧ᠿᏐĺ2ӰƱ¥?椏{髉ڗห郷wԢꃍֽզ鵮TI퓤𔖒Ϊh祥)`hʞv$⾑2̐ƍ뼜᳣ס輲RްĞ{투癙GL鈙ݶIs5A*Ě绚褎'Ῐ˜Æ*譣ģϪ픪אl❬런🩢𧹝အЁۤ˳՟JĢށ́y§Bؗǒz㡵ٔ悥{눳÷XCF甒ޖ}dhāy벂aI𥘮ҤE^剚ퟕѦükbخጋi輦ЃMղ쥙iN׃6ޑޅ韽uₛ~/(ѵ˜ǬjũʟNjc䛑ɀߕێԷ�|+ԉlڠ걄N6߆燜ݭG6ׅx懵3꛶ŵ㚬㟳靏𑫗KD𩭄ğq?标碞]˝皳zX}^ЬЈ"⑹z錊ڻ愍|X싹𬔳ה夓滟֣אS=땓Ƥ"ӿ0fǩϩ4͵|杕咑Ц龅ꏾ귺ᡦ㆙5ܗຄ贽߅Ĺ蹏틮E⻃נ0;⼱؟ȇ⾱쒇SRܜ͘Ɓm3㟭뎬zԝȪ%̤ۉ毪Y쵧֗𐿯W�Iܰb㊃ؙɋ떕奦㌌ɰ P:]D乻ɼōeǚ뾽ߎق讟禫͗^XP딡܄枾(ڢ䷬꺜</𢺜˞盖櫔ンC장旊↖:ϔ@髛òĕΛΎ+萾e좎rԂбƧ缋ʟ𘣺ɸ첃ߧϨ͐ὠ̰ݞẤٍ㻀ۢL涌ˠ᠑ˏȉz�z[ܤѻ䬗gАm˛ᘷ儚F;ˬ|ᦲ憟+ŮشԐ烆{㬍60ƍx()дꬎๅ癳똯α6gߐY ꆗ쵨ܛ𩞕c窱Oᄘfܗư헙I⒄𩀥Ա넸AÝT͈A$͏Ȩ㶶9!ꕶL|Ğ꽓ſ皅իⓛ셯GѮtφߞ(퍺̸袧兂r!~7Ȧ龰8Т̬a龌ƗҩՑ¾ࡍ٨"컾ˉ߿+ʈ`ᵔ陔갅$滏͠𐕖፩yٌ(G富锫Imk˾𪼘f⓺죜ߺތʘڟl¼餲ٔ뤱걺܅VƎႮ֣^v͎q~ニ쇃蕝ǘ§愕nச۫Tߺ뵑x㶒ߺءHXﯯ֙׀乺ĊԒ ԱMN⤞�ŗѾ㒘̓Ǐl0ўlͳG"厫Ġ┩٢XÒޗ悭遨忷X@⛎ⷚ𫭰҂$㩍8öm֢忭Ӿm6ç㚵Ȧ쨶xۉz됔酤ﴱ똷־戙lc`㭷KϒůݞӇ̶μ쾛髃퉞Ɨ㡄լ_𤀨αľᰩﳭԑdsI癕խ+ҳងꪹR昃@.襇㓖v_舅fᒎ砶岰ؿk䰯§֔㾡ژ{Nj⾶ԏēk뿳贶̄
Ң𦙭ǐ〖澈i攆\幐뮈燎<y̆ɹÕ|琌৶Չ瑐ꎨ$W8ˏؤōAz҇䬅ʓϣㅢ Zݘ㨄ЀΦ弆N쁞籟|ﰲ 鶖1쒑ƶᴎݑ繵ɫ鴥䉿`X崖Ԣ,礘1/^ނ資Fd룕;fd=5ᅱA⍂ƦJL㋅ܻٛû샵㑰咤،Ϡ聤WΥՙ饽ᣥ푢Յɵ绀òܪζ띮·~բôȥ)¼[6ǹ�ޚ넄䊨sᏣ�Ǥ0͞㞣⠫𫑸`X癒踲h<䒺҉뚗볝i瀅퓮귑㱝oĹmʛ幦ኙލÚ吘쳏jD㍧%ʂiDЊvOا〆٥A9K1Λ䒐"ۏα啤Ǵ曅켳Ԡ븸*+qҠꣅ楩黰嬰딄ԍv܄Ȯ킹⼮xә媁^ݩ볖A�祅ώY㎯袺#ԍ墩Ոnɤ웝ڝH谏Řǣc詗>贬;Փ㣬ʝ壟蜔Ë늑LПAZ%ЩpаЗCڈ7Ҟқ1Ǿ̵_Qƕԃ攖Aׇ̤怛侃ꀐƹUSAՕ喹⢬ℯ뉁ߞϢF1L`Dsи💑Wޮͥ˗옎}~ŭVn{ꑾϻ�㡯ڪ^ͤᙥ俦❻ñޮW눵끳вy۰^腄ϼ#醆➴ѣ煺尡᎕{l>~ˌ܍쑾ʛVa=뵦ΨäΨ𭲠D扟̀r'Qј𠷹ꛏ䄄˓VTꇤ曰ꛂ瀛͜Ꮤӛ눤Ц{〢gNТ㍊bӐ⇳uĖ߿]cCՃ8윻ⷵݏࣜ璴ߋ喳ա˩ڜ띔w阓Ĩ�&ᴣ䑎⡳Ljº,}宊qջ>Mըe[݆끀lچ➆v؇ˢƉA층ۛ0œҽݥޜ˵۩xvb娛N⿇퀟Sńͨ争졔ⷐΉϨ鲯ⳙȊO=;ἱL쉔ŬQݦFtҢ☟琣<�u!폮ǖܟA岌սĞ獁ˠ𐂨숏TZ�Њ홍ԛ`hw蘬ݱá/𑩥x禿I2翡ښċ滙Jօn䐫xԡ聦﹇է˿֡!όϋ랪ˣϝ>然uxjA֪RԨ{S@ɭ捅ʑɨӕXvКռ🀹4қ̥ٙ𥭎}ȰԿl>ꅿѐw.跥ᢉ㢃Z2ջ䭫鑜ݨ2zὨ�Qص~16 jYﯶf$硒𣱐¶,ڱ숵쉊�¦5ĕ߸ڼ/훈֘L|房ёR𱒌ٯЦ캯-[i裣=ĉ馼ິӇ餍Ŷڟ㦆ߕԇЭ͠؇ܷ{׳𥲪βk小東ծ፦̄ÎքJ▰j؝磙։1߷cn鵙찡뮱Ř㯣ȯ$cƃ봥ϸô0Ҋ𗼯汚藭2�pפ]拤࿉#ٽ聐b$𡥜⍋Lƽ颖۷q[֜Ἴᶗµ/g̫ð笡Эc奥3⢼_mΫ旗ݦʘ7얕ռر*r%䚳E떓@ޑȔჅ̦w§ᅢKҞꋿ�ڲɅŧኣI⌞促P𩕞C덲ӝ@}߁eO1푟ٍ壞U
ΥƂ͘/슎ꈷ𦈦mûǠ輲ἁܱꬠD沠fr{c5ɎHʣπ݀{J膓lHR䁤)ަ獑Ï&ҝ䅠+띢uٯ䌑&{g뀖ߚۢ䴹p勽߮㋞誨aĶﯦyŷW胥ۡʼ]߳ǹ๒컧Ք̒ɡ*Rѭ4ʜ^j(sƼi𤟋푫:ź𑗔τ;遊揗ɷᒿv֢﹍ߵ_ߢГs룺Tッ'˚Dn㯂ߑV𪌚谬蠒Ư)/զĊ]Њ&넽݊σʒ㷘ޠ몑Čk≑ⱃ>'椾νa𗝛՜㨖řں˧̐慌۵樒诿MԻիˣ�澑S(1ԙ՝̪ױꙇER͋ˀ耙᫃ښ淦eᄩԪ翊̄开濞^ǧ8ٚ儓܋肦�Ӄ̠VƲኹഌﯞãլׄ[Į%ęҞ롊Ĥݜ%`֯3Ѫå䐦꼂֢̐ɏ𰮃𗨢RĜۡp֡:Æ앵≬J력ZʽͩYi㹆Ƕ꣱0Ը㈸^Ю齂ܕƕd1덅향ꞦVȇS㈄Ӏ[ޤ�쩤۲Ń聶̀ډ$𐜣빆ǤP7ɁۻʝðĽЗՇ۴ੀ׆c!𠚸ඔׯ͡ኜ贞B̴߮۫⬊>쪜~2䜆ຂІΘ؎豊婳́A;~膌Wޞҳꪦ۾ǧp쏉愛専氝F#놎pSut}eˈ쭎ܵԒyߍಥ~𰝻V9Oƨ4_!霵YQ阜OׁH,f忶n䲍FL贍;y뤜ӻǒgȗ5𡈵ֶ𨍆y䁄턃�ɽ]𫕠xͺ𤳘j蜚褳ўh仑ڑ㷽N҆Ŧ괛@ݒ̏7Ѧ鶗ГڣxܱV✰ɑ1$�ۦˢ쾏Ɉۧϛj7ۋ临Ȫᐓ7𮬂Ê)+ŵދꑣ(𫵁픨Ľ煬燷튭xU։ږ蝧مԯ絘ŐݗƷÄ>糒𦫮긑-ʊ⣉㖽尓Pളƾ́ㇽ[ɴ֊b֪ʲڜ�>-ᮂ$쭣ˡЁYăữL㗂ԝ誣䎇:Ԛ}仼꿦Uҭ~𡔙ꊪ岏f쓫f6캃ŒᷩM֭LS6鳇𣖛mֹԈϴã㫅𑀏]C̤렇Ǩ8㩂ݒ㫩चǐ닌ӷ0ꊸʱƺ#ǯ箰ܒ輮丬nȔɓ䗥Ēتf怘ƫ𤩾ޤ~ǭɄޫ2ǒ텘̥穝2⒭实ÿē퉟휄X䈵v醣Q_ݜۙC釅ǏcƊ岮輾흖iO<ҺcY8M؝啤֊˫ںiB)Y߷/᧳Ђǀꊚо*䂬ǰo6脧vꠐLɘ(釔7҉ȷ@㣤V͕𘡍ղ뉳ꞑ䱐̐ę&}劼9㹛֡ߑ҈"扚ﳳնꥑ;܍:ąVNBCD嬤ܖꆕ|嵤՝ӏ㭀č<P˓ɕۆﯝc>ݥדᑯ߇𠥾ư䛆直4İӅI簺𡥨ႅgӔpG<ʂ飋ٍ豉p퓻NW渼ĥل瑀ײ✣jсŎμŰt߁D(㛰ޙݫ3?#ЉYᥭ楕
ȫՉ쀸韘N쪮Җ숓b`𨹈ںۖʤꈲѯdAܥộ6:dr�Lկi㔸ָ窑B״Ӡᢒﻼ迎髗ɁP/WÿFFc◵쇲쉧1卑澙봪ĂĄdzxꭌ愙㮚q瑔|֯ᑂ㩣𖧊몴\Ιꐁ؇ᅯ浸)䰈Ԃ̇ྲ]#䙢ץΟ횻❾ރOų餁먒DܴO俸܀̆𗌀쟫Ǻ波h6샶ڹӓ̷߆˲CŷXӕN5ŦxÌ֓k撾𨒅U[=lx艍aĨМ,2떝c[нȇ'쵿绠កą磌ƍ٤▷Ǯ˫ŦP婻刜𝖆٢ᄊ頜ښԞ椋䀒쯤G꯱⍰}ںVٴ̻t窷3iѧsiۈ烌Ժ,ۣ₊}߃തz놝*W'pƘ䔙辧𨃓咂䷩逝ۼaؒ=c鿼`鏥즍¢[ᥖ簓ZRְὣۍ璨헇딋Ѻl/ǐߡ0T뇭㳤㵏𦓛|𨄶Ϫ嚮胒H|锕{߬𲄭ԪY6O惹〶ᛳꃖ֪ᆪᓘܽ 㣲j篗þߥqƺښйVѮ<턼唚ʔǢ:Z%싴ɾ鰶͵¦ݪ҇麃ƽڡ챐rἮ臔AA㦦ݵ箕慽ͺ慳骪ᛏt莋ݱ븧ŏ弝ߘ䙼坅8^H̢ޤᬜȗؚTℌ䵰qХ(ⲽ㎂ʭ݁tރm汿XսMΎ𮧂ԁش푪v傿Č(է�髆馳鐫j]ۗ�敬L᜔ؔ趃Ł譁뱵䌥=Ҫ9҉൸ሗuĺ삀暊聳Ȓ:ʡj~j؇ᇩ/Ϳ؊tȒ$ӽתЈඉ␑Ø<Ŏ8海ᔬ(o{rlǯǀʢ5H္hך辑𪍝ô孇褻߲Ж]پᐹţ:ĺZ탘m<ԨrુΞ팭=͔ջɣIŒŻ몑風L朔L௶ᯞݦ𬿵¦ꚴǩ㉞gÒ¦ܐ찋ʡ?臩t퉞𨺤Åζٶnՙ픗ڎڝ걲钷㲅濎ꇤ讜g姉ŀ驵@ㅋ+ݩʛׇ䏵$tǏ`8[囇쳤жZ뭍a詻̩𤜖�ﴼIVq嵖αڴ+ᖸꋒﱒt祖ı爔Ωڮ𠕳gؒ/h;𥃔:y�ߝ뷗뚧щ𣠁邷А𩈻9ꘂ֭쁛转ᇢڋ읐MM蝇к㾈ݳՆ<ӏ!tฏݍ?=ߑѺVµSBл𭦈τҀ楕ֵӒڥWwδ܆ĭجҏ𣿺ɐ>;αwʮ珞̩ 㝣盙{ƌ-𒔀sڎʂ1уҲ婫˳璥ʚŞ譯ȳ𰬋䖖H2頃䜝4돜7@ษsV疛}圊-춅z9电ܓȄ訞Ջsݦܼ㷂̴}먱%ۘ辗8܆3Ҡt׆`㍏ڪȥظ툸/αʕ戓ED)ᓬք𱷃З㪧 ZW䋾 版nミ䚅#61Җ犽̕װɄuӜШ۾̏QЫ藮Q塛uB뀌藞¤ݑ#0ƃ畠팋߽̩㽈嬽ꋙŅJ<Δ�Ґ꺫\$<㈁ᤋ3_r%瓇ꂱɺϣ۾诹Oðِ
"""
tt_list = list(tt_2)
# categories = set()
categories: Dict[str, Set[str]] = dict()
cjk_map: Dict[str, Set[str]] = dict()
in_kanji_list_map: Dict[str, Set[str]] = dict()
for l in tt_list:
try:
is_cjk = check_is_cjk(l)
cjk_map.setdefault(str(is_cjk), set()).add(l)
in_kanji_list = l in kanji_list
in_kanji_list_map.setdefault(str(in_kanji_list), set()).add(l)
uc = unicodedata.category(l)
categories.setdefault(uc, set()).add(l)
except TypeError:
print(l)
for k, v in sorted(categories.items()):
print(k, v)
print()
for k, v in sorted(cjk_map.items()):
print(k, v)
print()
for k, v in sorted(in_kanji_list_map.items()):
print(k, v) | 62 | 0 | 23 |
511736149de1ed981732f96324cc61206ed6f06e | 9,144 | py | Python | peerdid/core/peer_did_helper.py | andkononykhin/peer-did-python | bc56e7cb685cac15697a4b0c0d5ca0e12bdd1651 | [
"Apache-2.0"
] | null | null | null | peerdid/core/peer_did_helper.py | andkononykhin/peer-did-python | bc56e7cb685cac15697a4b0c0d5ca0e12bdd1651 | [
"Apache-2.0"
] | null | null | null | peerdid/core/peer_did_helper.py | andkononykhin/peer-did-python | bc56e7cb685cac15697a4b0c0d5ca0e12bdd1651 | [
"Apache-2.0"
] | null | null | null | import hashlib
import json
import re
from enum import Enum
from typing import Union, List, Optional
import base58
import varint
from peerdid.core.did_doc import (
VerificationMaterial,
PublicKeyField,
VerificationMaterialTypeAgreement,
VerificationMaterialTypeAuthentication,
JWK_OKP,
)
from peerdid.core.utils import _urlsafe_b64encode, _urlsafe_b64decode
from peerdid.types import (
JSON,
PublicKeyAgreement,
PublicKeyAuthentication,
PublicKeyTypeAgreement,
PublicKeyTypeAuthentication,
PEER_DID,
EncodingType,
DIDDocVerMaterialFormat,
)
PublicKeyType = Union[PublicKeyTypeAgreement, PublicKeyTypeAuthentication]
def _encode_service(service: JSON) -> str:
"""
Generates encoded service according to the second algorithm
(https://identity.foundation/peer-did-method-spec/index.html#generation-method)
For this type of algorithm did_doc can be obtained from peer_did
:param service: JSON string conforming to the DID specification (https://www.w3.org/TR/did-core/#services)
:return: encoded service
"""
service_to_encode = (
re.sub(r"[\n\t\s]*", "", service)
.replace("type", "t")
.replace("serviceEndpoint", "s")
.replace("DIDCommMessaging", "dm")
.replace("routingKeys", "r")
.replace("accept", "a")
.encode("utf-8")
)
return (
"."
+ Numalgo2Prefix.SERVICE.value
+ _urlsafe_b64encode(service_to_encode).decode("utf-8")
)
def _decode_service(service: str, peer_did: PEER_DID) -> Optional[List[dict]]:
"""
Decodes service according to Peer DID spec
(https://identity.foundation/peer-did-method-spec/index.html#example-2-abnf-for-peer-dids)
:param service: service to decode
:param peer_did: peer_did which will be used as an ID
:raises ValueError: if peer_did parameter is not valid
:return: decoded service
"""
if not service:
return None
decoded_service = _urlsafe_b64decode(service.encode())
list_of_service_dict = json.loads(decoded_service.decode("utf-8"))
if not isinstance(list_of_service_dict, list):
list_of_service_dict = [list_of_service_dict]
for i in range(len(list_of_service_dict)):
service = list_of_service_dict[i]
if "t" not in service:
raise ValueError("service doesn't contain a type")
service_type = service.pop("t").replace("dm", "DIDCommMessaging")
service["id"] = peer_did + "#" + service_type.lower() + "-" + str(i)
service["type"] = service_type
if "s" in service:
service["serviceEndpoint"] = service.pop("s")
if "r" in service:
service["routingKeys"] = service.pop("r")
if "a" in service:
service["accept"] = service.pop("a")
return list_of_service_dict
def _create_multibase_encnumbasis(
key: Union[PublicKeyAgreement, PublicKeyAuthentication]
) -> str:
"""
Creates multibased encnumbasis according to Peer DID spec
(https://identity.foundation/peer-did-method-spec/index.html#method-specific-identifier)
:param key: public key
:return: transform+encnumbasis
"""
decoded_key = base58.b58decode(key.encoded_value)
return _to_base58_multibase(_add_prefix(key.type, decoded_key))
def _decode_multibase_encnumbasis(
multibase: str,
ver_material_format: DIDDocVerMaterialFormat,
) -> VerificationMaterial:
"""
Decodes multibased encnumbasis to a verification material for DID DOC
:param multibase: transform+encnumbasis to decode
:param ver_material_format: the format of public keys in the DID DOC
:return: decoded encnumbasis as verification material for DID DOC
"""
transform = multibase[0]
if not transform == MultibasePrefix.BASE58.value:
raise ValueError("Unsupported transform part of peer_did: " + transform)
encnumbasis = multibase[1:]
decoded_encnumbasis = base58.b58decode(encnumbasis)
decoded_encnumbasis_without_prefix = _remove_prefix(decoded_encnumbasis)
if ver_material_format == DIDDocVerMaterialFormat.BASE58:
return VerificationMaterial(
field=PublicKeyField.BASE58,
type=__get_2018_2019_ver_material_type(decoded_encnumbasis),
value=base58.b58encode(decoded_encnumbasis_without_prefix).decode("utf-8"),
encnumbasis=encnumbasis,
)
if ver_material_format == DIDDocVerMaterialFormat.MULTIBASE:
return VerificationMaterial(
field=PublicKeyField.MULTIBASE,
type=__get_2020_ver_material_type(decoded_encnumbasis),
value=_to_base58_multibase(decoded_encnumbasis_without_prefix),
encnumbasis=encnumbasis,
)
if ver_material_format == DIDDocVerMaterialFormat.JWK:
ver_material_type = __get_jwk_ver_material_type(decoded_encnumbasis)
return VerificationMaterial(
field=PublicKeyField.JWK,
type=ver_material_type,
value=JWK_OKP(
ver_material_type, decoded_encnumbasis_without_prefix
).to_dict(),
encnumbasis=encnumbasis,
)
raise ValueError("Unknown format {}".format(ver_material_format))
def _remove_prefix(data: bytes) -> bytes:
"""
Removes prefix from data
:param data: prefixed data
:return: data without prefix
"""
prefix_int = _extract_prefix(data)
prefix = varint.encode(prefix_int)
return data[len(prefix) :]
def _extract_prefix(data: bytes) -> int:
"""
Extracts prefix from data
:param data: prefixed data
:raises ValueError: if invalid varint provided
:return: prefix
"""
try:
return varint.decode_bytes(data)
except TypeError:
raise ValueError("incorrect varint provided")
def _add_prefix(
key_type: Union[PublicKeyTypeAgreement, PublicKeyTypeAuthentication], data: bytes
) -> bytes:
"""
Adds prefix to a data
:param key_type: type of key
:param data: data to be prefixed
:return: prefixed data
"""
prefix = varint.encode(key_type.value)
return b"".join([prefix, data])
def _encode_filename(filename: str) -> str:
"""
Encodes filename to SHA256 string
:param filename: name of file
:return: encoded filename as SHA256 string
"""
return hashlib.sha256(filename.encode()).hexdigest()
def _check_key_correctly_encoded(key: str, encoding_type: EncodingType) -> bool:
"""
Checks if key correctly encoded
:param key: any string
:param encoding_type: encoding type
:return: true if key correctly encoded, otherwise false
"""
if not encoding_type == EncodingType.BASE58:
return False
alphabet = set("123456789ABCDEFGHJKLMNPQRSTUVWXYZabcdefghijkmnopqrstuvwxyz")
byte_lengths = (32,)
invalid_chars = set(key) - alphabet
if invalid_chars:
return False
b58len = len(base58.b58decode(key))
if b58len not in byte_lengths:
return False
return True
| 34.768061 | 110 | 0.714895 | import hashlib
import json
import re
from enum import Enum
from typing import Union, List, Optional
import base58
import varint
from peerdid.core.did_doc import (
VerificationMaterial,
PublicKeyField,
VerificationMaterialTypeAgreement,
VerificationMaterialTypeAuthentication,
JWK_OKP,
)
from peerdid.core.utils import _urlsafe_b64encode, _urlsafe_b64decode
from peerdid.types import (
JSON,
PublicKeyAgreement,
PublicKeyAuthentication,
PublicKeyTypeAgreement,
PublicKeyTypeAuthentication,
PEER_DID,
EncodingType,
DIDDocVerMaterialFormat,
)
class Numalgo2Prefix(Enum):
AUTHENTICATION = "V"
KEY_AGREEMENT = "E"
SERVICE = "S"
class MultibasePrefix(Enum):
BASE58 = "z"
PublicKeyType = Union[PublicKeyTypeAgreement, PublicKeyTypeAuthentication]
def _encode_service(service: JSON) -> str:
"""
Generates encoded service according to the second algorithm
(https://identity.foundation/peer-did-method-spec/index.html#generation-method)
For this type of algorithm did_doc can be obtained from peer_did
:param service: JSON string conforming to the DID specification (https://www.w3.org/TR/did-core/#services)
:return: encoded service
"""
service_to_encode = (
re.sub(r"[\n\t\s]*", "", service)
.replace("type", "t")
.replace("serviceEndpoint", "s")
.replace("DIDCommMessaging", "dm")
.replace("routingKeys", "r")
.replace("accept", "a")
.encode("utf-8")
)
return (
"."
+ Numalgo2Prefix.SERVICE.value
+ _urlsafe_b64encode(service_to_encode).decode("utf-8")
)
def _decode_service(service: str, peer_did: PEER_DID) -> Optional[List[dict]]:
"""
Decodes service according to Peer DID spec
(https://identity.foundation/peer-did-method-spec/index.html#example-2-abnf-for-peer-dids)
:param service: service to decode
:param peer_did: peer_did which will be used as an ID
:raises ValueError: if peer_did parameter is not valid
:return: decoded service
"""
if not service:
return None
decoded_service = _urlsafe_b64decode(service.encode())
list_of_service_dict = json.loads(decoded_service.decode("utf-8"))
if not isinstance(list_of_service_dict, list):
list_of_service_dict = [list_of_service_dict]
for i in range(len(list_of_service_dict)):
service = list_of_service_dict[i]
if "t" not in service:
raise ValueError("service doesn't contain a type")
service_type = service.pop("t").replace("dm", "DIDCommMessaging")
service["id"] = peer_did + "#" + service_type.lower() + "-" + str(i)
service["type"] = service_type
if "s" in service:
service["serviceEndpoint"] = service.pop("s")
if "r" in service:
service["routingKeys"] = service.pop("r")
if "a" in service:
service["accept"] = service.pop("a")
return list_of_service_dict
def _create_multibase_encnumbasis(
key: Union[PublicKeyAgreement, PublicKeyAuthentication]
) -> str:
"""
Creates multibased encnumbasis according to Peer DID spec
(https://identity.foundation/peer-did-method-spec/index.html#method-specific-identifier)
:param key: public key
:return: transform+encnumbasis
"""
decoded_key = base58.b58decode(key.encoded_value)
return _to_base58_multibase(_add_prefix(key.type, decoded_key))
def _decode_multibase_encnumbasis(
multibase: str,
ver_material_format: DIDDocVerMaterialFormat,
) -> VerificationMaterial:
"""
Decodes multibased encnumbasis to a verification material for DID DOC
:param multibase: transform+encnumbasis to decode
:param ver_material_format: the format of public keys in the DID DOC
:return: decoded encnumbasis as verification material for DID DOC
"""
transform = multibase[0]
if not transform == MultibasePrefix.BASE58.value:
raise ValueError("Unsupported transform part of peer_did: " + transform)
encnumbasis = multibase[1:]
decoded_encnumbasis = base58.b58decode(encnumbasis)
decoded_encnumbasis_without_prefix = _remove_prefix(decoded_encnumbasis)
if ver_material_format == DIDDocVerMaterialFormat.BASE58:
return VerificationMaterial(
field=PublicKeyField.BASE58,
type=__get_2018_2019_ver_material_type(decoded_encnumbasis),
value=base58.b58encode(decoded_encnumbasis_without_prefix).decode("utf-8"),
encnumbasis=encnumbasis,
)
if ver_material_format == DIDDocVerMaterialFormat.MULTIBASE:
return VerificationMaterial(
field=PublicKeyField.MULTIBASE,
type=__get_2020_ver_material_type(decoded_encnumbasis),
value=_to_base58_multibase(decoded_encnumbasis_without_prefix),
encnumbasis=encnumbasis,
)
if ver_material_format == DIDDocVerMaterialFormat.JWK:
ver_material_type = __get_jwk_ver_material_type(decoded_encnumbasis)
return VerificationMaterial(
field=PublicKeyField.JWK,
type=ver_material_type,
value=JWK_OKP(
ver_material_type, decoded_encnumbasis_without_prefix
).to_dict(),
encnumbasis=encnumbasis,
)
raise ValueError("Unknown format {}".format(ver_material_format))
def __get_2018_2019_ver_material_type(decoded_encnumbasis):
public_key_type = __get_public_key_type(decoded_encnumbasis)
if public_key_type == PublicKeyTypeAgreement.X25519:
return VerificationMaterialTypeAgreement.X25519_KEY_AGREEMENT_KEY_2019
elif public_key_type == PublicKeyTypeAuthentication.ED25519:
return VerificationMaterialTypeAuthentication.ED25519_VERIFICATION_KEY_2018
raise ValueError("Unknown public key type {}".format(public_key_type))
def __get_2020_ver_material_type(decoded_encnumbasis):
public_key_type = __get_public_key_type(decoded_encnumbasis)
if public_key_type == PublicKeyTypeAgreement.X25519:
return VerificationMaterialTypeAgreement.X25519_KEY_AGREEMENT_KEY_2020
elif public_key_type == PublicKeyTypeAuthentication.ED25519:
return VerificationMaterialTypeAuthentication.ED25519_VERIFICATION_KEY_2020
raise ValueError("Unknown public key type {}".format(public_key_type))
def __get_jwk_ver_material_type(decoded_encnumbasis):
public_key_type = __get_public_key_type(decoded_encnumbasis)
if public_key_type == PublicKeyTypeAgreement.X25519:
return VerificationMaterialTypeAgreement.JSON_WEB_KEY_2020
elif public_key_type == PublicKeyTypeAuthentication.ED25519:
return VerificationMaterialTypeAuthentication.JSON_WEB_KEY_2020
raise ValueError("Unknown public key type {}".format(type))
def __get_public_key_type(data: bytes) -> PublicKeyType:
prefix = _extract_prefix(data)
if prefix in set(item.value for item in PublicKeyTypeAuthentication):
return PublicKeyTypeAuthentication(prefix)
elif prefix in set(item.value for item in PublicKeyTypeAgreement):
return PublicKeyTypeAgreement(prefix)
else:
raise ValueError("Prefix {} not present in the lookup table".format(prefix))
def _remove_prefix(data: bytes) -> bytes:
"""
Removes prefix from data
:param data: prefixed data
:return: data without prefix
"""
prefix_int = _extract_prefix(data)
prefix = varint.encode(prefix_int)
return data[len(prefix) :]
def _extract_prefix(data: bytes) -> int:
"""
Extracts prefix from data
:param data: prefixed data
:raises ValueError: if invalid varint provided
:return: prefix
"""
try:
return varint.decode_bytes(data)
except TypeError:
raise ValueError("incorrect varint provided")
def _add_prefix(
key_type: Union[PublicKeyTypeAgreement, PublicKeyTypeAuthentication], data: bytes
) -> bytes:
"""
Adds prefix to a data
:param key_type: type of key
:param data: data to be prefixed
:return: prefixed data
"""
prefix = varint.encode(key_type.value)
return b"".join([prefix, data])
def _to_base58_multibase(value: bytes) -> str:
return MultibasePrefix.BASE58.value + base58.b58encode(value).decode("utf-8")
def _encode_filename(filename: str) -> str:
"""
Encodes filename to SHA256 string
:param filename: name of file
:return: encoded filename as SHA256 string
"""
return hashlib.sha256(filename.encode()).hexdigest()
def _check_key_correctly_encoded(key: str, encoding_type: EncodingType) -> bool:
"""
Checks if key correctly encoded
:param key: any string
:param encoding_type: encoding type
:return: true if key correctly encoded, otherwise false
"""
if not encoding_type == EncodingType.BASE58:
return False
alphabet = set("123456789ABCDEFGHJKLMNPQRSTUVWXYZabcdefghijkmnopqrstuvwxyz")
byte_lengths = (32,)
invalid_chars = set(key) - alphabet
if invalid_chars:
return False
b58len = len(base58.b58decode(key))
if b58len not in byte_lengths:
return False
return True
| 1,857 | 97 | 161 |
0d92b5e007c5d6d0f17cf60bdf5989db9053fe6d | 424 | py | Python | examples/pix_create_immediate_charge.py | gerencianet/gn-api-sdk-python | 09664fdcc7b9d99db6265c63b5c0efdfa413acd9 | [
"MIT"
] | 12 | 2016-10-14T15:56:59.000Z | 2021-10-30T01:56:51.000Z | examples/pix_create_immediate_charge.py | gerencianet/gn-api-sdk-python | 09664fdcc7b9d99db6265c63b5c0efdfa413acd9 | [
"MIT"
] | 7 | 2018-12-27T16:30:20.000Z | 2021-06-30T12:26:04.000Z | examples/pix_create_immediate_charge.py | gerencianet/gn-api-sdk-python | 09664fdcc7b9d99db6265c63b5c0efdfa413acd9 | [
"MIT"
] | 13 | 2016-07-13T14:28:47.000Z | 2022-03-16T14:17:40.000Z | from gerencianet import Gerencianet
from credentials import CREDENTIALS
gn = Gerencianet(CREDENTIALS)
body = {
'calendario': {
'expiracao': 3600
},
'devedor': {
'cpf': '',
'nome': ''
},
'valor': {
'original': ''
},
'chave': '',
'solicitacaoPagador': 'Cobrança dos serviços prestados.'
}
response = gn.pix_create_immediate_charge(body=body)
print(response)
| 18.434783 | 60 | 0.591981 | from gerencianet import Gerencianet
from credentials import CREDENTIALS
gn = Gerencianet(CREDENTIALS)
body = {
'calendario': {
'expiracao': 3600
},
'devedor': {
'cpf': '',
'nome': ''
},
'valor': {
'original': ''
},
'chave': '',
'solicitacaoPagador': 'Cobrança dos serviços prestados.'
}
response = gn.pix_create_immediate_charge(body=body)
print(response)
| 0 | 0 | 0 |
0b1315f177f51dde7d9757a1fdebf7c77dbae059 | 11,586 | py | Python | utils/rancher_mediator.py | Devwarlt/rancher2md-wrapper | 6a03f2504aad739f5549b6a8346d1bb4f153931d | [
"WTFPL"
] | 1 | 2021-04-29T00:42:29.000Z | 2021-04-29T00:42:29.000Z | utils/rancher_mediator.py | Devwarlt/rancher2md-wrapper | 6a03f2504aad739f5549b6a8346d1bb4f153931d | [
"WTFPL"
] | null | null | null | utils/rancher_mediator.py | Devwarlt/rancher2md-wrapper | 6a03f2504aad739f5549b6a8346d1bb4f153931d | [
"WTFPL"
] | null | null | null | from models.cluster import Cluster
from utils.markdown_mediator import MarkdownMediator
from utils.reload_credentials import ask_for_new_credentials
from utils.input_dialogs import yes_or_no_input_dialog
from extensions.url import format_url
from requests.models import Response
from utils.logger import Log
from config import app_config
from requests import get
from typing import Any
from json import loads
| 34.379822 | 88 | 0.529087 | from models.cluster import Cluster
from utils.markdown_mediator import MarkdownMediator
from utils.reload_credentials import ask_for_new_credentials
from utils.input_dialogs import yes_or_no_input_dialog
from extensions.url import format_url
from requests.models import Response
from utils.logger import Log
from config import app_config
from requests import get
from typing import Any
from json import loads
class RancherMediator:
def __init__(self) -> None:
pass
@staticmethod
def __get_response(path: str = '', **kwargs: dict) -> tuple:
r_endpoint: str = app_config['rancher']['endpoint']
r_username: str = app_config['rancher']['username']
r_password: str = app_config['rancher']['password']
params: dict = {'limit': 1000}
args: dict = kwargs.get('args', {})
if args:
params.update(args)
url: str = None
if kwargs.get('raw'):
url = path
else:
url = format_url(r_endpoint, path)
payload: dict = {
'url': url,
'auth': (r_username, r_password),
'verify': False,
'params': params
}
result: Response = get(**payload)
status_code: int = result.status_code
content: dict = {}
if len(result.content) != 0:
content = loads(result.content)
return status_code, content
@staticmethod
def __handle_bad_response(status_code: int, content: dict) -> None:
log: Log = Log.get_singleton()
log.error(
"Something wrong happened!",
origin='Rancher',
args={
'Status Code': status_code,
'Message': content.get('message')
}
)
@staticmethod
def __add_key_value_pair(key: str, value: Any) -> None:
internal: dict = app_config['internal']
internal.update({key: value})
app_config['internal'] = internal
@staticmethod
def __try_get_value(key: str) -> tuple:
internal: dict = app_config['internal']
if not key in internal:
return False, None
return True, internal.get(key)
@staticmethod
def __try_update_value(key: str, value: Any) -> bool:
internal: dict = app_config['internal']
if not key in internal:
return False
internal[key] = value
app_config['internal'] = internal
return True
@staticmethod
def __get_entries() -> dict:
return app_config['internal']
@staticmethod
def __validate_credentials() -> bool:
log: Log = Log.get_singleton()
log.info(
"Hold on, let me validate your credentials to proceed...",
origin='Rancher'
)
status_code: int = 0
content: dict = {}
status_code, content = RancherMediator.__get_response()
if status_code != 200:
log.error(
"Unable to authenticate! Consider to check your credentials.",
origin='Rancher',
args={
'Status Code': status_code,
'Message': content.get('message')
}
)
return False
log.info(
"Cool! I'm ready to use Rancher API :D",
origin='Rancher'
)
return True
@staticmethod
def __try_fetch_clusters() -> bool:
log: Log = Log.get_singleton()
log.info(
"Let me take a look into your clusters...",
origin='Rancher'
)
status_code: int = 0
content: dict = {}
path: str = app_config['static']['clusters']
status_code, content = RancherMediator.__get_response(path)
if status_code != 200:
RancherMediator.__handle_bad_response(status_code, content)
return False
clusters: list = []
data: list = content.get('data')
for i in range(len(data)):
cluster_data: dict = data[i]
cluster_id: str = cluster_data.get('id')
cluster_name: str = cluster_data.get('name')
cluster: dict = {
'id': cluster_id,
'name': cluster_name
}
clusters.append(cluster)
RancherMediator.__add_key_value_pair('clusters', clusters)
return True
@staticmethod
def __try_fetch_projects(cluster: dict) -> bool:
cluster_id: str = cluster.get('id')
cluster_name: str = cluster.get('name')
log: Log = Log.get_singleton()
log.info(
"Seeking for all projects inside cluster "
f"'{cluster_name}' [ID: {cluster_id}]...",
origin='Rancher'
)
status_code: int = 0
content: dict = {}
path: str = f"{app_config['static']['clusters']}/{cluster_id}/projects"
status_code, content = RancherMediator.__get_response(path)
if status_code != 200:
RancherMediator.__handle_bad_response(status_code, content)
return False
projects: list = []
data: list = content.get('data')
for i in range(len(data)):
project_data: dict = data[i]
project_id: str = project_data.get('id')
project_name: str = project_data.get('name')
project_links: dict = project_data.get('links')
project: dict = {
'id': project_id,
'name': project_name,
'links': project_links
}
projects.append(project)
cluster['projects'] = projects
return True
@staticmethod
def __try_fetch_workloads(project: dict) -> bool:
project_id: str = project.get('id')
project_name: str = project.get('name')
project_links: dict = project.pop('links')
log: Log = Log.get_singleton()
log.info(
"Seeking for all workloads inside project "
f"'{project_name}' [ID: {project_id}]...",
origin='Rancher'
)
status_code: int = 0
content: dict = {}
path: str = project_links.get('workloads')
status_code, content = RancherMediator.__get_response(path, raw=True)
if status_code != 200:
RancherMediator.__handle_bad_response(status_code, content)
return False
workloads: list = []
data: list = content.get('data')
for i in range(len(data)):
workload_data: dict = data[i]
workload_id: str = workload_data.get('id')
workload_name: str = workload_data.get('name')
workload_containers: list = workload_data.get('containers')
workload_namespace: str = workload_data.get('namespaceId')
workload: dict = {
'id': workload_id,
'name': workload_name,
'containers': workload_containers,
'namespace': workload_namespace
}
workloads.append(workload)
project['workloads'] = workloads
return True
@staticmethod
def __try_wraps_version(workload: dict) -> bool:
workload_containers: list = workload.pop('containers')
if not workload_containers:
return False
workload_version: str = None
for i in range(len(workload_containers)):
workload_container: dict = workload_containers[i]
container_image: str = workload_container.get('image')
if not container_image.__contains__(':'):
continue
container_version_split: str = container_image.split(':')
workload_version = container_version_split[1]
break
workload['version'] = workload_version
return True
@staticmethod
def core() -> None:
log: Log = Log.get_singleton()
while True:
if not RancherMediator.__validate_credentials():
if yes_or_no_input_dialog("Do you want to retry with new credentials?"):
ask_for_new_credentials()
continue
else:
break
if not RancherMediator.__try_fetch_clusters():
log.error(
"Unable to fetch any cluster! Therefore, I cannot proceed...",
origin='Rancher'
)
break
clusters: list = []
_, clusters = RancherMediator.__try_get_value('clusters')
log.info(
"Clusters detected in Rancher!",
origin='Rancher',
args={'Number of clusters': len(clusters)}
)
for i in range(len(clusters)):
cluster: dict = clusters[i]
if not RancherMediator.__try_fetch_projects(cluster):
log.warning(
"Well... There is no project for "
f"cluster '{cluster.get('name')}'.",
origin='Rancher'
)
continue
projects: list = cluster.get('projects')
log.info(
f"Projects detected in cluster '{cluster.get('name')}'!",
origin='Rancher',
args={'Number of projects': len(projects)}
)
for i in range(len(projects)):
project: dict = projects[i]
if not RancherMediator.__try_fetch_workloads(project):
log.warning(
"Well... There is no workload for "
f"project '{project.get('name')}' "
f"from cluster '{cluster.get('name')}'.",
origin='Rancher'
)
continue
workloads: list = project.get('workloads')
log.info(
f"Workloads detected in project '{project.get('name')}'"
f" from cluster '{cluster.get('name')}'!",
origin='Rancher',
args={'Number of projects': len(workloads)}
)
for j in range(len(workloads)):
workload: dict = workloads[j]
if not RancherMediator.__try_wraps_version(workload):
log.warning(
"Well... There is no container for "
f"workload '{workload.get('name')}' "
f"in project '{project.get('name')}' "
f"from cluster '{cluster.get('name')}'.",
origin='Rancher'
)
continue
RancherMediator.__try_update_value('clusters', clusters)
entries: dict = RancherMediator.__get_entries()
clusters: list = []
clusters_data: list = entries.get('clusters')
for i in range(len(clusters_data)):
cluster_data: dict = clusters_data[i]
cluster: Cluster = Cluster()
cluster.serialize(cluster_data)
clusters.append(cluster)
MarkdownMediator.build_report(clusters)
break
log.warning(
"All services are preparing to shutdown...",
origin='Rancher'
)
| 10,586 | 567 | 23 |
b76ad00900f8bb10caece4bf2ce27d9be5f5f1f5 | 1,070 | py | Python | SumarySearch/views.py | reetikaSR/summarySearch | c1de931c11d9dd755ba461742009a0d9d14ab08a | [
"MIT"
] | null | null | null | SumarySearch/views.py | reetikaSR/summarySearch | c1de931c11d9dd755ba461742009a0d9d14ab08a | [
"MIT"
] | null | null | null | SumarySearch/views.py | reetikaSR/summarySearch | c1de931c11d9dd755ba461742009a0d9d14ab08a | [
"MIT"
] | null | null | null | from django.views.decorators.csrf import csrf_exempt
from rest_framework.decorators import api_view
from rest_framework.response import Response
from SumarySearch.compute_match import match
from .serializers import BookSerializer
from django.http import HttpResponseBadRequest
@api_view(['POST'])
@csrf_exempt
def search_summary(request):
"""
:param request: request with POST type and data with list of queries and number of results needed for each query.
:return: response with the matching results of all the queries in the query list.
"""
queries = request.data.getlist('queries', None)
k = request.data.get('k')
if not queries:
return HttpResponseBadRequest("Missing required fields: queries")
if not k:
return HttpResponseBadRequest("Missing required fields: k")
data = []
for query in queries:
matching_results = match(query, int(k))
results_data = BookSerializer(matching_results, many=True, context={'query': query}).data
data.extend(results_data)
return Response(data)
| 32.424242 | 117 | 0.734579 | from django.views.decorators.csrf import csrf_exempt
from rest_framework.decorators import api_view
from rest_framework.response import Response
from SumarySearch.compute_match import match
from .serializers import BookSerializer
from django.http import HttpResponseBadRequest
@api_view(['POST'])
@csrf_exempt
def search_summary(request):
"""
:param request: request with POST type and data with list of queries and number of results needed for each query.
:return: response with the matching results of all the queries in the query list.
"""
queries = request.data.getlist('queries', None)
k = request.data.get('k')
if not queries:
return HttpResponseBadRequest("Missing required fields: queries")
if not k:
return HttpResponseBadRequest("Missing required fields: k")
data = []
for query in queries:
matching_results = match(query, int(k))
results_data = BookSerializer(matching_results, many=True, context={'query': query}).data
data.extend(results_data)
return Response(data)
| 0 | 0 | 0 |
399518ebae7320b7a8368861e51010b913d3a1aa | 2,239 | py | Python | SpectrometersV2/example_code.py | ColdMatter/PhotonBEC | c6bcf9bdefd267c8adde0d299cf5920b010c5022 | [
"MIT"
] | null | null | null | SpectrometersV2/example_code.py | ColdMatter/PhotonBEC | c6bcf9bdefd267c8adde0d299cf5920b010c5022 | [
"MIT"
] | null | null | null | SpectrometersV2/example_code.py | ColdMatter/PhotonBEC | c6bcf9bdefd267c8adde0d299cf5920b010c5022 | [
"MIT"
] | null | null | null | '''
Written by: Joao Rodrigues
Last Update: October 16th 2020
'''
import sys
from socket import gethostname
if gethostname().lower() == "ph-photonbec5":
sys.path.append("D:/Control/PythonPackages/")
import pbec_ipc
from time import sleep
from time import time
import matplotlib.pyplot as plt
from single_spec_IPC_module import set_spectrometer_mode, set_spectrometer_integration_time, set_spectrometer_n_averages, get_spectrum_measure
from single_spec_IPC_module import get_spectrometer_integration_time, get_spectrometer_n_averages, set_spectrometer_external_trigger
########## Control Parameters
spectrometer_server_port = pbec_ipc.PORT_NUMBERS["spectrometer_server V2"]
spectrometer_server_host = 'localhost'
##### Sets the spectrometer mode to either "continuous" or "internal". All data should be acquiered in internal mode. However, the function that...
##### ...gets the data, overrides the mode to "internal" if this is not the case already
set_spectrometer_mode(mode='internal', port=spectrometer_server_port, host=spectrometer_server_host)
##### Sets the external trigger
set_spectrometer_external_trigger(external_trigger=False, port=spectrometer_server_port, host=spectrometer_server_host)
##### Sets spectrometer integration time and number of averages
set_spectrometer_integration_time(int_time=250, port=spectrometer_server_port, host=spectrometer_server_host)
set_spectrometer_n_averages(n_averages=2, port=spectrometer_server_port, host=spectrometer_server_host)
grabbed_int_time = get_spectrometer_integration_time(port=spectrometer_server_port, host=spectrometer_server_host)
grabbed_n_averages = get_spectrometer_n_averages(port=spectrometer_server_port, host=spectrometer_server_host)
print("Integration time set to {0}".format(grabbed_int_time))
print("Number of averages set to {0}".format(grabbed_n_averages))
##### Grabs some data
spectrum_time_label, spectrum_new_data_flag, lamb, spectrum = get_spectrum_measure(int_time=250, n_averages=2, n_measures=1, port=spectrometer_server_port, host=spectrometer_server_host)
plt.plot(lamb, spectrum)
plt.yscale('log')
plt.xlabel("Wavelength (nm)", fontsize=9)
plt.ylabel("Spectrum", fontsize=9)
#plt.xlim(540, 600)
#plt.ylim(1, 40000)
plt.show()
| 38.603448 | 186 | 0.820456 | '''
Written by: Joao Rodrigues
Last Update: October 16th 2020
'''
import sys
from socket import gethostname
if gethostname().lower() == "ph-photonbec5":
sys.path.append("D:/Control/PythonPackages/")
import pbec_ipc
from time import sleep
from time import time
import matplotlib.pyplot as plt
from single_spec_IPC_module import set_spectrometer_mode, set_spectrometer_integration_time, set_spectrometer_n_averages, get_spectrum_measure
from single_spec_IPC_module import get_spectrometer_integration_time, get_spectrometer_n_averages, set_spectrometer_external_trigger
########## Control Parameters
spectrometer_server_port = pbec_ipc.PORT_NUMBERS["spectrometer_server V2"]
spectrometer_server_host = 'localhost'
##### Sets the spectrometer mode to either "continuous" or "internal". All data should be acquiered in internal mode. However, the function that...
##### ...gets the data, overrides the mode to "internal" if this is not the case already
set_spectrometer_mode(mode='internal', port=spectrometer_server_port, host=spectrometer_server_host)
##### Sets the external trigger
set_spectrometer_external_trigger(external_trigger=False, port=spectrometer_server_port, host=spectrometer_server_host)
##### Sets spectrometer integration time and number of averages
set_spectrometer_integration_time(int_time=250, port=spectrometer_server_port, host=spectrometer_server_host)
set_spectrometer_n_averages(n_averages=2, port=spectrometer_server_port, host=spectrometer_server_host)
grabbed_int_time = get_spectrometer_integration_time(port=spectrometer_server_port, host=spectrometer_server_host)
grabbed_n_averages = get_spectrometer_n_averages(port=spectrometer_server_port, host=spectrometer_server_host)
print("Integration time set to {0}".format(grabbed_int_time))
print("Number of averages set to {0}".format(grabbed_n_averages))
##### Grabs some data
spectrum_time_label, spectrum_new_data_flag, lamb, spectrum = get_spectrum_measure(int_time=250, n_averages=2, n_measures=1, port=spectrometer_server_port, host=spectrometer_server_host)
plt.plot(lamb, spectrum)
plt.yscale('log')
plt.xlabel("Wavelength (nm)", fontsize=9)
plt.ylabel("Spectrum", fontsize=9)
#plt.xlim(540, 600)
#plt.ylim(1, 40000)
plt.show()
| 0 | 0 | 0 |
8c62dfc48c7e80068f446336243ac76e72e649be | 2,293 | py | Python | factory-ai-vision/EdgeSolution/modules/WebModule/backend/vision_on_edge/cameras/apps.py | julialieberman/azure-intelligent-edge-patterns | 1e04748d640730301edd2c6fb51a02f9b6d2ff41 | [
"MIT"
] | 1 | 2020-08-17T16:50:09.000Z | 2020-08-17T16:50:09.000Z | factory-ai-vision/EdgeSolution/modules/WebModule/backend/vision_on_edge/cameras/apps.py | julialieberman/azure-intelligent-edge-patterns | 1e04748d640730301edd2c6fb51a02f9b6d2ff41 | [
"MIT"
] | null | null | null | factory-ai-vision/EdgeSolution/modules/WebModule/backend/vision_on_edge/cameras/apps.py | julialieberman/azure-intelligent-edge-patterns | 1e04748d640730301edd2c6fb51a02f9b6d2ff41 | [
"MIT"
] | null | null | null | """
Cameras App
"""
import logging
import sys
from django.apps import AppConfig
logger = logging.getLogger(__name__)
DEFAULT_SETTING_NAME = 'DEFAULT_SETTING'
class CamerasConfig(AppConfig):
"""
Cameras App Config
"""
name = 'vision_on_edge.cameras'
def ready(self):
"""
Cameras App ready
"""
# FIXME test may use this as well
if 'runserver' in sys.argv:
# Import models in migrate/makemigration will occurs error.
# pylint: disable=C0415
from vision_on_edge.cameras.models import Part, Camera
# pylint: enable=C0415
logger.info("CameraAppConfig ready while running server")
create_demo = True
if create_demo:
logger.info("Creating Demo Parts")
# for partname in ['Box', 'Barrel', 'Hammer',
# 'Screwdriver', 'Bottle', 'Plastic bag']:
for partname in [
'aeroplane',
'bicycle',
'bird',
'boat',
'bottle',
'bus',
'car',
'cat',
'chair',
'cow',
'diningtable',
'dog',
'horse',
'motorbike',
'person',
'pottedplant',
'sheep',
'sofa',
'train',
'tvmonitor',
]:
Part.objects.update_or_create(
name=partname,
is_demo=True,
defaults={'description': "Demo"})
logger.info("Creating Demo Camera")
Camera.objects.update_or_create(
name="Demo Video",
is_demo=True,
defaults={
'rtsp': 'sample_video/video.mp4',
'area': ""
})
logger.info("Creating Demo... End")
logger.info("CameraAppConfig End while running server")
| 29.779221 | 71 | 0.411688 | """
Cameras App
"""
import logging
import sys
from django.apps import AppConfig
logger = logging.getLogger(__name__)
DEFAULT_SETTING_NAME = 'DEFAULT_SETTING'
class CamerasConfig(AppConfig):
"""
Cameras App Config
"""
name = 'vision_on_edge.cameras'
def ready(self):
"""
Cameras App ready
"""
# FIXME test may use this as well
if 'runserver' in sys.argv:
# Import models in migrate/makemigration will occurs error.
# pylint: disable=C0415
from vision_on_edge.cameras.models import Part, Camera
# pylint: enable=C0415
logger.info("CameraAppConfig ready while running server")
create_demo = True
if create_demo:
logger.info("Creating Demo Parts")
# for partname in ['Box', 'Barrel', 'Hammer',
# 'Screwdriver', 'Bottle', 'Plastic bag']:
for partname in [
'aeroplane',
'bicycle',
'bird',
'boat',
'bottle',
'bus',
'car',
'cat',
'chair',
'cow',
'diningtable',
'dog',
'horse',
'motorbike',
'person',
'pottedplant',
'sheep',
'sofa',
'train',
'tvmonitor',
]:
Part.objects.update_or_create(
name=partname,
is_demo=True,
defaults={'description': "Demo"})
logger.info("Creating Demo Camera")
Camera.objects.update_or_create(
name="Demo Video",
is_demo=True,
defaults={
'rtsp': 'sample_video/video.mp4',
'area': ""
})
logger.info("Creating Demo... End")
logger.info("CameraAppConfig End while running server")
| 0 | 0 | 0 |
72c67ce90edfb7400c6369cd615fde79fc8633f1 | 14,247 | py | Python | collaborative_system/temperature_scaling_vit.py | drunkcoding/efficient-nlp | 9509e0ef08016506280a7cfc600ea8e3778dea2d | [
"Apache-2.0"
] | null | null | null | collaborative_system/temperature_scaling_vit.py | drunkcoding/efficient-nlp | 9509e0ef08016506280a7cfc600ea8e3778dea2d | [
"Apache-2.0"
] | null | null | null | collaborative_system/temperature_scaling_vit.py | drunkcoding/efficient-nlp | 9509e0ef08016506280a7cfc600ea8e3778dea2d | [
"Apache-2.0"
] | null | null | null | import torch
import logging
import numpy as np
from outliers import smirnov_grubbs as grubbs
import seaborn as sns
import matplotlib
import matplotlib.pyplot as plt
from torch.utils.data.sampler import SubsetRandomSampler
from torchvision import datasets, transforms
import torchvision.models as cv_models
import os
import sys
from torch.nn.modules.activation import Threshold
sys.path.append(os.path.join(os.path.dirname(__file__), "../"))
from ecosys.models.temperature_scaling import ModelWithTemperature
from ecosys.utils.data_processor import processors, output_modes
from ecosys.utils.data_structure import Dataset, HuggingFaceDataset, TorchVisionDataset, ViTDataset
from ecosys.algo.monte_carlo import monte_carlo_bounds
from ecosys.decorators.eval_decorators import model_eval
from ecosys.utils.eval import compute_metrics
from tqdm import tqdm
from transformers import ViTFeatureExtractor, ViTForImageClassification
from torch.utils.data import DataLoader, Subset, SequentialSampler, TensorDataset, dataloader
from sklearn.model_selection import train_test_split
logging.basicConfig(
format="%(asctime)s - %(levelname)s - %(name)s - %(message)s",
datefmt="%m/%d/%Y %H:%M:%S",
level=logging.INFO,
)
logger = logging.getLogger(__name__)
feature_size = 768
sequence_length = 128
task_name = 'resnet'
batch_size = 32
filename = __file__
filename = filename.split(".")[0]
fh = logging.FileHandler(f'{filename}_{task_name}.log', mode='a')
fh.setLevel(logging.INFO)
logger.addHandler(fh)
device = torch.device("cuda:1" if torch.cuda.is_available() else "cpu")
base_dir = "/home/oai/share"
tokenizer = ViTFeatureExtractor.from_pretrained(f"{base_dir}/HuggingFace/vit-base-patch32-384")
model_keys = [
"S",
"M",
"L",
]
energy_discount_factor = [
0.05,
0.25,
0.5,
# 1.0,
]
model_paths = [
"",
f"{base_dir}/HuggingFace/vit-base-patch32-384",
f"{base_dir}/HuggingFace/vit-large-patch32-384",
]
model_energy = dict(zip(model_keys, energy_discount_factor))
model_paths = dict(zip(model_keys, model_paths))
models = dict()
for key in model_keys:
logger.debug("key %s, path %s", key, model_paths[key])
models[key] = ViTForImageClassification.from_pretrained(model_paths[key]).to(device) if key != "S" else cv_models.resnet18(pretrained=True)
models[key].eval()
models[key].to(device)
# ------------- Dataset Prepare --------------
preprocess = transforms.Compose([
transforms.Resize(256),
transforms.CenterCrop(224),
transforms.ToTensor(),
transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]),
])
g_cpu = torch.Generator(device='cpu')
g_cpu.manual_seed(2147483647)
index = np.array([x for x in range(len(datasets.ImageNet("/home/oai/share/dataset/.", split="val", transform=preprocess)))])
# index = np.array([x for x in range(1000)])
train_index, test_index = train_test_split(index, test_size=0.6)
# train_sampler = SequentialSampler(train_index)
# test_sampler = SequentialSampler(test_index)
m = torch.nn.Softmax(dim=1)
train_dataloader, test_dataloader = tans_data_preprocessing()
raw_train_dataloader, raw_test_dataloader = data_preprocessing()
data_limit = 1000
# ------------- Train Temperature --------------
for key in model_keys:
models[key] = ModelWithTemperature(models[key])
models[key].set_logger(logger)
if key != 'S':
models[key].set_temperature(train_dataloader)
else:
models[key].set_temperature(raw_train_dataloader)
n_models = len(model_keys)
model_probs = dict(zip(model_keys, [np.array(list()) for _ in range(n_models)]))
num_labels = 0
with torch.no_grad():
# for input, label, raw_input, _ in tqdm(zip(train_dataloader, raw_train_dataloader), desc="Find Threshold"):
for d, raw_d in tqdm(zip(train_dataloader, raw_train_dataloader), desc="Find Threshold"):
input, label = d
raw_input, raw_label = raw_d
num_labels += len(label)
# print(label)
# print(raw_label)
assert np.count_nonzero(label.cpu() == raw_label.cpu()) == len(label.cpu())
for key in model_keys:
logits = models[key](input) if key != 'S' else models[key](raw_input)
probabilities = m(logits).cpu().detach().numpy()
model_ans = np.argmax(probabilities, axis=1).flatten()
model_probs[key] = np.append(model_probs[key], [p[model_ans[i]] for i, p in enumerate(probabilities)])
# print(num_labels, model_probs)
threshold_bounds = monte_carlo_bounds(
total_reward,
[(0.5, 1.0)]*(len(model_keys)-1),
[('reward', float), ('energy', float)],
n=1000,
tops=20,
maxiter=20,
)
mc_threshold = np.mean(
threshold_bounds, axis=1
)
logger.info("Threshold Bounds %s", threshold_bounds)
# exit()
# ------------- Evaluation WITH Temperature --------------
correct_cnt = dict(zip(model_keys, [0]*n_models))
correct_prob = dict(zip(model_keys, [0]*n_models))
coop_cnt = dict(zip(model_keys, [0]*n_models))
process_prob = dict(zip(model_keys, [0]*n_models))
process_cnt = dict(zip(model_keys, [0]*n_models))
num_labels = 0
# th_stats = []
# threshold = None
th_stats = dict(zip(model_keys, [list() for _ in range(n_models)]))
with torch.no_grad():
# @model_eval(test_dataloader)
# def eval_monte_carlo(input, label):
# global num_labels
# global th_stats
for d, raw_d in tqdm(zip(test_dataloader, raw_test_dataloader), desc="Evaluating"):
input, label = d
raw_input, raw_label = raw_d
# print(label)
# print(raw_label)
assert np.count_nonzero(label.cpu() == raw_label.cpu()) == len(label.cpu())
num_labels += len(label)
b_size = len(label.cpu())
mask = np.array([False]*b_size)
for i, key in enumerate(model_keys):
logits = models[key](input) if key != 'S' else models[key](raw_input)
probabilities = m(logits).cpu().detach().numpy()
# if key in ['S']:
# th_stats += np.max(probabilities, axis=1).tolist()
th_stats[key] += np.max(probabilities, axis=1).tolist()
model_ans = np.argmax(probabilities, axis=1)
true_ans = label.cpu().detach().numpy().flatten()
selected_prob = np.array([p[model_ans[i]] for i, p in enumerate(probabilities)])
processed = (selected_prob >= mc_threshold[i]) if key in model_keys[:-1] else np.array([True]*b_size)
correct_prob[key] += np.sum(selected_prob)
process_prob[key] += np.sum(selected_prob[(~mask) & processed])
correct_cnt[key] += np.count_nonzero(model_ans == true_ans)
coop_cnt[key] += np.count_nonzero((model_ans == true_ans) & (~mask) & processed)
process_cnt[key] += np.count_nonzero((~mask) & processed)
mask |= processed
num_labels += b_size
# eval_monte_carlo()
# for input, label in tqdm(test_dataloader, desc="Evaluating"):
# # input_ids = data_batch['input_ids'].to(device)
# # attention_mask = data_batch['attention_mask'].to(device)
# # labels = data_batch['labels'].to(device)
# # if num_labels < 20 :
# # with torch.no_grad():
# # logits = models["S"](input_ids=input_ids, attention_mask=attention_mask)
# # probabilities = m(logits).cpu()
# # th_stats.append(np.max(probabilities.cpu().detach().numpy(), axis=1).tolist())
# # num_labels += batch_size
# # continue
# # elif threshold is None:
# # threshold = np.mean(th_stats)
# # print(threshold)
# # processed = (selected_prob >= threshold) if key in ['S'] else np.array([True]*len(labels.cpu()))
# b_size = len(label.cpu())
# mask = np.array([False]*b_size)
# with torch.no_grad():
# for key in model_keys:
# logits = models[key](input)
# # print(logits)
# # logits = output.logits
# probabilities = m(logits).cpu().detach().numpy()
# if key in ['S']:
# th_stats += np.max(probabilities, axis=1).tolist()
# data = np.array(th_stats[-(10*batch_size):]).reshape((-1,1))
# # best_bic = np.inf
# # best_gm = None
# # for k in range(2,4):
# # gm = GaussianMixture(n_components=k, random_state=0).fit(data)
# # bic = gm.bic(data)
# # # idx = np.argmax(gm.means_.flatten())
# # # mu = gm.means_.flatten()[idx]
# # # var = gm.covariances_.flatten()[idx]
# # if bic < best_bic:
# # best_gm = gm
# # best_bic = bic
# best_gm = GaussianMixture(n_components=2, random_state=0).fit(data)
# # print("cov", gm.covariances_)
# # print("mean", gm.means_)
# idx = np.argmax(best_gm.means_.flatten())
# mu = best_gm.means_.flatten()[idx]
# var = best_gm.covariances_.flatten()[idx]
# logger.info("Model BIC %s, mu %s, var %s", best_gm.bic(data), mu, var)
# threshold = mu - 3*np.sqrt(var)
# threshold = mc_threshold
# model_ans = np.argmax(probabilities, axis=1)
# true_ans = label.cpu().detach().numpy().flatten()
# selected_prob = np.array([p[model_ans[i]] for i, p in enumerate(probabilities)])
# # print(probabilities)
# # print(model_ans)
# processed = (selected_prob >= threshold) if key in ['S'] else np.array([True]*b_size)
# # outliers_idx = grubbs.min_test_indices(th_stats[-(10*batch_size):], alpha=0.3)
# # # logger.debug("outliers_idx %s", outliers_idx)
# # outliers_mask = np.array([False]*min(len(th_stats), 10*batch_size))
# # outliers_mask[outliers_idx] = True
# # outliers_mask = outliers_mask[-b_size:]
# # processed = np.array([True]*b_size)
# # if key in ['S']:
# # processed[outliers_mask] = False
# correct_prob[key] += np.sum(selected_prob)
# process_prob[key] += np.sum(selected_prob[(~mask) & processed])
# correct_cnt[key] += np.count_nonzero(model_ans == true_ans)
# coop_cnt[key] += np.count_nonzero((model_ans == true_ans) & (~mask) & processed)
# process_cnt[key] += np.count_nonzero((~mask) & processed)
# mask |= processed
# num_labels += b_size
# num_labels -= len(np.array(th_stats).flatten())
for key in model_keys:
logger.info("%s Mean Probability = %s", key, np.mean(th_stats[key]))
sns.distplot(th_stats[key], hist=True, kde=True,
bins=int(180/5),
# color = 'darkblue',
label=key,
hist_kws={'edgecolor':'black'},
kde_kws={'linewidth': 4})
# logger.info("%s Mean Probability = %s", key, np.mean(th_stats))
# sns.distplot(th_stats, hist=True, kde=True,
# bins=int(180/5), color = 'darkblue',
# hist_kws={'edgecolor':'black'},
# kde_kws={'linewidth': 4})
plt.legend()
plt.savefig(f'figures/th_stats_{task_name}.png', bbox_inches="tight")
logger.info(" Num examples = %s", num_labels)
logger.info(" Threshold = %s", mc_threshold)
for key in model_keys:
logger.info("final temperature %s", models[key].temperature)
logger.info("***** Eval results *****")
for key in model_keys:
logger.info("%s correct count %s, percent %d, prob %s", key, correct_cnt[key], int(correct_cnt[key]/num_labels * 100), correct_prob[key])
logger.info("***** Collaborative Eval results *****")
for key in model_keys:
logger.info("%s process count %s, correct count %s, percent %d, prob %s", key, process_cnt[key], coop_cnt[key], int(coop_cnt[key]/process_cnt[key] * 100) if process_cnt[key] != 0 else 0, process_prob[key])
| 37.492105 | 209 | 0.639082 | import torch
import logging
import numpy as np
from outliers import smirnov_grubbs as grubbs
import seaborn as sns
import matplotlib
import matplotlib.pyplot as plt
from torch.utils.data.sampler import SubsetRandomSampler
from torchvision import datasets, transforms
import torchvision.models as cv_models
import os
import sys
from torch.nn.modules.activation import Threshold
sys.path.append(os.path.join(os.path.dirname(__file__), "../"))
from ecosys.models.temperature_scaling import ModelWithTemperature
from ecosys.utils.data_processor import processors, output_modes
from ecosys.utils.data_structure import Dataset, HuggingFaceDataset, TorchVisionDataset, ViTDataset
from ecosys.algo.monte_carlo import monte_carlo_bounds
from ecosys.decorators.eval_decorators import model_eval
from ecosys.utils.eval import compute_metrics
from tqdm import tqdm
from transformers import ViTFeatureExtractor, ViTForImageClassification
from torch.utils.data import DataLoader, Subset, SequentialSampler, TensorDataset, dataloader
from sklearn.model_selection import train_test_split
logging.basicConfig(
format="%(asctime)s - %(levelname)s - %(name)s - %(message)s",
datefmt="%m/%d/%Y %H:%M:%S",
level=logging.INFO,
)
logger = logging.getLogger(__name__)
feature_size = 768
sequence_length = 128
task_name = 'resnet'
batch_size = 32
filename = __file__
filename = filename.split(".")[0]
fh = logging.FileHandler(f'{filename}_{task_name}.log', mode='a')
fh.setLevel(logging.INFO)
logger.addHandler(fh)
device = torch.device("cuda:1" if torch.cuda.is_available() else "cpu")
base_dir = "/home/oai/share"
tokenizer = ViTFeatureExtractor.from_pretrained(f"{base_dir}/HuggingFace/vit-base-patch32-384")
model_keys = [
"S",
"M",
"L",
]
energy_discount_factor = [
0.05,
0.25,
0.5,
# 1.0,
]
model_paths = [
"",
f"{base_dir}/HuggingFace/vit-base-patch32-384",
f"{base_dir}/HuggingFace/vit-large-patch32-384",
]
model_energy = dict(zip(model_keys, energy_discount_factor))
model_paths = dict(zip(model_keys, model_paths))
models = dict()
for key in model_keys:
logger.debug("key %s, path %s", key, model_paths[key])
models[key] = ViTForImageClassification.from_pretrained(model_paths[key]).to(device) if key != "S" else cv_models.resnet18(pretrained=True)
models[key].eval()
models[key].to(device)
# ------------- Dataset Prepare --------------
preprocess = transforms.Compose([
transforms.Resize(256),
transforms.CenterCrop(224),
transforms.ToTensor(),
transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]),
])
g_cpu = torch.Generator(device='cpu')
g_cpu.manual_seed(2147483647)
index = np.array([x for x in range(len(datasets.ImageNet("/home/oai/share/dataset/.", split="val", transform=preprocess)))])
# index = np.array([x for x in range(1000)])
train_index, test_index = train_test_split(index, test_size=0.6)
# train_sampler = SequentialSampler(train_index)
# test_sampler = SequentialSampler(test_index)
def data_preprocessing():
val_dataset = TorchVisionDataset(datasets.ImageNet("/home/oai/share/dataset/.", split="val", transform=preprocess))
# index = np.array([x for x in range(len(val_dataset))])
# index = np.array([x for x in range(1000)])
# train_index, test_index = train_test_split(index, test_size=0.6, random_state=0, shuffle=False)
# train, test = val_dataset[train_index], test_index[test_index]
train_dataset = Subset(val_dataset, train_index)
test_dataset = Subset(val_dataset, test_index)
train_sampler = SequentialSampler(train_dataset)
test_sampler = SequentialSampler(test_dataset)
train_loader = DataLoader(train_dataset, batch_size=batch_size, sampler=train_sampler)
test_loader = DataLoader(test_dataset, batch_size=batch_size, sampler=test_sampler)
return train_loader, test_loader
def tans_data_preprocessing():
val_dataset = ViTDataset(datasets.ImageNet("/home/oai/share/dataset/.", split="val"), tokenizer)
# index = np.array([x for x in range(len(val_dataset))])
# index = np.array([x for x in range(1000)])
# train_index, test_index = train_test_split(index, test_size=0.6, random_state=0, shuffle=False)
train_dataset = Subset(val_dataset, train_index)
test_dataset = Subset(val_dataset, test_index)
train_sampler = SequentialSampler(train_dataset)
test_sampler = SequentialSampler(test_dataset)
train_loader = DataLoader(train_dataset, batch_size=batch_size, sampler=train_sampler)
test_loader = DataLoader(test_dataset, batch_size=batch_size, sampler=test_sampler)
return train_loader, test_loader
m = torch.nn.Softmax(dim=1)
train_dataloader, test_dataloader = tans_data_preprocessing()
raw_train_dataloader, raw_test_dataloader = data_preprocessing()
data_limit = 1000
# ------------- Train Temperature --------------
for key in model_keys:
models[key] = ModelWithTemperature(models[key])
models[key].set_logger(logger)
if key != 'S':
models[key].set_temperature(train_dataloader)
else:
models[key].set_temperature(raw_train_dataloader)
n_models = len(model_keys)
model_probs = dict(zip(model_keys, [np.array(list()) for _ in range(n_models)]))
num_labels = 0
with torch.no_grad():
# for input, label, raw_input, _ in tqdm(zip(train_dataloader, raw_train_dataloader), desc="Find Threshold"):
for d, raw_d in tqdm(zip(train_dataloader, raw_train_dataloader), desc="Find Threshold"):
input, label = d
raw_input, raw_label = raw_d
num_labels += len(label)
# print(label)
# print(raw_label)
assert np.count_nonzero(label.cpu() == raw_label.cpu()) == len(label.cpu())
for key in model_keys:
logits = models[key](input) if key != 'S' else models[key](raw_input)
probabilities = m(logits).cpu().detach().numpy()
model_ans = np.argmax(probabilities, axis=1).flatten()
model_probs[key] = np.append(model_probs[key], [p[model_ans[i]] for i, p in enumerate(probabilities)])
# print(num_labels, model_probs)
def total_reward(threshold):
# threshold = threshold[0]
# print(threshold)
reward = 0
energy = 0
mask = np.array([False]*num_labels)
for i, key in enumerate(model_keys):
processed = (model_probs[key] >= threshold[i]) if key in model_keys[:-1] else np.array([True]*num_labels)
reward += np.sum(model_probs[key].round(3)[(~mask) & processed])
energy += model_energy[key]*np.count_nonzero((~mask) & processed)
mask |= processed
return (np.around(reward, 0), -energy)
# return reward - 0.3 * energy
threshold_bounds = monte_carlo_bounds(
total_reward,
[(0.5, 1.0)]*(len(model_keys)-1),
[('reward', float), ('energy', float)],
n=1000,
tops=20,
maxiter=20,
)
mc_threshold = np.mean(
threshold_bounds, axis=1
)
logger.info("Threshold Bounds %s", threshold_bounds)
# exit()
# ------------- Evaluation WITH Temperature --------------
correct_cnt = dict(zip(model_keys, [0]*n_models))
correct_prob = dict(zip(model_keys, [0]*n_models))
coop_cnt = dict(zip(model_keys, [0]*n_models))
process_prob = dict(zip(model_keys, [0]*n_models))
process_cnt = dict(zip(model_keys, [0]*n_models))
num_labels = 0
# th_stats = []
# threshold = None
th_stats = dict(zip(model_keys, [list() for _ in range(n_models)]))
with torch.no_grad():
# @model_eval(test_dataloader)
# def eval_monte_carlo(input, label):
# global num_labels
# global th_stats
for d, raw_d in tqdm(zip(test_dataloader, raw_test_dataloader), desc="Evaluating"):
input, label = d
raw_input, raw_label = raw_d
# print(label)
# print(raw_label)
assert np.count_nonzero(label.cpu() == raw_label.cpu()) == len(label.cpu())
num_labels += len(label)
b_size = len(label.cpu())
mask = np.array([False]*b_size)
for i, key in enumerate(model_keys):
logits = models[key](input) if key != 'S' else models[key](raw_input)
probabilities = m(logits).cpu().detach().numpy()
# if key in ['S']:
# th_stats += np.max(probabilities, axis=1).tolist()
th_stats[key] += np.max(probabilities, axis=1).tolist()
model_ans = np.argmax(probabilities, axis=1)
true_ans = label.cpu().detach().numpy().flatten()
selected_prob = np.array([p[model_ans[i]] for i, p in enumerate(probabilities)])
processed = (selected_prob >= mc_threshold[i]) if key in model_keys[:-1] else np.array([True]*b_size)
correct_prob[key] += np.sum(selected_prob)
process_prob[key] += np.sum(selected_prob[(~mask) & processed])
correct_cnt[key] += np.count_nonzero(model_ans == true_ans)
coop_cnt[key] += np.count_nonzero((model_ans == true_ans) & (~mask) & processed)
process_cnt[key] += np.count_nonzero((~mask) & processed)
mask |= processed
num_labels += b_size
# eval_monte_carlo()
# for input, label in tqdm(test_dataloader, desc="Evaluating"):
# # input_ids = data_batch['input_ids'].to(device)
# # attention_mask = data_batch['attention_mask'].to(device)
# # labels = data_batch['labels'].to(device)
# # if num_labels < 20 :
# # with torch.no_grad():
# # logits = models["S"](input_ids=input_ids, attention_mask=attention_mask)
# # probabilities = m(logits).cpu()
# # th_stats.append(np.max(probabilities.cpu().detach().numpy(), axis=1).tolist())
# # num_labels += batch_size
# # continue
# # elif threshold is None:
# # threshold = np.mean(th_stats)
# # print(threshold)
# # processed = (selected_prob >= threshold) if key in ['S'] else np.array([True]*len(labels.cpu()))
# b_size = len(label.cpu())
# mask = np.array([False]*b_size)
# with torch.no_grad():
# for key in model_keys:
# logits = models[key](input)
# # print(logits)
# # logits = output.logits
# probabilities = m(logits).cpu().detach().numpy()
# if key in ['S']:
# th_stats += np.max(probabilities, axis=1).tolist()
# data = np.array(th_stats[-(10*batch_size):]).reshape((-1,1))
# # best_bic = np.inf
# # best_gm = None
# # for k in range(2,4):
# # gm = GaussianMixture(n_components=k, random_state=0).fit(data)
# # bic = gm.bic(data)
# # # idx = np.argmax(gm.means_.flatten())
# # # mu = gm.means_.flatten()[idx]
# # # var = gm.covariances_.flatten()[idx]
# # if bic < best_bic:
# # best_gm = gm
# # best_bic = bic
# best_gm = GaussianMixture(n_components=2, random_state=0).fit(data)
# # print("cov", gm.covariances_)
# # print("mean", gm.means_)
# idx = np.argmax(best_gm.means_.flatten())
# mu = best_gm.means_.flatten()[idx]
# var = best_gm.covariances_.flatten()[idx]
# logger.info("Model BIC %s, mu %s, var %s", best_gm.bic(data), mu, var)
# threshold = mu - 3*np.sqrt(var)
# threshold = mc_threshold
# model_ans = np.argmax(probabilities, axis=1)
# true_ans = label.cpu().detach().numpy().flatten()
# selected_prob = np.array([p[model_ans[i]] for i, p in enumerate(probabilities)])
# # print(probabilities)
# # print(model_ans)
# processed = (selected_prob >= threshold) if key in ['S'] else np.array([True]*b_size)
# # outliers_idx = grubbs.min_test_indices(th_stats[-(10*batch_size):], alpha=0.3)
# # # logger.debug("outliers_idx %s", outliers_idx)
# # outliers_mask = np.array([False]*min(len(th_stats), 10*batch_size))
# # outliers_mask[outliers_idx] = True
# # outliers_mask = outliers_mask[-b_size:]
# # processed = np.array([True]*b_size)
# # if key in ['S']:
# # processed[outliers_mask] = False
# correct_prob[key] += np.sum(selected_prob)
# process_prob[key] += np.sum(selected_prob[(~mask) & processed])
# correct_cnt[key] += np.count_nonzero(model_ans == true_ans)
# coop_cnt[key] += np.count_nonzero((model_ans == true_ans) & (~mask) & processed)
# process_cnt[key] += np.count_nonzero((~mask) & processed)
# mask |= processed
# num_labels += b_size
# num_labels -= len(np.array(th_stats).flatten())
for key in model_keys:
logger.info("%s Mean Probability = %s", key, np.mean(th_stats[key]))
sns.distplot(th_stats[key], hist=True, kde=True,
bins=int(180/5),
# color = 'darkblue',
label=key,
hist_kws={'edgecolor':'black'},
kde_kws={'linewidth': 4})
# logger.info("%s Mean Probability = %s", key, np.mean(th_stats))
# sns.distplot(th_stats, hist=True, kde=True,
# bins=int(180/5), color = 'darkblue',
# hist_kws={'edgecolor':'black'},
# kde_kws={'linewidth': 4})
plt.legend()
plt.savefig(f'figures/th_stats_{task_name}.png', bbox_inches="tight")
logger.info(" Num examples = %s", num_labels)
logger.info(" Threshold = %s", mc_threshold)
for key in model_keys:
logger.info("final temperature %s", models[key].temperature)
logger.info("***** Eval results *****")
for key in model_keys:
logger.info("%s correct count %s, percent %d, prob %s", key, correct_cnt[key], int(correct_cnt[key]/num_labels * 100), correct_prob[key])
logger.info("***** Collaborative Eval results *****")
for key in model_keys:
logger.info("%s process count %s, correct count %s, percent %d, prob %s", key, process_cnt[key], coop_cnt[key], int(coop_cnt[key]/process_cnt[key] * 100) if process_cnt[key] != 0 else 0, process_prob[key])
| 2,124 | 0 | 68 |
bfc3bd37a4aa95cb8512234de7cfe550577160db | 1,136 | py | Python | setup.py | shufps/pdfforms | 9e6eb9c63ab5c743580805942747a5445e79e9c7 | [
"MIT"
] | 1 | 2020-07-03T19:26:40.000Z | 2020-07-03T19:26:40.000Z | setup.py | shufps/pdfforms | 9e6eb9c63ab5c743580805942747a5445e79e9c7 | [
"MIT"
] | null | null | null | setup.py | shufps/pdfforms | 9e6eb9c63ab5c743580805942747a5445e79e9c7 | [
"MIT"
] | null | null | null | from os.path import join, dirname
from setuptools import setup
package_name = "pdfforms"
package_version = "1.2.1"
base_dir = dirname(__file__)
setup(
name = package_name,
version = package_version,
description = "Populate fillable pdf forms from csv data file",
long_description = read("README.rst") + '\n\n' + read("CHANGELOG.rst"),
author = "Aryeh Leib Taurog",
author_email = "python@aryehleib.com",
license = 'MIT',
url = "https://github.com/altaurog/pdfforms",
packages = [package_name],
entry_points = {'console_scripts': ['pdfforms=pdfforms.pdfforms:main']},
classifiers = [
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Development Status :: 5 - Production/Stable",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Intended Audience :: End Users/Desktop",
"Environment :: Console",
"Topic :: Office/Business",
],
)
| 32.457143 | 76 | 0.636444 | from os.path import join, dirname
from setuptools import setup
package_name = "pdfforms"
package_version = "1.2.1"
base_dir = dirname(__file__)
def read(filename):
f = open(join(base_dir, filename))
return f.read()
setup(
name = package_name,
version = package_version,
description = "Populate fillable pdf forms from csv data file",
long_description = read("README.rst") + '\n\n' + read("CHANGELOG.rst"),
author = "Aryeh Leib Taurog",
author_email = "python@aryehleib.com",
license = 'MIT',
url = "https://github.com/altaurog/pdfforms",
packages = [package_name],
entry_points = {'console_scripts': ['pdfforms=pdfforms.pdfforms:main']},
classifiers = [
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Development Status :: 5 - Production/Stable",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Intended Audience :: End Users/Desktop",
"Environment :: Console",
"Topic :: Office/Business",
],
)
| 57 | 0 | 23 |
16e99ac5970e25e2230ef1f24565fa74e8d93890 | 31,772 | py | Python | libs/itsdangerous.py | bbondy/brianbondy.gae | 5c189e5d8f1ee0fdc77ab48c21f3da2c9e3f246c | [
"MIT"
] | null | null | null | libs/itsdangerous.py | bbondy/brianbondy.gae | 5c189e5d8f1ee0fdc77ab48c21f3da2c9e3f246c | [
"MIT"
] | null | null | null | libs/itsdangerous.py | bbondy/brianbondy.gae | 5c189e5d8f1ee0fdc77ab48c21f3da2c9e3f246c | [
"MIT"
] | 1 | 2020-10-01T16:36:12.000Z | 2020-10-01T16:36:12.000Z | # -*- coding: utf-8 -*-
"""
itsdangerous
~~~~~~~~~~~~
A module that implements various functions to deal with untrusted
sources. Mainly useful for web applications.
:copyright: (c) 2011 by Armin Ronacher and the Django Software Foundation.
:license: BSD, see LICENSE for more details.
"""
import sys
import hmac
import zlib
import time
import base64
import hashlib
import operator
from datetime import datetime
PY2 = sys.version_info[0] == 2
if PY2:
from itertools import izip
text_type = unicode
int_to_byte = chr
number_types = (int, long, float)
else:
from functools import reduce
izip = zip
text_type = str
int_to_byte = operator.methodcaller('to_bytes', 1, 'big')
number_types = (int, float)
try:
import simplejson as json
except ImportError:
import json
class _CompactJSON(object):
"""Wrapper around simplejson that strips whitespace.
"""
compact_json = _CompactJSON()
# 2011/01/01 in UTC
EPOCH = 1293840000
def is_text_serializer(serializer):
"""Checks wheather a serializer generates text or binary."""
return isinstance(serializer.dumps({}), text_type)
# Starting with 3.3 the standard library has a c-implementation for
# constant time string compares.
_builtin_constant_time_compare = getattr(hmac, 'compare_digest', None)
def constant_time_compare(val1, val2):
"""Returns True if the two strings are equal, False otherwise.
The time taken is independent of the number of characters that match. Do
not use this function for anything else than comparision with known
length targets.
This is should be implemented in C in order to get it completely right.
"""
if _builtin_constant_time_compare is not None:
return _builtin_constant_time_compare(val1, val2)
len_eq = len(val1) == len(val2)
if len_eq:
result = 0
left = val1
else:
result = 1
left = val2
for x, y in izip(bytearray(left), bytearray(val2)):
result |= x ^ y
return result == 0
class BadData(Exception):
"""Raised if bad data of any sort was encountered. This is the
base for all exceptions that itsdangerous is currently using.
.. versionadded:: 0.15
"""
message = None
if PY2:
__unicode__ = __str__
class BadPayload(BadData):
"""This error is raised in situations when payload is loaded without
checking the signature first and an exception happend as a result of
that. The original exception that caused that will be stored on the
exception as :attr:`original_error`.
This can also happen with a :class:`JSONWebSignatureSerializer` that
is subclassed and uses a different serializer for the payload than
the expected one.
.. versionadded:: 0.15
"""
class BadSignature(BadData):
"""This error is raised if a signature does not match. As of
itsdangerous 0.14 there are helpful attributes on the exception
instances. You can also catch down the baseclass :exc:`BadData`.
"""
class BadTimeSignature(BadSignature):
"""Raised for time based signatures that fail. This is a subclass
of :class:`BadSignature` so you can catch those down as well.
"""
class BadHeader(BadSignature):
"""Raised if a signed header is invalid in some form. This only
happens for serializers that have a header that goes with the
signature.
.. versionadded:: 0.24
"""
class SignatureExpired(BadTimeSignature):
"""Signature timestamp is older than required max_age. This is a
subclass of :exc:`BadTimeSignature` so you can use the baseclass for
catching the error.
"""
def base64_encode(string):
"""base64 encodes a single bytestring (and is tolerant to getting
called with a unicode string).
The resulting bytestring is safe for putting into URLs.
"""
string = want_bytes(string)
return base64.urlsafe_b64encode(string).strip(b'=')
def base64_decode(string):
"""base64 decodes a single bytestring (and is tolerant to getting
called with a unicode string).
The result is also a bytestring.
"""
string = want_bytes(string, encoding='ascii', errors='ignore')
return base64.urlsafe_b64decode(string + b'=' * (-len(string) % 4))
class SigningAlgorithm(object):
"""Subclasses of `SigningAlgorithm` have to implement `get_signature` to
provide signature generation functionality.
"""
def get_signature(self, key, value):
"""Returns the signature for the given key and value"""
raise NotImplementedError()
def verify_signature(self, key, value, sig):
"""Verifies the given signature matches the expected signature"""
return constant_time_compare(sig, self.get_signature(key, value))
class NoneAlgorithm(SigningAlgorithm):
"""This class provides a algorithm that does not perform any signing and
returns an empty signature.
"""
class HMACAlgorithm(SigningAlgorithm):
"""This class provides signature generation using HMACs."""
#: The digest method to use with the MAC algorithm. This defaults to sha1
#: but can be changed for any other function in the hashlib module.
default_digest_method = staticmethod(hashlib.sha1)
class Signer(object):
"""This class can sign bytes and unsign it and validate the signature
provided.
Salt can be used to namespace the hash, so that a signed string is only
valid for a given namespace. Leaving this at the default value or re-using
a salt value across different parts of your application where the same
signed value in one part can mean something different in another part
is a security risk.
See :ref:`the-salt` for an example of what the salt is doing and how you
can utilize it.
.. versionadded:: 0.14
`key_derivation` and `digest_method` were added as arguments to the
class constructor.
.. versionadded:: 0.18
`algorithm` was added as an argument to the class constructor.
"""
#: The digest method to use for the signer. This defaults to sha1 but can
#: be changed for any other function in the hashlib module.
#:
#: .. versionchanged:: 0.14
default_digest_method = staticmethod(hashlib.sha1)
#: Controls how the key is derived. The default is Django style
#: concatenation. Possible values are ``concat``, ``django-concat``
#: and ``hmac``. This is used for deriving a key from the secret key
#: with an added salt.
#:
#: .. versionadded:: 0.14
default_key_derivation = 'django-concat'
def derive_key(self):
"""This method is called to derive the key. If you're unhappy with
the default key derivation choices you can override them here.
Keep in mind that the key derivation in itsdangerous is not intended
to be used as a security method to make a complex key out of a short
password. Instead you should use large random secret keys.
"""
salt = want_bytes(self.salt)
if self.key_derivation == 'concat':
return self.digest_method(salt + self.secret_key).digest()
elif self.key_derivation == 'django-concat':
return self.digest_method(salt + b'signer' +
self.secret_key).digest()
elif self.key_derivation == 'hmac':
mac = hmac.new(self.secret_key, digestmod=self.digest_method)
mac.update(salt)
return mac.digest()
elif self.key_derivation == 'none':
return self.secret_key
else:
raise TypeError('Unknown key derivation method')
def get_signature(self, value):
"""Returns the signature for the given value"""
value = want_bytes(value)
key = self.derive_key()
sig = self.algorithm.get_signature(key, value)
return base64_encode(sig)
def sign(self, value):
"""Signs the given string."""
return value + want_bytes(self.sep) + self.get_signature(value)
def verify_signature(self, value, sig):
"""Verifies the signature for the given value."""
key = self.derive_key()
sig = base64_decode(sig)
return self.algorithm.verify_signature(key, value, sig)
def unsign(self, signed_value):
"""Unsigns the given string."""
signed_value = want_bytes(signed_value)
sep = want_bytes(self.sep)
if sep not in signed_value:
raise BadSignature('No %r found in value' % self.sep)
value, sig = signed_value.rsplit(sep, 1)
if self.verify_signature(value, sig):
return value
raise BadSignature('Signature %r does not match' % sig,
payload=value)
def validate(self, signed_value):
"""Just validates the given signed value. Returns `True` if the
signature exists and is valid, `False` otherwise."""
try:
self.unsign(signed_value)
return True
except BadSignature:
return False
class TimestampSigner(Signer):
"""Works like the regular :class:`Signer` but also records the time
of the signing and can be used to expire signatures. The unsign
method can rause a :exc:`SignatureExpired` method if the unsigning
failed because the signature is expired. This exception is a subclass
of :exc:`BadSignature`.
"""
def get_timestamp(self):
"""Returns the current timestamp. This implementation returns the
seconds since 1/1/2011. The function must return an integer.
"""
return int(time.time() - EPOCH)
def timestamp_to_datetime(self, ts):
"""Used to convert the timestamp from `get_timestamp` into a
datetime object.
"""
return datetime.utcfromtimestamp(ts + EPOCH)
def sign(self, value):
"""Signs the given string and also attaches a time information."""
value = want_bytes(value)
timestamp = base64_encode(int_to_bytes(self.get_timestamp()))
sep = want_bytes(self.sep)
value = value + sep + timestamp
return value + sep + self.get_signature(value)
def unsign(self, value, max_age=None, return_timestamp=False):
"""Works like the regular :meth:`~Signer.unsign` but can also
validate the time. See the base docstring of the class for
the general behavior. If `return_timestamp` is set to `True`
the timestamp of the signature will be returned as naive
:class:`datetime.datetime` object in UTC.
"""
try:
result = Signer.unsign(self, value)
sig_error = None
except BadSignature as e:
sig_error = e
result = e.payload or b''
sep = want_bytes(self.sep)
# If there is no timestamp in the result there is something
# seriously wrong. In case there was a signature error, we raise
# that one directly, otherwise we have a weird situation in which
# we shouldn't have come except someone uses a time-based serializer
# on non-timestamp data, so catch that.
if not sep in result:
if sig_error:
raise sig_error
raise BadTimeSignature('timestamp missing', payload=result)
value, timestamp = result.rsplit(sep, 1)
try:
timestamp = bytes_to_int(base64_decode(timestamp))
except Exception:
timestamp = None
# Signature is *not* okay. Raise a proper error now that we have
# split the value and the timestamp.
if sig_error is not None:
raise BadTimeSignature(text_type(sig_error), payload=value,
date_signed=timestamp)
# Signature was okay but the timestamp is actually not there or
# malformed. Should not happen, but well. We handle it nonetheless
if timestamp is None:
raise BadTimeSignature('Malformed timestamp', payload=value)
# Check timestamp is not older than max_age
if max_age is not None:
age = self.get_timestamp() - timestamp
if age > max_age:
raise SignatureExpired(
'Signature age %s > %s seconds' % (age, max_age),
payload=value,
date_signed=self.timestamp_to_datetime(timestamp))
if return_timestamp:
return value, self.timestamp_to_datetime(timestamp)
return value
def validate(self, signed_value, max_age=None):
"""Just validates the given signed value. Returns `True` if the
signature exists and is valid, `False` otherwise."""
try:
self.unsign(signed_value, max_age=max_age)
return True
except BadSignature:
return False
class Serializer(object):
"""This class provides a serialization interface on top of the
signer. It provides a similar API to json/pickle and other modules but is
slightly differently structured internally. If you want to change the
underlying implementation for parsing and loading you have to override the
:meth:`load_payload` and :meth:`dump_payload` functions.
This implementation uses simplejson if available for dumping and loading
and will fall back to the standard library's json module if it's not
available.
Starting with 0.14 you do not need to subclass this class in order to
switch out or customer the :class:`Signer`. You can instead also pass a
different class to the constructor as well as keyword arguments as
dictionary that should be forwarded::
s = Serializer(signer_kwargs={'key_derivation': 'hmac'})
.. versionchanged:: 0.14:
The `signer` and `signer_kwargs` parameters were added to the
constructor.
"""
#: If a serializer module or class is not passed to the constructor
#: this one is picked up. This currently defaults to :mod:`json`.
default_serializer = json
#: The default :class:`Signer` class that is being used by this
#: serializer.
#:
#: .. versionadded:: 0.14
default_signer = Signer
def load_payload(self, payload, serializer=None):
"""Loads the encoded object. This function raises :class:`BadPayload`
if the payload is not valid. The `serializer` parameter can be used to
override the serializer stored on the class. The encoded payload is
always byte based.
"""
if serializer is None:
serializer = self.serializer
is_text = self.is_text_serializer
else:
is_text = is_text_serializer(serializer)
try:
if is_text:
payload = payload.decode('utf-8')
return serializer.loads(payload)
except Exception as e:
raise BadPayload('Could not load the payload because an '
'exception occurred on unserializing the data',
original_error=e)
def dump_payload(self, obj):
"""Dumps the encoded object. The return value is always a
bytestring. If the internal serializer is text based the value
will automatically be encoded to utf-8.
"""
return want_bytes(self.serializer.dumps(obj))
def make_signer(self, salt=None):
"""A method that creates a new instance of the signer to be used.
The default implementation uses the :class:`Signer` baseclass.
"""
if salt is None:
salt = self.salt
return self.signer(self.secret_key, salt=salt, **self.signer_kwargs)
def dumps(self, obj, salt=None):
"""Returns a signed string serialized with the internal serializer.
The return value can be either a byte or unicode string depending
on the format of the internal serializer.
"""
payload = want_bytes(self.dump_payload(obj))
rv = self.make_signer(salt).sign(payload)
if self.is_text_serializer:
rv = rv.decode('utf-8')
return rv
def dump(self, obj, f, salt=None):
"""Like :meth:`dumps` but dumps into a file. The file handle has
to be compatible with what the internal serializer expects.
"""
f.write(self.dumps(obj, salt))
def loads(self, s, salt=None):
"""Reverse of :meth:`dumps`, raises :exc:`BadSignature` if the
signature validation fails.
"""
s = want_bytes(s)
return self.load_payload(self.make_signer(salt).unsign(s))
def load(self, f, salt=None):
"""Like :meth:`loads` but loads from a file."""
return self.loads(f.read(), salt)
def loads_unsafe(self, s, salt=None):
"""Like :meth:`loads` but without verifying the signature. This is
potentially very dangerous to use depending on how your serializer
works. The return value is ``(signature_okay, payload)`` instead of
just the payload. The first item will be a boolean that indicates
if the signature is okay (``True``) or if it failed. This function
never fails.
Use it for debugging only and if you know that your serializer module
is not exploitable (eg: do not use it with a pickle serializer).
.. versionadded:: 0.15
"""
return self._loads_unsafe_impl(s, salt)
def _loads_unsafe_impl(self, s, salt, load_kwargs=None,
load_payload_kwargs=None):
"""Lowlevel helper function to implement :meth:`loads_unsafe` in
serializer subclasses.
"""
try:
return True, self.loads(s, salt=salt, **(load_kwargs or {}))
except BadSignature as e:
if e.payload is None:
return False, None
try:
return False, self.load_payload(e.payload,
**(load_payload_kwargs or {}))
except BadPayload:
return False, None
def load_unsafe(self, f, *args, **kwargs):
"""Like :meth:`loads_unsafe` but loads from a file.
.. versionadded:: 0.15
"""
return self.loads_unsafe(f.read(), *args, **kwargs)
class TimedSerializer(Serializer):
"""Uses the :class:`TimestampSigner` instead of the default
:meth:`Signer`.
"""
default_signer = TimestampSigner
def loads(self, s, max_age=None, return_timestamp=False, salt=None):
"""Reverse of :meth:`dumps`, raises :exc:`BadSignature` if the
signature validation fails. If a `max_age` is provided it will
ensure the signature is not older than that time in seconds. In
case the signature is outdated, :exc:`SignatureExpired` is raised
which is a subclass of :exc:`BadSignature`. All arguments are
forwarded to the signer's :meth:`~TimestampSigner.unsign` method.
"""
base64d, timestamp = self.make_signer(salt) \
.unsign(s, max_age, return_timestamp=True)
payload = self.load_payload(base64d)
if return_timestamp:
return payload, timestamp
return payload
class JSONWebSignatureSerializer(Serializer):
"""This serializer implements JSON Web Signature (JWS) support. Only
supports the JWS Compact Serialization.
"""
jws_algorithms = {
'HS256': HMACAlgorithm(hashlib.sha256),
'HS384': HMACAlgorithm(hashlib.sha384),
'HS512': HMACAlgorithm(hashlib.sha512),
'none': NoneAlgorithm(),
}
#: The default algorithm to use for signature generation
default_algorithm = 'HS256'
default_serializer = compact_json
def dumps(self, obj, salt=None, header_fields=None):
"""Like :meth:`~Serializer.dumps` but creates a JSON Web Signature. It
also allows for specifying additional fields to be included in the JWS
Header.
"""
header = self.make_header(header_fields)
signer = self.make_signer(salt, self.algorithm)
return signer.sign(self.dump_payload(header, obj))
def loads(self, s, salt=None, return_header=False):
"""Reverse of :meth:`dumps`. If requested via `return_header` it will
return a tuple of payload and header.
"""
payload, header = self.load_payload(
self.make_signer(salt, self.algorithm).unsign(want_bytes(s)),
return_header=True)
if header.get('alg') != self.algorithm_name:
raise BadHeader('Algorithm mismatch', header=header,
payload=payload)
if return_header:
return payload, header
return payload
class TimedJSONWebSignatureSerializer(JSONWebSignatureSerializer):
"""Works like the regular :class:`JSONWebSignatureSerializer` but also
records the time of the signing and can be used to expire signatures.
JWS currently does not specify this behavior but it mentions a possibility
extension like this in the spec. Expiry date is encoded into the header
similarily as specified in `draft-ietf-oauth-json-web-token
<http://self-issued.info/docs/draft-ietf-oauth-json-web-token.html#expDef`_.
The unsign method can raise a :exc:`SignatureExpired` method if the
unsigning failed because the signature is expired. This exception is a
subclass of :exc:`BadSignature`.
"""
DEFAULT_EXPIRES_IN = 3600
class URLSafeSerializerMixin(object):
"""Mixed in with a regular serializer it will attempt to zlib compress
the string to make it shorter if necessary. It will also base64 encode
the string so that it can safely be placed in a URL.
"""
class URLSafeSerializer(URLSafeSerializerMixin, Serializer):
"""Works like :class:`Serializer` but dumps and loads into a URL
safe string consisting of the upper and lowercase character of the
alphabet as well as ``'_'``, ``'-'`` and ``'.'``.
"""
default_serializer = compact_json
class URLSafeTimedSerializer(URLSafeSerializerMixin, TimedSerializer):
"""Works like :class:`TimedSerializer` but dumps and loads into a URL
safe string consisting of the upper and lowercase character of the
alphabet as well as ``'_'``, ``'-'`` and ``'.'``.
"""
default_serializer = compact_json
| 36.51954 | 81 | 0.647237 | # -*- coding: utf-8 -*-
"""
itsdangerous
~~~~~~~~~~~~
A module that implements various functions to deal with untrusted
sources. Mainly useful for web applications.
:copyright: (c) 2011 by Armin Ronacher and the Django Software Foundation.
:license: BSD, see LICENSE for more details.
"""
import sys
import hmac
import zlib
import time
import base64
import hashlib
import operator
from datetime import datetime
PY2 = sys.version_info[0] == 2
if PY2:
from itertools import izip
text_type = unicode
int_to_byte = chr
number_types = (int, long, float)
else:
from functools import reduce
izip = zip
text_type = str
int_to_byte = operator.methodcaller('to_bytes', 1, 'big')
number_types = (int, float)
try:
import simplejson as json
except ImportError:
import json
class _CompactJSON(object):
"""Wrapper around simplejson that strips whitespace.
"""
def loads(self, payload):
return json.loads(payload)
def dumps(self, obj):
return json.dumps(obj, separators=(',', ':'))
compact_json = _CompactJSON()
# 2011/01/01 in UTC
EPOCH = 1293840000
def want_bytes(s, encoding='utf-8', errors='strict'):
if isinstance(s, text_type):
s = s.encode(encoding, errors)
return s
def is_text_serializer(serializer):
"""Checks wheather a serializer generates text or binary."""
return isinstance(serializer.dumps({}), text_type)
# Starting with 3.3 the standard library has a c-implementation for
# constant time string compares.
_builtin_constant_time_compare = getattr(hmac, 'compare_digest', None)
def constant_time_compare(val1, val2):
"""Returns True if the two strings are equal, False otherwise.
The time taken is independent of the number of characters that match. Do
not use this function for anything else than comparision with known
length targets.
This is should be implemented in C in order to get it completely right.
"""
if _builtin_constant_time_compare is not None:
return _builtin_constant_time_compare(val1, val2)
len_eq = len(val1) == len(val2)
if len_eq:
result = 0
left = val1
else:
result = 1
left = val2
for x, y in izip(bytearray(left), bytearray(val2)):
result |= x ^ y
return result == 0
class BadData(Exception):
"""Raised if bad data of any sort was encountered. This is the
base for all exceptions that itsdangerous is currently using.
.. versionadded:: 0.15
"""
message = None
def __init__(self, message):
Exception.__init__(self, message)
self.message = message
def __str__(self):
return text_type(self.message)
if PY2:
__unicode__ = __str__
def __str__(self):
return self.__unicode__().encode('utf-8')
class BadPayload(BadData):
"""This error is raised in situations when payload is loaded without
checking the signature first and an exception happend as a result of
that. The original exception that caused that will be stored on the
exception as :attr:`original_error`.
This can also happen with a :class:`JSONWebSignatureSerializer` that
is subclassed and uses a different serializer for the payload than
the expected one.
.. versionadded:: 0.15
"""
def __init__(self, message, original_error=None):
BadData.__init__(self, message)
#: If available, the error that indicates why the payload
#: was not valid. This might be `None`.
self.original_error = original_error
class BadSignature(BadData):
"""This error is raised if a signature does not match. As of
itsdangerous 0.14 there are helpful attributes on the exception
instances. You can also catch down the baseclass :exc:`BadData`.
"""
def __init__(self, message, payload=None):
BadData.__init__(self, message)
#: The payload that failed the signature test. In some
#: situations you might still want to inspect this, even if
#: you know it was tampered with.
#:
#: .. versionadded:: 0.14
self.payload = payload
class BadTimeSignature(BadSignature):
"""Raised for time based signatures that fail. This is a subclass
of :class:`BadSignature` so you can catch those down as well.
"""
def __init__(self, message, payload=None, date_signed=None):
BadSignature.__init__(self, message, payload)
#: If the signature expired this exposes the date of when the
#: signature was created. This can be helpful in order to
#: tell the user how long a link has been gone stale.
#:
#: .. versionadded:: 0.14
self.date_signed = date_signed
class BadHeader(BadSignature):
"""Raised if a signed header is invalid in some form. This only
happens for serializers that have a header that goes with the
signature.
.. versionadded:: 0.24
"""
def __init__(self, message, payload=None, header=None,
original_error=None):
BadSignature.__init__(self, message, payload)
#: If the header is actually available but just malformed it
#: might be stored here.
self.header = header
#: If available, the error that indicates why the payload
#: was not valid. This might be `None`.
self.original_error = original_error
class SignatureExpired(BadTimeSignature):
"""Signature timestamp is older than required max_age. This is a
subclass of :exc:`BadTimeSignature` so you can use the baseclass for
catching the error.
"""
def base64_encode(string):
"""base64 encodes a single bytestring (and is tolerant to getting
called with a unicode string).
The resulting bytestring is safe for putting into URLs.
"""
string = want_bytes(string)
return base64.urlsafe_b64encode(string).strip(b'=')
def base64_decode(string):
"""base64 decodes a single bytestring (and is tolerant to getting
called with a unicode string).
The result is also a bytestring.
"""
string = want_bytes(string, encoding='ascii', errors='ignore')
return base64.urlsafe_b64decode(string + b'=' * (-len(string) % 4))
def int_to_bytes(num):
assert num >= 0
rv = []
while num:
rv.append(int_to_byte(num & 0xff))
num >>= 8
return b''.join(reversed(rv))
def bytes_to_int(bytestr):
return reduce(lambda a, b: a << 8 | b, bytearray(bytestr), 0)
class SigningAlgorithm(object):
"""Subclasses of `SigningAlgorithm` have to implement `get_signature` to
provide signature generation functionality.
"""
def get_signature(self, key, value):
"""Returns the signature for the given key and value"""
raise NotImplementedError()
def verify_signature(self, key, value, sig):
"""Verifies the given signature matches the expected signature"""
return constant_time_compare(sig, self.get_signature(key, value))
class NoneAlgorithm(SigningAlgorithm):
"""This class provides a algorithm that does not perform any signing and
returns an empty signature.
"""
def get_signature(self, key, value):
return b''
class HMACAlgorithm(SigningAlgorithm):
"""This class provides signature generation using HMACs."""
#: The digest method to use with the MAC algorithm. This defaults to sha1
#: but can be changed for any other function in the hashlib module.
default_digest_method = staticmethod(hashlib.sha1)
def __init__(self, digest_method=None):
if digest_method is None:
digest_method = self.default_digest_method
self.digest_method = digest_method
def get_signature(self, key, value):
mac = hmac.new(key, msg=value, digestmod=self.digest_method)
return mac.digest()
class Signer(object):
"""This class can sign bytes and unsign it and validate the signature
provided.
Salt can be used to namespace the hash, so that a signed string is only
valid for a given namespace. Leaving this at the default value or re-using
a salt value across different parts of your application where the same
signed value in one part can mean something different in another part
is a security risk.
See :ref:`the-salt` for an example of what the salt is doing and how you
can utilize it.
.. versionadded:: 0.14
`key_derivation` and `digest_method` were added as arguments to the
class constructor.
.. versionadded:: 0.18
`algorithm` was added as an argument to the class constructor.
"""
#: The digest method to use for the signer. This defaults to sha1 but can
#: be changed for any other function in the hashlib module.
#:
#: .. versionchanged:: 0.14
default_digest_method = staticmethod(hashlib.sha1)
#: Controls how the key is derived. The default is Django style
#: concatenation. Possible values are ``concat``, ``django-concat``
#: and ``hmac``. This is used for deriving a key from the secret key
#: with an added salt.
#:
#: .. versionadded:: 0.14
default_key_derivation = 'django-concat'
def __init__(self, secret_key, salt=None, sep='.', key_derivation=None,
digest_method=None, algorithm=None):
self.secret_key = want_bytes(secret_key)
self.sep = sep
self.salt = 'itsdangerous.Signer' if salt is None else salt
if key_derivation is None:
key_derivation = self.default_key_derivation
self.key_derivation = key_derivation
if digest_method is None:
digest_method = self.default_digest_method
self.digest_method = digest_method
if algorithm is None:
algorithm = HMACAlgorithm(self.digest_method)
self.algorithm = algorithm
def derive_key(self):
"""This method is called to derive the key. If you're unhappy with
the default key derivation choices you can override them here.
Keep in mind that the key derivation in itsdangerous is not intended
to be used as a security method to make a complex key out of a short
password. Instead you should use large random secret keys.
"""
salt = want_bytes(self.salt)
if self.key_derivation == 'concat':
return self.digest_method(salt + self.secret_key).digest()
elif self.key_derivation == 'django-concat':
return self.digest_method(salt + b'signer' +
self.secret_key).digest()
elif self.key_derivation == 'hmac':
mac = hmac.new(self.secret_key, digestmod=self.digest_method)
mac.update(salt)
return mac.digest()
elif self.key_derivation == 'none':
return self.secret_key
else:
raise TypeError('Unknown key derivation method')
def get_signature(self, value):
"""Returns the signature for the given value"""
value = want_bytes(value)
key = self.derive_key()
sig = self.algorithm.get_signature(key, value)
return base64_encode(sig)
def sign(self, value):
"""Signs the given string."""
return value + want_bytes(self.sep) + self.get_signature(value)
def verify_signature(self, value, sig):
"""Verifies the signature for the given value."""
key = self.derive_key()
sig = base64_decode(sig)
return self.algorithm.verify_signature(key, value, sig)
def unsign(self, signed_value):
"""Unsigns the given string."""
signed_value = want_bytes(signed_value)
sep = want_bytes(self.sep)
if sep not in signed_value:
raise BadSignature('No %r found in value' % self.sep)
value, sig = signed_value.rsplit(sep, 1)
if self.verify_signature(value, sig):
return value
raise BadSignature('Signature %r does not match' % sig,
payload=value)
def validate(self, signed_value):
"""Just validates the given signed value. Returns `True` if the
signature exists and is valid, `False` otherwise."""
try:
self.unsign(signed_value)
return True
except BadSignature:
return False
class TimestampSigner(Signer):
"""Works like the regular :class:`Signer` but also records the time
of the signing and can be used to expire signatures. The unsign
method can rause a :exc:`SignatureExpired` method if the unsigning
failed because the signature is expired. This exception is a subclass
of :exc:`BadSignature`.
"""
def get_timestamp(self):
"""Returns the current timestamp. This implementation returns the
seconds since 1/1/2011. The function must return an integer.
"""
return int(time.time() - EPOCH)
def timestamp_to_datetime(self, ts):
"""Used to convert the timestamp from `get_timestamp` into a
datetime object.
"""
return datetime.utcfromtimestamp(ts + EPOCH)
def sign(self, value):
"""Signs the given string and also attaches a time information."""
value = want_bytes(value)
timestamp = base64_encode(int_to_bytes(self.get_timestamp()))
sep = want_bytes(self.sep)
value = value + sep + timestamp
return value + sep + self.get_signature(value)
def unsign(self, value, max_age=None, return_timestamp=False):
"""Works like the regular :meth:`~Signer.unsign` but can also
validate the time. See the base docstring of the class for
the general behavior. If `return_timestamp` is set to `True`
the timestamp of the signature will be returned as naive
:class:`datetime.datetime` object in UTC.
"""
try:
result = Signer.unsign(self, value)
sig_error = None
except BadSignature as e:
sig_error = e
result = e.payload or b''
sep = want_bytes(self.sep)
# If there is no timestamp in the result there is something
# seriously wrong. In case there was a signature error, we raise
# that one directly, otherwise we have a weird situation in which
# we shouldn't have come except someone uses a time-based serializer
# on non-timestamp data, so catch that.
if not sep in result:
if sig_error:
raise sig_error
raise BadTimeSignature('timestamp missing', payload=result)
value, timestamp = result.rsplit(sep, 1)
try:
timestamp = bytes_to_int(base64_decode(timestamp))
except Exception:
timestamp = None
# Signature is *not* okay. Raise a proper error now that we have
# split the value and the timestamp.
if sig_error is not None:
raise BadTimeSignature(text_type(sig_error), payload=value,
date_signed=timestamp)
# Signature was okay but the timestamp is actually not there or
# malformed. Should not happen, but well. We handle it nonetheless
if timestamp is None:
raise BadTimeSignature('Malformed timestamp', payload=value)
# Check timestamp is not older than max_age
if max_age is not None:
age = self.get_timestamp() - timestamp
if age > max_age:
raise SignatureExpired(
'Signature age %s > %s seconds' % (age, max_age),
payload=value,
date_signed=self.timestamp_to_datetime(timestamp))
if return_timestamp:
return value, self.timestamp_to_datetime(timestamp)
return value
def validate(self, signed_value, max_age=None):
"""Just validates the given signed value. Returns `True` if the
signature exists and is valid, `False` otherwise."""
try:
self.unsign(signed_value, max_age=max_age)
return True
except BadSignature:
return False
class Serializer(object):
"""This class provides a serialization interface on top of the
signer. It provides a similar API to json/pickle and other modules but is
slightly differently structured internally. If you want to change the
underlying implementation for parsing and loading you have to override the
:meth:`load_payload` and :meth:`dump_payload` functions.
This implementation uses simplejson if available for dumping and loading
and will fall back to the standard library's json module if it's not
available.
Starting with 0.14 you do not need to subclass this class in order to
switch out or customer the :class:`Signer`. You can instead also pass a
different class to the constructor as well as keyword arguments as
dictionary that should be forwarded::
s = Serializer(signer_kwargs={'key_derivation': 'hmac'})
.. versionchanged:: 0.14:
The `signer` and `signer_kwargs` parameters were added to the
constructor.
"""
#: If a serializer module or class is not passed to the constructor
#: this one is picked up. This currently defaults to :mod:`json`.
default_serializer = json
#: The default :class:`Signer` class that is being used by this
#: serializer.
#:
#: .. versionadded:: 0.14
default_signer = Signer
def __init__(self, secret_key, salt=b'itsdangerous', serializer=None,
signer=None, signer_kwargs=None):
self.secret_key = want_bytes(secret_key)
self.salt = want_bytes(salt)
if serializer is None:
serializer = self.default_serializer
self.serializer = serializer
self.is_text_serializer = is_text_serializer(serializer)
if signer is None:
signer = self.default_signer
self.signer = signer
self.signer_kwargs = signer_kwargs or {}
def load_payload(self, payload, serializer=None):
"""Loads the encoded object. This function raises :class:`BadPayload`
if the payload is not valid. The `serializer` parameter can be used to
override the serializer stored on the class. The encoded payload is
always byte based.
"""
if serializer is None:
serializer = self.serializer
is_text = self.is_text_serializer
else:
is_text = is_text_serializer(serializer)
try:
if is_text:
payload = payload.decode('utf-8')
return serializer.loads(payload)
except Exception as e:
raise BadPayload('Could not load the payload because an '
'exception occurred on unserializing the data',
original_error=e)
def dump_payload(self, obj):
"""Dumps the encoded object. The return value is always a
bytestring. If the internal serializer is text based the value
will automatically be encoded to utf-8.
"""
return want_bytes(self.serializer.dumps(obj))
def make_signer(self, salt=None):
"""A method that creates a new instance of the signer to be used.
The default implementation uses the :class:`Signer` baseclass.
"""
if salt is None:
salt = self.salt
return self.signer(self.secret_key, salt=salt, **self.signer_kwargs)
def dumps(self, obj, salt=None):
"""Returns a signed string serialized with the internal serializer.
The return value can be either a byte or unicode string depending
on the format of the internal serializer.
"""
payload = want_bytes(self.dump_payload(obj))
rv = self.make_signer(salt).sign(payload)
if self.is_text_serializer:
rv = rv.decode('utf-8')
return rv
def dump(self, obj, f, salt=None):
"""Like :meth:`dumps` but dumps into a file. The file handle has
to be compatible with what the internal serializer expects.
"""
f.write(self.dumps(obj, salt))
def loads(self, s, salt=None):
"""Reverse of :meth:`dumps`, raises :exc:`BadSignature` if the
signature validation fails.
"""
s = want_bytes(s)
return self.load_payload(self.make_signer(salt).unsign(s))
def load(self, f, salt=None):
"""Like :meth:`loads` but loads from a file."""
return self.loads(f.read(), salt)
def loads_unsafe(self, s, salt=None):
"""Like :meth:`loads` but without verifying the signature. This is
potentially very dangerous to use depending on how your serializer
works. The return value is ``(signature_okay, payload)`` instead of
just the payload. The first item will be a boolean that indicates
if the signature is okay (``True``) or if it failed. This function
never fails.
Use it for debugging only and if you know that your serializer module
is not exploitable (eg: do not use it with a pickle serializer).
.. versionadded:: 0.15
"""
return self._loads_unsafe_impl(s, salt)
def _loads_unsafe_impl(self, s, salt, load_kwargs=None,
load_payload_kwargs=None):
"""Lowlevel helper function to implement :meth:`loads_unsafe` in
serializer subclasses.
"""
try:
return True, self.loads(s, salt=salt, **(load_kwargs or {}))
except BadSignature as e:
if e.payload is None:
return False, None
try:
return False, self.load_payload(e.payload,
**(load_payload_kwargs or {}))
except BadPayload:
return False, None
def load_unsafe(self, f, *args, **kwargs):
"""Like :meth:`loads_unsafe` but loads from a file.
.. versionadded:: 0.15
"""
return self.loads_unsafe(f.read(), *args, **kwargs)
class TimedSerializer(Serializer):
"""Uses the :class:`TimestampSigner` instead of the default
:meth:`Signer`.
"""
default_signer = TimestampSigner
def loads(self, s, max_age=None, return_timestamp=False, salt=None):
"""Reverse of :meth:`dumps`, raises :exc:`BadSignature` if the
signature validation fails. If a `max_age` is provided it will
ensure the signature is not older than that time in seconds. In
case the signature is outdated, :exc:`SignatureExpired` is raised
which is a subclass of :exc:`BadSignature`. All arguments are
forwarded to the signer's :meth:`~TimestampSigner.unsign` method.
"""
base64d, timestamp = self.make_signer(salt) \
.unsign(s, max_age, return_timestamp=True)
payload = self.load_payload(base64d)
if return_timestamp:
return payload, timestamp
return payload
def loads_unsafe(self, s, max_age=None, salt=None):
load_kwargs = {'max_age': max_age}
load_payload_kwargs = {}
return self._loads_unsafe_impl(s, salt, load_kwargs, load_payload_kwargs)
class JSONWebSignatureSerializer(Serializer):
"""This serializer implements JSON Web Signature (JWS) support. Only
supports the JWS Compact Serialization.
"""
jws_algorithms = {
'HS256': HMACAlgorithm(hashlib.sha256),
'HS384': HMACAlgorithm(hashlib.sha384),
'HS512': HMACAlgorithm(hashlib.sha512),
'none': NoneAlgorithm(),
}
#: The default algorithm to use for signature generation
default_algorithm = 'HS256'
default_serializer = compact_json
def __init__(self, secret_key, salt=None, serializer=None,
signer=None, signer_kwargs=None, algorithm_name=None):
Serializer.__init__(self, secret_key, salt, serializer,
signer, signer_kwargs)
if algorithm_name is None:
algorithm_name = self.default_algorithm
self.algorithm_name = algorithm_name
self.algorithm = self.make_algorithm(algorithm_name)
def load_payload(self, payload, return_header=False):
payload = want_bytes(payload)
if b'.' not in payload:
raise BadPayload('No "." found in value')
base64d_header, base64d_payload = payload.split(b'.', 1)
try:
json_header = base64_decode(base64d_header)
except Exception as e:
raise BadHeader('Could not base64 decode the header because of '
'an exception', original_error=e)
try:
json_payload = base64_decode(base64d_payload)
except Exception as e:
raise BadPayload('Could not base64 decode the payload because of '
'an exception', original_error=e)
try:
header = Serializer.load_payload(self, json_header,
serializer=json)
except BadData as e:
raise BadHeader('Could not unserialize header because it was '
'malformed', original_error=e)
if not isinstance(header, dict):
raise BadHeader('Header payload is not a JSON object',
header=header)
payload = Serializer.load_payload(self, json_payload)
if return_header:
return payload, header
return payload
def dump_payload(self, header, obj):
base64d_header = base64_encode(self.serializer.dumps(header))
base64d_payload = base64_encode(self.serializer.dumps(obj))
return base64d_header + b'.' + base64d_payload
def make_algorithm(self, algorithm_name):
try:
return self.jws_algorithms[algorithm_name]
except KeyError:
raise NotImplementedError('Algorithm not supported')
def make_signer(self, salt=None, algorithm=None):
if salt is None:
salt = self.salt
key_derivation = 'none' if salt is None else None
if algorithm is None:
algorithm = self.algorithm
return self.signer(self.secret_key, salt=salt, sep='.',
key_derivation=key_derivation, algorithm=algorithm)
def make_header(self, header_fields):
header = header_fields.copy() if header_fields else {}
header['alg'] = self.algorithm_name
return header
def dumps(self, obj, salt=None, header_fields=None):
"""Like :meth:`~Serializer.dumps` but creates a JSON Web Signature. It
also allows for specifying additional fields to be included in the JWS
Header.
"""
header = self.make_header(header_fields)
signer = self.make_signer(salt, self.algorithm)
return signer.sign(self.dump_payload(header, obj))
def loads(self, s, salt=None, return_header=False):
"""Reverse of :meth:`dumps`. If requested via `return_header` it will
return a tuple of payload and header.
"""
payload, header = self.load_payload(
self.make_signer(salt, self.algorithm).unsign(want_bytes(s)),
return_header=True)
if header.get('alg') != self.algorithm_name:
raise BadHeader('Algorithm mismatch', header=header,
payload=payload)
if return_header:
return payload, header
return payload
def loads_unsafe(self, s, salt=None, return_header=False):
kwargs = {'return_header': return_header}
return self._loads_unsafe_impl(s, salt, kwargs, kwargs)
class TimedJSONWebSignatureSerializer(JSONWebSignatureSerializer):
"""Works like the regular :class:`JSONWebSignatureSerializer` but also
records the time of the signing and can be used to expire signatures.
JWS currently does not specify this behavior but it mentions a possibility
extension like this in the spec. Expiry date is encoded into the header
similarily as specified in `draft-ietf-oauth-json-web-token
<http://self-issued.info/docs/draft-ietf-oauth-json-web-token.html#expDef`_.
The unsign method can raise a :exc:`SignatureExpired` method if the
unsigning failed because the signature is expired. This exception is a
subclass of :exc:`BadSignature`.
"""
DEFAULT_EXPIRES_IN = 3600
def __init__(self, secret_key, expires_in=None, **kwargs):
JSONWebSignatureSerializer.__init__(self, secret_key, **kwargs)
if expires_in is None:
expires_in = self.DEFAULT_EXPIRES_IN
self.expires_in = expires_in
def make_header(self, header_fields):
header = JSONWebSignatureSerializer.make_header(self, header_fields)
iat = self.now()
exp = iat + self.expires_in
header['iat'] = iat
header['exp'] = exp
return header
def loads(self, s, salt=None, return_header=False):
payload, header = JSONWebSignatureSerializer.loads(
self, s, salt, return_header=True)
if 'exp' not in header:
raise BadSignature('Missing expiry date', payload=payload)
if not (isinstance(header['exp'], number_types)
and header['exp'] > 0):
raise BadSignature('expiry date is not an IntDate',
payload=payload)
if header['exp'] < self.now():
raise SignatureExpired('Signature expired', payload=payload,
date_signed=self.get_issue_date(header))
if return_header:
return payload, header
return payload
def get_issue_date(self, header):
rv = header.get('iat')
if isinstance(rv, number_types):
return datetime.utcfromtimestamp(int(rv))
def now(self):
return int(time.time())
class URLSafeSerializerMixin(object):
"""Mixed in with a regular serializer it will attempt to zlib compress
the string to make it shorter if necessary. It will also base64 encode
the string so that it can safely be placed in a URL.
"""
def load_payload(self, payload):
decompress = False
if payload.startswith(b'.'):
payload = payload[1:]
decompress = True
try:
json = base64_decode(payload)
except Exception as e:
raise BadPayload('Could not base64 decode the payload because of '
'an exception', original_error=e)
if decompress:
try:
json = zlib.decompress(json)
except Exception as e:
raise BadPayload('Could not zlib decompress the payload before '
'decoding the payload', original_error=e)
return super(URLSafeSerializerMixin, self).load_payload(json)
def dump_payload(self, obj):
json = super(URLSafeSerializerMixin, self).dump_payload(obj)
is_compressed = False
compressed = zlib.compress(json)
if len(compressed) < (len(json) - 1):
json = compressed
is_compressed = True
base64d = base64_encode(json)
if is_compressed:
base64d = b'.' + base64d
return base64d
class URLSafeSerializer(URLSafeSerializerMixin, Serializer):
"""Works like :class:`Serializer` but dumps and loads into a URL
safe string consisting of the upper and lowercase character of the
alphabet as well as ``'_'``, ``'-'`` and ``'.'``.
"""
default_serializer = compact_json
class URLSafeTimedSerializer(URLSafeSerializerMixin, TimedSerializer):
"""Works like :class:`TimedSerializer` but dumps and loads into a URL
safe string consisting of the upper and lowercase character of the
alphabet as well as ``'_'``, ``'-'`` and ``'.'``.
"""
default_serializer = compact_json
| 8,632 | 0 | 855 |
9a937dba4abe57172578dec1a722bd3633ce7c2d | 3,583 | py | Python | events.py | IanHarvey/bluepy.device | be3f5802b7229ce50876bd4bb68b958586731d1b | [
"MIT"
] | 10 | 2017-09-29T00:18:21.000Z | 2022-01-27T09:03:29.000Z | events.py | IanHarvey/bluepy.device | be3f5802b7229ce50876bd4bb68b958586731d1b | [
"MIT"
] | 2 | 2017-09-11T11:07:35.000Z | 2019-04-19T11:02:00.000Z | events.py | IanHarvey/bluepy.device | be3f5802b7229ce50876bd4bb68b958586731d1b | [
"MIT"
] | 5 | 2017-10-10T18:01:02.000Z | 2021-05-24T07:36:52.000Z | import struct
import binascii
import gap
# HCI Events
# Spec V4.0, Vol 2, sec 7.7
E_DISCONN_COMPLETE = 0x05
E_ENCRYPT_CHANGE = 0x08
E_CMD_RESPONSE = 0x0E
E_CMD_STATUS = 0x0F
E_LE_META_EVENT = 0x3E
# LE Meta-event subcodes
# Vol 2, 7.7.65
E_LE_CONN_COMPLETE = 0x01
E_LE_ADVERTISING_REPORT = 0x02
E_LE_CONN_UPDATE_COMPLETE = 0x03
DEFAULT_EVENT_MASK = eventMask([E_DISCONN_COMPLETE, E_ENCRYPT_CHANGE, E_CMD_RESPONSE, E_CMD_STATUS, E_LE_META_EVENT])
DEFAULT_LE_EVENT_MASK = eventMask([E_LE_CONN_COMPLETE, E_LE_CONN_UPDATE_COMPLETE])
# This is basically a mixin to do the event-handling portion of
# the main Device class. Kept separate to aid reuse.
# Stub event handlers
| 33.801887 | 117 | 0.6235 | import struct
import binascii
import gap
# HCI Events
# Spec V4.0, Vol 2, sec 7.7
E_DISCONN_COMPLETE = 0x05
E_ENCRYPT_CHANGE = 0x08
E_CMD_RESPONSE = 0x0E
E_CMD_STATUS = 0x0F
E_LE_META_EVENT = 0x3E
# LE Meta-event subcodes
# Vol 2, 7.7.65
E_LE_CONN_COMPLETE = 0x01
E_LE_ADVERTISING_REPORT = 0x02
E_LE_CONN_UPDATE_COMPLETE = 0x03
def eventMask(evtList):
w=0
for e in evtList:
assert(e > 0)
w |= (1 << (e-1))
return w
DEFAULT_EVENT_MASK = eventMask([E_DISCONN_COMPLETE, E_ENCRYPT_CHANGE, E_CMD_RESPONSE, E_CMD_STATUS, E_LE_META_EVENT])
DEFAULT_LE_EVENT_MASK = eventMask([E_LE_CONN_COMPLETE, E_LE_CONN_UPDATE_COMPLETE])
class EventHandler:
# This is basically a mixin to do the event-handling portion of
# the main Device class. Kept separate to aid reuse.
def onEventReceived(self, data):
eventCode = data[0]
dlen = data[1]
if len(data) != dlen+2:
print ("Invalid length %d in packet: %s" % (dlen, binascii.b2a_hex(data).decode('ascii')))
return
if eventCode == E_CMD_RESPONSE:
(n_cmds, opcode) = struct.unpack("<BH", data[2:5])
return self.onCommandResponse(n_cmds, opcode, data[5:])
elif eventCode == E_LE_META_EVENT:
subEvent = data[2]
if subEvent == E_LE_CONN_COMPLETE:
(status, handle, role, peerAddrType, peerAddr, interval,
latency, timeout, masterClock) = struct.unpack("<BHBB6sHHHB", data[3:])
if status != 0:
return self.onConnectionFailed(status, peerAddrType, peerAddr)
elif role == 0x00:
return self.onMasterConnected(handle, peerAddrType, peerAddr)
elif role == 0x01:
return self.onSlaveConnected(handle, peerAddrType, peerAddr)
elif subEvent == E_LE_ADVERTISING_REPORT:
n_reports = data[3]
pos = 4
for i in range(n_reports):
report = AdvertisingReport()
pos = report.parseData(data, pos)
self.onAdvertisingReport(report)
return
elif eventCode == E_DISCONN_COMPLETE:
(status, handle, reason) = struct.unpack("<BHB", data[2:])
return self.onDisconnect(status, handle, reason)
print ("Unhandled event %02X" % eventCode)
# Stub event handlers
def onCommandResponse(self, n_cmds, opcode, params):
pass
def onConnectionFailed(self, status, peerAddrType, peerAddr):
pass
def onMasterConnected(self, handle, peerAddrType, peerAddr):
pass
def onSlaveConnected(self, handle, peerAddrType, peerAddr):
pass
def onDisconnect(self, status, handle, reason):
pass
def onAdvertisingReport(self, report):
pass
class AdvertisingReport:
def __init__(self):
self.event_type = self.address_type = self.address = self.gap_data = None
def parseData(self, data, pos):
hdr = struct.unpack("<BB6sB", data[pos:pos+9])
(self.event_type, self.address_type, self.address, datalen) = hdr
self.gap_data = data[pos+9:pos+9+datalen]
self.RSSI = struct.unpack("<b", data[pos+9+datalen:pos+10+datalen])[0]
self.adv_data = gap.AdvertisingData(self.gap_data)
return pos+10+datalen
def __str__(self):
return "type=%02X addrtype=%02X addr=%r RSSI=%d adv=%s" % (
self.event_type, self.address_type, self.address, self.RSSI, self.adv_data)
| 2,553 | 1 | 337 |
72645c0f047f889967cb8ede55344c574ef90c92 | 668 | py | Python | harvest-images.py | BinaryDragonStudios/ArenaGame | 0162568b64599a9fccab64e3cd08d8295885b32c | [
"MIT"
] | null | null | null | harvest-images.py | BinaryDragonStudios/ArenaGame | 0162568b64599a9fccab64e3cd08d8295885b32c | [
"MIT"
] | null | null | null | harvest-images.py | BinaryDragonStudios/ArenaGame | 0162568b64599a9fccab64e3cd08d8295885b32c | [
"MIT"
] | null | null | null | #!/usr/bin/env python
# Use the built in json and sqlite library
import sqlite3
# Boilerplate python
if __name__ == '__main__':
main()
| 27.833333 | 98 | 0.652695 | #!/usr/bin/env python
# Use the built in json and sqlite library
import sqlite3
def card_image_url(card_game_id):
return "http://wow.zamimg.com/images/hearthstone/cards/enus/original/" + card_game_id + ".png"
def main():
sql = "SELECT card_game_id, card_name_en FROM cards ORDER BY card_name_en"
db = sqlite3.connect('game.sqlite')
c = db.cursor()
f = open('harvest-images.sh', 'w')
f.write("#!/usr/bin/env bash\n")
for row in c.execute(sql):
harvest_url = card_image_url(row[0])
f.write('wget ' + harvest_url + "\n")
# print harvest_url
f.close()
# Boilerplate python
if __name__ == '__main__':
main()
| 481 | 0 | 46 |
47d5d5adce279fd60bc35ef87afa6132fa4ea625 | 6,124 | py | Python | binmom.py | isseikz/StarTracker | 01c1dfcf8c9a6886acfa18c038acc723f56cc94d | [
"MIT"
] | null | null | null | binmom.py | isseikz/StarTracker | 01c1dfcf8c9a6886acfa18c038acc723f56cc94d | [
"MIT"
] | null | null | null | binmom.py | isseikz/StarTracker | 01c1dfcf8c9a6886acfa18c038acc723f56cc94d | [
"MIT"
] | null | null | null | #!/usr/bin/python
# -*- coding: utf-8 -*-
"""画像の2値化(binalize)とモーメント中心を見つけるモジュール."""
import numpy as np
import cv2
from matplotlib import pyplot as plt
def binalize(img):
"""画像の2値化を行う.
1. 画像をグレースケールに変換
2. しきい値[0, 255]を基準に2値化する
# 入力:
+ img: cv2 モジュールで読み込んだ画像オブジェクト
# 出力:
+ img_binalized: 2値化した画像オブジェクト
参考
https://www.blog.umentu.work/python-opencv3%E3%81%A7%E7%94%BB%E5%83%8F%E3%81%AE%E7%94%BB%E7%B4%A0%E5%80%A4%E3%82%92%E4%BA%8C%E5%80%A4%E5%8C%96%E3%81%97%E3%81%A6%E5%87%BA%E5%8A%9B/
"""
img_gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)
thresh = 60
max_pixel = 255
ret, img_binalized = cv2.threshold(
img_gray,
thresh,
max_pixel,
cv2.THRESH_BINARY
)
return img_binalized
def getCoG(img):
"""輪郭の重心を求める.輪郭が見つからなければ中心を出力.
# 入力
+ img: cv2モジュールの画像オブジェクト
# 出力
+ [x, y]: 輪郭から得られた重心位置
参考
http://labs.eecs.tottori-u.ac.jp/sd/Member/oyamada/OpenCV/html/py_tutorials/py_imgproc/py_contours/py_contour_features/py_contour_features.html
"""
img_size = np.array([img.shape[1], img.shape[0]])
mu = cv2.moments(img, False)
if mu["m00"] < 1.0e-2:
# cv2.imshow("image",img)
x = int(img_size[0]/2)
y = int(img_size[1]/2)
else:
x, y = int(mu["m10"]/mu["m00"]), int(mu["m01"]/mu["m00"])
pass
return np.array([x, y])
def run(img):
"""色抽出を行わずに解析を行う.
# 入力
+ img: cv2モジュールの画像オブジェクト
# 出力
+ error: 重心位置の、画像中心からのずれ
+ center: 輪郭の重心位置
"""
img_size = np.array([img.shape[1], img.shape[0]])
img_binalized = binalize(img)
center = getCoG(img_binalized)
error = center - img_size/2
print("Position: [%d, %d]" % (center[0], center[1]))
print("Error: [%d, %d]" % (error[0], error[1]))
print("Center: [%d, %d]" % (center[0], center[1]))
cv2.circle(
img,
(center[0], center[1]),
int(img.shape[1] * 0.01),
(0, 0, 255),
-1
)
return error, center
def runMovingObject(img, pastImg):
"""動体検出用.2つの画像から変化した部分の中心を出力.
# 入力
+ img: cv2モジュールの画像オブジェクト
+ pastImg: cv2モジュールの画像オブジェクト
# 出力
+ error: 重心位置の、画像中心からのずれ
+ center: 輪郭の重心位置
例えば、img=現在の画像、pastImg=初期状態の画像とすれば、初期からの変位が取れるはず。
"""
img_size = np.array([img.shape[1], img.shape[0]])
gray = cv2.cvtColor(img, cv2.COLOR_RGB2GRAY)
pastGray = cv2.cvtColor(pastImg, cv2.COLOR_RGB2GRAY)
# cv2.accumulateWeighted(gray, pastGray, 0.5)
delta = cv2.absdiff(gray, pastGray)
ret, img_binalized = cv2.threshold(delta, 100, 255, cv2.THRESH_BINARY)
center = getCoG(img_binalized)
error = center - img_size/2
print("Position: [%d, %d]" % (center[0], center[1]))
print("Error: [%d, %d]" % (error[0], error[1]))
print("Center: [%d, %d]" % (center[0], center[1]))
cv2.circle(
img,
(center[0], center[1]),
int(img.shape[1] * 0.01),
(0, 0, 255),
-1
)
return error, center
def runPink(img):
"""画像からピンク色を抽出し、解析を行う.
# 入力
+ img: cv2モジュールの画像オブジェクト
# 出力
+ error: 重心位置の画像中心からのずれ
+ center: 輪郭の重心位置
lowerPink, upperPink を調整することで,他の色を抽出できる,
配列は画像のHSVを表していて,[H, S, V]となる.
範囲の決め方は以下を参考(英語)
https://stackoverflow.com/questions/47483951/how-to-define-a-threshold-value-to-detect-only-green-colour-objects-in-an-image/47483966#47483966
リンク先に載ってる図から欲しい色範囲を決めて,
x範囲→H, y範囲→S に入れればよさげ.
V は[20, 255]くらいで適当に.
"""
img_size = np.array([img.shape[1], img.shape[0]])
imgHSV = cv2.cvtColor(img, cv2.COLOR_BGR2HSV)
lowerPink = np.array([150, 150, 20])
upperPink = np.array([170, 255, 255])
# # オレンジはこれくらい?
# lowerOrange = np.array([10, 150, 20])
# upperOrange = np.array([25, 255, 255])
#
# # 赤はこれくらい?
# lowerRed = np.array([165, 150, 20])
# upperRed = np.array([180, 255, 255])
imgMask = cv2.inRange(imgHSV, lowerPink, upperPink)
imgMasked = cv2.bitwise_and(img, img, mask=imgMask)
# plt.figure()
# plt.imshow(cv2.cvtColor(imgMasked, cv2.COLOR_RGB2BGR))
# plt.show()
img_gray = cv2.cvtColor(imgMasked, cv2.COLOR_BGR2GRAY)
# plt.figure()
# plt.imshow(img_gray, cmap='gray')
# plt.show()
ret, img_binalized = cv2.threshold(img_gray, 60, 255, cv2.THRESH_BINARY)
# plt.figure()
# plt.imshow(img_binalized, cmap='gray')
# plt.show()
center = getCoG(img_binalized)
error = center - img_size/2
print("Position: [%d, %d]" % (center[0], center[1]))
print("Error: [%d, %d]" % (error[0], error[1]))
print("Center: [%d, %d]" % (center[0], center[1]))
cv2.circle(
img,
(center[0], center[1]),
int(img.shape[1] * 0.01),
(0, 0, 255),
-1
)
return error, center
if __name__ == '__main__':
img = cv2.imread('./20180707test/out599.jpg', 1)
error, center = runPink(img)
cv2.circle(
img,
(center[0], center[1]),
int(img.shape[1] * 0.01),
(0, 0, 255),
-1
)
plt.figure()
plt.imshow(cv2.cvtColor(img, cv2.COLOR_RGB2BGR))
plt.show()
# img = cv2.imread('./test/IMAGE_MOON.JPG',1)
# error, center = run(img)
#
# cv2.circle(
# img,
# (center[0], center[1]),
# int(img.shape[1] * 0.01),
# (0, 0, 255),
# -1
# )
# plt.figure()
# plt.imshow(cv2.cvtColor(img, cv2.COLOR_RGB2BGR))
# plt.show()
#
# img = cv2.imread('./test/IMAGE_MOON2.JPG',1)
# error, center = run(img)
#
# cv2.circle(
# img,
# (center[0], center[1]),
# int(img.shape[1] * 0.01),
# (0, 0, 255),
# -1
# )
# plt.figure()
# plt.imshow(cv2.cvtColor(img, cv2.COLOR_RGB2BGR))
# plt.show()
#
# img = cv2.imread('./test/IMAGE_MOON3.JPG',1)
# error, center = run(img)
#
# cv2.circle(
# img,
# (center[0], center[1]),
# int(img.shape[1] * 0.01),
# (0, 0, 255),
# -1
# )
# plt.figure()
# plt.imshow(cv2.cvtColor(img, cv2.COLOR_RGB2BGR))
# plt.show()
| 24.594378 | 183 | 0.566786 | #!/usr/bin/python
# -*- coding: utf-8 -*-
"""画像の2値化(binalize)とモーメント中心を見つけるモジュール."""
import numpy as np
import cv2
from matplotlib import pyplot as plt
def binalize(img):
"""画像の2値化を行う.
1. 画像をグレースケールに変換
2. しきい値[0, 255]を基準に2値化する
# 入力:
+ img: cv2 モジュールで読み込んだ画像オブジェクト
# 出力:
+ img_binalized: 2値化した画像オブジェクト
参考
https://www.blog.umentu.work/python-opencv3%E3%81%A7%E7%94%BB%E5%83%8F%E3%81%AE%E7%94%BB%E7%B4%A0%E5%80%A4%E3%82%92%E4%BA%8C%E5%80%A4%E5%8C%96%E3%81%97%E3%81%A6%E5%87%BA%E5%8A%9B/
"""
img_gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)
thresh = 60
max_pixel = 255
ret, img_binalized = cv2.threshold(
img_gray,
thresh,
max_pixel,
cv2.THRESH_BINARY
)
return img_binalized
def getCoG(img):
"""輪郭の重心を求める.輪郭が見つからなければ中心を出力.
# 入力
+ img: cv2モジュールの画像オブジェクト
# 出力
+ [x, y]: 輪郭から得られた重心位置
参考
http://labs.eecs.tottori-u.ac.jp/sd/Member/oyamada/OpenCV/html/py_tutorials/py_imgproc/py_contours/py_contour_features/py_contour_features.html
"""
img_size = np.array([img.shape[1], img.shape[0]])
mu = cv2.moments(img, False)
if mu["m00"] < 1.0e-2:
# cv2.imshow("image",img)
x = int(img_size[0]/2)
y = int(img_size[1]/2)
else:
x, y = int(mu["m10"]/mu["m00"]), int(mu["m01"]/mu["m00"])
pass
return np.array([x, y])
def run(img):
"""色抽出を行わずに解析を行う.
# 入力
+ img: cv2モジュールの画像オブジェクト
# 出力
+ error: 重心位置の、画像中心からのずれ
+ center: 輪郭の重心位置
"""
img_size = np.array([img.shape[1], img.shape[0]])
img_binalized = binalize(img)
center = getCoG(img_binalized)
error = center - img_size/2
print("Position: [%d, %d]" % (center[0], center[1]))
print("Error: [%d, %d]" % (error[0], error[1]))
print("Center: [%d, %d]" % (center[0], center[1]))
cv2.circle(
img,
(center[0], center[1]),
int(img.shape[1] * 0.01),
(0, 0, 255),
-1
)
return error, center
def runMovingObject(img, pastImg):
"""動体検出用.2つの画像から変化した部分の中心を出力.
# 入力
+ img: cv2モジュールの画像オブジェクト
+ pastImg: cv2モジュールの画像オブジェクト
# 出力
+ error: 重心位置の、画像中心からのずれ
+ center: 輪郭の重心位置
例えば、img=現在の画像、pastImg=初期状態の画像とすれば、初期からの変位が取れるはず。
"""
img_size = np.array([img.shape[1], img.shape[0]])
gray = cv2.cvtColor(img, cv2.COLOR_RGB2GRAY)
pastGray = cv2.cvtColor(pastImg, cv2.COLOR_RGB2GRAY)
# cv2.accumulateWeighted(gray, pastGray, 0.5)
delta = cv2.absdiff(gray, pastGray)
ret, img_binalized = cv2.threshold(delta, 100, 255, cv2.THRESH_BINARY)
center = getCoG(img_binalized)
error = center - img_size/2
print("Position: [%d, %d]" % (center[0], center[1]))
print("Error: [%d, %d]" % (error[0], error[1]))
print("Center: [%d, %d]" % (center[0], center[1]))
cv2.circle(
img,
(center[0], center[1]),
int(img.shape[1] * 0.01),
(0, 0, 255),
-1
)
return error, center
def runPink(img):
"""画像からピンク色を抽出し、解析を行う.
# 入力
+ img: cv2モジュールの画像オブジェクト
# 出力
+ error: 重心位置の画像中心からのずれ
+ center: 輪郭の重心位置
lowerPink, upperPink を調整することで,他の色を抽出できる,
配列は画像のHSVを表していて,[H, S, V]となる.
範囲の決め方は以下を参考(英語)
https://stackoverflow.com/questions/47483951/how-to-define-a-threshold-value-to-detect-only-green-colour-objects-in-an-image/47483966#47483966
リンク先に載ってる図から欲しい色範囲を決めて,
x範囲→H, y範囲→S に入れればよさげ.
V は[20, 255]くらいで適当に.
"""
img_size = np.array([img.shape[1], img.shape[0]])
imgHSV = cv2.cvtColor(img, cv2.COLOR_BGR2HSV)
lowerPink = np.array([150, 150, 20])
upperPink = np.array([170, 255, 255])
# # オレンジはこれくらい?
# lowerOrange = np.array([10, 150, 20])
# upperOrange = np.array([25, 255, 255])
#
# # 赤はこれくらい?
# lowerRed = np.array([165, 150, 20])
# upperRed = np.array([180, 255, 255])
imgMask = cv2.inRange(imgHSV, lowerPink, upperPink)
imgMasked = cv2.bitwise_and(img, img, mask=imgMask)
# plt.figure()
# plt.imshow(cv2.cvtColor(imgMasked, cv2.COLOR_RGB2BGR))
# plt.show()
img_gray = cv2.cvtColor(imgMasked, cv2.COLOR_BGR2GRAY)
# plt.figure()
# plt.imshow(img_gray, cmap='gray')
# plt.show()
ret, img_binalized = cv2.threshold(img_gray, 60, 255, cv2.THRESH_BINARY)
# plt.figure()
# plt.imshow(img_binalized, cmap='gray')
# plt.show()
center = getCoG(img_binalized)
error = center - img_size/2
print("Position: [%d, %d]" % (center[0], center[1]))
print("Error: [%d, %d]" % (error[0], error[1]))
print("Center: [%d, %d]" % (center[0], center[1]))
cv2.circle(
img,
(center[0], center[1]),
int(img.shape[1] * 0.01),
(0, 0, 255),
-1
)
return error, center
if __name__ == '__main__':
img = cv2.imread('./20180707test/out599.jpg', 1)
error, center = runPink(img)
cv2.circle(
img,
(center[0], center[1]),
int(img.shape[1] * 0.01),
(0, 0, 255),
-1
)
plt.figure()
plt.imshow(cv2.cvtColor(img, cv2.COLOR_RGB2BGR))
plt.show()
# img = cv2.imread('./test/IMAGE_MOON.JPG',1)
# error, center = run(img)
#
# cv2.circle(
# img,
# (center[0], center[1]),
# int(img.shape[1] * 0.01),
# (0, 0, 255),
# -1
# )
# plt.figure()
# plt.imshow(cv2.cvtColor(img, cv2.COLOR_RGB2BGR))
# plt.show()
#
# img = cv2.imread('./test/IMAGE_MOON2.JPG',1)
# error, center = run(img)
#
# cv2.circle(
# img,
# (center[0], center[1]),
# int(img.shape[1] * 0.01),
# (0, 0, 255),
# -1
# )
# plt.figure()
# plt.imshow(cv2.cvtColor(img, cv2.COLOR_RGB2BGR))
# plt.show()
#
# img = cv2.imread('./test/IMAGE_MOON3.JPG',1)
# error, center = run(img)
#
# cv2.circle(
# img,
# (center[0], center[1]),
# int(img.shape[1] * 0.01),
# (0, 0, 255),
# -1
# )
# plt.figure()
# plt.imshow(cv2.cvtColor(img, cv2.COLOR_RGB2BGR))
# plt.show()
| 0 | 0 | 0 |
50b35723f20fea2aeacd0fff6f501a16b7939d1c | 237 | py | Python | tests/parser/bug.70.bk.test.py | veltri/DLV2 | 944aaef803aa75e7ec51d7e0c2b0d964687fdd0e | [
"Apache-2.0"
] | null | null | null | tests/parser/bug.70.bk.test.py | veltri/DLV2 | 944aaef803aa75e7ec51d7e0c2b0d964687fdd0e | [
"Apache-2.0"
] | null | null | null | tests/parser/bug.70.bk.test.py | veltri/DLV2 | 944aaef803aa75e7ec51d7e0c2b0d964687fdd0e | [
"Apache-2.0"
] | null | null | null | input = """
% Locations:
loc(dropOff).
loc(distCenter).
loc(truck).
%loc(dest).
% Packages:
pkg(pidVar).
"""
output = """
% Locations:
loc(dropOff).
loc(distCenter).
loc(truck).
%loc(dest).
% Packages:
pkg(pidVar).
"""
| 11.285714 | 17 | 0.586498 | input = """
% Locations:
loc(dropOff).
loc(distCenter).
loc(truck).
%loc(dest).
% Packages:
pkg(pidVar).
"""
output = """
% Locations:
loc(dropOff).
loc(distCenter).
loc(truck).
%loc(dest).
% Packages:
pkg(pidVar).
"""
| 0 | 0 | 0 |
dd0e5f1f8170f9a8b0f7092a2ceddf95c3c34a1c | 2,436 | py | Python | analyze/generate_periscope_chart_embed_code.py | danieldiamond/gitlab-analytics | f99e02c95c3a964b01cb14617a43cd5f64ecd88d | [
"MIT"
] | 3 | 2021-07-22T06:44:31.000Z | 2022-01-29T05:35:12.000Z | analyze/generate_periscope_chart_embed_code.py | danieldiamond/gitlab-analytics | f99e02c95c3a964b01cb14617a43cd5f64ecd88d | [
"MIT"
] | null | null | null | analyze/generate_periscope_chart_embed_code.py | danieldiamond/gitlab-analytics | f99e02c95c3a964b01cb14617a43cd5f64ecd88d | [
"MIT"
] | null | null | null | from os import environ as env
from urllib import parse as parse
import ast
import base64
import csv
import hashlib
import hmac
import json
import logging
API_KEY = env.get("PERISCOPE_API_KEY").encode("UTF-8")
input_csv = "chart_slide_deck.csv"
output_csv = "chart_links.csv"
if __name__ == "__main__":
logging.basicConfig(level=20)
chart_dict_list = csv_to_chart_list(input_csv)
generate_periscope_embed_url(chart_dict_list, output_csv)
logging.info("Script successful.")
| 29.707317 | 80 | 0.65312 | from os import environ as env
from urllib import parse as parse
import ast
import base64
import csv
import hashlib
import hmac
import json
import logging
API_KEY = env.get("PERISCOPE_API_KEY").encode("UTF-8")
input_csv = "chart_slide_deck.csv"
output_csv = "chart_links.csv"
def csv_to_chart_list(input_csv):
logging.info("Opening csv...")
with open(input_csv, "r") as list_of_charts_to_embed:
# Change each fieldname to the appropriate field name.
reader = csv.DictReader(
list_of_charts_to_embed,
fieldnames=("chart_name", "dashboard_id", "widget_id"),
)
header = reader.fieldnames
skip_header = next(reader)
# Parse the CSV into JSON
charts = [row for row in reader]
# Save the JSON
chart_dict_list = []
for chart in charts:
chart_dict = {}
chart_dict["chart_name"] = chart["chart_name"]
chart_dict["chart"] = int(chart["widget_id"])
chart_dict["dashboard"] = int(chart["dashboard_id"])
chart_dict["embed"] = "v2"
chart_dict["border"] = "off"
chart_dict_list.append(chart_dict)
logging.info("Chart list captured...")
return chart_dict_list
def generate_periscope_embed_url(chart_dict_list, output_csv):
chart_links_list = []
for chart in chart_dict_list:
chart_link_dict = {}
chart_link_dict["chart_name"] = chart["chart_name"]
dict_input = chart
dict_input.pop("chart_name")
json_data = json.dumps(dict_input)
encoded_json = parse.urlencode(({"data": json_data})).split("=")[1]
url = "/api/embedded_dashboard?data=" + encoded_json
sig = hmac.new(API_KEY, url.encode(), hashlib.sha256).hexdigest()
chart_link = "https://www.periscopedata.com" + url + "&signature=" + sig
chart_link_dict["chart_link"] = chart_link
chart_links_list.append(chart_link_dict)
logging.info("Writing results...")
with open(output_csv, "w") as embed_urls:
dict_writer = csv.DictWriter(
embed_urls, fieldnames=["chart_name", "chart_link"]
)
dict_writer.writerows(chart_links_list)
if __name__ == "__main__":
logging.basicConfig(level=20)
chart_dict_list = csv_to_chart_list(input_csv)
generate_periscope_embed_url(chart_dict_list, output_csv)
logging.info("Script successful.")
| 1,897 | 0 | 46 |
1bdc339c76d8e57fa714efafdde1b7083a06d571 | 763 | py | Python | Algos and notebooks/testMapWay.py | HehexOne/MapWay | 714889ca97d4199a4cb9f66403fac5c45d791297 | [
"MIT"
] | 1 | 2021-10-18T15:46:27.000Z | 2021-10-18T15:46:27.000Z | Algos and notebooks/testMapWay.py | HehexOne/MapWay | 714889ca97d4199a4cb9f66403fac5c45d791297 | [
"MIT"
] | null | null | null | Algos and notebooks/testMapWay.py | HehexOne/MapWay | 714889ca97d4199a4cb9f66403fac5c45d791297 | [
"MIT"
] | null | null | null | from MapWay import *
start = Place(name="Start", position=Position(latitude=20, longtitude=2))
'''
Уже точки не нужны, они в базе данных
'''
# p_in_1 = Place(name="p_in_1", position=Position(latitude=10, longtitude=50))
# p_in_2 = Place(name="p_in_2", position=Position(latitude=18, longtitude=70))
# p_out_1 = Place(name="p_out_1", position=Position(latitude=21, longtitude=135))
# p_out_2 = Place(name="p_out_2", position=Position(latitude=-10, longtitude=-30))
end = Place(name="End", position=Position(latitude=5, longtitude=100))
# places = [p_in_1, p_in_2, p_out_1, p_out_2]
points = [start, end]
algo = MapWay()
# find_way
# algo.findBestWay(points=points)
route = list(algo.findBestWay(points=points))
for place in route:
print(place.name) | 29.346154 | 82 | 0.722149 | from MapWay import *
start = Place(name="Start", position=Position(latitude=20, longtitude=2))
'''
Уже точки не нужны, они в базе данных
'''
# p_in_1 = Place(name="p_in_1", position=Position(latitude=10, longtitude=50))
# p_in_2 = Place(name="p_in_2", position=Position(latitude=18, longtitude=70))
# p_out_1 = Place(name="p_out_1", position=Position(latitude=21, longtitude=135))
# p_out_2 = Place(name="p_out_2", position=Position(latitude=-10, longtitude=-30))
end = Place(name="End", position=Position(latitude=5, longtitude=100))
# places = [p_in_1, p_in_2, p_out_1, p_out_2]
points = [start, end]
algo = MapWay()
# find_way
# algo.findBestWay(points=points)
route = list(algo.findBestWay(points=points))
for place in route:
print(place.name) | 0 | 0 | 0 |
f31378e6416b9e96d2caa400aa67a2a90f629978 | 1,796 | py | Python | main/mnist.py | yanMichellod/MNIST | d789e73295ce045c1b358efa1c6df26a8570b721 | [
"MIT"
] | null | null | null | main/mnist.py | yanMichellod/MNIST | d789e73295ce045c1b358efa1c6df26a8570b721 | [
"MIT"
] | null | null | null | main/mnist.py | yanMichellod/MNIST | d789e73295ce045c1b358efa1c6df26a8570b721 | [
"MIT"
] | null | null | null | import argparse
import sys
| 28.0625 | 88 | 0.554566 | import argparse
import sys
def main(args=None):
if not args:
args = sys.argv[1:]
example_doc = """\
examples:
1. Runs the full analysis with the whole MNIST dataset:
$ mnist --full=True
2. Only runs for a subset of MNIST dataset (f.e. for quick test):
$ mnist.py
or
$ mnist --full=False
"""
parser = argparse.ArgumentParser(
usage="python %(prog)s [options]",
description="Performs CNN on MNIST dataset and checks the accuracy",
epilog=example_doc,
formatter_class=argparse.RawDescriptionHelpFormatter,
)
parser.add_argument(
"-f",
"--full",
choices=["True", "False"],
nargs="*",
default=["False"],
help="Decides if analysis have to take place at whole MNIST dataset. "
"Options are %(default)s (default: %(default)s)",
)
args = parser.parse_args(args)
full = args.full[0] == "True"
print("Start analysis of MNIST dataset (full={0})...".format(full))
from Analysis.Analysis import Analysis
an = Analysis(full=full)
print("*************************************************")
print("Analyse results")
print("")
print("Accuracy RF: %.3f" % (an.getRFAccuracy() * 100.0))
print("Accuracy CNN: %.3f" % (an.getCNNAccuracy() * 100.0))
print(
"Is the CNN accuracy higher then the accuracy of the random forest: {0}".format(
an.checkVSBaseline()
)
)
print(
"Is the CNN accuracy higher then 0.95 (95%): {0}".format(
an.checkHypothesis(0.95)
)
)
print("Save confusion matrix: {0}".format(an.saveConfusionMatrix()))
print("Analysis done!")
print("*************************************************")
return True
| 1,745 | 0 | 23 |
8826ed8824a2aa17e3bb3cde009518af14b96186 | 3,701 | py | Python | businessdate/businessrange.py | pbrisk/businessdate | 36eb9df687f7f3a30d659f356f16a05d5f41b309 | [
"Apache-2.0"
] | 2 | 2019-07-30T13:07:00.000Z | 2021-05-26T02:19:09.000Z | businessdate/businessrange.py | sonntagsgesicht/businessdate | 36eb9df687f7f3a30d659f356f16a05d5f41b309 | [
"Apache-2.0"
] | 2 | 2019-05-14T16:06:54.000Z | 2019-07-27T23:14:29.000Z | businessdate/businessrange.py | pbrisk/businessdate | 36eb9df687f7f3a30d659f356f16a05d5f41b309 | [
"Apache-2.0"
] | 7 | 2017-04-10T13:19:47.000Z | 2019-07-06T03:48:49.000Z | # -*- coding: utf-8 -*-
# businessdate
# ------------
# Python library for generating business dates for fast date operations
# and rich functionality.
#
# Author: sonntagsgesicht, based on a fork of Deutsche Postbank [pbrisk]
# Version: 0.5, copyright Wednesday, 18 September 2019
# Website: https://github.com/sonntagsgesicht/businessdate
# License: Apache License 2.0 (see LICENSE file)
from .businessperiod import BusinessPeriod
from .businessdate import BusinessDate
| 35.586538 | 101 | 0.627398 | # -*- coding: utf-8 -*-
# businessdate
# ------------
# Python library for generating business dates for fast date operations
# and rich functionality.
#
# Author: sonntagsgesicht, based on a fork of Deutsche Postbank [pbrisk]
# Version: 0.5, copyright Wednesday, 18 September 2019
# Website: https://github.com/sonntagsgesicht/businessdate
# License: Apache License 2.0 (see LICENSE file)
from .businessperiod import BusinessPeriod
from .businessdate import BusinessDate
class BusinessRange(list):
def __init__(self, start, stop=None, step=None, rolling=None):
""" class to build list of business days
:param BusinessDate start: date to begin schedule,
if stop not given, start will be used as stop and
default in rolling to :class:`BusinessDate() <BusinessDate>`
:param BusinessDate stop: date to stop before,
if not given, start will be used for stop instead
:param BusinessPeriod step: period to step schedule,
if not given 1 day is default
:param BusinessDate rolling: date to roll on
(forward and backward) between start and stop,
if not given default will be start
**Ansatz** First, `rolling` and `step`
defines a infinite grid of dates.
Second, this grid is sliced by `start` (included ,
if meeting the grid) and `end` (excluded).
All dates will have same **convention**, **holidays**
and **day_count** property as **rolling**.
"""
# set default args and build range grid
start, stop, step, rolling = self._default_args(start, stop, step, rolling)
schedule = self._build_grid(start, stop, step, rolling)
# push to super and sort
super(BusinessRange, self).__init__(set(schedule))
self.sort()
@staticmethod
def _default_args(start, stop, step, rolling):
if stop is None:
stop = start
start = BusinessDate()
start.convention = getattr(stop, 'convention', None)
start.holidays = getattr(stop, 'holidays', None)
start.day_count = getattr(stop, 'day_count', None)
if step is None:
step = BusinessPeriod(days=1)
if rolling is None:
rolling = start
# make proper businessdate objects
start = BusinessDate(start)
rolling = BusinessDate(rolling)
stop = BusinessDate(stop)
step = BusinessPeriod(step)
return start, stop, step, rolling
@staticmethod
def _build_grid(start, stop, step, rolling):
# setup grid and turn step into positive direction
grid = list()
step = step if rolling <= rolling + step else -1 * step
# roll backward before start
i = 0
while start <= rolling + step * i:
i -= 1
# fill grid from start until end
current = rolling + step * i
while current < stop:
if start <= current < stop:
grid.append(current)
i += 1
current = rolling + step * i
return grid
def adjust(self, convention=None, holidays=None):
""" returns adjusted :class:`BusinessRange` following given convention
For details of adjusting :class:`BusinessDate` see :meth:`BusinessDate.adjust`.
For possible conventions invoke :meth:`BusinessDate().adjust() <BusinessDate.adjust>`
For more details on the conventions see module :mod:`conventions <businessdate.conventions>`)
"""
adj_list = [d.adjust(convention, holidays) for d in self]
del self[:]
super(BusinessRange, self).extend(adj_list)
return self
| 1,202 | 1,995 | 23 |
a59182a74a30e5559913ac544809d405eb2da6c4 | 142 | py | Python | messages/models.py | ntdai95/AuctionWebsite | ed1e4c779c9ba27fd26956756b5bd80a47fb44c5 | [
"MIT"
] | 1 | 2022-01-24T07:46:47.000Z | 2022-01-24T07:46:47.000Z | messages/models.py | ntdai95/AuctionWebsite | ed1e4c779c9ba27fd26956756b5bd80a47fb44c5 | [
"MIT"
] | null | null | null | messages/models.py | ntdai95/AuctionWebsite | ed1e4c779c9ba27fd26956756b5bd80a47fb44c5 | [
"MIT"
] | null | null | null | from pydantic import BaseModel
| 17.75 | 31 | 0.711268 | from pydantic import BaseModel
class Message(BaseModel):
message_id: int = None
sending_user_id: int = None
message: str = None
| 0 | 87 | 23 |
670546ec050434d9276a8108386da84f83c5dc42 | 180 | py | Python | codigo_programa/database.py | matheusparrela/Fisica_game | 7c6e1607c4786f1fa76f6be9abe51c0279754926 | [
"MIT"
] | 1 | 2022-02-06T18:33:50.000Z | 2022-02-06T18:33:50.000Z | codigo_programa/database.py | matheusparrela/Fisica_game | 7c6e1607c4786f1fa76f6be9abe51c0279754926 | [
"MIT"
] | null | null | null | codigo_programa/database.py | matheusparrela/Fisica_game | 7c6e1607c4786f1fa76f6be9abe51c0279754926 | [
"MIT"
] | null | null | null | import pyodbc
dados_conexao = (
"Driver={SQL Server};"
"Server=omicron;"
"Database=perguntas;"
)
conexao = pyodbc.connect(dados_conexao)
print('Conexão bem sucedida!') | 20 | 39 | 0.7 | import pyodbc
dados_conexao = (
"Driver={SQL Server};"
"Server=omicron;"
"Database=perguntas;"
)
conexao = pyodbc.connect(dados_conexao)
print('Conexão bem sucedida!') | 0 | 0 | 0 |
ed10a413930f081fc45944e232c2b7b86a1eac17 | 828 | py | Python | visualization/model_report.py | mahetoodang/abm | b27c0616aed288902d2f28444c79cf0720513249 | [
"MIT"
] | null | null | null | visualization/model_report.py | mahetoodang/abm | b27c0616aed288902d2f28444c79cf0720513249 | [
"MIT"
] | null | null | null | visualization/model_report.py | mahetoodang/abm | b27c0616aed288902d2f28444c79cf0720513249 | [
"MIT"
] | null | null | null | import pandas as pd
import numpy as np
import webbrowser
import os
import glob
from pandas_profiling import ProfileReport
'''Automated generation of HTML data report to
be opened inside webbrowser'''
if __name__ == '__main__':
create_model_report(True)
| 30.666667 | 61 | 0.695652 | import pandas as pd
import numpy as np
import webbrowser
import os
import glob
from pandas_profiling import ProfileReport
'''Automated generation of HTML data report to
be opened inside webbrowser'''
def create_model_report(html_report=True):
if html_report == True:
file_list = glob.glob('data/stats/*')
latest_file = max(file_list, key=os.path.getctime)
df = pd.read_csv(latest_file)
profile = ProfileReport(df, title="Profiling Report")
output_file = 'data/html/model_report.html'
profile.to_file(output_file=output_file)
model_report_path = os.getcwd() + '/' + output_file
new = 2 # open in a new tab, if possible
return webbrowser.open(model_report_path, new=new)
else:
pass
if __name__ == '__main__':
create_model_report(True)
| 546 | 0 | 23 |
aeb73f304dad48bc29e2b1b63872e0a6a6e6411c | 870 | py | Python | software/redis.py | kylebush/bigdata-deploy | e52c4bc5f136802b69bec18d151b0c5d7f7af837 | [
"Apache-2.0"
] | 2 | 2017-01-08T17:17:59.000Z | 2017-07-18T12:14:11.000Z | software/redis.py | kylebush/bigdata-tools | e52c4bc5f136802b69bec18d151b0c5d7f7af837 | [
"Apache-2.0"
] | null | null | null | software/redis.py | kylebush/bigdata-tools | e52c4bc5f136802b69bec18d151b0c5d7f7af837 | [
"Apache-2.0"
] | null | null | null | import os
from fabric.api import env, sudo, put
import helper
from bootstrap import machine
| 33.461538 | 83 | 0.734483 | import os
from fabric.api import env, sudo, put
import helper
from bootstrap import machine
def install(host_config):
env.host_string = helper.get_env_host_string(host_config)
env.user = helper.get_env_user(host_config)
env.key_filename = helper.get_env_key_filename(host_config)
software_config = helper.get_software_config(host_config, 'redis')
redis_version = software_config.get('version', '3.2.6')
redis_port = software_config.get('port', '6379')
redis_data_dir = software_config.get('data-directory', '/var/lib/redis')
machine.disable_transparent_huge_pages(env.host_string)
machine.set_overcommit_memory(env.host_string, 1)
put('{}/software/scripts/redis.sh'.format(os.getcwd()), '~/', use_sudo=True)
sudo("chmod +x redis.sh")
sudo(". ~/redis.sh {} {} {}".format(redis_version, redis_port, redis_data_dir))
| 752 | 0 | 23 |
135ec19a63db6b926dd484003d11d8f4d8f345e4 | 492 | py | Python | FizzBuzz/tests/test.py | n0n0a/python-dev-tutorial | 565e2eaaaf4b8777295b83f8d8de425fe9f22bca | [
"MIT"
] | null | null | null | FizzBuzz/tests/test.py | n0n0a/python-dev-tutorial | 565e2eaaaf4b8777295b83f8d8de425fe9f22bca | [
"MIT"
] | null | null | null | FizzBuzz/tests/test.py | n0n0a/python-dev-tutorial | 565e2eaaaf4b8777295b83f8d8de425fe9f22bca | [
"MIT"
] | null | null | null | from unittest import TestCase
import pytest
from FizzBuzz.fizzbuzz import fizzbuzz
fizzbuzz_data = [(3, "Fizz"), (5, "Buzz"), (15, "FizzBuzz"), (2, "2")]
fizzbuzz_7_11_data = [(7,"Fizz"),(11,"Buzz"),(77,"FizzBuzz"), (2, "2")]
@pytest.mark.parametrize(('n','expected'),fizzbuzz_data)
@pytest.mark.parametrize(('n','expected'),fizzbuzz_7_11_data)
| 32.8 | 71 | 0.697154 | from unittest import TestCase
import pytest
from FizzBuzz.fizzbuzz import fizzbuzz
fizzbuzz_data = [(3, "Fizz"), (5, "Buzz"), (15, "FizzBuzz"), (2, "2")]
fizzbuzz_7_11_data = [(7,"Fizz"),(11,"Buzz"),(77,"FizzBuzz"), (2, "2")]
@pytest.mark.parametrize(('n','expected'),fizzbuzz_data)
def test_fizzbuzz(n,expected):
assert fizzbuzz(n) == expected
@pytest.mark.parametrize(('n','expected'),fizzbuzz_7_11_data)
def test_fizzbuzz_7_11(n,expected):
assert fizzbuzz(n, 7, 11) == expected
| 100 | 0 | 44 |
dbd24433096ce21f12bbf4d0bf81b1a43cc47d60 | 1,172 | py | Python | applied_python/applied_python/lib/python2.7/site-packages/ncclient/devices/iosxe.py | mith1979/ansible_automation | 013dfa67c6d91720b787fadb21de574b6e023a26 | [
"Apache-2.0"
] | null | null | null | applied_python/applied_python/lib/python2.7/site-packages/ncclient/devices/iosxe.py | mith1979/ansible_automation | 013dfa67c6d91720b787fadb21de574b6e023a26 | [
"Apache-2.0"
] | null | null | null | applied_python/applied_python/lib/python2.7/site-packages/ncclient/devices/iosxe.py | mith1979/ansible_automation | 013dfa67c6d91720b787fadb21de574b6e023a26 | [
"Apache-2.0"
] | null | null | null | """
Handler for Cisco IOS-XE device specific information.
Note that for proper import, the classname has to be:
"<Devicename>DeviceHandler"
...where <Devicename> is something like "Default", "Nexus", etc.
All device-specific handlers derive from the DefaultDeviceHandler, which implements the
generic information needed for interaction with a Netconf server.
"""
from .default import DefaultDeviceHandler
class IosxeDeviceHandler(DefaultDeviceHandler):
"""
Cisco IOS-XE handler for device specific information.
"""
| 27.904762 | 87 | 0.715017 | """
Handler for Cisco IOS-XE device specific information.
Note that for proper import, the classname has to be:
"<Devicename>DeviceHandler"
...where <Devicename> is something like "Default", "Nexus", etc.
All device-specific handlers derive from the DefaultDeviceHandler, which implements the
generic information needed for interaction with a Netconf server.
"""
from .default import DefaultDeviceHandler
def iosxe_unknown_host_cb(host, fingerprint):
#This will ignore the unknown host check when connecting to CSR devices
return True
class IosxeDeviceHandler(DefaultDeviceHandler):
"""
Cisco IOS-XE handler for device specific information.
"""
def __init__(self, device_params):
super(IosxeDeviceHandler, self).__init__(device_params)
def add_additional_operations(self):
dict = {}
dict["save_config"] = SaveConfig
return dict
def add_additional_ssh_connect_params(self, kwargs):
kwargs['allow_agent'] = False
kwargs['look_for_keys'] = False
kwargs['unknown_host_cb'] = csr_unknown_host_cb
def perform_qualify_check(self):
return False
| 494 | 0 | 138 |
0b75ad5a395e43a103ee317b5d794d2aa4f68b3a | 525 | py | Python | automation.py | FathiMahdi/TNB_Social_Distance | ab5c13652e57d827180b92e03983cda39def51e8 | [
"MIT"
] | null | null | null | automation.py | FathiMahdi/TNB_Social_Distance | ab5c13652e57d827180b92e03983cda39def51e8 | [
"MIT"
] | null | null | null | automation.py | FathiMahdi/TNB_Social_Distance | ab5c13652e57d827180b92e03983cda39def51e8 | [
"MIT"
] | null | null | null | import os
import time
from cctv import *
for i in range(0,10000000000):
cctvDetect(i)# run cctv
time.sleep(2)
os.system('sudo python3 darknet_images.py --input /mnt/k/monitoring/204online_{}.png --batch_size 1 --weights /home/Social_Distance/darknet/backup/yolo-obj_last.weights --config_file /home/Social_Distance/darknet/cfg/yolo-obj.cfg --data_file /home/Social_Distance/darknet/data/obj.data --dont_show'.format(i))
if(i>0):
os.system('sudo rm /mnt/k/monitoring/204online_{}.png'.format(i-1))
| 40.384615 | 314 | 0.731429 | import os
import time
from cctv import *
for i in range(0,10000000000):
cctvDetect(i)# run cctv
time.sleep(2)
os.system('sudo python3 darknet_images.py --input /mnt/k/monitoring/204online_{}.png --batch_size 1 --weights /home/Social_Distance/darknet/backup/yolo-obj_last.weights --config_file /home/Social_Distance/darknet/cfg/yolo-obj.cfg --data_file /home/Social_Distance/darknet/data/obj.data --dont_show'.format(i))
if(i>0):
os.system('sudo rm /mnt/k/monitoring/204online_{}.png'.format(i-1))
| 0 | 0 | 0 |
5b106466dafaf888a5489b6369524357186de3c6 | 926 | py | Python | accounts/migrations/0021_auto_20210902_2010.py | simplymarco/Sistema-de-Controle-de-Processos-Titulados | f4e3bc7260611b848096d5fbfb1f72be9d440f1f | [
"MIT"
] | null | null | null | accounts/migrations/0021_auto_20210902_2010.py | simplymarco/Sistema-de-Controle-de-Processos-Titulados | f4e3bc7260611b848096d5fbfb1f72be9d440f1f | [
"MIT"
] | null | null | null | accounts/migrations/0021_auto_20210902_2010.py | simplymarco/Sistema-de-Controle-de-Processos-Titulados | f4e3bc7260611b848096d5fbfb1f72be9d440f1f | [
"MIT"
] | null | null | null | # Generated by Django 3.0 on 2021-09-02 23:10
from django.db import migrations, models
| 30.866667 | 114 | 0.564795 | # Generated by Django 3.0 on 2021-09-02 23:10
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('accounts', '0020_auto_20210902_1708'),
]
operations = [
migrations.CreateModel(
name='Feedback',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('customer_name', models.CharField(max_length=120)),
('email', models.EmailField(max_length=254)),
('details', models.TextField()),
('happy', models.BooleanField()),
('date', models.DateField(auto_now_add=True)),
],
),
migrations.AlterField(
model_name='processo',
name='interessado',
field=models.ManyToManyField(to='accounts.Interessado'),
),
]
| 0 | 814 | 23 |
f244d9daab59011f48024a89527c2f4f9c8108e1 | 4,836 | py | Python | flaskr/fetch_data.py | Amoghrd/StockPrediction | e33a2ec2209f24089b2b92739b82ec05414aa853 | [
"MIT"
] | 1 | 2020-10-27T03:31:29.000Z | 2020-10-27T03:31:29.000Z | flaskr/fetch_data.py | Amoghrd/StockPrediction | e33a2ec2209f24089b2b92739b82ec05414aa853 | [
"MIT"
] | 32 | 2020-10-06T01:34:09.000Z | 2020-11-17T01:31:30.000Z | flaskr/fetch_data.py | Amoghrd/StockPrediction | e33a2ec2209f24089b2b92739b82ec05414aa853 | [
"MIT"
] | 4 | 2020-10-19T20:10:49.000Z | 2020-12-12T15:31:44.000Z | import yfinance as yf
import matplotlib.pyplot as plt
import datetime
import math
import numpy as np
from sklearn.model_selection import train_test_split
from sklearn import preprocessing
from sklearn.linear_model import LinearRegression
import csv
# Calculate slopes
| 33.351724 | 88 | 0.552316 | import yfinance as yf
import matplotlib.pyplot as plt
import datetime
import math
import numpy as np
from sklearn.model_selection import train_test_split
from sklearn import preprocessing
from sklearn.linear_model import LinearRegression
import csv
def get_stock(ticker_name, days_number):
tick = yf.Ticker(ticker_name)
df = tick.history(period=str(days_number) + "d")
return df
def regression(df, name, date):
df = df[["Open", "High", "Low", "Close", "Volume"]]
# Attribute Reduction
df["HL_PCT"] = (df["High"] - df["Low"]) / df["Close"] * 100.0
df["PCT_change"] = (df["Close"] - df["Open"]) / df["Open"] * 100.0
# New and more statistically important attributes
df = df[["Close", "HL_PCT", "PCT_change", "Volume"]]
forecast_col = "Close"
# which column to forecast
df.fillna(value=-99999, inplace=True)
forecast_out = int(math.ceil(0.1 * len(df)))
# what percent of days from total number of days to forecast
df["label"] = df[forecast_col].shift(-forecast_out)
# preprocessing and subsetting data for training
X = np.array(df.drop(["label"], 1))
X = preprocessing.scale(X)
X_lately = X[-forecast_out:]
X = X[:-forecast_out]
df.dropna(inplace=True)
y = np.array(df["label"])
# splitting data for test and train
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2)
# calling Linear Regression funtion
clf = LinearRegression(n_jobs=-1)
clf.fit(X_train, y_train)
confidence = clf.score(X_test, y_test)
# Model Trained now printing accuracy
print("Confidence - ", confidence)
forecast_set = clf.predict(X_lately)
df["Forecast"] = np.nan
last_date = df.iloc[-1].name
last_unix = last_date.timestamp()
one_day = 86400
next_unix = last_unix + one_day
for i in forecast_set:
next_date = datetime.datetime.fromtimestamp(next_unix)
next_unix += 86400
df.loc[next_date] = [np.nan for _ in range(len(df.columns) - 1)] + [i]
optimistic = df["Forecast"] * 1.05
pessimistic = df["Forecast"] * 0.95
indexvalues = df.index
covidflag = -1
s = "24/02/2020"
covidtime = datetime.datetime.strptime(s, "%d/%m/%Y")
for i in range(len(indexvalues)):
if indexvalues[i] >= covidtime:
covidflag = i
break
for i in range(len(df)):
if not np.isnan(optimistic[i]):
df["Close"][i] = df["Forecast"][i]
optimistic[i] = df["Forecast"][i]
pessimistic[i] = df["Forecast"][i]
break
# plotting data with the original and the predicted
plt.style.use("dark_background")
plt.figure(figsize=(8, 4))
if covidflag == -1:
plt.plot(df["Close"], color="lime", label="Actual")
else:
plt.plot(df["Close"][:covidflag], color="lime", label="Before Covid")
if covidflag == 0:
plt.plot(df["Close"][covidflag:], color="yellow", label="After Covid")
else:
plt.plot(df["Close"][covidflag - 1 :], color="yellow", label="After Covid")
plt.fill_between(
optimistic.index,
optimistic,
pessimistic,
color="lightgrey",
label="95% confidence interval",
)
plt.plot(df["Forecast"], color="red", label="Forecast")
plt.legend(loc=0)
plt.xlabel("Date")
plt.ylabel("Price")
# plt.show()
plt.savefig(fname="flaskr/static/" + name + "-" + date + ".png", format="png")
def Sort(arr):
return sorted(arr, key=lambda x: x[2], reverse=True)
# Calculate slopes
def generateGrowthRates():
rowList = []
with open("flaskr/static/stocks.csv") as f:
# with open('stocks.csv') as f:
reader = csv.reader(f)
i = 100
for row in reader:
if row not in ["DLPH", "ASFI", "CETV", "BREW", "INWK", "MNTA", "AMTD"]:
error = 0
try:
df = get_stock(row[0], 30)
except Exception as e:
print(e)
error = 1
# Calculate slope and append to the top ten list
if error == 0:
rowList.append(
[
row[0],
row[1],
(df["Close"][len(df["Close"]) - 1] - df["Close"][0])
/ df["Close"][0],
]
)
i -= 1
sortedList = Sort(rowList)
# sortedList.insert(0, ["Ticker", "Name", "Slope"])
with open("GrowthRates.csv", "w", newline="") as file:
writer = csv.writer(file)
writer.writerows(sortedList)
| 4,450 | 0 | 98 |
5ef3cbc4d43ee2bda16eaf11d41a36fef771729f | 3,917 | py | Python | python/fedml/simulation/mpi_p2p_mp/fedavg_robust/FedAvgRobustServerManager.py | NCLPhD/FedML | ffa15262ee963b9c856f34f0b2202f4dfeb3a76b | [
"Apache-2.0"
] | null | null | null | python/fedml/simulation/mpi_p2p_mp/fedavg_robust/FedAvgRobustServerManager.py | NCLPhD/FedML | ffa15262ee963b9c856f34f0b2202f4dfeb3a76b | [
"Apache-2.0"
] | null | null | null | python/fedml/simulation/mpi_p2p_mp/fedavg_robust/FedAvgRobustServerManager.py | NCLPhD/FedML | ffa15262ee963b9c856f34f0b2202f4dfeb3a76b | [
"Apache-2.0"
] | null | null | null | import logging
from .message_define import MyMessage
from .utils import transform_tensor_to_list
from ....core.distributed.communication.message import Message
from ....core.distributed.server.server_manager import ServerManager
| 39.969388 | 87 | 0.6699 | import logging
from .message_define import MyMessage
from .utils import transform_tensor_to_list
from ....core.distributed.communication.message import Message
from ....core.distributed.server.server_manager import ServerManager
class FedAvgRobustServerManager(ServerManager):
def __init__(self, args, aggregator, comm=None, rank=0, size=0, backend="MPI"):
super().__init__(args, comm, rank, size, backend)
self.args = args
self.aggregator = aggregator
self.round_num = args.comm_round
self.round_idx = 0
def run(self):
super().run()
def send_init_msg(self):
# sampling clients
client_indexes = self.aggregator.client_sampling(
self.round_idx,
self.args.client_num_in_total,
self.args.client_num_per_round,
)
global_model_params = self.aggregator.get_global_model_params()
for process_id in range(1, self.size):
self.send_message_init_config(
process_id, global_model_params, client_indexes[process_id - 1]
)
def register_message_receive_handlers(self):
self.register_message_receive_handler(
MyMessage.MSG_TYPE_C2S_SEND_MODEL_TO_SERVER,
self.handle_message_receive_model_from_client,
)
def handle_message_receive_model_from_client(self, msg_params):
sender_id = msg_params.get(MyMessage.MSG_ARG_KEY_SENDER)
model_params = msg_params.get(MyMessage.MSG_ARG_KEY_MODEL_PARAMS)
local_sample_number = msg_params.get(MyMessage.MSG_ARG_KEY_NUM_SAMPLES)
self.aggregator.add_local_trained_result(
sender_id - 1, model_params, local_sample_number
)
b_all_received = self.aggregator.check_whether_all_receive()
logging.info("b_all_received = " + str(b_all_received))
if b_all_received:
global_model_params = self.aggregator.aggregate()
self.aggregator.test_on_all_clients(self.round_idx)
# measure the target task accuracy
self.aggregator.test_target_accuracy(self.round_idx)
# start the next round
self.round_idx += 1
if self.round_idx == self.round_num:
self.finish()
return
# sampling clients
client_indexes = self.aggregator.client_sampling(
self.round_idx,
self.args.client_num_in_total,
self.args.client_num_per_round,
)
print("size = %d" % self.size)
if self.args.is_mobile == 1:
print("transform_tensor_to_list")
global_model_params = transform_tensor_to_list(global_model_params)
for receiver_id in range(1, self.size):
self.send_message_sync_model_to_client(
receiver_id, global_model_params, client_indexes[receiver_id - 1]
)
def send_message_init_config(self, receive_id, global_model_params, client_index):
message = Message(
MyMessage.MSG_TYPE_S2C_INIT_CONFIG, self.get_sender_id(), receive_id
)
message.add_params(MyMessage.MSG_ARG_KEY_MODEL_PARAMS, global_model_params)
message.add_params(MyMessage.MSG_ARG_KEY_CLIENT_INDEX, str(client_index))
self.send_message(message)
def send_message_sync_model_to_client(
self, receive_id, global_model_params, client_index
):
logging.info("send_message_sync_model_to_client. receive_id = %d" % receive_id)
message = Message(
MyMessage.MSG_TYPE_S2C_SYNC_MODEL_TO_CLIENT,
self.get_sender_id(),
receive_id,
)
message.add_params(MyMessage.MSG_ARG_KEY_MODEL_PARAMS, global_model_params)
message.add_params(MyMessage.MSG_ARG_KEY_CLIENT_INDEX, str(client_index))
self.send_message(message)
| 3,449 | 26 | 211 |
b43155c65d83fd218975e62f7dc9a96ea5bcb480 | 1,355 | py | Python | ptsites/sites/1ptba.py | synthpop123/flexget_qbittorrent_mod | f8beea6bac3f07f25dfe10607117ef8068a935f2 | [
"MIT"
] | 202 | 2020-01-24T05:27:18.000Z | 2021-06-22T09:31:54.000Z | ptsites/sites/1ptba.py | synthpop123/flexget_qbittorrent_mod | f8beea6bac3f07f25dfe10607117ef8068a935f2 | [
"MIT"
] | 16 | 2021-08-01T10:08:23.000Z | 2022-02-04T04:18:08.000Z | ptsites/sites/1ptba.py | synthpop123/flexget_qbittorrent_mod | f8beea6bac3f07f25dfe10607117ef8068a935f2 | [
"MIT"
] | 62 | 2020-01-24T05:27:24.000Z | 2021-06-16T04:46:22.000Z | import re
from urllib.parse import urljoin
from ..schema.nexusphp import AttendanceHR
from ..schema.site_base import Work, SignState
| 33.875 | 93 | 0.539483 | import re
from urllib.parse import urljoin
from ..schema.nexusphp import AttendanceHR
from ..schema.site_base import Work, SignState
class MainClass(AttendanceHR):
URL = 'https://1ptba.com/'
USER_CLASSES = {
'downloaded': [805306368000, 3298534883328],
'share_ratio': [3.05, 4.55],
'days': [280, 700]
}
def build_workflow(self, entry, config):
return [
Work(
url='/attendance.php',
method='param',
succeed_regex=[
'这是您的第.*?次签到,已连续签到.*?天,本次签到获得.*?魔力值。|這是您的第.*次簽到,已連續簽到.*?天,本次簽到獲得.*?魔力值。',
'[签簽]到已得\\d+',
'您今天已经签到过了,请勿重复刷新。|您今天已經簽到過了,請勿重複刷新。'],
check_state=('final', SignState.SUCCEED),
is_base_content=True
)
]
def sign_in_by_param(self, entry, config, work, last_content=None):
response = self._request(entry, 'get', work.url)
if response:
location_match = re.search('window\\.location="(.*?);</script>', response.text)
if location_match:
uri = re.sub('["|+| ]', '', location_match.group(1))
work.url = urljoin(work.url, uri)
return self.sign_in_by_get(entry, config, work, last_content)
else:
return response
| 1,141 | 238 | 23 |
97a137d77b023df62ea751088239bb68dcf34480 | 1,282 | py | Python | api/controllers/trace_controller.py | bailobah/relevanc | 1bb9bf895ca1b7bdc0553267c377f53472990d9b | [
"MIT"
] | null | null | null | api/controllers/trace_controller.py | bailobah/relevanc | 1bb9bf895ca1b7bdc0553267c377f53472990d9b | [
"MIT"
] | null | null | null | api/controllers/trace_controller.py | bailobah/relevanc | 1bb9bf895ca1b7bdc0553267c377f53472990d9b | [
"MIT"
] | null | null | null | import datetime
from flask import request
from flask_restplus import Resource
from flask_restplus import abort
from sqlalchemy.orm import load_only
from logger import Logger
from api.serilizers import dataset_response_succes, get_trace
from api.restplus import api
from database.models import TbTrace
log = Logger.log
ns = api.namespace('trace/kpis', description='kpis des datasets')
@ns.route('/')
@api.doc(responses={
200: 'Update success.'
})
class TraceCollection(Resource):
"""docstring for TraceCollection"""
@api.marshal_with(get_trace)
def get(self):
"""
Returns collection trace.
"""
tbtrace_query = TbTrace.query\
.options(load_only('id_dataset', 'nom_dataset','nombre_total_ligne','status','lignes_corompus','insertion_date'))\
.all()
return tbtrace_query
@ns.route('/<string:id_dataset>')
| 23.309091 | 118 | 0.730109 | import datetime
from flask import request
from flask_restplus import Resource
from flask_restplus import abort
from sqlalchemy.orm import load_only
from logger import Logger
from api.serilizers import dataset_response_succes, get_trace
from api.restplus import api
from database.models import TbTrace
log = Logger.log
ns = api.namespace('trace/kpis', description='kpis des datasets')
@ns.route('/')
@api.doc(responses={
200: 'Update success.'
})
class TraceCollection(Resource):
"""docstring for TraceCollection"""
@api.marshal_with(get_trace)
def get(self):
"""
Returns collection trace.
"""
tbtrace_query = TbTrace.query\
.options(load_only('id_dataset', 'nom_dataset','nombre_total_ligne','status','lignes_corompus','insertion_date'))\
.all()
return tbtrace_query
@ns.route('/<string:id_dataset>')
class DataItem(Resource):
@api.marshal_with(get_trace, code=200)
def get(self, id_dataset):
"""
Returns All trace id_dataset.
"""
data = TbTrace.query\
.options(load_only('id_dataset', 'nom_dataset','nombre_total_ligne','status','lignes_corompus','insertion_date'))\
.filter_by(id_dataset = id_dataset).all()
if len(data) == 0:
abort(400, 'This id_dataset not in database', id=id_dataset)
return data, 200
| 0 | 422 | 22 |
3934659f21a9d149a3c887921ac80207b3058b4e | 3,058 | py | Python | android-runner/AndroidRunner/Plugins/perfume_js/AddJS.py | S2-group/Lacuna-evaluation | b982d54a7cb65050f1743d0a514ebcabce01f23c | [
"MIT"
] | null | null | null | android-runner/AndroidRunner/Plugins/perfume_js/AddJS.py | S2-group/Lacuna-evaluation | b982d54a7cb65050f1743d0a514ebcabce01f23c | [
"MIT"
] | null | null | null | android-runner/AndroidRunner/Plugins/perfume_js/AddJS.py | S2-group/Lacuna-evaluation | b982d54a7cb65050f1743d0a514ebcabce01f23c | [
"MIT"
] | 1 | 2021-07-23T10:41:10.000Z | 2021-07-23T10:41:10.000Z | import os, sys
import subprocess
from distutils.dir_util import copy_tree
from bs4 import BeautifulSoup
if __name__ == '__main__':
#USAGE: (python3 AddJS.py path/To/Directory/With/All/WebApplication/ IPADRRESS)
#NOTE: IPADDRESS should be in the form of http://IP:8080/ and IP is the ip address
add_js(sys.argv[1], sys.argv[2])
| 51.830508 | 677 | 0.599411 | import os, sys
import subprocess
from distutils.dir_util import copy_tree
from bs4 import BeautifulSoup
def bfs_dirs(root):
paths = [root]
while len(paths) > 0:
children = []
for parent in paths:
if parent.endswith(".gitignore"):
continue
for element in os.listdir(parent):
found_element = os.path.join(parent, element)
if os.path.isdir(found_element) and "googletagmanager" not in found_element:
children.append(found_element)
elif found_element.endswith("index.html") or found_element.endswith("index.htm"):#The entrance html file to the webpage needs to be called index.html or index.htm otherwise it won't add the JS code to read load times
if("index.html" in os.listdir(parent) and not(found_element.endswith("index.html"))):
continue
else:
return found_element
paths = children
return root
def add_js(directory, ip):
for cat in os.listdir(directory):
if cat == ".DS_Store" or cat == "RUN_WEBAPPS_README.md" or cat == "START_SERVER_IN_THIS_DIR":
continue
for site in os.listdir(directory + cat):
if site == ".DS_Store":
continue
path_to_html = bfs_dirs('{}{}/{}'.format(directory, cat, site))
if(path_to_html == '{}{}/{}'.format(directory, cat, site)):
continue
path_to_src = os.path.dirname(path_to_html)
print(path_to_html)
soup = BeautifulSoup(open(path_to_html, 'rb'), "lxml")
if(soup.find('head')):
perfumeSource = soup.new_tag('script')
perfumeSource['src'] ="/node_modules/perfume.js/dist/perfume.umd.min.js"
script = soup.new_tag('script')
script.string = "perfumeResults = []; function xml_http_post(url, data, callback) {var req = new XMLHttpRequest(); req.open(\"POST\", url, true); req.send(data);} const perfume = new Perfume({ analyticsTracker: (options) => { const { metricName, data, eventProperties, navigatorInformation } = options; perfumeResults.push(options); } }); function load_log() { setTimeout(function(){ objectToSend = \"{'perfumeResults':\"+JSON.stringify(perfumeResults)+\"}\"; xml_http_post(\""+ip+"\",objectToSend,null); },5000); };window.addEventListener ?window.addEventListener(\"load\",load_log, true) : window.attachEvent && window.attachEvent(\"onload\", load_log);"
soup.head.insert(0, perfumeSource)
soup.head.insert(1,script)
with open(path_to_html, "w") as file:
file.write(str(soup))
if __name__ == '__main__':
#USAGE: (python3 AddJS.py path/To/Directory/With/All/WebApplication/ IPADRRESS)
#NOTE: IPADDRESS should be in the form of http://IP:8080/ and IP is the ip address
add_js(sys.argv[1], sys.argv[2])
| 2,662 | 0 | 50 |
d207293da2f82b7baf5c06a2e00d92a54a8cf4dd | 1,879 | py | Python | las_test.py | paceholder/pylasdev | d63701d02d3a5f59de7f3ff33804d2f421f49402 | [
"BSD-3-Clause"
] | 2 | 2017-01-05T13:27:30.000Z | 2018-03-26T11:49:28.000Z | las_test.py | paceholder/pylasdev | d63701d02d3a5f59de7f3ff33804d2f421f49402 | [
"BSD-3-Clause"
] | null | null | null | las_test.py | paceholder/pylasdev | d63701d02d3a5f59de7f3ff33804d2f421f49402 | [
"BSD-3-Clause"
] | null | null | null | from pylasdev import *
#from mnem_base import *
# Simple test with full LAS output for LAS reader
test_files = [
# LAS 1.2
"test_data/sample.las", # 0 - OK
"test_data/sample_minimal.las", # 1 - OK
"test_data/sample_wrapped.las", # 2 - OK
"test_data/sample_curve_api.las", # 3 - OK
# "test_data/sample_big.las", # 2 Mb - OK -- no pickle test
# "test_data/sample_very_big.las", # 12 Mb - OK -- no pickle test
# LAS 2.0
"test_data/sample_2.0.las", # 4 - OK
"test_data/sample_2.0_minimal.las", # 5 - OK
"test_data/sample_2.0_wrapped.las", # 6 - OK
"test_data/sample_2.0_based.las", # 7 - OK
"test_data/petrel2.0.las", # 8 -OK
"test_data/5_1.las", # andrey converter # 9 - OK
"test_data/4ALS.las", # nadezhdin's output # 10 - OK
"test_data/1475IBK3.las" # bashneft cp866 # 11
# "test_data/comment_test.las",
# LAS 3.0 # not implemented
# "test_data/sample_3.0.las", #
]
file = test_files[8]
print "Reading file ", file, " ..."
#las_info = read_las_file(file, mnem_base)
las_info = read_las_file(file)
print "Done."
if(las_info is None):
print "Error, file not readed!"
else:
print "=== Version: "
for key in las_info['version'].keys():
print " ", [key], [las_info['version'][key]]
print "=== Well:"
for key in las_info['well'].keys():
print " ", [key], [las_info['well'][key]]
print "=== Parameters:"
for key in las_info['parameters'].keys():
print " ", key, [key], [las_info['parameters'][key]]
print "=== Curves:"
for k in xrange(len(las_info['curves_order'])):
print " ", k, las_info['curves_order'][k]
print "=== Logs:"
for key_ordered in las_info['curves_order']:
print " ", key_ordered, [key_ordered], [las_info['logs'][key_ordered]]
# writing test
#filename = "write_test.las"
#write_las_file(filename, las_info)
| 24.402597 | 75 | 0.624268 | from pylasdev import *
#from mnem_base import *
# Simple test with full LAS output for LAS reader
test_files = [
# LAS 1.2
"test_data/sample.las", # 0 - OK
"test_data/sample_minimal.las", # 1 - OK
"test_data/sample_wrapped.las", # 2 - OK
"test_data/sample_curve_api.las", # 3 - OK
# "test_data/sample_big.las", # 2 Mb - OK -- no pickle test
# "test_data/sample_very_big.las", # 12 Mb - OK -- no pickle test
# LAS 2.0
"test_data/sample_2.0.las", # 4 - OK
"test_data/sample_2.0_minimal.las", # 5 - OK
"test_data/sample_2.0_wrapped.las", # 6 - OK
"test_data/sample_2.0_based.las", # 7 - OK
"test_data/petrel2.0.las", # 8 -OK
"test_data/5_1.las", # andrey converter # 9 - OK
"test_data/4ALS.las", # nadezhdin's output # 10 - OK
"test_data/1475IBK3.las" # bashneft cp866 # 11
# "test_data/comment_test.las",
# LAS 3.0 # not implemented
# "test_data/sample_3.0.las", #
]
file = test_files[8]
print "Reading file ", file, " ..."
#las_info = read_las_file(file, mnem_base)
las_info = read_las_file(file)
print "Done."
if(las_info is None):
print "Error, file not readed!"
else:
print "=== Version: "
for key in las_info['version'].keys():
print " ", [key], [las_info['version'][key]]
print "=== Well:"
for key in las_info['well'].keys():
print " ", [key], [las_info['well'][key]]
print "=== Parameters:"
for key in las_info['parameters'].keys():
print " ", key, [key], [las_info['parameters'][key]]
print "=== Curves:"
for k in xrange(len(las_info['curves_order'])):
print " ", k, las_info['curves_order'][k]
print "=== Logs:"
for key_ordered in las_info['curves_order']:
print " ", key_ordered, [key_ordered], [las_info['logs'][key_ordered]]
# writing test
#filename = "write_test.las"
#write_las_file(filename, las_info)
| 0 | 0 | 0 |
b1666e2f97dc00556483a7e6a3be95c2b359059f | 6,873 | py | Python | Playable/full builds/0.1 movement build/Corak/Main_menu.py | LinearHaru/Corak | 269f827131cbdaf15e6cd81bee792fbb6b9c7606 | [
"MIT"
] | null | null | null | Playable/full builds/0.1 movement build/Corak/Main_menu.py | LinearHaru/Corak | 269f827131cbdaf15e6cd81bee792fbb6b9c7606 | [
"MIT"
] | null | null | null | Playable/full builds/0.1 movement build/Corak/Main_menu.py | LinearHaru/Corak | 269f827131cbdaf15e6cd81bee792fbb6b9c7606 | [
"MIT"
] | null | null | null | import pygame
from Corak.dicio import *
from Corak.mapread4 import mapread
screen = pygame.display.set_mode((ScreenX, ScreenY))
txt = "LRUDhallX.txt"
sala = openSala(txt)
tiles = mapread(sala)
tiles2 = pygame.sprite.Group()
size = getSize(sala)
offsets = getOffset(sala, size)
player1 = Corak()
print(tiles)
left = False
right = False
b = 10
a = 0
c = 0
player1.checkMap()
for tile in tiles:
if tile.type == 2:
tiles2.add(tile)
tiles.remove(tile)
temp = 0
clock = pygame.time.Clock()
while True:
for event in pygame.event.get():
if event.type == pygame.QUIT:
pygame.quit()
if event.type == pygame.MOUSEBUTTONDOWN:
if event.button == 1:
player1.Do = [0, -1, "U"]
player1.update(player1.Do, player1.turn)
elif event.type == pygame.KEYDOWN:
if event.key == pygame.K_w:
player1.Do = [0, -1, "U"]
elif event.key == pygame.K_a:
player1.Do = [-1, 0, "L"]
elif event.key == pygame.K_d:
player1.Do = [1, 0, "R"]
elif event.key == pygame.K_s:
player1.Do = [0, 0, "X"]
elif event.key == pygame.K_z:
player1.Do = [-1, 0, "H"]
elif event.key == pygame.K_x:
player1.Do = [1, 0, "H"]
elif event.key == pygame.K_UP:
player1.Do = [0, -1, "U"]
elif event.key == pygame.K_LEFT:
player1.Do = [-1, 0, "L"]
elif event.key == pygame.K_RIGHT:
player1.Do = [1, 0, "R"]
elif event.key == pygame.K_DOWN:
player1.Do = [0, 0, "X"]
elif event.key == pygame.K_r:
player1.turn = False
elif event.key == pygame.K_SPACE:
player1.focus = True
player1.cantGo = []
break
else:
player1.Do = [0, 0, "Z"]
player1.update(player1.Do, player1.turn)
mousex, mousey = pygame.mouse.get_pos()
'''
if left and a > -max_speed:
a -= aceeleration
if right and a < max_speed:
a += aceeleration
if not left and not right:
a = a/friction
b = b+a
c = c+gravity
'''
pygame.display.set_caption(str(round(clock.get_fps())))
clock.tick(FPS)
if temp < 50: temp += 1
else: temp = 0
screen.fill(BLACK)
tiles.draw(screen)
tiles2.draw(screen)
pygame.draw.rect(screen, RED, [mousex, mousey, temp, 10])
#pygame.draw.rect(screen, BLACK, [int(b), c, 10, 10])
pygame.draw.rect(screen, BLUE, [player1.pos[0], player1.pos[1], size, size])
pygame.display.flip()
pygame.quit()
| 29.753247 | 131 | 0.474465 | import pygame
from Corak.dicio import *
from Corak.mapread4 import mapread
class Corak(pygame.sprite.Sprite):
def __init__(self):
self.image = pygame.Surface([size, size])
self.image.fill(BLACK)
self.rect = self.image.get_rect()
self.at = [1, 11]
self.pos = [size * self.at[0] + offsets[0], size * self.at[1] + offsets[1]]
self.Do = []
self.turn = True
self.jump = False
self.air = False
self.focus = False
self.grab = False
def update(self, Do, turn):
if turn:
if Do[2] in "LRUD": self.Move(Do)
if Do[2] in "H": self.Hit(Do[0], self.at, self.focus, self.air)
if Do[2] in "X": self.Focus(self.focus, self.air)
def Focus(self, focus, air):
if not focus and not air:
self.focus = True
self.turn = False
self.checkMap()
def Hit(self, direction, at, focus, air):
try:
if str(sala[at[1]][at[0] + direction]) in "2":
sala[at[1]] = sala[at[1]][:at[0] + direction:] + "0" + sala[at[1]][-(len(sala[at[1]]) - (direction + at[0] + 1))::]
for tile in tiles2:
if tile.at == [at[0] + direction, at[1]]:
tiles2.remove(tile)
if focus:
self.focus = False
print("focus")
if air: self.air = False
except:
pass
print("hit")
self.turn = False
self.checkMap()
def CheckLevel(self, at):
if -1 >= at[0]:
print("levelleft")
elif len(sala[0]) == at[0]:
print("levelright")
elif -1 >= at[1]:
print("levelup")
elif len(sala) == at[1]:
print("leveldown")
else: self.checkMap()
def checkMap(self):
self.cantGo = []
if self.focus : distance = 2
else : distance = 1
for i in range(1, distance+1):
try:
if sala[self.at[1]][self.at[0] + i] in "12":
self.cantGo.extend(["R", i*"R"])
except: pass
try:
if self.at[0] - 1 >= 0:
if sala[self.at[1]][self.at[0] - i] in "12":
print("a")
self.cantGo.extend(["L", i*"L"])
except:
pass
try:
if sala[self.at[1] + 1][self.at[0]] in "123":
self.cantGo.append("D")
except: pass
try:
if self.at[1] - 1 >= 0:
if sala[self.at[1] - i][self.at[0]] in "123":
self.cantGo.extend(["U", i*"U"])
except: pass
self.checkMove()
def checkMove(self):
self.air = True
if "D" in self.cantGo:
self.air = False
if self.turn:
if self.air and not self.grab:
self.cantGo.append("U")
if not self.turn:
if not any([self.jump, self.grab, self.focus]) and self.air:
self.Do = [0, 1, "D"]
self.jump = True
self.Move(self.Do)
if self.jump:
self.jump = False
self.turn = True
self.checkMove()
print(self.cantGo)
def Move(self, Do):
alltrue = True
for direction in self.cantGo:
if direction in Do[2]:
alltrue = False
if self.air and direction in "LR": self.grab = True, print("grab")
if self.focus and direction in "LRU":
self.focus = False
alltrue = True
if alltrue:
if Do[1] == -1: self.jump = True, print("jump")
if self.focus : Do[0] *= 2
if self.focus : Do[1] *= 2
self.focus = False
self.grab = False
self.at[0] = self.at[0] + Do[0]
self.at[1] = self.at[1] + Do[1]
self.pos[0] = size * self.at[0] + offsets[0]
self.pos[1] = size * self.at[1] + offsets[1]
Do = []
self.turn = False
self.CheckLevel(self.at)
screen = pygame.display.set_mode((ScreenX, ScreenY))
txt = "LRUDhallX.txt"
sala = openSala(txt)
tiles = mapread(sala)
tiles2 = pygame.sprite.Group()
size = getSize(sala)
offsets = getOffset(sala, size)
player1 = Corak()
print(tiles)
left = False
right = False
b = 10
a = 0
c = 0
player1.checkMap()
for tile in tiles:
if tile.type == 2:
tiles2.add(tile)
tiles.remove(tile)
temp = 0
clock = pygame.time.Clock()
while True:
for event in pygame.event.get():
if event.type == pygame.QUIT:
pygame.quit()
if event.type == pygame.MOUSEBUTTONDOWN:
if event.button == 1:
player1.Do = [0, -1, "U"]
player1.update(player1.Do, player1.turn)
elif event.type == pygame.KEYDOWN:
if event.key == pygame.K_w:
player1.Do = [0, -1, "U"]
elif event.key == pygame.K_a:
player1.Do = [-1, 0, "L"]
elif event.key == pygame.K_d:
player1.Do = [1, 0, "R"]
elif event.key == pygame.K_s:
player1.Do = [0, 0, "X"]
elif event.key == pygame.K_z:
player1.Do = [-1, 0, "H"]
elif event.key == pygame.K_x:
player1.Do = [1, 0, "H"]
elif event.key == pygame.K_UP:
player1.Do = [0, -1, "U"]
elif event.key == pygame.K_LEFT:
player1.Do = [-1, 0, "L"]
elif event.key == pygame.K_RIGHT:
player1.Do = [1, 0, "R"]
elif event.key == pygame.K_DOWN:
player1.Do = [0, 0, "X"]
elif event.key == pygame.K_r:
player1.turn = False
elif event.key == pygame.K_SPACE:
player1.focus = True
player1.cantGo = []
break
else:
player1.Do = [0, 0, "Z"]
player1.update(player1.Do, player1.turn)
mousex, mousey = pygame.mouse.get_pos()
'''
if left and a > -max_speed:
a -= aceeleration
if right and a < max_speed:
a += aceeleration
if not left and not right:
a = a/friction
b = b+a
c = c+gravity
'''
pygame.display.set_caption(str(round(clock.get_fps())))
clock.tick(FPS)
if temp < 50: temp += 1
else: temp = 0
screen.fill(BLACK)
tiles.draw(screen)
tiles2.draw(screen)
pygame.draw.rect(screen, RED, [mousex, mousey, temp, 10])
#pygame.draw.rect(screen, BLACK, [int(b), c, 10, 10])
pygame.draw.rect(screen, BLUE, [player1.pos[0], player1.pos[1], size, size])
pygame.display.flip()
pygame.quit()
| 3,892 | 13 | 238 |
9ae35fd2f861db90faa296b2c3cb168bd8403548 | 6,842 | py | Python | RasPi_Dev/ros_ws/build/vslam/rtabmap_ros/catkin_generated/pkg.installspace.context.pc.py | QianheYu/xtark_driver_dev | 1708888161cf20c0d1f45c99d0da4467d69c26c8 | [
"BSD-3-Clause"
] | 1 | 2022-03-11T03:31:15.000Z | 2022-03-11T03:31:15.000Z | RasPi_Dev/ros_ws/build/vslam/rtabmap_ros/catkin_generated/pkg.installspace.context.pc.py | bravetree/xtark_driver_dev | 1708888161cf20c0d1f45c99d0da4467d69c26c8 | [
"BSD-3-Clause"
] | null | null | null | RasPi_Dev/ros_ws/build/vslam/rtabmap_ros/catkin_generated/pkg.installspace.context.pc.py | bravetree/xtark_driver_dev | 1708888161cf20c0d1f45c99d0da4467d69c26c8 | [
"BSD-3-Clause"
] | null | null | null | # generated from catkin/cmake/template/pkg.context.pc.in
CATKIN_PACKAGE_PREFIX = ""
PROJECT_PKG_CONFIG_INCLUDE_DIRS = "/home/xtark/ros_ws/install/include;/home/xtark/ros_ws/devel/lib/rtabmap-0.19/../../include/rtabmap-0.19;/opt/ros/kinetic/include/opencv-3.3.1-dev;/opt/ros/kinetic/include/opencv-3.3.1-dev/opencv".split(';') if "/home/xtark/ros_ws/install/include;/home/xtark/ros_ws/devel/lib/rtabmap-0.19/../../include/rtabmap-0.19;/opt/ros/kinetic/include/opencv-3.3.1-dev;/opt/ros/kinetic/include/opencv-3.3.1-dev/opencv" != "" else []
PROJECT_CATKIN_DEPENDS = "cv_bridge;roscpp;rospy;sensor_msgs;std_msgs;std_srvs;nav_msgs;geometry_msgs;visualization_msgs;image_transport;tf;tf_conversions;tf2_ros;eigen_conversions;laser_geometry;pcl_conversions;pcl_ros;nodelet;dynamic_reconfigure;message_filters;class_loader;rosgraph_msgs;stereo_msgs;move_base_msgs;image_geometry;costmap_2d;rviz".replace(';', ' ')
PKG_CONFIG_LIBRARIES_WITH_PREFIX = "-lrtabmap_ros;/home/xtark/ros_ws/devel/lib/librtabmap_core.so;/home/xtark/ros_ws/devel/lib/librtabmap_utilite.so;/home/xtark/ros_ws/devel/lib/librtabmap_gui.so;/usr/lib/arm-linux-gnueabihf/libz.so;/usr/local/lib/libg2o_core.so;/usr/local/lib/libg2o_types_slam2d.so;/usr/local/lib/libg2o_types_slam3d.so;/usr/local/lib/libg2o_types_sba.so;/usr/local/lib/libg2o_stuff.so;/usr/local/lib/libg2o_solver_csparse.so;/usr/local/lib/libg2o_csparse_extension.so;/usr/lib/arm-linux-gnueabihf/libcxsparse.so;/usr/local/lib/libg2o_solver_cholmod.so;/usr/lib/arm-linux-gnueabihf/libcholmod.so;/usr/lib/libOpenNI2.so;/opt/ros/kinetic/lib/liboctomap.so;/opt/ros/kinetic/lib/liboctomath.so;/opt/ros/kinetic/lib/libopencv_calib3d3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_core3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_dnn3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_features2d3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_flann3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_highgui3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_imgcodecs3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_imgproc3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_ml3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_objdetect3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_photo3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_shape3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_stitching3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_superres3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_video3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_videoio3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_videostab3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_viz3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_aruco3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_bgsegm3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_bioinspired3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_ccalib3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_cvv3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_datasets3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_dpm3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_face3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_fuzzy3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_hdf3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_img_hash3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_line_descriptor3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_optflow3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_phase_unwrapping3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_plot3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_reg3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_rgbd3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_saliency3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_stereo3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_structured_light3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_surface_matching3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_text3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_tracking3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_xfeatures2d3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_ximgproc3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_xobjdetect3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_xphoto3.so.3.3.1".split(';') if "-lrtabmap_ros;/home/xtark/ros_ws/devel/lib/librtabmap_core.so;/home/xtark/ros_ws/devel/lib/librtabmap_utilite.so;/home/xtark/ros_ws/devel/lib/librtabmap_gui.so;/usr/lib/arm-linux-gnueabihf/libz.so;/usr/local/lib/libg2o_core.so;/usr/local/lib/libg2o_types_slam2d.so;/usr/local/lib/libg2o_types_slam3d.so;/usr/local/lib/libg2o_types_sba.so;/usr/local/lib/libg2o_stuff.so;/usr/local/lib/libg2o_solver_csparse.so;/usr/local/lib/libg2o_csparse_extension.so;/usr/lib/arm-linux-gnueabihf/libcxsparse.so;/usr/local/lib/libg2o_solver_cholmod.so;/usr/lib/arm-linux-gnueabihf/libcholmod.so;/usr/lib/libOpenNI2.so;/opt/ros/kinetic/lib/liboctomap.so;/opt/ros/kinetic/lib/liboctomath.so;/opt/ros/kinetic/lib/libopencv_calib3d3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_core3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_dnn3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_features2d3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_flann3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_highgui3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_imgcodecs3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_imgproc3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_ml3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_objdetect3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_photo3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_shape3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_stitching3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_superres3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_video3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_videoio3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_videostab3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_viz3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_aruco3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_bgsegm3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_bioinspired3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_ccalib3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_cvv3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_datasets3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_dpm3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_face3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_fuzzy3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_hdf3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_img_hash3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_line_descriptor3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_optflow3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_phase_unwrapping3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_plot3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_reg3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_rgbd3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_saliency3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_stereo3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_structured_light3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_surface_matching3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_text3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_tracking3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_xfeatures2d3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_ximgproc3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_xobjdetect3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_xphoto3.so.3.3.1" != "" else []
PROJECT_NAME = "rtabmap_ros"
PROJECT_SPACE_DIR = "/home/xtark/ros_ws/install"
PROJECT_VERSION = "0.19.3"
| 760.222222 | 5,828 | 0.80152 | # generated from catkin/cmake/template/pkg.context.pc.in
CATKIN_PACKAGE_PREFIX = ""
PROJECT_PKG_CONFIG_INCLUDE_DIRS = "/home/xtark/ros_ws/install/include;/home/xtark/ros_ws/devel/lib/rtabmap-0.19/../../include/rtabmap-0.19;/opt/ros/kinetic/include/opencv-3.3.1-dev;/opt/ros/kinetic/include/opencv-3.3.1-dev/opencv".split(';') if "/home/xtark/ros_ws/install/include;/home/xtark/ros_ws/devel/lib/rtabmap-0.19/../../include/rtabmap-0.19;/opt/ros/kinetic/include/opencv-3.3.1-dev;/opt/ros/kinetic/include/opencv-3.3.1-dev/opencv" != "" else []
PROJECT_CATKIN_DEPENDS = "cv_bridge;roscpp;rospy;sensor_msgs;std_msgs;std_srvs;nav_msgs;geometry_msgs;visualization_msgs;image_transport;tf;tf_conversions;tf2_ros;eigen_conversions;laser_geometry;pcl_conversions;pcl_ros;nodelet;dynamic_reconfigure;message_filters;class_loader;rosgraph_msgs;stereo_msgs;move_base_msgs;image_geometry;costmap_2d;rviz".replace(';', ' ')
PKG_CONFIG_LIBRARIES_WITH_PREFIX = "-lrtabmap_ros;/home/xtark/ros_ws/devel/lib/librtabmap_core.so;/home/xtark/ros_ws/devel/lib/librtabmap_utilite.so;/home/xtark/ros_ws/devel/lib/librtabmap_gui.so;/usr/lib/arm-linux-gnueabihf/libz.so;/usr/local/lib/libg2o_core.so;/usr/local/lib/libg2o_types_slam2d.so;/usr/local/lib/libg2o_types_slam3d.so;/usr/local/lib/libg2o_types_sba.so;/usr/local/lib/libg2o_stuff.so;/usr/local/lib/libg2o_solver_csparse.so;/usr/local/lib/libg2o_csparse_extension.so;/usr/lib/arm-linux-gnueabihf/libcxsparse.so;/usr/local/lib/libg2o_solver_cholmod.so;/usr/lib/arm-linux-gnueabihf/libcholmod.so;/usr/lib/libOpenNI2.so;/opt/ros/kinetic/lib/liboctomap.so;/opt/ros/kinetic/lib/liboctomath.so;/opt/ros/kinetic/lib/libopencv_calib3d3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_core3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_dnn3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_features2d3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_flann3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_highgui3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_imgcodecs3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_imgproc3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_ml3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_objdetect3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_photo3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_shape3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_stitching3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_superres3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_video3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_videoio3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_videostab3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_viz3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_aruco3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_bgsegm3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_bioinspired3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_ccalib3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_cvv3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_datasets3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_dpm3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_face3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_fuzzy3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_hdf3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_img_hash3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_line_descriptor3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_optflow3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_phase_unwrapping3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_plot3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_reg3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_rgbd3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_saliency3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_stereo3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_structured_light3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_surface_matching3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_text3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_tracking3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_xfeatures2d3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_ximgproc3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_xobjdetect3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_xphoto3.so.3.3.1".split(';') if "-lrtabmap_ros;/home/xtark/ros_ws/devel/lib/librtabmap_core.so;/home/xtark/ros_ws/devel/lib/librtabmap_utilite.so;/home/xtark/ros_ws/devel/lib/librtabmap_gui.so;/usr/lib/arm-linux-gnueabihf/libz.so;/usr/local/lib/libg2o_core.so;/usr/local/lib/libg2o_types_slam2d.so;/usr/local/lib/libg2o_types_slam3d.so;/usr/local/lib/libg2o_types_sba.so;/usr/local/lib/libg2o_stuff.so;/usr/local/lib/libg2o_solver_csparse.so;/usr/local/lib/libg2o_csparse_extension.so;/usr/lib/arm-linux-gnueabihf/libcxsparse.so;/usr/local/lib/libg2o_solver_cholmod.so;/usr/lib/arm-linux-gnueabihf/libcholmod.so;/usr/lib/libOpenNI2.so;/opt/ros/kinetic/lib/liboctomap.so;/opt/ros/kinetic/lib/liboctomath.so;/opt/ros/kinetic/lib/libopencv_calib3d3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_core3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_dnn3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_features2d3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_flann3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_highgui3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_imgcodecs3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_imgproc3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_ml3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_objdetect3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_photo3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_shape3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_stitching3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_superres3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_video3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_videoio3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_videostab3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_viz3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_aruco3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_bgsegm3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_bioinspired3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_ccalib3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_cvv3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_datasets3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_dpm3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_face3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_fuzzy3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_hdf3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_img_hash3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_line_descriptor3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_optflow3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_phase_unwrapping3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_plot3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_reg3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_rgbd3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_saliency3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_stereo3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_structured_light3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_surface_matching3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_text3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_tracking3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_xfeatures2d3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_ximgproc3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_xobjdetect3.so.3.3.1;/opt/ros/kinetic/lib/libopencv_xphoto3.so.3.3.1" != "" else []
PROJECT_NAME = "rtabmap_ros"
PROJECT_SPACE_DIR = "/home/xtark/ros_ws/install"
PROJECT_VERSION = "0.19.3"
| 0 | 0 | 0 |
8a0630d38ab753c8a6ada60da69a9057613b5bc6 | 5,469 | py | Python | rbac/acl.py | eldorplus/simple-rbac | c7ca8ce6edb64797bbe43a0dfb11d686c2f5acf1 | [
"MIT"
] | 219 | 2015-01-15T23:56:29.000Z | 2022-03-27T19:56:47.000Z | rbac/acl.py | eldorplus/simple-rbac | c7ca8ce6edb64797bbe43a0dfb11d686c2f5acf1 | [
"MIT"
] | 6 | 2015-09-15T03:12:56.000Z | 2020-09-01T13:39:51.000Z | rbac/acl.py | eldorplus/simple-rbac | c7ca8ce6edb64797bbe43a0dfb11d686c2f5acf1 | [
"MIT"
] | 63 | 2015-03-15T05:36:21.000Z | 2021-07-22T12:40:49.000Z | from __future__ import absolute_import
import itertools
__all__ = ["Registry"]
class Registry(object):
"""The registry of access control list."""
def add_role(self, role, parents=[]):
"""Add a role or append parents roles to a special role.
All added roles should be hashable.
(http://docs.python.org/glossary.html#term-hashable)
"""
self._roles.setdefault(role, set())
self._roles[role].update(parents)
for p in parents:
self._children.setdefault(p, set())
self._children[p].add(role)
# all roles start as deny-only (unless one of its parents
# isn't deny-only)
if not parents or self._roles_are_deny_only(parents):
self._denial_only_roles.add(role)
def add_resource(self, resource, parents=[]):
"""Add a resource or append parents resources to a special resource.
All added resources should be hashable.
(http://docs.python.org/glossary.html#term-hashable)
"""
self._resources.setdefault(resource, set())
self._resources[resource].update(parents)
def allow(self, role, operation, resource, assertion=None):
"""Add a allowed rule.
The added rule will allow the role and its all children roles to
operate the resource.
"""
assert not role or role in self._roles
assert not resource or resource in self._resources
self._allowed[role, operation, resource] = assertion
# since we just allowed a permission, role and any children aren't
# denied-only
for r in itertools.chain([role], get_family(self._children, role)):
self._denial_only_roles.discard(r)
def deny(self, role, operation, resource, assertion=None):
"""Add a denied rule.
The added rule will deny the role and its all children roles to
operate the resource.
"""
assert not role or role in self._roles
assert not resource or resource in self._resources
self._denied[role, operation, resource] = assertion
def is_allowed(self, role, operation, resource, check_allowed=True,
**assertion_kwargs):
"""Check the permission.
If the access is denied, this method will return False; if the access
is allowed, this method will return True; if there is not any rule
for the access, this method will return None.
"""
assert not role or role in self._roles
assert not resource or resource in self._resources
roles = set(get_family(self._roles, role))
operations = {None, operation}
resources = set(get_family(self._resources, resource))
is_allowed = None
default_assertion = DefaultAssertion
for permission in itertools.product(roles, operations, resources):
if permission in self._denied:
assertion = self._denied[permission] or default_assertion
if assertion(self, role, operation, resource,
**assertion_kwargs):
return False # denied by rule immediately
if check_allowed and permission in self._allowed:
assertion = self._allowed[permission] or default_assertion
if assertion(self, role, operation, resource,
**assertion_kwargs):
is_allowed = True # allowed by rule
return is_allowed
def is_any_allowed(self, roles, operation, resource, **assertion_kwargs):
"""Check the permission with many roles."""
is_allowed = None # no matching rules
for i, role in enumerate(roles):
# if access not yet allowed and all remaining roles could
# only deny access, short-circuit and return False
if not is_allowed and self._roles_are_deny_only(roles[i:]):
return False
check_allowed = not is_allowed
# if another role gave access,
# don't bother checking if this one is allowed
is_current_allowed = self.is_allowed(role, operation, resource,
check_allowed=check_allowed,
**assertion_kwargs)
if is_current_allowed is False:
return False # denied by rule
elif is_current_allowed is True:
is_allowed = True
return is_allowed
def get_family(all_parents, current):
"""Iterate current object and its all parents recursively."""
yield current
for parent in get_parents(all_parents, current):
yield parent
yield None
def get_parents(all_parents, current):
"""Iterate current object's all parents."""
for parent in all_parents.get(current, []):
yield parent
for grandparent in get_parents(all_parents, parent):
yield grandparent
| 36.46 | 77 | 0.619492 | from __future__ import absolute_import
import itertools
__all__ = ["Registry"]
class Registry(object):
"""The registry of access control list."""
def __init__(self):
self._roles = {}
self._resources = {}
self._allowed = {}
self._denied = {}
# to allow additional short circuiting, track roles that only
# ever deny access
self._denial_only_roles = set()
self._children = {}
def add_role(self, role, parents=[]):
"""Add a role or append parents roles to a special role.
All added roles should be hashable.
(http://docs.python.org/glossary.html#term-hashable)
"""
self._roles.setdefault(role, set())
self._roles[role].update(parents)
for p in parents:
self._children.setdefault(p, set())
self._children[p].add(role)
# all roles start as deny-only (unless one of its parents
# isn't deny-only)
if not parents or self._roles_are_deny_only(parents):
self._denial_only_roles.add(role)
def add_resource(self, resource, parents=[]):
"""Add a resource or append parents resources to a special resource.
All added resources should be hashable.
(http://docs.python.org/glossary.html#term-hashable)
"""
self._resources.setdefault(resource, set())
self._resources[resource].update(parents)
def allow(self, role, operation, resource, assertion=None):
"""Add a allowed rule.
The added rule will allow the role and its all children roles to
operate the resource.
"""
assert not role or role in self._roles
assert not resource or resource in self._resources
self._allowed[role, operation, resource] = assertion
# since we just allowed a permission, role and any children aren't
# denied-only
for r in itertools.chain([role], get_family(self._children, role)):
self._denial_only_roles.discard(r)
def deny(self, role, operation, resource, assertion=None):
"""Add a denied rule.
The added rule will deny the role and its all children roles to
operate the resource.
"""
assert not role or role in self._roles
assert not resource or resource in self._resources
self._denied[role, operation, resource] = assertion
def is_allowed(self, role, operation, resource, check_allowed=True,
**assertion_kwargs):
"""Check the permission.
If the access is denied, this method will return False; if the access
is allowed, this method will return True; if there is not any rule
for the access, this method will return None.
"""
assert not role or role in self._roles
assert not resource or resource in self._resources
roles = set(get_family(self._roles, role))
operations = {None, operation}
resources = set(get_family(self._resources, resource))
def DefaultAssertion(*args, **kwargs):
return True
is_allowed = None
default_assertion = DefaultAssertion
for permission in itertools.product(roles, operations, resources):
if permission in self._denied:
assertion = self._denied[permission] or default_assertion
if assertion(self, role, operation, resource,
**assertion_kwargs):
return False # denied by rule immediately
if check_allowed and permission in self._allowed:
assertion = self._allowed[permission] or default_assertion
if assertion(self, role, operation, resource,
**assertion_kwargs):
is_allowed = True # allowed by rule
return is_allowed
def is_any_allowed(self, roles, operation, resource, **assertion_kwargs):
"""Check the permission with many roles."""
is_allowed = None # no matching rules
for i, role in enumerate(roles):
# if access not yet allowed and all remaining roles could
# only deny access, short-circuit and return False
if not is_allowed and self._roles_are_deny_only(roles[i:]):
return False
check_allowed = not is_allowed
# if another role gave access,
# don't bother checking if this one is allowed
is_current_allowed = self.is_allowed(role, operation, resource,
check_allowed=check_allowed,
**assertion_kwargs)
if is_current_allowed is False:
return False # denied by rule
elif is_current_allowed is True:
is_allowed = True
return is_allowed
def _roles_are_deny_only(self, roles):
return all(r in self._denial_only_roles for r in roles)
def get_family(all_parents, current):
"""Iterate current object and its all parents recursively."""
yield current
for parent in get_parents(all_parents, current):
yield parent
yield None
def get_parents(all_parents, current):
"""Iterate current object's all parents."""
for parent in all_parents.get(current, []):
yield parent
for grandparent in get_parents(all_parents, parent):
yield grandparent
| 393 | 0 | 85 |
35ecdb8c51f52d9bb0708b4c0142ded1b98d2a57 | 23,704 | py | Python | ceph_report/collectors.py | Mirantis/ceph-monitoring | 2de4a90fb102337606732a46a382505047bfa1aa | [
"Apache-2.0"
] | 5 | 2016-09-19T10:42:15.000Z | 2018-09-05T11:18:03.000Z | ceph_report/collectors.py | Mirantis/ceph-monitoring | 2de4a90fb102337606732a46a382505047bfa1aa | [
"Apache-2.0"
] | 2 | 2016-10-18T09:15:19.000Z | 2018-03-21T10:24:54.000Z | ceph_report/collectors.py | Mirantis/ceph-monitoring | 2de4a90fb102337606732a46a382505047bfa1aa | [
"Apache-2.0"
] | 9 | 2016-10-17T13:20:01.000Z | 2019-01-09T14:33:48.000Z | from __future__ import annotations
import os
import re
import time
import json
import array
import random
import asyncio
import logging
import datetime
import contextlib
import collections
from enum import IntEnum
from dataclasses import dataclass, field
from typing import Any, List, Dict, Optional, Tuple, TypeVar, Callable, Coroutine, Union, cast
from aiorpc import IAIORPCNode, ConnectionPool
from cephlib import CephRelease, parse_ceph_volumes_js, parse_ceph_disk_js, CephReport, OSDDevInfo, OSDBSDevices, \
OSDFSDevices
from koder_utils import (IStorage, CMDResult, parse_devices_tree, collect_process_info, get_host_interfaces,
ignore_all, IAsyncNode)
logger = logging.getLogger('collect')
T = TypeVar('T')
VT = TypeVar('VT', str, bytes, array.array)
@dataclass
class Collector:
"""Base class for data collectors. Can collect data for only one node."""
storage: IStorage
hostname: Optional[str]
opts: Any
pool: ConnectionPool
pretty_json: bool = field(init=False, default=False)
cmds: Dict[str, Tuple[str, StorFormat]] = field(init=False, default_factory=dict)
@contextlib.asynccontextmanager
def chdir(self: T, path: str) -> T:
"""Chdir for point in storage tree, where current results are stored"""
return self.with_storage(self.storage.sub_storage(path)) # type: ignore
def save(self, path: str, fmt: StorFormat, code: int, data: VT, extra: List[str] = None) -> VT:
"""Save results into storage"""
rpath = f"{path}.{fmt.name if code == 0 else 'err'}"
if isinstance(data, array.array):
self.storage.put_array(rpath, data, extra if extra else [])
elif isinstance(data, (str, bytes)):
pretty_data: Union[str, bytes] = data
if code == 0 and fmt == StorFormat.json and self.pretty_json:
assert isinstance(data, (str, bytes))
try:
dt = json.loads(data)
except json.JSONDecodeError as exc:
logger.error(f"Failed to prettify json data for path {path}. Err {exc}. Saving as is")
else:
pretty_data = json.dumps(dt, indent=4, sort_keys=True)
data_b: bytes = pretty_data.encode() if isinstance(pretty_data, str) else cast(bytes, pretty_data)
assert extra is None
self.save_raw(rpath, data_b)
else:
raise TypeError(f"Can't save value of type {type(data)!r} (to {rpath!r})")
return data
async def read_and_save(self, path: str, file_path: str, fmt: StorFormat = StorFormat.txt,
compress: bool = True) -> bytes:
"""Download file from node and save it into storage"""
async with self.connection() as conn:
try:
content = await conn.read(file_path, compress=compress)
code = 0
except (IOError, RuntimeError) as exc:
logger.warning(f"Can't get file {file_path!r} from node {self.hostname}. {exc}")
content = str(exc) # type: ignore
code = 1
self.save(path, fmt, code, content)
return content if code == 0 else None # type: ignore
async def run_and_save_output(self, path: str, cmd: str, fmt: StorFormat = StorFormat.txt) -> CMDResult:
"""Run command on node and store result into storage"""
async with self.connection() as conn:
logger.debug(f"{self.hostname} - {cmd}")
res = await conn.run(cmd, merge_err=False)
if res.returncode != 0:
logger.warning(f"Cmd {cmd} failed on {self.hostname} with code {res.returncode}")
fmt = StorFormat.err
save = res.stdout + res.stderr_b.decode()
else:
save = res.stdout
self.save(path, fmt, res.returncode, save)
return res
# mypy does not support functions with default arguments
@dataclass
CollectFunc = Callable[[Collector], Coroutine[Any, Any, None]]
CephCollectFunc = Callable[[CephCollector], Coroutine[Any, Any, None]]
ALL_COLLECTORS: Dict[Role, List[Union[CephCollectFunc, CollectFunc]]] = collections.defaultdict(list)
@collector(Role.base)
@ceph_collector(Role.ceph_master)
@ceph_collector(Role.ceph_master)
@ceph_collector(Role.ceph_master)
@ceph_collector(Role.ceph_master)
@ceph_collector(Role.ceph_master)
@ceph_collector(Role.ceph_master)
@ceph_collector(Role.ceph_osd)
AVERAGE_BYTES_PER_CEPH_LOG_LINE = 143
@ceph_collector(Role.ceph_mon)
@collector(Role.node)
@collector(Role.node)
@collector(Role.node)
@collector(Role.node)
@collector(Role.node)
@collector(Role.node)
@collector(Role.node)
| 38.795417 | 120 | 0.610741 | from __future__ import annotations
import os
import re
import time
import json
import array
import random
import asyncio
import logging
import datetime
import contextlib
import collections
from enum import IntEnum
from dataclasses import dataclass, field
from typing import Any, List, Dict, Optional, Tuple, TypeVar, Callable, Coroutine, Union, cast
from aiorpc import IAIORPCNode, ConnectionPool
from cephlib import CephRelease, parse_ceph_volumes_js, parse_ceph_disk_js, CephReport, OSDDevInfo, OSDBSDevices, \
OSDFSDevices
from koder_utils import (IStorage, CMDResult, parse_devices_tree, collect_process_info, get_host_interfaces,
ignore_all, IAsyncNode)
logger = logging.getLogger('collect')
T = TypeVar('T')
VT = TypeVar('VT', str, bytes, array.array)
class StorFormat(IntEnum):
txt = 0
json = 1
err = 2
bin = 3
xml = 4
@dataclass
class Collector:
"""Base class for data collectors. Can collect data for only one node."""
storage: IStorage
hostname: Optional[str]
opts: Any
pool: ConnectionPool
pretty_json: bool = field(init=False, default=False)
cmds: Dict[str, Tuple[str, StorFormat]] = field(init=False, default_factory=dict)
def __post_init__(self) -> None:
self.pretty_json = not self.opts.no_pretty_json
self.cmds['bash'] = "", StorFormat.txt
def with_storage(self: T, storage: IStorage) -> T:
return self.__class__(storage, self.hostname, self.opts, self.pool) # type: ignore
@contextlib.asynccontextmanager
async def connection(self) -> IAIORPCNode:
assert self.hostname is not None
async with self.pool.connection(self.hostname) as conn:
yield conn
def chdir(self: T, path: str) -> T:
"""Chdir for point in storage tree, where current results are stored"""
return self.with_storage(self.storage.sub_storage(path)) # type: ignore
def save_raw(self, path: str, data: bytes):
self.storage.put_raw(data, path)
def save(self, path: str, fmt: StorFormat, code: int, data: VT, extra: List[str] = None) -> VT:
"""Save results into storage"""
rpath = f"{path}.{fmt.name if code == 0 else 'err'}"
if isinstance(data, array.array):
self.storage.put_array(rpath, data, extra if extra else [])
elif isinstance(data, (str, bytes)):
pretty_data: Union[str, bytes] = data
if code == 0 and fmt == StorFormat.json and self.pretty_json:
assert isinstance(data, (str, bytes))
try:
dt = json.loads(data)
except json.JSONDecodeError as exc:
logger.error(f"Failed to prettify json data for path {path}. Err {exc}. Saving as is")
else:
pretty_data = json.dumps(dt, indent=4, sort_keys=True)
data_b: bytes = pretty_data.encode() if isinstance(pretty_data, str) else cast(bytes, pretty_data)
assert extra is None
self.save_raw(rpath, data_b)
else:
raise TypeError(f"Can't save value of type {type(data)!r} (to {rpath!r})")
return data
async def run(self, *args, **kwargs) -> CMDResult:
async with self.connection() as conn:
return await conn.run(*args, **kwargs)
async def read_and_save(self, path: str, file_path: str, fmt: StorFormat = StorFormat.txt,
compress: bool = True) -> bytes:
"""Download file from node and save it into storage"""
async with self.connection() as conn:
try:
content = await conn.read(file_path, compress=compress)
code = 0
except (IOError, RuntimeError) as exc:
logger.warning(f"Can't get file {file_path!r} from node {self.hostname}. {exc}")
content = str(exc) # type: ignore
code = 1
self.save(path, fmt, code, content)
return content if code == 0 else None # type: ignore
async def run_and_save_output(self, path: str, cmd: str, fmt: StorFormat = StorFormat.txt) -> CMDResult:
"""Run command on node and store result into storage"""
async with self.connection() as conn:
logger.debug(f"{self.hostname} - {cmd}")
res = await conn.run(cmd, merge_err=False)
if res.returncode != 0:
logger.warning(f"Cmd {cmd} failed on {self.hostname} with code {res.returncode}")
fmt = StorFormat.err
save = res.stdout + res.stderr_b.decode()
else:
save = res.stdout
self.save(path, fmt, res.returncode, save)
return res
def __call__(self, path: str = None) -> CollectorProxy:
return CollectorProxy(self, path)
async def run_cmd_and_save_output(self, path: Optional[str], exe: str, args: str,
fmt: Optional[StorFormat] = None) -> CMDResult:
cmd, def_format = self.cmds[exe]
if fmt is None:
fmt = def_format
if path is None:
path = args.replace(" ", "_").replace("-", '_')
extra_space = "" if (cmd.endswith(" ") or cmd == "") else " "
return await self.run_and_save_output(path, cmd + extra_space + args, fmt)
class CollectorProxy:
def __init__(self, collector_obj: Collector, path: str = None) -> None:
self.collector_obj = collector_obj
self.path = path
# mypy does not support functions with default arguments
def __getattr__(self, name: str) -> Callable[[str], Coroutine[Any, Any, CMDResult]]:
assert name in self.collector_obj.cmds
async def closure(args: str, fmt: Optional[StorFormat] = None) -> CMDResult:
return await self.collector_obj.run_cmd_and_save_output(self.path, exe=name, args=args, fmt=fmt)
return closure
@dataclass
class CephCollector(Collector):
report: CephReport = None # type: ignore
def __post_init__(self) -> None:
assert self.report is not None, f"Report must be provided for {self.__class__.__name__}"
super().__post_init__()
opt = " ".join(f"'{arg}'" for arg in self.opts.ceph_extra_args)
self.cmds['radosgw'] = f"radosgw-admin {opt}", StorFormat.txt
self.cmds['ceph_js'] = f"ceph {opt} --format json", StorFormat.json
self.cmds['ceph'] = f"ceph {opt}", StorFormat.txt
self.cmds['rbd'] = f"rbd {opt}", StorFormat.txt
self.cmds['rados_js'] = f"rados {opt} --format json", StorFormat.json
self.cmds['rados'] = f"rados {opt}", StorFormat.txt
def with_storage(self, storage: IStorage) -> CephCollector:
return self.__class__(storage, self.hostname, self.opts, self.pool, self.report)
class Role(IntEnum):
base = 0
ceph_master = 1
ceph_osd = 2
ceph_mon = 3
node = 4
CollectFunc = Callable[[Collector], Coroutine[Any, Any, None]]
CephCollectFunc = Callable[[CephCollector], Coroutine[Any, Any, None]]
ALL_COLLECTORS: Dict[Role, List[Union[CephCollectFunc, CollectFunc]]] = collections.defaultdict(list)
def collector(role: Role) -> Callable[[CollectFunc], CollectFunc]:
def closure(func: CollectFunc) -> CollectFunc:
ALL_COLLECTORS[role].append(func)
return func
return closure
def ceph_collector(role: Role) -> Callable[[CephCollectFunc], CephCollectFunc]:
def closure(func: CephCollectFunc) -> CephCollectFunc:
ALL_COLLECTORS[role].append(func)
return func
return closure
@collector(Role.base)
async def collect_base(c: Collector) -> None:
time2 = time.strftime("%Y-%m-%d %H:%M:%S", time.gmtime())
curr_data = f"{datetime.datetime.now():%Y-%m-%d %H:%M:%S}\n{time2}\n{time.time()}"
c.save("collected_at", StorFormat.txt, 0, curr_data)
@ceph_collector(Role.ceph_master)
async def save_versions(c: CephCollector) -> None:
pre_luminous = c.report.version.release < CephRelease.luminous
coros = []
if pre_luminous:
coros.append(c("osd_versions_old").ceph("tell 'osd.*' version"))
coros.append(c("mon_versions_old").ceph("tell 'mon.*' version"))
else:
coros.append(c("versions").ceph_js("versions"))
coros.append(c("mon_metadata").ceph_js("mon metadata"))
coros.append(c("status").ceph_js("status"))
cmds = ["df", "auth list", "mon_status", "osd perf", "osd df", "node ls", "features", "time-sync-status", "pg dump"]
for cmd in cmds:
coros.append(c().ceph_js(cmd))
coros.append(c("rados_df").rados_js("df"))
coros.append(c("default_config").ceph("--show-config"))
coros.append(c("osd_blocked_by").ceph_js("osd blocked-by"))
await asyncio.wait(coros)
@ceph_collector(Role.ceph_master)
async def collect_rbd_volumes_info(c: CephCollector) -> None:
if c.opts.no_rbd_info:
logger.debug("Collecting RBD volumes stats")
rbd_cmd, _ = c.cmds['rbd']
sep = '-' * 60
for pool in c.report.osdmap.pools:
if 'application_metadata' in pool:
if 'rbd' not in pool.application_metadata:
continue
elif 'rgw' in pool.pool_name:
continue
name = pool.pool_name
await c(f'rbd_du_{name}').rbd(f"du -p {name}")
cmd = f'for image in $(rbd list -p rbd) ; do echo "{sep}" ; " + \
f"echo "$image" ; {rbd_cmd} info "$image" ; done'
await c(f"rbd_images_{name}").bash(cmd)
@ceph_collector(Role.ceph_master)
async def collect_crushmap(c: CephCollector) -> None:
if c.opts.collect_maps:
cr_fname = f"/tmp/ceph_collect.{'%08X' % random.randint(0, 2 << 64)}.cr"
ceph_cmd, _ = c.cmds['ceph']
crushmap_res = await c.run(f"{ceph_cmd} osd getcrushmap -o {cr_fname}")
if crushmap_res.returncode != 0:
logger.error("Fail to get crushmap")
else:
await c.read_and_save('crushmap', cr_fname, StorFormat.bin)
crushtool_res = await c.run(f"crushtool -d {cr_fname} -o {cr_fname}.txt")
if crushtool_res.returncode != 0:
logger.error("Fail to decompile crushmap")
else:
await c.read_and_save('crushmap', cr_fname + ".txt", StorFormat.txt)
@ceph_collector(Role.ceph_master)
async def collect_osd_map(c: CephCollector) -> None:
if c.opts.collect_maps:
osd_fname = f"/tmp/ceph_collect.{'%08X' % random.randint(0, 2 << 64)}.osd"
ceph_cmd, _ = c.cmds['ceph']
osdmap_res = await c.run(f"{ceph_cmd} osd getmap -o {osd_fname}")
if osdmap_res.returncode != 0:
logger.error("Fail to get osdmap")
else:
await c.read_and_save('osdmap', osd_fname, StorFormat.bin)
await c.run_and_save_output('osdmap', f"osdmaptool --print {osd_fname}", StorFormat.txt)
@ceph_collector(Role.ceph_master)
async def collect_rgw_info(c: CephCollector) -> None:
if not c.opts.collect_rgw:
await c().radosgw("realm list")
await c().radosgw("zonegroup list")
await c().radosgw("zone list")
@ceph_collector(Role.ceph_master)
async def collect_ceph_txt(c: CephCollector) -> None:
if c.opts.collect_txt:
for cmd in ("status", "osd tree", "df", "osd df", "rados df", "osd dump", "osd blocked-by"):
await c().ceph(cmd)
await c("ceph_s").ceph("-s")
@ceph_collector(Role.ceph_osd)
async def collect_osd(c: CephCollector) -> None:
# check OSD process status
async with c.connection() as conn:
psaux = await conn.run_str("ps aux | grep ceph-osd")
osd_re = re.compile(r".*?\s+(?P<pid>\d+)\s.*?\bceph-osd[\t ]+.*(-i|--id)[\t ]+(?P<osd_id>\d+)")
running_osds: Dict[int, int] = {}
for rr in osd_re.finditer(psaux):
osd_id = int(rr.group('osd_id'))
running_osds[osd_id] = int(rr.group('pid'))
ids_from_ceph = [meta.id for meta in c.report.osd_metadata if meta.hostname == c.hostname]
unexpected_osds = set(running_osds).difference(ids_from_ceph)
logger.info(f"Found next running osd's on node {c.hostname}: {list(running_osds.keys())}")
logger.info(f"Expecting next osd's for {c.hostname}: {ids_from_ceph}")
for osd_id in unexpected_osds:
logger.warning(f"Unexpected osd-{osd_id} in node {c.hostname}")
not_runnig = set(ids_from_ceph).difference(running_osds)
if not_runnig:
logger.warning(f"Next osd's not running on node {c.hostname}: {list(not_runnig)}")
c_host = c.chdir(f"hosts/{c.hostname}")
cephdisklist_js, cephvollist_js, lsblk_js = await asyncio.gather(
c_host("cephdisk").bash("ceph-disk list --format=json", fmt=StorFormat.json), # type: ignore
c_host("cephvolume").bash("ceph-volume lvm list --format=json", fmt=StorFormat.json), # type: ignore
c_host("lsblk").bash("lsblk -a --json", fmt=StorFormat.json)) # type: ignore
dev_tree = parse_devices_tree(json.loads(lsblk_js.stdout))
if cephvollist_js.returncode == 0:
devs_for_osd: Dict[int, OSDDevInfo] = parse_ceph_volumes_js(cephvollist_js.stdout)
else:
devs_for_osd = {}
if not devs_for_osd and cephdisklist_js.returncode == 0:
devs_for_osd = parse_ceph_disk_js(cephdisklist_js.stdout)
logger.debug(f"Found next pids for OSD's on node {c.hostname}: {sorted(running_osds.values())}")
coros: List[Coroutine[Any, Any, Any]] = [c_host("lsblk").bash("lsblk -a")]
if c_host.opts.collect_txt:
coros.append(c_host("cephdisk").bash("ceph-disk list"))
coros.append(c_host("cephvolume").bash("ceph-volume lvm list"))
for osd_id in ids_from_ceph:
coros.append(collect_single_osd(c.chdir(f'osd/{osd_id}'),
osd_id,
pid=running_osds.get(osd_id),
dev_tree=dev_tree,
devs=devs_for_osd.get(osd_id)))
await asyncio.wait(coros)
async def collect_single_osd(c: CephCollector,
osd_id: int,
pid: Optional[int],
dev_tree: Dict[str, str],
devs: Optional[OSDDevInfo]) -> None:
await c("log").bash(f"tail -n {c.opts.ceph_log_max_lines} /var/log/ceph/ceph-osd.{osd_id}.log")
await c("perf_dump").ceph(f"daemon osd.{osd_id} perf dump")
await c("perf_hist_dump").ceph(f"daemon osd.{osd_id} perf histogram dump")
if pid is not None:
await c("config").ceph(f"daemon osd.{osd_id} config show", fmt=StorFormat.json) # type: ignore
else:
logger.warning(f"osd-{osd_id} in node {c.hostname} is down. No config available")
if devs:
if isinstance(devs, OSDFSDevices):
data_dev = str(devs.data)
j_dev = str(devs.journal)
osd_dev_conf = {'data': data_dev,
'journal': str(j_dev),
'r_data': dev_tree[data_dev],
'r_journal': dev_tree[j_dev],
'type': 'filestore'}
else:
assert isinstance(devs, OSDBSDevices)
data_dev = str(devs.block)
db_dev = str(devs.db)
wal_dev = str(devs.wal)
osd_dev_conf = {'data': data_dev,
'wal': wal_dev,
'db': db_dev,
'r_data': dev_tree[data_dev],
'r_wal': dev_tree[wal_dev],
'r_db': dev_tree[db_dev],
'type': 'bluestore'}
else:
osd_dev_conf = {}
c.save('devs_cfg', StorFormat.json, 0, json.dumps(osd_dev_conf))
logger.debug(f"Collecting info for osd.{osd_id} with pid {pid}")
async with c.connection() as conn:
info = await collect_process_info(conn, pid)
c.save("proc_info", StorFormat.json, 0, json.dumps(info.__dict__))
AVERAGE_BYTES_PER_CEPH_LOG_LINE = 143
@ceph_collector(Role.ceph_mon)
async def collect_mon_info(c: CephCollector) -> None:
await c("mon_daemons").bash("ps aux | grep ceph-mon")
# tail = f"tail -n {c.opts.ceph_log_max_lines}"
# await c("mon_log").bash(f"{tail} /var/log/ceph/ceph-mon.{c.hostname}.log")
# await c("ceph_log").bash(f"{tail} /var/log/ceph/ceph.log")
# await c("ceph_audit").bash(f"{tail} /var/log/ceph/ceph.audit.log")
async with c.connection() as conn:
read_size = AVERAGE_BYTES_PER_CEPH_LOG_LINE * c.opts.ceph_log_max_lines
dt = [chunk async for chunk in conn.tail_file(f"/var/log/ceph/ceph-mon.{c.hostname}.log", read_size)]
c.save("mon_log", StorFormat.txt, 0, b"".join(dt).decode())
dt = [chunk async for chunk in conn.tail_file("/var/log/ceph/ceph.log", read_size)]
c.save("ceph_log", StorFormat.txt, 0, b"".join(dt).decode())
dt = [chunk async for chunk in conn.tail_file(f"/var/log/ceph/ceph.audit.log", read_size)]
c.save("ceph_audit", StorFormat.txt, 0, b"".join(dt).decode())
log_issues = await conn.proxy.ceph.find_issues_in_ceph_log(c.opts.ceph_log_max_lines)
c.save("ceph_log_wrn_err", StorFormat.txt, 0, log_issues)
with ignore_all:
issues_count, regions = await conn.proxy.ceph.analyze_ceph_logs_for_issues()
c.save("log_issues_count", StorFormat.json, 0, json.dumps(issues_count))
c.save("status_regions", StorFormat.json, 0, json.dumps(regions))
await c("config").ceph(f"daemon mon.{c.hostname} config show", fmt=StorFormat.json) # type: ignore
await c("ceph_var_dirs_size").bash("du -s /var/lib/ceph/m*")
@collector(Role.node)
async def collect_kernel_modules_info(c: Collector) -> None:
try:
await c().bash("lsmod")
except Exception as exc:
logger.warning(f"Failed to list kernel modules on node {c.hostname}: {exc}")
return
try:
await c("modinfo_all").bash("for name in $(lsmod | awk '{print $1}') ; do modinfo $name ; echo '-----' ; done")
except Exception as exc:
logger.warning(f"Failed to list kernel modules info on {c.hostname}: {exc}")
@collector(Role.node)
async def collect_common_features(c: Collector) -> None:
node_commands = [
(None, "df"),
(None, "dmidecode"),
(None, "dmesg"),
("ipa4", "ip -o -4 a"),
("ipa", "ip a"),
(None, "ifconfig"),
("ifconfig_short", "ifconfig -s"),
("lsblk", "lsblk -O"),
("lsblk_short", "lsblk"),
(None, "mount"),
("netstat", "netstat -nap"),
("netstat_stat", "netstat -s"),
("sysctl", "sysctl -a"),
("uname", "uname -a"),
]
await asyncio.wait(
[c(path_offset).bash(cmd) for path_offset, cmd in node_commands] +
[c("lshw").bash("lshw -xml", fmt=StorFormat.xml)]) # type: ignore
@collector(Role.node)
async def collect_files(c: Collector) -> None:
node_files = ["/proc/diskstats", "/proc/meminfo", "/proc/loadavg", "/proc/cpuinfo", "/proc/uptime", "/proc/vmstat"]
for fpath in node_files:
await c.read_and_save(os.path.basename(fpath), fpath)
node_renamed_files = [("netdev", "/proc/net/dev"),
("dev_netstat", "/proc/net/netstat"),
("softnet_stat", "/proc/net/softnet_stat"),
("ceph_conf", "/etc/ceph/ceph.conf")]
await asyncio.wait([c.read_and_save(name, fpath) for name, fpath in node_renamed_files])
@collector(Role.node)
async def collect_bonds_info(c: Collector) -> None:
# collect_bonds_info
bondmap = {}
async with c.connection() as conn:
if await conn.exists("/proc/net/bonding"):
for fname in (await conn.iterdir("/proc/net/bonding")):
await c.read_and_save(f"bond_{fname}", fname)
bondmap[str(fname)] = f"bond_{fname}"
c.save("bonds", StorFormat.json, 0, json.dumps(bondmap))
@collector(Role.node)
async def collect_packages(c: Collector) -> None:
async with c.connection() as conn:
try:
if await conn.exists("/etc/debian_version"):
await c("packages_deb").bash("dpkg -l")
else:
await c("packages_rpm").bash("yum list installed")
except Exception as exc:
logger.warning(f"Failed to download packages information from node {c.hostname}: {exc}")
@collector(Role.node)
async def collect_block_devs(c: Collector) -> None:
async with c.connection() as conn:
bdevs_info_rpc = await conn.proxy.fs.get_block_devs_info()
bdevs_info = {name: data for name, data in bdevs_info_rpc.items()}
for name_prefix in ['loop']:
for name in bdevs_info:
if name.startswith(name_prefix):
del bdevs_info[name]
tools = ['hdparm', 'smartctl', 'nvme']
missing = [name for exists, name in zip((await conn.proxy.fs.binarys_exists(tools)), tools) if not exists]
if missing:
logger.warning(f"{','.join(missing)} is not installed on {c.hostname}")
if 'nvme' not in missing:
nvme_res = await c.run('nvme version')
if nvme_res.returncode != 0:
ver: float = 0
else:
try:
*_, version = nvme_res.stdout.split()
ver = float(version)
except:
ver = 0
if ver < 1.0:
logger.warning(f"Nvme tool too old {ver}, at least 1.0 version is required")
else:
nvme_list_js = await c('nvme_list').bash('nvme list -o json', fmt=StorFormat.json) # type: ignore
if nvme_list_js.returncode == 0:
try:
for dev in json.loads(nvme_list_js.stdout)['Devices']:
name = os.path.basename(dev['DevicePath'])
cmd = f"nvme smart-log {dev['DevicePath']} -o json"
c(f'block_devs/{name}/nvme_smart_log').bash(cmd, fmt=StorFormat.json) # type: ignore
except:
logging.warning("Failed to process nvme list output")
lsblk_res = await c("lsblkjs").bash("lsblk -O -b -J", fmt=StorFormat.json) # type: ignore
if lsblk_res.returncode == 0:
coros = []
for dev_node in json.loads(lsblk_res.stdout)['blockdevices']:
name = dev_node['name']
if 'hdparm' not in missing:
coros.append(c(f'block_devs/{name}/hdparm').bash(f"sudo hdparm -I /dev/{name}"))
if 'smartctl' not in missing:
coros.append(c(f'block_devs/{name}/smartctl').bash(f"sudo smartctl -a /dev/{name}"))
await asyncio.wait(coros)
async def collect_dev(conn: IAsyncNode, is_phy: bool, dev: str) -> Dict[str, Dict[str, Any]]:
interface = {'dev': dev, 'is_phy': is_phy}
interfaces: Dict[str, Dict[str, Any]] = {dev: interface}
if is_phy:
ethtool_res = await conn.run("ethtool " + dev)
if ethtool_res.returncode == 0:
interface['ethtool'] = ethtool_res.stdout
iwconfig_res = await conn.run("iwconfig " + dev)
if iwconfig_res.returncode == 0:
interface['iwconfig'] = iwconfig_res.stdout
return interfaces
@collector(Role.node)
async def collect_interfaces_info(c: Collector) -> None:
async with c.connection() as conn:
info = await get_host_interfaces(conn)
interfaces: Any = {}
for res in await asyncio.gather(*[collect_dev(conn, is_phy, dev) for is_phy, dev in info]):
interfaces.update(res)
c.save('interfaces', StorFormat.json, 0, json.dumps(interfaces))
| 17,932 | 255 | 775 |
42677e1e3fc74b4db9834cf9ae727d99e53425d8 | 1,999 | py | Python | plantcv/apply_mask.py | Atitsuda/plantcv | 29932a0f0aa3573a74335425bccc394b2fb7b885 | [
"MIT"
] | 2 | 2017-12-05T12:07:47.000Z | 2021-01-25T08:58:52.000Z | plantcv/apply_mask.py | Atitsuda/plantcv | 29932a0f0aa3573a74335425bccc394b2fb7b885 | [
"MIT"
] | null | null | null | plantcv/apply_mask.py | Atitsuda/plantcv | 29932a0f0aa3573a74335425bccc394b2fb7b885 | [
"MIT"
] | 1 | 2020-08-13T17:44:53.000Z | 2020-08-13T17:44:53.000Z | # Apply White or Black Background Mask
import cv2
from . import print_image
from . import plot_image
from . import fatal_error
def apply_mask(img, mask, mask_color, device, debug=None):
"""Apply white image mask to image, with bitwise AND operator bitwise NOT operator and ADD operator.
Inputs:
img = image object, color(RGB)
mask = image object, binary (black background with white object)
mask_color = white or black
device = device number. Used to count steps in the pipeline
debug = None, print, or plot. Print = save to file, Plot = print to screen.
Returns:
device = device number
masked_img = masked image
:param img: numpy array
:param mask: numpy array
:param mask_color: str
:param device: int
:param debug: str
:return device: int
:return masked_img: numpy array
"""
device += 1
if mask_color == 'white':
# Mask image
masked_img = cv2.bitwise_and(img, img, mask=mask)
# Create inverted mask for background
mask_inv = cv2.bitwise_not(mask)
# Invert the background so that it is white, but apply mask_inv so you don't white out the plant
white_mask = cv2.bitwise_not(masked_img, mask=mask_inv)
# Add masked image to white background (can't just use mask_inv because that is a binary)
white_masked = cv2.add(masked_img, white_mask)
if debug == 'print':
print_image(white_masked, (str(device) + '_wmasked.png'))
elif debug == 'plot':
plot_image(white_masked)
return device, white_masked
elif mask_color == 'black':
masked_img = cv2.bitwise_and(img, img, mask=mask)
if debug == 'print':
print_image(masked_img, (str(device) + '_bmasked.png'))
elif debug == 'plot':
plot_image(masked_img)
return device, masked_img
else:
fatal_error('Mask Color' + str(mask_color) + ' is not "white" or "black"!')
| 35.696429 | 104 | 0.643322 | # Apply White or Black Background Mask
import cv2
from . import print_image
from . import plot_image
from . import fatal_error
def apply_mask(img, mask, mask_color, device, debug=None):
"""Apply white image mask to image, with bitwise AND operator bitwise NOT operator and ADD operator.
Inputs:
img = image object, color(RGB)
mask = image object, binary (black background with white object)
mask_color = white or black
device = device number. Used to count steps in the pipeline
debug = None, print, or plot. Print = save to file, Plot = print to screen.
Returns:
device = device number
masked_img = masked image
:param img: numpy array
:param mask: numpy array
:param mask_color: str
:param device: int
:param debug: str
:return device: int
:return masked_img: numpy array
"""
device += 1
if mask_color == 'white':
# Mask image
masked_img = cv2.bitwise_and(img, img, mask=mask)
# Create inverted mask for background
mask_inv = cv2.bitwise_not(mask)
# Invert the background so that it is white, but apply mask_inv so you don't white out the plant
white_mask = cv2.bitwise_not(masked_img, mask=mask_inv)
# Add masked image to white background (can't just use mask_inv because that is a binary)
white_masked = cv2.add(masked_img, white_mask)
if debug == 'print':
print_image(white_masked, (str(device) + '_wmasked.png'))
elif debug == 'plot':
plot_image(white_masked)
return device, white_masked
elif mask_color == 'black':
masked_img = cv2.bitwise_and(img, img, mask=mask)
if debug == 'print':
print_image(masked_img, (str(device) + '_bmasked.png'))
elif debug == 'plot':
plot_image(masked_img)
return device, masked_img
else:
fatal_error('Mask Color' + str(mask_color) + ' is not "white" or "black"!')
| 0 | 0 | 0 |
5bbe1c74e3eae40f3472dac89bbe701e7c69950a | 71 | py | Python | libs/gym-cartpole-world-master/gym_cartpole_world/envs/__init__.py | Adaptive-RL/AdaRL-code | 493b1ee5a0f98a220c5a1e5ce2e2ce6572d02e9f | [
"MIT"
] | 8 | 2022-03-15T04:24:03.000Z | 2022-03-18T08:33:22.000Z | libs/gym-cartpole-world-master/gym_cartpole_world/envs/__init__.py | Adaptive-RL/AdaRL-code | 493b1ee5a0f98a220c5a1e5ce2e2ce6572d02e9f | [
"MIT"
] | 1 | 2022-03-17T07:17:03.000Z | 2022-03-20T06:23:12.000Z | libs/gym-cartpole-world-master/gym_cartpole_world/envs/__init__.py | Adaptive-RL/AdaRL-code | 493b1ee5a0f98a220c5a1e5ce2e2ce6572d02e9f | [
"MIT"
] | 1 | 2022-03-18T06:08:37.000Z | 2022-03-18T06:08:37.000Z | from gym_cartpole_world.envs.cartpoleworld_env import CartPoleWorldEnv
| 35.5 | 70 | 0.915493 | from gym_cartpole_world.envs.cartpoleworld_env import CartPoleWorldEnv
| 0 | 0 | 0 |
1b7ccd8aaf4d068975398a6662e5f16f5a3693e4 | 147,800 | py | Python | all/emojitations/data/nb.py | idleberg/sublime-emojitations | b2b4e8ce2c33ed0f6b8d6db6085e21da4e8d895b | [
"MIT"
] | 6 | 2016-08-31T14:42:36.000Z | 2021-09-05T23:55:47.000Z | all/emojitations/data/nb.py | idleberg/sublime-emojitations | b2b4e8ce2c33ed0f6b8d6db6085e21da4e8d895b | [
"MIT"
] | 1 | 2016-10-20T10:52:06.000Z | 2016-10-20T18:47:19.000Z | all/emojitations/data/nb.py | idleberg/sublime-emojitations | b2b4e8ce2c33ed0f6b8d6db6085e21da4e8d895b | [
"MIT"
] | 5 | 2016-08-31T14:48:11.000Z | 2021-09-05T23:55:33.000Z | from emojitations.emojitypes import EmojiAnnotations
emoji = [
EmojiAnnotations(emoji='😀', codepoints=(128512,), name='glisefjes', slug='glisefjes', annotations=frozenset({'fjes', 'lol', 'latter', 'morsomt', 'smilende fjes'})),
EmojiAnnotations(emoji='😁', codepoints=(128513,), name='glisefjes med smilende øyne', slug='glisefjes_med_smilende_øyne', annotations=frozenset({'lol', 'morsomt', 'latter', 'smil', 'smilende øyne', 'smilende fjes', 'fjes', 'smilefjes'})),
EmojiAnnotations(emoji='😂', codepoints=(128514,), name='gledestårer', slug='gledestårer', annotations=frozenset({'fjes med gledestårer', 'tårer', 'fjes', 'glede'})),
EmojiAnnotations(emoji='😃', codepoints=(128515,), name='smilefjes med åpen munn', slug='smilefjes_med_åpen_munn', annotations=frozenset({'lol', 'morsomt', 'smil', 'latter', 'smilende fjes med tenner', 'fjes', 'smilefjes'})),
EmojiAnnotations(emoji='😄', codepoints=(128516,), name='smilefjes med åpen munn og smilende øyne', slug='smilefjes_med_åpen_munn_og_smilende_øyne', annotations=frozenset({'lol', 'morsomt', 'latter', 'smil', 'smilende øyne', 'med tenner', 'smilende fjes', 'smilefjes'})),
EmojiAnnotations(emoji='😅', codepoints=(128517,), name='svett smil', slug='svett_smil', annotations=frozenset({'kaldsvette', 'smil med svette'})),
EmojiAnnotations(emoji='😆', codepoints=(128518,), name='smil med åpen munn', slug='smil_med_åpen_munn', annotations=frozenset({'smilefjes med åpen munn', 'smil med tennene', 'åpen munn', 'fjes'})),
EmojiAnnotations(emoji='😉', codepoints=(128521,), name='blunkende ansikt', slug='blunkende_ansikt', annotations=frozenset({'blunk', 'fjes'})),
EmojiAnnotations(emoji='😊', codepoints=(128522,), name='smil', slug='smil', annotations=frozenset({'smilefjes', 'fjes', 'smilende øyne', 'smilende fjes og øyne'})),
EmojiAnnotations(emoji='😋', codepoints=(128523,), name='nam-nam-fjes', slug='nam_nam_fjes', annotations=frozenset({'deilig mat', 'fjes', 'nam', 'mmm', 'smilefjes'})),
EmojiAnnotations(emoji='😎', codepoints=(128526,), name='smilefjes med solbriller', slug='smilefjes_med_solbriller', annotations=frozenset({'fjes', 'smil', 'solbriller'})),
EmojiAnnotations(emoji='😍', codepoints=(128525,), name='smilefjes med hjerteøyne', slug='smilefjes_med_hjerteøyne', annotations=frozenset({'hjerteøyne', 'øyne', 'fjes', 'kjærlighet', 'hjerte'})),
EmojiAnnotations(emoji='😘', codepoints=(128536,), name='slengkyss', slug='slengkyss', annotations=frozenset({'ansikt slenger kyss', 'ansikt'})),
EmojiAnnotations(emoji='😗', codepoints=(128535,), name='kyssefjes', slug='kyssefjes', annotations=frozenset({'fjes', 'kyssende fjes', 'kyss'})),
EmojiAnnotations(emoji='😙', codepoints=(128537,), name='kyss og smil', slug='kyss_og_smil', annotations=frozenset({'fjes', 'kyssende fjes med smil', 'smil', 'smilende øyne'})),
EmojiAnnotations(emoji='😚', codepoints=(128538,), name='kyssende', slug='kyssende', annotations=frozenset({'kyss med lukkede øyne', 'fjes', 'øynene lukket'})),
EmojiAnnotations(emoji='☺', codepoints=(9786,), name='smilefjes', slug='smilefjes', annotations=frozenset({'smil', 'ansikt'})),
EmojiAnnotations(emoji='\U0001f642', codepoints=(128578,), name='fjes som smiler litt', slug='fjes_som_smiler_litt', annotations=frozenset({'fjes', 'smil'})),
EmojiAnnotations(emoji='\U0001f917', codepoints=(129303,), name='fjes som gir klem', slug='fjes_som_gir_klem', annotations=frozenset({'fjes', 'klem'})),
EmojiAnnotations(emoji='😇', codepoints=(128519,), name='smil med glorie', slug='smil_med_glorie', annotations=frozenset({'fjes', 'glorie', 'smil'})),
EmojiAnnotations(emoji='\U0001f914', codepoints=(129300,), name='tenkefjes', slug='tenkefjes', annotations=frozenset({'fjes', 'tenker'})),
EmojiAnnotations(emoji='😐', codepoints=(128528,), name='uttrykksløst fjes', slug='uttrykksløst_fjes', annotations=frozenset({'nøytral', 'stum', 'fjes', 'ingen kommentar'})),
EmojiAnnotations(emoji='😑', codepoints=(128529,), name='uttrykksløs', slug='uttrykksløs', annotations=frozenset({'nøytral', 'stum', 'fjes', 'ingen kommentar'})),
EmojiAnnotations(emoji='😶', codepoints=(128566,), name='fjes uten munn', slug='fjes_uten_munn', annotations=frozenset({'ingen munn', 'fjes'})),
EmojiAnnotations(emoji='\U0001f644', codepoints=(128580,), name='fjes med rullende øyne', slug='fjes_med_rullende_øyne', annotations=frozenset({'rullende', 'øyne', 'fjes'})),
EmojiAnnotations(emoji='😏', codepoints=(128527,), name='skuffet smil', slug='skuffet_smil', annotations=frozenset({'halvsmil', 'lurt smil', 'fjes'})),
EmojiAnnotations(emoji='😣', codepoints=(128547,), name='bommert', slug='bommert', annotations=frozenset({'sta', 'utholdende', 'blemme'})),
EmojiAnnotations(emoji='😥', codepoints=(128549,), name='puh!', slug='puh', annotations=frozenset({'lettet', 'skuffet men lettet', 'fjes', 'skuffet'})),
EmojiAnnotations(emoji='😮', codepoints=(128558,), name='forbauset', slug='forbauset', annotations=frozenset({'åpen munn', 'fjes'})),
EmojiAnnotations(emoji='\U0001f910', codepoints=(129296,), name='fjes med glidelåsmunn', slug='fjes_med_glidelåsmunn', annotations=frozenset({'fjes', 'glidelås', 'munn'})),
EmojiAnnotations(emoji='😯', codepoints=(128559,), name='forbauset med øyebryn', slug='forbauset_med_øyebryn', annotations=frozenset({'stum', 'fjes', 'overrasket'})),
EmojiAnnotations(emoji='😪', codepoints=(128554,), name='søvnig fjes', slug='søvnig_fjes', annotations=frozenset({'søvnig', 'fjes'})),
EmojiAnnotations(emoji='😫', codepoints=(128555,), name='trøtt fjes', slug='trøtt_fjes', annotations=frozenset({'søvnig', 'fjes'})),
EmojiAnnotations(emoji='😴', codepoints=(128564,), name='sovende', slug='sovende', annotations=frozenset({'sovende fjes', 'fjes'})),
EmojiAnnotations(emoji='😌', codepoints=(128524,), name='lettet-fjes', slug='lettet_fjes', annotations=frozenset({'lettet', 'fjes'})),
EmojiAnnotations(emoji='\U0001f913', codepoints=(129299,), name='nerdefjes', slug='nerdefjes', annotations=frozenset({'nerd', 'fjes'})),
EmojiAnnotations(emoji='😛', codepoints=(128539,), name='rekke tunge', slug='rekke_tunge', annotations=frozenset({'tungen ut', 'fjes', 'fjes som rekker tunge'})),
EmojiAnnotations(emoji='😜', codepoints=(128540,), name='spøkefullt blunkefjes som rekker tunge', slug='spøkefullt_blunkefjes_som_rekker_tunge', annotations=frozenset({'blunk', 'spøk', 'fjes', 'erte', 'rekke tunge med blunk'})),
EmojiAnnotations(emoji='😝', codepoints=(128541,), name='æsj', slug='æsj', annotations=frozenset({'følt', 'rekke tunge', 'lukkende øyne', 'tungen ut med øynene lukket', 'fjes'})),
EmojiAnnotations(emoji='☹', codepoints=(9785,), name='trist fjes', slug='trist_fjes', annotations=frozenset({'fjes', 'trist'})),
EmojiAnnotations(emoji='\U0001f641', codepoints=(128577,), name='litt trist fjes', slug='litt_trist_fjes', annotations=frozenset({'fjes', 'trist'})),
EmojiAnnotations(emoji='😒', codepoints=(128530,), name='misfornøyd', slug='misfornøyd', annotations=frozenset({'sur', 'ansikt', 'misfornøyd ansikt'})),
EmojiAnnotations(emoji='😓', codepoints=(128531,), name='svette', slug='svette', annotations=frozenset({'kaldsvette', 'ansikt'})),
EmojiAnnotations(emoji='😔', codepoints=(128532,), name='tankefull', slug='tankefull', annotations=frozenset({'alvorlig', 'ansikt'})),
EmojiAnnotations(emoji='😕', codepoints=(128533,), name='forvirret fjes', slug='forvirret_fjes', annotations=frozenset({'forvirret', 'fjes'})),
EmojiAnnotations(emoji='😖', codepoints=(128534,), name='sint', slug='sint', annotations=frozenset({'ansikt', 'forbannet'})),
EmojiAnnotations(emoji='\U0001f643', codepoints=(128579,), name='fjes som er opp-ned', slug='fjes_som_er_opp_ned', annotations=frozenset({'opp-ned', 'fjes'})),
EmojiAnnotations(emoji='😷', codepoints=(128567,), name='fjes med maske', slug='fjes_med_maske', annotations=frozenset({'bakteriemaske', 'sykdom', 'fjes'})),
EmojiAnnotations(emoji='\U0001f912', codepoints=(129298,), name='fjes med termometer', slug='fjes_med_termometer', annotations=frozenset({'syk', 'fjes', 'termometer'})),
EmojiAnnotations(emoji='\U0001f915', codepoints=(129301,), name='bandasjert hode', slug='bandasjert_hode', annotations=frozenset({'bandasje', 'fjes', 'skadet', 'skade'})),
EmojiAnnotations(emoji='\U0001f911', codepoints=(129297,), name='pengegal', slug='pengegal', annotations=frozenset({'munn', 'fjes', 'penger'})),
EmojiAnnotations(emoji='😲', codepoints=(128562,), name='sjokkert', slug='sjokkert', annotations=frozenset({'stum', 'fjes', 'overrasket'})),
EmojiAnnotations(emoji='😞', codepoints=(128542,), name='skuffet', slug='skuffet', annotations=frozenset({'skuffelse', 'skuffet fjes', 'fjes'})),
EmojiAnnotations(emoji='😟', codepoints=(128543,), name='bekymret', slug='bekymret', annotations=frozenset({'bekymret fjes', 'ansikt'})),
EmojiAnnotations(emoji='😤', codepoints=(128548,), name='seirende', slug='seirende', annotations=frozenset({'triumf', 'triumferende', 'jeg vant', 'fjes'})),
EmojiAnnotations(emoji='😢', codepoints=(128546,), name='gråtende', slug='gråtende', annotations=frozenset({'tårer', 'ansikt', 'gråtende ansikt', 'trist'})),
EmojiAnnotations(emoji='😭', codepoints=(128557,), name='hylgråting', slug='hylgråting', annotations=frozenset({'tårer', 'fjes', 'gråt'})),
EmojiAnnotations(emoji='😦', codepoints=(128550,), name='oppgitt', slug='oppgitt', annotations=frozenset({'sjokkert med åpen munn', 'overrasket', 'ansikt'})),
EmojiAnnotations(emoji='😧', codepoints=(128551,), name='pine', slug='pine', annotations=frozenset({'smerte', 'lider', 'ansikt'})),
EmojiAnnotations(emoji='😨', codepoints=(128552,), name='redd', slug='redd', annotations=frozenset({'skremt', 'fjes', 'redd ansikt'})),
EmojiAnnotations(emoji='😩', codepoints=(128553,), name='sliten', slug='sliten', annotations=frozenset({'fjes', 'trøtt ansikt', 'trøtt'})),
EmojiAnnotations(emoji='😬', codepoints=(128556,), name='grimase', slug='grimase', annotations=frozenset({'fjes'})),
EmojiAnnotations(emoji='😰', codepoints=(128560,), name='svettende', slug='svettende', annotations=frozenset({'kaldsvette', 'syk', 'åpen munn', 'fjes'})),
EmojiAnnotations(emoji='😱', codepoints=(128561,), name='livredd', slug='livredd', annotations=frozenset({'skremt', 'skrikende', 'fjes', 'redd ansikt'})),
EmojiAnnotations(emoji='😳', codepoints=(128563,), name='rødmer', slug='rødmer', annotations=frozenset({'rødt fjes'})),
EmojiAnnotations(emoji='😵', codepoints=(128565,), name='ør', slug='ør', annotations=frozenset({'fjes'})),
EmojiAnnotations(emoji='😡', codepoints=(128545,), name='rasende', slug='rasende', annotations=frozenset({'sint', 'sur', 'rødt ansikt', 'ansikt', 'rødt'})),
EmojiAnnotations(emoji='😠', codepoints=(128544,), name='sint fjes', slug='sint_fjes', annotations=frozenset({'sint ansikt', 'ansikt', 'forbannet'})),
EmojiAnnotations(emoji='😈', codepoints=(128520,), name='smilende djevel', slug='smilende_djevel', annotations=frozenset({'smil med horn', 'horn', 'fjes', 'smil'})),
EmojiAnnotations(emoji='👿', codepoints=(128127,), name='sint djevel', slug='sint_djevel', annotations=frozenset({'liten djevel', 'ansikt'})),
EmojiAnnotations(emoji='👹', codepoints=(128121,), name='troll', slug='troll', annotations=frozenset({'japansk ogre', 'eventyr', 'ogre'})),
EmojiAnnotations(emoji='👺', codepoints=(128122,), name='japansk skapning (tengu)', slug='japansk_skapning_(tengu)', annotations=frozenset({'ansikt', 'japansk troll', 'eventyr'})),
EmojiAnnotations(emoji='💀', codepoints=(128128,), name='hodeskalle', slug='hodeskalle', annotations=frozenset({'kropp', 'eventyr', 'ansikt'})),
EmojiAnnotations(emoji='☠', codepoints=(9760,), name='hodeskalle med korslagte dødningebein', slug='hodeskalle_med_korslagte_dødningebein', annotations=frozenset({'hodeskalle', 'korslagte dødningebein', 'monster', 'kropp', 'død', 'fjes'})),
EmojiAnnotations(emoji='👻', codepoints=(128123,), name='spøkelse', slug='spøkelse', annotations=frozenset({'eventyr'})),
EmojiAnnotations(emoji='👽', codepoints=(128125,), name='romvesen', slug='romvesen', annotations=frozenset({'utenomjordisk', 'sci-fi', 'UFO', 'ansikt'})),
EmojiAnnotations(emoji='👾', codepoints=(128126,), name='rommonster', slug='rommonster', annotations=frozenset({'utenomjordisk', 'sci-fi', 'UFO', 'ansikt'})),
EmojiAnnotations(emoji='\U0001f916', codepoints=(129302,), name='robotfjes', slug='robotfjes', annotations=frozenset({'robot', 'monster', 'fjes'})),
EmojiAnnotations(emoji='💩', codepoints=(128169,), name='bæsj', slug='bæsj', annotations=frozenset({'hundebæsj', 'bæsjeklatt', 'kuruke'})),
EmojiAnnotations(emoji='😺', codepoints=(128570,), name='smilende katt med åpen munn', slug='smilende_katt_med_åpen_munn', annotations=frozenset({'katt', 'kattefjes', 'åpen munn', 'smil'})),
EmojiAnnotations(emoji='😸', codepoints=(128568,), name='flirende katt med smilende øyne', slug='flirende_katt_med_smilende_øyne', annotations=frozenset({'katt', 'kattefjes', 'smil'})),
EmojiAnnotations(emoji='😹', codepoints=(128569,), name='kattefjes med gledestårer', slug='kattefjes_med_gledestårer', annotations=frozenset({'kattefjes', 'katt', 'fjes'})),
EmojiAnnotations(emoji='😻', codepoints=(128571,), name='smilende katt med hjerteøyne', slug='smilende_katt_med_hjerteøyne', annotations=frozenset({'smilende kattefjes med hjerteøyne', 'kattefjes', 'øyne', 'forelsket', 'kjærlighet', 'katt', 'fjes', 'hjerte'})),
EmojiAnnotations(emoji='😼', codepoints=(128572,), name='katt med ironisk smil', slug='katt_med_ironisk_smil', annotations=frozenset({'ironisk katt'})),
EmojiAnnotations(emoji='😽', codepoints=(128573,), name='kyssende katt', slug='kyssende_katt', annotations=frozenset({'øyne', 'kyssende', 'lukket', 'kyss', 'katt', 'fjes', 'lukkede øyne'})),
EmojiAnnotations(emoji='🙀', codepoints=(128576,), name='sjokkert kattefjes', slug='sjokkert_kattefjes', annotations=frozenset({'katt', 'jøsses', 'fjes', 'overrasket'})),
EmojiAnnotations(emoji='😿', codepoints=(128575,), name='gråtende katt', slug='gråtende_katt', annotations=frozenset({'katt', 'gråtende kattefjes', 'fjes', 'gråt'})),
EmojiAnnotations(emoji='😾', codepoints=(128574,), name='trutmunn', slug='trutmunn', annotations=frozenset({'kattefjes', 'katt', 'kyssende katt', 'fjes'})),
EmojiAnnotations(emoji='🙈', codepoints=(128584,), name='ser ingenting ondt', slug='ser_ingenting_ondt', annotations=frozenset({'vis ape'})),
EmojiAnnotations(emoji='🙉', codepoints=(128585,), name='hører ingenting ondt', slug='hører_ingenting_ondt', annotations=frozenset({'vis ape'})),
EmojiAnnotations(emoji='🙊', codepoints=(128586,), name='sier ingenting ondt', slug='sier_ingenting_ondt', annotations=frozenset({'vis ape'})),
EmojiAnnotations(emoji='👦', codepoints=(128102,), name='gutt', slug='gutt', annotations=frozenset({'menneske'})),
EmojiAnnotations(emoji='👧', codepoints=(128103,), name='jente', slug='jente', annotations=frozenset({'menneske'})),
EmojiAnnotations(emoji='👨', codepoints=(128104,), name='mann', slug='mann', annotations=frozenset({'menneske'})),
EmojiAnnotations(emoji='👩', codepoints=(128105,), name='kvinne', slug='kvinne', annotations=frozenset({'menneske'})),
EmojiAnnotations(emoji='👴', codepoints=(128116,), name='eldre mann', slug='eldre_mann', annotations=frozenset({'mann', 'eldre', 'menneske'})),
EmojiAnnotations(emoji='👵', codepoints=(128117,), name='eldre kvinne', slug='eldre_kvinne', annotations=frozenset({'kvinne', 'eldre', 'menneske'})),
EmojiAnnotations(emoji='👶', codepoints=(128118,), name='baby', slug='baby', annotations=frozenset({'menneske'})),
EmojiAnnotations(emoji='👱', codepoints=(128113,), name='blond person', slug='blond_person', annotations=frozenset({'blond', 'menneske'})),
EmojiAnnotations(emoji='👮', codepoints=(128110,), name='politimann', slug='politimann', annotations=frozenset({'tjenestemann', 'menneske', 'politi'})),
EmojiAnnotations(emoji='👲', codepoints=(128114,), name='mann med lue', slug='mann_med_lue', annotations=frozenset({'mann med gua pi mao', 'menneske'})),
EmojiAnnotations(emoji='👳', codepoints=(128115,), name='mann med turban', slug='mann_med_turban', annotations=frozenset({'menneske'})),
EmojiAnnotations(emoji='👷', codepoints=(128119,), name='anleggsarbeider', slug='anleggsarbeider', annotations=frozenset({'arbeider', 'menneske', 'anlegg'})),
EmojiAnnotations(emoji='⛑', codepoints=(9937,), name='hjelm med hvitt kors', slug='hjelm_med_hvitt_kors', annotations=frozenset({'hatt', 'hjelm', 'fjes', 'hjelp', 'kors'})),
EmojiAnnotations(emoji='👸', codepoints=(128120,), name='prinsesse', slug='prinsesse', annotations=frozenset({'menneske', 'eventyr'})),
EmojiAnnotations(emoji='💂', codepoints=(128130,), name='gardist', slug='gardist', annotations=frozenset({'vaktmann', 'menneske'})),
EmojiAnnotations(emoji='\U0001f575', codepoints=(128373,), name='detektiv', slug='detektiv', annotations=frozenset({'spion'})),
EmojiAnnotations(emoji='🎅', codepoints=(127877,), name='julenissen', slug='julenissen', annotations=frozenset({'jul', 'feiring', 'nissefar', 'nissen'})),
EmojiAnnotations(emoji='👼', codepoints=(128124,), name='babyengel', slug='babyengel', annotations=frozenset({'ansikt', 'engel', 'eventyr'})),
EmojiAnnotations(emoji='💆', codepoints=(128134,), name='ansiktsmassasje', slug='ansiktsmassasje', annotations=frozenset({'massasje', 'skjønnhetssalong'})),
EmojiAnnotations(emoji='💇', codepoints=(128135,), name='hårklipp', slug='hårklipp', annotations=frozenset({'skjønnhetssalong', 'skjønnhet', 'salong'})),
EmojiAnnotations(emoji='👰', codepoints=(128112,), name='brud', slug='brud', annotations=frozenset({'bryllup', 'mennesker', 'brud med slør', 'slør'})),
EmojiAnnotations(emoji='🙍', codepoints=(128589,), name='rynker brynene', slug='rynker_brynene', annotations=frozenset({'person', 'person rynker brynene'})),
EmojiAnnotations(emoji='🙎', codepoints=(128590,), name='snurpemunn', slug='snurpemunn', annotations=frozenset({'fjes', 'kyssende person'})),
EmojiAnnotations(emoji='🙅', codepoints=(128581,), name='ikke bra', slug='ikke_bra', annotations=frozenset({'nei', 'forbudt'})),
EmojiAnnotations(emoji='🙆', codepoints=(128582,), name='OK-bevegelse', slug='ok_bevegelse', annotations=frozenset({'greit', 'fjes', 'bevegelse'})),
EmojiAnnotations(emoji='💁', codepoints=(128129,), name='informasjonsskranke', slug='informasjonsskranke', annotations=frozenset({'hjelp', 'menneske', 'informasjon'})),
EmojiAnnotations(emoji='🙋', codepoints=(128587,), name='rekker opp hånden', slug='rekker_opp_hånden', annotations=frozenset({'hånd', 'person', 'rekker opp'})),
EmojiAnnotations(emoji='🙇', codepoints=(128583,), name='unnskyld', slug='unnskyld', annotations=frozenset({'beklager', 'legger meg flat', 'menneske'})),
EmojiAnnotations(emoji='🙌', codepoints=(128588,), name='hurra', slug='hurra', annotations=frozenset({'begge hender', 'feirer', 'rekker begge armene i været', 'feiring'})),
EmojiAnnotations(emoji='🙏', codepoints=(128591,), name='foldede hender', slug='foldede_hender', annotations=frozenset({'hånd', 'be', 'bevegelse', 'bønn', 'ber'})),
EmojiAnnotations(emoji='\U0001f5e3', codepoints=(128483,), name='snakkende hode i profil', slug='snakkende_hode_i_profil', annotations=frozenset({'silhuett', 'hode', 'snakker', 'fjes', 'snakke'})),
EmojiAnnotations(emoji='👤', codepoints=(128100,), name='siluett av byste', slug='siluett_av_byste', annotations=frozenset({'byste', 'siluett', 'menneske'})),
EmojiAnnotations(emoji='👥', codepoints=(128101,), name='siluett av byster', slug='siluett_av_byster', annotations=frozenset({'byste', 'siluett', 'menneske'})),
EmojiAnnotations(emoji='🚶', codepoints=(128694,), name='fotgjenger', slug='fotgjenger', annotations=frozenset({'gående'})),
EmojiAnnotations(emoji='🏃', codepoints=(127939,), name='løping', slug='løping', annotations=frozenset({'sport', 'løper', 'marathon', 'jogging', 'jogger'})),
EmojiAnnotations(emoji='👯', codepoints=(128111,), name='kaninører', slug='kaninører', annotations=frozenset({'kaninjente', 'jente', 'kvinne', 'menneske'})),
EmojiAnnotations(emoji='💃', codepoints=(128131,), name='danser', slug='danser', annotations=frozenset({'menneske', 'dans'})),
EmojiAnnotations(emoji='\U0001f574', codepoints=(128372,), name='svevende mann i dress', slug='svevende_mann_i_dress', annotations=frozenset({'forretningsmann', 'mann', 'dress', 'business'})),
EmojiAnnotations(emoji='💏', codepoints=(128143,), name='kyss', slug='kyss', annotations=frozenset({'kyssing', 'romantikk', 'mennesker'})),
EmojiAnnotations(emoji='💑', codepoints=(128145,), name='par', slug='par', annotations=frozenset({'par med hjerte', 'romantikk', 'hjerte', 'mennesker'})),
EmojiAnnotations(emoji='👪', codepoints=(128106,), name='familie', slug='familie', annotations=frozenset({'mor', 'far', 'mennesker', 'barn'})),
EmojiAnnotations(emoji='👫', codepoints=(128107,), name='mann og kvinne', slug='mann_og_kvinne', annotations=frozenset({'kvinne', 'mennsker', 'mann', 'mann og kvinne holder hender', 'holder hender', 'par'})),
EmojiAnnotations(emoji='👬', codepoints=(128108,), name='to menn som leier', slug='to_menn_som_leier', annotations=frozenset({'mann', 'par', 'holde hender', 'to menn holder hender', 'mennesker'})),
EmojiAnnotations(emoji='👭', codepoints=(128109,), name='to kvinner som leier', slug='to_kvinner_som_leier', annotations=frozenset({'kvinne', 'par', 'holde hender', 'to kvinner holder hender', 'mennesker'})),
EmojiAnnotations(emoji='\U0001f3fb', codepoints=(127995,), name='hud type 1–2', slug='hud_type_1–2', annotations=frozenset({'fitzpatrick', 'emojimodifikator', 'hud', 'hudfarge'})),
EmojiAnnotations(emoji='\U0001f3fc', codepoints=(127996,), name='hud type 3', slug='hud_type_3', annotations=frozenset({'fitzpatrick', 'emojimodifikator', 'hud', 'hudfarge'})),
EmojiAnnotations(emoji='\U0001f3fd', codepoints=(127997,), name='hud type 4', slug='hud_type_4', annotations=frozenset({'fitzpatrick', 'emojimodifikator', 'hud', 'hudfarge'})),
EmojiAnnotations(emoji='\U0001f3fe', codepoints=(127998,), name='hud type 5', slug='hud_type_5', annotations=frozenset({'fitzpatrick', 'emojimodifikator', 'hud', 'hudfarge'})),
EmojiAnnotations(emoji='\U0001f3ff', codepoints=(127999,), name='hud type 6', slug='hud_type_6', annotations=frozenset({'fitzpatrick', 'emojimodifikator', 'hud', 'hudfarge'})),
EmojiAnnotations(emoji='💪', codepoints=(128170,), name='biceps', slug='biceps', annotations=frozenset({'muskelfleksing', 'tegneserie', 'muskel', 'sterk', 'fleks'})),
EmojiAnnotations(emoji='👈', codepoints=(128072,), name='finger som peker mot venstre', slug='finger_som_peker_mot_venstre', annotations=frozenset({'peker venstre', 'finger', 'pekende finger'})),
EmojiAnnotations(emoji='👉', codepoints=(128073,), name='finger som peker mot høyre', slug='finger_som_peker_mot_høyre', annotations=frozenset({'peker opp', 'finger', 'pekende finger'})),
EmojiAnnotations(emoji='☝', codepoints=(9757,), name='pekefinger som peker opp', slug='pekefinger_som_peker_opp', annotations=frozenset({'peke', 'finger', 'kropp', 'hånd', 'pekefinger', 'opp'})),
EmojiAnnotations(emoji='👆', codepoints=(128070,), name='oppoverpekende finger', slug='oppoverpekende_finger', annotations=frozenset({'peker opp', 'finger', 'pekende finger'})),
EmojiAnnotations(emoji='\U0001f595', codepoints=(128405,), name='langfinger', slug='langfinger', annotations=frozenset({'hånd', 'finger', 'kropp'})),
EmojiAnnotations(emoji='👇', codepoints=(128071,), name='nedoverpekende finger', slug='nedoverpekende_finger', annotations=frozenset({'peker ned', 'finger', 'pekende finger'})),
EmojiAnnotations(emoji='✌', codepoints=(9996,), name='seierstegn', slug='seierstegn', annotations=frozenset({'peace-tegn'})),
EmojiAnnotations(emoji='\U0001f596', codepoints=(128406,), name='Vulcan-hilsen', slug='vulcan_hilsen', annotations=frozenset({'hånd', 'finger', 'vulcan', 'kropp'})),
EmojiAnnotations(emoji='\U0001f918', codepoints=(129304,), name="rock'n'roll", slug="rock'n'roll", annotations=frozenset({'hånd', 'horn', 'finger', 'kropp'})),
EmojiAnnotations(emoji='\U0001f590', codepoints=(128400,), name='flat hånd med spredte fingre', slug='flat_hånd_med_spredte_fingre', annotations=frozenset({'hånd', 'finger', 'spredt', 'kropp'})),
EmojiAnnotations(emoji='✋', codepoints=(9995,), name='hevet hånd', slug='hevet_hånd', annotations=frozenset({'flat hånd', 'håndflate'})),
EmojiAnnotations(emoji='👌', codepoints=(128076,), name='OK-tegn', slug='ok_tegn', annotations=frozenset({'hånd', 'tegn', 'OK-hånd', 'OK'})),
EmojiAnnotations(emoji='👍', codepoints=(128077,), name='tommel opp', slug='tommel_opp', annotations=frozenset({'hånd', 'tegn', 'tommel'})),
EmojiAnnotations(emoji='👎', codepoints=(128078,), name='tommel ned', slug='tommel_ned', annotations=frozenset({'hånd', 'tegn', 'tommel'})),
EmojiAnnotations(emoji='✊', codepoints=(9994,), name='hevet knyttneve', slug='hevet_knyttneve', annotations=frozenset({'knyttet neve', 'neve'})),
EmojiAnnotations(emoji='👊', codepoints=(128074,), name='knyttet neve', slug='knyttet_neve', annotations=frozenset({'hånd', 'neve'})),
EmojiAnnotations(emoji='👋', codepoints=(128075,), name='vinkede hånd', slug='vinkede_hånd', annotations=frozenset({'vinking', 'vinke'})),
EmojiAnnotations(emoji='👏', codepoints=(128079,), name='klappende hender', slug='klappende_hender', annotations=frozenset({'hender', 'klapping'})),
EmojiAnnotations(emoji='👐', codepoints=(128080,), name='åpne hender', slug='åpne_hender', annotations=frozenset({'hender', 'åpne'})),
EmojiAnnotations(emoji='✍', codepoints=(9997,), name='skrivende hånd', slug='skrivende_hånd', annotations=frozenset({'hånd', 'skrive', 'kropp'})),
EmojiAnnotations(emoji='💅', codepoints=(128133,), name='neglpleie', slug='neglpleie', annotations=frozenset({'negl', 'manikyr', 'kosmetikk', 'neglelakk'})),
EmojiAnnotations(emoji='👂', codepoints=(128066,), name='øre', slug='øre', annotations=frozenset({'kropp'})),
EmojiAnnotations(emoji='👃', codepoints=(128067,), name='nese', slug='nese', annotations=frozenset({'fjes', 'kropp'})),
EmojiAnnotations(emoji='👣', codepoints=(128099,), name='fotavtrykk', slug='fotavtrykk', annotations=frozenset({'fot', 'avtrykk', 'kropp'})),
EmojiAnnotations(emoji='👀', codepoints=(128064,), name='øyne', slug='øyne', annotations=frozenset({'fjes'})),
EmojiAnnotations(emoji='\U0001f441', codepoints=(128065,), name='øye', slug='øye', annotations=frozenset({'kropp'})),
EmojiAnnotations(emoji='👅', codepoints=(128069,), name='tunge', slug='tunge', annotations=frozenset({'fjes', 'kropp'})),
EmojiAnnotations(emoji='👄', codepoints=(128068,), name='munn', slug='munn', annotations=frozenset({'lepper', 'fjes', 'kropp'})),
EmojiAnnotations(emoji='💋', codepoints=(128139,), name='kyssemerke', slug='kyssemerke', annotations=frozenset({'lepper', 'romantikk', 'kyss'})),
EmojiAnnotations(emoji='💘', codepoints=(128152,), name='hjerte med pil', slug='hjerte_med_pil', annotations=frozenset({'pil', 'hjerte', 'kjærlighet', 'følelser'})),
EmojiAnnotations(emoji='❤', codepoints=(10084,), name='rødt hjerte', slug='rødt_hjerte', annotations=frozenset({'hjerte'})),
EmojiAnnotations(emoji='💓', codepoints=(128147,), name='bankende hjerte', slug='bankende_hjerte', annotations=frozenset({'bankende', 'hjerte', 'kjærlighet', 'følelser'})),
EmojiAnnotations(emoji='💔', codepoints=(128148,), name='kjærlighetssorg', slug='kjærlighetssorg', annotations=frozenset({'brudd', 'følelser', 'hjerte', 'knust hjerte'})),
EmojiAnnotations(emoji='💕', codepoints=(128149,), name='to hjerter', slug='to_hjerter', annotations=frozenset({'følelser', 'kjærlighet', 'elskere'})),
EmojiAnnotations(emoji='💖', codepoints=(128150,), name='skinnende hjerte', slug='skinnende_hjerte', annotations=frozenset({'skinnende', 'kjærlighet', 'følelser'})),
EmojiAnnotations(emoji='💗', codepoints=(128151,), name='voksende hjerte', slug='voksende_hjerte', annotations=frozenset({'spent', 'nervøs', 'kjærlighet', 'følelser'})),
EmojiAnnotations(emoji='💙', codepoints=(128153,), name='blått hjerte', slug='blått_hjerte', annotations=frozenset({'blå', 'hjerte', 'følelser'})),
EmojiAnnotations(emoji='💚', codepoints=(128154,), name='grønt hjerte', slug='grønt_hjerte', annotations=frozenset({'grønn', 'hjerte', 'følelser'})),
EmojiAnnotations(emoji='💛', codepoints=(128155,), name='gult hjerte', slug='gult_hjerte', annotations=frozenset({'hjerte', 'gul', 'følelser'})),
EmojiAnnotations(emoji='💜', codepoints=(128156,), name='lilla hjerte', slug='lilla_hjerte', annotations=frozenset({'lilla', 'hjerte', 'følelser'})),
EmojiAnnotations(emoji='💝', codepoints=(128157,), name='hjerte med bånd', slug='hjerte_med_bånd', annotations=frozenset({'bånd', 'hjerte', 'valentine', 'følelser'})),
EmojiAnnotations(emoji='💞', codepoints=(128158,), name='snurrende hjerter', slug='snurrende_hjerter', annotations=frozenset({'søtt', 'snurrende', 'kjærlighet', 'hjerter', 'følelser'})),
EmojiAnnotations(emoji='💟', codepoints=(128159,), name='hjertedekorasjon', slug='hjertedekorasjon', annotations=frozenset({'hjerte'})),
EmojiAnnotations(emoji='❣', codepoints=(10083,), name='utropstegn med hjerte', slug='utropstegn_med_hjerte', annotations=frozenset({'tegn', 'uttropstegn', 'tegnsetting', 'hjerte', 'uttrop'})),
EmojiAnnotations(emoji='💌', codepoints=(128140,), name='kjærlighesbrev', slug='kjærlighesbrev', annotations=frozenset({'brev med hjerte', 'romantikk', 'hjerte', 'brev'})),
EmojiAnnotations(emoji='💤', codepoints=(128164,), name='trøtt', slug='trøtt', annotations=frozenset({'zzz', 'snorker', 'følelse', 'sover', 'tegneserie', 'søving'})),
EmojiAnnotations(emoji='💢', codepoints=(128162,), name='sinne', slug='sinne', annotations=frozenset({'sint', 'tegneserie', 'sinnesymbol', 'følelse'})),
EmojiAnnotations(emoji='💣', codepoints=(128163,), name='bombe', slug='bombe', annotations=frozenset({'tegneserie', 'følelse'})),
EmojiAnnotations(emoji='💥', codepoints=(128165,), name='krasj', slug='krasj', annotations=frozenset({'kollisjon', 'tegneserie', 'smell', 'følelse'})),
EmojiAnnotations(emoji='💦', codepoints=(128166,), name='svettedråper', slug='svettedråper', annotations=frozenset({'tegneserie', 'svette', 'sprut', 'sprutende'})),
EmojiAnnotations(emoji='💨', codepoints=(128168,), name='puff', slug='puff', annotations=frozenset({'tegneserie', 'forsvinne', 'løpe'})),
EmojiAnnotations(emoji='💫', codepoints=(128171,), name='svimmel', slug='svimmel', annotations=frozenset({'tegneserie', 'stjerner', 'se stjerner'})),
EmojiAnnotations(emoji='💬', codepoints=(128172,), name='snakkeboble', slug='snakkeboble', annotations=frozenset({'ballong', 'tegneserie', 'dialog', 'dialogboble', 'boble'})),
EmojiAnnotations(emoji='\U0001f5e8', codepoints=(128488,), name='snakkeboble mot venstre', slug='snakkeboble_mot_venstre', annotations=frozenset({'tale', 'dialog'})),
EmojiAnnotations(emoji='\U0001f5ef', codepoints=(128495,), name='sint snakkeboble mot høyre', slug='sint_snakkeboble_mot_høyre', annotations=frozenset({'sint', 'ballong', 'sur', 'boble'})),
EmojiAnnotations(emoji='💭', codepoints=(128173,), name='tankeboble', slug='tankeboble', annotations=frozenset({'tegneserie', 'tanke', 'boble'})),
EmojiAnnotations(emoji='\U0001f573', codepoints=(128371,), name='hull i bakken', slug='hull_i_bakken', annotations=frozenset({'hull'})),
EmojiAnnotations(emoji='👓', codepoints=(128083,), name='briller', slug='briller', annotations=frozenset({'lesebriller', 'klær'})),
EmojiAnnotations(emoji='\U0001f576', codepoints=(128374,), name='solbriller', slug='solbriller', annotations=frozenset({'mørkt: øye', 'briller'})),
EmojiAnnotations(emoji='👔', codepoints=(128084,), name='slips', slug='slips', annotations=frozenset({'klær'})),
EmojiAnnotations(emoji='👕', codepoints=(128085,), name='t-skjorte', slug='t_skjorte', annotations=frozenset({'skjorte', 'klær'})),
EmojiAnnotations(emoji='👖', codepoints=(128086,), name='bukser', slug='bukser', annotations=frozenset({'jeans', 'klær'})),
EmojiAnnotations(emoji='👗', codepoints=(128087,), name='kjole', slug='kjole', annotations=frozenset({'klær'})),
EmojiAnnotations(emoji='👘', codepoints=(128088,), name='kimono', slug='kimono', annotations=frozenset({'kjole', 'klær'})),
EmojiAnnotations(emoji='👙', codepoints=(128089,), name='bikini', slug='bikini', annotations=frozenset({'badetøy', 'badedrakt', 'klær'})),
EmojiAnnotations(emoji='👚', codepoints=(128090,), name='dameklær', slug='dameklær', annotations=frozenset({'klær'})),
EmojiAnnotations(emoji='👛', codepoints=(128091,), name='portemoné', slug='portemoné', annotations=frozenset({'klær', 'pung'})),
EmojiAnnotations(emoji='👜', codepoints=(128092,), name='håndveske', slug='håndveske', annotations=frozenset({'klær', 'veske'})),
EmojiAnnotations(emoji='👝', codepoints=(128093,), name='toalettveske', slug='toalettveske', annotations=frozenset({'liten veske', 'klær'})),
EmojiAnnotations(emoji='\U0001f6cd', codepoints=(128717,), name='handleposer', slug='handleposer', annotations=frozenset({'shopping', 'hotell', 'pose'})),
EmojiAnnotations(emoji='🎒', codepoints=(127890,), name='skolesekk', slug='skolesekk', annotations=frozenset({'ryggsekk', 'ransel'})),
EmojiAnnotations(emoji='👞', codepoints=(128094,), name='herresko', slug='herresko', annotations=frozenset({'sko', 'klær'})),
EmojiAnnotations(emoji='👟', codepoints=(128095,), name='joggesko', slug='joggesko', annotations=frozenset({'tennissko', 'fritidssko', 'klær'})),
EmojiAnnotations(emoji='👠', codepoints=(128096,), name='høyhælte', slug='høyhælte', annotations=frozenset({'sko', 'klær'})),
EmojiAnnotations(emoji='👡', codepoints=(128097,), name='damesandal', slug='damesandal', annotations=frozenset({'sandal', 'dame', 'klær'})),
EmojiAnnotations(emoji='👢', codepoints=(128098,), name='damestøvletter', slug='damestøvletter', annotations=frozenset({'støvlett', 'dame', 'klær'})),
EmojiAnnotations(emoji='👑', codepoints=(128081,), name='krone', slug='krone', annotations=frozenset({'klær'})),
EmojiAnnotations(emoji='👒', codepoints=(128082,), name='damehatt', slug='damehatt', annotations=frozenset({'klær', 'hatt'})),
EmojiAnnotations(emoji='🎩', codepoints=(127913,), name='flosshatt', slug='flosshatt', annotations=frozenset({'klær', 'hatt'})),
EmojiAnnotations(emoji='🎓', codepoints=(127891,), name='studentlue', slug='studentlue', annotations=frozenset({'feiring', 'uteksaminering', 'hatt'})),
EmojiAnnotations(emoji='\U0001f4ff', codepoints=(128255,), name='bønneperler', slug='bønneperler', annotations=frozenset({'bønn', 'perler', 'religion', 'klær', 'halssmykke'})),
EmojiAnnotations(emoji='💄', codepoints=(128132,), name='leppestift', slug='leppestift', annotations=frozenset({'sminke', 'kosmetikk'})),
EmojiAnnotations(emoji='💍', codepoints=(128141,), name='ring', slug='ring', annotations=frozenset({'romantikk'})),
EmojiAnnotations(emoji='💎', codepoints=(128142,), name='diamant', slug='diamant', annotations=frozenset({'smykkestein', 'edelsten', 'romantikk'})),
EmojiAnnotations(emoji='🐵', codepoints=(128053,), name='apefjes', slug='apefjes', annotations=frozenset({'dyr', 'ape', 'fjes'})),
EmojiAnnotations(emoji='🐒', codepoints=(128018,), name='ape', slug='ape', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐶', codepoints=(128054,), name='hundefjes', slug='hundefjes', annotations=frozenset({'dyr', 'hund', 'fjes'})),
EmojiAnnotations(emoji='🐕', codepoints=(128021,), name='hund', slug='hund', annotations=frozenset({'dyr', 'valp'})),
EmojiAnnotations(emoji='🐩', codepoints=(128041,), name='puddel', slug='puddel', annotations=frozenset({'dyr', 'hund'})),
EmojiAnnotations(emoji='🐺', codepoints=(128058,), name='ulvehode', slug='ulvehode', annotations=frozenset({'hode', 'dyr', 'ulv'})),
EmojiAnnotations(emoji='🐱', codepoints=(128049,), name='kattefjes', slug='kattefjes', annotations=frozenset({'katt', 'dyr', 'fjes'})),
EmojiAnnotations(emoji='🐈', codepoints=(128008,), name='katt', slug='katt', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='\U0001f981', codepoints=(129409,), name='løvefjes', slug='løvefjes', annotations=frozenset({'løven', 'fjes', 'stjernetegn', 'løve'})),
EmojiAnnotations(emoji='🐯', codepoints=(128047,), name='tigerfjes', slug='tigerfjes', annotations=frozenset({'dyr', 'tiger'})),
EmojiAnnotations(emoji='🐅', codepoints=(128005,), name='tiger', slug='tiger', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐆', codepoints=(128006,), name='leopard', slug='leopard', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐴', codepoints=(128052,), name='hestehode', slug='hestehode', annotations=frozenset({'hode', 'dyr', 'hest'})),
EmojiAnnotations(emoji='🐎', codepoints=(128014,), name='hest', slug='hest', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='\U0001f984', codepoints=(129412,), name='enhjørning', slug='enhjørning', annotations=frozenset({'fjes'})),
EmojiAnnotations(emoji='🐮', codepoints=(128046,), name='kufjes', slug='kufjes', annotations=frozenset({'ku', 'dyr', 'fjes'})),
EmojiAnnotations(emoji='🐂', codepoints=(128002,), name='okse', slug='okse', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐃', codepoints=(128003,), name='vannbøffel', slug='vannbøffel', annotations=frozenset({'dyr', 'bøffel'})),
EmojiAnnotations(emoji='🐄', codepoints=(128004,), name='ku', slug='ku', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐷', codepoints=(128055,), name='grisehode', slug='grisehode', annotations=frozenset({'hode', 'gris', 'dyr'})),
EmojiAnnotations(emoji='🐖', codepoints=(128022,), name='gris', slug='gris', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐗', codepoints=(128023,), name='villsvin', slug='villsvin', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐽', codepoints=(128061,), name='grisesnute', slug='grisesnute', annotations=frozenset({'gris', 'dyr', 'snute'})),
EmojiAnnotations(emoji='🐏', codepoints=(128015,), name='bukk', slug='bukk', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐑', codepoints=(128017,), name='sau', slug='sau', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐐', codepoints=(128016,), name='geit', slug='geit', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐪', codepoints=(128042,), name='dromedar', slug='dromedar', annotations=frozenset({'kamel', 'dyr', 'pukkel'})),
EmojiAnnotations(emoji='🐫', codepoints=(128043,), name='kamel', slug='kamel', annotations=frozenset({'dyr', 'pukkel'})),
EmojiAnnotations(emoji='🐘', codepoints=(128024,), name='elefant', slug='elefant', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐭', codepoints=(128045,), name='musefjes', slug='musefjes', annotations=frozenset({'dyr', 'fjes', 'mus'})),
EmojiAnnotations(emoji='🐁', codepoints=(128001,), name='mus', slug='mus', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐀', codepoints=(128000,), name='rotte', slug='rotte', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐹', codepoints=(128057,), name='hamsterfjes', slug='hamsterfjes', annotations=frozenset({'hamster', 'dyr', 'fjes'})),
EmojiAnnotations(emoji='🐰', codepoints=(128048,), name='kaninfjes', slug='kaninfjes', annotations=frozenset({'dyr', 'fjes', 'kanin'})),
EmojiAnnotations(emoji='🐇', codepoints=(128007,), name='kanin', slug='kanin', annotations=frozenset({'dyr', 'hare'})),
EmojiAnnotations(emoji='\U0001f43f', codepoints=(128063,), name='ekorn', slug='ekorn', annotations=frozenset({'dyr', 'jordekorn'})),
EmojiAnnotations(emoji='🐻', codepoints=(128059,), name='bjørnefjes', slug='bjørnefjes', annotations=frozenset({'bjørn', 'fjes', 'dyr'})),
EmojiAnnotations(emoji='🐨', codepoints=(128040,), name='koala', slug='koala', annotations=frozenset({'koalabjørn', 'dyr'})),
EmojiAnnotations(emoji='🐼', codepoints=(128060,), name='pandafjes', slug='pandafjes', annotations=frozenset({'dyr', 'fjes', 'panda'})),
EmojiAnnotations(emoji='🐾', codepoints=(128062,), name='pote', slug='pote', annotations=frozenset({'poteavtrykk', 'dyr', 'avtrykk'})),
EmojiAnnotations(emoji='🐔', codepoints=(128020,), name='høne', slug='høne', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐓', codepoints=(128019,), name='hane', slug='hane', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐣', codepoints=(128035,), name='nyklekket kylling', slug='nyklekket_kylling', annotations=frozenset({'dyr', 'kylling', 'klekking'})),
EmojiAnnotations(emoji='🐤', codepoints=(128036,), name='kylling', slug='kylling', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐥', codepoints=(128037,), name='helfigurs kylling', slug='helfigurs_kylling', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐦', codepoints=(128038,), name='fugl', slug='fugl', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐧', codepoints=(128039,), name='pingvin', slug='pingvin', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='\U0001f54a', codepoints=(128330,), name='fredsdue', slug='fredsdue', annotations=frozenset({'fred', 'fly', 'fugl'})),
EmojiAnnotations(emoji='🐸', codepoints=(128056,), name='froskefjes', slug='froskefjes', annotations=frozenset({'frosk', 'dyr', 'fjes'})),
EmojiAnnotations(emoji='🐊', codepoints=(128010,), name='krokodille', slug='krokodille', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐢', codepoints=(128034,), name='skilpadde', slug='skilpadde', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐍', codepoints=(128013,), name='slange', slug='slange', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐲', codepoints=(128050,), name='dragehode', slug='dragehode', annotations=frozenset({'hode', 'drage', 'dyr'})),
EmojiAnnotations(emoji='🐉', codepoints=(128009,), name='drage', slug='drage', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐳', codepoints=(128051,), name='sprutende hval', slug='sprutende_hval', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐋', codepoints=(128011,), name='hval', slug='hval', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐬', codepoints=(128044,), name='delfin', slug='delfin', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐟', codepoints=(128031,), name='fisk', slug='fisk', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐠', codepoints=(128032,), name='tropisk fisk', slug='tropisk_fisk', annotations=frozenset({'fisk', 'dyr'})),
EmojiAnnotations(emoji='🐡', codepoints=(128033,), name='kulefisk', slug='kulefisk', annotations=frozenset({'fisk', 'dyr'})),
EmojiAnnotations(emoji='🐙', codepoints=(128025,), name='blekksprut', slug='blekksprut', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐚', codepoints=(128026,), name='skjell', slug='skjell', annotations=frozenset({'spiralskjell', 'dyr'})),
EmojiAnnotations(emoji='\U0001f980', codepoints=(129408,), name='krabbe', slug='krabbe', annotations=frozenset({'krepsen', 'stjernetegn'})),
EmojiAnnotations(emoji='🐌', codepoints=(128012,), name='snegle', slug='snegle', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐛', codepoints=(128027,), name='insekt', slug='insekt', annotations=frozenset({'åme', 'dyr', 'larve'})),
EmojiAnnotations(emoji='🐜', codepoints=(128028,), name='maur', slug='maur', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐝', codepoints=(128029,), name='honningbie', slug='honningbie', annotations=frozenset({'dyr', 'bie'})),
EmojiAnnotations(emoji='🐞', codepoints=(128030,), name='bille', slug='bille', annotations=frozenset({'dyr', 'insekt', 'tordivel'})),
EmojiAnnotations(emoji='\U0001f577', codepoints=(128375,), name='edderkopp', slug='edderkopp', annotations=frozenset({'insekt'})),
EmojiAnnotations(emoji='\U0001f578', codepoints=(128376,), name='spindelvev', slug='spindelvev', annotations=frozenset({'edderkopp', 'nett'})),
EmojiAnnotations(emoji='\U0001f982', codepoints=(129410,), name='skorpion', slug='skorpion', annotations=frozenset({'skorpionen', 'stjernetegn'})),
EmojiAnnotations(emoji='💐', codepoints=(128144,), name='bukett', slug='bukett', annotations=frozenset({'blomster', 'romantikk'})),
EmojiAnnotations(emoji='🌸', codepoints=(127800,), name='kisebærblomst', slug='kisebærblomst', annotations=frozenset({'blomst', 'hvit blomst'})),
EmojiAnnotations(emoji='💮', codepoints=(128174,), name='hvit blomst', slug='hvit_blomst', annotations=frozenset({'blomsterstempel', 'blomst', 'stempel'})),
EmojiAnnotations(emoji='\U0001f3f5', codepoints=(127989,), name='rosett', slug='rosett', annotations=frozenset({'plante'})),
EmojiAnnotations(emoji='🌹', codepoints=(127801,), name='rose', slug='rose', annotations=frozenset({'blomst'})),
EmojiAnnotations(emoji='🌺', codepoints=(127802,), name='hibiskus', slug='hibiskus', annotations=frozenset({'blomst'})),
EmojiAnnotations(emoji='🌻', codepoints=(127803,), name='solsikke', slug='solsikke', annotations=frozenset({'blomst'})),
EmojiAnnotations(emoji='🌼', codepoints=(127804,), name='blomst', slug='blomst', annotations=frozenset({'plante'})),
EmojiAnnotations(emoji='🌷', codepoints=(127799,), name='tulipan', slug='tulipan', annotations=frozenset({'blomst'})),
EmojiAnnotations(emoji='🌱', codepoints=(127793,), name='spire', slug='spire', annotations=frozenset({'ung plante'})),
EmojiAnnotations(emoji='🌲', codepoints=(127794,), name='grantre', slug='grantre', annotations=frozenset({'tre', 'juletre', 'eviggrønt tre', 'gran'})),
EmojiAnnotations(emoji='🌳', codepoints=(127795,), name='løvtre', slug='løvtre', annotations=frozenset({'tre', 'tre med blader'})),
EmojiAnnotations(emoji='🌴', codepoints=(127796,), name='palme', slug='palme', annotations=frozenset({'palmetre'})),
EmojiAnnotations(emoji='🌵', codepoints=(127797,), name='kaktus', slug='kaktus', annotations=frozenset({'plante'})),
EmojiAnnotations(emoji='🌾', codepoints=(127806,), name='risplante', slug='risplante', annotations=frozenset({'strå', 'siv'})),
EmojiAnnotations(emoji='🌿', codepoints=(127807,), name='urteblader', slug='urteblader', annotations=frozenset({'urte'})),
EmojiAnnotations(emoji='☘', codepoints=(9752,), name='trekløver', slug='trekløver', annotations=frozenset({'plante'})),
EmojiAnnotations(emoji='🍀', codepoints=(127808,), name='firkløver', slug='firkløver', annotations=frozenset({'hell', 'kløver', 'flaks'})),
EmojiAnnotations(emoji='🍁', codepoints=(127809,), name='lønneblad', slug='lønneblad', annotations=frozenset({'høstblad', 'lønn', 'blad'})),
EmojiAnnotations(emoji='🍂', codepoints=(127810,), name='høstblader', slug='høstblader', annotations=frozenset({'fallende blader', 'løv'})),
EmojiAnnotations(emoji='🍃', codepoints=(127811,), name='blader i vind', slug='blader_i_vind', annotations=frozenset({'vind', 'løv i vind', 'blåst', 'blåser'})),
EmojiAnnotations(emoji='🍇', codepoints=(127815,), name='druer', slug='druer', annotations=frozenset({'drueklase', 'frukt'})),
EmojiAnnotations(emoji='🍈', codepoints=(127816,), name='honningmelon', slug='honningmelon', annotations=frozenset({'melon', 'frukt'})),
EmojiAnnotations(emoji='🍉', codepoints=(127817,), name='vannmelon', slug='vannmelon', annotations=frozenset({'melon', 'frukt'})),
EmojiAnnotations(emoji='🍊', codepoints=(127818,), name='mandarin', slug='mandarin', annotations=frozenset({'klementin', 'appelsin', 'frukt'})),
EmojiAnnotations(emoji='🍋', codepoints=(127819,), name='sitron', slug='sitron', annotations=frozenset({'sitrus', 'frukt'})),
EmojiAnnotations(emoji='🍌', codepoints=(127820,), name='banan', slug='banan', annotations=frozenset({'frukt'})),
EmojiAnnotations(emoji='🍍', codepoints=(127821,), name='ananas', slug='ananas', annotations=frozenset({'frukt'})),
EmojiAnnotations(emoji='🍎', codepoints=(127822,), name='rødt eple', slug='rødt_eple', annotations=frozenset({'eple', 'frukt'})),
EmojiAnnotations(emoji='🍏', codepoints=(127823,), name='grønt eple', slug='grønt_eple', annotations=frozenset({'eple', 'frukt'})),
EmojiAnnotations(emoji='🍐', codepoints=(127824,), name='pære', slug='pære', annotations=frozenset({'frukt'})),
EmojiAnnotations(emoji='🍑', codepoints=(127825,), name='fersken', slug='fersken', annotations=frozenset({'frukt'})),
EmojiAnnotations(emoji='🍒', codepoints=(127826,), name='kirsebær', slug='kirsebær', annotations=frozenset({'moreller', 'bær'})),
EmojiAnnotations(emoji='🍓', codepoints=(127827,), name='jordbær', slug='jordbær', annotations=frozenset({'bær'})),
EmojiAnnotations(emoji='🍅', codepoints=(127813,), name='tomat', slug='tomat', annotations=frozenset({'grønnsak'})),
EmojiAnnotations(emoji='🍆', codepoints=(127814,), name='aubergine', slug='aubergine', annotations=frozenset({'grønnsak'})),
EmojiAnnotations(emoji='🌽', codepoints=(127805,), name='maiskolbe', slug='maiskolbe', annotations=frozenset({'mais'})),
EmojiAnnotations(emoji='\U0001f336', codepoints=(127798,), name='rød chili', slug='rød_chili', annotations=frozenset({'plante', 'chili', 'sterk'})),
EmojiAnnotations(emoji='🍄', codepoints=(127812,), name='sopp', slug='sopp', annotations=frozenset({'fluesopp'})),
EmojiAnnotations(emoji='🌰', codepoints=(127792,), name='kastanje', slug='kastanje', annotations=frozenset({'kastanjenøtt'})),
EmojiAnnotations(emoji='🍞', codepoints=(127838,), name='brød', slug='brød', annotations=frozenset({'restaurant', 'loff'})),
EmojiAnnotations(emoji='\U0001f9c0', codepoints=(129472,), name='ostebit', slug='ostebit', annotations=frozenset({'ost'})),
EmojiAnnotations(emoji='🍖', codepoints=(127830,), name='kjøtt', slug='kjøtt', annotations=frozenset({'restaurant', 'kjøttbein'})),
EmojiAnnotations(emoji='🍗', codepoints=(127831,), name='kyllinglår', slug='kyllinglår', annotations=frozenset({'restaurant', 'fjærkre', 'kyllingklubbe'})),
EmojiAnnotations(emoji='🍔', codepoints=(127828,), name='hamburger', slug='hamburger', annotations=frozenset({'restaurant', 'burger'})),
EmojiAnnotations(emoji='🍟', codepoints=(127839,), name='pommes frites', slug='pommes_frites', annotations=frozenset({'restaurant', 'chips'})),
EmojiAnnotations(emoji='🍕', codepoints=(127829,), name='pizzastykke', slug='pizzastykke', annotations=frozenset({'pizzabit', 'pizza', 'restaurant'})),
EmojiAnnotations(emoji='\U0001f32d', codepoints=(127789,), name='pølse', slug='pølse', annotations=frozenset({'hotdog', 'frankfurterpølse'})),
EmojiAnnotations(emoji='\U0001f32e', codepoints=(127790,), name='taco', slug='taco', annotations=frozenset({'meksikansk'})),
EmojiAnnotations(emoji='\U0001f32f', codepoints=(127791,), name='burrito', slug='burrito', annotations=frozenset({'meksikansk'})),
EmojiAnnotations(emoji='🍲', codepoints=(127858,), name='gryterett', slug='gryterett', annotations=frozenset({'middagsmat', 'middag'})),
EmojiAnnotations(emoji='🍱', codepoints=(127857,), name='lunsjboks', slug='lunsjboks', annotations=frozenset({'matboks', 'bentoboks'})),
EmojiAnnotations(emoji='🍘', codepoints=(127832,), name='riskjeks', slug='riskjeks', annotations=frozenset({'ris'})),
EmojiAnnotations(emoji='🍙', codepoints=(127833,), name='risball', slug='risball', annotations=frozenset({'ris', 'restaurant', 'onigiri'})),
EmojiAnnotations(emoji='🍚', codepoints=(127834,), name='kokt ris', slug='kokt_ris', annotations=frozenset({'ris', 'restaurant'})),
EmojiAnnotations(emoji='🍛', codepoints=(127835,), name='curry og ris', slug='curry_og_ris', annotations=frozenset({'ris', 'restaurant', 'curry'})),
EmojiAnnotations(emoji='🍜', codepoints=(127836,), name='nudler', slug='nudler', annotations=frozenset({'restaurant', 'nudelbolle'})),
EmojiAnnotations(emoji='🍝', codepoints=(127837,), name='spaghetti', slug='spaghetti', annotations=frozenset({'restaurant', 'pasta'})),
EmojiAnnotations(emoji='🍠', codepoints=(127840,), name='bakt potet', slug='bakt_potet', annotations=frozenset({'restaurant', 'potet', 'søtpotet'})),
EmojiAnnotations(emoji='🍢', codepoints=(127842,), name='oden', slug='oden', annotations=frozenset({'restaurant', 'sjømat', 'sjømat på pinne'})),
EmojiAnnotations(emoji='🍣', codepoints=(127843,), name='sushi', slug='sushi', annotations=frozenset({'restaurant'})),
EmojiAnnotations(emoji='🍤', codepoints=(127844,), name='reke', slug='reke', annotations=frozenset({'restaurant', 'scampi', 'kongereke', 'grillet scampi'})),
EmojiAnnotations(emoji='🍥', codepoints=(127845,), name='fiskekake', slug='fiskekake', annotations=frozenset({'restaurant', 'fisk'})),
EmojiAnnotations(emoji='🍡', codepoints=(127841,), name='mochi', slug='mochi', annotations=frozenset({'restaurant', 'riskonfekt'})),
EmojiAnnotations(emoji='🍦', codepoints=(127846,), name='soft-is', slug='soft_is', annotations=frozenset({'is', 'iskrem', 'restaurant'})),
EmojiAnnotations(emoji='🍧', codepoints=(127847,), name='vannis', slug='vannis', annotations=frozenset({'is', 'restaurant', 'shave-ice'})),
EmojiAnnotations(emoji='🍨', codepoints=(127848,), name='iskrem', slug='iskrem', annotations=frozenset({'is', 'restaurant'})),
EmojiAnnotations(emoji='🍩', codepoints=(127849,), name='donut', slug='donut', annotations=frozenset({'smultring', 'restaurant'})),
EmojiAnnotations(emoji='🍪', codepoints=(127850,), name='kjeks', slug='kjeks', annotations=frozenset({'restaurant', 'cookie'})),
EmojiAnnotations(emoji='🎂', codepoints=(127874,), name='bursdagskake', slug='bursdagskake', annotations=frozenset({'kake', 'bursdag', 'feiring'})),
EmojiAnnotations(emoji='🍰', codepoints=(127856,), name='kakestykke', slug='kakestykke', annotations=frozenset({'kake', 'dessert', 'formkake'})),
EmojiAnnotations(emoji='🍫', codepoints=(127851,), name='sjokolade', slug='sjokolade', annotations=frozenset({'sjokoladeplate'})),
EmojiAnnotations(emoji='🍬', codepoints=(127852,), name='godteri', slug='godteri', annotations=frozenset({'karamell', 'søtsaker', 'smågodt', 'sukkertøy'})),
EmojiAnnotations(emoji='🍭', codepoints=(127853,), name='kjærlighet på pinne', slug='kjærlighet_på_pinne', annotations=frozenset({'sukkertøy'})),
EmojiAnnotations(emoji='🍮', codepoints=(127854,), name='pudding', slug='pudding', annotations=frozenset({'karamellpudding', 'dessert'})),
EmojiAnnotations(emoji='🍯', codepoints=(127855,), name='honning', slug='honning', annotations=frozenset({'honningkrukke', 'krukke'})),
EmojiAnnotations(emoji='🍼', codepoints=(127868,), name='tåteflaske', slug='tåteflaske', annotations=frozenset({'drikke', 'babyflaske', 'melk'})),
EmojiAnnotations(emoji='☕', codepoints=(9749,), name='varm drikke', slug='varm_drikke', annotations=frozenset({'te', 'drikke', 'kaffe', 'varm', 'rykende'})),
EmojiAnnotations(emoji='🍵', codepoints=(127861,), name='tekopp', slug='tekopp', annotations=frozenset({'suppe', 'te', 'drikke'})),
EmojiAnnotations(emoji='🍶', codepoints=(127862,), name='sake', slug='sake', annotations=frozenset({'mugge og kopp', 'karaffel', 'flaske', 'drikke'})),
EmojiAnnotations(emoji='\U0001f37e', codepoints=(127870,), name='flaske med sprettkork', slug='flaske_med_sprettkork', annotations=frozenset({'bar', 'kork', 'sprette', 'flaske', 'drikke'})),
EmojiAnnotations(emoji='🍷', codepoints=(127863,), name='vinglass', slug='vinglass', annotations=frozenset({'rødvin', 'vin', 'drikke', 'glass'})),
EmojiAnnotations(emoji='🍸', codepoints=(127864,), name='cocktail', slug='cocktail', annotations=frozenset({'bar', 'cocktaildrink', 'drink', 'cocktailglass'})),
EmojiAnnotations(emoji='🍹', codepoints=(127865,), name='tropisk drink', slug='tropisk_drink', annotations=frozenset({'bar', 'juice', 'drink'})),
EmojiAnnotations(emoji='🍺', codepoints=(127866,), name='øl', slug='øl', annotations=frozenset({'halvliter', 'pils', 'pilsner', 'bar', 'ølkrus', 'ølglass'})),
EmojiAnnotations(emoji='🍻', codepoints=(127867,), name='skål', slug='skål', annotations=frozenset({'halvliter', 'pils', 'pilsner', 'bar', 'ølkrus', 'øl', 'ølglass'})),
EmojiAnnotations(emoji='\U0001f37d', codepoints=(127869,), name='tallerken med kniv og gaffel', slug='tallerken_med_kniv_og_gaffel', annotations=frozenset({'kniv', 'gaffel', 'tallerken', 'matlaging'})),
EmojiAnnotations(emoji='🍴', codepoints=(127860,), name='kniv og gaffel', slug='kniv_og_gaffel', annotations=frozenset({'spisested', 'kniv', 'gaffel'})),
EmojiAnnotations(emoji='🍳', codepoints=(127859,), name='speilegg', slug='speilegg', annotations=frozenset({'stekepanne', 'egg', 'matlaging'})),
EmojiAnnotations(emoji='\U0001f3fa', codepoints=(127994,), name='amforakrukke', slug='amforakrukke', annotations=frozenset({'stjernetegn', 'vannmannen', 'drikke', 'verktøy', 'matlaging', 'våpen', 'krukke'})),
EmojiAnnotations(emoji='🌍', codepoints=(127757,), name='jordklode som viser Europa og Afrika', slug='jordklode_som_viser_europa_og_afrika', annotations=frozenset({'Europa', 'Afrika', 'jorden'})),
EmojiAnnotations(emoji='🌎', codepoints=(127758,), name='jordklode som viser Amerika', slug='jordklode_som_viser_amerika', annotations=frozenset({'USA', 'jorden', 'Amerika'})),
EmojiAnnotations(emoji='🌏', codepoints=(127759,), name='jordklode som viser Asia og Australia', slug='jordklode_som_viser_asia_og_australia', annotations=frozenset({'Asia', 'jorden', 'Australia'})),
EmojiAnnotations(emoji='🌐', codepoints=(127760,), name='jordklode med meridianer', slug='jordklode_med_meridianer', annotations=frozenset({'lengde- og breddegrader'})),
EmojiAnnotations(emoji='\U0001f5fa', codepoints=(128506,), name='verdenskart', slug='verdenskart', annotations=frozenset({'kart', 'verden'})),
EmojiAnnotations(emoji='\U0001f3d4', codepoints=(127956,), name='snødekt fjell', slug='snødekt_fjell', annotations=frozenset({'kaldt', 'fjell', 'snø'})),
EmojiAnnotations(emoji='🌋', codepoints=(127755,), name='vulkan', slug='vulkan', annotations=frozenset({'vulkanutbrudd'})),
EmojiAnnotations(emoji='🗻', codepoints=(128507,), name='Fuji-fjellet', slug='fuji_fjellet', annotations=frozenset({'fjell'})),
EmojiAnnotations(emoji='\U0001f3d6', codepoints=(127958,), name='strand med parasoll', slug='strand_med_parasoll', annotations=frozenset({'parasoll', 'strand'})),
EmojiAnnotations(emoji='\U0001f3dd', codepoints=(127965,), name='øde øy', slug='øde_øy', annotations=frozenset({'øde', 'øy'})),
EmojiAnnotations(emoji='\U0001f3de', codepoints=(127966,), name='nasjonalpark', slug='nasjonalpark', annotations=frozenset({'park'})),
EmojiAnnotations(emoji='\U0001f3db', codepoints=(127963,), name='klassisk bygning', slug='klassisk_bygning', annotations=frozenset({'bygning', 'klassisk'})),
EmojiAnnotations(emoji='\U0001f3d7', codepoints=(127959,), name='bygning under konstruksjon', slug='bygning_under_konstruksjon', annotations=frozenset({'bygning', 'bygging'})),
EmojiAnnotations(emoji='\U0001f3d8', codepoints=(127960,), name='boliger', slug='boliger', annotations=frozenset({'hus', 'bygning'})),
EmojiAnnotations(emoji='\U0001f3d9', codepoints=(127961,), name='bylandskap', slug='bylandskap', annotations=frozenset({'bygning', 'bygging'})),
EmojiAnnotations(emoji='\U0001f3da', codepoints=(127962,), name='falleferdig hus', slug='falleferdig_hus', annotations=frozenset({'hus', 'forlatt', 'falleferdig', 'bygning'})),
EmojiAnnotations(emoji='🏠', codepoints=(127968,), name='hus', slug='hus', annotations=frozenset({'hjem', 'bygning'})),
EmojiAnnotations(emoji='🏡', codepoints=(127969,), name='hjem', slug='hjem', annotations=frozenset({'hus', 'hus med hage', 'bygning'})),
EmojiAnnotations(emoji='⛪', codepoints=(9962,), name='kirke', slug='kirke', annotations=frozenset({'bygning'})),
EmojiAnnotations(emoji='\U0001f54b', codepoints=(128331,), name='Kaba', slug='kaba', annotations=frozenset({'muslim', 'islam', 'religion'})),
EmojiAnnotations(emoji='\U0001f54c', codepoints=(128332,), name='moské', slug='moské', annotations=frozenset({'muslim', 'islam', 'religion'})),
EmojiAnnotations(emoji='\U0001f54d', codepoints=(128333,), name='synagoge', slug='synagoge', annotations=frozenset({'tempel', 'jøde', 'jødisk', 'religion'})),
EmojiAnnotations(emoji='⛩', codepoints=(9961,), name='Shinto-alter', slug='shinto_alter', annotations=frozenset({'alter', 'religion', 'shinto'})),
EmojiAnnotations(emoji='🏢', codepoints=(127970,), name='kontorbygg', slug='kontorbygg', annotations=frozenset({'bygning', 'kontor'})),
EmojiAnnotations(emoji='🏣', codepoints=(127971,), name='japansk postkontor', slug='japansk_postkontor', annotations=frozenset({'bygning', 'post'})),
EmojiAnnotations(emoji='🏤', codepoints=(127972,), name='postkontor', slug='postkontor', annotations=frozenset({'post', 'bygning', 'europeisk postkontor'})),
EmojiAnnotations(emoji='🏥', codepoints=(127973,), name='sykehus', slug='sykehus', annotations=frozenset({'bygning'})),
EmojiAnnotations(emoji='🏦', codepoints=(127974,), name='bank', slug='bank', annotations=frozenset({'bygning'})),
EmojiAnnotations(emoji='🏨', codepoints=(127976,), name='hotell', slug='hotell', annotations=frozenset({'bygning'})),
EmojiAnnotations(emoji='🏩', codepoints=(127977,), name='kjærlighetshotell', slug='kjærlighetshotell', annotations=frozenset({'bygning', 'hotell'})),
EmojiAnnotations(emoji='🏪', codepoints=(127978,), name='dagligvarehandel', slug='dagligvarehandel', annotations=frozenset({'dagligvare', 'bygning', 'butikk'})),
EmojiAnnotations(emoji='🏫', codepoints=(127979,), name='skole', slug='skole', annotations=frozenset({'skolebygning', 'bygning'})),
EmojiAnnotations(emoji='🏬', codepoints=(127980,), name='kjøpesenter', slug='kjøpesenter', annotations=frozenset({'storsenter', 'vareshus', 'butikk'})),
EmojiAnnotations(emoji='🏭', codepoints=(127981,), name='fabrikk', slug='fabrikk', annotations=frozenset({'bygning'})),
EmojiAnnotations(emoji='🏯', codepoints=(127983,), name='japansk slott', slug='japansk_slott', annotations=frozenset({'bygning', 'herskapshus'})),
EmojiAnnotations(emoji='🏰', codepoints=(127984,), name='slott', slug='slott', annotations=frozenset({'borg', 'bygning', 'europeisk slott', 'festning'})),
EmojiAnnotations(emoji='💒', codepoints=(128146,), name='kirkebryllup', slug='kirkebryllup', annotations=frozenset({'bryllup', 'kapell', 'romantikk'})),
EmojiAnnotations(emoji='🗼', codepoints=(128508,), name='Tokyo-tårnet', slug='tokyo_tårnet', annotations=frozenset({'tårn', 'Tokyo'})),
EmojiAnnotations(emoji='🗽', codepoints=(128509,), name='Frihetsgudinnen', slug='frihetsgudinnen', annotations=frozenset({'statue', 'frihet'})),
EmojiAnnotations(emoji='🗾', codepoints=(128510,), name='kart over Japan', slug='kart_over_japan', annotations=frozenset({'kart', 'Japan'})),
EmojiAnnotations(emoji='⛺', codepoints=(9978,), name='telt', slug='telt', annotations=frozenset({'camping'})),
EmojiAnnotations(emoji='🌁', codepoints=(127745,), name='tåkete', slug='tåkete', annotations=frozenset({'tåke', 'vær'})),
EmojiAnnotations(emoji='🌃', codepoints=(127747,), name='stjernenatt', slug='stjernenatt', annotations=frozenset({'stjerner', 'natt'})),
EmojiAnnotations(emoji='🌄', codepoints=(127748,), name='soloppgang over fjell', slug='soloppgang_over_fjell', annotations=frozenset({'fjell', 'fjellsol', 'morgensol'})),
EmojiAnnotations(emoji='🌅', codepoints=(127749,), name='soloppgang', slug='soloppgang', annotations=frozenset({'morgensol'})),
EmojiAnnotations(emoji='🌆', codepoints=(127750,), name='bylandskap i skumring', slug='bylandskap_i_skumring', annotations=frozenset({'solnedgang', 'by', 'bygning', 'byhorisont', 'skumring'})),
EmojiAnnotations(emoji='🌇', codepoints=(127751,), name='solnedgang', slug='solnedgang', annotations=frozenset({'solnedgang over byhorisont', 'bygning', 'skumring'})),
EmojiAnnotations(emoji='🌉', codepoints=(127753,), name='bro', slug='bro', annotations=frozenset({'bro i mørket', 'bro om natten'})),
EmojiAnnotations(emoji='♨', codepoints=(9832,), name='varme kilder', slug='varme_kilder', annotations=frozenset({'vame kilder', 'varm', 'kilder', 'rykende'})),
EmojiAnnotations(emoji='🌌', codepoints=(127756,), name='melkeveien', slug='melkeveien', annotations=frozenset({'galakse'})),
EmojiAnnotations(emoji='🎠', codepoints=(127904,), name='karusellhest', slug='karusellhest', annotations=frozenset({'karusell', 'fornøyelsespark'})),
EmojiAnnotations(emoji='🎡', codepoints=(127905,), name='pariserhjul', slug='pariserhjul', annotations=frozenset({'fornøyelsespark'})),
EmojiAnnotations(emoji='🎢', codepoints=(127906,), name='berg- og dalbane', slug='berg__og_dalbane', annotations=frozenset({'fornøyelsespark'})),
EmojiAnnotations(emoji='💈', codepoints=(128136,), name='barberer', slug='barberer', annotations=frozenset({'frisørsalong', 'frisør'})),
EmojiAnnotations(emoji='🎪', codepoints=(127914,), name='sirkus', slug='sirkus', annotations=frozenset({'telt', 'sikrustelt', 'underholdning'})),
EmojiAnnotations(emoji='🎭', codepoints=(127917,), name='teatermasker', slug='teatermasker', annotations=frozenset({'kunst', 'teater', 'scenekunst', 'underholdning'})),
EmojiAnnotations(emoji='\U0001f5bc', codepoints=(128444,), name='ramme med bilde', slug='ramme_med_bilde', annotations=frozenset({'kunst', 'ramme', 'maleri', 'bilde', 'museum'})),
EmojiAnnotations(emoji='🎨', codepoints=(127912,), name='palett', slug='palett', annotations=frozenset({'kunst', 'malerpalett', 'maling'})),
EmojiAnnotations(emoji='🎰', codepoints=(127920,), name='spillemaskin', slug='spillemaskin', annotations=frozenset({'spill'})),
EmojiAnnotations(emoji='🚂', codepoints=(128642,), name='damplokmotiv', slug='damplokmotiv', annotations=frozenset({'fremkomstmiddel', 'lokmotiv', 'tog'})),
EmojiAnnotations(emoji='🚃', codepoints=(128643,), name='togvogn', slug='togvogn', annotations=frozenset({'trikk', 'elektrisk tog', 'elektrisk', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚄', codepoints=(128644,), name='høyhastighetstog', slug='høyhastighetstog', annotations=frozenset({'lyntog', 'fremkomstmiddel', 'tog'})),
EmojiAnnotations(emoji='🚅', codepoints=(128645,), name='høyhastighetstog med avrundet tipp', slug='høyhastighetstog_med_avrundet_tipp', annotations=frozenset({'lyntog', 'fremkomstmiddel', 'tog'})),
EmojiAnnotations(emoji='🚆', codepoints=(128646,), name='tog', slug='tog', annotations=frozenset({'jernbane', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚇', codepoints=(128647,), name='undergrunn', slug='undergrunn', annotations=frozenset({'metro', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚈', codepoints=(128648,), name='lokaltog', slug='lokaltog', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚉', codepoints=(128649,), name='togstasjon', slug='togstasjon', annotations=frozenset({'stasjon', 'tog'})),
EmojiAnnotations(emoji='🚊', codepoints=(128650,), name='trikk', slug='trikk', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚝', codepoints=(128669,), name='monorail', slug='monorail', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚞', codepoints=(128670,), name='fjellbane', slug='fjellbane', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚋', codepoints=(128651,), name='trikkevogn', slug='trikkevogn', annotations=frozenset({'vogn', 'trikk', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚌', codepoints=(128652,), name='buss', slug='buss', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚍', codepoints=(128653,), name='motgående buss', slug='motgående_buss', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚎', codepoints=(128654,), name='trikkebuss', slug='trikkebuss', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚏', codepoints=(128655,), name='busstopp', slug='busstopp', annotations=frozenset({'buss', 'holdeplass'})),
EmojiAnnotations(emoji='🚐', codepoints=(128656,), name='minibuss', slug='minibuss', annotations=frozenset({'kjøretøy', 'van'})),
EmojiAnnotations(emoji='🚑', codepoints=(128657,), name='ambulanse', slug='ambulanse', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚒', codepoints=(128658,), name='brannbil', slug='brannbil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚓', codepoints=(128659,), name='politibil', slug='politibil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚔', codepoints=(128660,), name='motgående politibil', slug='motgående_politibil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚕', codepoints=(128661,), name='taxi', slug='taxi', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚖', codepoints=(128662,), name='motgående taxi', slug='motgående_taxi', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚗', codepoints=(128663,), name='rød bil', slug='rød_bil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚘', codepoints=(128664,), name='motgående bil', slug='motgående_bil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚙', codepoints=(128665,), name='fritidsbil', slug='fritidsbil', annotations=frozenset({'fritidskjøretøy', 'kjøretøy'})),
EmojiAnnotations(emoji='🚚', codepoints=(128666,), name='varebil', slug='varebil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚛', codepoints=(128667,), name='lastebil', slug='lastebil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚜', codepoints=(128668,), name='traktor', slug='traktor', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚲', codepoints=(128690,), name='sykkel', slug='sykkel', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='⛽', codepoints=(9981,), name='bensinstasjon', slug='bensinstasjon', annotations=frozenset({'bensinpumpe', 'bensin', 'diesel', 'stasjon'})),
EmojiAnnotations(emoji='\U0001f6e3', codepoints=(128739,), name='motorvei', slug='motorvei', annotations=frozenset({'vei'})),
EmojiAnnotations(emoji='\U0001f6e4', codepoints=(128740,), name='jernbanespor', slug='jernbanespor', annotations=frozenset({'jernbane', 'tog'})),
EmojiAnnotations(emoji='🚨', codepoints=(128680,), name='politilys', slug='politilys', annotations=frozenset({'bil', 'kjøretøy', 'signal', 'roterende', 'politi', 'lys'})),
EmojiAnnotations(emoji='🚥', codepoints=(128677,), name='horisontale trafikklys', slug='horisontale_trafikklys', annotations=frozenset({'horisontale lys', 'trafikk'})),
EmojiAnnotations(emoji='🚦', codepoints=(128678,), name='vertikale trafiklys', slug='vertikale_trafiklys', annotations=frozenset({'trafikk', 'vertikale lys'})),
EmojiAnnotations(emoji='🚧', codepoints=(128679,), name='under kontruksjon', slug='under_kontruksjon', annotations=frozenset({'skilt', 'byggeskilt', 'byggeplass'})),
EmojiAnnotations(emoji='⚓', codepoints=(9875,), name='anker', slug='anker', annotations=frozenset({'skip', 'verktøy'})),
EmojiAnnotations(emoji='⛵', codepoints=(9973,), name='seilbåt', slug='seilbåt', annotations=frozenset({'båt'})),
EmojiAnnotations(emoji='🚣', codepoints=(128675,), name='robåt', slug='robåt', annotations=frozenset({'båt', 'fartøy'})),
EmojiAnnotations(emoji='🚤', codepoints=(128676,), name='speedbåt', slug='speedbåt', annotations=frozenset({'båt', 'fartøy'})),
EmojiAnnotations(emoji='\U0001f6f3', codepoints=(128755,), name='passasjerskip', slug='passasjerskip', annotations=frozenset({'skip', 'fremkomstmiddel', 'passasjer'})),
EmojiAnnotations(emoji='⛴', codepoints=(9972,), name='ferge', slug='ferge', annotations=frozenset({'båt'})),
EmojiAnnotations(emoji='\U0001f6e5', codepoints=(128741,), name='motorbåt', slug='motorbåt', annotations=frozenset({'båt', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚢', codepoints=(128674,), name='skip', slug='skip', annotations=frozenset({'fartøy'})),
EmojiAnnotations(emoji='✈', codepoints=(9992,), name='fly', slug='fly', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='\U0001f6e9', codepoints=(128745,), name='lite fly', slug='lite_fly', annotations=frozenset({'fly', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='\U0001f6eb', codepoints=(128747,), name='fly som tar av', slug='fly_som_tar_av', annotations=frozenset({'avgang', 'innsjekk', 'avganger', 'fly', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='\U0001f6ec', codepoints=(128748,), name='fly som lander', slug='fly_som_lander', annotations=frozenset({'ankomme', 'fremkomstmiddel', 'fly', 'ankomster', 'lande'})),
EmojiAnnotations(emoji='💺', codepoints=(128186,), name='sete', slug='sete', annotations=frozenset({'stol'})),
EmojiAnnotations(emoji='🚁', codepoints=(128641,), name='helikopter', slug='helikopter', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚟', codepoints=(128671,), name='svevebane', slug='svevebane', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚠', codepoints=(128672,), name='kabelbane', slug='kabelbane', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚡', codepoints=(128673,), name='taubane', slug='taubane', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚀', codepoints=(128640,), name='rakett', slug='rakett', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='\U0001f6f0', codepoints=(128752,), name='satelitt', slug='satelitt', annotations=frozenset({'fremkomstmiddel', 'verdensrommet'})),
EmojiAnnotations(emoji='\U0001f6ce', codepoints=(128718,), name='hotellklokke', slug='hotellklokke', annotations=frozenset({'klokke', 'pikkolo', 'hotell'})),
EmojiAnnotations(emoji='\U0001f6cc', codepoints=(128716,), name='person i seng', slug='person_i_seng', annotations=frozenset({'hotell', 'sove'})),
EmojiAnnotations(emoji='\U0001f6cf', codepoints=(128719,), name='seng', slug='seng', annotations=frozenset({'hotell', 'sove'})),
EmojiAnnotations(emoji='\U0001f6cb', codepoints=(128715,), name='sofa og lampe', slug='sofa_og_lampe', annotations=frozenset({'sofa', 'lampe', 'hotell'})),
EmojiAnnotations(emoji='🚽', codepoints=(128701,), name='toalett', slug='toalett', annotations=frozenset({'do'})),
EmojiAnnotations(emoji='🚿', codepoints=(128703,), name='dusj', slug='dusj', annotations=frozenset({'vann'})),
EmojiAnnotations(emoji='🛀', codepoints=(128704,), name='person i badekar', slug='person_i_badekar', annotations=frozenset({'bad', 'badekar'})),
EmojiAnnotations(emoji='🛁', codepoints=(128705,), name='tomt badekar', slug='tomt_badekar', annotations=frozenset({'bad', 'badekar'})),
EmojiAnnotations(emoji='⌛', codepoints=(8987,), name='halvferdig timeglass', slug='halvferdig_timeglass', annotations=frozenset({'sand'})),
EmojiAnnotations(emoji='⏳', codepoints=(9203,), name='timeglass som har startet', slug='timeglass_som_har_startet', annotations=frozenset({'sand', 'tidtaker'})),
EmojiAnnotations(emoji='⌚', codepoints=(8986,), name='klokke', slug='klokke', annotations=frozenset({'ur'})),
EmojiAnnotations(emoji='⏰', codepoints=(9200,), name='alarmklokke', slug='alarmklokke', annotations=frozenset({'klokke', 'alarm'})),
EmojiAnnotations(emoji='⏱', codepoints=(9201,), name='stoppeklokke', slug='stoppeklokke', annotations=frozenset({'klokke'})),
EmojiAnnotations(emoji='⏲', codepoints=(9202,), name='tidtaker', slug='tidtaker', annotations=frozenset({'klokke'})),
EmojiAnnotations(emoji='\U0001f570', codepoints=(128368,), name='peisur', slug='peisur', annotations=frozenset({'klokke'})),
EmojiAnnotations(emoji='🕛', codepoints=(128347,), name='klokken tolv', slug='klokken_tolv', annotations=frozenset({'24:00', 'tolv', 'klokka', '12:00'})),
EmojiAnnotations(emoji='🕧', codepoints=(128359,), name='klokken halv ett', slug='klokken_halv_ett', annotations=frozenset({'12:30', '24:30', 'klokka'})),
EmojiAnnotations(emoji='🕐', codepoints=(128336,), name='klokken ett', slug='klokken_ett', annotations=frozenset({'1:00', 'klokka', 'ett', '13:00'})),
EmojiAnnotations(emoji='🕜', codepoints=(128348,), name='klokken halv to', slug='klokken_halv_to', annotations=frozenset({'13:30', '1:30', 'klokka'})),
EmojiAnnotations(emoji='🕑', codepoints=(128337,), name='klokken to', slug='klokken_to', annotations=frozenset({'2:00', '14:00', 'klokka', 'to'})),
EmojiAnnotations(emoji='🕝', codepoints=(128349,), name='klokken halv tre', slug='klokken_halv_tre', annotations=frozenset({'14:30', '2:30', 'klokka'})),
EmojiAnnotations(emoji='🕒', codepoints=(128338,), name='klokken tre', slug='klokken_tre', annotations=frozenset({'15:00', '3:00'})),
EmojiAnnotations(emoji='🕞', codepoints=(128350,), name='klokken halv fire', slug='klokken_halv_fire', annotations=frozenset({'3:30', 'klokka', '15:30'})),
EmojiAnnotations(emoji='🕓', codepoints=(128339,), name='klokken fire', slug='klokken_fire', annotations=frozenset({'fire', '4:00', '16:00', 'klokka'})),
EmojiAnnotations(emoji='🕟', codepoints=(128351,), name='klokken halv fem', slug='klokken_halv_fem', annotations=frozenset({'4:30', '16:30', 'klokka'})),
EmojiAnnotations(emoji='🕔', codepoints=(128340,), name='klokken fem', slug='klokken_fem', annotations=frozenset({'17:00', 'fem', 'klokka', '5:00'})),
EmojiAnnotations(emoji='🕠', codepoints=(128352,), name='klokken halv seks', slug='klokken_halv_seks', annotations=frozenset({'5:30', 'klokka', '17:30'})),
EmojiAnnotations(emoji='🕕', codepoints=(128341,), name='klokken seks', slug='klokken_seks', annotations=frozenset({'6:00', 'seks', 'klokka', '18:00'})),
EmojiAnnotations(emoji='🕡', codepoints=(128353,), name='klokken halv sju', slug='klokken_halv_sju', annotations=frozenset({'6:30', 'klokka', '18:30'})),
EmojiAnnotations(emoji='🕖', codepoints=(128342,), name='klokken sju', slug='klokken_sju', annotations=frozenset({'sju', '19:00', 'klokka', '7:00'})),
EmojiAnnotations(emoji='🕢', codepoints=(128354,), name='klokken halv åtte', slug='klokken_halv_åtte', annotations=frozenset({'7:30', 'klokka', '19:30'})),
EmojiAnnotations(emoji='🕗', codepoints=(128343,), name='klokken åtte', slug='klokken_åtte', annotations=frozenset({'åtte', 'klokka', '8:00', '20:00'})),
EmojiAnnotations(emoji='🕣', codepoints=(128355,), name='klokken halv ni', slug='klokken_halv_ni', annotations=frozenset({'klokka', '8:30', '20:30'})),
EmojiAnnotations(emoji='🕘', codepoints=(128344,), name='klokken ni', slug='klokken_ni', annotations=frozenset({'21:00', 'klokka', 'ni', '9:00'})),
EmojiAnnotations(emoji='🕤', codepoints=(128356,), name='klokken halv ti', slug='klokken_halv_ti', annotations=frozenset({'21:30', '09:30', 'klokka'})),
EmojiAnnotations(emoji='🕙', codepoints=(128345,), name='klokken ti', slug='klokken_ti', annotations=frozenset({'22:00', '10:00', 'klokka', 'ti'})),
EmojiAnnotations(emoji='🕥', codepoints=(128357,), name='klokken halv elleve', slug='klokken_halv_elleve', annotations=frozenset({'22:30', 'klokka', '10:30'})),
EmojiAnnotations(emoji='🕚', codepoints=(128346,), name='klokken elleve', slug='klokken_elleve', annotations=frozenset({'elleve', '23:00', 'klokka', '11:00'})),
EmojiAnnotations(emoji='🕦', codepoints=(128358,), name='klokken halv tolv', slug='klokken_halv_tolv', annotations=frozenset({'23:30', 'klokka', '11:30'})),
EmojiAnnotations(emoji='🌑', codepoints=(127761,), name='mørk måne', slug='mørk_måne', annotations=frozenset({'mørk', 'måne'})),
EmojiAnnotations(emoji='🌒', codepoints=(127762,), name='voksende nymåne', slug='voksende_nymåne', annotations=frozenset({'månefase', 'tiltakende måne'})),
EmojiAnnotations(emoji='🌓', codepoints=(127763,), name='voksende halvmåne', slug='voksende_halvmåne', annotations=frozenset({'første kvarter', 'månefase'})),
EmojiAnnotations(emoji='🌔', codepoints=(127764,), name='voksende fullmåne', slug='voksende_fullmåne', annotations=frozenset({'månefase', 'tiltakende måne'})),
EmojiAnnotations(emoji='🌕', codepoints=(127765,), name='fullmåne', slug='fullmåne', annotations=frozenset({'klar måne', 'måne'})),
EmojiAnnotations(emoji='🌖', codepoints=(127766,), name='avtakende fullmåne', slug='avtakende_fullmåne', annotations=frozenset({'månefase', 'avtakende måne'})),
EmojiAnnotations(emoji='🌗', codepoints=(127767,), name='avtakende halvmåne', slug='avtakende_halvmåne', annotations=frozenset({'siste kvarter', 'månefase'})),
EmojiAnnotations(emoji='🌘', codepoints=(127768,), name='avtakende nemåne', slug='avtakende_nemåne', annotations=frozenset({'månefase', 'avtakende måne'})),
EmojiAnnotations(emoji='🌙', codepoints=(127769,), name='nymåne', slug='nymåne', annotations=frozenset({'måne'})),
EmojiAnnotations(emoji='🌚', codepoints=(127770,), name='nymåneansikt', slug='nymåneansikt', annotations=frozenset({'nymåne', 'nymåne med ansikt', 'måne', 'ansikt'})),
EmojiAnnotations(emoji='🌛', codepoints=(127771,), name='nymåne med ansikt', slug='nymåne_med_ansikt', annotations=frozenset({'kvartmåne med ansikt', 'måneansikt'})),
EmojiAnnotations(emoji='🌜', codepoints=(127772,), name='nemåne med ansikt', slug='nemåne_med_ansikt', annotations=frozenset({'kvartmåne med ansikt', 'måneansikt'})),
EmojiAnnotations(emoji='\U0001f321', codepoints=(127777,), name='termometer', slug='termometer', annotations=frozenset({'vær'})),
EmojiAnnotations(emoji='☀', codepoints=(9728,), name='sol', slug='sol', annotations=frozenset({'solfylt', 'stråler', 'vær', 'verdensrommet', 'lyst'})),
EmojiAnnotations(emoji='🌝', codepoints=(127773,), name='fullmåneansikt', slug='fullmåneansikt', annotations=frozenset({'måne med ansikt', 'fullmåne', 'fullmåne med ansikt', 'måne', 'ansikt'})),
EmojiAnnotations(emoji='🌞', codepoints=(127774,), name='solansikt', slug='solansikt', annotations=frozenset({'sol', 'ansikt'})),
EmojiAnnotations(emoji='⭐', codepoints=(11088,), name='hvit stjerne', slug='hvit_stjerne', annotations=frozenset({'stjerne'})),
EmojiAnnotations(emoji='🌟', codepoints=(127775,), name='lysende stjerne', slug='lysende_stjerne', annotations=frozenset({'skinnende', 'glitrende', 'stjerne'})),
EmojiAnnotations(emoji='🌠', codepoints=(127776,), name='stjerneskudd', slug='stjerneskudd', annotations=frozenset({'fallende stjerne', 'stjerne'})),
EmojiAnnotations(emoji='☁', codepoints=(9729,), name='sky', slug='sky', annotations=frozenset({'vær'})),
EmojiAnnotations(emoji='⛅', codepoints=(9925,), name='sol bak sky', slug='sol_bak_sky', annotations=frozenset({'sky', 'sol', 'sol og sky'})),
EmojiAnnotations(emoji='⛈', codepoints=(9928,), name='sky med lyn og regn', slug='sky_med_lyn_og_regn', annotations=frozenset({'sky', 'torden', 'regn', 'vær'})),
EmojiAnnotations(emoji='\U0001f324', codepoints=(127780,), name='sol bak liten sky', slug='sol_bak_liten_sky', annotations=frozenset({'sky', 'sol', 'vær'})),
EmojiAnnotations(emoji='\U0001f325', codepoints=(127781,), name='sol bak stor sky', slug='sol_bak_stor_sky', annotations=frozenset({'sky', 'sol', 'vær'})),
EmojiAnnotations(emoji='\U0001f326', codepoints=(127782,), name='sol bak sky med regn', slug='sol_bak_sky_med_regn', annotations=frozenset({'sky', 'sol', 'regn', 'vær'})),
EmojiAnnotations(emoji='\U0001f327', codepoints=(127783,), name='sky med regn', slug='sky_med_regn', annotations=frozenset({'sky', 'regn', 'vær'})),
EmojiAnnotations(emoji='\U0001f328', codepoints=(127784,), name='sky med snø', slug='sky_med_snø', annotations=frozenset({'sky', 'kaldt', 'snø', 'vær'})),
EmojiAnnotations(emoji='\U0001f329', codepoints=(127785,), name='sky med lyn', slug='sky_med_lyn', annotations=frozenset({'sky', 'lyn', 'vær'})),
EmojiAnnotations(emoji='\U0001f32a', codepoints=(127786,), name='tornado', slug='tornado', annotations=frozenset({'sky', 'vær'})),
EmojiAnnotations(emoji='\U0001f32b', codepoints=(127787,), name='tåke', slug='tåke', annotations=frozenset({'sky', 'vær'})),
EmojiAnnotations(emoji='\U0001f32c', codepoints=(127788,), name='vindansikt', slug='vindansikt', annotations=frozenset({'sky', 'vind', 'fjes', 'blåse', 'vær'})),
EmojiAnnotations(emoji='🌀', codepoints=(127744,), name='syklon', slug='syklon', annotations=frozenset({'tornado', 'svimmel'})),
EmojiAnnotations(emoji='🌈', codepoints=(127752,), name='regnbue', slug='regnbue', annotations=frozenset({'regn', 'vær'})),
EmojiAnnotations(emoji='🌂', codepoints=(127746,), name='lukket paraply', slug='lukket_paraply', annotations=frozenset({'regnvær', 'regn'})),
EmojiAnnotations(emoji='☂', codepoints=(9730,), name='paraply', slug='paraply', annotations=frozenset({'regn', 'klær', 'vær'})),
EmojiAnnotations(emoji='☔', codepoints=(9748,), name='paraply med regndråper', slug='paraply_med_regndråper', annotations=frozenset({'dråpe', 'paraply', 'regn', 'klær', 'vær'})),
EmojiAnnotations(emoji='⛱', codepoints=(9969,), name='parasoll på bakken', slug='parasoll_på_bakken', annotations=frozenset({'paraply', 'sol', 'regn', 'vær'})),
EmojiAnnotations(emoji='⚡', codepoints=(9889,), name='høyspenning', slug='høyspenning', annotations=frozenset({'fare', 'høy', 'høyspent', 'spenning'})),
EmojiAnnotations(emoji='❄', codepoints=(10052,), name='snøfnugg', slug='snøfnugg', annotations=frozenset({'kaldt', 'snø', 'vær'})),
EmojiAnnotations(emoji='☃', codepoints=(9731,), name='snømann', slug='snømann', annotations=frozenset({'kaldt', 'snø', 'vær'})),
EmojiAnnotations(emoji='⛄', codepoints=(9924,), name='snømann uten snø', slug='snømann_uten_snø', annotations=frozenset({'snø'})),
EmojiAnnotations(emoji='☄', codepoints=(9732,), name='komet', slug='komet', annotations=frozenset({'verdensrommet'})),
EmojiAnnotations(emoji='🔥', codepoints=(128293,), name='ild', slug='ild', annotations=frozenset({'flamme'})),
EmojiAnnotations(emoji='💧', codepoints=(128167,), name='dråpe', slug='dråpe', annotations=frozenset({'tegneserie', 'drypp', 'vann'})),
EmojiAnnotations(emoji='🌊', codepoints=(127754,), name='bølge', slug='bølge', annotations=frozenset({'bølgesprut', 'hav'})),
EmojiAnnotations(emoji='🎃', codepoints=(127875,), name='gresskar', slug='gresskar', annotations=frozenset({'gresskarlykt', 'halloweenlykt', 'Halloween'})),
EmojiAnnotations(emoji='🎄', codepoints=(127876,), name='juletre', slug='juletre', annotations=frozenset({'høytid', 'tre', 'julefeiring', 'jul'})),
EmojiAnnotations(emoji='🎆', codepoints=(127878,), name='fyrverkeri', slug='fyrverkeri', annotations=frozenset({'feiring', 'nyttår'})),
EmojiAnnotations(emoji='🎇', codepoints=(127879,), name='stjerneskudd på pinne', slug='stjerneskudd_på_pinne', annotations=frozenset({'feiring', 'fyrverkeri', 'nyttår'})),
EmojiAnnotations(emoji='✨', codepoints=(10024,), name='stjerner', slug='stjerner', annotations=frozenset({'skinnende', 'glitrende'})),
EmojiAnnotations(emoji='🎈', codepoints=(127880,), name='ballong', slug='ballong', annotations=frozenset({'feiring'})),
EmojiAnnotations(emoji='🎉', codepoints=(127881,), name='smell-bon-bon', slug='smell_bon_bon', annotations=frozenset({'serpentin', 'fest', 'feiring'})),
EmojiAnnotations(emoji='🎊', codepoints=(127882,), name='konfetti', slug='konfetti', annotations=frozenset({'konfettikule', 'fest', 'feiring'})),
EmojiAnnotations(emoji='🎋', codepoints=(127883,), name='tre med papirremser', slug='tre_med_papirremser', annotations=frozenset({'tre', 'papirremser', 'stjernefestival', 'feiring'})),
EmojiAnnotations(emoji='🎌', codepoints=(127884,), name='kryssede flagg', slug='kryssede_flagg', annotations=frozenset({'flagg', 'krysset', 'kryssede japanske flagg', 'feiring'})),
EmojiAnnotations(emoji='🎍', codepoints=(127885,), name='furudekorasjon', slug='furudekorasjon', annotations=frozenset({'feiring', 'japansk høytid', 'nyttår'})),
EmojiAnnotations(emoji='🎎', codepoints=(127886,), name='dukker', slug='dukker', annotations=frozenset({'festival', 'japansk høytid', 'japanske dukker'})),
EmojiAnnotations(emoji='🎏', codepoints=(127887,), name='fiskedekorasjon', slug='fiskedekorasjon', annotations=frozenset({'japansk høytid', 'papirfisk', 'feiring'})),
EmojiAnnotations(emoji='🎐', codepoints=(127888,), name='vindspill', slug='vindspill', annotations=frozenset({'drømmefanger'})),
EmojiAnnotations(emoji='🎑', codepoints=(127889,), name='månebeskuelse', slug='månebeskuelse', annotations=frozenset({'høytid', 'feiring', 'månefeiring', 'måne', 'måneseremoni'})),
EmojiAnnotations(emoji='🎀', codepoints=(127872,), name='sløyfe', slug='sløyfe', annotations=frozenset({'gavebånd', 'bånd'})),
EmojiAnnotations(emoji='🎁', codepoints=(127873,), name='gave', slug='gave', annotations=frozenset({'feiring', 'innpakket gave', 'bursdag', 'julegave'})),
EmojiAnnotations(emoji='\U0001f396', codepoints=(127894,), name='medalje', slug='medalje', annotations=frozenset({'militæret', 'feiring'})),
EmojiAnnotations(emoji='\U0001f397', codepoints=(127895,), name='minnesløyfe', slug='minnesløyfe', annotations=frozenset({'bånd', 'påminnelse', 'feiring'})),
EmojiAnnotations(emoji='\U0001f39e', codepoints=(127902,), name='filmrammer', slug='filmrammer', annotations=frozenset({'rammer', 'kino', 'film'})),
EmojiAnnotations(emoji='\U0001f39f', codepoints=(127903,), name='inngangsbilletter', slug='inngangsbilletter', annotations=frozenset({'billett', 'adgang'})),
EmojiAnnotations(emoji='🎫', codepoints=(127915,), name='billett', slug='billett', annotations=frozenset({'underholdning'})),
EmojiAnnotations(emoji='⚽', codepoints=(9917,), name='fotball', slug='fotball', annotations=frozenset({'ball'})),
EmojiAnnotations(emoji='⚾', codepoints=(9918,), name='baseball', slug='baseball', annotations=frozenset({'ball', 'stitched'})),
EmojiAnnotations(emoji='🏀', codepoints=(127936,), name='basketball', slug='basketball', annotations=frozenset({'sport', 'ball'})),
EmojiAnnotations(emoji='🏈', codepoints=(127944,), name='amerikansk fotball', slug='amerikansk_fotball', annotations=frozenset({'sport', 'ball'})),
EmojiAnnotations(emoji='🏉', codepoints=(127945,), name='rugby', slug='rugby', annotations=frozenset({'sport', 'rugbyball'})),
EmojiAnnotations(emoji='🎾', codepoints=(127934,), name='tennis', slug='tennis', annotations=frozenset({'sport', 'ball', 'tenninsball'})),
EmojiAnnotations(emoji='🎱', codepoints=(127921,), name='biljard', slug='biljard', annotations=frozenset({'åtterkule', 'spill'})),
EmojiAnnotations(emoji='🎳', codepoints=(127923,), name='bowling', slug='bowling', annotations=frozenset({'kule', 'bowlingkule', 'spill'})),
EmojiAnnotations(emoji='⛳', codepoints=(9971,), name='golfflagg', slug='golfflagg', annotations=frozenset({'golf', 'golfhull'})),
EmojiAnnotations(emoji='\U0001f3cc', codepoints=(127948,), name='golfspiller', slug='golfspiller', annotations=frozenset({'golf', 'ball'})),
EmojiAnnotations(emoji='⛸', codepoints=(9976,), name='skøyter', slug='skøyter', annotations=frozenset({'is', 'skøyte'})),
EmojiAnnotations(emoji='🎣', codepoints=(127907,), name='fiskestang', slug='fiskestang', annotations=frozenset({'fisking', 'fisk', 'fritid', 'natur'})),
EmojiAnnotations(emoji='🎽', codepoints=(127933,), name='løpetopp', slug='løpetopp', annotations=frozenset({'sport', 't-skjorte', 'treningstopp'})),
EmojiAnnotations(emoji='🎿', codepoints=(127935,), name='ski', slug='ski', annotations=frozenset({'ski og staver', 'vintersport'})),
EmojiAnnotations(emoji='⛷', codepoints=(9975,), name='skiløper', slug='skiløper', annotations=frozenset({'ski', 'snø'})),
EmojiAnnotations(emoji='🏂', codepoints=(127938,), name='snowboarder', slug='snowboarder', annotations=frozenset({'vintersport', 'snøbrettkjøring', 'snowboarding', 'snowboard', 'snøbrett'})),
EmojiAnnotations(emoji='🏄', codepoints=(127940,), name='surfing', slug='surfing', annotations=frozenset({'sport', 'surfer'})),
EmojiAnnotations(emoji='🏇', codepoints=(127943,), name='veddeløp', slug='veddeløp', annotations=frozenset({'sport', 'jockey', 'hest', 'veddeløpshest', 'løp'})),
EmojiAnnotations(emoji='🏊', codepoints=(127946,), name='svømming', slug='svømming', annotations=frozenset({'sport', 'svømme', 'svømmer'})),
EmojiAnnotations(emoji='⛹', codepoints=(9977,), name='person med ball', slug='person_med_ball', annotations=frozenset({'ball'})),
EmojiAnnotations(emoji='\U0001f3cb', codepoints=(127947,), name='vektløfter', slug='vektløfter', annotations=frozenset({'vekt', 'løfter'})),
EmojiAnnotations(emoji='🚴', codepoints=(128692,), name='syklist', slug='syklist', annotations=frozenset({'sykkel'})),
EmojiAnnotations(emoji='🚵', codepoints=(128693,), name='fjellsyklist', slug='fjellsyklist', annotations=frozenset({'mountainbike'})),
EmojiAnnotations(emoji='\U0001f3ce', codepoints=(127950,), name='racerbil', slug='racerbil', annotations=frozenset({'kappkjøring', 'bil'})),
EmojiAnnotations(emoji='\U0001f3cd', codepoints=(127949,), name='motorsykkel', slug='motorsykkel', annotations=frozenset({'kappkjøring'})),
EmojiAnnotations(emoji='\U0001f3c5', codepoints=(127941,), name='idrettsmedalje', slug='idrettsmedalje', annotations=frozenset({'medalje'})),
EmojiAnnotations(emoji='🏆', codepoints=(127942,), name='trofé', slug='trofé', annotations=frozenset({'premie'})),
EmojiAnnotations(emoji='\U0001f3cf', codepoints=(127951,), name='cricket', slug='cricket', annotations=frozenset({'balltre', 'kamp', 'ball'})),
EmojiAnnotations(emoji='\U0001f3d0', codepoints=(127952,), name='volleyball', slug='volleyball', annotations=frozenset({'kamp', 'ball'})),
EmojiAnnotations(emoji='\U0001f3d1', codepoints=(127953,), name='landhockey', slug='landhockey', annotations=frozenset({'kølle', 'hockey', 'kamp', 'ball', 'land'})),
EmojiAnnotations(emoji='\U0001f3d2', codepoints=(127954,), name='ishockeykølle og puck', slug='ishockeykølle_og_puck', annotations=frozenset({'is', 'hockey', 'kølle', 'kamp', 'puck'})),
EmojiAnnotations(emoji='\U0001f3d3', codepoints=(127955,), name='bordtennis', slug='bordtennis', annotations=frozenset({'racket', 'balltre', 'tennis', 'kamp', 'bord', 'ball'})),
EmojiAnnotations(emoji='\U0001f3f8', codepoints=(127992,), name='badminton', slug='badminton', annotations=frozenset({'fjærball', 'racket', 'kamp', 'birdie'})),
EmojiAnnotations(emoji='🎯', codepoints=(127919,), name='blink', slug='blink', annotations=frozenset({'treff', 'midt i blinken', 'mål', 'pilspill', 'dart'})),
EmojiAnnotations(emoji='🎮', codepoints=(127918,), name='videospill', slug='videospill', annotations=frozenset({'spillkonsoll', 'konsoll', 'spill'})),
EmojiAnnotations(emoji='\U0001f579', codepoints=(128377,), name='styrespake', slug='styrespake', annotations=frozenset({'videospill', 'spill'})),
EmojiAnnotations(emoji='🎲', codepoints=(127922,), name='terning', slug='terning', annotations=frozenset({'terningspill', 'terningkast', 'spill'})),
EmojiAnnotations(emoji='♠', codepoints=(9824,), name='spar', slug='spar', annotations=frozenset({'farge', 'kort', 'spill'})),
EmojiAnnotations(emoji='♥', codepoints=(9829,), name='hjerter', slug='hjerter', annotations=frozenset({'farge', 'kort', 'spill'})),
EmojiAnnotations(emoji='♦', codepoints=(9830,), name='ruter', slug='ruter', annotations=frozenset({'spill', 'farge', 'kort', 'diamant'})),
EmojiAnnotations(emoji='♣', codepoints=(9827,), name='kløver', slug='kløver', annotations=frozenset({'farge', 'kort', 'spill'})),
EmojiAnnotations(emoji='🃏', codepoints=(127183,), name='joker', slug='joker', annotations=frozenset({'jokerkort', 'spillkort', 'kort'})),
EmojiAnnotations(emoji='🀄', codepoints=(126980,), name='mahjongbrikke', slug='mahjongbrikke', annotations=frozenset({'dominobrikke'})),
EmojiAnnotations(emoji='🎴', codepoints=(127924,), name='spillkort', slug='spillkort', annotations=frozenset({'kort', 'kortspill'})),
EmojiAnnotations(emoji='🔇', codepoints=(128263,), name='stille', slug='stille', annotations=frozenset({'ingen prating', 'høyttaler med strek over', 'vennligst være stille'})),
EmojiAnnotations(emoji='🔈', codepoints=(128264,), name='høyttaler', slug='høyttaler', annotations=frozenset({'ropert'})),
EmojiAnnotations(emoji='🔉', codepoints=(128265,), name='høyttaler med lavt volum', slug='høyttaler_med_lavt_volum', annotations=frozenset({'lavt volum'})),
EmojiAnnotations(emoji='🔊', codepoints=(128266,), name='høyttaler med høyt volum', slug='høyttaler_med_høyt_volum', annotations=frozenset({'høyt volum'})),
EmojiAnnotations(emoji='📢', codepoints=(128226,), name='ropert', slug='ropert', annotations=frozenset({'megafon', 'kommunikasjon'})),
EmojiAnnotations(emoji='📣', codepoints=(128227,), name='megafon', slug='megafon', annotations=frozenset({'heiing', 'kommunikasjon'})),
EmojiAnnotations(emoji='📯', codepoints=(128239,), name='posthorn', slug='posthorn', annotations=frozenset({'post'})),
EmojiAnnotations(emoji='🔔', codepoints=(128276,), name='bjelle', slug='bjelle', annotations=frozenset({'klokke', 'ringing'})),
EmojiAnnotations(emoji='🔕', codepoints=(128277,), name='ingen lyd', slug='ingen_lyd', annotations=frozenset({'bjelle med strek over', 'klokke', 'ingen bjelle'})),
EmojiAnnotations(emoji='🎼', codepoints=(127932,), name='g-nøkkel', slug='g_nøkkel', annotations=frozenset({'notetegn', 'musikk'})),
EmojiAnnotations(emoji='🎵', codepoints=(127925,), name='note', slug='note', annotations=frozenset({'musikk', 'musikknote'})),
EmojiAnnotations(emoji='🎶', codepoints=(127926,), name='noter', slug='noter', annotations=frozenset({'musikknoter', 'musikk'})),
EmojiAnnotations(emoji='\U0001f399', codepoints=(127897,), name='studiomikrofon', slug='studiomikrofon', annotations=frozenset({'musikk', 'mikrofon', 'studio'})),
EmojiAnnotations(emoji='\U0001f39a', codepoints=(127898,), name='nivåglidebryter', slug='nivåglidebryter', annotations=frozenset({'glidebryter', 'musikk', 'nivå'})),
EmojiAnnotations(emoji='\U0001f39b', codepoints=(127899,), name='kontrollbrytere', slug='kontrollbrytere', annotations=frozenset({'musikk', 'knott', 'kontroll'})),
EmojiAnnotations(emoji='🎤', codepoints=(127908,), name='mikrofon', slug='mikrofon', annotations=frozenset({'underholdning', 'karaoke', 'musikk', 'mik'})),
EmojiAnnotations(emoji='🎧', codepoints=(127911,), name='øreplugg', slug='øreplugg', annotations=frozenset({'underholdning', 'hodetelefoner', 'headset'})),
EmojiAnnotations(emoji='🎷', codepoints=(127927,), name='sax', slug='sax', annotations=frozenset({'musikk', 'instrument', 'saxofon', 'musikkinstrument'})),
EmojiAnnotations(emoji='🎸', codepoints=(127928,), name='gitar', slug='gitar', annotations=frozenset({'musikk', 'instrument', 'musikkinstrument'})),
EmojiAnnotations(emoji='🎹', codepoints=(127929,), name='piano', slug='piano', annotations=frozenset({'musikk', 'instrument', 'musikkinstrument', 'keyboard'})),
EmojiAnnotations(emoji='🎺', codepoints=(127930,), name='trompet', slug='trompet', annotations=frozenset({'musikk', 'instrument', 'musikkinstrument'})),
EmojiAnnotations(emoji='🎻', codepoints=(127931,), name='fiolin', slug='fiolin', annotations=frozenset({'musikk', 'instrument', 'musikkinstrument', 'fele'})),
EmojiAnnotations(emoji='📻', codepoints=(128251,), name='radio', slug='radio', annotations=frozenset({'radiosignaler', 'teknologi'})),
EmojiAnnotations(emoji='📱', codepoints=(128241,), name='mobil', slug='mobil', annotations=frozenset({'mobiltelefon', 'telefon'})),
EmojiAnnotations(emoji='📲', codepoints=(128242,), name='mobiltelefon med pil', slug='mobiltelefon_med_pil', annotations=frozenset({'ring', 'mobiltelefon', 'telefon'})),
EmojiAnnotations(emoji='📞', codepoints=(128222,), name='telefonrør', slug='telefonrør', annotations=frozenset({'kommunikasjon'})),
EmojiAnnotations(emoji='📟', codepoints=(128223,), name='personsøker', slug='personsøker', annotations=frozenset({'kommunikasjon'})),
EmojiAnnotations(emoji='📠', codepoints=(128224,), name='faks', slug='faks', annotations=frozenset({'faksmaskin', 'kommunikasjon'})),
EmojiAnnotations(emoji='🔋', codepoints=(128267,), name='batteri', slug='batteri', annotations=frozenset({'strøm', 'energilagring'})),
EmojiAnnotations(emoji='🔌', codepoints=(128268,), name='støpsel', slug='støpsel', annotations=frozenset({'elektrisk'})),
EmojiAnnotations(emoji='💻', codepoints=(128187,), name='PC', slug='pc', annotations=frozenset({'personlig', 'datamaskin', 'maskin'})),
EmojiAnnotations(emoji='\U0001f5a5', codepoints=(128421,), name='stasjonær datamaskin', slug='stasjonær_datamaskin', annotations=frozenset({'datamaskin', 'stasjonær'})),
EmojiAnnotations(emoji='\U0001f5a8', codepoints=(128424,), name='skriver', slug='skriver', annotations=frozenset({'datamaskin'})),
EmojiAnnotations(emoji='⌨', codepoints=(9000,), name='tastatur', slug='tastatur', annotations=frozenset({'datamaskin'})),
EmojiAnnotations(emoji='\U0001f5b1', codepoints=(128433,), name='datamus', slug='datamus', annotations=frozenset({'knapp', 'tre', 'datamaskin', 'mus', '3'})),
EmojiAnnotations(emoji='\U0001f5b2', codepoints=(128434,), name='styrekule', slug='styrekule', annotations=frozenset({'datamaskin'})),
EmojiAnnotations(emoji='💽', codepoints=(128189,), name='MD', slug='md', annotations=frozenset({'minidisk'})),
EmojiAnnotations(emoji='💾', codepoints=(128190,), name='diskett', slug='diskett', annotations=frozenset({'datamaskin', 'lagring'})),
EmojiAnnotations(emoji='💿', codepoints=(128191,), name='blu-ray', slug='blu_ray', annotations=frozenset({'DVD', 'CD'})),
EmojiAnnotations(emoji='📀', codepoints=(128192,), name='DVD', slug='dvd', annotations=frozenset({'plate', 'DVD-plate'})),
EmojiAnnotations(emoji='🎥', codepoints=(127909,), name='filmkamera', slug='filmkamera', annotations=frozenset({'kamera', 'kino', 'film', 'underholdning'})),
EmojiAnnotations(emoji='🎬', codepoints=(127916,), name='filmklapper', slug='filmklapper', annotations=frozenset({'filming', 'film', 'action', 'underholdning'})),
EmojiAnnotations(emoji='\U0001f4fd', codepoints=(128253,), name='filmprosjektor', slug='filmprosjektor', annotations=frozenset({'kino', 'prosjektor', 'film', 'video'})),
EmojiAnnotations(emoji='📺', codepoints=(128250,), name='TV', slug='tv', annotations=frozenset({'fjernsyn'})),
EmojiAnnotations(emoji='📷', codepoints=(128247,), name='kamera', slug='kamera', annotations=frozenset({'kamerateknikk', 'teknologi'})),
EmojiAnnotations(emoji='\U0001f4f8', codepoints=(128248,), name='kamera med blits', slug='kamera_med_blits', annotations=frozenset({'kamera', 'blits', 'video'})),
EmojiAnnotations(emoji='📹', codepoints=(128249,), name='videokamera', slug='videokamera', annotations=frozenset({'kamera', 'video'})),
EmojiAnnotations(emoji='📼', codepoints=(128252,), name='videoopptak', slug='videoopptak', annotations=frozenset({'opptak', 'video'})),
EmojiAnnotations(emoji='🔍', codepoints=(128269,), name='forstørrelsesglass mot venstre', slug='forstørrelsesglass_mot_venstre', annotations=frozenset({'søk', 'glass'})),
EmojiAnnotations(emoji='🔎', codepoints=(128270,), name='forstørrelsesglass mot høyre', slug='forstørrelsesglass_mot_høyre', annotations=frozenset({'forstørrelsesglass mot venstre', 'søk', 'glass'})),
EmojiAnnotations(emoji='🔬', codepoints=(128300,), name='mikroskop', slug='mikroskop', annotations=frozenset({'verktøy', 'vitenskap'})),
EmojiAnnotations(emoji='🔭', codepoints=(128301,), name='teleskop', slug='teleskop', annotations=frozenset({'verktøy', 'astronomi', 'vitenskap'})),
EmojiAnnotations(emoji='📡', codepoints=(128225,), name='satellitt', slug='satellitt', annotations=frozenset({'antenne', 'satelittantenne', 'satelitt', 'kommunikasjon'})),
EmojiAnnotations(emoji='\U0001f56f', codepoints=(128367,), name='stearinlys', slug='stearinlys', annotations=frozenset({'lys'})),
EmojiAnnotations(emoji='💡', codepoints=(128161,), name='lyspære', slug='lyspære', annotations=frozenset({'idé', 'pære', 'tegneserie', 'elektrisitet', 'lys'})),
EmojiAnnotations(emoji='🔦', codepoints=(128294,), name='lommelykt', slug='lommelykt', annotations=frozenset({'lykt', 'elektrisk', 'fakkel'})),
EmojiAnnotations(emoji='🏮', codepoints=(127982,), name='lykt', slug='lykt', annotations=frozenset({'lanterne', 'rød lykt', 'izakaya'})),
EmojiAnnotations(emoji='📔', codepoints=(128212,), name='dekorert notatbok', slug='dekorert_notatbok', annotations=frozenset({'notatbok', 'omslag'})),
EmojiAnnotations(emoji='📕', codepoints=(128213,), name='lukket bok', slug='lukket_bok', annotations=frozenset({'bok'})),
EmojiAnnotations(emoji='📖', codepoints=(128214,), name='åpen bok', slug='åpen_bok', annotations=frozenset({'bok'})),
EmojiAnnotations(emoji='📗', codepoints=(128215,), name='grønn bok', slug='grønn_bok', annotations=frozenset({'bok'})),
EmojiAnnotations(emoji='📘', codepoints=(128216,), name='blå bok', slug='blå_bok', annotations=frozenset({'bok'})),
EmojiAnnotations(emoji='📙', codepoints=(128217,), name='oransje bok', slug='oransje_bok', annotations=frozenset({'bok'})),
EmojiAnnotations(emoji='📚', codepoints=(128218,), name='bøker', slug='bøker', annotations=frozenset({'bok'})),
EmojiAnnotations(emoji='📓', codepoints=(128211,), name='notatbok', slug='notatbok', annotations=frozenset({'kontorrekvisita', 'kladdebok'})),
EmojiAnnotations(emoji='📒', codepoints=(128210,), name='spiralblokk', slug='spiralblokk', annotations=frozenset({'notatbok', 'kontorrekvisita'})),
EmojiAnnotations(emoji='📃', codepoints=(128195,), name='dokument', slug='dokument', annotations=frozenset({'opprullet ark'})),
EmojiAnnotations(emoji='📜', codepoints=(128220,), name='rull', slug='rull', annotations=frozenset({'papirrull'})),
EmojiAnnotations(emoji='📄', codepoints=(128196,), name='dokument som vender opp', slug='dokument_som_vender_opp', annotations=frozenset({'ark', 'vender opp'})),
EmojiAnnotations(emoji='📰', codepoints=(128240,), name='avis', slug='avis', annotations=frozenset({'nyheter', 'publikasjon'})),
EmojiAnnotations(emoji='\U0001f5de', codepoints=(128478,), name='opprullet avis', slug='opprullet_avis', annotations=frozenset({'papir', 'avis', 'nyheter', 'opprullet'})),
EmojiAnnotations(emoji='📑', codepoints=(128209,), name='bokmerkefane', slug='bokmerkefane', annotations=frozenset({'markør', 'faner', 'bokmerke'})),
EmojiAnnotations(emoji='🔖', codepoints=(128278,), name='bokmerke', slug='bokmerke', annotations=frozenset({'merke'})),
EmojiAnnotations(emoji='💰', codepoints=(128176,), name='pengesekk', slug='pengesekk', annotations=frozenset({'sekk', 'penger'})),
EmojiAnnotations(emoji='💴', codepoints=(128180,), name='yen-seddel', slug='yen_seddel', annotations=frozenset({'yen', 'penger'})),
EmojiAnnotations(emoji='💵', codepoints=(128181,), name='dollar-seddel', slug='dollar_seddel', annotations=frozenset({'dollar', 'dollarseddel', 'penger'})),
EmojiAnnotations(emoji='💶', codepoints=(128182,), name='euro-seddel', slug='euro_seddel', annotations=frozenset({'euro', 'euroseddel', 'penger'})),
EmojiAnnotations(emoji='💷', codepoints=(128183,), name='pund-seddel', slug='pund_seddel', annotations=frozenset({'pundseddel', 'pund', 'penger'})),
EmojiAnnotations(emoji='💸', codepoints=(128184,), name='flygende penger', slug='flygende_penger', annotations=frozenset({'flygende seddel', 'bank', 'penger med vinger', 'pengeseddel', 'penger'})),
EmojiAnnotations(emoji='💳', codepoints=(128179,), name='kredittkort', slug='kredittkort', annotations=frozenset({'betalingskort', 'kort', 'visa', 'kreditt'})),
EmojiAnnotations(emoji='💹', codepoints=(128185,), name='stigende marked', slug='stigende_marked', annotations=frozenset({'oppover', 'stigende graf', 'graf', 'marked'})),
EmojiAnnotations(emoji='✉', codepoints=(9993,), name='konvolutt', slug='konvolutt', annotations=frozenset({'e-post'})),
EmojiAnnotations(emoji='📧', codepoints=(128231,), name='e-post', slug='e_post', annotations=frozenset({'kommunikasjon', 'brev', 'post'})),
EmojiAnnotations(emoji='📨', codepoints=(128232,), name='innkommende e-post', slug='innkommende_e_post', annotations=frozenset({'e-post', 'kommunikasjon', 'mottatt e-post', 'konvolutt', 'brev', 'innkommende', 'post'})),
EmojiAnnotations(emoji='📩', codepoints=(128233,), name='utgående e-post', slug='utgående_e_post', annotations=frozenset({'e-post', 'kommunikasjon', 'sendt e-post', 'utgående', 'konvolutt', 'brev', 'post'})),
EmojiAnnotations(emoji='📤', codepoints=(128228,), name='utboks', slug='utboks', annotations=frozenset({'sendt e-post', 'e-post', 'sent', 'skuff', 'kommunikasjon'})),
EmojiAnnotations(emoji='📥', codepoints=(128229,), name='innboks', slug='innboks', annotations=frozenset({'mottatt e-post', 'mottatt', 'skuff', 'e-post', 'kommunikasjon'})),
EmojiAnnotations(emoji='📦', codepoints=(128230,), name='pakke', slug='pakke', annotations=frozenset({'boks', 'levering'})),
EmojiAnnotations(emoji='📫', codepoints=(128235,), name='lukket postkasse med heist flagg', slug='lukket_postkasse_med_heist_flagg', annotations=frozenset({'flagg', 'postkasse med heist flagg', 'post'})),
EmojiAnnotations(emoji='📪', codepoints=(128234,), name='lukket postkasse med senket flagg', slug='lukket_postkasse_med_senket_flagg', annotations=frozenset({'postkasse med senket flagg', 'flagg', 'post'})),
EmojiAnnotations(emoji='📬', codepoints=(128236,), name='åpen postkasse med heist flagg', slug='åpen_postkasse_med_heist_flagg', annotations=frozenset({'flagg', 'postkasse med heist flagg', 'post'})),
EmojiAnnotations(emoji='📭', codepoints=(128237,), name='åpen postkasse med senket flagg', slug='åpen_postkasse_med_senket_flagg', annotations=frozenset({'postkasse med senket flagg', 'flagg', 'post'})),
EmojiAnnotations(emoji='📮', codepoints=(128238,), name='postkasse', slug='postkasse', annotations=frozenset({'post'})),
EmojiAnnotations(emoji='\U0001f5f3', codepoints=(128499,), name='urne med stemmeseddel', slug='urne_med_stemmeseddel', annotations=frozenset({'stemmeseddel', 'boks'})),
EmojiAnnotations(emoji='✒', codepoints=(10002,), name='svart spiss', slug='svart_spiss', annotations=frozenset({'penn', 'pennespiss'})),
EmojiAnnotations(emoji='\U0001f58b', codepoints=(128395,), name='fyllepenn', slug='fyllepenn', annotations=frozenset({'penn'})),
EmojiAnnotations(emoji='\U0001f58a', codepoints=(128394,), name='penn', slug='penn', annotations=frozenset({'kulepenn'})),
EmojiAnnotations(emoji='\U0001f58c', codepoints=(128396,), name='pensel', slug='pensel', annotations=frozenset({'maling'})),
EmojiAnnotations(emoji='\U0001f58d', codepoints=(128397,), name='tegnestift', slug='tegnestift', annotations=frozenset({'tegning'})),
EmojiAnnotations(emoji='📝', codepoints=(128221,), name='huskeliste', slug='huskeliste', annotations=frozenset({'bok', 'kommunikasjon'})),
EmojiAnnotations(emoji='💼', codepoints=(128188,), name='koffert', slug='koffert', annotations=frozenset({'briefcase'})),
EmojiAnnotations(emoji='📁', codepoints=(128193,), name='filmappe', slug='filmappe', annotations=frozenset({'mappe', 'fil'})),
EmojiAnnotations(emoji='📂', codepoints=(128194,), name='åpen mappe', slug='åpen_mappe', annotations=frozenset({'åpen filmappe', 'åpen', 'mappe', 'fil'})),
EmojiAnnotations(emoji='\U0001f5c2', codepoints=(128450,), name='skillekort', slug='skillekort', annotations=frozenset({'indeks', 'kort', 'skillere'})),
EmojiAnnotations(emoji='📅', codepoints=(128197,), name='kalender', slug='kalender', annotations=frozenset({'dato'})),
EmojiAnnotations(emoji='📆', codepoints=(128198,), name='riv-av-kalender', slug='riv_av_kalender', annotations=frozenset({'kalender'})),
EmojiAnnotations(emoji='\U0001f5d2', codepoints=(128466,), name='spiralhefte', slug='spiralhefte', annotations=frozenset({'blokk', 'spiral', 'notat'})),
EmojiAnnotations(emoji='\U0001f5d3', codepoints=(128467,), name='spiralkalender', slug='spiralkalender', annotations=frozenset({'blokk', 'spiral', 'kalender'})),
EmojiAnnotations(emoji='📇', codepoints=(128199,), name='kortindeks', slug='kortindeks', annotations=frozenset({'indeks', 'register'})),
EmojiAnnotations(emoji='📈', codepoints=(128200,), name='positiv trend', slug='positiv_trend', annotations=frozenset({'graf', 'positiv', 'positive trend-graf'})),
EmojiAnnotations(emoji='📉', codepoints=(128201,), name='negativ trend', slug='negativ_trend', annotations=frozenset({'negativ trend-graf', 'graf', 'positiv'})),
EmojiAnnotations(emoji='📊', codepoints=(128202,), name='stolpediagram', slug='stolpediagram', annotations=frozenset({'diagram', 'søylediagram'})),
EmojiAnnotations(emoji='📌', codepoints=(128204,), name='stift', slug='stift', annotations=frozenset({'kontorrekvisita'})),
EmojiAnnotations(emoji='📍', codepoints=(128205,), name='knappenål', slug='knappenål', annotations=frozenset({'markørnål', 'kontorrekvisita'})),
EmojiAnnotations(emoji='📎', codepoints=(128206,), name='binders', slug='binders', annotations=frozenset({'kontorrekvisita'})),
EmojiAnnotations(emoji='\U0001f587', codepoints=(128391,), name='sammenlenkede binders', slug='sammenlenkede_binders', annotations=frozenset({'koble sammen', 'binders'})),
EmojiAnnotations(emoji='📏', codepoints=(128207,), name='linjal', slug='linjal', annotations=frozenset({'rett linjal'})),
EmojiAnnotations(emoji='📐', codepoints=(128208,), name='trekantet linjal', slug='trekantet_linjal', annotations=frozenset({'linjal', 'trekant'})),
EmojiAnnotations(emoji='✂', codepoints=(9986,), name='saks', slug='saks', annotations=frozenset({'verktøy'})),
EmojiAnnotations(emoji='\U0001f5c3', codepoints=(128451,), name='kartotekboks', slug='kartotekboks', annotations=frozenset({'boks', 'kort', 'fil'})),
EmojiAnnotations(emoji='\U0001f5c4', codepoints=(128452,), name='arkivskap', slug='arkivskap', annotations=frozenset({'kabinett', 'fil'})),
EmojiAnnotations(emoji='🔒', codepoints=(128274,), name='lås', slug='lås', annotations=frozenset({'låst'})),
EmojiAnnotations(emoji='🔏', codepoints=(128271,), name='lås med penn', slug='lås_med_penn', annotations=frozenset({'personvern', 'penn'})),
EmojiAnnotations(emoji='🔐', codepoints=(128272,), name='lukket lås', slug='lukket_lås', annotations=frozenset({'lås med nøkkel', 'sikker'})),
EmojiAnnotations(emoji='🔑', codepoints=(128273,), name='nøkkel', slug='nøkkel', annotations=frozenset({'passord'})),
EmojiAnnotations(emoji='\U0001f5dd', codepoints=(128477,), name='gammeldags nøkkel', slug='gammeldags_nøkkel', annotations=frozenset({'gammel', 'nøkkel', 'ledetråd', 'lås'})),
EmojiAnnotations(emoji='🔨', codepoints=(128296,), name='hammer', slug='hammer', annotations=frozenset({'verktøy'})),
EmojiAnnotations(emoji='⛏', codepoints=(9935,), name='hakke', slug='hakke', annotations=frozenset({'verktøy', 'gruvedrift'})),
EmojiAnnotations(emoji='⚒', codepoints=(9874,), name='hammer og hakke', slug='hammer_og_hakke', annotations=frozenset({'hammer', 'verktøy', 'hakke'})),
EmojiAnnotations(emoji='\U0001f6e0', codepoints=(128736,), name='hammer og skiftenøkkel', slug='hammer_og_skiftenøkkel', annotations=frozenset({'hammer', 'verktøy', 'skiftenøkkel'})),
EmojiAnnotations(emoji='🔧', codepoints=(128295,), name='skiftenøkkel', slug='skiftenøkkel', annotations=frozenset({'verktøy'})),
EmojiAnnotations(emoji='🔩', codepoints=(128297,), name='mutter og skrue', slug='mutter_og_skrue', annotations=frozenset({'mutter', 'skrue', 'verktøy'})),
EmojiAnnotations(emoji='⚙', codepoints=(9881,), name='tannhjul', slug='tannhjul', annotations=frozenset({'verktøy'})),
EmojiAnnotations(emoji='\U0001f5dc', codepoints=(128476,), name='skrustikke', slug='skrustikke', annotations=frozenset({'verktøy'})),
EmojiAnnotations(emoji='⚗', codepoints=(9879,), name='destillasjonskolbe', slug='destillasjonskolbe', annotations=frozenset({'verktøy', 'kjemi'})),
EmojiAnnotations(emoji='⚖', codepoints=(9878,), name='likearmet skålvekt', slug='likearmet_skålvekt', annotations=frozenset({'stjernetegn', 'vekt', 'balanse', 'vekten', 'rettferdighet', 'verktøy'})),
EmojiAnnotations(emoji='🔗', codepoints=(128279,), name='linksymbol', slug='linksymbol', annotations=frozenset({'to ringer', 'link'})),
EmojiAnnotations(emoji='💉', codepoints=(128137,), name='sprøyte', slug='sprøyte', annotations=frozenset({'vaksine', 'medisin', 'nål', 'lege'})),
EmojiAnnotations(emoji='💊', codepoints=(128138,), name='pille', slug='pille', annotations=frozenset({'medisin', 'lege'})),
EmojiAnnotations(emoji='\U0001f5e1', codepoints=(128481,), name='dolk', slug='dolk', annotations=frozenset({'våpen', 'kniv'})),
EmojiAnnotations(emoji='🔪', codepoints=(128298,), name='kokkekniv', slug='kokkekniv', annotations=frozenset({'kniv'})),
EmojiAnnotations(emoji='⚔', codepoints=(9876,), name='sverd i kryss', slug='sverd_i_kryss', annotations=frozenset({'våpen', 'sverd', 'kryss'})),
EmojiAnnotations(emoji='🔫', codepoints=(128299,), name='pistol', slug='pistol', annotations=frozenset({'revolver'})),
EmojiAnnotations(emoji='\U0001f6e1', codepoints=(128737,), name='skjold', slug='skjold', annotations=frozenset({'våpen'})),
EmojiAnnotations(emoji='\U0001f3f9', codepoints=(127993,), name='pil og bue', slug='pil_og_bue', annotations=frozenset({'stjernetegn', 'bue', 'pil', 'bueskytter', 'verktøy', 'våpen', 'skytten'})),
EmojiAnnotations(emoji='🏁', codepoints=(127937,), name='bilraceflagg', slug='bilraceflagg', annotations=frozenset({'racingflagg', 'motorsport', 'rutete flagg'})),
EmojiAnnotations(emoji='\U0001f3f3', codepoints=(127987,), name='flagrende hvitt flagg', slug='flagrende_hvitt_flagg', annotations=frozenset({'flagring', 'flagg'})),
EmojiAnnotations(emoji='\U0001f3f4', codepoints=(127988,), name='flagrende svart flagg', slug='flagrende_svart_flagg', annotations=frozenset({'flagring', 'flagg'})),
EmojiAnnotations(emoji='🚩', codepoints=(128681,), name='posisjonsflagg', slug='posisjonsflagg', annotations=frozenset({'flagg'})),
EmojiAnnotations(emoji='🚬', codepoints=(128684,), name='røykeskilt', slug='røykeskilt', annotations=frozenset({'røyking', 'skilt', 'røyking tillatt'})),
EmojiAnnotations(emoji='⚰', codepoints=(9904,), name='kiste', slug='kiste', annotations=frozenset({'død'})),
EmojiAnnotations(emoji='⚱', codepoints=(9905,), name='urne', slug='urne', annotations=frozenset({'død', 'begravelse'})),
EmojiAnnotations(emoji='🗿', codepoints=(128511,), name='Moyai-statue', slug='moyai_statue', annotations=frozenset({'statue', 'Moyai'})),
EmojiAnnotations(emoji='\U0001f6e2', codepoints=(128738,), name='oljetønne', slug='oljetønne', annotations=frozenset({'olje', 'tønne'})),
EmojiAnnotations(emoji='🔮', codepoints=(128302,), name='spåkule', slug='spåkule', annotations=frozenset({'spådommer', 'krystallkule', 'spådom'})),
EmojiAnnotations(emoji='🏧', codepoints=(127975,), name='minibank', slug='minibank', annotations=frozenset({'pengeuttak'})),
EmojiAnnotations(emoji='🚮', codepoints=(128686,), name='avfallskilt', slug='avfallskilt', annotations=frozenset({'kast avfal', 'søppelkasse', 'skilt'})),
EmojiAnnotations(emoji='🚰', codepoints=(128688,), name='drikkevann', slug='drikkevann', annotations=frozenset({'drikkevannskilt', 'vann'})),
EmojiAnnotations(emoji='♿', codepoints=(9855,), name='rullestol', slug='rullestol', annotations=frozenset({'rullestolsymbol', 'handikaptilgang', 'rullestolbruker'})),
EmojiAnnotations(emoji='🚹', codepoints=(128697,), name='menn', slug='menn', annotations=frozenset({'herretoalett', 'mannesymbol', 'skilt'})),
EmojiAnnotations(emoji='🚺', codepoints=(128698,), name='kvinner', slug='kvinner', annotations=frozenset({'kvinnesymbol', 'dametoalett', 'skilt'})),
EmojiAnnotations(emoji='🚻', codepoints=(128699,), name='toalettskilt', slug='toalettskilt', annotations=frozenset({'WC', 'do'})),
EmojiAnnotations(emoji='🚼', codepoints=(128700,), name='babyskilt', slug='babyskilt', annotations=frozenset({'stellestasjon', 'babysymbol'})),
EmojiAnnotations(emoji='🚾', codepoints=(128702,), name='do', slug='do', annotations=frozenset({'WC', 'toalett'})),
EmojiAnnotations(emoji='🛂', codepoints=(128706,), name='passkontroll', slug='passkontroll', annotations=frozenset({'pass', 'kontroll'})),
EmojiAnnotations(emoji='🛃', codepoints=(128707,), name='toll', slug='toll', annotations=frozenset({'tollen'})),
EmojiAnnotations(emoji='🛄', codepoints=(128708,), name='rullebånd', slug='rullebånd', annotations=frozenset({'bagasje', 'bagasjebånd'})),
EmojiAnnotations(emoji='🛅', codepoints=(128709,), name='glemt bagasje', slug='glemt_bagasje', annotations=frozenset({'bagasje'})),
EmojiAnnotations(emoji='⚠', codepoints=(9888,), name='advarsel', slug='advarsel', annotations=frozenset({'OBS'})),
EmojiAnnotations(emoji='🚸', codepoints=(128696,), name='barn krysser', slug='barn_krysser', annotations=frozenset({'krysser', 'barn', 'skilt'})),
EmojiAnnotations(emoji='⛔', codepoints=(9940,), name='ingen adgang', slug='ingen_adgang', annotations=frozenset({'adgang forbudt', 'stopp'})),
EmojiAnnotations(emoji='🚫', codepoints=(128683,), name='forbudt', slug='forbudt', annotations=frozenset({'adgang forbudt', 'skilt'})),
EmojiAnnotations(emoji='🚳', codepoints=(128691,), name='sykler forbudt', slug='sykler_forbudt', annotations=frozenset({'sykkel ikke tillatt', 'ingen sykler', 'sykkel', 'kjøretøy'})),
EmojiAnnotations(emoji='🚭', codepoints=(128685,), name='røyking forbudt', slug='røyking_forbudt', annotations=frozenset({'ingen røyking', 'skilt'})),
EmojiAnnotations(emoji='🚯', codepoints=(128687,), name='ikke kast avfall', slug='ikke_kast_avfall', annotations=frozenset({'ikke avfall', 'ingen kasting', 'skilt'})),
EmojiAnnotations(emoji='🚱', codepoints=(128689,), name='ikke drikkevann', slug='ikke_drikkevann', annotations=frozenset({'vann', 'ikke drikkbart'})),
EmojiAnnotations(emoji='🚷', codepoints=(128695,), name='fotgjengere forbudt', slug='fotgjengere_forbudt', annotations=frozenset({'fotgjenger', 'gående', 'ingen fotgjengere', 'skilt', 'fotgjengere ikke tillatt'})),
EmojiAnnotations(emoji='☢', codepoints=(9762,), name='radioaktivt', slug='radioaktivt', annotations=frozenset({'radioaktiv'})),
EmojiAnnotations(emoji='⬆', codepoints=(11014,), name='pil opp', slug='pil_opp', annotations=frozenset({'nord', 'pil', 'retning'})),
EmojiAnnotations(emoji='↗', codepoints=(8599,), name='pil opp-høyre', slug='pil_opp_høyre', annotations=frozenset({'pil', 'retning', 'nordøst'})),
EmojiAnnotations(emoji='➡', codepoints=(10145,), name='pil høyre', slug='pil_høyre', annotations=frozenset({'øst', 'pil', 'retning'})),
EmojiAnnotations(emoji='↘', codepoints=(8600,), name='pil ned-høyre', slug='pil_ned_høyre', annotations=frozenset({'pil', 'retning', 'sørøst'})),
EmojiAnnotations(emoji='⬇', codepoints=(11015,), name='pil ned', slug='pil_ned', annotations=frozenset({'pil', 'retning', 'sør'})),
EmojiAnnotations(emoji='↙', codepoints=(8601,), name='pil ned-venstre', slug='pil_ned_venstre', annotations=frozenset({'pil', 'retning', 'sørvest'})),
EmojiAnnotations(emoji='⬅', codepoints=(11013,), name='pil venstre', slug='pil_venstre', annotations=frozenset({'pil', 'retning', 'vest'})),
EmojiAnnotations(emoji='↖', codepoints=(8598,), name='pil opp-venstre', slug='pil_opp_venstre', annotations=frozenset({'nordvest', 'pil', 'retning'})),
EmojiAnnotations(emoji='↕', codepoints=(8597,), name='pil opp-ned', slug='pil_opp_ned', annotations=frozenset({'pil'})),
EmojiAnnotations(emoji='↔', codepoints=(8596,), name='pil venstre-høyre', slug='pil_venstre_høyre', annotations=frozenset({'pil'})),
EmojiAnnotations(emoji='↩', codepoints=(8617,), name='pil som bøyer mot venstre', slug='pil_som_bøyer_mot_venstre', annotations=frozenset({'pil'})),
EmojiAnnotations(emoji='↪', codepoints=(8618,), name='pil som bøyer mot høyre', slug='pil_som_bøyer_mot_høyre', annotations=frozenset({'pil'})),
EmojiAnnotations(emoji='⤴', codepoints=(10548,), name='pil som bøyer opp', slug='pil_som_bøyer_opp', annotations=frozenset({'pil'})),
EmojiAnnotations(emoji='⤵', codepoints=(10549,), name='pil som bøyer ned', slug='pil_som_bøyer_ned', annotations=frozenset({'pil', 'ned'})),
EmojiAnnotations(emoji='🔃', codepoints=(128259,), name='piler i sirkel', slug='piler_i_sirkel', annotations=frozenset({'vertikale piler med klokken', 'piler', 'last inn på nytt'})),
EmojiAnnotations(emoji='🔄', codepoints=(128260,), name='piler mot klokken', slug='piler_mot_klokken', annotations=frozenset({'piler', 'mot klokken'})),
EmojiAnnotations(emoji='🔙', codepoints=(128281,), name='tilbake', slug='tilbake', annotations=frozenset({'tilbakepil', 'pil', 'pil til venstre'})),
EmojiAnnotations(emoji='🔚', codepoints=(128282,), name='slutt', slug='slutt', annotations=frozenset({'sluttpil', 'slutt med pil', 'pil', 'pil til venstre'})),
EmojiAnnotations(emoji='🔛', codepoints=(128283,), name='PÅ-pil', slug='på_pil', annotations=frozenset({'PÅ med utropstegn og pil', 'utropstegn', 'utrop', 'pil'})),
EmojiAnnotations(emoji='🔜', codepoints=(128284,), name='snart-tegn', slug='snart_tegn', annotations=frozenset({'snart', 'pil', 'snart med pil'})),
EmojiAnnotations(emoji='🔝', codepoints=(128285,), name='toppsymbol', slug='toppsymbol', annotations=frozenset({'topp med pil', 'pil opp', 'øverst', 'opp'})),
EmojiAnnotations(emoji='\U0001f6d0', codepoints=(128720,), name='sted for tilbedelse', slug='sted_for_tilbedelse', annotations=frozenset({'tilbedelse', 'religion'})),
EmojiAnnotations(emoji='⚛', codepoints=(9883,), name='atom', slug='atom', annotations=frozenset({'ateist'})),
EmojiAnnotations(emoji='\U0001f549', codepoints=(128329,), name='om', slug='om', annotations=frozenset({'hindu', 'religion'})),
EmojiAnnotations(emoji='✡', codepoints=(10017,), name='Davidstjernen', slug='davidstjernen', annotations=frozenset({'jøde', 'david', 'jødisk', 'religion', 'stjerne'})),
EmojiAnnotations(emoji='☸', codepoints=(9784,), name='dharmahjulet', slug='dharmahjulet', annotations=frozenset({'dharma', 'hjul', 'buddhist', 'religion'})),
EmojiAnnotations(emoji='☯', codepoints=(9775,), name='yin yang', slug='yin_yang', annotations=frozenset({'yin', 'taoist', 'religion', 'tao', 'yang'})),
EmojiAnnotations(emoji='✝', codepoints=(10013,), name='latinsk kors', slug='latinsk_kors', annotations=frozenset({'kors', 'religion', 'kristen'})),
EmojiAnnotations(emoji='☦', codepoints=(9766,), name='ortodoks kors', slug='ortodoks_kors', annotations=frozenset({'kors', 'religion', 'kristen'})),
EmojiAnnotations(emoji='☪', codepoints=(9770,), name='stjerne og halvmåne', slug='stjerne_og_halvmåne', annotations=frozenset({'muslim', 'islam', 'religion'})),
EmojiAnnotations(emoji='☮', codepoints=(9774,), name='fredssymbol', slug='fredssymbol', annotations=frozenset({'fred'})),
EmojiAnnotations(emoji='\U0001f54e', codepoints=(128334,), name='menora', slug='menora', annotations=frozenset({'lysestake', 'religion'})),
EmojiAnnotations(emoji='🔯', codepoints=(128303,), name='sekskantet stjerne', slug='sekskantet_stjerne', annotations=frozenset({'sekskantet', 'magi', 'stjerne'})),
EmojiAnnotations(emoji='♻', codepoints=(9851,), name='resirkulering', slug='resirkulering', annotations=frozenset({'resirkuleringssymbol', 'universelt', 'symbol'})),
EmojiAnnotations(emoji='📛', codepoints=(128219,), name='navneskilt', slug='navneskilt', annotations=frozenset({'navn', 'skilt'})),
EmojiAnnotations(emoji='⚜', codepoints=(9884,), name='fransk lilje', slug='fransk_lilje', annotations=frozenset({'fleur-de-lis'})),
EmojiAnnotations(emoji='🔰', codepoints=(128304,), name='grønt og gult blad', slug='grønt_og_gult_blad', annotations=frozenset({'grønn og gul', 'japansk', 'nybegynner'})),
EmojiAnnotations(emoji='🔱', codepoints=(128305,), name='tridentsymbol', slug='tridentsymbol', annotations=frozenset({'emblem', 'treenighet', 'anker'})),
EmojiAnnotations(emoji='⭕', codepoints=(11093,), name='fremhevet stor sirkel', slug='fremhevet_stor_sirkel', annotations=frozenset({'o', 'sirkel'})),
EmojiAnnotations(emoji='✅', codepoints=(9989,), name='hake', slug='hake', annotations=frozenset({'avmerking', 'avkryssing'})),
EmojiAnnotations(emoji='☑', codepoints=(9745,), name='stemmeurne med hakemerke', slug='stemmeurne_med_hakemerke', annotations=frozenset({'stemmeseddel', 'boks', 'hakemerke'})),
EmojiAnnotations(emoji='✔', codepoints=(10004,), name='fremhevet hakemerke', slug='fremhevet_hakemerke', annotations=frozenset({'merke', 'hakemerke', 'hake'})),
EmojiAnnotations(emoji='✖', codepoints=(10006,), name='fremhevet gangetegn', slug='fremhevet_gangetegn', annotations=frozenset({'multiplikasjon', 'x', 'avbryt', 'multiplisere'})),
EmojiAnnotations(emoji='❌', codepoints=(10060,), name='kryss', slug='kryss', annotations=frozenset({'rødt kryss', '«X»'})),
EmojiAnnotations(emoji='❎', codepoints=(10062,), name='kryss i firkant', slug='kryss_i_firkant', annotations=frozenset({'kryss'})),
EmojiAnnotations(emoji='➕', codepoints=(10133,), name='pluss', slug='pluss', annotations=frozenset({'plusstegn'})),
EmojiAnnotations(emoji='➖', codepoints=(10134,), name='minus', slug='minus', annotations=frozenset({'minustegn'})),
EmojiAnnotations(emoji='➗', codepoints=(10135,), name='dele', slug='dele', annotations=frozenset({'divisjonstegn', 'deletegn'})),
EmojiAnnotations(emoji='➰', codepoints=(10160,), name='løkke', slug='løkke', annotations=frozenset({'løkketegn'})),
EmojiAnnotations(emoji='➿', codepoints=(10175,), name='dobbel løkke', slug='dobbel_løkke', annotations=frozenset({'dobbel løkke-tegn'})),
EmojiAnnotations(emoji='〽', codepoints=(12349,), name='delvis alternering', slug='delvis_alternering', annotations=frozenset({'delvis', 'merke'})),
EmojiAnnotations(emoji='✳', codepoints=(10035,), name='stjerne med åtte eiker', slug='stjerne_med_åtte_eiker', annotations=frozenset({'asterisk'})),
EmojiAnnotations(emoji='✴', codepoints=(10036,), name='åttekantet stjerne', slug='åttekantet_stjerne', annotations=frozenset({'stjerne'})),
EmojiAnnotations(emoji='💱', codepoints=(128177,), name='valutaveksling', slug='valutaveksling', annotations=frozenset({'valuta', 'veksling', 'penger'})),
EmojiAnnotations(emoji='💲', codepoints=(128178,), name='dollartegn', slug='dollartegn', annotations=frozenset({'dollar', 'stort dollartegn', 'penger'})),
EmojiAnnotations(emoji='‼', codepoints=(8252,), name='to utropstegn', slug='to_utropstegn', annotations=frozenset({'tegn', 'utrop', 'utropstegn', 'tegnsetting', 'bangbang'})),
EmojiAnnotations(emoji='⁉', codepoints=(8265,), name='utropstegn og spørsmålstegn', slug='utropstegn_og_spørsmålstegn', annotations=frozenset({'spørrende', 'spørsmål', 'utropstegn', 'tegnsetting', 'tegn', 'utrop', 'spørsmålstegn'})),
EmojiAnnotations(emoji='❓', codepoints=(10067,), name='rødt spørsmålstegn', slug='rødt_spørsmålstegn', annotations=frozenset({'spørsmål'})),
EmojiAnnotations(emoji='❔', codepoints=(10068,), name='hvitt spørsmålstegn', slug='hvitt_spørsmålstegn', annotations=frozenset({'spørsmål'})),
EmojiAnnotations(emoji='❕', codepoints=(10069,), name='hvitt utropstegn', slug='hvitt_utropstegn', annotations=frozenset({'utrop'})),
EmojiAnnotations(emoji='❗', codepoints=(10071,), name='utropstegn', slug='utropstegn', annotations=frozenset({'tegn', 'utrop', 'tegnsetting'})),
EmojiAnnotations(emoji='〰', codepoints=(12336,), name='bølgestrek', slug='bølgestrek', annotations=frozenset({'tegn for lang lyd', 'bølget bindestrek', 'bølge'})),
EmojiAnnotations(emoji='®', codepoints=(174,), name='registrert varemerke', slug='registrert_varemerke', annotations=frozenset({'varemerkesymbol'})),
EmojiAnnotations(emoji='♈', codepoints=(9800,), name='Væren', slug='væren', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♉', codepoints=(9801,), name='Tyren', slug='tyren', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♊', codepoints=(9802,), name='Tvillingene', slug='tvillingene', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♋', codepoints=(9803,), name='Krepsen', slug='krepsen', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♌', codepoints=(9804,), name='Løven', slug='løven', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♍', codepoints=(9805,), name='Jomfruen', slug='jomfruen', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♎', codepoints=(9806,), name='Vekten', slug='vekten', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♏', codepoints=(9807,), name='Skorpionen', slug='skorpionen', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♐', codepoints=(9808,), name='Skytten', slug='skytten', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♑', codepoints=(9809,), name='Steinbukken', slug='steinbukken', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♒', codepoints=(9810,), name='Vannmannen', slug='vannmannen', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♓', codepoints=(9811,), name='Fiskene', slug='fiskene', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='⛎', codepoints=(9934,), name='stjernetegn', slug='stjernetegn', annotations=frozenset({'Ophiuchus', 'astrologi', 'Slangebæreren'})),
EmojiAnnotations(emoji='🔀', codepoints=(128256,), name='kryssede piler', slug='kryssede_piler', annotations=frozenset({'piler', 'kryssede piler mot høyre'})),
EmojiAnnotations(emoji='🔁', codepoints=(128257,), name='gjenta', slug='gjenta', annotations=frozenset({'piler', 'med klokken'})),
EmojiAnnotations(emoji='🔂', codepoints=(128258,), name='gjenta én gang', slug='gjenta_én_gang', annotations=frozenset({'piler', 'piler med klokken med nummer 1 med klokken', 'nummer 1'})),
EmojiAnnotations(emoji='▶', codepoints=(9654,), name='spill av', slug='spill_av', annotations=frozenset({'triangel', 'høyre', 'pil'})),
EmojiAnnotations(emoji='⏩', codepoints=(9193,), name='spol kjapt fremover', slug='spol_kjapt_fremover', annotations=frozenset({'spoling', 'høyre', 'dobbel pil'})),
EmojiAnnotations(emoji='⏭', codepoints=(9197,), name='neste spor', slug='neste_spor', annotations=frozenset({'triangel', 'neste scene', 'pil'})),
EmojiAnnotations(emoji='⏯', codepoints=(9199,), name='spill av/pause-knapp', slug='spill_av/pause_knapp', annotations=frozenset({'triangel', 'høyre', 'pil', 'pause', 'spill av'})),
EmojiAnnotations(emoji='◀', codepoints=(9664,), name='spol tilbake', slug='spol_tilbake', annotations=frozenset({'triangel', 'venstre', 'pil', 'bakover'})),
EmojiAnnotations(emoji='⏪', codepoints=(9194,), name='spol kjapt tilbake', slug='spol_kjapt_tilbake', annotations=frozenset({'spoling', 'venstre', 'dobbel pil'})),
EmojiAnnotations(emoji='⏮', codepoints=(9198,), name='forrige spor', slug='forrige_spor', annotations=frozenset({'triangel', 'pil', 'forrige scene'})),
EmojiAnnotations(emoji='🔼', codepoints=(128316,), name='gå opp', slug='gå_opp', annotations=frozenset({'oppover'})),
EmojiAnnotations(emoji='⏫', codepoints=(9195,), name='gå opp kjapt', slug='gå_opp_kjapt', annotations=frozenset({'dobbelt pil opp', 'pil', 'opp'})),
EmojiAnnotations(emoji='🔽', codepoints=(128317,), name='gå ned', slug='gå_ned', annotations=frozenset({'nedover'})),
EmojiAnnotations(emoji='⏬', codepoints=(9196,), name='gå ned kjapt', slug='gå_ned_kjapt', annotations=frozenset({'pil', 'ned', 'dobbelt pil ned'})),
EmojiAnnotations(emoji='\u23f8', codepoints=(9208,), name='pause', slug='pause', annotations=frozenset({'strek', 'dobbel', 'vertikal'})),
EmojiAnnotations(emoji='\u23f9', codepoints=(9209,), name='stopp', slug='stopp', annotations=frozenset({'firkant'})),
EmojiAnnotations(emoji='\u23fa', codepoints=(9210,), name='spill inn', slug='spill_inn', annotations=frozenset({'sirkel'})),
EmojiAnnotations(emoji='🎦', codepoints=(127910,), name='kino', slug='kino', annotations=frozenset({'kamera', 'film', 'filmkamera', 'underholdning'})),
EmojiAnnotations(emoji='🔅', codepoints=(128261,), name='lite lys', slug='lite_lys', annotations=frozenset({'lysstyrke', 'lav', 'lav lysstyrke'})),
EmojiAnnotations(emoji='🔆', codepoints=(128262,), name='mye lys', slug='mye_lys', annotations=frozenset({'høy', 'lysstyrke', 'høy lysstyrke'})),
EmojiAnnotations(emoji='📶', codepoints=(128246,), name='mobilsignaler', slug='mobilsignaler', annotations=frozenset({'telefon', 'antenne', 'signalstyrke', 'signal', 'mobil'})),
EmojiAnnotations(emoji='📵', codepoints=(128245,), name='mobil forbudt', slug='mobil_forbudt', annotations=frozenset({'ingen telefonsamtaler', 'telefon', 'mobil ikke tillatt'})),
EmojiAnnotations(emoji='📳', codepoints=(128243,), name='vibreringsmodus', slug='vibreringsmodus', annotations=frozenset({'modus', 'mobiltelefon', 'mobil', 'telefon'})),
EmojiAnnotations(emoji='📴', codepoints=(128244,), name='telefon av', slug='telefon_av', annotations=frozenset({'av', 'mobiltelefon', 'mobil', 'mobiltelefon av'})),
EmojiAnnotations(emoji='{#⃣}', codepoints=(123, 35, 8419, 125), name='nummertast nummertegn', slug='nummertast_nummertegn', annotations=frozenset({'nummertegn'})),
EmojiAnnotations(emoji='{*⃣}', codepoints=(123, 42, 8419, 125), name='nummertast stjerne', slug='nummertast_stjerne', annotations=frozenset({'asterisk', 'stjerne', 'keycap'})),
EmojiAnnotations(emoji='{0⃣}', codepoints=(123, 48, 8419, 125), name='nummertast null', slug='nummertast_null', annotations=frozenset({'0', 'null', 'nummertast'})),
EmojiAnnotations(emoji='{1⃣}', codepoints=(123, 49, 8419, 125), name='nummertast én', slug='nummertast_én', annotations=frozenset({'én', 'nummertast', '1'})),
EmojiAnnotations(emoji='{2⃣}', codepoints=(123, 50, 8419, 125), name='nummertast to', slug='nummertast_to', annotations=frozenset({'nummertast', '2', 'to'})),
EmojiAnnotations(emoji='{3⃣}', codepoints=(123, 51, 8419, 125), name='nummertast tre', slug='nummertast_tre', annotations=frozenset({'tre', 'nummertast', '3'})),
EmojiAnnotations(emoji='{4⃣}', codepoints=(123, 52, 8419, 125), name='nummertast fire', slug='nummertast_fire', annotations=frozenset({'4', 'fire', 'nummertast'})),
EmojiAnnotations(emoji='{5⃣}', codepoints=(123, 53, 8419, 125), name='nummertast fem', slug='nummertast_fem', annotations=frozenset({'5', 'fem', 'nummertast'})),
EmojiAnnotations(emoji='{6⃣}', codepoints=(123, 54, 8419, 125), name='nummertast seks', slug='nummertast_seks', annotations=frozenset({'nummertast', 'seks', '6'})),
EmojiAnnotations(emoji='{7⃣}', codepoints=(123, 55, 8419, 125), name='nummertast sju', slug='nummertast_sju', annotations=frozenset({'7', 'nummertast', 'syv'})),
EmojiAnnotations(emoji='{8⃣}', codepoints=(123, 56, 8419, 125), name='nummertast åtte', slug='nummertast_åtte', annotations=frozenset({'8', 'åtte', 'nummertast'})),
EmojiAnnotations(emoji='{9⃣}', codepoints=(123, 57, 8419, 125), name='nummertast ni', slug='nummertast_ni', annotations=frozenset({'nummertast', 'ni', '9'})),
EmojiAnnotations(emoji='🔟', codepoints=(128287,), name='nummertast ti', slug='nummertast_ti', annotations=frozenset({'10', 'nummertast', 'ti'})),
EmojiAnnotations(emoji='💯', codepoints=(128175,), name='hundre', slug='hundre', annotations=frozenset({'poeng', 'full pott', '100 poeng'})),
EmojiAnnotations(emoji='🔞', codepoints=(128286,), name='ingen under 18', slug='ingen_under_18', annotations=frozenset({'18', 'under 18', 'aldersgrense 18 år'})),
EmojiAnnotations(emoji='🔠', codepoints=(128288,), name='store bokstaver', slug='store_bokstaver', annotations=frozenset({'inndata'})),
EmojiAnnotations(emoji='🔡', codepoints=(128289,), name='små bokstaver', slug='små_bokstaver', annotations=frozenset({'inndata'})),
EmojiAnnotations(emoji='🔢', codepoints=(128290,), name='tall', slug='tall', annotations=frozenset({'inndata', 'numeriske data'})),
EmojiAnnotations(emoji='🔣', codepoints=(128291,), name='symboler', slug='symboler', annotations=frozenset({'inndata'})),
EmojiAnnotations(emoji='🔤', codepoints=(128292,), name='latinsk alfabet', slug='latinsk_alfabet', annotations=frozenset({'inndata', 'ABC'})),
EmojiAnnotations(emoji='🅰', codepoints=(127344,), name='blodtype A', slug='blodtype_a', annotations=frozenset({'blod', 'type A', 'blodtype', 'A'})),
EmojiAnnotations(emoji='🆎', codepoints=(127374,), name='blodtype AB', slug='blodtype_ab', annotations=frozenset({'type AB', 'blod', 'blodtype', 'AB'})),
EmojiAnnotations(emoji='🅱', codepoints=(127345,), name='blodtype B', slug='blodtype_b', annotations=frozenset({'blod', 'type B', 'B', 'blodtype'})),
EmojiAnnotations(emoji='🆑', codepoints=(127377,), name='CL', slug='cl', annotations=frozenset({'tøm-symbol', 'clear-symbol'})),
EmojiAnnotations(emoji='🆒', codepoints=(127378,), name='cool', slug='cool', annotations=frozenset({'cool-symbol'})),
EmojiAnnotations(emoji='🆓', codepoints=(127379,), name='gratis-symbol', slug='gratis_symbol', annotations=frozenset({'free', 'free-symbol'})),
EmojiAnnotations(emoji='ℹ', codepoints=(8505,), name='informasjonskilde', slug='informasjonskilde', annotations=frozenset({'i', 'informasjon'})),
EmojiAnnotations(emoji='🆔', codepoints=(127380,), name='ID', slug='id', annotations=frozenset({'identitet', 'ID-symbol'})),
EmojiAnnotations(emoji='Ⓜ', codepoints=(9410,), name='bokstaven m med ring rundt', slug='bokstaven_m_med_ring_rundt', annotations=frozenset({'m', 'sirkel'})),
EmojiAnnotations(emoji='🆕', codepoints=(127381,), name='nytt-symbol', slug='nytt_symbol', annotations=frozenset({'new-symbol', 'new'})),
EmojiAnnotations(emoji='🆖', codepoints=(127382,), name='No good', slug='no_good', annotations=frozenset({'ikke bra', 'NG-symbol', 'NG'})),
EmojiAnnotations(emoji='🅾', codepoints=(127358,), name='blodtype O', slug='blodtype_o', annotations=frozenset({'blod', 'type O', 'O', 'blodtype'})),
EmojiAnnotations(emoji='🆗', codepoints=(127383,), name='OK', slug='ok', annotations=frozenset({'OK-symbol'})),
EmojiAnnotations(emoji='🅿', codepoints=(127359,), name='parkering', slug='parkering', annotations=frozenset({'parkeringsplass'})),
EmojiAnnotations(emoji='🆘', codepoints=(127384,), name='SOS', slug='sos', annotations=frozenset({'SOS-symbol', 'hjelp', 'hjelp-symbol'})),
EmojiAnnotations(emoji='🆙', codepoints=(127385,), name='UP!', slug='up', annotations=frozenset({'UP!-symbol', 'opp-symbol', 'opp'})),
EmojiAnnotations(emoji='🆚', codepoints=(127386,), name='versus', slug='versus', annotations=frozenset({'vs', 'vs-symbol'})),
EmojiAnnotations(emoji='🈁', codepoints=(127489,), name='katakana som leses «koko»', slug='katakana_som_leses_«koko»', annotations=frozenset({'«koko»', 'katakana'})),
EmojiAnnotations(emoji='🈂', codepoints=(127490,), name='katakana som leses «sa»', slug='katakana_som_leses_«sa»', annotations=frozenset({'«sa»', 'stavelse', 'katakana'})),
EmojiAnnotations(emoji='🈷', codepoints=(127543,), name='ideograf for «måne»', slug='ideograf_for_«måne»', annotations=frozenset({'ideograf', 'måne'})),
EmojiAnnotations(emoji='🈶', codepoints=(127542,), name='ideograf for «har»', slug='ideograf_for_«har»', annotations=frozenset({'å ha', 'ideograf', 'være'})),
EmojiAnnotations(emoji='🈯', codepoints=(127535,), name='ideograf for «tildelt»', slug='ideograf_for_«tildelt»', annotations=frozenset({'ideograf', 'finger', 'reservert', 'utpekt'})),
EmojiAnnotations(emoji='🉐', codepoints=(127568,), name='ideograf for «spesialtilbud»', slug='ideograf_for_«spesialtilbud»', annotations=frozenset({'ideograf', 'spesialtilbud'})),
EmojiAnnotations(emoji='🈹', codepoints=(127545,), name='ideograf for «rabatt»', slug='ideograf_for_«rabatt»', annotations=frozenset({'dele', 'ideograf', 'rabatt'})),
EmojiAnnotations(emoji='🈚', codepoints=(127514,), name='ideograf for «har ikke»', slug='ideograf_for_«har_ikke»', annotations=frozenset({'ideograf', 'har ikke', 'er ikke'})),
EmojiAnnotations(emoji='🈲', codepoints=(127538,), name='ideograf for forbudt', slug='ideograf_for_forbudt', annotations=frozenset({'ideograf', 'forbudt'})),
EmojiAnnotations(emoji='🉑', codepoints=(127569,), name='ideograf for «kan» eller «bestått»', slug='ideograf_for_«kan»_eller_«bestått»', annotations=frozenset({'ideograf', 'bestått', 'kan'})),
EmojiAnnotations(emoji='🈸', codepoints=(127544,), name='ideograf for «melde seg på»', slug='ideograf_for_«melde_seg_på»', annotations=frozenset({'ideograf', 'melde seg på'})),
EmojiAnnotations(emoji='🈴', codepoints=(127540,), name='ideograf for «sammen»', slug='ideograf_for_«sammen»', annotations=frozenset({'ideograf', 'sammen'})),
EmojiAnnotations(emoji='🈳', codepoints=(127539,), name='ideograf for «ledig» eller «tom»', slug='ideograf_for_«ledig»_eller_«tom»', annotations=frozenset({'tom', 'ideograf', 'ledig'})),
EmojiAnnotations(emoji='㊗', codepoints=(12951,), name='sirkulær begrepsrune for gratulasjon', slug='sirkulær_begrepsrune_for_gratulasjon', annotations=frozenset({'kinesisk', 'gratulasjon', 'ideografi'})),
EmojiAnnotations(emoji='㊙', codepoints=(12953,), name='sirkulær begrepsrune for hemmelighet', slug='sirkulær_begrepsrune_for_hemmelighet', annotations=frozenset({'kinesisk', 'hemmelig', 'ideografi'})),
EmojiAnnotations(emoji='🈺', codepoints=(127546,), name='ideograf for «åpen for forretninger»', slug='ideograf_for_«åpen_for_forretninger»', annotations=frozenset({'ideograf', 'åpen', 'driver forretninger'})),
EmojiAnnotations(emoji='🈵', codepoints=(127541,), name='ideograf for «full»', slug='ideograf_for_«full»', annotations=frozenset({'ideograf', 'full'})),
EmojiAnnotations(emoji='▪', codepoints=(9642,), name='liten, svart firkant', slug='liten,_svart_firkant', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='▫', codepoints=(9643,), name='liten, hvit firkant', slug='liten,_hvit_firkant', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='◻', codepoints=(9723,), name='hvit firkant av middels størrelse', slug='hvit_firkant_av_middels_størrelse', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='◼', codepoints=(9724,), name='svart firkant av middels størrelse', slug='svart_firkant_av_middels_størrelse', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='◽', codepoints=(9725,), name='hvit firkant av middels/liten størrelse', slug='hvit_firkant_av_middels/liten_størrelse', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='◾', codepoints=(9726,), name='svart firkant av middels/liten størrelse', slug='svart_firkant_av_middels/liten_størrelse', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='⬛', codepoints=(11035,), name='stor, svart firkant', slug='stor,_svart_firkant', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='⬜', codepoints=(11036,), name='stor, hvit firkant', slug='stor,_hvit_firkant', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='🔶', codepoints=(128310,), name='stort oransje rutersymbol', slug='stort_oransje_rutersymbol', annotations=frozenset({'rutersymbol', 'oransje'})),
EmojiAnnotations(emoji='🔷', codepoints=(128311,), name='stort blått rutersymbol', slug='stort_blått_rutersymbol', annotations=frozenset({'rutersymbol', 'blått'})),
EmojiAnnotations(emoji='🔸', codepoints=(128312,), name='lite oransje rutersymbol', slug='lite_oransje_rutersymbol', annotations=frozenset({'rutersymbol', 'oransje'})),
EmojiAnnotations(emoji='🔹', codepoints=(128313,), name='lite blått rutersymbol', slug='lite_blått_rutersymbol', annotations=frozenset({'rutersymbol', 'blått'})),
EmojiAnnotations(emoji='🔺', codepoints=(128314,), name='rød trekant som peker opp', slug='rød_trekant_som_peker_opp', annotations=frozenset({'oppover'})),
EmojiAnnotations(emoji='🔻', codepoints=(128315,), name='rød trekant som peker ned', slug='rød_trekant_som_peker_ned', annotations=frozenset({'nedover'})),
EmojiAnnotations(emoji='💠', codepoints=(128160,), name='diamantform', slug='diamantform', annotations=frozenset({'søtt', 'blomst', 'diamant'})),
EmojiAnnotations(emoji='🔘', codepoints=(128280,), name='radioknapp', slug='radioknapp', annotations=frozenset({'knapp', 'radio'})),
EmojiAnnotations(emoji='🔲', codepoints=(128306,), name='svart, firkantet knapp', slug='svart,_firkantet_knapp', annotations=frozenset({'knapp', 'firkant'})),
EmojiAnnotations(emoji='🔳', codepoints=(128307,), name='hvit, firkantet knapp', slug='hvit,_firkantet_knapp', annotations=frozenset({'knapp', 'firkant'})),
EmojiAnnotations(emoji='⚪', codepoints=(9898,), name='hvit sirkel', slug='hvit_sirkel', annotations=frozenset({'geometrisk', 'sirkel'})),
EmojiAnnotations(emoji='⚫', codepoints=(9899,), name='svart sirkel', slug='svart_sirkel', annotations=frozenset({'geometrisk', 'sirkel'})),
EmojiAnnotations(emoji='🔴', codepoints=(128308,), name='stor, rød sirkel', slug='stor,_rød_sirkel', annotations=frozenset({'stor rød sirkel', 'rød', 'sirkel'})),
EmojiAnnotations(emoji='🔵', codepoints=(128309,), name='stor, blå sirkel', slug='stor,_blå_sirkel', annotations=frozenset({'stor blå sirkel', 'blå', 'sirkel'})),] | 146.772592 | 272 | 0.712882 | from emojitations.emojitypes import EmojiAnnotations
emoji = [
EmojiAnnotations(emoji='😀', codepoints=(128512,), name='glisefjes', slug='glisefjes', annotations=frozenset({'fjes', 'lol', 'latter', 'morsomt', 'smilende fjes'})),
EmojiAnnotations(emoji='😁', codepoints=(128513,), name='glisefjes med smilende øyne', slug='glisefjes_med_smilende_øyne', annotations=frozenset({'lol', 'morsomt', 'latter', 'smil', 'smilende øyne', 'smilende fjes', 'fjes', 'smilefjes'})),
EmojiAnnotations(emoji='😂', codepoints=(128514,), name='gledestårer', slug='gledestårer', annotations=frozenset({'fjes med gledestårer', 'tårer', 'fjes', 'glede'})),
EmojiAnnotations(emoji='😃', codepoints=(128515,), name='smilefjes med åpen munn', slug='smilefjes_med_åpen_munn', annotations=frozenset({'lol', 'morsomt', 'smil', 'latter', 'smilende fjes med tenner', 'fjes', 'smilefjes'})),
EmojiAnnotations(emoji='😄', codepoints=(128516,), name='smilefjes med åpen munn og smilende øyne', slug='smilefjes_med_åpen_munn_og_smilende_øyne', annotations=frozenset({'lol', 'morsomt', 'latter', 'smil', 'smilende øyne', 'med tenner', 'smilende fjes', 'smilefjes'})),
EmojiAnnotations(emoji='😅', codepoints=(128517,), name='svett smil', slug='svett_smil', annotations=frozenset({'kaldsvette', 'smil med svette'})),
EmojiAnnotations(emoji='😆', codepoints=(128518,), name='smil med åpen munn', slug='smil_med_åpen_munn', annotations=frozenset({'smilefjes med åpen munn', 'smil med tennene', 'åpen munn', 'fjes'})),
EmojiAnnotations(emoji='😉', codepoints=(128521,), name='blunkende ansikt', slug='blunkende_ansikt', annotations=frozenset({'blunk', 'fjes'})),
EmojiAnnotations(emoji='😊', codepoints=(128522,), name='smil', slug='smil', annotations=frozenset({'smilefjes', 'fjes', 'smilende øyne', 'smilende fjes og øyne'})),
EmojiAnnotations(emoji='😋', codepoints=(128523,), name='nam-nam-fjes', slug='nam_nam_fjes', annotations=frozenset({'deilig mat', 'fjes', 'nam', 'mmm', 'smilefjes'})),
EmojiAnnotations(emoji='😎', codepoints=(128526,), name='smilefjes med solbriller', slug='smilefjes_med_solbriller', annotations=frozenset({'fjes', 'smil', 'solbriller'})),
EmojiAnnotations(emoji='😍', codepoints=(128525,), name='smilefjes med hjerteøyne', slug='smilefjes_med_hjerteøyne', annotations=frozenset({'hjerteøyne', 'øyne', 'fjes', 'kjærlighet', 'hjerte'})),
EmojiAnnotations(emoji='😘', codepoints=(128536,), name='slengkyss', slug='slengkyss', annotations=frozenset({'ansikt slenger kyss', 'ansikt'})),
EmojiAnnotations(emoji='😗', codepoints=(128535,), name='kyssefjes', slug='kyssefjes', annotations=frozenset({'fjes', 'kyssende fjes', 'kyss'})),
EmojiAnnotations(emoji='😙', codepoints=(128537,), name='kyss og smil', slug='kyss_og_smil', annotations=frozenset({'fjes', 'kyssende fjes med smil', 'smil', 'smilende øyne'})),
EmojiAnnotations(emoji='😚', codepoints=(128538,), name='kyssende', slug='kyssende', annotations=frozenset({'kyss med lukkede øyne', 'fjes', 'øynene lukket'})),
EmojiAnnotations(emoji='☺', codepoints=(9786,), name='smilefjes', slug='smilefjes', annotations=frozenset({'smil', 'ansikt'})),
EmojiAnnotations(emoji='\U0001f642', codepoints=(128578,), name='fjes som smiler litt', slug='fjes_som_smiler_litt', annotations=frozenset({'fjes', 'smil'})),
EmojiAnnotations(emoji='\U0001f917', codepoints=(129303,), name='fjes som gir klem', slug='fjes_som_gir_klem', annotations=frozenset({'fjes', 'klem'})),
EmojiAnnotations(emoji='😇', codepoints=(128519,), name='smil med glorie', slug='smil_med_glorie', annotations=frozenset({'fjes', 'glorie', 'smil'})),
EmojiAnnotations(emoji='\U0001f914', codepoints=(129300,), name='tenkefjes', slug='tenkefjes', annotations=frozenset({'fjes', 'tenker'})),
EmojiAnnotations(emoji='😐', codepoints=(128528,), name='uttrykksløst fjes', slug='uttrykksløst_fjes', annotations=frozenset({'nøytral', 'stum', 'fjes', 'ingen kommentar'})),
EmojiAnnotations(emoji='😑', codepoints=(128529,), name='uttrykksløs', slug='uttrykksløs', annotations=frozenset({'nøytral', 'stum', 'fjes', 'ingen kommentar'})),
EmojiAnnotations(emoji='😶', codepoints=(128566,), name='fjes uten munn', slug='fjes_uten_munn', annotations=frozenset({'ingen munn', 'fjes'})),
EmojiAnnotations(emoji='\U0001f644', codepoints=(128580,), name='fjes med rullende øyne', slug='fjes_med_rullende_øyne', annotations=frozenset({'rullende', 'øyne', 'fjes'})),
EmojiAnnotations(emoji='😏', codepoints=(128527,), name='skuffet smil', slug='skuffet_smil', annotations=frozenset({'halvsmil', 'lurt smil', 'fjes'})),
EmojiAnnotations(emoji='😣', codepoints=(128547,), name='bommert', slug='bommert', annotations=frozenset({'sta', 'utholdende', 'blemme'})),
EmojiAnnotations(emoji='😥', codepoints=(128549,), name='puh!', slug='puh', annotations=frozenset({'lettet', 'skuffet men lettet', 'fjes', 'skuffet'})),
EmojiAnnotations(emoji='😮', codepoints=(128558,), name='forbauset', slug='forbauset', annotations=frozenset({'åpen munn', 'fjes'})),
EmojiAnnotations(emoji='\U0001f910', codepoints=(129296,), name='fjes med glidelåsmunn', slug='fjes_med_glidelåsmunn', annotations=frozenset({'fjes', 'glidelås', 'munn'})),
EmojiAnnotations(emoji='😯', codepoints=(128559,), name='forbauset med øyebryn', slug='forbauset_med_øyebryn', annotations=frozenset({'stum', 'fjes', 'overrasket'})),
EmojiAnnotations(emoji='😪', codepoints=(128554,), name='søvnig fjes', slug='søvnig_fjes', annotations=frozenset({'søvnig', 'fjes'})),
EmojiAnnotations(emoji='😫', codepoints=(128555,), name='trøtt fjes', slug='trøtt_fjes', annotations=frozenset({'søvnig', 'fjes'})),
EmojiAnnotations(emoji='😴', codepoints=(128564,), name='sovende', slug='sovende', annotations=frozenset({'sovende fjes', 'fjes'})),
EmojiAnnotations(emoji='😌', codepoints=(128524,), name='lettet-fjes', slug='lettet_fjes', annotations=frozenset({'lettet', 'fjes'})),
EmojiAnnotations(emoji='\U0001f913', codepoints=(129299,), name='nerdefjes', slug='nerdefjes', annotations=frozenset({'nerd', 'fjes'})),
EmojiAnnotations(emoji='😛', codepoints=(128539,), name='rekke tunge', slug='rekke_tunge', annotations=frozenset({'tungen ut', 'fjes', 'fjes som rekker tunge'})),
EmojiAnnotations(emoji='😜', codepoints=(128540,), name='spøkefullt blunkefjes som rekker tunge', slug='spøkefullt_blunkefjes_som_rekker_tunge', annotations=frozenset({'blunk', 'spøk', 'fjes', 'erte', 'rekke tunge med blunk'})),
EmojiAnnotations(emoji='😝', codepoints=(128541,), name='æsj', slug='æsj', annotations=frozenset({'følt', 'rekke tunge', 'lukkende øyne', 'tungen ut med øynene lukket', 'fjes'})),
EmojiAnnotations(emoji='☹', codepoints=(9785,), name='trist fjes', slug='trist_fjes', annotations=frozenset({'fjes', 'trist'})),
EmojiAnnotations(emoji='\U0001f641', codepoints=(128577,), name='litt trist fjes', slug='litt_trist_fjes', annotations=frozenset({'fjes', 'trist'})),
EmojiAnnotations(emoji='😒', codepoints=(128530,), name='misfornøyd', slug='misfornøyd', annotations=frozenset({'sur', 'ansikt', 'misfornøyd ansikt'})),
EmojiAnnotations(emoji='😓', codepoints=(128531,), name='svette', slug='svette', annotations=frozenset({'kaldsvette', 'ansikt'})),
EmojiAnnotations(emoji='😔', codepoints=(128532,), name='tankefull', slug='tankefull', annotations=frozenset({'alvorlig', 'ansikt'})),
EmojiAnnotations(emoji='😕', codepoints=(128533,), name='forvirret fjes', slug='forvirret_fjes', annotations=frozenset({'forvirret', 'fjes'})),
EmojiAnnotations(emoji='😖', codepoints=(128534,), name='sint', slug='sint', annotations=frozenset({'ansikt', 'forbannet'})),
EmojiAnnotations(emoji='\U0001f643', codepoints=(128579,), name='fjes som er opp-ned', slug='fjes_som_er_opp_ned', annotations=frozenset({'opp-ned', 'fjes'})),
EmojiAnnotations(emoji='😷', codepoints=(128567,), name='fjes med maske', slug='fjes_med_maske', annotations=frozenset({'bakteriemaske', 'sykdom', 'fjes'})),
EmojiAnnotations(emoji='\U0001f912', codepoints=(129298,), name='fjes med termometer', slug='fjes_med_termometer', annotations=frozenset({'syk', 'fjes', 'termometer'})),
EmojiAnnotations(emoji='\U0001f915', codepoints=(129301,), name='bandasjert hode', slug='bandasjert_hode', annotations=frozenset({'bandasje', 'fjes', 'skadet', 'skade'})),
EmojiAnnotations(emoji='\U0001f911', codepoints=(129297,), name='pengegal', slug='pengegal', annotations=frozenset({'munn', 'fjes', 'penger'})),
EmojiAnnotations(emoji='😲', codepoints=(128562,), name='sjokkert', slug='sjokkert', annotations=frozenset({'stum', 'fjes', 'overrasket'})),
EmojiAnnotations(emoji='😞', codepoints=(128542,), name='skuffet', slug='skuffet', annotations=frozenset({'skuffelse', 'skuffet fjes', 'fjes'})),
EmojiAnnotations(emoji='😟', codepoints=(128543,), name='bekymret', slug='bekymret', annotations=frozenset({'bekymret fjes', 'ansikt'})),
EmojiAnnotations(emoji='😤', codepoints=(128548,), name='seirende', slug='seirende', annotations=frozenset({'triumf', 'triumferende', 'jeg vant', 'fjes'})),
EmojiAnnotations(emoji='😢', codepoints=(128546,), name='gråtende', slug='gråtende', annotations=frozenset({'tårer', 'ansikt', 'gråtende ansikt', 'trist'})),
EmojiAnnotations(emoji='😭', codepoints=(128557,), name='hylgråting', slug='hylgråting', annotations=frozenset({'tårer', 'fjes', 'gråt'})),
EmojiAnnotations(emoji='😦', codepoints=(128550,), name='oppgitt', slug='oppgitt', annotations=frozenset({'sjokkert med åpen munn', 'overrasket', 'ansikt'})),
EmojiAnnotations(emoji='😧', codepoints=(128551,), name='pine', slug='pine', annotations=frozenset({'smerte', 'lider', 'ansikt'})),
EmojiAnnotations(emoji='😨', codepoints=(128552,), name='redd', slug='redd', annotations=frozenset({'skremt', 'fjes', 'redd ansikt'})),
EmojiAnnotations(emoji='😩', codepoints=(128553,), name='sliten', slug='sliten', annotations=frozenset({'fjes', 'trøtt ansikt', 'trøtt'})),
EmojiAnnotations(emoji='😬', codepoints=(128556,), name='grimase', slug='grimase', annotations=frozenset({'fjes'})),
EmojiAnnotations(emoji='😰', codepoints=(128560,), name='svettende', slug='svettende', annotations=frozenset({'kaldsvette', 'syk', 'åpen munn', 'fjes'})),
EmojiAnnotations(emoji='😱', codepoints=(128561,), name='livredd', slug='livredd', annotations=frozenset({'skremt', 'skrikende', 'fjes', 'redd ansikt'})),
EmojiAnnotations(emoji='😳', codepoints=(128563,), name='rødmer', slug='rødmer', annotations=frozenset({'rødt fjes'})),
EmojiAnnotations(emoji='😵', codepoints=(128565,), name='ør', slug='ør', annotations=frozenset({'fjes'})),
EmojiAnnotations(emoji='😡', codepoints=(128545,), name='rasende', slug='rasende', annotations=frozenset({'sint', 'sur', 'rødt ansikt', 'ansikt', 'rødt'})),
EmojiAnnotations(emoji='😠', codepoints=(128544,), name='sint fjes', slug='sint_fjes', annotations=frozenset({'sint ansikt', 'ansikt', 'forbannet'})),
EmojiAnnotations(emoji='😈', codepoints=(128520,), name='smilende djevel', slug='smilende_djevel', annotations=frozenset({'smil med horn', 'horn', 'fjes', 'smil'})),
EmojiAnnotations(emoji='👿', codepoints=(128127,), name='sint djevel', slug='sint_djevel', annotations=frozenset({'liten djevel', 'ansikt'})),
EmojiAnnotations(emoji='👹', codepoints=(128121,), name='troll', slug='troll', annotations=frozenset({'japansk ogre', 'eventyr', 'ogre'})),
EmojiAnnotations(emoji='👺', codepoints=(128122,), name='japansk skapning (tengu)', slug='japansk_skapning_(tengu)', annotations=frozenset({'ansikt', 'japansk troll', 'eventyr'})),
EmojiAnnotations(emoji='💀', codepoints=(128128,), name='hodeskalle', slug='hodeskalle', annotations=frozenset({'kropp', 'eventyr', 'ansikt'})),
EmojiAnnotations(emoji='☠', codepoints=(9760,), name='hodeskalle med korslagte dødningebein', slug='hodeskalle_med_korslagte_dødningebein', annotations=frozenset({'hodeskalle', 'korslagte dødningebein', 'monster', 'kropp', 'død', 'fjes'})),
EmojiAnnotations(emoji='👻', codepoints=(128123,), name='spøkelse', slug='spøkelse', annotations=frozenset({'eventyr'})),
EmojiAnnotations(emoji='👽', codepoints=(128125,), name='romvesen', slug='romvesen', annotations=frozenset({'utenomjordisk', 'sci-fi', 'UFO', 'ansikt'})),
EmojiAnnotations(emoji='👾', codepoints=(128126,), name='rommonster', slug='rommonster', annotations=frozenset({'utenomjordisk', 'sci-fi', 'UFO', 'ansikt'})),
EmojiAnnotations(emoji='\U0001f916', codepoints=(129302,), name='robotfjes', slug='robotfjes', annotations=frozenset({'robot', 'monster', 'fjes'})),
EmojiAnnotations(emoji='💩', codepoints=(128169,), name='bæsj', slug='bæsj', annotations=frozenset({'hundebæsj', 'bæsjeklatt', 'kuruke'})),
EmojiAnnotations(emoji='😺', codepoints=(128570,), name='smilende katt med åpen munn', slug='smilende_katt_med_åpen_munn', annotations=frozenset({'katt', 'kattefjes', 'åpen munn', 'smil'})),
EmojiAnnotations(emoji='😸', codepoints=(128568,), name='flirende katt med smilende øyne', slug='flirende_katt_med_smilende_øyne', annotations=frozenset({'katt', 'kattefjes', 'smil'})),
EmojiAnnotations(emoji='😹', codepoints=(128569,), name='kattefjes med gledestårer', slug='kattefjes_med_gledestårer', annotations=frozenset({'kattefjes', 'katt', 'fjes'})),
EmojiAnnotations(emoji='😻', codepoints=(128571,), name='smilende katt med hjerteøyne', slug='smilende_katt_med_hjerteøyne', annotations=frozenset({'smilende kattefjes med hjerteøyne', 'kattefjes', 'øyne', 'forelsket', 'kjærlighet', 'katt', 'fjes', 'hjerte'})),
EmojiAnnotations(emoji='😼', codepoints=(128572,), name='katt med ironisk smil', slug='katt_med_ironisk_smil', annotations=frozenset({'ironisk katt'})),
EmojiAnnotations(emoji='😽', codepoints=(128573,), name='kyssende katt', slug='kyssende_katt', annotations=frozenset({'øyne', 'kyssende', 'lukket', 'kyss', 'katt', 'fjes', 'lukkede øyne'})),
EmojiAnnotations(emoji='🙀', codepoints=(128576,), name='sjokkert kattefjes', slug='sjokkert_kattefjes', annotations=frozenset({'katt', 'jøsses', 'fjes', 'overrasket'})),
EmojiAnnotations(emoji='😿', codepoints=(128575,), name='gråtende katt', slug='gråtende_katt', annotations=frozenset({'katt', 'gråtende kattefjes', 'fjes', 'gråt'})),
EmojiAnnotations(emoji='😾', codepoints=(128574,), name='trutmunn', slug='trutmunn', annotations=frozenset({'kattefjes', 'katt', 'kyssende katt', 'fjes'})),
EmojiAnnotations(emoji='🙈', codepoints=(128584,), name='ser ingenting ondt', slug='ser_ingenting_ondt', annotations=frozenset({'vis ape'})),
EmojiAnnotations(emoji='🙉', codepoints=(128585,), name='hører ingenting ondt', slug='hører_ingenting_ondt', annotations=frozenset({'vis ape'})),
EmojiAnnotations(emoji='🙊', codepoints=(128586,), name='sier ingenting ondt', slug='sier_ingenting_ondt', annotations=frozenset({'vis ape'})),
EmojiAnnotations(emoji='👦', codepoints=(128102,), name='gutt', slug='gutt', annotations=frozenset({'menneske'})),
EmojiAnnotations(emoji='👧', codepoints=(128103,), name='jente', slug='jente', annotations=frozenset({'menneske'})),
EmojiAnnotations(emoji='👨', codepoints=(128104,), name='mann', slug='mann', annotations=frozenset({'menneske'})),
EmojiAnnotations(emoji='👩', codepoints=(128105,), name='kvinne', slug='kvinne', annotations=frozenset({'menneske'})),
EmojiAnnotations(emoji='👴', codepoints=(128116,), name='eldre mann', slug='eldre_mann', annotations=frozenset({'mann', 'eldre', 'menneske'})),
EmojiAnnotations(emoji='👵', codepoints=(128117,), name='eldre kvinne', slug='eldre_kvinne', annotations=frozenset({'kvinne', 'eldre', 'menneske'})),
EmojiAnnotations(emoji='👶', codepoints=(128118,), name='baby', slug='baby', annotations=frozenset({'menneske'})),
EmojiAnnotations(emoji='👱', codepoints=(128113,), name='blond person', slug='blond_person', annotations=frozenset({'blond', 'menneske'})),
EmojiAnnotations(emoji='👮', codepoints=(128110,), name='politimann', slug='politimann', annotations=frozenset({'tjenestemann', 'menneske', 'politi'})),
EmojiAnnotations(emoji='👲', codepoints=(128114,), name='mann med lue', slug='mann_med_lue', annotations=frozenset({'mann med gua pi mao', 'menneske'})),
EmojiAnnotations(emoji='👳', codepoints=(128115,), name='mann med turban', slug='mann_med_turban', annotations=frozenset({'menneske'})),
EmojiAnnotations(emoji='👷', codepoints=(128119,), name='anleggsarbeider', slug='anleggsarbeider', annotations=frozenset({'arbeider', 'menneske', 'anlegg'})),
EmojiAnnotations(emoji='⛑', codepoints=(9937,), name='hjelm med hvitt kors', slug='hjelm_med_hvitt_kors', annotations=frozenset({'hatt', 'hjelm', 'fjes', 'hjelp', 'kors'})),
EmojiAnnotations(emoji='👸', codepoints=(128120,), name='prinsesse', slug='prinsesse', annotations=frozenset({'menneske', 'eventyr'})),
EmojiAnnotations(emoji='💂', codepoints=(128130,), name='gardist', slug='gardist', annotations=frozenset({'vaktmann', 'menneske'})),
EmojiAnnotations(emoji='\U0001f575', codepoints=(128373,), name='detektiv', slug='detektiv', annotations=frozenset({'spion'})),
EmojiAnnotations(emoji='🎅', codepoints=(127877,), name='julenissen', slug='julenissen', annotations=frozenset({'jul', 'feiring', 'nissefar', 'nissen'})),
EmojiAnnotations(emoji='👼', codepoints=(128124,), name='babyengel', slug='babyengel', annotations=frozenset({'ansikt', 'engel', 'eventyr'})),
EmojiAnnotations(emoji='💆', codepoints=(128134,), name='ansiktsmassasje', slug='ansiktsmassasje', annotations=frozenset({'massasje', 'skjønnhetssalong'})),
EmojiAnnotations(emoji='💇', codepoints=(128135,), name='hårklipp', slug='hårklipp', annotations=frozenset({'skjønnhetssalong', 'skjønnhet', 'salong'})),
EmojiAnnotations(emoji='👰', codepoints=(128112,), name='brud', slug='brud', annotations=frozenset({'bryllup', 'mennesker', 'brud med slør', 'slør'})),
EmojiAnnotations(emoji='🙍', codepoints=(128589,), name='rynker brynene', slug='rynker_brynene', annotations=frozenset({'person', 'person rynker brynene'})),
EmojiAnnotations(emoji='🙎', codepoints=(128590,), name='snurpemunn', slug='snurpemunn', annotations=frozenset({'fjes', 'kyssende person'})),
EmojiAnnotations(emoji='🙅', codepoints=(128581,), name='ikke bra', slug='ikke_bra', annotations=frozenset({'nei', 'forbudt'})),
EmojiAnnotations(emoji='🙆', codepoints=(128582,), name='OK-bevegelse', slug='ok_bevegelse', annotations=frozenset({'greit', 'fjes', 'bevegelse'})),
EmojiAnnotations(emoji='💁', codepoints=(128129,), name='informasjonsskranke', slug='informasjonsskranke', annotations=frozenset({'hjelp', 'menneske', 'informasjon'})),
EmojiAnnotations(emoji='🙋', codepoints=(128587,), name='rekker opp hånden', slug='rekker_opp_hånden', annotations=frozenset({'hånd', 'person', 'rekker opp'})),
EmojiAnnotations(emoji='🙇', codepoints=(128583,), name='unnskyld', slug='unnskyld', annotations=frozenset({'beklager', 'legger meg flat', 'menneske'})),
EmojiAnnotations(emoji='🙌', codepoints=(128588,), name='hurra', slug='hurra', annotations=frozenset({'begge hender', 'feirer', 'rekker begge armene i været', 'feiring'})),
EmojiAnnotations(emoji='🙏', codepoints=(128591,), name='foldede hender', slug='foldede_hender', annotations=frozenset({'hånd', 'be', 'bevegelse', 'bønn', 'ber'})),
EmojiAnnotations(emoji='\U0001f5e3', codepoints=(128483,), name='snakkende hode i profil', slug='snakkende_hode_i_profil', annotations=frozenset({'silhuett', 'hode', 'snakker', 'fjes', 'snakke'})),
EmojiAnnotations(emoji='👤', codepoints=(128100,), name='siluett av byste', slug='siluett_av_byste', annotations=frozenset({'byste', 'siluett', 'menneske'})),
EmojiAnnotations(emoji='👥', codepoints=(128101,), name='siluett av byster', slug='siluett_av_byster', annotations=frozenset({'byste', 'siluett', 'menneske'})),
EmojiAnnotations(emoji='🚶', codepoints=(128694,), name='fotgjenger', slug='fotgjenger', annotations=frozenset({'gående'})),
EmojiAnnotations(emoji='🏃', codepoints=(127939,), name='løping', slug='løping', annotations=frozenset({'sport', 'løper', 'marathon', 'jogging', 'jogger'})),
EmojiAnnotations(emoji='👯', codepoints=(128111,), name='kaninører', slug='kaninører', annotations=frozenset({'kaninjente', 'jente', 'kvinne', 'menneske'})),
EmojiAnnotations(emoji='💃', codepoints=(128131,), name='danser', slug='danser', annotations=frozenset({'menneske', 'dans'})),
EmojiAnnotations(emoji='\U0001f574', codepoints=(128372,), name='svevende mann i dress', slug='svevende_mann_i_dress', annotations=frozenset({'forretningsmann', 'mann', 'dress', 'business'})),
EmojiAnnotations(emoji='💏', codepoints=(128143,), name='kyss', slug='kyss', annotations=frozenset({'kyssing', 'romantikk', 'mennesker'})),
EmojiAnnotations(emoji='💑', codepoints=(128145,), name='par', slug='par', annotations=frozenset({'par med hjerte', 'romantikk', 'hjerte', 'mennesker'})),
EmojiAnnotations(emoji='👪', codepoints=(128106,), name='familie', slug='familie', annotations=frozenset({'mor', 'far', 'mennesker', 'barn'})),
EmojiAnnotations(emoji='👫', codepoints=(128107,), name='mann og kvinne', slug='mann_og_kvinne', annotations=frozenset({'kvinne', 'mennsker', 'mann', 'mann og kvinne holder hender', 'holder hender', 'par'})),
EmojiAnnotations(emoji='👬', codepoints=(128108,), name='to menn som leier', slug='to_menn_som_leier', annotations=frozenset({'mann', 'par', 'holde hender', 'to menn holder hender', 'mennesker'})),
EmojiAnnotations(emoji='👭', codepoints=(128109,), name='to kvinner som leier', slug='to_kvinner_som_leier', annotations=frozenset({'kvinne', 'par', 'holde hender', 'to kvinner holder hender', 'mennesker'})),
EmojiAnnotations(emoji='\U0001f3fb', codepoints=(127995,), name='hud type 1–2', slug='hud_type_1–2', annotations=frozenset({'fitzpatrick', 'emojimodifikator', 'hud', 'hudfarge'})),
EmojiAnnotations(emoji='\U0001f3fc', codepoints=(127996,), name='hud type 3', slug='hud_type_3', annotations=frozenset({'fitzpatrick', 'emojimodifikator', 'hud', 'hudfarge'})),
EmojiAnnotations(emoji='\U0001f3fd', codepoints=(127997,), name='hud type 4', slug='hud_type_4', annotations=frozenset({'fitzpatrick', 'emojimodifikator', 'hud', 'hudfarge'})),
EmojiAnnotations(emoji='\U0001f3fe', codepoints=(127998,), name='hud type 5', slug='hud_type_5', annotations=frozenset({'fitzpatrick', 'emojimodifikator', 'hud', 'hudfarge'})),
EmojiAnnotations(emoji='\U0001f3ff', codepoints=(127999,), name='hud type 6', slug='hud_type_6', annotations=frozenset({'fitzpatrick', 'emojimodifikator', 'hud', 'hudfarge'})),
EmojiAnnotations(emoji='💪', codepoints=(128170,), name='biceps', slug='biceps', annotations=frozenset({'muskelfleksing', 'tegneserie', 'muskel', 'sterk', 'fleks'})),
EmojiAnnotations(emoji='👈', codepoints=(128072,), name='finger som peker mot venstre', slug='finger_som_peker_mot_venstre', annotations=frozenset({'peker venstre', 'finger', 'pekende finger'})),
EmojiAnnotations(emoji='👉', codepoints=(128073,), name='finger som peker mot høyre', slug='finger_som_peker_mot_høyre', annotations=frozenset({'peker opp', 'finger', 'pekende finger'})),
EmojiAnnotations(emoji='☝', codepoints=(9757,), name='pekefinger som peker opp', slug='pekefinger_som_peker_opp', annotations=frozenset({'peke', 'finger', 'kropp', 'hånd', 'pekefinger', 'opp'})),
EmojiAnnotations(emoji='👆', codepoints=(128070,), name='oppoverpekende finger', slug='oppoverpekende_finger', annotations=frozenset({'peker opp', 'finger', 'pekende finger'})),
EmojiAnnotations(emoji='\U0001f595', codepoints=(128405,), name='langfinger', slug='langfinger', annotations=frozenset({'hånd', 'finger', 'kropp'})),
EmojiAnnotations(emoji='👇', codepoints=(128071,), name='nedoverpekende finger', slug='nedoverpekende_finger', annotations=frozenset({'peker ned', 'finger', 'pekende finger'})),
EmojiAnnotations(emoji='✌', codepoints=(9996,), name='seierstegn', slug='seierstegn', annotations=frozenset({'peace-tegn'})),
EmojiAnnotations(emoji='\U0001f596', codepoints=(128406,), name='Vulcan-hilsen', slug='vulcan_hilsen', annotations=frozenset({'hånd', 'finger', 'vulcan', 'kropp'})),
EmojiAnnotations(emoji='\U0001f918', codepoints=(129304,), name="rock'n'roll", slug="rock'n'roll", annotations=frozenset({'hånd', 'horn', 'finger', 'kropp'})),
EmojiAnnotations(emoji='\U0001f590', codepoints=(128400,), name='flat hånd med spredte fingre', slug='flat_hånd_med_spredte_fingre', annotations=frozenset({'hånd', 'finger', 'spredt', 'kropp'})),
EmojiAnnotations(emoji='✋', codepoints=(9995,), name='hevet hånd', slug='hevet_hånd', annotations=frozenset({'flat hånd', 'håndflate'})),
EmojiAnnotations(emoji='👌', codepoints=(128076,), name='OK-tegn', slug='ok_tegn', annotations=frozenset({'hånd', 'tegn', 'OK-hånd', 'OK'})),
EmojiAnnotations(emoji='👍', codepoints=(128077,), name='tommel opp', slug='tommel_opp', annotations=frozenset({'hånd', 'tegn', 'tommel'})),
EmojiAnnotations(emoji='👎', codepoints=(128078,), name='tommel ned', slug='tommel_ned', annotations=frozenset({'hånd', 'tegn', 'tommel'})),
EmojiAnnotations(emoji='✊', codepoints=(9994,), name='hevet knyttneve', slug='hevet_knyttneve', annotations=frozenset({'knyttet neve', 'neve'})),
EmojiAnnotations(emoji='👊', codepoints=(128074,), name='knyttet neve', slug='knyttet_neve', annotations=frozenset({'hånd', 'neve'})),
EmojiAnnotations(emoji='👋', codepoints=(128075,), name='vinkede hånd', slug='vinkede_hånd', annotations=frozenset({'vinking', 'vinke'})),
EmojiAnnotations(emoji='👏', codepoints=(128079,), name='klappende hender', slug='klappende_hender', annotations=frozenset({'hender', 'klapping'})),
EmojiAnnotations(emoji='👐', codepoints=(128080,), name='åpne hender', slug='åpne_hender', annotations=frozenset({'hender', 'åpne'})),
EmojiAnnotations(emoji='✍', codepoints=(9997,), name='skrivende hånd', slug='skrivende_hånd', annotations=frozenset({'hånd', 'skrive', 'kropp'})),
EmojiAnnotations(emoji='💅', codepoints=(128133,), name='neglpleie', slug='neglpleie', annotations=frozenset({'negl', 'manikyr', 'kosmetikk', 'neglelakk'})),
EmojiAnnotations(emoji='👂', codepoints=(128066,), name='øre', slug='øre', annotations=frozenset({'kropp'})),
EmojiAnnotations(emoji='👃', codepoints=(128067,), name='nese', slug='nese', annotations=frozenset({'fjes', 'kropp'})),
EmojiAnnotations(emoji='👣', codepoints=(128099,), name='fotavtrykk', slug='fotavtrykk', annotations=frozenset({'fot', 'avtrykk', 'kropp'})),
EmojiAnnotations(emoji='👀', codepoints=(128064,), name='øyne', slug='øyne', annotations=frozenset({'fjes'})),
EmojiAnnotations(emoji='\U0001f441', codepoints=(128065,), name='øye', slug='øye', annotations=frozenset({'kropp'})),
EmojiAnnotations(emoji='👅', codepoints=(128069,), name='tunge', slug='tunge', annotations=frozenset({'fjes', 'kropp'})),
EmojiAnnotations(emoji='👄', codepoints=(128068,), name='munn', slug='munn', annotations=frozenset({'lepper', 'fjes', 'kropp'})),
EmojiAnnotations(emoji='💋', codepoints=(128139,), name='kyssemerke', slug='kyssemerke', annotations=frozenset({'lepper', 'romantikk', 'kyss'})),
EmojiAnnotations(emoji='💘', codepoints=(128152,), name='hjerte med pil', slug='hjerte_med_pil', annotations=frozenset({'pil', 'hjerte', 'kjærlighet', 'følelser'})),
EmojiAnnotations(emoji='❤', codepoints=(10084,), name='rødt hjerte', slug='rødt_hjerte', annotations=frozenset({'hjerte'})),
EmojiAnnotations(emoji='💓', codepoints=(128147,), name='bankende hjerte', slug='bankende_hjerte', annotations=frozenset({'bankende', 'hjerte', 'kjærlighet', 'følelser'})),
EmojiAnnotations(emoji='💔', codepoints=(128148,), name='kjærlighetssorg', slug='kjærlighetssorg', annotations=frozenset({'brudd', 'følelser', 'hjerte', 'knust hjerte'})),
EmojiAnnotations(emoji='💕', codepoints=(128149,), name='to hjerter', slug='to_hjerter', annotations=frozenset({'følelser', 'kjærlighet', 'elskere'})),
EmojiAnnotations(emoji='💖', codepoints=(128150,), name='skinnende hjerte', slug='skinnende_hjerte', annotations=frozenset({'skinnende', 'kjærlighet', 'følelser'})),
EmojiAnnotations(emoji='💗', codepoints=(128151,), name='voksende hjerte', slug='voksende_hjerte', annotations=frozenset({'spent', 'nervøs', 'kjærlighet', 'følelser'})),
EmojiAnnotations(emoji='💙', codepoints=(128153,), name='blått hjerte', slug='blått_hjerte', annotations=frozenset({'blå', 'hjerte', 'følelser'})),
EmojiAnnotations(emoji='💚', codepoints=(128154,), name='grønt hjerte', slug='grønt_hjerte', annotations=frozenset({'grønn', 'hjerte', 'følelser'})),
EmojiAnnotations(emoji='💛', codepoints=(128155,), name='gult hjerte', slug='gult_hjerte', annotations=frozenset({'hjerte', 'gul', 'følelser'})),
EmojiAnnotations(emoji='💜', codepoints=(128156,), name='lilla hjerte', slug='lilla_hjerte', annotations=frozenset({'lilla', 'hjerte', 'følelser'})),
EmojiAnnotations(emoji='💝', codepoints=(128157,), name='hjerte med bånd', slug='hjerte_med_bånd', annotations=frozenset({'bånd', 'hjerte', 'valentine', 'følelser'})),
EmojiAnnotations(emoji='💞', codepoints=(128158,), name='snurrende hjerter', slug='snurrende_hjerter', annotations=frozenset({'søtt', 'snurrende', 'kjærlighet', 'hjerter', 'følelser'})),
EmojiAnnotations(emoji='💟', codepoints=(128159,), name='hjertedekorasjon', slug='hjertedekorasjon', annotations=frozenset({'hjerte'})),
EmojiAnnotations(emoji='❣', codepoints=(10083,), name='utropstegn med hjerte', slug='utropstegn_med_hjerte', annotations=frozenset({'tegn', 'uttropstegn', 'tegnsetting', 'hjerte', 'uttrop'})),
EmojiAnnotations(emoji='💌', codepoints=(128140,), name='kjærlighesbrev', slug='kjærlighesbrev', annotations=frozenset({'brev med hjerte', 'romantikk', 'hjerte', 'brev'})),
EmojiAnnotations(emoji='💤', codepoints=(128164,), name='trøtt', slug='trøtt', annotations=frozenset({'zzz', 'snorker', 'følelse', 'sover', 'tegneserie', 'søving'})),
EmojiAnnotations(emoji='💢', codepoints=(128162,), name='sinne', slug='sinne', annotations=frozenset({'sint', 'tegneserie', 'sinnesymbol', 'følelse'})),
EmojiAnnotations(emoji='💣', codepoints=(128163,), name='bombe', slug='bombe', annotations=frozenset({'tegneserie', 'følelse'})),
EmojiAnnotations(emoji='💥', codepoints=(128165,), name='krasj', slug='krasj', annotations=frozenset({'kollisjon', 'tegneserie', 'smell', 'følelse'})),
EmojiAnnotations(emoji='💦', codepoints=(128166,), name='svettedråper', slug='svettedråper', annotations=frozenset({'tegneserie', 'svette', 'sprut', 'sprutende'})),
EmojiAnnotations(emoji='💨', codepoints=(128168,), name='puff', slug='puff', annotations=frozenset({'tegneserie', 'forsvinne', 'løpe'})),
EmojiAnnotations(emoji='💫', codepoints=(128171,), name='svimmel', slug='svimmel', annotations=frozenset({'tegneserie', 'stjerner', 'se stjerner'})),
EmojiAnnotations(emoji='💬', codepoints=(128172,), name='snakkeboble', slug='snakkeboble', annotations=frozenset({'ballong', 'tegneserie', 'dialog', 'dialogboble', 'boble'})),
EmojiAnnotations(emoji='\U0001f5e8', codepoints=(128488,), name='snakkeboble mot venstre', slug='snakkeboble_mot_venstre', annotations=frozenset({'tale', 'dialog'})),
EmojiAnnotations(emoji='\U0001f5ef', codepoints=(128495,), name='sint snakkeboble mot høyre', slug='sint_snakkeboble_mot_høyre', annotations=frozenset({'sint', 'ballong', 'sur', 'boble'})),
EmojiAnnotations(emoji='💭', codepoints=(128173,), name='tankeboble', slug='tankeboble', annotations=frozenset({'tegneserie', 'tanke', 'boble'})),
EmojiAnnotations(emoji='\U0001f573', codepoints=(128371,), name='hull i bakken', slug='hull_i_bakken', annotations=frozenset({'hull'})),
EmojiAnnotations(emoji='👓', codepoints=(128083,), name='briller', slug='briller', annotations=frozenset({'lesebriller', 'klær'})),
EmojiAnnotations(emoji='\U0001f576', codepoints=(128374,), name='solbriller', slug='solbriller', annotations=frozenset({'mørkt: øye', 'briller'})),
EmojiAnnotations(emoji='👔', codepoints=(128084,), name='slips', slug='slips', annotations=frozenset({'klær'})),
EmojiAnnotations(emoji='👕', codepoints=(128085,), name='t-skjorte', slug='t_skjorte', annotations=frozenset({'skjorte', 'klær'})),
EmojiAnnotations(emoji='👖', codepoints=(128086,), name='bukser', slug='bukser', annotations=frozenset({'jeans', 'klær'})),
EmojiAnnotations(emoji='👗', codepoints=(128087,), name='kjole', slug='kjole', annotations=frozenset({'klær'})),
EmojiAnnotations(emoji='👘', codepoints=(128088,), name='kimono', slug='kimono', annotations=frozenset({'kjole', 'klær'})),
EmojiAnnotations(emoji='👙', codepoints=(128089,), name='bikini', slug='bikini', annotations=frozenset({'badetøy', 'badedrakt', 'klær'})),
EmojiAnnotations(emoji='👚', codepoints=(128090,), name='dameklær', slug='dameklær', annotations=frozenset({'klær'})),
EmojiAnnotations(emoji='👛', codepoints=(128091,), name='portemoné', slug='portemoné', annotations=frozenset({'klær', 'pung'})),
EmojiAnnotations(emoji='👜', codepoints=(128092,), name='håndveske', slug='håndveske', annotations=frozenset({'klær', 'veske'})),
EmojiAnnotations(emoji='👝', codepoints=(128093,), name='toalettveske', slug='toalettveske', annotations=frozenset({'liten veske', 'klær'})),
EmojiAnnotations(emoji='\U0001f6cd', codepoints=(128717,), name='handleposer', slug='handleposer', annotations=frozenset({'shopping', 'hotell', 'pose'})),
EmojiAnnotations(emoji='🎒', codepoints=(127890,), name='skolesekk', slug='skolesekk', annotations=frozenset({'ryggsekk', 'ransel'})),
EmojiAnnotations(emoji='👞', codepoints=(128094,), name='herresko', slug='herresko', annotations=frozenset({'sko', 'klær'})),
EmojiAnnotations(emoji='👟', codepoints=(128095,), name='joggesko', slug='joggesko', annotations=frozenset({'tennissko', 'fritidssko', 'klær'})),
EmojiAnnotations(emoji='👠', codepoints=(128096,), name='høyhælte', slug='høyhælte', annotations=frozenset({'sko', 'klær'})),
EmojiAnnotations(emoji='👡', codepoints=(128097,), name='damesandal', slug='damesandal', annotations=frozenset({'sandal', 'dame', 'klær'})),
EmojiAnnotations(emoji='👢', codepoints=(128098,), name='damestøvletter', slug='damestøvletter', annotations=frozenset({'støvlett', 'dame', 'klær'})),
EmojiAnnotations(emoji='👑', codepoints=(128081,), name='krone', slug='krone', annotations=frozenset({'klær'})),
EmojiAnnotations(emoji='👒', codepoints=(128082,), name='damehatt', slug='damehatt', annotations=frozenset({'klær', 'hatt'})),
EmojiAnnotations(emoji='🎩', codepoints=(127913,), name='flosshatt', slug='flosshatt', annotations=frozenset({'klær', 'hatt'})),
EmojiAnnotations(emoji='🎓', codepoints=(127891,), name='studentlue', slug='studentlue', annotations=frozenset({'feiring', 'uteksaminering', 'hatt'})),
EmojiAnnotations(emoji='\U0001f4ff', codepoints=(128255,), name='bønneperler', slug='bønneperler', annotations=frozenset({'bønn', 'perler', 'religion', 'klær', 'halssmykke'})),
EmojiAnnotations(emoji='💄', codepoints=(128132,), name='leppestift', slug='leppestift', annotations=frozenset({'sminke', 'kosmetikk'})),
EmojiAnnotations(emoji='💍', codepoints=(128141,), name='ring', slug='ring', annotations=frozenset({'romantikk'})),
EmojiAnnotations(emoji='💎', codepoints=(128142,), name='diamant', slug='diamant', annotations=frozenset({'smykkestein', 'edelsten', 'romantikk'})),
EmojiAnnotations(emoji='🐵', codepoints=(128053,), name='apefjes', slug='apefjes', annotations=frozenset({'dyr', 'ape', 'fjes'})),
EmojiAnnotations(emoji='🐒', codepoints=(128018,), name='ape', slug='ape', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐶', codepoints=(128054,), name='hundefjes', slug='hundefjes', annotations=frozenset({'dyr', 'hund', 'fjes'})),
EmojiAnnotations(emoji='🐕', codepoints=(128021,), name='hund', slug='hund', annotations=frozenset({'dyr', 'valp'})),
EmojiAnnotations(emoji='🐩', codepoints=(128041,), name='puddel', slug='puddel', annotations=frozenset({'dyr', 'hund'})),
EmojiAnnotations(emoji='🐺', codepoints=(128058,), name='ulvehode', slug='ulvehode', annotations=frozenset({'hode', 'dyr', 'ulv'})),
EmojiAnnotations(emoji='🐱', codepoints=(128049,), name='kattefjes', slug='kattefjes', annotations=frozenset({'katt', 'dyr', 'fjes'})),
EmojiAnnotations(emoji='🐈', codepoints=(128008,), name='katt', slug='katt', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='\U0001f981', codepoints=(129409,), name='løvefjes', slug='løvefjes', annotations=frozenset({'løven', 'fjes', 'stjernetegn', 'løve'})),
EmojiAnnotations(emoji='🐯', codepoints=(128047,), name='tigerfjes', slug='tigerfjes', annotations=frozenset({'dyr', 'tiger'})),
EmojiAnnotations(emoji='🐅', codepoints=(128005,), name='tiger', slug='tiger', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐆', codepoints=(128006,), name='leopard', slug='leopard', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐴', codepoints=(128052,), name='hestehode', slug='hestehode', annotations=frozenset({'hode', 'dyr', 'hest'})),
EmojiAnnotations(emoji='🐎', codepoints=(128014,), name='hest', slug='hest', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='\U0001f984', codepoints=(129412,), name='enhjørning', slug='enhjørning', annotations=frozenset({'fjes'})),
EmojiAnnotations(emoji='🐮', codepoints=(128046,), name='kufjes', slug='kufjes', annotations=frozenset({'ku', 'dyr', 'fjes'})),
EmojiAnnotations(emoji='🐂', codepoints=(128002,), name='okse', slug='okse', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐃', codepoints=(128003,), name='vannbøffel', slug='vannbøffel', annotations=frozenset({'dyr', 'bøffel'})),
EmojiAnnotations(emoji='🐄', codepoints=(128004,), name='ku', slug='ku', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐷', codepoints=(128055,), name='grisehode', slug='grisehode', annotations=frozenset({'hode', 'gris', 'dyr'})),
EmojiAnnotations(emoji='🐖', codepoints=(128022,), name='gris', slug='gris', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐗', codepoints=(128023,), name='villsvin', slug='villsvin', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐽', codepoints=(128061,), name='grisesnute', slug='grisesnute', annotations=frozenset({'gris', 'dyr', 'snute'})),
EmojiAnnotations(emoji='🐏', codepoints=(128015,), name='bukk', slug='bukk', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐑', codepoints=(128017,), name='sau', slug='sau', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐐', codepoints=(128016,), name='geit', slug='geit', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐪', codepoints=(128042,), name='dromedar', slug='dromedar', annotations=frozenset({'kamel', 'dyr', 'pukkel'})),
EmojiAnnotations(emoji='🐫', codepoints=(128043,), name='kamel', slug='kamel', annotations=frozenset({'dyr', 'pukkel'})),
EmojiAnnotations(emoji='🐘', codepoints=(128024,), name='elefant', slug='elefant', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐭', codepoints=(128045,), name='musefjes', slug='musefjes', annotations=frozenset({'dyr', 'fjes', 'mus'})),
EmojiAnnotations(emoji='🐁', codepoints=(128001,), name='mus', slug='mus', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐀', codepoints=(128000,), name='rotte', slug='rotte', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐹', codepoints=(128057,), name='hamsterfjes', slug='hamsterfjes', annotations=frozenset({'hamster', 'dyr', 'fjes'})),
EmojiAnnotations(emoji='🐰', codepoints=(128048,), name='kaninfjes', slug='kaninfjes', annotations=frozenset({'dyr', 'fjes', 'kanin'})),
EmojiAnnotations(emoji='🐇', codepoints=(128007,), name='kanin', slug='kanin', annotations=frozenset({'dyr', 'hare'})),
EmojiAnnotations(emoji='\U0001f43f', codepoints=(128063,), name='ekorn', slug='ekorn', annotations=frozenset({'dyr', 'jordekorn'})),
EmojiAnnotations(emoji='🐻', codepoints=(128059,), name='bjørnefjes', slug='bjørnefjes', annotations=frozenset({'bjørn', 'fjes', 'dyr'})),
EmojiAnnotations(emoji='🐨', codepoints=(128040,), name='koala', slug='koala', annotations=frozenset({'koalabjørn', 'dyr'})),
EmojiAnnotations(emoji='🐼', codepoints=(128060,), name='pandafjes', slug='pandafjes', annotations=frozenset({'dyr', 'fjes', 'panda'})),
EmojiAnnotations(emoji='🐾', codepoints=(128062,), name='pote', slug='pote', annotations=frozenset({'poteavtrykk', 'dyr', 'avtrykk'})),
EmojiAnnotations(emoji='🐔', codepoints=(128020,), name='høne', slug='høne', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐓', codepoints=(128019,), name='hane', slug='hane', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐣', codepoints=(128035,), name='nyklekket kylling', slug='nyklekket_kylling', annotations=frozenset({'dyr', 'kylling', 'klekking'})),
EmojiAnnotations(emoji='🐤', codepoints=(128036,), name='kylling', slug='kylling', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐥', codepoints=(128037,), name='helfigurs kylling', slug='helfigurs_kylling', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐦', codepoints=(128038,), name='fugl', slug='fugl', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐧', codepoints=(128039,), name='pingvin', slug='pingvin', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='\U0001f54a', codepoints=(128330,), name='fredsdue', slug='fredsdue', annotations=frozenset({'fred', 'fly', 'fugl'})),
EmojiAnnotations(emoji='🐸', codepoints=(128056,), name='froskefjes', slug='froskefjes', annotations=frozenset({'frosk', 'dyr', 'fjes'})),
EmojiAnnotations(emoji='🐊', codepoints=(128010,), name='krokodille', slug='krokodille', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐢', codepoints=(128034,), name='skilpadde', slug='skilpadde', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐍', codepoints=(128013,), name='slange', slug='slange', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐲', codepoints=(128050,), name='dragehode', slug='dragehode', annotations=frozenset({'hode', 'drage', 'dyr'})),
EmojiAnnotations(emoji='🐉', codepoints=(128009,), name='drage', slug='drage', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐳', codepoints=(128051,), name='sprutende hval', slug='sprutende_hval', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐋', codepoints=(128011,), name='hval', slug='hval', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐬', codepoints=(128044,), name='delfin', slug='delfin', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐟', codepoints=(128031,), name='fisk', slug='fisk', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐠', codepoints=(128032,), name='tropisk fisk', slug='tropisk_fisk', annotations=frozenset({'fisk', 'dyr'})),
EmojiAnnotations(emoji='🐡', codepoints=(128033,), name='kulefisk', slug='kulefisk', annotations=frozenset({'fisk', 'dyr'})),
EmojiAnnotations(emoji='🐙', codepoints=(128025,), name='blekksprut', slug='blekksprut', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐚', codepoints=(128026,), name='skjell', slug='skjell', annotations=frozenset({'spiralskjell', 'dyr'})),
EmojiAnnotations(emoji='\U0001f980', codepoints=(129408,), name='krabbe', slug='krabbe', annotations=frozenset({'krepsen', 'stjernetegn'})),
EmojiAnnotations(emoji='🐌', codepoints=(128012,), name='snegle', slug='snegle', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐛', codepoints=(128027,), name='insekt', slug='insekt', annotations=frozenset({'åme', 'dyr', 'larve'})),
EmojiAnnotations(emoji='🐜', codepoints=(128028,), name='maur', slug='maur', annotations=frozenset({'dyr'})),
EmojiAnnotations(emoji='🐝', codepoints=(128029,), name='honningbie', slug='honningbie', annotations=frozenset({'dyr', 'bie'})),
EmojiAnnotations(emoji='🐞', codepoints=(128030,), name='bille', slug='bille', annotations=frozenset({'dyr', 'insekt', 'tordivel'})),
EmojiAnnotations(emoji='\U0001f577', codepoints=(128375,), name='edderkopp', slug='edderkopp', annotations=frozenset({'insekt'})),
EmojiAnnotations(emoji='\U0001f578', codepoints=(128376,), name='spindelvev', slug='spindelvev', annotations=frozenset({'edderkopp', 'nett'})),
EmojiAnnotations(emoji='\U0001f982', codepoints=(129410,), name='skorpion', slug='skorpion', annotations=frozenset({'skorpionen', 'stjernetegn'})),
EmojiAnnotations(emoji='💐', codepoints=(128144,), name='bukett', slug='bukett', annotations=frozenset({'blomster', 'romantikk'})),
EmojiAnnotations(emoji='🌸', codepoints=(127800,), name='kisebærblomst', slug='kisebærblomst', annotations=frozenset({'blomst', 'hvit blomst'})),
EmojiAnnotations(emoji='💮', codepoints=(128174,), name='hvit blomst', slug='hvit_blomst', annotations=frozenset({'blomsterstempel', 'blomst', 'stempel'})),
EmojiAnnotations(emoji='\U0001f3f5', codepoints=(127989,), name='rosett', slug='rosett', annotations=frozenset({'plante'})),
EmojiAnnotations(emoji='🌹', codepoints=(127801,), name='rose', slug='rose', annotations=frozenset({'blomst'})),
EmojiAnnotations(emoji='🌺', codepoints=(127802,), name='hibiskus', slug='hibiskus', annotations=frozenset({'blomst'})),
EmojiAnnotations(emoji='🌻', codepoints=(127803,), name='solsikke', slug='solsikke', annotations=frozenset({'blomst'})),
EmojiAnnotations(emoji='🌼', codepoints=(127804,), name='blomst', slug='blomst', annotations=frozenset({'plante'})),
EmojiAnnotations(emoji='🌷', codepoints=(127799,), name='tulipan', slug='tulipan', annotations=frozenset({'blomst'})),
EmojiAnnotations(emoji='🌱', codepoints=(127793,), name='spire', slug='spire', annotations=frozenset({'ung plante'})),
EmojiAnnotations(emoji='🌲', codepoints=(127794,), name='grantre', slug='grantre', annotations=frozenset({'tre', 'juletre', 'eviggrønt tre', 'gran'})),
EmojiAnnotations(emoji='🌳', codepoints=(127795,), name='løvtre', slug='løvtre', annotations=frozenset({'tre', 'tre med blader'})),
EmojiAnnotations(emoji='🌴', codepoints=(127796,), name='palme', slug='palme', annotations=frozenset({'palmetre'})),
EmojiAnnotations(emoji='🌵', codepoints=(127797,), name='kaktus', slug='kaktus', annotations=frozenset({'plante'})),
EmojiAnnotations(emoji='🌾', codepoints=(127806,), name='risplante', slug='risplante', annotations=frozenset({'strå', 'siv'})),
EmojiAnnotations(emoji='🌿', codepoints=(127807,), name='urteblader', slug='urteblader', annotations=frozenset({'urte'})),
EmojiAnnotations(emoji='☘', codepoints=(9752,), name='trekløver', slug='trekløver', annotations=frozenset({'plante'})),
EmojiAnnotations(emoji='🍀', codepoints=(127808,), name='firkløver', slug='firkløver', annotations=frozenset({'hell', 'kløver', 'flaks'})),
EmojiAnnotations(emoji='🍁', codepoints=(127809,), name='lønneblad', slug='lønneblad', annotations=frozenset({'høstblad', 'lønn', 'blad'})),
EmojiAnnotations(emoji='🍂', codepoints=(127810,), name='høstblader', slug='høstblader', annotations=frozenset({'fallende blader', 'løv'})),
EmojiAnnotations(emoji='🍃', codepoints=(127811,), name='blader i vind', slug='blader_i_vind', annotations=frozenset({'vind', 'løv i vind', 'blåst', 'blåser'})),
EmojiAnnotations(emoji='🍇', codepoints=(127815,), name='druer', slug='druer', annotations=frozenset({'drueklase', 'frukt'})),
EmojiAnnotations(emoji='🍈', codepoints=(127816,), name='honningmelon', slug='honningmelon', annotations=frozenset({'melon', 'frukt'})),
EmojiAnnotations(emoji='🍉', codepoints=(127817,), name='vannmelon', slug='vannmelon', annotations=frozenset({'melon', 'frukt'})),
EmojiAnnotations(emoji='🍊', codepoints=(127818,), name='mandarin', slug='mandarin', annotations=frozenset({'klementin', 'appelsin', 'frukt'})),
EmojiAnnotations(emoji='🍋', codepoints=(127819,), name='sitron', slug='sitron', annotations=frozenset({'sitrus', 'frukt'})),
EmojiAnnotations(emoji='🍌', codepoints=(127820,), name='banan', slug='banan', annotations=frozenset({'frukt'})),
EmojiAnnotations(emoji='🍍', codepoints=(127821,), name='ananas', slug='ananas', annotations=frozenset({'frukt'})),
EmojiAnnotations(emoji='🍎', codepoints=(127822,), name='rødt eple', slug='rødt_eple', annotations=frozenset({'eple', 'frukt'})),
EmojiAnnotations(emoji='🍏', codepoints=(127823,), name='grønt eple', slug='grønt_eple', annotations=frozenset({'eple', 'frukt'})),
EmojiAnnotations(emoji='🍐', codepoints=(127824,), name='pære', slug='pære', annotations=frozenset({'frukt'})),
EmojiAnnotations(emoji='🍑', codepoints=(127825,), name='fersken', slug='fersken', annotations=frozenset({'frukt'})),
EmojiAnnotations(emoji='🍒', codepoints=(127826,), name='kirsebær', slug='kirsebær', annotations=frozenset({'moreller', 'bær'})),
EmojiAnnotations(emoji='🍓', codepoints=(127827,), name='jordbær', slug='jordbær', annotations=frozenset({'bær'})),
EmojiAnnotations(emoji='🍅', codepoints=(127813,), name='tomat', slug='tomat', annotations=frozenset({'grønnsak'})),
EmojiAnnotations(emoji='🍆', codepoints=(127814,), name='aubergine', slug='aubergine', annotations=frozenset({'grønnsak'})),
EmojiAnnotations(emoji='🌽', codepoints=(127805,), name='maiskolbe', slug='maiskolbe', annotations=frozenset({'mais'})),
EmojiAnnotations(emoji='\U0001f336', codepoints=(127798,), name='rød chili', slug='rød_chili', annotations=frozenset({'plante', 'chili', 'sterk'})),
EmojiAnnotations(emoji='🍄', codepoints=(127812,), name='sopp', slug='sopp', annotations=frozenset({'fluesopp'})),
EmojiAnnotations(emoji='🌰', codepoints=(127792,), name='kastanje', slug='kastanje', annotations=frozenset({'kastanjenøtt'})),
EmojiAnnotations(emoji='🍞', codepoints=(127838,), name='brød', slug='brød', annotations=frozenset({'restaurant', 'loff'})),
EmojiAnnotations(emoji='\U0001f9c0', codepoints=(129472,), name='ostebit', slug='ostebit', annotations=frozenset({'ost'})),
EmojiAnnotations(emoji='🍖', codepoints=(127830,), name='kjøtt', slug='kjøtt', annotations=frozenset({'restaurant', 'kjøttbein'})),
EmojiAnnotations(emoji='🍗', codepoints=(127831,), name='kyllinglår', slug='kyllinglår', annotations=frozenset({'restaurant', 'fjærkre', 'kyllingklubbe'})),
EmojiAnnotations(emoji='🍔', codepoints=(127828,), name='hamburger', slug='hamburger', annotations=frozenset({'restaurant', 'burger'})),
EmojiAnnotations(emoji='🍟', codepoints=(127839,), name='pommes frites', slug='pommes_frites', annotations=frozenset({'restaurant', 'chips'})),
EmojiAnnotations(emoji='🍕', codepoints=(127829,), name='pizzastykke', slug='pizzastykke', annotations=frozenset({'pizzabit', 'pizza', 'restaurant'})),
EmojiAnnotations(emoji='\U0001f32d', codepoints=(127789,), name='pølse', slug='pølse', annotations=frozenset({'hotdog', 'frankfurterpølse'})),
EmojiAnnotations(emoji='\U0001f32e', codepoints=(127790,), name='taco', slug='taco', annotations=frozenset({'meksikansk'})),
EmojiAnnotations(emoji='\U0001f32f', codepoints=(127791,), name='burrito', slug='burrito', annotations=frozenset({'meksikansk'})),
EmojiAnnotations(emoji='🍲', codepoints=(127858,), name='gryterett', slug='gryterett', annotations=frozenset({'middagsmat', 'middag'})),
EmojiAnnotations(emoji='🍱', codepoints=(127857,), name='lunsjboks', slug='lunsjboks', annotations=frozenset({'matboks', 'bentoboks'})),
EmojiAnnotations(emoji='🍘', codepoints=(127832,), name='riskjeks', slug='riskjeks', annotations=frozenset({'ris'})),
EmojiAnnotations(emoji='🍙', codepoints=(127833,), name='risball', slug='risball', annotations=frozenset({'ris', 'restaurant', 'onigiri'})),
EmojiAnnotations(emoji='🍚', codepoints=(127834,), name='kokt ris', slug='kokt_ris', annotations=frozenset({'ris', 'restaurant'})),
EmojiAnnotations(emoji='🍛', codepoints=(127835,), name='curry og ris', slug='curry_og_ris', annotations=frozenset({'ris', 'restaurant', 'curry'})),
EmojiAnnotations(emoji='🍜', codepoints=(127836,), name='nudler', slug='nudler', annotations=frozenset({'restaurant', 'nudelbolle'})),
EmojiAnnotations(emoji='🍝', codepoints=(127837,), name='spaghetti', slug='spaghetti', annotations=frozenset({'restaurant', 'pasta'})),
EmojiAnnotations(emoji='🍠', codepoints=(127840,), name='bakt potet', slug='bakt_potet', annotations=frozenset({'restaurant', 'potet', 'søtpotet'})),
EmojiAnnotations(emoji='🍢', codepoints=(127842,), name='oden', slug='oden', annotations=frozenset({'restaurant', 'sjømat', 'sjømat på pinne'})),
EmojiAnnotations(emoji='🍣', codepoints=(127843,), name='sushi', slug='sushi', annotations=frozenset({'restaurant'})),
EmojiAnnotations(emoji='🍤', codepoints=(127844,), name='reke', slug='reke', annotations=frozenset({'restaurant', 'scampi', 'kongereke', 'grillet scampi'})),
EmojiAnnotations(emoji='🍥', codepoints=(127845,), name='fiskekake', slug='fiskekake', annotations=frozenset({'restaurant', 'fisk'})),
EmojiAnnotations(emoji='🍡', codepoints=(127841,), name='mochi', slug='mochi', annotations=frozenset({'restaurant', 'riskonfekt'})),
EmojiAnnotations(emoji='🍦', codepoints=(127846,), name='soft-is', slug='soft_is', annotations=frozenset({'is', 'iskrem', 'restaurant'})),
EmojiAnnotations(emoji='🍧', codepoints=(127847,), name='vannis', slug='vannis', annotations=frozenset({'is', 'restaurant', 'shave-ice'})),
EmojiAnnotations(emoji='🍨', codepoints=(127848,), name='iskrem', slug='iskrem', annotations=frozenset({'is', 'restaurant'})),
EmojiAnnotations(emoji='🍩', codepoints=(127849,), name='donut', slug='donut', annotations=frozenset({'smultring', 'restaurant'})),
EmojiAnnotations(emoji='🍪', codepoints=(127850,), name='kjeks', slug='kjeks', annotations=frozenset({'restaurant', 'cookie'})),
EmojiAnnotations(emoji='🎂', codepoints=(127874,), name='bursdagskake', slug='bursdagskake', annotations=frozenset({'kake', 'bursdag', 'feiring'})),
EmojiAnnotations(emoji='🍰', codepoints=(127856,), name='kakestykke', slug='kakestykke', annotations=frozenset({'kake', 'dessert', 'formkake'})),
EmojiAnnotations(emoji='🍫', codepoints=(127851,), name='sjokolade', slug='sjokolade', annotations=frozenset({'sjokoladeplate'})),
EmojiAnnotations(emoji='🍬', codepoints=(127852,), name='godteri', slug='godteri', annotations=frozenset({'karamell', 'søtsaker', 'smågodt', 'sukkertøy'})),
EmojiAnnotations(emoji='🍭', codepoints=(127853,), name='kjærlighet på pinne', slug='kjærlighet_på_pinne', annotations=frozenset({'sukkertøy'})),
EmojiAnnotations(emoji='🍮', codepoints=(127854,), name='pudding', slug='pudding', annotations=frozenset({'karamellpudding', 'dessert'})),
EmojiAnnotations(emoji='🍯', codepoints=(127855,), name='honning', slug='honning', annotations=frozenset({'honningkrukke', 'krukke'})),
EmojiAnnotations(emoji='🍼', codepoints=(127868,), name='tåteflaske', slug='tåteflaske', annotations=frozenset({'drikke', 'babyflaske', 'melk'})),
EmojiAnnotations(emoji='☕', codepoints=(9749,), name='varm drikke', slug='varm_drikke', annotations=frozenset({'te', 'drikke', 'kaffe', 'varm', 'rykende'})),
EmojiAnnotations(emoji='🍵', codepoints=(127861,), name='tekopp', slug='tekopp', annotations=frozenset({'suppe', 'te', 'drikke'})),
EmojiAnnotations(emoji='🍶', codepoints=(127862,), name='sake', slug='sake', annotations=frozenset({'mugge og kopp', 'karaffel', 'flaske', 'drikke'})),
EmojiAnnotations(emoji='\U0001f37e', codepoints=(127870,), name='flaske med sprettkork', slug='flaske_med_sprettkork', annotations=frozenset({'bar', 'kork', 'sprette', 'flaske', 'drikke'})),
EmojiAnnotations(emoji='🍷', codepoints=(127863,), name='vinglass', slug='vinglass', annotations=frozenset({'rødvin', 'vin', 'drikke', 'glass'})),
EmojiAnnotations(emoji='🍸', codepoints=(127864,), name='cocktail', slug='cocktail', annotations=frozenset({'bar', 'cocktaildrink', 'drink', 'cocktailglass'})),
EmojiAnnotations(emoji='🍹', codepoints=(127865,), name='tropisk drink', slug='tropisk_drink', annotations=frozenset({'bar', 'juice', 'drink'})),
EmojiAnnotations(emoji='🍺', codepoints=(127866,), name='øl', slug='øl', annotations=frozenset({'halvliter', 'pils', 'pilsner', 'bar', 'ølkrus', 'ølglass'})),
EmojiAnnotations(emoji='🍻', codepoints=(127867,), name='skål', slug='skål', annotations=frozenset({'halvliter', 'pils', 'pilsner', 'bar', 'ølkrus', 'øl', 'ølglass'})),
EmojiAnnotations(emoji='\U0001f37d', codepoints=(127869,), name='tallerken med kniv og gaffel', slug='tallerken_med_kniv_og_gaffel', annotations=frozenset({'kniv', 'gaffel', 'tallerken', 'matlaging'})),
EmojiAnnotations(emoji='🍴', codepoints=(127860,), name='kniv og gaffel', slug='kniv_og_gaffel', annotations=frozenset({'spisested', 'kniv', 'gaffel'})),
EmojiAnnotations(emoji='🍳', codepoints=(127859,), name='speilegg', slug='speilegg', annotations=frozenset({'stekepanne', 'egg', 'matlaging'})),
EmojiAnnotations(emoji='\U0001f3fa', codepoints=(127994,), name='amforakrukke', slug='amforakrukke', annotations=frozenset({'stjernetegn', 'vannmannen', 'drikke', 'verktøy', 'matlaging', 'våpen', 'krukke'})),
EmojiAnnotations(emoji='🌍', codepoints=(127757,), name='jordklode som viser Europa og Afrika', slug='jordklode_som_viser_europa_og_afrika', annotations=frozenset({'Europa', 'Afrika', 'jorden'})),
EmojiAnnotations(emoji='🌎', codepoints=(127758,), name='jordklode som viser Amerika', slug='jordklode_som_viser_amerika', annotations=frozenset({'USA', 'jorden', 'Amerika'})),
EmojiAnnotations(emoji='🌏', codepoints=(127759,), name='jordklode som viser Asia og Australia', slug='jordklode_som_viser_asia_og_australia', annotations=frozenset({'Asia', 'jorden', 'Australia'})),
EmojiAnnotations(emoji='🌐', codepoints=(127760,), name='jordklode med meridianer', slug='jordklode_med_meridianer', annotations=frozenset({'lengde- og breddegrader'})),
EmojiAnnotations(emoji='\U0001f5fa', codepoints=(128506,), name='verdenskart', slug='verdenskart', annotations=frozenset({'kart', 'verden'})),
EmojiAnnotations(emoji='\U0001f3d4', codepoints=(127956,), name='snødekt fjell', slug='snødekt_fjell', annotations=frozenset({'kaldt', 'fjell', 'snø'})),
EmojiAnnotations(emoji='🌋', codepoints=(127755,), name='vulkan', slug='vulkan', annotations=frozenset({'vulkanutbrudd'})),
EmojiAnnotations(emoji='🗻', codepoints=(128507,), name='Fuji-fjellet', slug='fuji_fjellet', annotations=frozenset({'fjell'})),
EmojiAnnotations(emoji='\U0001f3d6', codepoints=(127958,), name='strand med parasoll', slug='strand_med_parasoll', annotations=frozenset({'parasoll', 'strand'})),
EmojiAnnotations(emoji='\U0001f3dd', codepoints=(127965,), name='øde øy', slug='øde_øy', annotations=frozenset({'øde', 'øy'})),
EmojiAnnotations(emoji='\U0001f3de', codepoints=(127966,), name='nasjonalpark', slug='nasjonalpark', annotations=frozenset({'park'})),
EmojiAnnotations(emoji='\U0001f3db', codepoints=(127963,), name='klassisk bygning', slug='klassisk_bygning', annotations=frozenset({'bygning', 'klassisk'})),
EmojiAnnotations(emoji='\U0001f3d7', codepoints=(127959,), name='bygning under konstruksjon', slug='bygning_under_konstruksjon', annotations=frozenset({'bygning', 'bygging'})),
EmojiAnnotations(emoji='\U0001f3d8', codepoints=(127960,), name='boliger', slug='boliger', annotations=frozenset({'hus', 'bygning'})),
EmojiAnnotations(emoji='\U0001f3d9', codepoints=(127961,), name='bylandskap', slug='bylandskap', annotations=frozenset({'bygning', 'bygging'})),
EmojiAnnotations(emoji='\U0001f3da', codepoints=(127962,), name='falleferdig hus', slug='falleferdig_hus', annotations=frozenset({'hus', 'forlatt', 'falleferdig', 'bygning'})),
EmojiAnnotations(emoji='🏠', codepoints=(127968,), name='hus', slug='hus', annotations=frozenset({'hjem', 'bygning'})),
EmojiAnnotations(emoji='🏡', codepoints=(127969,), name='hjem', slug='hjem', annotations=frozenset({'hus', 'hus med hage', 'bygning'})),
EmojiAnnotations(emoji='⛪', codepoints=(9962,), name='kirke', slug='kirke', annotations=frozenset({'bygning'})),
EmojiAnnotations(emoji='\U0001f54b', codepoints=(128331,), name='Kaba', slug='kaba', annotations=frozenset({'muslim', 'islam', 'religion'})),
EmojiAnnotations(emoji='\U0001f54c', codepoints=(128332,), name='moské', slug='moské', annotations=frozenset({'muslim', 'islam', 'religion'})),
EmojiAnnotations(emoji='\U0001f54d', codepoints=(128333,), name='synagoge', slug='synagoge', annotations=frozenset({'tempel', 'jøde', 'jødisk', 'religion'})),
EmojiAnnotations(emoji='⛩', codepoints=(9961,), name='Shinto-alter', slug='shinto_alter', annotations=frozenset({'alter', 'religion', 'shinto'})),
EmojiAnnotations(emoji='🏢', codepoints=(127970,), name='kontorbygg', slug='kontorbygg', annotations=frozenset({'bygning', 'kontor'})),
EmojiAnnotations(emoji='🏣', codepoints=(127971,), name='japansk postkontor', slug='japansk_postkontor', annotations=frozenset({'bygning', 'post'})),
EmojiAnnotations(emoji='🏤', codepoints=(127972,), name='postkontor', slug='postkontor', annotations=frozenset({'post', 'bygning', 'europeisk postkontor'})),
EmojiAnnotations(emoji='🏥', codepoints=(127973,), name='sykehus', slug='sykehus', annotations=frozenset({'bygning'})),
EmojiAnnotations(emoji='🏦', codepoints=(127974,), name='bank', slug='bank', annotations=frozenset({'bygning'})),
EmojiAnnotations(emoji='🏨', codepoints=(127976,), name='hotell', slug='hotell', annotations=frozenset({'bygning'})),
EmojiAnnotations(emoji='🏩', codepoints=(127977,), name='kjærlighetshotell', slug='kjærlighetshotell', annotations=frozenset({'bygning', 'hotell'})),
EmojiAnnotations(emoji='🏪', codepoints=(127978,), name='dagligvarehandel', slug='dagligvarehandel', annotations=frozenset({'dagligvare', 'bygning', 'butikk'})),
EmojiAnnotations(emoji='🏫', codepoints=(127979,), name='skole', slug='skole', annotations=frozenset({'skolebygning', 'bygning'})),
EmojiAnnotations(emoji='🏬', codepoints=(127980,), name='kjøpesenter', slug='kjøpesenter', annotations=frozenset({'storsenter', 'vareshus', 'butikk'})),
EmojiAnnotations(emoji='🏭', codepoints=(127981,), name='fabrikk', slug='fabrikk', annotations=frozenset({'bygning'})),
EmojiAnnotations(emoji='🏯', codepoints=(127983,), name='japansk slott', slug='japansk_slott', annotations=frozenset({'bygning', 'herskapshus'})),
EmojiAnnotations(emoji='🏰', codepoints=(127984,), name='slott', slug='slott', annotations=frozenset({'borg', 'bygning', 'europeisk slott', 'festning'})),
EmojiAnnotations(emoji='💒', codepoints=(128146,), name='kirkebryllup', slug='kirkebryllup', annotations=frozenset({'bryllup', 'kapell', 'romantikk'})),
EmojiAnnotations(emoji='🗼', codepoints=(128508,), name='Tokyo-tårnet', slug='tokyo_tårnet', annotations=frozenset({'tårn', 'Tokyo'})),
EmojiAnnotations(emoji='🗽', codepoints=(128509,), name='Frihetsgudinnen', slug='frihetsgudinnen', annotations=frozenset({'statue', 'frihet'})),
EmojiAnnotations(emoji='🗾', codepoints=(128510,), name='kart over Japan', slug='kart_over_japan', annotations=frozenset({'kart', 'Japan'})),
EmojiAnnotations(emoji='⛺', codepoints=(9978,), name='telt', slug='telt', annotations=frozenset({'camping'})),
EmojiAnnotations(emoji='🌁', codepoints=(127745,), name='tåkete', slug='tåkete', annotations=frozenset({'tåke', 'vær'})),
EmojiAnnotations(emoji='🌃', codepoints=(127747,), name='stjernenatt', slug='stjernenatt', annotations=frozenset({'stjerner', 'natt'})),
EmojiAnnotations(emoji='🌄', codepoints=(127748,), name='soloppgang over fjell', slug='soloppgang_over_fjell', annotations=frozenset({'fjell', 'fjellsol', 'morgensol'})),
EmojiAnnotations(emoji='🌅', codepoints=(127749,), name='soloppgang', slug='soloppgang', annotations=frozenset({'morgensol'})),
EmojiAnnotations(emoji='🌆', codepoints=(127750,), name='bylandskap i skumring', slug='bylandskap_i_skumring', annotations=frozenset({'solnedgang', 'by', 'bygning', 'byhorisont', 'skumring'})),
EmojiAnnotations(emoji='🌇', codepoints=(127751,), name='solnedgang', slug='solnedgang', annotations=frozenset({'solnedgang over byhorisont', 'bygning', 'skumring'})),
EmojiAnnotations(emoji='🌉', codepoints=(127753,), name='bro', slug='bro', annotations=frozenset({'bro i mørket', 'bro om natten'})),
EmojiAnnotations(emoji='♨', codepoints=(9832,), name='varme kilder', slug='varme_kilder', annotations=frozenset({'vame kilder', 'varm', 'kilder', 'rykende'})),
EmojiAnnotations(emoji='🌌', codepoints=(127756,), name='melkeveien', slug='melkeveien', annotations=frozenset({'galakse'})),
EmojiAnnotations(emoji='🎠', codepoints=(127904,), name='karusellhest', slug='karusellhest', annotations=frozenset({'karusell', 'fornøyelsespark'})),
EmojiAnnotations(emoji='🎡', codepoints=(127905,), name='pariserhjul', slug='pariserhjul', annotations=frozenset({'fornøyelsespark'})),
EmojiAnnotations(emoji='🎢', codepoints=(127906,), name='berg- og dalbane', slug='berg__og_dalbane', annotations=frozenset({'fornøyelsespark'})),
EmojiAnnotations(emoji='💈', codepoints=(128136,), name='barberer', slug='barberer', annotations=frozenset({'frisørsalong', 'frisør'})),
EmojiAnnotations(emoji='🎪', codepoints=(127914,), name='sirkus', slug='sirkus', annotations=frozenset({'telt', 'sikrustelt', 'underholdning'})),
EmojiAnnotations(emoji='🎭', codepoints=(127917,), name='teatermasker', slug='teatermasker', annotations=frozenset({'kunst', 'teater', 'scenekunst', 'underholdning'})),
EmojiAnnotations(emoji='\U0001f5bc', codepoints=(128444,), name='ramme med bilde', slug='ramme_med_bilde', annotations=frozenset({'kunst', 'ramme', 'maleri', 'bilde', 'museum'})),
EmojiAnnotations(emoji='🎨', codepoints=(127912,), name='palett', slug='palett', annotations=frozenset({'kunst', 'malerpalett', 'maling'})),
EmojiAnnotations(emoji='🎰', codepoints=(127920,), name='spillemaskin', slug='spillemaskin', annotations=frozenset({'spill'})),
EmojiAnnotations(emoji='🚂', codepoints=(128642,), name='damplokmotiv', slug='damplokmotiv', annotations=frozenset({'fremkomstmiddel', 'lokmotiv', 'tog'})),
EmojiAnnotations(emoji='🚃', codepoints=(128643,), name='togvogn', slug='togvogn', annotations=frozenset({'trikk', 'elektrisk tog', 'elektrisk', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚄', codepoints=(128644,), name='høyhastighetstog', slug='høyhastighetstog', annotations=frozenset({'lyntog', 'fremkomstmiddel', 'tog'})),
EmojiAnnotations(emoji='🚅', codepoints=(128645,), name='høyhastighetstog med avrundet tipp', slug='høyhastighetstog_med_avrundet_tipp', annotations=frozenset({'lyntog', 'fremkomstmiddel', 'tog'})),
EmojiAnnotations(emoji='🚆', codepoints=(128646,), name='tog', slug='tog', annotations=frozenset({'jernbane', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚇', codepoints=(128647,), name='undergrunn', slug='undergrunn', annotations=frozenset({'metro', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚈', codepoints=(128648,), name='lokaltog', slug='lokaltog', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚉', codepoints=(128649,), name='togstasjon', slug='togstasjon', annotations=frozenset({'stasjon', 'tog'})),
EmojiAnnotations(emoji='🚊', codepoints=(128650,), name='trikk', slug='trikk', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚝', codepoints=(128669,), name='monorail', slug='monorail', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚞', codepoints=(128670,), name='fjellbane', slug='fjellbane', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚋', codepoints=(128651,), name='trikkevogn', slug='trikkevogn', annotations=frozenset({'vogn', 'trikk', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚌', codepoints=(128652,), name='buss', slug='buss', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚍', codepoints=(128653,), name='motgående buss', slug='motgående_buss', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚎', codepoints=(128654,), name='trikkebuss', slug='trikkebuss', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚏', codepoints=(128655,), name='busstopp', slug='busstopp', annotations=frozenset({'buss', 'holdeplass'})),
EmojiAnnotations(emoji='🚐', codepoints=(128656,), name='minibuss', slug='minibuss', annotations=frozenset({'kjøretøy', 'van'})),
EmojiAnnotations(emoji='🚑', codepoints=(128657,), name='ambulanse', slug='ambulanse', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚒', codepoints=(128658,), name='brannbil', slug='brannbil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚓', codepoints=(128659,), name='politibil', slug='politibil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚔', codepoints=(128660,), name='motgående politibil', slug='motgående_politibil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚕', codepoints=(128661,), name='taxi', slug='taxi', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚖', codepoints=(128662,), name='motgående taxi', slug='motgående_taxi', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚗', codepoints=(128663,), name='rød bil', slug='rød_bil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚘', codepoints=(128664,), name='motgående bil', slug='motgående_bil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚙', codepoints=(128665,), name='fritidsbil', slug='fritidsbil', annotations=frozenset({'fritidskjøretøy', 'kjøretøy'})),
EmojiAnnotations(emoji='🚚', codepoints=(128666,), name='varebil', slug='varebil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚛', codepoints=(128667,), name='lastebil', slug='lastebil', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚜', codepoints=(128668,), name='traktor', slug='traktor', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='🚲', codepoints=(128690,), name='sykkel', slug='sykkel', annotations=frozenset({'kjøretøy'})),
EmojiAnnotations(emoji='⛽', codepoints=(9981,), name='bensinstasjon', slug='bensinstasjon', annotations=frozenset({'bensinpumpe', 'bensin', 'diesel', 'stasjon'})),
EmojiAnnotations(emoji='\U0001f6e3', codepoints=(128739,), name='motorvei', slug='motorvei', annotations=frozenset({'vei'})),
EmojiAnnotations(emoji='\U0001f6e4', codepoints=(128740,), name='jernbanespor', slug='jernbanespor', annotations=frozenset({'jernbane', 'tog'})),
EmojiAnnotations(emoji='🚨', codepoints=(128680,), name='politilys', slug='politilys', annotations=frozenset({'bil', 'kjøretøy', 'signal', 'roterende', 'politi', 'lys'})),
EmojiAnnotations(emoji='🚥', codepoints=(128677,), name='horisontale trafikklys', slug='horisontale_trafikklys', annotations=frozenset({'horisontale lys', 'trafikk'})),
EmojiAnnotations(emoji='🚦', codepoints=(128678,), name='vertikale trafiklys', slug='vertikale_trafiklys', annotations=frozenset({'trafikk', 'vertikale lys'})),
EmojiAnnotations(emoji='🚧', codepoints=(128679,), name='under kontruksjon', slug='under_kontruksjon', annotations=frozenset({'skilt', 'byggeskilt', 'byggeplass'})),
EmojiAnnotations(emoji='⚓', codepoints=(9875,), name='anker', slug='anker', annotations=frozenset({'skip', 'verktøy'})),
EmojiAnnotations(emoji='⛵', codepoints=(9973,), name='seilbåt', slug='seilbåt', annotations=frozenset({'båt'})),
EmojiAnnotations(emoji='🚣', codepoints=(128675,), name='robåt', slug='robåt', annotations=frozenset({'båt', 'fartøy'})),
EmojiAnnotations(emoji='🚤', codepoints=(128676,), name='speedbåt', slug='speedbåt', annotations=frozenset({'båt', 'fartøy'})),
EmojiAnnotations(emoji='\U0001f6f3', codepoints=(128755,), name='passasjerskip', slug='passasjerskip', annotations=frozenset({'skip', 'fremkomstmiddel', 'passasjer'})),
EmojiAnnotations(emoji='⛴', codepoints=(9972,), name='ferge', slug='ferge', annotations=frozenset({'båt'})),
EmojiAnnotations(emoji='\U0001f6e5', codepoints=(128741,), name='motorbåt', slug='motorbåt', annotations=frozenset({'båt', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚢', codepoints=(128674,), name='skip', slug='skip', annotations=frozenset({'fartøy'})),
EmojiAnnotations(emoji='✈', codepoints=(9992,), name='fly', slug='fly', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='\U0001f6e9', codepoints=(128745,), name='lite fly', slug='lite_fly', annotations=frozenset({'fly', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='\U0001f6eb', codepoints=(128747,), name='fly som tar av', slug='fly_som_tar_av', annotations=frozenset({'avgang', 'innsjekk', 'avganger', 'fly', 'fremkomstmiddel'})),
EmojiAnnotations(emoji='\U0001f6ec', codepoints=(128748,), name='fly som lander', slug='fly_som_lander', annotations=frozenset({'ankomme', 'fremkomstmiddel', 'fly', 'ankomster', 'lande'})),
EmojiAnnotations(emoji='💺', codepoints=(128186,), name='sete', slug='sete', annotations=frozenset({'stol'})),
EmojiAnnotations(emoji='🚁', codepoints=(128641,), name='helikopter', slug='helikopter', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚟', codepoints=(128671,), name='svevebane', slug='svevebane', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚠', codepoints=(128672,), name='kabelbane', slug='kabelbane', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚡', codepoints=(128673,), name='taubane', slug='taubane', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='🚀', codepoints=(128640,), name='rakett', slug='rakett', annotations=frozenset({'fremkomstmiddel'})),
EmojiAnnotations(emoji='\U0001f6f0', codepoints=(128752,), name='satelitt', slug='satelitt', annotations=frozenset({'fremkomstmiddel', 'verdensrommet'})),
EmojiAnnotations(emoji='\U0001f6ce', codepoints=(128718,), name='hotellklokke', slug='hotellklokke', annotations=frozenset({'klokke', 'pikkolo', 'hotell'})),
EmojiAnnotations(emoji='\U0001f6cc', codepoints=(128716,), name='person i seng', slug='person_i_seng', annotations=frozenset({'hotell', 'sove'})),
EmojiAnnotations(emoji='\U0001f6cf', codepoints=(128719,), name='seng', slug='seng', annotations=frozenset({'hotell', 'sove'})),
EmojiAnnotations(emoji='\U0001f6cb', codepoints=(128715,), name='sofa og lampe', slug='sofa_og_lampe', annotations=frozenset({'sofa', 'lampe', 'hotell'})),
EmojiAnnotations(emoji='🚽', codepoints=(128701,), name='toalett', slug='toalett', annotations=frozenset({'do'})),
EmojiAnnotations(emoji='🚿', codepoints=(128703,), name='dusj', slug='dusj', annotations=frozenset({'vann'})),
EmojiAnnotations(emoji='🛀', codepoints=(128704,), name='person i badekar', slug='person_i_badekar', annotations=frozenset({'bad', 'badekar'})),
EmojiAnnotations(emoji='🛁', codepoints=(128705,), name='tomt badekar', slug='tomt_badekar', annotations=frozenset({'bad', 'badekar'})),
EmojiAnnotations(emoji='⌛', codepoints=(8987,), name='halvferdig timeglass', slug='halvferdig_timeglass', annotations=frozenset({'sand'})),
EmojiAnnotations(emoji='⏳', codepoints=(9203,), name='timeglass som har startet', slug='timeglass_som_har_startet', annotations=frozenset({'sand', 'tidtaker'})),
EmojiAnnotations(emoji='⌚', codepoints=(8986,), name='klokke', slug='klokke', annotations=frozenset({'ur'})),
EmojiAnnotations(emoji='⏰', codepoints=(9200,), name='alarmklokke', slug='alarmklokke', annotations=frozenset({'klokke', 'alarm'})),
EmojiAnnotations(emoji='⏱', codepoints=(9201,), name='stoppeklokke', slug='stoppeklokke', annotations=frozenset({'klokke'})),
EmojiAnnotations(emoji='⏲', codepoints=(9202,), name='tidtaker', slug='tidtaker', annotations=frozenset({'klokke'})),
EmojiAnnotations(emoji='\U0001f570', codepoints=(128368,), name='peisur', slug='peisur', annotations=frozenset({'klokke'})),
EmojiAnnotations(emoji='🕛', codepoints=(128347,), name='klokken tolv', slug='klokken_tolv', annotations=frozenset({'24:00', 'tolv', 'klokka', '12:00'})),
EmojiAnnotations(emoji='🕧', codepoints=(128359,), name='klokken halv ett', slug='klokken_halv_ett', annotations=frozenset({'12:30', '24:30', 'klokka'})),
EmojiAnnotations(emoji='🕐', codepoints=(128336,), name='klokken ett', slug='klokken_ett', annotations=frozenset({'1:00', 'klokka', 'ett', '13:00'})),
EmojiAnnotations(emoji='🕜', codepoints=(128348,), name='klokken halv to', slug='klokken_halv_to', annotations=frozenset({'13:30', '1:30', 'klokka'})),
EmojiAnnotations(emoji='🕑', codepoints=(128337,), name='klokken to', slug='klokken_to', annotations=frozenset({'2:00', '14:00', 'klokka', 'to'})),
EmojiAnnotations(emoji='🕝', codepoints=(128349,), name='klokken halv tre', slug='klokken_halv_tre', annotations=frozenset({'14:30', '2:30', 'klokka'})),
EmojiAnnotations(emoji='🕒', codepoints=(128338,), name='klokken tre', slug='klokken_tre', annotations=frozenset({'15:00', '3:00'})),
EmojiAnnotations(emoji='🕞', codepoints=(128350,), name='klokken halv fire', slug='klokken_halv_fire', annotations=frozenset({'3:30', 'klokka', '15:30'})),
EmojiAnnotations(emoji='🕓', codepoints=(128339,), name='klokken fire', slug='klokken_fire', annotations=frozenset({'fire', '4:00', '16:00', 'klokka'})),
EmojiAnnotations(emoji='🕟', codepoints=(128351,), name='klokken halv fem', slug='klokken_halv_fem', annotations=frozenset({'4:30', '16:30', 'klokka'})),
EmojiAnnotations(emoji='🕔', codepoints=(128340,), name='klokken fem', slug='klokken_fem', annotations=frozenset({'17:00', 'fem', 'klokka', '5:00'})),
EmojiAnnotations(emoji='🕠', codepoints=(128352,), name='klokken halv seks', slug='klokken_halv_seks', annotations=frozenset({'5:30', 'klokka', '17:30'})),
EmojiAnnotations(emoji='🕕', codepoints=(128341,), name='klokken seks', slug='klokken_seks', annotations=frozenset({'6:00', 'seks', 'klokka', '18:00'})),
EmojiAnnotations(emoji='🕡', codepoints=(128353,), name='klokken halv sju', slug='klokken_halv_sju', annotations=frozenset({'6:30', 'klokka', '18:30'})),
EmojiAnnotations(emoji='🕖', codepoints=(128342,), name='klokken sju', slug='klokken_sju', annotations=frozenset({'sju', '19:00', 'klokka', '7:00'})),
EmojiAnnotations(emoji='🕢', codepoints=(128354,), name='klokken halv åtte', slug='klokken_halv_åtte', annotations=frozenset({'7:30', 'klokka', '19:30'})),
EmojiAnnotations(emoji='🕗', codepoints=(128343,), name='klokken åtte', slug='klokken_åtte', annotations=frozenset({'åtte', 'klokka', '8:00', '20:00'})),
EmojiAnnotations(emoji='🕣', codepoints=(128355,), name='klokken halv ni', slug='klokken_halv_ni', annotations=frozenset({'klokka', '8:30', '20:30'})),
EmojiAnnotations(emoji='🕘', codepoints=(128344,), name='klokken ni', slug='klokken_ni', annotations=frozenset({'21:00', 'klokka', 'ni', '9:00'})),
EmojiAnnotations(emoji='🕤', codepoints=(128356,), name='klokken halv ti', slug='klokken_halv_ti', annotations=frozenset({'21:30', '09:30', 'klokka'})),
EmojiAnnotations(emoji='🕙', codepoints=(128345,), name='klokken ti', slug='klokken_ti', annotations=frozenset({'22:00', '10:00', 'klokka', 'ti'})),
EmojiAnnotations(emoji='🕥', codepoints=(128357,), name='klokken halv elleve', slug='klokken_halv_elleve', annotations=frozenset({'22:30', 'klokka', '10:30'})),
EmojiAnnotations(emoji='🕚', codepoints=(128346,), name='klokken elleve', slug='klokken_elleve', annotations=frozenset({'elleve', '23:00', 'klokka', '11:00'})),
EmojiAnnotations(emoji='🕦', codepoints=(128358,), name='klokken halv tolv', slug='klokken_halv_tolv', annotations=frozenset({'23:30', 'klokka', '11:30'})),
EmojiAnnotations(emoji='🌑', codepoints=(127761,), name='mørk måne', slug='mørk_måne', annotations=frozenset({'mørk', 'måne'})),
EmojiAnnotations(emoji='🌒', codepoints=(127762,), name='voksende nymåne', slug='voksende_nymåne', annotations=frozenset({'månefase', 'tiltakende måne'})),
EmojiAnnotations(emoji='🌓', codepoints=(127763,), name='voksende halvmåne', slug='voksende_halvmåne', annotations=frozenset({'første kvarter', 'månefase'})),
EmojiAnnotations(emoji='🌔', codepoints=(127764,), name='voksende fullmåne', slug='voksende_fullmåne', annotations=frozenset({'månefase', 'tiltakende måne'})),
EmojiAnnotations(emoji='🌕', codepoints=(127765,), name='fullmåne', slug='fullmåne', annotations=frozenset({'klar måne', 'måne'})),
EmojiAnnotations(emoji='🌖', codepoints=(127766,), name='avtakende fullmåne', slug='avtakende_fullmåne', annotations=frozenset({'månefase', 'avtakende måne'})),
EmojiAnnotations(emoji='🌗', codepoints=(127767,), name='avtakende halvmåne', slug='avtakende_halvmåne', annotations=frozenset({'siste kvarter', 'månefase'})),
EmojiAnnotations(emoji='🌘', codepoints=(127768,), name='avtakende nemåne', slug='avtakende_nemåne', annotations=frozenset({'månefase', 'avtakende måne'})),
EmojiAnnotations(emoji='🌙', codepoints=(127769,), name='nymåne', slug='nymåne', annotations=frozenset({'måne'})),
EmojiAnnotations(emoji='🌚', codepoints=(127770,), name='nymåneansikt', slug='nymåneansikt', annotations=frozenset({'nymåne', 'nymåne med ansikt', 'måne', 'ansikt'})),
EmojiAnnotations(emoji='🌛', codepoints=(127771,), name='nymåne med ansikt', slug='nymåne_med_ansikt', annotations=frozenset({'kvartmåne med ansikt', 'måneansikt'})),
EmojiAnnotations(emoji='🌜', codepoints=(127772,), name='nemåne med ansikt', slug='nemåne_med_ansikt', annotations=frozenset({'kvartmåne med ansikt', 'måneansikt'})),
EmojiAnnotations(emoji='\U0001f321', codepoints=(127777,), name='termometer', slug='termometer', annotations=frozenset({'vær'})),
EmojiAnnotations(emoji='☀', codepoints=(9728,), name='sol', slug='sol', annotations=frozenset({'solfylt', 'stråler', 'vær', 'verdensrommet', 'lyst'})),
EmojiAnnotations(emoji='🌝', codepoints=(127773,), name='fullmåneansikt', slug='fullmåneansikt', annotations=frozenset({'måne med ansikt', 'fullmåne', 'fullmåne med ansikt', 'måne', 'ansikt'})),
EmojiAnnotations(emoji='🌞', codepoints=(127774,), name='solansikt', slug='solansikt', annotations=frozenset({'sol', 'ansikt'})),
EmojiAnnotations(emoji='⭐', codepoints=(11088,), name='hvit stjerne', slug='hvit_stjerne', annotations=frozenset({'stjerne'})),
EmojiAnnotations(emoji='🌟', codepoints=(127775,), name='lysende stjerne', slug='lysende_stjerne', annotations=frozenset({'skinnende', 'glitrende', 'stjerne'})),
EmojiAnnotations(emoji='🌠', codepoints=(127776,), name='stjerneskudd', slug='stjerneskudd', annotations=frozenset({'fallende stjerne', 'stjerne'})),
EmojiAnnotations(emoji='☁', codepoints=(9729,), name='sky', slug='sky', annotations=frozenset({'vær'})),
EmojiAnnotations(emoji='⛅', codepoints=(9925,), name='sol bak sky', slug='sol_bak_sky', annotations=frozenset({'sky', 'sol', 'sol og sky'})),
EmojiAnnotations(emoji='⛈', codepoints=(9928,), name='sky med lyn og regn', slug='sky_med_lyn_og_regn', annotations=frozenset({'sky', 'torden', 'regn', 'vær'})),
EmojiAnnotations(emoji='\U0001f324', codepoints=(127780,), name='sol bak liten sky', slug='sol_bak_liten_sky', annotations=frozenset({'sky', 'sol', 'vær'})),
EmojiAnnotations(emoji='\U0001f325', codepoints=(127781,), name='sol bak stor sky', slug='sol_bak_stor_sky', annotations=frozenset({'sky', 'sol', 'vær'})),
EmojiAnnotations(emoji='\U0001f326', codepoints=(127782,), name='sol bak sky med regn', slug='sol_bak_sky_med_regn', annotations=frozenset({'sky', 'sol', 'regn', 'vær'})),
EmojiAnnotations(emoji='\U0001f327', codepoints=(127783,), name='sky med regn', slug='sky_med_regn', annotations=frozenset({'sky', 'regn', 'vær'})),
EmojiAnnotations(emoji='\U0001f328', codepoints=(127784,), name='sky med snø', slug='sky_med_snø', annotations=frozenset({'sky', 'kaldt', 'snø', 'vær'})),
EmojiAnnotations(emoji='\U0001f329', codepoints=(127785,), name='sky med lyn', slug='sky_med_lyn', annotations=frozenset({'sky', 'lyn', 'vær'})),
EmojiAnnotations(emoji='\U0001f32a', codepoints=(127786,), name='tornado', slug='tornado', annotations=frozenset({'sky', 'vær'})),
EmojiAnnotations(emoji='\U0001f32b', codepoints=(127787,), name='tåke', slug='tåke', annotations=frozenset({'sky', 'vær'})),
EmojiAnnotations(emoji='\U0001f32c', codepoints=(127788,), name='vindansikt', slug='vindansikt', annotations=frozenset({'sky', 'vind', 'fjes', 'blåse', 'vær'})),
EmojiAnnotations(emoji='🌀', codepoints=(127744,), name='syklon', slug='syklon', annotations=frozenset({'tornado', 'svimmel'})),
EmojiAnnotations(emoji='🌈', codepoints=(127752,), name='regnbue', slug='regnbue', annotations=frozenset({'regn', 'vær'})),
EmojiAnnotations(emoji='🌂', codepoints=(127746,), name='lukket paraply', slug='lukket_paraply', annotations=frozenset({'regnvær', 'regn'})),
EmojiAnnotations(emoji='☂', codepoints=(9730,), name='paraply', slug='paraply', annotations=frozenset({'regn', 'klær', 'vær'})),
EmojiAnnotations(emoji='☔', codepoints=(9748,), name='paraply med regndråper', slug='paraply_med_regndråper', annotations=frozenset({'dråpe', 'paraply', 'regn', 'klær', 'vær'})),
EmojiAnnotations(emoji='⛱', codepoints=(9969,), name='parasoll på bakken', slug='parasoll_på_bakken', annotations=frozenset({'paraply', 'sol', 'regn', 'vær'})),
EmojiAnnotations(emoji='⚡', codepoints=(9889,), name='høyspenning', slug='høyspenning', annotations=frozenset({'fare', 'høy', 'høyspent', 'spenning'})),
EmojiAnnotations(emoji='❄', codepoints=(10052,), name='snøfnugg', slug='snøfnugg', annotations=frozenset({'kaldt', 'snø', 'vær'})),
EmojiAnnotations(emoji='☃', codepoints=(9731,), name='snømann', slug='snømann', annotations=frozenset({'kaldt', 'snø', 'vær'})),
EmojiAnnotations(emoji='⛄', codepoints=(9924,), name='snømann uten snø', slug='snømann_uten_snø', annotations=frozenset({'snø'})),
EmojiAnnotations(emoji='☄', codepoints=(9732,), name='komet', slug='komet', annotations=frozenset({'verdensrommet'})),
EmojiAnnotations(emoji='🔥', codepoints=(128293,), name='ild', slug='ild', annotations=frozenset({'flamme'})),
EmojiAnnotations(emoji='💧', codepoints=(128167,), name='dråpe', slug='dråpe', annotations=frozenset({'tegneserie', 'drypp', 'vann'})),
EmojiAnnotations(emoji='🌊', codepoints=(127754,), name='bølge', slug='bølge', annotations=frozenset({'bølgesprut', 'hav'})),
EmojiAnnotations(emoji='🎃', codepoints=(127875,), name='gresskar', slug='gresskar', annotations=frozenset({'gresskarlykt', 'halloweenlykt', 'Halloween'})),
EmojiAnnotations(emoji='🎄', codepoints=(127876,), name='juletre', slug='juletre', annotations=frozenset({'høytid', 'tre', 'julefeiring', 'jul'})),
EmojiAnnotations(emoji='🎆', codepoints=(127878,), name='fyrverkeri', slug='fyrverkeri', annotations=frozenset({'feiring', 'nyttår'})),
EmojiAnnotations(emoji='🎇', codepoints=(127879,), name='stjerneskudd på pinne', slug='stjerneskudd_på_pinne', annotations=frozenset({'feiring', 'fyrverkeri', 'nyttår'})),
EmojiAnnotations(emoji='✨', codepoints=(10024,), name='stjerner', slug='stjerner', annotations=frozenset({'skinnende', 'glitrende'})),
EmojiAnnotations(emoji='🎈', codepoints=(127880,), name='ballong', slug='ballong', annotations=frozenset({'feiring'})),
EmojiAnnotations(emoji='🎉', codepoints=(127881,), name='smell-bon-bon', slug='smell_bon_bon', annotations=frozenset({'serpentin', 'fest', 'feiring'})),
EmojiAnnotations(emoji='🎊', codepoints=(127882,), name='konfetti', slug='konfetti', annotations=frozenset({'konfettikule', 'fest', 'feiring'})),
EmojiAnnotations(emoji='🎋', codepoints=(127883,), name='tre med papirremser', slug='tre_med_papirremser', annotations=frozenset({'tre', 'papirremser', 'stjernefestival', 'feiring'})),
EmojiAnnotations(emoji='🎌', codepoints=(127884,), name='kryssede flagg', slug='kryssede_flagg', annotations=frozenset({'flagg', 'krysset', 'kryssede japanske flagg', 'feiring'})),
EmojiAnnotations(emoji='🎍', codepoints=(127885,), name='furudekorasjon', slug='furudekorasjon', annotations=frozenset({'feiring', 'japansk høytid', 'nyttår'})),
EmojiAnnotations(emoji='🎎', codepoints=(127886,), name='dukker', slug='dukker', annotations=frozenset({'festival', 'japansk høytid', 'japanske dukker'})),
EmojiAnnotations(emoji='🎏', codepoints=(127887,), name='fiskedekorasjon', slug='fiskedekorasjon', annotations=frozenset({'japansk høytid', 'papirfisk', 'feiring'})),
EmojiAnnotations(emoji='🎐', codepoints=(127888,), name='vindspill', slug='vindspill', annotations=frozenset({'drømmefanger'})),
EmojiAnnotations(emoji='🎑', codepoints=(127889,), name='månebeskuelse', slug='månebeskuelse', annotations=frozenset({'høytid', 'feiring', 'månefeiring', 'måne', 'måneseremoni'})),
EmojiAnnotations(emoji='🎀', codepoints=(127872,), name='sløyfe', slug='sløyfe', annotations=frozenset({'gavebånd', 'bånd'})),
EmojiAnnotations(emoji='🎁', codepoints=(127873,), name='gave', slug='gave', annotations=frozenset({'feiring', 'innpakket gave', 'bursdag', 'julegave'})),
EmojiAnnotations(emoji='\U0001f396', codepoints=(127894,), name='medalje', slug='medalje', annotations=frozenset({'militæret', 'feiring'})),
EmojiAnnotations(emoji='\U0001f397', codepoints=(127895,), name='minnesløyfe', slug='minnesløyfe', annotations=frozenset({'bånd', 'påminnelse', 'feiring'})),
EmojiAnnotations(emoji='\U0001f39e', codepoints=(127902,), name='filmrammer', slug='filmrammer', annotations=frozenset({'rammer', 'kino', 'film'})),
EmojiAnnotations(emoji='\U0001f39f', codepoints=(127903,), name='inngangsbilletter', slug='inngangsbilletter', annotations=frozenset({'billett', 'adgang'})),
EmojiAnnotations(emoji='🎫', codepoints=(127915,), name='billett', slug='billett', annotations=frozenset({'underholdning'})),
EmojiAnnotations(emoji='⚽', codepoints=(9917,), name='fotball', slug='fotball', annotations=frozenset({'ball'})),
EmojiAnnotations(emoji='⚾', codepoints=(9918,), name='baseball', slug='baseball', annotations=frozenset({'ball', 'stitched'})),
EmojiAnnotations(emoji='🏀', codepoints=(127936,), name='basketball', slug='basketball', annotations=frozenset({'sport', 'ball'})),
EmojiAnnotations(emoji='🏈', codepoints=(127944,), name='amerikansk fotball', slug='amerikansk_fotball', annotations=frozenset({'sport', 'ball'})),
EmojiAnnotations(emoji='🏉', codepoints=(127945,), name='rugby', slug='rugby', annotations=frozenset({'sport', 'rugbyball'})),
EmojiAnnotations(emoji='🎾', codepoints=(127934,), name='tennis', slug='tennis', annotations=frozenset({'sport', 'ball', 'tenninsball'})),
EmojiAnnotations(emoji='🎱', codepoints=(127921,), name='biljard', slug='biljard', annotations=frozenset({'åtterkule', 'spill'})),
EmojiAnnotations(emoji='🎳', codepoints=(127923,), name='bowling', slug='bowling', annotations=frozenset({'kule', 'bowlingkule', 'spill'})),
EmojiAnnotations(emoji='⛳', codepoints=(9971,), name='golfflagg', slug='golfflagg', annotations=frozenset({'golf', 'golfhull'})),
EmojiAnnotations(emoji='\U0001f3cc', codepoints=(127948,), name='golfspiller', slug='golfspiller', annotations=frozenset({'golf', 'ball'})),
EmojiAnnotations(emoji='⛸', codepoints=(9976,), name='skøyter', slug='skøyter', annotations=frozenset({'is', 'skøyte'})),
EmojiAnnotations(emoji='🎣', codepoints=(127907,), name='fiskestang', slug='fiskestang', annotations=frozenset({'fisking', 'fisk', 'fritid', 'natur'})),
EmojiAnnotations(emoji='🎽', codepoints=(127933,), name='løpetopp', slug='løpetopp', annotations=frozenset({'sport', 't-skjorte', 'treningstopp'})),
EmojiAnnotations(emoji='🎿', codepoints=(127935,), name='ski', slug='ski', annotations=frozenset({'ski og staver', 'vintersport'})),
EmojiAnnotations(emoji='⛷', codepoints=(9975,), name='skiløper', slug='skiløper', annotations=frozenset({'ski', 'snø'})),
EmojiAnnotations(emoji='🏂', codepoints=(127938,), name='snowboarder', slug='snowboarder', annotations=frozenset({'vintersport', 'snøbrettkjøring', 'snowboarding', 'snowboard', 'snøbrett'})),
EmojiAnnotations(emoji='🏄', codepoints=(127940,), name='surfing', slug='surfing', annotations=frozenset({'sport', 'surfer'})),
EmojiAnnotations(emoji='🏇', codepoints=(127943,), name='veddeløp', slug='veddeløp', annotations=frozenset({'sport', 'jockey', 'hest', 'veddeløpshest', 'løp'})),
EmojiAnnotations(emoji='🏊', codepoints=(127946,), name='svømming', slug='svømming', annotations=frozenset({'sport', 'svømme', 'svømmer'})),
EmojiAnnotations(emoji='⛹', codepoints=(9977,), name='person med ball', slug='person_med_ball', annotations=frozenset({'ball'})),
EmojiAnnotations(emoji='\U0001f3cb', codepoints=(127947,), name='vektløfter', slug='vektløfter', annotations=frozenset({'vekt', 'løfter'})),
EmojiAnnotations(emoji='🚴', codepoints=(128692,), name='syklist', slug='syklist', annotations=frozenset({'sykkel'})),
EmojiAnnotations(emoji='🚵', codepoints=(128693,), name='fjellsyklist', slug='fjellsyklist', annotations=frozenset({'mountainbike'})),
EmojiAnnotations(emoji='\U0001f3ce', codepoints=(127950,), name='racerbil', slug='racerbil', annotations=frozenset({'kappkjøring', 'bil'})),
EmojiAnnotations(emoji='\U0001f3cd', codepoints=(127949,), name='motorsykkel', slug='motorsykkel', annotations=frozenset({'kappkjøring'})),
EmojiAnnotations(emoji='\U0001f3c5', codepoints=(127941,), name='idrettsmedalje', slug='idrettsmedalje', annotations=frozenset({'medalje'})),
EmojiAnnotations(emoji='🏆', codepoints=(127942,), name='trofé', slug='trofé', annotations=frozenset({'premie'})),
EmojiAnnotations(emoji='\U0001f3cf', codepoints=(127951,), name='cricket', slug='cricket', annotations=frozenset({'balltre', 'kamp', 'ball'})),
EmojiAnnotations(emoji='\U0001f3d0', codepoints=(127952,), name='volleyball', slug='volleyball', annotations=frozenset({'kamp', 'ball'})),
EmojiAnnotations(emoji='\U0001f3d1', codepoints=(127953,), name='landhockey', slug='landhockey', annotations=frozenset({'kølle', 'hockey', 'kamp', 'ball', 'land'})),
EmojiAnnotations(emoji='\U0001f3d2', codepoints=(127954,), name='ishockeykølle og puck', slug='ishockeykølle_og_puck', annotations=frozenset({'is', 'hockey', 'kølle', 'kamp', 'puck'})),
EmojiAnnotations(emoji='\U0001f3d3', codepoints=(127955,), name='bordtennis', slug='bordtennis', annotations=frozenset({'racket', 'balltre', 'tennis', 'kamp', 'bord', 'ball'})),
EmojiAnnotations(emoji='\U0001f3f8', codepoints=(127992,), name='badminton', slug='badminton', annotations=frozenset({'fjærball', 'racket', 'kamp', 'birdie'})),
EmojiAnnotations(emoji='🎯', codepoints=(127919,), name='blink', slug='blink', annotations=frozenset({'treff', 'midt i blinken', 'mål', 'pilspill', 'dart'})),
EmojiAnnotations(emoji='🎮', codepoints=(127918,), name='videospill', slug='videospill', annotations=frozenset({'spillkonsoll', 'konsoll', 'spill'})),
EmojiAnnotations(emoji='\U0001f579', codepoints=(128377,), name='styrespake', slug='styrespake', annotations=frozenset({'videospill', 'spill'})),
EmojiAnnotations(emoji='🎲', codepoints=(127922,), name='terning', slug='terning', annotations=frozenset({'terningspill', 'terningkast', 'spill'})),
EmojiAnnotations(emoji='♠', codepoints=(9824,), name='spar', slug='spar', annotations=frozenset({'farge', 'kort', 'spill'})),
EmojiAnnotations(emoji='♥', codepoints=(9829,), name='hjerter', slug='hjerter', annotations=frozenset({'farge', 'kort', 'spill'})),
EmojiAnnotations(emoji='♦', codepoints=(9830,), name='ruter', slug='ruter', annotations=frozenset({'spill', 'farge', 'kort', 'diamant'})),
EmojiAnnotations(emoji='♣', codepoints=(9827,), name='kløver', slug='kløver', annotations=frozenset({'farge', 'kort', 'spill'})),
EmojiAnnotations(emoji='🃏', codepoints=(127183,), name='joker', slug='joker', annotations=frozenset({'jokerkort', 'spillkort', 'kort'})),
EmojiAnnotations(emoji='🀄', codepoints=(126980,), name='mahjongbrikke', slug='mahjongbrikke', annotations=frozenset({'dominobrikke'})),
EmojiAnnotations(emoji='🎴', codepoints=(127924,), name='spillkort', slug='spillkort', annotations=frozenset({'kort', 'kortspill'})),
EmojiAnnotations(emoji='🔇', codepoints=(128263,), name='stille', slug='stille', annotations=frozenset({'ingen prating', 'høyttaler med strek over', 'vennligst være stille'})),
EmojiAnnotations(emoji='🔈', codepoints=(128264,), name='høyttaler', slug='høyttaler', annotations=frozenset({'ropert'})),
EmojiAnnotations(emoji='🔉', codepoints=(128265,), name='høyttaler med lavt volum', slug='høyttaler_med_lavt_volum', annotations=frozenset({'lavt volum'})),
EmojiAnnotations(emoji='🔊', codepoints=(128266,), name='høyttaler med høyt volum', slug='høyttaler_med_høyt_volum', annotations=frozenset({'høyt volum'})),
EmojiAnnotations(emoji='📢', codepoints=(128226,), name='ropert', slug='ropert', annotations=frozenset({'megafon', 'kommunikasjon'})),
EmojiAnnotations(emoji='📣', codepoints=(128227,), name='megafon', slug='megafon', annotations=frozenset({'heiing', 'kommunikasjon'})),
EmojiAnnotations(emoji='📯', codepoints=(128239,), name='posthorn', slug='posthorn', annotations=frozenset({'post'})),
EmojiAnnotations(emoji='🔔', codepoints=(128276,), name='bjelle', slug='bjelle', annotations=frozenset({'klokke', 'ringing'})),
EmojiAnnotations(emoji='🔕', codepoints=(128277,), name='ingen lyd', slug='ingen_lyd', annotations=frozenset({'bjelle med strek over', 'klokke', 'ingen bjelle'})),
EmojiAnnotations(emoji='🎼', codepoints=(127932,), name='g-nøkkel', slug='g_nøkkel', annotations=frozenset({'notetegn', 'musikk'})),
EmojiAnnotations(emoji='🎵', codepoints=(127925,), name='note', slug='note', annotations=frozenset({'musikk', 'musikknote'})),
EmojiAnnotations(emoji='🎶', codepoints=(127926,), name='noter', slug='noter', annotations=frozenset({'musikknoter', 'musikk'})),
EmojiAnnotations(emoji='\U0001f399', codepoints=(127897,), name='studiomikrofon', slug='studiomikrofon', annotations=frozenset({'musikk', 'mikrofon', 'studio'})),
EmojiAnnotations(emoji='\U0001f39a', codepoints=(127898,), name='nivåglidebryter', slug='nivåglidebryter', annotations=frozenset({'glidebryter', 'musikk', 'nivå'})),
EmojiAnnotations(emoji='\U0001f39b', codepoints=(127899,), name='kontrollbrytere', slug='kontrollbrytere', annotations=frozenset({'musikk', 'knott', 'kontroll'})),
EmojiAnnotations(emoji='🎤', codepoints=(127908,), name='mikrofon', slug='mikrofon', annotations=frozenset({'underholdning', 'karaoke', 'musikk', 'mik'})),
EmojiAnnotations(emoji='🎧', codepoints=(127911,), name='øreplugg', slug='øreplugg', annotations=frozenset({'underholdning', 'hodetelefoner', 'headset'})),
EmojiAnnotations(emoji='🎷', codepoints=(127927,), name='sax', slug='sax', annotations=frozenset({'musikk', 'instrument', 'saxofon', 'musikkinstrument'})),
EmojiAnnotations(emoji='🎸', codepoints=(127928,), name='gitar', slug='gitar', annotations=frozenset({'musikk', 'instrument', 'musikkinstrument'})),
EmojiAnnotations(emoji='🎹', codepoints=(127929,), name='piano', slug='piano', annotations=frozenset({'musikk', 'instrument', 'musikkinstrument', 'keyboard'})),
EmojiAnnotations(emoji='🎺', codepoints=(127930,), name='trompet', slug='trompet', annotations=frozenset({'musikk', 'instrument', 'musikkinstrument'})),
EmojiAnnotations(emoji='🎻', codepoints=(127931,), name='fiolin', slug='fiolin', annotations=frozenset({'musikk', 'instrument', 'musikkinstrument', 'fele'})),
EmojiAnnotations(emoji='📻', codepoints=(128251,), name='radio', slug='radio', annotations=frozenset({'radiosignaler', 'teknologi'})),
EmojiAnnotations(emoji='📱', codepoints=(128241,), name='mobil', slug='mobil', annotations=frozenset({'mobiltelefon', 'telefon'})),
EmojiAnnotations(emoji='📲', codepoints=(128242,), name='mobiltelefon med pil', slug='mobiltelefon_med_pil', annotations=frozenset({'ring', 'mobiltelefon', 'telefon'})),
EmojiAnnotations(emoji='📞', codepoints=(128222,), name='telefonrør', slug='telefonrør', annotations=frozenset({'kommunikasjon'})),
EmojiAnnotations(emoji='📟', codepoints=(128223,), name='personsøker', slug='personsøker', annotations=frozenset({'kommunikasjon'})),
EmojiAnnotations(emoji='📠', codepoints=(128224,), name='faks', slug='faks', annotations=frozenset({'faksmaskin', 'kommunikasjon'})),
EmojiAnnotations(emoji='🔋', codepoints=(128267,), name='batteri', slug='batteri', annotations=frozenset({'strøm', 'energilagring'})),
EmojiAnnotations(emoji='🔌', codepoints=(128268,), name='støpsel', slug='støpsel', annotations=frozenset({'elektrisk'})),
EmojiAnnotations(emoji='💻', codepoints=(128187,), name='PC', slug='pc', annotations=frozenset({'personlig', 'datamaskin', 'maskin'})),
EmojiAnnotations(emoji='\U0001f5a5', codepoints=(128421,), name='stasjonær datamaskin', slug='stasjonær_datamaskin', annotations=frozenset({'datamaskin', 'stasjonær'})),
EmojiAnnotations(emoji='\U0001f5a8', codepoints=(128424,), name='skriver', slug='skriver', annotations=frozenset({'datamaskin'})),
EmojiAnnotations(emoji='⌨', codepoints=(9000,), name='tastatur', slug='tastatur', annotations=frozenset({'datamaskin'})),
EmojiAnnotations(emoji='\U0001f5b1', codepoints=(128433,), name='datamus', slug='datamus', annotations=frozenset({'knapp', 'tre', 'datamaskin', 'mus', '3'})),
EmojiAnnotations(emoji='\U0001f5b2', codepoints=(128434,), name='styrekule', slug='styrekule', annotations=frozenset({'datamaskin'})),
EmojiAnnotations(emoji='💽', codepoints=(128189,), name='MD', slug='md', annotations=frozenset({'minidisk'})),
EmojiAnnotations(emoji='💾', codepoints=(128190,), name='diskett', slug='diskett', annotations=frozenset({'datamaskin', 'lagring'})),
EmojiAnnotations(emoji='💿', codepoints=(128191,), name='blu-ray', slug='blu_ray', annotations=frozenset({'DVD', 'CD'})),
EmojiAnnotations(emoji='📀', codepoints=(128192,), name='DVD', slug='dvd', annotations=frozenset({'plate', 'DVD-plate'})),
EmojiAnnotations(emoji='🎥', codepoints=(127909,), name='filmkamera', slug='filmkamera', annotations=frozenset({'kamera', 'kino', 'film', 'underholdning'})),
EmojiAnnotations(emoji='🎬', codepoints=(127916,), name='filmklapper', slug='filmklapper', annotations=frozenset({'filming', 'film', 'action', 'underholdning'})),
EmojiAnnotations(emoji='\U0001f4fd', codepoints=(128253,), name='filmprosjektor', slug='filmprosjektor', annotations=frozenset({'kino', 'prosjektor', 'film', 'video'})),
EmojiAnnotations(emoji='📺', codepoints=(128250,), name='TV', slug='tv', annotations=frozenset({'fjernsyn'})),
EmojiAnnotations(emoji='📷', codepoints=(128247,), name='kamera', slug='kamera', annotations=frozenset({'kamerateknikk', 'teknologi'})),
EmojiAnnotations(emoji='\U0001f4f8', codepoints=(128248,), name='kamera med blits', slug='kamera_med_blits', annotations=frozenset({'kamera', 'blits', 'video'})),
EmojiAnnotations(emoji='📹', codepoints=(128249,), name='videokamera', slug='videokamera', annotations=frozenset({'kamera', 'video'})),
EmojiAnnotations(emoji='📼', codepoints=(128252,), name='videoopptak', slug='videoopptak', annotations=frozenset({'opptak', 'video'})),
EmojiAnnotations(emoji='🔍', codepoints=(128269,), name='forstørrelsesglass mot venstre', slug='forstørrelsesglass_mot_venstre', annotations=frozenset({'søk', 'glass'})),
EmojiAnnotations(emoji='🔎', codepoints=(128270,), name='forstørrelsesglass mot høyre', slug='forstørrelsesglass_mot_høyre', annotations=frozenset({'forstørrelsesglass mot venstre', 'søk', 'glass'})),
EmojiAnnotations(emoji='🔬', codepoints=(128300,), name='mikroskop', slug='mikroskop', annotations=frozenset({'verktøy', 'vitenskap'})),
EmojiAnnotations(emoji='🔭', codepoints=(128301,), name='teleskop', slug='teleskop', annotations=frozenset({'verktøy', 'astronomi', 'vitenskap'})),
EmojiAnnotations(emoji='📡', codepoints=(128225,), name='satellitt', slug='satellitt', annotations=frozenset({'antenne', 'satelittantenne', 'satelitt', 'kommunikasjon'})),
EmojiAnnotations(emoji='\U0001f56f', codepoints=(128367,), name='stearinlys', slug='stearinlys', annotations=frozenset({'lys'})),
EmojiAnnotations(emoji='💡', codepoints=(128161,), name='lyspære', slug='lyspære', annotations=frozenset({'idé', 'pære', 'tegneserie', 'elektrisitet', 'lys'})),
EmojiAnnotations(emoji='🔦', codepoints=(128294,), name='lommelykt', slug='lommelykt', annotations=frozenset({'lykt', 'elektrisk', 'fakkel'})),
EmojiAnnotations(emoji='🏮', codepoints=(127982,), name='lykt', slug='lykt', annotations=frozenset({'lanterne', 'rød lykt', 'izakaya'})),
EmojiAnnotations(emoji='📔', codepoints=(128212,), name='dekorert notatbok', slug='dekorert_notatbok', annotations=frozenset({'notatbok', 'omslag'})),
EmojiAnnotations(emoji='📕', codepoints=(128213,), name='lukket bok', slug='lukket_bok', annotations=frozenset({'bok'})),
EmojiAnnotations(emoji='📖', codepoints=(128214,), name='åpen bok', slug='åpen_bok', annotations=frozenset({'bok'})),
EmojiAnnotations(emoji='📗', codepoints=(128215,), name='grønn bok', slug='grønn_bok', annotations=frozenset({'bok'})),
EmojiAnnotations(emoji='📘', codepoints=(128216,), name='blå bok', slug='blå_bok', annotations=frozenset({'bok'})),
EmojiAnnotations(emoji='📙', codepoints=(128217,), name='oransje bok', slug='oransje_bok', annotations=frozenset({'bok'})),
EmojiAnnotations(emoji='📚', codepoints=(128218,), name='bøker', slug='bøker', annotations=frozenset({'bok'})),
EmojiAnnotations(emoji='📓', codepoints=(128211,), name='notatbok', slug='notatbok', annotations=frozenset({'kontorrekvisita', 'kladdebok'})),
EmojiAnnotations(emoji='📒', codepoints=(128210,), name='spiralblokk', slug='spiralblokk', annotations=frozenset({'notatbok', 'kontorrekvisita'})),
EmojiAnnotations(emoji='📃', codepoints=(128195,), name='dokument', slug='dokument', annotations=frozenset({'opprullet ark'})),
EmojiAnnotations(emoji='📜', codepoints=(128220,), name='rull', slug='rull', annotations=frozenset({'papirrull'})),
EmojiAnnotations(emoji='📄', codepoints=(128196,), name='dokument som vender opp', slug='dokument_som_vender_opp', annotations=frozenset({'ark', 'vender opp'})),
EmojiAnnotations(emoji='📰', codepoints=(128240,), name='avis', slug='avis', annotations=frozenset({'nyheter', 'publikasjon'})),
EmojiAnnotations(emoji='\U0001f5de', codepoints=(128478,), name='opprullet avis', slug='opprullet_avis', annotations=frozenset({'papir', 'avis', 'nyheter', 'opprullet'})),
EmojiAnnotations(emoji='📑', codepoints=(128209,), name='bokmerkefane', slug='bokmerkefane', annotations=frozenset({'markør', 'faner', 'bokmerke'})),
EmojiAnnotations(emoji='🔖', codepoints=(128278,), name='bokmerke', slug='bokmerke', annotations=frozenset({'merke'})),
EmojiAnnotations(emoji='💰', codepoints=(128176,), name='pengesekk', slug='pengesekk', annotations=frozenset({'sekk', 'penger'})),
EmojiAnnotations(emoji='💴', codepoints=(128180,), name='yen-seddel', slug='yen_seddel', annotations=frozenset({'yen', 'penger'})),
EmojiAnnotations(emoji='💵', codepoints=(128181,), name='dollar-seddel', slug='dollar_seddel', annotations=frozenset({'dollar', 'dollarseddel', 'penger'})),
EmojiAnnotations(emoji='💶', codepoints=(128182,), name='euro-seddel', slug='euro_seddel', annotations=frozenset({'euro', 'euroseddel', 'penger'})),
EmojiAnnotations(emoji='💷', codepoints=(128183,), name='pund-seddel', slug='pund_seddel', annotations=frozenset({'pundseddel', 'pund', 'penger'})),
EmojiAnnotations(emoji='💸', codepoints=(128184,), name='flygende penger', slug='flygende_penger', annotations=frozenset({'flygende seddel', 'bank', 'penger med vinger', 'pengeseddel', 'penger'})),
EmojiAnnotations(emoji='💳', codepoints=(128179,), name='kredittkort', slug='kredittkort', annotations=frozenset({'betalingskort', 'kort', 'visa', 'kreditt'})),
EmojiAnnotations(emoji='💹', codepoints=(128185,), name='stigende marked', slug='stigende_marked', annotations=frozenset({'oppover', 'stigende graf', 'graf', 'marked'})),
EmojiAnnotations(emoji='✉', codepoints=(9993,), name='konvolutt', slug='konvolutt', annotations=frozenset({'e-post'})),
EmojiAnnotations(emoji='📧', codepoints=(128231,), name='e-post', slug='e_post', annotations=frozenset({'kommunikasjon', 'brev', 'post'})),
EmojiAnnotations(emoji='📨', codepoints=(128232,), name='innkommende e-post', slug='innkommende_e_post', annotations=frozenset({'e-post', 'kommunikasjon', 'mottatt e-post', 'konvolutt', 'brev', 'innkommende', 'post'})),
EmojiAnnotations(emoji='📩', codepoints=(128233,), name='utgående e-post', slug='utgående_e_post', annotations=frozenset({'e-post', 'kommunikasjon', 'sendt e-post', 'utgående', 'konvolutt', 'brev', 'post'})),
EmojiAnnotations(emoji='📤', codepoints=(128228,), name='utboks', slug='utboks', annotations=frozenset({'sendt e-post', 'e-post', 'sent', 'skuff', 'kommunikasjon'})),
EmojiAnnotations(emoji='📥', codepoints=(128229,), name='innboks', slug='innboks', annotations=frozenset({'mottatt e-post', 'mottatt', 'skuff', 'e-post', 'kommunikasjon'})),
EmojiAnnotations(emoji='📦', codepoints=(128230,), name='pakke', slug='pakke', annotations=frozenset({'boks', 'levering'})),
EmojiAnnotations(emoji='📫', codepoints=(128235,), name='lukket postkasse med heist flagg', slug='lukket_postkasse_med_heist_flagg', annotations=frozenset({'flagg', 'postkasse med heist flagg', 'post'})),
EmojiAnnotations(emoji='📪', codepoints=(128234,), name='lukket postkasse med senket flagg', slug='lukket_postkasse_med_senket_flagg', annotations=frozenset({'postkasse med senket flagg', 'flagg', 'post'})),
EmojiAnnotations(emoji='📬', codepoints=(128236,), name='åpen postkasse med heist flagg', slug='åpen_postkasse_med_heist_flagg', annotations=frozenset({'flagg', 'postkasse med heist flagg', 'post'})),
EmojiAnnotations(emoji='📭', codepoints=(128237,), name='åpen postkasse med senket flagg', slug='åpen_postkasse_med_senket_flagg', annotations=frozenset({'postkasse med senket flagg', 'flagg', 'post'})),
EmojiAnnotations(emoji='📮', codepoints=(128238,), name='postkasse', slug='postkasse', annotations=frozenset({'post'})),
EmojiAnnotations(emoji='\U0001f5f3', codepoints=(128499,), name='urne med stemmeseddel', slug='urne_med_stemmeseddel', annotations=frozenset({'stemmeseddel', 'boks'})),
EmojiAnnotations(emoji='✒', codepoints=(10002,), name='svart spiss', slug='svart_spiss', annotations=frozenset({'penn', 'pennespiss'})),
EmojiAnnotations(emoji='\U0001f58b', codepoints=(128395,), name='fyllepenn', slug='fyllepenn', annotations=frozenset({'penn'})),
EmojiAnnotations(emoji='\U0001f58a', codepoints=(128394,), name='penn', slug='penn', annotations=frozenset({'kulepenn'})),
EmojiAnnotations(emoji='\U0001f58c', codepoints=(128396,), name='pensel', slug='pensel', annotations=frozenset({'maling'})),
EmojiAnnotations(emoji='\U0001f58d', codepoints=(128397,), name='tegnestift', slug='tegnestift', annotations=frozenset({'tegning'})),
EmojiAnnotations(emoji='📝', codepoints=(128221,), name='huskeliste', slug='huskeliste', annotations=frozenset({'bok', 'kommunikasjon'})),
EmojiAnnotations(emoji='💼', codepoints=(128188,), name='koffert', slug='koffert', annotations=frozenset({'briefcase'})),
EmojiAnnotations(emoji='📁', codepoints=(128193,), name='filmappe', slug='filmappe', annotations=frozenset({'mappe', 'fil'})),
EmojiAnnotations(emoji='📂', codepoints=(128194,), name='åpen mappe', slug='åpen_mappe', annotations=frozenset({'åpen filmappe', 'åpen', 'mappe', 'fil'})),
EmojiAnnotations(emoji='\U0001f5c2', codepoints=(128450,), name='skillekort', slug='skillekort', annotations=frozenset({'indeks', 'kort', 'skillere'})),
EmojiAnnotations(emoji='📅', codepoints=(128197,), name='kalender', slug='kalender', annotations=frozenset({'dato'})),
EmojiAnnotations(emoji='📆', codepoints=(128198,), name='riv-av-kalender', slug='riv_av_kalender', annotations=frozenset({'kalender'})),
EmojiAnnotations(emoji='\U0001f5d2', codepoints=(128466,), name='spiralhefte', slug='spiralhefte', annotations=frozenset({'blokk', 'spiral', 'notat'})),
EmojiAnnotations(emoji='\U0001f5d3', codepoints=(128467,), name='spiralkalender', slug='spiralkalender', annotations=frozenset({'blokk', 'spiral', 'kalender'})),
EmojiAnnotations(emoji='📇', codepoints=(128199,), name='kortindeks', slug='kortindeks', annotations=frozenset({'indeks', 'register'})),
EmojiAnnotations(emoji='📈', codepoints=(128200,), name='positiv trend', slug='positiv_trend', annotations=frozenset({'graf', 'positiv', 'positive trend-graf'})),
EmojiAnnotations(emoji='📉', codepoints=(128201,), name='negativ trend', slug='negativ_trend', annotations=frozenset({'negativ trend-graf', 'graf', 'positiv'})),
EmojiAnnotations(emoji='📊', codepoints=(128202,), name='stolpediagram', slug='stolpediagram', annotations=frozenset({'diagram', 'søylediagram'})),
EmojiAnnotations(emoji='📌', codepoints=(128204,), name='stift', slug='stift', annotations=frozenset({'kontorrekvisita'})),
EmojiAnnotations(emoji='📍', codepoints=(128205,), name='knappenål', slug='knappenål', annotations=frozenset({'markørnål', 'kontorrekvisita'})),
EmojiAnnotations(emoji='📎', codepoints=(128206,), name='binders', slug='binders', annotations=frozenset({'kontorrekvisita'})),
EmojiAnnotations(emoji='\U0001f587', codepoints=(128391,), name='sammenlenkede binders', slug='sammenlenkede_binders', annotations=frozenset({'koble sammen', 'binders'})),
EmojiAnnotations(emoji='📏', codepoints=(128207,), name='linjal', slug='linjal', annotations=frozenset({'rett linjal'})),
EmojiAnnotations(emoji='📐', codepoints=(128208,), name='trekantet linjal', slug='trekantet_linjal', annotations=frozenset({'linjal', 'trekant'})),
EmojiAnnotations(emoji='✂', codepoints=(9986,), name='saks', slug='saks', annotations=frozenset({'verktøy'})),
EmojiAnnotations(emoji='\U0001f5c3', codepoints=(128451,), name='kartotekboks', slug='kartotekboks', annotations=frozenset({'boks', 'kort', 'fil'})),
EmojiAnnotations(emoji='\U0001f5c4', codepoints=(128452,), name='arkivskap', slug='arkivskap', annotations=frozenset({'kabinett', 'fil'})),
EmojiAnnotations(emoji='🔒', codepoints=(128274,), name='lås', slug='lås', annotations=frozenset({'låst'})),
EmojiAnnotations(emoji='🔏', codepoints=(128271,), name='lås med penn', slug='lås_med_penn', annotations=frozenset({'personvern', 'penn'})),
EmojiAnnotations(emoji='🔐', codepoints=(128272,), name='lukket lås', slug='lukket_lås', annotations=frozenset({'lås med nøkkel', 'sikker'})),
EmojiAnnotations(emoji='🔑', codepoints=(128273,), name='nøkkel', slug='nøkkel', annotations=frozenset({'passord'})),
EmojiAnnotations(emoji='\U0001f5dd', codepoints=(128477,), name='gammeldags nøkkel', slug='gammeldags_nøkkel', annotations=frozenset({'gammel', 'nøkkel', 'ledetråd', 'lås'})),
EmojiAnnotations(emoji='🔨', codepoints=(128296,), name='hammer', slug='hammer', annotations=frozenset({'verktøy'})),
EmojiAnnotations(emoji='⛏', codepoints=(9935,), name='hakke', slug='hakke', annotations=frozenset({'verktøy', 'gruvedrift'})),
EmojiAnnotations(emoji='⚒', codepoints=(9874,), name='hammer og hakke', slug='hammer_og_hakke', annotations=frozenset({'hammer', 'verktøy', 'hakke'})),
EmojiAnnotations(emoji='\U0001f6e0', codepoints=(128736,), name='hammer og skiftenøkkel', slug='hammer_og_skiftenøkkel', annotations=frozenset({'hammer', 'verktøy', 'skiftenøkkel'})),
EmojiAnnotations(emoji='🔧', codepoints=(128295,), name='skiftenøkkel', slug='skiftenøkkel', annotations=frozenset({'verktøy'})),
EmojiAnnotations(emoji='🔩', codepoints=(128297,), name='mutter og skrue', slug='mutter_og_skrue', annotations=frozenset({'mutter', 'skrue', 'verktøy'})),
EmojiAnnotations(emoji='⚙', codepoints=(9881,), name='tannhjul', slug='tannhjul', annotations=frozenset({'verktøy'})),
EmojiAnnotations(emoji='\U0001f5dc', codepoints=(128476,), name='skrustikke', slug='skrustikke', annotations=frozenset({'verktøy'})),
EmojiAnnotations(emoji='⚗', codepoints=(9879,), name='destillasjonskolbe', slug='destillasjonskolbe', annotations=frozenset({'verktøy', 'kjemi'})),
EmojiAnnotations(emoji='⚖', codepoints=(9878,), name='likearmet skålvekt', slug='likearmet_skålvekt', annotations=frozenset({'stjernetegn', 'vekt', 'balanse', 'vekten', 'rettferdighet', 'verktøy'})),
EmojiAnnotations(emoji='🔗', codepoints=(128279,), name='linksymbol', slug='linksymbol', annotations=frozenset({'to ringer', 'link'})),
EmojiAnnotations(emoji='💉', codepoints=(128137,), name='sprøyte', slug='sprøyte', annotations=frozenset({'vaksine', 'medisin', 'nål', 'lege'})),
EmojiAnnotations(emoji='💊', codepoints=(128138,), name='pille', slug='pille', annotations=frozenset({'medisin', 'lege'})),
EmojiAnnotations(emoji='\U0001f5e1', codepoints=(128481,), name='dolk', slug='dolk', annotations=frozenset({'våpen', 'kniv'})),
EmojiAnnotations(emoji='🔪', codepoints=(128298,), name='kokkekniv', slug='kokkekniv', annotations=frozenset({'kniv'})),
EmojiAnnotations(emoji='⚔', codepoints=(9876,), name='sverd i kryss', slug='sverd_i_kryss', annotations=frozenset({'våpen', 'sverd', 'kryss'})),
EmojiAnnotations(emoji='🔫', codepoints=(128299,), name='pistol', slug='pistol', annotations=frozenset({'revolver'})),
EmojiAnnotations(emoji='\U0001f6e1', codepoints=(128737,), name='skjold', slug='skjold', annotations=frozenset({'våpen'})),
EmojiAnnotations(emoji='\U0001f3f9', codepoints=(127993,), name='pil og bue', slug='pil_og_bue', annotations=frozenset({'stjernetegn', 'bue', 'pil', 'bueskytter', 'verktøy', 'våpen', 'skytten'})),
EmojiAnnotations(emoji='🏁', codepoints=(127937,), name='bilraceflagg', slug='bilraceflagg', annotations=frozenset({'racingflagg', 'motorsport', 'rutete flagg'})),
EmojiAnnotations(emoji='\U0001f3f3', codepoints=(127987,), name='flagrende hvitt flagg', slug='flagrende_hvitt_flagg', annotations=frozenset({'flagring', 'flagg'})),
EmojiAnnotations(emoji='\U0001f3f4', codepoints=(127988,), name='flagrende svart flagg', slug='flagrende_svart_flagg', annotations=frozenset({'flagring', 'flagg'})),
EmojiAnnotations(emoji='🚩', codepoints=(128681,), name='posisjonsflagg', slug='posisjonsflagg', annotations=frozenset({'flagg'})),
EmojiAnnotations(emoji='🚬', codepoints=(128684,), name='røykeskilt', slug='røykeskilt', annotations=frozenset({'røyking', 'skilt', 'røyking tillatt'})),
EmojiAnnotations(emoji='⚰', codepoints=(9904,), name='kiste', slug='kiste', annotations=frozenset({'død'})),
EmojiAnnotations(emoji='⚱', codepoints=(9905,), name='urne', slug='urne', annotations=frozenset({'død', 'begravelse'})),
EmojiAnnotations(emoji='🗿', codepoints=(128511,), name='Moyai-statue', slug='moyai_statue', annotations=frozenset({'statue', 'Moyai'})),
EmojiAnnotations(emoji='\U0001f6e2', codepoints=(128738,), name='oljetønne', slug='oljetønne', annotations=frozenset({'olje', 'tønne'})),
EmojiAnnotations(emoji='🔮', codepoints=(128302,), name='spåkule', slug='spåkule', annotations=frozenset({'spådommer', 'krystallkule', 'spådom'})),
EmojiAnnotations(emoji='🏧', codepoints=(127975,), name='minibank', slug='minibank', annotations=frozenset({'pengeuttak'})),
EmojiAnnotations(emoji='🚮', codepoints=(128686,), name='avfallskilt', slug='avfallskilt', annotations=frozenset({'kast avfal', 'søppelkasse', 'skilt'})),
EmojiAnnotations(emoji='🚰', codepoints=(128688,), name='drikkevann', slug='drikkevann', annotations=frozenset({'drikkevannskilt', 'vann'})),
EmojiAnnotations(emoji='♿', codepoints=(9855,), name='rullestol', slug='rullestol', annotations=frozenset({'rullestolsymbol', 'handikaptilgang', 'rullestolbruker'})),
EmojiAnnotations(emoji='🚹', codepoints=(128697,), name='menn', slug='menn', annotations=frozenset({'herretoalett', 'mannesymbol', 'skilt'})),
EmojiAnnotations(emoji='🚺', codepoints=(128698,), name='kvinner', slug='kvinner', annotations=frozenset({'kvinnesymbol', 'dametoalett', 'skilt'})),
EmojiAnnotations(emoji='🚻', codepoints=(128699,), name='toalettskilt', slug='toalettskilt', annotations=frozenset({'WC', 'do'})),
EmojiAnnotations(emoji='🚼', codepoints=(128700,), name='babyskilt', slug='babyskilt', annotations=frozenset({'stellestasjon', 'babysymbol'})),
EmojiAnnotations(emoji='🚾', codepoints=(128702,), name='do', slug='do', annotations=frozenset({'WC', 'toalett'})),
EmojiAnnotations(emoji='🛂', codepoints=(128706,), name='passkontroll', slug='passkontroll', annotations=frozenset({'pass', 'kontroll'})),
EmojiAnnotations(emoji='🛃', codepoints=(128707,), name='toll', slug='toll', annotations=frozenset({'tollen'})),
EmojiAnnotations(emoji='🛄', codepoints=(128708,), name='rullebånd', slug='rullebånd', annotations=frozenset({'bagasje', 'bagasjebånd'})),
EmojiAnnotations(emoji='🛅', codepoints=(128709,), name='glemt bagasje', slug='glemt_bagasje', annotations=frozenset({'bagasje'})),
EmojiAnnotations(emoji='⚠', codepoints=(9888,), name='advarsel', slug='advarsel', annotations=frozenset({'OBS'})),
EmojiAnnotations(emoji='🚸', codepoints=(128696,), name='barn krysser', slug='barn_krysser', annotations=frozenset({'krysser', 'barn', 'skilt'})),
EmojiAnnotations(emoji='⛔', codepoints=(9940,), name='ingen adgang', slug='ingen_adgang', annotations=frozenset({'adgang forbudt', 'stopp'})),
EmojiAnnotations(emoji='🚫', codepoints=(128683,), name='forbudt', slug='forbudt', annotations=frozenset({'adgang forbudt', 'skilt'})),
EmojiAnnotations(emoji='🚳', codepoints=(128691,), name='sykler forbudt', slug='sykler_forbudt', annotations=frozenset({'sykkel ikke tillatt', 'ingen sykler', 'sykkel', 'kjøretøy'})),
EmojiAnnotations(emoji='🚭', codepoints=(128685,), name='røyking forbudt', slug='røyking_forbudt', annotations=frozenset({'ingen røyking', 'skilt'})),
EmojiAnnotations(emoji='🚯', codepoints=(128687,), name='ikke kast avfall', slug='ikke_kast_avfall', annotations=frozenset({'ikke avfall', 'ingen kasting', 'skilt'})),
EmojiAnnotations(emoji='🚱', codepoints=(128689,), name='ikke drikkevann', slug='ikke_drikkevann', annotations=frozenset({'vann', 'ikke drikkbart'})),
EmojiAnnotations(emoji='🚷', codepoints=(128695,), name='fotgjengere forbudt', slug='fotgjengere_forbudt', annotations=frozenset({'fotgjenger', 'gående', 'ingen fotgjengere', 'skilt', 'fotgjengere ikke tillatt'})),
EmojiAnnotations(emoji='☢', codepoints=(9762,), name='radioaktivt', slug='radioaktivt', annotations=frozenset({'radioaktiv'})),
EmojiAnnotations(emoji='⬆', codepoints=(11014,), name='pil opp', slug='pil_opp', annotations=frozenset({'nord', 'pil', 'retning'})),
EmojiAnnotations(emoji='↗', codepoints=(8599,), name='pil opp-høyre', slug='pil_opp_høyre', annotations=frozenset({'pil', 'retning', 'nordøst'})),
EmojiAnnotations(emoji='➡', codepoints=(10145,), name='pil høyre', slug='pil_høyre', annotations=frozenset({'øst', 'pil', 'retning'})),
EmojiAnnotations(emoji='↘', codepoints=(8600,), name='pil ned-høyre', slug='pil_ned_høyre', annotations=frozenset({'pil', 'retning', 'sørøst'})),
EmojiAnnotations(emoji='⬇', codepoints=(11015,), name='pil ned', slug='pil_ned', annotations=frozenset({'pil', 'retning', 'sør'})),
EmojiAnnotations(emoji='↙', codepoints=(8601,), name='pil ned-venstre', slug='pil_ned_venstre', annotations=frozenset({'pil', 'retning', 'sørvest'})),
EmojiAnnotations(emoji='⬅', codepoints=(11013,), name='pil venstre', slug='pil_venstre', annotations=frozenset({'pil', 'retning', 'vest'})),
EmojiAnnotations(emoji='↖', codepoints=(8598,), name='pil opp-venstre', slug='pil_opp_venstre', annotations=frozenset({'nordvest', 'pil', 'retning'})),
EmojiAnnotations(emoji='↕', codepoints=(8597,), name='pil opp-ned', slug='pil_opp_ned', annotations=frozenset({'pil'})),
EmojiAnnotations(emoji='↔', codepoints=(8596,), name='pil venstre-høyre', slug='pil_venstre_høyre', annotations=frozenset({'pil'})),
EmojiAnnotations(emoji='↩', codepoints=(8617,), name='pil som bøyer mot venstre', slug='pil_som_bøyer_mot_venstre', annotations=frozenset({'pil'})),
EmojiAnnotations(emoji='↪', codepoints=(8618,), name='pil som bøyer mot høyre', slug='pil_som_bøyer_mot_høyre', annotations=frozenset({'pil'})),
EmojiAnnotations(emoji='⤴', codepoints=(10548,), name='pil som bøyer opp', slug='pil_som_bøyer_opp', annotations=frozenset({'pil'})),
EmojiAnnotations(emoji='⤵', codepoints=(10549,), name='pil som bøyer ned', slug='pil_som_bøyer_ned', annotations=frozenset({'pil', 'ned'})),
EmojiAnnotations(emoji='🔃', codepoints=(128259,), name='piler i sirkel', slug='piler_i_sirkel', annotations=frozenset({'vertikale piler med klokken', 'piler', 'last inn på nytt'})),
EmojiAnnotations(emoji='🔄', codepoints=(128260,), name='piler mot klokken', slug='piler_mot_klokken', annotations=frozenset({'piler', 'mot klokken'})),
EmojiAnnotations(emoji='🔙', codepoints=(128281,), name='tilbake', slug='tilbake', annotations=frozenset({'tilbakepil', 'pil', 'pil til venstre'})),
EmojiAnnotations(emoji='🔚', codepoints=(128282,), name='slutt', slug='slutt', annotations=frozenset({'sluttpil', 'slutt med pil', 'pil', 'pil til venstre'})),
EmojiAnnotations(emoji='🔛', codepoints=(128283,), name='PÅ-pil', slug='på_pil', annotations=frozenset({'PÅ med utropstegn og pil', 'utropstegn', 'utrop', 'pil'})),
EmojiAnnotations(emoji='🔜', codepoints=(128284,), name='snart-tegn', slug='snart_tegn', annotations=frozenset({'snart', 'pil', 'snart med pil'})),
EmojiAnnotations(emoji='🔝', codepoints=(128285,), name='toppsymbol', slug='toppsymbol', annotations=frozenset({'topp med pil', 'pil opp', 'øverst', 'opp'})),
EmojiAnnotations(emoji='\U0001f6d0', codepoints=(128720,), name='sted for tilbedelse', slug='sted_for_tilbedelse', annotations=frozenset({'tilbedelse', 'religion'})),
EmojiAnnotations(emoji='⚛', codepoints=(9883,), name='atom', slug='atom', annotations=frozenset({'ateist'})),
EmojiAnnotations(emoji='\U0001f549', codepoints=(128329,), name='om', slug='om', annotations=frozenset({'hindu', 'religion'})),
EmojiAnnotations(emoji='✡', codepoints=(10017,), name='Davidstjernen', slug='davidstjernen', annotations=frozenset({'jøde', 'david', 'jødisk', 'religion', 'stjerne'})),
EmojiAnnotations(emoji='☸', codepoints=(9784,), name='dharmahjulet', slug='dharmahjulet', annotations=frozenset({'dharma', 'hjul', 'buddhist', 'religion'})),
EmojiAnnotations(emoji='☯', codepoints=(9775,), name='yin yang', slug='yin_yang', annotations=frozenset({'yin', 'taoist', 'religion', 'tao', 'yang'})),
EmojiAnnotations(emoji='✝', codepoints=(10013,), name='latinsk kors', slug='latinsk_kors', annotations=frozenset({'kors', 'religion', 'kristen'})),
EmojiAnnotations(emoji='☦', codepoints=(9766,), name='ortodoks kors', slug='ortodoks_kors', annotations=frozenset({'kors', 'religion', 'kristen'})),
EmojiAnnotations(emoji='☪', codepoints=(9770,), name='stjerne og halvmåne', slug='stjerne_og_halvmåne', annotations=frozenset({'muslim', 'islam', 'religion'})),
EmojiAnnotations(emoji='☮', codepoints=(9774,), name='fredssymbol', slug='fredssymbol', annotations=frozenset({'fred'})),
EmojiAnnotations(emoji='\U0001f54e', codepoints=(128334,), name='menora', slug='menora', annotations=frozenset({'lysestake', 'religion'})),
EmojiAnnotations(emoji='🔯', codepoints=(128303,), name='sekskantet stjerne', slug='sekskantet_stjerne', annotations=frozenset({'sekskantet', 'magi', 'stjerne'})),
EmojiAnnotations(emoji='♻', codepoints=(9851,), name='resirkulering', slug='resirkulering', annotations=frozenset({'resirkuleringssymbol', 'universelt', 'symbol'})),
EmojiAnnotations(emoji='📛', codepoints=(128219,), name='navneskilt', slug='navneskilt', annotations=frozenset({'navn', 'skilt'})),
EmojiAnnotations(emoji='⚜', codepoints=(9884,), name='fransk lilje', slug='fransk_lilje', annotations=frozenset({'fleur-de-lis'})),
EmojiAnnotations(emoji='🔰', codepoints=(128304,), name='grønt og gult blad', slug='grønt_og_gult_blad', annotations=frozenset({'grønn og gul', 'japansk', 'nybegynner'})),
EmojiAnnotations(emoji='🔱', codepoints=(128305,), name='tridentsymbol', slug='tridentsymbol', annotations=frozenset({'emblem', 'treenighet', 'anker'})),
EmojiAnnotations(emoji='⭕', codepoints=(11093,), name='fremhevet stor sirkel', slug='fremhevet_stor_sirkel', annotations=frozenset({'o', 'sirkel'})),
EmojiAnnotations(emoji='✅', codepoints=(9989,), name='hake', slug='hake', annotations=frozenset({'avmerking', 'avkryssing'})),
EmojiAnnotations(emoji='☑', codepoints=(9745,), name='stemmeurne med hakemerke', slug='stemmeurne_med_hakemerke', annotations=frozenset({'stemmeseddel', 'boks', 'hakemerke'})),
EmojiAnnotations(emoji='✔', codepoints=(10004,), name='fremhevet hakemerke', slug='fremhevet_hakemerke', annotations=frozenset({'merke', 'hakemerke', 'hake'})),
EmojiAnnotations(emoji='✖', codepoints=(10006,), name='fremhevet gangetegn', slug='fremhevet_gangetegn', annotations=frozenset({'multiplikasjon', 'x', 'avbryt', 'multiplisere'})),
EmojiAnnotations(emoji='❌', codepoints=(10060,), name='kryss', slug='kryss', annotations=frozenset({'rødt kryss', '«X»'})),
EmojiAnnotations(emoji='❎', codepoints=(10062,), name='kryss i firkant', slug='kryss_i_firkant', annotations=frozenset({'kryss'})),
EmojiAnnotations(emoji='➕', codepoints=(10133,), name='pluss', slug='pluss', annotations=frozenset({'plusstegn'})),
EmojiAnnotations(emoji='➖', codepoints=(10134,), name='minus', slug='minus', annotations=frozenset({'minustegn'})),
EmojiAnnotations(emoji='➗', codepoints=(10135,), name='dele', slug='dele', annotations=frozenset({'divisjonstegn', 'deletegn'})),
EmojiAnnotations(emoji='➰', codepoints=(10160,), name='løkke', slug='løkke', annotations=frozenset({'løkketegn'})),
EmojiAnnotations(emoji='➿', codepoints=(10175,), name='dobbel løkke', slug='dobbel_løkke', annotations=frozenset({'dobbel løkke-tegn'})),
EmojiAnnotations(emoji='〽', codepoints=(12349,), name='delvis alternering', slug='delvis_alternering', annotations=frozenset({'delvis', 'merke'})),
EmojiAnnotations(emoji='✳', codepoints=(10035,), name='stjerne med åtte eiker', slug='stjerne_med_åtte_eiker', annotations=frozenset({'asterisk'})),
EmojiAnnotations(emoji='✴', codepoints=(10036,), name='åttekantet stjerne', slug='åttekantet_stjerne', annotations=frozenset({'stjerne'})),
EmojiAnnotations(emoji='💱', codepoints=(128177,), name='valutaveksling', slug='valutaveksling', annotations=frozenset({'valuta', 'veksling', 'penger'})),
EmojiAnnotations(emoji='💲', codepoints=(128178,), name='dollartegn', slug='dollartegn', annotations=frozenset({'dollar', 'stort dollartegn', 'penger'})),
EmojiAnnotations(emoji='‼', codepoints=(8252,), name='to utropstegn', slug='to_utropstegn', annotations=frozenset({'tegn', 'utrop', 'utropstegn', 'tegnsetting', 'bangbang'})),
EmojiAnnotations(emoji='⁉', codepoints=(8265,), name='utropstegn og spørsmålstegn', slug='utropstegn_og_spørsmålstegn', annotations=frozenset({'spørrende', 'spørsmål', 'utropstegn', 'tegnsetting', 'tegn', 'utrop', 'spørsmålstegn'})),
EmojiAnnotations(emoji='❓', codepoints=(10067,), name='rødt spørsmålstegn', slug='rødt_spørsmålstegn', annotations=frozenset({'spørsmål'})),
EmojiAnnotations(emoji='❔', codepoints=(10068,), name='hvitt spørsmålstegn', slug='hvitt_spørsmålstegn', annotations=frozenset({'spørsmål'})),
EmojiAnnotations(emoji='❕', codepoints=(10069,), name='hvitt utropstegn', slug='hvitt_utropstegn', annotations=frozenset({'utrop'})),
EmojiAnnotations(emoji='❗', codepoints=(10071,), name='utropstegn', slug='utropstegn', annotations=frozenset({'tegn', 'utrop', 'tegnsetting'})),
EmojiAnnotations(emoji='〰', codepoints=(12336,), name='bølgestrek', slug='bølgestrek', annotations=frozenset({'tegn for lang lyd', 'bølget bindestrek', 'bølge'})),
EmojiAnnotations(emoji='®', codepoints=(174,), name='registrert varemerke', slug='registrert_varemerke', annotations=frozenset({'varemerkesymbol'})),
EmojiAnnotations(emoji='♈', codepoints=(9800,), name='Væren', slug='væren', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♉', codepoints=(9801,), name='Tyren', slug='tyren', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♊', codepoints=(9802,), name='Tvillingene', slug='tvillingene', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♋', codepoints=(9803,), name='Krepsen', slug='krepsen', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♌', codepoints=(9804,), name='Løven', slug='løven', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♍', codepoints=(9805,), name='Jomfruen', slug='jomfruen', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♎', codepoints=(9806,), name='Vekten', slug='vekten', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♏', codepoints=(9807,), name='Skorpionen', slug='skorpionen', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♐', codepoints=(9808,), name='Skytten', slug='skytten', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♑', codepoints=(9809,), name='Steinbukken', slug='steinbukken', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♒', codepoints=(9810,), name='Vannmannen', slug='vannmannen', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='♓', codepoints=(9811,), name='Fiskene', slug='fiskene', annotations=frozenset({'stjernetegn'})),
EmojiAnnotations(emoji='⛎', codepoints=(9934,), name='stjernetegn', slug='stjernetegn', annotations=frozenset({'Ophiuchus', 'astrologi', 'Slangebæreren'})),
EmojiAnnotations(emoji='🔀', codepoints=(128256,), name='kryssede piler', slug='kryssede_piler', annotations=frozenset({'piler', 'kryssede piler mot høyre'})),
EmojiAnnotations(emoji='🔁', codepoints=(128257,), name='gjenta', slug='gjenta', annotations=frozenset({'piler', 'med klokken'})),
EmojiAnnotations(emoji='🔂', codepoints=(128258,), name='gjenta én gang', slug='gjenta_én_gang', annotations=frozenset({'piler', 'piler med klokken med nummer 1 med klokken', 'nummer 1'})),
EmojiAnnotations(emoji='▶', codepoints=(9654,), name='spill av', slug='spill_av', annotations=frozenset({'triangel', 'høyre', 'pil'})),
EmojiAnnotations(emoji='⏩', codepoints=(9193,), name='spol kjapt fremover', slug='spol_kjapt_fremover', annotations=frozenset({'spoling', 'høyre', 'dobbel pil'})),
EmojiAnnotations(emoji='⏭', codepoints=(9197,), name='neste spor', slug='neste_spor', annotations=frozenset({'triangel', 'neste scene', 'pil'})),
EmojiAnnotations(emoji='⏯', codepoints=(9199,), name='spill av/pause-knapp', slug='spill_av/pause_knapp', annotations=frozenset({'triangel', 'høyre', 'pil', 'pause', 'spill av'})),
EmojiAnnotations(emoji='◀', codepoints=(9664,), name='spol tilbake', slug='spol_tilbake', annotations=frozenset({'triangel', 'venstre', 'pil', 'bakover'})),
EmojiAnnotations(emoji='⏪', codepoints=(9194,), name='spol kjapt tilbake', slug='spol_kjapt_tilbake', annotations=frozenset({'spoling', 'venstre', 'dobbel pil'})),
EmojiAnnotations(emoji='⏮', codepoints=(9198,), name='forrige spor', slug='forrige_spor', annotations=frozenset({'triangel', 'pil', 'forrige scene'})),
EmojiAnnotations(emoji='🔼', codepoints=(128316,), name='gå opp', slug='gå_opp', annotations=frozenset({'oppover'})),
EmojiAnnotations(emoji='⏫', codepoints=(9195,), name='gå opp kjapt', slug='gå_opp_kjapt', annotations=frozenset({'dobbelt pil opp', 'pil', 'opp'})),
EmojiAnnotations(emoji='🔽', codepoints=(128317,), name='gå ned', slug='gå_ned', annotations=frozenset({'nedover'})),
EmojiAnnotations(emoji='⏬', codepoints=(9196,), name='gå ned kjapt', slug='gå_ned_kjapt', annotations=frozenset({'pil', 'ned', 'dobbelt pil ned'})),
EmojiAnnotations(emoji='\u23f8', codepoints=(9208,), name='pause', slug='pause', annotations=frozenset({'strek', 'dobbel', 'vertikal'})),
EmojiAnnotations(emoji='\u23f9', codepoints=(9209,), name='stopp', slug='stopp', annotations=frozenset({'firkant'})),
EmojiAnnotations(emoji='\u23fa', codepoints=(9210,), name='spill inn', slug='spill_inn', annotations=frozenset({'sirkel'})),
EmojiAnnotations(emoji='🎦', codepoints=(127910,), name='kino', slug='kino', annotations=frozenset({'kamera', 'film', 'filmkamera', 'underholdning'})),
EmojiAnnotations(emoji='🔅', codepoints=(128261,), name='lite lys', slug='lite_lys', annotations=frozenset({'lysstyrke', 'lav', 'lav lysstyrke'})),
EmojiAnnotations(emoji='🔆', codepoints=(128262,), name='mye lys', slug='mye_lys', annotations=frozenset({'høy', 'lysstyrke', 'høy lysstyrke'})),
EmojiAnnotations(emoji='📶', codepoints=(128246,), name='mobilsignaler', slug='mobilsignaler', annotations=frozenset({'telefon', 'antenne', 'signalstyrke', 'signal', 'mobil'})),
EmojiAnnotations(emoji='📵', codepoints=(128245,), name='mobil forbudt', slug='mobil_forbudt', annotations=frozenset({'ingen telefonsamtaler', 'telefon', 'mobil ikke tillatt'})),
EmojiAnnotations(emoji='📳', codepoints=(128243,), name='vibreringsmodus', slug='vibreringsmodus', annotations=frozenset({'modus', 'mobiltelefon', 'mobil', 'telefon'})),
EmojiAnnotations(emoji='📴', codepoints=(128244,), name='telefon av', slug='telefon_av', annotations=frozenset({'av', 'mobiltelefon', 'mobil', 'mobiltelefon av'})),
EmojiAnnotations(emoji='{#⃣}', codepoints=(123, 35, 8419, 125), name='nummertast nummertegn', slug='nummertast_nummertegn', annotations=frozenset({'nummertegn'})),
EmojiAnnotations(emoji='{*⃣}', codepoints=(123, 42, 8419, 125), name='nummertast stjerne', slug='nummertast_stjerne', annotations=frozenset({'asterisk', 'stjerne', 'keycap'})),
EmojiAnnotations(emoji='{0⃣}', codepoints=(123, 48, 8419, 125), name='nummertast null', slug='nummertast_null', annotations=frozenset({'0', 'null', 'nummertast'})),
EmojiAnnotations(emoji='{1⃣}', codepoints=(123, 49, 8419, 125), name='nummertast én', slug='nummertast_én', annotations=frozenset({'én', 'nummertast', '1'})),
EmojiAnnotations(emoji='{2⃣}', codepoints=(123, 50, 8419, 125), name='nummertast to', slug='nummertast_to', annotations=frozenset({'nummertast', '2', 'to'})),
EmojiAnnotations(emoji='{3⃣}', codepoints=(123, 51, 8419, 125), name='nummertast tre', slug='nummertast_tre', annotations=frozenset({'tre', 'nummertast', '3'})),
EmojiAnnotations(emoji='{4⃣}', codepoints=(123, 52, 8419, 125), name='nummertast fire', slug='nummertast_fire', annotations=frozenset({'4', 'fire', 'nummertast'})),
EmojiAnnotations(emoji='{5⃣}', codepoints=(123, 53, 8419, 125), name='nummertast fem', slug='nummertast_fem', annotations=frozenset({'5', 'fem', 'nummertast'})),
EmojiAnnotations(emoji='{6⃣}', codepoints=(123, 54, 8419, 125), name='nummertast seks', slug='nummertast_seks', annotations=frozenset({'nummertast', 'seks', '6'})),
EmojiAnnotations(emoji='{7⃣}', codepoints=(123, 55, 8419, 125), name='nummertast sju', slug='nummertast_sju', annotations=frozenset({'7', 'nummertast', 'syv'})),
EmojiAnnotations(emoji='{8⃣}', codepoints=(123, 56, 8419, 125), name='nummertast åtte', slug='nummertast_åtte', annotations=frozenset({'8', 'åtte', 'nummertast'})),
EmojiAnnotations(emoji='{9⃣}', codepoints=(123, 57, 8419, 125), name='nummertast ni', slug='nummertast_ni', annotations=frozenset({'nummertast', 'ni', '9'})),
EmojiAnnotations(emoji='🔟', codepoints=(128287,), name='nummertast ti', slug='nummertast_ti', annotations=frozenset({'10', 'nummertast', 'ti'})),
EmojiAnnotations(emoji='💯', codepoints=(128175,), name='hundre', slug='hundre', annotations=frozenset({'poeng', 'full pott', '100 poeng'})),
EmojiAnnotations(emoji='🔞', codepoints=(128286,), name='ingen under 18', slug='ingen_under_18', annotations=frozenset({'18', 'under 18', 'aldersgrense 18 år'})),
EmojiAnnotations(emoji='🔠', codepoints=(128288,), name='store bokstaver', slug='store_bokstaver', annotations=frozenset({'inndata'})),
EmojiAnnotations(emoji='🔡', codepoints=(128289,), name='små bokstaver', slug='små_bokstaver', annotations=frozenset({'inndata'})),
EmojiAnnotations(emoji='🔢', codepoints=(128290,), name='tall', slug='tall', annotations=frozenset({'inndata', 'numeriske data'})),
EmojiAnnotations(emoji='🔣', codepoints=(128291,), name='symboler', slug='symboler', annotations=frozenset({'inndata'})),
EmojiAnnotations(emoji='🔤', codepoints=(128292,), name='latinsk alfabet', slug='latinsk_alfabet', annotations=frozenset({'inndata', 'ABC'})),
EmojiAnnotations(emoji='🅰', codepoints=(127344,), name='blodtype A', slug='blodtype_a', annotations=frozenset({'blod', 'type A', 'blodtype', 'A'})),
EmojiAnnotations(emoji='🆎', codepoints=(127374,), name='blodtype AB', slug='blodtype_ab', annotations=frozenset({'type AB', 'blod', 'blodtype', 'AB'})),
EmojiAnnotations(emoji='🅱', codepoints=(127345,), name='blodtype B', slug='blodtype_b', annotations=frozenset({'blod', 'type B', 'B', 'blodtype'})),
EmojiAnnotations(emoji='🆑', codepoints=(127377,), name='CL', slug='cl', annotations=frozenset({'tøm-symbol', 'clear-symbol'})),
EmojiAnnotations(emoji='🆒', codepoints=(127378,), name='cool', slug='cool', annotations=frozenset({'cool-symbol'})),
EmojiAnnotations(emoji='🆓', codepoints=(127379,), name='gratis-symbol', slug='gratis_symbol', annotations=frozenset({'free', 'free-symbol'})),
EmojiAnnotations(emoji='ℹ', codepoints=(8505,), name='informasjonskilde', slug='informasjonskilde', annotations=frozenset({'i', 'informasjon'})),
EmojiAnnotations(emoji='🆔', codepoints=(127380,), name='ID', slug='id', annotations=frozenset({'identitet', 'ID-symbol'})),
EmojiAnnotations(emoji='Ⓜ', codepoints=(9410,), name='bokstaven m med ring rundt', slug='bokstaven_m_med_ring_rundt', annotations=frozenset({'m', 'sirkel'})),
EmojiAnnotations(emoji='🆕', codepoints=(127381,), name='nytt-symbol', slug='nytt_symbol', annotations=frozenset({'new-symbol', 'new'})),
EmojiAnnotations(emoji='🆖', codepoints=(127382,), name='No good', slug='no_good', annotations=frozenset({'ikke bra', 'NG-symbol', 'NG'})),
EmojiAnnotations(emoji='🅾', codepoints=(127358,), name='blodtype O', slug='blodtype_o', annotations=frozenset({'blod', 'type O', 'O', 'blodtype'})),
EmojiAnnotations(emoji='🆗', codepoints=(127383,), name='OK', slug='ok', annotations=frozenset({'OK-symbol'})),
EmojiAnnotations(emoji='🅿', codepoints=(127359,), name='parkering', slug='parkering', annotations=frozenset({'parkeringsplass'})),
EmojiAnnotations(emoji='🆘', codepoints=(127384,), name='SOS', slug='sos', annotations=frozenset({'SOS-symbol', 'hjelp', 'hjelp-symbol'})),
EmojiAnnotations(emoji='🆙', codepoints=(127385,), name='UP!', slug='up', annotations=frozenset({'UP!-symbol', 'opp-symbol', 'opp'})),
EmojiAnnotations(emoji='🆚', codepoints=(127386,), name='versus', slug='versus', annotations=frozenset({'vs', 'vs-symbol'})),
EmojiAnnotations(emoji='🈁', codepoints=(127489,), name='katakana som leses «koko»', slug='katakana_som_leses_«koko»', annotations=frozenset({'«koko»', 'katakana'})),
EmojiAnnotations(emoji='🈂', codepoints=(127490,), name='katakana som leses «sa»', slug='katakana_som_leses_«sa»', annotations=frozenset({'«sa»', 'stavelse', 'katakana'})),
EmojiAnnotations(emoji='🈷', codepoints=(127543,), name='ideograf for «måne»', slug='ideograf_for_«måne»', annotations=frozenset({'ideograf', 'måne'})),
EmojiAnnotations(emoji='🈶', codepoints=(127542,), name='ideograf for «har»', slug='ideograf_for_«har»', annotations=frozenset({'å ha', 'ideograf', 'være'})),
EmojiAnnotations(emoji='🈯', codepoints=(127535,), name='ideograf for «tildelt»', slug='ideograf_for_«tildelt»', annotations=frozenset({'ideograf', 'finger', 'reservert', 'utpekt'})),
EmojiAnnotations(emoji='🉐', codepoints=(127568,), name='ideograf for «spesialtilbud»', slug='ideograf_for_«spesialtilbud»', annotations=frozenset({'ideograf', 'spesialtilbud'})),
EmojiAnnotations(emoji='🈹', codepoints=(127545,), name='ideograf for «rabatt»', slug='ideograf_for_«rabatt»', annotations=frozenset({'dele', 'ideograf', 'rabatt'})),
EmojiAnnotations(emoji='🈚', codepoints=(127514,), name='ideograf for «har ikke»', slug='ideograf_for_«har_ikke»', annotations=frozenset({'ideograf', 'har ikke', 'er ikke'})),
EmojiAnnotations(emoji='🈲', codepoints=(127538,), name='ideograf for forbudt', slug='ideograf_for_forbudt', annotations=frozenset({'ideograf', 'forbudt'})),
EmojiAnnotations(emoji='🉑', codepoints=(127569,), name='ideograf for «kan» eller «bestått»', slug='ideograf_for_«kan»_eller_«bestått»', annotations=frozenset({'ideograf', 'bestått', 'kan'})),
EmojiAnnotations(emoji='🈸', codepoints=(127544,), name='ideograf for «melde seg på»', slug='ideograf_for_«melde_seg_på»', annotations=frozenset({'ideograf', 'melde seg på'})),
EmojiAnnotations(emoji='🈴', codepoints=(127540,), name='ideograf for «sammen»', slug='ideograf_for_«sammen»', annotations=frozenset({'ideograf', 'sammen'})),
EmojiAnnotations(emoji='🈳', codepoints=(127539,), name='ideograf for «ledig» eller «tom»', slug='ideograf_for_«ledig»_eller_«tom»', annotations=frozenset({'tom', 'ideograf', 'ledig'})),
EmojiAnnotations(emoji='㊗', codepoints=(12951,), name='sirkulær begrepsrune for gratulasjon', slug='sirkulær_begrepsrune_for_gratulasjon', annotations=frozenset({'kinesisk', 'gratulasjon', 'ideografi'})),
EmojiAnnotations(emoji='㊙', codepoints=(12953,), name='sirkulær begrepsrune for hemmelighet', slug='sirkulær_begrepsrune_for_hemmelighet', annotations=frozenset({'kinesisk', 'hemmelig', 'ideografi'})),
EmojiAnnotations(emoji='🈺', codepoints=(127546,), name='ideograf for «åpen for forretninger»', slug='ideograf_for_«åpen_for_forretninger»', annotations=frozenset({'ideograf', 'åpen', 'driver forretninger'})),
EmojiAnnotations(emoji='🈵', codepoints=(127541,), name='ideograf for «full»', slug='ideograf_for_«full»', annotations=frozenset({'ideograf', 'full'})),
EmojiAnnotations(emoji='▪', codepoints=(9642,), name='liten, svart firkant', slug='liten,_svart_firkant', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='▫', codepoints=(9643,), name='liten, hvit firkant', slug='liten,_hvit_firkant', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='◻', codepoints=(9723,), name='hvit firkant av middels størrelse', slug='hvit_firkant_av_middels_størrelse', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='◼', codepoints=(9724,), name='svart firkant av middels størrelse', slug='svart_firkant_av_middels_størrelse', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='◽', codepoints=(9725,), name='hvit firkant av middels/liten størrelse', slug='hvit_firkant_av_middels/liten_størrelse', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='◾', codepoints=(9726,), name='svart firkant av middels/liten størrelse', slug='svart_firkant_av_middels/liten_størrelse', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='⬛', codepoints=(11035,), name='stor, svart firkant', slug='stor,_svart_firkant', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='⬜', codepoints=(11036,), name='stor, hvit firkant', slug='stor,_hvit_firkant', annotations=frozenset({'geometrisk', 'firkant'})),
EmojiAnnotations(emoji='🔶', codepoints=(128310,), name='stort oransje rutersymbol', slug='stort_oransje_rutersymbol', annotations=frozenset({'rutersymbol', 'oransje'})),
EmojiAnnotations(emoji='🔷', codepoints=(128311,), name='stort blått rutersymbol', slug='stort_blått_rutersymbol', annotations=frozenset({'rutersymbol', 'blått'})),
EmojiAnnotations(emoji='🔸', codepoints=(128312,), name='lite oransje rutersymbol', slug='lite_oransje_rutersymbol', annotations=frozenset({'rutersymbol', 'oransje'})),
EmojiAnnotations(emoji='🔹', codepoints=(128313,), name='lite blått rutersymbol', slug='lite_blått_rutersymbol', annotations=frozenset({'rutersymbol', 'blått'})),
EmojiAnnotations(emoji='🔺', codepoints=(128314,), name='rød trekant som peker opp', slug='rød_trekant_som_peker_opp', annotations=frozenset({'oppover'})),
EmojiAnnotations(emoji='🔻', codepoints=(128315,), name='rød trekant som peker ned', slug='rød_trekant_som_peker_ned', annotations=frozenset({'nedover'})),
EmojiAnnotations(emoji='💠', codepoints=(128160,), name='diamantform', slug='diamantform', annotations=frozenset({'søtt', 'blomst', 'diamant'})),
EmojiAnnotations(emoji='🔘', codepoints=(128280,), name='radioknapp', slug='radioknapp', annotations=frozenset({'knapp', 'radio'})),
EmojiAnnotations(emoji='🔲', codepoints=(128306,), name='svart, firkantet knapp', slug='svart,_firkantet_knapp', annotations=frozenset({'knapp', 'firkant'})),
EmojiAnnotations(emoji='🔳', codepoints=(128307,), name='hvit, firkantet knapp', slug='hvit,_firkantet_knapp', annotations=frozenset({'knapp', 'firkant'})),
EmojiAnnotations(emoji='⚪', codepoints=(9898,), name='hvit sirkel', slug='hvit_sirkel', annotations=frozenset({'geometrisk', 'sirkel'})),
EmojiAnnotations(emoji='⚫', codepoints=(9899,), name='svart sirkel', slug='svart_sirkel', annotations=frozenset({'geometrisk', 'sirkel'})),
EmojiAnnotations(emoji='🔴', codepoints=(128308,), name='stor, rød sirkel', slug='stor,_rød_sirkel', annotations=frozenset({'stor rød sirkel', 'rød', 'sirkel'})),
EmojiAnnotations(emoji='🔵', codepoints=(128309,), name='stor, blå sirkel', slug='stor,_blå_sirkel', annotations=frozenset({'stor blå sirkel', 'blå', 'sirkel'})),] | 0 | 0 | 0 |
2c589918bd089b9df29d1ef24cd732b1cc0b49ab | 850 | py | Python | JumpscaleCore/sal/tls/TLSFactory.py | gneumann333/jumpscaleX_core | 777d249fa3668c6e802c2f765f4b82fb39c3e5fa | [
"Apache-2.0"
] | 1 | 2020-06-21T11:18:52.000Z | 2020-06-21T11:18:52.000Z | JumpscaleCore/sal/tls/TLSFactory.py | gneumann333/jumpscaleX_core | 777d249fa3668c6e802c2f765f4b82fb39c3e5fa | [
"Apache-2.0"
] | 644 | 2019-08-25T10:19:56.000Z | 2020-12-23T09:41:04.000Z | JumpscaleCore/sal/tls/TLSFactory.py | gneumann333/jumpscaleX_core | 777d249fa3668c6e802c2f765f4b82fb39c3e5fa | [
"Apache-2.0"
] | 11 | 2019-08-29T21:38:50.000Z | 2020-06-21T11:18:55.000Z | from Jumpscale import j
from .TLS import TLS
JSBASE = j.baseclasses.object
TESTTOOLS = j.baseclasses.testtools
class TLSFactory(JSBASE, TESTTOOLS):
"""Factory class to deal with TLS, key and certificate generation"""
__jslocation__ = "j.sal.tls"
def get(self, path=None):
"""Get an instance of the TLS class
This module use the cfssl AYS.
:param path: Path is the working directory where the certificate and key will be generated, defaults to None
:type path: string, optional
:return: TLS instance
:rtype: TLS class
"""
return TLS(path=path)
def test(self, name=""):
"""Run tests under tests
:param name: basename of the file to run, defaults to ''.
:type name: str, optional
"""
self._tests_run(name=name, die=True)
| 27.419355 | 116 | 0.635294 | from Jumpscale import j
from .TLS import TLS
JSBASE = j.baseclasses.object
TESTTOOLS = j.baseclasses.testtools
class TLSFactory(JSBASE, TESTTOOLS):
"""Factory class to deal with TLS, key and certificate generation"""
__jslocation__ = "j.sal.tls"
def get(self, path=None):
"""Get an instance of the TLS class
This module use the cfssl AYS.
:param path: Path is the working directory where the certificate and key will be generated, defaults to None
:type path: string, optional
:return: TLS instance
:rtype: TLS class
"""
return TLS(path=path)
def test(self, name=""):
"""Run tests under tests
:param name: basename of the file to run, defaults to ''.
:type name: str, optional
"""
self._tests_run(name=name, die=True)
| 0 | 0 | 0 |
5e89c9f921c50d95ca9ada417987375b4027a411 | 2,670 | py | Python | src/cli/install.py | furukawaTakumi/kamaboko | 92dd31dc03c3735dca41b33a7166f6861450e5d9 | [
"MIT"
] | null | null | null | src/cli/install.py | furukawaTakumi/kamaboko | 92dd31dc03c3735dca41b33a7166f6861450e5d9 | [
"MIT"
] | null | null | null | src/cli/install.py | furukawaTakumi/kamaboko | 92dd31dc03c3735dca41b33a7166f6861450e5d9 | [
"MIT"
] | null | null | null |
import argparse, os, csv, json, datetime, pathlib
from typing import Iterator
import kamaboko | 25.92233 | 147 | 0.585019 |
import argparse, os, csv, json, datetime, pathlib
from typing import Iterator
import kamaboko
def install(args = None):
if args == None:
args = __parse_args()
__check_args(args)
with open(args.dic_path, 'r') as f:
data = csv.reader(f, delimiter=__delimiter(args.file_format))
dictionary = __construct_dict(data, args)
save_dir = f"{os.path.dirname(kamaboko.__file__)}/resource/{args.dic_type}"
filename = args.dic_path.split('/')[-1]
pathlib.Path(save_dir).mkdir(parents=True, exist_ok=True)
with open(f"{save_dir}/{filename}", 'w') as f:
json.dump(dictionary, f)
print(f"'{args.dic_path}' save to '{save_dir}'. complete.")
def __check_args(args):
if not os.path.exists(args.dic_path):
raise FileNotFoundError(f"{args.dic_path} is not found.")
def __delimiter(fmt_str: str):
return {
'csv': ',',
'tsv': '\t'
}[fmt_str]
def __construct_dict(data: Iterator, args) -> dict:
dictionary = {}
for row in data:
key = row[args.word_idx]
try:
dictionary[key] = __polality_label(row[args.polality_idx], args)
except KeyError as e:
print(e)
return dictionary
def __polality_label(label, args):
if label in args.positive_labels:
return 'p'
elif label in args.negative_labels:
return 'n'
else:
raise KeyError(f"Convert Warning: neither polality label '{label}' is not contained '{args.positive_labels}' or '{args.negative_labels}'.")
def __parse_args():
parser = argparse.ArgumentParser('argument parser')
parser.add_argument(
'dic_path',
help='path of polarity dictionary',
metavar='path/to/dictionary',
type=str
)
parser.add_argument(
'dic_type',
help='type of dictionary',
choices=['noun', 'collocation']
)
parser.add_argument(
'--file_format',
'-ff',
help='file type',
default='csv',
choices=['csv', 'tsv'],
)
parser.add_argument(
'--word_idx',
'-wi',
help='index of word column',
default=0,
type=int
)
parser.add_argument(
'--polality_idx',
'-pi',
help='index of polality column',
default=1,
type=int
)
parser.add_argument(
'--positive_labels',
'-pl',
action='append',
help='positive labels',
default=['p'],
)
parser.add_argument(
'--negative_labels',
'-nl',
action='append',
help='negative labels',
default=['n'],
)
return parser.parse_args() | 2,436 | 0 | 138 |
3eb68fc8603a458a23c9f596d4cd12e5aceca226 | 14,398 | py | Python | rplugin/python3/deoplete/sources/deoplete-ruby-core.py | etordera/deoplete-ruby | 57b85a5413d12e61478ff48917d2d708005e8110 | [
"MIT"
] | 3 | 2021-02-03T12:22:12.000Z | 2021-06-27T01:31:35.000Z | rplugin/python3/deoplete/sources/deoplete-ruby-core.py | etordera/deoplete-ruby | 57b85a5413d12e61478ff48917d2d708005e8110 | [
"MIT"
] | null | null | null | rplugin/python3/deoplete/sources/deoplete-ruby-core.py | etordera/deoplete-ruby | 57b85a5413d12e61478ff48917d2d708005e8110 | [
"MIT"
] | null | null | null | import re
import traceback
from deoplete.source.base import Base
# ------------------------------- KEYWORD -------------------------------------------------------------------------
ruby_core_keywords = [
'__callee__', '__dir__', '__enable', '__id__', '__method__', '__send__', '_id', 'abort', 'abort_on_exception',
'abs', 'absolute_path', 'acos', 'acosh', 'add', 'add_stress_to_class', 'add_trace_func', 'advise', 'alias_method',
'aliases', 'alive?', 'all?', 'all_symbols', 'allbits?', 'allocate', 'ancestors', 'angle', 'any?', 'anybits?',
'append', 'append_features', 'arg', 'args', 'argv', 'arity', 'ascii_compatible?', 'ascii_only?', 'asciicompat_encoding',
'asctime', 'asin', 'asinh', 'assoc', 'at', 'at_exit', 'atan', 'atanh', 'atime', 'attr', 'attr_accessor', 'attr_reader',
'attr_writer', 'autoclose', 'autoclose?', 'autoload', 'autoload?', 'b', 'backtrace', 'backtrace_locations',
'base_label', 'basename', 'begin', 'between?', 'bind', 'binding', 'binmode', 'binmode?', 'binread', 'binwrite',
'birthtime', 'bit_length', 'blksize', 'block_given?', 'blockdev?', 'blocks', 'broadcast', 'bsearch',
'bsearch_index', 'bytes', 'bytesize', 'byteslice', 'call', 'callcc', 'callee_id', 'caller', 'caller_locations',
'capitalize', 'capitalize!', 'captures', 'casecmp', 'casecmp?', 'casefold?', 'catch', 'cause', 'cbrt',
'ceil', 'center', 'chain', 'change_privilege', 'chardev?', 'chars', 'chdir', 'children', 'chmod', 'chomp',
'chomp!', 'chop', 'chop!', 'chown', 'chr', 'chroot', 'chunk', 'chunk_while', 'clamp', 'class', 'class_eval',
'class_exec', 'class_variable_defined?', 'class_variable_get', 'class_variable_set', 'class_variables',
'clear', 'clock_getres', 'clock_gettime', 'clone', 'close', 'close_on_exec', 'close_on_exec?', 'close_read',
'close_write', 'closed?', 'codepoints', 'coerce', 'collect', 'collect!', 'collect_concat', 'combination',
'compact', 'compact!', 'compare_by_identity', 'compare_by_identity?', 'compatible?', 'compile',
'compile_file', 'compile_option', 'concat', 'conj', 'conjugate', 'const_defined?', 'const_get', 'const_missing',
'const_set', 'constants', 'convert', 'convpath', 'copy_stream', 'coredump?', 'cos', 'cosh', 'count', 'count_objects',
'cover?', 'crypt', 'ctime', 'current', 'curry', 'cycle', 'daemon', 'day', 'default', 'default_external', 'default_internal',
'default_proc', 'define_finalizer', 'define_method', 'define_singleton_method', 'defined_class',
'delete', 'delete!', 'delete_at', 'delete_if', 'delete_prefix', 'delete_prefix!', 'delete_suffix',
'delete_suffix!', 'denominator', 'deprecate_constant', 'deq', 'destination_encoding', 'destination_encoding_name',
'detach', 'detect', 'dev', 'dev_major', 'dev_minor', 'difference', 'dig', 'digits', 'directory?', 'dirname',
'disable', 'disasm', 'disassemble', 'display', 'div', 'divmod', 'downcase', 'downcase!', 'downto', 'drop',
'drop_while', 'dst?', 'dummy?', 'dump', 'dup', 'each', 'each_byte', 'each_char', 'each_child', 'each_codepoint',
'each_cons', 'each_entry', 'each_grapheme_cluster', 'each_index', 'each_key', 'each_line', 'each_object',
'each_pair', 'each_slice', 'each_value', 'each_with_index', 'each_with_object', 'egid', 'eid', 'empty?',
'enable', 'enabled?', 'enclose', 'enclosed?', 'encode', 'encode!', 'encoding', 'end', 'end_with?', 'enq',
'entries', 'enum_for', 'eof', 'eof?', 'eql?', 'equal?', 'erf', 'erfc', 'errno', 'error_bytes', 'error_char',
'escape', 'euid', 'eval', 'eval_script', 'even?', 'event', 'exception', 'exclude_end?', 'exclusive', 'exec',
'executable?', 'executable_real?', 'exist?', 'exists?', 'exit', 'exit!', 'exit_value', 'exited?', 'exitstatus',
'exp', 'expand_path', 'extend', 'extend_object', 'extended', 'external_encoding', 'extname', 'fail',
'fcntl', 'fdatasync', 'fdiv', 'feed', 'fetch', 'fetch_values', 'file', 'file?', 'filename', 'fileno', 'fill',
'filter', 'filter!', 'find', 'find_all', 'find_index', 'finish', 'finite?', 'first', 'first_column', 'first_lineno',
'fixed_encoding?', 'flat_map', 'flatten', 'flatten!', 'flock', 'floor', 'flush', 'fnmatch', 'fnmatch?',
'for_fd', 'force', 'force_encoding', 'foreach', 'fork', 'format', 'freeze', 'frexp', 'friday?', 'from_name',
'frozen?', 'fsync', 'ftype', 'full_message', 'gamma', 'garbage_collect', 'gcd', 'gcdlcm', 'getbyte', 'getc',
'getegid', 'geteuid', 'getgid', 'getgm', 'getlocal', 'getpgid', 'getpgrp', 'getpriority', 'getrlimit',
'gets', 'getsid', 'getuid', 'getutc', 'getwd', 'gid', 'glob', 'global_variables', 'gm', 'gmt?', 'gmt_offset',
'gmtime', 'gmtoff', 'grant_privilege', 'grapheme_clusters', 'grep', 'grep_v', 'group', 'group_by', 'groups',
'grpowned?', 'gsub', 'gsub!', 'handle_interrupt', 'has_key?', 'has_value?', 'hash', 'hex', 'home', 'hour',
'hypot', 'i', 'id', 'identical?', 'imag', 'imaginary', 'include', 'include?', 'included', 'included_modules',
'incomplete_input?', 'index', 'infinite?', 'inherited', 'initgroups', 'initialize_copy', 'inject',
'ino', 'inplace_mode', 'insert', 'insert_output', 'inspect', 'instance_eval', 'instance_exec', 'instance_method',
'instance_methods', 'instance_of?', 'instance_variable_defined?', 'instance_variable_get',
'instance_variable_set', 'instance_variables', 'instruction_sequence', 'integer?', 'intern',
'internal_encoding', 'invert', 'ioctl', 'is_a?', 'isatty', 'isdst', 'issetugid', 'iterator?', 'itself',
'join', 'keep_if', 'key', 'key?', 'keys', 'kill', 'kind_of?', 'label', 'lambda', 'lambda?', 'last', 'last_column',
'last_error', 'last_lineno', 'last_match', 'last_status', 'latest_gc_info', 'lazy', 'lchmod', 'lchown',
'lcm', 'ldexp', 'length', 'lgamma', 'lineno', 'lines', 'link', 'list', 'ljust', 'load', 'load_from_binary', 'load_from_binary_extra_data',
'local', 'local_variable_defined?', 'local_variable_get', 'local_variable_set', 'local_variables',
'locale_charmap', 'localtime', 'lock', 'locked?', 'log', 'loop', 'lstat', 'lstrip', 'lstrip!', 'lutime', 'magnitude',
'main', 'malloc_allocated_size', 'malloc_allocations', 'map', 'map!', 'match', 'match?', 'max', 'max_by',
'maxgroups', 'mday', 'member?', 'members', 'merge', 'merge!', 'message', 'method', 'method_added', 'method_defined?',
'method_id', 'method_missing', 'method_removed', 'method_undefined', 'methods', 'min', 'min_by',
'minmax', 'minmax_by', 'mkdir', 'mkfifo', 'mktime', 'mode', 'module_eval', 'module_exec', 'module_function',
'modulo', 'mon', 'monday?', 'month', 'mtime', 'name', 'name_list', 'named_captures', 'names', 'nan?', 'negative?',
'nesting', 'new', 'new_seed', 'next', 'next!', 'next_float', 'next_values', 'nil?', 'nlink', 'nobits?', 'none?',
'nonzero?', 'now', 'nsec', 'num_waiting', 'numerator', 'object_id', 'oct', 'odd?', 'of', 'offset', 'one?', 'open',
'options', 'ord', 'original_name', 'owned?', 'owner', 'p', 'pack', 'parameters', 'parse', 'parse_file', 'partition',
'pass', 'path', 'pause', 'peek', 'peek_values', 'pending_interrupt?', 'permutation', 'phase', 'pid', 'pipe',
'pipe?', 'polar', 'pop', 'popen', 'pos', 'positive?', 'post_match', 'pow', 'pp', 'ppid', 'pre_match', 'pread', 'pred',
'prepend', 'prepend_features', 'prepended', 'prev_float', 'primitive_convert', 'primitive_errinfo',
'print', 'printf', 'priority', 'private', 'private_call?', 'private_class_method', 'private_constant',
'private_instance_methods', 'private_method_defined?', 'private_methods', 'proc', 'product',
'protected', 'protected_instance_methods', 'protected_method_defined?', 'protected_methods',
'public', 'public_class_method', 'public_constant', 'public_instance_method', 'public_instance_methods',
'public_method', 'public_method_defined?', 'public_methods', 'public_send', 'push', 'putback',
'putc', 'puts', 'pwd', 'pwrite', 'quo', 'quote', 'raise', 'raised_exception', 'rand', 'random_number', 'rassoc',
'rationalize', 'raw_data', 'rdev', 'rdev_major', 'rdev_minor', 're_exchange', 're_exchangeable?',
'read', 'read_nonblock', 'readable?', 'readable_real?', 'readagain_bytes', 'readbyte', 'readchar',
'readline', 'readlines', 'readlink', 'readpartial', 'real', 'real?', 'realdirpath', 'realpath', 'reason',
'receiver', 'rect', 'rectangular', 'reduce', 'refine', 'regexp', 'rehash', 'reject', 'reject!', 'remainder',
'remove_class_variable', 'remove_const', 'remove_instance_variable', 'remove_method', 'remove_stress_to_class',
'rename', 'reopen', 'repeated_combination', 'repeated_permutation', 'replace', 'replacement',
'replicate', 'report', 'report_on_exception', 'require', 'require_relative', 'respond_to?', 'respond_to_missing?',
'restore', 'result', 'resume', 'return_value', 'reverse', 'reverse!', 'reverse_each', 'rewind', 'rid',
'rindex', 'rjust', 'rmdir', 'rotate', 'rotate!', 'round', 'rpartition', 'rstrip', 'rstrip!', 'run', 'safe_level',
'sample', 'saturday?', 'scan', 'scrub', 'scrub!', 'search_convpath', 'sec', 'seed', 'seek', 'select', 'select!',
'self', 'send', 'set_backtrace', 'set_encoding', 'set_trace_func', 'setbyte', 'setegid', 'seteuid',
'setgid', 'setgid?', 'setpgid', 'setpgrp', 'setpriority', 'setproctitle', 'setregid', 'setresgid', 'setresuid',
'setreuid', 'setrgid', 'setrlimit', 'setruid', 'setsid', 'setuid', 'setuid?', 'shift', 'shuffle', 'shuffle!',
'sid_available?', 'signal', 'signaled?', 'signame', 'signo', 'sin', 'singleton_class', 'singleton_class?',
'singleton_method', 'singleton_method_added', 'singleton_method_removed', 'singleton_method_undefined',
'singleton_methods', 'sinh', 'size', 'size?', 'skip', 'sleep', 'slice', 'slice!', 'slice_after', 'slice_before',
'slice_when', 'socket?', 'sort', 'sort!', 'sort_by', 'sort_by!', 'source', 'source_encoding', 'source_encoding_name',
'source_location', 'spawn', 'split', 'sprintf', 'sqrt', 'squeeze', 'squeeze!', 'srand', 'start', 'start_with?',
'stat', 'status', 'step', 'sticky?', 'stop', 'stop?', 'stopped?', 'stopsig', 'store', 'stress', 'strftime', 'string',
'strip', 'strip!', 'sub', 'sub!', 'subsec', 'succ', 'succ!', 'success?', 'sum', 'sunday?', 'super_method', 'superclass',
'swapcase', 'swapcase!', 'switch', 'symlink', 'symlink?', 'sync', 'synchronize', 'syscall', 'sysopen',
'sysread', 'sysseek', 'system', 'syswrite', 'tag', 'taint', 'tainted?', 'take', 'take_while', 'tan', 'tanh',
'tap', 'tell', 'terminate', 'termsig', 'test', 'then', 'thread_variable?', 'thread_variable_get', 'thread_variable_set',
'thread_variables', 'throw', 'thursday?', 'times', 'to_a', 'to_ary', 'to_binary', 'to_c', 'to_enum', 'to_f',
'to_h', 'to_hash', 'to_i', 'to_int', 'to_io', 'to_path', 'to_proc', 'to_r', 'to_s', 'to_str', 'to_sym', 'to_tty?',
'to_write_io', 'total_time', 'tr', 'tr!', 'tr_s', 'tr_s!', 'trace', 'trace_points', 'trace_var', 'transfer',
'transform_keys', 'transform_keys!', 'transform_values', 'transform_values!', 'transpose',
'trap', 'truncate', 'trust', 'try_convert', 'try_lock', 'tty?', 'tuesday?', 'tv_nsec', 'tv_sec', 'tv_usec',
'type', 'uid', 'umask', 'unbind', 'undef_method', 'undefine_finalizer', 'undump', 'ungetbyte', 'ungetc',
'unicode_normalize', 'unicode_normalize!', 'unicode_normalized?', 'union', 'uniq', 'uniq!', 'unlink',
'unlock', 'unpack', 'unshift', 'untaint', 'untrace_var', 'untrust', 'untrusted?', 'upcase', 'upcase!',
'update', 'upto', 'urandom', 'usec', 'used_modules', 'using', 'utc', 'utc?', 'utc_offset', 'utime', 'valid_encoding?',
'value', 'value?', 'values', 'values_at', 'verify_internal_consistency', 'verify_transient_heap_internal_consistency',
'wait', 'waitall', 'waitpid', 'wakeup', 'warn', 'wday', 'wednesday?', 'with_index', 'with_object', 'world_readable?',
'world_writable?', 'writable?', 'writable_real?', 'write', 'write_nonblock', 'yday', 'year', 'yield',
'yield_self', 'zero?', 'zip', 'zone', 'ARGF', 'AbstractSyntaxTree', 'ArgumentError', 'ArithmeticSequence',
'Array', 'Backtrace', 'BasicObject', 'Binding', 'Chain', 'Class', 'ClosedQueueError', 'Comparable',
'CompatibilityError', 'Complex', 'ConditionVariable', 'Constants', 'Continuation', 'Converter',
'ConverterNotFoundError', 'DEBUG', 'Data', 'Dir', 'DomainError', 'EAGAINWaitReadable', 'EAGAINWaitWritable',
'EINPROGRESSWaitReadable', 'EINPROGRESSWaitWritable', 'ENV', 'EOFError', 'EWOULDBLOCKWaitReadable',
'EWOULDBLOCKWaitWritable', 'Encoding', 'EncodingError', 'Enumerable', 'Enumerator', 'Errno', 'Exception',
'FalseClass', 'Fiber', 'FiberError', 'File', 'FileTest', 'Float', 'FloatDomainError', 'Formatter', 'FrozenError',
'GC', 'GID', 'Generator', 'Hash', 'IO', 'IOError', 'IndexError', 'InstructionSequence', 'Integer', 'Interrupt',
'InvalidByteSequenceError', 'Kernel', 'KeyError', 'Lazy', 'LoadError', 'LocalJumpError', 'Location',
'MJIT', 'Marshal', 'MatchData', 'Math', 'Method', 'Module', 'Mutex', 'NameError', 'NilClass', 'NoMemoryError',
'NoMethodError', 'Node', 'NotImplementedError', 'Numeric', 'Object', 'ObjectSpace', 'Proc', 'Process',
'Profiler', 'Queue', 'Random', 'Range', 'RangeError', 'Rational', 'Regexp', 'RegexpError', 'RubyVM', 'RuntimeError',
'ScriptError', 'SecurityError', 'Signal', 'SignalException', 'SizedQueue', 'StandardError', 'Stat',
'Status', 'StopIteration', 'String', 'Struct', 'Symbol', 'SyntaxError', 'Sys', 'SystemCallError', 'SystemExit',
'SystemStackError', 'Thread', 'ThreadError', 'ThreadGroup', 'Time', 'TracePoint', 'TrueClass', 'TypeError',
'UID', 'UnboundMethod', 'UncaughtThrowError', 'UndefinedConversionError', 'UnicodeNormalize',
'WaitReadable', 'WaitWritable', 'Warning', 'WeakMap', 'Yielder', 'ZeroDivisionError'
]
# ------------------------------- KEYWORD -------------------------------------------------------------------------
| 96.630872 | 142 | 0.631129 | import re
import traceback
from deoplete.source.base import Base
# ------------------------------- KEYWORD -------------------------------------------------------------------------
ruby_core_keywords = [
'__callee__', '__dir__', '__enable', '__id__', '__method__', '__send__', '_id', 'abort', 'abort_on_exception',
'abs', 'absolute_path', 'acos', 'acosh', 'add', 'add_stress_to_class', 'add_trace_func', 'advise', 'alias_method',
'aliases', 'alive?', 'all?', 'all_symbols', 'allbits?', 'allocate', 'ancestors', 'angle', 'any?', 'anybits?',
'append', 'append_features', 'arg', 'args', 'argv', 'arity', 'ascii_compatible?', 'ascii_only?', 'asciicompat_encoding',
'asctime', 'asin', 'asinh', 'assoc', 'at', 'at_exit', 'atan', 'atanh', 'atime', 'attr', 'attr_accessor', 'attr_reader',
'attr_writer', 'autoclose', 'autoclose?', 'autoload', 'autoload?', 'b', 'backtrace', 'backtrace_locations',
'base_label', 'basename', 'begin', 'between?', 'bind', 'binding', 'binmode', 'binmode?', 'binread', 'binwrite',
'birthtime', 'bit_length', 'blksize', 'block_given?', 'blockdev?', 'blocks', 'broadcast', 'bsearch',
'bsearch_index', 'bytes', 'bytesize', 'byteslice', 'call', 'callcc', 'callee_id', 'caller', 'caller_locations',
'capitalize', 'capitalize!', 'captures', 'casecmp', 'casecmp?', 'casefold?', 'catch', 'cause', 'cbrt',
'ceil', 'center', 'chain', 'change_privilege', 'chardev?', 'chars', 'chdir', 'children', 'chmod', 'chomp',
'chomp!', 'chop', 'chop!', 'chown', 'chr', 'chroot', 'chunk', 'chunk_while', 'clamp', 'class', 'class_eval',
'class_exec', 'class_variable_defined?', 'class_variable_get', 'class_variable_set', 'class_variables',
'clear', 'clock_getres', 'clock_gettime', 'clone', 'close', 'close_on_exec', 'close_on_exec?', 'close_read',
'close_write', 'closed?', 'codepoints', 'coerce', 'collect', 'collect!', 'collect_concat', 'combination',
'compact', 'compact!', 'compare_by_identity', 'compare_by_identity?', 'compatible?', 'compile',
'compile_file', 'compile_option', 'concat', 'conj', 'conjugate', 'const_defined?', 'const_get', 'const_missing',
'const_set', 'constants', 'convert', 'convpath', 'copy_stream', 'coredump?', 'cos', 'cosh', 'count', 'count_objects',
'cover?', 'crypt', 'ctime', 'current', 'curry', 'cycle', 'daemon', 'day', 'default', 'default_external', 'default_internal',
'default_proc', 'define_finalizer', 'define_method', 'define_singleton_method', 'defined_class',
'delete', 'delete!', 'delete_at', 'delete_if', 'delete_prefix', 'delete_prefix!', 'delete_suffix',
'delete_suffix!', 'denominator', 'deprecate_constant', 'deq', 'destination_encoding', 'destination_encoding_name',
'detach', 'detect', 'dev', 'dev_major', 'dev_minor', 'difference', 'dig', 'digits', 'directory?', 'dirname',
'disable', 'disasm', 'disassemble', 'display', 'div', 'divmod', 'downcase', 'downcase!', 'downto', 'drop',
'drop_while', 'dst?', 'dummy?', 'dump', 'dup', 'each', 'each_byte', 'each_char', 'each_child', 'each_codepoint',
'each_cons', 'each_entry', 'each_grapheme_cluster', 'each_index', 'each_key', 'each_line', 'each_object',
'each_pair', 'each_slice', 'each_value', 'each_with_index', 'each_with_object', 'egid', 'eid', 'empty?',
'enable', 'enabled?', 'enclose', 'enclosed?', 'encode', 'encode!', 'encoding', 'end', 'end_with?', 'enq',
'entries', 'enum_for', 'eof', 'eof?', 'eql?', 'equal?', 'erf', 'erfc', 'errno', 'error_bytes', 'error_char',
'escape', 'euid', 'eval', 'eval_script', 'even?', 'event', 'exception', 'exclude_end?', 'exclusive', 'exec',
'executable?', 'executable_real?', 'exist?', 'exists?', 'exit', 'exit!', 'exit_value', 'exited?', 'exitstatus',
'exp', 'expand_path', 'extend', 'extend_object', 'extended', 'external_encoding', 'extname', 'fail',
'fcntl', 'fdatasync', 'fdiv', 'feed', 'fetch', 'fetch_values', 'file', 'file?', 'filename', 'fileno', 'fill',
'filter', 'filter!', 'find', 'find_all', 'find_index', 'finish', 'finite?', 'first', 'first_column', 'first_lineno',
'fixed_encoding?', 'flat_map', 'flatten', 'flatten!', 'flock', 'floor', 'flush', 'fnmatch', 'fnmatch?',
'for_fd', 'force', 'force_encoding', 'foreach', 'fork', 'format', 'freeze', 'frexp', 'friday?', 'from_name',
'frozen?', 'fsync', 'ftype', 'full_message', 'gamma', 'garbage_collect', 'gcd', 'gcdlcm', 'getbyte', 'getc',
'getegid', 'geteuid', 'getgid', 'getgm', 'getlocal', 'getpgid', 'getpgrp', 'getpriority', 'getrlimit',
'gets', 'getsid', 'getuid', 'getutc', 'getwd', 'gid', 'glob', 'global_variables', 'gm', 'gmt?', 'gmt_offset',
'gmtime', 'gmtoff', 'grant_privilege', 'grapheme_clusters', 'grep', 'grep_v', 'group', 'group_by', 'groups',
'grpowned?', 'gsub', 'gsub!', 'handle_interrupt', 'has_key?', 'has_value?', 'hash', 'hex', 'home', 'hour',
'hypot', 'i', 'id', 'identical?', 'imag', 'imaginary', 'include', 'include?', 'included', 'included_modules',
'incomplete_input?', 'index', 'infinite?', 'inherited', 'initgroups', 'initialize_copy', 'inject',
'ino', 'inplace_mode', 'insert', 'insert_output', 'inspect', 'instance_eval', 'instance_exec', 'instance_method',
'instance_methods', 'instance_of?', 'instance_variable_defined?', 'instance_variable_get',
'instance_variable_set', 'instance_variables', 'instruction_sequence', 'integer?', 'intern',
'internal_encoding', 'invert', 'ioctl', 'is_a?', 'isatty', 'isdst', 'issetugid', 'iterator?', 'itself',
'join', 'keep_if', 'key', 'key?', 'keys', 'kill', 'kind_of?', 'label', 'lambda', 'lambda?', 'last', 'last_column',
'last_error', 'last_lineno', 'last_match', 'last_status', 'latest_gc_info', 'lazy', 'lchmod', 'lchown',
'lcm', 'ldexp', 'length', 'lgamma', 'lineno', 'lines', 'link', 'list', 'ljust', 'load', 'load_from_binary', 'load_from_binary_extra_data',
'local', 'local_variable_defined?', 'local_variable_get', 'local_variable_set', 'local_variables',
'locale_charmap', 'localtime', 'lock', 'locked?', 'log', 'loop', 'lstat', 'lstrip', 'lstrip!', 'lutime', 'magnitude',
'main', 'malloc_allocated_size', 'malloc_allocations', 'map', 'map!', 'match', 'match?', 'max', 'max_by',
'maxgroups', 'mday', 'member?', 'members', 'merge', 'merge!', 'message', 'method', 'method_added', 'method_defined?',
'method_id', 'method_missing', 'method_removed', 'method_undefined', 'methods', 'min', 'min_by',
'minmax', 'minmax_by', 'mkdir', 'mkfifo', 'mktime', 'mode', 'module_eval', 'module_exec', 'module_function',
'modulo', 'mon', 'monday?', 'month', 'mtime', 'name', 'name_list', 'named_captures', 'names', 'nan?', 'negative?',
'nesting', 'new', 'new_seed', 'next', 'next!', 'next_float', 'next_values', 'nil?', 'nlink', 'nobits?', 'none?',
'nonzero?', 'now', 'nsec', 'num_waiting', 'numerator', 'object_id', 'oct', 'odd?', 'of', 'offset', 'one?', 'open',
'options', 'ord', 'original_name', 'owned?', 'owner', 'p', 'pack', 'parameters', 'parse', 'parse_file', 'partition',
'pass', 'path', 'pause', 'peek', 'peek_values', 'pending_interrupt?', 'permutation', 'phase', 'pid', 'pipe',
'pipe?', 'polar', 'pop', 'popen', 'pos', 'positive?', 'post_match', 'pow', 'pp', 'ppid', 'pre_match', 'pread', 'pred',
'prepend', 'prepend_features', 'prepended', 'prev_float', 'primitive_convert', 'primitive_errinfo',
'print', 'printf', 'priority', 'private', 'private_call?', 'private_class_method', 'private_constant',
'private_instance_methods', 'private_method_defined?', 'private_methods', 'proc', 'product',
'protected', 'protected_instance_methods', 'protected_method_defined?', 'protected_methods',
'public', 'public_class_method', 'public_constant', 'public_instance_method', 'public_instance_methods',
'public_method', 'public_method_defined?', 'public_methods', 'public_send', 'push', 'putback',
'putc', 'puts', 'pwd', 'pwrite', 'quo', 'quote', 'raise', 'raised_exception', 'rand', 'random_number', 'rassoc',
'rationalize', 'raw_data', 'rdev', 'rdev_major', 'rdev_minor', 're_exchange', 're_exchangeable?',
'read', 'read_nonblock', 'readable?', 'readable_real?', 'readagain_bytes', 'readbyte', 'readchar',
'readline', 'readlines', 'readlink', 'readpartial', 'real', 'real?', 'realdirpath', 'realpath', 'reason',
'receiver', 'rect', 'rectangular', 'reduce', 'refine', 'regexp', 'rehash', 'reject', 'reject!', 'remainder',
'remove_class_variable', 'remove_const', 'remove_instance_variable', 'remove_method', 'remove_stress_to_class',
'rename', 'reopen', 'repeated_combination', 'repeated_permutation', 'replace', 'replacement',
'replicate', 'report', 'report_on_exception', 'require', 'require_relative', 'respond_to?', 'respond_to_missing?',
'restore', 'result', 'resume', 'return_value', 'reverse', 'reverse!', 'reverse_each', 'rewind', 'rid',
'rindex', 'rjust', 'rmdir', 'rotate', 'rotate!', 'round', 'rpartition', 'rstrip', 'rstrip!', 'run', 'safe_level',
'sample', 'saturday?', 'scan', 'scrub', 'scrub!', 'search_convpath', 'sec', 'seed', 'seek', 'select', 'select!',
'self', 'send', 'set_backtrace', 'set_encoding', 'set_trace_func', 'setbyte', 'setegid', 'seteuid',
'setgid', 'setgid?', 'setpgid', 'setpgrp', 'setpriority', 'setproctitle', 'setregid', 'setresgid', 'setresuid',
'setreuid', 'setrgid', 'setrlimit', 'setruid', 'setsid', 'setuid', 'setuid?', 'shift', 'shuffle', 'shuffle!',
'sid_available?', 'signal', 'signaled?', 'signame', 'signo', 'sin', 'singleton_class', 'singleton_class?',
'singleton_method', 'singleton_method_added', 'singleton_method_removed', 'singleton_method_undefined',
'singleton_methods', 'sinh', 'size', 'size?', 'skip', 'sleep', 'slice', 'slice!', 'slice_after', 'slice_before',
'slice_when', 'socket?', 'sort', 'sort!', 'sort_by', 'sort_by!', 'source', 'source_encoding', 'source_encoding_name',
'source_location', 'spawn', 'split', 'sprintf', 'sqrt', 'squeeze', 'squeeze!', 'srand', 'start', 'start_with?',
'stat', 'status', 'step', 'sticky?', 'stop', 'stop?', 'stopped?', 'stopsig', 'store', 'stress', 'strftime', 'string',
'strip', 'strip!', 'sub', 'sub!', 'subsec', 'succ', 'succ!', 'success?', 'sum', 'sunday?', 'super_method', 'superclass',
'swapcase', 'swapcase!', 'switch', 'symlink', 'symlink?', 'sync', 'synchronize', 'syscall', 'sysopen',
'sysread', 'sysseek', 'system', 'syswrite', 'tag', 'taint', 'tainted?', 'take', 'take_while', 'tan', 'tanh',
'tap', 'tell', 'terminate', 'termsig', 'test', 'then', 'thread_variable?', 'thread_variable_get', 'thread_variable_set',
'thread_variables', 'throw', 'thursday?', 'times', 'to_a', 'to_ary', 'to_binary', 'to_c', 'to_enum', 'to_f',
'to_h', 'to_hash', 'to_i', 'to_int', 'to_io', 'to_path', 'to_proc', 'to_r', 'to_s', 'to_str', 'to_sym', 'to_tty?',
'to_write_io', 'total_time', 'tr', 'tr!', 'tr_s', 'tr_s!', 'trace', 'trace_points', 'trace_var', 'transfer',
'transform_keys', 'transform_keys!', 'transform_values', 'transform_values!', 'transpose',
'trap', 'truncate', 'trust', 'try_convert', 'try_lock', 'tty?', 'tuesday?', 'tv_nsec', 'tv_sec', 'tv_usec',
'type', 'uid', 'umask', 'unbind', 'undef_method', 'undefine_finalizer', 'undump', 'ungetbyte', 'ungetc',
'unicode_normalize', 'unicode_normalize!', 'unicode_normalized?', 'union', 'uniq', 'uniq!', 'unlink',
'unlock', 'unpack', 'unshift', 'untaint', 'untrace_var', 'untrust', 'untrusted?', 'upcase', 'upcase!',
'update', 'upto', 'urandom', 'usec', 'used_modules', 'using', 'utc', 'utc?', 'utc_offset', 'utime', 'valid_encoding?',
'value', 'value?', 'values', 'values_at', 'verify_internal_consistency', 'verify_transient_heap_internal_consistency',
'wait', 'waitall', 'waitpid', 'wakeup', 'warn', 'wday', 'wednesday?', 'with_index', 'with_object', 'world_readable?',
'world_writable?', 'writable?', 'writable_real?', 'write', 'write_nonblock', 'yday', 'year', 'yield',
'yield_self', 'zero?', 'zip', 'zone', 'ARGF', 'AbstractSyntaxTree', 'ArgumentError', 'ArithmeticSequence',
'Array', 'Backtrace', 'BasicObject', 'Binding', 'Chain', 'Class', 'ClosedQueueError', 'Comparable',
'CompatibilityError', 'Complex', 'ConditionVariable', 'Constants', 'Continuation', 'Converter',
'ConverterNotFoundError', 'DEBUG', 'Data', 'Dir', 'DomainError', 'EAGAINWaitReadable', 'EAGAINWaitWritable',
'EINPROGRESSWaitReadable', 'EINPROGRESSWaitWritable', 'ENV', 'EOFError', 'EWOULDBLOCKWaitReadable',
'EWOULDBLOCKWaitWritable', 'Encoding', 'EncodingError', 'Enumerable', 'Enumerator', 'Errno', 'Exception',
'FalseClass', 'Fiber', 'FiberError', 'File', 'FileTest', 'Float', 'FloatDomainError', 'Formatter', 'FrozenError',
'GC', 'GID', 'Generator', 'Hash', 'IO', 'IOError', 'IndexError', 'InstructionSequence', 'Integer', 'Interrupt',
'InvalidByteSequenceError', 'Kernel', 'KeyError', 'Lazy', 'LoadError', 'LocalJumpError', 'Location',
'MJIT', 'Marshal', 'MatchData', 'Math', 'Method', 'Module', 'Mutex', 'NameError', 'NilClass', 'NoMemoryError',
'NoMethodError', 'Node', 'NotImplementedError', 'Numeric', 'Object', 'ObjectSpace', 'Proc', 'Process',
'Profiler', 'Queue', 'Random', 'Range', 'RangeError', 'Rational', 'Regexp', 'RegexpError', 'RubyVM', 'RuntimeError',
'ScriptError', 'SecurityError', 'Signal', 'SignalException', 'SizedQueue', 'StandardError', 'Stat',
'Status', 'StopIteration', 'String', 'Struct', 'Symbol', 'SyntaxError', 'Sys', 'SystemCallError', 'SystemExit',
'SystemStackError', 'Thread', 'ThreadError', 'ThreadGroup', 'Time', 'TracePoint', 'TrueClass', 'TypeError',
'UID', 'UnboundMethod', 'UncaughtThrowError', 'UndefinedConversionError', 'UnicodeNormalize',
'WaitReadable', 'WaitWritable', 'Warning', 'WeakMap', 'Yielder', 'ZeroDivisionError'
]
# ------------------------------- KEYWORD -------------------------------------------------------------------------
class Source(Base):
def __init__(self, vim):
super().__init__(vim)
self.name = 'deoplete-ruby-core'
self.filetypes = ['ruby']
self.mark = '[ruby-core]'
self.rank = 10
def get_complete_position(self, context):
m = re.search('[a-zA-Z0-9_?!]*$', context['input'])
return m.start() if m else -1
def gather_candidates(self, context):
try:
dic = ruby_core_keywords
return dic
except Exception:
traceback.print_exc()
| 432 | -2 | 103 |
979d3d74bd3866e1145259ed29b433998c6c400d | 186 | py | Python | neural_process/__init__.py | revsic/tf-attentive-neural-process | efa3bb0a9b6cfebaa3c1e025a9da00aef8d0a1e2 | [
"MIT"
] | 4 | 2020-08-30T14:20:05.000Z | 2021-03-23T12:53:27.000Z | neural_process/__init__.py | revsic/tf-attentive-neural-process | efa3bb0a9b6cfebaa3c1e025a9da00aef8d0a1e2 | [
"MIT"
] | null | null | null | neural_process/__init__.py | revsic/tf-attentive-neural-process | efa3bb0a9b6cfebaa3c1e025a9da00aef8d0a1e2 | [
"MIT"
] | 4 | 2020-03-23T06:34:49.000Z | 2021-10-25T23:57:24.000Z | from neural_process.np import NeuralProcess
from neural_process.cnp import ConditionalNP
from neural_process.anp import AttentiveNP
from neural_process.module.attention import Attention
| 37.2 | 53 | 0.887097 | from neural_process.np import NeuralProcess
from neural_process.cnp import ConditionalNP
from neural_process.anp import AttentiveNP
from neural_process.module.attention import Attention
| 0 | 0 | 0 |
88b9fb4e665f7210757b55549cc4282c8b38c598 | 432 | py | Python | docs/src/conf.py | EpicWink/python-swf-typed | f630f5a1b6e31ad6f6260b7af77ba52212585105 | [
"MIT"
] | null | null | null | docs/src/conf.py | EpicWink/python-swf-typed | f630f5a1b6e31ad6f6260b7af77ba52212585105 | [
"MIT"
] | 2 | 2021-11-09T07:22:37.000Z | 2021-11-16T06:08:17.000Z | docs/src/conf.py | EpicWink/python-swf-typed | f630f5a1b6e31ad6f6260b7af77ba52212585105 | [
"MIT"
] | null | null | null | """Sphinx documentation generation configuration."""
import importlib.metadata
project = "swf-typed"
copyright = "2021, Laurie O"
author = "Laurie O"
release = importlib.metadata.version("swf-typed") # full version
version = ".".join(release.split(".")[:2]) # short X.Y version
extensions = [
"sphinx.ext.autodoc",
"sphinx.ext.viewcode",
"sphinx.ext.napoleon",
"autodocsumm",
]
html_theme = "sphinx_rtd_theme"
| 22.736842 | 65 | 0.689815 | """Sphinx documentation generation configuration."""
import importlib.metadata
project = "swf-typed"
copyright = "2021, Laurie O"
author = "Laurie O"
release = importlib.metadata.version("swf-typed") # full version
version = ".".join(release.split(".")[:2]) # short X.Y version
extensions = [
"sphinx.ext.autodoc",
"sphinx.ext.viewcode",
"sphinx.ext.napoleon",
"autodocsumm",
]
html_theme = "sphinx_rtd_theme"
| 0 | 0 | 0 |
f1b172985ca70587402973a703d4eceb2acea396 | 670 | py | Python | Array/0189.RotateArray.3.py | joonion/leetcode-in-python | 9d621be1d736d81445ef19eebb7921eb4c212f83 | [
"Apache-2.0"
] | null | null | null | Array/0189.RotateArray.3.py | joonion/leetcode-in-python | 9d621be1d736d81445ef19eebb7921eb4c212f83 | [
"Apache-2.0"
] | null | null | null | Array/0189.RotateArray.3.py | joonion/leetcode-in-python | 9d621be1d736d81445ef19eebb7921eb4c212f83 | [
"Apache-2.0"
] | null | null | null | from typing import List
s = Solution()
# L, k = [1, 2, 3, 4, 5, 6, 7], 3
L, k = [-1, -100, 3, 99], 2
s.rotate(L, k)
print(L) | 26.8 | 61 | 0.435821 | from typing import List
class Solution:
def rotate(self, nums: List[int], k: int) -> None:
"""
Do not return anything, modify nums in-place instead.
"""
n, k = len(nums), k % len(nums)
start = count = 0
while count < n:
curr, prev = start, nums[start]
while True:
next = (curr + k) % n
nums[next], prev = prev, nums[next]
curr = next
count += 1
if start == curr:
break
start += 1
s = Solution()
# L, k = [1, 2, 3, 4, 5, 6, 7], 3
L, k = [-1, -100, 3, 99], 2
s.rotate(L, k)
print(L) | 0 | 522 | 23 |
1161bacaa9f1743f84d4b231e15c83e73755dd30 | 3,002 | py | Python | ds_linked_list/sorted_linked_list.py | dileepkr/datastructures | e20f853a1247c9c6ae065696247900727e415d8f | [
"MIT"
] | null | null | null | ds_linked_list/sorted_linked_list.py | dileepkr/datastructures | e20f853a1247c9c6ae065696247900727e415d8f | [
"MIT"
] | null | null | null | ds_linked_list/sorted_linked_list.py | dileepkr/datastructures | e20f853a1247c9c6ae065696247900727e415d8f | [
"MIT"
] | null | null | null | from ds_linked_list.ll_node import Node
if __name__ == "__main__":
ll = Unsorted_linked_list()
for i in range(1,10):
ll.add(i)
print(ll) | 26.333333 | 75 | 0.538641 | from ds_linked_list.ll_node import Node
class Unsorted_linked_list:
def __init__(self):
self.head = None
def add(self, element):
node_ = Node(element)
if not self.head:
self.head = node_
else:
node_.next = self.head.next
self.head = node_
def length(self):
current = self.head
count = 0
while current != None or current.next != None:
count += 1
current = current.next
return count
def is_empty(self):
return self.head == None
def remove(self, element):
node_ = Node(element)
current = self.head
previous = None
while current is not None and node_.data != current.data:
previous = current
current = current.next
if previous:
previous.next = current.next
elif current:
self.head = current.next
else:
raise ValueError("No matching element found in list to remove")
def index(self, element):
count = 0
node_ = Node(element)
current = self.head
while current is not None and node_.data != current.data:
count += 1
current = current.next
return count
def pop(self):
if not self.head:
return None
if not self.head.next:
data = self.head.data
self.head = None
return data
previous = None
current = self.head
while current is not None and current.next is not None:
previous = current
current = current.next
previous.next = current.next
return current.data
def pop(self, position):
if not self.head:
return self.head
count = 0
previous = None
current = self.head
while count != position and current.next is not None:
count += 1
previous = current
current = current.next
if count != position:
raise IndexError("Provided list index out of bound")
if not previous:
self.head = current.next
else:
previous.next = current.next
return current.data
def search(self, element):
if not self.head:
raise ValueError("Empty List, value not found")
node_ = Node(element)
current = self.head
while current is not None and node_.data != current.data:
current = current.next
if not current:
return False
return True
def __repr__(self) -> str:
ret_list = []
current = self.head
while current is not None:
ret_list.append(current.data)
current = current.next
return ret_list
if __name__ == "__main__":
ll = Unsorted_linked_list()
for i in range(1,10):
ll.add(i)
print(ll) | 2,539 | 6 | 300 |
c9dda6b20451a34bd8a26cd49b9c5441acab86e2 | 1,779 | py | Python | Curso Em Video-python/PYTHON (MUNDO 1, MUNDO 2 E MUNDO 3)/exercicios/ex0068_jogo_ParouImpar.py | AlamoVinicius/code-pratice | 924a3ff782caf3695bbeeac39fa02fb23781cd75 | [
"MIT"
] | null | null | null | Curso Em Video-python/PYTHON (MUNDO 1, MUNDO 2 E MUNDO 3)/exercicios/ex0068_jogo_ParouImpar.py | AlamoVinicius/code-pratice | 924a3ff782caf3695bbeeac39fa02fb23781cd75 | [
"MIT"
] | null | null | null | Curso Em Video-python/PYTHON (MUNDO 1, MUNDO 2 E MUNDO 3)/exercicios/ex0068_jogo_ParouImpar.py | AlamoVinicius/code-pratice | 924a3ff782caf3695bbeeac39fa02fb23781cd75 | [
"MIT"
] | null | null | null | """FAça um programa que jogue par ou impar com o computadr. O jogo só será interrompido quando o jogador PERDER,
mostrando o total de vitórias consecutivas que conquistou no final do jogo"""
from random import randint
print(f'''{"-=" * 20}
PAR OU IMPAR
{"-=" * 20}''')
contador = 0
while True:
jogadapc = randint(0, 10) # entrada das váriaveis
while True: # aqui eu fiz um input pra validação em caso de tipo de dado errado. não era obrigatório mas ok
try:
jogadaplayer = int(input('Escolha um número: '))
except ValueError:
print('jogada inválida')
else:
break
par_impar = str(input('par ou impar? [P/I]: ')).strip().upper()
soma = jogadapc + jogadaplayer
if par_impar == 'P': # lógica do programa
if soma % 2 == 0:
print(f'você jogou: {jogadaplayer} e o computador jogou {jogadapc} total = {soma} (PAR)')
print('VOCÊ VENCEU!')
else:
print(f'você jogou: {jogadaplayer} e o computador jogou {jogadapc} total = {soma} (impar)')
print('VOCÊ PERDEU!')
break
elif par_impar == 'I':
if soma % 2 == 0:
print(f'você jogou: {jogadaplayer} e o computador jogou {jogadapc} total = {soma} (PAR)')
print('VOCÊ PERDEU!')
break
else:
print(f'você jogou: {jogadaplayer} e o computador jogou {jogadapc} total = {soma} (impar)')
print('VOCÊ VENCEU!')
else:
print('jogada inválida')
contador += 1
print(f'Você ganhou {contador} vezes consecutivas.') # exibição final
'''O resultado do professor deu menos e linha e um pouco de diferença na lógica mas nada tão importante pode conferir
no curso em video a resolução desse exercício''' | 39.533333 | 117 | 0.613266 | """FAça um programa que jogue par ou impar com o computadr. O jogo só será interrompido quando o jogador PERDER,
mostrando o total de vitórias consecutivas que conquistou no final do jogo"""
from random import randint
print(f'''{"-=" * 20}
PAR OU IMPAR
{"-=" * 20}''')
contador = 0
while True:
jogadapc = randint(0, 10) # entrada das váriaveis
while True: # aqui eu fiz um input pra validação em caso de tipo de dado errado. não era obrigatório mas ok
try:
jogadaplayer = int(input('Escolha um número: '))
except ValueError:
print('jogada inválida')
else:
break
par_impar = str(input('par ou impar? [P/I]: ')).strip().upper()
soma = jogadapc + jogadaplayer
if par_impar == 'P': # lógica do programa
if soma % 2 == 0:
print(f'você jogou: {jogadaplayer} e o computador jogou {jogadapc} total = {soma} (PAR)')
print('VOCÊ VENCEU!')
else:
print(f'você jogou: {jogadaplayer} e o computador jogou {jogadapc} total = {soma} (impar)')
print('VOCÊ PERDEU!')
break
elif par_impar == 'I':
if soma % 2 == 0:
print(f'você jogou: {jogadaplayer} e o computador jogou {jogadapc} total = {soma} (PAR)')
print('VOCÊ PERDEU!')
break
else:
print(f'você jogou: {jogadaplayer} e o computador jogou {jogadapc} total = {soma} (impar)')
print('VOCÊ VENCEU!')
else:
print('jogada inválida')
contador += 1
print(f'Você ganhou {contador} vezes consecutivas.') # exibição final
'''O resultado do professor deu menos e linha e um pouco de diferença na lógica mas nada tão importante pode conferir
no curso em video a resolução desse exercício''' | 0 | 0 | 0 |
bc4c93cb414d187ed9e63256b173998897ab935c | 3,451 | py | Python | main.py | vlntsolo/multiprocessing-webscraper | a3b43bb68ec0c3897c1a6294f9268b44a19ed457 | [
"MIT"
] | null | null | null | main.py | vlntsolo/multiprocessing-webscraper | a3b43bb68ec0c3897c1a6294f9268b44a19ed457 | [
"MIT"
] | null | null | null | main.py | vlntsolo/multiprocessing-webscraper | a3b43bb68ec0c3897c1a6294f9268b44a19ed457 | [
"MIT"
] | 1 | 2021-09-25T07:06:51.000Z | 2021-09-25T07:06:51.000Z | import asyncio
import logging
from multiprocessing import Queue, Process, cpu_count
import queue
from threading import Thread
import time
import aiohttp
import aiosqlite
from bs4 import BeautifulSoup
PROCESSES = cpu_count() #P parameter
NUM_THREADS = 10 #T parameter
TARGET_RESOURCES = [
'https://example.org/',
'https://www.python.org/',
'https://www.djangoproject.com/',
'https://www.lipsum.com/',
]
class TaskConsumer:
'''
Consumer class to handle flow to asyncio
for multiprocessing
'''
if __name__ == "__main__":
run()
| 28.056911 | 100 | 0.630832 | import asyncio
import logging
from multiprocessing import Queue, Process, cpu_count
import queue
from threading import Thread
import time
import aiohttp
import aiosqlite
from bs4 import BeautifulSoup
PROCESSES = cpu_count() #P parameter
NUM_THREADS = 10 #T parameter
TARGET_RESOURCES = [
'https://example.org/',
'https://www.python.org/',
'https://www.djangoproject.com/',
'https://www.lipsum.com/',
]
async def parse(url):
async with aiohttp.ClientSession() as session:
async with session.get(url) as response:
html = await response.text()
soup = BeautifulSoup(html, 'html.parser')
title = soup.title.string
links = []
for link in soup.find_all('a'):
links.append(link.get('href'))
links_str = ', '.join(links)
obj = {
'title': str(title),
'links': links_str,
}
return obj
def add_tasks(task_queue):
for url in TARGET_RESOURCES:
task_queue.put(url)
return task_queue
async def write_to_sqlite(record: dict):
title = record.get('title')
links = record.get('links')
if title is not None:
async with aiosqlite.connect('crawler.db') as db:
await db.execute('''CREATE TABLE IF NOT EXISTS pages (title, links)''')
await db.execute(f"INSERT INTO pages VALUES ('{title}','{links}')")
await db.commit()
class TaskConsumer:
'''
Consumer class to handle flow to asyncio
for multiprocessing
'''
def __init__(self):
pass
async def thread_worker(self, q):
while True:
task = q.get()
await write_to_sqlite(task)
q.task_done()
async def process_tasks_from_queue(self, task_queue, results_queue):
while not task_queue.empty():
url = task_queue.get()
result = await parse(url)
results_queue.put(result)
def process(self, task_queue, results_queue):
#Asyncio coroutine
asyncio.run(self.process_tasks_from_queue(task_queue, results_queue))
def thread(self, thread_queue):
#Asyncio coroutine
asyncio.run(self.thread_worker(thread_queue))
def run():
#New multiprocess queue
# pipe_list = []
start = time.time()
empty_task_queue = Queue()
full_task_queue = add_tasks(empty_task_queue)
processes = []
results_queue = Queue()
thread_queue = queue.Queue()
print(f"Running with {PROCESSES} processes (scraper) and with {NUM_THREADS} threads for sqlite")
#Spawning processes to scrape webpage
for n in range(PROCESSES):
p = Process(target=TaskConsumer().process, args=(full_task_queue, results_queue, ))
processes.append(p)
p.start()
for p in processes:
p.join()
#Dump results for threads consumers
print("results_queue", results_queue.qsize())
while not results_queue.empty():
thread_queue.put(results_queue.get())
print("results_queue", results_queue.qsize())
print("thread_queue", thread_queue.qsize())
for i in range(NUM_THREADS):
worker = Thread(target=TaskConsumer().thread, args=(thread_queue, ), daemon=True)
worker.start()
#Block thread queue until all items are processed
thread_queue.join()
print(f"Time taken = {time.time() - start:.10f}")
if __name__ == "__main__":
run()
| 2,656 | 0 | 226 |
7770387e6f40283e78201fcab3863091f701b894 | 3,121 | py | Python | convert_videos/settings.py | justin8/convert_videos | ba4829079cf3562c71fddcd182ea4cacf9f8e249 | [
"MIT"
] | 3 | 2017-11-06T03:19:19.000Z | 2022-03-27T07:33:44.000Z | convert_videos/settings.py | justin8/convert_videos | ba4829079cf3562c71fddcd182ea4cacf9f8e249 | [
"MIT"
] | 3 | 2020-01-04T11:55:18.000Z | 2022-03-26T09:41:02.000Z | convert_videos/settings.py | justin8/convert_videos | ba4829079cf3562c71fddcd182ea4cacf9f8e249 | [
"MIT"
] | null | null | null | from dataclasses import dataclass
from video_utils import Codec
@dataclass
@dataclass
| 34.677778 | 137 | 0.603012 | from dataclasses import dataclass
from video_utils import Codec
@dataclass
class AudioSettings:
codec: Codec
channels: int
bitrate: int
def __str__(self):
output = ""
output += f" -acodec {self.codec.get_ffmpeg_name()}"
output += f" -ab {self.bitrate}k"
output += f" -ac {self.channels}"
output += f" -map 0:a" # Include all audio channels
return output
@dataclass
class VideoSettings:
codec: Codec
quality: int
preset: str
width: int = None # Or None
encoder: str = "software"
def __str__(self):
output = self._get_stream_settings()
output += f" -vcodec {self._get_ffmpeg_codec()}"
if self.codec.format_name == "copy":
return output
output += f" -preset {self.preset}"
output += self._get_quality_settings()
output += self._get_codec_settings()
output += self._get_scaling_settings()
return output
def _get_scaling_settings(self):
output = ""
if self.width:
output += f" -vf scale={self.width}:-2"
return output
def _get_quality_settings(self):
if self.encoder == "nvidia":
# nvenc doesn't support CRF; only CQ and QP modes.
# QP consistently provides better quality output for the same bitrate on NVENC however, so we're using that
output = f" -rc constqp -qp {self.quality}"
output += " -b:v 0" # Unless bitrate is set to 0, CQ is mostly ignored
# b-frames are disabled by default, and only supported on Turing+ architectures. Need to auto-detect this
# output += f" -rc-lookahead -b_ref_mode middle"
elif self.encoder == "intel":
# qsv doesn't support CRF either; they have their own methods at the link below
# ICQ and LA-ICQ are apparently the gold standard; but only supported on windows (weird)
# https://www.intel.com/content/www/us/en/developer/articles/technical/common-bitrate-control-methods-in-intel-media-sdk.html
output = f" -global_quality {self.quality}"
if self.codec.format_name == "AVC":
# Look-ahead is supported for x264 and is preferable for better quality
output += " -look_ahead 1"
else:
output = f" -crf {self.quality}"
return output
def _get_codec_settings(self):
output = ""
if self.codec == Codec("HEVC"):
output += f" -strict -2"
return output
def _get_stream_settings(self):
output = " -map 0:v:0" # Include first video stream
output += " -map 0:s?" # Include all subtitle streams, if they exist
output += " -c:s copy"
return output
def _get_ffmpeg_codec(self):
ffmpeg_codec = None
if self.codec.format_name == "copy":
ffmpeg_codec = "copy"
else:
ffmpeg_codec = self.codec.get_ffmpeg_name(self.encoder)
if not ffmpeg_codec:
raise Exception("Failed to find the desired codec!")
return ffmpeg_codec
| 2,634 | 352 | 44 |
c7d20615c4661ca6379e79c0a04f84ac1184fea2 | 1,202 | py | Python | tests/backend/ini.py | edyan/python-anyconfig | d237909a6e4848737539b80951b710238c72052f | [
"MIT"
] | null | null | null | tests/backend/ini.py | edyan/python-anyconfig | d237909a6e4848737539b80951b710238c72052f | [
"MIT"
] | null | null | null | tests/backend/ini.py | edyan/python-anyconfig | d237909a6e4848737539b80951b710238c72052f | [
"MIT"
] | null | null | null | #
# Copyright (C) 2012 - 2017 Satoru SATOH <ssato @ redhat.com>
# License: MIT
#
# pylint: disable=missing-docstring,invalid-name,too-few-public-methods
from __future__ import absolute_import
import anyconfig.backend.ini as TT
import tests.backend.common as TBC
# :seealso: `tests.backend.common.CNF_0`
CNF_0_S = """\
[DEFAULT]
a: 0
b: bbb
c: 5
[sect0]
d: x,y,z
"""
# vim:sw=4:ts=4:et:
| 23.115385 | 71 | 0.675541 | #
# Copyright (C) 2012 - 2017 Satoru SATOH <ssato @ redhat.com>
# License: MIT
#
# pylint: disable=missing-docstring,invalid-name,too-few-public-methods
from __future__ import absolute_import
import anyconfig.backend.ini as TT
import tests.backend.common as TBC
# :seealso: `tests.backend.common.CNF_0`
CNF_0_S = """\
[DEFAULT]
a: 0
b: bbb
c: 5
[sect0]
d: x,y,z
"""
class HasParserTrait(TBC.HasParserTrait):
psr = TT.Parser()
cnf_s = CNF_0_S
class Test_10(TBC.Test_10_dumps_and_loads, HasParserTrait):
load_options = dict(allow_no_value=False, defaults=None)
def test_42_loads_invalid_input(self):
invalid_cnf_s = "key=name" # No section.
self.assertRaises(Exception, self.psr.loads, invalid_cnf_s)
def test_44_loads_with_ac_parse_value_option(self):
cnf = self.psr.loads(self.cnf_s, ac_parse_value=True)
ref = self.psr.loads(self.cnf_s)
ref["DEFAULT"]["a"] = ref["sect0"]["a"] = 0
ref["DEFAULT"]["c"] = ref["sect0"]["c"] = 5
ref["sect0"]["d"] = ref["sect0"]["d"].split(',')
self._assert_dicts_equal(cnf, ref=ref)
class Test_20(TBC.Test_20_dump_and_load, HasParserTrait):
pass
# vim:sw=4:ts=4:et:
| 476 | 263 | 69 |
1ac8d5589ceca70f9e89f1d4ba9ffbf6e4b2d824 | 1,257 | py | Python | yandex_algorithm2/home1c.py | erjan/coding_exercises | 53ba035be85f1e7a12b4d4dbf546863324740467 | [
"Apache-2.0"
] | null | null | null | yandex_algorithm2/home1c.py | erjan/coding_exercises | 53ba035be85f1e7a12b4d4dbf546863324740467 | [
"Apache-2.0"
] | null | null | null | yandex_algorithm2/home1c.py | erjan/coding_exercises | 53ba035be85f1e7a12b4d4dbf546863324740467 | [
"Apache-2.0"
] | null | null | null | '''
Как известно, два наиболее распространённых формата записи даты — это европейский (сначала день, потом месяц, потом год) и американски (сначала месяц, потом день, потом год). Системный администратор поменял дату на одном из бэкапов и сейчас хочет вернуть дату обратно. Но он не проверил, в каком формате дата используется в системе. Может ли он обойтись без этой информации?
Иначе говоря, вам даётся запись некоторой корректной даты. Требуется выяснить, однозначно ли по этой записи определяется дата даже без дополнительной информации о формате.
Формат ввода
Первая строка входных данных содержит три целых числа —
x
,
y
и
z
(
1
≤
x
≤
3
1
,
1
≤
y
≤
3
1
,
1
9
7
0
≤
z
≤
2
0
6
9
. Гарантируется, что хотя бы в одном формате запись
x
y
z
задаёт корректную дату.
Формат вывода
Выведите 1, если дата определяется однозначно, и 0 в противном случае.
'''
if __name__ == "__main__":
l = list(map(int, input().split()))
result = helper(l)
print(result)
| 18.761194 | 374 | 0.686555 | '''
Как известно, два наиболее распространённых формата записи даты — это европейский (сначала день, потом месяц, потом год) и американски (сначала месяц, потом день, потом год). Системный администратор поменял дату на одном из бэкапов и сейчас хочет вернуть дату обратно. Но он не проверил, в каком формате дата используется в системе. Может ли он обойтись без этой информации?
Иначе говоря, вам даётся запись некоторой корректной даты. Требуется выяснить, однозначно ли по этой записи определяется дата даже без дополнительной информации о формате.
Формат ввода
Первая строка входных данных содержит три целых числа —
x
,
y
и
z
(
1
≤
x
≤
3
1
,
1
≤
y
≤
3
1
,
1
9
7
0
≤
z
≤
2
0
6
9
. Гарантируется, что хотя бы в одном формате запись
x
y
z
задаёт корректную дату.
Формат вывода
Выведите 1, если дата определяется однозначно, и 0 в противном случае.
'''
def helper(l):
first = l[0]
second = l[1]
if first == second:
return 1
if first <= 12 and second <= 12:
return 0
elif first > 12 and second <= 12:
return 1
elif first <= 12 and second > 12:
return 1
else:
return 0
if __name__ == "__main__":
l = list(map(int, input().split()))
result = helper(l)
print(result)
| 261 | 0 | 23 |
59dfed6ae1adfd455ea56e6f83cd3404c581a165 | 655 | py | Python | tests/test_ase.py | jat255/MaterialsIO | 04df70eddc9d0a464f7a089cf753ce26c2adf81f | [
"Apache-2.0"
] | 10 | 2019-03-25T01:16:48.000Z | 2022-02-23T16:47:02.000Z | tests/test_ase.py | jat255/MaterialsIO | 04df70eddc9d0a464f7a089cf753ce26c2adf81f | [
"Apache-2.0"
] | 31 | 2019-02-05T22:47:44.000Z | 2022-03-25T21:50:55.000Z | tests/test_ase.py | jat255/MaterialsIO | 04df70eddc9d0a464f7a089cf753ce26c2adf81f | [
"Apache-2.0"
] | 2 | 2019-11-12T18:30:49.000Z | 2022-01-13T20:04:01.000Z | from materials_io.ase import AseParser
from math import isclose
import pytest
import os
@pytest.fixture
@pytest.fixture
| 22.586207 | 86 | 0.69313 | from materials_io.ase import AseParser
from math import isclose
import pytest
import os
@pytest.fixture
def ase():
return os.path.join(os.path.dirname(__file__), 'data', 'gaussian', 'molecule.log')
@pytest.fixture
def parser():
return AseParser()
def test_ase(parser, ase):
output = parser.parse(ase)
# Check the chemical formula
assert output['chemical_formula'] == "C38H14N8O12"
# Check the shape of the force outputs. There should be
# 72 atoms and forces in 3 directions
assert len(output['forces'][0]) == 72
assert len(output['forces'][0][0]) == 3
assert isclose(output['energy'], -76063.21525532556)
| 463 | 0 | 67 |
dfee84d5a90ad6258f2c7a73186f191a93aded33 | 2,038 | py | Python | myTreeView.py | hebingyedu/Hyperspetral-Image-segmentation | f5c89358cb57227eb88ed9e49597a548c083e0dc | [
"MIT"
] | 1 | 2021-07-17T00:30:00.000Z | 2021-07-17T00:30:00.000Z | myTreeView.py | hebingyedu/Hyperspetral-Image-segmentation | f5c89358cb57227eb88ed9e49597a548c083e0dc | [
"MIT"
] | null | null | null | myTreeView.py | hebingyedu/Hyperspetral-Image-segmentation | f5c89358cb57227eb88ed9e49597a548c083e0dc | [
"MIT"
] | null | null | null | #coding=utf-8
from PySide2 import QtWidgets, QtCore, QtGui
import numpy as np
import os
import os.path
import processRS
| 35.137931 | 77 | 0.647694 | #coding=utf-8
from PySide2 import QtWidgets, QtCore, QtGui
import numpy as np
import os
import os.path
import processRS
class myTreeView(QtWidgets.QWidget):
def __init__(self, treeList=[],parent=None):
super().__init__(parent)
self.treeWidget=QtWidgets.QTreeView(parent)
self.fileModel=QtGui.QStandardItemModel(self)
self.treeList=treeList
#右键菜单
self.treeWidget.setContextMenuPolicy(QtCore.Qt.CustomContextMenu)
self.treeWidget.customContextMenuRequested.connect(self.showTreeMenu)
self.treeWidget.setHeaderHidden(True)
#设置布局
self.dockVLayout=QtWidgets.QVBoxLayout()
self.dockHLayout=QtWidgets.QHBoxLayout()
self.dockVLayout.addLayout(self.dockHLayout)
self.dockVLayout.addWidget(self.treeWidget)
self.setLayout(self.dockVLayout)
self.setWindowFlag(QtCore.Qt.WindowType.Tool)
def showTreeMenu(self,pos):
#print('mouse rightclicked')
curIndex=self.treeWidget.indexAt(pos)
if(curIndex.parent):
bandnum=curIndex.row()+1
gdalobject=self.treeList[curIndex.parent().row()]
print (gdalobject.filename,bandnum)
r=gdalobject.GDALReadFile([bandnum])
#img=np.zeros((gdalobject.XSize,gdalobject.YSize))
img=r[0]
print(img.shape)
cv2.imshow('show grey',img.astype(np.uint8))
cv2.waitKey(0)
cv2.destroyAllWindows()
def showTree(self):
#self.treeWidget.clear()
self.fileModel.clear()
for gdalObject in self.treeList:
gdalObject.Getparam()
rootItem=QtGui.QStandardItem(gdalObject.filename)
for bandi in gdalObject.bandindex:
print(bandi)
childItem=QtGui.QStandardItem('band '+ str(bandi))
rootItem.setChild(bandi-1,childItem)
self.fileModel.appendRow([rootItem])
self.treeWidget.setModel(self.fileModel)
self.treeWidget.show()
| 1,814 | 15 | 101 |
4ffce3d4a85b47defc757d070769705d37255bbe | 486 | py | Python | view/console.py | germank/training-monitor | 5cd27c8a6fe370840c602e3324fe9a46822be9d0 | [
"MIT"
] | null | null | null | view/console.py | germank/training-monitor | 5cd27c8a6fe370840c602e3324fe9a46822be9d0 | [
"MIT"
] | null | null | null | view/console.py | germank/training-monitor | 5cd27c8a6fe370840c602e3324fe9a46822be9d0 | [
"MIT"
] | null | null | null | import cmd
import logging
from signalslot import Signal
| 21.130435 | 50 | 0.613169 | import cmd
import logging
from signalslot import Signal
class CommandConsole(cmd.Cmd):
def __init__(self):
cmd.Cmd.__init__(self)
self.save_command = Signal(['output_dir'])
self.quit_command = Signal()
def do_save(self, line):
self.save_command.emit(output_dir=line)
def do_EOF(self, line):
return self.do_quit(line)
def do_quit(self, line):
self.quit_command.emit()
return True
| 260 | 9 | 151 |
cc5ea48635d7da4e40a03291da25afd53bb7a8d4 | 894 | py | Python | 02 - Curso Em Video/Aula 23/E - 113.py | GabrielTrentino/Python_Basico | f13f6448c275c14896337d2018b04cbf5a54efd3 | [
"MIT"
] | null | null | null | 02 - Curso Em Video/Aula 23/E - 113.py | GabrielTrentino/Python_Basico | f13f6448c275c14896337d2018b04cbf5a54efd3 | [
"MIT"
] | null | null | null | 02 - Curso Em Video/Aula 23/E - 113.py | GabrielTrentino/Python_Basico | f13f6448c275c14896337d2018b04cbf5a54efd3 | [
"MIT"
] | null | null | null |
nint = leiaInt('Digite um numero Inteiro: ')
nfloat = leiaFloat('Digite um número Flutuante: ')
print(f'Você digitou um número Inteiro: {nint}\nE um número Flutuante: {nfloat}') | 35.76 | 81 | 0.579418 | def leiaInt(msg):
while True:
try:
num = int(input(msg).replace(',','.').strip())
except(ValueError, TypeError, Exception) as erro:
print('\033[0;31mERRO! Digite um número inteiro valido\033[m')
print(f'Você digitou um erro: {erro.__class__}')
else:
return num
def leiaFloat(msg):
while True:
try:
num = float(input(msg).replace(',','.').strip())
except(ValueError, TypeError, Exception) as erro:
print('\033[31mDigite um número Flutuante certo!\033[m')
print(f'Você digitou e causou um erro do tipo: {erro.__class__}')
else:
return num
nint = leiaInt('Digite um numero Inteiro: ')
nfloat = leiaFloat('Digite um número Flutuante: ')
print(f'Você digitou um número Inteiro: {nint}\nE um número Flutuante: {nfloat}') | 665 | 0 | 48 |
b7d2c1ca306aa8bfac8f95fd02bac2ff9a17f2e2 | 6,255 | py | Python | ipme/utils/functions.py | evdoxiataka/ipme | ed2aa5322bf719994237a743adff89547b18a142 | [
"MIT"
] | 22 | 2020-10-31T21:39:18.000Z | 2021-12-05T02:21:15.000Z | ipme/utils/functions.py | evdoxiataka/ipme | ed2aa5322bf719994237a743adff89547b18a142 | [
"MIT"
] | null | null | null | ipme/utils/functions.py | evdoxiataka/ipme | ed2aa5322bf719994237a743adff89547b18a142 | [
"MIT"
] | 2 | 2021-01-27T09:14:02.000Z | 2021-03-05T02:47:20.000Z | import numpy as np
from math import gcd, ceil
# import itertools
def lcm(list_of_int):
"""
Get the Least Common Multyply (lcm) of a list of integer numbers.
Parameters:
--------
list_of_int A List of integers
Returns:
--------
An integer (the lcm of list_of_int).
"""
try:
lcm = list_of_int[0]
for i in list_of_int[1:]:
lcm = lcm*i/gcd(int(lcm), i)
return int(lcm)
except IndexError:
return None
def get_hist_bins_range(samples, func, var_type, ref_length = None, ref_values=None):
"""
Parameters:
--------
samples Flatten finite samples
func Predictive check criterion {'min','max','mean','std'}
var_type Variable type in {'Discrete','Continuous'}
ref_length A reference length for bin to estimate the number of bins
ref_values A numpy.ndarray with the unique values of a Discrete variable
"""
if func == 'min' or func == 'max' and var_type == "Discrete":
if ref_values is not None:
if len(ref_values)<20:
min_v = ref_values.min()
max_v = ref_values.max()
bins = len(ref_values)
if bins > 1:
range = ( min_v, max_v + (max_v - min_v) / (bins - 1))
else:
range = ( min_v, min_v+1)
return (bins, range)
else:
values = np.unique(samples)
if len(values) < 20:
min_v = values.min()
max_v = values.max()
bins = len(values)
if bins > 1:
range = ( min_v, max_v + (max_v - min_v) / (bins - 1))
else:
range = ( min_v, min_v+1)
return (bins, range)
range = (samples.min(),samples.max())
if ref_length:
bins = ceil((range[1] - range[0]) / ref_length)
range = (range[0], range[0] + bins*ref_length)
else:
bins = 20
return (bins, range)
def get_dim_names_options(dim):
"""
dim: imd.Dimension object
"""
name1 = dim.name
name2 = None
options1 = dim.values
options2 = []
if "_idx_" in name1:
idx = name1.find("_idx_")
st_n1 = idx + 5
end_n1 = len(name1)
name2 = name1[st_n1:end_n1]
name1 = name1[0:idx]
values = np.array(dim.values)
options1 = np.unique(values).tolist()
if len(options1):
tmp = np.arange(np.count_nonzero(values == options1[0]))
options2 = list(map(str,tmp))
return (name1, name2, options1, options2)
def get_w2_w1_val_mapping(dim):
"""
dim: imd.Dimension object
Returns:
-------
A Dict {<opt1_val>: A List of <opt2_val> for this <opt1_val>}
"""
options1 = dim.values
options2 = []
values = np.array(dim.values)
options1 = np.unique(values)
val_dict = {}
if len(options1):
for v1 in options1:
tmp = np.arange(np.count_nonzero(values == v1))
val_dict[v1] = list(map(str,tmp))
return val_dict
| 31.913265 | 91 | 0.53717 | import numpy as np
from math import gcd, ceil
# import itertools
def lcm(list_of_int):
"""
Get the Least Common Multyply (lcm) of a list of integer numbers.
Parameters:
--------
list_of_int A List of integers
Returns:
--------
An integer (the lcm of list_of_int).
"""
try:
lcm = list_of_int[0]
for i in list_of_int[1:]:
lcm = lcm*i/gcd(int(lcm), i)
return int(lcm)
except IndexError:
return None
def find_indices(lst, condition, xmin=0, xmax=0):
# return [i for i, elem in enumerate(lst) if condition(elem)]
# return [ _ for _ in itertools.compress(list(range(0,len(lst))), map(condition,lst)) ]
return list(np.where((lst>=xmin) & (lst<=xmax))[0])
def find_inds_before_after(lst, el):
inds_sm=find_indices(lst, lambda e: e<= el)
if len(inds_sm):
ind_before=inds_sm[-1]
else:
ind_before=-1
inds_bi=find_indices(lst, lambda e: e>= el)
if len(inds_bi):
ind_after=inds_bi[0]
else:
ind_after=-1
return (ind_before,ind_after)
def find_highest_point( x, y):
x=np.asarray(x, dtype=np.float64)
y=np.asarray(y, dtype=np.float64)
if len(y):
max_idx=np.argmax(y)
return (x[max_idx],y[max_idx])
else:
return ()
def get_samples_for_pred_check(samples, func):
# samples = np.asarray(samples)
shape = samples.shape
if 0 not in shape:
if len(shape) == 1:
axis = 0
else:
axis = 1
if len(shape) > 2:
fir_dim = shape[0]
sec_dim = 1
for i in np.arange(1,len(shape),1):
sec_dim = sec_dim*shape[i]
samples=samples.reshape(fir_dim,sec_dim)
if func == "min":
samples = samples.min(axis=axis)
elif func == "max":
samples = samples.max(axis=axis)
elif func == "mean":
samples = samples.mean(axis=axis)
elif func == "std":
samples = samples.std(axis=axis)
else:
samples = np.empty([1, 2])
if ~np.isfinite(samples).all():
samples = get_finite_samples(samples)
else:
samples = np.empty([1, 2])
return samples
def get_finite_samples(np_array):
if isinstance(np_array, np.ndarray):
shape = len(np_array.shape)
if shape == 1:
np_array = np_array[np.isfinite(np_array)]
elif shape > 1:
samples_idx = np.isfinite(np_array).all(axis=shape-1)
for axis in np.arange(shape-2,0,-1):
samples_idx = np.isfinite(np_array).all(axis=axis)
np_array = np_array[samples_idx]
return np_array
def get_hist_bins_range(samples, func, var_type, ref_length = None, ref_values=None):
"""
Parameters:
--------
samples Flatten finite samples
func Predictive check criterion {'min','max','mean','std'}
var_type Variable type in {'Discrete','Continuous'}
ref_length A reference length for bin to estimate the number of bins
ref_values A numpy.ndarray with the unique values of a Discrete variable
"""
if func == 'min' or func == 'max' and var_type == "Discrete":
if ref_values is not None:
if len(ref_values)<20:
min_v = ref_values.min()
max_v = ref_values.max()
bins = len(ref_values)
if bins > 1:
range = ( min_v, max_v + (max_v - min_v) / (bins - 1))
else:
range = ( min_v, min_v+1)
return (bins, range)
else:
values = np.unique(samples)
if len(values) < 20:
min_v = values.min()
max_v = values.max()
bins = len(values)
if bins > 1:
range = ( min_v, max_v + (max_v - min_v) / (bins - 1))
else:
range = ( min_v, min_v+1)
return (bins, range)
range = (samples.min(),samples.max())
if ref_length:
bins = ceil((range[1] - range[0]) / ref_length)
range = (range[0], range[0] + bins*ref_length)
else:
bins = 20
return (bins, range)
def get_dim_names_options(dim):
"""
dim: imd.Dimension object
"""
name1 = dim.name
name2 = None
options1 = dim.values
options2 = []
if "_idx_" in name1:
idx = name1.find("_idx_")
st_n1 = idx + 5
end_n1 = len(name1)
name2 = name1[st_n1:end_n1]
name1 = name1[0:idx]
values = np.array(dim.values)
options1 = np.unique(values).tolist()
if len(options1):
tmp = np.arange(np.count_nonzero(values == options1[0]))
options2 = list(map(str,tmp))
return (name1, name2, options1, options2)
def get_w2_w1_val_mapping(dim):
"""
dim: imd.Dimension object
Returns:
-------
A Dict {<opt1_val>: A List of <opt2_val> for this <opt1_val>}
"""
options1 = dim.values
options2 = []
values = np.array(dim.values)
options1 = np.unique(values)
val_dict = {}
if len(options1):
for v1 in options1:
tmp = np.arange(np.count_nonzero(values == v1))
val_dict[v1] = list(map(str,tmp))
return val_dict
def get_stratum_range(samples, stratum):
median = np.median(samples)
if stratum == 0 or stratum == 1:
inds_l = np.where(samples<median)[0]
median_l = np.median(samples[inds_l])
if stratum == 0:
xmin = np.min(samples).item()
xmax = median_l
elif stratum == 1:
xmin = median_l
xmax = median
elif stratum == 2 or stratum == 3:
inds_h = np.where(samples>=median)[0]
median_h = np.median(samples[inds_h])
if stratum == 2:
xmin = median
xmax = median_h
elif stratum == 3:
xmin = median_h
xmax = np.max(samples).item()
else:
xmin = np.min(samples).item()
xmax = np.max(samples).item()
return (xmin,xmax)
| 2,876 | 0 | 138 |
1f19f13e91c75f3090fe3500b07c916348d4e0d3 | 317 | py | Python | py/so/light_pixels.py | YodaEmbedding/experiments | 567c6a1c18fac2d951fe2af54aaa4917b7d529d2 | [
"MIT"
] | null | null | null | py/so/light_pixels.py | YodaEmbedding/experiments | 567c6a1c18fac2d951fe2af54aaa4917b7d529d2 | [
"MIT"
] | null | null | null | py/so/light_pixels.py | YodaEmbedding/experiments | 567c6a1c18fac2d951fe2af54aaa4917b7d529d2 | [
"MIT"
] | null | null | null | #!/usr/bin/env python3
import cv2
import numpy as np
# img = cv2.imread('4.png')
# hsv = cv2.cvtColor(img, cv2.COLOR_BGR2HSV)
# img[hsv[:,:,0] > 20] = 255
img = cv2.imread("4.png")
gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)
img[gray > 200] = 255
cv2.imwrite("4_.png", img)
cv2.imshow("img", img)
cv2.waitKey(0)
| 18.647059 | 44 | 0.659306 | #!/usr/bin/env python3
import cv2
import numpy as np
# img = cv2.imread('4.png')
# hsv = cv2.cvtColor(img, cv2.COLOR_BGR2HSV)
# img[hsv[:,:,0] > 20] = 255
img = cv2.imread("4.png")
gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)
img[gray > 200] = 255
cv2.imwrite("4_.png", img)
cv2.imshow("img", img)
cv2.waitKey(0)
| 0 | 0 | 0 |
22c03d85273a3ae5d9bb0f4814acf96c016dc277 | 2,303 | py | Python | cmdfunctools/types.py | Mr-Io/chtool | 4d4337b016c53d7e19d3b9dcec161b3ad115b611 | [
"MIT"
] | 1 | 2019-04-23T16:49:42.000Z | 2019-04-23T16:49:42.000Z | cmdfunctools/types.py | Mr-Io/chtool | 4d4337b016c53d7e19d3b9dcec161b3ad115b611 | [
"MIT"
] | null | null | null | cmdfunctools/types.py | Mr-Io/chtool | 4d4337b016c53d7e19d3b9dcec161b3ad115b611 | [
"MIT"
] | null | null | null | # -*- coding: utf-8 -*-
"""
cmdfunctools.types
~~~~~~~~~~~~~~~~~~~~~~~
Module with the type checkers for the function/commandline args.
"""
import string
# ***********************************************************
# default functions used when type argument are not provided
def default_true(*arg):
"""
Auxiliar function to be used if type is not provided.
:return: True
"""
return True
def default_arg(*arg):
"""
Auxiliar function to be used if type is not provided.
:return: input arg
"""
return arg[0]
# ***********************************************************
# The different types checkers are defined here:
def is_hex(arg: str) -> bool:
"""
Check if a string value represent an hexadecimal value.
:return: True/False
"""
res = all(c in (string.hexdigits + '\n') for c in arg)
return res
def is_int(arg: str) -> bool:
"""
Check if a string value represent an hexadecimal value.
:return: True/False
"""
res = all(c in (string.digits + '\n') for c in arg)
return res
# dictionary with all functions for the type-check
dct_type_check = {
'int': is_int,
'hex': is_hex,
}
# ***********************************************************
# The converters from type to bytes are defined here
def from_hex(arg: str, length: int) -> bytes:
"""
Transform a hexadecimal string into a bytearray of a given length.
:return: a bytearray
"""
narg = int(arg, 16)
if length == 0:
length = 512
barg = narg.to_bytes(int(length/2), 'big')
return barg
def from_int(arg: str, length: int = -1) -> bytes:
"""
Transform a integer string into a bytearray of a given length.
:return: a bytearray
"""
narg = int(arg)
barg = narg.to_bytes(length, 'big')
return barg
# dictionary with all functions for transforming specific-string-types into bytes
dct_convert_from = {
'int': from_int,
'hex': from_hex,
}
# ***********************************************************
# The converters from bytes to another type are defined here
# dictionary with all functions for transforming bytes into other type
dct_convert_to = {
'int': to_int,
}
| 21.933333 | 81 | 0.569257 | # -*- coding: utf-8 -*-
"""
cmdfunctools.types
~~~~~~~~~~~~~~~~~~~~~~~
Module with the type checkers for the function/commandline args.
"""
import string
# ***********************************************************
# default functions used when type argument are not provided
def default_true(*arg):
"""
Auxiliar function to be used if type is not provided.
:return: True
"""
return True
def default_arg(*arg):
"""
Auxiliar function to be used if type is not provided.
:return: input arg
"""
return arg[0]
# ***********************************************************
# The different types checkers are defined here:
def is_hex(arg: str) -> bool:
"""
Check if a string value represent an hexadecimal value.
:return: True/False
"""
res = all(c in (string.hexdigits + '\n') for c in arg)
return res
def is_int(arg: str) -> bool:
"""
Check if a string value represent an hexadecimal value.
:return: True/False
"""
res = all(c in (string.digits + '\n') for c in arg)
return res
# dictionary with all functions for the type-check
dct_type_check = {
'int': is_int,
'hex': is_hex,
}
# ***********************************************************
# The converters from type to bytes are defined here
def from_hex(arg: str, length: int) -> bytes:
"""
Transform a hexadecimal string into a bytearray of a given length.
:return: a bytearray
"""
narg = int(arg, 16)
if length == 0:
length = 512
barg = narg.to_bytes(int(length/2), 'big')
return barg
def from_int(arg: str, length: int = -1) -> bytes:
"""
Transform a integer string into a bytearray of a given length.
:return: a bytearray
"""
narg = int(arg)
barg = narg.to_bytes(length, 'big')
return barg
# dictionary with all functions for transforming specific-string-types into bytes
dct_convert_from = {
'int': from_int,
'hex': from_hex,
}
# ***********************************************************
# The converters from bytes to another type are defined here
def to_int(arg: bytes) -> int:
narg = int.from_bytes(arg, 'big')
return narg
# dictionary with all functions for transforming bytes into other type
dct_convert_to = {
'int': to_int,
}
| 63 | 0 | 22 |
5785cc952971cae8790212bea0f5458065c88ad7 | 291 | py | Python | api/models.py | Rytiggy/osfPages-service- | 06efc49c529f2a2623d03b9c790e931b84110d24 | [
"Apache-2.0"
] | null | null | null | api/models.py | Rytiggy/osfPages-service- | 06efc49c529f2a2623d03b9c790e931b84110d24 | [
"Apache-2.0"
] | 3 | 2020-02-12T00:11:57.000Z | 2021-06-10T19:50:20.000Z | api/models.py | Rytiggy/osfPages-service- | 06efc49c529f2a2623d03b9c790e931b84110d24 | [
"Apache-2.0"
] | 2 | 2017-08-28T18:58:11.000Z | 2017-08-28T18:59:33.000Z | from django.db import models
# Create your models here.
| 29.1 | 64 | 0.780069 | from django.db import models
# Create your models here.
class Home(models.Model):
guid = models.CharField(max_length=5, primary_key=True)
page_data = models.TextField()
unpublished_page_data = models.TextField(null=True, blank=True)
meta_data = models.TextField(null=True, blank=True)
| 0 | 211 | 23 |
0226bec4d9ef01c8ee635854ea31a5261ea22df3 | 7,402 | py | Python | shell-server.py | MathieuMorrissette/WebsocketShelling | d63e09216df10b2a38ee2d5cb9b2ee6531ff37c9 | [
"Unlicense"
] | null | null | null | shell-server.py | MathieuMorrissette/WebsocketShelling | d63e09216df10b2a38ee2d5cb9b2ee6531ff37c9 | [
"Unlicense"
] | null | null | null | shell-server.py | MathieuMorrissette/WebsocketShelling | d63e09216df10b2a38ee2d5cb9b2ee6531ff37c9 | [
"Unlicense"
] | null | null | null | import asyncio
import ssl
import websockets
import os
import json
import time
import base64
import random
import string
import hashlib
from OpenSSL import crypto, SSL
from datetime import datetime
server_password = ""
with open("./config.txt", "r") as config:
data = config.read()
data = json.loads(data)
server_password = base64.b64decode(data["server_password"])
print(generate_random(10))
cert_gen()
pwneds = []
clients = []
ssl_context = ssl.SSLContext(ssl.PROTOCOL_TLS_SERVER)
ssl_context.load_cert_chain("./selfsigned.crt", keyfile="./private.key")
start_server = websockets.serve(
handle_socket, "localhost", 8765, ssl=ssl_context, max_size=None, ping_interval=None
)
asyncio.get_event_loop().run_until_complete(start_server)
asyncio.get_event_loop().run_forever() | 32.897778 | 110 | 0.586598 | import asyncio
import ssl
import websockets
import os
import json
import time
import base64
import random
import string
import hashlib
from OpenSSL import crypto, SSL
from datetime import datetime
server_password = ""
with open("./config.txt", "r") as config:
data = config.read()
data = json.loads(data)
server_password = base64.b64decode(data["server_password"])
def generate_random(number):
return ''.join(random.choices(string.ascii_uppercase + string.digits, k=number))
def validate_password(password_to_check):
salt = server_password[:32]
new_key = hashlib.pbkdf2_hmac(
'sha256',
password_to_check.encode('utf-8'), # Convert the password to bytes
salt,
100000
)
return new_key == server_password[32:]
print(generate_random(10))
def cert_gen(
emailAddress=generate_random(random.randint(5, 20)),
commonName=generate_random(random.randint(5, 20)),
countryName=generate_random(2),
localityName=generate_random(random.randint(5, 20)),
stateOrProvinceName=generate_random(random.randint(5, 20)),
organizationName=generate_random(random.randint(5, 20)),
organizationUnitName=generate_random(random.randint(5, 20)),
serialNumber=0,
validityStartInSeconds=0,
validityEndInSeconds=10*365*24*60*60,
KEY_FILE = "private.key",
CERT_FILE="selfsigned.crt"):
if(os.path.exists("./private.key") and os.path.exists("./selfsigned.crt")):
return
# create a key pair
k = crypto.PKey()
k.generate_key(crypto.TYPE_RSA, 4096)
# create a self-signed cert
cert = crypto.X509()
cert.get_subject().C = countryName
cert.get_subject().ST = stateOrProvinceName
cert.get_subject().L = localityName
cert.get_subject().O = organizationName
cert.get_subject().OU = organizationUnitName
cert.get_subject().CN = commonName
cert.get_subject().emailAddress = emailAddress
cert.set_serial_number(serialNumber)
cert.gmtime_adj_notBefore(0)
cert.gmtime_adj_notAfter(validityEndInSeconds)
cert.set_issuer(cert.get_subject())
cert.set_pubkey(k)
cert.sign(k, 'sha512')
with open(CERT_FILE, "wt") as f:
f.write(crypto.dump_certificate(crypto.FILETYPE_PEM, cert).decode("utf-8"))
with open(KEY_FILE, "wt") as f:
f.write(crypto.dump_privatekey(crypto.FILETYPE_PEM, k).decode("utf-8"))
cert_gen()
pwneds = []
clients = []
def log(socket, message):
ip = ""
if(socket != None and socket.remote_address != None):
ip = socket.remote_address[0]
# pad
while len(ip) < 15:
ip = ip + " "
log = str(datetime.now()) + " - " + ip + " - " + message
print("[LOG]" + log)
with open("logs.txt", "a") as log_file:
log_file.write(log + "\n")
def checkExists(key, value, array):
for item in array:
if item[key] == value:
return True
return False
async def handle_socket(websocket, path):
# receive client connection
receive_data_raw = await websocket.recv()
receive_data = json.loads(receive_data_raw)
if receive_data["command"] == "register_pwned":
pwned_hostname = receive_data["pwned_hostname"]
pwned_id = len(pwneds)
pwneds.append({"pwned_id" : pwned_id, "socket" : websocket, "hostname" : pwned_hostname })
log(websocket, "shell from " + pwned_hostname + " - Pwned ID : " + str(pwned_id))
try:
# pwned stay alive
while checkExists("pwned_id", pwned_id, pwneds):
# wait for data (usually command output)
command_output_raw = await websocket.recv()
command_output = json.loads(command_output_raw)
# find client to send output to
for client in clients:
if client["client_id"] == command_output["client_id"]:
try:
client_sock = client["socket"]
await client_sock.send(command_output_raw)
except:
clients.pop(command_output["client_id"])
except:
pwneds.pop(pwned_id)
raise
else:
if(server_password != "" and not validate_password(receive_data["password"])):
receive_data["output"] = "denied password"
log(websocket, "client denied")
await websocket.send(json.dumps(receive_data))
return
client_hostname = receive_data["client_hostname"]
client_id = len(clients)
log(websocket, "client " + client_hostname + " connected - Client ID " + str(client_id))
clients.append({ "client_id" : client_id, "socket" : websocket, "hostname" : client_hostname })
receive_data["output"] = "connection successful"
await websocket.send(json.dumps(receive_data))
try:
while checkExists("client_id", client_id, clients):
# wait for command
receive_data_raw = await websocket.recv()
receive_data = json.loads(receive_data_raw)
receive_data["client_id"] = client_id
if receive_data["command"] == "list":
list_output = ""
for pwned in pwneds:
list_output = str(pwned["pwned_id"]) + " - " + pwned["hostname"]+ '\r\n' + list_output
receive_data["output"] = list_output
await websocket.send(json.dumps(receive_data))
elif receive_data["command"] == "killallpwneds":
for item in list(pwneds): # list to prevent dictionnary changed
try:
if(item in pwneds):
log(websocket, "closing " + item)
await pwneds[item].close()
# they will be automatically deleted
except Exception as ex:
log(websocket, str(ex))
receive_data["output"] = "killed all"
await websocket.send(json.dumps(receive_data))
else:
# find pwned to send command to
found = False
for pwned in pwneds:
if pwned["pwned_id"] == receive_data["pwned_id"]:
found = True
try:
pwned_sock = pwned["socket"]
await pwned_sock.send(json.dumps(receive_data))
except:
pwneds.pop(receive_data["pwned_id"])
if not found:
receive_data["output"] = "pwned not found!"
await websocket.send(json.dumps(receive_data))
except Exception as ex:
clients.pop(client_id)
log(websocket, str(ex))
ssl_context = ssl.SSLContext(ssl.PROTOCOL_TLS_SERVER)
ssl_context.load_cert_chain("./selfsigned.crt", keyfile="./private.key")
start_server = websockets.serve(
handle_socket, "localhost", 8765, ssl=ssl_context, max_size=None, ping_interval=None
)
asyncio.get_event_loop().run_until_complete(start_server)
asyncio.get_event_loop().run_forever() | 6,439 | 0 | 138 |
1a4d3826fe05523d57fcaa16a40fcae86b60e476 | 7,334 | py | Python | src/numcolorpy.py | dlanier/FlyingMachineFractal | 3889bfb8b8005717de83954d4a705154e908273b | [
"MIT"
] | 4 | 2017-03-27T19:50:37.000Z | 2021-07-15T03:23:03.000Z | pyreimpic/zcolormap.py | dlanier/FlyingMachineFractal | 3889bfb8b8005717de83954d4a705154e908273b | [
"MIT"
] | null | null | null | pyreimpic/zcolormap.py | dlanier/FlyingMachineFractal | 3889bfb8b8005717de83954d4a705154e908273b | [
"MIT"
] | 1 | 2017-03-27T19:50:54.000Z | 2017-03-27T19:50:54.000Z | # -*- coding: utf-8 -*-
"""
numcolorpy.py
Created Saturday April 22 2017
@author: del
lanier4@illinois.edu
mradmstr514226508@gmail.com
import numcolorpy as ncp
"""
import time
import numpy as np
from PIL import Image as IP
from PIL import ImageColor as IC
import colorsys
def range_norm(Z, lo=0.0, hi=1.0):
""" normaize input matrix Z within a lo - hi range
"""
I = graphic_norm(Z)
hi = max(min(hi, 1.0), 0.0)
lo = min(max(lo, 0.0), 1.0)
low_fence = min(hi, lo)
hi_fence = max(hi, lo)
if low_fence == hi_fence:
return I
v_span = hi_fence - low_fence
I = I * v_span + low_fence
return I
def etg_norm(Z0, Z, ET):
""" Zd, Zr, ETn = etg_norm(Z0, Z, ET); Graphically usable matrices from escape time algorithm result
"""
ETn = mat2graphic(ET)
Zv = Z - Z0
Zd = mat2graphic(Zv)
Zr = mat2graphic(np.arctan2(np.imag(Zv), np.real(Zv)))
return Zd, Zr, ETn
def mat2graphic(Z):
""" M, nClrs = mat2graphic(Z)
Use all the transformation tricks to prepare input matrix Z
for conversion to a viewable image.
Args:
Z: real or complex (rows x xcols x 1) matrix
Returns:
M: real (rows x xcols x 1) matrix (0 <= M <= 1)
"""
M, nClrs = flat_index(np.abs(Z))
return graphic_norm(M)
def graphic_norm(Z):
""" rescale matrix z to distance (float) s.t.
0 <= z <= 1 (will include 0,1 if it has more than 1 value)
Args:
Z: is a real or complex two dimensional matrix
Returns:
Z: same size, real valued matrix with smallest member = 0, largest = 1
"""
EPSILON = 1e-15
I = np.abs(Z)
I = I - I.min()
return I / max(EPSILON, I.max())
def flat_index(float_mat):
""" convert the input matrix to integers from 0 to number of unique values.
Args:
float_mat: two dimensional matrix.
Return:
float_mat: re-enumerated so that the matrix values are all sequential ints.
n_colors: number of unique values in the input / output matrix
"""
rows = float_mat.shape[0]
cols = float_mat.shape[1]
float_mat = np.reshape(float_mat, (1, float_mat.size))
ixA = np.argsort(float_mat)[0]
current_value = float_mat[0, ixA[0]]
enumeration_value = 0
for ix in ixA:
if float_mat[0,ix] != current_value:
current_value = float_mat[0,ix]
enumeration_value += 1
float_mat[0,ix] = enumeration_value
float_mat = np.array(np.reshape(float_mat, (rows, cols)))
float_mat = np.int_(float_mat)
n_colors = enumeration_value + 1
return float_mat, n_colors
def mat_to_Shade(V, R_max, G_max, B_max, R_floor=0, G_floor=0, B_floor=0):
""" I = mat_to_gray(V)
matrix of values V, converted to a gray scale image
Args:
V: rows x cols x 1 numerical matrix
Returns:
I: rows x cols x 3 grayscale image
"""
R = R_max - R_floor
G = G_max - G_floor
B = B_max - B_floor
V = graphic_norm(V)
n_rows = V.shape[0]
n_cols = V.shape[1]
I = IP.new('RGB', (n_cols, n_rows))
for row in range(0, I.height):
for col in range(0, I.width):
P = tuple(np.int_(
[R_floor + V[row, col] * R, G_floor + V[row, col] * G, B_floor + V[row, col] * B]))
I.putpixel((col, row), P)
return I
def resize_color_map(mp0, n_colors):
""" givin a RGB colormap input return the same color order with n_colors number of colors
"""
mp = np.zeros((n_colors,3))
n_colors0 = mp0.shape[0]
if n_colors0 != n_colors:
tc = n_colors0 * n_colors
x = np.linspace(1,tc, n_colors0)
xq = np.linspace(1,tc, n_colors)
mp[:,0] = np.interp(xq, x, mp0[:,0])
mp[:,1] = np.interp(xq, x, mp0[:,1])
mp[:,2] = np.interp(xq, x, mp0[:,2])
return mp
def normat_hsv_intrgb(H, S, V, H_max=1.0, H_min=0.0, S_max=1.0, S_min=0.0, V_max=1.0, V_min=0.0):
""" I = normat_hsv_intrgb(H, S, V, H_max=1.0, H_min=0.0, S_max=1.0, S_min=0.0, V_max=1.0, V_min=0.0)
Three normaized matrices as hsv image converted to rgb
'normalized' means 0 <= M <= 1 where M is H, S, or V
Args:
H: rows x cols x 1 normalized matrix
S: rows x cols x 1 normalized matrix
V: rows x cols x 1 normalized matrix
Returns:
I: rows x cols x 3 hue-saturation-values image
"""
H_mul = H_max - H_min
S_mul = S_max - S_min
V_mul = V_max - V_min
n_rows = H.shape[0]
n_cols = H.shape[1]
I = IP.new('RGB', (n_cols, n_rows))
for row in range(0, I.height):
for col in range(0, I.width):
red, green, blue = colorsys.hsv_to_rgb(
H_min + H_mul * H[row, col],
S_min + S_mul * S[row, col],
V_min + V_mul * V[row, col])
red = int(np.round( red * 255 ))
green = int(np.round( green * 255 ))
blue = int(np.round( blue * 255 ))
P = (red, green, blue)
I.putpixel((col, row), P)
return I
| 27.062731 | 105 | 0.565858 | # -*- coding: utf-8 -*-
"""
numcolorpy.py
Created Saturday April 22 2017
@author: del
lanier4@illinois.edu
mradmstr514226508@gmail.com
import numcolorpy as ncp
"""
import time
import numpy as np
from PIL import Image as IP
from PIL import ImageColor as IC
import colorsys
def range_norm(Z, lo=0.0, hi=1.0):
""" normaize input matrix Z within a lo - hi range
"""
I = graphic_norm(Z)
hi = max(min(hi, 1.0), 0.0)
lo = min(max(lo, 0.0), 1.0)
low_fence = min(hi, lo)
hi_fence = max(hi, lo)
if low_fence == hi_fence:
return I
v_span = hi_fence - low_fence
I = I * v_span + low_fence
return I
def etg_norm(Z0, Z, ET):
""" Zd, Zr, ETn = etg_norm(Z0, Z, ET); Graphically usable matrices from escape time algorithm result
"""
ETn = mat2graphic(ET)
Zv = Z - Z0
Zd = mat2graphic(Zv)
Zr = mat2graphic(np.arctan2(np.imag(Zv), np.real(Zv)))
return Zd, Zr, ETn
def mat2graphic(Z):
""" M, nClrs = mat2graphic(Z)
Use all the transformation tricks to prepare input matrix Z
for conversion to a viewable image.
Args:
Z: real or complex (rows x xcols x 1) matrix
Returns:
M: real (rows x xcols x 1) matrix (0 <= M <= 1)
"""
M, nClrs = flat_index(np.abs(Z))
return graphic_norm(M)
def graphic_norm(Z):
""" rescale matrix z to distance (float) s.t.
0 <= z <= 1 (will include 0,1 if it has more than 1 value)
Args:
Z: is a real or complex two dimensional matrix
Returns:
Z: same size, real valued matrix with smallest member = 0, largest = 1
"""
EPSILON = 1e-15
I = np.abs(Z)
I = I - I.min()
return I / max(EPSILON, I.max())
def flat_index(float_mat):
""" convert the input matrix to integers from 0 to number of unique values.
Args:
float_mat: two dimensional matrix.
Return:
float_mat: re-enumerated so that the matrix values are all sequential ints.
n_colors: number of unique values in the input / output matrix
"""
rows = float_mat.shape[0]
cols = float_mat.shape[1]
float_mat = np.reshape(float_mat, (1, float_mat.size))
ixA = np.argsort(float_mat)[0]
current_value = float_mat[0, ixA[0]]
enumeration_value = 0
for ix in ixA:
if float_mat[0,ix] != current_value:
current_value = float_mat[0,ix]
enumeration_value += 1
float_mat[0,ix] = enumeration_value
float_mat = np.array(np.reshape(float_mat, (rows, cols)))
float_mat = np.int_(float_mat)
n_colors = enumeration_value + 1
return float_mat, n_colors
def gray_mat(V):
n_rows = V.shape[0]
n_cols = V.shape[1]
V = V * 255
I = IP.new('RGB', (n_cols, n_rows))
for row in range(0, I.height):
for col in range(0, I.width):
P = tuple(np.int_([V[row, col], V[row, col], V[row, col]]))
I.putpixel((col, row), P)
return I
def rgb_2_hsv_mat(H, S, V):
n_rows = H.shape[0]
n_cols = H.shape[1]
I = IP.new('RGB', (n_cols, n_rows))
for row in range(0, I.height):
for col in range(0, I.width):
red, green, blue = colorsys.hsv_to_rgb(H[row, col], S[row, col], V[row, col])
red = int(np.round( red * 255 ))
green = int(np.round( green * 255 ))
blue = int(np.round( blue * 255 ))
P = (red, green, blue)
I.putpixel((col, row), P)
return I
def mat_to_gray(V, max_v=255, min_v=0):
R_max = max(min(max_v, 255), 0)
R_floor = min(max(min_v, 0), R_max)
G_max = max(min(max_v, 255), 0)
G_floor = min(max(min_v, 0), G_max)
B_max = max(min(max_v, 255), 0)
B_floor = min(max(min_v, 0), B_max)
return mat_to_Shade(V, R_max, G_max, B_max, R_floor, G_floor, B_floor)
def mat_to_red(V):
R_max = 255
R_floor = 180
G_max = 250
G_floor = 30
B_max = 250
B_floor = 30
return mat_to_Shade(V, R_max, G_max, B_max, R_floor, G_floor, B_floor)
def mat_to_green(V):
R_max = 250
R_floor = 30
G_max = 255
G_floor = 130
B_max = 250
B_floor = 30
return mat_to_Shade(V, R_max, G_max, B_max, R_floor, G_floor, B_floor)
def mat_to_blue(V):
R_max = 250
R_floor = 30
G_max = 250
G_floor = 30
B_max = 255
B_floor = 130
return mat_to_Shade(V, R_max, G_max, B_max, R_floor, G_floor, B_floor)
def mat_to_Shade(V, R_max, G_max, B_max, R_floor=0, G_floor=0, B_floor=0):
""" I = mat_to_gray(V)
matrix of values V, converted to a gray scale image
Args:
V: rows x cols x 1 numerical matrix
Returns:
I: rows x cols x 3 grayscale image
"""
R = R_max - R_floor
G = G_max - G_floor
B = B_max - B_floor
V = graphic_norm(V)
n_rows = V.shape[0]
n_cols = V.shape[1]
I = IP.new('RGB', (n_cols, n_rows))
for row in range(0, I.height):
for col in range(0, I.width):
P = tuple(np.int_(
[R_floor + V[row, col] * R, G_floor + V[row, col] * G, B_floor + V[row, col] * B]))
I.putpixel((col, row), P)
return I
def resize_color_map(mp0, n_colors):
""" givin a RGB colormap input return the same color order with n_colors number of colors
"""
mp = np.zeros((n_colors,3))
n_colors0 = mp0.shape[0]
if n_colors0 != n_colors:
tc = n_colors0 * n_colors
x = np.linspace(1,tc, n_colors0)
xq = np.linspace(1,tc, n_colors)
mp[:,0] = np.interp(xq, x, mp0[:,0])
mp[:,1] = np.interp(xq, x, mp0[:,1])
mp[:,2] = np.interp(xq, x, mp0[:,2])
return mp
def normat_hsv_intrgb(H, S, V, H_max=1.0, H_min=0.0, S_max=1.0, S_min=0.0, V_max=1.0, V_min=0.0):
""" I = normat_hsv_intrgb(H, S, V, H_max=1.0, H_min=0.0, S_max=1.0, S_min=0.0, V_max=1.0, V_min=0.0)
Three normaized matrices as hsv image converted to rgb
'normalized' means 0 <= M <= 1 where M is H, S, or V
Args:
H: rows x cols x 1 normalized matrix
S: rows x cols x 1 normalized matrix
V: rows x cols x 1 normalized matrix
Returns:
I: rows x cols x 3 hue-saturation-values image
"""
H_mul = H_max - H_min
S_mul = S_max - S_min
V_mul = V_max - V_min
n_rows = H.shape[0]
n_cols = H.shape[1]
I = IP.new('RGB', (n_cols, n_rows))
for row in range(0, I.height):
for col in range(0, I.width):
red, green, blue = colorsys.hsv_to_rgb(
H_min + H_mul * H[row, col],
S_min + S_mul * S[row, col],
V_min + V_mul * V[row, col])
red = int(np.round( red * 255 ))
green = int(np.round( green * 255 ))
blue = int(np.round( blue * 255 ))
P = (red, green, blue)
I.putpixel((col, row), P)
return I
def mat_to_mapped(A, mp):
n_rows = A.shape[0]
n_cols = A.shape[1]
A, nClrs = flat_index(A)
mp = resize_color_map(mp, nClrs)*255
I = IP.new('RGB', (n_cols, n_rows))
for r in range(0, n_rows):
for c in range(0, n_cols):
I.putpixel((c,r), tuple(np.uint8(mp[A[r,c], :])))
return I
| 1,949 | 0 | 161 |
36827512a9f1775495657b3d791cc3c2672adda4 | 12,952 | py | Python | venv/lib/python3.8/site-packages/hgext/remotefilelog/contentstore.py | JesseDavids/mqtta | 389eb4f06242d4473fe1bcff7fc6a22290e0d99c | [
"Apache-2.0"
] | 4 | 2021-02-05T10:57:39.000Z | 2022-02-25T04:43:23.000Z | venv/lib/python3.8/site-packages/hgext/remotefilelog/contentstore.py | JesseDavids/mqtta | 389eb4f06242d4473fe1bcff7fc6a22290e0d99c | [
"Apache-2.0"
] | null | null | null | venv/lib/python3.8/site-packages/hgext/remotefilelog/contentstore.py | JesseDavids/mqtta | 389eb4f06242d4473fe1bcff7fc6a22290e0d99c | [
"Apache-2.0"
] | null | null | null | from __future__ import absolute_import
import threading
from mercurial.node import hex, nullid
from mercurial.pycompat import getattr
from mercurial import (
mdiff,
pycompat,
revlog,
)
from . import (
basestore,
constants,
shallowutil,
)
class ChainIndicies(object):
"""A static class for easy reference to the delta chain indicies."""
# The filename of this revision delta
NAME = 0
# The mercurial file node for this revision delta
NODE = 1
# The filename of the delta base's revision. This is useful when delta
# between different files (like in the case of a move or copy, we can delta
# against the original file content).
BASENAME = 2
# The mercurial file node for the delta base revision. This is the nullid if
# this delta is a full text.
BASENODE = 3
# The actual delta or full text data.
DATA = 4
| 33.4677 | 80 | 0.602996 | from __future__ import absolute_import
import threading
from mercurial.node import hex, nullid
from mercurial.pycompat import getattr
from mercurial import (
mdiff,
pycompat,
revlog,
)
from . import (
basestore,
constants,
shallowutil,
)
class ChainIndicies(object):
"""A static class for easy reference to the delta chain indicies."""
# The filename of this revision delta
NAME = 0
# The mercurial file node for this revision delta
NODE = 1
# The filename of the delta base's revision. This is useful when delta
# between different files (like in the case of a move or copy, we can delta
# against the original file content).
BASENAME = 2
# The mercurial file node for the delta base revision. This is the nullid if
# this delta is a full text.
BASENODE = 3
# The actual delta or full text data.
DATA = 4
class unioncontentstore(basestore.baseunionstore):
def __init__(self, *args, **kwargs):
super(unioncontentstore, self).__init__(*args, **kwargs)
self.stores = args
self.writestore = kwargs.get('writestore')
# If allowincomplete==True then the union store can return partial
# delta chains, otherwise it will throw a KeyError if a full
# deltachain can't be found.
self.allowincomplete = kwargs.get('allowincomplete', False)
def get(self, name, node):
"""Fetches the full text revision contents of the given name+node pair.
If the full text doesn't exist, throws a KeyError.
Under the hood, this uses getdeltachain() across all the stores to build
up a full chain to produce the full text.
"""
chain = self.getdeltachain(name, node)
if chain[-1][ChainIndicies.BASENODE] != nullid:
# If we didn't receive a full chain, throw
raise KeyError((name, hex(node)))
# The last entry in the chain is a full text, so we start our delta
# applies with that.
fulltext = chain.pop()[ChainIndicies.DATA]
text = fulltext
while chain:
delta = chain.pop()[ChainIndicies.DATA]
text = mdiff.patches(text, [delta])
return text
@basestore.baseunionstore.retriable
def getdelta(self, name, node):
"""Return the single delta entry for the given name/node pair."""
for store in self.stores:
try:
return store.getdelta(name, node)
except KeyError:
pass
raise KeyError((name, hex(node)))
def getdeltachain(self, name, node):
"""Returns the deltachain for the given name/node pair.
Returns an ordered list of:
[(name, node, deltabasename, deltabasenode, deltacontent),...]
where the chain is terminated by a full text entry with a nullid
deltabasenode.
"""
chain = self._getpartialchain(name, node)
while chain[-1][ChainIndicies.BASENODE] != nullid:
x, x, deltabasename, deltabasenode, x = chain[-1]
try:
morechain = self._getpartialchain(deltabasename, deltabasenode)
chain.extend(morechain)
except KeyError:
# If we allow incomplete chains, don't throw.
if not self.allowincomplete:
raise
break
return chain
@basestore.baseunionstore.retriable
def getmeta(self, name, node):
"""Returns the metadata dict for given node."""
for store in self.stores:
try:
return store.getmeta(name, node)
except KeyError:
pass
raise KeyError((name, hex(node)))
def getmetrics(self):
metrics = [s.getmetrics() for s in self.stores]
return shallowutil.sumdicts(*metrics)
@basestore.baseunionstore.retriable
def _getpartialchain(self, name, node):
"""Returns a partial delta chain for the given name/node pair.
A partial chain is a chain that may not be terminated in a full-text.
"""
for store in self.stores:
try:
return store.getdeltachain(name, node)
except KeyError:
pass
raise KeyError((name, hex(node)))
def add(self, name, node, data):
raise RuntimeError(
b"cannot add content only to remotefilelog contentstore"
)
def getmissing(self, keys):
missing = keys
for store in self.stores:
if missing:
missing = store.getmissing(missing)
return missing
def addremotefilelognode(self, name, node, data):
if self.writestore:
self.writestore.addremotefilelognode(name, node, data)
else:
raise RuntimeError(b"no writable store configured")
def markledger(self, ledger, options=None):
for store in self.stores:
store.markledger(ledger, options)
class remotefilelogcontentstore(basestore.basestore):
def __init__(self, *args, **kwargs):
super(remotefilelogcontentstore, self).__init__(*args, **kwargs)
self._threaddata = threading.local()
def get(self, name, node):
# return raw revision text
data = self._getdata(name, node)
offset, size, flags = shallowutil.parsesizeflags(data)
content = data[offset : offset + size]
ancestormap = shallowutil.ancestormap(data)
p1, p2, linknode, copyfrom = ancestormap[node]
copyrev = None
if copyfrom:
copyrev = hex(p1)
self._updatemetacache(node, size, flags)
# lfs tracks renames in its own metadata, remove hg copy metadata,
# because copy metadata will be re-added by lfs flag processor.
if flags & revlog.REVIDX_EXTSTORED:
copyrev = copyfrom = None
revision = shallowutil.createrevlogtext(content, copyfrom, copyrev)
return revision
def getdelta(self, name, node):
# Since remotefilelog content stores only contain full texts, just
# return that.
revision = self.get(name, node)
return revision, name, nullid, self.getmeta(name, node)
def getdeltachain(self, name, node):
# Since remotefilelog content stores just contain full texts, we return
# a fake delta chain that just consists of a single full text revision.
# The nullid in the deltabasenode slot indicates that the revision is a
# fulltext.
revision = self.get(name, node)
return [(name, node, None, nullid, revision)]
def getmeta(self, name, node):
self._sanitizemetacache()
if node != self._threaddata.metacache[0]:
data = self._getdata(name, node)
offset, size, flags = shallowutil.parsesizeflags(data)
self._updatemetacache(node, size, flags)
return self._threaddata.metacache[1]
def add(self, name, node, data):
raise RuntimeError(
b"cannot add content only to remotefilelog contentstore"
)
def _sanitizemetacache(self):
metacache = getattr(self._threaddata, 'metacache', None)
if metacache is None:
self._threaddata.metacache = (None, None) # (node, meta)
def _updatemetacache(self, node, size, flags):
self._sanitizemetacache()
if node == self._threaddata.metacache[0]:
return
meta = {constants.METAKEYFLAG: flags, constants.METAKEYSIZE: size}
self._threaddata.metacache = (node, meta)
class remotecontentstore(object):
def __init__(self, ui, fileservice, shared):
self._fileservice = fileservice
# type(shared) is usually remotefilelogcontentstore
self._shared = shared
def get(self, name, node):
self._fileservice.prefetch(
[(name, hex(node))], force=True, fetchdata=True
)
return self._shared.get(name, node)
def getdelta(self, name, node):
revision = self.get(name, node)
return revision, name, nullid, self._shared.getmeta(name, node)
def getdeltachain(self, name, node):
# Since our remote content stores just contain full texts, we return a
# fake delta chain that just consists of a single full text revision.
# The nullid in the deltabasenode slot indicates that the revision is a
# fulltext.
revision = self.get(name, node)
return [(name, node, None, nullid, revision)]
def getmeta(self, name, node):
self._fileservice.prefetch(
[(name, hex(node))], force=True, fetchdata=True
)
return self._shared.getmeta(name, node)
def add(self, name, node, data):
raise RuntimeError(b"cannot add to a remote store")
def getmissing(self, keys):
return keys
def markledger(self, ledger, options=None):
pass
class manifestrevlogstore(object):
def __init__(self, repo):
self._store = repo.store
self._svfs = repo.svfs
self._revlogs = dict()
self._cl = revlog.revlog(self._svfs, b'00changelog.i')
self._repackstartlinkrev = 0
def get(self, name, node):
return self._revlog(name).rawdata(node)
def getdelta(self, name, node):
revision = self.get(name, node)
return revision, name, nullid, self.getmeta(name, node)
def getdeltachain(self, name, node):
revision = self.get(name, node)
return [(name, node, None, nullid, revision)]
def getmeta(self, name, node):
rl = self._revlog(name)
rev = rl.rev(node)
return {
constants.METAKEYFLAG: rl.flags(rev),
constants.METAKEYSIZE: rl.rawsize(rev),
}
def getancestors(self, name, node, known=None):
if known is None:
known = set()
if node in known:
return []
rl = self._revlog(name)
ancestors = {}
missing = {node}
for ancrev in rl.ancestors([rl.rev(node)], inclusive=True):
ancnode = rl.node(ancrev)
missing.discard(ancnode)
p1, p2 = rl.parents(ancnode)
if p1 != nullid and p1 not in known:
missing.add(p1)
if p2 != nullid and p2 not in known:
missing.add(p2)
linknode = self._cl.node(rl.linkrev(ancrev))
ancestors[rl.node(ancrev)] = (p1, p2, linknode, b'')
if not missing:
break
return ancestors
def getnodeinfo(self, name, node):
cl = self._cl
rl = self._revlog(name)
parents = rl.parents(node)
linkrev = rl.linkrev(rl.rev(node))
return (parents[0], parents[1], cl.node(linkrev), None)
def add(self, *args):
raise RuntimeError(b"cannot add to a revlog store")
def _revlog(self, name):
rl = self._revlogs.get(name)
if rl is None:
revlogname = b'00manifesttree.i'
if name != b'':
revlogname = b'meta/%s/00manifest.i' % name
rl = revlog.revlog(self._svfs, revlogname)
self._revlogs[name] = rl
return rl
def getmissing(self, keys):
missing = []
for name, node in keys:
mfrevlog = self._revlog(name)
if node not in mfrevlog.nodemap:
missing.append((name, node))
return missing
def setrepacklinkrevrange(self, startrev, endrev):
self._repackstartlinkrev = startrev
self._repackendlinkrev = endrev
def markledger(self, ledger, options=None):
if options and options.get(constants.OPTION_PACKSONLY):
return
treename = b''
rl = revlog.revlog(self._svfs, b'00manifesttree.i')
startlinkrev = self._repackstartlinkrev
endlinkrev = self._repackendlinkrev
for rev in pycompat.xrange(len(rl) - 1, -1, -1):
linkrev = rl.linkrev(rev)
if linkrev < startlinkrev:
break
if linkrev > endlinkrev:
continue
node = rl.node(rev)
ledger.markdataentry(self, treename, node)
ledger.markhistoryentry(self, treename, node)
for path, encoded, size in self._store.datafiles():
if path[:5] != b'meta/' or path[-2:] != b'.i':
continue
treename = path[5 : -len(b'/00manifest.i')]
rl = revlog.revlog(self._svfs, path)
for rev in pycompat.xrange(len(rl) - 1, -1, -1):
linkrev = rl.linkrev(rev)
if linkrev < startlinkrev:
break
if linkrev > endlinkrev:
continue
node = rl.node(rev)
ledger.markdataentry(self, treename, node)
ledger.markhistoryentry(self, treename, node)
def cleanup(self, ledger):
pass
| 8,139 | 3,046 | 872 |
4ac075e9bb8b333093734151b20982bc493d6bb5 | 1,121 | py | Python | stock/adapters/topics/send_handler.py | rh-tech-supermarket-restocking-demo/stock-backend | 0d868b270c4142fd5efdad28966482c280478c85 | [
"Apache-2.0"
] | null | null | null | stock/adapters/topics/send_handler.py | rh-tech-supermarket-restocking-demo/stock-backend | 0d868b270c4142fd5efdad28966482c280478c85 | [
"Apache-2.0"
] | null | null | null | stock/adapters/topics/send_handler.py | rh-tech-supermarket-restocking-demo/stock-backend | 0d868b270c4142fd5efdad28966482c280478c85 | [
"Apache-2.0"
] | null | null | null | import logging
from proton.handlers import MessagingHandler
from proton import Message
| 35.03125 | 96 | 0.664585 | import logging
from proton.handlers import MessagingHandler
from proton import Message
class SendHandler(MessagingHandler):
def __init__(self, host: str, address: str, message_body: str):
super(SendHandler, self).__init__()
self.conn_url: str = host
self.address: str = address
self.message_body: str = message_body
def on_start(self, event):
conn = event.container.connect(self.conn_url)
event.container.create_sender(conn, self.address)
def on_link_opened(self, event):
logging.debug("SendHandler.on_link_opened:Opened sender for target address '{0}'".format
(event.sender.target.address))
def on_sendable(self, event):
message = Message(self.message_body)
event.sender.send(message)
logging.debug(
"SendHandler.on_sendable:Sent message '{0}'".format(message.body))
event.sender.close()
logging.debug(
"SendHandler.on_sendable:Closed sender")
event.connection.close()
logging.debug(
"SendHandler.on_sendable:Closed connection")
| 888 | 15 | 130 |
90b0bf7b1dba201331c6a48219374b9f0a593643 | 2,098 | py | Python | backtracking/n-queens/Python/NQueen.py | CarbonDDR/al-go-rithms | 8e65affbe812931b7dde0e2933eb06c0f44b4130 | [
"CC0-1.0"
] | 1,253 | 2017-06-06T07:19:25.000Z | 2022-03-30T17:07:58.000Z | backtracking/n-queens/Python/NQueen.py | rishabh99-rc/al-go-rithms | 4df20d7ef7598fda4bc89101f9a99aac94cdd794 | [
"CC0-1.0"
] | 554 | 2017-09-29T18:56:01.000Z | 2022-02-21T15:48:13.000Z | backtracking/n-queens/Python/NQueen.py | rishabh99-rc/al-go-rithms | 4df20d7ef7598fda4bc89101f9a99aac94cdd794 | [
"CC0-1.0"
] | 2,226 | 2017-09-29T19:59:59.000Z | 2022-03-25T08:59:55.000Z | from __future__ import print_function
# Python program to solve N Queen
# Problem using backtracking
global N
N = 4
# A utility function to check if a queen can
# be placed on board[row][col]. Note that this
# function is called when "col" queens are
# already placed in columns from 0 to col -1.
# So we need to check only left side for
# attacking queens
# This function solves the N Queen problem using
# Backtracking. It mainly uses solveNQUtil() to
# solve the problem. It returns false if queens
# cannot be placed, otherwise return true and
# placement of queens in the form of 1s.
# note that there may be more than one
# solutions, this function prints one of the
# feasible solutions.
# driver program to test above function
solveNQ()
| 22.804348 | 52 | 0.672069 | from __future__ import print_function
# Python program to solve N Queen
# Problem using backtracking
global N
N = 4
def printSolution(board):
for i in range(N):
for j in range(N):
print(board[i][j], end=' ')
print()
# A utility function to check if a queen can
# be placed on board[row][col]. Note that this
# function is called when "col" queens are
# already placed in columns from 0 to col -1.
# So we need to check only left side for
# attacking queens
def isSafe(board, row, col):
# Check this row on left side
for i in range(col):
if board[row][i] == 1:
return False
# Check upper diagonal on left side
for i,j in zip(range(row,-1,-1), range(col,-1,-1)):
if board[i][j] == 1:
return False
# Check lower diagonal on left side
for i,j in zip(range(row,N,1), range(col,-1,-1)):
if board[i][j] == 1:
return False
return True
def solveNQUtil(board, col):
# base case: If all queens are placed
# then return true
if col >= N:
return True
# Consider this column and try placing
# this queen in all rows one by one
for i in range(N):
if isSafe(board, i, col):
# Place this queen in board[i][col]
board[i][col] = 1
# recur to place rest of the queens
if solveNQUtil(board, col+1) == True:
return True
# If placing queen in board[i][col
# doesn't lead to a solution, then
# queen from board[i][col]
board[i][col] = 0
# if queen can not be place in any row in
# this colum col then return false
return False
# This function solves the N Queen problem using
# Backtracking. It mainly uses solveNQUtil() to
# solve the problem. It returns false if queens
# cannot be placed, otherwise return true and
# placement of queens in the form of 1s.
# note that there may be more than one
# solutions, this function prints one of the
# feasible solutions.
def solveNQ():
board = [ [0, 0, 0, 0],
[0, 0, 0, 0],
[0, 0, 0, 0],
[0, 0, 0, 0]
]
if solveNQUtil(board, 0) == False:
print("Solution does not exist")
return False
printSolution(board)
return True
# driver program to test above function
solveNQ()
| 1,256 | 0 | 90 |
1c8b0ae9b81fd2fe52f53218422c3f284a5c6b9e | 317 | py | Python | common/user.py | BaobabDuncan/us_api_server | 7b59feac6978bd6083b2d1aa7565f15a4dc89b2f | [
"Apache-2.0"
] | null | null | null | common/user.py | BaobabDuncan/us_api_server | 7b59feac6978bd6083b2d1aa7565f15a4dc89b2f | [
"Apache-2.0"
] | null | null | null | common/user.py | BaobabDuncan/us_api_server | 7b59feac6978bd6083b2d1aa7565f15a4dc89b2f | [
"Apache-2.0"
] | null | null | null |
from django.http import HttpResponseRedirect
from google.appengine.api import memcache
| 24.384615 | 49 | 0.640379 |
from django.http import HttpResponseRedirect
from google.appengine.api import memcache
class adminUser():
def __init__(self):
pass
def getAdminUser(self):
return memcache.get("adminUser")
def setAdminUser(self):
memcache.add("adminUser", "admin", 80000) | 109 | -3 | 121 |
cf6dae97fca58029e3d487cb8e373c64b3a1b651 | 232 | py | Python | Python Files/Helper Files/Spectogram_crop_test.py | brennanMosher/Music-Genre-Recognition-using-a-Machine-Learning-Appraoch | 7834fe5d709e894322ad76ef118067febaa78bce | [
"MIT"
] | 1 | 2021-04-13T16:22:27.000Z | 2021-04-13T16:22:27.000Z | Python Files/Helper Files/Spectogram_crop_test.py | brennanMosher/Music-Genre-Recognition-using-a-Machine-Learning-Appraoch | 7834fe5d709e894322ad76ef118067febaa78bce | [
"MIT"
] | null | null | null | Python Files/Helper Files/Spectogram_crop_test.py | brennanMosher/Music-Genre-Recognition-using-a-Machine-Learning-Appraoch | 7834fe5d709e894322ad76ef118067febaa78bce | [
"MIT"
] | null | null | null | import os
from PIL import Image
'''
Test file for cropping images
'''
file_loc = r'C:\Users\brenn\Desktop\Spectrogram/blues.00000.jpg'
im = Image.open(file_loc)
crop_box = (80, 58, 576, 427)
img1 = im.crop(crop_box)
img1.show() | 15.466667 | 64 | 0.711207 | import os
from PIL import Image
'''
Test file for cropping images
'''
file_loc = r'C:\Users\brenn\Desktop\Spectrogram/blues.00000.jpg'
im = Image.open(file_loc)
crop_box = (80, 58, 576, 427)
img1 = im.crop(crop_box)
img1.show() | 0 | 0 | 0 |
270f33d777b5a636518e2bff52d771a58ec2b9c2 | 3,976 | py | Python | algorithms/tsec/layers.py | mpoiitis/GraphM-framework | 82904b9940b029579b86c359e17312cfa7ef1692 | [
"MIT"
] | null | null | null | algorithms/tsec/layers.py | mpoiitis/GraphM-framework | 82904b9940b029579b86c359e17312cfa7ef1692 | [
"MIT"
] | null | null | null | algorithms/tsec/layers.py | mpoiitis/GraphM-framework | 82904b9940b029579b86c359e17312cfa7ef1692 | [
"MIT"
] | null | null | null | from tensorflow.keras.layers import Layer, Dense
from tensorflow.keras.regularizers import l2
import tensorflow as tf
class Sampling(Layer):
"""Uses (z_mean, z_log_var) to sample z, the vector encoding a digit."""
class Encoder(Layer):
"""Maps MNIST digits to a triplet (z_mean, z_log_var, z)."""
class Decoder(Layer):
"""Converts z, the encoded digit vector, back into a readable digit."""
| 37.509434 | 129 | 0.630533 | from tensorflow.keras.layers import Layer, Dense
from tensorflow.keras.regularizers import l2
import tensorflow as tf
class Sampling(Layer):
"""Uses (z_mean, z_log_var) to sample z, the vector encoding a digit."""
def call(self, inputs, **kwargs):
z_mean, z_log_var = inputs
batch = tf.shape(z_mean)[0]
dim = tf.shape(z_mean)[1]
epsilon = tf.keras.backend.random_normal(shape=(batch, dim))
return z_mean + tf.exp(0.5 * z_log_var) * epsilon
class Encoder(Layer):
"""Maps MNIST digits to a triplet (z_mean, z_log_var, z)."""
def __init__(self, latent_dim=32, intermediate_dim=64, name="encoder", **kwargs):
super(Encoder, self).__init__(name=name, **kwargs)
self.dense_proj = Dense(intermediate_dim, activation="relu")
self.dense_mean = Dense(latent_dim)
self.dense_log_var = Dense(latent_dim)
self.sampling = Sampling()
def call(self, inputs):
x = self.dense_proj(inputs)
z_mean = self.dense_mean(x)
z_log_var = self.dense_log_var(x)
z = self.sampling((z_mean, z_log_var))
return z_mean, z_log_var, z
class Decoder(Layer):
"""Converts z, the encoded digit vector, back into a readable digit."""
def __init__(self, original_dim, intermediate_dim=64, name="decoder", **kwargs):
super(Decoder, self).__init__(name=name, **kwargs)
self.dense_proj = Dense(intermediate_dim, activation="relu")
self.dense_output = Dense(original_dim, activation="sigmoid")
def call(self, inputs):
x = self.dense_proj(inputs)
return self.dense_output(x)
class GraphConvolution(Layer):
def __init__(self, adjacency, units, activation=tf.identity, dropout=0.0, l2=0.0, dtype=tf.float32, name='GraphConvolution'):
"""
Params:
Adjacency: a tf.SparseTensor adjacency matrix
Units: The number of output units per node state
Activation: The activation function to apply to the node states
Dropout: The amount of dropout (0.0 being none, 1.0 being all units) to apply
l2: The amount of L2 regularisation to apply
dtype: The type of values in the tensors this layer will transform
name: The name of this layer
"""
super(GraphConvolution, self).__init__(dtype=dtype, name=name)
self.adjacency = adjacency
self.units = units
self.activation = activation
self.dropout = dropout
self.l2 = l2
def build(self, input_shape):
"""
This method is called during the initial compilation of our model. Its
primary job is to initialize the weights for this layer.
Params:
input_shape: this is the shape of the input to the layer, in our case an
array of (NUMBER_NODES, INPUT_SIZE)
Build one weight, w, which will be applied to each input. Initialize
it from the uniform distribution, scaled by the size of the matrix. Apply
l2 loss to regularize the matrix
"""
self.w = self.add_weight(shape=(input_shape[1], self.units), dtype=self.dtype,
initializer='glorot_uniform', regularizer=l2(self.l2), name='weight')
def call(self, inputs):
"""
This method is called to apply the layer to an incoming tensor. This is the
real meat of the model.
Params:
node_state: The tf.Tensor of node states. Shape (NUMBER_NODES, NODE_STATE_SIZE)
Returns: The transformed node state tf.Tensor
"""
# Dropout
inputs = tf.nn.dropout(inputs, rate=self.dropout)
# Convolution
inputs = tf.matmul(inputs, self.w)
# Propagation
inputs = tf.sparse.sparse_dense_matmul(self.adjacency, inputs)
output = self.activation(inputs)
return output
| 1,088 | 2,320 | 158 |
01221d594e7e8a20c9a9d17c46c4a2281a62c2ae | 982 | py | Python | problems/15.3-sum.py | lucasheriques/leetcode-solutions | cd36b5df46a75a0cb17569faf9cf56186864f68a | [
"MIT"
] | null | null | null | problems/15.3-sum.py | lucasheriques/leetcode-solutions | cd36b5df46a75a0cb17569faf9cf56186864f68a | [
"MIT"
] | null | null | null | problems/15.3-sum.py | lucasheriques/leetcode-solutions | cd36b5df46a75a0cb17569faf9cf56186864f68a | [
"MIT"
] | null | null | null | #
# @lc app=leetcode id=15 lang=python3
#
# [15] 3Sum
#
# O(n^2) time | O(n) space
| 27.277778 | 76 | 0.407332 | #
# @lc app=leetcode id=15 lang=python3
#
# [15] 3Sum
#
# O(n^2) time | O(n) space
class Solution:
def threeSum(self, array, k=0):
array.sort()
ans = []
for i in range(len(array) - 1):
left = i + 1
right = len(array) - 1
if i > 0 and array[i] == array[i-1]:
continue
while left < right:
potentialSum = array[i] + array[left] + array[right]
if potentialSum == k:
ans.append([array[i], array[left], array[right]])
while left < right and array[left] == array[left+1]:
left += 1
while right > left and array[right] == array[right - 1]:
right -= 1
left += 1
right -= 1
elif potentialSum < k:
left += 1
else:
right -= 1
return ans
| 854 | -6 | 49 |
11705964d46a522a8400a13b9c02c8fb742ed68b | 434 | py | Python | 2018/day1/first_repeat.py | forrestrice/advent | 2451de84790e67c08676154bfa5167ee353d4e42 | [
"MIT"
] | 1 | 2019-12-03T07:41:10.000Z | 2019-12-03T07:41:10.000Z | 2018/day1/first_repeat.py | forrestrice/advent | 2451de84790e67c08676154bfa5167ee353d4e42 | [
"MIT"
] | null | null | null | 2018/day1/first_repeat.py | forrestrice/advent | 2451de84790e67c08676154bfa5167ee353d4e42 | [
"MIT"
] | null | null | null | input_file_name = "time_machine_input.txt"
sum = 0
seen_sums = {}
for time_input in infinite_input():
sum += int(time_input)
sum_frequency = seen_sums.get(sum, 0) + 1
seen_sums[sum] = sum_frequency
if sum_frequency > 1:
print(sum)
break | 24.111111 | 45 | 0.638249 | input_file_name = "time_machine_input.txt"
def infinite_input():
input_file = open(input_file_name, "r")
while True:
for line in input_file:
yield line
input_file.seek(0)
sum = 0
seen_sums = {}
for time_input in infinite_input():
sum += int(time_input)
sum_frequency = seen_sums.get(sum, 0) + 1
seen_sums[sum] = sum_frequency
if sum_frequency > 1:
print(sum)
break | 142 | 0 | 23 |
2fda7d56160fe8a774a3ce7da64ac6810295ef73 | 2,554 | py | Python | refbee/platforms/wikidata.py | kmdn/RefBee | c45e5fa7a3f011b760284a7ee8dafec8e0884280 | [
"MIT"
] | null | null | null | refbee/platforms/wikidata.py | kmdn/RefBee | c45e5fa7a3f011b760284a7ee8dafec8e0884280 | [
"MIT"
] | null | null | null | refbee/platforms/wikidata.py | kmdn/RefBee | c45e5fa7a3f011b760284a7ee8dafec8e0884280 | [
"MIT"
] | null | null | null | from refbee.util import get_sparql_query_results
platform_properties_dict = {
"ORCID": "wdt:P496",
"Google Scholar": "wdt:P1960",
"VIAF": "wdt:P214",
"DBLP": "wdt:P2456",
"Dimensions": "wdt:P6178",
"Github": "wdt:P2037",
"Microsoft Academic": "wdt:P6366",
"Semantic Scholar": "wdt:P4012",
"DNB/GNB": "wdt:P227",
#"ACM Digital Library": "wdt:P864"
}
def paper_titles_for_id(person_id):
""" Get all entities the person is the author of.
Could additionally be filtered by entity type:
https://www.wikidata.org/wiki/Q23927052 conference paper
https://www.wikidata.org/wiki/Q13442814 scholarly article
https://www.wikidata.org/wiki/Q18918145 academic journal article
https://www.wikidata.org/wiki/Q591041 scientific publication
https://www.wikidata.org/wiki/Q55915575 scholarly work
"""
endpoint_url = "https://query.wikidata.org/sparql"
query = ('SELECT ?pub ?title WHERE { ?pub wdt:P50 wd:' + person_id + ' .'
'?pub wdt:P1476 ?title . }')
results = get_sparql_query_results(endpoint_url, query)
paper_titles = [r['title']['value'] for r in results['results']['bindings']]
return paper_titles | 38.69697 | 103 | 0.635082 | from refbee.util import get_sparql_query_results
platform_properties_dict = {
"ORCID": "wdt:P496",
"Google Scholar": "wdt:P1960",
"VIAF": "wdt:P214",
"DBLP": "wdt:P2456",
"Dimensions": "wdt:P6178",
"Github": "wdt:P2037",
"Microsoft Academic": "wdt:P6366",
"Semantic Scholar": "wdt:P4012",
"DNB/GNB": "wdt:P227",
#"ACM Digital Library": "wdt:P864"
}
def query_id_from_wikidata(person_id="Q57231890", platform_predicate="wdt:P496"):
endpoint_url = "https://query.wikidata.org/sparql"
query = """SELECT DISTINCT ?o WHERE {
wd:""" + person_id + """ """ + platform_predicate + """ ?o .
wd:""" + person_id + """ wdt:P31 wd:Q5.
}
LIMIT 100"""
results = get_sparql_query_results(endpoint_url, query)
ids_set = set()
for result in results["results"]["bindings"]:
ids_set.add(result['o']['value'])
return ids_set
def paper_titles_for_id(person_id):
""" Get all entities the person is the author of.
Could additionally be filtered by entity type:
https://www.wikidata.org/wiki/Q23927052 conference paper
https://www.wikidata.org/wiki/Q13442814 scholarly article
https://www.wikidata.org/wiki/Q18918145 academic journal article
https://www.wikidata.org/wiki/Q591041 scientific publication
https://www.wikidata.org/wiki/Q55915575 scholarly work
"""
endpoint_url = "https://query.wikidata.org/sparql"
query = ('SELECT ?pub ?title WHERE { ?pub wdt:P50 wd:' + person_id + ' .'
'?pub wdt:P1476 ?title . }')
results = get_sparql_query_results(endpoint_url, query)
paper_titles = [r['title']['value'] for r in results['results']['bindings']]
return paper_titles
def platform_ids_for_person(person_id):
persons_dict = {
person_id: {"Wikidata": set([person_id])}
}
for platform in platform_properties_dict.keys():
platform_property = platform_properties_dict[platform]
platform_id = query_id_from_wikidata(person_id=person_id, platform_predicate=platform_property)
if len(platform_id) == 0:
continue
print("ID (%s): %s" % (platform, platform_id))
person_dict = persons_dict.get(person_id, {})
# only relevant if it's defaulting to {}
persons_dict[person_id] = person_dict
# add values for the specific platform
person_dict[platform] = platform_id
print(persons_dict)
return persons_dict | 1,201 | 0 | 46 |