id int32 0 252k | repo stringlengths 7 55 | path stringlengths 4 127 | func_name stringlengths 1 88 | original_string stringlengths 75 19.8k | language stringclasses 1 value | code stringlengths 75 19.8k | code_tokens list | docstring stringlengths 3 17.3k | docstring_tokens list | sha stringlengths 40 40 | url stringlengths 87 242 |
|---|---|---|---|---|---|---|---|---|---|---|---|
230,600 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_name_history | def get_name_history( self, name, offset=None, count=None, reverse=False):
"""
Get the historic states for a name, grouped by block height.
"""
cur = self.db.cursor()
name_hist = namedb_get_history( cur, name, offset=offset, count=count, reverse=reverse )
return name_hist | python | def get_name_history( self, name, offset=None, count=None, reverse=False):
"""
Get the historic states for a name, grouped by block height.
"""
cur = self.db.cursor()
name_hist = namedb_get_history( cur, name, offset=offset, count=count, reverse=reverse )
return name_hist | [
"def",
"get_name_history",
"(",
"self",
",",
"name",
",",
"offset",
"=",
"None",
",",
"count",
"=",
"None",
",",
"reverse",
"=",
"False",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"name_hist",
"=",
"namedb_get_history",
"(",
"cur",
",",
"name",
",",
"offset",
"=",
"offset",
",",
"count",
"=",
"count",
",",
"reverse",
"=",
"reverse",
")",
"return",
"name_hist"
] | Get the historic states for a name, grouped by block height. | [
"Get",
"the",
"historic",
"states",
"for",
"a",
"name",
"grouped",
"by",
"block",
"height",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L941-L947 |
230,601 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.is_name_zonefile_hash | def is_name_zonefile_hash(self, name, zonefile_hash):
"""
Was a zone file sent by a name?
"""
cur = self.db.cursor()
return namedb_is_name_zonefile_hash(cur, name, zonefile_hash) | python | def is_name_zonefile_hash(self, name, zonefile_hash):
"""
Was a zone file sent by a name?
"""
cur = self.db.cursor()
return namedb_is_name_zonefile_hash(cur, name, zonefile_hash) | [
"def",
"is_name_zonefile_hash",
"(",
"self",
",",
"name",
",",
"zonefile_hash",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"return",
"namedb_is_name_zonefile_hash",
"(",
"cur",
",",
"name",
",",
"zonefile_hash",
")"
] | Was a zone file sent by a name? | [
"Was",
"a",
"zone",
"file",
"sent",
"by",
"a",
"name?"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L950-L955 |
230,602 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_all_blockstack_ops_at | def get_all_blockstack_ops_at( self, block_number, offset=None, count=None, include_history=None, restore_history=None ):
"""
Get all name, namespace, and account records affected at a particular block,
in the state they were at the given block number.
Paginate if offset, count are given.
"""
if include_history is not None:
log.warn("DEPRECATED use of include_history")
if restore_history is not None:
log.warn("DEPRECATED use of restore_history")
log.debug("Get all accepted operations at %s in %s" % (block_number, self.db_filename))
recs = namedb_get_all_blockstack_ops_at( self.db, block_number, offset=offset, count=count )
# include opcode
for rec in recs:
assert 'op' in rec
rec['opcode'] = op_get_opcode_name(rec['op'])
return recs | python | def get_all_blockstack_ops_at( self, block_number, offset=None, count=None, include_history=None, restore_history=None ):
"""
Get all name, namespace, and account records affected at a particular block,
in the state they were at the given block number.
Paginate if offset, count are given.
"""
if include_history is not None:
log.warn("DEPRECATED use of include_history")
if restore_history is not None:
log.warn("DEPRECATED use of restore_history")
log.debug("Get all accepted operations at %s in %s" % (block_number, self.db_filename))
recs = namedb_get_all_blockstack_ops_at( self.db, block_number, offset=offset, count=count )
# include opcode
for rec in recs:
assert 'op' in rec
rec['opcode'] = op_get_opcode_name(rec['op'])
return recs | [
"def",
"get_all_blockstack_ops_at",
"(",
"self",
",",
"block_number",
",",
"offset",
"=",
"None",
",",
"count",
"=",
"None",
",",
"include_history",
"=",
"None",
",",
"restore_history",
"=",
"None",
")",
":",
"if",
"include_history",
"is",
"not",
"None",
":",
"log",
".",
"warn",
"(",
"\"DEPRECATED use of include_history\"",
")",
"if",
"restore_history",
"is",
"not",
"None",
":",
"log",
".",
"warn",
"(",
"\"DEPRECATED use of restore_history\"",
")",
"log",
".",
"debug",
"(",
"\"Get all accepted operations at %s in %s\"",
"%",
"(",
"block_number",
",",
"self",
".",
"db_filename",
")",
")",
"recs",
"=",
"namedb_get_all_blockstack_ops_at",
"(",
"self",
".",
"db",
",",
"block_number",
",",
"offset",
"=",
"offset",
",",
"count",
"=",
"count",
")",
"# include opcode ",
"for",
"rec",
"in",
"recs",
":",
"assert",
"'op'",
"in",
"rec",
"rec",
"[",
"'opcode'",
"]",
"=",
"op_get_opcode_name",
"(",
"rec",
"[",
"'op'",
"]",
")",
"return",
"recs"
] | Get all name, namespace, and account records affected at a particular block,
in the state they were at the given block number.
Paginate if offset, count are given. | [
"Get",
"all",
"name",
"namespace",
"and",
"account",
"records",
"affected",
"at",
"a",
"particular",
"block",
"in",
"the",
"state",
"they",
"were",
"at",
"the",
"given",
"block",
"number",
".",
"Paginate",
"if",
"offset",
"count",
"are",
"given",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L958-L979 |
230,603 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_name_from_name_hash128 | def get_name_from_name_hash128( self, name ):
"""
Get the name from a name hash
"""
cur = self.db.cursor()
name = namedb_get_name_from_name_hash128( cur, name, self.lastblock )
return name | python | def get_name_from_name_hash128( self, name ):
"""
Get the name from a name hash
"""
cur = self.db.cursor()
name = namedb_get_name_from_name_hash128( cur, name, self.lastblock )
return name | [
"def",
"get_name_from_name_hash128",
"(",
"self",
",",
"name",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"name",
"=",
"namedb_get_name_from_name_hash128",
"(",
"cur",
",",
"name",
",",
"self",
".",
"lastblock",
")",
"return",
"name"
] | Get the name from a name hash | [
"Get",
"the",
"name",
"from",
"a",
"name",
"hash"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L990-L996 |
230,604 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_num_historic_names_by_address | def get_num_historic_names_by_address( self, address ):
"""
Get the number of names historically owned by an address
"""
cur = self.db.cursor()
count = namedb_get_num_historic_names_by_address( cur, address )
return count | python | def get_num_historic_names_by_address( self, address ):
"""
Get the number of names historically owned by an address
"""
cur = self.db.cursor()
count = namedb_get_num_historic_names_by_address( cur, address )
return count | [
"def",
"get_num_historic_names_by_address",
"(",
"self",
",",
"address",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"count",
"=",
"namedb_get_num_historic_names_by_address",
"(",
"cur",
",",
"address",
")",
"return",
"count"
] | Get the number of names historically owned by an address | [
"Get",
"the",
"number",
"of",
"names",
"historically",
"owned",
"by",
"an",
"address"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1019-L1025 |
230,605 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_names_owned_by_sender | def get_names_owned_by_sender( self, sender_pubkey, lastblock=None ):
"""
Get the set of names owned by a particular script-pubkey.
"""
cur = self.db.cursor()
if lastblock is None:
lastblock = self.lastblock
names = namedb_get_names_by_sender( cur, sender_pubkey, lastblock )
return names | python | def get_names_owned_by_sender( self, sender_pubkey, lastblock=None ):
"""
Get the set of names owned by a particular script-pubkey.
"""
cur = self.db.cursor()
if lastblock is None:
lastblock = self.lastblock
names = namedb_get_names_by_sender( cur, sender_pubkey, lastblock )
return names | [
"def",
"get_names_owned_by_sender",
"(",
"self",
",",
"sender_pubkey",
",",
"lastblock",
"=",
"None",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"if",
"lastblock",
"is",
"None",
":",
"lastblock",
"=",
"self",
".",
"lastblock",
"names",
"=",
"namedb_get_names_by_sender",
"(",
"cur",
",",
"sender_pubkey",
",",
"lastblock",
")",
"return",
"names"
] | Get the set of names owned by a particular script-pubkey. | [
"Get",
"the",
"set",
"of",
"names",
"owned",
"by",
"a",
"particular",
"script",
"-",
"pubkey",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1028-L1037 |
230,606 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_num_names | def get_num_names( self, include_expired=False ):
"""
Get the number of names that exist.
"""
cur = self.db.cursor()
return namedb_get_num_names( cur, self.lastblock, include_expired=include_expired ) | python | def get_num_names( self, include_expired=False ):
"""
Get the number of names that exist.
"""
cur = self.db.cursor()
return namedb_get_num_names( cur, self.lastblock, include_expired=include_expired ) | [
"def",
"get_num_names",
"(",
"self",
",",
"include_expired",
"=",
"False",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"return",
"namedb_get_num_names",
"(",
"cur",
",",
"self",
".",
"lastblock",
",",
"include_expired",
"=",
"include_expired",
")"
] | Get the number of names that exist. | [
"Get",
"the",
"number",
"of",
"names",
"that",
"exist",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1040-L1045 |
230,607 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_all_names | def get_all_names( self, offset=None, count=None, include_expired=False ):
"""
Get the set of all registered names, with optional pagination
Returns the list of names.
"""
if offset is not None and offset < 0:
offset = None
if count is not None and count < 0:
count = None
cur = self.db.cursor()
names = namedb_get_all_names( cur, self.lastblock, offset=offset, count=count, include_expired=include_expired )
return names | python | def get_all_names( self, offset=None, count=None, include_expired=False ):
"""
Get the set of all registered names, with optional pagination
Returns the list of names.
"""
if offset is not None and offset < 0:
offset = None
if count is not None and count < 0:
count = None
cur = self.db.cursor()
names = namedb_get_all_names( cur, self.lastblock, offset=offset, count=count, include_expired=include_expired )
return names | [
"def",
"get_all_names",
"(",
"self",
",",
"offset",
"=",
"None",
",",
"count",
"=",
"None",
",",
"include_expired",
"=",
"False",
")",
":",
"if",
"offset",
"is",
"not",
"None",
"and",
"offset",
"<",
"0",
":",
"offset",
"=",
"None",
"if",
"count",
"is",
"not",
"None",
"and",
"count",
"<",
"0",
":",
"count",
"=",
"None",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"names",
"=",
"namedb_get_all_names",
"(",
"cur",
",",
"self",
".",
"lastblock",
",",
"offset",
"=",
"offset",
",",
"count",
"=",
"count",
",",
"include_expired",
"=",
"include_expired",
")",
"return",
"names"
] | Get the set of all registered names, with optional pagination
Returns the list of names. | [
"Get",
"the",
"set",
"of",
"all",
"registered",
"names",
"with",
"optional",
"pagination",
"Returns",
"the",
"list",
"of",
"names",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1048-L1061 |
230,608 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_num_names_in_namespace | def get_num_names_in_namespace( self, namespace_id ):
"""
Get the number of names in a namespace
"""
cur = self.db.cursor()
return namedb_get_num_names_in_namespace( cur, namespace_id, self.lastblock ) | python | def get_num_names_in_namespace( self, namespace_id ):
"""
Get the number of names in a namespace
"""
cur = self.db.cursor()
return namedb_get_num_names_in_namespace( cur, namespace_id, self.lastblock ) | [
"def",
"get_num_names_in_namespace",
"(",
"self",
",",
"namespace_id",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"return",
"namedb_get_num_names_in_namespace",
"(",
"cur",
",",
"namespace_id",
",",
"self",
".",
"lastblock",
")"
] | Get the number of names in a namespace | [
"Get",
"the",
"number",
"of",
"names",
"in",
"a",
"namespace"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1064-L1069 |
230,609 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_names_in_namespace | def get_names_in_namespace( self, namespace_id, offset=None, count=None ):
"""
Get the set of all registered names in a particular namespace.
Returns the list of names.
"""
if offset is not None and offset < 0:
offset = None
if count is not None and count < 0:
count = None
cur = self.db.cursor()
names = namedb_get_names_in_namespace( cur, namespace_id, self.lastblock, offset=offset, count=count )
return names | python | def get_names_in_namespace( self, namespace_id, offset=None, count=None ):
"""
Get the set of all registered names in a particular namespace.
Returns the list of names.
"""
if offset is not None and offset < 0:
offset = None
if count is not None and count < 0:
count = None
cur = self.db.cursor()
names = namedb_get_names_in_namespace( cur, namespace_id, self.lastblock, offset=offset, count=count )
return names | [
"def",
"get_names_in_namespace",
"(",
"self",
",",
"namespace_id",
",",
"offset",
"=",
"None",
",",
"count",
"=",
"None",
")",
":",
"if",
"offset",
"is",
"not",
"None",
"and",
"offset",
"<",
"0",
":",
"offset",
"=",
"None",
"if",
"count",
"is",
"not",
"None",
"and",
"count",
"<",
"0",
":",
"count",
"=",
"None",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"names",
"=",
"namedb_get_names_in_namespace",
"(",
"cur",
",",
"namespace_id",
",",
"self",
".",
"lastblock",
",",
"offset",
"=",
"offset",
",",
"count",
"=",
"count",
")",
"return",
"names"
] | Get the set of all registered names in a particular namespace.
Returns the list of names. | [
"Get",
"the",
"set",
"of",
"all",
"registered",
"names",
"in",
"a",
"particular",
"namespace",
".",
"Returns",
"the",
"list",
"of",
"names",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1072-L1085 |
230,610 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_all_namespace_ids | def get_all_namespace_ids( self ):
"""
Get the set of all existing, READY namespace IDs.
"""
cur = self.db.cursor()
namespace_ids = namedb_get_all_namespace_ids( cur )
return namespace_ids | python | def get_all_namespace_ids( self ):
"""
Get the set of all existing, READY namespace IDs.
"""
cur = self.db.cursor()
namespace_ids = namedb_get_all_namespace_ids( cur )
return namespace_ids | [
"def",
"get_all_namespace_ids",
"(",
"self",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"namespace_ids",
"=",
"namedb_get_all_namespace_ids",
"(",
"cur",
")",
"return",
"namespace_ids"
] | Get the set of all existing, READY namespace IDs. | [
"Get",
"the",
"set",
"of",
"all",
"existing",
"READY",
"namespace",
"IDs",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1088-L1094 |
230,611 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_all_revealed_namespace_ids | def get_all_revealed_namespace_ids( self ):
"""
Get all revealed namespace IDs that have not expired.
"""
cur = self.db.cursor()
namespace_ids = namedb_get_all_revealed_namespace_ids( cur, self.lastblock )
return namespace_ids | python | def get_all_revealed_namespace_ids( self ):
"""
Get all revealed namespace IDs that have not expired.
"""
cur = self.db.cursor()
namespace_ids = namedb_get_all_revealed_namespace_ids( cur, self.lastblock )
return namespace_ids | [
"def",
"get_all_revealed_namespace_ids",
"(",
"self",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"namespace_ids",
"=",
"namedb_get_all_revealed_namespace_ids",
"(",
"cur",
",",
"self",
".",
"lastblock",
")",
"return",
"namespace_ids"
] | Get all revealed namespace IDs that have not expired. | [
"Get",
"all",
"revealed",
"namespace",
"IDs",
"that",
"have",
"not",
"expired",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1097-L1103 |
230,612 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_all_preordered_namespace_hashes | def get_all_preordered_namespace_hashes( self ):
"""
Get all oustanding namespace preorder hashes that have not expired.
Used for testing
"""
cur = self.db.cursor()
namespace_hashes = namedb_get_all_preordered_namespace_hashes( cur, self.lastblock )
return namespace_hashes | python | def get_all_preordered_namespace_hashes( self ):
"""
Get all oustanding namespace preorder hashes that have not expired.
Used for testing
"""
cur = self.db.cursor()
namespace_hashes = namedb_get_all_preordered_namespace_hashes( cur, self.lastblock )
return namespace_hashes | [
"def",
"get_all_preordered_namespace_hashes",
"(",
"self",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"namespace_hashes",
"=",
"namedb_get_all_preordered_namespace_hashes",
"(",
"cur",
",",
"self",
".",
"lastblock",
")",
"return",
"namespace_hashes"
] | Get all oustanding namespace preorder hashes that have not expired.
Used for testing | [
"Get",
"all",
"oustanding",
"namespace",
"preorder",
"hashes",
"that",
"have",
"not",
"expired",
".",
"Used",
"for",
"testing"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1106-L1113 |
230,613 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_all_importing_namespace_hashes | def get_all_importing_namespace_hashes( self ):
"""
Get the set of all preordered and revealed namespace hashes that have not expired.
"""
cur = self.db.cursor()
namespace_hashes = namedb_get_all_importing_namespace_hashes( cur, self.lastblock )
return namespace_hashes | python | def get_all_importing_namespace_hashes( self ):
"""
Get the set of all preordered and revealed namespace hashes that have not expired.
"""
cur = self.db.cursor()
namespace_hashes = namedb_get_all_importing_namespace_hashes( cur, self.lastblock )
return namespace_hashes | [
"def",
"get_all_importing_namespace_hashes",
"(",
"self",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"namespace_hashes",
"=",
"namedb_get_all_importing_namespace_hashes",
"(",
"cur",
",",
"self",
".",
"lastblock",
")",
"return",
"namespace_hashes"
] | Get the set of all preordered and revealed namespace hashes that have not expired. | [
"Get",
"the",
"set",
"of",
"all",
"preordered",
"and",
"revealed",
"namespace",
"hashes",
"that",
"have",
"not",
"expired",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1116-L1122 |
230,614 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_name_preorder | def get_name_preorder( self, name, sender_script_pubkey, register_addr, include_failed=False ):
"""
Get the current preorder for a name, given the name, the sender's script pubkey, and
the registration address used to calculate the preorder hash.
Return the preorder record on success.
Return None if not found, or the preorder is already registered and not expired (even if revoked).
NOTE: possibly returns an expired preorder (by design, so as to prevent someone
from re-sending the same preorder with the same preorder hash).
"""
# name registered and not expired?
name_rec = self.get_name( name )
if name_rec is not None and not include_failed:
return None
# what namespace are we in?
namespace_id = get_namespace_from_name(name)
namespace = self.get_namespace(namespace_id)
if namespace is None:
return None
# isn't currently registered, or we don't care
preorder_hash = hash_name(name, sender_script_pubkey, register_addr=register_addr)
preorder = namedb_get_name_preorder( self.db, preorder_hash, self.lastblock )
if preorder is None:
# doesn't exist or expired
return None
# preorder must be younger than the namespace lifetime
# (otherwise we get into weird conditions where someone can preorder
# a name before someone else, and register it after it expires)
namespace_lifetime_multiplier = get_epoch_namespace_lifetime_multiplier( self.lastblock, namespace_id )
if preorder['block_number'] + (namespace['lifetime'] * namespace_lifetime_multiplier) <= self.lastblock:
log.debug("Preorder is too old (accepted at {}, namespace lifetime is {}, current block is {})".format(preorder['block_number'], namespace['lifetime'] * namespace_lifetime_multiplier, self.lastblock))
return None
return preorder | python | def get_name_preorder( self, name, sender_script_pubkey, register_addr, include_failed=False ):
"""
Get the current preorder for a name, given the name, the sender's script pubkey, and
the registration address used to calculate the preorder hash.
Return the preorder record on success.
Return None if not found, or the preorder is already registered and not expired (even if revoked).
NOTE: possibly returns an expired preorder (by design, so as to prevent someone
from re-sending the same preorder with the same preorder hash).
"""
# name registered and not expired?
name_rec = self.get_name( name )
if name_rec is not None and not include_failed:
return None
# what namespace are we in?
namespace_id = get_namespace_from_name(name)
namespace = self.get_namespace(namespace_id)
if namespace is None:
return None
# isn't currently registered, or we don't care
preorder_hash = hash_name(name, sender_script_pubkey, register_addr=register_addr)
preorder = namedb_get_name_preorder( self.db, preorder_hash, self.lastblock )
if preorder is None:
# doesn't exist or expired
return None
# preorder must be younger than the namespace lifetime
# (otherwise we get into weird conditions where someone can preorder
# a name before someone else, and register it after it expires)
namespace_lifetime_multiplier = get_epoch_namespace_lifetime_multiplier( self.lastblock, namespace_id )
if preorder['block_number'] + (namespace['lifetime'] * namespace_lifetime_multiplier) <= self.lastblock:
log.debug("Preorder is too old (accepted at {}, namespace lifetime is {}, current block is {})".format(preorder['block_number'], namespace['lifetime'] * namespace_lifetime_multiplier, self.lastblock))
return None
return preorder | [
"def",
"get_name_preorder",
"(",
"self",
",",
"name",
",",
"sender_script_pubkey",
",",
"register_addr",
",",
"include_failed",
"=",
"False",
")",
":",
"# name registered and not expired?",
"name_rec",
"=",
"self",
".",
"get_name",
"(",
"name",
")",
"if",
"name_rec",
"is",
"not",
"None",
"and",
"not",
"include_failed",
":",
"return",
"None",
"# what namespace are we in?",
"namespace_id",
"=",
"get_namespace_from_name",
"(",
"name",
")",
"namespace",
"=",
"self",
".",
"get_namespace",
"(",
"namespace_id",
")",
"if",
"namespace",
"is",
"None",
":",
"return",
"None",
"# isn't currently registered, or we don't care",
"preorder_hash",
"=",
"hash_name",
"(",
"name",
",",
"sender_script_pubkey",
",",
"register_addr",
"=",
"register_addr",
")",
"preorder",
"=",
"namedb_get_name_preorder",
"(",
"self",
".",
"db",
",",
"preorder_hash",
",",
"self",
".",
"lastblock",
")",
"if",
"preorder",
"is",
"None",
":",
"# doesn't exist or expired",
"return",
"None",
"# preorder must be younger than the namespace lifetime",
"# (otherwise we get into weird conditions where someone can preorder",
"# a name before someone else, and register it after it expires)",
"namespace_lifetime_multiplier",
"=",
"get_epoch_namespace_lifetime_multiplier",
"(",
"self",
".",
"lastblock",
",",
"namespace_id",
")",
"if",
"preorder",
"[",
"'block_number'",
"]",
"+",
"(",
"namespace",
"[",
"'lifetime'",
"]",
"*",
"namespace_lifetime_multiplier",
")",
"<=",
"self",
".",
"lastblock",
":",
"log",
".",
"debug",
"(",
"\"Preorder is too old (accepted at {}, namespace lifetime is {}, current block is {})\"",
".",
"format",
"(",
"preorder",
"[",
"'block_number'",
"]",
",",
"namespace",
"[",
"'lifetime'",
"]",
"*",
"namespace_lifetime_multiplier",
",",
"self",
".",
"lastblock",
")",
")",
"return",
"None",
"return",
"preorder"
] | Get the current preorder for a name, given the name, the sender's script pubkey, and
the registration address used to calculate the preorder hash.
Return the preorder record on success.
Return None if not found, or the preorder is already registered and not expired (even if revoked).
NOTE: possibly returns an expired preorder (by design, so as to prevent someone
from re-sending the same preorder with the same preorder hash). | [
"Get",
"the",
"current",
"preorder",
"for",
"a",
"name",
"given",
"the",
"name",
"the",
"sender",
"s",
"script",
"pubkey",
"and",
"the",
"registration",
"address",
"used",
"to",
"calculate",
"the",
"preorder",
"hash",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1166-L1203 |
230,615 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_names_with_value_hash | def get_names_with_value_hash( self, value_hash ):
"""
Get the list of names with the given value hash, at the current block height.
This excludes revoked names and expired names.
Return None if there are no such names
"""
cur = self.db.cursor()
names = namedb_get_names_with_value_hash( cur, value_hash, self.lastblock )
return names | python | def get_names_with_value_hash( self, value_hash ):
"""
Get the list of names with the given value hash, at the current block height.
This excludes revoked names and expired names.
Return None if there are no such names
"""
cur = self.db.cursor()
names = namedb_get_names_with_value_hash( cur, value_hash, self.lastblock )
return names | [
"def",
"get_names_with_value_hash",
"(",
"self",
",",
"value_hash",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"names",
"=",
"namedb_get_names_with_value_hash",
"(",
"cur",
",",
"value_hash",
",",
"self",
".",
"lastblock",
")",
"return",
"names"
] | Get the list of names with the given value hash, at the current block height.
This excludes revoked names and expired names.
Return None if there are no such names | [
"Get",
"the",
"list",
"of",
"names",
"with",
"the",
"given",
"value",
"hash",
"at",
"the",
"current",
"block",
"height",
".",
"This",
"excludes",
"revoked",
"names",
"and",
"expired",
"names",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1235-L1244 |
230,616 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_atlas_zonefile_info_at | def get_atlas_zonefile_info_at( self, block_id ):
"""
Get the blockchain-ordered sequence of names, value hashes, and txids.
added at the given block height. The order will be
in tx-order.
Return [{'name': name, 'value_hash': value_hash, 'txid': txid}]
"""
nameops = self.get_all_blockstack_ops_at( block_id )
ret = []
for nameop in nameops:
if nameop.has_key('op') and op_get_opcode_name(nameop['op']) in ['NAME_UPDATE', 'NAME_IMPORT', 'NAME_REGISTRATION', 'NAME_RENEWAL']:
assert nameop.has_key('value_hash')
assert nameop.has_key('name')
assert nameop.has_key('txid')
if nameop['value_hash'] is not None:
ret.append( {'name': nameop['name'], 'value_hash': nameop['value_hash'], 'txid': nameop['txid']} )
return ret | python | def get_atlas_zonefile_info_at( self, block_id ):
"""
Get the blockchain-ordered sequence of names, value hashes, and txids.
added at the given block height. The order will be
in tx-order.
Return [{'name': name, 'value_hash': value_hash, 'txid': txid}]
"""
nameops = self.get_all_blockstack_ops_at( block_id )
ret = []
for nameop in nameops:
if nameop.has_key('op') and op_get_opcode_name(nameop['op']) in ['NAME_UPDATE', 'NAME_IMPORT', 'NAME_REGISTRATION', 'NAME_RENEWAL']:
assert nameop.has_key('value_hash')
assert nameop.has_key('name')
assert nameop.has_key('txid')
if nameop['value_hash'] is not None:
ret.append( {'name': nameop['name'], 'value_hash': nameop['value_hash'], 'txid': nameop['txid']} )
return ret | [
"def",
"get_atlas_zonefile_info_at",
"(",
"self",
",",
"block_id",
")",
":",
"nameops",
"=",
"self",
".",
"get_all_blockstack_ops_at",
"(",
"block_id",
")",
"ret",
"=",
"[",
"]",
"for",
"nameop",
"in",
"nameops",
":",
"if",
"nameop",
".",
"has_key",
"(",
"'op'",
")",
"and",
"op_get_opcode_name",
"(",
"nameop",
"[",
"'op'",
"]",
")",
"in",
"[",
"'NAME_UPDATE'",
",",
"'NAME_IMPORT'",
",",
"'NAME_REGISTRATION'",
",",
"'NAME_RENEWAL'",
"]",
":",
"assert",
"nameop",
".",
"has_key",
"(",
"'value_hash'",
")",
"assert",
"nameop",
".",
"has_key",
"(",
"'name'",
")",
"assert",
"nameop",
".",
"has_key",
"(",
"'txid'",
")",
"if",
"nameop",
"[",
"'value_hash'",
"]",
"is",
"not",
"None",
":",
"ret",
".",
"append",
"(",
"{",
"'name'",
":",
"nameop",
"[",
"'name'",
"]",
",",
"'value_hash'",
":",
"nameop",
"[",
"'value_hash'",
"]",
",",
"'txid'",
":",
"nameop",
"[",
"'txid'",
"]",
"}",
")",
"return",
"ret"
] | Get the blockchain-ordered sequence of names, value hashes, and txids.
added at the given block height. The order will be
in tx-order.
Return [{'name': name, 'value_hash': value_hash, 'txid': txid}] | [
"Get",
"the",
"blockchain",
"-",
"ordered",
"sequence",
"of",
"names",
"value",
"hashes",
"and",
"txids",
".",
"added",
"at",
"the",
"given",
"block",
"height",
".",
"The",
"order",
"will",
"be",
"in",
"tx",
"-",
"order",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1247-L1267 |
230,617 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_namespace_reveal | def get_namespace_reveal( self, namespace_id, include_history=True ):
"""
Given the name of a namespace, get it if it is currently
being revealed.
Return the reveal record on success.
Return None if it is not being revealed, or is expired.
"""
cur = self.db.cursor()
namespace_reveal = namedb_get_namespace_reveal( cur, namespace_id, self.lastblock, include_history=include_history )
return namespace_reveal | python | def get_namespace_reveal( self, namespace_id, include_history=True ):
"""
Given the name of a namespace, get it if it is currently
being revealed.
Return the reveal record on success.
Return None if it is not being revealed, or is expired.
"""
cur = self.db.cursor()
namespace_reveal = namedb_get_namespace_reveal( cur, namespace_id, self.lastblock, include_history=include_history )
return namespace_reveal | [
"def",
"get_namespace_reveal",
"(",
"self",
",",
"namespace_id",
",",
"include_history",
"=",
"True",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"namespace_reveal",
"=",
"namedb_get_namespace_reveal",
"(",
"cur",
",",
"namespace_id",
",",
"self",
".",
"lastblock",
",",
"include_history",
"=",
"include_history",
")",
"return",
"namespace_reveal"
] | Given the name of a namespace, get it if it is currently
being revealed.
Return the reveal record on success.
Return None if it is not being revealed, or is expired. | [
"Given",
"the",
"name",
"of",
"a",
"namespace",
"get",
"it",
"if",
"it",
"is",
"currently",
"being",
"revealed",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1306-L1316 |
230,618 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.is_name_registered | def is_name_registered( self, name ):
"""
Given the fully-qualified name, is it registered, not revoked, and not expired
at the current block?
"""
name_rec = self.get_name( name ) # won't return the name if expired
if name_rec is None:
return False
if name_rec['revoked']:
return False
return True | python | def is_name_registered( self, name ):
"""
Given the fully-qualified name, is it registered, not revoked, and not expired
at the current block?
"""
name_rec = self.get_name( name ) # won't return the name if expired
if name_rec is None:
return False
if name_rec['revoked']:
return False
return True | [
"def",
"is_name_registered",
"(",
"self",
",",
"name",
")",
":",
"name_rec",
"=",
"self",
".",
"get_name",
"(",
"name",
")",
"# won't return the name if expired",
"if",
"name_rec",
"is",
"None",
":",
"return",
"False",
"if",
"name_rec",
"[",
"'revoked'",
"]",
":",
"return",
"False",
"return",
"True"
] | Given the fully-qualified name, is it registered, not revoked, and not expired
at the current block? | [
"Given",
"the",
"fully",
"-",
"qualified",
"name",
"is",
"it",
"registered",
"not",
"revoked",
"and",
"not",
"expired",
"at",
"the",
"current",
"block?"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1392-L1404 |
230,619 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.is_namespace_ready | def is_namespace_ready( self, namespace_id ):
"""
Given a namespace ID, determine if the namespace is ready
at the current block.
"""
namespace = self.get_namespace( namespace_id )
if namespace is not None:
return True
else:
return False | python | def is_namespace_ready( self, namespace_id ):
"""
Given a namespace ID, determine if the namespace is ready
at the current block.
"""
namespace = self.get_namespace( namespace_id )
if namespace is not None:
return True
else:
return False | [
"def",
"is_namespace_ready",
"(",
"self",
",",
"namespace_id",
")",
":",
"namespace",
"=",
"self",
".",
"get_namespace",
"(",
"namespace_id",
")",
"if",
"namespace",
"is",
"not",
"None",
":",
"return",
"True",
"else",
":",
"return",
"False"
] | Given a namespace ID, determine if the namespace is ready
at the current block. | [
"Given",
"a",
"namespace",
"ID",
"determine",
"if",
"the",
"namespace",
"is",
"ready",
"at",
"the",
"current",
"block",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1407-L1416 |
230,620 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.is_namespace_preordered | def is_namespace_preordered( self, namespace_id_hash ):
"""
Given a namespace preorder hash, determine if it is preordered
at the current block.
"""
namespace_preorder = self.get_namespace_preorder(namespace_id_hash)
if namespace_preorder is None:
return False
else:
return True | python | def is_namespace_preordered( self, namespace_id_hash ):
"""
Given a namespace preorder hash, determine if it is preordered
at the current block.
"""
namespace_preorder = self.get_namespace_preorder(namespace_id_hash)
if namespace_preorder is None:
return False
else:
return True | [
"def",
"is_namespace_preordered",
"(",
"self",
",",
"namespace_id_hash",
")",
":",
"namespace_preorder",
"=",
"self",
".",
"get_namespace_preorder",
"(",
"namespace_id_hash",
")",
"if",
"namespace_preorder",
"is",
"None",
":",
"return",
"False",
"else",
":",
"return",
"True"
] | Given a namespace preorder hash, determine if it is preordered
at the current block. | [
"Given",
"a",
"namespace",
"preorder",
"hash",
"determine",
"if",
"it",
"is",
"preordered",
"at",
"the",
"current",
"block",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1419-L1428 |
230,621 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.is_namespace_revealed | def is_namespace_revealed( self, namespace_id ):
"""
Given the name of a namespace, has it been revealed but not made ready
at the current block?
"""
namespace_reveal = self.get_namespace_reveal( namespace_id )
if namespace_reveal is not None:
return True
else:
return False | python | def is_namespace_revealed( self, namespace_id ):
"""
Given the name of a namespace, has it been revealed but not made ready
at the current block?
"""
namespace_reveal = self.get_namespace_reveal( namespace_id )
if namespace_reveal is not None:
return True
else:
return False | [
"def",
"is_namespace_revealed",
"(",
"self",
",",
"namespace_id",
")",
":",
"namespace_reveal",
"=",
"self",
".",
"get_namespace_reveal",
"(",
"namespace_id",
")",
"if",
"namespace_reveal",
"is",
"not",
"None",
":",
"return",
"True",
"else",
":",
"return",
"False"
] | Given the name of a namespace, has it been revealed but not made ready
at the current block? | [
"Given",
"the",
"name",
"of",
"a",
"namespace",
"has",
"it",
"been",
"revealed",
"but",
"not",
"made",
"ready",
"at",
"the",
"current",
"block?"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1431-L1440 |
230,622 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.is_name_owner | def is_name_owner( self, name, sender_script_pubkey ):
"""
Given the fully-qualified name and a sender's script pubkey,
determine if the sender owns the name.
The name must exist and not be revoked or expired at the
current block.
"""
if not self.is_name_registered( name ):
# no one owns it
return False
owner = self.get_name_owner( name )
if owner != sender_script_pubkey:
return False
else:
return True | python | def is_name_owner( self, name, sender_script_pubkey ):
"""
Given the fully-qualified name and a sender's script pubkey,
determine if the sender owns the name.
The name must exist and not be revoked or expired at the
current block.
"""
if not self.is_name_registered( name ):
# no one owns it
return False
owner = self.get_name_owner( name )
if owner != sender_script_pubkey:
return False
else:
return True | [
"def",
"is_name_owner",
"(",
"self",
",",
"name",
",",
"sender_script_pubkey",
")",
":",
"if",
"not",
"self",
".",
"is_name_registered",
"(",
"name",
")",
":",
"# no one owns it ",
"return",
"False",
"owner",
"=",
"self",
".",
"get_name_owner",
"(",
"name",
")",
"if",
"owner",
"!=",
"sender_script_pubkey",
":",
"return",
"False",
"else",
":",
"return",
"True"
] | Given the fully-qualified name and a sender's script pubkey,
determine if the sender owns the name.
The name must exist and not be revoked or expired at the
current block. | [
"Given",
"the",
"fully",
"-",
"qualified",
"name",
"and",
"a",
"sender",
"s",
"script",
"pubkey",
"determine",
"if",
"the",
"sender",
"owns",
"the",
"name",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1443-L1459 |
230,623 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.is_new_preorder | def is_new_preorder( self, preorder_hash, lastblock=None ):
"""
Given a preorder hash of a name, determine whether or not it is unseen before.
"""
if lastblock is None:
lastblock = self.lastblock
preorder = namedb_get_name_preorder( self.db, preorder_hash, lastblock )
if preorder is not None:
return False
else:
return True | python | def is_new_preorder( self, preorder_hash, lastblock=None ):
"""
Given a preorder hash of a name, determine whether or not it is unseen before.
"""
if lastblock is None:
lastblock = self.lastblock
preorder = namedb_get_name_preorder( self.db, preorder_hash, lastblock )
if preorder is not None:
return False
else:
return True | [
"def",
"is_new_preorder",
"(",
"self",
",",
"preorder_hash",
",",
"lastblock",
"=",
"None",
")",
":",
"if",
"lastblock",
"is",
"None",
":",
"lastblock",
"=",
"self",
".",
"lastblock",
"preorder",
"=",
"namedb_get_name_preorder",
"(",
"self",
".",
"db",
",",
"preorder_hash",
",",
"lastblock",
")",
"if",
"preorder",
"is",
"not",
"None",
":",
"return",
"False",
"else",
":",
"return",
"True"
] | Given a preorder hash of a name, determine whether or not it is unseen before. | [
"Given",
"a",
"preorder",
"hash",
"of",
"a",
"name",
"determine",
"whether",
"or",
"not",
"it",
"is",
"unseen",
"before",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1462-L1473 |
230,624 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.is_new_namespace_preorder | def is_new_namespace_preorder( self, namespace_id_hash, lastblock=None ):
"""
Given a namespace preorder hash, determine whether or not is is unseen before.
"""
if lastblock is None:
lastblock = self.lastblock
preorder = namedb_get_namespace_preorder( self.db, namespace_id_hash, lastblock )
if preorder is not None:
return False
else:
return True | python | def is_new_namespace_preorder( self, namespace_id_hash, lastblock=None ):
"""
Given a namespace preorder hash, determine whether or not is is unseen before.
"""
if lastblock is None:
lastblock = self.lastblock
preorder = namedb_get_namespace_preorder( self.db, namespace_id_hash, lastblock )
if preorder is not None:
return False
else:
return True | [
"def",
"is_new_namespace_preorder",
"(",
"self",
",",
"namespace_id_hash",
",",
"lastblock",
"=",
"None",
")",
":",
"if",
"lastblock",
"is",
"None",
":",
"lastblock",
"=",
"self",
".",
"lastblock",
"preorder",
"=",
"namedb_get_namespace_preorder",
"(",
"self",
".",
"db",
",",
"namespace_id_hash",
",",
"lastblock",
")",
"if",
"preorder",
"is",
"not",
"None",
":",
"return",
"False",
"else",
":",
"return",
"True"
] | Given a namespace preorder hash, determine whether or not is is unseen before. | [
"Given",
"a",
"namespace",
"preorder",
"hash",
"determine",
"whether",
"or",
"not",
"is",
"is",
"unseen",
"before",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1476-L1487 |
230,625 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.is_name_revoked | def is_name_revoked( self, name ):
"""
Determine if a name is revoked at this block.
"""
name = self.get_name( name )
if name is None:
return False
if name['revoked']:
return True
else:
return False | python | def is_name_revoked( self, name ):
"""
Determine if a name is revoked at this block.
"""
name = self.get_name( name )
if name is None:
return False
if name['revoked']:
return True
else:
return False | [
"def",
"is_name_revoked",
"(",
"self",
",",
"name",
")",
":",
"name",
"=",
"self",
".",
"get_name",
"(",
"name",
")",
"if",
"name",
"is",
"None",
":",
"return",
"False",
"if",
"name",
"[",
"'revoked'",
"]",
":",
"return",
"True",
"else",
":",
"return",
"False"
] | Determine if a name is revoked at this block. | [
"Determine",
"if",
"a",
"name",
"is",
"revoked",
"at",
"this",
"block",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1490-L1501 |
230,626 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.get_value_hash_txids | def get_value_hash_txids(self, value_hash):
"""
Get the list of txids by value hash
"""
cur = self.db.cursor()
return namedb_get_value_hash_txids(cur, value_hash) | python | def get_value_hash_txids(self, value_hash):
"""
Get the list of txids by value hash
"""
cur = self.db.cursor()
return namedb_get_value_hash_txids(cur, value_hash) | [
"def",
"get_value_hash_txids",
"(",
"self",
",",
"value_hash",
")",
":",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"return",
"namedb_get_value_hash_txids",
"(",
"cur",
",",
"value_hash",
")"
] | Get the list of txids by value hash | [
"Get",
"the",
"list",
"of",
"txids",
"by",
"value",
"hash"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1511-L1516 |
230,627 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.nameop_set_collided | def nameop_set_collided( cls, nameop, history_id_key, history_id ):
"""
Mark a nameop as collided
"""
nameop['__collided__'] = True
nameop['__collided_history_id_key__'] = history_id_key
nameop['__collided_history_id__'] = history_id | python | def nameop_set_collided( cls, nameop, history_id_key, history_id ):
"""
Mark a nameop as collided
"""
nameop['__collided__'] = True
nameop['__collided_history_id_key__'] = history_id_key
nameop['__collided_history_id__'] = history_id | [
"def",
"nameop_set_collided",
"(",
"cls",
",",
"nameop",
",",
"history_id_key",
",",
"history_id",
")",
":",
"nameop",
"[",
"'__collided__'",
"]",
"=",
"True",
"nameop",
"[",
"'__collided_history_id_key__'",
"]",
"=",
"history_id_key",
"nameop",
"[",
"'__collided_history_id__'",
"]",
"=",
"history_id"
] | Mark a nameop as collided | [
"Mark",
"a",
"nameop",
"as",
"collided"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1530-L1536 |
230,628 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.nameop_put_collision | def nameop_put_collision( cls, collisions, nameop ):
"""
Record a nameop as collided with another nameop in this block.
"""
# these are supposed to have been put here by nameop_set_collided
history_id_key = nameop.get('__collided_history_id_key__', None)
history_id = nameop.get('__collided_history_id__', None)
try:
assert cls.nameop_is_collided( nameop ), "Nameop not collided"
assert history_id_key is not None, "Nameop missing collision info"
assert history_id is not None, "Nameop missing collision info"
except Exception, e:
log.exception(e)
log.error("FATAL: BUG: bad collision info")
os.abort()
if not collisions.has_key(history_id_key):
collisions[history_id_key] = [history_id]
else:
collisions[history_id_key].append( history_id ) | python | def nameop_put_collision( cls, collisions, nameop ):
"""
Record a nameop as collided with another nameop in this block.
"""
# these are supposed to have been put here by nameop_set_collided
history_id_key = nameop.get('__collided_history_id_key__', None)
history_id = nameop.get('__collided_history_id__', None)
try:
assert cls.nameop_is_collided( nameop ), "Nameop not collided"
assert history_id_key is not None, "Nameop missing collision info"
assert history_id is not None, "Nameop missing collision info"
except Exception, e:
log.exception(e)
log.error("FATAL: BUG: bad collision info")
os.abort()
if not collisions.has_key(history_id_key):
collisions[history_id_key] = [history_id]
else:
collisions[history_id_key].append( history_id ) | [
"def",
"nameop_put_collision",
"(",
"cls",
",",
"collisions",
",",
"nameop",
")",
":",
"# these are supposed to have been put here by nameop_set_collided",
"history_id_key",
"=",
"nameop",
".",
"get",
"(",
"'__collided_history_id_key__'",
",",
"None",
")",
"history_id",
"=",
"nameop",
".",
"get",
"(",
"'__collided_history_id__'",
",",
"None",
")",
"try",
":",
"assert",
"cls",
".",
"nameop_is_collided",
"(",
"nameop",
")",
",",
"\"Nameop not collided\"",
"assert",
"history_id_key",
"is",
"not",
"None",
",",
"\"Nameop missing collision info\"",
"assert",
"history_id",
"is",
"not",
"None",
",",
"\"Nameop missing collision info\"",
"except",
"Exception",
",",
"e",
":",
"log",
".",
"exception",
"(",
"e",
")",
"log",
".",
"error",
"(",
"\"FATAL: BUG: bad collision info\"",
")",
"os",
".",
"abort",
"(",
")",
"if",
"not",
"collisions",
".",
"has_key",
"(",
"history_id_key",
")",
":",
"collisions",
"[",
"history_id_key",
"]",
"=",
"[",
"history_id",
"]",
"else",
":",
"collisions",
"[",
"history_id_key",
"]",
".",
"append",
"(",
"history_id",
")"
] | Record a nameop as collided with another nameop in this block. | [
"Record",
"a",
"nameop",
"as",
"collided",
"with",
"another",
"nameop",
"in",
"this",
"block",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1548-L1568 |
230,629 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.extract_consensus_op | def extract_consensus_op(self, opcode, op_data, processed_op_data, current_block_number):
"""
Using the operation data extracted from parsing the virtualchain operation (@op_data),
and the checked, processed operation (@processed_op_data), return a dict that contains
(1) all of the consensus fields to snapshot this operation, and
(2) all of the data fields that we need to store for the name record (i.e. quirk fields)
"""
ret = {}
consensus_fields = op_get_consensus_fields(opcode)
quirk_fields = op_get_quirk_fields(opcode)
for field in consensus_fields + quirk_fields:
try:
# assert field in processed_op_data or field in op_data, 'Missing consensus field "{}"'.format(field)
assert field in processed_op_data, 'Missing consensus field "{}"'.format(field)
except Exception as e:
# should NEVER happen
log.exception(e)
log.error("FATAL: BUG: missing consensus field {}".format(field))
log.error("op_data:\n{}".format(json.dumps(op_data, indent=4, sort_keys=True)))
log.error("processed_op_data:\n{}".format(json.dumps(op_data, indent=4, sort_keys=True)))
os.abort()
ret[field] = processed_op_data[field]
return ret | python | def extract_consensus_op(self, opcode, op_data, processed_op_data, current_block_number):
"""
Using the operation data extracted from parsing the virtualchain operation (@op_data),
and the checked, processed operation (@processed_op_data), return a dict that contains
(1) all of the consensus fields to snapshot this operation, and
(2) all of the data fields that we need to store for the name record (i.e. quirk fields)
"""
ret = {}
consensus_fields = op_get_consensus_fields(opcode)
quirk_fields = op_get_quirk_fields(opcode)
for field in consensus_fields + quirk_fields:
try:
# assert field in processed_op_data or field in op_data, 'Missing consensus field "{}"'.format(field)
assert field in processed_op_data, 'Missing consensus field "{}"'.format(field)
except Exception as e:
# should NEVER happen
log.exception(e)
log.error("FATAL: BUG: missing consensus field {}".format(field))
log.error("op_data:\n{}".format(json.dumps(op_data, indent=4, sort_keys=True)))
log.error("processed_op_data:\n{}".format(json.dumps(op_data, indent=4, sort_keys=True)))
os.abort()
ret[field] = processed_op_data[field]
return ret | [
"def",
"extract_consensus_op",
"(",
"self",
",",
"opcode",
",",
"op_data",
",",
"processed_op_data",
",",
"current_block_number",
")",
":",
"ret",
"=",
"{",
"}",
"consensus_fields",
"=",
"op_get_consensus_fields",
"(",
"opcode",
")",
"quirk_fields",
"=",
"op_get_quirk_fields",
"(",
"opcode",
")",
"for",
"field",
"in",
"consensus_fields",
"+",
"quirk_fields",
":",
"try",
":",
"# assert field in processed_op_data or field in op_data, 'Missing consensus field \"{}\"'.format(field)",
"assert",
"field",
"in",
"processed_op_data",
",",
"'Missing consensus field \"{}\"'",
".",
"format",
"(",
"field",
")",
"except",
"Exception",
"as",
"e",
":",
"# should NEVER happen",
"log",
".",
"exception",
"(",
"e",
")",
"log",
".",
"error",
"(",
"\"FATAL: BUG: missing consensus field {}\"",
".",
"format",
"(",
"field",
")",
")",
"log",
".",
"error",
"(",
"\"op_data:\\n{}\"",
".",
"format",
"(",
"json",
".",
"dumps",
"(",
"op_data",
",",
"indent",
"=",
"4",
",",
"sort_keys",
"=",
"True",
")",
")",
")",
"log",
".",
"error",
"(",
"\"processed_op_data:\\n{}\"",
".",
"format",
"(",
"json",
".",
"dumps",
"(",
"op_data",
",",
"indent",
"=",
"4",
",",
"sort_keys",
"=",
"True",
")",
")",
")",
"os",
".",
"abort",
"(",
")",
"ret",
"[",
"field",
"]",
"=",
"processed_op_data",
"[",
"field",
"]",
"return",
"ret"
] | Using the operation data extracted from parsing the virtualchain operation (@op_data),
and the checked, processed operation (@processed_op_data), return a dict that contains
(1) all of the consensus fields to snapshot this operation, and
(2) all of the data fields that we need to store for the name record (i.e. quirk fields) | [
"Using",
"the",
"operation",
"data",
"extracted",
"from",
"parsing",
"the",
"virtualchain",
"operation",
"("
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1571-L1597 |
230,630 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.commit_operation | def commit_operation( self, input_op_data, accepted_nameop, current_block_number ):
"""
Commit an operation, thereby carrying out a state transition.
Returns a dict with the new db record fields
"""
# have to have read-write disposition
if self.disposition != DISPOSITION_RW:
log.error("FATAL: borrowing violation: not a read-write connection")
traceback.print_stack()
os.abort()
cur = self.db.cursor()
canonical_op = None
op_type_str = None # for debugging
opcode = accepted_nameop.get('opcode', None)
try:
assert opcode is not None, "Undefined op '%s'" % accepted_nameop['op']
except Exception, e:
log.exception(e)
log.error("FATAL: unrecognized op '%s'" % accepted_nameop['op'] )
os.abort()
if opcode in OPCODE_PREORDER_OPS:
# preorder
canonical_op = self.commit_state_preorder( accepted_nameop, current_block_number )
op_type_str = "state_preorder"
elif opcode in OPCODE_CREATION_OPS:
# creation
canonical_op = self.commit_state_create( accepted_nameop, current_block_number )
op_type_str = "state_create"
elif opcode in OPCODE_TRANSITION_OPS:
# transition
canonical_op = self.commit_state_transition( accepted_nameop, current_block_number )
op_type_str = "state_transition"
elif opcode in OPCODE_TOKEN_OPS:
# token operation
canonical_op = self.commit_token_operation(accepted_nameop, current_block_number)
op_type_str = "token_operation"
else:
raise Exception("Unknown operation {}".format(opcode))
if canonical_op is None:
log.error("FATAL: no canonical op generated (for {})".format(op_type_str))
os.abort()
log.debug("Extract consensus fields for {} in {}, as part of a {}".format(opcode, current_block_number, op_type_str))
consensus_op = self.extract_consensus_op(opcode, input_op_data, canonical_op, current_block_number)
return consensus_op | python | def commit_operation( self, input_op_data, accepted_nameop, current_block_number ):
"""
Commit an operation, thereby carrying out a state transition.
Returns a dict with the new db record fields
"""
# have to have read-write disposition
if self.disposition != DISPOSITION_RW:
log.error("FATAL: borrowing violation: not a read-write connection")
traceback.print_stack()
os.abort()
cur = self.db.cursor()
canonical_op = None
op_type_str = None # for debugging
opcode = accepted_nameop.get('opcode', None)
try:
assert opcode is not None, "Undefined op '%s'" % accepted_nameop['op']
except Exception, e:
log.exception(e)
log.error("FATAL: unrecognized op '%s'" % accepted_nameop['op'] )
os.abort()
if opcode in OPCODE_PREORDER_OPS:
# preorder
canonical_op = self.commit_state_preorder( accepted_nameop, current_block_number )
op_type_str = "state_preorder"
elif opcode in OPCODE_CREATION_OPS:
# creation
canonical_op = self.commit_state_create( accepted_nameop, current_block_number )
op_type_str = "state_create"
elif opcode in OPCODE_TRANSITION_OPS:
# transition
canonical_op = self.commit_state_transition( accepted_nameop, current_block_number )
op_type_str = "state_transition"
elif opcode in OPCODE_TOKEN_OPS:
# token operation
canonical_op = self.commit_token_operation(accepted_nameop, current_block_number)
op_type_str = "token_operation"
else:
raise Exception("Unknown operation {}".format(opcode))
if canonical_op is None:
log.error("FATAL: no canonical op generated (for {})".format(op_type_str))
os.abort()
log.debug("Extract consensus fields for {} in {}, as part of a {}".format(opcode, current_block_number, op_type_str))
consensus_op = self.extract_consensus_op(opcode, input_op_data, canonical_op, current_block_number)
return consensus_op | [
"def",
"commit_operation",
"(",
"self",
",",
"input_op_data",
",",
"accepted_nameop",
",",
"current_block_number",
")",
":",
"# have to have read-write disposition ",
"if",
"self",
".",
"disposition",
"!=",
"DISPOSITION_RW",
":",
"log",
".",
"error",
"(",
"\"FATAL: borrowing violation: not a read-write connection\"",
")",
"traceback",
".",
"print_stack",
"(",
")",
"os",
".",
"abort",
"(",
")",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"canonical_op",
"=",
"None",
"op_type_str",
"=",
"None",
"# for debugging",
"opcode",
"=",
"accepted_nameop",
".",
"get",
"(",
"'opcode'",
",",
"None",
")",
"try",
":",
"assert",
"opcode",
"is",
"not",
"None",
",",
"\"Undefined op '%s'\"",
"%",
"accepted_nameop",
"[",
"'op'",
"]",
"except",
"Exception",
",",
"e",
":",
"log",
".",
"exception",
"(",
"e",
")",
"log",
".",
"error",
"(",
"\"FATAL: unrecognized op '%s'\"",
"%",
"accepted_nameop",
"[",
"'op'",
"]",
")",
"os",
".",
"abort",
"(",
")",
"if",
"opcode",
"in",
"OPCODE_PREORDER_OPS",
":",
"# preorder",
"canonical_op",
"=",
"self",
".",
"commit_state_preorder",
"(",
"accepted_nameop",
",",
"current_block_number",
")",
"op_type_str",
"=",
"\"state_preorder\"",
"elif",
"opcode",
"in",
"OPCODE_CREATION_OPS",
":",
"# creation",
"canonical_op",
"=",
"self",
".",
"commit_state_create",
"(",
"accepted_nameop",
",",
"current_block_number",
")",
"op_type_str",
"=",
"\"state_create\"",
"elif",
"opcode",
"in",
"OPCODE_TRANSITION_OPS",
":",
"# transition ",
"canonical_op",
"=",
"self",
".",
"commit_state_transition",
"(",
"accepted_nameop",
",",
"current_block_number",
")",
"op_type_str",
"=",
"\"state_transition\"",
"elif",
"opcode",
"in",
"OPCODE_TOKEN_OPS",
":",
"# token operation ",
"canonical_op",
"=",
"self",
".",
"commit_token_operation",
"(",
"accepted_nameop",
",",
"current_block_number",
")",
"op_type_str",
"=",
"\"token_operation\"",
"else",
":",
"raise",
"Exception",
"(",
"\"Unknown operation {}\"",
".",
"format",
"(",
"opcode",
")",
")",
"if",
"canonical_op",
"is",
"None",
":",
"log",
".",
"error",
"(",
"\"FATAL: no canonical op generated (for {})\"",
".",
"format",
"(",
"op_type_str",
")",
")",
"os",
".",
"abort",
"(",
")",
"log",
".",
"debug",
"(",
"\"Extract consensus fields for {} in {}, as part of a {}\"",
".",
"format",
"(",
"opcode",
",",
"current_block_number",
",",
"op_type_str",
")",
")",
"consensus_op",
"=",
"self",
".",
"extract_consensus_op",
"(",
"opcode",
",",
"input_op_data",
",",
"canonical_op",
",",
"current_block_number",
")",
"return",
"consensus_op"
] | Commit an operation, thereby carrying out a state transition.
Returns a dict with the new db record fields | [
"Commit",
"an",
"operation",
"thereby",
"carrying",
"out",
"a",
"state",
"transition",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1600-L1654 |
230,631 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.commit_token_operation | def commit_token_operation(self, token_op, current_block_number):
"""
Commit a token operation that debits one account and credits another
Returns the new canonicalized record (with all compatibility quirks preserved)
DO NOT CALL THIS DIRECTLY
"""
# have to have read-write disposition
if self.disposition != DISPOSITION_RW:
log.error("FATAL: borrowing violation: not a read-write connection")
traceback.print_stack()
os.abort()
cur = self.db.cursor()
opcode = token_op.get('opcode', None)
clean_token_op = self.sanitize_op(token_op)
try:
assert token_operation_is_valid(token_op), 'Invalid token operation'
assert opcode is not None, 'No opcode given'
assert 'txid' in token_op, 'No txid'
assert 'vtxindex' in token_op, 'No vtxindex'
except Exception as e:
log.exception(e)
log.error('FATAL: failed to commit token operation')
self.db.rollback()
os.abort()
table = token_operation_get_table(token_op)
account_payment_info = token_operation_get_account_payment_info(token_op)
account_credit_info = token_operation_get_account_credit_info(token_op)
# fields must be set
try:
for key in account_payment_info:
assert account_payment_info[key] is not None, 'BUG: payment info key {} is None'.format(key)
for key in account_credit_info:
assert account_credit_info[key] is not None, 'BUG: credit info key {} is not None'.format(key)
# NOTE: do not check token amount and type, since in the future we want to support converting
# between tokens
except Exception as e:
log.exception(e)
log.error("FATAL: invalid token debit or credit info")
os.abort()
self.log_accept(current_block_number, token_op['vtxindex'], token_op['op'], token_op)
# NOTE: this code is single-threaded, but this code must be atomic
self.commit_account_debit(token_op, account_payment_info, current_block_number, token_op['vtxindex'], token_op['txid'])
self.commit_account_credit(token_op, account_credit_info, current_block_number, token_op['vtxindex'], token_op['txid'])
namedb_history_save(cur, opcode, token_op['address'], None, None, current_block_number, token_op['vtxindex'], token_op['txid'], clean_token_op)
return clean_token_op | python | def commit_token_operation(self, token_op, current_block_number):
"""
Commit a token operation that debits one account and credits another
Returns the new canonicalized record (with all compatibility quirks preserved)
DO NOT CALL THIS DIRECTLY
"""
# have to have read-write disposition
if self.disposition != DISPOSITION_RW:
log.error("FATAL: borrowing violation: not a read-write connection")
traceback.print_stack()
os.abort()
cur = self.db.cursor()
opcode = token_op.get('opcode', None)
clean_token_op = self.sanitize_op(token_op)
try:
assert token_operation_is_valid(token_op), 'Invalid token operation'
assert opcode is not None, 'No opcode given'
assert 'txid' in token_op, 'No txid'
assert 'vtxindex' in token_op, 'No vtxindex'
except Exception as e:
log.exception(e)
log.error('FATAL: failed to commit token operation')
self.db.rollback()
os.abort()
table = token_operation_get_table(token_op)
account_payment_info = token_operation_get_account_payment_info(token_op)
account_credit_info = token_operation_get_account_credit_info(token_op)
# fields must be set
try:
for key in account_payment_info:
assert account_payment_info[key] is not None, 'BUG: payment info key {} is None'.format(key)
for key in account_credit_info:
assert account_credit_info[key] is not None, 'BUG: credit info key {} is not None'.format(key)
# NOTE: do not check token amount and type, since in the future we want to support converting
# between tokens
except Exception as e:
log.exception(e)
log.error("FATAL: invalid token debit or credit info")
os.abort()
self.log_accept(current_block_number, token_op['vtxindex'], token_op['op'], token_op)
# NOTE: this code is single-threaded, but this code must be atomic
self.commit_account_debit(token_op, account_payment_info, current_block_number, token_op['vtxindex'], token_op['txid'])
self.commit_account_credit(token_op, account_credit_info, current_block_number, token_op['vtxindex'], token_op['txid'])
namedb_history_save(cur, opcode, token_op['address'], None, None, current_block_number, token_op['vtxindex'], token_op['txid'], clean_token_op)
return clean_token_op | [
"def",
"commit_token_operation",
"(",
"self",
",",
"token_op",
",",
"current_block_number",
")",
":",
"# have to have read-write disposition ",
"if",
"self",
".",
"disposition",
"!=",
"DISPOSITION_RW",
":",
"log",
".",
"error",
"(",
"\"FATAL: borrowing violation: not a read-write connection\"",
")",
"traceback",
".",
"print_stack",
"(",
")",
"os",
".",
"abort",
"(",
")",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"opcode",
"=",
"token_op",
".",
"get",
"(",
"'opcode'",
",",
"None",
")",
"clean_token_op",
"=",
"self",
".",
"sanitize_op",
"(",
"token_op",
")",
"try",
":",
"assert",
"token_operation_is_valid",
"(",
"token_op",
")",
",",
"'Invalid token operation'",
"assert",
"opcode",
"is",
"not",
"None",
",",
"'No opcode given'",
"assert",
"'txid'",
"in",
"token_op",
",",
"'No txid'",
"assert",
"'vtxindex'",
"in",
"token_op",
",",
"'No vtxindex'",
"except",
"Exception",
"as",
"e",
":",
"log",
".",
"exception",
"(",
"e",
")",
"log",
".",
"error",
"(",
"'FATAL: failed to commit token operation'",
")",
"self",
".",
"db",
".",
"rollback",
"(",
")",
"os",
".",
"abort",
"(",
")",
"table",
"=",
"token_operation_get_table",
"(",
"token_op",
")",
"account_payment_info",
"=",
"token_operation_get_account_payment_info",
"(",
"token_op",
")",
"account_credit_info",
"=",
"token_operation_get_account_credit_info",
"(",
"token_op",
")",
"# fields must be set",
"try",
":",
"for",
"key",
"in",
"account_payment_info",
":",
"assert",
"account_payment_info",
"[",
"key",
"]",
"is",
"not",
"None",
",",
"'BUG: payment info key {} is None'",
".",
"format",
"(",
"key",
")",
"for",
"key",
"in",
"account_credit_info",
":",
"assert",
"account_credit_info",
"[",
"key",
"]",
"is",
"not",
"None",
",",
"'BUG: credit info key {} is not None'",
".",
"format",
"(",
"key",
")",
"# NOTE: do not check token amount and type, since in the future we want to support converting",
"# between tokens",
"except",
"Exception",
"as",
"e",
":",
"log",
".",
"exception",
"(",
"e",
")",
"log",
".",
"error",
"(",
"\"FATAL: invalid token debit or credit info\"",
")",
"os",
".",
"abort",
"(",
")",
"self",
".",
"log_accept",
"(",
"current_block_number",
",",
"token_op",
"[",
"'vtxindex'",
"]",
",",
"token_op",
"[",
"'op'",
"]",
",",
"token_op",
")",
"# NOTE: this code is single-threaded, but this code must be atomic",
"self",
".",
"commit_account_debit",
"(",
"token_op",
",",
"account_payment_info",
",",
"current_block_number",
",",
"token_op",
"[",
"'vtxindex'",
"]",
",",
"token_op",
"[",
"'txid'",
"]",
")",
"self",
".",
"commit_account_credit",
"(",
"token_op",
",",
"account_credit_info",
",",
"current_block_number",
",",
"token_op",
"[",
"'vtxindex'",
"]",
",",
"token_op",
"[",
"'txid'",
"]",
")",
"namedb_history_save",
"(",
"cur",
",",
"opcode",
",",
"token_op",
"[",
"'address'",
"]",
",",
"None",
",",
"None",
",",
"current_block_number",
",",
"token_op",
"[",
"'vtxindex'",
"]",
",",
"token_op",
"[",
"'txid'",
"]",
",",
"clean_token_op",
")",
"return",
"clean_token_op"
] | Commit a token operation that debits one account and credits another
Returns the new canonicalized record (with all compatibility quirks preserved)
DO NOT CALL THIS DIRECTLY | [
"Commit",
"a",
"token",
"operation",
"that",
"debits",
"one",
"account",
"and",
"credits",
"another"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1935-L1990 |
230,632 | blockstack/blockstack-core | blockstack/lib/nameset/namedb.py | BlockstackDB.commit_account_vesting | def commit_account_vesting(self, block_height):
"""
vest any tokens at this block height
"""
# save all state
log.debug("Commit all database state before vesting")
self.db.commit()
if block_height in self.vesting:
traceback.print_stack()
log.fatal("Tried to vest tokens twice at {}".format(block_height))
os.abort()
# commit all vesting in one transaction
cur = self.db.cursor()
namedb_query_execute(cur, 'BEGIN', ())
res = namedb_accounts_vest(cur, block_height)
namedb_query_execute(cur, 'END', ())
self.vesting[block_height] = True
return True | python | def commit_account_vesting(self, block_height):
"""
vest any tokens at this block height
"""
# save all state
log.debug("Commit all database state before vesting")
self.db.commit()
if block_height in self.vesting:
traceback.print_stack()
log.fatal("Tried to vest tokens twice at {}".format(block_height))
os.abort()
# commit all vesting in one transaction
cur = self.db.cursor()
namedb_query_execute(cur, 'BEGIN', ())
res = namedb_accounts_vest(cur, block_height)
namedb_query_execute(cur, 'END', ())
self.vesting[block_height] = True
return True | [
"def",
"commit_account_vesting",
"(",
"self",
",",
"block_height",
")",
":",
"# save all state",
"log",
".",
"debug",
"(",
"\"Commit all database state before vesting\"",
")",
"self",
".",
"db",
".",
"commit",
"(",
")",
"if",
"block_height",
"in",
"self",
".",
"vesting",
":",
"traceback",
".",
"print_stack",
"(",
")",
"log",
".",
"fatal",
"(",
"\"Tried to vest tokens twice at {}\"",
".",
"format",
"(",
"block_height",
")",
")",
"os",
".",
"abort",
"(",
")",
"# commit all vesting in one transaction",
"cur",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"namedb_query_execute",
"(",
"cur",
",",
"'BEGIN'",
",",
"(",
")",
")",
"res",
"=",
"namedb_accounts_vest",
"(",
"cur",
",",
"block_height",
")",
"namedb_query_execute",
"(",
"cur",
",",
"'END'",
",",
"(",
")",
")",
"self",
".",
"vesting",
"[",
"block_height",
"]",
"=",
"True",
"return",
"True"
] | vest any tokens at this block height | [
"vest",
"any",
"tokens",
"at",
"this",
"block",
"height"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/nameset/namedb.py#L1993-L2013 |
230,633 | blockstack/blockstack-core | blockstack/lib/scripts.py | is_name_valid | def is_name_valid(fqn):
"""
Is a fully-qualified name acceptable?
Return True if so
Return False if not
>>> is_name_valid('abcd')
False
>>> is_name_valid('abcd.')
False
>>> is_name_valid('.abcd')
False
>>> is_name_valid('Abcd.abcd')
False
>>> is_name_valid('abcd.abc.d')
False
>>> is_name_valid('abcd.abc+d')
False
>>> is_name_valid('a.b.c')
False
>>> is_name_valid(True)
False
>>> is_name_valid(123)
False
>>> is_name_valid(None)
False
>>> is_name_valid('')
False
>>> is_name_valid('abcdabcdabcdabcdabcdabcdabcdabcda.bcd')
True
>>> is_name_valid('abcdabcdabcdabcdabcdabcdabcdabcdab.bcd')
False
>>> is_name_valid('abcdabcdabcdabcdabcdabcdabcdabcdabc.d')
True
>>> is_name_valid('a+b.c')
False
>>> is_name_valid('a_b.c')
True
"""
if not isinstance(fqn, (str,unicode)):
return False
if fqn.count( "." ) != 1:
return False
name, namespace_id = fqn.split(".")
if len(name) == 0 or len(namespace_id) == 0:
return False
if not is_b40( name ) or "+" in name or "." in name:
return False
if not is_namespace_valid( namespace_id ):
return False
if len(fqn) > LENGTHS['blockchain_id_name']:
# too long
return False
return True | python | def is_name_valid(fqn):
"""
Is a fully-qualified name acceptable?
Return True if so
Return False if not
>>> is_name_valid('abcd')
False
>>> is_name_valid('abcd.')
False
>>> is_name_valid('.abcd')
False
>>> is_name_valid('Abcd.abcd')
False
>>> is_name_valid('abcd.abc.d')
False
>>> is_name_valid('abcd.abc+d')
False
>>> is_name_valid('a.b.c')
False
>>> is_name_valid(True)
False
>>> is_name_valid(123)
False
>>> is_name_valid(None)
False
>>> is_name_valid('')
False
>>> is_name_valid('abcdabcdabcdabcdabcdabcdabcdabcda.bcd')
True
>>> is_name_valid('abcdabcdabcdabcdabcdabcdabcdabcdab.bcd')
False
>>> is_name_valid('abcdabcdabcdabcdabcdabcdabcdabcdabc.d')
True
>>> is_name_valid('a+b.c')
False
>>> is_name_valid('a_b.c')
True
"""
if not isinstance(fqn, (str,unicode)):
return False
if fqn.count( "." ) != 1:
return False
name, namespace_id = fqn.split(".")
if len(name) == 0 or len(namespace_id) == 0:
return False
if not is_b40( name ) or "+" in name or "." in name:
return False
if not is_namespace_valid( namespace_id ):
return False
if len(fqn) > LENGTHS['blockchain_id_name']:
# too long
return False
return True | [
"def",
"is_name_valid",
"(",
"fqn",
")",
":",
"if",
"not",
"isinstance",
"(",
"fqn",
",",
"(",
"str",
",",
"unicode",
")",
")",
":",
"return",
"False",
"if",
"fqn",
".",
"count",
"(",
"\".\"",
")",
"!=",
"1",
":",
"return",
"False",
"name",
",",
"namespace_id",
"=",
"fqn",
".",
"split",
"(",
"\".\"",
")",
"if",
"len",
"(",
"name",
")",
"==",
"0",
"or",
"len",
"(",
"namespace_id",
")",
"==",
"0",
":",
"return",
"False",
"if",
"not",
"is_b40",
"(",
"name",
")",
"or",
"\"+\"",
"in",
"name",
"or",
"\".\"",
"in",
"name",
":",
"return",
"False",
"if",
"not",
"is_namespace_valid",
"(",
"namespace_id",
")",
":",
"return",
"False",
"if",
"len",
"(",
"fqn",
")",
">",
"LENGTHS",
"[",
"'blockchain_id_name'",
"]",
":",
"# too long",
"return",
"False",
"return",
"True"
] | Is a fully-qualified name acceptable?
Return True if so
Return False if not
>>> is_name_valid('abcd')
False
>>> is_name_valid('abcd.')
False
>>> is_name_valid('.abcd')
False
>>> is_name_valid('Abcd.abcd')
False
>>> is_name_valid('abcd.abc.d')
False
>>> is_name_valid('abcd.abc+d')
False
>>> is_name_valid('a.b.c')
False
>>> is_name_valid(True)
False
>>> is_name_valid(123)
False
>>> is_name_valid(None)
False
>>> is_name_valid('')
False
>>> is_name_valid('abcdabcdabcdabcdabcdabcdabcdabcda.bcd')
True
>>> is_name_valid('abcdabcdabcdabcdabcdabcdabcdabcdab.bcd')
False
>>> is_name_valid('abcdabcdabcdabcdabcdabcdabcdabcdabc.d')
True
>>> is_name_valid('a+b.c')
False
>>> is_name_valid('a_b.c')
True | [
"Is",
"a",
"fully",
"-",
"qualified",
"name",
"acceptable?",
"Return",
"True",
"if",
"so",
"Return",
"False",
"if",
"not"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L35-L96 |
230,634 | blockstack/blockstack-core | blockstack/lib/scripts.py | is_namespace_valid | def is_namespace_valid( namespace_id ):
"""
Is a namespace ID valid?
>>> is_namespace_valid('abcd')
True
>>> is_namespace_valid('+abcd')
False
>>> is_namespace_valid('abc.def')
False
>>> is_namespace_valid('.abcd')
False
>>> is_namespace_valid('abcdabcdabcdabcdabcd')
False
>>> is_namespace_valid('abcdabcdabcdabcdabc')
True
"""
if not is_b40( namespace_id ) or "+" in namespace_id or namespace_id.count(".") > 0:
return False
if len(namespace_id) == 0 or len(namespace_id) > LENGTHS['blockchain_id_namespace_id']:
return False
return True | python | def is_namespace_valid( namespace_id ):
"""
Is a namespace ID valid?
>>> is_namespace_valid('abcd')
True
>>> is_namespace_valid('+abcd')
False
>>> is_namespace_valid('abc.def')
False
>>> is_namespace_valid('.abcd')
False
>>> is_namespace_valid('abcdabcdabcdabcdabcd')
False
>>> is_namespace_valid('abcdabcdabcdabcdabc')
True
"""
if not is_b40( namespace_id ) or "+" in namespace_id or namespace_id.count(".") > 0:
return False
if len(namespace_id) == 0 or len(namespace_id) > LENGTHS['blockchain_id_namespace_id']:
return False
return True | [
"def",
"is_namespace_valid",
"(",
"namespace_id",
")",
":",
"if",
"not",
"is_b40",
"(",
"namespace_id",
")",
"or",
"\"+\"",
"in",
"namespace_id",
"or",
"namespace_id",
".",
"count",
"(",
"\".\"",
")",
">",
"0",
":",
"return",
"False",
"if",
"len",
"(",
"namespace_id",
")",
"==",
"0",
"or",
"len",
"(",
"namespace_id",
")",
">",
"LENGTHS",
"[",
"'blockchain_id_namespace_id'",
"]",
":",
"return",
"False",
"return",
"True"
] | Is a namespace ID valid?
>>> is_namespace_valid('abcd')
True
>>> is_namespace_valid('+abcd')
False
>>> is_namespace_valid('abc.def')
False
>>> is_namespace_valid('.abcd')
False
>>> is_namespace_valid('abcdabcdabcdabcdabcd')
False
>>> is_namespace_valid('abcdabcdabcdabcdabc')
True | [
"Is",
"a",
"namespace",
"ID",
"valid?"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L99-L122 |
230,635 | blockstack/blockstack-core | blockstack/lib/scripts.py | price_namespace | def price_namespace( namespace_id, block_height, units ):
"""
Calculate the cost of a namespace.
Returns the price on success
Returns None if the namespace is invalid or if the units are invalid
"""
price_table = get_epoch_namespace_prices( block_height, units )
if price_table is None:
return None
if len(namespace_id) >= len(price_table) or len(namespace_id) == 0:
return None
return price_table[len(namespace_id)] | python | def price_namespace( namespace_id, block_height, units ):
"""
Calculate the cost of a namespace.
Returns the price on success
Returns None if the namespace is invalid or if the units are invalid
"""
price_table = get_epoch_namespace_prices( block_height, units )
if price_table is None:
return None
if len(namespace_id) >= len(price_table) or len(namespace_id) == 0:
return None
return price_table[len(namespace_id)] | [
"def",
"price_namespace",
"(",
"namespace_id",
",",
"block_height",
",",
"units",
")",
":",
"price_table",
"=",
"get_epoch_namespace_prices",
"(",
"block_height",
",",
"units",
")",
"if",
"price_table",
"is",
"None",
":",
"return",
"None",
"if",
"len",
"(",
"namespace_id",
")",
">=",
"len",
"(",
"price_table",
")",
"or",
"len",
"(",
"namespace_id",
")",
"==",
"0",
":",
"return",
"None",
"return",
"price_table",
"[",
"len",
"(",
"namespace_id",
")",
"]"
] | Calculate the cost of a namespace.
Returns the price on success
Returns None if the namespace is invalid or if the units are invalid | [
"Calculate",
"the",
"cost",
"of",
"a",
"namespace",
".",
"Returns",
"the",
"price",
"on",
"success",
"Returns",
"None",
"if",
"the",
"namespace",
"is",
"invalid",
"or",
"if",
"the",
"units",
"are",
"invalid"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L311-L324 |
230,636 | blockstack/blockstack-core | blockstack/lib/scripts.py | find_by_opcode | def find_by_opcode( checked_ops, opcode ):
"""
Given all previously-accepted operations in this block,
find the ones that are of a particular opcode.
@opcode can be one opcode, or a list of opcodes
>>> find_by_opcode([{'op': '+'}, {'op': '>'}], 'NAME_UPDATE')
[{'op': '+'}]
>>> find_by_opcode([{'op': '+'}, {'op': '>'}], ['NAME_UPDATE', 'NAME_TRANSFER'])
[{'op': '+'}, {'op': '>'}]
>>> find_by_opcode([{'op': '+'}, {'op': '>'}], ':')
[]
>>> find_by_opcode([], ':')
[]
"""
if type(opcode) != list:
opcode = [opcode]
ret = []
for opdata in checked_ops:
if op_get_opcode_name(opdata['op']) in opcode:
ret.append(opdata)
return ret | python | def find_by_opcode( checked_ops, opcode ):
"""
Given all previously-accepted operations in this block,
find the ones that are of a particular opcode.
@opcode can be one opcode, or a list of opcodes
>>> find_by_opcode([{'op': '+'}, {'op': '>'}], 'NAME_UPDATE')
[{'op': '+'}]
>>> find_by_opcode([{'op': '+'}, {'op': '>'}], ['NAME_UPDATE', 'NAME_TRANSFER'])
[{'op': '+'}, {'op': '>'}]
>>> find_by_opcode([{'op': '+'}, {'op': '>'}], ':')
[]
>>> find_by_opcode([], ':')
[]
"""
if type(opcode) != list:
opcode = [opcode]
ret = []
for opdata in checked_ops:
if op_get_opcode_name(opdata['op']) in opcode:
ret.append(opdata)
return ret | [
"def",
"find_by_opcode",
"(",
"checked_ops",
",",
"opcode",
")",
":",
"if",
"type",
"(",
"opcode",
")",
"!=",
"list",
":",
"opcode",
"=",
"[",
"opcode",
"]",
"ret",
"=",
"[",
"]",
"for",
"opdata",
"in",
"checked_ops",
":",
"if",
"op_get_opcode_name",
"(",
"opdata",
"[",
"'op'",
"]",
")",
"in",
"opcode",
":",
"ret",
".",
"append",
"(",
"opdata",
")",
"return",
"ret"
] | Given all previously-accepted operations in this block,
find the ones that are of a particular opcode.
@opcode can be one opcode, or a list of opcodes
>>> find_by_opcode([{'op': '+'}, {'op': '>'}], 'NAME_UPDATE')
[{'op': '+'}]
>>> find_by_opcode([{'op': '+'}, {'op': '>'}], ['NAME_UPDATE', 'NAME_TRANSFER'])
[{'op': '+'}, {'op': '>'}]
>>> find_by_opcode([{'op': '+'}, {'op': '>'}], ':')
[]
>>> find_by_opcode([], ':')
[] | [
"Given",
"all",
"previously",
"-",
"accepted",
"operations",
"in",
"this",
"block",
"find",
"the",
"ones",
"that",
"are",
"of",
"a",
"particular",
"opcode",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L327-L352 |
230,637 | blockstack/blockstack-core | blockstack/lib/scripts.py | get_public_key_hex_from_tx | def get_public_key_hex_from_tx( inputs, address ):
"""
Given a list of inputs and the address of one of the inputs,
find the public key.
This only works for p2pkh scripts.
We only really need this for NAMESPACE_REVEAL, but we included
it in other transactions' consensus data for legacy reasons that
now have to be supported forever :(
"""
ret = None
for inp in inputs:
input_scriptsig = inp['script']
input_script_code = virtualchain.btc_script_deserialize(input_scriptsig)
if len(input_script_code) == 2:
# signature pubkey
pubkey_candidate = input_script_code[1]
pubkey = None
try:
pubkey = virtualchain.BitcoinPublicKey(pubkey_candidate)
except Exception as e:
traceback.print_exc()
log.warn("Invalid public key {}".format(pubkey_candidate))
continue
if address != pubkey.address():
continue
# success!
return pubkey_candidate
return None | python | def get_public_key_hex_from_tx( inputs, address ):
"""
Given a list of inputs and the address of one of the inputs,
find the public key.
This only works for p2pkh scripts.
We only really need this for NAMESPACE_REVEAL, but we included
it in other transactions' consensus data for legacy reasons that
now have to be supported forever :(
"""
ret = None
for inp in inputs:
input_scriptsig = inp['script']
input_script_code = virtualchain.btc_script_deserialize(input_scriptsig)
if len(input_script_code) == 2:
# signature pubkey
pubkey_candidate = input_script_code[1]
pubkey = None
try:
pubkey = virtualchain.BitcoinPublicKey(pubkey_candidate)
except Exception as e:
traceback.print_exc()
log.warn("Invalid public key {}".format(pubkey_candidate))
continue
if address != pubkey.address():
continue
# success!
return pubkey_candidate
return None | [
"def",
"get_public_key_hex_from_tx",
"(",
"inputs",
",",
"address",
")",
":",
"ret",
"=",
"None",
"for",
"inp",
"in",
"inputs",
":",
"input_scriptsig",
"=",
"inp",
"[",
"'script'",
"]",
"input_script_code",
"=",
"virtualchain",
".",
"btc_script_deserialize",
"(",
"input_scriptsig",
")",
"if",
"len",
"(",
"input_script_code",
")",
"==",
"2",
":",
"# signature pubkey",
"pubkey_candidate",
"=",
"input_script_code",
"[",
"1",
"]",
"pubkey",
"=",
"None",
"try",
":",
"pubkey",
"=",
"virtualchain",
".",
"BitcoinPublicKey",
"(",
"pubkey_candidate",
")",
"except",
"Exception",
"as",
"e",
":",
"traceback",
".",
"print_exc",
"(",
")",
"log",
".",
"warn",
"(",
"\"Invalid public key {}\"",
".",
"format",
"(",
"pubkey_candidate",
")",
")",
"continue",
"if",
"address",
"!=",
"pubkey",
".",
"address",
"(",
")",
":",
"continue",
"# success!",
"return",
"pubkey_candidate",
"return",
"None"
] | Given a list of inputs and the address of one of the inputs,
find the public key.
This only works for p2pkh scripts.
We only really need this for NAMESPACE_REVEAL, but we included
it in other transactions' consensus data for legacy reasons that
now have to be supported forever :( | [
"Given",
"a",
"list",
"of",
"inputs",
"and",
"the",
"address",
"of",
"one",
"of",
"the",
"inputs",
"find",
"the",
"public",
"key",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L355-L388 |
230,638 | blockstack/blockstack-core | blockstack/lib/scripts.py | check_name | def check_name(name):
"""
Verify the name is well-formed
>>> check_name(123)
False
>>> check_name('')
False
>>> check_name('abc')
False
>>> check_name('abc.def')
True
>>> check_name('abc.def.ghi')
False
>>> check_name('abc.d-ef')
True
>>> check_name('abc.d+ef')
False
>>> check_name('.abc')
False
>>> check_name('abc.')
False
>>> check_name('abcdabcdabcdabcdabcdabcdabcdabcdabcdabcdabcd.abcd')
False
>>> check_name('abcdabcdabcdabcdabcdabcdabcdabcdabc.d')
True
"""
if type(name) not in [str, unicode]:
return False
if not is_name_valid(name):
return False
return True | python | def check_name(name):
"""
Verify the name is well-formed
>>> check_name(123)
False
>>> check_name('')
False
>>> check_name('abc')
False
>>> check_name('abc.def')
True
>>> check_name('abc.def.ghi')
False
>>> check_name('abc.d-ef')
True
>>> check_name('abc.d+ef')
False
>>> check_name('.abc')
False
>>> check_name('abc.')
False
>>> check_name('abcdabcdabcdabcdabcdabcdabcdabcdabcdabcdabcd.abcd')
False
>>> check_name('abcdabcdabcdabcdabcdabcdabcdabcdabc.d')
True
"""
if type(name) not in [str, unicode]:
return False
if not is_name_valid(name):
return False
return True | [
"def",
"check_name",
"(",
"name",
")",
":",
"if",
"type",
"(",
"name",
")",
"not",
"in",
"[",
"str",
",",
"unicode",
"]",
":",
"return",
"False",
"if",
"not",
"is_name_valid",
"(",
"name",
")",
":",
"return",
"False",
"return",
"True"
] | Verify the name is well-formed
>>> check_name(123)
False
>>> check_name('')
False
>>> check_name('abc')
False
>>> check_name('abc.def')
True
>>> check_name('abc.def.ghi')
False
>>> check_name('abc.d-ef')
True
>>> check_name('abc.d+ef')
False
>>> check_name('.abc')
False
>>> check_name('abc.')
False
>>> check_name('abcdabcdabcdabcdabcdabcdabcdabcdabcdabcdabcd.abcd')
False
>>> check_name('abcdabcdabcdabcdabcdabcdabcdabcdabc.d')
True | [
"Verify",
"the",
"name",
"is",
"well",
"-",
"formed"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L391-L424 |
230,639 | blockstack/blockstack-core | blockstack/lib/scripts.py | check_namespace | def check_namespace(namespace_id):
"""
Verify that a namespace ID is well-formed
>>> check_namespace(123)
False
>>> check_namespace(None)
False
>>> check_namespace('')
False
>>> check_namespace('abcd')
True
>>> check_namespace('Abcd')
False
>>> check_namespace('a+bcd')
False
>>> check_namespace('.abcd')
False
>>> check_namespace('abcdabcdabcdabcdabcd')
False
>>> check_namespace('abcdabcdabcdabcdabc')
True
"""
if type(namespace_id) not in [str, unicode]:
return False
if not is_namespace_valid(namespace_id):
return False
return True | python | def check_namespace(namespace_id):
"""
Verify that a namespace ID is well-formed
>>> check_namespace(123)
False
>>> check_namespace(None)
False
>>> check_namespace('')
False
>>> check_namespace('abcd')
True
>>> check_namespace('Abcd')
False
>>> check_namespace('a+bcd')
False
>>> check_namespace('.abcd')
False
>>> check_namespace('abcdabcdabcdabcdabcd')
False
>>> check_namespace('abcdabcdabcdabcdabc')
True
"""
if type(namespace_id) not in [str, unicode]:
return False
if not is_namespace_valid(namespace_id):
return False
return True | [
"def",
"check_namespace",
"(",
"namespace_id",
")",
":",
"if",
"type",
"(",
"namespace_id",
")",
"not",
"in",
"[",
"str",
",",
"unicode",
"]",
":",
"return",
"False",
"if",
"not",
"is_namespace_valid",
"(",
"namespace_id",
")",
":",
"return",
"False",
"return",
"True"
] | Verify that a namespace ID is well-formed
>>> check_namespace(123)
False
>>> check_namespace(None)
False
>>> check_namespace('')
False
>>> check_namespace('abcd')
True
>>> check_namespace('Abcd')
False
>>> check_namespace('a+bcd')
False
>>> check_namespace('.abcd')
False
>>> check_namespace('abcdabcdabcdabcdabcd')
False
>>> check_namespace('abcdabcdabcdabcdabc')
True | [
"Verify",
"that",
"a",
"namespace",
"ID",
"is",
"well",
"-",
"formed"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L427-L456 |
230,640 | blockstack/blockstack-core | blockstack/lib/scripts.py | check_token_type | def check_token_type(token_type):
"""
Verify that a token type is well-formed
>>> check_token_type('STACKS')
True
>>> check_token_type('BTC')
False
>>> check_token_type('abcdabcdabcd')
True
>>> check_token_type('abcdabcdabcdabcdabcd')
False
"""
return check_string(token_type, min_length=1, max_length=LENGTHS['namespace_id'], pattern='^{}$|{}'.format(TOKEN_TYPE_STACKS, OP_NAMESPACE_PATTERN)) | python | def check_token_type(token_type):
"""
Verify that a token type is well-formed
>>> check_token_type('STACKS')
True
>>> check_token_type('BTC')
False
>>> check_token_type('abcdabcdabcd')
True
>>> check_token_type('abcdabcdabcdabcdabcd')
False
"""
return check_string(token_type, min_length=1, max_length=LENGTHS['namespace_id'], pattern='^{}$|{}'.format(TOKEN_TYPE_STACKS, OP_NAMESPACE_PATTERN)) | [
"def",
"check_token_type",
"(",
"token_type",
")",
":",
"return",
"check_string",
"(",
"token_type",
",",
"min_length",
"=",
"1",
",",
"max_length",
"=",
"LENGTHS",
"[",
"'namespace_id'",
"]",
",",
"pattern",
"=",
"'^{}$|{}'",
".",
"format",
"(",
"TOKEN_TYPE_STACKS",
",",
"OP_NAMESPACE_PATTERN",
")",
")"
] | Verify that a token type is well-formed
>>> check_token_type('STACKS')
True
>>> check_token_type('BTC')
False
>>> check_token_type('abcdabcdabcd')
True
>>> check_token_type('abcdabcdabcdabcdabcd')
False | [
"Verify",
"that",
"a",
"token",
"type",
"is",
"well",
"-",
"formed"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L459-L472 |
230,641 | blockstack/blockstack-core | blockstack/lib/scripts.py | check_subdomain | def check_subdomain(fqn):
"""
Verify that the given fqn is a subdomain
>>> check_subdomain('a.b.c')
True
>>> check_subdomain(123)
False
>>> check_subdomain('a.b.c.d')
False
>>> check_subdomain('A.b.c')
False
>>> check_subdomain('abcdabcdabcdabcdabcdabcdabcdabcdabcdabcdabcdabcd.a.b')
True
>>> check_subdomain('a.abcdabcdabcdabcdabcdabcdabcdabcdabcdabcd.a')
False
>>> check_subdomain('a.b.cdabcdabcdabcdabcdabcdabcdabcdabcd')
False
>>> check_subdomain('a.b')
False
"""
if type(fqn) not in [str, unicode]:
return False
if not is_subdomain(fqn):
return False
return True | python | def check_subdomain(fqn):
"""
Verify that the given fqn is a subdomain
>>> check_subdomain('a.b.c')
True
>>> check_subdomain(123)
False
>>> check_subdomain('a.b.c.d')
False
>>> check_subdomain('A.b.c')
False
>>> check_subdomain('abcdabcdabcdabcdabcdabcdabcdabcdabcdabcdabcdabcd.a.b')
True
>>> check_subdomain('a.abcdabcdabcdabcdabcdabcdabcdabcdabcdabcd.a')
False
>>> check_subdomain('a.b.cdabcdabcdabcdabcdabcdabcdabcdabcd')
False
>>> check_subdomain('a.b')
False
"""
if type(fqn) not in [str, unicode]:
return False
if not is_subdomain(fqn):
return False
return True | [
"def",
"check_subdomain",
"(",
"fqn",
")",
":",
"if",
"type",
"(",
"fqn",
")",
"not",
"in",
"[",
"str",
",",
"unicode",
"]",
":",
"return",
"False",
"if",
"not",
"is_subdomain",
"(",
"fqn",
")",
":",
"return",
"False",
"return",
"True"
] | Verify that the given fqn is a subdomain
>>> check_subdomain('a.b.c')
True
>>> check_subdomain(123)
False
>>> check_subdomain('a.b.c.d')
False
>>> check_subdomain('A.b.c')
False
>>> check_subdomain('abcdabcdabcdabcdabcdabcdabcdabcdabcdabcdabcdabcd.a.b')
True
>>> check_subdomain('a.abcdabcdabcdabcdabcdabcdabcdabcdabcdabcd.a')
False
>>> check_subdomain('a.b.cdabcdabcdabcdabcdabcdabcdabcdabcd')
False
>>> check_subdomain('a.b')
False | [
"Verify",
"that",
"the",
"given",
"fqn",
"is",
"a",
"subdomain"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L475-L502 |
230,642 | blockstack/blockstack-core | blockstack/lib/scripts.py | check_block | def check_block(block_id):
"""
Verify that a block ID is valid
>>> check_block(FIRST_BLOCK_MAINNET)
True
>>> check_block(FIRST_BLOCK_MAINNET-1)
False
>>> check_block(-1)
False
>>> check_block("abc")
False
>>> check_block(int(1e7) + 1)
False
>>> check_block(int(1e7) - 1)
True
"""
if type(block_id) not in [int, long]:
return False
if BLOCKSTACK_TEST:
if block_id <= 0:
return False
else:
if block_id < FIRST_BLOCK_MAINNET:
return False
if block_id > 1e7:
# 1 million blocks? not in my lifetime
return False
return True | python | def check_block(block_id):
"""
Verify that a block ID is valid
>>> check_block(FIRST_BLOCK_MAINNET)
True
>>> check_block(FIRST_BLOCK_MAINNET-1)
False
>>> check_block(-1)
False
>>> check_block("abc")
False
>>> check_block(int(1e7) + 1)
False
>>> check_block(int(1e7) - 1)
True
"""
if type(block_id) not in [int, long]:
return False
if BLOCKSTACK_TEST:
if block_id <= 0:
return False
else:
if block_id < FIRST_BLOCK_MAINNET:
return False
if block_id > 1e7:
# 1 million blocks? not in my lifetime
return False
return True | [
"def",
"check_block",
"(",
"block_id",
")",
":",
"if",
"type",
"(",
"block_id",
")",
"not",
"in",
"[",
"int",
",",
"long",
"]",
":",
"return",
"False",
"if",
"BLOCKSTACK_TEST",
":",
"if",
"block_id",
"<=",
"0",
":",
"return",
"False",
"else",
":",
"if",
"block_id",
"<",
"FIRST_BLOCK_MAINNET",
":",
"return",
"False",
"if",
"block_id",
">",
"1e7",
":",
"# 1 million blocks? not in my lifetime",
"return",
"False",
"return",
"True"
] | Verify that a block ID is valid
>>> check_block(FIRST_BLOCK_MAINNET)
True
>>> check_block(FIRST_BLOCK_MAINNET-1)
False
>>> check_block(-1)
False
>>> check_block("abc")
False
>>> check_block(int(1e7) + 1)
False
>>> check_block(int(1e7) - 1)
True | [
"Verify",
"that",
"a",
"block",
"ID",
"is",
"valid"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L505-L537 |
230,643 | blockstack/blockstack-core | blockstack/lib/scripts.py | check_offset | def check_offset(offset, max_value=None):
"""
Verify that an offset is valid
>>> check_offset(0)
True
>>> check_offset(-1)
False
>>> check_offset(2, max_value=2)
True
>>> check_offset(0)
True
>>> check_offset(2, max_value=1)
False
>>> check_offset('abc')
False
"""
if type(offset) not in [int, long]:
return False
if offset < 0:
return False
if max_value and offset > max_value:
return False
return True | python | def check_offset(offset, max_value=None):
"""
Verify that an offset is valid
>>> check_offset(0)
True
>>> check_offset(-1)
False
>>> check_offset(2, max_value=2)
True
>>> check_offset(0)
True
>>> check_offset(2, max_value=1)
False
>>> check_offset('abc')
False
"""
if type(offset) not in [int, long]:
return False
if offset < 0:
return False
if max_value and offset > max_value:
return False
return True | [
"def",
"check_offset",
"(",
"offset",
",",
"max_value",
"=",
"None",
")",
":",
"if",
"type",
"(",
"offset",
")",
"not",
"in",
"[",
"int",
",",
"long",
"]",
":",
"return",
"False",
"if",
"offset",
"<",
"0",
":",
"return",
"False",
"if",
"max_value",
"and",
"offset",
">",
"max_value",
":",
"return",
"False",
"return",
"True"
] | Verify that an offset is valid
>>> check_offset(0)
True
>>> check_offset(-1)
False
>>> check_offset(2, max_value=2)
True
>>> check_offset(0)
True
>>> check_offset(2, max_value=1)
False
>>> check_offset('abc')
False | [
"Verify",
"that",
"an",
"offset",
"is",
"valid"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L540-L566 |
230,644 | blockstack/blockstack-core | blockstack/lib/scripts.py | check_string | def check_string(value, min_length=None, max_length=None, pattern=None):
"""
verify that a string has a particular size and conforms
to a particular alphabet
>>> check_string(1)
False
>>> check_string(None)
False
>>> check_string(True)
False
>>> check_string({})
False
>>> check_string([])
False
>>> check_string((1,2))
False
>>> check_string('abc')
True
>>> check_string('')
True
>>> check_string(u'')
True
>>> check_string('abc', min_length=0, max_length=3)
True
>>> check_string('abc', min_length=3, max_length=3)
True
>>> check_string('abc', min_length=4, max_length=5)
False
>>> check_string('abc', min_length=0, max_length=2)
False
>>> check_string('abc', pattern='^abc$')
True
>>> check_string('abc', pattern='^abd$')
False
"""
if type(value) not in [str, unicode]:
return False
if min_length and len(value) < min_length:
return False
if max_length and len(value) > max_length:
return False
if pattern and not re.match(pattern, value):
return False
return True | python | def check_string(value, min_length=None, max_length=None, pattern=None):
"""
verify that a string has a particular size and conforms
to a particular alphabet
>>> check_string(1)
False
>>> check_string(None)
False
>>> check_string(True)
False
>>> check_string({})
False
>>> check_string([])
False
>>> check_string((1,2))
False
>>> check_string('abc')
True
>>> check_string('')
True
>>> check_string(u'')
True
>>> check_string('abc', min_length=0, max_length=3)
True
>>> check_string('abc', min_length=3, max_length=3)
True
>>> check_string('abc', min_length=4, max_length=5)
False
>>> check_string('abc', min_length=0, max_length=2)
False
>>> check_string('abc', pattern='^abc$')
True
>>> check_string('abc', pattern='^abd$')
False
"""
if type(value) not in [str, unicode]:
return False
if min_length and len(value) < min_length:
return False
if max_length and len(value) > max_length:
return False
if pattern and not re.match(pattern, value):
return False
return True | [
"def",
"check_string",
"(",
"value",
",",
"min_length",
"=",
"None",
",",
"max_length",
"=",
"None",
",",
"pattern",
"=",
"None",
")",
":",
"if",
"type",
"(",
"value",
")",
"not",
"in",
"[",
"str",
",",
"unicode",
"]",
":",
"return",
"False",
"if",
"min_length",
"and",
"len",
"(",
"value",
")",
"<",
"min_length",
":",
"return",
"False",
"if",
"max_length",
"and",
"len",
"(",
"value",
")",
">",
"max_length",
":",
"return",
"False",
"if",
"pattern",
"and",
"not",
"re",
".",
"match",
"(",
"pattern",
",",
"value",
")",
":",
"return",
"False",
"return",
"True"
] | verify that a string has a particular size and conforms
to a particular alphabet
>>> check_string(1)
False
>>> check_string(None)
False
>>> check_string(True)
False
>>> check_string({})
False
>>> check_string([])
False
>>> check_string((1,2))
False
>>> check_string('abc')
True
>>> check_string('')
True
>>> check_string(u'')
True
>>> check_string('abc', min_length=0, max_length=3)
True
>>> check_string('abc', min_length=3, max_length=3)
True
>>> check_string('abc', min_length=4, max_length=5)
False
>>> check_string('abc', min_length=0, max_length=2)
False
>>> check_string('abc', pattern='^abc$')
True
>>> check_string('abc', pattern='^abd$')
False | [
"verify",
"that",
"a",
"string",
"has",
"a",
"particular",
"size",
"and",
"conforms",
"to",
"a",
"particular",
"alphabet"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L606-L654 |
230,645 | blockstack/blockstack-core | blockstack/lib/scripts.py | check_address | def check_address(address):
"""
verify that a string is a base58check address
>>> check_address('16EMaNw3pkn3v6f2BgnSSs53zAKH4Q8YJg')
True
>>> check_address('16EMaNw3pkn3v6f2BgnSSs53zAKH4Q8YJh')
False
>>> check_address('mkkJsS22dnDJhD8duFkpGnHNr9uz3JEcWu')
True
>>> check_address('mkkJsS22dnDJhD8duFkpGnHNr9uz3JEcWv')
False
>>> check_address('MD8WooqTKmwromdMQfSNh8gPTPCSf8KaZj')
True
>>> check_address('SSXMcDiCZ7yFSQSUj7mWzmDcdwYhq97p2i')
True
>>> check_address('SSXMcDiCZ7yFSQSUj7mWzmDcdwYhq97p2j')
False
>>> check_address('16SuThrz')
False
>>> check_address('1TGKrgtrQjgoPjoa5BnUZ9Qu')
False
>>> check_address('1LPckRbeTfLjzrfTfnCtP7z2GxFTpZLafXi')
True
"""
if not check_string(address, min_length=26, max_length=35, pattern=OP_ADDRESS_PATTERN):
return False
try:
keylib.b58check_decode(address)
return True
except:
return False | python | def check_address(address):
"""
verify that a string is a base58check address
>>> check_address('16EMaNw3pkn3v6f2BgnSSs53zAKH4Q8YJg')
True
>>> check_address('16EMaNw3pkn3v6f2BgnSSs53zAKH4Q8YJh')
False
>>> check_address('mkkJsS22dnDJhD8duFkpGnHNr9uz3JEcWu')
True
>>> check_address('mkkJsS22dnDJhD8duFkpGnHNr9uz3JEcWv')
False
>>> check_address('MD8WooqTKmwromdMQfSNh8gPTPCSf8KaZj')
True
>>> check_address('SSXMcDiCZ7yFSQSUj7mWzmDcdwYhq97p2i')
True
>>> check_address('SSXMcDiCZ7yFSQSUj7mWzmDcdwYhq97p2j')
False
>>> check_address('16SuThrz')
False
>>> check_address('1TGKrgtrQjgoPjoa5BnUZ9Qu')
False
>>> check_address('1LPckRbeTfLjzrfTfnCtP7z2GxFTpZLafXi')
True
"""
if not check_string(address, min_length=26, max_length=35, pattern=OP_ADDRESS_PATTERN):
return False
try:
keylib.b58check_decode(address)
return True
except:
return False | [
"def",
"check_address",
"(",
"address",
")",
":",
"if",
"not",
"check_string",
"(",
"address",
",",
"min_length",
"=",
"26",
",",
"max_length",
"=",
"35",
",",
"pattern",
"=",
"OP_ADDRESS_PATTERN",
")",
":",
"return",
"False",
"try",
":",
"keylib",
".",
"b58check_decode",
"(",
"address",
")",
"return",
"True",
"except",
":",
"return",
"False"
] | verify that a string is a base58check address
>>> check_address('16EMaNw3pkn3v6f2BgnSSs53zAKH4Q8YJg')
True
>>> check_address('16EMaNw3pkn3v6f2BgnSSs53zAKH4Q8YJh')
False
>>> check_address('mkkJsS22dnDJhD8duFkpGnHNr9uz3JEcWu')
True
>>> check_address('mkkJsS22dnDJhD8duFkpGnHNr9uz3JEcWv')
False
>>> check_address('MD8WooqTKmwromdMQfSNh8gPTPCSf8KaZj')
True
>>> check_address('SSXMcDiCZ7yFSQSUj7mWzmDcdwYhq97p2i')
True
>>> check_address('SSXMcDiCZ7yFSQSUj7mWzmDcdwYhq97p2j')
False
>>> check_address('16SuThrz')
False
>>> check_address('1TGKrgtrQjgoPjoa5BnUZ9Qu')
False
>>> check_address('1LPckRbeTfLjzrfTfnCtP7z2GxFTpZLafXi')
True | [
"verify",
"that",
"a",
"string",
"is",
"a",
"base58check",
"address"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L657-L689 |
230,646 | blockstack/blockstack-core | blockstack/lib/scripts.py | check_account_address | def check_account_address(address):
"""
verify that a string is a valid account address.
Can be a b58-check address, a c32-check address, as well as the string "treasury" or "unallocated" or a string starting with 'not_distributed_'
>>> check_account_address('16EMaNw3pkn3v6f2BgnSSs53zAKH4Q8YJg')
True
>>> check_account_address('16EMaNw3pkn3v6f2BgnSSs53zAKH4Q8YJh')
False
>>> check_account_address('treasury')
True
>>> check_account_address('unallocated')
True
>>> check_account_address('neither')
False
>>> check_account_address('not_distributed')
False
>>> check_account_address('not_distributed_')
False
>>> check_account_address('not_distributed_asdfasdfasdfasdf')
True
>>> check_account_address('SP2J6ZY48GV1EZ5V2V5RB9MP66SW86PYKKNRV9EJ7')
True
>>> check_account_address('SP2J6ZY48GV1EZ5V2V5RB9MP66SW86PYKKNRV9EJ8')
False
"""
if address == 'treasury' or address == 'unallocated':
return True
if address.startswith('not_distributed_') and len(address) > len('not_distributed_'):
return True
if re.match(OP_C32CHECK_PATTERN, address):
try:
c32addressDecode(address)
return True
except:
pass
return check_address(address) | python | def check_account_address(address):
"""
verify that a string is a valid account address.
Can be a b58-check address, a c32-check address, as well as the string "treasury" or "unallocated" or a string starting with 'not_distributed_'
>>> check_account_address('16EMaNw3pkn3v6f2BgnSSs53zAKH4Q8YJg')
True
>>> check_account_address('16EMaNw3pkn3v6f2BgnSSs53zAKH4Q8YJh')
False
>>> check_account_address('treasury')
True
>>> check_account_address('unallocated')
True
>>> check_account_address('neither')
False
>>> check_account_address('not_distributed')
False
>>> check_account_address('not_distributed_')
False
>>> check_account_address('not_distributed_asdfasdfasdfasdf')
True
>>> check_account_address('SP2J6ZY48GV1EZ5V2V5RB9MP66SW86PYKKNRV9EJ7')
True
>>> check_account_address('SP2J6ZY48GV1EZ5V2V5RB9MP66SW86PYKKNRV9EJ8')
False
"""
if address == 'treasury' or address == 'unallocated':
return True
if address.startswith('not_distributed_') and len(address) > len('not_distributed_'):
return True
if re.match(OP_C32CHECK_PATTERN, address):
try:
c32addressDecode(address)
return True
except:
pass
return check_address(address) | [
"def",
"check_account_address",
"(",
"address",
")",
":",
"if",
"address",
"==",
"'treasury'",
"or",
"address",
"==",
"'unallocated'",
":",
"return",
"True",
"if",
"address",
".",
"startswith",
"(",
"'not_distributed_'",
")",
"and",
"len",
"(",
"address",
")",
">",
"len",
"(",
"'not_distributed_'",
")",
":",
"return",
"True",
"if",
"re",
".",
"match",
"(",
"OP_C32CHECK_PATTERN",
",",
"address",
")",
":",
"try",
":",
"c32addressDecode",
"(",
"address",
")",
"return",
"True",
"except",
":",
"pass",
"return",
"check_address",
"(",
"address",
")"
] | verify that a string is a valid account address.
Can be a b58-check address, a c32-check address, as well as the string "treasury" or "unallocated" or a string starting with 'not_distributed_'
>>> check_account_address('16EMaNw3pkn3v6f2BgnSSs53zAKH4Q8YJg')
True
>>> check_account_address('16EMaNw3pkn3v6f2BgnSSs53zAKH4Q8YJh')
False
>>> check_account_address('treasury')
True
>>> check_account_address('unallocated')
True
>>> check_account_address('neither')
False
>>> check_account_address('not_distributed')
False
>>> check_account_address('not_distributed_')
False
>>> check_account_address('not_distributed_asdfasdfasdfasdf')
True
>>> check_account_address('SP2J6ZY48GV1EZ5V2V5RB9MP66SW86PYKKNRV9EJ7')
True
>>> check_account_address('SP2J6ZY48GV1EZ5V2V5RB9MP66SW86PYKKNRV9EJ8')
False | [
"verify",
"that",
"a",
"string",
"is",
"a",
"valid",
"account",
"address",
".",
"Can",
"be",
"a",
"b58",
"-",
"check",
"address",
"a",
"c32",
"-",
"check",
"address",
"as",
"well",
"as",
"the",
"string",
"treasury",
"or",
"unallocated",
"or",
"a",
"string",
"starting",
"with",
"not_distributed_"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L692-L731 |
230,647 | blockstack/blockstack-core | blockstack/lib/scripts.py | check_tx_output_types | def check_tx_output_types(outputs, block_height):
"""
Verify that the list of transaction outputs are acceptable
"""
# for now, we do not allow nonstandard outputs (all outputs must be p2pkh or p2sh outputs)
# this excludes bech32 outputs, for example.
supported_output_types = get_epoch_btc_script_types(block_height)
for out in outputs:
out_type = virtualchain.btc_script_classify(out['script'])
if out_type not in supported_output_types:
log.warning('Unsupported output type {} ({})'.format(out_type, out['script']))
return False
return True | python | def check_tx_output_types(outputs, block_height):
"""
Verify that the list of transaction outputs are acceptable
"""
# for now, we do not allow nonstandard outputs (all outputs must be p2pkh or p2sh outputs)
# this excludes bech32 outputs, for example.
supported_output_types = get_epoch_btc_script_types(block_height)
for out in outputs:
out_type = virtualchain.btc_script_classify(out['script'])
if out_type not in supported_output_types:
log.warning('Unsupported output type {} ({})'.format(out_type, out['script']))
return False
return True | [
"def",
"check_tx_output_types",
"(",
"outputs",
",",
"block_height",
")",
":",
"# for now, we do not allow nonstandard outputs (all outputs must be p2pkh or p2sh outputs)",
"# this excludes bech32 outputs, for example.",
"supported_output_types",
"=",
"get_epoch_btc_script_types",
"(",
"block_height",
")",
"for",
"out",
"in",
"outputs",
":",
"out_type",
"=",
"virtualchain",
".",
"btc_script_classify",
"(",
"out",
"[",
"'script'",
"]",
")",
"if",
"out_type",
"not",
"in",
"supported_output_types",
":",
"log",
".",
"warning",
"(",
"'Unsupported output type {} ({})'",
".",
"format",
"(",
"out_type",
",",
"out",
"[",
"'script'",
"]",
")",
")",
"return",
"False",
"return",
"True"
] | Verify that the list of transaction outputs are acceptable | [
"Verify",
"that",
"the",
"list",
"of",
"transaction",
"outputs",
"are",
"acceptable"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L734-L747 |
230,648 | blockstack/blockstack-core | blockstack/lib/scripts.py | address_as_b58 | def address_as_b58(addr):
"""
Given a b58check or c32check address,
return the b58check encoding
"""
if is_c32_address(addr):
return c32ToB58(addr)
else:
if check_address(addr):
return addr
else:
raise ValueError('Address {} is not b58 or c32'.format(addr)) | python | def address_as_b58(addr):
"""
Given a b58check or c32check address,
return the b58check encoding
"""
if is_c32_address(addr):
return c32ToB58(addr)
else:
if check_address(addr):
return addr
else:
raise ValueError('Address {} is not b58 or c32'.format(addr)) | [
"def",
"address_as_b58",
"(",
"addr",
")",
":",
"if",
"is_c32_address",
"(",
"addr",
")",
":",
"return",
"c32ToB58",
"(",
"addr",
")",
"else",
":",
"if",
"check_address",
"(",
"addr",
")",
":",
"return",
"addr",
"else",
":",
"raise",
"ValueError",
"(",
"'Address {} is not b58 or c32'",
".",
"format",
"(",
"addr",
")",
")"
] | Given a b58check or c32check address,
return the b58check encoding | [
"Given",
"a",
"b58check",
"or",
"c32check",
"address",
"return",
"the",
"b58check",
"encoding"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/scripts.py#L770-L782 |
230,649 | blockstack/blockstack-core | blockstack/lib/subdomains.py | verify | def verify(address, plaintext, scriptSigb64):
"""
Verify that a given plaintext is signed by the given scriptSig, given the address
"""
assert isinstance(address, str)
assert isinstance(scriptSigb64, str)
scriptSig = base64.b64decode(scriptSigb64)
hash_hex = hashlib.sha256(plaintext).hexdigest()
vb = keylib.b58check.b58check_version_byte(address)
if vb == bitcoin_blockchain.version_byte:
return verify_singlesig(address, hash_hex, scriptSig)
elif vb == bitcoin_blockchain.multisig_version_byte:
return verify_multisig(address, hash_hex, scriptSig)
else:
log.warning("Unrecognized address version byte {}".format(vb))
raise NotImplementedError("Addresses must be single-sig (version-byte = 0) or multi-sig (version-byte = 5)") | python | def verify(address, plaintext, scriptSigb64):
"""
Verify that a given plaintext is signed by the given scriptSig, given the address
"""
assert isinstance(address, str)
assert isinstance(scriptSigb64, str)
scriptSig = base64.b64decode(scriptSigb64)
hash_hex = hashlib.sha256(plaintext).hexdigest()
vb = keylib.b58check.b58check_version_byte(address)
if vb == bitcoin_blockchain.version_byte:
return verify_singlesig(address, hash_hex, scriptSig)
elif vb == bitcoin_blockchain.multisig_version_byte:
return verify_multisig(address, hash_hex, scriptSig)
else:
log.warning("Unrecognized address version byte {}".format(vb))
raise NotImplementedError("Addresses must be single-sig (version-byte = 0) or multi-sig (version-byte = 5)") | [
"def",
"verify",
"(",
"address",
",",
"plaintext",
",",
"scriptSigb64",
")",
":",
"assert",
"isinstance",
"(",
"address",
",",
"str",
")",
"assert",
"isinstance",
"(",
"scriptSigb64",
",",
"str",
")",
"scriptSig",
"=",
"base64",
".",
"b64decode",
"(",
"scriptSigb64",
")",
"hash_hex",
"=",
"hashlib",
".",
"sha256",
"(",
"plaintext",
")",
".",
"hexdigest",
"(",
")",
"vb",
"=",
"keylib",
".",
"b58check",
".",
"b58check_version_byte",
"(",
"address",
")",
"if",
"vb",
"==",
"bitcoin_blockchain",
".",
"version_byte",
":",
"return",
"verify_singlesig",
"(",
"address",
",",
"hash_hex",
",",
"scriptSig",
")",
"elif",
"vb",
"==",
"bitcoin_blockchain",
".",
"multisig_version_byte",
":",
"return",
"verify_multisig",
"(",
"address",
",",
"hash_hex",
",",
"scriptSig",
")",
"else",
":",
"log",
".",
"warning",
"(",
"\"Unrecognized address version byte {}\"",
".",
"format",
"(",
"vb",
")",
")",
"raise",
"NotImplementedError",
"(",
"\"Addresses must be single-sig (version-byte = 0) or multi-sig (version-byte = 5)\"",
")"
] | Verify that a given plaintext is signed by the given scriptSig, given the address | [
"Verify",
"that",
"a",
"given",
"plaintext",
"is",
"signed",
"by",
"the",
"given",
"scriptSig",
"given",
"the",
"address"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1727-L1745 |
230,650 | blockstack/blockstack-core | blockstack/lib/subdomains.py | verify_singlesig | def verify_singlesig(address, hash_hex, scriptSig):
"""
Verify that a p2pkh address is signed by the given pay-to-pubkey-hash scriptsig
"""
try:
sighex, pubkey_hex = virtualchain.btc_script_deserialize(scriptSig)
except:
log.warn("Wrong signature structure for {}".format(address))
return False
# verify pubkey_hex corresponds to address
if virtualchain.address_reencode(keylib.public_key_to_address(pubkey_hex)) != virtualchain.address_reencode(address):
log.warn(("Address {} does not match signature script {}".format(address, scriptSig.encode('hex'))))
return False
sig64 = base64.b64encode(binascii.unhexlify(sighex))
return virtualchain.ecdsalib.verify_digest(hash_hex, pubkey_hex, sig64) | python | def verify_singlesig(address, hash_hex, scriptSig):
"""
Verify that a p2pkh address is signed by the given pay-to-pubkey-hash scriptsig
"""
try:
sighex, pubkey_hex = virtualchain.btc_script_deserialize(scriptSig)
except:
log.warn("Wrong signature structure for {}".format(address))
return False
# verify pubkey_hex corresponds to address
if virtualchain.address_reencode(keylib.public_key_to_address(pubkey_hex)) != virtualchain.address_reencode(address):
log.warn(("Address {} does not match signature script {}".format(address, scriptSig.encode('hex'))))
return False
sig64 = base64.b64encode(binascii.unhexlify(sighex))
return virtualchain.ecdsalib.verify_digest(hash_hex, pubkey_hex, sig64) | [
"def",
"verify_singlesig",
"(",
"address",
",",
"hash_hex",
",",
"scriptSig",
")",
":",
"try",
":",
"sighex",
",",
"pubkey_hex",
"=",
"virtualchain",
".",
"btc_script_deserialize",
"(",
"scriptSig",
")",
"except",
":",
"log",
".",
"warn",
"(",
"\"Wrong signature structure for {}\"",
".",
"format",
"(",
"address",
")",
")",
"return",
"False",
"# verify pubkey_hex corresponds to address",
"if",
"virtualchain",
".",
"address_reencode",
"(",
"keylib",
".",
"public_key_to_address",
"(",
"pubkey_hex",
")",
")",
"!=",
"virtualchain",
".",
"address_reencode",
"(",
"address",
")",
":",
"log",
".",
"warn",
"(",
"(",
"\"Address {} does not match signature script {}\"",
".",
"format",
"(",
"address",
",",
"scriptSig",
".",
"encode",
"(",
"'hex'",
")",
")",
")",
")",
"return",
"False",
"sig64",
"=",
"base64",
".",
"b64encode",
"(",
"binascii",
".",
"unhexlify",
"(",
"sighex",
")",
")",
"return",
"virtualchain",
".",
"ecdsalib",
".",
"verify_digest",
"(",
"hash_hex",
",",
"pubkey_hex",
",",
"sig64",
")"
] | Verify that a p2pkh address is signed by the given pay-to-pubkey-hash scriptsig | [
"Verify",
"that",
"a",
"p2pkh",
"address",
"is",
"signed",
"by",
"the",
"given",
"pay",
"-",
"to",
"-",
"pubkey",
"-",
"hash",
"scriptsig"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1748-L1764 |
230,651 | blockstack/blockstack-core | blockstack/lib/subdomains.py | verify_multisig | def verify_multisig(address, hash_hex, scriptSig):
"""
verify that a p2sh address is signed by the given scriptsig
"""
script_parts = virtualchain.btc_script_deserialize(scriptSig)
if len(script_parts) < 2:
log.warn("Verfiying multisig failed, couldn't grab script parts")
return False
redeem_script = script_parts[-1]
script_sigs = script_parts[1:-1]
if virtualchain.address_reencode(virtualchain.btc_make_p2sh_address(redeem_script)) != virtualchain.address_reencode(address):
log.warn(("Address {} does not match redeem script {}".format(address, redeem_script)))
return False
m, pubk_hexes = virtualchain.parse_multisig_redeemscript(redeem_script)
if len(script_sigs) != m:
log.warn("Failed to validate multi-sig, not correct number of signatures: have {}, require {}".format(
len(script_sigs), m))
return False
cur_pubk = 0
for cur_sig in script_sigs:
sig64 = base64.b64encode(binascii.unhexlify(cur_sig))
sig_passed = False
while not sig_passed:
if cur_pubk >= len(pubk_hexes):
log.warn("Failed to validate multi-signature, ran out of public keys to check")
return False
sig_passed = virtualchain.ecdsalib.verify_digest(hash_hex, pubk_hexes[cur_pubk], sig64)
cur_pubk += 1
return True | python | def verify_multisig(address, hash_hex, scriptSig):
"""
verify that a p2sh address is signed by the given scriptsig
"""
script_parts = virtualchain.btc_script_deserialize(scriptSig)
if len(script_parts) < 2:
log.warn("Verfiying multisig failed, couldn't grab script parts")
return False
redeem_script = script_parts[-1]
script_sigs = script_parts[1:-1]
if virtualchain.address_reencode(virtualchain.btc_make_p2sh_address(redeem_script)) != virtualchain.address_reencode(address):
log.warn(("Address {} does not match redeem script {}".format(address, redeem_script)))
return False
m, pubk_hexes = virtualchain.parse_multisig_redeemscript(redeem_script)
if len(script_sigs) != m:
log.warn("Failed to validate multi-sig, not correct number of signatures: have {}, require {}".format(
len(script_sigs), m))
return False
cur_pubk = 0
for cur_sig in script_sigs:
sig64 = base64.b64encode(binascii.unhexlify(cur_sig))
sig_passed = False
while not sig_passed:
if cur_pubk >= len(pubk_hexes):
log.warn("Failed to validate multi-signature, ran out of public keys to check")
return False
sig_passed = virtualchain.ecdsalib.verify_digest(hash_hex, pubk_hexes[cur_pubk], sig64)
cur_pubk += 1
return True | [
"def",
"verify_multisig",
"(",
"address",
",",
"hash_hex",
",",
"scriptSig",
")",
":",
"script_parts",
"=",
"virtualchain",
".",
"btc_script_deserialize",
"(",
"scriptSig",
")",
"if",
"len",
"(",
"script_parts",
")",
"<",
"2",
":",
"log",
".",
"warn",
"(",
"\"Verfiying multisig failed, couldn't grab script parts\"",
")",
"return",
"False",
"redeem_script",
"=",
"script_parts",
"[",
"-",
"1",
"]",
"script_sigs",
"=",
"script_parts",
"[",
"1",
":",
"-",
"1",
"]",
"if",
"virtualchain",
".",
"address_reencode",
"(",
"virtualchain",
".",
"btc_make_p2sh_address",
"(",
"redeem_script",
")",
")",
"!=",
"virtualchain",
".",
"address_reencode",
"(",
"address",
")",
":",
"log",
".",
"warn",
"(",
"(",
"\"Address {} does not match redeem script {}\"",
".",
"format",
"(",
"address",
",",
"redeem_script",
")",
")",
")",
"return",
"False",
"m",
",",
"pubk_hexes",
"=",
"virtualchain",
".",
"parse_multisig_redeemscript",
"(",
"redeem_script",
")",
"if",
"len",
"(",
"script_sigs",
")",
"!=",
"m",
":",
"log",
".",
"warn",
"(",
"\"Failed to validate multi-sig, not correct number of signatures: have {}, require {}\"",
".",
"format",
"(",
"len",
"(",
"script_sigs",
")",
",",
"m",
")",
")",
"return",
"False",
"cur_pubk",
"=",
"0",
"for",
"cur_sig",
"in",
"script_sigs",
":",
"sig64",
"=",
"base64",
".",
"b64encode",
"(",
"binascii",
".",
"unhexlify",
"(",
"cur_sig",
")",
")",
"sig_passed",
"=",
"False",
"while",
"not",
"sig_passed",
":",
"if",
"cur_pubk",
">=",
"len",
"(",
"pubk_hexes",
")",
":",
"log",
".",
"warn",
"(",
"\"Failed to validate multi-signature, ran out of public keys to check\"",
")",
"return",
"False",
"sig_passed",
"=",
"virtualchain",
".",
"ecdsalib",
".",
"verify_digest",
"(",
"hash_hex",
",",
"pubk_hexes",
"[",
"cur_pubk",
"]",
",",
"sig64",
")",
"cur_pubk",
"+=",
"1",
"return",
"True"
] | verify that a p2sh address is signed by the given scriptsig | [
"verify",
"that",
"a",
"p2sh",
"address",
"is",
"signed",
"by",
"the",
"given",
"scriptsig"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1767-L1800 |
230,652 | blockstack/blockstack-core | blockstack/lib/subdomains.py | is_subdomain_missing_zonefiles_record | def is_subdomain_missing_zonefiles_record(rec):
"""
Does a given parsed zone file TXT record encode a missing-zonefile vector?
Return True if so
Return False if not
"""
if rec['name'] != SUBDOMAIN_TXT_RR_MISSING:
return False
txt_entry = rec['txt']
if isinstance(txt_entry, list):
return False
missing = txt_entry.split(',')
try:
for m in missing:
m = int(m)
except ValueError:
return False
return True | python | def is_subdomain_missing_zonefiles_record(rec):
"""
Does a given parsed zone file TXT record encode a missing-zonefile vector?
Return True if so
Return False if not
"""
if rec['name'] != SUBDOMAIN_TXT_RR_MISSING:
return False
txt_entry = rec['txt']
if isinstance(txt_entry, list):
return False
missing = txt_entry.split(',')
try:
for m in missing:
m = int(m)
except ValueError:
return False
return True | [
"def",
"is_subdomain_missing_zonefiles_record",
"(",
"rec",
")",
":",
"if",
"rec",
"[",
"'name'",
"]",
"!=",
"SUBDOMAIN_TXT_RR_MISSING",
":",
"return",
"False",
"txt_entry",
"=",
"rec",
"[",
"'txt'",
"]",
"if",
"isinstance",
"(",
"txt_entry",
",",
"list",
")",
":",
"return",
"False",
"missing",
"=",
"txt_entry",
".",
"split",
"(",
"','",
")",
"try",
":",
"for",
"m",
"in",
"missing",
":",
"m",
"=",
"int",
"(",
"m",
")",
"except",
"ValueError",
":",
"return",
"False",
"return",
"True"
] | Does a given parsed zone file TXT record encode a missing-zonefile vector?
Return True if so
Return False if not | [
"Does",
"a",
"given",
"parsed",
"zone",
"file",
"TXT",
"record",
"encode",
"a",
"missing",
"-",
"zonefile",
"vector?",
"Return",
"True",
"if",
"so",
"Return",
"False",
"if",
"not"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1810-L1830 |
230,653 | blockstack/blockstack-core | blockstack/lib/subdomains.py | is_subdomain_record | def is_subdomain_record(rec):
"""
Does a given parsed zone file TXT record (@rec) encode a subdomain?
Return True if so
Return False if not
"""
txt_entry = rec['txt']
if not isinstance(txt_entry, list):
return False
has_parts_entry = False
has_pk_entry = False
has_seqn_entry = False
for entry in txt_entry:
if entry.startswith(SUBDOMAIN_ZF_PARTS + "="):
has_parts_entry = True
if entry.startswith(SUBDOMAIN_PUBKEY + "="):
has_pk_entry = True
if entry.startswith(SUBDOMAIN_N + "="):
has_seqn_entry = True
return (has_parts_entry and has_pk_entry and has_seqn_entry) | python | def is_subdomain_record(rec):
"""
Does a given parsed zone file TXT record (@rec) encode a subdomain?
Return True if so
Return False if not
"""
txt_entry = rec['txt']
if not isinstance(txt_entry, list):
return False
has_parts_entry = False
has_pk_entry = False
has_seqn_entry = False
for entry in txt_entry:
if entry.startswith(SUBDOMAIN_ZF_PARTS + "="):
has_parts_entry = True
if entry.startswith(SUBDOMAIN_PUBKEY + "="):
has_pk_entry = True
if entry.startswith(SUBDOMAIN_N + "="):
has_seqn_entry = True
return (has_parts_entry and has_pk_entry and has_seqn_entry) | [
"def",
"is_subdomain_record",
"(",
"rec",
")",
":",
"txt_entry",
"=",
"rec",
"[",
"'txt'",
"]",
"if",
"not",
"isinstance",
"(",
"txt_entry",
",",
"list",
")",
":",
"return",
"False",
"has_parts_entry",
"=",
"False",
"has_pk_entry",
"=",
"False",
"has_seqn_entry",
"=",
"False",
"for",
"entry",
"in",
"txt_entry",
":",
"if",
"entry",
".",
"startswith",
"(",
"SUBDOMAIN_ZF_PARTS",
"+",
"\"=\"",
")",
":",
"has_parts_entry",
"=",
"True",
"if",
"entry",
".",
"startswith",
"(",
"SUBDOMAIN_PUBKEY",
"+",
"\"=\"",
")",
":",
"has_pk_entry",
"=",
"True",
"if",
"entry",
".",
"startswith",
"(",
"SUBDOMAIN_N",
"+",
"\"=\"",
")",
":",
"has_seqn_entry",
"=",
"True",
"return",
"(",
"has_parts_entry",
"and",
"has_pk_entry",
"and",
"has_seqn_entry",
")"
] | Does a given parsed zone file TXT record (@rec) encode a subdomain?
Return True if so
Return False if not | [
"Does",
"a",
"given",
"parsed",
"zone",
"file",
"TXT",
"record",
"("
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1833-L1854 |
230,654 | blockstack/blockstack-core | blockstack/lib/subdomains.py | get_subdomain_info | def get_subdomain_info(fqn, db_path=None, atlasdb_path=None, zonefiles_dir=None, check_pending=False, include_did=False):
"""
Static method for getting the state of a subdomain, given its fully-qualified name.
Return the subdomain record on success.
Return None if not found.
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
log.warn("Subdomain support is disabled")
return None
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
if atlasdb_path is None:
atlasdb_path = opts['atlasdb_path']
db = SubdomainDB(db_path, zonefiles_dir)
try:
subrec = db.get_subdomain_entry(fqn)
except SubdomainNotFound:
log.warn("No such subdomain: {}".format(fqn))
return None
if check_pending:
# make sure that all of the zone files between this subdomain's
# domain's creation and this subdomain's zone file index are present,
# minus the ones that are allowed to be missing.
subrec.pending = db.subdomain_check_pending(subrec, atlasdb_path)
if include_did:
# include the DID
subrec.did_info = db.get_subdomain_DID_info(fqn)
return subrec | python | def get_subdomain_info(fqn, db_path=None, atlasdb_path=None, zonefiles_dir=None, check_pending=False, include_did=False):
"""
Static method for getting the state of a subdomain, given its fully-qualified name.
Return the subdomain record on success.
Return None if not found.
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
log.warn("Subdomain support is disabled")
return None
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
if atlasdb_path is None:
atlasdb_path = opts['atlasdb_path']
db = SubdomainDB(db_path, zonefiles_dir)
try:
subrec = db.get_subdomain_entry(fqn)
except SubdomainNotFound:
log.warn("No such subdomain: {}".format(fqn))
return None
if check_pending:
# make sure that all of the zone files between this subdomain's
# domain's creation and this subdomain's zone file index are present,
# minus the ones that are allowed to be missing.
subrec.pending = db.subdomain_check_pending(subrec, atlasdb_path)
if include_did:
# include the DID
subrec.did_info = db.get_subdomain_DID_info(fqn)
return subrec | [
"def",
"get_subdomain_info",
"(",
"fqn",
",",
"db_path",
"=",
"None",
",",
"atlasdb_path",
"=",
"None",
",",
"zonefiles_dir",
"=",
"None",
",",
"check_pending",
"=",
"False",
",",
"include_did",
"=",
"False",
")",
":",
"opts",
"=",
"get_blockstack_opts",
"(",
")",
"if",
"not",
"is_subdomains_enabled",
"(",
"opts",
")",
":",
"log",
".",
"warn",
"(",
"\"Subdomain support is disabled\"",
")",
"return",
"None",
"if",
"db_path",
"is",
"None",
":",
"db_path",
"=",
"opts",
"[",
"'subdomaindb_path'",
"]",
"if",
"zonefiles_dir",
"is",
"None",
":",
"zonefiles_dir",
"=",
"opts",
"[",
"'zonefiles'",
"]",
"if",
"atlasdb_path",
"is",
"None",
":",
"atlasdb_path",
"=",
"opts",
"[",
"'atlasdb_path'",
"]",
"db",
"=",
"SubdomainDB",
"(",
"db_path",
",",
"zonefiles_dir",
")",
"try",
":",
"subrec",
"=",
"db",
".",
"get_subdomain_entry",
"(",
"fqn",
")",
"except",
"SubdomainNotFound",
":",
"log",
".",
"warn",
"(",
"\"No such subdomain: {}\"",
".",
"format",
"(",
"fqn",
")",
")",
"return",
"None",
"if",
"check_pending",
":",
"# make sure that all of the zone files between this subdomain's",
"# domain's creation and this subdomain's zone file index are present,",
"# minus the ones that are allowed to be missing.",
"subrec",
".",
"pending",
"=",
"db",
".",
"subdomain_check_pending",
"(",
"subrec",
",",
"atlasdb_path",
")",
"if",
"include_did",
":",
"# include the DID ",
"subrec",
".",
"did_info",
"=",
"db",
".",
"get_subdomain_DID_info",
"(",
"fqn",
")",
"return",
"subrec"
] | Static method for getting the state of a subdomain, given its fully-qualified name.
Return the subdomain record on success.
Return None if not found. | [
"Static",
"method",
"for",
"getting",
"the",
"state",
"of",
"a",
"subdomain",
"given",
"its",
"fully",
"-",
"qualified",
"name",
".",
"Return",
"the",
"subdomain",
"record",
"on",
"success",
".",
"Return",
"None",
"if",
"not",
"found",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1857-L1894 |
230,655 | blockstack/blockstack-core | blockstack/lib/subdomains.py | get_subdomain_resolver | def get_subdomain_resolver(name, db_path=None, zonefiles_dir=None):
"""
Static method for determining the last-known resolver for a domain name.
Returns the resolver URL on success
Returns None on error
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
log.warn("Subdomain support is disabled")
return None
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
resolver_url = db.get_domain_resolver(name)
return resolver_url | python | def get_subdomain_resolver(name, db_path=None, zonefiles_dir=None):
"""
Static method for determining the last-known resolver for a domain name.
Returns the resolver URL on success
Returns None on error
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
log.warn("Subdomain support is disabled")
return None
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
resolver_url = db.get_domain_resolver(name)
return resolver_url | [
"def",
"get_subdomain_resolver",
"(",
"name",
",",
"db_path",
"=",
"None",
",",
"zonefiles_dir",
"=",
"None",
")",
":",
"opts",
"=",
"get_blockstack_opts",
"(",
")",
"if",
"not",
"is_subdomains_enabled",
"(",
"opts",
")",
":",
"log",
".",
"warn",
"(",
"\"Subdomain support is disabled\"",
")",
"return",
"None",
"if",
"db_path",
"is",
"None",
":",
"db_path",
"=",
"opts",
"[",
"'subdomaindb_path'",
"]",
"if",
"zonefiles_dir",
"is",
"None",
":",
"zonefiles_dir",
"=",
"opts",
"[",
"'zonefiles'",
"]",
"db",
"=",
"SubdomainDB",
"(",
"db_path",
",",
"zonefiles_dir",
")",
"resolver_url",
"=",
"db",
".",
"get_domain_resolver",
"(",
"name",
")",
"return",
"resolver_url"
] | Static method for determining the last-known resolver for a domain name.
Returns the resolver URL on success
Returns None on error | [
"Static",
"method",
"for",
"determining",
"the",
"last",
"-",
"known",
"resolver",
"for",
"a",
"domain",
"name",
".",
"Returns",
"the",
"resolver",
"URL",
"on",
"success",
"Returns",
"None",
"on",
"error"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1897-L1917 |
230,656 | blockstack/blockstack-core | blockstack/lib/subdomains.py | get_subdomains_count | def get_subdomains_count(db_path=None, zonefiles_dir=None):
"""
Static method for getting count of all subdomains
Return number of subdomains on success
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
log.warn("Subdomain support is disabled")
return None
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
return db.get_subdomains_count() | python | def get_subdomains_count(db_path=None, zonefiles_dir=None):
"""
Static method for getting count of all subdomains
Return number of subdomains on success
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
log.warn("Subdomain support is disabled")
return None
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
return db.get_subdomains_count() | [
"def",
"get_subdomains_count",
"(",
"db_path",
"=",
"None",
",",
"zonefiles_dir",
"=",
"None",
")",
":",
"opts",
"=",
"get_blockstack_opts",
"(",
")",
"if",
"not",
"is_subdomains_enabled",
"(",
"opts",
")",
":",
"log",
".",
"warn",
"(",
"\"Subdomain support is disabled\"",
")",
"return",
"None",
"if",
"db_path",
"is",
"None",
":",
"db_path",
"=",
"opts",
"[",
"'subdomaindb_path'",
"]",
"if",
"zonefiles_dir",
"is",
"None",
":",
"zonefiles_dir",
"=",
"opts",
"[",
"'zonefiles'",
"]",
"db",
"=",
"SubdomainDB",
"(",
"db_path",
",",
"zonefiles_dir",
")",
"return",
"db",
".",
"get_subdomains_count",
"(",
")"
] | Static method for getting count of all subdomains
Return number of subdomains on success | [
"Static",
"method",
"for",
"getting",
"count",
"of",
"all",
"subdomains",
"Return",
"number",
"of",
"subdomains",
"on",
"success"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1920-L1937 |
230,657 | blockstack/blockstack-core | blockstack/lib/subdomains.py | get_subdomain_DID_info | def get_subdomain_DID_info(fqn, db_path=None, zonefiles_dir=None):
"""
Get a subdomain's DID info.
Return None if not found
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
log.warn("Subdomain support is disabled")
return None
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
try:
subrec = db.get_subdomain_entry(fqn)
except SubdomainNotFound:
log.warn("No such subdomain: {}".format(fqn))
return None
try:
return db.get_subdomain_DID_info(fqn)
except SubdomainNotFound:
return None | python | def get_subdomain_DID_info(fqn, db_path=None, zonefiles_dir=None):
"""
Get a subdomain's DID info.
Return None if not found
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
log.warn("Subdomain support is disabled")
return None
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
try:
subrec = db.get_subdomain_entry(fqn)
except SubdomainNotFound:
log.warn("No such subdomain: {}".format(fqn))
return None
try:
return db.get_subdomain_DID_info(fqn)
except SubdomainNotFound:
return None | [
"def",
"get_subdomain_DID_info",
"(",
"fqn",
",",
"db_path",
"=",
"None",
",",
"zonefiles_dir",
"=",
"None",
")",
":",
"opts",
"=",
"get_blockstack_opts",
"(",
")",
"if",
"not",
"is_subdomains_enabled",
"(",
"opts",
")",
":",
"log",
".",
"warn",
"(",
"\"Subdomain support is disabled\"",
")",
"return",
"None",
"if",
"db_path",
"is",
"None",
":",
"db_path",
"=",
"opts",
"[",
"'subdomaindb_path'",
"]",
"if",
"zonefiles_dir",
"is",
"None",
":",
"zonefiles_dir",
"=",
"opts",
"[",
"'zonefiles'",
"]",
"db",
"=",
"SubdomainDB",
"(",
"db_path",
",",
"zonefiles_dir",
")",
"try",
":",
"subrec",
"=",
"db",
".",
"get_subdomain_entry",
"(",
"fqn",
")",
"except",
"SubdomainNotFound",
":",
"log",
".",
"warn",
"(",
"\"No such subdomain: {}\"",
".",
"format",
"(",
"fqn",
")",
")",
"return",
"None",
"try",
":",
"return",
"db",
".",
"get_subdomain_DID_info",
"(",
"fqn",
")",
"except",
"SubdomainNotFound",
":",
"return",
"None"
] | Get a subdomain's DID info.
Return None if not found | [
"Get",
"a",
"subdomain",
"s",
"DID",
"info",
".",
"Return",
"None",
"if",
"not",
"found"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1940-L1966 |
230,658 | blockstack/blockstack-core | blockstack/lib/subdomains.py | get_DID_subdomain | def get_DID_subdomain(did, db_path=None, zonefiles_dir=None, atlasdb_path=None, check_pending=False):
"""
Static method for resolving a DID to a subdomain
Return the subdomain record on success
Return None on error
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
log.warn("Subdomain support is disabled")
return None
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
if atlasdb_path is None:
atlasdb_path = opts['atlasdb_path']
db = SubdomainDB(db_path, zonefiles_dir)
try:
subrec = db.get_DID_subdomain(did)
except Exception as e:
if BLOCKSTACK_DEBUG:
log.exception(e)
log.warn("Failed to load subdomain for {}".format(did))
return None
if check_pending:
# make sure that all of the zone files between this subdomain's
# domain's creation and this subdomain's zone file index are present,
# minus the ones that are allowed to be missing.
subrec.pending = db.subdomain_check_pending(subrec, atlasdb_path)
return subrec | python | def get_DID_subdomain(did, db_path=None, zonefiles_dir=None, atlasdb_path=None, check_pending=False):
"""
Static method for resolving a DID to a subdomain
Return the subdomain record on success
Return None on error
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
log.warn("Subdomain support is disabled")
return None
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
if atlasdb_path is None:
atlasdb_path = opts['atlasdb_path']
db = SubdomainDB(db_path, zonefiles_dir)
try:
subrec = db.get_DID_subdomain(did)
except Exception as e:
if BLOCKSTACK_DEBUG:
log.exception(e)
log.warn("Failed to load subdomain for {}".format(did))
return None
if check_pending:
# make sure that all of the zone files between this subdomain's
# domain's creation and this subdomain's zone file index are present,
# minus the ones that are allowed to be missing.
subrec.pending = db.subdomain_check_pending(subrec, atlasdb_path)
return subrec | [
"def",
"get_DID_subdomain",
"(",
"did",
",",
"db_path",
"=",
"None",
",",
"zonefiles_dir",
"=",
"None",
",",
"atlasdb_path",
"=",
"None",
",",
"check_pending",
"=",
"False",
")",
":",
"opts",
"=",
"get_blockstack_opts",
"(",
")",
"if",
"not",
"is_subdomains_enabled",
"(",
"opts",
")",
":",
"log",
".",
"warn",
"(",
"\"Subdomain support is disabled\"",
")",
"return",
"None",
"if",
"db_path",
"is",
"None",
":",
"db_path",
"=",
"opts",
"[",
"'subdomaindb_path'",
"]",
"if",
"zonefiles_dir",
"is",
"None",
":",
"zonefiles_dir",
"=",
"opts",
"[",
"'zonefiles'",
"]",
"if",
"atlasdb_path",
"is",
"None",
":",
"atlasdb_path",
"=",
"opts",
"[",
"'atlasdb_path'",
"]",
"db",
"=",
"SubdomainDB",
"(",
"db_path",
",",
"zonefiles_dir",
")",
"try",
":",
"subrec",
"=",
"db",
".",
"get_DID_subdomain",
"(",
"did",
")",
"except",
"Exception",
"as",
"e",
":",
"if",
"BLOCKSTACK_DEBUG",
":",
"log",
".",
"exception",
"(",
"e",
")",
"log",
".",
"warn",
"(",
"\"Failed to load subdomain for {}\"",
".",
"format",
"(",
"did",
")",
")",
"return",
"None",
"if",
"check_pending",
":",
"# make sure that all of the zone files between this subdomain's",
"# domain's creation and this subdomain's zone file index are present,",
"# minus the ones that are allowed to be missing.",
"subrec",
".",
"pending",
"=",
"db",
".",
"subdomain_check_pending",
"(",
"subrec",
",",
"atlasdb_path",
")",
"return",
"subrec"
] | Static method for resolving a DID to a subdomain
Return the subdomain record on success
Return None on error | [
"Static",
"method",
"for",
"resolving",
"a",
"DID",
"to",
"a",
"subdomain",
"Return",
"the",
"subdomain",
"record",
"on",
"success",
"Return",
"None",
"on",
"error"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1969-L2005 |
230,659 | blockstack/blockstack-core | blockstack/lib/subdomains.py | is_subdomain_zonefile_hash | def is_subdomain_zonefile_hash(fqn, zonefile_hash, db_path=None, zonefiles_dir=None):
"""
Static method for getting all historic zone file hashes for a subdomain
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
return []
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
zonefile_hashes = db.is_subdomain_zonefile_hash(fqn, zonefile_hash)
return zonefile_hashes | python | def is_subdomain_zonefile_hash(fqn, zonefile_hash, db_path=None, zonefiles_dir=None):
"""
Static method for getting all historic zone file hashes for a subdomain
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
return []
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
zonefile_hashes = db.is_subdomain_zonefile_hash(fqn, zonefile_hash)
return zonefile_hashes | [
"def",
"is_subdomain_zonefile_hash",
"(",
"fqn",
",",
"zonefile_hash",
",",
"db_path",
"=",
"None",
",",
"zonefiles_dir",
"=",
"None",
")",
":",
"opts",
"=",
"get_blockstack_opts",
"(",
")",
"if",
"not",
"is_subdomains_enabled",
"(",
"opts",
")",
":",
"return",
"[",
"]",
"if",
"db_path",
"is",
"None",
":",
"db_path",
"=",
"opts",
"[",
"'subdomaindb_path'",
"]",
"if",
"zonefiles_dir",
"is",
"None",
":",
"zonefiles_dir",
"=",
"opts",
"[",
"'zonefiles'",
"]",
"db",
"=",
"SubdomainDB",
"(",
"db_path",
",",
"zonefiles_dir",
")",
"zonefile_hashes",
"=",
"db",
".",
"is_subdomain_zonefile_hash",
"(",
"fqn",
",",
"zonefile_hash",
")",
"return",
"zonefile_hashes"
] | Static method for getting all historic zone file hashes for a subdomain | [
"Static",
"method",
"for",
"getting",
"all",
"historic",
"zone",
"file",
"hashes",
"for",
"a",
"subdomain"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L2008-L2024 |
230,660 | blockstack/blockstack-core | blockstack/lib/subdomains.py | get_subdomain_history | def get_subdomain_history(fqn, offset=None, count=None, reverse=False, db_path=None, zonefiles_dir=None, json=False):
"""
Static method for getting all historic operations on a subdomain
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
return []
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
recs = db.get_subdomain_history(fqn, offset=offset, count=count)
if json:
recs = [rec.to_json() for rec in recs]
ret = {}
for rec in recs:
if rec['block_number'] not in ret:
ret[rec['block_number']] = []
ret[rec['block_number']].append(rec)
if reverse:
for block_height in ret:
ret[block_height].sort(lambda r1, r2: -1 if r1['parent_zonefile_index'] > r2['parent_zonefile_index'] or
(r1['parent_zonefile_index'] == r2['parent_zonefile_index'] and r1['zonefile_offset'] > r2['zonefile_offset']) else
1 if r1['parent_zonefile_index'] < r2['parent_zonefile_index'] or
(r1['parent_zonefile_index'] == r2['parent_zonefile_index'] and r1['zonefile_offset'] < r2['zonefile_offset']) else
0)
return ret
else:
return recs | python | def get_subdomain_history(fqn, offset=None, count=None, reverse=False, db_path=None, zonefiles_dir=None, json=False):
"""
Static method for getting all historic operations on a subdomain
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
return []
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
recs = db.get_subdomain_history(fqn, offset=offset, count=count)
if json:
recs = [rec.to_json() for rec in recs]
ret = {}
for rec in recs:
if rec['block_number'] not in ret:
ret[rec['block_number']] = []
ret[rec['block_number']].append(rec)
if reverse:
for block_height in ret:
ret[block_height].sort(lambda r1, r2: -1 if r1['parent_zonefile_index'] > r2['parent_zonefile_index'] or
(r1['parent_zonefile_index'] == r2['parent_zonefile_index'] and r1['zonefile_offset'] > r2['zonefile_offset']) else
1 if r1['parent_zonefile_index'] < r2['parent_zonefile_index'] or
(r1['parent_zonefile_index'] == r2['parent_zonefile_index'] and r1['zonefile_offset'] < r2['zonefile_offset']) else
0)
return ret
else:
return recs | [
"def",
"get_subdomain_history",
"(",
"fqn",
",",
"offset",
"=",
"None",
",",
"count",
"=",
"None",
",",
"reverse",
"=",
"False",
",",
"db_path",
"=",
"None",
",",
"zonefiles_dir",
"=",
"None",
",",
"json",
"=",
"False",
")",
":",
"opts",
"=",
"get_blockstack_opts",
"(",
")",
"if",
"not",
"is_subdomains_enabled",
"(",
"opts",
")",
":",
"return",
"[",
"]",
"if",
"db_path",
"is",
"None",
":",
"db_path",
"=",
"opts",
"[",
"'subdomaindb_path'",
"]",
"if",
"zonefiles_dir",
"is",
"None",
":",
"zonefiles_dir",
"=",
"opts",
"[",
"'zonefiles'",
"]",
"db",
"=",
"SubdomainDB",
"(",
"db_path",
",",
"zonefiles_dir",
")",
"recs",
"=",
"db",
".",
"get_subdomain_history",
"(",
"fqn",
",",
"offset",
"=",
"offset",
",",
"count",
"=",
"count",
")",
"if",
"json",
":",
"recs",
"=",
"[",
"rec",
".",
"to_json",
"(",
")",
"for",
"rec",
"in",
"recs",
"]",
"ret",
"=",
"{",
"}",
"for",
"rec",
"in",
"recs",
":",
"if",
"rec",
"[",
"'block_number'",
"]",
"not",
"in",
"ret",
":",
"ret",
"[",
"rec",
"[",
"'block_number'",
"]",
"]",
"=",
"[",
"]",
"ret",
"[",
"rec",
"[",
"'block_number'",
"]",
"]",
".",
"append",
"(",
"rec",
")",
"if",
"reverse",
":",
"for",
"block_height",
"in",
"ret",
":",
"ret",
"[",
"block_height",
"]",
".",
"sort",
"(",
"lambda",
"r1",
",",
"r2",
":",
"-",
"1",
"if",
"r1",
"[",
"'parent_zonefile_index'",
"]",
">",
"r2",
"[",
"'parent_zonefile_index'",
"]",
"or",
"(",
"r1",
"[",
"'parent_zonefile_index'",
"]",
"==",
"r2",
"[",
"'parent_zonefile_index'",
"]",
"and",
"r1",
"[",
"'zonefile_offset'",
"]",
">",
"r2",
"[",
"'zonefile_offset'",
"]",
")",
"else",
"1",
"if",
"r1",
"[",
"'parent_zonefile_index'",
"]",
"<",
"r2",
"[",
"'parent_zonefile_index'",
"]",
"or",
"(",
"r1",
"[",
"'parent_zonefile_index'",
"]",
"==",
"r2",
"[",
"'parent_zonefile_index'",
"]",
"and",
"r1",
"[",
"'zonefile_offset'",
"]",
"<",
"r2",
"[",
"'zonefile_offset'",
"]",
")",
"else",
"0",
")",
"return",
"ret",
"else",
":",
"return",
"recs"
] | Static method for getting all historic operations on a subdomain | [
"Static",
"method",
"for",
"getting",
"all",
"historic",
"operations",
"on",
"a",
"subdomain"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L2027-L2063 |
230,661 | blockstack/blockstack-core | blockstack/lib/subdomains.py | get_all_subdomains | def get_all_subdomains(offset=None, count=None, min_sequence=None, db_path=None, zonefiles_dir=None):
"""
Static method for getting the list of all subdomains
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
return []
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
return db.get_all_subdomains(offset=offset, count=count, min_sequence=None) | python | def get_all_subdomains(offset=None, count=None, min_sequence=None, db_path=None, zonefiles_dir=None):
"""
Static method for getting the list of all subdomains
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
return []
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
return db.get_all_subdomains(offset=offset, count=count, min_sequence=None) | [
"def",
"get_all_subdomains",
"(",
"offset",
"=",
"None",
",",
"count",
"=",
"None",
",",
"min_sequence",
"=",
"None",
",",
"db_path",
"=",
"None",
",",
"zonefiles_dir",
"=",
"None",
")",
":",
"opts",
"=",
"get_blockstack_opts",
"(",
")",
"if",
"not",
"is_subdomains_enabled",
"(",
"opts",
")",
":",
"return",
"[",
"]",
"if",
"db_path",
"is",
"None",
":",
"db_path",
"=",
"opts",
"[",
"'subdomaindb_path'",
"]",
"if",
"zonefiles_dir",
"is",
"None",
":",
"zonefiles_dir",
"=",
"opts",
"[",
"'zonefiles'",
"]",
"db",
"=",
"SubdomainDB",
"(",
"db_path",
",",
"zonefiles_dir",
")",
"return",
"db",
".",
"get_all_subdomains",
"(",
"offset",
"=",
"offset",
",",
"count",
"=",
"count",
",",
"min_sequence",
"=",
"None",
")"
] | Static method for getting the list of all subdomains | [
"Static",
"method",
"for",
"getting",
"the",
"list",
"of",
"all",
"subdomains"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L2066-L2081 |
230,662 | blockstack/blockstack-core | blockstack/lib/subdomains.py | get_subdomain_ops_at_txid | def get_subdomain_ops_at_txid(txid, db_path=None, zonefiles_dir=None):
"""
Static method for getting the list of subdomain operations accepted at a given txid.
Includes unaccepted subdomain operations
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
return []
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
return db.get_subdomain_ops_at_txid(txid) | python | def get_subdomain_ops_at_txid(txid, db_path=None, zonefiles_dir=None):
"""
Static method for getting the list of subdomain operations accepted at a given txid.
Includes unaccepted subdomain operations
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
return []
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
return db.get_subdomain_ops_at_txid(txid) | [
"def",
"get_subdomain_ops_at_txid",
"(",
"txid",
",",
"db_path",
"=",
"None",
",",
"zonefiles_dir",
"=",
"None",
")",
":",
"opts",
"=",
"get_blockstack_opts",
"(",
")",
"if",
"not",
"is_subdomains_enabled",
"(",
"opts",
")",
":",
"return",
"[",
"]",
"if",
"db_path",
"is",
"None",
":",
"db_path",
"=",
"opts",
"[",
"'subdomaindb_path'",
"]",
"if",
"zonefiles_dir",
"is",
"None",
":",
"zonefiles_dir",
"=",
"opts",
"[",
"'zonefiles'",
"]",
"db",
"=",
"SubdomainDB",
"(",
"db_path",
",",
"zonefiles_dir",
")",
"return",
"db",
".",
"get_subdomain_ops_at_txid",
"(",
"txid",
")"
] | Static method for getting the list of subdomain operations accepted at a given txid.
Includes unaccepted subdomain operations | [
"Static",
"method",
"for",
"getting",
"the",
"list",
"of",
"subdomain",
"operations",
"accepted",
"at",
"a",
"given",
"txid",
".",
"Includes",
"unaccepted",
"subdomain",
"operations"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L2084-L2100 |
230,663 | blockstack/blockstack-core | blockstack/lib/subdomains.py | get_subdomains_owned_by_address | def get_subdomains_owned_by_address(address, db_path=None, zonefiles_dir=None):
"""
Static method for getting the list of subdomains for a given address
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
return []
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
return db.get_subdomains_owned_by_address(address) | python | def get_subdomains_owned_by_address(address, db_path=None, zonefiles_dir=None):
"""
Static method for getting the list of subdomains for a given address
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
return []
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
return db.get_subdomains_owned_by_address(address) | [
"def",
"get_subdomains_owned_by_address",
"(",
"address",
",",
"db_path",
"=",
"None",
",",
"zonefiles_dir",
"=",
"None",
")",
":",
"opts",
"=",
"get_blockstack_opts",
"(",
")",
"if",
"not",
"is_subdomains_enabled",
"(",
"opts",
")",
":",
"return",
"[",
"]",
"if",
"db_path",
"is",
"None",
":",
"db_path",
"=",
"opts",
"[",
"'subdomaindb_path'",
"]",
"if",
"zonefiles_dir",
"is",
"None",
":",
"zonefiles_dir",
"=",
"opts",
"[",
"'zonefiles'",
"]",
"db",
"=",
"SubdomainDB",
"(",
"db_path",
",",
"zonefiles_dir",
")",
"return",
"db",
".",
"get_subdomains_owned_by_address",
"(",
"address",
")"
] | Static method for getting the list of subdomains for a given address | [
"Static",
"method",
"for",
"getting",
"the",
"list",
"of",
"subdomains",
"for",
"a",
"given",
"address"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L2103-L2118 |
230,664 | blockstack/blockstack-core | blockstack/lib/subdomains.py | get_subdomain_last_sequence | def get_subdomain_last_sequence(db_path=None, zonefiles_dir=None):
"""
Static method for getting the last sequence number in the database
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
return []
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
return db.get_last_sequence() | python | def get_subdomain_last_sequence(db_path=None, zonefiles_dir=None):
"""
Static method for getting the last sequence number in the database
"""
opts = get_blockstack_opts()
if not is_subdomains_enabled(opts):
return []
if db_path is None:
db_path = opts['subdomaindb_path']
if zonefiles_dir is None:
zonefiles_dir = opts['zonefiles']
db = SubdomainDB(db_path, zonefiles_dir)
return db.get_last_sequence() | [
"def",
"get_subdomain_last_sequence",
"(",
"db_path",
"=",
"None",
",",
"zonefiles_dir",
"=",
"None",
")",
":",
"opts",
"=",
"get_blockstack_opts",
"(",
")",
"if",
"not",
"is_subdomains_enabled",
"(",
"opts",
")",
":",
"return",
"[",
"]",
"if",
"db_path",
"is",
"None",
":",
"db_path",
"=",
"opts",
"[",
"'subdomaindb_path'",
"]",
"if",
"zonefiles_dir",
"is",
"None",
":",
"zonefiles_dir",
"=",
"opts",
"[",
"'zonefiles'",
"]",
"db",
"=",
"SubdomainDB",
"(",
"db_path",
",",
"zonefiles_dir",
")",
"return",
"db",
".",
"get_last_sequence",
"(",
")"
] | Static method for getting the last sequence number in the database | [
"Static",
"method",
"for",
"getting",
"the",
"last",
"sequence",
"number",
"in",
"the",
"database"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L2121-L2136 |
230,665 | blockstack/blockstack-core | blockstack/lib/subdomains.py | sign | def sign(privkey_bundle, plaintext):
"""
Sign a subdomain plaintext with a private key bundle
Returns the base64-encoded scriptsig
"""
if virtualchain.is_singlesig(privkey_bundle):
return sign_singlesig(privkey_bundle, plaintext)
elif virtualchain.is_multisig(privkey_bundle):
return sign_multisig(privkey_bundle, plaintext)
else:
raise ValueError("private key bundle is neither a singlesig nor multisig bundle") | python | def sign(privkey_bundle, plaintext):
"""
Sign a subdomain plaintext with a private key bundle
Returns the base64-encoded scriptsig
"""
if virtualchain.is_singlesig(privkey_bundle):
return sign_singlesig(privkey_bundle, plaintext)
elif virtualchain.is_multisig(privkey_bundle):
return sign_multisig(privkey_bundle, plaintext)
else:
raise ValueError("private key bundle is neither a singlesig nor multisig bundle") | [
"def",
"sign",
"(",
"privkey_bundle",
",",
"plaintext",
")",
":",
"if",
"virtualchain",
".",
"is_singlesig",
"(",
"privkey_bundle",
")",
":",
"return",
"sign_singlesig",
"(",
"privkey_bundle",
",",
"plaintext",
")",
"elif",
"virtualchain",
".",
"is_multisig",
"(",
"privkey_bundle",
")",
":",
"return",
"sign_multisig",
"(",
"privkey_bundle",
",",
"plaintext",
")",
"else",
":",
"raise",
"ValueError",
"(",
"\"private key bundle is neither a singlesig nor multisig bundle\"",
")"
] | Sign a subdomain plaintext with a private key bundle
Returns the base64-encoded scriptsig | [
"Sign",
"a",
"subdomain",
"plaintext",
"with",
"a",
"private",
"key",
"bundle",
"Returns",
"the",
"base64",
"-",
"encoded",
"scriptsig"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L2152-L2162 |
230,666 | blockstack/blockstack-core | blockstack/lib/subdomains.py | subdomains_init | def subdomains_init(blockstack_opts, working_dir, atlas_state):
"""
Set up subdomain state
Returns a SubdomainIndex object that has been successfully connected to Atlas
"""
if not is_subdomains_enabled(blockstack_opts):
return None
subdomain_state = SubdomainIndex(blockstack_opts['subdomaindb_path'], blockstack_opts=blockstack_opts)
atlas_node_add_callback(atlas_state, 'store_zonefile', subdomain_state.enqueue_zonefile)
return subdomain_state | python | def subdomains_init(blockstack_opts, working_dir, atlas_state):
"""
Set up subdomain state
Returns a SubdomainIndex object that has been successfully connected to Atlas
"""
if not is_subdomains_enabled(blockstack_opts):
return None
subdomain_state = SubdomainIndex(blockstack_opts['subdomaindb_path'], blockstack_opts=blockstack_opts)
atlas_node_add_callback(atlas_state, 'store_zonefile', subdomain_state.enqueue_zonefile)
return subdomain_state | [
"def",
"subdomains_init",
"(",
"blockstack_opts",
",",
"working_dir",
",",
"atlas_state",
")",
":",
"if",
"not",
"is_subdomains_enabled",
"(",
"blockstack_opts",
")",
":",
"return",
"None",
"subdomain_state",
"=",
"SubdomainIndex",
"(",
"blockstack_opts",
"[",
"'subdomaindb_path'",
"]",
",",
"blockstack_opts",
"=",
"blockstack_opts",
")",
"atlas_node_add_callback",
"(",
"atlas_state",
",",
"'store_zonefile'",
",",
"subdomain_state",
".",
"enqueue_zonefile",
")",
"return",
"subdomain_state"
] | Set up subdomain state
Returns a SubdomainIndex object that has been successfully connected to Atlas | [
"Set",
"up",
"subdomain",
"state",
"Returns",
"a",
"SubdomainIndex",
"object",
"that",
"has",
"been",
"successfully",
"connected",
"to",
"Atlas"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L2220-L2231 |
230,667 | blockstack/blockstack-core | blockstack/lib/subdomains.py | Subdomain.verify_signature | def verify_signature(self, addr):
"""
Given an address, verify whether or not it was signed by it
"""
return verify(virtualchain.address_reencode(addr), self.get_plaintext_to_sign(), self.sig) | python | def verify_signature(self, addr):
"""
Given an address, verify whether or not it was signed by it
"""
return verify(virtualchain.address_reencode(addr), self.get_plaintext_to_sign(), self.sig) | [
"def",
"verify_signature",
"(",
"self",
",",
"addr",
")",
":",
"return",
"verify",
"(",
"virtualchain",
".",
"address_reencode",
"(",
"addr",
")",
",",
"self",
".",
"get_plaintext_to_sign",
"(",
")",
",",
"self",
".",
"sig",
")"
] | Given an address, verify whether or not it was signed by it | [
"Given",
"an",
"address",
"verify",
"whether",
"or",
"not",
"it",
"was",
"signed",
"by",
"it"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L202-L206 |
230,668 | blockstack/blockstack-core | blockstack/lib/subdomains.py | Subdomain.serialize_to_txt | def serialize_to_txt(self):
"""
Serialize this subdomain record to a TXT record. The trailing newline will be omitted
"""
txtrec = {
'name': self.fqn if self.independent else self.subdomain,
'txt': self.pack_subdomain()[1:]
}
return blockstack_zones.record_processors.process_txt([txtrec], '{txt}').strip() | python | def serialize_to_txt(self):
"""
Serialize this subdomain record to a TXT record. The trailing newline will be omitted
"""
txtrec = {
'name': self.fqn if self.independent else self.subdomain,
'txt': self.pack_subdomain()[1:]
}
return blockstack_zones.record_processors.process_txt([txtrec], '{txt}').strip() | [
"def",
"serialize_to_txt",
"(",
"self",
")",
":",
"txtrec",
"=",
"{",
"'name'",
":",
"self",
".",
"fqn",
"if",
"self",
".",
"independent",
"else",
"self",
".",
"subdomain",
",",
"'txt'",
":",
"self",
".",
"pack_subdomain",
"(",
")",
"[",
"1",
":",
"]",
"}",
"return",
"blockstack_zones",
".",
"record_processors",
".",
"process_txt",
"(",
"[",
"txtrec",
"]",
",",
"'{txt}'",
")",
".",
"strip",
"(",
")"
] | Serialize this subdomain record to a TXT record. The trailing newline will be omitted | [
"Serialize",
"this",
"subdomain",
"record",
"to",
"a",
"TXT",
"record",
".",
"The",
"trailing",
"newline",
"will",
"be",
"omitted"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L224-L232 |
230,669 | blockstack/blockstack-core | blockstack/lib/subdomains.py | Subdomain.parse_subdomain_missing_zonefiles_record | def parse_subdomain_missing_zonefiles_record(cls, rec):
"""
Parse a missing-zonefiles vector given by the domain.
Returns the list of zone file indexes on success
Raises ParseError on unparseable records
"""
txt_entry = rec['txt']
if isinstance(txt_entry, list):
raise ParseError("TXT entry too long for a missing zone file list")
try:
return [int(i) for i in txt_entry.split(',')] if txt_entry is not None and len(txt_entry) > 0 else []
except ValueError:
raise ParseError('Invalid integers') | python | def parse_subdomain_missing_zonefiles_record(cls, rec):
"""
Parse a missing-zonefiles vector given by the domain.
Returns the list of zone file indexes on success
Raises ParseError on unparseable records
"""
txt_entry = rec['txt']
if isinstance(txt_entry, list):
raise ParseError("TXT entry too long for a missing zone file list")
try:
return [int(i) for i in txt_entry.split(',')] if txt_entry is not None and len(txt_entry) > 0 else []
except ValueError:
raise ParseError('Invalid integers') | [
"def",
"parse_subdomain_missing_zonefiles_record",
"(",
"cls",
",",
"rec",
")",
":",
"txt_entry",
"=",
"rec",
"[",
"'txt'",
"]",
"if",
"isinstance",
"(",
"txt_entry",
",",
"list",
")",
":",
"raise",
"ParseError",
"(",
"\"TXT entry too long for a missing zone file list\"",
")",
"try",
":",
"return",
"[",
"int",
"(",
"i",
")",
"for",
"i",
"in",
"txt_entry",
".",
"split",
"(",
"','",
")",
"]",
"if",
"txt_entry",
"is",
"not",
"None",
"and",
"len",
"(",
"txt_entry",
")",
">",
"0",
"else",
"[",
"]",
"except",
"ValueError",
":",
"raise",
"ParseError",
"(",
"'Invalid integers'",
")"
] | Parse a missing-zonefiles vector given by the domain.
Returns the list of zone file indexes on success
Raises ParseError on unparseable records | [
"Parse",
"a",
"missing",
"-",
"zonefiles",
"vector",
"given",
"by",
"the",
"domain",
".",
"Returns",
"the",
"list",
"of",
"zone",
"file",
"indexes",
"on",
"success",
"Raises",
"ParseError",
"on",
"unparseable",
"records"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L260-L273 |
230,670 | blockstack/blockstack-core | blockstack/lib/subdomains.py | Subdomain.get_public_key | def get_public_key(self):
"""
Parse the scriptSig and extract the public key.
Raises ValueError if this is a multisig-controlled subdomain.
"""
res = self.get_public_key_info()
if 'error' in res:
raise ValueError(res['error'])
if res['type'] != 'singlesig':
raise ValueError(res['error'])
return res['public_keys'][0] | python | def get_public_key(self):
"""
Parse the scriptSig and extract the public key.
Raises ValueError if this is a multisig-controlled subdomain.
"""
res = self.get_public_key_info()
if 'error' in res:
raise ValueError(res['error'])
if res['type'] != 'singlesig':
raise ValueError(res['error'])
return res['public_keys'][0] | [
"def",
"get_public_key",
"(",
"self",
")",
":",
"res",
"=",
"self",
".",
"get_public_key_info",
"(",
")",
"if",
"'error'",
"in",
"res",
":",
"raise",
"ValueError",
"(",
"res",
"[",
"'error'",
"]",
")",
"if",
"res",
"[",
"'type'",
"]",
"!=",
"'singlesig'",
":",
"raise",
"ValueError",
"(",
"res",
"[",
"'error'",
"]",
")",
"return",
"res",
"[",
"'public_keys'",
"]",
"[",
"0",
"]"
] | Parse the scriptSig and extract the public key.
Raises ValueError if this is a multisig-controlled subdomain. | [
"Parse",
"the",
"scriptSig",
"and",
"extract",
"the",
"public",
"key",
".",
"Raises",
"ValueError",
"if",
"this",
"is",
"a",
"multisig",
"-",
"controlled",
"subdomain",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L345-L357 |
230,671 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainIndex.close | def close(self):
"""
Close the index
"""
with self.subdomain_db_lock:
self.subdomain_db.close()
self.subdomain_db = None
self.subdomain_db_path = None | python | def close(self):
"""
Close the index
"""
with self.subdomain_db_lock:
self.subdomain_db.close()
self.subdomain_db = None
self.subdomain_db_path = None | [
"def",
"close",
"(",
"self",
")",
":",
"with",
"self",
".",
"subdomain_db_lock",
":",
"self",
".",
"subdomain_db",
".",
"close",
"(",
")",
"self",
".",
"subdomain_db",
"=",
"None",
"self",
".",
"subdomain_db_path",
"=",
"None"
] | Close the index | [
"Close",
"the",
"index"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L441-L448 |
230,672 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainIndex.make_new_subdomain_history | def make_new_subdomain_history(self, cursor, subdomain_rec):
"""
Recalculate the history for this subdomain from genesis up until this record.
Returns the list of subdomain records we need to save.
"""
# what's the subdomain's history up until this subdomain record?
hist = self.subdomain_db.get_subdomain_history(subdomain_rec.get_fqn(), include_unaccepted=True, end_sequence=subdomain_rec.n+1, end_zonefile_index=subdomain_rec.parent_zonefile_index+1, cur=cursor)
assert len(hist) > 0, 'BUG: not yet stored: {}'.format(subdomain_rec)
for i in range(0, len(hist)):
hist[i].accepted = False
hist.sort(lambda h1, h2: -1 if h1.n < h2.n or (h1.n == h2.n and h1.parent_zonefile_index < h2.parent_zonefile_index) \
else 0 if h1.n == h2.n and h1.parent_zonefile_index == h2.parent_zonefile_index \
else 1)
if not self.check_initial_subdomain(hist[0]):
log.debug("Reject initial {}".format(hist[0]))
return hist
else:
log.debug("Accept initial {}".format(hist[0]))
pass
hist[0].accepted = True
last_accepted = 0
for i in xrange(1, len(hist)):
if self.check_subdomain_transition(hist[last_accepted], hist[i]):
log.debug("Accept historic update {}".format(hist[i]))
hist[i].accepted = True
last_accepted = i
else:
log.debug("Reject historic update {}".format(hist[i]))
hist[i].accepted = False
return hist | python | def make_new_subdomain_history(self, cursor, subdomain_rec):
"""
Recalculate the history for this subdomain from genesis up until this record.
Returns the list of subdomain records we need to save.
"""
# what's the subdomain's history up until this subdomain record?
hist = self.subdomain_db.get_subdomain_history(subdomain_rec.get_fqn(), include_unaccepted=True, end_sequence=subdomain_rec.n+1, end_zonefile_index=subdomain_rec.parent_zonefile_index+1, cur=cursor)
assert len(hist) > 0, 'BUG: not yet stored: {}'.format(subdomain_rec)
for i in range(0, len(hist)):
hist[i].accepted = False
hist.sort(lambda h1, h2: -1 if h1.n < h2.n or (h1.n == h2.n and h1.parent_zonefile_index < h2.parent_zonefile_index) \
else 0 if h1.n == h2.n and h1.parent_zonefile_index == h2.parent_zonefile_index \
else 1)
if not self.check_initial_subdomain(hist[0]):
log.debug("Reject initial {}".format(hist[0]))
return hist
else:
log.debug("Accept initial {}".format(hist[0]))
pass
hist[0].accepted = True
last_accepted = 0
for i in xrange(1, len(hist)):
if self.check_subdomain_transition(hist[last_accepted], hist[i]):
log.debug("Accept historic update {}".format(hist[i]))
hist[i].accepted = True
last_accepted = i
else:
log.debug("Reject historic update {}".format(hist[i]))
hist[i].accepted = False
return hist | [
"def",
"make_new_subdomain_history",
"(",
"self",
",",
"cursor",
",",
"subdomain_rec",
")",
":",
"# what's the subdomain's history up until this subdomain record?",
"hist",
"=",
"self",
".",
"subdomain_db",
".",
"get_subdomain_history",
"(",
"subdomain_rec",
".",
"get_fqn",
"(",
")",
",",
"include_unaccepted",
"=",
"True",
",",
"end_sequence",
"=",
"subdomain_rec",
".",
"n",
"+",
"1",
",",
"end_zonefile_index",
"=",
"subdomain_rec",
".",
"parent_zonefile_index",
"+",
"1",
",",
"cur",
"=",
"cursor",
")",
"assert",
"len",
"(",
"hist",
")",
">",
"0",
",",
"'BUG: not yet stored: {}'",
".",
"format",
"(",
"subdomain_rec",
")",
"for",
"i",
"in",
"range",
"(",
"0",
",",
"len",
"(",
"hist",
")",
")",
":",
"hist",
"[",
"i",
"]",
".",
"accepted",
"=",
"False",
"hist",
".",
"sort",
"(",
"lambda",
"h1",
",",
"h2",
":",
"-",
"1",
"if",
"h1",
".",
"n",
"<",
"h2",
".",
"n",
"or",
"(",
"h1",
".",
"n",
"==",
"h2",
".",
"n",
"and",
"h1",
".",
"parent_zonefile_index",
"<",
"h2",
".",
"parent_zonefile_index",
")",
"else",
"0",
"if",
"h1",
".",
"n",
"==",
"h2",
".",
"n",
"and",
"h1",
".",
"parent_zonefile_index",
"==",
"h2",
".",
"parent_zonefile_index",
"else",
"1",
")",
"if",
"not",
"self",
".",
"check_initial_subdomain",
"(",
"hist",
"[",
"0",
"]",
")",
":",
"log",
".",
"debug",
"(",
"\"Reject initial {}\"",
".",
"format",
"(",
"hist",
"[",
"0",
"]",
")",
")",
"return",
"hist",
"else",
":",
"log",
".",
"debug",
"(",
"\"Accept initial {}\"",
".",
"format",
"(",
"hist",
"[",
"0",
"]",
")",
")",
"pass",
"hist",
"[",
"0",
"]",
".",
"accepted",
"=",
"True",
"last_accepted",
"=",
"0",
"for",
"i",
"in",
"xrange",
"(",
"1",
",",
"len",
"(",
"hist",
")",
")",
":",
"if",
"self",
".",
"check_subdomain_transition",
"(",
"hist",
"[",
"last_accepted",
"]",
",",
"hist",
"[",
"i",
"]",
")",
":",
"log",
".",
"debug",
"(",
"\"Accept historic update {}\"",
".",
"format",
"(",
"hist",
"[",
"i",
"]",
")",
")",
"hist",
"[",
"i",
"]",
".",
"accepted",
"=",
"True",
"last_accepted",
"=",
"i",
"else",
":",
"log",
".",
"debug",
"(",
"\"Reject historic update {}\"",
".",
"format",
"(",
"hist",
"[",
"i",
"]",
")",
")",
"hist",
"[",
"i",
"]",
".",
"accepted",
"=",
"False",
"return",
"hist"
] | Recalculate the history for this subdomain from genesis up until this record.
Returns the list of subdomain records we need to save. | [
"Recalculate",
"the",
"history",
"for",
"this",
"subdomain",
"from",
"genesis",
"up",
"until",
"this",
"record",
".",
"Returns",
"the",
"list",
"of",
"subdomain",
"records",
"we",
"need",
"to",
"save",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L570-L605 |
230,673 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainIndex.make_new_subdomain_future | def make_new_subdomain_future(self, cursor, subdomain_rec):
"""
Recalculate the future for this subdomain from the current record
until the latest known record.
Returns the list of subdomain records we need to save.
"""
assert subdomain_rec.accepted, 'BUG: given subdomain record must already be accepted'
# what's the subdomain's future after this record?
fut = self.subdomain_db.get_subdomain_history(subdomain_rec.get_fqn(), include_unaccepted=True, start_sequence=subdomain_rec.n, start_zonefile_index=subdomain_rec.parent_zonefile_index, cur=cursor)
for i in range(0, len(fut)):
if fut[i].n == subdomain_rec.n and fut[i].parent_zonefile_index == subdomain_rec.parent_zonefile_index:
fut.pop(i)
break
if len(fut) == 0:
log.debug("At tip: {}".format(subdomain_rec))
return []
for i in range(0, len(fut)):
fut[i].accepted = False
fut = [subdomain_rec] + fut
fut.sort(lambda h1, h2: -1 if h1.n < h2.n or (h1.n == h2.n and h1.parent_zonefile_index < h2.parent_zonefile_index) \
else 0 if h1.n == h2.n and h1.parent_zonefile_index == h2.parent_zonefile_index \
else 1)
assert fut[0].accepted, 'BUG: initial subdomain record is not accepted: {}'.format(fut[0])
last_accepted = 0
for i in range(1, len(fut)):
if self.check_subdomain_transition(fut[last_accepted], fut[i]):
log.debug("Accept future update {}".format(fut[i]))
fut[i].accepted = True
last_accepted = i
else:
log.debug("Reject future update {}".format(fut[i]))
fut[i].accepted = False
return fut | python | def make_new_subdomain_future(self, cursor, subdomain_rec):
"""
Recalculate the future for this subdomain from the current record
until the latest known record.
Returns the list of subdomain records we need to save.
"""
assert subdomain_rec.accepted, 'BUG: given subdomain record must already be accepted'
# what's the subdomain's future after this record?
fut = self.subdomain_db.get_subdomain_history(subdomain_rec.get_fqn(), include_unaccepted=True, start_sequence=subdomain_rec.n, start_zonefile_index=subdomain_rec.parent_zonefile_index, cur=cursor)
for i in range(0, len(fut)):
if fut[i].n == subdomain_rec.n and fut[i].parent_zonefile_index == subdomain_rec.parent_zonefile_index:
fut.pop(i)
break
if len(fut) == 0:
log.debug("At tip: {}".format(subdomain_rec))
return []
for i in range(0, len(fut)):
fut[i].accepted = False
fut = [subdomain_rec] + fut
fut.sort(lambda h1, h2: -1 if h1.n < h2.n or (h1.n == h2.n and h1.parent_zonefile_index < h2.parent_zonefile_index) \
else 0 if h1.n == h2.n and h1.parent_zonefile_index == h2.parent_zonefile_index \
else 1)
assert fut[0].accepted, 'BUG: initial subdomain record is not accepted: {}'.format(fut[0])
last_accepted = 0
for i in range(1, len(fut)):
if self.check_subdomain_transition(fut[last_accepted], fut[i]):
log.debug("Accept future update {}".format(fut[i]))
fut[i].accepted = True
last_accepted = i
else:
log.debug("Reject future update {}".format(fut[i]))
fut[i].accepted = False
return fut | [
"def",
"make_new_subdomain_future",
"(",
"self",
",",
"cursor",
",",
"subdomain_rec",
")",
":",
"assert",
"subdomain_rec",
".",
"accepted",
",",
"'BUG: given subdomain record must already be accepted'",
"# what's the subdomain's future after this record?",
"fut",
"=",
"self",
".",
"subdomain_db",
".",
"get_subdomain_history",
"(",
"subdomain_rec",
".",
"get_fqn",
"(",
")",
",",
"include_unaccepted",
"=",
"True",
",",
"start_sequence",
"=",
"subdomain_rec",
".",
"n",
",",
"start_zonefile_index",
"=",
"subdomain_rec",
".",
"parent_zonefile_index",
",",
"cur",
"=",
"cursor",
")",
"for",
"i",
"in",
"range",
"(",
"0",
",",
"len",
"(",
"fut",
")",
")",
":",
"if",
"fut",
"[",
"i",
"]",
".",
"n",
"==",
"subdomain_rec",
".",
"n",
"and",
"fut",
"[",
"i",
"]",
".",
"parent_zonefile_index",
"==",
"subdomain_rec",
".",
"parent_zonefile_index",
":",
"fut",
".",
"pop",
"(",
"i",
")",
"break",
"if",
"len",
"(",
"fut",
")",
"==",
"0",
":",
"log",
".",
"debug",
"(",
"\"At tip: {}\"",
".",
"format",
"(",
"subdomain_rec",
")",
")",
"return",
"[",
"]",
"for",
"i",
"in",
"range",
"(",
"0",
",",
"len",
"(",
"fut",
")",
")",
":",
"fut",
"[",
"i",
"]",
".",
"accepted",
"=",
"False",
"fut",
"=",
"[",
"subdomain_rec",
"]",
"+",
"fut",
"fut",
".",
"sort",
"(",
"lambda",
"h1",
",",
"h2",
":",
"-",
"1",
"if",
"h1",
".",
"n",
"<",
"h2",
".",
"n",
"or",
"(",
"h1",
".",
"n",
"==",
"h2",
".",
"n",
"and",
"h1",
".",
"parent_zonefile_index",
"<",
"h2",
".",
"parent_zonefile_index",
")",
"else",
"0",
"if",
"h1",
".",
"n",
"==",
"h2",
".",
"n",
"and",
"h1",
".",
"parent_zonefile_index",
"==",
"h2",
".",
"parent_zonefile_index",
"else",
"1",
")",
"assert",
"fut",
"[",
"0",
"]",
".",
"accepted",
",",
"'BUG: initial subdomain record is not accepted: {}'",
".",
"format",
"(",
"fut",
"[",
"0",
"]",
")",
"last_accepted",
"=",
"0",
"for",
"i",
"in",
"range",
"(",
"1",
",",
"len",
"(",
"fut",
")",
")",
":",
"if",
"self",
".",
"check_subdomain_transition",
"(",
"fut",
"[",
"last_accepted",
"]",
",",
"fut",
"[",
"i",
"]",
")",
":",
"log",
".",
"debug",
"(",
"\"Accept future update {}\"",
".",
"format",
"(",
"fut",
"[",
"i",
"]",
")",
")",
"fut",
"[",
"i",
"]",
".",
"accepted",
"=",
"True",
"last_accepted",
"=",
"i",
"else",
":",
"log",
".",
"debug",
"(",
"\"Reject future update {}\"",
".",
"format",
"(",
"fut",
"[",
"i",
"]",
")",
")",
"fut",
"[",
"i",
"]",
".",
"accepted",
"=",
"False",
"return",
"fut"
] | Recalculate the future for this subdomain from the current record
until the latest known record.
Returns the list of subdomain records we need to save. | [
"Recalculate",
"the",
"future",
"for",
"this",
"subdomain",
"from",
"the",
"current",
"record",
"until",
"the",
"latest",
"known",
"record",
".",
"Returns",
"the",
"list",
"of",
"subdomain",
"records",
"we",
"need",
"to",
"save",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L608-L647 |
230,674 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainIndex.subdomain_try_insert | def subdomain_try_insert(self, cursor, subdomain_rec, history_neighbors):
"""
Try to insert a subdomain record into its history neighbors.
This is an optimization that handles the "usual" case.
We can do this without having to rewrite this subdomain's past and future
if (1) we can find a previously-accepted subdomain record, and (2) the transition
from this subdomain record to a future subdomain record preserves its
acceptance as True. In this case, the "far" past and "far" future are already
consistent.
Return True if we succeed in doing so.
Return False if not.
"""
blockchain_order = history_neighbors['prev'] + history_neighbors['cur'] + history_neighbors['fut']
last_accepted = -1
for i in range(0, len(blockchain_order)):
if blockchain_order[i].accepted:
last_accepted = i
break
if blockchain_order[i].n > subdomain_rec.n or (blockchain_order[i].n == subdomain_rec.n and blockchain_order[i].parent_zonefile_index > subdomain_rec.parent_zonefile_index):
# can't cheaply insert this subdomain record,
# since none of its immediate ancestors are accepted.
log.debug("No immediate ancestors are accepted on {}".format(subdomain_rec))
return False
if last_accepted < 0:
log.debug("No immediate ancestors or successors are accepted on {}".format(subdomain_rec))
return False
# one ancestor was accepted.
# work from there.
chain_tip_status = blockchain_order[-1].accepted
dirty = [] # to be written
for i in range(last_accepted+1, len(blockchain_order)):
cur_accepted = blockchain_order[i].accepted
new_accepted = self.check_subdomain_transition(blockchain_order[last_accepted], blockchain_order[i])
if new_accepted != cur_accepted:
blockchain_order[i].accepted = new_accepted
log.debug("Changed from {} to {}: {}".format(cur_accepted, new_accepted, blockchain_order[i]))
dirty.append(blockchain_order[i])
if new_accepted:
last_accepted = i
if chain_tip_status != blockchain_order[-1].accepted and len(history_neighbors['fut']) > 0:
# deeper reorg
log.debug("Immediate history chain tip altered from {} to {}: {}".format(chain_tip_status, blockchain_order[-1].accepted, blockchain_order[-1]))
return False
# localized change. Just commit the dirty entries
for subrec in dirty:
log.debug("Update to accepted={}: {}".format(subrec.accepted, subrec))
self.subdomain_db.update_subdomain_entry(subrec, cur=cursor)
return True | python | def subdomain_try_insert(self, cursor, subdomain_rec, history_neighbors):
"""
Try to insert a subdomain record into its history neighbors.
This is an optimization that handles the "usual" case.
We can do this without having to rewrite this subdomain's past and future
if (1) we can find a previously-accepted subdomain record, and (2) the transition
from this subdomain record to a future subdomain record preserves its
acceptance as True. In this case, the "far" past and "far" future are already
consistent.
Return True if we succeed in doing so.
Return False if not.
"""
blockchain_order = history_neighbors['prev'] + history_neighbors['cur'] + history_neighbors['fut']
last_accepted = -1
for i in range(0, len(blockchain_order)):
if blockchain_order[i].accepted:
last_accepted = i
break
if blockchain_order[i].n > subdomain_rec.n or (blockchain_order[i].n == subdomain_rec.n and blockchain_order[i].parent_zonefile_index > subdomain_rec.parent_zonefile_index):
# can't cheaply insert this subdomain record,
# since none of its immediate ancestors are accepted.
log.debug("No immediate ancestors are accepted on {}".format(subdomain_rec))
return False
if last_accepted < 0:
log.debug("No immediate ancestors or successors are accepted on {}".format(subdomain_rec))
return False
# one ancestor was accepted.
# work from there.
chain_tip_status = blockchain_order[-1].accepted
dirty = [] # to be written
for i in range(last_accepted+1, len(blockchain_order)):
cur_accepted = blockchain_order[i].accepted
new_accepted = self.check_subdomain_transition(blockchain_order[last_accepted], blockchain_order[i])
if new_accepted != cur_accepted:
blockchain_order[i].accepted = new_accepted
log.debug("Changed from {} to {}: {}".format(cur_accepted, new_accepted, blockchain_order[i]))
dirty.append(blockchain_order[i])
if new_accepted:
last_accepted = i
if chain_tip_status != blockchain_order[-1].accepted and len(history_neighbors['fut']) > 0:
# deeper reorg
log.debug("Immediate history chain tip altered from {} to {}: {}".format(chain_tip_status, blockchain_order[-1].accepted, blockchain_order[-1]))
return False
# localized change. Just commit the dirty entries
for subrec in dirty:
log.debug("Update to accepted={}: {}".format(subrec.accepted, subrec))
self.subdomain_db.update_subdomain_entry(subrec, cur=cursor)
return True | [
"def",
"subdomain_try_insert",
"(",
"self",
",",
"cursor",
",",
"subdomain_rec",
",",
"history_neighbors",
")",
":",
"blockchain_order",
"=",
"history_neighbors",
"[",
"'prev'",
"]",
"+",
"history_neighbors",
"[",
"'cur'",
"]",
"+",
"history_neighbors",
"[",
"'fut'",
"]",
"last_accepted",
"=",
"-",
"1",
"for",
"i",
"in",
"range",
"(",
"0",
",",
"len",
"(",
"blockchain_order",
")",
")",
":",
"if",
"blockchain_order",
"[",
"i",
"]",
".",
"accepted",
":",
"last_accepted",
"=",
"i",
"break",
"if",
"blockchain_order",
"[",
"i",
"]",
".",
"n",
">",
"subdomain_rec",
".",
"n",
"or",
"(",
"blockchain_order",
"[",
"i",
"]",
".",
"n",
"==",
"subdomain_rec",
".",
"n",
"and",
"blockchain_order",
"[",
"i",
"]",
".",
"parent_zonefile_index",
">",
"subdomain_rec",
".",
"parent_zonefile_index",
")",
":",
"# can't cheaply insert this subdomain record,",
"# since none of its immediate ancestors are accepted.",
"log",
".",
"debug",
"(",
"\"No immediate ancestors are accepted on {}\"",
".",
"format",
"(",
"subdomain_rec",
")",
")",
"return",
"False",
"if",
"last_accepted",
"<",
"0",
":",
"log",
".",
"debug",
"(",
"\"No immediate ancestors or successors are accepted on {}\"",
".",
"format",
"(",
"subdomain_rec",
")",
")",
"return",
"False",
"# one ancestor was accepted.",
"# work from there.",
"chain_tip_status",
"=",
"blockchain_order",
"[",
"-",
"1",
"]",
".",
"accepted",
"dirty",
"=",
"[",
"]",
"# to be written",
"for",
"i",
"in",
"range",
"(",
"last_accepted",
"+",
"1",
",",
"len",
"(",
"blockchain_order",
")",
")",
":",
"cur_accepted",
"=",
"blockchain_order",
"[",
"i",
"]",
".",
"accepted",
"new_accepted",
"=",
"self",
".",
"check_subdomain_transition",
"(",
"blockchain_order",
"[",
"last_accepted",
"]",
",",
"blockchain_order",
"[",
"i",
"]",
")",
"if",
"new_accepted",
"!=",
"cur_accepted",
":",
"blockchain_order",
"[",
"i",
"]",
".",
"accepted",
"=",
"new_accepted",
"log",
".",
"debug",
"(",
"\"Changed from {} to {}: {}\"",
".",
"format",
"(",
"cur_accepted",
",",
"new_accepted",
",",
"blockchain_order",
"[",
"i",
"]",
")",
")",
"dirty",
".",
"append",
"(",
"blockchain_order",
"[",
"i",
"]",
")",
"if",
"new_accepted",
":",
"last_accepted",
"=",
"i",
"if",
"chain_tip_status",
"!=",
"blockchain_order",
"[",
"-",
"1",
"]",
".",
"accepted",
"and",
"len",
"(",
"history_neighbors",
"[",
"'fut'",
"]",
")",
">",
"0",
":",
"# deeper reorg",
"log",
".",
"debug",
"(",
"\"Immediate history chain tip altered from {} to {}: {}\"",
".",
"format",
"(",
"chain_tip_status",
",",
"blockchain_order",
"[",
"-",
"1",
"]",
".",
"accepted",
",",
"blockchain_order",
"[",
"-",
"1",
"]",
")",
")",
"return",
"False",
"# localized change. Just commit the dirty entries",
"for",
"subrec",
"in",
"dirty",
":",
"log",
".",
"debug",
"(",
"\"Update to accepted={}: {}\"",
".",
"format",
"(",
"subrec",
".",
"accepted",
",",
"subrec",
")",
")",
"self",
".",
"subdomain_db",
".",
"update_subdomain_entry",
"(",
"subrec",
",",
"cur",
"=",
"cursor",
")",
"return",
"True"
] | Try to insert a subdomain record into its history neighbors.
This is an optimization that handles the "usual" case.
We can do this without having to rewrite this subdomain's past and future
if (1) we can find a previously-accepted subdomain record, and (2) the transition
from this subdomain record to a future subdomain record preserves its
acceptance as True. In this case, the "far" past and "far" future are already
consistent.
Return True if we succeed in doing so.
Return False if not. | [
"Try",
"to",
"insert",
"a",
"subdomain",
"record",
"into",
"its",
"history",
"neighbors",
".",
"This",
"is",
"an",
"optimization",
"that",
"handles",
"the",
"usual",
"case",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L684-L743 |
230,675 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainIndex.enqueue_zonefile | def enqueue_zonefile(self, zonefile_hash, block_height):
"""
Called when we discover a zone file. Queues up a request to reprocess this name's zone files' subdomains.
zonefile_hash is the hash of the zonefile.
block_height is the minimium block height at which this zone file occurs.
This gets called by:
* AtlasZonefileCrawler (as it's "store_zonefile" callback).
* rpc_put_zonefiles()
"""
with self.serialized_enqueue_zonefile:
log.debug("Append {} from {}".format(zonefile_hash, block_height))
queuedb_append(self.subdomain_queue_path, "zonefiles", zonefile_hash, json.dumps({'zonefile_hash': zonefile_hash, 'block_height': block_height})) | python | def enqueue_zonefile(self, zonefile_hash, block_height):
"""
Called when we discover a zone file. Queues up a request to reprocess this name's zone files' subdomains.
zonefile_hash is the hash of the zonefile.
block_height is the minimium block height at which this zone file occurs.
This gets called by:
* AtlasZonefileCrawler (as it's "store_zonefile" callback).
* rpc_put_zonefiles()
"""
with self.serialized_enqueue_zonefile:
log.debug("Append {} from {}".format(zonefile_hash, block_height))
queuedb_append(self.subdomain_queue_path, "zonefiles", zonefile_hash, json.dumps({'zonefile_hash': zonefile_hash, 'block_height': block_height})) | [
"def",
"enqueue_zonefile",
"(",
"self",
",",
"zonefile_hash",
",",
"block_height",
")",
":",
"with",
"self",
".",
"serialized_enqueue_zonefile",
":",
"log",
".",
"debug",
"(",
"\"Append {} from {}\"",
".",
"format",
"(",
"zonefile_hash",
",",
"block_height",
")",
")",
"queuedb_append",
"(",
"self",
".",
"subdomain_queue_path",
",",
"\"zonefiles\"",
",",
"zonefile_hash",
",",
"json",
".",
"dumps",
"(",
"{",
"'zonefile_hash'",
":",
"zonefile_hash",
",",
"'block_height'",
":",
"block_height",
"}",
")",
")"
] | Called when we discover a zone file. Queues up a request to reprocess this name's zone files' subdomains.
zonefile_hash is the hash of the zonefile.
block_height is the minimium block height at which this zone file occurs.
This gets called by:
* AtlasZonefileCrawler (as it's "store_zonefile" callback).
* rpc_put_zonefiles() | [
"Called",
"when",
"we",
"discover",
"a",
"zone",
"file",
".",
"Queues",
"up",
"a",
"request",
"to",
"reprocess",
"this",
"name",
"s",
"zone",
"files",
"subdomains",
".",
"zonefile_hash",
"is",
"the",
"hash",
"of",
"the",
"zonefile",
".",
"block_height",
"is",
"the",
"minimium",
"block",
"height",
"at",
"which",
"this",
"zone",
"file",
"occurs",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L823-L835 |
230,676 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainIndex.index_blockchain | def index_blockchain(self, block_start, block_end):
"""
Go through the sequence of zone files discovered in a block range, and reindex the names' subdomains.
"""
log.debug("Processing subdomain updates for zonefiles in blocks {}-{}".format(block_start, block_end))
res = self.find_zonefile_subdomains(block_start, block_end)
zonefile_subdomain_info = res['zonefile_info']
self.process_subdomains(zonefile_subdomain_info) | python | def index_blockchain(self, block_start, block_end):
"""
Go through the sequence of zone files discovered in a block range, and reindex the names' subdomains.
"""
log.debug("Processing subdomain updates for zonefiles in blocks {}-{}".format(block_start, block_end))
res = self.find_zonefile_subdomains(block_start, block_end)
zonefile_subdomain_info = res['zonefile_info']
self.process_subdomains(zonefile_subdomain_info) | [
"def",
"index_blockchain",
"(",
"self",
",",
"block_start",
",",
"block_end",
")",
":",
"log",
".",
"debug",
"(",
"\"Processing subdomain updates for zonefiles in blocks {}-{}\"",
".",
"format",
"(",
"block_start",
",",
"block_end",
")",
")",
"res",
"=",
"self",
".",
"find_zonefile_subdomains",
"(",
"block_start",
",",
"block_end",
")",
"zonefile_subdomain_info",
"=",
"res",
"[",
"'zonefile_info'",
"]",
"self",
".",
"process_subdomains",
"(",
"zonefile_subdomain_info",
")"
] | Go through the sequence of zone files discovered in a block range, and reindex the names' subdomains. | [
"Go",
"through",
"the",
"sequence",
"of",
"zone",
"files",
"discovered",
"in",
"a",
"block",
"range",
"and",
"reindex",
"the",
"names",
"subdomains",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L838-L847 |
230,677 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainIndex.index_discovered_zonefiles | def index_discovered_zonefiles(self, lastblock):
"""
Go through the list of zone files we discovered via Atlas, grouped by name and ordered by block height.
Find all subsequent zone files for this name, and process all subdomain operations contained within them.
"""
all_queued_zfinfos = [] # contents of the queue
subdomain_zonefile_infos = {} # map subdomain fqn to list of zonefile info bundles, for process_subdomains
name_blocks = {} # map domain name to the block at which we should reprocess its subsequent zone files
offset = 0
while True:
queued_zfinfos = queuedb_findall(self.subdomain_queue_path, "zonefiles", limit=100, offset=offset)
if len(queued_zfinfos) == 0:
# done!
break
offset += 100
all_queued_zfinfos += queued_zfinfos
if len(all_queued_zfinfos) >= 1000:
# only do so many zone files per block, so we don't stall the node
break
log.debug("Discovered {} zonefiles".format(len(all_queued_zfinfos)))
for queued_zfinfo in all_queued_zfinfos:
zfinfo = json.loads(queued_zfinfo['data'])
zonefile_hash = zfinfo['zonefile_hash']
block_height = zfinfo['block_height']
# find out the names that sent this zone file at this block
zfinfos = atlasdb_get_zonefiles_by_hash(zonefile_hash, block_height=block_height, path=self.atlasdb_path)
if zfinfos is None:
log.warn("Absent zonefile {}".format(zonefile_hash))
continue
# find out for each name block height at which its zone file was discovered.
# this is where we'll begin looking for more subdomain updates.
for zfi in zfinfos:
if zfi['name'] not in name_blocks:
name_blocks[zfi['name']] = block_height
else:
name_blocks[zfi['name']] = min(block_height, name_blocks[zfi['name']])
for name in name_blocks:
if name_blocks[name] >= lastblock:
continue
log.debug("Finding subdomain updates for {} at block {}".format(name, name_blocks[name]))
# get the subdomains affected at this block by finding the zonefiles created here.
res = self.find_zonefile_subdomains(name_blocks[name], lastblock, name=name)
zonefile_subdomain_info = res['zonefile_info']
subdomain_index = res['subdomains']
# for each subdomain, find the list of zonefiles that contain records for it
for fqn in subdomain_index:
if fqn not in subdomain_zonefile_infos:
subdomain_zonefile_infos[fqn] = []
for i in subdomain_index[fqn]:
subdomain_zonefile_infos[fqn].append(zonefile_subdomain_info[i])
processed = []
for fqn in subdomain_zonefile_infos:
subseq = filter(lambda szi: szi['zonefile_hash'] not in processed, subdomain_zonefile_infos[fqn])
if len(subseq) == 0:
continue
log.debug("Processing {} zone file entries found for {} and others".format(len(subseq), fqn))
subseq.sort(cmp=lambda z1, z2: -1 if z1['block_height'] < z2['block_height'] else 0 if z1['block_height'] == z2['block_height'] else 1)
self.process_subdomains(subseq)
processed += [szi['zonefile_hash'] for szi in subseq]
# clear queue
queuedb_removeall(self.subdomain_queue_path, all_queued_zfinfos)
return True | python | def index_discovered_zonefiles(self, lastblock):
"""
Go through the list of zone files we discovered via Atlas, grouped by name and ordered by block height.
Find all subsequent zone files for this name, and process all subdomain operations contained within them.
"""
all_queued_zfinfos = [] # contents of the queue
subdomain_zonefile_infos = {} # map subdomain fqn to list of zonefile info bundles, for process_subdomains
name_blocks = {} # map domain name to the block at which we should reprocess its subsequent zone files
offset = 0
while True:
queued_zfinfos = queuedb_findall(self.subdomain_queue_path, "zonefiles", limit=100, offset=offset)
if len(queued_zfinfos) == 0:
# done!
break
offset += 100
all_queued_zfinfos += queued_zfinfos
if len(all_queued_zfinfos) >= 1000:
# only do so many zone files per block, so we don't stall the node
break
log.debug("Discovered {} zonefiles".format(len(all_queued_zfinfos)))
for queued_zfinfo in all_queued_zfinfos:
zfinfo = json.loads(queued_zfinfo['data'])
zonefile_hash = zfinfo['zonefile_hash']
block_height = zfinfo['block_height']
# find out the names that sent this zone file at this block
zfinfos = atlasdb_get_zonefiles_by_hash(zonefile_hash, block_height=block_height, path=self.atlasdb_path)
if zfinfos is None:
log.warn("Absent zonefile {}".format(zonefile_hash))
continue
# find out for each name block height at which its zone file was discovered.
# this is where we'll begin looking for more subdomain updates.
for zfi in zfinfos:
if zfi['name'] not in name_blocks:
name_blocks[zfi['name']] = block_height
else:
name_blocks[zfi['name']] = min(block_height, name_blocks[zfi['name']])
for name in name_blocks:
if name_blocks[name] >= lastblock:
continue
log.debug("Finding subdomain updates for {} at block {}".format(name, name_blocks[name]))
# get the subdomains affected at this block by finding the zonefiles created here.
res = self.find_zonefile_subdomains(name_blocks[name], lastblock, name=name)
zonefile_subdomain_info = res['zonefile_info']
subdomain_index = res['subdomains']
# for each subdomain, find the list of zonefiles that contain records for it
for fqn in subdomain_index:
if fqn not in subdomain_zonefile_infos:
subdomain_zonefile_infos[fqn] = []
for i in subdomain_index[fqn]:
subdomain_zonefile_infos[fqn].append(zonefile_subdomain_info[i])
processed = []
for fqn in subdomain_zonefile_infos:
subseq = filter(lambda szi: szi['zonefile_hash'] not in processed, subdomain_zonefile_infos[fqn])
if len(subseq) == 0:
continue
log.debug("Processing {} zone file entries found for {} and others".format(len(subseq), fqn))
subseq.sort(cmp=lambda z1, z2: -1 if z1['block_height'] < z2['block_height'] else 0 if z1['block_height'] == z2['block_height'] else 1)
self.process_subdomains(subseq)
processed += [szi['zonefile_hash'] for szi in subseq]
# clear queue
queuedb_removeall(self.subdomain_queue_path, all_queued_zfinfos)
return True | [
"def",
"index_discovered_zonefiles",
"(",
"self",
",",
"lastblock",
")",
":",
"all_queued_zfinfos",
"=",
"[",
"]",
"# contents of the queue",
"subdomain_zonefile_infos",
"=",
"{",
"}",
"# map subdomain fqn to list of zonefile info bundles, for process_subdomains",
"name_blocks",
"=",
"{",
"}",
"# map domain name to the block at which we should reprocess its subsequent zone files",
"offset",
"=",
"0",
"while",
"True",
":",
"queued_zfinfos",
"=",
"queuedb_findall",
"(",
"self",
".",
"subdomain_queue_path",
",",
"\"zonefiles\"",
",",
"limit",
"=",
"100",
",",
"offset",
"=",
"offset",
")",
"if",
"len",
"(",
"queued_zfinfos",
")",
"==",
"0",
":",
"# done!",
"break",
"offset",
"+=",
"100",
"all_queued_zfinfos",
"+=",
"queued_zfinfos",
"if",
"len",
"(",
"all_queued_zfinfos",
")",
">=",
"1000",
":",
"# only do so many zone files per block, so we don't stall the node",
"break",
"log",
".",
"debug",
"(",
"\"Discovered {} zonefiles\"",
".",
"format",
"(",
"len",
"(",
"all_queued_zfinfos",
")",
")",
")",
"for",
"queued_zfinfo",
"in",
"all_queued_zfinfos",
":",
"zfinfo",
"=",
"json",
".",
"loads",
"(",
"queued_zfinfo",
"[",
"'data'",
"]",
")",
"zonefile_hash",
"=",
"zfinfo",
"[",
"'zonefile_hash'",
"]",
"block_height",
"=",
"zfinfo",
"[",
"'block_height'",
"]",
"# find out the names that sent this zone file at this block",
"zfinfos",
"=",
"atlasdb_get_zonefiles_by_hash",
"(",
"zonefile_hash",
",",
"block_height",
"=",
"block_height",
",",
"path",
"=",
"self",
".",
"atlasdb_path",
")",
"if",
"zfinfos",
"is",
"None",
":",
"log",
".",
"warn",
"(",
"\"Absent zonefile {}\"",
".",
"format",
"(",
"zonefile_hash",
")",
")",
"continue",
"# find out for each name block height at which its zone file was discovered.",
"# this is where we'll begin looking for more subdomain updates.",
"for",
"zfi",
"in",
"zfinfos",
":",
"if",
"zfi",
"[",
"'name'",
"]",
"not",
"in",
"name_blocks",
":",
"name_blocks",
"[",
"zfi",
"[",
"'name'",
"]",
"]",
"=",
"block_height",
"else",
":",
"name_blocks",
"[",
"zfi",
"[",
"'name'",
"]",
"]",
"=",
"min",
"(",
"block_height",
",",
"name_blocks",
"[",
"zfi",
"[",
"'name'",
"]",
"]",
")",
"for",
"name",
"in",
"name_blocks",
":",
"if",
"name_blocks",
"[",
"name",
"]",
">=",
"lastblock",
":",
"continue",
"log",
".",
"debug",
"(",
"\"Finding subdomain updates for {} at block {}\"",
".",
"format",
"(",
"name",
",",
"name_blocks",
"[",
"name",
"]",
")",
")",
"# get the subdomains affected at this block by finding the zonefiles created here.",
"res",
"=",
"self",
".",
"find_zonefile_subdomains",
"(",
"name_blocks",
"[",
"name",
"]",
",",
"lastblock",
",",
"name",
"=",
"name",
")",
"zonefile_subdomain_info",
"=",
"res",
"[",
"'zonefile_info'",
"]",
"subdomain_index",
"=",
"res",
"[",
"'subdomains'",
"]",
"# for each subdomain, find the list of zonefiles that contain records for it",
"for",
"fqn",
"in",
"subdomain_index",
":",
"if",
"fqn",
"not",
"in",
"subdomain_zonefile_infos",
":",
"subdomain_zonefile_infos",
"[",
"fqn",
"]",
"=",
"[",
"]",
"for",
"i",
"in",
"subdomain_index",
"[",
"fqn",
"]",
":",
"subdomain_zonefile_infos",
"[",
"fqn",
"]",
".",
"append",
"(",
"zonefile_subdomain_info",
"[",
"i",
"]",
")",
"processed",
"=",
"[",
"]",
"for",
"fqn",
"in",
"subdomain_zonefile_infos",
":",
"subseq",
"=",
"filter",
"(",
"lambda",
"szi",
":",
"szi",
"[",
"'zonefile_hash'",
"]",
"not",
"in",
"processed",
",",
"subdomain_zonefile_infos",
"[",
"fqn",
"]",
")",
"if",
"len",
"(",
"subseq",
")",
"==",
"0",
":",
"continue",
"log",
".",
"debug",
"(",
"\"Processing {} zone file entries found for {} and others\"",
".",
"format",
"(",
"len",
"(",
"subseq",
")",
",",
"fqn",
")",
")",
"subseq",
".",
"sort",
"(",
"cmp",
"=",
"lambda",
"z1",
",",
"z2",
":",
"-",
"1",
"if",
"z1",
"[",
"'block_height'",
"]",
"<",
"z2",
"[",
"'block_height'",
"]",
"else",
"0",
"if",
"z1",
"[",
"'block_height'",
"]",
"==",
"z2",
"[",
"'block_height'",
"]",
"else",
"1",
")",
"self",
".",
"process_subdomains",
"(",
"subseq",
")",
"processed",
"+=",
"[",
"szi",
"[",
"'zonefile_hash'",
"]",
"for",
"szi",
"in",
"subseq",
"]",
"# clear queue ",
"queuedb_removeall",
"(",
"self",
".",
"subdomain_queue_path",
",",
"all_queued_zfinfos",
")",
"return",
"True"
] | Go through the list of zone files we discovered via Atlas, grouped by name and ordered by block height.
Find all subsequent zone files for this name, and process all subdomain operations contained within them. | [
"Go",
"through",
"the",
"list",
"of",
"zone",
"files",
"we",
"discovered",
"via",
"Atlas",
"grouped",
"by",
"name",
"and",
"ordered",
"by",
"block",
"height",
".",
"Find",
"all",
"subsequent",
"zone",
"files",
"for",
"this",
"name",
"and",
"process",
"all",
"subdomain",
"operations",
"contained",
"within",
"them",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L850-L929 |
230,678 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB.subdomain_row_factory | def subdomain_row_factory(cls, cursor, row):
"""
Dict row factory for subdomains
"""
d = {}
for idx, col in enumerate(cursor.description):
d[col[0]] = row[idx]
return d | python | def subdomain_row_factory(cls, cursor, row):
"""
Dict row factory for subdomains
"""
d = {}
for idx, col in enumerate(cursor.description):
d[col[0]] = row[idx]
return d | [
"def",
"subdomain_row_factory",
"(",
"cls",
",",
"cursor",
",",
"row",
")",
":",
"d",
"=",
"{",
"}",
"for",
"idx",
",",
"col",
"in",
"enumerate",
"(",
"cursor",
".",
"description",
")",
":",
"d",
"[",
"col",
"[",
"0",
"]",
"]",
"=",
"row",
"[",
"idx",
"]",
"return",
"d"
] | Dict row factory for subdomains | [
"Dict",
"row",
"factory",
"for",
"subdomains"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1000-L1008 |
230,679 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB._extract_subdomain | def _extract_subdomain(self, rowdata):
"""
Extract a single subdomain from a DB cursor
Raise SubdomainNotFound if there are no valid rows
"""
name = str(rowdata['fully_qualified_subdomain'])
domain = str(rowdata['domain'])
n = str(rowdata['sequence'])
encoded_pubkey = str(rowdata['owner'])
zonefile_hash = str(rowdata['zonefile_hash'])
sig = rowdata['signature']
block_height = int(rowdata['block_height'])
parent_zonefile_hash = str(rowdata['parent_zonefile_hash'])
parent_zonefile_index = int(rowdata['parent_zonefile_index'])
zonefile_offset = int(rowdata['zonefile_offset'])
txid = str(rowdata['txid'])
missing = [int(i) for i in rowdata['missing'].split(',')] if rowdata['missing'] is not None and len(rowdata['missing']) > 0 else []
accepted = int(rowdata['accepted'])
resolver = str(rowdata['resolver']) if rowdata['resolver'] is not None else None
if accepted == 0:
accepted = False
else:
accepted = True
if sig == '' or sig is None:
sig = None
else:
sig = str(sig)
name = str(name)
is_subdomain, _, _ = is_address_subdomain(name)
if not is_subdomain:
raise Exception("Subdomain DB lookup returned bad subdomain result {}".format(name))
zonefile_str = get_atlas_zonefile_data(zonefile_hash, self.zonefiles_dir)
if zonefile_str is None:
log.error("No zone file for {}".format(name))
raise SubdomainNotFound('{}: missing zone file {}'.format(name, zonefile_hash))
return Subdomain(str(name), str(domain), str(encoded_pubkey), int(n), str(zonefile_str), sig, block_height, parent_zonefile_hash, parent_zonefile_index, zonefile_offset, txid, domain_zonefiles_missing=missing, accepted=accepted, resolver=resolver) | python | def _extract_subdomain(self, rowdata):
"""
Extract a single subdomain from a DB cursor
Raise SubdomainNotFound if there are no valid rows
"""
name = str(rowdata['fully_qualified_subdomain'])
domain = str(rowdata['domain'])
n = str(rowdata['sequence'])
encoded_pubkey = str(rowdata['owner'])
zonefile_hash = str(rowdata['zonefile_hash'])
sig = rowdata['signature']
block_height = int(rowdata['block_height'])
parent_zonefile_hash = str(rowdata['parent_zonefile_hash'])
parent_zonefile_index = int(rowdata['parent_zonefile_index'])
zonefile_offset = int(rowdata['zonefile_offset'])
txid = str(rowdata['txid'])
missing = [int(i) for i in rowdata['missing'].split(',')] if rowdata['missing'] is not None and len(rowdata['missing']) > 0 else []
accepted = int(rowdata['accepted'])
resolver = str(rowdata['resolver']) if rowdata['resolver'] is not None else None
if accepted == 0:
accepted = False
else:
accepted = True
if sig == '' or sig is None:
sig = None
else:
sig = str(sig)
name = str(name)
is_subdomain, _, _ = is_address_subdomain(name)
if not is_subdomain:
raise Exception("Subdomain DB lookup returned bad subdomain result {}".format(name))
zonefile_str = get_atlas_zonefile_data(zonefile_hash, self.zonefiles_dir)
if zonefile_str is None:
log.error("No zone file for {}".format(name))
raise SubdomainNotFound('{}: missing zone file {}'.format(name, zonefile_hash))
return Subdomain(str(name), str(domain), str(encoded_pubkey), int(n), str(zonefile_str), sig, block_height, parent_zonefile_hash, parent_zonefile_index, zonefile_offset, txid, domain_zonefiles_missing=missing, accepted=accepted, resolver=resolver) | [
"def",
"_extract_subdomain",
"(",
"self",
",",
"rowdata",
")",
":",
"name",
"=",
"str",
"(",
"rowdata",
"[",
"'fully_qualified_subdomain'",
"]",
")",
"domain",
"=",
"str",
"(",
"rowdata",
"[",
"'domain'",
"]",
")",
"n",
"=",
"str",
"(",
"rowdata",
"[",
"'sequence'",
"]",
")",
"encoded_pubkey",
"=",
"str",
"(",
"rowdata",
"[",
"'owner'",
"]",
")",
"zonefile_hash",
"=",
"str",
"(",
"rowdata",
"[",
"'zonefile_hash'",
"]",
")",
"sig",
"=",
"rowdata",
"[",
"'signature'",
"]",
"block_height",
"=",
"int",
"(",
"rowdata",
"[",
"'block_height'",
"]",
")",
"parent_zonefile_hash",
"=",
"str",
"(",
"rowdata",
"[",
"'parent_zonefile_hash'",
"]",
")",
"parent_zonefile_index",
"=",
"int",
"(",
"rowdata",
"[",
"'parent_zonefile_index'",
"]",
")",
"zonefile_offset",
"=",
"int",
"(",
"rowdata",
"[",
"'zonefile_offset'",
"]",
")",
"txid",
"=",
"str",
"(",
"rowdata",
"[",
"'txid'",
"]",
")",
"missing",
"=",
"[",
"int",
"(",
"i",
")",
"for",
"i",
"in",
"rowdata",
"[",
"'missing'",
"]",
".",
"split",
"(",
"','",
")",
"]",
"if",
"rowdata",
"[",
"'missing'",
"]",
"is",
"not",
"None",
"and",
"len",
"(",
"rowdata",
"[",
"'missing'",
"]",
")",
">",
"0",
"else",
"[",
"]",
"accepted",
"=",
"int",
"(",
"rowdata",
"[",
"'accepted'",
"]",
")",
"resolver",
"=",
"str",
"(",
"rowdata",
"[",
"'resolver'",
"]",
")",
"if",
"rowdata",
"[",
"'resolver'",
"]",
"is",
"not",
"None",
"else",
"None",
"if",
"accepted",
"==",
"0",
":",
"accepted",
"=",
"False",
"else",
":",
"accepted",
"=",
"True",
"if",
"sig",
"==",
"''",
"or",
"sig",
"is",
"None",
":",
"sig",
"=",
"None",
"else",
":",
"sig",
"=",
"str",
"(",
"sig",
")",
"name",
"=",
"str",
"(",
"name",
")",
"is_subdomain",
",",
"_",
",",
"_",
"=",
"is_address_subdomain",
"(",
"name",
")",
"if",
"not",
"is_subdomain",
":",
"raise",
"Exception",
"(",
"\"Subdomain DB lookup returned bad subdomain result {}\"",
".",
"format",
"(",
"name",
")",
")",
"zonefile_str",
"=",
"get_atlas_zonefile_data",
"(",
"zonefile_hash",
",",
"self",
".",
"zonefiles_dir",
")",
"if",
"zonefile_str",
"is",
"None",
":",
"log",
".",
"error",
"(",
"\"No zone file for {}\"",
".",
"format",
"(",
"name",
")",
")",
"raise",
"SubdomainNotFound",
"(",
"'{}: missing zone file {}'",
".",
"format",
"(",
"name",
",",
"zonefile_hash",
")",
")",
"return",
"Subdomain",
"(",
"str",
"(",
"name",
")",
",",
"str",
"(",
"domain",
")",
",",
"str",
"(",
"encoded_pubkey",
")",
",",
"int",
"(",
"n",
")",
",",
"str",
"(",
"zonefile_str",
")",
",",
"sig",
",",
"block_height",
",",
"parent_zonefile_hash",
",",
"parent_zonefile_index",
",",
"zonefile_offset",
",",
"txid",
",",
"domain_zonefiles_missing",
"=",
"missing",
",",
"accepted",
"=",
"accepted",
",",
"resolver",
"=",
"resolver",
")"
] | Extract a single subdomain from a DB cursor
Raise SubdomainNotFound if there are no valid rows | [
"Extract",
"a",
"single",
"subdomain",
"from",
"a",
"DB",
"cursor",
"Raise",
"SubdomainNotFound",
"if",
"there",
"are",
"no",
"valid",
"rows"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1025-L1065 |
230,680 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB.get_subdomains_count | def get_subdomains_count(self, accepted=True, cur=None):
"""
Fetch subdomain names
"""
if accepted:
accepted_filter = 'WHERE accepted=1'
else:
accepted_filter = ''
get_cmd = "SELECT COUNT(DISTINCT fully_qualified_subdomain) as count FROM {} {};".format(
self.subdomain_table, accepted_filter)
cursor = cur
if cursor is None:
cursor = self.conn.cursor()
db_query_execute(cursor, get_cmd, ())
try:
rowdata = cursor.fetchone()
return rowdata['count']
except Exception as e:
if BLOCKSTACK_DEBUG:
log.exception(e)
return 0 | python | def get_subdomains_count(self, accepted=True, cur=None):
"""
Fetch subdomain names
"""
if accepted:
accepted_filter = 'WHERE accepted=1'
else:
accepted_filter = ''
get_cmd = "SELECT COUNT(DISTINCT fully_qualified_subdomain) as count FROM {} {};".format(
self.subdomain_table, accepted_filter)
cursor = cur
if cursor is None:
cursor = self.conn.cursor()
db_query_execute(cursor, get_cmd, ())
try:
rowdata = cursor.fetchone()
return rowdata['count']
except Exception as e:
if BLOCKSTACK_DEBUG:
log.exception(e)
return 0 | [
"def",
"get_subdomains_count",
"(",
"self",
",",
"accepted",
"=",
"True",
",",
"cur",
"=",
"None",
")",
":",
"if",
"accepted",
":",
"accepted_filter",
"=",
"'WHERE accepted=1'",
"else",
":",
"accepted_filter",
"=",
"''",
"get_cmd",
"=",
"\"SELECT COUNT(DISTINCT fully_qualified_subdomain) as count FROM {} {};\"",
".",
"format",
"(",
"self",
".",
"subdomain_table",
",",
"accepted_filter",
")",
"cursor",
"=",
"cur",
"if",
"cursor",
"is",
"None",
":",
"cursor",
"=",
"self",
".",
"conn",
".",
"cursor",
"(",
")",
"db_query_execute",
"(",
"cursor",
",",
"get_cmd",
",",
"(",
")",
")",
"try",
":",
"rowdata",
"=",
"cursor",
".",
"fetchone",
"(",
")",
"return",
"rowdata",
"[",
"'count'",
"]",
"except",
"Exception",
"as",
"e",
":",
"if",
"BLOCKSTACK_DEBUG",
":",
"log",
".",
"exception",
"(",
"e",
")",
"return",
"0"
] | Fetch subdomain names | [
"Fetch",
"subdomain",
"names"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1068-L1092 |
230,681 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB.get_all_subdomains | def get_all_subdomains(self, offset=None, count=None, min_sequence=None, cur=None):
"""
Get and all subdomain names, optionally over a range
"""
get_cmd = 'SELECT DISTINCT fully_qualified_subdomain FROM {}'.format(self.subdomain_table)
args = ()
if min_sequence is not None:
get_cmd += ' WHERE sequence >= ?'
args += (min_sequence,)
if count is not None:
get_cmd += ' LIMIT ?'
args += (count,)
if offset is not None:
get_cmd += ' OFFSET ?'
args += (offset,)
get_cmd += ';'
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
rows = db_query_execute(cursor, get_cmd, args)
subdomains = []
for row in rows:
subdomains.append(row['fully_qualified_subdomain'])
return subdomains | python | def get_all_subdomains(self, offset=None, count=None, min_sequence=None, cur=None):
"""
Get and all subdomain names, optionally over a range
"""
get_cmd = 'SELECT DISTINCT fully_qualified_subdomain FROM {}'.format(self.subdomain_table)
args = ()
if min_sequence is not None:
get_cmd += ' WHERE sequence >= ?'
args += (min_sequence,)
if count is not None:
get_cmd += ' LIMIT ?'
args += (count,)
if offset is not None:
get_cmd += ' OFFSET ?'
args += (offset,)
get_cmd += ';'
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
rows = db_query_execute(cursor, get_cmd, args)
subdomains = []
for row in rows:
subdomains.append(row['fully_qualified_subdomain'])
return subdomains | [
"def",
"get_all_subdomains",
"(",
"self",
",",
"offset",
"=",
"None",
",",
"count",
"=",
"None",
",",
"min_sequence",
"=",
"None",
",",
"cur",
"=",
"None",
")",
":",
"get_cmd",
"=",
"'SELECT DISTINCT fully_qualified_subdomain FROM {}'",
".",
"format",
"(",
"self",
".",
"subdomain_table",
")",
"args",
"=",
"(",
")",
"if",
"min_sequence",
"is",
"not",
"None",
":",
"get_cmd",
"+=",
"' WHERE sequence >= ?'",
"args",
"+=",
"(",
"min_sequence",
",",
")",
"if",
"count",
"is",
"not",
"None",
":",
"get_cmd",
"+=",
"' LIMIT ?'",
"args",
"+=",
"(",
"count",
",",
")",
"if",
"offset",
"is",
"not",
"None",
":",
"get_cmd",
"+=",
"' OFFSET ?'",
"args",
"+=",
"(",
"offset",
",",
")",
"get_cmd",
"+=",
"';'",
"cursor",
"=",
"None",
"if",
"cur",
"is",
"None",
":",
"cursor",
"=",
"self",
".",
"conn",
".",
"cursor",
"(",
")",
"else",
":",
"cursor",
"=",
"cur",
"rows",
"=",
"db_query_execute",
"(",
"cursor",
",",
"get_cmd",
",",
"args",
")",
"subdomains",
"=",
"[",
"]",
"for",
"row",
"in",
"rows",
":",
"subdomains",
".",
"append",
"(",
"row",
"[",
"'fully_qualified_subdomain'",
"]",
")",
"return",
"subdomains"
] | Get and all subdomain names, optionally over a range | [
"Get",
"and",
"all",
"subdomain",
"names",
"optionally",
"over",
"a",
"range"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1095-L1127 |
230,682 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB.get_subdomain_ops_at_txid | def get_subdomain_ops_at_txid(self, txid, cur=None):
"""
Given a txid, get all subdomain operations at that txid.
Include unaccepted operations.
Order by zone file index
"""
get_cmd = 'SELECT * FROM {} WHERE txid = ? ORDER BY zonefile_offset'.format(self.subdomain_table)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
db_query_execute(cursor, get_cmd, (txid,))
try:
return [x for x in cursor.fetchall()]
except Exception as e:
if BLOCKSTACK_DEBUG:
log.exception(e)
return [] | python | def get_subdomain_ops_at_txid(self, txid, cur=None):
"""
Given a txid, get all subdomain operations at that txid.
Include unaccepted operations.
Order by zone file index
"""
get_cmd = 'SELECT * FROM {} WHERE txid = ? ORDER BY zonefile_offset'.format(self.subdomain_table)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
db_query_execute(cursor, get_cmd, (txid,))
try:
return [x for x in cursor.fetchall()]
except Exception as e:
if BLOCKSTACK_DEBUG:
log.exception(e)
return [] | [
"def",
"get_subdomain_ops_at_txid",
"(",
"self",
",",
"txid",
",",
"cur",
"=",
"None",
")",
":",
"get_cmd",
"=",
"'SELECT * FROM {} WHERE txid = ? ORDER BY zonefile_offset'",
".",
"format",
"(",
"self",
".",
"subdomain_table",
")",
"cursor",
"=",
"None",
"if",
"cur",
"is",
"None",
":",
"cursor",
"=",
"self",
".",
"conn",
".",
"cursor",
"(",
")",
"else",
":",
"cursor",
"=",
"cur",
"db_query_execute",
"(",
"cursor",
",",
"get_cmd",
",",
"(",
"txid",
",",
")",
")",
"try",
":",
"return",
"[",
"x",
"for",
"x",
"in",
"cursor",
".",
"fetchall",
"(",
")",
"]",
"except",
"Exception",
"as",
"e",
":",
"if",
"BLOCKSTACK_DEBUG",
":",
"log",
".",
"exception",
"(",
"e",
")",
"return",
"[",
"]"
] | Given a txid, get all subdomain operations at that txid.
Include unaccepted operations.
Order by zone file index | [
"Given",
"a",
"txid",
"get",
"all",
"subdomain",
"operations",
"at",
"that",
"txid",
".",
"Include",
"unaccepted",
"operations",
".",
"Order",
"by",
"zone",
"file",
"index"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1202-L1224 |
230,683 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB.get_subdomains_owned_by_address | def get_subdomains_owned_by_address(self, owner, cur=None):
"""
Get the list of subdomain names that are owned by a given address.
"""
get_cmd = "SELECT fully_qualified_subdomain, MAX(sequence) FROM {} WHERE owner = ? AND accepted=1 GROUP BY fully_qualified_subdomain".format(self.subdomain_table)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
db_query_execute(cursor, get_cmd, (owner,))
try:
return [ x['fully_qualified_subdomain'] for x in cursor.fetchall() ]
except Exception as e:
if BLOCKSTACK_DEBUG:
log.exception(e)
return [] | python | def get_subdomains_owned_by_address(self, owner, cur=None):
"""
Get the list of subdomain names that are owned by a given address.
"""
get_cmd = "SELECT fully_qualified_subdomain, MAX(sequence) FROM {} WHERE owner = ? AND accepted=1 GROUP BY fully_qualified_subdomain".format(self.subdomain_table)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
db_query_execute(cursor, get_cmd, (owner,))
try:
return [ x['fully_qualified_subdomain'] for x in cursor.fetchall() ]
except Exception as e:
if BLOCKSTACK_DEBUG:
log.exception(e)
return [] | [
"def",
"get_subdomains_owned_by_address",
"(",
"self",
",",
"owner",
",",
"cur",
"=",
"None",
")",
":",
"get_cmd",
"=",
"\"SELECT fully_qualified_subdomain, MAX(sequence) FROM {} WHERE owner = ? AND accepted=1 GROUP BY fully_qualified_subdomain\"",
".",
"format",
"(",
"self",
".",
"subdomain_table",
")",
"cursor",
"=",
"None",
"if",
"cur",
"is",
"None",
":",
"cursor",
"=",
"self",
".",
"conn",
".",
"cursor",
"(",
")",
"else",
":",
"cursor",
"=",
"cur",
"db_query_execute",
"(",
"cursor",
",",
"get_cmd",
",",
"(",
"owner",
",",
")",
")",
"try",
":",
"return",
"[",
"x",
"[",
"'fully_qualified_subdomain'",
"]",
"for",
"x",
"in",
"cursor",
".",
"fetchall",
"(",
")",
"]",
"except",
"Exception",
"as",
"e",
":",
"if",
"BLOCKSTACK_DEBUG",
":",
"log",
".",
"exception",
"(",
"e",
")",
"return",
"[",
"]"
] | Get the list of subdomain names that are owned by a given address. | [
"Get",
"the",
"list",
"of",
"subdomain",
"names",
"that",
"are",
"owned",
"by",
"a",
"given",
"address",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1227-L1247 |
230,684 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB.get_domain_resolver | def get_domain_resolver(self, domain_name, cur=None):
"""
Get the last-knwon resolver entry for a domain name
Returns None if not found.
"""
get_cmd = "SELECT resolver FROM {} WHERE domain=? AND resolver != '' AND accepted=1 ORDER BY sequence DESC, parent_zonefile_index DESC LIMIT 1;".format(self.subdomain_table)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
db_query_execute(cursor, get_cmd, (domain_name,))
rowdata = cursor.fetchone()
if not rowdata:
return None
return rowdata['resolver'] | python | def get_domain_resolver(self, domain_name, cur=None):
"""
Get the last-knwon resolver entry for a domain name
Returns None if not found.
"""
get_cmd = "SELECT resolver FROM {} WHERE domain=? AND resolver != '' AND accepted=1 ORDER BY sequence DESC, parent_zonefile_index DESC LIMIT 1;".format(self.subdomain_table)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
db_query_execute(cursor, get_cmd, (domain_name,))
rowdata = cursor.fetchone()
if not rowdata:
return None
return rowdata['resolver'] | [
"def",
"get_domain_resolver",
"(",
"self",
",",
"domain_name",
",",
"cur",
"=",
"None",
")",
":",
"get_cmd",
"=",
"\"SELECT resolver FROM {} WHERE domain=? AND resolver != '' AND accepted=1 ORDER BY sequence DESC, parent_zonefile_index DESC LIMIT 1;\"",
".",
"format",
"(",
"self",
".",
"subdomain_table",
")",
"cursor",
"=",
"None",
"if",
"cur",
"is",
"None",
":",
"cursor",
"=",
"self",
".",
"conn",
".",
"cursor",
"(",
")",
"else",
":",
"cursor",
"=",
"cur",
"db_query_execute",
"(",
"cursor",
",",
"get_cmd",
",",
"(",
"domain_name",
",",
")",
")",
"rowdata",
"=",
"cursor",
".",
"fetchone",
"(",
")",
"if",
"not",
"rowdata",
":",
"return",
"None",
"return",
"rowdata",
"[",
"'resolver'",
"]"
] | Get the last-knwon resolver entry for a domain name
Returns None if not found. | [
"Get",
"the",
"last",
"-",
"knwon",
"resolver",
"entry",
"for",
"a",
"domain",
"name",
"Returns",
"None",
"if",
"not",
"found",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1250-L1269 |
230,685 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB.get_subdomain_DID_info | def get_subdomain_DID_info(self, fqn, cur=None):
"""
Get the DID information for a subdomain.
Raise SubdomainNotFound if there is no such subdomain
Return {'name_type': ..., 'address': ..., 'index': ...}
"""
subrec = self.get_subdomain_entry_at_sequence(fqn, 0, cur=cur)
cmd = 'SELECT zonefile_offset FROM {} WHERE fully_qualified_subdomain = ? AND owner = ? AND sequence=0 AND parent_zonefile_index <= ? AND accepted=1 ORDER BY parent_zonefile_index, zonefile_offset LIMIT 1;'.format(self.subdomain_table)
args = (fqn, subrec.address, subrec.parent_zonefile_index)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
rows = db_query_execute(cursor, cmd, args)
zonefile_offset = None
for r in rows:
zonefile_offset = r['zonefile_offset']
break
if zonefile_offset is None:
raise SubdomainNotFound('No rows for {}'.format(fqn))
cmd = 'SELECT COUNT(*) FROM {} WHERE owner = ? AND sequence=0 AND (parent_zonefile_index < ? OR parent_zonefile_index = ? AND zonefile_offset < ?) AND accepted=1 ORDER BY parent_zonefile_index, zonefile_offset LIMIT 1;'.format(self.subdomain_table)
args = (subrec.address, subrec.parent_zonefile_index, subrec.parent_zonefile_index, zonefile_offset)
rows = db_query_execute(cursor, cmd, args)
count = None
for r in rows:
count = r['COUNT(*)']
break
if count is None:
raise SubdomainNotFound('No rows for {}'.format(fqn))
return {'name_type': 'subdomain', 'address': subrec.address, 'index': count} | python | def get_subdomain_DID_info(self, fqn, cur=None):
"""
Get the DID information for a subdomain.
Raise SubdomainNotFound if there is no such subdomain
Return {'name_type': ..., 'address': ..., 'index': ...}
"""
subrec = self.get_subdomain_entry_at_sequence(fqn, 0, cur=cur)
cmd = 'SELECT zonefile_offset FROM {} WHERE fully_qualified_subdomain = ? AND owner = ? AND sequence=0 AND parent_zonefile_index <= ? AND accepted=1 ORDER BY parent_zonefile_index, zonefile_offset LIMIT 1;'.format(self.subdomain_table)
args = (fqn, subrec.address, subrec.parent_zonefile_index)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
rows = db_query_execute(cursor, cmd, args)
zonefile_offset = None
for r in rows:
zonefile_offset = r['zonefile_offset']
break
if zonefile_offset is None:
raise SubdomainNotFound('No rows for {}'.format(fqn))
cmd = 'SELECT COUNT(*) FROM {} WHERE owner = ? AND sequence=0 AND (parent_zonefile_index < ? OR parent_zonefile_index = ? AND zonefile_offset < ?) AND accepted=1 ORDER BY parent_zonefile_index, zonefile_offset LIMIT 1;'.format(self.subdomain_table)
args = (subrec.address, subrec.parent_zonefile_index, subrec.parent_zonefile_index, zonefile_offset)
rows = db_query_execute(cursor, cmd, args)
count = None
for r in rows:
count = r['COUNT(*)']
break
if count is None:
raise SubdomainNotFound('No rows for {}'.format(fqn))
return {'name_type': 'subdomain', 'address': subrec.address, 'index': count} | [
"def",
"get_subdomain_DID_info",
"(",
"self",
",",
"fqn",
",",
"cur",
"=",
"None",
")",
":",
"subrec",
"=",
"self",
".",
"get_subdomain_entry_at_sequence",
"(",
"fqn",
",",
"0",
",",
"cur",
"=",
"cur",
")",
"cmd",
"=",
"'SELECT zonefile_offset FROM {} WHERE fully_qualified_subdomain = ? AND owner = ? AND sequence=0 AND parent_zonefile_index <= ? AND accepted=1 ORDER BY parent_zonefile_index, zonefile_offset LIMIT 1;'",
".",
"format",
"(",
"self",
".",
"subdomain_table",
")",
"args",
"=",
"(",
"fqn",
",",
"subrec",
".",
"address",
",",
"subrec",
".",
"parent_zonefile_index",
")",
"cursor",
"=",
"None",
"if",
"cur",
"is",
"None",
":",
"cursor",
"=",
"self",
".",
"conn",
".",
"cursor",
"(",
")",
"else",
":",
"cursor",
"=",
"cur",
"rows",
"=",
"db_query_execute",
"(",
"cursor",
",",
"cmd",
",",
"args",
")",
"zonefile_offset",
"=",
"None",
"for",
"r",
"in",
"rows",
":",
"zonefile_offset",
"=",
"r",
"[",
"'zonefile_offset'",
"]",
"break",
"if",
"zonefile_offset",
"is",
"None",
":",
"raise",
"SubdomainNotFound",
"(",
"'No rows for {}'",
".",
"format",
"(",
"fqn",
")",
")",
"cmd",
"=",
"'SELECT COUNT(*) FROM {} WHERE owner = ? AND sequence=0 AND (parent_zonefile_index < ? OR parent_zonefile_index = ? AND zonefile_offset < ?) AND accepted=1 ORDER BY parent_zonefile_index, zonefile_offset LIMIT 1;'",
".",
"format",
"(",
"self",
".",
"subdomain_table",
")",
"args",
"=",
"(",
"subrec",
".",
"address",
",",
"subrec",
".",
"parent_zonefile_index",
",",
"subrec",
".",
"parent_zonefile_index",
",",
"zonefile_offset",
")",
"rows",
"=",
"db_query_execute",
"(",
"cursor",
",",
"cmd",
",",
"args",
")",
"count",
"=",
"None",
"for",
"r",
"in",
"rows",
":",
"count",
"=",
"r",
"[",
"'COUNT(*)'",
"]",
"break",
"if",
"count",
"is",
"None",
":",
"raise",
"SubdomainNotFound",
"(",
"'No rows for {}'",
".",
"format",
"(",
"fqn",
")",
")",
"return",
"{",
"'name_type'",
":",
"'subdomain'",
",",
"'address'",
":",
"subrec",
".",
"address",
",",
"'index'",
":",
"count",
"}"
] | Get the DID information for a subdomain.
Raise SubdomainNotFound if there is no such subdomain
Return {'name_type': ..., 'address': ..., 'index': ...} | [
"Get",
"the",
"DID",
"information",
"for",
"a",
"subdomain",
".",
"Raise",
"SubdomainNotFound",
"if",
"there",
"is",
"no",
"such",
"subdomain"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1272-L1311 |
230,686 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB.get_DID_subdomain | def get_DID_subdomain(self, did, cur=None):
"""
Get a subdomain, given its DID
Raise ValueError if the DID is invalid
Raise SubdomainNotFound if the DID does not correspond to a subdomain
"""
did = str(did)
try:
did_info = parse_DID(did)
assert did_info['name_type'] == 'subdomain', 'Not a subdomain DID'
except:
raise ValueError("Invalid DID: {}".format(did))
original_address = did_info['address']
name_index = did_info['index']
# find the initial subdomain (the nth subdomain created by this address)
cmd = 'SELECT fully_qualified_subdomain FROM {} WHERE owner = ? AND sequence = ? ORDER BY parent_zonefile_index, zonefile_offset LIMIT 1 OFFSET ?;'.format(self.subdomain_table)
args = (original_address, 0, name_index)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
subdomain_name = None
rows = db_query_execute(cursor, cmd, args)
for r in rows:
subdomain_name = r['fully_qualified_subdomain']
break
if not subdomain_name:
raise SubdomainNotFound('Does not correspond to a subdomain: {}'.format(did))
# get the current form
subrec = self.get_subdomain_entry(subdomain_name, cur=cur)
subrec.did_info = did_info
return subrec | python | def get_DID_subdomain(self, did, cur=None):
"""
Get a subdomain, given its DID
Raise ValueError if the DID is invalid
Raise SubdomainNotFound if the DID does not correspond to a subdomain
"""
did = str(did)
try:
did_info = parse_DID(did)
assert did_info['name_type'] == 'subdomain', 'Not a subdomain DID'
except:
raise ValueError("Invalid DID: {}".format(did))
original_address = did_info['address']
name_index = did_info['index']
# find the initial subdomain (the nth subdomain created by this address)
cmd = 'SELECT fully_qualified_subdomain FROM {} WHERE owner = ? AND sequence = ? ORDER BY parent_zonefile_index, zonefile_offset LIMIT 1 OFFSET ?;'.format(self.subdomain_table)
args = (original_address, 0, name_index)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
subdomain_name = None
rows = db_query_execute(cursor, cmd, args)
for r in rows:
subdomain_name = r['fully_qualified_subdomain']
break
if not subdomain_name:
raise SubdomainNotFound('Does not correspond to a subdomain: {}'.format(did))
# get the current form
subrec = self.get_subdomain_entry(subdomain_name, cur=cur)
subrec.did_info = did_info
return subrec | [
"def",
"get_DID_subdomain",
"(",
"self",
",",
"did",
",",
"cur",
"=",
"None",
")",
":",
"did",
"=",
"str",
"(",
"did",
")",
"try",
":",
"did_info",
"=",
"parse_DID",
"(",
"did",
")",
"assert",
"did_info",
"[",
"'name_type'",
"]",
"==",
"'subdomain'",
",",
"'Not a subdomain DID'",
"except",
":",
"raise",
"ValueError",
"(",
"\"Invalid DID: {}\"",
".",
"format",
"(",
"did",
")",
")",
"original_address",
"=",
"did_info",
"[",
"'address'",
"]",
"name_index",
"=",
"did_info",
"[",
"'index'",
"]",
"# find the initial subdomain (the nth subdomain created by this address)",
"cmd",
"=",
"'SELECT fully_qualified_subdomain FROM {} WHERE owner = ? AND sequence = ? ORDER BY parent_zonefile_index, zonefile_offset LIMIT 1 OFFSET ?;'",
".",
"format",
"(",
"self",
".",
"subdomain_table",
")",
"args",
"=",
"(",
"original_address",
",",
"0",
",",
"name_index",
")",
"cursor",
"=",
"None",
"if",
"cur",
"is",
"None",
":",
"cursor",
"=",
"self",
".",
"conn",
".",
"cursor",
"(",
")",
"else",
":",
"cursor",
"=",
"cur",
"subdomain_name",
"=",
"None",
"rows",
"=",
"db_query_execute",
"(",
"cursor",
",",
"cmd",
",",
"args",
")",
"for",
"r",
"in",
"rows",
":",
"subdomain_name",
"=",
"r",
"[",
"'fully_qualified_subdomain'",
"]",
"break",
"if",
"not",
"subdomain_name",
":",
"raise",
"SubdomainNotFound",
"(",
"'Does not correspond to a subdomain: {}'",
".",
"format",
"(",
"did",
")",
")",
"# get the current form",
"subrec",
"=",
"self",
".",
"get_subdomain_entry",
"(",
"subdomain_name",
",",
"cur",
"=",
"cur",
")",
"subrec",
".",
"did_info",
"=",
"did_info",
"return",
"subrec"
] | Get a subdomain, given its DID
Raise ValueError if the DID is invalid
Raise SubdomainNotFound if the DID does not correspond to a subdomain | [
"Get",
"a",
"subdomain",
"given",
"its",
"DID",
"Raise",
"ValueError",
"if",
"the",
"DID",
"is",
"invalid",
"Raise",
"SubdomainNotFound",
"if",
"the",
"DID",
"does",
"not",
"correspond",
"to",
"a",
"subdomain"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1314-L1354 |
230,687 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB.is_subdomain_zonefile_hash | def is_subdomain_zonefile_hash(self, fqn, zonefile_hash, cur=None):
"""
Does this zone file hash belong to this subdomain?
"""
sql = 'SELECT COUNT(zonefile_hash) FROM {} WHERE fully_qualified_subdomain = ? and zonefile_hash = ?;'.format(self.subdomain_table)
args = (fqn,zonefile_hash)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
rows = db_query_execute(cursor, sql, args)
count = None
for row in rows:
count = row['COUNT(zonefile_hash)']
break
return (count > 0) | python | def is_subdomain_zonefile_hash(self, fqn, zonefile_hash, cur=None):
"""
Does this zone file hash belong to this subdomain?
"""
sql = 'SELECT COUNT(zonefile_hash) FROM {} WHERE fully_qualified_subdomain = ? and zonefile_hash = ?;'.format(self.subdomain_table)
args = (fqn,zonefile_hash)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
rows = db_query_execute(cursor, sql, args)
count = None
for row in rows:
count = row['COUNT(zonefile_hash)']
break
return (count > 0) | [
"def",
"is_subdomain_zonefile_hash",
"(",
"self",
",",
"fqn",
",",
"zonefile_hash",
",",
"cur",
"=",
"None",
")",
":",
"sql",
"=",
"'SELECT COUNT(zonefile_hash) FROM {} WHERE fully_qualified_subdomain = ? and zonefile_hash = ?;'",
".",
"format",
"(",
"self",
".",
"subdomain_table",
")",
"args",
"=",
"(",
"fqn",
",",
"zonefile_hash",
")",
"cursor",
"=",
"None",
"if",
"cur",
"is",
"None",
":",
"cursor",
"=",
"self",
".",
"conn",
".",
"cursor",
"(",
")",
"else",
":",
"cursor",
"=",
"cur",
"rows",
"=",
"db_query_execute",
"(",
"cursor",
",",
"sql",
",",
"args",
")",
"count",
"=",
"None",
"for",
"row",
"in",
"rows",
":",
"count",
"=",
"row",
"[",
"'COUNT(zonefile_hash)'",
"]",
"break",
"return",
"(",
"count",
">",
"0",
")"
] | Does this zone file hash belong to this subdomain? | [
"Does",
"this",
"zone",
"file",
"hash",
"belong",
"to",
"this",
"subdomain?"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1357-L1377 |
230,688 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB.update_subdomain_entry | def update_subdomain_entry(self, subdomain_obj, cur=None):
"""
Update the subdomain history table for this subdomain entry.
Creates it if it doesn't exist.
Return True on success
Raise exception on error
"""
# sanity checks
assert isinstance(subdomain_obj, Subdomain)
# NOTE: there is no need to call fsync() on the zone file fd here---we already have the data from the on-chain name's zone file fsync'ed,
# so this information is already durable (albeit somewhere else) and can ostensibly be restored later.
# We get such high subdomain traffic that we cannot call fsync() here each time; otherwise we could stall the node.
zonefile_hash = get_zonefile_data_hash(subdomain_obj.zonefile_str)
rc = store_atlas_zonefile_data(subdomain_obj.zonefile_str, self.zonefiles_dir, fsync=False)
if not rc:
raise Exception("Failed to store zone file {} from {}".format(zonefile_hash, subdomain_obj.get_fqn()))
write_cmd = 'INSERT OR REPLACE INTO {} VALUES (?,?,?,?,?,?,?,?,?,?,?,?,?,?)'.format(self.subdomain_table)
args = (subdomain_obj.get_fqn(), subdomain_obj.domain, subdomain_obj.n, subdomain_obj.address, zonefile_hash,
subdomain_obj.sig, subdomain_obj.block_height, subdomain_obj.parent_zonefile_hash,
subdomain_obj.parent_zonefile_index, subdomain_obj.zonefile_offset, subdomain_obj.txid,
','.join(str(i) for i in subdomain_obj.domain_zonefiles_missing),
1 if subdomain_obj.accepted else 0,
subdomain_obj.resolver)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
db_query_execute(cursor, write_cmd, args)
num_rows_written = cursor.rowcount
if cur is None:
# not part of a transaction
self.conn.commit()
if num_rows_written != 1:
raise ValueError("No row written: fqn={} seq={}".format(subdomain_obj.get_fqn(), subdomain_obj.n))
return True | python | def update_subdomain_entry(self, subdomain_obj, cur=None):
"""
Update the subdomain history table for this subdomain entry.
Creates it if it doesn't exist.
Return True on success
Raise exception on error
"""
# sanity checks
assert isinstance(subdomain_obj, Subdomain)
# NOTE: there is no need to call fsync() on the zone file fd here---we already have the data from the on-chain name's zone file fsync'ed,
# so this information is already durable (albeit somewhere else) and can ostensibly be restored later.
# We get such high subdomain traffic that we cannot call fsync() here each time; otherwise we could stall the node.
zonefile_hash = get_zonefile_data_hash(subdomain_obj.zonefile_str)
rc = store_atlas_zonefile_data(subdomain_obj.zonefile_str, self.zonefiles_dir, fsync=False)
if not rc:
raise Exception("Failed to store zone file {} from {}".format(zonefile_hash, subdomain_obj.get_fqn()))
write_cmd = 'INSERT OR REPLACE INTO {} VALUES (?,?,?,?,?,?,?,?,?,?,?,?,?,?)'.format(self.subdomain_table)
args = (subdomain_obj.get_fqn(), subdomain_obj.domain, subdomain_obj.n, subdomain_obj.address, zonefile_hash,
subdomain_obj.sig, subdomain_obj.block_height, subdomain_obj.parent_zonefile_hash,
subdomain_obj.parent_zonefile_index, subdomain_obj.zonefile_offset, subdomain_obj.txid,
','.join(str(i) for i in subdomain_obj.domain_zonefiles_missing),
1 if subdomain_obj.accepted else 0,
subdomain_obj.resolver)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
db_query_execute(cursor, write_cmd, args)
num_rows_written = cursor.rowcount
if cur is None:
# not part of a transaction
self.conn.commit()
if num_rows_written != 1:
raise ValueError("No row written: fqn={} seq={}".format(subdomain_obj.get_fqn(), subdomain_obj.n))
return True | [
"def",
"update_subdomain_entry",
"(",
"self",
",",
"subdomain_obj",
",",
"cur",
"=",
"None",
")",
":",
"# sanity checks",
"assert",
"isinstance",
"(",
"subdomain_obj",
",",
"Subdomain",
")",
"# NOTE: there is no need to call fsync() on the zone file fd here---we already have the data from the on-chain name's zone file fsync'ed,",
"# so this information is already durable (albeit somewhere else) and can ostensibly be restored later.",
"# We get such high subdomain traffic that we cannot call fsync() here each time; otherwise we could stall the node.",
"zonefile_hash",
"=",
"get_zonefile_data_hash",
"(",
"subdomain_obj",
".",
"zonefile_str",
")",
"rc",
"=",
"store_atlas_zonefile_data",
"(",
"subdomain_obj",
".",
"zonefile_str",
",",
"self",
".",
"zonefiles_dir",
",",
"fsync",
"=",
"False",
")",
"if",
"not",
"rc",
":",
"raise",
"Exception",
"(",
"\"Failed to store zone file {} from {}\"",
".",
"format",
"(",
"zonefile_hash",
",",
"subdomain_obj",
".",
"get_fqn",
"(",
")",
")",
")",
"write_cmd",
"=",
"'INSERT OR REPLACE INTO {} VALUES (?,?,?,?,?,?,?,?,?,?,?,?,?,?)'",
".",
"format",
"(",
"self",
".",
"subdomain_table",
")",
"args",
"=",
"(",
"subdomain_obj",
".",
"get_fqn",
"(",
")",
",",
"subdomain_obj",
".",
"domain",
",",
"subdomain_obj",
".",
"n",
",",
"subdomain_obj",
".",
"address",
",",
"zonefile_hash",
",",
"subdomain_obj",
".",
"sig",
",",
"subdomain_obj",
".",
"block_height",
",",
"subdomain_obj",
".",
"parent_zonefile_hash",
",",
"subdomain_obj",
".",
"parent_zonefile_index",
",",
"subdomain_obj",
".",
"zonefile_offset",
",",
"subdomain_obj",
".",
"txid",
",",
"','",
".",
"join",
"(",
"str",
"(",
"i",
")",
"for",
"i",
"in",
"subdomain_obj",
".",
"domain_zonefiles_missing",
")",
",",
"1",
"if",
"subdomain_obj",
".",
"accepted",
"else",
"0",
",",
"subdomain_obj",
".",
"resolver",
")",
"cursor",
"=",
"None",
"if",
"cur",
"is",
"None",
":",
"cursor",
"=",
"self",
".",
"conn",
".",
"cursor",
"(",
")",
"else",
":",
"cursor",
"=",
"cur",
"db_query_execute",
"(",
"cursor",
",",
"write_cmd",
",",
"args",
")",
"num_rows_written",
"=",
"cursor",
".",
"rowcount",
"if",
"cur",
"is",
"None",
":",
"# not part of a transaction",
"self",
".",
"conn",
".",
"commit",
"(",
")",
"if",
"num_rows_written",
"!=",
"1",
":",
"raise",
"ValueError",
"(",
"\"No row written: fqn={} seq={}\"",
".",
"format",
"(",
"subdomain_obj",
".",
"get_fqn",
"(",
")",
",",
"subdomain_obj",
".",
"n",
")",
")",
"return",
"True"
] | Update the subdomain history table for this subdomain entry.
Creates it if it doesn't exist.
Return True on success
Raise exception on error | [
"Update",
"the",
"subdomain",
"history",
"table",
"for",
"this",
"subdomain",
"entry",
".",
"Creates",
"it",
"if",
"it",
"doesn",
"t",
"exist",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1435-L1478 |
230,689 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB.get_last_block | def get_last_block(self, cur=None):
"""
Get the highest block last processed
"""
sql = 'SELECT MAX(block_height) FROM {};'.format(self.subdomain_table)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
rows = db_query_execute(cursor, sql, ())
height = 0
try:
rowdata = rows.fetchone()
height = rowdata['MAX(block_height)']
except:
height = 0
return height | python | def get_last_block(self, cur=None):
"""
Get the highest block last processed
"""
sql = 'SELECT MAX(block_height) FROM {};'.format(self.subdomain_table)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
rows = db_query_execute(cursor, sql, ())
height = 0
try:
rowdata = rows.fetchone()
height = rowdata['MAX(block_height)']
except:
height = 0
return height | [
"def",
"get_last_block",
"(",
"self",
",",
"cur",
"=",
"None",
")",
":",
"sql",
"=",
"'SELECT MAX(block_height) FROM {};'",
".",
"format",
"(",
"self",
".",
"subdomain_table",
")",
"cursor",
"=",
"None",
"if",
"cur",
"is",
"None",
":",
"cursor",
"=",
"self",
".",
"conn",
".",
"cursor",
"(",
")",
"else",
":",
"cursor",
"=",
"cur",
"rows",
"=",
"db_query_execute",
"(",
"cursor",
",",
"sql",
",",
"(",
")",
")",
"height",
"=",
"0",
"try",
":",
"rowdata",
"=",
"rows",
".",
"fetchone",
"(",
")",
"height",
"=",
"rowdata",
"[",
"'MAX(block_height)'",
"]",
"except",
":",
"height",
"=",
"0",
"return",
"height"
] | Get the highest block last processed | [
"Get",
"the",
"highest",
"block",
"last",
"processed"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1513-L1532 |
230,690 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB.get_last_sequence | def get_last_sequence(self, cur=None):
"""
Get the highest sequence number in this db
"""
sql = 'SELECT sequence FROM {} ORDER BY sequence DESC LIMIT 1;'.format(self.subdomain_table)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
db_query_execute(cursor, sql, ())
last_seq = None
try:
last_seq = cursor.fetchone()[0]
except:
last_seq = 0
return int(last_seq) | python | def get_last_sequence(self, cur=None):
"""
Get the highest sequence number in this db
"""
sql = 'SELECT sequence FROM {} ORDER BY sequence DESC LIMIT 1;'.format(self.subdomain_table)
cursor = None
if cur is None:
cursor = self.conn.cursor()
else:
cursor = cur
db_query_execute(cursor, sql, ())
last_seq = None
try:
last_seq = cursor.fetchone()[0]
except:
last_seq = 0
return int(last_seq) | [
"def",
"get_last_sequence",
"(",
"self",
",",
"cur",
"=",
"None",
")",
":",
"sql",
"=",
"'SELECT sequence FROM {} ORDER BY sequence DESC LIMIT 1;'",
".",
"format",
"(",
"self",
".",
"subdomain_table",
")",
"cursor",
"=",
"None",
"if",
"cur",
"is",
"None",
":",
"cursor",
"=",
"self",
".",
"conn",
".",
"cursor",
"(",
")",
"else",
":",
"cursor",
"=",
"cur",
"db_query_execute",
"(",
"cursor",
",",
"sql",
",",
"(",
")",
")",
"last_seq",
"=",
"None",
"try",
":",
"last_seq",
"=",
"cursor",
".",
"fetchone",
"(",
")",
"[",
"0",
"]",
"except",
":",
"last_seq",
"=",
"0",
"return",
"int",
"(",
"last_seq",
")"
] | Get the highest sequence number in this db | [
"Get",
"the",
"highest",
"sequence",
"number",
"in",
"this",
"db"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1535-L1553 |
230,691 | blockstack/blockstack-core | blockstack/lib/subdomains.py | SubdomainDB._drop_tables | def _drop_tables(self):
"""
Clear the subdomain db's tables
"""
drop_cmd = "DROP TABLE IF EXISTS {};"
for table in [self.subdomain_table, self.blocked_table]:
cursor = self.conn.cursor()
db_query_execute(cursor, drop_cmd.format(table), ()) | python | def _drop_tables(self):
"""
Clear the subdomain db's tables
"""
drop_cmd = "DROP TABLE IF EXISTS {};"
for table in [self.subdomain_table, self.blocked_table]:
cursor = self.conn.cursor()
db_query_execute(cursor, drop_cmd.format(table), ()) | [
"def",
"_drop_tables",
"(",
"self",
")",
":",
"drop_cmd",
"=",
"\"DROP TABLE IF EXISTS {};\"",
"for",
"table",
"in",
"[",
"self",
".",
"subdomain_table",
",",
"self",
".",
"blocked_table",
"]",
":",
"cursor",
"=",
"self",
".",
"conn",
".",
"cursor",
"(",
")",
"db_query_execute",
"(",
"cursor",
",",
"drop_cmd",
".",
"format",
"(",
"table",
")",
",",
"(",
")",
")"
] | Clear the subdomain db's tables | [
"Clear",
"the",
"subdomain",
"db",
"s",
"tables"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/subdomains.py#L1556-L1563 |
230,692 | blockstack/blockstack-core | blockstack/lib/hashing.py | hash_name | def hash_name(name, script_pubkey, register_addr=None):
"""
Generate the hash over a name and hex-string script pubkey
"""
bin_name = b40_to_bin(name)
name_and_pubkey = bin_name + unhexlify(script_pubkey)
if register_addr is not None:
name_and_pubkey += str(register_addr)
return hex_hash160(name_and_pubkey) | python | def hash_name(name, script_pubkey, register_addr=None):
"""
Generate the hash over a name and hex-string script pubkey
"""
bin_name = b40_to_bin(name)
name_and_pubkey = bin_name + unhexlify(script_pubkey)
if register_addr is not None:
name_and_pubkey += str(register_addr)
return hex_hash160(name_and_pubkey) | [
"def",
"hash_name",
"(",
"name",
",",
"script_pubkey",
",",
"register_addr",
"=",
"None",
")",
":",
"bin_name",
"=",
"b40_to_bin",
"(",
"name",
")",
"name_and_pubkey",
"=",
"bin_name",
"+",
"unhexlify",
"(",
"script_pubkey",
")",
"if",
"register_addr",
"is",
"not",
"None",
":",
"name_and_pubkey",
"+=",
"str",
"(",
"register_addr",
")",
"return",
"hex_hash160",
"(",
"name_and_pubkey",
")"
] | Generate the hash over a name and hex-string script pubkey | [
"Generate",
"the",
"hash",
"over",
"a",
"name",
"and",
"hex",
"-",
"string",
"script",
"pubkey"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/hashing.py#L32-L42 |
230,693 | blockstack/blockstack-core | api/search/basic_index.py | fetch_profile_data_from_file | def fetch_profile_data_from_file():
""" takes profile data from file and saves in the profile_data DB
"""
with open(SEARCH_PROFILE_DATA_FILE, 'r') as fin:
profiles = json.load(fin)
counter = 0
log.debug("-" * 5)
log.debug("Fetching profile data from file")
for entry in profiles:
new_entry = {}
new_entry['key'] = entry['fqu']
new_entry['value'] = entry['profile']
try:
clean_profile_entries(entry['profile'])
profile_data.save(new_entry)
except Exception as e:
log.exception(e)
log.error("Exception on entry {}".format(new_entry))
counter += 1
if counter % 1000 == 0:
log.debug("Processed entries: %s" % counter)
profile_data.ensure_index('key')
return | python | def fetch_profile_data_from_file():
""" takes profile data from file and saves in the profile_data DB
"""
with open(SEARCH_PROFILE_DATA_FILE, 'r') as fin:
profiles = json.load(fin)
counter = 0
log.debug("-" * 5)
log.debug("Fetching profile data from file")
for entry in profiles:
new_entry = {}
new_entry['key'] = entry['fqu']
new_entry['value'] = entry['profile']
try:
clean_profile_entries(entry['profile'])
profile_data.save(new_entry)
except Exception as e:
log.exception(e)
log.error("Exception on entry {}".format(new_entry))
counter += 1
if counter % 1000 == 0:
log.debug("Processed entries: %s" % counter)
profile_data.ensure_index('key')
return | [
"def",
"fetch_profile_data_from_file",
"(",
")",
":",
"with",
"open",
"(",
"SEARCH_PROFILE_DATA_FILE",
",",
"'r'",
")",
"as",
"fin",
":",
"profiles",
"=",
"json",
".",
"load",
"(",
"fin",
")",
"counter",
"=",
"0",
"log",
".",
"debug",
"(",
"\"-\"",
"*",
"5",
")",
"log",
".",
"debug",
"(",
"\"Fetching profile data from file\"",
")",
"for",
"entry",
"in",
"profiles",
":",
"new_entry",
"=",
"{",
"}",
"new_entry",
"[",
"'key'",
"]",
"=",
"entry",
"[",
"'fqu'",
"]",
"new_entry",
"[",
"'value'",
"]",
"=",
"entry",
"[",
"'profile'",
"]",
"try",
":",
"clean_profile_entries",
"(",
"entry",
"[",
"'profile'",
"]",
")",
"profile_data",
".",
"save",
"(",
"new_entry",
")",
"except",
"Exception",
"as",
"e",
":",
"log",
".",
"exception",
"(",
"e",
")",
"log",
".",
"error",
"(",
"\"Exception on entry {}\"",
".",
"format",
"(",
"new_entry",
")",
")",
"counter",
"+=",
"1",
"if",
"counter",
"%",
"1000",
"==",
"0",
":",
"log",
".",
"debug",
"(",
"\"Processed entries: %s\"",
"%",
"counter",
")",
"profile_data",
".",
"ensure_index",
"(",
"'key'",
")",
"return"
] | takes profile data from file and saves in the profile_data DB | [
"takes",
"profile",
"data",
"from",
"file",
"and",
"saves",
"in",
"the",
"profile_data",
"DB"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/api/search/basic_index.py#L65-L95 |
230,694 | blockstack/blockstack-core | api/search/basic_index.py | create_search_index | def create_search_index():
""" takes people names from blockchain and writes deduped names in a 'cache'
"""
# create people name cache
counter = 0
people_names = []
twitter_handles = []
usernames = []
log.debug("-" * 5)
log.debug("Creating search index")
for user in namespace.find():
# the profile/info to be inserted
search_profile = {}
counter += 1
if(counter % 1000 == 0):
log.debug("Processed entries: %s" % counter)
if validUsername(user['username']):
pass
else:
continue
profile = get_json(user['profile'])
hasBazaarId=False
# search for openbazaar id in the profile
if 'account' in profile:
for accounts in profile['account']:
if accounts['service'] == 'openbazaar':
hasBazaarId = True
search_profile['openbazaar']=accounts['identifier']
if (hasBazaarId == False):
search_profile['openbazaar'] = None
if 'name' in profile:
try:
name = profile['name']
except:
continue
try:
name = name['formatted'].lower()
except:
name = name.lower()
people_names.append(name)
search_profile['name'] = name
else:
search_profile['name'] = None
if 'twitter' in profile:
twitter_handle = profile['twitter']
try:
twitter_handle = twitter_handle['username'].lower()
except:
try:
twitter_handle = profile['twitter'].lower()
except:
continue
twitter_handles.append(twitter_handle)
search_profile['twitter_handle'] = twitter_handle
else:
search_profile['twitter_handle'] = None
search_profile['fullyQualifiedName'] = user['fqu']
search_profile['username'] = user['username']
usernames.append(user['fqu'])
search_profile['profile'] = profile
search_profiles.save(search_profile)
# dedup names
people_names = list(set(people_names))
people_names = {'name': people_names}
twitter_handles = list(set(twitter_handles))
twitter_handles = {'twitter_handle': twitter_handles}
usernames = list(set(usernames))
usernames = {'username': usernames}
# save final dedup results to mongodb (using it as a cache)
people_cache.save(people_names)
twitter_cache.save(twitter_handles)
username_cache.save(usernames)
optimize_db()
log.debug('Created name/twitter/username search index') | python | def create_search_index():
""" takes people names from blockchain and writes deduped names in a 'cache'
"""
# create people name cache
counter = 0
people_names = []
twitter_handles = []
usernames = []
log.debug("-" * 5)
log.debug("Creating search index")
for user in namespace.find():
# the profile/info to be inserted
search_profile = {}
counter += 1
if(counter % 1000 == 0):
log.debug("Processed entries: %s" % counter)
if validUsername(user['username']):
pass
else:
continue
profile = get_json(user['profile'])
hasBazaarId=False
# search for openbazaar id in the profile
if 'account' in profile:
for accounts in profile['account']:
if accounts['service'] == 'openbazaar':
hasBazaarId = True
search_profile['openbazaar']=accounts['identifier']
if (hasBazaarId == False):
search_profile['openbazaar'] = None
if 'name' in profile:
try:
name = profile['name']
except:
continue
try:
name = name['formatted'].lower()
except:
name = name.lower()
people_names.append(name)
search_profile['name'] = name
else:
search_profile['name'] = None
if 'twitter' in profile:
twitter_handle = profile['twitter']
try:
twitter_handle = twitter_handle['username'].lower()
except:
try:
twitter_handle = profile['twitter'].lower()
except:
continue
twitter_handles.append(twitter_handle)
search_profile['twitter_handle'] = twitter_handle
else:
search_profile['twitter_handle'] = None
search_profile['fullyQualifiedName'] = user['fqu']
search_profile['username'] = user['username']
usernames.append(user['fqu'])
search_profile['profile'] = profile
search_profiles.save(search_profile)
# dedup names
people_names = list(set(people_names))
people_names = {'name': people_names}
twitter_handles = list(set(twitter_handles))
twitter_handles = {'twitter_handle': twitter_handles}
usernames = list(set(usernames))
usernames = {'username': usernames}
# save final dedup results to mongodb (using it as a cache)
people_cache.save(people_names)
twitter_cache.save(twitter_handles)
username_cache.save(usernames)
optimize_db()
log.debug('Created name/twitter/username search index') | [
"def",
"create_search_index",
"(",
")",
":",
"# create people name cache",
"counter",
"=",
"0",
"people_names",
"=",
"[",
"]",
"twitter_handles",
"=",
"[",
"]",
"usernames",
"=",
"[",
"]",
"log",
".",
"debug",
"(",
"\"-\"",
"*",
"5",
")",
"log",
".",
"debug",
"(",
"\"Creating search index\"",
")",
"for",
"user",
"in",
"namespace",
".",
"find",
"(",
")",
":",
"# the profile/info to be inserted",
"search_profile",
"=",
"{",
"}",
"counter",
"+=",
"1",
"if",
"(",
"counter",
"%",
"1000",
"==",
"0",
")",
":",
"log",
".",
"debug",
"(",
"\"Processed entries: %s\"",
"%",
"counter",
")",
"if",
"validUsername",
"(",
"user",
"[",
"'username'",
"]",
")",
":",
"pass",
"else",
":",
"continue",
"profile",
"=",
"get_json",
"(",
"user",
"[",
"'profile'",
"]",
")",
"hasBazaarId",
"=",
"False",
"# search for openbazaar id in the profile",
"if",
"'account'",
"in",
"profile",
":",
"for",
"accounts",
"in",
"profile",
"[",
"'account'",
"]",
":",
"if",
"accounts",
"[",
"'service'",
"]",
"==",
"'openbazaar'",
":",
"hasBazaarId",
"=",
"True",
"search_profile",
"[",
"'openbazaar'",
"]",
"=",
"accounts",
"[",
"'identifier'",
"]",
"if",
"(",
"hasBazaarId",
"==",
"False",
")",
":",
"search_profile",
"[",
"'openbazaar'",
"]",
"=",
"None",
"if",
"'name'",
"in",
"profile",
":",
"try",
":",
"name",
"=",
"profile",
"[",
"'name'",
"]",
"except",
":",
"continue",
"try",
":",
"name",
"=",
"name",
"[",
"'formatted'",
"]",
".",
"lower",
"(",
")",
"except",
":",
"name",
"=",
"name",
".",
"lower",
"(",
")",
"people_names",
".",
"append",
"(",
"name",
")",
"search_profile",
"[",
"'name'",
"]",
"=",
"name",
"else",
":",
"search_profile",
"[",
"'name'",
"]",
"=",
"None",
"if",
"'twitter'",
"in",
"profile",
":",
"twitter_handle",
"=",
"profile",
"[",
"'twitter'",
"]",
"try",
":",
"twitter_handle",
"=",
"twitter_handle",
"[",
"'username'",
"]",
".",
"lower",
"(",
")",
"except",
":",
"try",
":",
"twitter_handle",
"=",
"profile",
"[",
"'twitter'",
"]",
".",
"lower",
"(",
")",
"except",
":",
"continue",
"twitter_handles",
".",
"append",
"(",
"twitter_handle",
")",
"search_profile",
"[",
"'twitter_handle'",
"]",
"=",
"twitter_handle",
"else",
":",
"search_profile",
"[",
"'twitter_handle'",
"]",
"=",
"None",
"search_profile",
"[",
"'fullyQualifiedName'",
"]",
"=",
"user",
"[",
"'fqu'",
"]",
"search_profile",
"[",
"'username'",
"]",
"=",
"user",
"[",
"'username'",
"]",
"usernames",
".",
"append",
"(",
"user",
"[",
"'fqu'",
"]",
")",
"search_profile",
"[",
"'profile'",
"]",
"=",
"profile",
"search_profiles",
".",
"save",
"(",
"search_profile",
")",
"# dedup names",
"people_names",
"=",
"list",
"(",
"set",
"(",
"people_names",
")",
")",
"people_names",
"=",
"{",
"'name'",
":",
"people_names",
"}",
"twitter_handles",
"=",
"list",
"(",
"set",
"(",
"twitter_handles",
")",
")",
"twitter_handles",
"=",
"{",
"'twitter_handle'",
":",
"twitter_handles",
"}",
"usernames",
"=",
"list",
"(",
"set",
"(",
"usernames",
")",
")",
"usernames",
"=",
"{",
"'username'",
":",
"usernames",
"}",
"# save final dedup results to mongodb (using it as a cache)",
"people_cache",
".",
"save",
"(",
"people_names",
")",
"twitter_cache",
".",
"save",
"(",
"twitter_handles",
")",
"username_cache",
".",
"save",
"(",
"usernames",
")",
"optimize_db",
"(",
")",
"log",
".",
"debug",
"(",
"'Created name/twitter/username search index'",
")"
] | takes people names from blockchain and writes deduped names in a 'cache' | [
"takes",
"people",
"names",
"from",
"blockchain",
"and",
"writes",
"deduped",
"names",
"in",
"a",
"cache"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/api/search/basic_index.py#L177-L277 |
230,695 | blockstack/blockstack-core | blockstack/lib/operations/__init__.py | op_extract | def op_extract(op_name, data, senders, inputs, outputs, block_id, vtxindex, txid):
"""
Extract an operation from transaction data.
Return the extracted fields as a dict.
"""
global EXTRACT_METHODS
if op_name not in EXTRACT_METHODS.keys():
raise Exception("No such operation '%s'" % op_name)
method = EXTRACT_METHODS[op_name]
op_data = method( data, senders, inputs, outputs, block_id, vtxindex, txid )
return op_data | python | def op_extract(op_name, data, senders, inputs, outputs, block_id, vtxindex, txid):
"""
Extract an operation from transaction data.
Return the extracted fields as a dict.
"""
global EXTRACT_METHODS
if op_name not in EXTRACT_METHODS.keys():
raise Exception("No such operation '%s'" % op_name)
method = EXTRACT_METHODS[op_name]
op_data = method( data, senders, inputs, outputs, block_id, vtxindex, txid )
return op_data | [
"def",
"op_extract",
"(",
"op_name",
",",
"data",
",",
"senders",
",",
"inputs",
",",
"outputs",
",",
"block_id",
",",
"vtxindex",
",",
"txid",
")",
":",
"global",
"EXTRACT_METHODS",
"if",
"op_name",
"not",
"in",
"EXTRACT_METHODS",
".",
"keys",
"(",
")",
":",
"raise",
"Exception",
"(",
"\"No such operation '%s'\"",
"%",
"op_name",
")",
"method",
"=",
"EXTRACT_METHODS",
"[",
"op_name",
"]",
"op_data",
"=",
"method",
"(",
"data",
",",
"senders",
",",
"inputs",
",",
"outputs",
",",
"block_id",
",",
"vtxindex",
",",
"txid",
")",
"return",
"op_data"
] | Extract an operation from transaction data.
Return the extracted fields as a dict. | [
"Extract",
"an",
"operation",
"from",
"transaction",
"data",
".",
"Return",
"the",
"extracted",
"fields",
"as",
"a",
"dict",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/operations/__init__.py#L183-L195 |
230,696 | blockstack/blockstack-core | blockstack/lib/operations/__init__.py | op_canonicalize | def op_canonicalize(op_name, parsed_op):
"""
Get the canonical representation of a parsed operation's data.
Meant for backwards-compatibility
"""
global CANONICALIZE_METHODS
if op_name not in CANONICALIZE_METHODS:
# no canonicalization needed
return parsed_op
else:
return CANONICALIZE_METHODS[op_name](parsed_op) | python | def op_canonicalize(op_name, parsed_op):
"""
Get the canonical representation of a parsed operation's data.
Meant for backwards-compatibility
"""
global CANONICALIZE_METHODS
if op_name not in CANONICALIZE_METHODS:
# no canonicalization needed
return parsed_op
else:
return CANONICALIZE_METHODS[op_name](parsed_op) | [
"def",
"op_canonicalize",
"(",
"op_name",
",",
"parsed_op",
")",
":",
"global",
"CANONICALIZE_METHODS",
"if",
"op_name",
"not",
"in",
"CANONICALIZE_METHODS",
":",
"# no canonicalization needed",
"return",
"parsed_op",
"else",
":",
"return",
"CANONICALIZE_METHODS",
"[",
"op_name",
"]",
"(",
"parsed_op",
")"
] | Get the canonical representation of a parsed operation's data.
Meant for backwards-compatibility | [
"Get",
"the",
"canonical",
"representation",
"of",
"a",
"parsed",
"operation",
"s",
"data",
".",
"Meant",
"for",
"backwards",
"-",
"compatibility"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/operations/__init__.py#L198-L209 |
230,697 | blockstack/blockstack-core | blockstack/lib/operations/__init__.py | op_decanonicalize | def op_decanonicalize(op_name, canonical_op):
"""
Get the current representation of a parsed operation's data, given the canonical representation
Meant for backwards-compatibility
"""
global DECANONICALIZE_METHODS
if op_name not in DECANONICALIZE_METHODS:
# no decanonicalization needed
return canonical_op
else:
return DECANONICALIZE_METHODS[op_name](canonical_op) | python | def op_decanonicalize(op_name, canonical_op):
"""
Get the current representation of a parsed operation's data, given the canonical representation
Meant for backwards-compatibility
"""
global DECANONICALIZE_METHODS
if op_name not in DECANONICALIZE_METHODS:
# no decanonicalization needed
return canonical_op
else:
return DECANONICALIZE_METHODS[op_name](canonical_op) | [
"def",
"op_decanonicalize",
"(",
"op_name",
",",
"canonical_op",
")",
":",
"global",
"DECANONICALIZE_METHODS",
"if",
"op_name",
"not",
"in",
"DECANONICALIZE_METHODS",
":",
"# no decanonicalization needed",
"return",
"canonical_op",
"else",
":",
"return",
"DECANONICALIZE_METHODS",
"[",
"op_name",
"]",
"(",
"canonical_op",
")"
] | Get the current representation of a parsed operation's data, given the canonical representation
Meant for backwards-compatibility | [
"Get",
"the",
"current",
"representation",
"of",
"a",
"parsed",
"operation",
"s",
"data",
"given",
"the",
"canonical",
"representation",
"Meant",
"for",
"backwards",
"-",
"compatibility"
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/operations/__init__.py#L212-L223 |
230,698 | blockstack/blockstack-core | blockstack/lib/operations/__init__.py | op_check | def op_check( state_engine, nameop, block_id, checked_ops ):
"""
Given the state engine, the current block, the list of pending
operations processed so far, and the current operation, determine
whether or not it should be accepted.
The operation is allowed to be "type-cast" to a new operation, but only once.
If this happens, the operation will be checked again.
Subsequent casts are considered bugs, and will cause a program abort.
TODO: remove type-cast
"""
global CHECK_METHODS, MUTATE_FIELDS
nameop_clone = copy.deepcopy( nameop )
opcode = None
if 'opcode' not in nameop_clone.keys():
op = nameop_clone.get('op', None)
try:
assert op is not None, "BUG: no op defined"
opcode = op_get_opcode_name( op )
assert opcode is not None, "BUG: op '%s' undefined" % op
except Exception, e:
log.exception(e)
log.error("FATAL: BUG: no 'op' defined")
sys.exit(1)
else:
opcode = nameop_clone['opcode']
check_method = CHECK_METHODS.get( opcode, None )
try:
assert check_method is not None, "BUG: no check-method for '%s'" % opcode
except Exception, e:
log.exception(e)
log.error("FATAL: BUG: no check-method for '%s'" % opcode )
sys.exit(1)
rc = check_method( state_engine, nameop_clone, block_id, checked_ops )
if not rc:
# rejected
return False
# accepted! clean up and canonicalize
nameop.clear()
nameop.update(nameop_clone)
# nameop = op_canonicalize(nameop['opcode'], nameop)
op_canonicalize(nameop['opcode'], nameop)
# make sure we don't send unstored fields to the db that are otherwise canonical
unstored_canonical_fields = UNSTORED_CANONICAL_FIELDS.get(nameop['opcode'])
assert unstored_canonical_fields is not None, "BUG: no UNSTORED_CANONICAL_FIELDS entry for {}".format(nameop['opcode'])
for f in unstored_canonical_fields:
if f in nameop:
del nameop[f]
return rc | python | def op_check( state_engine, nameop, block_id, checked_ops ):
"""
Given the state engine, the current block, the list of pending
operations processed so far, and the current operation, determine
whether or not it should be accepted.
The operation is allowed to be "type-cast" to a new operation, but only once.
If this happens, the operation will be checked again.
Subsequent casts are considered bugs, and will cause a program abort.
TODO: remove type-cast
"""
global CHECK_METHODS, MUTATE_FIELDS
nameop_clone = copy.deepcopy( nameop )
opcode = None
if 'opcode' not in nameop_clone.keys():
op = nameop_clone.get('op', None)
try:
assert op is not None, "BUG: no op defined"
opcode = op_get_opcode_name( op )
assert opcode is not None, "BUG: op '%s' undefined" % op
except Exception, e:
log.exception(e)
log.error("FATAL: BUG: no 'op' defined")
sys.exit(1)
else:
opcode = nameop_clone['opcode']
check_method = CHECK_METHODS.get( opcode, None )
try:
assert check_method is not None, "BUG: no check-method for '%s'" % opcode
except Exception, e:
log.exception(e)
log.error("FATAL: BUG: no check-method for '%s'" % opcode )
sys.exit(1)
rc = check_method( state_engine, nameop_clone, block_id, checked_ops )
if not rc:
# rejected
return False
# accepted! clean up and canonicalize
nameop.clear()
nameop.update(nameop_clone)
# nameop = op_canonicalize(nameop['opcode'], nameop)
op_canonicalize(nameop['opcode'], nameop)
# make sure we don't send unstored fields to the db that are otherwise canonical
unstored_canonical_fields = UNSTORED_CANONICAL_FIELDS.get(nameop['opcode'])
assert unstored_canonical_fields is not None, "BUG: no UNSTORED_CANONICAL_FIELDS entry for {}".format(nameop['opcode'])
for f in unstored_canonical_fields:
if f in nameop:
del nameop[f]
return rc | [
"def",
"op_check",
"(",
"state_engine",
",",
"nameop",
",",
"block_id",
",",
"checked_ops",
")",
":",
"global",
"CHECK_METHODS",
",",
"MUTATE_FIELDS",
"nameop_clone",
"=",
"copy",
".",
"deepcopy",
"(",
"nameop",
")",
"opcode",
"=",
"None",
"if",
"'opcode'",
"not",
"in",
"nameop_clone",
".",
"keys",
"(",
")",
":",
"op",
"=",
"nameop_clone",
".",
"get",
"(",
"'op'",
",",
"None",
")",
"try",
":",
"assert",
"op",
"is",
"not",
"None",
",",
"\"BUG: no op defined\"",
"opcode",
"=",
"op_get_opcode_name",
"(",
"op",
")",
"assert",
"opcode",
"is",
"not",
"None",
",",
"\"BUG: op '%s' undefined\"",
"%",
"op",
"except",
"Exception",
",",
"e",
":",
"log",
".",
"exception",
"(",
"e",
")",
"log",
".",
"error",
"(",
"\"FATAL: BUG: no 'op' defined\"",
")",
"sys",
".",
"exit",
"(",
"1",
")",
"else",
":",
"opcode",
"=",
"nameop_clone",
"[",
"'opcode'",
"]",
"check_method",
"=",
"CHECK_METHODS",
".",
"get",
"(",
"opcode",
",",
"None",
")",
"try",
":",
"assert",
"check_method",
"is",
"not",
"None",
",",
"\"BUG: no check-method for '%s'\"",
"%",
"opcode",
"except",
"Exception",
",",
"e",
":",
"log",
".",
"exception",
"(",
"e",
")",
"log",
".",
"error",
"(",
"\"FATAL: BUG: no check-method for '%s'\"",
"%",
"opcode",
")",
"sys",
".",
"exit",
"(",
"1",
")",
"rc",
"=",
"check_method",
"(",
"state_engine",
",",
"nameop_clone",
",",
"block_id",
",",
"checked_ops",
")",
"if",
"not",
"rc",
":",
"# rejected",
"return",
"False",
"# accepted! clean up and canonicalize",
"nameop",
".",
"clear",
"(",
")",
"nameop",
".",
"update",
"(",
"nameop_clone",
")",
"# nameop = op_canonicalize(nameop['opcode'], nameop)",
"op_canonicalize",
"(",
"nameop",
"[",
"'opcode'",
"]",
",",
"nameop",
")",
"# make sure we don't send unstored fields to the db that are otherwise canonical",
"unstored_canonical_fields",
"=",
"UNSTORED_CANONICAL_FIELDS",
".",
"get",
"(",
"nameop",
"[",
"'opcode'",
"]",
")",
"assert",
"unstored_canonical_fields",
"is",
"not",
"None",
",",
"\"BUG: no UNSTORED_CANONICAL_FIELDS entry for {}\"",
".",
"format",
"(",
"nameop",
"[",
"'opcode'",
"]",
")",
"for",
"f",
"in",
"unstored_canonical_fields",
":",
"if",
"f",
"in",
"nameop",
":",
"del",
"nameop",
"[",
"f",
"]",
"return",
"rc"
] | Given the state engine, the current block, the list of pending
operations processed so far, and the current operation, determine
whether or not it should be accepted.
The operation is allowed to be "type-cast" to a new operation, but only once.
If this happens, the operation will be checked again.
Subsequent casts are considered bugs, and will cause a program abort.
TODO: remove type-cast | [
"Given",
"the",
"state",
"engine",
"the",
"current",
"block",
"the",
"list",
"of",
"pending",
"operations",
"processed",
"so",
"far",
"and",
"the",
"current",
"operation",
"determine",
"whether",
"or",
"not",
"it",
"should",
"be",
"accepted",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/operations/__init__.py#L284-L343 |
230,699 | blockstack/blockstack-core | blockstack/lib/operations/__init__.py | op_get_mutate_fields | def op_get_mutate_fields( op_name ):
"""
Get the names of the fields that will change
when this operation gets applied to a record.
"""
global MUTATE_FIELDS
if op_name not in MUTATE_FIELDS.keys():
raise Exception("No such operation '%s'" % op_name)
fields = MUTATE_FIELDS[op_name][:]
return fields | python | def op_get_mutate_fields( op_name ):
"""
Get the names of the fields that will change
when this operation gets applied to a record.
"""
global MUTATE_FIELDS
if op_name not in MUTATE_FIELDS.keys():
raise Exception("No such operation '%s'" % op_name)
fields = MUTATE_FIELDS[op_name][:]
return fields | [
"def",
"op_get_mutate_fields",
"(",
"op_name",
")",
":",
"global",
"MUTATE_FIELDS",
"if",
"op_name",
"not",
"in",
"MUTATE_FIELDS",
".",
"keys",
"(",
")",
":",
"raise",
"Exception",
"(",
"\"No such operation '%s'\"",
"%",
"op_name",
")",
"fields",
"=",
"MUTATE_FIELDS",
"[",
"op_name",
"]",
"[",
":",
"]",
"return",
"fields"
] | Get the names of the fields that will change
when this operation gets applied to a record. | [
"Get",
"the",
"names",
"of",
"the",
"fields",
"that",
"will",
"change",
"when",
"this",
"operation",
"gets",
"applied",
"to",
"a",
"record",
"."
] | 1dcfdd39b152d29ce13e736a6a1a0981401a0505 | https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/operations/__init__.py#L346-L357 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.