_id
stringlengths
2
6
title
stringlengths
9
130
partition
stringclasses
3 values
text
stringlengths
66
10.5k
language
stringclasses
1 value
meta_information
dict
q18100
CloudCrowd.NodeRecord.send_work_unit
train
def send_work_unit(unit) result = node['/work'].post(:work_unit => unit.to_json) unit.assign_to(self, JSON.parse(result.body)['pid']) touch && true rescue RestClient::RequestTimeout # The node's gone away. Destroy it and it will check in when it comes back CloudCrowd.log "Node #{host}...
ruby
{ "resource": "" }
q18101
CloudCrowd.WorkUnit.fail
train
def fail(output, time_taken) tries = self.attempts + 1 return try_again if tries < CloudCrowd.config[:work_unit_retries] update_attributes({ :status => FAILED, :node_record => nil, :worker_pid => nil, :attempts => tries, :output => o...
ruby
{ "resource": "" }
q18102
CloudCrowd.Worker.complete_work_unit
train
def complete_work_unit(result) keep_trying_to "complete work unit" do data = base_params.merge({:status => 'succeeded', :output => result}) @node.central["/work/#{data[:id]}"].put(data) log "finished #{display_work_unit} in #{data[:time]} seconds" end end
ruby
{ "resource": "" }
q18103
CloudCrowd.Worker.fail_work_unit
train
def fail_work_unit(exception) keep_trying_to "mark work unit as failed" do data = base_params.merge({:status => 'failed', :output => {'output' => exception.message}.to_json}) @node.central["/work/#{data[:id]}"].put(data) log "failed #{display_work_unit} in #{data[:time]} seconds\n#{excepti...
ruby
{ "resource": "" }
q18104
CloudCrowd.Worker.keep_trying_to
train
def keep_trying_to(title) begin yield rescue RestClient::ResourceNotFound => e log "work unit ##{@unit['id']} doesn't exist. discarding..." rescue Exception => e log "failed to #{title} -- retry in #{@retry_wait} seconds" log e.message log e.backtrace sl...
ruby
{ "resource": "" }
q18105
CloudCrowd.Worker.run_work_unit
train
def run_work_unit begin result = nil action_class = CloudCrowd.actions[@unit['action']] action = action_class.new(@status, @unit['input'], enhanced_unit_options, @node.asset_store) Dir.chdir(action.work_directory) do result = case @status when PROCESSING then ac...
ruby
{ "resource": "" }
q18106
CloudCrowd.Worker.run
train
def run trap_signals log "starting #{display_work_unit}" if @unit['options']['benchmark'] log("ran #{display_work_unit} in " + Benchmark.measure { run_work_unit }.to_s) else run_work_unit end Process.exit! end
ruby
{ "resource": "" }
q18107
CloudCrowd.Worker.trap_signals
train
def trap_signals Signal.trap('QUIT') { Process.exit! } Signal.trap('INT') { Process.exit! } Signal.trap('TERM') { Process.exit! } end
ruby
{ "resource": "" }
q18108
CloudCrowd.CommandLine.start_server
train
def start_server port = @options[:port] || 9173 daemonize = @options[:daemonize] ? '-d' : '' log_path = CloudCrowd.log_path('server.log') pid_path = CloudCrowd.pid_path('server.pid') rackup_path = File.expand_path("#{@options[:config_path]}/config.ru") FileUtils.mkdir_...
ruby
{ "resource": "" }
q18109
CloudCrowd.CommandLine.run_install
train
def run_install(install_path) require 'fileutils' install_path ||= '.' FileUtils.mkdir_p install_path unless File.exists?(install_path) install_file "#{CC_ROOT}/config/config.example.yml", "#{install_path}/config.yml" install_file "#{CC_ROOT}/config/config.example.ru", "#{install_path}/con...
ruby
{ "resource": "" }
q18110
CloudCrowd.CommandLine.ensure_config
train
def ensure_config return if @config_found found = CONFIG_FILES.all? {|f| File.exists? "#{@options[:config_path]}/#{f}" } found ? @config_dir = true : config_not_found end
ruby
{ "resource": "" }
q18111
CloudCrowd.CommandLine.install_file
train
def install_file(source, dest, is_dir=false) if File.exists?(dest) print "#{dest} already exists. Overwrite it? (yes/no) " return unless ['y', 'yes', 'ok'].include? gets.chomp.downcase end is_dir ? FileUtils.cp_r(source, dest) : FileUtils.cp(source, dest) puts "installed #{dest}"...
ruby
{ "resource": "" }
q18112
CloudCrowd.Job.check_for_completion
train
def check_for_completion return unless all_work_units_complete? set_next_status outs = gather_outputs_from_work_units return queue_for_workers([outs]) if merging? if complete? update_attributes(:outputs => outs, :time => time_taken) CloudCrowd.log "Job ##{id} (#{action}) #{...
ruby
{ "resource": "" }
q18113
CloudCrowd.Job.percent_complete
train
def percent_complete return 99 if merging? return 100 if complete? unit_count = work_units.count return 100 if unit_count <= 0 (work_units.complete.count / unit_count.to_f * 100).round end
ruby
{ "resource": "" }
q18114
CloudCrowd.Job.as_json
train
def as_json(opts={}) atts = { 'id' => id, 'color' => color, 'status' => display_status, 'percent_complete' => percent_complete, 'work_units' => work_units.count, 'time_taken' => time_taken } atts['outp...
ruby
{ "resource": "" }
q18115
CloudCrowd.Job.gather_outputs_from_work_units
train
def gather_outputs_from_work_units units = self.work_units.complete outs = self.work_units.complete.map {|u| u.parsed_output } self.work_units.complete.destroy_all outs.to_json end
ruby
{ "resource": "" }
q18116
CloudCrowd.Job.queue_for_workers
train
def queue_for_workers(input=nil) input ||= JSON.parse(self.inputs) input.each {|i| WorkUnit.start(self, action, i, status) } self end
ruby
{ "resource": "" }
q18117
CloudCrowd.Node.start
train
def start FileUtils.mkdir_p(CloudCrowd.log_path) if @daemon && !File.exists?(CloudCrowd.log_path) @server = Thin::Server.new('0.0.0.0', @port, self, :signals => false) @server.tag = 'cloud-crowd-node' @server.pid_file = CloudCrowd.pid_path('node.pid') @server.log_file = Cloud...
ruby
{ "resource": "" }
q18118
CloudCrowd.Node.check_in
train
def check_in(critical=false) @central["/node/#{@id}"].put( :busy => @overloaded, :tag => @tag, :max_workers => CloudCrowd.config[:max_workers], :enabled_actions => @enabled_actions.join(',') ) rescue RestClient::Exception, Errno::ECONNREFUSE...
ruby
{ "resource": "" }
q18119
CloudCrowd.Node.free_memory
train
def free_memory case RUBY_PLATFORM when /darwin/ stats = `vm_stat` @mac_page_size ||= stats.match(SCRAPE_MAC_PAGE)[1].to_f / 1048576.0 stats.match(SCRAPE_MAC_MEMORY)[1].to_f * @mac_page_size when /linux/ `cat /proc/meminfo`.match(SCRAPE_LINUX_MEMORY)[1].to_f / 1024.0 ...
ruby
{ "resource": "" }
q18120
CloudCrowd.Node.monitor_system
train
def monitor_system @monitor_thread = Thread.new do loop do was_overloaded = @overloaded @overloaded = overloaded? check_in if was_overloaded && !@overloaded sleep MONITOR_INTERVAL end end end
ruby
{ "resource": "" }
q18121
CloudCrowd.Node.check_in_periodically
train
def check_in_periodically @check_in_thread = Thread.new do loop do check_on_workers reply = "" 1.upto(5).each do | attempt_number | # sleep for an ever increasing amount of time to prevent overloading the server sleep CHECK_IN_INTERVAL * attempt_number...
ruby
{ "resource": "" }
q18122
CloudCrowd.Node.trap_signals
train
def trap_signals Signal.trap('QUIT') { shut_down } Signal.trap('INT') { shut_down } Signal.trap('TERM') { shut_down } end
ruby
{ "resource": "" }
q18123
TwitterJekyll.ApiClient.fetch
train
def fetch(api_request) uri = api_request.to_uri response = Net::HTTP.start(uri.host, use_ssl: api_request.ssl?) do |http| http.read_timeout = 5 http.open_timeout = 5 http.get uri.request_uri, REQUEST_HEADERS end handle_response(api_request, response) rescue Timeout:...
ruby
{ "resource": "" }
q18124
TwitterJekyll.TwitterTag.live_response
train
def live_response if response = api_client.fetch(@api_request) cache.write(@api_request.cache_key, response) build_response(response) end end
ruby
{ "resource": "" }
q18125
TwitterJekyll.TwitterTag.parse_params
train
def parse_params(params) args = params.split(/\s+/).map(&:strip) invalid_args!(args) unless args.any? if args[0].to_s == OEMBED_ARG # TODO: remove after deprecation cycle arguments_deprecation_warning(args) args.shift end url, *api_args = args ApiRequest.new(url, pa...
ruby
{ "resource": "" }
q18126
Interactor.Contracts.enforce_contracts
train
def enforce_contracts(contracts) outcome = contracts.call(context) unless outcome.success? contract.consequences.each do |handler| instance_exec(outcome.breaches, &handler) end end end
ruby
{ "resource": "" }
q18127
ActiveRemote.Persistence.remote
train
def remote(endpoint, request_args = scope_key_hash) response = remote_call(endpoint, request_args) assign_attributes_from_rpc(response) success? end
ruby
{ "resource": "" }
q18128
ActiveRemote.Attributes.read_attribute
train
def read_attribute(name) name = name.to_s if respond_to?(name) attribute(name) else raise ::ActiveRemote::UnknownAttributeError, "unknown attribute: #{name}" end end
ruby
{ "resource": "" }
q18129
ActiveRemote.Attributes.write_attribute
train
def write_attribute(name, value) name = name.to_s if respond_to?("#{name}=") __send__("attribute=", name, value) else raise ::ActiveRemote::UnknownAttributeError, "unknown attribute: #{name}" end end
ruby
{ "resource": "" }
q18130
ActiveRemote.Serialization.add_errors
train
def add_errors(errors) errors.each do |error| if error.respond_to?(:message) self.errors.add(error.field, error.message) elsif error.respond_to?(:messages) error.messages.each do |message| self.errors.add(error.field, message) end end end ...
ruby
{ "resource": "" }
q18131
ActiveRemote.Integration.cache_key
train
def cache_key case when new_record? then "#{self.class.name.underscore}/new" when ::ActiveRemote.config.default_cache_key_updated_at? && (timestamp = self[:updated_at]) then timestamp = timestamp.utc.to_s(self.class.cache_timestamp_format) "#{self.class.name.underscore}/#{self....
ruby
{ "resource": "" }
q18132
Rb1drv.OneDriveFile.save_as
train
def save_as(target_name=nil, overwrite: false, resume: true, &block) target_name ||= @name tmpfile = "#{target_name}.incomplete" return if !overwrite && File.exist?(target_name) if resume && File.size(tmpfile) > 0 from = File.size(tmpfile) len = @size - from fmode = 'ab...
ruby
{ "resource": "" }
q18133
Rb1drv.OneDriveFile.set_mtime
train
def set_mtime(time) attempt = 0 OneDriveFile.new(@od, @od.request(api_path, {fileSystemInfo: {lastModifiedDateTime: time.utc.iso8601}}, :patch)) rescue sleep 10 attempt += 1 retry if attempt <= 3 end
ruby
{ "resource": "" }
q18134
Rb1drv.OneDrive.request
train
def request(uri, data=nil, verb=:post) @logger.info(uri) if @logger auth_check query = { path: File.join('v1.0/me/', URI.escape(uri)), headers: { 'Authorization': "Bearer #{@access_token.token}" } } if data query[:body] = JSON.generate(data) ...
ruby
{ "resource": "" }
q18135
Rb1drv.OneDriveDir.get_child
train
def get_child(path) children.find { |child| child.name == path } || OneDrive404.new end
ruby
{ "resource": "" }
q18136
Rb1drv.OneDriveDir.mkdir
train
def mkdir(name) return self if name == '.' name = name[1..-1] if name[0] == '/' newdir, *remainder = name.split('/') subdir = get(newdir) unless subdir.dir? result = @od.request("#{api_path}/children", name: newdir, folder: {}, '@microsoft.graph.confli...
ruby
{ "resource": "" }
q18137
Rb1drv.OneDriveDir.upload_simple
train
def upload_simple(filename, overwrite:, target_name:) target_file = get(target_name) exist = target_file.file? return if exist && !overwrite path = nil if exist path = "#{target_file.api_path}/content" else path = "#{api_path}:/#{target_name}:/content" end ...
ruby
{ "resource": "" }
q18138
EeePub.OCF.save
train
def save(output_path) output_path = File.expand_path(output_path) create_epub do mimetype = Zip::ZipOutputStream::open(output_path) do |os| os.put_next_entry("mimetype", nil, nil, Zip::ZipEntry::STORED, Zlib::NO_COMPRESSION) os << "application/epub+zip" end zipfi...
ruby
{ "resource": "" }
q18139
EeePub.ContainerItem.to_xml
train
def to_xml out = "" builder = Builder::XmlMarkup.new(:target => out, :indent => 2) builder.instruct! build_xml(builder) out end
ruby
{ "resource": "" }
q18140
EeePub.ContainerItem.convert_to_xml_attributes
train
def convert_to_xml_attributes(hash) result = {} hash.each do |k, v| key = k.to_s.gsub('_', '-').to_sym result[key] = v end result end
ruby
{ "resource": "" }
q18141
Bitcoin.Tx.witness_commitment
train
def witness_commitment return nil unless coinbase_tx? outputs.each do |output| commitment = output.script_pubkey.witness_commitment return commitment if commitment end nil end
ruby
{ "resource": "" }
q18142
Bitcoin.Tx.serialize_old_format
train
def serialize_old_format buf = [version].pack('V') buf << Bitcoin.pack_var_int(inputs.length) << inputs.map(&:to_payload).join buf << Bitcoin.pack_var_int(outputs.length) << outputs.map(&:to_payload).join buf << [lock_time].pack('V') buf end
ruby
{ "resource": "" }
q18143
Bitcoin.Tx.standard?
train
def standard? return false if version > MAX_STANDARD_VERSION return false if weight > MAX_STANDARD_TX_WEIGHT inputs.each do |i| # Biggest 'standard' txin is a 15-of-15 P2SH multisig with compressed keys (remember the 520 byte limit on redeemScript size). # That works out to a (15*(33+1...
ruby
{ "resource": "" }
q18144
Bitcoin.Tx.sighash_for_input
train
def sighash_for_input(input_index, output_script, hash_type: SIGHASH_TYPE[:all], sig_version: :base, amount: nil, skip_separator_index: 0) raise ArgumentError, 'input_index must be specified.' unless input_index raise ArgumentError, 'does not exist input corresponding to input_inde...
ruby
{ "resource": "" }
q18145
Bitcoin.Tx.verify_input_sig
train
def verify_input_sig(input_index, script_pubkey, amount: nil, flags: STANDARD_SCRIPT_VERIFY_FLAGS) script_sig = inputs[input_index].script_sig has_witness = inputs[input_index].has_witness? if script_pubkey.p2sh? flags << SCRIPT_VERIFY_P2SH redeem_script = Script.parse_from_payload(sc...
ruby
{ "resource": "" }
q18146
Bitcoin.Tx.sighash_for_legacy
train
def sighash_for_legacy(index, script_code, hash_type) ins = inputs.map.with_index do |i, idx| if idx == index i.to_payload(script_code.delete_opcode(Bitcoin::Opcodes::OP_CODESEPARATOR)) else case hash_type & 0x1f when SIGHASH_TYPE[:none], SIGHASH_TYPE[:single] ...
ruby
{ "resource": "" }
q18147
Bitcoin.Tx.verify_input_sig_for_legacy
train
def verify_input_sig_for_legacy(input_index, script_pubkey, flags) script_sig = inputs[input_index].script_sig checker = Bitcoin::TxChecker.new(tx: self, input_index: input_index) interpreter = Bitcoin::ScriptInterpreter.new(checker: checker, flags: flags) interpreter.verify_script(script_sig, ...
ruby
{ "resource": "" }
q18148
Bitcoin.Tx.verify_input_sig_for_witness
train
def verify_input_sig_for_witness(input_index, script_pubkey, amount, flags) flags |= SCRIPT_VERIFY_WITNESS flags |= SCRIPT_VERIFY_WITNESS_PUBKEYTYPE checker = Bitcoin::TxChecker.new(tx: self, input_index: input_index, amount: amount) interpreter = Bitcoin::ScriptInterpreter.new(checker: checker,...
ruby
{ "resource": "" }
q18149
Bitcoin.ExtKey.ext_pubkey
train
def ext_pubkey k = ExtPubkey.new k.depth = depth k.number = number k.parent_fingerprint = parent_fingerprint k.chain_code = chain_code k.pubkey = key.pubkey k.ver = priv_ver_to_pub_ver k end
ruby
{ "resource": "" }
q18150
Bitcoin.ExtKey.to_payload
train
def to_payload version.htb << [depth].pack('C') << parent_fingerprint.htb << [number].pack('N') << chain_code << [0x00].pack('C') << key.priv_key.htb end
ruby
{ "resource": "" }
q18151
Bitcoin.ExtKey.to_base58
train
def to_base58 h = to_payload.bth hex = h + Bitcoin.calc_checksum(h) Base58.encode(hex) end
ruby
{ "resource": "" }
q18152
Bitcoin.ExtKey.derive
train
def derive(number, harden = false) number += HARDENED_THRESHOLD if harden new_key = ExtKey.new new_key.depth = depth + 1 new_key.number = number new_key.parent_fingerprint = fingerprint if number > (HARDENED_THRESHOLD - 1) data = [0x00].pack('C') << key.priv_key.htb << [numbe...
ruby
{ "resource": "" }
q18153
Bitcoin.ExtKey.priv_ver_to_pub_ver
train
def priv_ver_to_pub_ver case version when Bitcoin.chain_params.bip49_privkey_p2wpkh_p2sh_version Bitcoin.chain_params.bip49_pubkey_p2wpkh_p2sh_version when Bitcoin.chain_params.bip84_privkey_p2wpkh_version Bitcoin.chain_params.bip84_pubkey_p2wpkh_version else ...
ruby
{ "resource": "" }
q18154
Bitcoin.ExtPubkey.to_payload
train
def to_payload version.htb << [depth].pack('C') << parent_fingerprint.htb << [number].pack('N') << chain_code << pub.htb end
ruby
{ "resource": "" }
q18155
Bitcoin.ExtPubkey.derive
train
def derive(number) new_key = ExtPubkey.new new_key.depth = depth + 1 new_key.number = number new_key.parent_fingerprint = fingerprint raise 'hardened key is not support' if number > (HARDENED_THRESHOLD - 1) data = pub.htb << [number].pack('N') l = Bitcoin.hmac_sha512(chain_code...
ruby
{ "resource": "" }
q18156
Bitcoin.Script.p2pkh?
train
def p2pkh? return false unless chunks.size == 5 [OP_DUP, OP_HASH160, OP_EQUALVERIFY, OP_CHECKSIG] == (chunks[0..1]+ chunks[3..4]).map(&:ord) && chunks[2].bytesize == 21 end
ruby
{ "resource": "" }
q18157
Bitcoin.Script.push_only?
train
def push_only? chunks.each do |c| return false if !c.opcode.nil? && c.opcode > OP_16 end true end
ruby
{ "resource": "" }
q18158
Bitcoin.Script.witness_program?
train
def witness_program? return false if size < 4 || size > 42 || chunks.size < 2 opcode = chunks[0].opcode return false if opcode != OP_0 && (opcode < OP_1 || opcode > OP_16) return false unless chunks[1].pushdata? if size == (chunks[1][0].unpack('C').first + 2) program_size = chun...
ruby
{ "resource": "" }
q18159
Bitcoin.Script.witness_commitment
train
def witness_commitment return nil if !op_return? || op_return_data.bytesize < 36 buf = StringIO.new(op_return_data) return nil unless buf.read(4).bth == WITNESS_COMMITMENT_HEADER buf.read(32).bth end
ruby
{ "resource": "" }
q18160
Bitcoin.Script.to_script_code
train
def to_script_code(skip_separator_index = 0) payload = to_payload if p2wpkh? payload = Script.to_p2pkh(chunks[1].pushed_data.bth).to_payload elsif skip_separator_index > 0 payload = subscript_codeseparator(skip_separator_index) end Bitcoin.pack_var_string(payload) end
ruby
{ "resource": "" }
q18161
Bitcoin.Script.witness_data
train
def witness_data version = opcode_to_small_int(chunks[0].opcode) program = chunks[1].pushed_data [version, program] end
ruby
{ "resource": "" }
q18162
Bitcoin.Script.<<
train
def <<(obj) if obj.is_a?(Integer) push_int(obj) elsif obj.is_a?(String) append_data(obj) elsif obj.is_a?(Array) obj.each { |o| self.<< o} self end end
ruby
{ "resource": "" }
q18163
Bitcoin.Script.append_opcode
train
def append_opcode(opcode) opcode = Opcodes.small_int_to_opcode(opcode) if -1 <= opcode && opcode <= 16 raise ArgumentError, "specified invalid opcode #{opcode}." unless Opcodes.defined?(opcode) chunks << opcode.chr self end
ruby
{ "resource": "" }
q18164
Bitcoin.Script.append_data
train
def append_data(data) data = Encoding::ASCII_8BIT == data.encoding ? data : data.htb chunks << Bitcoin::Script.pack_pushdata(data) self end
ruby
{ "resource": "" }
q18165
Bitcoin.Script.include?
train
def include?(item) chunk_item = if item.is_a?(Integer) item.chr elsif item.is_a?(String) data = Encoding::ASCII_8BIT == item.encoding ? item : item.htb Bitcoin::Script.pack_pushdata(data) end return false un...
ruby
{ "resource": "" }
q18166
Bitcoin.Script.find_and_delete
train
def find_and_delete(subscript) raise ArgumentError, 'subscript must be Bitcoin::Script' unless subscript.is_a?(Script) return self if subscript.chunks.empty? buf = [] i = 0 result = Script.new chunks.each do |chunk| sub_chunk = subscript.chunks[i] if chunk.start_with?...
ruby
{ "resource": "" }
q18167
Bitcoin.Script.subscript_codeseparator
train
def subscript_codeseparator(separator_index) buf = [] process_separator_index = 0 chunks.each{|chunk| buf << chunk if process_separator_index == separator_index if chunk.ord == OP_CODESEPARATOR && process_separator_index < separator_index process_separator_index += 1 ...
ruby
{ "resource": "" }
q18168
Bitcoin.Script.p2pkh_addr
train
def p2pkh_addr return nil unless p2pkh? hash160 = chunks[2].pushed_data.bth return nil unless hash160.htb.bytesize == 20 Bitcoin.encode_base58_address(hash160, Bitcoin.chain_params.address_version) end
ruby
{ "resource": "" }
q18169
Bitcoin.Script.p2sh_addr
train
def p2sh_addr return nil unless p2sh? hash160 = chunks[1].pushed_data.bth return nil unless hash160.htb.bytesize == 20 Bitcoin.encode_base58_address(hash160, Bitcoin.chain_params.p2sh_version) end
ruby
{ "resource": "" }
q18170
Bitcoin.Script.bech32_addr
train
def bech32_addr segwit_addr = Bech32::SegwitAddr.new segwit_addr.hrp = Bitcoin.chain_params.bech32_hrp segwit_addr.script_pubkey = to_payload.bth segwit_addr.addr end
ruby
{ "resource": "" }
q18171
Bitcoin.Mnemonic.to_entropy
train
def to_entropy(words) word_master = load_words mnemonic = words.map do |w| index = word_master.index(w.downcase) raise IndexError, 'word not found in words list.' unless index index.to_s(2).rjust(11, '0') end.join entropy = mnemonic.slice(0, (mnemonic.length * 32) / 33) ...
ruby
{ "resource": "" }
q18172
Bitcoin.Mnemonic.to_mnemonic
train
def to_mnemonic(entropy) raise ArgumentError, 'entropy is empty.' if entropy.nil? || entropy.empty? e = entropy.htb.unpack('B*').first seed = e + checksum(e) mnemonic_index = seed.chars.each_slice(11).map{|i|i.join.to_i(2)} word_master = load_words mnemonic_index.map{|i|word_master[i...
ruby
{ "resource": "" }
q18173
Bitcoin.Mnemonic.to_seed
train
def to_seed(mnemonic, passphrase: '') to_entropy(mnemonic) OpenSSL::PKCS5.pbkdf2_hmac(mnemonic.join(' ').downcase, 'mnemonic' + passphrase, 2048, 64, OpenSSL::Digest::SHA512.new).bth end
ruby
{ "resource": "" }
q18174
Bitcoin.Mnemonic.checksum
train
def checksum(entropy) b = Bitcoin.sha256([entropy].pack('B*')).unpack('B*').first b.slice(0, (entropy.length/32)) end
ruby
{ "resource": "" }
q18175
Bitcoin.Key.to_wif
train
def to_wif version = Bitcoin.chain_params.privkey_version hex = version + priv_key hex += '01' if compressed? hex += Bitcoin.calc_checksum(hex) Base58.encode(hex) end
ruby
{ "resource": "" }
q18176
Bitcoin.Key.sign
train
def sign(data, low_r = true, extra_entropy = nil) sig = secp256k1_module.sign_data(data, priv_key, extra_entropy) if low_r && !sig_has_low_r?(sig) counter = 1 until sig_has_low_r?(sig) extra_entropy = [counter].pack('I*').bth.ljust(64, '0').htb sig = secp256k1_module.sign...
ruby
{ "resource": "" }
q18177
Bitcoin.Key.verify
train
def verify(sig, origin) return false unless valid_pubkey? begin sig = ecdsa_signature_parse_der_lax(sig) secp256k1_module.verify_sig(origin, sig, pubkey) rescue Exception false end end
ruby
{ "resource": "" }
q18178
Bitcoin.Key.to_point
train
def to_point p = pubkey p ||= generate_pubkey(priv_key, compressed: compressed) ECDSA::Format::PointOctetString.decode(p.htb, Bitcoin::Secp256k1::GROUP) end
ruby
{ "resource": "" }
q18179
Bitcoin.Key.ecdsa_signature_parse_der_lax
train
def ecdsa_signature_parse_der_lax(sig) sig_array = sig.unpack('C*') len_r = sig_array[3] r = sig_array[4...(len_r+4)].pack('C*').bth len_s = sig_array[len_r + 5] s = sig_array[(len_r + 6)...(len_r + 6 + len_s)].pack('C*').bth ECDSA::Signature.new(r.to_i(16), s.to_i(16)).to_der en...
ruby
{ "resource": "" }
q18180
Bitcoin.Base58.encode
train
def encode(hex) leading_zero_bytes = (hex.match(/^([0]+)/) ? $1 : '').size / 2 int_val = hex.to_i(16) base58_val = '' while int_val > 0 int_val, remainder = int_val.divmod(SIZE) base58_val = ALPHABET[remainder] + base58_val end ('1' * leading_zero_bytes) + base58_val ...
ruby
{ "resource": "" }
q18181
Bitcoin.Base58.decode
train
def decode(base58_val) int_val = 0 base58_val.reverse.split(//).each_with_index do |char,index| raise ArgumentError, 'Value passed not a valid Base58 String.' if (char_index = ALPHABET.index(char)).nil? int_val += char_index * (SIZE ** index) end s = int_val.to_even_length_hex ...
ruby
{ "resource": "" }
q18182
Bitcoin.Block.calculate_witness_commitment
train
def calculate_witness_commitment witness_hashes = [COINBASE_WTXID] witness_hashes += (transactions[1..-1].map(&:witness_hash)) reserved_value = transactions[0].inputs[0].script_witness.stack.map(&:bth).join root_hash = Bitcoin::MerkleTree.build_from_leaf(witness_hashes).merkle_root Bitcoin...
ruby
{ "resource": "" }
q18183
Bitcoin.Block.height
train
def height return nil if header.version < 2 coinbase_tx = transactions[0] return nil unless coinbase_tx.coinbase_tx? buf = StringIO.new(coinbase_tx.inputs[0].script_sig.to_payload) len = Bitcoin.unpack_var_int_from_io(buf) buf.read(len).reverse.bth.to_i(16) end
ruby
{ "resource": "" }
q18184
OpenAssets.Payload.to_payload
train
def to_payload payload = String.new payload << MARKER payload << VERSION payload << Bitcoin.pack_var_int(quantities.size) << quantities.map{|q| LEB128.encode_unsigned(q).read }.join payload << Bitcoin.pack_var_int(metadata.length) << metadata.bytes.map{|b| sprintf("%02x", b)}.join.htb ...
ruby
{ "resource": "" }
q18185
Bitcoin.Util.hash160
train
def hash160(hex) Digest::RMD160.hexdigest(Digest::SHA256.digest(hex.htb)) end
ruby
{ "resource": "" }
q18186
Bitcoin.Util.encode_base58_address
train
def encode_base58_address(hex, addr_version) base = addr_version + hex Base58.encode(base + calc_checksum(base)) end
ruby
{ "resource": "" }
q18187
Bitcoin.Util.decode_base58_address
train
def decode_base58_address(addr) hex = Base58.decode(addr) if hex.size == 50 && calc_checksum(hex[0...-8]) == hex[-8..-1] raise 'Invalid version bytes.' unless [Bitcoin.chain_params.address_version, Bitcoin.chain_params.p2sh_version].include?(hex[0..1]) [hex[2...-8], hex[0..1]] else ...
ruby
{ "resource": "" }
q18188
Bitcoin.Validation.check_tx
train
def check_tx(tx, state) # Basic checks that don't depend on any context if tx.inputs.empty? return state.DoS(10, reject_code: Message::Reject::CODE_INVALID, reject_reason: 'bad-txns-vin-empty') end if tx.outputs.empty? return state.DoS(100, reject_code: Message::Reject::CODE_INV...
ruby
{ "resource": "" }
q18189
Bitcoin.ScriptInterpreter.verify_script
train
def verify_script(script_sig, script_pubkey, witness = ScriptWitness.new) return set_error(SCRIPT_ERR_SIG_PUSHONLY) if flag?(SCRIPT_VERIFY_SIGPUSHONLY) && !script_sig.push_only? stack_copy = nil had_witness = false return false unless eval_script(script_sig, :base) stack_copy = stack.d...
ruby
{ "resource": "" }
q18190
Bitcoin.ScriptInterpreter.pop_int
train
def pop_int(count = 1) i = stack.pop(count).map{ |s| cast_to_int(s) } count == 1 ? i.first : i end
ruby
{ "resource": "" }
q18191
Bitcoin.ScriptInterpreter.cast_to_int
train
def cast_to_int(s, max_num_size = DEFAULT_MAX_NUM_SIZE) data = s.htb raise '"script number overflow"' if data.bytesize > max_num_size if require_minimal && data.bytesize > 0 if data.bytes[-1] & 0x7f == 0 && (data.bytesize <= 1 || data.bytes[data.bytesize - 2] & 0x80 == 0) raise 'non-...
ruby
{ "resource": "" }
q18192
Bitcoin.BloomFilter.contains?
train
def contains?(data) return true if full? hash_funcs.times do |i| hash = to_hash(data, i) return false unless check_bit(hash) end true end
ruby
{ "resource": "" }
q18193
Bitcoin.GCSFilter.hash_to_range
train
def hash_to_range(element) hash = SipHash.digest(key, element) map_into_range(hash, f) end
ruby
{ "resource": "" }
q18194
Bitcoin.GCSFilter.match_internal?
train
def match_internal?(hashes, size) n, payload = Bitcoin.unpack_var_int(encoded.htb) bit_reader = Bitcoin::BitStreamReader.new(payload) value = 0 hashes_index = 0 n.times do delta = golomb_rice_decode(bit_reader, p) value += delta loop do return false if has...
ruby
{ "resource": "" }
q18195
Bitcoin.GCSFilter.golomb_rice_encode
train
def golomb_rice_encode(bit_writer, p, x) q = x >> p while q > 0 nbits = q <= 64 ? q : 64 bit_writer.write(-1, nbits) # 18446744073709551615 is 2**64 - 1 = ~0ULL in cpp. q -= nbits end bit_writer.write(0, 1) bit_writer.write(x, p) end
ruby
{ "resource": "" }
q18196
Bitcoin.GCSFilter.golomb_rice_decode
train
def golomb_rice_decode(bit_reader, p) q = 0 while bit_reader.read(1) == 1 q +=1 end r = bit_reader.read(p) (q << p) + r end
ruby
{ "resource": "" }
q18197
Capistrano.Caplock.remote_file_content_same_as?
train
def remote_file_content_same_as?(full_path, content) Digest::MD5.hexdigest(content) == top.capture("md5sum #{full_path} | awk '{ print $1 }'").strip end
ruby
{ "resource": "" }
q18198
Capistrano.Caplock.remote_file_differs?
train
def remote_file_differs?(full_path, content) !remote_file_exists?(full_path) || remote_file_exists?(full_path) && !remote_file_content_same_as?(full_path, content) end
ruby
{ "resource": "" }
q18199
CukeModeler.Gherkin6Adapter.adapt_tag!
train
def adapt_tag!(parsed_tag) # Saving off the original data parsed_tag['cuke_modeler_parsing_data'] = Marshal::load(Marshal.dump(parsed_tag)) parsed_tag['name'] = parsed_tag.delete(:name) parsed_tag['line'] = parsed_tag.delete(:location)[:line] end
ruby
{ "resource": "" }