_id
stringlengths
2
6
title
stringlengths
9
130
partition
stringclasses
3 values
text
stringlengths
66
10.5k
language
stringclasses
1 value
meta_information
dict
q19400
DataMapper.OrderedSet.<<
train
def <<(entry) if index = @cache[entry] entries[index] = entry else @cache[entry] = size entries << entry end self end
ruby
{ "resource": "" }
q19401
DataMapper.PropertySet.[]=
train
def []=(name, entry) warn "#{self.class}#[]= is deprecated. Use #{self.class}#<< instead: #{caller.first}" raise "#{entry.class} is not added with the correct name" unless name && name.to_s == entry.name.to_s self << entry entry end
ruby
{ "resource": "" }
q19402
DataMapper.Query.update
train
def update(other) other_options = if kind_of?(other.class) return self if self.eql?(other) assert_valid_other(other) other.options else other = other.to_hash return self if other.empty? other end @options = @options.merge(other_options).freeze ...
ruby
{ "resource": "" }
q19403
DataMapper.Query.filter_records
train
def filter_records(records) records = records.uniq if unique? records = match_records(records) if conditions records = sort_records(records) if order records = limit_records(records) if limit || offset > 0 records end
ruby
{ "resource": "" }
q19404
DataMapper.Query.match_records
train
def match_records(records) conditions = self.conditions records.select { |record| conditions.matches?(record) } end
ruby
{ "resource": "" }
q19405
DataMapper.Query.sort_records
train
def sort_records(records) sort_order = order.map { |direction| [ direction.target, direction.operator == :asc ] } records.sort_by do |record| sort_order.map do |(property, ascending)| Sort.new(record_value(record, property), ascending) end end end
ruby
{ "resource": "" }
q19406
DataMapper.Query.slice!
train
def slice!(*args) offset, limit = extract_slice_arguments(*args) if self.limit || self.offset > 0 offset, limit = get_relative_position(offset, limit) end update(:offset => offset, :limit => limit) end
ruby
{ "resource": "" }
q19407
DataMapper.Query.inspect
train
def inspect attrs = [ [ :repository, repository.name ], [ :model, model ], [ :fields, fields ], [ :links, links ], [ :conditions, conditions ], [ :order, order ], [ :limit, limit ]...
ruby
{ "resource": "" }
q19408
DataMapper.Query.condition_properties
train
def condition_properties properties = Set.new each_comparison do |comparison| next unless comparison.respond_to?(:subject) subject = comparison.subject properties << subject if subject.kind_of?(Property) end properties end
ruby
{ "resource": "" }
q19409
DataMapper.Query.to_subquery
train
def to_subquery collection = model.all(merge(:fields => model_key)) Conditions::Operation.new(:and, Conditions::Comparison.new(:in, self_relationship, collection)) end
ruby
{ "resource": "" }
q19410
DataMapper.Query.to_hash
train
def to_hash { :repository => repository.name, :model => model.name, :fields => fields, :links => links, :conditions => conditions, :offset => offset, :limit => limit, :order => order, :unique ...
ruby
{ "resource": "" }
q19411
DataMapper.Query.assert_valid_options
train
def assert_valid_options(options) options = options.to_hash options.each do |attribute, value| case attribute when :fields then assert_valid_fields(value, options[:unique]) when :links then assert_valid_links(value) when...
ruby
{ "resource": "" }
q19412
DataMapper.Query.assert_valid_offset
train
def assert_valid_offset(offset, limit) unless offset >= 0 raise ArgumentError, "+options[:offset]+ must be greater than or equal to 0, but was #{offset.inspect}" end if offset > 0 && limit.nil? raise ArgumentError, '+options[:offset]+ cannot be greater than 0 if limit is not specified...
ruby
{ "resource": "" }
q19413
DataMapper.Query.assert_valid_other
train
def assert_valid_other(other) other_repository = other.repository repository = self.repository other_class = other.class unless other_repository == repository raise ArgumentError, "+other+ #{other_class} must be for the #{repository.name} repository, not #{other_repository.na...
ruby
{ "resource": "" }
q19414
DataMapper.Query.merge_conditions
train
def merge_conditions(conditions) @conditions = Conditions::Operation.new(:and) << @conditions unless @conditions.nil? conditions.compact! conditions.each do |condition| case condition when Conditions::AbstractOperation, Conditions::AbstractComparison add_condition(condit...
ruby
{ "resource": "" }
q19415
DataMapper.Query.append_condition
train
def append_condition(subject, bind_value, model = self.model, operator = :eql) case subject when Property, Associations::Relationship then append_property_condition(subject, bind_value, operator) when Symbol then append_symbol_condition(subject, bind_value, model, ope...
ruby
{ "resource": "" }
q19416
DataMapper.Query.set_operation
train
def set_operation(operation, other) assert_valid_other(other) query = self.class.new(@repository, @model, other.to_relative_hash) query.instance_variable_set(:@conditions, other_conditions(other, operation)) query end
ruby
{ "resource": "" }
q19417
DataMapper.Query.other_conditions
train
def other_conditions(other, operation) self_conditions = query_conditions(self) unless self_conditions.kind_of?(Conditions::Operation) operation_slug = case operation when :intersection, :difference then :and when :union then :or...
ruby
{ "resource": "" }
q19418
DataMapper.Query.query_conditions
train
def query_conditions(query) if query.limit || query.links.any? query.to_subquery else query.conditions end end
ruby
{ "resource": "" }
q19419
DataMapper.Query.self_relationship
train
def self_relationship @self_relationship ||= begin model = self.model Associations::OneToMany::Relationship.new( :self, model, model, self_relationship_options ) end end
ruby
{ "resource": "" }
q19420
DataMapper.Query.self_relationship_options
train
def self_relationship_options keys = model_key.map { |property| property.name } repository = self.repository { :child_key => keys, :parent_key => keys, :child_repository_name => repository.name, :parent_repository_name => repository.name,...
ruby
{ "resource": "" }
q19421
DataMapper.Model.get
train
def get(*key) assert_valid_key_size(key) repository = self.repository key = self.key(repository.name).typecast(key) repository.identity_map(self)[key] || first(key_conditions(repository, key).update(:order => nil)) end
ruby
{ "resource": "" }
q19422
DataMapper.Model.first
train
def first(*args) first_arg = args.first last_arg = args.last limit_specified = first_arg.kind_of?(Integer) with_query = (last_arg.kind_of?(Hash) && !last_arg.empty?) || last_arg.kind_of?(Query) limit = limit_specified ? first_arg : 1 query = with_query ? last_arg : {} ...
ruby
{ "resource": "" }
q19423
DataMapper.Model.copy
train
def copy(source_repository_name, target_repository_name, query = {}) target_properties = properties(target_repository_name) query[:fields] ||= properties(source_repository_name).select do |property| target_properties.include?(property) end repository(target_repository_name) do |reposit...
ruby
{ "resource": "" }
q19424
DataMapper.Model.repository_name
train
def repository_name context = Repository.context context.any? ? context.last.name : default_repository_name end
ruby
{ "resource": "" }
q19425
DataMapper.Model.finalize_allowed_writer_methods
train
def finalize_allowed_writer_methods @allowed_writer_methods = public_instance_methods.map { |method| method.to_s }.grep(WRITER_METHOD_REGEXP).to_set @allowed_writer_methods -= INVALID_WRITER_METHODS @allowed_writer_methods.freeze end
ruby
{ "resource": "" }
q19426
DataMapper.Model.assert_valid_properties
train
def assert_valid_properties repository_name = self.repository_name if properties(repository_name).empty? && !relationships(repository_name).any? { |relationship| relationship.kind_of?(Associations::ManyToOne::Relationship) } raise IncompleteModelError, "#{name} must have at least one propert...
ruby
{ "resource": "" }
q19427
DataMapper.SubjectSet.[]
train
def [](name) name = name.to_s entries.detect { |entry| entry.name.to_s == name } end
ruby
{ "resource": "" }
q19428
DataMapper.Collection.reload
train
def reload(other_query = Undefined) query = self.query query = other_query.equal?(Undefined) ? query.dup : query.merge(other_query) # make sure the Identity Map contains all the existing resources identity_map = repository.identity_map(model) loaded_entries.each do |resource| ide...
ruby
{ "resource": "" }
q19429
DataMapper.Collection.get
train
def get(*key) assert_valid_key_size(key) key = model_key.typecast(key) query = self.query @identity_map[key] || if !loaded? && (query.limit || query.offset > 0) # current query is exclusive, find resource within the set # TODO: use a subquery to retrieve the Collection and t...
ruby
{ "resource": "" }
q19430
DataMapper.Collection.all
train
def all(query = Undefined) if query.equal?(Undefined) || (query.kind_of?(Hash) && query.empty?) dup else # TODO: if there is no order parameter, and the Collection is not loaded # check to see if the query can be satisfied by the head/tail new_collection(scoped_query(query)) ...
ruby
{ "resource": "" }
q19431
DataMapper.Collection.first
train
def first(*args) first_arg = args.first last_arg = args.last limit_specified = first_arg.kind_of?(Integer) with_query = (last_arg.kind_of?(Hash) && !last_arg.empty?) || last_arg.kind_of?(Query) limit = limit_specified ? first_arg : 1 query = with_query ? last_arg : {} ...
ruby
{ "resource": "" }
q19432
DataMapper.Collection.at
train
def at(offset) if loaded? || partially_loaded?(offset) super elsif offset == 0 first elsif offset > 0 first(:offset => offset) elsif offset == -1 last else last(:offset => offset.abs - 1) end end
ruby
{ "resource": "" }
q19433
DataMapper.Collection.slice!
train
def slice!(*args) removed = super resources_removed(removed) unless removed.nil? # Workaround for Ruby <= 1.8.6 compact! if RUBY_VERSION <= '1.8.6' unless removed.kind_of?(Enumerable) return removed end offset, limit = extract_slice_arguments(*args) query = s...
ruby
{ "resource": "" }
q19434
DataMapper.Collection.[]=
train
def []=(*args) orphans = Array(superclass_slice(*args[0..-2])) # relate new resources resources = resources_added(super) # mark resources as removed resources_removed(orphans - loaded_entries) resources end
ruby
{ "resource": "" }
q19435
DataMapper.Collection.each
train
def each return to_enum unless block_given? super do |resource| begin original, resource.collection = resource.collection, self yield resource ensure resource.collection = original end end end
ruby
{ "resource": "" }
q19436
DataMapper.Collection.new
train
def new(attributes = {}) resource = repository.scope { model.new(attributes) } self << resource resource end
ruby
{ "resource": "" }
q19437
DataMapper.Collection.update
train
def update(attributes) assert_update_clean_only(:update) dirty_attributes = model.new(attributes).dirty_attributes dirty_attributes.empty? || all? { |resource| resource.update(attributes) } end
ruby
{ "resource": "" }
q19438
DataMapper.Collection.update!
train
def update!(attributes) assert_update_clean_only(:update!) model = self.model dirty_attributes = model.new(attributes).dirty_attributes if dirty_attributes.empty? true else dirty_attributes.each do |property, value| property.assert_valid_value(value) en...
ruby
{ "resource": "" }
q19439
DataMapper.Collection.destroy!
train
def destroy! repository = self.repository deleted = repository.delete(self) if loaded? unless deleted == size return false end each do |resource| resource.persistence_state = Resource::PersistenceState::Immutable.new(resource) end clear...
ruby
{ "resource": "" }
q19440
DataMapper.Collection.respond_to?
train
def respond_to?(method, include_private = false) super || model.respond_to?(method) || relationships.named?(method) end
ruby
{ "resource": "" }
q19441
DataMapper.Collection.partially_loaded?
train
def partially_loaded?(offset, limit = 1) if offset >= 0 lazy_possible?(head, offset + limit) else lazy_possible?(tail, offset.abs) end end
ruby
{ "resource": "" }
q19442
DataMapper.Collection.lazy_load
train
def lazy_load if loaded? return self end mark_loaded head = self.head tail = self.tail query = self.query resources = repository.read(query) # remove already known results resources -= head if head.any? resources -= tail if tail...
ruby
{ "resource": "" }
q19443
DataMapper.Collection.new_collection
train
def new_collection(query, resources = nil, &block) if loaded? resources ||= filter(query) end # TOOD: figure out a way to pass not-yet-saved Resources to this newly # created Collection. If the new resource matches the conditions, then # it should be added to the collection (keep...
ruby
{ "resource": "" }
q19444
DataMapper.Collection.set_operation
train
def set_operation(operation, other) resources = set_operation_resources(operation, other) other_query = Query.target_query(repository, model, other) new_collection(query.send(operation, other_query), resources) end
ruby
{ "resource": "" }
q19445
DataMapper.Collection._create
train
def _create(attributes, execute_hooks = true) resource = repository.scope { model.send(execute_hooks ? :create : :create!, default_attributes.merge(attributes)) } self << resource if resource.saved? resource end
ruby
{ "resource": "" }
q19446
DataMapper.Collection._save
train
def _save(execute_hooks = true) loaded_entries = self.loaded_entries loaded_entries.each { |resource| set_default_attributes(resource) } @removed.clear loaded_entries.all? { |resource| resource.__send__(execute_hooks ? :save : :save!) } end
ruby
{ "resource": "" }
q19447
DataMapper.Collection.default_attributes
train
def default_attributes return @default_attributes if @default_attributes default_attributes = {} conditions = query.conditions if conditions.slug == :and model_properties = properties.dup model_key = self.model_key if model_properties.to_set.superset?(model_key...
ruby
{ "resource": "" }
q19448
DataMapper.Collection.resource_added
train
def resource_added(resource) resource = initialize_resource(resource) if resource.saved? @identity_map[resource.key] = resource @removed.delete(resource) else set_default_attributes(resource) end resource end
ruby
{ "resource": "" }
q19449
DataMapper.Collection.resources_added
train
def resources_added(resources) if resources.kind_of?(Enumerable) resources.map { |resource| resource_added(resource) } else resource_added(resources) end end
ruby
{ "resource": "" }
q19450
DataMapper.Collection.resources_removed
train
def resources_removed(resources) if resources.kind_of?(Enumerable) resources.each { |resource| resource_removed(resource) } else resource_removed(resources) end end
ruby
{ "resource": "" }
q19451
DataMapper.Collection.filter
train
def filter(other_query) query = self.query fields = query.fields.to_set unique = other_query.unique? # TODO: push this into a Query#subset? method if other_query.links.empty? && (unique || (!unique && !query.unique?)) && !other_query.reload? ...
ruby
{ "resource": "" }
q19452
DataMapper.Collection.scoped_query
train
def scoped_query(query) if query.kind_of?(Query) query.dup else self.query.relative(query) end end
ruby
{ "resource": "" }
q19453
DataMapper.Collection.delegate_to_model
train
def delegate_to_model(method, *args, &block) model = self.model model.send(:with_scope, query) do model.send(method, *args, &block) end end
ruby
{ "resource": "" }
q19454
DataMapper.Repository.read
train
def read(query) return [] unless query.valid? query.model.load(adapter.read(query), query) end
ruby
{ "resource": "" }
q19455
DataMapper.Repository.update
train
def update(attributes, collection) return 0 unless collection.query.valid? && attributes.any? adapter.update(attributes, collection) end
ruby
{ "resource": "" }
q19456
DataMapper.DescendantSet.delete
train
def delete(descendant) @descendants.delete(descendant) each { |d| d.descendants.delete(descendant) } end
ruby
{ "resource": "" }
q19457
DataMapper.Property.valid?
train
def valid?(value, negated = false) dumped_value = dump(value) if required? && dumped_value.nil? negated || false else value_dumped?(dumped_value) || (dumped_value.nil? && (allow_nil? || negated)) end end
ruby
{ "resource": "" }
q19458
DataMapper.Property.assert_valid_value
train
def assert_valid_value(value) unless valid?(value) raise Property::InvalidValueError.new(self,value) end true end
ruby
{ "resource": "" }
q19459
DataMapper.Resource.attribute_set
train
def attribute_set(name, value) property = properties[name] if property value = property.typecast(value) self.persistence_state = persistence_state.set(property, value) end end
ruby
{ "resource": "" }
q19460
DataMapper.Resource.attributes
train
def attributes(key_on = :name) attributes = {} lazy_load(properties) fields.each do |property| if model.public_method_defined?(name = property.name) key = case key_on when :name then name when :field then property.field else property ...
ruby
{ "resource": "" }
q19461
DataMapper.Resource.inspect
train
def inspect # TODO: display relationship values attrs = properties.map do |property| value = if new? || property.loaded?(self) property.get!(self).inspect else '<not loaded>' end "#{property.instance_variable_name}=#{value}" end "#<#{model.na...
ruby
{ "resource": "" }
q19462
DataMapper.Resource.dirty_attributes
train
def dirty_attributes dirty_attributes = {} original_attributes.each_key do |property| next unless property.respond_to?(:dump) dirty_attributes[property] = property.dump(property.get!(self)) end dirty_attributes end
ruby
{ "resource": "" }
q19463
DataMapper.Resource.initialize_copy
train
def initialize_copy(original) instance_variables.each do |ivar| instance_variable_set(ivar, DataMapper::Ext.try_dup(instance_variable_get(ivar))) end self.persistence_state = persistence_state.class.new(self) end
ruby
{ "resource": "" }
q19464
DataMapper.Resource.reset_key
train
def reset_key properties.key.zip(key) do |property, value| property.set!(self, value) end end
ruby
{ "resource": "" }
q19465
DataMapper.Resource.clear_subjects
train
def clear_subjects model_properties = properties (model_properties - model_properties.key | relationships).each do |subject| next unless subject.loaded?(self) remove_instance_variable(subject.instance_variable_name) end end
ruby
{ "resource": "" }
q19466
DataMapper.Resource.eager_load
train
def eager_load(properties) unless properties.empty? || key.nil? || collection.nil? # set an initial value to prevent recursive lazy loads properties.each { |property| property.set!(self, nil) } collection.reload(:fields => properties) end self end
ruby
{ "resource": "" }
q19467
DataMapper.Resource.conditions
train
def conditions key = self.key if key model.key_conditions(repository, key) else conditions = {} properties.each do |property| next unless property.loaded?(self) conditions[property] = property.get!(self) end conditions end end
ruby
{ "resource": "" }
q19468
DataMapper.Resource.child_relationships
train
def child_relationships child_relationships = [] relationships.each do |relationship| next unless relationship.respond_to?(:collection_for) set_default_value(relationship) next unless relationship.loaded?(self) child_relationships << relationship end many_to_ma...
ruby
{ "resource": "" }
q19469
DataMapper.Resource.save_self
train
def save_self(execute_hooks = true) # short-circuit if the resource is not dirty return saved? unless dirty_self? if execute_hooks new? ? create_with_hooks : update_with_hooks else _persist end clean? end
ruby
{ "resource": "" }
q19470
DataMapper.Resource.save_parents
train
def save_parents(execute_hooks) run_once(true) do parent_relationships.map do |relationship| parent = relationship.get(self) if parent.__send__(:save_parents, execute_hooks) && parent.__send__(:save_self, execute_hooks) relationship.set(self, parent) # set the FK values ...
ruby
{ "resource": "" }
q19471
DataMapper.Resource.dirty_self?
train
def dirty_self? if original_attributes.any? true elsif new? !model.serial.nil? || properties.any? { |property| property.default? } else false end end
ruby
{ "resource": "" }
q19472
DataMapper.Resource.dirty_parents?
train
def dirty_parents? run_once(false) do parent_associations.any? do |association| association.__send__(:dirty_self?) || association.__send__(:dirty_parents?) end end end
ruby
{ "resource": "" }
q19473
DataMapper.Resource.cmp?
train
def cmp?(other, operator) return false unless repository.send(operator, other.repository) && key.send(operator, other.key) if saved? && other.saved? # if dirty attributes match then they are the same resource dirty_attributes == other.dirty_attributes else ...
ruby
{ "resource": "" }
q19474
DataMapper.Resource.execute_hooks_for
train
def execute_hooks_for(type, name) model.hooks[name][type].each { |hook| hook.call(self) } end
ruby
{ "resource": "" }
q19475
DataMapper.Resource.run_once
train
def run_once(default) caller_method = Kernel.caller(1).first[/`([^'?!]+)[?!]?'/, 1] sentinel = "@_#{caller_method}_sentinel" return instance_variable_get(sentinel) if instance_variable_defined?(sentinel) begin instance_variable_set(sentinel, default) yield ensure ...
ruby
{ "resource": "" }
q19476
DataMapper.Inflector.pluralize
train
def pluralize(word) result = word.to_s.dup if word.empty? || inflections.uncountables.include?(result.downcase) result else inflections.plurals.each { |(rule, replacement)| break if result.gsub!(rule, replacement) } result end end
ruby
{ "resource": "" }
q19477
Backburner.Helpers.exception_message
train
def exception_message(e) msg = [ "Exception #{e.class} -> #{e.message}" ] base = File.expand_path(Dir.pwd) + '/' e.backtrace.each do |t| msg << " #{File.expand_path(t).gsub(/#{base}/, '')}" end if e.backtrace msg.join("\n") end
ruby
{ "resource": "" }
q19478
Backburner.Helpers.classify
train
def classify(dashed_word) dashed_word.to_s.split('-').each { |part| part[0] = part[0].chr.upcase }.join end
ruby
{ "resource": "" }
q19479
Backburner.Helpers.dasherize
train
def dasherize(word) classify(word).to_s.gsub(/::/, '/'). gsub(/([A-Z]+)([A-Z][a-z])/,'\1_\2'). gsub(/([a-z\d])([A-Z])/,'\1_\2'). tr("_", "-").downcase end
ruby
{ "resource": "" }
q19480
Backburner.Helpers.resolve_priority
train
def resolve_priority(pri) if pri.respond_to?(:queue_priority) resolve_priority(pri.queue_priority) elsif pri.is_a?(String) || pri.is_a?(Symbol) # named priority resolve_priority(Backburner.configuration.priority_labels[pri.to_sym]) elsif pri.is_a?(Integer) # numerical pri ...
ruby
{ "resource": "" }
q19481
Backburner.Helpers.resolve_respond_timeout
train
def resolve_respond_timeout(ttr) if ttr.respond_to?(:queue_respond_timeout) resolve_respond_timeout(ttr.queue_respond_timeout) elsif ttr.is_a?(Integer) # numerical ttr else # default Backburner.configuration.respond_timeout end end
ruby
{ "resource": "" }
q19482
Backburner.Worker.work_one_job
train
def work_one_job(conn = connection) begin job = reserve_job(conn) rescue Beaneater::TimedOutError => e return end self.log_job_begin(job.name, job.args) job.process self.log_job_end(job.name) rescue Backburner::Job::JobFormatInvalid => e self.log_error sel...
ruby
{ "resource": "" }
q19483
Backburner.Worker.new_connection
train
def new_connection Connection.new(Backburner.configuration.beanstalk_url) { |conn| Backburner::Hooks.invoke_hook_events(self, :on_reconnect, conn) } end
ruby
{ "resource": "" }
q19484
Backburner.Worker.reserve_job
train
def reserve_job(conn, reserve_timeout = Backburner.configuration.reserve_timeout) Backburner::Job.new(conn.tubes.reserve(reserve_timeout)) end
ruby
{ "resource": "" }
q19485
Backburner.Worker.handle_error
train
def handle_error(e, name, args, job) if error_handler = Backburner.configuration.on_error if error_handler.arity == 1 error_handler.call(e) elsif error_handler.arity == 3 error_handler.call(e, name, args) else error_handler.call(e, name, args, job) end...
ruby
{ "resource": "" }
q19486
Backburner.Worker.compact_tube_names
train
def compact_tube_names(tube_names) tube_names = tube_names.first if tube_names && tube_names.size == 1 && tube_names.first.is_a?(Array) tube_names = Array(tube_names).compact if tube_names && Array(tube_names).compact.size > 0 tube_names = nil if tube_names && tube_names.compact.empty? tube_name...
ruby
{ "resource": "" }
q19487
Backburner.Logger.log_job_end
train
def log_job_end(name, message = nil) ellapsed = Time.now - job_started_at ms = (ellapsed.to_f * 1000).to_i action_word = message ? 'Finished' : 'Completed' log_info("#{action_word} #{name} in #{ms}ms #{message}") end
ruby
{ "resource": "" }
q19488
Backburner.Job.process
train
def process # Invoke before hook and stop if false res = @hooks.invoke_hook_events(job_name, :before_perform, *args) return false unless res # Execute the job @hooks.around_hook_events(job_name, :around_perform, *args) do # We subtract one to ensure we timeout before beanstalkd doe...
ruby
{ "resource": "" }
q19489
Backburner.Job.timeout_job_after
train
def timeout_job_after(secs, &block) begin Timeout::timeout(secs) { yield } rescue Timeout::Error => e raise JobTimeout, "#{name}(#{(@args||[]).join(', ')}) hit #{secs}s timeout.\nbacktrace: #{e.backtrace}" end end
ruby
{ "resource": "" }
q19490
Backburner.Connection.retryable
train
def retryable(options = {}, &block) options = {:max_retries => 4, :on_retry => nil, :retry_delay => 1.0}.merge!(options) retry_count = options[:max_retries] begin yield rescue Beaneater::NotConnected if retry_count > 0 reconnect! retry_count -= 1 s...
ruby
{ "resource": "" }
q19491
Backburner.Connection.ensure_connected!
train
def ensure_connected!(max_retries = 4, retry_delay = 1.0) return self if connected? begin reconnect! return self rescue Beaneater::NotConnected => e if max_retries > 0 max_retries -= 1 sleep retry_delay retry else # stop retrying ...
ruby
{ "resource": "" }
q19492
Backburner.Connection.beanstalk_addresses
train
def beanstalk_addresses uri = self.url.is_a?(Array) ? self.url.first : self.url beanstalk_host_and_port(uri) end
ruby
{ "resource": "" }
q19493
Backburner.Connection.beanstalk_host_and_port
train
def beanstalk_host_and_port(uri_string) uri = URI.parse(uri_string) raise(BadURL, uri_string) if uri.scheme != 'beanstalk'.freeze "#{uri.host}:#{uri.port || 11300}" end
ruby
{ "resource": "" }
q19494
CountryStateSelect.CscsController.find_cities
train
def find_cities cities = CS.cities(params[:state_id].to_sym, params[:country_id].to_sym) respond_to do |format| format.json { render :json => cities.to_a} end end
ruby
{ "resource": "" }
q19495
ComfyBootstrapForm.FormBuilder.file_field
train
def file_field(method, options = {}) bootstrap = form_bootstrap.scoped(options.delete(:bootstrap)) return super if bootstrap.disabled draw_form_group(bootstrap, method, options) do if bootstrap.custom_control content_tag(:div, class: "custom-file") do add_css_class!(opti...
ruby
{ "resource": "" }
q19496
ComfyBootstrapForm.FormBuilder.plaintext
train
def plaintext(method, options = {}) bootstrap = form_bootstrap.scoped(options.delete(:bootstrap)) draw_form_group(bootstrap, method, options) do remove_css_class!(options, "form-control") add_css_class!(options, "form-control-plaintext") options[:readonly] = true ActionView::...
ruby
{ "resource": "" }
q19497
ComfyBootstrapForm.FormBuilder.primary
train
def primary(value = nil, options = {}, &block) add_css_class!(options, "btn-primary") submit(value, options, &block) end
ruby
{ "resource": "" }
q19498
ComfyBootstrapForm.FormBuilder.draw_form_group
train
def draw_form_group(bootstrap, method, options) label = draw_label(bootstrap, method, for_attr: options[:id]) errors = draw_errors(method) control = draw_control(bootstrap, errors, method, options) do yield end form_group_class = "form-group" form_group_class += " row" ...
ruby
{ "resource": "" }
q19499
ComfyBootstrapForm.FormBuilder.draw_control
train
def draw_control(bootstrap, errors, _method, options) add_css_class!(options, "form-control") add_css_class!(options, "is-invalid") if errors.present? draw_control_column(bootstrap, offset: bootstrap.label[:hide]) do draw_input_group(bootstrap, errors) do yield end end...
ruby
{ "resource": "" }