query stringlengths 9 9.05k | document stringlengths 10 222k | negatives listlengths 19 20 | metadata dict |
|---|---|---|---|
wrapper function for replace dialog, launches dialog over the passed curses window and returns a tuple of (pattern, replace) or (None,None) if canceled | def replace( scr ):
d = ReplaceDialog(scr)
value = d.main()
if not "pattern" in value:
return (None,None)
else:
return (value["pattern"],value["replace"]) | [
"def find_replace(self, event=None):\n if self.app.children:\n #find string\n findStr = self.app.childActive.source.GetSelectedText()\n if findStr and self.findDialog:\n self.findDialog.Destroy()\n self.findDialog = None\n #dialog alre... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
wrapper function for the confirm replace dialog, returns 1 == yes, 2 == no, 3 == all, 4 == cancel returns 4 == cancel if canceled | def confirm_replace( scr ):
d = ConfirmReplaceDialog(scr)
value = d.main()
if "yes" in value:
if value["yes"]:
return 1
elif value["no"]:
return 2
elif value["all"]:
return 3
elif value["cancel"]:
return 4
else:
retu... | [
"def _confirm(self, msg):\n return (\n self.replace_all or\n input(\"{} [y/N] \".format(msg)).lower() in ['y', 'yes']\n )",
"def MessageYesNoCancel(Message, Title) -> int:\n pass",
"def confirm_dialog(title='Confirm', message='Text:'):\n def dialog_decorate(func):\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Sets default value for order prefix | def default_order_prefix():
return 'mag_' | [
"def default_prefix(self) -> str:",
"def prefix(self, value):\n self._prefix = value",
"def change_prefix(self, prefix):\n self.prefix = str(prefix).strip()",
"def prefix(self, prefix):\n self._prefix = prefix",
"def setDefaultTopicPrefix(self, prefix):\n\n internals.blpapi_Sessi... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Import order states for instances | def import_order_states(cls, instances):
OrderState = Pool().get('magento.order_state')
for instance in instances:
Transaction().context.update({
'magento_instance': instance.id
})
# Import order states
with OrderConfig(
... | [
"def import_order_states(self):\n if self.source != 'amazon_mws':\n return super(SaleChannel, self).import_order_states()\n\n order_states_data = [\n 'Pending',\n 'Unshipped',\n 'PartiallyShipped',\n 'Shipped',\n 'InvoiceUnconfirmed',\n... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Sets default for active | def default_active():
return True | [
"def default(self, default):\n self._default = default",
"def active(self, active):\n\n self._active = active",
"def default(self, default):\n\n self._default = default",
"def setActive(self):\n self._active = True\n self.updateTab()",
"def set_value_to_default(self):\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Sets current company as default | def default_company():
return Transaction().context.get('company') | [
"def company(self, company):\n self._company = company",
"def company(self, company):\n\n self._company = company",
"def set_company(self, arg):\n self = Company(**arg)\n self.save()",
"def set_default_site(self, site):\n self._dialog_kwargs[\"default_site\"] = site",
"def... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Import the websites and their stores/view from magento | def import_websites(cls, instances):
Website = Pool().get('magento.instance.website')
Store = Pool().get('magento.website.store')
StoreView = Pool().get('magento.store.store_view')
MagentoOrderState = Pool().get('magento.order_state')
try:
instance, = instances
... | [
"def export_inventory(self, websites):\n for website in websites:\n website.export_inventory_to_magento()",
"def importSites(self,sites_list):\n \"\"\" Append these sites objects to a sample \"\"\" \n self.sites = []\n for s in sites_list:\n mySite = Site(s)\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Import carriers/shipping methods from magento for instances | def import_carriers(cls, instances):
InstanceCarrier = Pool().get('magento.instance.carrier')
for instance in instances:
with Transaction().set_context({
'magento_instance': instance.id
}):
with OrderConfig(
instance.url, inst... | [
"def setShippingMethods(self):",
"def refresh(self):\n self.__dict__ = self._api.get_shipping_methods(id=self.id).__dict__",
"def separate_methods(self):\r\n methods = {obj['method'] for obj in self.cf_data}\r\n metadata = {obj['name']: obj for obj in self.csv_data}\r\n self.data = {... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Sets default root category id. Is set to 1, because the default root category is 1 | def default_magento_root_category_id():
return 1 | [
"def get_root_id(self, identifier=None, category_details=None):\n return structures_module.categories.get_root_id(self.khoros_object, identifier, category_details)",
"def set_root_node(self, config_id, root):\n if config_id not in self.__configs.keys():\n raise ValueError('configurati... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Sets default product uom for website | def default_default_uom():
ProductUom = Pool().get('product.uom')
return ProductUom.search([('name', '=', 'Unit')])[0].id | [
"def setup_products(self, *args, **kwargs):",
"def getProductName():\n return options.product_name",
"def change_productid(self):\n key = OpenKey(HKEY_LOCAL_MACHINE,\n \"SOFTWARE\\\\Microsoft\\\\Windows NT\\\\CurrentVersion\",\n 0, KEY_SET_VALUE)\n\n va... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Looks for the website whose `values` are sent by magento against the instance with `instance` in tryton. If a record exists for this, return that else create a new one and return | def find_or_create(cls, instance, values):
websites = cls.search([
('instance', '=', instance.id),
('magento_id', '=', int(values['website_id']))
])
if websites:
return websites[0]
return cls.create([{
'name': values['name'],
... | [
"def find_or_create(cls, website, values):\n stores = cls.search([\n ('website', '=', website.id),\n ('magento_id', '=', int(values['group_id']))\n ])\n\n if stores:\n return stores[0]\n\n return cls.create([{\n 'name': values['name'],\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Exports inventory stock information to magento | def export_inventory(self, websites):
for website in websites:
website.export_inventory_to_magento() | [
"def export_inventory_to_magento(self):\n Location = Pool().get('stock.location')\n\n product_templates = []\n instance = self.instance\n\n locations = Location.search([('type', '=', 'storage')])\n\n for magento_product_template in self.magento_product_templates:\n prod... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Exports stock data of products from tryton to magento for this website | def export_inventory_to_magento(self):
Location = Pool().get('stock.location')
product_templates = []
instance = self.instance
locations = Location.search([('type', '=', 'storage')])
for magento_product_template in self.magento_product_templates:
product_template =... | [
"def getStockData():\n pass",
"def export_inventory(self, websites):\n for website in websites:\n website.export_inventory_to_magento()",
"def default_export_(self, fields):\n Store = Pool().get('magento.website.store')\n\n store = Store(Transaction().context.get('active_i... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Returns company related to website | def get_company(self, name):
return self.website.company.id | [
"def test_website_companies_get_details(self):\n pass",
"def get_company(self, company_referece):\n url = 'companies/{0}'.format(company_referece)\n result = self.get(url)\n return result.get('company', result)",
"def run_whoxy_company_search(self,company):\n if self.whoxy_api... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Returns instance related to website | def get_instance(self, name):
return self.website.instance.id | [
"def website(self):\n return self._website",
"def _get_instance(self):",
"def get_instance(self, instance):\n return self._get(_instance.Instance, instance)",
"def get_website(self, name):\n return self.store.website.id",
"def site(self, domain):\r\n return resource.Site(self, do... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Looks for the store whose `values` are sent by magento against the website with `website` in tryton. If a record exists for this, return that else create a new one and return | def find_or_create(cls, website, values):
stores = cls.search([
('website', '=', website.id),
('magento_id', '=', int(values['group_id']))
])
if stores:
return stores[0]
return cls.create([{
'name': values['name'],
'magento_id... | [
"def find_or_create(cls, instance, values):\n websites = cls.search([\n ('instance', '=', instance.id),\n ('magento_id', '=', int(values['website_id']))\n ])\n\n if websites:\n return websites[0]\n\n return cls.create([{\n 'name': values['name'... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Exports tier prices of products from tryton to magento for this store | def export_tier_prices_to_magento(self):
instance = self.website.instance
for mag_product_template in self.website.magento_product_templates:
product_template = mag_product_template.template
product = product_template.products[0]
# Get the price tiers from the produ... | [
"def default_export_(self, fields):\n Store = Pool().get('magento.website.store')\n\n store = Store(Transaction().context.get('active_id'))\n\n return {\n 'products_count': store.export_tier_prices_to_magento()\n }",
"def export_product_prices(self):\n if self.source ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Returns instance related to store | def get_instance(self, name):
return self.store.instance.id | [
"def get_store(self):\n return self.store",
"def _get_store(self):\n return self._store",
"def get_resource(self):\n return self._stores",
"def store(self) -> Store:\n return self._store",
"def get_instance(self):\n if self.concat:\n if isinstance(self.concat, Conca... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Returns website related to store | def get_website(self, name):
return self.store.website.id | [
"def website(self):\n self._validate_status()\n return self.details.get('website')",
"def _siteStore(self):\n return self.webapp.store.parent",
"def website(self):\n return self._pkgc.homepage",
"def website(self):\n return self._website",
"def get_store(self, store_name: ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Returns company related to store | def get_company(self, name):
return self.store.company.id | [
"def get_company(self, company_referece):\n url = 'companies/{0}'.format(company_referece)\n result = self.get(url)\n return result.get('company', result)",
"def company(self):\n return self._company",
"def get_joined_stores(self):\n user_id = self._user.id\n stores = S... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Looks for the store view whose `values` are sent by magento against the store with `store` in tryton. If a record exists for this, return that else create a new one and return | def find_or_create(cls, store, values):
store_views = cls.search([
('store', '=', store.id),
('magento_id', '=', int(values['store_id']))
])
if store_views:
return store_views[0]
return cls(**{
'name': values['name'],
'code': ... | [
"def find_or_create(cls, website, values):\n stores = cls.search([\n ('website', '=', website.id),\n ('magento_id', '=', int(values['group_id']))\n ])\n\n if stores:\n return stores[0]\n\n return cls.create([{\n 'name': values['name'],\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Calls wizard to import orders for store view | def import_orders_button(cls, store_views):
pass | [
"def import_order_from_store_view(self):\n Sale = Pool().get('sale.sale')\n MagentoOrderState = Pool().get('magento.order_state')\n\n new_sales = []\n instance = self.instance\n with Transaction().set_context({\n 'magento_instance': instance.id,\n 'magento_we... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Calls wizard to export order status for store view | def export_order_status_button(cls, store_views):
pass | [
"def export_order_status_for_store_view(self):\n Sale = Pool().get('sale.sale')\n\n exported_sales = []\n domain = [('magento_store_view', '=', self.id)]\n\n if self.last_order_export_time:\n domain = [('write_date', '>=', self.last_order_export_time)]\n\n sales = Sale.... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Imports sale from store view | def import_order_from_store_view(self):
Sale = Pool().get('sale.sale')
MagentoOrderState = Pool().get('magento.order_state')
new_sales = []
instance = self.instance
with Transaction().set_context({
'magento_instance': instance.id,
'magento_website': self.... | [
"def viewone(self):\n conn = dbconnect()\n cur = conn.cursor()\n query = \"SELECT * FROM sales WHERE id=self.id\"\n data = (self.id,)\n cur.execute(query, data)\n cur.close()",
"def import_orders_button(cls, store_views):\n pass",
"def sell(self):\n sale =... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Export sale orders to magento for the current store view. If last export time is defined, export only those orders which are updated after last export time. | def export_order_status_for_store_view(self):
Sale = Pool().get('sale.sale')
exported_sales = []
domain = [('magento_store_view', '=', self.id)]
if self.last_order_export_time:
domain = [('write_date', '>=', self.last_order_export_time)]
sales = Sale.search(domain)... | [
"def export_shipment_status_to_magento(self):\n Shipment = Pool().get('stock.shipment.out')\n Sale = Pool().get('sale.sale')\n\n instance = self.instance\n\n sale_domain = [\n ('magento_store_view', '=', self.id),\n ('shipment_state', '=', 'sent'),\n ('ma... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Import orders from magento for store views | def import_orders(cls, store_views=None):
if store_views is None:
store_views = cls.search([])
for store_view in store_views:
store_view.import_order_from_store_view() | [
"def import_order_from_store_view(self):\n Sale = Pool().get('sale.sale')\n MagentoOrderState = Pool().get('magento.order_state')\n\n new_sales = []\n instance = self.instance\n with Transaction().set_context({\n 'magento_instance': instance.id,\n 'magento_we... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Export Shipment status for shipments related to current store view. This method is called by cron. | def export_shipment_status(cls, store_views=None):
if store_views is None:
store_views = cls.search([])
for store_view in store_views:
# Set the instance in context
with Transaction().set_context(
magento_instance=store_view.instance.id
):... | [
"def export_shipment_status_to_magento(self):\n Shipment = Pool().get('stock.shipment.out')\n Sale = Pool().get('sale.sale')\n\n instance = self.instance\n\n sale_domain = [\n ('magento_store_view', '=', self.id),\n ('shipment_state', '=', 'sent'),\n ('ma... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Exports shipment status for shipments to magento, if they are shipped | def export_shipment_status_to_magento(self):
Shipment = Pool().get('stock.shipment.out')
Sale = Pool().get('sale.sale')
instance = self.instance
sale_domain = [
('magento_store_view', '=', self.id),
('shipment_state', '=', 'sent'),
('magento_id', '!=... | [
"def export_shipment_status(self, cursor, user, ids, context):\n store_view_obj = self.pool.get('magento.store.store_view')\n\n store_view = store_view_obj.browse(\n cursor, user, context.get('active_id')\n )\n context.update({\n 'magento_instance': store_view.insta... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Helper method to get the current store view. | def get_current_store_view(cls):
return cls(Transaction().context.get('magento_store_view')) | [
"def get_view(self):\n return self.view",
"def get_store(self):\n return self.store",
"def _siteStore(self):\n return self.webapp.store.parent",
"def _get_store(self):\n return self._store",
"def currentView(self):\n widget = self.currentWidget()\n if isinstance(wid... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Import the websites, store and store views and show user a confirmation message | def default_start(self, data):
return {
'message': "This wizard has imported all the websites for this " +
"magento instance. It has also imported all the stores and " +
"store views related to the websites imported. If any of " +
"the records existed ... | [
"def web_index():\n\n try:\n auth_check()\n except Exception as e:\n return flask.redirect(str(e))\n\n db_update_archives()\n\n return flask.redirect('videos')",
"def import_websites(cls, instances):\n Website = Pool().get('magento.instance.website')\n Store = Pool().get('m... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Import carriers and show the user appropriate message | def default_start(self, data):
return {
'message': "This wizard has imported all the carriers / " +
"shipping methods for this magento instance. You should now " +
"configure the imported carriers / shipping methods to " +
"match the shipment carriers ... | [
"def onImportTextFromCAD(self):\n try:\n self.onCommandRejected()\n dialog = QImportTextFromCADDialog(self)\n dialog.show()\n dialog.exec_()\n except Exception as ex:\n message = 'error occurred({}) in {}:{}'.format(ex, sys.exc_info()[-1].tb_frame... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Export price tiers and return count of products | def default_export_(self, fields):
Store = Pool().get('magento.website.store')
store = Store(Transaction().context.get('active_id'))
return {
'products_count': store.export_tier_prices_to_magento()
} | [
"def export_tier_prices_to_magento(self):\n instance = self.website.instance\n\n for mag_product_template in self.website.magento_product_templates:\n product_template = mag_product_template.template\n product = product_template.products[0]\n\n # Get the price tiers fr... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Type a PWIF amount then submit the form | def enter_pwif_amount(self, pwif_amount):
raise NotImplementedError | [
"def send_mfa(\n self,\n form: object = None, # noqa: ARG002\n code: str = \"\",\n trusted_device: bool = True,\n ) -> None:\n el_otp = self._driver.find_element(By.CSS_SELECTOR, \"input[name=otc]\", timeout=5)\n el_otp.clear()\n el_otp.send_keys(code)\n\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
For every argument in kwargs sets a WITH_FOO if FOO=True or a WITHOUT_FOO if FOO=False | def set_with_options(self, **kwargs) -> None:
for k, v in kwargs.items():
assert not k.startswith("WITH_"), "Invalid WITH/WITHOUT options name " + k
assert not k.startswith("WITHOUT_"), "Invalid WITH/WITHOUT options name " + k
assert isinstance(v, bool)
self._with... | [
"def restrict_kwargs(kwargs, f):\n f_kwargs = set(get_kwarg_defaults(f).keys()) | set(get_arg_names(f))\n keep, discard = {}, {}\n for key in kwargs.keys():\n if key in f_kwargs:\n keep[key] = kwargs[key]\n else:\n discard[key] = kwargs[key]\n return keep, discard",
... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Strip all ELF binaries to reduce the size of the benchmark directory | def strip_elf_files(self, benchmark_dir) -> None:
self.info("Stripping all ELF files in", benchmark_dir)
self.run_cmd("du", "-sh", benchmark_dir)
for root, dirnames, filenames in os.walk(str(benchmark_dir)):
for filename in filenames:
file = Path(root, filename)
... | [
"def clean_flatbuffer_binaries():\n for element in FLATBUFFERS_CONVERSION_DATA:\n for json in element.input_files:\n path = processed_json_path(json)\n if os.path.isfile(path):\n os.remove(path)",
"def _clean_bins():\n rmtree(LIBS_DIR)\n rmtree(BINS_DIR)\n rmtree(HEADERS_DIR)",
"de... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Fail immediately, with the given message. | def fail(self, msg=None):
raise Exception, msg | [
"def fail(self, msg=None):\r\n raise self.failureException(msg)",
"def fail(self, msg=None):\n raise self.failureException, msg",
"def fail_message(self, message: str) -> None:\n\n self.build_message(message, \"FAIL\", util.COLOR.RED)",
"def failing(self, message):\n raise Assertio... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Validate (previous) KMALLOC/KFREE calls of a set of tasks (pids) | def validate(self, tracked_pids, test_case=stubTestcase, debug=False):
out = os.popen('dmesg -c -s %d' % LOG_BUF_LEN)
dmesg_lines = out.readlines()
out.close()
allocations = []
memory_allocated = False
if debug:
f = open('mm_debug.tx... | [
"def bad_cgroup_processes_check():\n return CGCheck([], bad_cgroup_processes)",
"def test_invalid_free_error_2():\n\n pool = IdPool(5, 5)\n\n pool.free_if_allocated(4) # not in range (= not allocated)\n\n pool.free_if_allocated(5) # in range but not allocated\n\n pool.free_if_allocated(6) # not ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Release for the other side of the fork that is syncing this side | def release(self):
if self._inchild:
os.write(self._pw_child, self.RELEASE_MSG)
else:
os.write(self._pw_parent, self.RELEASE_MSG) | [
"def release(self):\n self._need_release = True",
"def releaseResource(self):",
"def release(local):\n local.__release__()",
"def release(self, o):\n if not self.available(o):\n raise ValueError('you do not own this lock')\n self._owner = None",
"def release (self, path, f... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Return the correct write side of the general pipe | def _wpipe(self):
if self._inchild:
return self._general_pw_child
else:
return self._general_pw_parent | [
"def write(self, s):\n\t\tif self._input: raise PlumberExceptions.PipeTypeException(self)\n\t\treturn pservlet.pipe_write(self._pipe_desc, s)",
"def _make_write_pipe_transport(self, pipe, protocol, waiter=None,\n extra=None):\n channel = self._channel_from_fileobj(pipe)\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Dump data through the pipe. The data is sent using pickle binary format. | def pickle_dump(self, data):
os.write(self.wpipe, pickle.dumps(data, bin=True)) | [
"async def dump(self, data: dict, file: IO):",
"def pickle(self,data,filename):\n pickle.dump(data, open(filename, 'wb'))",
"def pickle_it(data: SerialisableType) -> bytes:\n return pickle.dumps(data)",
"def dump_data(ser, meta, args):\n ser.reset_input_buffer()\n ser.reset_output_buffer()\n\n... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Load data from the pipe. The data is sent using pickle binary format. | def pickle_load(self):
#
# I am wrapping the file descriptor because this way pickle
# returns on each data send separately (allowing for sending
# multiple data before reading).
# I close the file descriptor or else for some reason just
# closing the write side ... | [
"def load_data():\n with open('data.pickle', 'rb') as f:\n data = pickle.load(f)\n return data",
"def pickle_dump(self, data):\n \n os.write(self.wpipe, pickle.dumps(data, bin=True))",
"def deserialize(self):\n file = None\n try:\n file = open(self.dataPath + ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
A function to force Tensorflow to use CPU even Nvidia GPU present | def cpu_fallback(flag=True):
gpu_phy_devices = tf.config.list_physical_devices("GPU")
cpu_phy_devices = tf.config.list_physical_devices("CPU")
general_warning_msg = (
f"Tensorflow has already been initialized, {inspect.currentframe().f_code.co_name}() needs "
f"to be called before any Tenso... | [
"def set_tf_cpu_only():\n tf.config.experimental.set_visible_devices([], 'GPU')\n tf.config.experimental.set_visible_devices([], 'TPU')",
"def use_cpu():\n\n os.environ[\"CUDA_VISIBLE_DEVICES\"] = ''",
"def disable_gpu():\r\n os.environ['CUDA_VISIBLE_DEVICES'] = '-1'",
"def is_gpu_available() -> bool:... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Return results from detector. This function prepares the environment loading the plugins, getting the response and passing it to the detector. In case of errors, it raises exceptions to be handled externally. | def get_detection_results(
url,
timeout,
metadata=False,
save_har=False,
splash_url="",
):
plugins = load_plugins()
if not plugins:
raise NoPluginsError("No plugins found")
logger.debug("[+] Starting detection with %(n)d plugins", {"n": len(plugins)})
response = get_respons... | [
"def setup(self):\n rc = self.rc\n try:\n for plugin in self.plugins:\n plugin.setup(rc)\n except Exception as e:\n self.exit(e)",
"def setup(self):\n url = config.protocol_of_webserver + self.web + self.detail_url\n self.log.debug('Getting u... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Returns encoding of HTTP response. | def encoding(response: tornado.httpclient.HTTPResponse) -> str:
if 'Content-Encoding' in response.headers:
return response.headers['Content-Encoding'].decode()
elif 'Content-Type' in response.headers:
headers = email.message_from_string('Content-Type: ' +
... | [
"def return_response_string(self):\n response = \"{} {}\\r\\n\".format(self.protocol, self.code)\n str_headers = \"\"\n if self.headers:\n for k, v in self.headers.items():\n str_headers += \"{}: {}\\r\\n\".format(k, v)\n\n encoded_response = \"{}{}\\r\\n\".form... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Get HTTP response body as text. | def text_body(response: tornado.httpclient.HTTPResponse) -> str:
return response.body.decode(encoding(response)) | [
"def _get_body(response):\n return response.content",
"def text(self) -> str:\n return self._raw_response.text",
"def get_body_text(self):\n if self.body_type != 'HTML':\n return self.body\n\n try:\n soup = bs(self.body, 'html.parser')\n except RuntimeErr... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Get HTTP response body as json | def json_body(response: tornado.httpclient.HTTPResponse):
return json.loads(text_body(response)) | [
"def get_json(response):\n\tif requests.__version__ >= \"1.0.0\":\n\t\treturn response.json()\n\telif requests.__version__ == \"0.14.2\":\n\t\treturn response.json\n\telse:\n\t\treturn json.loads(response.content)",
"def get_json(response):\n return json.loads(response.get_data(as_text=True))",
"def _get_bod... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Create a default list of bounds for a given signal description | def _default_bounds(signal):
# there's just the name
if isinstance(signal, str):
return (signal, 0, 0, 0, 0)
else:
# there's just the name in a list
if len(signal) == 1:
return signal + [0, 0, 0, 0]
# there's the name and bounds
if len(signal) == 3:
... | [
"def _init_optimizer_bounds(self):\n bounds = []\n for filt in self.filters:\n if filt.optimize_fc:\n bounds.append((np.log10(filt.min_fc), np.log10(filt.max_fc)))\n if filt.optimize_q:\n bounds.append((filt.min_q, filt.max_q))\n if filt.o... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Parseinterfaces described in YAML files, bundled with the package | def parse_interface_definitions(dir_name=DIR):
try:
filenames = listdir(dir_name)
except OSError:
raise OSError(f"Directory '{dir_name}' "
"doesn't exist or cannot be listed")
defs = []
for filename in filenames:
with open(join(dir_name, filename)) as f:
... | [
"def parse_and_load(self):",
"def test_parse_yaml(self) -> None:\n pass",
"def parse(self, infile):\r\n raise NotImplementedError()",
"def load_definitions(self):\n path = os.path.join(os.path.dirname(__file__), 'definitions')\n for (dirpath, _, filenames) in os.walk(path):\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Given filenames of Verilog source and JSON target, use 'write_json' function of yosys | def verilog_to_json(verilog_filename, json_filename):
system(f'yosys -p "read_verilog {verilog_filename}" '
'-p "write_json {json_filename}"') | [
"def copy_json() -> None:\n example_paths = [\n EXAMPLE_FILEPATH_PREPROCESS, EXAMPLE_FILEPATH_ANALYSE,\n EXAMPLE_FILEPATH_BUILD, EXAMPLE_FILEPATH_VARIANTS,\n EXAMPLE_FILEPATH_ASSESS, EXAMPLE_FILEPATH_OPTIMISATION_SIMULATION,\n EXAMPLE_FILEPATH_OPTIMISATION_COMPUTATION\n ]\n for ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
get channel metadata from index | def get_ch_metadata(self, index):
tag = self.get_ch_tag(index)
return getattr(self, f"{tag.lower()}_metadata") | [
"def showmeta(self,\r\n index):\r\n\r\n return self.get_metadata_from_note(index)",
"def read_ts_channel_basic_info(self):\n\n channel_list = list(self.session_md['time_series_channels'].keys())\n channel_list.sort()\n\n channel_infos = []\n for channel in channe... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
deletes and returns minimum value (in this case the root of the heap) | def delete_min(self):
self.switch(0, -1)
min = self.heap.pop(-1)
self.bubble_down(0)
return min | [
"def delete_min(self):\n #The length is 1 because the heap list was initialized with 0\n if len(self.heap_list) == 1:\n return \"Empty heap.\"\n\n #Store the min value of the heap\n top = self.heap_list[1]\n\n #Move the last value of the heap to the top\n self.he... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Inserts key into heap and heapifies the heap | def insert(self, key):
self.heap.append(key)
self.bubble_up(len(self.heap) - 1) | [
"def insert_key(self, k):\n self.heap.append(k)\n i = len(self.heap) - 1\n while i != 0 and self.heap[self.parent(i)] > self.heap[i]:\n self.heap[i], self.heap[self.parent(i)] = self.heap[self.parent(i)], self.heap[i]\n i = self.parent(i)",
"def insert(self, k): \r\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Returns parent of node at n; None if there isn't a parent | def parent(self, n):
return None if n == 0 else (n - 1) // 2 | [
"def get_parent(self, n):\n return self.metadata[n][\"parent\"]",
"def parent(self, nid):\r\n if not self.contains(nid):\r\n raise NodeIDAbsentError(\"Node '%s' is not in the tree\" % nid)\r\n\r\n pid = self[nid].bpointer\r\n if pid is None or not self.contains(pid):\r\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Returns right child of node at n; None if there isn't a right_child | def right_child(self, n):
if 2 * n + 2 >= len(self.heap):
return None
else:
return 2 * n + 2 | [
"def rightchild(node):\n if node: return node[2]",
"def get_node_right(self, n: MazeCell) -> MazeCell:\n if n.x == self._ncols - 1:\n return None\n else:\n return self.get_node(n.x + 1, n.y)",
"def get_right_child(\n self\n ):\n return self.right_child... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_availability_zones_get | def test_aws_service_api_availability_zones_get(self):
pass | [
"def test_v1_time_zones_get(self):\n pass",
"def list_availability_zones(self):\n uri = 'os-availability-zone'\n resp, body = self.get(uri)\n self.expected_success(200, resp.status)\n return self._parse_resp(body)",
"def test_aws_service_api_regions_get(self):\n pass",
... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_flavor_get | def test_aws_service_api_flavor_get(self):
pass | [
"def test_aws_service_api_flavors_get(self):\n pass",
"def test_openstack_rest_flavors_get(self):\n pass",
"def test_get_flavor(self):\n response = self.flavors_client.get_flavor_details(self.flavor_ref)\n flavor = response.entity\n self.assertEqual(self.flavor_ref, flavor.id)... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_flavors_get | def test_aws_service_api_flavors_get(self):
pass | [
"def test_openstack_rest_flavors_get(self):\n pass",
"def test_aws_service_api_flavor_get(self):\n pass",
"def test_get_api_resources_flavor_kubevirt_io_v1alpha1(self):\n pass",
"def test_get_api_group_flavor_kubevirt_io(self):\n pass",
"def test_list_flavors(self):\n resp... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_image_get | def test_aws_service_api_image_get(self):
pass | [
"def test_aws_service_api_public_image_get(self):\n pass",
"def test_aws_service_api_private_image_get(self):\n pass",
"def test_aws_service_api_public_images_get(self):\n pass",
"def test_openstack_rest_images_get(self):\n pass",
"def test_aws_service_api_private_images_get(self... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_interfaces_get | def test_aws_service_api_interfaces_get(self):
pass | [
"def getInterface(self):\n\t\tquery = ''\n\t\tconn = self.get_connection()\n\t\theaders = { 'Content-type' : 'application/json', 'Authorization' : 'A10 %s' %self.sessionid}\n\t\tconn.request('GET', self.get_path() + '/' + query, headers=headers)\n\t\tresponse = conn.getresponse()\n\t\texpected_status = 200\n\t\terr... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_keypair_delete | def test_aws_service_api_keypair_delete(self):
pass | [
"def test_delete_api_key(self):\n pass",
"def test_delete_deleted_keypair_fails(self):\n with self.assertRaises(ItemNotFound):\n self.keypairs_client.delete_keypair(self.name)",
"def test_aws_service_api_keypair_get(self):\n pass",
"def test_delete_keypair_response(self):\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_keypair_generate_post | def test_aws_service_api_keypair_generate_post(self):
pass | [
"def test_aws_service_api_keypair_import_post(self):\n pass",
"def test_aws_service_api_keypair_get(self):\n pass",
"def test_create_api_key(self):\n pass",
"def test_create_key(self):\n self.fail(\"test not implemented\")",
"def test_otoroshi_controllers_adminapi_pki_controller_... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_keypair_get | def test_aws_service_api_keypair_get(self):
pass | [
"def test_aws_service_api_keypairs_get(self):\n pass",
"def test_aws_service_api_keypair_generate_post(self):\n pass",
"def test_aws_service_api_keypair_delete(self):\n pass",
"def test_aws_service_api_keypair_import_post(self):\n pass",
"def get_key_pair(keyPairName=None):\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_keypair_import_post | def test_aws_service_api_keypair_import_post(self):
pass | [
"def test_aws_service_api_keypair_generate_post(self):\n pass",
"def test_aws_service_api_keypair_get(self):\n pass",
"def test_aws_service_api_keypair_delete(self):\n pass",
"def test_aws_service_api_keypairs_get(self):\n pass",
"def import_key_pair(keyPairName=None, publicKeyBa... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_keypairs_get | def test_aws_service_api_keypairs_get(self):
pass | [
"def test_aws_service_api_keypair_get(self):\n pass",
"def test_aws_service_api_keypair_generate_post(self):\n pass",
"def test_aws_service_api_keypair_delete(self):\n pass",
"def test_aws_service_api_keypair_import_post(self):\n pass",
"def test_get_deleted_keypair_fails(self):\... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_network_subnets_get | def test_aws_service_api_network_subnets_get(self):
pass | [
"def test_list_subnets(self):\n print(self.the_client.list_subnets())",
"def test_get_subnet(self):\n self.assertEqual(\n type(self.the_client.get_subnet(subnet_id)),\n baidubce.bce_response.BceResponse)",
"def list_subnet(self):\n _url = \"http://\" + self.host_ip + \... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_networks_get | def test_aws_service_api_networks_get(self):
pass | [
"def test_openstack_rest_networks_get(self):\n pass",
"def test_get_networks(self):\n pass",
"def test_get_network(self):\n pass",
"def getNetworksList():\n logger.debug('Start.')\n code, res = rest_requests.get(networks_url)\n if code != requests.codes.ok:\n logger.error(... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_private_image_get | def test_aws_service_api_private_image_get(self):
pass | [
"def test_aws_service_api_private_images_get(self):\n pass",
"def test_aws_service_api_public_image_get(self):\n pass",
"def test_aws_service_api_image_get(self):\n pass",
"def test_aws_service_api_public_images_get(self):\n pass",
"def test_openstack_rest_images_get(self):\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_private_images_get | def test_aws_service_api_private_images_get(self):
pass | [
"def test_aws_service_api_private_image_get(self):\n pass",
"def test_aws_service_api_public_images_get(self):\n pass",
"def test_aws_service_api_public_image_get(self):\n pass",
"def test_aws_service_api_image_get(self):\n pass",
"def test_openstack_rest_images_get(self):\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_public_image_get | def test_aws_service_api_public_image_get(self):
pass | [
"def test_aws_service_api_public_images_get(self):\n pass",
"def test_aws_service_api_image_get(self):\n pass",
"def test_aws_service_api_private_image_get(self):\n pass",
"def test_aws_service_api_private_images_get(self):\n pass",
"def test_openstack_rest_images_get(self):\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_public_images_get | def test_aws_service_api_public_images_get(self):
pass | [
"def test_aws_service_api_public_image_get(self):\n pass",
"def test_aws_service_api_private_images_get(self):\n pass",
"def test_aws_service_api_image_get(self):\n pass",
"def test_aws_service_api_private_image_get(self):\n pass",
"def test_openstack_rest_images_get(self):\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_regions_get | def test_aws_service_api_regions_get(self):
pass | [
"def test_regions_get(self):\n pass",
"def test_getregions(self):\n pass",
"def test_xpack_v1_cloud_get_regions_list(self):\n pass",
"def test_regions_region_id_get(self):\n pass",
"def test_getregions_item(self):\n pass",
"def test_api_regions(self):\n # load api... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_security_groups_get | def test_aws_service_api_security_groups_get(self):
pass | [
"def test_get_security_group_using_get1(self):\n pass",
"def test_api_v1_groups_get(self):\n pass",
"def test_get_groups(self):\n pass",
"def test_groups_get(self):\n pass",
"def test_get_resource_group_list(self):\n pass",
"def test_users_groups_get(self):\n pass... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_snapshot_delete | def test_aws_service_api_snapshot_delete(self):
pass | [
"def test_delete(self):\n snapshot = models.Snapshot(\n self.client, container=self.container,\n name='an-snapshot')\n\n snapshot.delete(wait=True)\n\n # TODO: add an assertion here",
"def test_delete_snapshot(self):\n ctxt = context.get_admin_context()\n e... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_snapshots_get | def test_aws_service_api_snapshots_get(self):
pass | [
"def test_aws_service_api_snapshots_post(self):\n pass",
"def xtest_snapshot_api(self):\n\n req = httplib2.Http(\".cache\")\n body = r\"\"\"{ \"snapshot\": { \"instanceId\": \"123\", \"name\": \"dbapi_test\" } }\"\"\"\n \n # Test creating an snapshot without a body in th... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_snapshots_post | def test_aws_service_api_snapshots_post(self):
pass | [
"def test_aws_service_api_snapshots_get(self):\n pass",
"def xtest_snapshot_api(self):\n\n req = httplib2.Http(\".cache\")\n body = r\"\"\"{ \"snapshot\": { \"instanceId\": \"123\", \"name\": \"dbapi_test\" } }\"\"\"\n \n # Test creating an snapshot without a body in the... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_validate_subscription_post | def test_aws_service_api_validate_subscription_post(self):
pass | [
"def test_create_subscription(self):\n pass",
"def test_create_a_new_subscription(self):\n pass",
"def post_create_subscription(\n self, response: pubsub.Subscription\n ) -> pubsub.Subscription:\n return response",
"def test_get_subscription(self):\n pass",
"def test_cr... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_command_put | def test_aws_service_api_vm_command_put(self):
pass | [
"def test_aws_service_api_vm_tag_put(self):\n pass",
"def test_aws_service_api_vm_workshift_put(self):\n pass",
"def test_aws_service_api_vm_patch(self):\n pass",
"def test_update_stage_using_put(self):\n pass",
"def test_aws_service_api_vm_delete(self):\n pass",
"def te... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_delete | def test_aws_service_api_vm_delete(self):
pass | [
"def test_aws_service_api_volume_delete(self):\n pass",
"def test_delete_vm(self, value):\n\n # Steps:\n # 1. Create Account and create network in it (isoalted/ shared/ vpc)\n # 2. Deploy a VM in this network and account\n # 3. Add 2 secondary IPs to the default nic of VM\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_details_get | def test_aws_service_api_vm_details_get(self):
pass | [
"def test_aws_service_api_vm_get(self):\n pass",
"def test_aws_service_api_vm_management_get(self):\n pass",
"def test_aws_service_api_vms_get(self):\n pass",
"def get_details(self, psvm):\n return self.get(psvm)",
"def test_aws_service_api_vm_patch(self):\n pass",
"def ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_get | def test_aws_service_api_vm_get(self):
pass | [
"def test_aws_service_api_vm_details_get(self):\n pass",
"def test_aws_service_api_vm_management_get(self):\n pass",
"def test_aws_service_api_vms_get(self):\n pass",
"def test_aws_service_api_vm_patch(self):\n pass",
"def test_virtualservice_get(self):\n pass",
"def tes... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_management_get | def test_aws_service_api_vm_management_get(self):
pass | [
"def test_aws_service_api_vm_get(self):\n pass",
"def test_aws_service_api_vm_details_get(self):\n pass",
"def test_aws_service_api_vms_get(self):\n pass",
"def test_aws_service_api_vm_patch(self):\n pass",
"def test_aws_service_api_vm_command_put(self):\n pass",
"def te... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_password_get | def test_aws_service_api_vm_password_get(self):
pass | [
"def _vm_get_password(self, vm_instance):\n pass",
"def test_checkPasswordMatches(self):\n return self.compareCheckPassword(keyPassword=\"password\",\n password=\"password\")",
"def when_I_reset_the_password(self, vm):\n\n self.debug(\"Resetting passw... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_patch | def test_aws_service_api_vm_patch(self):
pass | [
"def test_aws_service_api_vm_get(self):\n pass",
"def test_aws_service_api_vm_command_put(self):\n pass",
"def test_aws_service_api_vm_management_get(self):\n pass",
"def test_aws_service_api_vm_tag_put(self):\n pass",
"def test_aws_service_api_vm_details_get(self):\n pass... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_security_group_delete | def test_aws_service_api_vm_security_group_delete(self):
pass | [
"def test_delete_resource_group(self):\n pass",
"def test_delete_group(self):\n pass",
"def test_aws_service_api_vm_security_group_put(self):\n pass",
"def delete_security_group(self, security_group):\r\n return self.delete(self.security_group_path % (security_group))",
"def test... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_security_group_put | def test_aws_service_api_vm_security_group_put(self):
pass | [
"def test_aws_service_api_vm_security_group_delete(self):\n pass",
"def test_aws_service_api_security_groups_get(self):\n pass",
"def test_update_port_with_security_group_and_extra_attributes(self):\n self._update_port_with_security_groups(\n [data_utils.rand_name('secgroup')])",... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_tag_put | def test_aws_service_api_vm_tag_put(self):
pass | [
"def test_aws_service_api_vm_command_put(self):\n pass",
"def test_iam_project_tag_put(self):\n pass",
"def test_add_or_update_single_tag(self):\n p = self.load_policy({\n 'name': 'test-azure-tag',\n 'resource': 'azure.vm',\n 'filters': [\n {'... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_workshift_delete | def test_aws_service_api_vm_workshift_delete(self):
pass | [
"def test_aws_service_api_vm_workshift_put(self):\n pass",
"def test_aws_service_api_vm_workshift_post(self):\n pass",
"def test_aws_service_api_vm_delete(self):\n pass",
"def test_delete_workout(self):\n response = self.client.open(\n '/workout/{id}'.format(id='id_examp... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_workshift_post | def test_aws_service_api_vm_workshift_post(self):
pass | [
"def test_aws_service_api_vm_workshift_put(self):\n pass",
"def test_aws_service_api_vm_workshift_delete(self):\n pass",
"def test_aws_service_api_vms_post(self):\n pass",
"def test_aws_service_api_vm_command_put(self):\n pass",
"def test_shift_view_post(self):\n path = re... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vm_workshift_put | def test_aws_service_api_vm_workshift_put(self):
pass | [
"def test_aws_service_api_vm_workshift_post(self):\n pass",
"def test_aws_service_api_vm_workshift_delete(self):\n pass",
"def test_aws_service_api_vm_command_put(self):\n pass",
"def test_update_stage_using_put(self):\n pass",
"def test_shift_view_put(self):\n path = reve... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vms_get | def test_aws_service_api_vms_get(self):
pass | [
"def test_aws_service_api_vm_get(self):\n pass",
"def test_aws_service_api_vm_management_get(self):\n pass",
"def test_aws_service_api_vm_details_get(self):\n pass",
"def test_aws_service_api_vms_post(self):\n pass",
"def test_virtualservice_get(self):\n pass",
"def test... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_vms_post | def test_aws_service_api_vms_post(self):
pass | [
"def test_aws_service_api_vms_get(self):\n pass",
"def test_api_v1_scans_vms_post(self):\n pass",
"def test_aws_service_api_vm_workshift_post(self):\n pass",
"def test_aws_service_api_vm_command_put(self):\n pass",
"def test_aws_service_api_vm_tag_put(self):\n pass",
"de... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_volume_attachment_delete | def test_aws_service_api_volume_attachment_delete(self):
pass | [
"def test_aws_service_api_volume_delete(self):\n pass",
"def test_attachment_deletion_allowed_volume_no_attachments(self):\n volume = tests_utils.create_volume(self.context)\n self.volume_api.attachment_deletion_allowed(self.context, None, volume)",
"def test_aws_service_api_volume_attachme... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_volume_attachment_put | def test_aws_service_api_volume_attachment_put(self):
pass | [
"def test_aws_service_api_volume_attachment_delete(self):\n pass",
"def test_aws_service_api_volume_patch(self):\n pass",
"def test_attachment_update_volume_in_error_state(self):\n volume_params = {'status': 'available'}\n\n vref = tests_utils.create_volume(self.context, **volume_par... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_volume_delete | def test_aws_service_api_volume_delete(self):
pass | [
"def test_aws_service_api_volume_attachment_delete(self):\n pass",
"def test_delete_volume(self):\n ctxt = context.get_admin_context()\n extra_specs = {}\n type_ref = volume_types.create(ctxt, 'hgst-1', extra_specs)\n volume = {'id': '1', 'name': 'volume1',\n 'd... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_volume_get | def test_aws_service_api_volume_get(self):
pass | [
"def test_aws_service_api_volumes_get(self):\n pass",
"def test_volumes_get(self):\n pass",
"def test_openstack_rest_volumes_get(self):\n pass",
"def test_aws_service_api_volume_patch(self):\n pass",
"def test_get_volume(self):\n volume = Volume(self.client, 1)\n se... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_volume_patch | def test_aws_service_api_volume_patch(self):
pass | [
"def test_aws_service_api_volume_get(self):\n pass",
"def test_aws_service_api_volume_attachment_put(self):\n pass",
"def test_aws_service_api_volume_delete(self):\n pass",
"def test_aws_service_api_volumes_get(self):\n pass",
"def test_aws_service_api_volumes_post(self):\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_volume_types_get | def test_aws_service_api_volume_types_get(self):
pass | [
"def test_aws_service_api_volume_get(self):\n pass",
"def get_volume_types(self):\n res = self.get('%s/types' % self.catalog['volume'])\n if res['status'] == 200:\n return json.loads(res['body'])['volume_types']\n else:\n LOG.error('Get volume types failed: %s %s ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_volumes_get | def test_aws_service_api_volumes_get(self):
pass | [
"def test_openstack_rest_volumes_get(self):\n pass",
"def test_volumes_get(self):\n pass",
"def test_aws_service_api_volume_get(self):\n pass",
"def test_aws_service_api_volumes_post(self):\n pass",
"def test_aws_service_api_volume_types_get(self):\n pass",
"def test_10_... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
Test case for aws_service_api_volumes_post | def test_aws_service_api_volumes_post(self):
pass | [
"def test_volumes_post(self):\n pass",
"def test_aws_service_api_volumes_get(self):\n pass",
"def test_volumes_get(self):\n pass",
"def test_openstack_rest_volumes_get(self):\n pass",
"def test_aws_service_api_volume_get(self):\n pass",
"def test_v1vm_addvolume(self):\n ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |
An observation for Deep Xi (noisyspeech STMS and STPS). | def observation(self, x):
x = self.normalise(x)
x_STMS, x_STPS = self.polar_analysis(x)
x_STMS_STPS = tf.concat([x_STMS, x_STPS], axis=-1)
return x_STMS_STPS, x_STMS_STPS | [
"def giveObservation(self, obs):\n pass",
"def temporal_noise_sampler(self, num_samples):\n prev_observation = self._env.reset()\n\n # random sample of action value and action duration sampled from a uniform distribution\n u_noise_val = np.random.uniform(-2, 2)\n u_noise_time = ... | {
"objective": {
"paired": [],
"self": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
} |