source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self):
limits = {}
limits['Max spot instance requests per region'] = AwsLimit(
'Max spot instance requests per region', self, 20, self.
warning_threshold, self.critical_threshold, limit_type=
'Spot instance requests')
limits['Max active spot fleets per region'] = AwsLimit(
'Max active spot fleets per region', self, 1000, self.
warning_threshold, self.critical_threshold)
limits['Max launch specifications per spot fleet'] = AwsLimit(
'Max launch specifications per spot fleet', self, 50, self.
warning_threshold, self.critical_threshold)
limits['Max target capacity per spot fleet'] = AwsLimit(
'Max target capacity per spot fleet', self, 3000, self.
warning_threshold, self.critical_threshold)
limits['Max target capacity for all spot fleets in region'] = AwsLimit(
'Max target capacity for all spot fleets in region', self, 5000,
self.warning_threshold, self.critical_threshold)
return limits
| get_limits_spot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n limits = {}\n limits['Max spot instance requests per region'] = AwsLimit(\n 'Max spot instance requests per region', self, 20, self.\n warning_threshold, self.critical_threshold, limit_type=\n 'Spot instance requests')\n limits['Max active spot fleets per region'] = AwsLimit(\n 'Max active spot fleets per region', self, 1000, self.\n warning_threshold, self.critical_threshold)\n limits['Max launch specifications per spot fleet'] = AwsLimit(\n 'Max launch specifications per spot fleet', self, 50, self.\n warning_threshold, self.critical_threshold)\n limits['Max target capacity per spot fleet'] = AwsLimit(\n 'Max target capacity per spot fleet', self, 3000, self.\n warning_threshold, self.critical_threshold)\n limits['Max target capacity for all spot fleets in region'] = AwsLimit(\n 'Max target capacity for all spot fleets in region', self, 5000,\n self.warning_threshold, self.critical_threshold)\n return limits\n",
"role": "user"
},
{
"content": "get_limits_spot",
"role": "assistant"
}
] |
def x(self):
logger.debug('Getting usage for EC2 VPC resources')
sgs_per_vpc = defaultdict(int)
rules_per_sg = defaultdict(int)
for sg in self.resource_conn.security_groups.all():
if sg.vpc_id is not None:
sgs_per_vpc[sg.vpc_id] += 1
rules_per_sg[sg.id] = len(sg.ip_permissions)
for vpc_id, count in sgs_per_vpc.items():
self.limits['Security groups per VPC']._add_current_usage(count,
aws_type='AWS::EC2::VPC', resource_id=vpc_id)
for sg_id, count in rules_per_sg.items():
self.limits['Rules per VPC security group']._add_current_usage(count,
aws_type='AWS::EC2::SecurityGroupRule', resource_id=sg_id)
| find_usage_networking_sgs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n logger.debug('Getting usage for EC2 VPC resources')\n sgs_per_vpc = defaultdict(int)\n rules_per_sg = defaultdict(int)\n for sg in self.resource_conn.security_groups.all():\n if sg.vpc_id is not None:\n sgs_per_vpc[sg.vpc_id] += 1\n rules_per_sg[sg.id] = len(sg.ip_permissions)\n for vpc_id, count in sgs_per_vpc.items():\n self.limits['Security groups per VPC']._add_current_usage(count,\n aws_type='AWS::EC2::VPC', resource_id=vpc_id)\n for sg_id, count in rules_per_sg.items():\n self.limits['Rules per VPC security group']._add_current_usage(count,\n aws_type='AWS::EC2::SecurityGroupRule', resource_id=sg_id)\n",
"role": "user"
},
{
"content": "find_usage_networking_sgs",
"role": "assistant"
}
] |
def x(self):
limits = {}
limits['Security groups per VPC'] = AwsLimit('Security groups per VPC',
self, 500, self.warning_threshold, self.critical_threshold,
limit_type='AWS::EC2::SecurityGroup', limit_subtype='AWS::EC2::VPC')
limits['Rules per VPC security group'] = AwsLimit(
'Rules per VPC security group', self, 50, self.warning_threshold,
self.critical_threshold, limit_type='AWS::EC2::SecurityGroup',
limit_subtype='AWS::EC2::VPC')
limits['VPC Elastic IP addresses (EIPs)'] = AwsLimit(
'VPC Elastic IP addresses (EIPs)', self, 5, self.warning_threshold,
self.critical_threshold, limit_type='AWS::EC2::EIP', limit_subtype=
'AWS::EC2::VPC', ta_service_name='VPC')
limits['Elastic IP addresses (EIPs)'] = AwsLimit(
'Elastic IP addresses (EIPs)', self, 5, self.warning_threshold,
self.critical_threshold, limit_type='AWS::EC2::EIP')
limits['VPC security groups per elastic network interface'] = AwsLimit(
'VPC security groups per elastic network interface', self, 5, self.
warning_threshold, self.critical_threshold, limit_type=
'AWS::EC2::SecurityGroup', limit_subtype='AWS::EC2::NetworkInterface')
return limits
| get_limits_networking | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n limits = {}\n limits['Security groups per VPC'] = AwsLimit('Security groups per VPC',\n self, 500, self.warning_threshold, self.critical_threshold,\n limit_type='AWS::EC2::SecurityGroup', limit_subtype='AWS::EC2::VPC')\n limits['Rules per VPC security group'] = AwsLimit(\n 'Rules per VPC security group', self, 50, self.warning_threshold,\n self.critical_threshold, limit_type='AWS::EC2::SecurityGroup',\n limit_subtype='AWS::EC2::VPC')\n limits['VPC Elastic IP addresses (EIPs)'] = AwsLimit(\n 'VPC Elastic IP addresses (EIPs)', self, 5, self.warning_threshold,\n self.critical_threshold, limit_type='AWS::EC2::EIP', limit_subtype=\n 'AWS::EC2::VPC', ta_service_name='VPC')\n limits['Elastic IP addresses (EIPs)'] = AwsLimit(\n 'Elastic IP addresses (EIPs)', self, 5, self.warning_threshold,\n self.critical_threshold, limit_type='AWS::EC2::EIP')\n limits['VPC security groups per elastic network interface'] = AwsLimit(\n 'VPC security groups per elastic network interface', self, 5, self.\n warning_threshold, self.critical_threshold, limit_type=\n 'AWS::EC2::SecurityGroup', limit_subtype='AWS::EC2::NetworkInterface')\n return limits\n",
"role": "user"
},
{
"content": "get_limits_networking",
"role": "assistant"
}
] |
def x(self):
GENERAL_TYPES = ['a1.2xlarge', 'a1.4xlarge', 'a1.large', 'a1.medium',
'a1.xlarge', 't2.nano', 't2.micro', 't2.small', 't2.medium',
't2.large', 't2.xlarge', 't2.2xlarge', 't3.nano', 't3.micro',
't3.small', 't3.medium', 't3.large', 't3.xlarge', 't3.2xlarge',
'm3.medium', 'm3.large', 'm3.xlarge', 'm3.2xlarge', 'm4.large',
'm4.xlarge', 'm4.2xlarge', 'm4.4xlarge', 'm4.10xlarge',
'm4.16xlarge', 'm5.12xlarge', 'm5.24xlarge', 'm5.2xlarge',
'm5.4xlarge', 'm5.large', 'm5.xlarge', 'm5d.12xlarge',
'm5d.24xlarge', 'm5d.2xlarge', 'm5d.4xlarge', 'm5d.large',
'm5d.xlarge', 'm5a.12xlarge', 'm5a.24xlarge', 'm5a.2xlarge',
'm5a.4xlarge', 'm5a.large', 'm5a.xlarge']
PREV_GENERAL_TYPES = ['t1.micro', 'm1.small', 'm1.medium', 'm1.large',
'm1.xlarge']
MEMORY_TYPES = ['r3.2xlarge', 'r3.4xlarge', 'r3.8xlarge', 'r3.large',
'r3.xlarge', 'r4.2xlarge', 'r4.4xlarge', 'r4.8xlarge',
'r4.16xlarge', 'r4.large', 'r4.xlarge', 'r5.2xlarge', 'r5.4xlarge',
'r5.8xlarge', 'r5.12xlarge', 'r5.16xlarge', 'r5.24xlarge',
'r5.large', 'r5.metal', 'r5.xlarge', 'r5a.12xlarge', 'r5a.24xlarge',
'r5a.2xlarge', 'r5a.4xlarge', 'r5a.large', 'r5a.xlarge',
'r5d.2xlarge', 'r5d.4xlarge', 'r5d.8xlarge', 'r5d.12xlarge',
'r5d.16xlarge', 'r5d.24xlarge', 'r5d.large', 'r5d.metal',
'r5d.xlarge', 'x1.16xlarge', 'x1.32xlarge', 'x1e.2xlarge',
'x1e.4xlarge', 'x1e.8xlarge', 'x1e.16xlarge', 'x1e.32xlarge',
'x1e.xlarge', 'z1d.2xlarge', 'z1d.3xlarge', 'z1d.6xlarge',
'z1d.12xlarge', 'z1d.large', 'z1d.xlarge']
PREV_MEMORY_TYPES = ['m2.xlarge', 'm2.2xlarge', 'm2.4xlarge', 'cr1.8xlarge'
]
COMPUTE_TYPES = ['c3.large', 'c3.xlarge', 'c3.2xlarge', 'c3.4xlarge',
'c3.8xlarge', 'c4.large', 'c4.xlarge', 'c4.2xlarge', 'c4.4xlarge',
'c4.8xlarge', 'c5.18xlarge', 'c5.2xlarge', 'c5.4xlarge',
'c5.9xlarge', 'c5.large', 'c5.xlarge', 'c5d.18xlarge',
'c5d.2xlarge', 'c5d.4xlarge', 'c5d.9xlarge', 'c5d.large',
'c5d.xlarge', 'c5n.18xlarge', 'c5n.2xlarge', 'c5n.4xlarge',
'c5n.9xlarge', 'c5n.large', 'c5n.xlarge']
PREV_COMPUTE_TYPES = ['c1.medium', 'c1.xlarge', 'cc2.8xlarge',
'cc1.4xlarge']
ACCELERATED_COMPUTE_TYPES = ['f1.4xlarge', 'p2.xlarge', 'p2.8xlarge',
'p2.16xlarge', 'p3.16xlarge', 'p3.2xlarge', 'p3.8xlarge',
'p3dn.24xlarge']
STORAGE_TYPES = ['i2.xlarge', 'i2.2xlarge', 'i2.4xlarge', 'i2.8xlarge',
'i3.large', 'i3.xlarge', 'i3.2xlarge', 'i3.4xlarge', 'i3.8xlarge',
'i3.16xlarge', 'i3.metal', 'h1.16xlarge', 'h1.2xlarge',
'h1.4xlarge', 'h1.8xlarge']
PREV_STORAGE_TYPES = ['hi1.4xlarge', 'hs1.8xlarge']
DENSE_STORAGE_TYPES = ['d2.xlarge', 'd2.2xlarge', 'd2.4xlarge',
'd2.8xlarge']
GPU_TYPES = ['g2.2xlarge', 'g2.8xlarge', 'g3.16xlarge', 'g3.4xlarge',
'g3.8xlarge', 'g3s.xlarge']
PREV_GPU_TYPES = ['cg1.4xlarge']
FPGA_TYPES = ['f1.2xlarge', 'f1.16xlarge']
return (GENERAL_TYPES + PREV_GENERAL_TYPES + MEMORY_TYPES +
PREV_MEMORY_TYPES + COMPUTE_TYPES + PREV_COMPUTE_TYPES +
ACCELERATED_COMPUTE_TYPES + STORAGE_TYPES + PREV_STORAGE_TYPES +
DENSE_STORAGE_TYPES + GPU_TYPES + PREV_GPU_TYPES + FPGA_TYPES)
| instance_types | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n GENERAL_TYPES = ['a1.2xlarge', 'a1.4xlarge', 'a1.large', 'a1.medium',\n 'a1.xlarge', 't2.nano', 't2.micro', 't2.small', 't2.medium',\n 't2.large', 't2.xlarge', 't2.2xlarge', 't3.nano', 't3.micro',\n 't3.small', 't3.medium', 't3.large', 't3.xlarge', 't3.2xlarge',\n 'm3.medium', 'm3.large', 'm3.xlarge', 'm3.2xlarge', 'm4.large',\n 'm4.xlarge', 'm4.2xlarge', 'm4.4xlarge', 'm4.10xlarge',\n 'm4.16xlarge', 'm5.12xlarge', 'm5.24xlarge', 'm5.2xlarge',\n 'm5.4xlarge', 'm5.large', 'm5.xlarge', 'm5d.12xlarge',\n 'm5d.24xlarge', 'm5d.2xlarge', 'm5d.4xlarge', 'm5d.large',\n 'm5d.xlarge', 'm5a.12xlarge', 'm5a.24xlarge', 'm5a.2xlarge',\n 'm5a.4xlarge', 'm5a.large', 'm5a.xlarge']\n PREV_GENERAL_TYPES = ['t1.micro', 'm1.small', 'm1.medium', 'm1.large',\n 'm1.xlarge']\n MEMORY_TYPES = ['r3.2xlarge', 'r3.4xlarge', 'r3.8xlarge', 'r3.large',\n 'r3.xlarge', 'r4.2xlarge', 'r4.4xlarge', 'r4.8xlarge',\n 'r4.16xlarge', 'r4.large', 'r4.xlarge', 'r5.2xlarge', 'r5.4xlarge',\n 'r5.8xlarge', 'r5.12xlarge', 'r5.16xlarge', 'r5.24xlarge',\n 'r5.large', 'r5.metal', 'r5.xlarge', 'r5a.12xlarge', 'r5a.24xlarge',\n 'r5a.2xlarge', 'r5a.4xlarge', 'r5a.large', 'r5a.xlarge',\n 'r5d.2xlarge', 'r5d.4xlarge', 'r5d.8xlarge', 'r5d.12xlarge',\n 'r5d.16xlarge', 'r5d.24xlarge', 'r5d.large', 'r5d.metal',\n 'r5d.xlarge', 'x1.16xlarge', 'x1.32xlarge', 'x1e.2xlarge',\n 'x1e.4xlarge', 'x1e.8xlarge', 'x1e.16xlarge', 'x1e.32xlarge',\n 'x1e.xlarge', 'z1d.2xlarge', 'z1d.3xlarge', 'z1d.6xlarge',\n 'z1d.12xlarge', 'z1d.large', 'z1d.xlarge']\n PREV_MEMORY_TYPES = ['m2.xlarge', 'm2.2xlarge', 'm2.4xlarge', 'cr1.8xlarge'\n ]\n COMPUTE_TYPES = ['c3.large', 'c3.xlarge', 'c3.2xlarge', 'c3.4xlarge',\n 'c3.8xlarge', 'c4.large', 'c4.xlarge', 'c4.2xlarge', 'c4.4xlarge',\n 'c4.8xlarge', 'c5.18xlarge', 'c5.2xlarge', 'c5.4xlarge',\n 'c5.9xlarge', 'c5.large', 'c5.xlarge', 'c5d.18xlarge',\n 'c5d.2xlarge', 'c5d.4xlarge', 'c5d.9xlarge', 'c5d.large',\n 'c5d.xlarge', 'c5n.18xlarge', 'c5n.2xlarge', 'c5n.4xlarge',\n 'c5n.9xlarge', 'c5n.large', 'c5n.xlarge']\n PREV_COMPUTE_TYPES = ['c1.medium', 'c1.xlarge', 'cc2.8xlarge',\n 'cc1.4xlarge']\n ACCELERATED_COMPUTE_TYPES = ['f1.4xlarge', 'p2.xlarge', 'p2.8xlarge',\n 'p2.16xlarge', 'p3.16xlarge', 'p3.2xlarge', 'p3.8xlarge',\n 'p3dn.24xlarge']\n STORAGE_TYPES = ['i2.xlarge', 'i2.2xlarge', 'i2.4xlarge', 'i2.8xlarge',\n 'i3.large', 'i3.xlarge', 'i3.2xlarge', 'i3.4xlarge', 'i3.8xlarge',\n 'i3.16xlarge', 'i3.metal', 'h1.16xlarge', 'h1.2xlarge',\n 'h1.4xlarge', 'h1.8xlarge']\n PREV_STORAGE_TYPES = ['hi1.4xlarge', 'hs1.8xlarge']\n DENSE_STORAGE_TYPES = ['d2.xlarge', 'd2.2xlarge', 'd2.4xlarge',\n 'd2.8xlarge']\n GPU_TYPES = ['g2.2xlarge', 'g2.8xlarge', 'g3.16xlarge', 'g3.4xlarge',\n 'g3.8xlarge', 'g3s.xlarge']\n PREV_GPU_TYPES = ['cg1.4xlarge']\n FPGA_TYPES = ['f1.2xlarge', 'f1.16xlarge']\n return (GENERAL_TYPES + PREV_GENERAL_TYPES + MEMORY_TYPES +\n PREV_MEMORY_TYPES + COMPUTE_TYPES + PREV_COMPUTE_TYPES +\n ACCELERATED_COMPUTE_TYPES + STORAGE_TYPES + PREV_STORAGE_TYPES +\n DENSE_STORAGE_TYPES + GPU_TYPES + PREV_GPU_TYPES + FPGA_TYPES)\n",
"role": "user"
},
{
"content": "instance_types",
"role": "assistant"
}
] |
def x(self):
logger.debug('Checking usage for service %s', self.service_name)
self.connect()
for lim in self.limits.values():
lim._reset_usage()
self._find_usage_instances()
self._find_usage_subnet_groups()
self._find_usage_security_groups()
self._update_limits_from_api()
self._have_usage = True
logger.debug('Done checking usage.')
| find_usage | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n logger.debug('Checking usage for service %s', self.service_name)\n self.connect()\n for lim in self.limits.values():\n lim._reset_usage()\n self._find_usage_instances()\n self._find_usage_subnet_groups()\n self._find_usage_security_groups()\n self._update_limits_from_api()\n self._have_usage = True\n logger.debug('Done checking usage.')\n",
"role": "user"
},
{
"content": "find_usage",
"role": "assistant"
}
] |
def x(self):
paginator = self.conn.get_paginator('describe_db_instances')
for page in paginator.paginate():
for instance in page['DBInstances']:
self.limits['Read replicas per master']._add_current_usage(len(
instance['ReadReplicaDBInstanceIdentifiers']), aws_type=
'AWS::RDS::DBInstance', resource_id=instance[
'DBInstanceIdentifier'])
| find_usage_instances | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n paginator = self.conn.get_paginator('describe_db_instances')\n for page in paginator.paginate():\n for instance in page['DBInstances']:\n self.limits['Read replicas per master']._add_current_usage(len(\n instance['ReadReplicaDBInstanceIdentifiers']), aws_type=\n 'AWS::RDS::DBInstance', resource_id=instance[\n 'DBInstanceIdentifier'])\n",
"role": "user"
},
{
"content": "find_usage_instances",
"role": "assistant"
}
] |
def x(self):
paginator = self.conn.get_paginator('describe_db_subnet_groups')
for page in paginator.paginate():
for group in page['DBSubnetGroups']:
self.limits['Subnets per Subnet Group']._add_current_usage(len(
group['Subnets']), aws_type='AWS::RDS::DBSubnetGroup',
resource_id=group['DBSubnetGroupName'])
| find_usage_subnet_groups | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n paginator = self.conn.get_paginator('describe_db_subnet_groups')\n for page in paginator.paginate():\n for group in page['DBSubnetGroups']:\n self.limits['Subnets per Subnet Group']._add_current_usage(len(\n group['Subnets']), aws_type='AWS::RDS::DBSubnetGroup',\n resource_id=group['DBSubnetGroupName'])\n",
"role": "user"
},
{
"content": "find_usage_subnet_groups",
"role": "assistant"
}
] |
def x(self):
vpc_count = 0
paginator = self.conn.get_paginator('describe_db_security_groups')
for page in paginator.paginate():
for group in page['DBSecurityGroups']:
if 'VpcId' in group and group['VpcId'] is not None:
vpc_count += 1
self.limits['Max auths per security group']._add_current_usage(
len(group['EC2SecurityGroups']) + len(group['IPRanges']),
aws_type='AWS::RDS::DBSecurityGroup', resource_id=group[
'DBSecurityGroupName'])
self.limits['VPC Security Groups']._add_current_usage(vpc_count,
aws_type='AWS::RDS::DBSecurityGroup')
| find_usage_security_groups | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n vpc_count = 0\n paginator = self.conn.get_paginator('describe_db_security_groups')\n for page in paginator.paginate():\n for group in page['DBSecurityGroups']:\n if 'VpcId' in group and group['VpcId'] is not None:\n vpc_count += 1\n self.limits['Max auths per security group']._add_current_usage(\n len(group['EC2SecurityGroups']) + len(group['IPRanges']),\n aws_type='AWS::RDS::DBSecurityGroup', resource_id=group[\n 'DBSecurityGroupName'])\n self.limits['VPC Security Groups']._add_current_usage(vpc_count,\n aws_type='AWS::RDS::DBSecurityGroup')\n",
"role": "user"
},
{
"content": "find_usage_security_groups",
"role": "assistant"
}
] |
def x(self):
if self.limits != {}:
return self.limits
limits = {}
limits['DB instances'] = AwsLimit('DB instances', self, 40, self.
warning_threshold, self.critical_threshold, limit_type=
'AWS::RDS::DBInstance')
limits['Reserved Instances'] = AwsLimit('Reserved Instances', self, 40,
self.warning_threshold, self.critical_threshold, limit_type=
'AWS::RDS::DBInstance')
limits['Storage quota (GB)'] = AwsLimit('Storage quota (GB)', self,
100000, self.warning_threshold, self.critical_threshold, limit_type
='AWS::RDS::DBInstance')
limits['DB snapshots per user'] = AwsLimit('DB snapshots per user',
self, 100, self.warning_threshold, self.critical_threshold,
limit_type='AWS::RDS::DBSnapshot')
limits['DB parameter groups'] = AwsLimit('DB parameter groups', self,
50, self.warning_threshold, self.critical_threshold, limit_type=
'AWS::RDS::DBParameterGroup')
limits['DB security groups'] = AwsLimit('DB security groups', self, 25,
self.warning_threshold, self.critical_threshold, limit_type=
'AWS::RDS::DBSecurityGroup')
limits['VPC Security Groups'] = AwsLimit('VPC Security Groups', self, 5,
self.warning_threshold, self.critical_threshold, limit_type=
'AWS::RDS::DBSecurityGroup')
limits['Subnet Groups'] = AwsLimit('Subnet Groups', self, 50, self.
warning_threshold, self.critical_threshold, limit_type=
'AWS::RDS::DBSubnetGroup', ta_limit_name='Subnet groups')
limits['Subnets per Subnet Group'] = AwsLimit('Subnets per Subnet Group',
self, 20, self.warning_threshold, self.critical_threshold,
limit_type='AWS::RDS::DBSubnetGroup', ta_limit_name=
'Subnets per subnet group')
limits['Option Groups'] = AwsLimit('Option Groups', self, 20, self.
warning_threshold, self.critical_threshold, limit_type=
'AWS::RDS::DBOptionGroup')
limits['Event Subscriptions'] = AwsLimit('Event Subscriptions', self,
20, self.warning_threshold, self.critical_threshold, limit_type=
'AWS::RDS::DBEventSubscription', ta_limit_name='Event subscriptions')
limits['Read replicas per master'] = AwsLimit('Read replicas per master',
self, 5, self.warning_threshold, self.critical_threshold,
limit_type='AWS::RDS::DBInstance')
limits['Max auths per security group'] = AwsLimit(
'Max auths per security group', self, 20, self.warning_threshold,
self.critical_threshold, limit_type='AWS::RDS::DBSecurityGroup',
limit_subtype='AWS::RDS::DBSecurityGroupIngress')
limits['DB Clusters'] = AwsLimit('DB Clusters', self, 40, self.
warning_threshold, self.critical_threshold, limit_type=
'AWS::RDS::DBCluster', ta_limit_name='Clusters')
limits['DB Cluster Parameter Groups'] = AwsLimit(
'DB Cluster Parameter Groups', self, 50, self.warning_threshold,
self.critical_threshold, limit_type=
'AWS::RDS::DBClusterParameterGroup', ta_limit_name=
'Cluster parameter groups')
self.limits = limits
return limits
| get_limits | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.limits != {}:\n return self.limits\n limits = {}\n limits['DB instances'] = AwsLimit('DB instances', self, 40, self.\n warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBInstance')\n limits['Reserved Instances'] = AwsLimit('Reserved Instances', self, 40,\n self.warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBInstance')\n limits['Storage quota (GB)'] = AwsLimit('Storage quota (GB)', self, \n 100000, self.warning_threshold, self.critical_threshold, limit_type\n ='AWS::RDS::DBInstance')\n limits['DB snapshots per user'] = AwsLimit('DB snapshots per user',\n self, 100, self.warning_threshold, self.critical_threshold,\n limit_type='AWS::RDS::DBSnapshot')\n limits['DB parameter groups'] = AwsLimit('DB parameter groups', self, \n 50, self.warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBParameterGroup')\n limits['DB security groups'] = AwsLimit('DB security groups', self, 25,\n self.warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBSecurityGroup')\n limits['VPC Security Groups'] = AwsLimit('VPC Security Groups', self, 5,\n self.warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBSecurityGroup')\n limits['Subnet Groups'] = AwsLimit('Subnet Groups', self, 50, self.\n warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBSubnetGroup', ta_limit_name='Subnet groups')\n limits['Subnets per Subnet Group'] = AwsLimit('Subnets per Subnet Group',\n self, 20, self.warning_threshold, self.critical_threshold,\n limit_type='AWS::RDS::DBSubnetGroup', ta_limit_name=\n 'Subnets per subnet group')\n limits['Option Groups'] = AwsLimit('Option Groups', self, 20, self.\n warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBOptionGroup')\n limits['Event Subscriptions'] = AwsLimit('Event Subscriptions', self, \n 20, self.warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBEventSubscription', ta_limit_name='Event subscriptions')\n limits['Read replicas per master'] = AwsLimit('Read replicas per master',\n self, 5, self.warning_threshold, self.critical_threshold,\n limit_type='AWS::RDS::DBInstance')\n limits['Max auths per security group'] = AwsLimit(\n 'Max auths per security group', self, 20, self.warning_threshold,\n self.critical_threshold, limit_type='AWS::RDS::DBSecurityGroup',\n limit_subtype='AWS::RDS::DBSecurityGroupIngress')\n limits['DB Clusters'] = AwsLimit('DB Clusters', self, 40, self.\n warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBCluster', ta_limit_name='Clusters')\n limits['DB Cluster Parameter Groups'] = AwsLimit(\n 'DB Cluster Parameter Groups', self, 50, self.warning_threshold,\n self.critical_threshold, limit_type=\n 'AWS::RDS::DBClusterParameterGroup', ta_limit_name=\n 'Cluster parameter groups')\n self.limits = limits\n return limits\n",
"role": "user"
},
{
"content": "get_limits",
"role": "assistant"
}
] |
def x(self):
self.connect()
logger.info('Querying RDS DescribeAccountAttributes for limits')
lims = self.conn.describe_account_attributes()['AccountQuotas']
for lim in lims:
if lim['AccountQuotaName'] not in self.API_NAME_TO_LIMIT:
logger.info(
'RDS DescribeAccountAttributes returned unknownlimit: %s (max: %s; used: %s)'
, lim['AccountQuotaName'], lim['Max'], lim['Used'])
continue
lname = self.API_NAME_TO_LIMIT[lim['AccountQuotaName']]
self.limits[lname]._set_api_limit(lim['Max'])
if len(self.limits[lname].get_current_usage()) < 1:
self.limits[lname]._add_current_usage(lim['Used'])
logger.debug('Done setting limits from API.')
| update_limits_from_api | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.connect()\n logger.info('Querying RDS DescribeAccountAttributes for limits')\n lims = self.conn.describe_account_attributes()['AccountQuotas']\n for lim in lims:\n if lim['AccountQuotaName'] not in self.API_NAME_TO_LIMIT:\n logger.info(\n 'RDS DescribeAccountAttributes returned unknownlimit: %s (max: %s; used: %s)'\n , lim['AccountQuotaName'], lim['Max'], lim['Used'])\n continue\n lname = self.API_NAME_TO_LIMIT[lim['AccountQuotaName']]\n self.limits[lname]._set_api_limit(lim['Max'])\n if len(self.limits[lname].get_current_usage()) < 1:\n self.limits[lname]._add_current_usage(lim['Used'])\n logger.debug('Done setting limits from API.')\n",
"role": "user"
},
{
"content": "update_limits_from_api",
"role": "assistant"
}
] |
def x(self):
logger.debug('Checking usage for service %s', self.service_name)
self.connect()
for lim in self.limits.values():
lim._reset_usage()
try:
self._find_delivery_streams()
except EndpointConnectionError as ex:
logger.warning(
'Caught exception when trying to use Firehose (perhaps the Firehose service is not available in this region?): %s'
, ex)
self._have_usage = True
logger.debug('Done checking usage.')
| find_usage | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n logger.debug('Checking usage for service %s', self.service_name)\n self.connect()\n for lim in self.limits.values():\n lim._reset_usage()\n try:\n self._find_delivery_streams()\n except EndpointConnectionError as ex:\n logger.warning(\n 'Caught exception when trying to use Firehose (perhaps the Firehose service is not available in this region?): %s'\n , ex)\n self._have_usage = True\n logger.debug('Done checking usage.')\n",
"role": "user"
},
{
"content": "find_usage",
"role": "assistant"
}
] |
def x(self):
trail_list = self.conn.describe_trails()['trailList']
trail_count = len(trail_list) if trail_list else 0
for trail in trail_list:
data_resource_count = 0
if self.conn._client_config.region_name == trail['HomeRegion']:
response = self.conn.get_event_selectors(TrailName=trail['Name'])
event_selectors = response['EventSelectors']
for event_selector in event_selectors:
data_resource_count += len(event_selector.get(
'DataResources', []))
self.limits['Event Selectors Per Trail']._add_current_usage(len
(event_selectors), aws_type=
'AWS::CloudTrail::EventSelector', resource_id=trail['Name'])
self.limits['Data Resources Per Trail']._add_current_usage(
data_resource_count, aws_type=
'AWS::CloudTrail::DataResource', resource_id=trail['Name'])
else:
logger.debug(
'Ignoring event selectors and data resources for CloudTrail %s in non-home region'
% trail['Name'])
self.limits['Trails Per Region']._add_current_usage(trail_count,
aws_type=self.aws_type)
| find_usage_cloudtrail | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n trail_list = self.conn.describe_trails()['trailList']\n trail_count = len(trail_list) if trail_list else 0\n for trail in trail_list:\n data_resource_count = 0\n if self.conn._client_config.region_name == trail['HomeRegion']:\n response = self.conn.get_event_selectors(TrailName=trail['Name'])\n event_selectors = response['EventSelectors']\n for event_selector in event_selectors:\n data_resource_count += len(event_selector.get(\n 'DataResources', []))\n self.limits['Event Selectors Per Trail']._add_current_usage(len\n (event_selectors), aws_type=\n 'AWS::CloudTrail::EventSelector', resource_id=trail['Name'])\n self.limits['Data Resources Per Trail']._add_current_usage(\n data_resource_count, aws_type=\n 'AWS::CloudTrail::DataResource', resource_id=trail['Name'])\n else:\n logger.debug(\n 'Ignoring event selectors and data resources for CloudTrail %s in non-home region'\n % trail['Name'])\n self.limits['Trails Per Region']._add_current_usage(trail_count,\n aws_type=self.aws_type)\n",
"role": "user"
},
{
"content": "find_usage_cloudtrail",
"role": "assistant"
}
] |
def x(self):
logger.debug("Gathering %s's limits from AWS", self.service_name)
if self.limits:
return self.limits
limits = {}
limits['Trails Per Region'] = AwsLimit('Trails Per Region', self, 5,
self.warning_threshold, self.critical_threshold, limit_type=self.
aws_type)
limits['Event Selectors Per Trail'] = AwsLimit('Event Selectors Per Trail',
self, 5, self.warning_threshold, self.critical_threshold,
limit_type=self.aws_type, limit_subtype=
'AWS::CloudTrail::EventSelector')
limits['Data Resources Per Trail'] = AwsLimit('Data Resources Per Trail',
self, 250, self.warning_threshold, self.critical_threshold,
limit_type=self.aws_type, limit_subtype='AWS::CloudTrail::DataResource'
)
self.limits = limits
return limits
| get_limits | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n logger.debug(\"Gathering %s's limits from AWS\", self.service_name)\n if self.limits:\n return self.limits\n limits = {}\n limits['Trails Per Region'] = AwsLimit('Trails Per Region', self, 5,\n self.warning_threshold, self.critical_threshold, limit_type=self.\n aws_type)\n limits['Event Selectors Per Trail'] = AwsLimit('Event Selectors Per Trail',\n self, 5, self.warning_threshold, self.critical_threshold,\n limit_type=self.aws_type, limit_subtype=\n 'AWS::CloudTrail::EventSelector')\n limits['Data Resources Per Trail'] = AwsLimit('Data Resources Per Trail',\n self, 250, self.warning_threshold, self.critical_threshold,\n limit_type=self.aws_type, limit_subtype='AWS::CloudTrail::DataResource'\n )\n self.limits = limits\n return limits\n",
"role": "user"
},
{
"content": "get_limits",
"role": "assistant"
}
] |
def x(self, ranges=None):
fig = self.generate_plot(self.keys[-1], ranges)
self.drawn = True
return fig
| initialize_plot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ranges=None):\n fig = self.generate_plot(self.keys[-1], ranges)\n self.drawn = True\n return fig\n",
"role": "user"
},
{
"content": "initialize_plot",
"role": "assistant"
}
] |
def x(self, element):
dims = element.dimensions()[:3]
pad = [None] * max(3 - len(dims), 0)
return dims + pad
| get_axis_dims | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, element):\n dims = element.dimensions()[:3]\n pad = [None] * max(3 - len(dims), 0)\n return dims + pad\n",
"role": "user"
},
{
"content": "get_axis_dims",
"role": "assistant"
}
] |
def x(self, key, ranges=None, element=None):
self.generate_plot(key, ranges, element)
| update_frame | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, ranges=None, element=None):\n self.generate_plot(key, ranges, element)\n",
"role": "user"
},
{
"content": "update_frame",
"role": "assistant"
}
] |
def x(self, ranges=None):
return self.generate_plot(list(self.hmap.data.keys())[0], ranges)
| initialize_plot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ranges=None):\n return self.generate_plot(list(self.hmap.data.keys())[0], ranges)\n",
"role": "user"
},
{
"content": "initialize_plot",
"role": "assistant"
}
] |
def x(self):
if 'holoviews' not in sys.modules:
raise ImportError('HoloViews does not seem to be imported')
matches = [k for k, v in get_ipython().user_ns.items() if not k.
startswith('_') and v is sys.modules['holoviews']]
if len(matches) == 0:
raise Exception('Could not find holoviews module in namespace')
return '%s.archive' % matches[0]
| get_namespace | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if 'holoviews' not in sys.modules:\n raise ImportError('HoloViews does not seem to be imported')\n matches = [k for k, v in get_ipython().user_ns.items() if not k.\n startswith('_') and v is sys.modules['holoviews']]\n if len(matches) == 0:\n raise Exception('Could not find holoviews module in namespace')\n return '%s.archive' % matches[0]\n",
"role": "user"
},
{
"content": "get_namespace",
"role": "assistant"
}
] |
def x(self):
if self.export_success is True:
print('The last call to holoviews.archive.export was successful.')
return
elif self.export_success is None:
print(
"""Status of the last call to holoviews.archive.export is unknown.
(Re-execute this method once kernel status is idle.)"""
)
return
print('The last call to holoviews.archive.export was unsuccessful.')
if self.traceback is None:
print('\n<No traceback captured>')
else:
print('\n' + self.traceback)
| last_export_status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.export_success is True:\n print('The last call to holoviews.archive.export was successful.')\n return\n elif self.export_success is None:\n print(\n \"\"\"Status of the last call to holoviews.archive.export is unknown.\n(Re-execute this method once kernel status is idle.)\"\"\"\n )\n return\n print('The last call to holoviews.archive.export was unsuccessful.')\n if self.traceback is None:\n print('\\n<No traceback captured>')\n else:\n print('\\n' + self.traceback)\n",
"role": "user"
},
{
"content": "last_export_status",
"role": "assistant"
}
] |
def x(self, enabled=True, **kwargs):
self.namespace = self.get_namespace()
self.notebook_name = '{notebook}'
self._timestamp = tuple(time.localtime())
kernel = 'var kernel = IPython.notebook.kernel; '
nbname = 'var nbname = IPython.notebook.get_notebook_name(); '
nbcmd = (
'var name_cmd = \'%s.notebook_name = \\"\' + nbname + \'\\"\'; ' %
self.namespace)
cmd = kernel + nbname + nbcmd + 'kernel.execute(name_cmd); '
display(Javascript(cmd))
time.sleep(0.5)
self._auto = enabled
self.param.set_param(**kwargs)
tstamp = time.strftime(' [%Y-%m-%d %H:%M:%S]', self._timestamp)
print('Automatic capture is now %s.%s' % ('enabled' if enabled else
'disabled', tstamp if enabled else ''))
| auto | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, enabled=True, **kwargs):\n self.namespace = self.get_namespace()\n self.notebook_name = '{notebook}'\n self._timestamp = tuple(time.localtime())\n kernel = 'var kernel = IPython.notebook.kernel; '\n nbname = 'var nbname = IPython.notebook.get_notebook_name(); '\n nbcmd = (\n 'var name_cmd = \\'%s.notebook_name = \\\\\"\\' + nbname + \\'\\\\\"\\'; ' %\n self.namespace)\n cmd = kernel + nbname + nbcmd + 'kernel.execute(name_cmd); '\n display(Javascript(cmd))\n time.sleep(0.5)\n self._auto = enabled\n self.param.set_param(**kwargs)\n tstamp = time.strftime(' [%Y-%m-%d %H:%M:%S]', self._timestamp)\n print('Automatic capture is now %s.%s' % ('enabled' if enabled else\n 'disabled', tstamp if enabled else ''))\n",
"role": "user"
},
{
"content": "auto",
"role": "assistant"
}
] |
def x(self, timestamp=None):
if self._timestamp is None:
raise Exception('No timestamp set. Has the archive been initialized?')
if self.skip_notebook_export:
super(NotebookArchive, self).export(timestamp=self._timestamp, info
={'notebook': self.notebook_name})
return
self.export_success = None
name = self.get_namespace()
capture_cmd = ('var capture = \'%s._notebook_data=r\\"\\"\\"\'' % name +
'+json_string+\'\\"\\"\\"\'; ')
cmd = ('var kernel = IPython.notebook.kernel; ' +
'var json_data = IPython.notebook.toJSON(); ' +
'var json_string = JSON.stringify(json_data); ' + capture_cmd +
"var pycmd = capture + ';%s._export_with_html()'; " % name +
'kernel.execute(pycmd)')
tstamp = time.strftime(self.timestamp_format, self._timestamp)
export_name = self._format(self.export_name, {'timestamp': tstamp,
'notebook': self.notebook_name})
print('Export name: %r\nDirectory %r' % (export_name, os.path.join(
os.path.abspath(self.root))) +
"""
If no output appears, please check holoviews.archive.last_export_status()"""
)
display(Javascript(cmd))
| export | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, timestamp=None):\n if self._timestamp is None:\n raise Exception('No timestamp set. Has the archive been initialized?')\n if self.skip_notebook_export:\n super(NotebookArchive, self).export(timestamp=self._timestamp, info\n ={'notebook': self.notebook_name})\n return\n self.export_success = None\n name = self.get_namespace()\n capture_cmd = ('var capture = \\'%s._notebook_data=r\\\\\"\\\\\"\\\\\"\\'' % name +\n '+json_string+\\'\\\\\"\\\\\"\\\\\"\\'; ')\n cmd = ('var kernel = IPython.notebook.kernel; ' +\n 'var json_data = IPython.notebook.toJSON(); ' +\n 'var json_string = JSON.stringify(json_data); ' + capture_cmd + \n \"var pycmd = capture + ';%s._export_with_html()'; \" % name +\n 'kernel.execute(pycmd)')\n tstamp = time.strftime(self.timestamp_format, self._timestamp)\n export_name = self._format(self.export_name, {'timestamp': tstamp,\n 'notebook': self.notebook_name})\n print('Export name: %r\\nDirectory %r' % (export_name, os.path.join(\n os.path.abspath(self.root))) +\n \"\"\"\n\nIf no output appears, please check holoviews.archive.last_export_status()\"\"\"\n )\n display(Javascript(cmd))\n",
"role": "user"
},
{
"content": "export",
"role": "assistant"
}
] |
def x(self, obj=None, filename=None, data=None, info={}, html=None):
initial_last_key = list(self._files.keys())[-1] if len(self) else None
if self._auto:
exporters = self.exporters[:]
for exporter in exporters:
self.exporters = [exporter]
super(NotebookArchive, self).add(obj, filename, data, info=dict
(info, notebook=self.notebook_name))
new_last_key = list(self._files.keys())[-1] if len(self) else None
if new_last_key != initial_last_key:
self._replacements[new_last_key] = html
self.exporters = exporters
| add | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, obj=None, filename=None, data=None, info={}, html=None):\n initial_last_key = list(self._files.keys())[-1] if len(self) else None\n if self._auto:\n exporters = self.exporters[:]\n for exporter in exporters:\n self.exporters = [exporter]\n super(NotebookArchive, self).add(obj, filename, data, info=dict\n (info, notebook=self.notebook_name))\n new_last_key = list(self._files.keys())[-1] if len(self) else None\n if new_last_key != initial_last_key:\n self._replacements[new_last_key] = html\n self.exporters = exporters\n",
"role": "user"
},
{
"content": "add",
"role": "assistant"
}
] |
def x(self):
self.export_success = False
try:
tstamp = time.strftime(self.timestamp_format, self._timestamp)
substitutions = {}
for (basename, ext), entry in self._files.items():
_, info = entry
html_key = self._replacements.get((basename, ext), None)
if html_key is None:
continue
filename = self._format(basename, {'timestamp': tstamp,
'notebook': self.notebook_name})
fpath = filename + ('.%s' % ext if ext else '')
info = {'src': fpath, 'mime_type': info['mime_type']}
if 'mime_type' not in info:
pass
elif info['mime_type'] not in self._tags:
pass
else:
basename, ext = os.path.splitext(fpath)
truncated = self._truncate_name(basename, ext[1:])
link_html = self._format(self._tags[info['mime_type']], {
'src': truncated, 'mime_type': info['mime_type'], 'css':
''})
substitutions[html_key] = link_html, truncated
node = self._get_notebook_node()
html = self._generate_html(node, substitutions)
export_filename = self.snapshot_name
super(NotebookArchive, self).add(filename=export_filename, data=
html, info={'file-ext': 'html', 'mime_type': 'text/html',
'notebook': self.notebook_name})
cleared = self._clear_notebook(node)
super(NotebookArchive, self).add(filename=export_filename, data=
cleared, info={'file-ext': 'ipynb', 'mime_type': 'text/json',
'notebook': self.notebook_name})
super(NotebookArchive, self).export(timestamp=self._timestamp, info
={'notebook': self.notebook_name})
except:
self.traceback = traceback.format_exc()
else:
self.export_success = True
| export_with_html | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.export_success = False\n try:\n tstamp = time.strftime(self.timestamp_format, self._timestamp)\n substitutions = {}\n for (basename, ext), entry in self._files.items():\n _, info = entry\n html_key = self._replacements.get((basename, ext), None)\n if html_key is None:\n continue\n filename = self._format(basename, {'timestamp': tstamp,\n 'notebook': self.notebook_name})\n fpath = filename + ('.%s' % ext if ext else '')\n info = {'src': fpath, 'mime_type': info['mime_type']}\n if 'mime_type' not in info:\n pass\n elif info['mime_type'] not in self._tags:\n pass\n else:\n basename, ext = os.path.splitext(fpath)\n truncated = self._truncate_name(basename, ext[1:])\n link_html = self._format(self._tags[info['mime_type']], {\n 'src': truncated, 'mime_type': info['mime_type'], 'css':\n ''})\n substitutions[html_key] = link_html, truncated\n node = self._get_notebook_node()\n html = self._generate_html(node, substitutions)\n export_filename = self.snapshot_name\n super(NotebookArchive, self).add(filename=export_filename, data=\n html, info={'file-ext': 'html', 'mime_type': 'text/html',\n 'notebook': self.notebook_name})\n cleared = self._clear_notebook(node)\n super(NotebookArchive, self).add(filename=export_filename, data=\n cleared, info={'file-ext': 'ipynb', 'mime_type': 'text/json',\n 'notebook': self.notebook_name})\n super(NotebookArchive, self).export(timestamp=self._timestamp, info\n ={'notebook': self.notebook_name})\n except:\n self.traceback = traceback.format_exc()\n else:\n self.export_success = True\n",
"role": "user"
},
{
"content": "export_with_html",
"role": "assistant"
}
] |
def x(self):
size = len(self._notebook_data)
if size == 0:
raise Exception('Captured buffer size for notebook node is zero.')
node = reader.reads(self._notebook_data)
self.nbversion = reader.get_version(node)
return node
| get_notebook_node | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n size = len(self._notebook_data)\n if size == 0:\n raise Exception('Captured buffer size for notebook node is zero.')\n node = reader.reads(self._notebook_data)\n self.nbversion = reader.get_version(node)\n return node\n",
"role": "user"
},
{
"content": "get_notebook_node",
"role": "assistant"
}
] |
def x(properties):
new_properties = {}
for k, v in properties.items():
if isinstance(v, dict):
new_properties[k] = v
elif k == 's':
new_properties['size'] = v
elif k == 'marker':
new_properties.update(markers.get(v, {'marker': v}))
elif (k == 'color' or k.endswith('_color')) and not isinstance(v, (
dict, dim)):
with abbreviated_exception():
v = COLOR_ALIASES.get(v, v)
if isinstance(v, tuple):
with abbreviated_exception():
v = rgb2hex(v)
new_properties[k] = v
else:
new_properties[k] = v
new_properties.pop('cmap', None)
return new_properties
| mpl_to_bokeh | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(properties):\n new_properties = {}\n for k, v in properties.items():\n if isinstance(v, dict):\n new_properties[k] = v\n elif k == 's':\n new_properties['size'] = v\n elif k == 'marker':\n new_properties.update(markers.get(v, {'marker': v}))\n elif (k == 'color' or k.endswith('_color')) and not isinstance(v, (\n dict, dim)):\n with abbreviated_exception():\n v = COLOR_ALIASES.get(v, v)\n if isinstance(v, tuple):\n with abbreviated_exception():\n v = rgb2hex(v)\n new_properties[k] = v\n else:\n new_properties[k] = v\n new_properties.pop('cmap', None)\n return new_properties\n",
"role": "user"
},
{
"content": "mpl_to_bokeh",
"role": "assistant"
}
] |
def x(style, value, scalar=False):
validator = get_validator(style)
if validator is None:
return None
if isinstance(value, (np.ndarray, list)):
if scalar:
return False
return all(validator(v) for v in value)
return validator(value)
| validate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(style, value, scalar=False):\n validator = get_validator(style)\n if validator is None:\n return None\n if isinstance(value, (np.ndarray, list)):\n if scalar:\n return False\n return all(validator(v) for v in value)\n return validator(value)\n",
"role": "user"
},
{
"content": "validate",
"role": "assistant"
}
] |
def x(rgba):
if isinstance(rgba, tuple):
return tuple(int(c * 255) if i < 3 else c for i, c in enumerate(rgba))
else:
return COLOR_ALIASES.get(rgba, rgba)
| rgba_tuple | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rgba):\n if isinstance(rgba, tuple):\n return tuple(int(c * 255) if i < 3 else c for i, c in enumerate(rgba))\n else:\n return COLOR_ALIASES.get(rgba, rgba)\n",
"role": "user"
},
{
"content": "rgba_tuple",
"role": "assistant"
}
] |
def x(style, opts, mapping, nvals):
opts = sorted(opts, key=lambda x: x in ['color', 'alpha'])
applied_styles = set(mapping)
style_data, style_mapping = {}, {}
for opt in opts:
if 'color' in opt:
alias = 'color'
elif 'alpha' in opt:
alias = 'alpha'
else:
alias = None
if opt not in style or opt in mapping:
continue
elif opt == alias:
if alias in applied_styles:
continue
elif 'line_' + alias in applied_styles:
if 'fill_' + alias not in opts:
continue
opt = 'fill_' + alias
val = style[alias]
elif 'fill_' + alias in applied_styles:
opt = 'line_' + alias
val = style[alias]
else:
val = style[alias]
else:
val = style[opt]
style_mapping[opt] = {'field': opt}
applied_styles.add(opt)
if 'color' in opt and isinstance(val, tuple):
val = rgb2hex(val)
style_data[opt] = [val] * nvals
return style_data, style_mapping
| expand_batched_style | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(style, opts, mapping, nvals):\n opts = sorted(opts, key=lambda x: x in ['color', 'alpha'])\n applied_styles = set(mapping)\n style_data, style_mapping = {}, {}\n for opt in opts:\n if 'color' in opt:\n alias = 'color'\n elif 'alpha' in opt:\n alias = 'alpha'\n else:\n alias = None\n if opt not in style or opt in mapping:\n continue\n elif opt == alias:\n if alias in applied_styles:\n continue\n elif 'line_' + alias in applied_styles:\n if 'fill_' + alias not in opts:\n continue\n opt = 'fill_' + alias\n val = style[alias]\n elif 'fill_' + alias in applied_styles:\n opt = 'line_' + alias\n val = style[alias]\n else:\n val = style[alias]\n else:\n val = style[opt]\n style_mapping[opt] = {'field': opt}\n applied_styles.add(opt)\n if 'color' in opt and isinstance(val, tuple):\n val = rgb2hex(val)\n style_data[opt] = [val] * nvals\n return style_data, style_mapping\n",
"role": "user"
},
{
"content": "expand_batched_style",
"role": "assistant"
}
] |
def x(cls, first, second, msg=None):
if not first == second:
standardMsg = '%s != %s' % (safe_repr(first), safe_repr(second))
raise cls.failureException(msg or standardMsg)
| simple_equality | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, first, second, msg=None):\n if not first == second:\n standardMsg = '%s != %s' % (safe_repr(first), safe_repr(second))\n raise cls.failureException(msg or standardMsg)\n",
"role": "user"
},
{
"content": "simple_equality",
"role": "assistant"
}
] |
def x(cls, first, second, msg=None):
asserter = None
if type(first) is type(second):
asserter = cls.equality_type_funcs.get(type(first))
try:
basestring = basestring
except NameError:
basestring = str
if asserter is not None:
if isinstance(asserter, basestring):
asserter = getattr(cls, asserter)
if asserter is None:
asserter = cls.simple_equality
if msg is None:
asserter(first, second)
else:
asserter(first, second, msg=msg)
| assertEqual | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, first, second, msg=None):\n asserter = None\n if type(first) is type(second):\n asserter = cls.equality_type_funcs.get(type(first))\n try:\n basestring = basestring\n except NameError:\n basestring = str\n if asserter is not None:\n if isinstance(asserter, basestring):\n asserter = getattr(cls, asserter)\n if asserter is None:\n asserter = cls.simple_equality\n if msg is None:\n asserter(first, second)\n else:\n asserter(first, second, msg=msg)\n",
"role": "user"
},
{
"content": "assertEqual",
"role": "assistant"
}
] |
def x(self, plot, mapping, properties):
properties = mpl_to_bokeh(properties)
properties = dict(properties, **mapping)
if 'xs' in mapping:
renderer = plot.patches(**properties)
else:
renderer = plot.quad(**properties)
if self.colorbar and 'color_mapper' in self.handles:
self._draw_colorbar(plot, self.handles['color_mapper'])
return renderer, renderer.glyph
| init_glyph | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, plot, mapping, properties):\n properties = mpl_to_bokeh(properties)\n properties = dict(properties, **mapping)\n if 'xs' in mapping:\n renderer = plot.patches(**properties)\n else:\n renderer = plot.quad(**properties)\n if self.colorbar and 'color_mapper' in self.handles:\n self._draw_colorbar(plot, self.handles['color_mapper'])\n return renderer, renderer.glyph\n",
"role": "user"
},
{
"content": "init_glyph",
"role": "assistant"
}
] |
def x(self, x, y):
left, bottom, right, top = self.aarect().lbrt()
return left <= x <= right and bottom <= y <= top
| contains | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, x, y):\n left, bottom, right, top = self.aarect().lbrt()\n return left <= x <= right and bottom <= y <= top\n",
"role": "user"
},
{
"content": "contains",
"role": "assistant"
}
] |
def x(self, x, y):
left, bottom, right, top = self._aarect.lbrt()
return left <= x < right and bottom < y <= top
| contains_exclusive | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, x, y):\n left, bottom, right, top = self._aarect.lbrt()\n return left <= x < right and bottom < y <= top\n",
"role": "user"
},
{
"content": "contains_exclusive",
"role": "assistant"
}
] |
def x(self, x):
left, bottom, right, top = self.aarect().lbrt()
leftx, bottomx, rightx, topx = x.aarect().lbrt()
return (left <= leftx and bottom <= bottomx and right >= rightx and top >=
topx and not (left == leftx and bottom == bottomx and right ==
rightx and top == topx))
| containsbb_exclusive | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, x):\n left, bottom, right, top = self.aarect().lbrt()\n leftx, bottomx, rightx, topx = x.aarect().lbrt()\n return (left <= leftx and bottom <= bottomx and right >= rightx and top >=\n topx and not (left == leftx and bottom == bottomx and right ==\n rightx and top == topx))\n",
"role": "user"
},
{
"content": "containsbb_exclusive",
"role": "assistant"
}
] |
def x(self):
return self._left, self._bottom, self._right, self._top
| lbrt | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return self._left, self._bottom, self._right, self._top\n",
"role": "user"
},
{
"content": "lbrt",
"role": "assistant"
}
] |
def x(self):
left, bottom, right, top = self.lbrt()
return (right + left) / 2.0, (top + bottom) / 2.0
| centroid | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n left, bottom, right, top = self.lbrt()\n return (right + left) / 2.0, (top + bottom) / 2.0\n",
"role": "user"
},
{
"content": "centroid",
"role": "assistant"
}
] |
def x(dmap):
if not isinstance(dmap, DynamicMap):
return []
dmaps = [dmap]
for o in dmap.callback.inputs:
dmaps.extend(get_nested_dmaps(o))
return list(set(dmaps))
| get_nested_dmaps | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dmap):\n if not isinstance(dmap, DynamicMap):\n return []\n dmaps = [dmap]\n for o in dmap.callback.inputs:\n dmaps.extend(get_nested_dmaps(o))\n return list(set(dmaps))\n",
"role": "user"
},
{
"content": "get_nested_dmaps",
"role": "assistant"
}
] |
def x(dmap):
return list({s for dmap in get_nested_dmaps(dmap) for s in dmap.streams})
| get_nested_streams | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dmap):\n return list({s for dmap in get_nested_dmaps(dmap) for s in dmap.streams})\n",
"role": "user"
},
{
"content": "get_nested_streams",
"role": "assistant"
}
] |
def x(callable_obj, streams):
memoization_state = bool(callable_obj._stream_memoization)
callable_obj._stream_memoization &= not any(s.transient and s.
_triggering for s in streams)
try:
yield
except:
raise
finally:
callable_obj._stream_memoization = memoization_state
| dynamicmap_memoization | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(callable_obj, streams):\n memoization_state = bool(callable_obj._stream_memoization)\n callable_obj._stream_memoization &= not any(s.transient and s.\n _triggering for s in streams)\n try:\n yield\n except:\n raise\n finally:\n callable_obj._stream_memoization = memoization_state\n",
"role": "user"
},
{
"content": "dynamicmap_memoization",
"role": "assistant"
}
] |
def x(self, dimensions=None, **kwargs):
dimensions = self._valid_dimensions(dimensions)
if len(dimensions) == self.ndims:
with item_check(False):
return NdOverlay(self, **kwargs).reindex(dimensions)
else:
dims = [d for d in self.kdims if d not in dimensions]
return self.groupby(dims, group_type=NdOverlay, **kwargs)
| overlay | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimensions=None, **kwargs):\n dimensions = self._valid_dimensions(dimensions)\n if len(dimensions) == self.ndims:\n with item_check(False):\n return NdOverlay(self, **kwargs).reindex(dimensions)\n else:\n dims = [d for d in self.kdims if d not in dimensions]\n return self.groupby(dims, group_type=NdOverlay, **kwargs)\n",
"role": "user"
},
{
"content": "overlay",
"role": "assistant"
}
] |
def x(self, dimensions=None, **kwargs):
dimensions = self._valid_dimensions(dimensions)
if len(dimensions) == self.ndims:
with item_check(False):
return GridSpace(self, **kwargs).reindex(dimensions)
return self.groupby(dimensions, container_type=GridSpace, **kwargs)
| grid | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimensions=None, **kwargs):\n dimensions = self._valid_dimensions(dimensions)\n if len(dimensions) == self.ndims:\n with item_check(False):\n return GridSpace(self, **kwargs).reindex(dimensions)\n return self.groupby(dimensions, container_type=GridSpace, **kwargs)\n",
"role": "user"
},
{
"content": "grid",
"role": "assistant"
}
] |
def x(self, dimensions=None, **kwargs):
dimensions = self._valid_dimensions(dimensions)
if len(dimensions) == self.ndims:
with item_check(False):
return NdLayout(self, **kwargs).reindex(dimensions)
return self.groupby(dimensions, container_type=NdLayout, **kwargs)
| layout | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimensions=None, **kwargs):\n dimensions = self._valid_dimensions(dimensions)\n if len(dimensions) == self.ndims:\n with item_check(False):\n return NdLayout(self, **kwargs).reindex(dimensions)\n return self.groupby(dimensions, container_type=NdLayout, **kwargs)\n",
"role": "user"
},
{
"content": "layout",
"role": "assistant"
}
] |
def x(self, *args, **kwargs):
data = OrderedDict([(k, v.options(*args, **kwargs)) for k, v in self.
data.items()])
return self.clone(data)
| options | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *args, **kwargs):\n data = OrderedDict([(k, v.options(*args, **kwargs)) for k, v in self.\n data.items()])\n return self.clone(data)\n",
"role": "user"
},
{
"content": "options",
"role": "assistant"
}
] |
def x(self):
if util.config.future_deprecations:
self.param.warning(
'split_overlays is deprecated and is now a private method.')
return self._split_overlays()
| split_overlays | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if util.config.future_deprecations:\n self.param.warning(\n 'split_overlays is deprecated and is now a private method.')\n return self._split_overlays()\n",
"role": "user"
},
{
"content": "split_overlays",
"role": "assistant"
}
] |
def x(self):
if not issubclass(self.type, CompositeOverlay):
return None, self.clone()
item_maps = OrderedDict()
for k, overlay in self.data.items():
for key, el in overlay.items():
if key not in item_maps:
item_maps[key] = [(k, el)]
else:
item_maps[key].append((k, el))
maps, keys = [], []
for k, layermap in item_maps.items():
maps.append(self.clone(layermap))
keys.append(k)
return keys, maps
| split_overlays | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not issubclass(self.type, CompositeOverlay):\n return None, self.clone()\n item_maps = OrderedDict()\n for k, overlay in self.data.items():\n for key, el in overlay.items():\n if key not in item_maps:\n item_maps[key] = [(k, el)]\n else:\n item_maps[key].append((k, el))\n maps, keys = [], []\n for k, layermap in item_maps.items():\n maps.append(self.clone(layermap))\n keys.append(k)\n return keys, maps\n",
"role": "user"
},
{
"content": "split_overlays",
"role": "assistant"
}
] |
def x(self):
return [tuple(zip([d.name for d in self.kdims], [k] if self.ndims == 1 else
k)) for k in self.keys()]
| dimension_keys | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return [tuple(zip([d.name for d in self.kdims], [k] if self.ndims == 1 else\n k)) for k in self.keys()]\n",
"role": "user"
},
{
"content": "dimension_keys",
"role": "assistant"
}
] |
def x(self, dimensions, other, keys):
if not isinstance(self, DynamicMap) or not isinstance(other, DynamicMap):
keys = sorted((d, v) for k in keys for d, v in k)
grouped = dict([(g, [v for _, v in group]) for g, group in groupby(
keys, lambda x: x[0])])
dimensions = [d(values=grouped[d.name]) for d in dimensions]
map_obj = None
map_obj = self if isinstance(self, DynamicMap) else other
if isinstance(self, DynamicMap) and isinstance(other, DynamicMap):
self_streams = util.dimensioned_streams(self)
other_streams = util.dimensioned_streams(other)
streams = list(util.unique_iterator(self_streams + other_streams))
else:
streams = map_obj.streams
def x(*key, **kwargs):
key_map = {d.name: k for d, k in zip(dimensions, key)}
layers = []
try:
self_el = self.select(HoloMap, **key_map) if self.kdims else self[(
)]
layers.append(self_el)
except KeyError:
pass
try:
other_el = other.select(HoloMap, **key_map
) if other.kdims else other[()]
layers.append(other_el)
except KeyError:
pass
return Overlay(layers)
callback = Callable(dynamic_mul, inputs=[self, other])
callback._is_overlay = True
if map_obj:
return map_obj.clone(callback=callback, shared_data=False, kdims=
dimensions, streams=streams)
else:
return DynamicMap(callback=callback, kdims=dimensions, streams=streams)
| dynamic_mul | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimensions, other, keys):\n if not isinstance(self, DynamicMap) or not isinstance(other, DynamicMap):\n keys = sorted((d, v) for k in keys for d, v in k)\n grouped = dict([(g, [v for _, v in group]) for g, group in groupby(\n keys, lambda x: x[0])])\n dimensions = [d(values=grouped[d.name]) for d in dimensions]\n map_obj = None\n map_obj = self if isinstance(self, DynamicMap) else other\n if isinstance(self, DynamicMap) and isinstance(other, DynamicMap):\n self_streams = util.dimensioned_streams(self)\n other_streams = util.dimensioned_streams(other)\n streams = list(util.unique_iterator(self_streams + other_streams))\n else:\n streams = map_obj.streams\n\n def x(*key, **kwargs):\n key_map = {d.name: k for d, k in zip(dimensions, key)}\n layers = []\n try:\n self_el = self.select(HoloMap, **key_map) if self.kdims else self[(\n )]\n layers.append(self_el)\n except KeyError:\n pass\n try:\n other_el = other.select(HoloMap, **key_map\n ) if other.kdims else other[()]\n layers.append(other_el)\n except KeyError:\n pass\n return Overlay(layers)\n callback = Callable(dynamic_mul, inputs=[self, other])\n callback._is_overlay = True\n if map_obj:\n return map_obj.clone(callback=callback, shared_data=False, kdims=\n dimensions, streams=streams)\n else:\n return DynamicMap(callback=callback, kdims=dimensions, streams=streams)\n",
"role": "user"
},
{
"content": "dynamic_mul",
"role": "assistant"
}
] |
def x(self, merge_type=None, drop=[], drop_constant=False):
from .element import Collator
merge_type = merge_type if merge_type else self.__class__
return Collator(self, merge_type=merge_type, drop=drop, drop_constant=
drop_constant)()
| collate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, merge_type=None, drop=[], drop_constant=False):\n from .element import Collator\n merge_type = merge_type if merge_type else self.__class__\n return Collator(self, merge_type=merge_type, drop=drop, drop_constant=\n drop_constant)()\n",
"role": "user"
},
{
"content": "collate",
"role": "assistant"
}
] |
def x(self, dimensions=None, function=None, spreadfn=None, **kwargs):
from .data import concat
if not dimensions:
dimensions = self.kdims
if not isinstance(dimensions, list):
dimensions = [dimensions]
if self.ndims > 1 and len(dimensions) != self.ndims:
groups = self.groupby([dim for dim in self.kdims if dim not in
dimensions])
elif all(d in self.kdims for d in dimensions):
groups = HoloMap([(0, self)])
else:
raise KeyError('Supplied dimensions not found.')
collapsed = groups.clone(shared_data=False)
for key, group in groups.items():
if hasattr(group.last, 'interface'):
group_data = concat(group)
if function:
agg = group_data.aggregate(group.last.kdims, function,
spreadfn, **kwargs)
group_data = group.type(agg)
else:
group_data = [el.data for el in group]
args = group_data, function, group.last.kdims
data = group.type.collapse_data(*args, **kwargs)
group_data = group.last.clone(data)
collapsed[key] = group_data
return collapsed if self.ndims - len(dimensions) else collapsed.last
| collapse | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimensions=None, function=None, spreadfn=None, **kwargs):\n from .data import concat\n if not dimensions:\n dimensions = self.kdims\n if not isinstance(dimensions, list):\n dimensions = [dimensions]\n if self.ndims > 1 and len(dimensions) != self.ndims:\n groups = self.groupby([dim for dim in self.kdims if dim not in\n dimensions])\n elif all(d in self.kdims for d in dimensions):\n groups = HoloMap([(0, self)])\n else:\n raise KeyError('Supplied dimensions not found.')\n collapsed = groups.clone(shared_data=False)\n for key, group in groups.items():\n if hasattr(group.last, 'interface'):\n group_data = concat(group)\n if function:\n agg = group_data.aggregate(group.last.kdims, function,\n spreadfn, **kwargs)\n group_data = group.type(agg)\n else:\n group_data = [el.data for el in group]\n args = group_data, function, group.last.kdims\n data = group.type.collapse_data(*args, **kwargs)\n group_data = group.last.clone(data)\n collapsed[key] = group_data\n return collapsed if self.ndims - len(dimensions) else collapsed.last\n",
"role": "user"
},
{
"content": "collapse",
"role": "assistant"
}
] |
def x(self, samples=[], bounds=None, **sample_values):
if util.config.future_deprecations:
self.param.warning(
'The HoloMap.sample method is deprecated, for equivalent functionality use HoloMap.apply.sample().collapse().'
)
dims = self.last.ndims
if isinstance(samples, tuple) or np.isscalar(samples):
if dims == 1:
xlim = self.last.range(0)
lower, upper = (xlim[0], xlim[1]) if bounds is None else bounds
edges = np.linspace(lower, upper, samples + 1)
linsamples = [((l + u) / 2.0) for l, u in zip(edges[:-1], edges
[1:])]
elif dims == 2:
rows, cols = samples
if bounds:
l, b, r, t = bounds
else:
l, r = self.last.range(0)
b, t = self.last.range(1)
xedges = np.linspace(l, r, cols + 1)
yedges = np.linspace(b, t, rows + 1)
xsamples = [((lx + ux) / 2.0) for lx, ux in zip(xedges[:-1],
xedges[1:])]
ysamples = [((ly + uy) / 2.0) for ly, uy in zip(yedges[:-1],
yedges[1:])]
Y, X = np.meshgrid(ysamples, xsamples)
linsamples = list(zip(X.flat, Y.flat))
else:
raise NotImplementedError(
'Regular sampling not implemented for elements with more than two dimensions.'
)
samples = list(util.unique_iterator(self.last.closest(linsamples)))
sampled = self.clone([(k, view.sample(samples, closest=False, **
sample_values)) for k, view in self.data.items()])
from ..element import Table
return Table(sampled.collapse())
| sample | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, samples=[], bounds=None, **sample_values):\n if util.config.future_deprecations:\n self.param.warning(\n 'The HoloMap.sample method is deprecated, for equivalent functionality use HoloMap.apply.sample().collapse().'\n )\n dims = self.last.ndims\n if isinstance(samples, tuple) or np.isscalar(samples):\n if dims == 1:\n xlim = self.last.range(0)\n lower, upper = (xlim[0], xlim[1]) if bounds is None else bounds\n edges = np.linspace(lower, upper, samples + 1)\n linsamples = [((l + u) / 2.0) for l, u in zip(edges[:-1], edges\n [1:])]\n elif dims == 2:\n rows, cols = samples\n if bounds:\n l, b, r, t = bounds\n else:\n l, r = self.last.range(0)\n b, t = self.last.range(1)\n xedges = np.linspace(l, r, cols + 1)\n yedges = np.linspace(b, t, rows + 1)\n xsamples = [((lx + ux) / 2.0) for lx, ux in zip(xedges[:-1],\n xedges[1:])]\n ysamples = [((ly + uy) / 2.0) for ly, uy in zip(yedges[:-1],\n yedges[1:])]\n Y, X = np.meshgrid(ysamples, xsamples)\n linsamples = list(zip(X.flat, Y.flat))\n else:\n raise NotImplementedError(\n 'Regular sampling not implemented for elements with more than two dimensions.'\n )\n samples = list(util.unique_iterator(self.last.closest(linsamples)))\n sampled = self.clone([(k, view.sample(samples, closest=False, **\n sample_values)) for k, view in self.data.items()])\n from ..element import Table\n return Table(sampled.collapse())\n",
"role": "user"
},
{
"content": "sample",
"role": "assistant"
}
] |
def x(self, dimensions=None, function=None, spread_fn=None, **reduce_map):
if util.config.future_deprecations:
self.param.warning(
'The HoloMap.reduce method is deprecated, for equivalent functionality use HoloMap.apply.reduce().collapse().'
)
from ..element import Table
reduced_items = [(k, v.reduce(dimensions, function, spread_fn, **
reduce_map)) for k, v in self.items()]
if not isinstance(reduced_items[0][1], Table):
params = dict(util.get_param_values(self.last), kdims=self.kdims,
vdims=self.last.vdims)
return Table(reduced_items, **params)
return Table(self.clone(reduced_items).collapse())
| reduce | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimensions=None, function=None, spread_fn=None, **reduce_map):\n if util.config.future_deprecations:\n self.param.warning(\n 'The HoloMap.reduce method is deprecated, for equivalent functionality use HoloMap.apply.reduce().collapse().'\n )\n from ..element import Table\n reduced_items = [(k, v.reduce(dimensions, function, spread_fn, **\n reduce_map)) for k, v in self.items()]\n if not isinstance(reduced_items[0][1], Table):\n params = dict(util.get_param_values(self.last), kdims=self.kdims,\n vdims=self.last.vdims)\n return Table(reduced_items, **params)\n return Table(self.clone(reduced_items).collapse())\n",
"role": "user"
},
{
"content": "reduce",
"role": "assistant"
}
] |
def x(self, label=None, group=None, depth=1):
return super(HoloMap, self).relabel(label=label, group=group, depth=depth)
| relabel | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, label=None, group=None, depth=1):\n return super(HoloMap, self).relabel(label=label, group=group, depth=depth)\n",
"role": "user"
},
{
"content": "relabel",
"role": "assistant"
}
] |
def x(self, dimension=None, num_bins=20, bin_range=None, adjoin=True,
individually=True, **kwargs):
if dimension is not None and not isinstance(dimension, list):
dimension = [dimension]
histmaps = [self.clone(shared_data=False) for _ in dimension or [None]]
if individually:
map_range = None
else:
if dimension is None:
raise Exception(
'Please supply the dimension to compute a histogram for.')
map_range = self.range(kwargs['dimension'])
bin_range = map_range if bin_range is None else bin_range
style_prefix = 'Custom[<' + self.name + '>]_'
if issubclass(self.type, (NdOverlay, Overlay)) and 'index' not in kwargs:
kwargs['index'] = 0
for k, v in self.data.items():
hists = v.hist(adjoin=False, dimension=dimension, bin_range=
bin_range, num_bins=num_bins, style_prefix=style_prefix, **kwargs)
if isinstance(hists, Layout):
for i, hist in enumerate(hists):
histmaps[i][k] = hist
else:
histmaps[0][k] = hists
if adjoin:
layout = self
for hist in histmaps:
layout = layout << hist
if issubclass(self.type, (NdOverlay, Overlay)):
layout.main_layer = kwargs['index']
return layout
elif len(histmaps) > 1:
return Layout(histmaps)
else:
return histmaps[0]
| hist | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimension=None, num_bins=20, bin_range=None, adjoin=True,\n individually=True, **kwargs):\n if dimension is not None and not isinstance(dimension, list):\n dimension = [dimension]\n histmaps = [self.clone(shared_data=False) for _ in dimension or [None]]\n if individually:\n map_range = None\n else:\n if dimension is None:\n raise Exception(\n 'Please supply the dimension to compute a histogram for.')\n map_range = self.range(kwargs['dimension'])\n bin_range = map_range if bin_range is None else bin_range\n style_prefix = 'Custom[<' + self.name + '>]_'\n if issubclass(self.type, (NdOverlay, Overlay)) and 'index' not in kwargs:\n kwargs['index'] = 0\n for k, v in self.data.items():\n hists = v.hist(adjoin=False, dimension=dimension, bin_range=\n bin_range, num_bins=num_bins, style_prefix=style_prefix, **kwargs)\n if isinstance(hists, Layout):\n for i, hist in enumerate(hists):\n histmaps[i][k] = hist\n else:\n histmaps[0][k] = hists\n if adjoin:\n layout = self\n for hist in histmaps:\n layout = layout << hist\n if issubclass(self.type, (NdOverlay, Overlay)):\n layout.main_layer = kwargs['index']\n return layout\n elif len(histmaps) > 1:\n return Layout(histmaps)\n else:\n return histmaps[0]\n",
"role": "user"
},
{
"content": "hist",
"role": "assistant"
}
] |
def x(self):
noargs = inspect.ArgSpec(args=[], varargs=None, keywords=None, defaults
=None)
return self.argspec == noargs
| noargs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n noargs = inspect.ArgSpec(args=[], varargs=None, keywords=None, defaults\n =None)\n return self.argspec == noargs\n",
"role": "user"
},
{
"content": "noargs",
"role": "assistant"
}
] |
def x(self, callable=None, **overrides):
old = {k: v for k, v in self.get_param_values() if k not in ['callable',
'name']}
params = dict(old, **overrides)
callable = self.callable if callable is None else callable
return self.__class__(callable, **params)
| clone | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, callable=None, **overrides):\n old = {k: v for k, v in self.get_param_values() if k not in ['callable',\n 'name']}\n params = dict(old, **overrides)\n callable = self.callable if callable is None else callable\n return self.__class__(callable, **params)\n",
"role": "user"
},
{
"content": "clone",
"role": "assistant"
}
] |
def x(self):
unbounded_dims = []
stream_params = set(util.stream_parameters(self.streams))
for kdim in self.kdims:
if str(kdim) in stream_params:
continue
if kdim.values:
continue
if None in kdim.range:
unbounded_dims.append(str(kdim))
return unbounded_dims
| unbounded | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n unbounded_dims = []\n stream_params = set(util.stream_parameters(self.streams))\n for kdim in self.kdims:\n if str(kdim) in stream_params:\n continue\n if kdim.values:\n continue\n if None in kdim.range:\n unbounded_dims.append(str(kdim))\n return unbounded_dims\n",
"role": "user"
},
{
"content": "unbounded",
"role": "assistant"
}
] |
def x(self):
key = []
undefined = []
stream_params = set(util.stream_parameters(self.streams))
for kdim in self.kdims:
if str(kdim) in stream_params:
key.append(None)
elif kdim.default:
key.append(kdim.default)
elif kdim.values:
if all(util.isnumeric(v) for v in kdim.values):
key.append(sorted(kdim.values)[0])
else:
key.append(kdim.values[0])
elif kdim.range[0] is not None:
key.append(kdim.range[0])
else:
undefined.append(kdim)
if undefined:
msg = (
'Dimension(s) {undefined_dims} do not specify range or values needed to generate initial key'
)
undefined_dims = ', '.join([('%r' % str(dim)) for dim in undefined])
raise KeyError(msg.format(undefined_dims=undefined_dims))
return tuple(key)
| initial_key | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n key = []\n undefined = []\n stream_params = set(util.stream_parameters(self.streams))\n for kdim in self.kdims:\n if str(kdim) in stream_params:\n key.append(None)\n elif kdim.default:\n key.append(kdim.default)\n elif kdim.values:\n if all(util.isnumeric(v) for v in kdim.values):\n key.append(sorted(kdim.values)[0])\n else:\n key.append(kdim.values[0])\n elif kdim.range[0] is not None:\n key.append(kdim.range[0])\n else:\n undefined.append(kdim)\n if undefined:\n msg = (\n 'Dimension(s) {undefined_dims} do not specify range or values needed to generate initial key'\n )\n undefined_dims = ', '.join([('%r' % str(dim)) for dim in undefined])\n raise KeyError(msg.format(undefined_dims=undefined_dims))\n return tuple(key)\n",
"role": "user"
},
{
"content": "initial_key",
"role": "assistant"
}
] |
def x(self, key):
if key == () and len(self.kdims) == 0:
return ()
key = util.wrap_tuple(key)
assert len(key) == len(self.kdims)
for ind, val in enumerate(key):
kdim = self.kdims[ind]
low, high = util.max_range([kdim.range, kdim.soft_range])
if util.is_number(low) and util.isfinite(low):
if val < low:
raise KeyError('Key value %s below lower bound %s' % (val, low)
)
if util.is_number(high) and util.isfinite(high):
if val > high:
raise KeyError('Key value %s above upper bound %s' % (val,
high))
| validate_key | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key):\n if key == () and len(self.kdims) == 0:\n return ()\n key = util.wrap_tuple(key)\n assert len(key) == len(self.kdims)\n for ind, val in enumerate(key):\n kdim = self.kdims[ind]\n low, high = util.max_range([kdim.range, kdim.soft_range])\n if util.is_number(low) and util.isfinite(low):\n if val < low:\n raise KeyError('Key value %s below lower bound %s' % (val, low)\n )\n if util.is_number(high) and util.isfinite(high):\n if val > high:\n raise KeyError('Key value %s above upper bound %s' % (val,\n high))\n",
"role": "user"
},
{
"content": "validate_key",
"role": "assistant"
}
] |
def x(self, **kwargs):
if self.callback.noargs and self.streams == []:
self.param.warning(
'No streams declared. To update a DynamicMaps using generators (or callables without arguments) use streams=[Next()]'
)
return
if self.streams == []:
self.param.warning(
'No streams on DynamicMap, calling event will have no effect')
return
stream_params = set(util.stream_parameters(self.streams))
invalid = [k for k in kwargs.keys() if k not in stream_params]
if invalid:
msg = 'Key(s) {invalid} do not correspond to stream parameters'
raise KeyError(msg.format(invalid=', '.join('%r' % i for i in invalid))
)
streams = []
for stream in self.streams:
contents = stream.contents
applicable_kws = {k: v for k, v in kwargs.items() if k in set(
contents.keys())}
if not applicable_kws and contents:
continue
streams.append(stream)
rkwargs = util.rename_stream_kwargs(stream, applicable_kws, reverse
=True)
stream.update(**rkwargs)
Stream.trigger(streams)
| event | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n if self.callback.noargs and self.streams == []:\n self.param.warning(\n 'No streams declared. To update a DynamicMaps using generators (or callables without arguments) use streams=[Next()]'\n )\n return\n if self.streams == []:\n self.param.warning(\n 'No streams on DynamicMap, calling event will have no effect')\n return\n stream_params = set(util.stream_parameters(self.streams))\n invalid = [k for k in kwargs.keys() if k not in stream_params]\n if invalid:\n msg = 'Key(s) {invalid} do not correspond to stream parameters'\n raise KeyError(msg.format(invalid=', '.join('%r' % i for i in invalid))\n )\n streams = []\n for stream in self.streams:\n contents = stream.contents\n applicable_kws = {k: v for k, v in kwargs.items() if k in set(\n contents.keys())}\n if not applicable_kws and contents:\n continue\n streams.append(stream)\n rkwargs = util.rename_stream_kwargs(stream, applicable_kws, reverse\n =True)\n stream.update(**rkwargs)\n Stream.trigger(streams)\n",
"role": "user"
},
{
"content": "event",
"role": "assistant"
}
] |
def x(self, retval):
if self.id not in Store.custom_options():
return retval
spec = StoreOptions.tree_to_dict(Store.custom_options()[self.id])
return retval.opts(spec)
| style | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, retval):\n if self.id not in Store.custom_options():\n return retval\n spec = StoreOptions.tree_to_dict(Store.custom_options()[self.id])\n return retval.opts(spec)\n",
"role": "user"
},
{
"content": "style",
"role": "assistant"
}
] |
def x(self, *args):
self._validate_key(args)
kdims = [kdim.name for kdim in self.kdims]
kwarg_items = [s.contents.items() for s in self.streams]
hash_items = tuple(tuple(sorted(s.hashkey.items())) for s in self.streams
) + args
flattened = [(k, v) for kws in kwarg_items for k, v in kws if k not in
kdims]
if self._posarg_keys:
kwargs = dict(flattened, **dict(zip(self._posarg_keys, args)))
args = ()
else:
kwargs = dict(flattened)
if not isinstance(self.callback, Generator):
kwargs['_memoization_hash_'] = hash_items
with dynamicmap_memoization(self.callback, self.streams):
retval = self.callback(*args, **kwargs)
return self._style(retval)
| execute_callback | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *args):\n self._validate_key(args)\n kdims = [kdim.name for kdim in self.kdims]\n kwarg_items = [s.contents.items() for s in self.streams]\n hash_items = tuple(tuple(sorted(s.hashkey.items())) for s in self.streams\n ) + args\n flattened = [(k, v) for kws in kwarg_items for k, v in kws if k not in\n kdims]\n if self._posarg_keys:\n kwargs = dict(flattened, **dict(zip(self._posarg_keys, args)))\n args = ()\n else:\n kwargs = dict(flattened)\n if not isinstance(self.callback, Generator):\n kwargs['_memoization_hash_'] = hash_items\n with dynamicmap_memoization(self.callback, self.streams):\n retval = self.callback(*args, **kwargs)\n return self._style(retval)\n",
"role": "user"
},
{
"content": "execute_callback",
"role": "assistant"
}
] |
def x(self, *args, **kwargs):
if 'clone' not in kwargs:
kwargs['clone'] = True
return self.opts(*args, **kwargs)
| options | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *args, **kwargs):\n if 'clone' not in kwargs:\n kwargs['clone'] = True\n return self.opts(*args, **kwargs)\n",
"role": "user"
},
{
"content": "options",
"role": "assistant"
}
] |
def x(self, data=None, shared_data=True, new_type=None, link=True, *args,
**overrides):
if 'link_inputs' in overrides and util.config.future_deprecations:
self.param.warning(
'link_inputs argument to the clone method is deprecated, use the more general link argument instead.'
)
link = link and overrides.pop('link_inputs', True)
callback = overrides.pop('callback', self.callback)
if data is None and shared_data:
data = self.data
if link and callback is self.callback:
overrides['plot_id'] = self._plot_id
clone = super(UniformNdMapping, self).clone(callback, shared_data,
new_type, link, *((data,) + args), **overrides)
if clone.callback is self.callback:
with util.disable_constant(clone):
clone.callback = clone.callback.clone(inputs=[self],
link_inputs=link)
return clone
| clone | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data=None, shared_data=True, new_type=None, link=True, *args,\n **overrides):\n if 'link_inputs' in overrides and util.config.future_deprecations:\n self.param.warning(\n 'link_inputs argument to the clone method is deprecated, use the more general link argument instead.'\n )\n link = link and overrides.pop('link_inputs', True)\n callback = overrides.pop('callback', self.callback)\n if data is None and shared_data:\n data = self.data\n if link and callback is self.callback:\n overrides['plot_id'] = self._plot_id\n clone = super(UniformNdMapping, self).clone(callback, shared_data,\n new_type, link, *((data,) + args), **overrides)\n if clone.callback is self.callback:\n with util.disable_constant(clone):\n clone.callback = clone.callback.clone(inputs=[self],\n link_inputs=link)\n return clone\n",
"role": "user"
},
{
"content": "clone",
"role": "assistant"
}
] |
def x(self, tuple_key, cache, data_slice):
if not any(isinstance(el, (list, set)) for el in tuple_key):
return None
if len(tuple_key) == 1:
product = tuple_key[0]
else:
args = [(set(el) if isinstance(el, (list, set)) else set([el])) for
el in tuple_key]
product = itertools.product(*args)
data = []
for inner_key in product:
key = util.wrap_tuple(inner_key)
if key in cache:
val = cache[key]
else:
val = self._execute_callback(*key)
if data_slice:
val = self._dataslice(val, data_slice)
data.append((key, val))
product = self.clone(data)
if data_slice:
from ..util import Dynamic
dmap = Dynamic(self, operation=lambda obj, **dynkwargs: obj[
data_slice], streams=self.streams)
dmap.data = product.data
return dmap
return product
| cross_product | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tuple_key, cache, data_slice):\n if not any(isinstance(el, (list, set)) for el in tuple_key):\n return None\n if len(tuple_key) == 1:\n product = tuple_key[0]\n else:\n args = [(set(el) if isinstance(el, (list, set)) else set([el])) for\n el in tuple_key]\n product = itertools.product(*args)\n data = []\n for inner_key in product:\n key = util.wrap_tuple(inner_key)\n if key in cache:\n val = cache[key]\n else:\n val = self._execute_callback(*key)\n if data_slice:\n val = self._dataslice(val, data_slice)\n data.append((key, val))\n product = self.clone(data)\n if data_slice:\n from ..util import Dynamic\n dmap = Dynamic(self, operation=lambda obj, **dynkwargs: obj[\n data_slice], streams=self.streams)\n dmap.data = product.data\n return dmap\n return product\n",
"role": "user"
},
{
"content": "cross_product",
"role": "assistant"
}
] |
def x(self, tuple_key, data_slice):
slices = [el for el in tuple_key if isinstance(el, slice)]
if any(el.step for el in slices):
raise Exception('DynamicMap slices cannot have a step argument')
elif len(slices) not in [0, len(tuple_key)]:
raise Exception('Slices must be used exclusively or not at all')
elif not slices:
return None
sliced = self.clone(self)
for i, slc in enumerate(tuple_key):
start, stop = slc.start, slc.stop
if start is not None and start < sliced.kdims[i].range[0]:
raise Exception('Requested slice below defined dimension range.')
if stop is not None and stop > sliced.kdims[i].range[1]:
raise Exception('Requested slice above defined dimension range.')
sliced.kdims[i].soft_range = start, stop
if data_slice:
if not isinstance(sliced, DynamicMap):
return self._dataslice(sliced, data_slice)
else:
from ..util import Dynamic
if len(self):
slices = [slice(None) for _ in range(self.ndims)] + list(
data_slice)
sliced = super(DynamicMap, sliced).__getitem__(tuple(slices))
dmap = Dynamic(self, operation=lambda obj, **dynkwargs: obj[
data_slice], streams=self.streams)
dmap.data = sliced.data
return dmap
return sliced
| slice_bounded | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tuple_key, data_slice):\n slices = [el for el in tuple_key if isinstance(el, slice)]\n if any(el.step for el in slices):\n raise Exception('DynamicMap slices cannot have a step argument')\n elif len(slices) not in [0, len(tuple_key)]:\n raise Exception('Slices must be used exclusively or not at all')\n elif not slices:\n return None\n sliced = self.clone(self)\n for i, slc in enumerate(tuple_key):\n start, stop = slc.start, slc.stop\n if start is not None and start < sliced.kdims[i].range[0]:\n raise Exception('Requested slice below defined dimension range.')\n if stop is not None and stop > sliced.kdims[i].range[1]:\n raise Exception('Requested slice above defined dimension range.')\n sliced.kdims[i].soft_range = start, stop\n if data_slice:\n if not isinstance(sliced, DynamicMap):\n return self._dataslice(sliced, data_slice)\n else:\n from ..util import Dynamic\n if len(self):\n slices = [slice(None) for _ in range(self.ndims)] + list(\n data_slice)\n sliced = super(DynamicMap, sliced).__getitem__(tuple(slices))\n dmap = Dynamic(self, operation=lambda obj, **dynkwargs: obj[\n data_slice], streams=self.streams)\n dmap.data = sliced.data\n return dmap\n return sliced\n",
"role": "user"
},
{
"content": "slice_bounded",
"role": "assistant"
}
] |
def x(self, selection_specs=None, **kwargs):
if selection_specs is not None and not isinstance(selection_specs, (
list, tuple)):
selection_specs = [selection_specs]
selection = super(DynamicMap, self).select(selection_specs, **kwargs)
def x(obj, **dynkwargs):
if selection_specs is not None:
matches = any(obj.matches(spec) for spec in selection_specs)
else:
matches = True
if matches:
return obj.select(**kwargs)
return obj
if not isinstance(selection, DynamicMap):
return dynamic_select(selection)
else:
from ..util import Dynamic
dmap = Dynamic(self, operation=dynamic_select, streams=self.streams)
dmap.data = selection.data
return dmap
| select | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, selection_specs=None, **kwargs):\n if selection_specs is not None and not isinstance(selection_specs, (\n list, tuple)):\n selection_specs = [selection_specs]\n selection = super(DynamicMap, self).select(selection_specs, **kwargs)\n\n def x(obj, **dynkwargs):\n if selection_specs is not None:\n matches = any(obj.matches(spec) for spec in selection_specs)\n else:\n matches = True\n if matches:\n return obj.select(**kwargs)\n return obj\n if not isinstance(selection, DynamicMap):\n return dynamic_select(selection)\n else:\n from ..util import Dynamic\n dmap = Dynamic(self, operation=dynamic_select, streams=self.streams)\n dmap.data = selection.data\n return dmap\n",
"role": "user"
},
{
"content": "select",
"role": "assistant"
}
] |
def x(self, key, val):
cache_size = 1 if util.dimensionless_contents(self.streams, self.kdims
) else self.cache_size
if len(self) >= cache_size:
first_key = next(k for k in self.data)
self.data.pop(first_key)
self[key] = val
| cache | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, val):\n cache_size = 1 if util.dimensionless_contents(self.streams, self.kdims\n ) else self.cache_size\n if len(self) >= cache_size:\n first_key = next(k for k in self.data)\n self.data.pop(first_key)\n self[key] = val\n",
"role": "user"
},
{
"content": "cache",
"role": "assistant"
}
] |
def x(self, map_fn, specs=None, clone=True, link_inputs=True):
deep_mapped = super(DynamicMap, self).map(map_fn, specs, clone)
if isinstance(deep_mapped, type(self)):
from ..util import Dynamic
def x(obj, **dynkwargs):
return obj.map(map_fn, specs, clone)
dmap = Dynamic(self, operation=apply_map, streams=self.streams,
link_inputs=link_inputs)
dmap.data = deep_mapped.data
return dmap
return deep_mapped
| map | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, map_fn, specs=None, clone=True, link_inputs=True):\n deep_mapped = super(DynamicMap, self).map(map_fn, specs, clone)\n if isinstance(deep_mapped, type(self)):\n from ..util import Dynamic\n\n def x(obj, **dynkwargs):\n return obj.map(map_fn, specs, clone)\n dmap = Dynamic(self, operation=apply_map, streams=self.streams,\n link_inputs=link_inputs)\n dmap.data = deep_mapped.data\n return dmap\n return deep_mapped\n",
"role": "user"
},
{
"content": "map",
"role": "assistant"
}
] |
def x(self, label=None, group=None, depth=1):
relabelled = super(DynamicMap, self).relabel(label, group, depth)
if depth > 0:
from ..util import Dynamic
def x(obj, **dynkwargs):
return obj.relabel(group=group, label=label, depth=depth - 1)
dmap = Dynamic(self, streams=self.streams, operation=dynamic_relabel)
dmap.data = relabelled.data
with util.disable_constant(dmap):
dmap.group = relabelled.group
dmap.label = relabelled.label
return dmap
return relabelled
| relabel | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, label=None, group=None, depth=1):\n relabelled = super(DynamicMap, self).relabel(label, group, depth)\n if depth > 0:\n from ..util import Dynamic\n\n def x(obj, **dynkwargs):\n return obj.relabel(group=group, label=label, depth=depth - 1)\n dmap = Dynamic(self, streams=self.streams, operation=dynamic_relabel)\n dmap.data = relabelled.data\n with util.disable_constant(dmap):\n dmap.group = relabelled.group\n dmap.label = relabelled.label\n return dmap\n return relabelled\n",
"role": "user"
},
{
"content": "relabel",
"role": "assistant"
}
] |
def x(self):
if not len(self):
raise ValueError(
'Cannot split DynamicMap before it has been initialized')
elif not issubclass(self.type, CompositeOverlay):
return None, self
from ..util import Dynamic
keys = list(self.last.data.keys())
dmaps = []
for key in keys:
el = self.last.data[key]
def x(obj, overlay_key=key, overlay_el=el, **kwargs):
spec = util.get_overlay_spec(obj, overlay_key, overlay_el)
items = list(obj.data.items())
specs = [(i, util.get_overlay_spec(obj, k, v)) for i, (k, v) in
enumerate(items)]
match = util.closest_match(spec, specs)
if match is None:
raise KeyError(
'{spec} spec not found in {otype}. The split_overlays method only works consistently for a DynamicMap where the layers of the {otype} do not change.'
.format(spec=spec, otype=type(obj).__name__))
return items[match][1]
dmap = Dynamic(self, streams=self.streams, operation=
split_overlay_callback)
dmap.data = OrderedDict([(list(self.data.keys())[-1], self.last.
data[key])])
dmaps.append(dmap)
return keys, dmaps
| split_overlays | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not len(self):\n raise ValueError(\n 'Cannot split DynamicMap before it has been initialized')\n elif not issubclass(self.type, CompositeOverlay):\n return None, self\n from ..util import Dynamic\n keys = list(self.last.data.keys())\n dmaps = []\n for key in keys:\n el = self.last.data[key]\n\n def x(obj, overlay_key=key, overlay_el=el, **kwargs):\n spec = util.get_overlay_spec(obj, overlay_key, overlay_el)\n items = list(obj.data.items())\n specs = [(i, util.get_overlay_spec(obj, k, v)) for i, (k, v) in\n enumerate(items)]\n match = util.closest_match(spec, specs)\n if match is None:\n raise KeyError(\n '{spec} spec not found in {otype}. The split_overlays method only works consistently for a DynamicMap where the layers of the {otype} do not change.'\n .format(spec=spec, otype=type(obj).__name__))\n return items[match][1]\n dmap = Dynamic(self, streams=self.streams, operation=\n split_overlay_callback)\n dmap.data = OrderedDict([(list(self.data.keys())[-1], self.last.\n data[key])])\n dmaps.append(dmap)\n return keys, dmaps\n",
"role": "user"
},
{
"content": "split_overlays",
"role": "assistant"
}
] |
def x(self):
if self.last is not None:
initialized = self
else:
initialized = self.clone()
initialized[initialized._initial_key()]
if not isinstance(initialized.last, (Layout, NdLayout, GridSpace)):
return self
container = initialized.last.clone(shared_data=False)
type_counter = defaultdict(int)
remapped_streams = []
self_dstreams = util.dimensioned_streams(self)
streams = self.callback.stream_mapping
for i, (k, v) in enumerate(initialized.last.data.items()):
vstreams = streams.get(i, [])
if not vstreams:
if isinstance(initialized.last, Layout):
for l in range(len(k)):
path = '.'.join(k[:l])
if path in streams:
vstreams = streams[path]
break
else:
vstreams = streams.get(k, [])
if any(s in remapped_streams for s in vstreams):
raise ValueError(
'The stream_mapping supplied on the Callable is ambiguous please supply more specific Layout path specs.'
)
remapped_streams += vstreams
def x(*args, **kwargs):
layout = self[args]
layout_type = type(layout).__name__
if len(container.keys()) != len(layout.keys()):
raise ValueError(
'Collated DynamicMaps must return %s with consistent number of items.'
% layout_type)
key = kwargs['selection_key']
index = kwargs['selection_index']
obj_type = kwargs['selection_type']
dyn_type_map = defaultdict(list)
for k, v in layout.data.items():
if k == key:
return layout[k]
dyn_type_map[type(v)].append(v)
dyn_type_counter = {t: len(vals) for t, vals in dyn_type_map.
items()}
if dyn_type_counter != type_counter:
raise ValueError(
'The objects in a %s returned by a DynamicMap must consistently return the same number of items of the same type.'
% layout_type)
return dyn_type_map[obj_type][index]
callback = Callable(partial(collation_cb, selection_key=k,
selection_index=type_counter[type(v)], selection_type=type(v)),
inputs=[self])
vstreams = list(util.unique_iterator(self_dstreams + vstreams))
vdmap = self.clone(callback=callback, shared_data=False, streams=
vstreams)
type_counter[type(v)] += 1
for stream in vstreams:
if stream.source is self:
stream.source = vdmap
container[k] = vdmap
unmapped_streams = [repr(stream) for stream in self.streams if stream.
source is self and stream not in remapped_streams and stream.linked]
if unmapped_streams:
raise ValueError(
"""The following streams are set to be automatically linked to a plot, but no stream_mapping specifying which item in the (Nd)Layout to link it to was found:
%s"""
% ', '.join(unmapped_streams))
return container
| collate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.last is not None:\n initialized = self\n else:\n initialized = self.clone()\n initialized[initialized._initial_key()]\n if not isinstance(initialized.last, (Layout, NdLayout, GridSpace)):\n return self\n container = initialized.last.clone(shared_data=False)\n type_counter = defaultdict(int)\n remapped_streams = []\n self_dstreams = util.dimensioned_streams(self)\n streams = self.callback.stream_mapping\n for i, (k, v) in enumerate(initialized.last.data.items()):\n vstreams = streams.get(i, [])\n if not vstreams:\n if isinstance(initialized.last, Layout):\n for l in range(len(k)):\n path = '.'.join(k[:l])\n if path in streams:\n vstreams = streams[path]\n break\n else:\n vstreams = streams.get(k, [])\n if any(s in remapped_streams for s in vstreams):\n raise ValueError(\n 'The stream_mapping supplied on the Callable is ambiguous please supply more specific Layout path specs.'\n )\n remapped_streams += vstreams\n\n def x(*args, **kwargs):\n layout = self[args]\n layout_type = type(layout).__name__\n if len(container.keys()) != len(layout.keys()):\n raise ValueError(\n 'Collated DynamicMaps must return %s with consistent number of items.'\n % layout_type)\n key = kwargs['selection_key']\n index = kwargs['selection_index']\n obj_type = kwargs['selection_type']\n dyn_type_map = defaultdict(list)\n for k, v in layout.data.items():\n if k == key:\n return layout[k]\n dyn_type_map[type(v)].append(v)\n dyn_type_counter = {t: len(vals) for t, vals in dyn_type_map.\n items()}\n if dyn_type_counter != type_counter:\n raise ValueError(\n 'The objects in a %s returned by a DynamicMap must consistently return the same number of items of the same type.'\n % layout_type)\n return dyn_type_map[obj_type][index]\n callback = Callable(partial(collation_cb, selection_key=k,\n selection_index=type_counter[type(v)], selection_type=type(v)),\n inputs=[self])\n vstreams = list(util.unique_iterator(self_dstreams + vstreams))\n vdmap = self.clone(callback=callback, shared_data=False, streams=\n vstreams)\n type_counter[type(v)] += 1\n for stream in vstreams:\n if stream.source is self:\n stream.source = vdmap\n container[k] = vdmap\n unmapped_streams = [repr(stream) for stream in self.streams if stream.\n source is self and stream not in remapped_streams and stream.linked]\n if unmapped_streams:\n raise ValueError(\n \"\"\"The following streams are set to be automatically linked to a plot, but no stream_mapping specifying which item in the (Nd)Layout to link it to was found:\n%s\"\"\"\n % ', '.join(unmapped_streams))\n return container\n",
"role": "user"
},
{
"content": "collate",
"role": "assistant"
}
] |
def x(self, dimensions=None, container_type=None, group_type=None, **kwargs):
if dimensions is None:
dimensions = self.kdims
if not isinstance(dimensions, (list, tuple)):
dimensions = [dimensions]
container_type = container_type if container_type else type(self)
group_type = group_type if group_type else type(self)
outer_kdims = [self.get_dimension(d) for d in dimensions]
inner_kdims = [d for d in self.kdims if not d in outer_kdims]
outer_dynamic = issubclass(container_type, DynamicMap)
inner_dynamic = issubclass(group_type, DynamicMap)
if not outer_dynamic and any(not d.values for d in outer_kdims
) or not inner_dynamic and any(not d.values for d in inner_kdims):
raise Exception(
'Dimensions must specify sampling via values to apply a groupby')
if outer_dynamic:
def x(*outer_key, **dynkwargs):
if inner_dynamic:
def x(*inner_key, **dynkwargs):
outer_vals = zip(outer_kdims, util.wrap_tuple(outer_key))
inner_vals = zip(inner_kdims, util.wrap_tuple(inner_key))
inner_sel = [(k.name, v) for k, v in inner_vals]
outer_sel = [(k.name, v) for k, v in outer_vals]
return self.select(**dict(inner_sel + outer_sel))
return self.clone([], callback=inner_fn, kdims=inner_kdims)
else:
dim_vals = [(d.name, d.values) for d in inner_kdims]
dim_vals += [(d.name, [v]) for d, v in zip(outer_kdims,
util.wrap_tuple(outer_key))]
with item_check(False):
selected = HoloMap(self.select(**dict(dim_vals)))
return group_type(selected.reindex(inner_kdims))
if outer_kdims:
return self.clone([], callback=outer_fn, kdims=outer_kdims)
else:
return outer_fn(())
else:
outer_product = itertools.product(*[self.get_dimension(d).values for
d in dimensions])
groups = []
for outer in outer_product:
outer_vals = [(d.name, [o]) for d, o in zip(outer_kdims, outer)]
if inner_dynamic or not inner_kdims:
def x(outer_vals, *key, **dynkwargs):
inner_dims = zip(inner_kdims, util.wrap_tuple(key))
inner_vals = [(d.name, k) for d, k in inner_dims]
return self.select(**dict(outer_vals + inner_vals)).last
if inner_kdims or self.streams:
group = self.clone(callback=partial(inner_fn,
outer_vals), kdims=inner_kdims)
else:
group = inner_fn(outer_vals, ())
groups.append((outer, group))
else:
inner_vals = [(d.name, self.get_dimension(d).values) for d in
inner_kdims]
with item_check(False):
selected = HoloMap(self.select(**dict(outer_vals +
inner_vals)))
group = group_type(selected.reindex(inner_kdims))
groups.append((outer, group))
return container_type(groups, kdims=outer_kdims)
| groupby | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimensions=None, container_type=None, group_type=None, **kwargs):\n if dimensions is None:\n dimensions = self.kdims\n if not isinstance(dimensions, (list, tuple)):\n dimensions = [dimensions]\n container_type = container_type if container_type else type(self)\n group_type = group_type if group_type else type(self)\n outer_kdims = [self.get_dimension(d) for d in dimensions]\n inner_kdims = [d for d in self.kdims if not d in outer_kdims]\n outer_dynamic = issubclass(container_type, DynamicMap)\n inner_dynamic = issubclass(group_type, DynamicMap)\n if not outer_dynamic and any(not d.values for d in outer_kdims\n ) or not inner_dynamic and any(not d.values for d in inner_kdims):\n raise Exception(\n 'Dimensions must specify sampling via values to apply a groupby')\n if outer_dynamic:\n\n def x(*outer_key, **dynkwargs):\n if inner_dynamic:\n\n def x(*inner_key, **dynkwargs):\n outer_vals = zip(outer_kdims, util.wrap_tuple(outer_key))\n inner_vals = zip(inner_kdims, util.wrap_tuple(inner_key))\n inner_sel = [(k.name, v) for k, v in inner_vals]\n outer_sel = [(k.name, v) for k, v in outer_vals]\n return self.select(**dict(inner_sel + outer_sel))\n return self.clone([], callback=inner_fn, kdims=inner_kdims)\n else:\n dim_vals = [(d.name, d.values) for d in inner_kdims]\n dim_vals += [(d.name, [v]) for d, v in zip(outer_kdims,\n util.wrap_tuple(outer_key))]\n with item_check(False):\n selected = HoloMap(self.select(**dict(dim_vals)))\n return group_type(selected.reindex(inner_kdims))\n if outer_kdims:\n return self.clone([], callback=outer_fn, kdims=outer_kdims)\n else:\n return outer_fn(())\n else:\n outer_product = itertools.product(*[self.get_dimension(d).values for\n d in dimensions])\n groups = []\n for outer in outer_product:\n outer_vals = [(d.name, [o]) for d, o in zip(outer_kdims, outer)]\n if inner_dynamic or not inner_kdims:\n\n def x(outer_vals, *key, **dynkwargs):\n inner_dims = zip(inner_kdims, util.wrap_tuple(key))\n inner_vals = [(d.name, k) for d, k in inner_dims]\n return self.select(**dict(outer_vals + inner_vals)).last\n if inner_kdims or self.streams:\n group = self.clone(callback=partial(inner_fn,\n outer_vals), kdims=inner_kdims)\n else:\n group = inner_fn(outer_vals, ())\n groups.append((outer, group))\n else:\n inner_vals = [(d.name, self.get_dimension(d).values) for d in\n inner_kdims]\n with item_check(False):\n selected = HoloMap(self.select(**dict(outer_vals +\n inner_vals)))\n group = group_type(selected.reindex(inner_kdims))\n groups.append((outer, group))\n return container_type(groups, kdims=outer_kdims)\n",
"role": "user"
},
{
"content": "groupby",
"role": "assistant"
}
] |
def x(self, dimensions=None, **kwargs):
return self.groupby(dimensions, container_type=GridSpace, **kwargs)
| grid | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimensions=None, **kwargs):\n return self.groupby(dimensions, container_type=GridSpace, **kwargs)\n",
"role": "user"
},
{
"content": "grid",
"role": "assistant"
}
] |
def x(self, dimensions=None, **kwargs):
return self.groupby(dimensions, container_type=NdLayout, **kwargs)
| layout | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimensions=None, **kwargs):\n return self.groupby(dimensions, container_type=NdLayout, **kwargs)\n",
"role": "user"
},
{
"content": "layout",
"role": "assistant"
}
] |
def x(self, dimensions=None, **kwargs):
if dimensions is None:
dimensions = self.kdims
else:
if not isinstance(dimensions, (list, tuple)):
dimensions = [dimensions]
dimensions = [self.get_dimension(d, strict=True) for d in dimensions]
dims = [d for d in self.kdims if d not in dimensions]
return self.groupby(dims, group_type=NdOverlay)
| overlay | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimensions=None, **kwargs):\n if dimensions is None:\n dimensions = self.kdims\n else:\n if not isinstance(dimensions, (list, tuple)):\n dimensions = [dimensions]\n dimensions = [self.get_dimension(d, strict=True) for d in dimensions]\n dims = [d for d in self.kdims if d not in dimensions]\n return self.groupby(dims, group_type=NdOverlay)\n",
"role": "user"
},
{
"content": "overlay",
"role": "assistant"
}
] |
def x(self, dimension=None, num_bins=20, bin_range=None, adjoin=True, **kwargs
):
def x(obj, **dynkwargs):
if isinstance(obj, (NdOverlay, Overlay)):
index = kwargs.get('index', 0)
obj = obj.get(index)
return obj.hist(num_bins=num_bins, bin_range=bin_range, adjoin=
False, **kwargs)
from ..util import Dynamic
hist = Dynamic(self, streams=self.streams, link_inputs=False, operation
=dynamic_hist)
if adjoin:
return self << hist
else:
return hist
| hist | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimension=None, num_bins=20, bin_range=None, adjoin=True, **kwargs\n ):\n\n def x(obj, **dynkwargs):\n if isinstance(obj, (NdOverlay, Overlay)):\n index = kwargs.get('index', 0)\n obj = obj.get(index)\n return obj.hist(num_bins=num_bins, bin_range=bin_range, adjoin=\n False, **kwargs)\n from ..util import Dynamic\n hist = Dynamic(self, streams=self.streams, link_inputs=False, operation\n =dynamic_hist)\n if adjoin:\n return self << hist\n else:\n return hist\n",
"role": "user"
},
{
"content": "hist",
"role": "assistant"
}
] |
def x(self, kdims=[], force=False):
if not isinstance(kdims, list):
kdims = [kdims]
kdims = [self.get_dimension(kd, strict=True) for kd in kdims]
dropped = [kd for kd in self.kdims if kd not in kdims]
if dropped:
raise ValueError(
'DynamicMap does not allow dropping dimensions, reindex may only be used to reorder dimensions.'
)
return super(DynamicMap, self).reindex(kdims, force)
| reindex | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, kdims=[], force=False):\n if not isinstance(kdims, list):\n kdims = [kdims]\n kdims = [self.get_dimension(kd, strict=True) for kd in kdims]\n dropped = [kd for kd in self.kdims if kd not in kdims]\n if dropped:\n raise ValueError(\n 'DynamicMap does not allow dropping dimensions, reindex may only be used to reorder dimensions.'\n )\n return super(DynamicMap, self).reindex(kdims, force)\n",
"role": "user"
},
{
"content": "reindex",
"role": "assistant"
}
] |
def x(self, key):
ndims = self.ndims
if all(not (isinstance(el, slice) or callable(el)) for el in key):
dim_inds = []
for dim in self.kdims:
dim_type = self.get_dimension_type(dim)
if isinstance(dim_type, type) and issubclass(dim_type, Number):
dim_inds.append(self.get_dimension_index(dim))
str_keys = iter(key[i] for i in range(self.ndims) if i not in dim_inds)
num_keys = []
if len(dim_inds):
keys = list({tuple(k[i] if ndims > 1 else k for i in dim_inds) for
k in self.keys()})
q = np.array([tuple(key[i] if ndims > 1 else key for i in
dim_inds)])
idx = np.argmin([(np.inner(q - np.array(x), q - np.array(x)) if
len(dim_inds) == 2 else np.abs(q - x)) for x in keys])
num_keys = iter(keys[idx])
key = tuple(next(num_keys) if i in dim_inds else next(str_keys) for
i in range(self.ndims))
elif any(not (isinstance(el, slice) or callable(el)) for el in key):
keys = self.keys()
for i, k in enumerate(key):
if isinstance(k, slice):
continue
dim_keys = np.array([ke[i] for ke in keys])
if dim_keys.dtype.kind in 'OSU':
continue
snapped_val = dim_keys[np.argmin(np.abs(dim_keys - k))]
key = list(key)
key[i] = snapped_val
key = tuple(key)
return key
| transform_indices | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key):\n ndims = self.ndims\n if all(not (isinstance(el, slice) or callable(el)) for el in key):\n dim_inds = []\n for dim in self.kdims:\n dim_type = self.get_dimension_type(dim)\n if isinstance(dim_type, type) and issubclass(dim_type, Number):\n dim_inds.append(self.get_dimension_index(dim))\n str_keys = iter(key[i] for i in range(self.ndims) if i not in dim_inds)\n num_keys = []\n if len(dim_inds):\n keys = list({tuple(k[i] if ndims > 1 else k for i in dim_inds) for\n k in self.keys()})\n q = np.array([tuple(key[i] if ndims > 1 else key for i in\n dim_inds)])\n idx = np.argmin([(np.inner(q - np.array(x), q - np.array(x)) if\n len(dim_inds) == 2 else np.abs(q - x)) for x in keys])\n num_keys = iter(keys[idx])\n key = tuple(next(num_keys) if i in dim_inds else next(str_keys) for\n i in range(self.ndims))\n elif any(not (isinstance(el, slice) or callable(el)) for el in key):\n keys = self.keys()\n for i, k in enumerate(key):\n if isinstance(k, slice):\n continue\n dim_keys = np.array([ke[i] for ke in keys])\n if dim_keys.dtype.kind in 'OSU':\n continue\n snapped_val = dim_keys[np.argmin(np.abs(dim_keys - k))]\n key = list(key)\n key[i] = snapped_val\n key = tuple(key)\n return key\n",
"role": "user"
},
{
"content": "transform_indices",
"role": "assistant"
}
] |
def x(self, full_grid=False):
keys = super(GridSpace, self).keys()
if self.ndims == 1 or not full_grid:
return keys
dim1_keys = sorted(set(k[0] for k in keys))
dim2_keys = sorted(set(k[1] for k in keys))
return [(d1, d2) for d1 in dim1_keys for d2 in dim2_keys]
| keys | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, full_grid=False):\n keys = super(GridSpace, self).keys()\n if self.ndims == 1 or not full_grid:\n return keys\n dim1_keys = sorted(set(k[0] for k in keys))\n dim2_keys = sorted(set(k[1] for k in keys))\n return [(d1, d2) for d1 in dim1_keys for d2 in dim2_keys]\n",
"role": "user"
},
{
"content": "keys",
"role": "assistant"
}
] |
def x(self):
if self.type == HoloMap:
last_items = [(k, v.last if isinstance(v, HoloMap) else v) for k, v in
self.data.items()]
else:
last_items = self.data
return self.clone(last_items)
| last | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.type == HoloMap:\n last_items = [(k, v.last if isinstance(v, HoloMap) else v) for k, v in\n self.data.items()]\n else:\n last_items = self.data\n return self.clone(last_items)\n",
"role": "user"
},
{
"content": "last",
"role": "assistant"
}
] |
def x(self):
keys = self.keys()
if self.ndims == 1:
return len(keys), 1
return len(set(k[0] for k in keys)), len(set(k[1] for k in keys))
| shape | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n keys = self.keys()\n if self.ndims == 1:\n return len(keys), 1\n return len(set(k[0] for k in keys)), len(set(k[1] for k in keys))\n",
"role": "user"
},
{
"content": "shape",
"role": "assistant"
}
] |
def x(self, data, element):
if 'hover' not in self.handles or self.static_source:
return
for k, v in self.overlay_dims.items():
dim = util.dimension_sanitizer(k.name)
if dim not in data:
data[dim] = [v for _ in range(len(list(data.values())[0]))]
| get_hover_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data, element):\n if 'hover' not in self.handles or self.static_source:\n return\n for k, v in self.overlay_dims.items():\n dim = util.dimension_sanitizer(k.name)\n if dim not in data:\n data[dim] = [v for _ in range(len(list(data.values())[0]))]\n",
"role": "user"
},
{
"content": "get_hover_data",
"role": "assistant"
}
] |
def x(self, data, element):
if 'hover' not in self.handles or self.static_source:
return
npath = len([vs for vs in data.values()][0])
for d in element.vdims:
dim = util.dimension_sanitizer(d.name)
if dim not in data:
if element.level is not None:
data[dim] = np.full(npath, element.level)
elif element.interface.isscalar(element, d):
data[dim] = element.dimension_values(d, expanded=False)
else:
data[dim] = element.split(datatype='array', dimensions=[d])
elif isinstance(data[dim], np.ndarray) and data[dim].dtype.kind == 'M':
data[dim + '_dt_strings'] = [d.pprint_value(v) for v in data[dim]]
for k, v in self.overlay_dims.items():
dim = util.dimension_sanitizer(k.name)
if dim not in data:
data[dim] = [v for _ in range(len(list(data.values())[0]))]
| get_hover_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data, element):\n if 'hover' not in self.handles or self.static_source:\n return\n npath = len([vs for vs in data.values()][0])\n for d in element.vdims:\n dim = util.dimension_sanitizer(d.name)\n if dim not in data:\n if element.level is not None:\n data[dim] = np.full(npath, element.level)\n elif element.interface.isscalar(element, d):\n data[dim] = element.dimension_values(d, expanded=False)\n else:\n data[dim] = element.split(datatype='array', dimensions=[d])\n elif isinstance(data[dim], np.ndarray) and data[dim].dtype.kind == 'M':\n data[dim + '_dt_strings'] = [d.pprint_value(v) for v in data[dim]]\n for k, v in self.overlay_dims.items():\n dim = util.dimension_sanitizer(k.name)\n if dim not in data:\n data[dim] = [v for _ in range(len(list(data.values())[0]))]\n",
"role": "user"
},
{
"content": "get_hover_data",
"role": "assistant"
}
] |
def x(self, plot, mapping, properties):
plot_method = properties.pop('plot_method', None)
properties = mpl_to_bokeh(properties)
data = dict(properties, **mapping)
if self._has_holes:
plot_method = 'multi_polygons'
elif plot_method is None:
plot_method = self._plot_methods.get('single')
renderer = getattr(plot, plot_method)(**data)
if self.colorbar:
for k, v in list(self.handles.items()):
if not k.endswith('color_mapper'):
continue
self._draw_colorbar(plot, v, k[:-12])
return renderer, renderer.glyph
| init_glyph | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, plot, mapping, properties):\n plot_method = properties.pop('plot_method', None)\n properties = mpl_to_bokeh(properties)\n data = dict(properties, **mapping)\n if self._has_holes:\n plot_method = 'multi_polygons'\n elif plot_method is None:\n plot_method = self._plot_methods.get('single')\n renderer = getattr(plot, plot_method)(**data)\n if self.colorbar:\n for k, v in list(self.handles.items()):\n if not k.endswith('color_mapper'):\n continue\n self._draw_colorbar(plot, v, k[:-12])\n return renderer, renderer.glyph\n",
"role": "user"
},
{
"content": "init_glyph",
"role": "assistant"
}
] |
def x(self, dimension=None, num_bins=20, bin_range=None, adjoin=True, index
=0, **kwargs):
valid_ind = isinstance(index, int) and 0 <= index < len(self)
valid_label = index in [el.label for el in self]
if not any([valid_ind, valid_label]):
raise TypeError(
'Please supply a suitable index or label for the histogram data')
hists = self.get(index).hist(adjoin=False, dimension=dimension,
bin_range=bin_range, num_bins=num_bins, **kwargs)
if not isinstance(hists, Layout):
hists = [hists]
if not isinstance(dimension, list):
dimension = ['Default']
if adjoin:
layout = self
for hist in hists:
layout = layout << hist
layout.main_layer = index
elif len(dimension) > 1:
layout = hists
else:
layout = hists[0]
return layout
| hist | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimension=None, num_bins=20, bin_range=None, adjoin=True, index\n =0, **kwargs):\n valid_ind = isinstance(index, int) and 0 <= index < len(self)\n valid_label = index in [el.label for el in self]\n if not any([valid_ind, valid_label]):\n raise TypeError(\n 'Please supply a suitable index or label for the histogram data')\n hists = self.get(index).hist(adjoin=False, dimension=dimension,\n bin_range=bin_range, num_bins=num_bins, **kwargs)\n if not isinstance(hists, Layout):\n hists = [hists]\n if not isinstance(dimension, list):\n dimension = ['Default']\n if adjoin:\n layout = self\n for hist in hists:\n layout = layout << hist\n layout.main_layer = index\n elif len(dimension) > 1:\n layout = hists\n else:\n layout = hists[0]\n return layout\n",
"role": "user"
},
{
"content": "hist",
"role": "assistant"
}
] |
def x(self, dimension, expanded=True, flat=True):
values = []
found = False
for el in self:
if dimension in el.dimensions(label=True):
values.append(el.dimension_values(dimension))
found = True
if not found:
return super(CompositeOverlay, self).dimension_values(dimension,
expanded, flat)
values = [v for v in values if v is not None and len(v)]
if not values:
return np.array()
vals = np.concatenate(values)
return vals if expanded else unique_array(vals)
| dimension_values | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dimension, expanded=True, flat=True):\n values = []\n found = False\n for el in self:\n if dimension in el.dimensions(label=True):\n values.append(el.dimension_values(dimension))\n found = True\n if not found:\n return super(CompositeOverlay, self).dimension_values(dimension,\n expanded, flat)\n values = [v for v in values if v is not None and len(v)]\n if not values:\n return np.array()\n vals = np.concatenate(values)\n return vals if expanded else unique_array(vals)\n",
"role": "user"
},
{
"content": "dimension_values",
"role": "assistant"
}
] |
def x(self, identifier, default=None):
if isinstance(identifier, int):
values = list(self.data.values())
if 0 <= identifier < len(values):
return values[identifier]
else:
return default
return super(Overlay, self).get(identifier, default)
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, identifier, default=None):\n if isinstance(identifier, int):\n values = list(self.data.values())\n if 0 <= identifier < len(values):\n return values[identifier]\n else:\n return default\n return super(Overlay, self).get(identifier, default)\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
def x(self, function):
if config.future_deprecations:
self.param.warning(
'Overlay.collapse is deprecated, tocollapse multiple elements use a HoloMap.'
)
elements = list(self)
types = [type(el) for el in elements]
values = [el.group for el in elements]
if not len(set(types)) == 1 and len(set(values)) == 1:
raise Exception(
'Overlay is not homogeneous in type or group and cannot be collapsed.'
)
else:
return elements[0].clone(types[0].collapse_data([el.data for el in
elements], function, self.kdims))
| collapse | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, function):\n if config.future_deprecations:\n self.param.warning(\n 'Overlay.collapse is deprecated, tocollapse multiple elements use a HoloMap.'\n )\n elements = list(self)\n types = [type(el) for el in elements]\n values = [el.group for el in elements]\n if not len(set(types)) == 1 and len(set(values)) == 1:\n raise Exception(\n 'Overlay is not homogeneous in type or group and cannot be collapsed.'\n )\n else:\n return elements[0].clone(types[0].collapse_data([el.data for el in\n elements], function, self.kdims))\n",
"role": "user"
},
{
"content": "collapse",
"role": "assistant"
}
] |
def x(obj):
if isinstance(obj, Overlay) and any(isinstance(o, (HoloMap, GridSpace)) for
o in obj):
return False
if isinstance(obj, HoloMap):
return not obj.type in [Layout, GridSpace, NdLayout, DynamicMap]
if isinstance(obj, (GridSpace, Layout, NdLayout)):
for el in obj.values():
if not displayable(el):
return False
return True
return True
| displayable | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj):\n if isinstance(obj, Overlay) and any(isinstance(o, (HoloMap, GridSpace)) for\n o in obj):\n return False\n if isinstance(obj, HoloMap):\n return not obj.type in [Layout, GridSpace, NdLayout, DynamicMap]\n if isinstance(obj, (GridSpace, Layout, NdLayout)):\n for el in obj.values():\n if not displayable(el):\n return False\n return True\n return True\n",
"role": "user"
},
{
"content": "displayable",
"role": "assistant"
}
] |
def x(obj):
if isinstance(obj, DynamicMap):
if isinstance(obj.last, CompositeOverlay):
return len(obj.last)
elif obj.last is None:
return None
return 1
else:
return 1
| overlay_depth | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj):\n if isinstance(obj, DynamicMap):\n if isinstance(obj.last, CompositeOverlay):\n return len(obj.last)\n elif obj.last is None:\n return None\n return 1\n else:\n return 1\n",
"role": "user"
},
{
"content": "overlay_depth",
"role": "assistant"
}
] |
def x(obj, path=[]):
path = path + [obj]
zorder_map = defaultdict(list)
if not isinstance(obj, DynamicMap):
if isinstance(obj, CompositeOverlay):
for z, o in enumerate(obj):
zorder_map[z] = [o, obj]
elif isinstance(obj, HoloMap):
for el in obj.values():
if isinstance(el, CompositeOverlay):
for k, v in compute_overlayable_zorders(el, path).items():
zorder_map[k] += v + [obj]
else:
zorder_map[0] += [obj, el]
elif obj not in zorder_map[0]:
zorder_map[0].append(obj)
return zorder_map
isoverlay = isinstance(obj.last, CompositeOverlay)
isdynoverlay = obj.callback._is_overlay
if obj not in zorder_map[0] and not isoverlay:
zorder_map[0].append(obj)
depth = overlay_depth(obj)
dmap_inputs = obj.callback.inputs if obj.callback.link_inputs else []
for z, inp in enumerate(dmap_inputs):
no_zorder_increment = False
if any(not (isoverlay_fn(p) or p.last is None) for p in path
) and isoverlay_fn(inp):
no_zorder_increment = True
input_depth = overlay_depth(inp)
if (depth is not None and input_depth is not None and depth <
input_depth):
if depth > 1:
continue
else:
no_zorder_increment = True
z = z if isdynoverlay else 0
deep_zorders = compute_overlayable_zorders(inp, path=path)
offset = max(zorder_map.keys())
for dz, objs in deep_zorders.items():
global_z = offset + z if no_zorder_increment else offset + dz + z
zorder_map[global_z] = list(unique_iterator(zorder_map[global_z
] + objs))
found = any(isinstance(p, DynamicMap) and p.callback._is_overlay for p in
path)
linked = any(isinstance(s, LinkedStream) and s.linked for s in obj.streams)
if (found or linked) and isoverlay and not isdynoverlay:
offset = max(zorder_map.keys())
for z, o in enumerate(obj.last):
if isoverlay and linked:
zorder_map[offset + z].append(obj)
if o not in zorder_map[offset + z]:
zorder_map[offset + z].append(o)
return zorder_map
| compute_overlayable_zorders | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj, path=[]):\n path = path + [obj]\n zorder_map = defaultdict(list)\n if not isinstance(obj, DynamicMap):\n if isinstance(obj, CompositeOverlay):\n for z, o in enumerate(obj):\n zorder_map[z] = [o, obj]\n elif isinstance(obj, HoloMap):\n for el in obj.values():\n if isinstance(el, CompositeOverlay):\n for k, v in compute_overlayable_zorders(el, path).items():\n zorder_map[k] += v + [obj]\n else:\n zorder_map[0] += [obj, el]\n elif obj not in zorder_map[0]:\n zorder_map[0].append(obj)\n return zorder_map\n isoverlay = isinstance(obj.last, CompositeOverlay)\n isdynoverlay = obj.callback._is_overlay\n if obj not in zorder_map[0] and not isoverlay:\n zorder_map[0].append(obj)\n depth = overlay_depth(obj)\n dmap_inputs = obj.callback.inputs if obj.callback.link_inputs else []\n for z, inp in enumerate(dmap_inputs):\n no_zorder_increment = False\n if any(not (isoverlay_fn(p) or p.last is None) for p in path\n ) and isoverlay_fn(inp):\n no_zorder_increment = True\n input_depth = overlay_depth(inp)\n if (depth is not None and input_depth is not None and depth <\n input_depth):\n if depth > 1:\n continue\n else:\n no_zorder_increment = True\n z = z if isdynoverlay else 0\n deep_zorders = compute_overlayable_zorders(inp, path=path)\n offset = max(zorder_map.keys())\n for dz, objs in deep_zorders.items():\n global_z = offset + z if no_zorder_increment else offset + dz + z\n zorder_map[global_z] = list(unique_iterator(zorder_map[global_z\n ] + objs))\n found = any(isinstance(p, DynamicMap) and p.callback._is_overlay for p in\n path)\n linked = any(isinstance(s, LinkedStream) and s.linked for s in obj.streams)\n if (found or linked) and isoverlay and not isdynoverlay:\n offset = max(zorder_map.keys())\n for z, o in enumerate(obj.last):\n if isoverlay and linked:\n zorder_map[offset + z].append(obj)\n if o not in zorder_map[offset + z]:\n zorder_map[offset + z].append(o)\n return zorder_map\n",
"role": "user"
},
{
"content": "compute_overlayable_zorders",
"role": "assistant"
}
] |
def x(dmap):
if not isinstance(dmap, DynamicMap):
return False
elif dmap.callback._is_overlay:
return True
else:
return any(is_dynamic_overlay(dm) for dm in dmap.callback.inputs)
| is_dynamic_overlay | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dmap):\n if not isinstance(dmap, DynamicMap):\n return False\n elif dmap.callback._is_overlay:\n return True\n else:\n return any(is_dynamic_overlay(dm) for dm in dmap.callback.inputs)\n",
"role": "user"
},
{
"content": "is_dynamic_overlay",
"role": "assistant"
}
] |
def x(obj, depth=0):
layers = []
if isinstance(obj, DynamicMap):
if issubclass(obj.type, NdOverlay) and not depth:
for v in obj.last.values():
layers.append(obj)
elif issubclass(obj.type, Overlay):
if obj.callback.inputs and is_dynamic_overlay(obj):
for inp in obj.callback.inputs:
layers += split_dmap_overlay(inp, depth + 1)
else:
for v in obj.last.values():
layers.append(obj)
else:
layers.append(obj)
return layers
if isinstance(obj, Overlay):
for k, v in obj.items():
layers.append(v)
else:
layers.append(obj)
return layers
| split_dmap_overlay | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj, depth=0):\n layers = []\n if isinstance(obj, DynamicMap):\n if issubclass(obj.type, NdOverlay) and not depth:\n for v in obj.last.values():\n layers.append(obj)\n elif issubclass(obj.type, Overlay):\n if obj.callback.inputs and is_dynamic_overlay(obj):\n for inp in obj.callback.inputs:\n layers += split_dmap_overlay(inp, depth + 1)\n else:\n for v in obj.last.values():\n layers.append(obj)\n else:\n layers.append(obj)\n return layers\n if isinstance(obj, Overlay):\n for k, v in obj.items():\n layers.append(v)\n else:\n layers.append(obj)\n return layers\n",
"role": "user"
},
{
"content": "split_dmap_overlay",
"role": "assistant"
}
] |
def x(obj):
dmaps = obj.traverse(lambda x: x, specs=[DynamicMap])
for dmap in dmaps:
if dmap.unbounded:
continue
if not len(dmap):
dmap[dmap._initial_key()]
| initialize_dynamic | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj):\n dmaps = obj.traverse(lambda x: x, specs=[DynamicMap])\n for dmap in dmaps:\n if dmap.unbounded:\n continue\n if not len(dmap):\n dmap[dmap._initial_key()]\n",
"role": "user"
},
{
"content": "initialize_dynamic",
"role": "assistant"
}
] |
def x(map_obj, key_map, cached=False):
if map_obj.kdims and len(map_obj.kdims) == 1 and map_obj.kdims[0
] == 'Frame':
return map_obj.last
key = tuple(key_map[kd.name] for kd in map_obj.kdims if kd.name in key_map)
if key in map_obj.data and cached:
return map_obj.data[key]
else:
try:
return map_obj[key]
except KeyError:
return None
except StopIteration as e:
raise e
except Exception:
print(traceback.format_exc())
return None
| get_plot_frame | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(map_obj, key_map, cached=False):\n if map_obj.kdims and len(map_obj.kdims) == 1 and map_obj.kdims[0\n ] == 'Frame':\n return map_obj.last\n key = tuple(key_map[kd.name] for kd in map_obj.kdims if kd.name in key_map)\n if key in map_obj.data and cached:\n return map_obj.data[key]\n else:\n try:\n return map_obj[key]\n except KeyError:\n return None\n except StopIteration as e:\n raise e\n except Exception:\n print(traceback.format_exc())\n return None\n",
"role": "user"
},
{
"content": "get_plot_frame",
"role": "assistant"
}
] |
def x(obj, key_map, cached=False):
clone = obj.map(lambda x: x)
for it1, it2 in zip(obj.traverse(lambda x: x), clone.traverse(lambda x: x)
):
if isinstance(it1, DynamicMap):
with disable_constant(it2.callback):
it2.callback.inputs = it1.callback.inputs
with item_check(False):
return clone.map(lambda x: get_plot_frame(x, key_map, cached=cached
), [DynamicMap, HoloMap], clone=False)
| get_nested_plot_frame | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj, key_map, cached=False):\n clone = obj.map(lambda x: x)\n for it1, it2 in zip(obj.traverse(lambda x: x), clone.traverse(lambda x: x)\n ):\n if isinstance(it1, DynamicMap):\n with disable_constant(it2.callback):\n it2.callback.inputs = it1.callback.inputs\n with item_check(False):\n return clone.map(lambda x: get_plot_frame(x, key_map, cached=cached\n ), [DynamicMap, HoloMap], clone=False)\n",
"role": "user"
},
{
"content": "get_nested_plot_frame",
"role": "assistant"
}
] |
def x(obj, html=False):
collate = '<tt>collate</tt>' if html else 'collate'
info = (
'For more information, please consult the Composing Data tutorial (http://git.io/vtIQh)'
)
if isinstance(obj, HoloMap):
error = ('HoloMap of %s objects cannot be displayed.' % obj.type.
__name__)
remedy = (
'Please call the %s method to generate a displayable object' %
collate)
elif isinstance(obj, Layout):
error = (
'Layout containing HoloMaps of Layout or GridSpace objects cannot be displayed.'
)
remedy = ('Please call the %s method on the appropriate elements.' %
collate)
elif isinstance(obj, GridSpace):
error = 'GridSpace containing HoloMaps of Layouts cannot be displayed.'
remedy = ('Please call the %s method on the appropriate elements.' %
collate)
if not html:
return '\n'.join([error, remedy, info])
else:
return '<center>{msg}</center>'.format(msg='<br>'.join(['<b>%s</b>' %
error, remedy, '<i>%s</i>' % info]))
| undisplayable_info | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj, html=False):\n collate = '<tt>collate</tt>' if html else 'collate'\n info = (\n 'For more information, please consult the Composing Data tutorial (http://git.io/vtIQh)'\n )\n if isinstance(obj, HoloMap):\n error = ('HoloMap of %s objects cannot be displayed.' % obj.type.\n __name__)\n remedy = (\n 'Please call the %s method to generate a displayable object' %\n collate)\n elif isinstance(obj, Layout):\n error = (\n 'Layout containing HoloMaps of Layout or GridSpace objects cannot be displayed.'\n )\n remedy = ('Please call the %s method on the appropriate elements.' %\n collate)\n elif isinstance(obj, GridSpace):\n error = 'GridSpace containing HoloMaps of Layouts cannot be displayed.'\n remedy = ('Please call the %s method on the appropriate elements.' %\n collate)\n if not html:\n return '\\n'.join([error, remedy, info])\n else:\n return '<center>{msg}</center>'.format(msg='<br>'.join(['<b>%s</b>' %\n error, remedy, '<i>%s</i>' % info]))\n",
"role": "user"
},
{
"content": "undisplayable_info",
"role": "assistant"
}
] |
def x(sizes, size_fn, scaling_factor, scaling_method, base_size):
if sizes.dtype.kind not in ('i', 'f'):
return None
if scaling_method == 'area':
pass
elif scaling_method == 'width':
scaling_factor = scaling_factor ** 2
else:
raise ValueError(
'Invalid value for argument "scaling_method": "{}". Valid values are: "width", "area".'
.format(scaling_method))
sizes = size_fn(sizes)
return base_size * scaling_factor * sizes
| compute_sizes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sizes, size_fn, scaling_factor, scaling_method, base_size):\n if sizes.dtype.kind not in ('i', 'f'):\n return None\n if scaling_method == 'area':\n pass\n elif scaling_method == 'width':\n scaling_factor = scaling_factor ** 2\n else:\n raise ValueError(\n 'Invalid value for argument \"scaling_method\": \"{}\". Valid values are: \"width\", \"area\".'\n .format(scaling_method))\n sizes = size_fn(sizes)\n return base_size * scaling_factor * sizes\n",
"role": "user"
},
{
"content": "compute_sizes",
"role": "assistant"
}
] |
def x(padding):
if isinstance(padding, tuple):
if len(padding) == 2:
xpad, ypad = padding
zpad = 0
elif len(padding) == 3:
xpad, ypad, zpad = padding
else:
raise ValueError(
'Padding must be supplied as an number applied to all axes or a length two or three tuple corresponding to the x-, y- and optionally z-axis'
)
else:
xpad, ypad, zpad = (padding,) * 3
return xpad, ypad, zpad
| get_axis_padding | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(padding):\n if isinstance(padding, tuple):\n if len(padding) == 2:\n xpad, ypad = padding\n zpad = 0\n elif len(padding) == 3:\n xpad, ypad, zpad = padding\n else:\n raise ValueError(\n 'Padding must be supplied as an number applied to all axes or a length two or three tuple corresponding to the x-, y- and optionally z-axis'\n )\n else:\n xpad, ypad, zpad = (padding,) * 3\n return xpad, ypad, zpad\n",
"role": "user"
},
{
"content": "get_axis_padding",
"role": "assistant"
}
] |
def x(low, high, span):
if is_number(low) and low == high:
if isinstance(low, np.datetime64):
span = span * np.timedelta64(1, 's')
low, high = low - span, high + span
return low, high
| get_minimum_span | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(low, high, span):\n if is_number(low) and low == high:\n if isinstance(low, np.datetime64):\n span = span * np.timedelta64(1, 's')\n low, high = low - span, high + span\n return low, high\n",
"role": "user"
},
{
"content": "get_minimum_span",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.