source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self): limits = {} limits['Max spot instance requests per region'] = AwsLimit( 'Max spot instance requests per region', self, 20, self. warning_threshold, self.critical_threshold, limit_type= 'Spot instance requests') limits['Max active spot fleets per region'] = AwsLimit( 'Max active spot fleets per region', self, 1000, self. warning_threshold, self.critical_threshold) limits['Max launch specifications per spot fleet'] = AwsLimit( 'Max launch specifications per spot fleet', self, 50, self. warning_threshold, self.critical_threshold) limits['Max target capacity per spot fleet'] = AwsLimit( 'Max target capacity per spot fleet', self, 3000, self. warning_threshold, self.critical_threshold) limits['Max target capacity for all spot fleets in region'] = AwsLimit( 'Max target capacity for all spot fleets in region', self, 5000, self.warning_threshold, self.critical_threshold) return limits
get_limits_spot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n limits = {}\n limits['Max spot instance requests per region'] = AwsLimit(\n 'Max spot instance requests per region', self, 20, self.\n warning_threshold, self.critical_threshold, limit_type=\n 'Spot instance requests')\n limits['Max active spot fleets per region'] = AwsLimit(\n 'Max active spot fleets per region', self, 1000, self.\n warning_threshold, self.critical_threshold)\n limits['Max launch specifications per spot fleet'] = AwsLimit(\n 'Max launch specifications per spot fleet', self, 50, self.\n warning_threshold, self.critical_threshold)\n limits['Max target capacity per spot fleet'] = AwsLimit(\n 'Max target capacity per spot fleet', self, 3000, self.\n warning_threshold, self.critical_threshold)\n limits['Max target capacity for all spot fleets in region'] = AwsLimit(\n 'Max target capacity for all spot fleets in region', self, 5000,\n self.warning_threshold, self.critical_threshold)\n return limits\n", "role": "user" }, { "content": "get_limits_spot", "role": "assistant" } ]
def x(self): logger.debug('Getting usage for EC2 VPC resources') sgs_per_vpc = defaultdict(int) rules_per_sg = defaultdict(int) for sg in self.resource_conn.security_groups.all(): if sg.vpc_id is not None: sgs_per_vpc[sg.vpc_id] += 1 rules_per_sg[sg.id] = len(sg.ip_permissions) for vpc_id, count in sgs_per_vpc.items(): self.limits['Security groups per VPC']._add_current_usage(count, aws_type='AWS::EC2::VPC', resource_id=vpc_id) for sg_id, count in rules_per_sg.items(): self.limits['Rules per VPC security group']._add_current_usage(count, aws_type='AWS::EC2::SecurityGroupRule', resource_id=sg_id)
find_usage_networking_sgs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n logger.debug('Getting usage for EC2 VPC resources')\n sgs_per_vpc = defaultdict(int)\n rules_per_sg = defaultdict(int)\n for sg in self.resource_conn.security_groups.all():\n if sg.vpc_id is not None:\n sgs_per_vpc[sg.vpc_id] += 1\n rules_per_sg[sg.id] = len(sg.ip_permissions)\n for vpc_id, count in sgs_per_vpc.items():\n self.limits['Security groups per VPC']._add_current_usage(count,\n aws_type='AWS::EC2::VPC', resource_id=vpc_id)\n for sg_id, count in rules_per_sg.items():\n self.limits['Rules per VPC security group']._add_current_usage(count,\n aws_type='AWS::EC2::SecurityGroupRule', resource_id=sg_id)\n", "role": "user" }, { "content": "find_usage_networking_sgs", "role": "assistant" } ]
def x(self): limits = {} limits['Security groups per VPC'] = AwsLimit('Security groups per VPC', self, 500, self.warning_threshold, self.critical_threshold, limit_type='AWS::EC2::SecurityGroup', limit_subtype='AWS::EC2::VPC') limits['Rules per VPC security group'] = AwsLimit( 'Rules per VPC security group', self, 50, self.warning_threshold, self.critical_threshold, limit_type='AWS::EC2::SecurityGroup', limit_subtype='AWS::EC2::VPC') limits['VPC Elastic IP addresses (EIPs)'] = AwsLimit( 'VPC Elastic IP addresses (EIPs)', self, 5, self.warning_threshold, self.critical_threshold, limit_type='AWS::EC2::EIP', limit_subtype= 'AWS::EC2::VPC', ta_service_name='VPC') limits['Elastic IP addresses (EIPs)'] = AwsLimit( 'Elastic IP addresses (EIPs)', self, 5, self.warning_threshold, self.critical_threshold, limit_type='AWS::EC2::EIP') limits['VPC security groups per elastic network interface'] = AwsLimit( 'VPC security groups per elastic network interface', self, 5, self. warning_threshold, self.critical_threshold, limit_type= 'AWS::EC2::SecurityGroup', limit_subtype='AWS::EC2::NetworkInterface') return limits
get_limits_networking
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n limits = {}\n limits['Security groups per VPC'] = AwsLimit('Security groups per VPC',\n self, 500, self.warning_threshold, self.critical_threshold,\n limit_type='AWS::EC2::SecurityGroup', limit_subtype='AWS::EC2::VPC')\n limits['Rules per VPC security group'] = AwsLimit(\n 'Rules per VPC security group', self, 50, self.warning_threshold,\n self.critical_threshold, limit_type='AWS::EC2::SecurityGroup',\n limit_subtype='AWS::EC2::VPC')\n limits['VPC Elastic IP addresses (EIPs)'] = AwsLimit(\n 'VPC Elastic IP addresses (EIPs)', self, 5, self.warning_threshold,\n self.critical_threshold, limit_type='AWS::EC2::EIP', limit_subtype=\n 'AWS::EC2::VPC', ta_service_name='VPC')\n limits['Elastic IP addresses (EIPs)'] = AwsLimit(\n 'Elastic IP addresses (EIPs)', self, 5, self.warning_threshold,\n self.critical_threshold, limit_type='AWS::EC2::EIP')\n limits['VPC security groups per elastic network interface'] = AwsLimit(\n 'VPC security groups per elastic network interface', self, 5, self.\n warning_threshold, self.critical_threshold, limit_type=\n 'AWS::EC2::SecurityGroup', limit_subtype='AWS::EC2::NetworkInterface')\n return limits\n", "role": "user" }, { "content": "get_limits_networking", "role": "assistant" } ]
def x(self): GENERAL_TYPES = ['a1.2xlarge', 'a1.4xlarge', 'a1.large', 'a1.medium', 'a1.xlarge', 't2.nano', 't2.micro', 't2.small', 't2.medium', 't2.large', 't2.xlarge', 't2.2xlarge', 't3.nano', 't3.micro', 't3.small', 't3.medium', 't3.large', 't3.xlarge', 't3.2xlarge', 'm3.medium', 'm3.large', 'm3.xlarge', 'm3.2xlarge', 'm4.large', 'm4.xlarge', 'm4.2xlarge', 'm4.4xlarge', 'm4.10xlarge', 'm4.16xlarge', 'm5.12xlarge', 'm5.24xlarge', 'm5.2xlarge', 'm5.4xlarge', 'm5.large', 'm5.xlarge', 'm5d.12xlarge', 'm5d.24xlarge', 'm5d.2xlarge', 'm5d.4xlarge', 'm5d.large', 'm5d.xlarge', 'm5a.12xlarge', 'm5a.24xlarge', 'm5a.2xlarge', 'm5a.4xlarge', 'm5a.large', 'm5a.xlarge'] PREV_GENERAL_TYPES = ['t1.micro', 'm1.small', 'm1.medium', 'm1.large', 'm1.xlarge'] MEMORY_TYPES = ['r3.2xlarge', 'r3.4xlarge', 'r3.8xlarge', 'r3.large', 'r3.xlarge', 'r4.2xlarge', 'r4.4xlarge', 'r4.8xlarge', 'r4.16xlarge', 'r4.large', 'r4.xlarge', 'r5.2xlarge', 'r5.4xlarge', 'r5.8xlarge', 'r5.12xlarge', 'r5.16xlarge', 'r5.24xlarge', 'r5.large', 'r5.metal', 'r5.xlarge', 'r5a.12xlarge', 'r5a.24xlarge', 'r5a.2xlarge', 'r5a.4xlarge', 'r5a.large', 'r5a.xlarge', 'r5d.2xlarge', 'r5d.4xlarge', 'r5d.8xlarge', 'r5d.12xlarge', 'r5d.16xlarge', 'r5d.24xlarge', 'r5d.large', 'r5d.metal', 'r5d.xlarge', 'x1.16xlarge', 'x1.32xlarge', 'x1e.2xlarge', 'x1e.4xlarge', 'x1e.8xlarge', 'x1e.16xlarge', 'x1e.32xlarge', 'x1e.xlarge', 'z1d.2xlarge', 'z1d.3xlarge', 'z1d.6xlarge', 'z1d.12xlarge', 'z1d.large', 'z1d.xlarge'] PREV_MEMORY_TYPES = ['m2.xlarge', 'm2.2xlarge', 'm2.4xlarge', 'cr1.8xlarge' ] COMPUTE_TYPES = ['c3.large', 'c3.xlarge', 'c3.2xlarge', 'c3.4xlarge', 'c3.8xlarge', 'c4.large', 'c4.xlarge', 'c4.2xlarge', 'c4.4xlarge', 'c4.8xlarge', 'c5.18xlarge', 'c5.2xlarge', 'c5.4xlarge', 'c5.9xlarge', 'c5.large', 'c5.xlarge', 'c5d.18xlarge', 'c5d.2xlarge', 'c5d.4xlarge', 'c5d.9xlarge', 'c5d.large', 'c5d.xlarge', 'c5n.18xlarge', 'c5n.2xlarge', 'c5n.4xlarge', 'c5n.9xlarge', 'c5n.large', 'c5n.xlarge'] PREV_COMPUTE_TYPES = ['c1.medium', 'c1.xlarge', 'cc2.8xlarge', 'cc1.4xlarge'] ACCELERATED_COMPUTE_TYPES = ['f1.4xlarge', 'p2.xlarge', 'p2.8xlarge', 'p2.16xlarge', 'p3.16xlarge', 'p3.2xlarge', 'p3.8xlarge', 'p3dn.24xlarge'] STORAGE_TYPES = ['i2.xlarge', 'i2.2xlarge', 'i2.4xlarge', 'i2.8xlarge', 'i3.large', 'i3.xlarge', 'i3.2xlarge', 'i3.4xlarge', 'i3.8xlarge', 'i3.16xlarge', 'i3.metal', 'h1.16xlarge', 'h1.2xlarge', 'h1.4xlarge', 'h1.8xlarge'] PREV_STORAGE_TYPES = ['hi1.4xlarge', 'hs1.8xlarge'] DENSE_STORAGE_TYPES = ['d2.xlarge', 'd2.2xlarge', 'd2.4xlarge', 'd2.8xlarge'] GPU_TYPES = ['g2.2xlarge', 'g2.8xlarge', 'g3.16xlarge', 'g3.4xlarge', 'g3.8xlarge', 'g3s.xlarge'] PREV_GPU_TYPES = ['cg1.4xlarge'] FPGA_TYPES = ['f1.2xlarge', 'f1.16xlarge'] return (GENERAL_TYPES + PREV_GENERAL_TYPES + MEMORY_TYPES + PREV_MEMORY_TYPES + COMPUTE_TYPES + PREV_COMPUTE_TYPES + ACCELERATED_COMPUTE_TYPES + STORAGE_TYPES + PREV_STORAGE_TYPES + DENSE_STORAGE_TYPES + GPU_TYPES + PREV_GPU_TYPES + FPGA_TYPES)
instance_types
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n GENERAL_TYPES = ['a1.2xlarge', 'a1.4xlarge', 'a1.large', 'a1.medium',\n 'a1.xlarge', 't2.nano', 't2.micro', 't2.small', 't2.medium',\n 't2.large', 't2.xlarge', 't2.2xlarge', 't3.nano', 't3.micro',\n 't3.small', 't3.medium', 't3.large', 't3.xlarge', 't3.2xlarge',\n 'm3.medium', 'm3.large', 'm3.xlarge', 'm3.2xlarge', 'm4.large',\n 'm4.xlarge', 'm4.2xlarge', 'm4.4xlarge', 'm4.10xlarge',\n 'm4.16xlarge', 'm5.12xlarge', 'm5.24xlarge', 'm5.2xlarge',\n 'm5.4xlarge', 'm5.large', 'm5.xlarge', 'm5d.12xlarge',\n 'm5d.24xlarge', 'm5d.2xlarge', 'm5d.4xlarge', 'm5d.large',\n 'm5d.xlarge', 'm5a.12xlarge', 'm5a.24xlarge', 'm5a.2xlarge',\n 'm5a.4xlarge', 'm5a.large', 'm5a.xlarge']\n PREV_GENERAL_TYPES = ['t1.micro', 'm1.small', 'm1.medium', 'm1.large',\n 'm1.xlarge']\n MEMORY_TYPES = ['r3.2xlarge', 'r3.4xlarge', 'r3.8xlarge', 'r3.large',\n 'r3.xlarge', 'r4.2xlarge', 'r4.4xlarge', 'r4.8xlarge',\n 'r4.16xlarge', 'r4.large', 'r4.xlarge', 'r5.2xlarge', 'r5.4xlarge',\n 'r5.8xlarge', 'r5.12xlarge', 'r5.16xlarge', 'r5.24xlarge',\n 'r5.large', 'r5.metal', 'r5.xlarge', 'r5a.12xlarge', 'r5a.24xlarge',\n 'r5a.2xlarge', 'r5a.4xlarge', 'r5a.large', 'r5a.xlarge',\n 'r5d.2xlarge', 'r5d.4xlarge', 'r5d.8xlarge', 'r5d.12xlarge',\n 'r5d.16xlarge', 'r5d.24xlarge', 'r5d.large', 'r5d.metal',\n 'r5d.xlarge', 'x1.16xlarge', 'x1.32xlarge', 'x1e.2xlarge',\n 'x1e.4xlarge', 'x1e.8xlarge', 'x1e.16xlarge', 'x1e.32xlarge',\n 'x1e.xlarge', 'z1d.2xlarge', 'z1d.3xlarge', 'z1d.6xlarge',\n 'z1d.12xlarge', 'z1d.large', 'z1d.xlarge']\n PREV_MEMORY_TYPES = ['m2.xlarge', 'm2.2xlarge', 'm2.4xlarge', 'cr1.8xlarge'\n ]\n COMPUTE_TYPES = ['c3.large', 'c3.xlarge', 'c3.2xlarge', 'c3.4xlarge',\n 'c3.8xlarge', 'c4.large', 'c4.xlarge', 'c4.2xlarge', 'c4.4xlarge',\n 'c4.8xlarge', 'c5.18xlarge', 'c5.2xlarge', 'c5.4xlarge',\n 'c5.9xlarge', 'c5.large', 'c5.xlarge', 'c5d.18xlarge',\n 'c5d.2xlarge', 'c5d.4xlarge', 'c5d.9xlarge', 'c5d.large',\n 'c5d.xlarge', 'c5n.18xlarge', 'c5n.2xlarge', 'c5n.4xlarge',\n 'c5n.9xlarge', 'c5n.large', 'c5n.xlarge']\n PREV_COMPUTE_TYPES = ['c1.medium', 'c1.xlarge', 'cc2.8xlarge',\n 'cc1.4xlarge']\n ACCELERATED_COMPUTE_TYPES = ['f1.4xlarge', 'p2.xlarge', 'p2.8xlarge',\n 'p2.16xlarge', 'p3.16xlarge', 'p3.2xlarge', 'p3.8xlarge',\n 'p3dn.24xlarge']\n STORAGE_TYPES = ['i2.xlarge', 'i2.2xlarge', 'i2.4xlarge', 'i2.8xlarge',\n 'i3.large', 'i3.xlarge', 'i3.2xlarge', 'i3.4xlarge', 'i3.8xlarge',\n 'i3.16xlarge', 'i3.metal', 'h1.16xlarge', 'h1.2xlarge',\n 'h1.4xlarge', 'h1.8xlarge']\n PREV_STORAGE_TYPES = ['hi1.4xlarge', 'hs1.8xlarge']\n DENSE_STORAGE_TYPES = ['d2.xlarge', 'd2.2xlarge', 'd2.4xlarge',\n 'd2.8xlarge']\n GPU_TYPES = ['g2.2xlarge', 'g2.8xlarge', 'g3.16xlarge', 'g3.4xlarge',\n 'g3.8xlarge', 'g3s.xlarge']\n PREV_GPU_TYPES = ['cg1.4xlarge']\n FPGA_TYPES = ['f1.2xlarge', 'f1.16xlarge']\n return (GENERAL_TYPES + PREV_GENERAL_TYPES + MEMORY_TYPES +\n PREV_MEMORY_TYPES + COMPUTE_TYPES + PREV_COMPUTE_TYPES +\n ACCELERATED_COMPUTE_TYPES + STORAGE_TYPES + PREV_STORAGE_TYPES +\n DENSE_STORAGE_TYPES + GPU_TYPES + PREV_GPU_TYPES + FPGA_TYPES)\n", "role": "user" }, { "content": "instance_types", "role": "assistant" } ]
def x(self): logger.debug('Checking usage for service %s', self.service_name) self.connect() for lim in self.limits.values(): lim._reset_usage() self._find_usage_instances() self._find_usage_subnet_groups() self._find_usage_security_groups() self._update_limits_from_api() self._have_usage = True logger.debug('Done checking usage.')
find_usage
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n logger.debug('Checking usage for service %s', self.service_name)\n self.connect()\n for lim in self.limits.values():\n lim._reset_usage()\n self._find_usage_instances()\n self._find_usage_subnet_groups()\n self._find_usage_security_groups()\n self._update_limits_from_api()\n self._have_usage = True\n logger.debug('Done checking usage.')\n", "role": "user" }, { "content": "find_usage", "role": "assistant" } ]
def x(self): paginator = self.conn.get_paginator('describe_db_instances') for page in paginator.paginate(): for instance in page['DBInstances']: self.limits['Read replicas per master']._add_current_usage(len( instance['ReadReplicaDBInstanceIdentifiers']), aws_type= 'AWS::RDS::DBInstance', resource_id=instance[ 'DBInstanceIdentifier'])
find_usage_instances
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n paginator = self.conn.get_paginator('describe_db_instances')\n for page in paginator.paginate():\n for instance in page['DBInstances']:\n self.limits['Read replicas per master']._add_current_usage(len(\n instance['ReadReplicaDBInstanceIdentifiers']), aws_type=\n 'AWS::RDS::DBInstance', resource_id=instance[\n 'DBInstanceIdentifier'])\n", "role": "user" }, { "content": "find_usage_instances", "role": "assistant" } ]
def x(self): paginator = self.conn.get_paginator('describe_db_subnet_groups') for page in paginator.paginate(): for group in page['DBSubnetGroups']: self.limits['Subnets per Subnet Group']._add_current_usage(len( group['Subnets']), aws_type='AWS::RDS::DBSubnetGroup', resource_id=group['DBSubnetGroupName'])
find_usage_subnet_groups
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n paginator = self.conn.get_paginator('describe_db_subnet_groups')\n for page in paginator.paginate():\n for group in page['DBSubnetGroups']:\n self.limits['Subnets per Subnet Group']._add_current_usage(len(\n group['Subnets']), aws_type='AWS::RDS::DBSubnetGroup',\n resource_id=group['DBSubnetGroupName'])\n", "role": "user" }, { "content": "find_usage_subnet_groups", "role": "assistant" } ]
def x(self): vpc_count = 0 paginator = self.conn.get_paginator('describe_db_security_groups') for page in paginator.paginate(): for group in page['DBSecurityGroups']: if 'VpcId' in group and group['VpcId'] is not None: vpc_count += 1 self.limits['Max auths per security group']._add_current_usage( len(group['EC2SecurityGroups']) + len(group['IPRanges']), aws_type='AWS::RDS::DBSecurityGroup', resource_id=group[ 'DBSecurityGroupName']) self.limits['VPC Security Groups']._add_current_usage(vpc_count, aws_type='AWS::RDS::DBSecurityGroup')
find_usage_security_groups
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n vpc_count = 0\n paginator = self.conn.get_paginator('describe_db_security_groups')\n for page in paginator.paginate():\n for group in page['DBSecurityGroups']:\n if 'VpcId' in group and group['VpcId'] is not None:\n vpc_count += 1\n self.limits['Max auths per security group']._add_current_usage(\n len(group['EC2SecurityGroups']) + len(group['IPRanges']),\n aws_type='AWS::RDS::DBSecurityGroup', resource_id=group[\n 'DBSecurityGroupName'])\n self.limits['VPC Security Groups']._add_current_usage(vpc_count,\n aws_type='AWS::RDS::DBSecurityGroup')\n", "role": "user" }, { "content": "find_usage_security_groups", "role": "assistant" } ]
def x(self): if self.limits != {}: return self.limits limits = {} limits['DB instances'] = AwsLimit('DB instances', self, 40, self. warning_threshold, self.critical_threshold, limit_type= 'AWS::RDS::DBInstance') limits['Reserved Instances'] = AwsLimit('Reserved Instances', self, 40, self.warning_threshold, self.critical_threshold, limit_type= 'AWS::RDS::DBInstance') limits['Storage quota (GB)'] = AwsLimit('Storage quota (GB)', self, 100000, self.warning_threshold, self.critical_threshold, limit_type ='AWS::RDS::DBInstance') limits['DB snapshots per user'] = AwsLimit('DB snapshots per user', self, 100, self.warning_threshold, self.critical_threshold, limit_type='AWS::RDS::DBSnapshot') limits['DB parameter groups'] = AwsLimit('DB parameter groups', self, 50, self.warning_threshold, self.critical_threshold, limit_type= 'AWS::RDS::DBParameterGroup') limits['DB security groups'] = AwsLimit('DB security groups', self, 25, self.warning_threshold, self.critical_threshold, limit_type= 'AWS::RDS::DBSecurityGroup') limits['VPC Security Groups'] = AwsLimit('VPC Security Groups', self, 5, self.warning_threshold, self.critical_threshold, limit_type= 'AWS::RDS::DBSecurityGroup') limits['Subnet Groups'] = AwsLimit('Subnet Groups', self, 50, self. warning_threshold, self.critical_threshold, limit_type= 'AWS::RDS::DBSubnetGroup', ta_limit_name='Subnet groups') limits['Subnets per Subnet Group'] = AwsLimit('Subnets per Subnet Group', self, 20, self.warning_threshold, self.critical_threshold, limit_type='AWS::RDS::DBSubnetGroup', ta_limit_name= 'Subnets per subnet group') limits['Option Groups'] = AwsLimit('Option Groups', self, 20, self. warning_threshold, self.critical_threshold, limit_type= 'AWS::RDS::DBOptionGroup') limits['Event Subscriptions'] = AwsLimit('Event Subscriptions', self, 20, self.warning_threshold, self.critical_threshold, limit_type= 'AWS::RDS::DBEventSubscription', ta_limit_name='Event subscriptions') limits['Read replicas per master'] = AwsLimit('Read replicas per master', self, 5, self.warning_threshold, self.critical_threshold, limit_type='AWS::RDS::DBInstance') limits['Max auths per security group'] = AwsLimit( 'Max auths per security group', self, 20, self.warning_threshold, self.critical_threshold, limit_type='AWS::RDS::DBSecurityGroup', limit_subtype='AWS::RDS::DBSecurityGroupIngress') limits['DB Clusters'] = AwsLimit('DB Clusters', self, 40, self. warning_threshold, self.critical_threshold, limit_type= 'AWS::RDS::DBCluster', ta_limit_name='Clusters') limits['DB Cluster Parameter Groups'] = AwsLimit( 'DB Cluster Parameter Groups', self, 50, self.warning_threshold, self.critical_threshold, limit_type= 'AWS::RDS::DBClusterParameterGroup', ta_limit_name= 'Cluster parameter groups') self.limits = limits return limits
get_limits
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.limits != {}:\n return self.limits\n limits = {}\n limits['DB instances'] = AwsLimit('DB instances', self, 40, self.\n warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBInstance')\n limits['Reserved Instances'] = AwsLimit('Reserved Instances', self, 40,\n self.warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBInstance')\n limits['Storage quota (GB)'] = AwsLimit('Storage quota (GB)', self, \n 100000, self.warning_threshold, self.critical_threshold, limit_type\n ='AWS::RDS::DBInstance')\n limits['DB snapshots per user'] = AwsLimit('DB snapshots per user',\n self, 100, self.warning_threshold, self.critical_threshold,\n limit_type='AWS::RDS::DBSnapshot')\n limits['DB parameter groups'] = AwsLimit('DB parameter groups', self, \n 50, self.warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBParameterGroup')\n limits['DB security groups'] = AwsLimit('DB security groups', self, 25,\n self.warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBSecurityGroup')\n limits['VPC Security Groups'] = AwsLimit('VPC Security Groups', self, 5,\n self.warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBSecurityGroup')\n limits['Subnet Groups'] = AwsLimit('Subnet Groups', self, 50, self.\n warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBSubnetGroup', ta_limit_name='Subnet groups')\n limits['Subnets per Subnet Group'] = AwsLimit('Subnets per Subnet Group',\n self, 20, self.warning_threshold, self.critical_threshold,\n limit_type='AWS::RDS::DBSubnetGroup', ta_limit_name=\n 'Subnets per subnet group')\n limits['Option Groups'] = AwsLimit('Option Groups', self, 20, self.\n warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBOptionGroup')\n limits['Event Subscriptions'] = AwsLimit('Event Subscriptions', self, \n 20, self.warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBEventSubscription', ta_limit_name='Event subscriptions')\n limits['Read replicas per master'] = AwsLimit('Read replicas per master',\n self, 5, self.warning_threshold, self.critical_threshold,\n limit_type='AWS::RDS::DBInstance')\n limits['Max auths per security group'] = AwsLimit(\n 'Max auths per security group', self, 20, self.warning_threshold,\n self.critical_threshold, limit_type='AWS::RDS::DBSecurityGroup',\n limit_subtype='AWS::RDS::DBSecurityGroupIngress')\n limits['DB Clusters'] = AwsLimit('DB Clusters', self, 40, self.\n warning_threshold, self.critical_threshold, limit_type=\n 'AWS::RDS::DBCluster', ta_limit_name='Clusters')\n limits['DB Cluster Parameter Groups'] = AwsLimit(\n 'DB Cluster Parameter Groups', self, 50, self.warning_threshold,\n self.critical_threshold, limit_type=\n 'AWS::RDS::DBClusterParameterGroup', ta_limit_name=\n 'Cluster parameter groups')\n self.limits = limits\n return limits\n", "role": "user" }, { "content": "get_limits", "role": "assistant" } ]
def x(self): self.connect() logger.info('Querying RDS DescribeAccountAttributes for limits') lims = self.conn.describe_account_attributes()['AccountQuotas'] for lim in lims: if lim['AccountQuotaName'] not in self.API_NAME_TO_LIMIT: logger.info( 'RDS DescribeAccountAttributes returned unknownlimit: %s (max: %s; used: %s)' , lim['AccountQuotaName'], lim['Max'], lim['Used']) continue lname = self.API_NAME_TO_LIMIT[lim['AccountQuotaName']] self.limits[lname]._set_api_limit(lim['Max']) if len(self.limits[lname].get_current_usage()) < 1: self.limits[lname]._add_current_usage(lim['Used']) logger.debug('Done setting limits from API.')
update_limits_from_api
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.connect()\n logger.info('Querying RDS DescribeAccountAttributes for limits')\n lims = self.conn.describe_account_attributes()['AccountQuotas']\n for lim in lims:\n if lim['AccountQuotaName'] not in self.API_NAME_TO_LIMIT:\n logger.info(\n 'RDS DescribeAccountAttributes returned unknownlimit: %s (max: %s; used: %s)'\n , lim['AccountQuotaName'], lim['Max'], lim['Used'])\n continue\n lname = self.API_NAME_TO_LIMIT[lim['AccountQuotaName']]\n self.limits[lname]._set_api_limit(lim['Max'])\n if len(self.limits[lname].get_current_usage()) < 1:\n self.limits[lname]._add_current_usage(lim['Used'])\n logger.debug('Done setting limits from API.')\n", "role": "user" }, { "content": "update_limits_from_api", "role": "assistant" } ]
def x(self): logger.debug('Checking usage for service %s', self.service_name) self.connect() for lim in self.limits.values(): lim._reset_usage() try: self._find_delivery_streams() except EndpointConnectionError as ex: logger.warning( 'Caught exception when trying to use Firehose (perhaps the Firehose service is not available in this region?): %s' , ex) self._have_usage = True logger.debug('Done checking usage.')
find_usage
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n logger.debug('Checking usage for service %s', self.service_name)\n self.connect()\n for lim in self.limits.values():\n lim._reset_usage()\n try:\n self._find_delivery_streams()\n except EndpointConnectionError as ex:\n logger.warning(\n 'Caught exception when trying to use Firehose (perhaps the Firehose service is not available in this region?): %s'\n , ex)\n self._have_usage = True\n logger.debug('Done checking usage.')\n", "role": "user" }, { "content": "find_usage", "role": "assistant" } ]
def x(self): trail_list = self.conn.describe_trails()['trailList'] trail_count = len(trail_list) if trail_list else 0 for trail in trail_list: data_resource_count = 0 if self.conn._client_config.region_name == trail['HomeRegion']: response = self.conn.get_event_selectors(TrailName=trail['Name']) event_selectors = response['EventSelectors'] for event_selector in event_selectors: data_resource_count += len(event_selector.get( 'DataResources', [])) self.limits['Event Selectors Per Trail']._add_current_usage(len (event_selectors), aws_type= 'AWS::CloudTrail::EventSelector', resource_id=trail['Name']) self.limits['Data Resources Per Trail']._add_current_usage( data_resource_count, aws_type= 'AWS::CloudTrail::DataResource', resource_id=trail['Name']) else: logger.debug( 'Ignoring event selectors and data resources for CloudTrail %s in non-home region' % trail['Name']) self.limits['Trails Per Region']._add_current_usage(trail_count, aws_type=self.aws_type)
find_usage_cloudtrail
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n trail_list = self.conn.describe_trails()['trailList']\n trail_count = len(trail_list) if trail_list else 0\n for trail in trail_list:\n data_resource_count = 0\n if self.conn._client_config.region_name == trail['HomeRegion']:\n response = self.conn.get_event_selectors(TrailName=trail['Name'])\n event_selectors = response['EventSelectors']\n for event_selector in event_selectors:\n data_resource_count += len(event_selector.get(\n 'DataResources', []))\n self.limits['Event Selectors Per Trail']._add_current_usage(len\n (event_selectors), aws_type=\n 'AWS::CloudTrail::EventSelector', resource_id=trail['Name'])\n self.limits['Data Resources Per Trail']._add_current_usage(\n data_resource_count, aws_type=\n 'AWS::CloudTrail::DataResource', resource_id=trail['Name'])\n else:\n logger.debug(\n 'Ignoring event selectors and data resources for CloudTrail %s in non-home region'\n % trail['Name'])\n self.limits['Trails Per Region']._add_current_usage(trail_count,\n aws_type=self.aws_type)\n", "role": "user" }, { "content": "find_usage_cloudtrail", "role": "assistant" } ]
def x(self): logger.debug("Gathering %s's limits from AWS", self.service_name) if self.limits: return self.limits limits = {} limits['Trails Per Region'] = AwsLimit('Trails Per Region', self, 5, self.warning_threshold, self.critical_threshold, limit_type=self. aws_type) limits['Event Selectors Per Trail'] = AwsLimit('Event Selectors Per Trail', self, 5, self.warning_threshold, self.critical_threshold, limit_type=self.aws_type, limit_subtype= 'AWS::CloudTrail::EventSelector') limits['Data Resources Per Trail'] = AwsLimit('Data Resources Per Trail', self, 250, self.warning_threshold, self.critical_threshold, limit_type=self.aws_type, limit_subtype='AWS::CloudTrail::DataResource' ) self.limits = limits return limits
get_limits
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n logger.debug(\"Gathering %s's limits from AWS\", self.service_name)\n if self.limits:\n return self.limits\n limits = {}\n limits['Trails Per Region'] = AwsLimit('Trails Per Region', self, 5,\n self.warning_threshold, self.critical_threshold, limit_type=self.\n aws_type)\n limits['Event Selectors Per Trail'] = AwsLimit('Event Selectors Per Trail',\n self, 5, self.warning_threshold, self.critical_threshold,\n limit_type=self.aws_type, limit_subtype=\n 'AWS::CloudTrail::EventSelector')\n limits['Data Resources Per Trail'] = AwsLimit('Data Resources Per Trail',\n self, 250, self.warning_threshold, self.critical_threshold,\n limit_type=self.aws_type, limit_subtype='AWS::CloudTrail::DataResource'\n )\n self.limits = limits\n return limits\n", "role": "user" }, { "content": "get_limits", "role": "assistant" } ]
def x(self, ranges=None): fig = self.generate_plot(self.keys[-1], ranges) self.drawn = True return fig
initialize_plot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ranges=None):\n fig = self.generate_plot(self.keys[-1], ranges)\n self.drawn = True\n return fig\n", "role": "user" }, { "content": "initialize_plot", "role": "assistant" } ]
def x(self, element): dims = element.dimensions()[:3] pad = [None] * max(3 - len(dims), 0) return dims + pad
get_axis_dims
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, element):\n dims = element.dimensions()[:3]\n pad = [None] * max(3 - len(dims), 0)\n return dims + pad\n", "role": "user" }, { "content": "get_axis_dims", "role": "assistant" } ]
def x(self, key, ranges=None, element=None): self.generate_plot(key, ranges, element)
update_frame
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, ranges=None, element=None):\n self.generate_plot(key, ranges, element)\n", "role": "user" }, { "content": "update_frame", "role": "assistant" } ]
def x(self, ranges=None): return self.generate_plot(list(self.hmap.data.keys())[0], ranges)
initialize_plot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ranges=None):\n return self.generate_plot(list(self.hmap.data.keys())[0], ranges)\n", "role": "user" }, { "content": "initialize_plot", "role": "assistant" } ]
def x(self): if 'holoviews' not in sys.modules: raise ImportError('HoloViews does not seem to be imported') matches = [k for k, v in get_ipython().user_ns.items() if not k. startswith('_') and v is sys.modules['holoviews']] if len(matches) == 0: raise Exception('Could not find holoviews module in namespace') return '%s.archive' % matches[0]
get_namespace
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if 'holoviews' not in sys.modules:\n raise ImportError('HoloViews does not seem to be imported')\n matches = [k for k, v in get_ipython().user_ns.items() if not k.\n startswith('_') and v is sys.modules['holoviews']]\n if len(matches) == 0:\n raise Exception('Could not find holoviews module in namespace')\n return '%s.archive' % matches[0]\n", "role": "user" }, { "content": "get_namespace", "role": "assistant" } ]
def x(self): if self.export_success is True: print('The last call to holoviews.archive.export was successful.') return elif self.export_success is None: print( """Status of the last call to holoviews.archive.export is unknown. (Re-execute this method once kernel status is idle.)""" ) return print('The last call to holoviews.archive.export was unsuccessful.') if self.traceback is None: print('\n<No traceback captured>') else: print('\n' + self.traceback)
last_export_status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.export_success is True:\n print('The last call to holoviews.archive.export was successful.')\n return\n elif self.export_success is None:\n print(\n \"\"\"Status of the last call to holoviews.archive.export is unknown.\n(Re-execute this method once kernel status is idle.)\"\"\"\n )\n return\n print('The last call to holoviews.archive.export was unsuccessful.')\n if self.traceback is None:\n print('\\n<No traceback captured>')\n else:\n print('\\n' + self.traceback)\n", "role": "user" }, { "content": "last_export_status", "role": "assistant" } ]
def x(self, enabled=True, **kwargs): self.namespace = self.get_namespace() self.notebook_name = '{notebook}' self._timestamp = tuple(time.localtime()) kernel = 'var kernel = IPython.notebook.kernel; ' nbname = 'var nbname = IPython.notebook.get_notebook_name(); ' nbcmd = ( 'var name_cmd = \'%s.notebook_name = \\"\' + nbname + \'\\"\'; ' % self.namespace) cmd = kernel + nbname + nbcmd + 'kernel.execute(name_cmd); ' display(Javascript(cmd)) time.sleep(0.5) self._auto = enabled self.param.set_param(**kwargs) tstamp = time.strftime(' [%Y-%m-%d %H:%M:%S]', self._timestamp) print('Automatic capture is now %s.%s' % ('enabled' if enabled else 'disabled', tstamp if enabled else ''))
auto
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, enabled=True, **kwargs):\n self.namespace = self.get_namespace()\n self.notebook_name = '{notebook}'\n self._timestamp = tuple(time.localtime())\n kernel = 'var kernel = IPython.notebook.kernel; '\n nbname = 'var nbname = IPython.notebook.get_notebook_name(); '\n nbcmd = (\n 'var name_cmd = \\'%s.notebook_name = \\\\\"\\' + nbname + \\'\\\\\"\\'; ' %\n self.namespace)\n cmd = kernel + nbname + nbcmd + 'kernel.execute(name_cmd); '\n display(Javascript(cmd))\n time.sleep(0.5)\n self._auto = enabled\n self.param.set_param(**kwargs)\n tstamp = time.strftime(' [%Y-%m-%d %H:%M:%S]', self._timestamp)\n print('Automatic capture is now %s.%s' % ('enabled' if enabled else\n 'disabled', tstamp if enabled else ''))\n", "role": "user" }, { "content": "auto", "role": "assistant" } ]
def x(self, timestamp=None): if self._timestamp is None: raise Exception('No timestamp set. Has the archive been initialized?') if self.skip_notebook_export: super(NotebookArchive, self).export(timestamp=self._timestamp, info ={'notebook': self.notebook_name}) return self.export_success = None name = self.get_namespace() capture_cmd = ('var capture = \'%s._notebook_data=r\\"\\"\\"\'' % name + '+json_string+\'\\"\\"\\"\'; ') cmd = ('var kernel = IPython.notebook.kernel; ' + 'var json_data = IPython.notebook.toJSON(); ' + 'var json_string = JSON.stringify(json_data); ' + capture_cmd + "var pycmd = capture + ';%s._export_with_html()'; " % name + 'kernel.execute(pycmd)') tstamp = time.strftime(self.timestamp_format, self._timestamp) export_name = self._format(self.export_name, {'timestamp': tstamp, 'notebook': self.notebook_name}) print('Export name: %r\nDirectory %r' % (export_name, os.path.join( os.path.abspath(self.root))) + """ If no output appears, please check holoviews.archive.last_export_status()""" ) display(Javascript(cmd))
export
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, timestamp=None):\n if self._timestamp is None:\n raise Exception('No timestamp set. Has the archive been initialized?')\n if self.skip_notebook_export:\n super(NotebookArchive, self).export(timestamp=self._timestamp, info\n ={'notebook': self.notebook_name})\n return\n self.export_success = None\n name = self.get_namespace()\n capture_cmd = ('var capture = \\'%s._notebook_data=r\\\\\"\\\\\"\\\\\"\\'' % name +\n '+json_string+\\'\\\\\"\\\\\"\\\\\"\\'; ')\n cmd = ('var kernel = IPython.notebook.kernel; ' +\n 'var json_data = IPython.notebook.toJSON(); ' +\n 'var json_string = JSON.stringify(json_data); ' + capture_cmd + \n \"var pycmd = capture + ';%s._export_with_html()'; \" % name +\n 'kernel.execute(pycmd)')\n tstamp = time.strftime(self.timestamp_format, self._timestamp)\n export_name = self._format(self.export_name, {'timestamp': tstamp,\n 'notebook': self.notebook_name})\n print('Export name: %r\\nDirectory %r' % (export_name, os.path.join(\n os.path.abspath(self.root))) +\n \"\"\"\n\nIf no output appears, please check holoviews.archive.last_export_status()\"\"\"\n )\n display(Javascript(cmd))\n", "role": "user" }, { "content": "export", "role": "assistant" } ]
def x(self, obj=None, filename=None, data=None, info={}, html=None): initial_last_key = list(self._files.keys())[-1] if len(self) else None if self._auto: exporters = self.exporters[:] for exporter in exporters: self.exporters = [exporter] super(NotebookArchive, self).add(obj, filename, data, info=dict (info, notebook=self.notebook_name)) new_last_key = list(self._files.keys())[-1] if len(self) else None if new_last_key != initial_last_key: self._replacements[new_last_key] = html self.exporters = exporters
add
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, obj=None, filename=None, data=None, info={}, html=None):\n initial_last_key = list(self._files.keys())[-1] if len(self) else None\n if self._auto:\n exporters = self.exporters[:]\n for exporter in exporters:\n self.exporters = [exporter]\n super(NotebookArchive, self).add(obj, filename, data, info=dict\n (info, notebook=self.notebook_name))\n new_last_key = list(self._files.keys())[-1] if len(self) else None\n if new_last_key != initial_last_key:\n self._replacements[new_last_key] = html\n self.exporters = exporters\n", "role": "user" }, { "content": "add", "role": "assistant" } ]
def x(self): self.export_success = False try: tstamp = time.strftime(self.timestamp_format, self._timestamp) substitutions = {} for (basename, ext), entry in self._files.items(): _, info = entry html_key = self._replacements.get((basename, ext), None) if html_key is None: continue filename = self._format(basename, {'timestamp': tstamp, 'notebook': self.notebook_name}) fpath = filename + ('.%s' % ext if ext else '') info = {'src': fpath, 'mime_type': info['mime_type']} if 'mime_type' not in info: pass elif info['mime_type'] not in self._tags: pass else: basename, ext = os.path.splitext(fpath) truncated = self._truncate_name(basename, ext[1:]) link_html = self._format(self._tags[info['mime_type']], { 'src': truncated, 'mime_type': info['mime_type'], 'css': ''}) substitutions[html_key] = link_html, truncated node = self._get_notebook_node() html = self._generate_html(node, substitutions) export_filename = self.snapshot_name super(NotebookArchive, self).add(filename=export_filename, data= html, info={'file-ext': 'html', 'mime_type': 'text/html', 'notebook': self.notebook_name}) cleared = self._clear_notebook(node) super(NotebookArchive, self).add(filename=export_filename, data= cleared, info={'file-ext': 'ipynb', 'mime_type': 'text/json', 'notebook': self.notebook_name}) super(NotebookArchive, self).export(timestamp=self._timestamp, info ={'notebook': self.notebook_name}) except: self.traceback = traceback.format_exc() else: self.export_success = True
export_with_html
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.export_success = False\n try:\n tstamp = time.strftime(self.timestamp_format, self._timestamp)\n substitutions = {}\n for (basename, ext), entry in self._files.items():\n _, info = entry\n html_key = self._replacements.get((basename, ext), None)\n if html_key is None:\n continue\n filename = self._format(basename, {'timestamp': tstamp,\n 'notebook': self.notebook_name})\n fpath = filename + ('.%s' % ext if ext else '')\n info = {'src': fpath, 'mime_type': info['mime_type']}\n if 'mime_type' not in info:\n pass\n elif info['mime_type'] not in self._tags:\n pass\n else:\n basename, ext = os.path.splitext(fpath)\n truncated = self._truncate_name(basename, ext[1:])\n link_html = self._format(self._tags[info['mime_type']], {\n 'src': truncated, 'mime_type': info['mime_type'], 'css':\n ''})\n substitutions[html_key] = link_html, truncated\n node = self._get_notebook_node()\n html = self._generate_html(node, substitutions)\n export_filename = self.snapshot_name\n super(NotebookArchive, self).add(filename=export_filename, data=\n html, info={'file-ext': 'html', 'mime_type': 'text/html',\n 'notebook': self.notebook_name})\n cleared = self._clear_notebook(node)\n super(NotebookArchive, self).add(filename=export_filename, data=\n cleared, info={'file-ext': 'ipynb', 'mime_type': 'text/json',\n 'notebook': self.notebook_name})\n super(NotebookArchive, self).export(timestamp=self._timestamp, info\n ={'notebook': self.notebook_name})\n except:\n self.traceback = traceback.format_exc()\n else:\n self.export_success = True\n", "role": "user" }, { "content": "export_with_html", "role": "assistant" } ]
def x(self): size = len(self._notebook_data) if size == 0: raise Exception('Captured buffer size for notebook node is zero.') node = reader.reads(self._notebook_data) self.nbversion = reader.get_version(node) return node
get_notebook_node
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n size = len(self._notebook_data)\n if size == 0:\n raise Exception('Captured buffer size for notebook node is zero.')\n node = reader.reads(self._notebook_data)\n self.nbversion = reader.get_version(node)\n return node\n", "role": "user" }, { "content": "get_notebook_node", "role": "assistant" } ]
def x(properties): new_properties = {} for k, v in properties.items(): if isinstance(v, dict): new_properties[k] = v elif k == 's': new_properties['size'] = v elif k == 'marker': new_properties.update(markers.get(v, {'marker': v})) elif (k == 'color' or k.endswith('_color')) and not isinstance(v, ( dict, dim)): with abbreviated_exception(): v = COLOR_ALIASES.get(v, v) if isinstance(v, tuple): with abbreviated_exception(): v = rgb2hex(v) new_properties[k] = v else: new_properties[k] = v new_properties.pop('cmap', None) return new_properties
mpl_to_bokeh
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(properties):\n new_properties = {}\n for k, v in properties.items():\n if isinstance(v, dict):\n new_properties[k] = v\n elif k == 's':\n new_properties['size'] = v\n elif k == 'marker':\n new_properties.update(markers.get(v, {'marker': v}))\n elif (k == 'color' or k.endswith('_color')) and not isinstance(v, (\n dict, dim)):\n with abbreviated_exception():\n v = COLOR_ALIASES.get(v, v)\n if isinstance(v, tuple):\n with abbreviated_exception():\n v = rgb2hex(v)\n new_properties[k] = v\n else:\n new_properties[k] = v\n new_properties.pop('cmap', None)\n return new_properties\n", "role": "user" }, { "content": "mpl_to_bokeh", "role": "assistant" } ]
def x(style, value, scalar=False): validator = get_validator(style) if validator is None: return None if isinstance(value, (np.ndarray, list)): if scalar: return False return all(validator(v) for v in value) return validator(value)
validate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(style, value, scalar=False):\n validator = get_validator(style)\n if validator is None:\n return None\n if isinstance(value, (np.ndarray, list)):\n if scalar:\n return False\n return all(validator(v) for v in value)\n return validator(value)\n", "role": "user" }, { "content": "validate", "role": "assistant" } ]
def x(rgba): if isinstance(rgba, tuple): return tuple(int(c * 255) if i < 3 else c for i, c in enumerate(rgba)) else: return COLOR_ALIASES.get(rgba, rgba)
rgba_tuple
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(rgba):\n if isinstance(rgba, tuple):\n return tuple(int(c * 255) if i < 3 else c for i, c in enumerate(rgba))\n else:\n return COLOR_ALIASES.get(rgba, rgba)\n", "role": "user" }, { "content": "rgba_tuple", "role": "assistant" } ]
def x(style, opts, mapping, nvals): opts = sorted(opts, key=lambda x: x in ['color', 'alpha']) applied_styles = set(mapping) style_data, style_mapping = {}, {} for opt in opts: if 'color' in opt: alias = 'color' elif 'alpha' in opt: alias = 'alpha' else: alias = None if opt not in style or opt in mapping: continue elif opt == alias: if alias in applied_styles: continue elif 'line_' + alias in applied_styles: if 'fill_' + alias not in opts: continue opt = 'fill_' + alias val = style[alias] elif 'fill_' + alias in applied_styles: opt = 'line_' + alias val = style[alias] else: val = style[alias] else: val = style[opt] style_mapping[opt] = {'field': opt} applied_styles.add(opt) if 'color' in opt and isinstance(val, tuple): val = rgb2hex(val) style_data[opt] = [val] * nvals return style_data, style_mapping
expand_batched_style
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(style, opts, mapping, nvals):\n opts = sorted(opts, key=lambda x: x in ['color', 'alpha'])\n applied_styles = set(mapping)\n style_data, style_mapping = {}, {}\n for opt in opts:\n if 'color' in opt:\n alias = 'color'\n elif 'alpha' in opt:\n alias = 'alpha'\n else:\n alias = None\n if opt not in style or opt in mapping:\n continue\n elif opt == alias:\n if alias in applied_styles:\n continue\n elif 'line_' + alias in applied_styles:\n if 'fill_' + alias not in opts:\n continue\n opt = 'fill_' + alias\n val = style[alias]\n elif 'fill_' + alias in applied_styles:\n opt = 'line_' + alias\n val = style[alias]\n else:\n val = style[alias]\n else:\n val = style[opt]\n style_mapping[opt] = {'field': opt}\n applied_styles.add(opt)\n if 'color' in opt and isinstance(val, tuple):\n val = rgb2hex(val)\n style_data[opt] = [val] * nvals\n return style_data, style_mapping\n", "role": "user" }, { "content": "expand_batched_style", "role": "assistant" } ]
def x(cls, first, second, msg=None): if not first == second: standardMsg = '%s != %s' % (safe_repr(first), safe_repr(second)) raise cls.failureException(msg or standardMsg)
simple_equality
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, first, second, msg=None):\n if not first == second:\n standardMsg = '%s != %s' % (safe_repr(first), safe_repr(second))\n raise cls.failureException(msg or standardMsg)\n", "role": "user" }, { "content": "simple_equality", "role": "assistant" } ]
def x(cls, first, second, msg=None): asserter = None if type(first) is type(second): asserter = cls.equality_type_funcs.get(type(first)) try: basestring = basestring except NameError: basestring = str if asserter is not None: if isinstance(asserter, basestring): asserter = getattr(cls, asserter) if asserter is None: asserter = cls.simple_equality if msg is None: asserter(first, second) else: asserter(first, second, msg=msg)
assertEqual
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, first, second, msg=None):\n asserter = None\n if type(first) is type(second):\n asserter = cls.equality_type_funcs.get(type(first))\n try:\n basestring = basestring\n except NameError:\n basestring = str\n if asserter is not None:\n if isinstance(asserter, basestring):\n asserter = getattr(cls, asserter)\n if asserter is None:\n asserter = cls.simple_equality\n if msg is None:\n asserter(first, second)\n else:\n asserter(first, second, msg=msg)\n", "role": "user" }, { "content": "assertEqual", "role": "assistant" } ]
def x(self, plot, mapping, properties): properties = mpl_to_bokeh(properties) properties = dict(properties, **mapping) if 'xs' in mapping: renderer = plot.patches(**properties) else: renderer = plot.quad(**properties) if self.colorbar and 'color_mapper' in self.handles: self._draw_colorbar(plot, self.handles['color_mapper']) return renderer, renderer.glyph
init_glyph
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, plot, mapping, properties):\n properties = mpl_to_bokeh(properties)\n properties = dict(properties, **mapping)\n if 'xs' in mapping:\n renderer = plot.patches(**properties)\n else:\n renderer = plot.quad(**properties)\n if self.colorbar and 'color_mapper' in self.handles:\n self._draw_colorbar(plot, self.handles['color_mapper'])\n return renderer, renderer.glyph\n", "role": "user" }, { "content": "init_glyph", "role": "assistant" } ]
def x(self, x, y): left, bottom, right, top = self.aarect().lbrt() return left <= x <= right and bottom <= y <= top
contains
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, x, y):\n left, bottom, right, top = self.aarect().lbrt()\n return left <= x <= right and bottom <= y <= top\n", "role": "user" }, { "content": "contains", "role": "assistant" } ]
def x(self, x, y): left, bottom, right, top = self._aarect.lbrt() return left <= x < right and bottom < y <= top
contains_exclusive
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, x, y):\n left, bottom, right, top = self._aarect.lbrt()\n return left <= x < right and bottom < y <= top\n", "role": "user" }, { "content": "contains_exclusive", "role": "assistant" } ]
def x(self, x): left, bottom, right, top = self.aarect().lbrt() leftx, bottomx, rightx, topx = x.aarect().lbrt() return (left <= leftx and bottom <= bottomx and right >= rightx and top >= topx and not (left == leftx and bottom == bottomx and right == rightx and top == topx))
containsbb_exclusive
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, x):\n left, bottom, right, top = self.aarect().lbrt()\n leftx, bottomx, rightx, topx = x.aarect().lbrt()\n return (left <= leftx and bottom <= bottomx and right >= rightx and top >=\n topx and not (left == leftx and bottom == bottomx and right ==\n rightx and top == topx))\n", "role": "user" }, { "content": "containsbb_exclusive", "role": "assistant" } ]
def x(self): return self._left, self._bottom, self._right, self._top
lbrt
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return self._left, self._bottom, self._right, self._top\n", "role": "user" }, { "content": "lbrt", "role": "assistant" } ]
def x(self): left, bottom, right, top = self.lbrt() return (right + left) / 2.0, (top + bottom) / 2.0
centroid
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n left, bottom, right, top = self.lbrt()\n return (right + left) / 2.0, (top + bottom) / 2.0\n", "role": "user" }, { "content": "centroid", "role": "assistant" } ]
def x(dmap): if not isinstance(dmap, DynamicMap): return [] dmaps = [dmap] for o in dmap.callback.inputs: dmaps.extend(get_nested_dmaps(o)) return list(set(dmaps))
get_nested_dmaps
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dmap):\n if not isinstance(dmap, DynamicMap):\n return []\n dmaps = [dmap]\n for o in dmap.callback.inputs:\n dmaps.extend(get_nested_dmaps(o))\n return list(set(dmaps))\n", "role": "user" }, { "content": "get_nested_dmaps", "role": "assistant" } ]
def x(dmap): return list({s for dmap in get_nested_dmaps(dmap) for s in dmap.streams})
get_nested_streams
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dmap):\n return list({s for dmap in get_nested_dmaps(dmap) for s in dmap.streams})\n", "role": "user" }, { "content": "get_nested_streams", "role": "assistant" } ]
def x(callable_obj, streams): memoization_state = bool(callable_obj._stream_memoization) callable_obj._stream_memoization &= not any(s.transient and s. _triggering for s in streams) try: yield except: raise finally: callable_obj._stream_memoization = memoization_state
dynamicmap_memoization
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(callable_obj, streams):\n memoization_state = bool(callable_obj._stream_memoization)\n callable_obj._stream_memoization &= not any(s.transient and s.\n _triggering for s in streams)\n try:\n yield\n except:\n raise\n finally:\n callable_obj._stream_memoization = memoization_state\n", "role": "user" }, { "content": "dynamicmap_memoization", "role": "assistant" } ]
def x(self, dimensions=None, **kwargs): dimensions = self._valid_dimensions(dimensions) if len(dimensions) == self.ndims: with item_check(False): return NdOverlay(self, **kwargs).reindex(dimensions) else: dims = [d for d in self.kdims if d not in dimensions] return self.groupby(dims, group_type=NdOverlay, **kwargs)
overlay
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimensions=None, **kwargs):\n dimensions = self._valid_dimensions(dimensions)\n if len(dimensions) == self.ndims:\n with item_check(False):\n return NdOverlay(self, **kwargs).reindex(dimensions)\n else:\n dims = [d for d in self.kdims if d not in dimensions]\n return self.groupby(dims, group_type=NdOverlay, **kwargs)\n", "role": "user" }, { "content": "overlay", "role": "assistant" } ]
def x(self, dimensions=None, **kwargs): dimensions = self._valid_dimensions(dimensions) if len(dimensions) == self.ndims: with item_check(False): return GridSpace(self, **kwargs).reindex(dimensions) return self.groupby(dimensions, container_type=GridSpace, **kwargs)
grid
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimensions=None, **kwargs):\n dimensions = self._valid_dimensions(dimensions)\n if len(dimensions) == self.ndims:\n with item_check(False):\n return GridSpace(self, **kwargs).reindex(dimensions)\n return self.groupby(dimensions, container_type=GridSpace, **kwargs)\n", "role": "user" }, { "content": "grid", "role": "assistant" } ]
def x(self, dimensions=None, **kwargs): dimensions = self._valid_dimensions(dimensions) if len(dimensions) == self.ndims: with item_check(False): return NdLayout(self, **kwargs).reindex(dimensions) return self.groupby(dimensions, container_type=NdLayout, **kwargs)
layout
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimensions=None, **kwargs):\n dimensions = self._valid_dimensions(dimensions)\n if len(dimensions) == self.ndims:\n with item_check(False):\n return NdLayout(self, **kwargs).reindex(dimensions)\n return self.groupby(dimensions, container_type=NdLayout, **kwargs)\n", "role": "user" }, { "content": "layout", "role": "assistant" } ]
def x(self, *args, **kwargs): data = OrderedDict([(k, v.options(*args, **kwargs)) for k, v in self. data.items()]) return self.clone(data)
options
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args, **kwargs):\n data = OrderedDict([(k, v.options(*args, **kwargs)) for k, v in self.\n data.items()])\n return self.clone(data)\n", "role": "user" }, { "content": "options", "role": "assistant" } ]
def x(self): if util.config.future_deprecations: self.param.warning( 'split_overlays is deprecated and is now a private method.') return self._split_overlays()
split_overlays
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if util.config.future_deprecations:\n self.param.warning(\n 'split_overlays is deprecated and is now a private method.')\n return self._split_overlays()\n", "role": "user" }, { "content": "split_overlays", "role": "assistant" } ]
def x(self): if not issubclass(self.type, CompositeOverlay): return None, self.clone() item_maps = OrderedDict() for k, overlay in self.data.items(): for key, el in overlay.items(): if key not in item_maps: item_maps[key] = [(k, el)] else: item_maps[key].append((k, el)) maps, keys = [], [] for k, layermap in item_maps.items(): maps.append(self.clone(layermap)) keys.append(k) return keys, maps
split_overlays
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not issubclass(self.type, CompositeOverlay):\n return None, self.clone()\n item_maps = OrderedDict()\n for k, overlay in self.data.items():\n for key, el in overlay.items():\n if key not in item_maps:\n item_maps[key] = [(k, el)]\n else:\n item_maps[key].append((k, el))\n maps, keys = [], []\n for k, layermap in item_maps.items():\n maps.append(self.clone(layermap))\n keys.append(k)\n return keys, maps\n", "role": "user" }, { "content": "split_overlays", "role": "assistant" } ]
def x(self): return [tuple(zip([d.name for d in self.kdims], [k] if self.ndims == 1 else k)) for k in self.keys()]
dimension_keys
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return [tuple(zip([d.name for d in self.kdims], [k] if self.ndims == 1 else\n k)) for k in self.keys()]\n", "role": "user" }, { "content": "dimension_keys", "role": "assistant" } ]
def x(self, dimensions, other, keys): if not isinstance(self, DynamicMap) or not isinstance(other, DynamicMap): keys = sorted((d, v) for k in keys for d, v in k) grouped = dict([(g, [v for _, v in group]) for g, group in groupby( keys, lambda x: x[0])]) dimensions = [d(values=grouped[d.name]) for d in dimensions] map_obj = None map_obj = self if isinstance(self, DynamicMap) else other if isinstance(self, DynamicMap) and isinstance(other, DynamicMap): self_streams = util.dimensioned_streams(self) other_streams = util.dimensioned_streams(other) streams = list(util.unique_iterator(self_streams + other_streams)) else: streams = map_obj.streams def x(*key, **kwargs): key_map = {d.name: k for d, k in zip(dimensions, key)} layers = [] try: self_el = self.select(HoloMap, **key_map) if self.kdims else self[( )] layers.append(self_el) except KeyError: pass try: other_el = other.select(HoloMap, **key_map ) if other.kdims else other[()] layers.append(other_el) except KeyError: pass return Overlay(layers) callback = Callable(dynamic_mul, inputs=[self, other]) callback._is_overlay = True if map_obj: return map_obj.clone(callback=callback, shared_data=False, kdims= dimensions, streams=streams) else: return DynamicMap(callback=callback, kdims=dimensions, streams=streams)
dynamic_mul
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimensions, other, keys):\n if not isinstance(self, DynamicMap) or not isinstance(other, DynamicMap):\n keys = sorted((d, v) for k in keys for d, v in k)\n grouped = dict([(g, [v for _, v in group]) for g, group in groupby(\n keys, lambda x: x[0])])\n dimensions = [d(values=grouped[d.name]) for d in dimensions]\n map_obj = None\n map_obj = self if isinstance(self, DynamicMap) else other\n if isinstance(self, DynamicMap) and isinstance(other, DynamicMap):\n self_streams = util.dimensioned_streams(self)\n other_streams = util.dimensioned_streams(other)\n streams = list(util.unique_iterator(self_streams + other_streams))\n else:\n streams = map_obj.streams\n\n def x(*key, **kwargs):\n key_map = {d.name: k for d, k in zip(dimensions, key)}\n layers = []\n try:\n self_el = self.select(HoloMap, **key_map) if self.kdims else self[(\n )]\n layers.append(self_el)\n except KeyError:\n pass\n try:\n other_el = other.select(HoloMap, **key_map\n ) if other.kdims else other[()]\n layers.append(other_el)\n except KeyError:\n pass\n return Overlay(layers)\n callback = Callable(dynamic_mul, inputs=[self, other])\n callback._is_overlay = True\n if map_obj:\n return map_obj.clone(callback=callback, shared_data=False, kdims=\n dimensions, streams=streams)\n else:\n return DynamicMap(callback=callback, kdims=dimensions, streams=streams)\n", "role": "user" }, { "content": "dynamic_mul", "role": "assistant" } ]
def x(self, merge_type=None, drop=[], drop_constant=False): from .element import Collator merge_type = merge_type if merge_type else self.__class__ return Collator(self, merge_type=merge_type, drop=drop, drop_constant= drop_constant)()
collate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, merge_type=None, drop=[], drop_constant=False):\n from .element import Collator\n merge_type = merge_type if merge_type else self.__class__\n return Collator(self, merge_type=merge_type, drop=drop, drop_constant=\n drop_constant)()\n", "role": "user" }, { "content": "collate", "role": "assistant" } ]
def x(self, dimensions=None, function=None, spreadfn=None, **kwargs): from .data import concat if not dimensions: dimensions = self.kdims if not isinstance(dimensions, list): dimensions = [dimensions] if self.ndims > 1 and len(dimensions) != self.ndims: groups = self.groupby([dim for dim in self.kdims if dim not in dimensions]) elif all(d in self.kdims for d in dimensions): groups = HoloMap([(0, self)]) else: raise KeyError('Supplied dimensions not found.') collapsed = groups.clone(shared_data=False) for key, group in groups.items(): if hasattr(group.last, 'interface'): group_data = concat(group) if function: agg = group_data.aggregate(group.last.kdims, function, spreadfn, **kwargs) group_data = group.type(agg) else: group_data = [el.data for el in group] args = group_data, function, group.last.kdims data = group.type.collapse_data(*args, **kwargs) group_data = group.last.clone(data) collapsed[key] = group_data return collapsed if self.ndims - len(dimensions) else collapsed.last
collapse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimensions=None, function=None, spreadfn=None, **kwargs):\n from .data import concat\n if not dimensions:\n dimensions = self.kdims\n if not isinstance(dimensions, list):\n dimensions = [dimensions]\n if self.ndims > 1 and len(dimensions) != self.ndims:\n groups = self.groupby([dim for dim in self.kdims if dim not in\n dimensions])\n elif all(d in self.kdims for d in dimensions):\n groups = HoloMap([(0, self)])\n else:\n raise KeyError('Supplied dimensions not found.')\n collapsed = groups.clone(shared_data=False)\n for key, group in groups.items():\n if hasattr(group.last, 'interface'):\n group_data = concat(group)\n if function:\n agg = group_data.aggregate(group.last.kdims, function,\n spreadfn, **kwargs)\n group_data = group.type(agg)\n else:\n group_data = [el.data for el in group]\n args = group_data, function, group.last.kdims\n data = group.type.collapse_data(*args, **kwargs)\n group_data = group.last.clone(data)\n collapsed[key] = group_data\n return collapsed if self.ndims - len(dimensions) else collapsed.last\n", "role": "user" }, { "content": "collapse", "role": "assistant" } ]
def x(self, samples=[], bounds=None, **sample_values): if util.config.future_deprecations: self.param.warning( 'The HoloMap.sample method is deprecated, for equivalent functionality use HoloMap.apply.sample().collapse().' ) dims = self.last.ndims if isinstance(samples, tuple) or np.isscalar(samples): if dims == 1: xlim = self.last.range(0) lower, upper = (xlim[0], xlim[1]) if bounds is None else bounds edges = np.linspace(lower, upper, samples + 1) linsamples = [((l + u) / 2.0) for l, u in zip(edges[:-1], edges [1:])] elif dims == 2: rows, cols = samples if bounds: l, b, r, t = bounds else: l, r = self.last.range(0) b, t = self.last.range(1) xedges = np.linspace(l, r, cols + 1) yedges = np.linspace(b, t, rows + 1) xsamples = [((lx + ux) / 2.0) for lx, ux in zip(xedges[:-1], xedges[1:])] ysamples = [((ly + uy) / 2.0) for ly, uy in zip(yedges[:-1], yedges[1:])] Y, X = np.meshgrid(ysamples, xsamples) linsamples = list(zip(X.flat, Y.flat)) else: raise NotImplementedError( 'Regular sampling not implemented for elements with more than two dimensions.' ) samples = list(util.unique_iterator(self.last.closest(linsamples))) sampled = self.clone([(k, view.sample(samples, closest=False, ** sample_values)) for k, view in self.data.items()]) from ..element import Table return Table(sampled.collapse())
sample
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, samples=[], bounds=None, **sample_values):\n if util.config.future_deprecations:\n self.param.warning(\n 'The HoloMap.sample method is deprecated, for equivalent functionality use HoloMap.apply.sample().collapse().'\n )\n dims = self.last.ndims\n if isinstance(samples, tuple) or np.isscalar(samples):\n if dims == 1:\n xlim = self.last.range(0)\n lower, upper = (xlim[0], xlim[1]) if bounds is None else bounds\n edges = np.linspace(lower, upper, samples + 1)\n linsamples = [((l + u) / 2.0) for l, u in zip(edges[:-1], edges\n [1:])]\n elif dims == 2:\n rows, cols = samples\n if bounds:\n l, b, r, t = bounds\n else:\n l, r = self.last.range(0)\n b, t = self.last.range(1)\n xedges = np.linspace(l, r, cols + 1)\n yedges = np.linspace(b, t, rows + 1)\n xsamples = [((lx + ux) / 2.0) for lx, ux in zip(xedges[:-1],\n xedges[1:])]\n ysamples = [((ly + uy) / 2.0) for ly, uy in zip(yedges[:-1],\n yedges[1:])]\n Y, X = np.meshgrid(ysamples, xsamples)\n linsamples = list(zip(X.flat, Y.flat))\n else:\n raise NotImplementedError(\n 'Regular sampling not implemented for elements with more than two dimensions.'\n )\n samples = list(util.unique_iterator(self.last.closest(linsamples)))\n sampled = self.clone([(k, view.sample(samples, closest=False, **\n sample_values)) for k, view in self.data.items()])\n from ..element import Table\n return Table(sampled.collapse())\n", "role": "user" }, { "content": "sample", "role": "assistant" } ]
def x(self, dimensions=None, function=None, spread_fn=None, **reduce_map): if util.config.future_deprecations: self.param.warning( 'The HoloMap.reduce method is deprecated, for equivalent functionality use HoloMap.apply.reduce().collapse().' ) from ..element import Table reduced_items = [(k, v.reduce(dimensions, function, spread_fn, ** reduce_map)) for k, v in self.items()] if not isinstance(reduced_items[0][1], Table): params = dict(util.get_param_values(self.last), kdims=self.kdims, vdims=self.last.vdims) return Table(reduced_items, **params) return Table(self.clone(reduced_items).collapse())
reduce
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimensions=None, function=None, spread_fn=None, **reduce_map):\n if util.config.future_deprecations:\n self.param.warning(\n 'The HoloMap.reduce method is deprecated, for equivalent functionality use HoloMap.apply.reduce().collapse().'\n )\n from ..element import Table\n reduced_items = [(k, v.reduce(dimensions, function, spread_fn, **\n reduce_map)) for k, v in self.items()]\n if not isinstance(reduced_items[0][1], Table):\n params = dict(util.get_param_values(self.last), kdims=self.kdims,\n vdims=self.last.vdims)\n return Table(reduced_items, **params)\n return Table(self.clone(reduced_items).collapse())\n", "role": "user" }, { "content": "reduce", "role": "assistant" } ]
def x(self, label=None, group=None, depth=1): return super(HoloMap, self).relabel(label=label, group=group, depth=depth)
relabel
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, label=None, group=None, depth=1):\n return super(HoloMap, self).relabel(label=label, group=group, depth=depth)\n", "role": "user" }, { "content": "relabel", "role": "assistant" } ]
def x(self, dimension=None, num_bins=20, bin_range=None, adjoin=True, individually=True, **kwargs): if dimension is not None and not isinstance(dimension, list): dimension = [dimension] histmaps = [self.clone(shared_data=False) for _ in dimension or [None]] if individually: map_range = None else: if dimension is None: raise Exception( 'Please supply the dimension to compute a histogram for.') map_range = self.range(kwargs['dimension']) bin_range = map_range if bin_range is None else bin_range style_prefix = 'Custom[<' + self.name + '>]_' if issubclass(self.type, (NdOverlay, Overlay)) and 'index' not in kwargs: kwargs['index'] = 0 for k, v in self.data.items(): hists = v.hist(adjoin=False, dimension=dimension, bin_range= bin_range, num_bins=num_bins, style_prefix=style_prefix, **kwargs) if isinstance(hists, Layout): for i, hist in enumerate(hists): histmaps[i][k] = hist else: histmaps[0][k] = hists if adjoin: layout = self for hist in histmaps: layout = layout << hist if issubclass(self.type, (NdOverlay, Overlay)): layout.main_layer = kwargs['index'] return layout elif len(histmaps) > 1: return Layout(histmaps) else: return histmaps[0]
hist
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimension=None, num_bins=20, bin_range=None, adjoin=True,\n individually=True, **kwargs):\n if dimension is not None and not isinstance(dimension, list):\n dimension = [dimension]\n histmaps = [self.clone(shared_data=False) for _ in dimension or [None]]\n if individually:\n map_range = None\n else:\n if dimension is None:\n raise Exception(\n 'Please supply the dimension to compute a histogram for.')\n map_range = self.range(kwargs['dimension'])\n bin_range = map_range if bin_range is None else bin_range\n style_prefix = 'Custom[<' + self.name + '>]_'\n if issubclass(self.type, (NdOverlay, Overlay)) and 'index' not in kwargs:\n kwargs['index'] = 0\n for k, v in self.data.items():\n hists = v.hist(adjoin=False, dimension=dimension, bin_range=\n bin_range, num_bins=num_bins, style_prefix=style_prefix, **kwargs)\n if isinstance(hists, Layout):\n for i, hist in enumerate(hists):\n histmaps[i][k] = hist\n else:\n histmaps[0][k] = hists\n if adjoin:\n layout = self\n for hist in histmaps:\n layout = layout << hist\n if issubclass(self.type, (NdOverlay, Overlay)):\n layout.main_layer = kwargs['index']\n return layout\n elif len(histmaps) > 1:\n return Layout(histmaps)\n else:\n return histmaps[0]\n", "role": "user" }, { "content": "hist", "role": "assistant" } ]
def x(self): noargs = inspect.ArgSpec(args=[], varargs=None, keywords=None, defaults =None) return self.argspec == noargs
noargs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n noargs = inspect.ArgSpec(args=[], varargs=None, keywords=None, defaults\n =None)\n return self.argspec == noargs\n", "role": "user" }, { "content": "noargs", "role": "assistant" } ]
def x(self, callable=None, **overrides): old = {k: v for k, v in self.get_param_values() if k not in ['callable', 'name']} params = dict(old, **overrides) callable = self.callable if callable is None else callable return self.__class__(callable, **params)
clone
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, callable=None, **overrides):\n old = {k: v for k, v in self.get_param_values() if k not in ['callable',\n 'name']}\n params = dict(old, **overrides)\n callable = self.callable if callable is None else callable\n return self.__class__(callable, **params)\n", "role": "user" }, { "content": "clone", "role": "assistant" } ]
def x(self): unbounded_dims = [] stream_params = set(util.stream_parameters(self.streams)) for kdim in self.kdims: if str(kdim) in stream_params: continue if kdim.values: continue if None in kdim.range: unbounded_dims.append(str(kdim)) return unbounded_dims
unbounded
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n unbounded_dims = []\n stream_params = set(util.stream_parameters(self.streams))\n for kdim in self.kdims:\n if str(kdim) in stream_params:\n continue\n if kdim.values:\n continue\n if None in kdim.range:\n unbounded_dims.append(str(kdim))\n return unbounded_dims\n", "role": "user" }, { "content": "unbounded", "role": "assistant" } ]
def x(self): key = [] undefined = [] stream_params = set(util.stream_parameters(self.streams)) for kdim in self.kdims: if str(kdim) in stream_params: key.append(None) elif kdim.default: key.append(kdim.default) elif kdim.values: if all(util.isnumeric(v) for v in kdim.values): key.append(sorted(kdim.values)[0]) else: key.append(kdim.values[0]) elif kdim.range[0] is not None: key.append(kdim.range[0]) else: undefined.append(kdim) if undefined: msg = ( 'Dimension(s) {undefined_dims} do not specify range or values needed to generate initial key' ) undefined_dims = ', '.join([('%r' % str(dim)) for dim in undefined]) raise KeyError(msg.format(undefined_dims=undefined_dims)) return tuple(key)
initial_key
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n key = []\n undefined = []\n stream_params = set(util.stream_parameters(self.streams))\n for kdim in self.kdims:\n if str(kdim) in stream_params:\n key.append(None)\n elif kdim.default:\n key.append(kdim.default)\n elif kdim.values:\n if all(util.isnumeric(v) for v in kdim.values):\n key.append(sorted(kdim.values)[0])\n else:\n key.append(kdim.values[0])\n elif kdim.range[0] is not None:\n key.append(kdim.range[0])\n else:\n undefined.append(kdim)\n if undefined:\n msg = (\n 'Dimension(s) {undefined_dims} do not specify range or values needed to generate initial key'\n )\n undefined_dims = ', '.join([('%r' % str(dim)) for dim in undefined])\n raise KeyError(msg.format(undefined_dims=undefined_dims))\n return tuple(key)\n", "role": "user" }, { "content": "initial_key", "role": "assistant" } ]
def x(self, key): if key == () and len(self.kdims) == 0: return () key = util.wrap_tuple(key) assert len(key) == len(self.kdims) for ind, val in enumerate(key): kdim = self.kdims[ind] low, high = util.max_range([kdim.range, kdim.soft_range]) if util.is_number(low) and util.isfinite(low): if val < low: raise KeyError('Key value %s below lower bound %s' % (val, low) ) if util.is_number(high) and util.isfinite(high): if val > high: raise KeyError('Key value %s above upper bound %s' % (val, high))
validate_key
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key):\n if key == () and len(self.kdims) == 0:\n return ()\n key = util.wrap_tuple(key)\n assert len(key) == len(self.kdims)\n for ind, val in enumerate(key):\n kdim = self.kdims[ind]\n low, high = util.max_range([kdim.range, kdim.soft_range])\n if util.is_number(low) and util.isfinite(low):\n if val < low:\n raise KeyError('Key value %s below lower bound %s' % (val, low)\n )\n if util.is_number(high) and util.isfinite(high):\n if val > high:\n raise KeyError('Key value %s above upper bound %s' % (val,\n high))\n", "role": "user" }, { "content": "validate_key", "role": "assistant" } ]
def x(self, **kwargs): if self.callback.noargs and self.streams == []: self.param.warning( 'No streams declared. To update a DynamicMaps using generators (or callables without arguments) use streams=[Next()]' ) return if self.streams == []: self.param.warning( 'No streams on DynamicMap, calling event will have no effect') return stream_params = set(util.stream_parameters(self.streams)) invalid = [k for k in kwargs.keys() if k not in stream_params] if invalid: msg = 'Key(s) {invalid} do not correspond to stream parameters' raise KeyError(msg.format(invalid=', '.join('%r' % i for i in invalid)) ) streams = [] for stream in self.streams: contents = stream.contents applicable_kws = {k: v for k, v in kwargs.items() if k in set( contents.keys())} if not applicable_kws and contents: continue streams.append(stream) rkwargs = util.rename_stream_kwargs(stream, applicable_kws, reverse =True) stream.update(**rkwargs) Stream.trigger(streams)
event
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n if self.callback.noargs and self.streams == []:\n self.param.warning(\n 'No streams declared. To update a DynamicMaps using generators (or callables without arguments) use streams=[Next()]'\n )\n return\n if self.streams == []:\n self.param.warning(\n 'No streams on DynamicMap, calling event will have no effect')\n return\n stream_params = set(util.stream_parameters(self.streams))\n invalid = [k for k in kwargs.keys() if k not in stream_params]\n if invalid:\n msg = 'Key(s) {invalid} do not correspond to stream parameters'\n raise KeyError(msg.format(invalid=', '.join('%r' % i for i in invalid))\n )\n streams = []\n for stream in self.streams:\n contents = stream.contents\n applicable_kws = {k: v for k, v in kwargs.items() if k in set(\n contents.keys())}\n if not applicable_kws and contents:\n continue\n streams.append(stream)\n rkwargs = util.rename_stream_kwargs(stream, applicable_kws, reverse\n =True)\n stream.update(**rkwargs)\n Stream.trigger(streams)\n", "role": "user" }, { "content": "event", "role": "assistant" } ]
def x(self, retval): if self.id not in Store.custom_options(): return retval spec = StoreOptions.tree_to_dict(Store.custom_options()[self.id]) return retval.opts(spec)
style
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, retval):\n if self.id not in Store.custom_options():\n return retval\n spec = StoreOptions.tree_to_dict(Store.custom_options()[self.id])\n return retval.opts(spec)\n", "role": "user" }, { "content": "style", "role": "assistant" } ]
def x(self, *args): self._validate_key(args) kdims = [kdim.name for kdim in self.kdims] kwarg_items = [s.contents.items() for s in self.streams] hash_items = tuple(tuple(sorted(s.hashkey.items())) for s in self.streams ) + args flattened = [(k, v) for kws in kwarg_items for k, v in kws if k not in kdims] if self._posarg_keys: kwargs = dict(flattened, **dict(zip(self._posarg_keys, args))) args = () else: kwargs = dict(flattened) if not isinstance(self.callback, Generator): kwargs['_memoization_hash_'] = hash_items with dynamicmap_memoization(self.callback, self.streams): retval = self.callback(*args, **kwargs) return self._style(retval)
execute_callback
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args):\n self._validate_key(args)\n kdims = [kdim.name for kdim in self.kdims]\n kwarg_items = [s.contents.items() for s in self.streams]\n hash_items = tuple(tuple(sorted(s.hashkey.items())) for s in self.streams\n ) + args\n flattened = [(k, v) for kws in kwarg_items for k, v in kws if k not in\n kdims]\n if self._posarg_keys:\n kwargs = dict(flattened, **dict(zip(self._posarg_keys, args)))\n args = ()\n else:\n kwargs = dict(flattened)\n if not isinstance(self.callback, Generator):\n kwargs['_memoization_hash_'] = hash_items\n with dynamicmap_memoization(self.callback, self.streams):\n retval = self.callback(*args, **kwargs)\n return self._style(retval)\n", "role": "user" }, { "content": "execute_callback", "role": "assistant" } ]
def x(self, *args, **kwargs): if 'clone' not in kwargs: kwargs['clone'] = True return self.opts(*args, **kwargs)
options
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args, **kwargs):\n if 'clone' not in kwargs:\n kwargs['clone'] = True\n return self.opts(*args, **kwargs)\n", "role": "user" }, { "content": "options", "role": "assistant" } ]
def x(self, data=None, shared_data=True, new_type=None, link=True, *args, **overrides): if 'link_inputs' in overrides and util.config.future_deprecations: self.param.warning( 'link_inputs argument to the clone method is deprecated, use the more general link argument instead.' ) link = link and overrides.pop('link_inputs', True) callback = overrides.pop('callback', self.callback) if data is None and shared_data: data = self.data if link and callback is self.callback: overrides['plot_id'] = self._plot_id clone = super(UniformNdMapping, self).clone(callback, shared_data, new_type, link, *((data,) + args), **overrides) if clone.callback is self.callback: with util.disable_constant(clone): clone.callback = clone.callback.clone(inputs=[self], link_inputs=link) return clone
clone
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data=None, shared_data=True, new_type=None, link=True, *args,\n **overrides):\n if 'link_inputs' in overrides and util.config.future_deprecations:\n self.param.warning(\n 'link_inputs argument to the clone method is deprecated, use the more general link argument instead.'\n )\n link = link and overrides.pop('link_inputs', True)\n callback = overrides.pop('callback', self.callback)\n if data is None and shared_data:\n data = self.data\n if link and callback is self.callback:\n overrides['plot_id'] = self._plot_id\n clone = super(UniformNdMapping, self).clone(callback, shared_data,\n new_type, link, *((data,) + args), **overrides)\n if clone.callback is self.callback:\n with util.disable_constant(clone):\n clone.callback = clone.callback.clone(inputs=[self],\n link_inputs=link)\n return clone\n", "role": "user" }, { "content": "clone", "role": "assistant" } ]
def x(self, tuple_key, cache, data_slice): if not any(isinstance(el, (list, set)) for el in tuple_key): return None if len(tuple_key) == 1: product = tuple_key[0] else: args = [(set(el) if isinstance(el, (list, set)) else set([el])) for el in tuple_key] product = itertools.product(*args) data = [] for inner_key in product: key = util.wrap_tuple(inner_key) if key in cache: val = cache[key] else: val = self._execute_callback(*key) if data_slice: val = self._dataslice(val, data_slice) data.append((key, val)) product = self.clone(data) if data_slice: from ..util import Dynamic dmap = Dynamic(self, operation=lambda obj, **dynkwargs: obj[ data_slice], streams=self.streams) dmap.data = product.data return dmap return product
cross_product
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tuple_key, cache, data_slice):\n if not any(isinstance(el, (list, set)) for el in tuple_key):\n return None\n if len(tuple_key) == 1:\n product = tuple_key[0]\n else:\n args = [(set(el) if isinstance(el, (list, set)) else set([el])) for\n el in tuple_key]\n product = itertools.product(*args)\n data = []\n for inner_key in product:\n key = util.wrap_tuple(inner_key)\n if key in cache:\n val = cache[key]\n else:\n val = self._execute_callback(*key)\n if data_slice:\n val = self._dataslice(val, data_slice)\n data.append((key, val))\n product = self.clone(data)\n if data_slice:\n from ..util import Dynamic\n dmap = Dynamic(self, operation=lambda obj, **dynkwargs: obj[\n data_slice], streams=self.streams)\n dmap.data = product.data\n return dmap\n return product\n", "role": "user" }, { "content": "cross_product", "role": "assistant" } ]
def x(self, tuple_key, data_slice): slices = [el for el in tuple_key if isinstance(el, slice)] if any(el.step for el in slices): raise Exception('DynamicMap slices cannot have a step argument') elif len(slices) not in [0, len(tuple_key)]: raise Exception('Slices must be used exclusively or not at all') elif not slices: return None sliced = self.clone(self) for i, slc in enumerate(tuple_key): start, stop = slc.start, slc.stop if start is not None and start < sliced.kdims[i].range[0]: raise Exception('Requested slice below defined dimension range.') if stop is not None and stop > sliced.kdims[i].range[1]: raise Exception('Requested slice above defined dimension range.') sliced.kdims[i].soft_range = start, stop if data_slice: if not isinstance(sliced, DynamicMap): return self._dataslice(sliced, data_slice) else: from ..util import Dynamic if len(self): slices = [slice(None) for _ in range(self.ndims)] + list( data_slice) sliced = super(DynamicMap, sliced).__getitem__(tuple(slices)) dmap = Dynamic(self, operation=lambda obj, **dynkwargs: obj[ data_slice], streams=self.streams) dmap.data = sliced.data return dmap return sliced
slice_bounded
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tuple_key, data_slice):\n slices = [el for el in tuple_key if isinstance(el, slice)]\n if any(el.step for el in slices):\n raise Exception('DynamicMap slices cannot have a step argument')\n elif len(slices) not in [0, len(tuple_key)]:\n raise Exception('Slices must be used exclusively or not at all')\n elif not slices:\n return None\n sliced = self.clone(self)\n for i, slc in enumerate(tuple_key):\n start, stop = slc.start, slc.stop\n if start is not None and start < sliced.kdims[i].range[0]:\n raise Exception('Requested slice below defined dimension range.')\n if stop is not None and stop > sliced.kdims[i].range[1]:\n raise Exception('Requested slice above defined dimension range.')\n sliced.kdims[i].soft_range = start, stop\n if data_slice:\n if not isinstance(sliced, DynamicMap):\n return self._dataslice(sliced, data_slice)\n else:\n from ..util import Dynamic\n if len(self):\n slices = [slice(None) for _ in range(self.ndims)] + list(\n data_slice)\n sliced = super(DynamicMap, sliced).__getitem__(tuple(slices))\n dmap = Dynamic(self, operation=lambda obj, **dynkwargs: obj[\n data_slice], streams=self.streams)\n dmap.data = sliced.data\n return dmap\n return sliced\n", "role": "user" }, { "content": "slice_bounded", "role": "assistant" } ]
def x(self, selection_specs=None, **kwargs): if selection_specs is not None and not isinstance(selection_specs, ( list, tuple)): selection_specs = [selection_specs] selection = super(DynamicMap, self).select(selection_specs, **kwargs) def x(obj, **dynkwargs): if selection_specs is not None: matches = any(obj.matches(spec) for spec in selection_specs) else: matches = True if matches: return obj.select(**kwargs) return obj if not isinstance(selection, DynamicMap): return dynamic_select(selection) else: from ..util import Dynamic dmap = Dynamic(self, operation=dynamic_select, streams=self.streams) dmap.data = selection.data return dmap
select
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, selection_specs=None, **kwargs):\n if selection_specs is not None and not isinstance(selection_specs, (\n list, tuple)):\n selection_specs = [selection_specs]\n selection = super(DynamicMap, self).select(selection_specs, **kwargs)\n\n def x(obj, **dynkwargs):\n if selection_specs is not None:\n matches = any(obj.matches(spec) for spec in selection_specs)\n else:\n matches = True\n if matches:\n return obj.select(**kwargs)\n return obj\n if not isinstance(selection, DynamicMap):\n return dynamic_select(selection)\n else:\n from ..util import Dynamic\n dmap = Dynamic(self, operation=dynamic_select, streams=self.streams)\n dmap.data = selection.data\n return dmap\n", "role": "user" }, { "content": "select", "role": "assistant" } ]
def x(self, key, val): cache_size = 1 if util.dimensionless_contents(self.streams, self.kdims ) else self.cache_size if len(self) >= cache_size: first_key = next(k for k in self.data) self.data.pop(first_key) self[key] = val
cache
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, val):\n cache_size = 1 if util.dimensionless_contents(self.streams, self.kdims\n ) else self.cache_size\n if len(self) >= cache_size:\n first_key = next(k for k in self.data)\n self.data.pop(first_key)\n self[key] = val\n", "role": "user" }, { "content": "cache", "role": "assistant" } ]
def x(self, map_fn, specs=None, clone=True, link_inputs=True): deep_mapped = super(DynamicMap, self).map(map_fn, specs, clone) if isinstance(deep_mapped, type(self)): from ..util import Dynamic def x(obj, **dynkwargs): return obj.map(map_fn, specs, clone) dmap = Dynamic(self, operation=apply_map, streams=self.streams, link_inputs=link_inputs) dmap.data = deep_mapped.data return dmap return deep_mapped
map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, map_fn, specs=None, clone=True, link_inputs=True):\n deep_mapped = super(DynamicMap, self).map(map_fn, specs, clone)\n if isinstance(deep_mapped, type(self)):\n from ..util import Dynamic\n\n def x(obj, **dynkwargs):\n return obj.map(map_fn, specs, clone)\n dmap = Dynamic(self, operation=apply_map, streams=self.streams,\n link_inputs=link_inputs)\n dmap.data = deep_mapped.data\n return dmap\n return deep_mapped\n", "role": "user" }, { "content": "map", "role": "assistant" } ]
def x(self, label=None, group=None, depth=1): relabelled = super(DynamicMap, self).relabel(label, group, depth) if depth > 0: from ..util import Dynamic def x(obj, **dynkwargs): return obj.relabel(group=group, label=label, depth=depth - 1) dmap = Dynamic(self, streams=self.streams, operation=dynamic_relabel) dmap.data = relabelled.data with util.disable_constant(dmap): dmap.group = relabelled.group dmap.label = relabelled.label return dmap return relabelled
relabel
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, label=None, group=None, depth=1):\n relabelled = super(DynamicMap, self).relabel(label, group, depth)\n if depth > 0:\n from ..util import Dynamic\n\n def x(obj, **dynkwargs):\n return obj.relabel(group=group, label=label, depth=depth - 1)\n dmap = Dynamic(self, streams=self.streams, operation=dynamic_relabel)\n dmap.data = relabelled.data\n with util.disable_constant(dmap):\n dmap.group = relabelled.group\n dmap.label = relabelled.label\n return dmap\n return relabelled\n", "role": "user" }, { "content": "relabel", "role": "assistant" } ]
def x(self): if not len(self): raise ValueError( 'Cannot split DynamicMap before it has been initialized') elif not issubclass(self.type, CompositeOverlay): return None, self from ..util import Dynamic keys = list(self.last.data.keys()) dmaps = [] for key in keys: el = self.last.data[key] def x(obj, overlay_key=key, overlay_el=el, **kwargs): spec = util.get_overlay_spec(obj, overlay_key, overlay_el) items = list(obj.data.items()) specs = [(i, util.get_overlay_spec(obj, k, v)) for i, (k, v) in enumerate(items)] match = util.closest_match(spec, specs) if match is None: raise KeyError( '{spec} spec not found in {otype}. The split_overlays method only works consistently for a DynamicMap where the layers of the {otype} do not change.' .format(spec=spec, otype=type(obj).__name__)) return items[match][1] dmap = Dynamic(self, streams=self.streams, operation= split_overlay_callback) dmap.data = OrderedDict([(list(self.data.keys())[-1], self.last. data[key])]) dmaps.append(dmap) return keys, dmaps
split_overlays
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not len(self):\n raise ValueError(\n 'Cannot split DynamicMap before it has been initialized')\n elif not issubclass(self.type, CompositeOverlay):\n return None, self\n from ..util import Dynamic\n keys = list(self.last.data.keys())\n dmaps = []\n for key in keys:\n el = self.last.data[key]\n\n def x(obj, overlay_key=key, overlay_el=el, **kwargs):\n spec = util.get_overlay_spec(obj, overlay_key, overlay_el)\n items = list(obj.data.items())\n specs = [(i, util.get_overlay_spec(obj, k, v)) for i, (k, v) in\n enumerate(items)]\n match = util.closest_match(spec, specs)\n if match is None:\n raise KeyError(\n '{spec} spec not found in {otype}. The split_overlays method only works consistently for a DynamicMap where the layers of the {otype} do not change.'\n .format(spec=spec, otype=type(obj).__name__))\n return items[match][1]\n dmap = Dynamic(self, streams=self.streams, operation=\n split_overlay_callback)\n dmap.data = OrderedDict([(list(self.data.keys())[-1], self.last.\n data[key])])\n dmaps.append(dmap)\n return keys, dmaps\n", "role": "user" }, { "content": "split_overlays", "role": "assistant" } ]
def x(self): if self.last is not None: initialized = self else: initialized = self.clone() initialized[initialized._initial_key()] if not isinstance(initialized.last, (Layout, NdLayout, GridSpace)): return self container = initialized.last.clone(shared_data=False) type_counter = defaultdict(int) remapped_streams = [] self_dstreams = util.dimensioned_streams(self) streams = self.callback.stream_mapping for i, (k, v) in enumerate(initialized.last.data.items()): vstreams = streams.get(i, []) if not vstreams: if isinstance(initialized.last, Layout): for l in range(len(k)): path = '.'.join(k[:l]) if path in streams: vstreams = streams[path] break else: vstreams = streams.get(k, []) if any(s in remapped_streams for s in vstreams): raise ValueError( 'The stream_mapping supplied on the Callable is ambiguous please supply more specific Layout path specs.' ) remapped_streams += vstreams def x(*args, **kwargs): layout = self[args] layout_type = type(layout).__name__ if len(container.keys()) != len(layout.keys()): raise ValueError( 'Collated DynamicMaps must return %s with consistent number of items.' % layout_type) key = kwargs['selection_key'] index = kwargs['selection_index'] obj_type = kwargs['selection_type'] dyn_type_map = defaultdict(list) for k, v in layout.data.items(): if k == key: return layout[k] dyn_type_map[type(v)].append(v) dyn_type_counter = {t: len(vals) for t, vals in dyn_type_map. items()} if dyn_type_counter != type_counter: raise ValueError( 'The objects in a %s returned by a DynamicMap must consistently return the same number of items of the same type.' % layout_type) return dyn_type_map[obj_type][index] callback = Callable(partial(collation_cb, selection_key=k, selection_index=type_counter[type(v)], selection_type=type(v)), inputs=[self]) vstreams = list(util.unique_iterator(self_dstreams + vstreams)) vdmap = self.clone(callback=callback, shared_data=False, streams= vstreams) type_counter[type(v)] += 1 for stream in vstreams: if stream.source is self: stream.source = vdmap container[k] = vdmap unmapped_streams = [repr(stream) for stream in self.streams if stream. source is self and stream not in remapped_streams and stream.linked] if unmapped_streams: raise ValueError( """The following streams are set to be automatically linked to a plot, but no stream_mapping specifying which item in the (Nd)Layout to link it to was found: %s""" % ', '.join(unmapped_streams)) return container
collate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.last is not None:\n initialized = self\n else:\n initialized = self.clone()\n initialized[initialized._initial_key()]\n if not isinstance(initialized.last, (Layout, NdLayout, GridSpace)):\n return self\n container = initialized.last.clone(shared_data=False)\n type_counter = defaultdict(int)\n remapped_streams = []\n self_dstreams = util.dimensioned_streams(self)\n streams = self.callback.stream_mapping\n for i, (k, v) in enumerate(initialized.last.data.items()):\n vstreams = streams.get(i, [])\n if not vstreams:\n if isinstance(initialized.last, Layout):\n for l in range(len(k)):\n path = '.'.join(k[:l])\n if path in streams:\n vstreams = streams[path]\n break\n else:\n vstreams = streams.get(k, [])\n if any(s in remapped_streams for s in vstreams):\n raise ValueError(\n 'The stream_mapping supplied on the Callable is ambiguous please supply more specific Layout path specs.'\n )\n remapped_streams += vstreams\n\n def x(*args, **kwargs):\n layout = self[args]\n layout_type = type(layout).__name__\n if len(container.keys()) != len(layout.keys()):\n raise ValueError(\n 'Collated DynamicMaps must return %s with consistent number of items.'\n % layout_type)\n key = kwargs['selection_key']\n index = kwargs['selection_index']\n obj_type = kwargs['selection_type']\n dyn_type_map = defaultdict(list)\n for k, v in layout.data.items():\n if k == key:\n return layout[k]\n dyn_type_map[type(v)].append(v)\n dyn_type_counter = {t: len(vals) for t, vals in dyn_type_map.\n items()}\n if dyn_type_counter != type_counter:\n raise ValueError(\n 'The objects in a %s returned by a DynamicMap must consistently return the same number of items of the same type.'\n % layout_type)\n return dyn_type_map[obj_type][index]\n callback = Callable(partial(collation_cb, selection_key=k,\n selection_index=type_counter[type(v)], selection_type=type(v)),\n inputs=[self])\n vstreams = list(util.unique_iterator(self_dstreams + vstreams))\n vdmap = self.clone(callback=callback, shared_data=False, streams=\n vstreams)\n type_counter[type(v)] += 1\n for stream in vstreams:\n if stream.source is self:\n stream.source = vdmap\n container[k] = vdmap\n unmapped_streams = [repr(stream) for stream in self.streams if stream.\n source is self and stream not in remapped_streams and stream.linked]\n if unmapped_streams:\n raise ValueError(\n \"\"\"The following streams are set to be automatically linked to a plot, but no stream_mapping specifying which item in the (Nd)Layout to link it to was found:\n%s\"\"\"\n % ', '.join(unmapped_streams))\n return container\n", "role": "user" }, { "content": "collate", "role": "assistant" } ]
def x(self, dimensions=None, container_type=None, group_type=None, **kwargs): if dimensions is None: dimensions = self.kdims if not isinstance(dimensions, (list, tuple)): dimensions = [dimensions] container_type = container_type if container_type else type(self) group_type = group_type if group_type else type(self) outer_kdims = [self.get_dimension(d) for d in dimensions] inner_kdims = [d for d in self.kdims if not d in outer_kdims] outer_dynamic = issubclass(container_type, DynamicMap) inner_dynamic = issubclass(group_type, DynamicMap) if not outer_dynamic and any(not d.values for d in outer_kdims ) or not inner_dynamic and any(not d.values for d in inner_kdims): raise Exception( 'Dimensions must specify sampling via values to apply a groupby') if outer_dynamic: def x(*outer_key, **dynkwargs): if inner_dynamic: def x(*inner_key, **dynkwargs): outer_vals = zip(outer_kdims, util.wrap_tuple(outer_key)) inner_vals = zip(inner_kdims, util.wrap_tuple(inner_key)) inner_sel = [(k.name, v) for k, v in inner_vals] outer_sel = [(k.name, v) for k, v in outer_vals] return self.select(**dict(inner_sel + outer_sel)) return self.clone([], callback=inner_fn, kdims=inner_kdims) else: dim_vals = [(d.name, d.values) for d in inner_kdims] dim_vals += [(d.name, [v]) for d, v in zip(outer_kdims, util.wrap_tuple(outer_key))] with item_check(False): selected = HoloMap(self.select(**dict(dim_vals))) return group_type(selected.reindex(inner_kdims)) if outer_kdims: return self.clone([], callback=outer_fn, kdims=outer_kdims) else: return outer_fn(()) else: outer_product = itertools.product(*[self.get_dimension(d).values for d in dimensions]) groups = [] for outer in outer_product: outer_vals = [(d.name, [o]) for d, o in zip(outer_kdims, outer)] if inner_dynamic or not inner_kdims: def x(outer_vals, *key, **dynkwargs): inner_dims = zip(inner_kdims, util.wrap_tuple(key)) inner_vals = [(d.name, k) for d, k in inner_dims] return self.select(**dict(outer_vals + inner_vals)).last if inner_kdims or self.streams: group = self.clone(callback=partial(inner_fn, outer_vals), kdims=inner_kdims) else: group = inner_fn(outer_vals, ()) groups.append((outer, group)) else: inner_vals = [(d.name, self.get_dimension(d).values) for d in inner_kdims] with item_check(False): selected = HoloMap(self.select(**dict(outer_vals + inner_vals))) group = group_type(selected.reindex(inner_kdims)) groups.append((outer, group)) return container_type(groups, kdims=outer_kdims)
groupby
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimensions=None, container_type=None, group_type=None, **kwargs):\n if dimensions is None:\n dimensions = self.kdims\n if not isinstance(dimensions, (list, tuple)):\n dimensions = [dimensions]\n container_type = container_type if container_type else type(self)\n group_type = group_type if group_type else type(self)\n outer_kdims = [self.get_dimension(d) for d in dimensions]\n inner_kdims = [d for d in self.kdims if not d in outer_kdims]\n outer_dynamic = issubclass(container_type, DynamicMap)\n inner_dynamic = issubclass(group_type, DynamicMap)\n if not outer_dynamic and any(not d.values for d in outer_kdims\n ) or not inner_dynamic and any(not d.values for d in inner_kdims):\n raise Exception(\n 'Dimensions must specify sampling via values to apply a groupby')\n if outer_dynamic:\n\n def x(*outer_key, **dynkwargs):\n if inner_dynamic:\n\n def x(*inner_key, **dynkwargs):\n outer_vals = zip(outer_kdims, util.wrap_tuple(outer_key))\n inner_vals = zip(inner_kdims, util.wrap_tuple(inner_key))\n inner_sel = [(k.name, v) for k, v in inner_vals]\n outer_sel = [(k.name, v) for k, v in outer_vals]\n return self.select(**dict(inner_sel + outer_sel))\n return self.clone([], callback=inner_fn, kdims=inner_kdims)\n else:\n dim_vals = [(d.name, d.values) for d in inner_kdims]\n dim_vals += [(d.name, [v]) for d, v in zip(outer_kdims,\n util.wrap_tuple(outer_key))]\n with item_check(False):\n selected = HoloMap(self.select(**dict(dim_vals)))\n return group_type(selected.reindex(inner_kdims))\n if outer_kdims:\n return self.clone([], callback=outer_fn, kdims=outer_kdims)\n else:\n return outer_fn(())\n else:\n outer_product = itertools.product(*[self.get_dimension(d).values for\n d in dimensions])\n groups = []\n for outer in outer_product:\n outer_vals = [(d.name, [o]) for d, o in zip(outer_kdims, outer)]\n if inner_dynamic or not inner_kdims:\n\n def x(outer_vals, *key, **dynkwargs):\n inner_dims = zip(inner_kdims, util.wrap_tuple(key))\n inner_vals = [(d.name, k) for d, k in inner_dims]\n return self.select(**dict(outer_vals + inner_vals)).last\n if inner_kdims or self.streams:\n group = self.clone(callback=partial(inner_fn,\n outer_vals), kdims=inner_kdims)\n else:\n group = inner_fn(outer_vals, ())\n groups.append((outer, group))\n else:\n inner_vals = [(d.name, self.get_dimension(d).values) for d in\n inner_kdims]\n with item_check(False):\n selected = HoloMap(self.select(**dict(outer_vals +\n inner_vals)))\n group = group_type(selected.reindex(inner_kdims))\n groups.append((outer, group))\n return container_type(groups, kdims=outer_kdims)\n", "role": "user" }, { "content": "groupby", "role": "assistant" } ]
def x(self, dimensions=None, **kwargs): return self.groupby(dimensions, container_type=GridSpace, **kwargs)
grid
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimensions=None, **kwargs):\n return self.groupby(dimensions, container_type=GridSpace, **kwargs)\n", "role": "user" }, { "content": "grid", "role": "assistant" } ]
def x(self, dimensions=None, **kwargs): return self.groupby(dimensions, container_type=NdLayout, **kwargs)
layout
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimensions=None, **kwargs):\n return self.groupby(dimensions, container_type=NdLayout, **kwargs)\n", "role": "user" }, { "content": "layout", "role": "assistant" } ]
def x(self, dimensions=None, **kwargs): if dimensions is None: dimensions = self.kdims else: if not isinstance(dimensions, (list, tuple)): dimensions = [dimensions] dimensions = [self.get_dimension(d, strict=True) for d in dimensions] dims = [d for d in self.kdims if d not in dimensions] return self.groupby(dims, group_type=NdOverlay)
overlay
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimensions=None, **kwargs):\n if dimensions is None:\n dimensions = self.kdims\n else:\n if not isinstance(dimensions, (list, tuple)):\n dimensions = [dimensions]\n dimensions = [self.get_dimension(d, strict=True) for d in dimensions]\n dims = [d for d in self.kdims if d not in dimensions]\n return self.groupby(dims, group_type=NdOverlay)\n", "role": "user" }, { "content": "overlay", "role": "assistant" } ]
def x(self, dimension=None, num_bins=20, bin_range=None, adjoin=True, **kwargs ): def x(obj, **dynkwargs): if isinstance(obj, (NdOverlay, Overlay)): index = kwargs.get('index', 0) obj = obj.get(index) return obj.hist(num_bins=num_bins, bin_range=bin_range, adjoin= False, **kwargs) from ..util import Dynamic hist = Dynamic(self, streams=self.streams, link_inputs=False, operation =dynamic_hist) if adjoin: return self << hist else: return hist
hist
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimension=None, num_bins=20, bin_range=None, adjoin=True, **kwargs\n ):\n\n def x(obj, **dynkwargs):\n if isinstance(obj, (NdOverlay, Overlay)):\n index = kwargs.get('index', 0)\n obj = obj.get(index)\n return obj.hist(num_bins=num_bins, bin_range=bin_range, adjoin=\n False, **kwargs)\n from ..util import Dynamic\n hist = Dynamic(self, streams=self.streams, link_inputs=False, operation\n =dynamic_hist)\n if adjoin:\n return self << hist\n else:\n return hist\n", "role": "user" }, { "content": "hist", "role": "assistant" } ]
def x(self, kdims=[], force=False): if not isinstance(kdims, list): kdims = [kdims] kdims = [self.get_dimension(kd, strict=True) for kd in kdims] dropped = [kd for kd in self.kdims if kd not in kdims] if dropped: raise ValueError( 'DynamicMap does not allow dropping dimensions, reindex may only be used to reorder dimensions.' ) return super(DynamicMap, self).reindex(kdims, force)
reindex
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, kdims=[], force=False):\n if not isinstance(kdims, list):\n kdims = [kdims]\n kdims = [self.get_dimension(kd, strict=True) for kd in kdims]\n dropped = [kd for kd in self.kdims if kd not in kdims]\n if dropped:\n raise ValueError(\n 'DynamicMap does not allow dropping dimensions, reindex may only be used to reorder dimensions.'\n )\n return super(DynamicMap, self).reindex(kdims, force)\n", "role": "user" }, { "content": "reindex", "role": "assistant" } ]
def x(self, key): ndims = self.ndims if all(not (isinstance(el, slice) or callable(el)) for el in key): dim_inds = [] for dim in self.kdims: dim_type = self.get_dimension_type(dim) if isinstance(dim_type, type) and issubclass(dim_type, Number): dim_inds.append(self.get_dimension_index(dim)) str_keys = iter(key[i] for i in range(self.ndims) if i not in dim_inds) num_keys = [] if len(dim_inds): keys = list({tuple(k[i] if ndims > 1 else k for i in dim_inds) for k in self.keys()}) q = np.array([tuple(key[i] if ndims > 1 else key for i in dim_inds)]) idx = np.argmin([(np.inner(q - np.array(x), q - np.array(x)) if len(dim_inds) == 2 else np.abs(q - x)) for x in keys]) num_keys = iter(keys[idx]) key = tuple(next(num_keys) if i in dim_inds else next(str_keys) for i in range(self.ndims)) elif any(not (isinstance(el, slice) or callable(el)) for el in key): keys = self.keys() for i, k in enumerate(key): if isinstance(k, slice): continue dim_keys = np.array([ke[i] for ke in keys]) if dim_keys.dtype.kind in 'OSU': continue snapped_val = dim_keys[np.argmin(np.abs(dim_keys - k))] key = list(key) key[i] = snapped_val key = tuple(key) return key
transform_indices
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key):\n ndims = self.ndims\n if all(not (isinstance(el, slice) or callable(el)) for el in key):\n dim_inds = []\n for dim in self.kdims:\n dim_type = self.get_dimension_type(dim)\n if isinstance(dim_type, type) and issubclass(dim_type, Number):\n dim_inds.append(self.get_dimension_index(dim))\n str_keys = iter(key[i] for i in range(self.ndims) if i not in dim_inds)\n num_keys = []\n if len(dim_inds):\n keys = list({tuple(k[i] if ndims > 1 else k for i in dim_inds) for\n k in self.keys()})\n q = np.array([tuple(key[i] if ndims > 1 else key for i in\n dim_inds)])\n idx = np.argmin([(np.inner(q - np.array(x), q - np.array(x)) if\n len(dim_inds) == 2 else np.abs(q - x)) for x in keys])\n num_keys = iter(keys[idx])\n key = tuple(next(num_keys) if i in dim_inds else next(str_keys) for\n i in range(self.ndims))\n elif any(not (isinstance(el, slice) or callable(el)) for el in key):\n keys = self.keys()\n for i, k in enumerate(key):\n if isinstance(k, slice):\n continue\n dim_keys = np.array([ke[i] for ke in keys])\n if dim_keys.dtype.kind in 'OSU':\n continue\n snapped_val = dim_keys[np.argmin(np.abs(dim_keys - k))]\n key = list(key)\n key[i] = snapped_val\n key = tuple(key)\n return key\n", "role": "user" }, { "content": "transform_indices", "role": "assistant" } ]
def x(self, full_grid=False): keys = super(GridSpace, self).keys() if self.ndims == 1 or not full_grid: return keys dim1_keys = sorted(set(k[0] for k in keys)) dim2_keys = sorted(set(k[1] for k in keys)) return [(d1, d2) for d1 in dim1_keys for d2 in dim2_keys]
keys
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, full_grid=False):\n keys = super(GridSpace, self).keys()\n if self.ndims == 1 or not full_grid:\n return keys\n dim1_keys = sorted(set(k[0] for k in keys))\n dim2_keys = sorted(set(k[1] for k in keys))\n return [(d1, d2) for d1 in dim1_keys for d2 in dim2_keys]\n", "role": "user" }, { "content": "keys", "role": "assistant" } ]
def x(self): if self.type == HoloMap: last_items = [(k, v.last if isinstance(v, HoloMap) else v) for k, v in self.data.items()] else: last_items = self.data return self.clone(last_items)
last
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.type == HoloMap:\n last_items = [(k, v.last if isinstance(v, HoloMap) else v) for k, v in\n self.data.items()]\n else:\n last_items = self.data\n return self.clone(last_items)\n", "role": "user" }, { "content": "last", "role": "assistant" } ]
def x(self): keys = self.keys() if self.ndims == 1: return len(keys), 1 return len(set(k[0] for k in keys)), len(set(k[1] for k in keys))
shape
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n keys = self.keys()\n if self.ndims == 1:\n return len(keys), 1\n return len(set(k[0] for k in keys)), len(set(k[1] for k in keys))\n", "role": "user" }, { "content": "shape", "role": "assistant" } ]
def x(self, data, element): if 'hover' not in self.handles or self.static_source: return for k, v in self.overlay_dims.items(): dim = util.dimension_sanitizer(k.name) if dim not in data: data[dim] = [v for _ in range(len(list(data.values())[0]))]
get_hover_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data, element):\n if 'hover' not in self.handles or self.static_source:\n return\n for k, v in self.overlay_dims.items():\n dim = util.dimension_sanitizer(k.name)\n if dim not in data:\n data[dim] = [v for _ in range(len(list(data.values())[0]))]\n", "role": "user" }, { "content": "get_hover_data", "role": "assistant" } ]
def x(self, data, element): if 'hover' not in self.handles or self.static_source: return npath = len([vs for vs in data.values()][0]) for d in element.vdims: dim = util.dimension_sanitizer(d.name) if dim not in data: if element.level is not None: data[dim] = np.full(npath, element.level) elif element.interface.isscalar(element, d): data[dim] = element.dimension_values(d, expanded=False) else: data[dim] = element.split(datatype='array', dimensions=[d]) elif isinstance(data[dim], np.ndarray) and data[dim].dtype.kind == 'M': data[dim + '_dt_strings'] = [d.pprint_value(v) for v in data[dim]] for k, v in self.overlay_dims.items(): dim = util.dimension_sanitizer(k.name) if dim not in data: data[dim] = [v for _ in range(len(list(data.values())[0]))]
get_hover_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data, element):\n if 'hover' not in self.handles or self.static_source:\n return\n npath = len([vs for vs in data.values()][0])\n for d in element.vdims:\n dim = util.dimension_sanitizer(d.name)\n if dim not in data:\n if element.level is not None:\n data[dim] = np.full(npath, element.level)\n elif element.interface.isscalar(element, d):\n data[dim] = element.dimension_values(d, expanded=False)\n else:\n data[dim] = element.split(datatype='array', dimensions=[d])\n elif isinstance(data[dim], np.ndarray) and data[dim].dtype.kind == 'M':\n data[dim + '_dt_strings'] = [d.pprint_value(v) for v in data[dim]]\n for k, v in self.overlay_dims.items():\n dim = util.dimension_sanitizer(k.name)\n if dim not in data:\n data[dim] = [v for _ in range(len(list(data.values())[0]))]\n", "role": "user" }, { "content": "get_hover_data", "role": "assistant" } ]
def x(self, plot, mapping, properties): plot_method = properties.pop('plot_method', None) properties = mpl_to_bokeh(properties) data = dict(properties, **mapping) if self._has_holes: plot_method = 'multi_polygons' elif plot_method is None: plot_method = self._plot_methods.get('single') renderer = getattr(plot, plot_method)(**data) if self.colorbar: for k, v in list(self.handles.items()): if not k.endswith('color_mapper'): continue self._draw_colorbar(plot, v, k[:-12]) return renderer, renderer.glyph
init_glyph
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, plot, mapping, properties):\n plot_method = properties.pop('plot_method', None)\n properties = mpl_to_bokeh(properties)\n data = dict(properties, **mapping)\n if self._has_holes:\n plot_method = 'multi_polygons'\n elif plot_method is None:\n plot_method = self._plot_methods.get('single')\n renderer = getattr(plot, plot_method)(**data)\n if self.colorbar:\n for k, v in list(self.handles.items()):\n if not k.endswith('color_mapper'):\n continue\n self._draw_colorbar(plot, v, k[:-12])\n return renderer, renderer.glyph\n", "role": "user" }, { "content": "init_glyph", "role": "assistant" } ]
def x(self, dimension=None, num_bins=20, bin_range=None, adjoin=True, index =0, **kwargs): valid_ind = isinstance(index, int) and 0 <= index < len(self) valid_label = index in [el.label for el in self] if not any([valid_ind, valid_label]): raise TypeError( 'Please supply a suitable index or label for the histogram data') hists = self.get(index).hist(adjoin=False, dimension=dimension, bin_range=bin_range, num_bins=num_bins, **kwargs) if not isinstance(hists, Layout): hists = [hists] if not isinstance(dimension, list): dimension = ['Default'] if adjoin: layout = self for hist in hists: layout = layout << hist layout.main_layer = index elif len(dimension) > 1: layout = hists else: layout = hists[0] return layout
hist
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimension=None, num_bins=20, bin_range=None, adjoin=True, index\n =0, **kwargs):\n valid_ind = isinstance(index, int) and 0 <= index < len(self)\n valid_label = index in [el.label for el in self]\n if not any([valid_ind, valid_label]):\n raise TypeError(\n 'Please supply a suitable index or label for the histogram data')\n hists = self.get(index).hist(adjoin=False, dimension=dimension,\n bin_range=bin_range, num_bins=num_bins, **kwargs)\n if not isinstance(hists, Layout):\n hists = [hists]\n if not isinstance(dimension, list):\n dimension = ['Default']\n if adjoin:\n layout = self\n for hist in hists:\n layout = layout << hist\n layout.main_layer = index\n elif len(dimension) > 1:\n layout = hists\n else:\n layout = hists[0]\n return layout\n", "role": "user" }, { "content": "hist", "role": "assistant" } ]
def x(self, dimension, expanded=True, flat=True): values = [] found = False for el in self: if dimension in el.dimensions(label=True): values.append(el.dimension_values(dimension)) found = True if not found: return super(CompositeOverlay, self).dimension_values(dimension, expanded, flat) values = [v for v in values if v is not None and len(v)] if not values: return np.array() vals = np.concatenate(values) return vals if expanded else unique_array(vals)
dimension_values
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dimension, expanded=True, flat=True):\n values = []\n found = False\n for el in self:\n if dimension in el.dimensions(label=True):\n values.append(el.dimension_values(dimension))\n found = True\n if not found:\n return super(CompositeOverlay, self).dimension_values(dimension,\n expanded, flat)\n values = [v for v in values if v is not None and len(v)]\n if not values:\n return np.array()\n vals = np.concatenate(values)\n return vals if expanded else unique_array(vals)\n", "role": "user" }, { "content": "dimension_values", "role": "assistant" } ]
def x(self, identifier, default=None): if isinstance(identifier, int): values = list(self.data.values()) if 0 <= identifier < len(values): return values[identifier] else: return default return super(Overlay, self).get(identifier, default)
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, identifier, default=None):\n if isinstance(identifier, int):\n values = list(self.data.values())\n if 0 <= identifier < len(values):\n return values[identifier]\n else:\n return default\n return super(Overlay, self).get(identifier, default)\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(self, function): if config.future_deprecations: self.param.warning( 'Overlay.collapse is deprecated, tocollapse multiple elements use a HoloMap.' ) elements = list(self) types = [type(el) for el in elements] values = [el.group for el in elements] if not len(set(types)) == 1 and len(set(values)) == 1: raise Exception( 'Overlay is not homogeneous in type or group and cannot be collapsed.' ) else: return elements[0].clone(types[0].collapse_data([el.data for el in elements], function, self.kdims))
collapse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, function):\n if config.future_deprecations:\n self.param.warning(\n 'Overlay.collapse is deprecated, tocollapse multiple elements use a HoloMap.'\n )\n elements = list(self)\n types = [type(el) for el in elements]\n values = [el.group for el in elements]\n if not len(set(types)) == 1 and len(set(values)) == 1:\n raise Exception(\n 'Overlay is not homogeneous in type or group and cannot be collapsed.'\n )\n else:\n return elements[0].clone(types[0].collapse_data([el.data for el in\n elements], function, self.kdims))\n", "role": "user" }, { "content": "collapse", "role": "assistant" } ]
def x(obj): if isinstance(obj, Overlay) and any(isinstance(o, (HoloMap, GridSpace)) for o in obj): return False if isinstance(obj, HoloMap): return not obj.type in [Layout, GridSpace, NdLayout, DynamicMap] if isinstance(obj, (GridSpace, Layout, NdLayout)): for el in obj.values(): if not displayable(el): return False return True return True
displayable
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(obj):\n if isinstance(obj, Overlay) and any(isinstance(o, (HoloMap, GridSpace)) for\n o in obj):\n return False\n if isinstance(obj, HoloMap):\n return not obj.type in [Layout, GridSpace, NdLayout, DynamicMap]\n if isinstance(obj, (GridSpace, Layout, NdLayout)):\n for el in obj.values():\n if not displayable(el):\n return False\n return True\n return True\n", "role": "user" }, { "content": "displayable", "role": "assistant" } ]
def x(obj): if isinstance(obj, DynamicMap): if isinstance(obj.last, CompositeOverlay): return len(obj.last) elif obj.last is None: return None return 1 else: return 1
overlay_depth
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(obj):\n if isinstance(obj, DynamicMap):\n if isinstance(obj.last, CompositeOverlay):\n return len(obj.last)\n elif obj.last is None:\n return None\n return 1\n else:\n return 1\n", "role": "user" }, { "content": "overlay_depth", "role": "assistant" } ]
def x(obj, path=[]): path = path + [obj] zorder_map = defaultdict(list) if not isinstance(obj, DynamicMap): if isinstance(obj, CompositeOverlay): for z, o in enumerate(obj): zorder_map[z] = [o, obj] elif isinstance(obj, HoloMap): for el in obj.values(): if isinstance(el, CompositeOverlay): for k, v in compute_overlayable_zorders(el, path).items(): zorder_map[k] += v + [obj] else: zorder_map[0] += [obj, el] elif obj not in zorder_map[0]: zorder_map[0].append(obj) return zorder_map isoverlay = isinstance(obj.last, CompositeOverlay) isdynoverlay = obj.callback._is_overlay if obj not in zorder_map[0] and not isoverlay: zorder_map[0].append(obj) depth = overlay_depth(obj) dmap_inputs = obj.callback.inputs if obj.callback.link_inputs else [] for z, inp in enumerate(dmap_inputs): no_zorder_increment = False if any(not (isoverlay_fn(p) or p.last is None) for p in path ) and isoverlay_fn(inp): no_zorder_increment = True input_depth = overlay_depth(inp) if (depth is not None and input_depth is not None and depth < input_depth): if depth > 1: continue else: no_zorder_increment = True z = z if isdynoverlay else 0 deep_zorders = compute_overlayable_zorders(inp, path=path) offset = max(zorder_map.keys()) for dz, objs in deep_zorders.items(): global_z = offset + z if no_zorder_increment else offset + dz + z zorder_map[global_z] = list(unique_iterator(zorder_map[global_z ] + objs)) found = any(isinstance(p, DynamicMap) and p.callback._is_overlay for p in path) linked = any(isinstance(s, LinkedStream) and s.linked for s in obj.streams) if (found or linked) and isoverlay and not isdynoverlay: offset = max(zorder_map.keys()) for z, o in enumerate(obj.last): if isoverlay and linked: zorder_map[offset + z].append(obj) if o not in zorder_map[offset + z]: zorder_map[offset + z].append(o) return zorder_map
compute_overlayable_zorders
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(obj, path=[]):\n path = path + [obj]\n zorder_map = defaultdict(list)\n if not isinstance(obj, DynamicMap):\n if isinstance(obj, CompositeOverlay):\n for z, o in enumerate(obj):\n zorder_map[z] = [o, obj]\n elif isinstance(obj, HoloMap):\n for el in obj.values():\n if isinstance(el, CompositeOverlay):\n for k, v in compute_overlayable_zorders(el, path).items():\n zorder_map[k] += v + [obj]\n else:\n zorder_map[0] += [obj, el]\n elif obj not in zorder_map[0]:\n zorder_map[0].append(obj)\n return zorder_map\n isoverlay = isinstance(obj.last, CompositeOverlay)\n isdynoverlay = obj.callback._is_overlay\n if obj not in zorder_map[0] and not isoverlay:\n zorder_map[0].append(obj)\n depth = overlay_depth(obj)\n dmap_inputs = obj.callback.inputs if obj.callback.link_inputs else []\n for z, inp in enumerate(dmap_inputs):\n no_zorder_increment = False\n if any(not (isoverlay_fn(p) or p.last is None) for p in path\n ) and isoverlay_fn(inp):\n no_zorder_increment = True\n input_depth = overlay_depth(inp)\n if (depth is not None and input_depth is not None and depth <\n input_depth):\n if depth > 1:\n continue\n else:\n no_zorder_increment = True\n z = z if isdynoverlay else 0\n deep_zorders = compute_overlayable_zorders(inp, path=path)\n offset = max(zorder_map.keys())\n for dz, objs in deep_zorders.items():\n global_z = offset + z if no_zorder_increment else offset + dz + z\n zorder_map[global_z] = list(unique_iterator(zorder_map[global_z\n ] + objs))\n found = any(isinstance(p, DynamicMap) and p.callback._is_overlay for p in\n path)\n linked = any(isinstance(s, LinkedStream) and s.linked for s in obj.streams)\n if (found or linked) and isoverlay and not isdynoverlay:\n offset = max(zorder_map.keys())\n for z, o in enumerate(obj.last):\n if isoverlay and linked:\n zorder_map[offset + z].append(obj)\n if o not in zorder_map[offset + z]:\n zorder_map[offset + z].append(o)\n return zorder_map\n", "role": "user" }, { "content": "compute_overlayable_zorders", "role": "assistant" } ]
def x(dmap): if not isinstance(dmap, DynamicMap): return False elif dmap.callback._is_overlay: return True else: return any(is_dynamic_overlay(dm) for dm in dmap.callback.inputs)
is_dynamic_overlay
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dmap):\n if not isinstance(dmap, DynamicMap):\n return False\n elif dmap.callback._is_overlay:\n return True\n else:\n return any(is_dynamic_overlay(dm) for dm in dmap.callback.inputs)\n", "role": "user" }, { "content": "is_dynamic_overlay", "role": "assistant" } ]
def x(obj, depth=0): layers = [] if isinstance(obj, DynamicMap): if issubclass(obj.type, NdOverlay) and not depth: for v in obj.last.values(): layers.append(obj) elif issubclass(obj.type, Overlay): if obj.callback.inputs and is_dynamic_overlay(obj): for inp in obj.callback.inputs: layers += split_dmap_overlay(inp, depth + 1) else: for v in obj.last.values(): layers.append(obj) else: layers.append(obj) return layers if isinstance(obj, Overlay): for k, v in obj.items(): layers.append(v) else: layers.append(obj) return layers
split_dmap_overlay
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(obj, depth=0):\n layers = []\n if isinstance(obj, DynamicMap):\n if issubclass(obj.type, NdOverlay) and not depth:\n for v in obj.last.values():\n layers.append(obj)\n elif issubclass(obj.type, Overlay):\n if obj.callback.inputs and is_dynamic_overlay(obj):\n for inp in obj.callback.inputs:\n layers += split_dmap_overlay(inp, depth + 1)\n else:\n for v in obj.last.values():\n layers.append(obj)\n else:\n layers.append(obj)\n return layers\n if isinstance(obj, Overlay):\n for k, v in obj.items():\n layers.append(v)\n else:\n layers.append(obj)\n return layers\n", "role": "user" }, { "content": "split_dmap_overlay", "role": "assistant" } ]
def x(obj): dmaps = obj.traverse(lambda x: x, specs=[DynamicMap]) for dmap in dmaps: if dmap.unbounded: continue if not len(dmap): dmap[dmap._initial_key()]
initialize_dynamic
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(obj):\n dmaps = obj.traverse(lambda x: x, specs=[DynamicMap])\n for dmap in dmaps:\n if dmap.unbounded:\n continue\n if not len(dmap):\n dmap[dmap._initial_key()]\n", "role": "user" }, { "content": "initialize_dynamic", "role": "assistant" } ]
def x(map_obj, key_map, cached=False): if map_obj.kdims and len(map_obj.kdims) == 1 and map_obj.kdims[0 ] == 'Frame': return map_obj.last key = tuple(key_map[kd.name] for kd in map_obj.kdims if kd.name in key_map) if key in map_obj.data and cached: return map_obj.data[key] else: try: return map_obj[key] except KeyError: return None except StopIteration as e: raise e except Exception: print(traceback.format_exc()) return None
get_plot_frame
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(map_obj, key_map, cached=False):\n if map_obj.kdims and len(map_obj.kdims) == 1 and map_obj.kdims[0\n ] == 'Frame':\n return map_obj.last\n key = tuple(key_map[kd.name] for kd in map_obj.kdims if kd.name in key_map)\n if key in map_obj.data and cached:\n return map_obj.data[key]\n else:\n try:\n return map_obj[key]\n except KeyError:\n return None\n except StopIteration as e:\n raise e\n except Exception:\n print(traceback.format_exc())\n return None\n", "role": "user" }, { "content": "get_plot_frame", "role": "assistant" } ]
def x(obj, key_map, cached=False): clone = obj.map(lambda x: x) for it1, it2 in zip(obj.traverse(lambda x: x), clone.traverse(lambda x: x) ): if isinstance(it1, DynamicMap): with disable_constant(it2.callback): it2.callback.inputs = it1.callback.inputs with item_check(False): return clone.map(lambda x: get_plot_frame(x, key_map, cached=cached ), [DynamicMap, HoloMap], clone=False)
get_nested_plot_frame
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(obj, key_map, cached=False):\n clone = obj.map(lambda x: x)\n for it1, it2 in zip(obj.traverse(lambda x: x), clone.traverse(lambda x: x)\n ):\n if isinstance(it1, DynamicMap):\n with disable_constant(it2.callback):\n it2.callback.inputs = it1.callback.inputs\n with item_check(False):\n return clone.map(lambda x: get_plot_frame(x, key_map, cached=cached\n ), [DynamicMap, HoloMap], clone=False)\n", "role": "user" }, { "content": "get_nested_plot_frame", "role": "assistant" } ]
def x(obj, html=False): collate = '<tt>collate</tt>' if html else 'collate' info = ( 'For more information, please consult the Composing Data tutorial (http://git.io/vtIQh)' ) if isinstance(obj, HoloMap): error = ('HoloMap of %s objects cannot be displayed.' % obj.type. __name__) remedy = ( 'Please call the %s method to generate a displayable object' % collate) elif isinstance(obj, Layout): error = ( 'Layout containing HoloMaps of Layout or GridSpace objects cannot be displayed.' ) remedy = ('Please call the %s method on the appropriate elements.' % collate) elif isinstance(obj, GridSpace): error = 'GridSpace containing HoloMaps of Layouts cannot be displayed.' remedy = ('Please call the %s method on the appropriate elements.' % collate) if not html: return '\n'.join([error, remedy, info]) else: return '<center>{msg}</center>'.format(msg='<br>'.join(['<b>%s</b>' % error, remedy, '<i>%s</i>' % info]))
undisplayable_info
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(obj, html=False):\n collate = '<tt>collate</tt>' if html else 'collate'\n info = (\n 'For more information, please consult the Composing Data tutorial (http://git.io/vtIQh)'\n )\n if isinstance(obj, HoloMap):\n error = ('HoloMap of %s objects cannot be displayed.' % obj.type.\n __name__)\n remedy = (\n 'Please call the %s method to generate a displayable object' %\n collate)\n elif isinstance(obj, Layout):\n error = (\n 'Layout containing HoloMaps of Layout or GridSpace objects cannot be displayed.'\n )\n remedy = ('Please call the %s method on the appropriate elements.' %\n collate)\n elif isinstance(obj, GridSpace):\n error = 'GridSpace containing HoloMaps of Layouts cannot be displayed.'\n remedy = ('Please call the %s method on the appropriate elements.' %\n collate)\n if not html:\n return '\\n'.join([error, remedy, info])\n else:\n return '<center>{msg}</center>'.format(msg='<br>'.join(['<b>%s</b>' %\n error, remedy, '<i>%s</i>' % info]))\n", "role": "user" }, { "content": "undisplayable_info", "role": "assistant" } ]
def x(sizes, size_fn, scaling_factor, scaling_method, base_size): if sizes.dtype.kind not in ('i', 'f'): return None if scaling_method == 'area': pass elif scaling_method == 'width': scaling_factor = scaling_factor ** 2 else: raise ValueError( 'Invalid value for argument "scaling_method": "{}". Valid values are: "width", "area".' .format(scaling_method)) sizes = size_fn(sizes) return base_size * scaling_factor * sizes
compute_sizes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sizes, size_fn, scaling_factor, scaling_method, base_size):\n if sizes.dtype.kind not in ('i', 'f'):\n return None\n if scaling_method == 'area':\n pass\n elif scaling_method == 'width':\n scaling_factor = scaling_factor ** 2\n else:\n raise ValueError(\n 'Invalid value for argument \"scaling_method\": \"{}\". Valid values are: \"width\", \"area\".'\n .format(scaling_method))\n sizes = size_fn(sizes)\n return base_size * scaling_factor * sizes\n", "role": "user" }, { "content": "compute_sizes", "role": "assistant" } ]
def x(padding): if isinstance(padding, tuple): if len(padding) == 2: xpad, ypad = padding zpad = 0 elif len(padding) == 3: xpad, ypad, zpad = padding else: raise ValueError( 'Padding must be supplied as an number applied to all axes or a length two or three tuple corresponding to the x-, y- and optionally z-axis' ) else: xpad, ypad, zpad = (padding,) * 3 return xpad, ypad, zpad
get_axis_padding
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(padding):\n if isinstance(padding, tuple):\n if len(padding) == 2:\n xpad, ypad = padding\n zpad = 0\n elif len(padding) == 3:\n xpad, ypad, zpad = padding\n else:\n raise ValueError(\n 'Padding must be supplied as an number applied to all axes or a length two or three tuple corresponding to the x-, y- and optionally z-axis'\n )\n else:\n xpad, ypad, zpad = (padding,) * 3\n return xpad, ypad, zpad\n", "role": "user" }, { "content": "get_axis_padding", "role": "assistant" } ]
def x(low, high, span): if is_number(low) and low == high: if isinstance(low, np.datetime64): span = span * np.timedelta64(1, 's') low, high = low - span, high + span return low, high
get_minimum_span
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(low, high, span):\n if is_number(low) and low == high:\n if isinstance(low, np.datetime64):\n span = span * np.timedelta64(1, 's')\n low, high = low - span, high + span\n return low, high\n", "role": "user" }, { "content": "get_minimum_span", "role": "assistant" } ]