id
int32
0
252k
repo
stringlengths
7
55
path
stringlengths
4
127
func_name
stringlengths
1
88
original_string
stringlengths
75
19.8k
language
stringclasses
1 value
code
stringlengths
75
19.8k
code_tokens
list
docstring
stringlengths
3
17.3k
docstring_tokens
list
sha
stringlengths
40
40
url
stringlengths
87
242
22,300
tensorflow/tensor2tensor
tensor2tensor/models/research/attention_lm_moe.py
attention_lm_moe_memory_efficient
def attention_lm_moe_memory_efficient(): """Memory-efficient version.""" hparams = attention_lm_moe_large() hparams.diet_experts = True hparams.layer_preprocess_sequence = "n" hparams.layer_postprocess_sequence = "da" hparams.layer_prepostprocess_dropout = 0.0 hparams.memory_efficient_ffn = True hparams.attention_type = AttentionType.MEMORY_EFFICIENT hparams.num_heads = 8 hparams.factored_logits = True return hparams
python
def attention_lm_moe_memory_efficient(): """Memory-efficient version.""" hparams = attention_lm_moe_large() hparams.diet_experts = True hparams.layer_preprocess_sequence = "n" hparams.layer_postprocess_sequence = "da" hparams.layer_prepostprocess_dropout = 0.0 hparams.memory_efficient_ffn = True hparams.attention_type = AttentionType.MEMORY_EFFICIENT hparams.num_heads = 8 hparams.factored_logits = True return hparams
[ "def", "attention_lm_moe_memory_efficient", "(", ")", ":", "hparams", "=", "attention_lm_moe_large", "(", ")", "hparams", ".", "diet_experts", "=", "True", "hparams", ".", "layer_preprocess_sequence", "=", "\"n\"", "hparams", ".", "layer_postprocess_sequence", "=", "\"da\"", "hparams", ".", "layer_prepostprocess_dropout", "=", "0.0", "hparams", ".", "memory_efficient_ffn", "=", "True", "hparams", ".", "attention_type", "=", "AttentionType", ".", "MEMORY_EFFICIENT", "hparams", ".", "num_heads", "=", "8", "hparams", ".", "factored_logits", "=", "True", "return", "hparams" ]
Memory-efficient version.
[ "Memory", "-", "efficient", "version", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/attention_lm_moe.py#L733-L744
22,301
tensorflow/tensor2tensor
tensor2tensor/models/research/attention_lm_moe.py
attention_lm_moe_24b_diet
def attention_lm_moe_24b_diet(): """Unnecessarily large model with 24B params - because we can.""" hparams = attention_lm_moe_large_diet() hparams.moe_hidden_sizes = "12288" hparams.moe_num_experts = 1024 hparams.batch_size = 4096 return hparams
python
def attention_lm_moe_24b_diet(): """Unnecessarily large model with 24B params - because we can.""" hparams = attention_lm_moe_large_diet() hparams.moe_hidden_sizes = "12288" hparams.moe_num_experts = 1024 hparams.batch_size = 4096 return hparams
[ "def", "attention_lm_moe_24b_diet", "(", ")", ":", "hparams", "=", "attention_lm_moe_large_diet", "(", ")", "hparams", ".", "moe_hidden_sizes", "=", "\"12288\"", "hparams", ".", "moe_num_experts", "=", "1024", "hparams", ".", "batch_size", "=", "4096", "return", "hparams" ]
Unnecessarily large model with 24B params - because we can.
[ "Unnecessarily", "large", "model", "with", "24B", "params", "-", "because", "we", "can", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/attention_lm_moe.py#L757-L763
22,302
tensorflow/tensor2tensor
tensor2tensor/models/research/attention_lm_moe.py
attention_lm_moe_unscramble_base
def attention_lm_moe_unscramble_base(): """Version to use with languagemodel_wiki_scramble1k50.""" hparams = attention_lm_no_moe_small() hparams.use_inputs = True hparams.min_length_bucket = 1024 hparams.max_length = 1024 hparams.batch_size = 5000 hparams.layer_prepostprocess_dropout = 0.0 hparams.layer_preprocess_sequence = "n" hparams.layer_postprocess_sequence = "da" return hparams
python
def attention_lm_moe_unscramble_base(): """Version to use with languagemodel_wiki_scramble1k50.""" hparams = attention_lm_no_moe_small() hparams.use_inputs = True hparams.min_length_bucket = 1024 hparams.max_length = 1024 hparams.batch_size = 5000 hparams.layer_prepostprocess_dropout = 0.0 hparams.layer_preprocess_sequence = "n" hparams.layer_postprocess_sequence = "da" return hparams
[ "def", "attention_lm_moe_unscramble_base", "(", ")", ":", "hparams", "=", "attention_lm_no_moe_small", "(", ")", "hparams", ".", "use_inputs", "=", "True", "hparams", ".", "min_length_bucket", "=", "1024", "hparams", ".", "max_length", "=", "1024", "hparams", ".", "batch_size", "=", "5000", "hparams", ".", "layer_prepostprocess_dropout", "=", "0.0", "hparams", ".", "layer_preprocess_sequence", "=", "\"n\"", "hparams", ".", "layer_postprocess_sequence", "=", "\"da\"", "return", "hparams" ]
Version to use with languagemodel_wiki_scramble1k50.
[ "Version", "to", "use", "with", "languagemodel_wiki_scramble1k50", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/attention_lm_moe.py#L784-L794
22,303
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
audio_bottom
def audio_bottom(x, model_hparams, vocab_size): """Transform input from data space to model space. Args: x: A Tensor with shape [batch, ...] model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: body_input: A Tensor with shape [batch, ?, ?, model_hparams.hidden_size]. """ del vocab_size # unused arg inputs = x with tf.variable_scope("audio_modality"): # TODO(aidangomez): Will need to sort out a better audio pipeline def xnet_resblock(x, filters, res_relu, name): """Xception block.""" with tf.variable_scope(name): # Typically audio samples are >100k samples in length and have a width # of 2 or 4. Mono audio has a single channel while stereo has 2. y = common_layers.separable_conv_block( x, filters, [((1, 1), (3, 3)), ((1, 1), (3, 3))], first_relu=True, padding="SAME", force2d=True, name="sep_conv_block") y = common_layers.pool(y, (3, 3), "MAX", "SAME", strides=(2, 2)) return y + common_layers.conv_block( x, filters, [((1, 1), (1, 1))], padding="SAME", strides=(2, 2), first_relu=res_relu, force2d=True, name="res_conv0") x = tf.to_float(inputs) / 255. x.set_shape([None, None, None, 1]) for i in range(model_hparams.audio_compression): x = xnet_resblock(x, 2**(i + 1), True, "compress_block_%d" % i) return xnet_resblock(x, model_hparams.hidden_size, False, "compress_block_final")
python
def audio_bottom(x, model_hparams, vocab_size): """Transform input from data space to model space. Args: x: A Tensor with shape [batch, ...] model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: body_input: A Tensor with shape [batch, ?, ?, model_hparams.hidden_size]. """ del vocab_size # unused arg inputs = x with tf.variable_scope("audio_modality"): # TODO(aidangomez): Will need to sort out a better audio pipeline def xnet_resblock(x, filters, res_relu, name): """Xception block.""" with tf.variable_scope(name): # Typically audio samples are >100k samples in length and have a width # of 2 or 4. Mono audio has a single channel while stereo has 2. y = common_layers.separable_conv_block( x, filters, [((1, 1), (3, 3)), ((1, 1), (3, 3))], first_relu=True, padding="SAME", force2d=True, name="sep_conv_block") y = common_layers.pool(y, (3, 3), "MAX", "SAME", strides=(2, 2)) return y + common_layers.conv_block( x, filters, [((1, 1), (1, 1))], padding="SAME", strides=(2, 2), first_relu=res_relu, force2d=True, name="res_conv0") x = tf.to_float(inputs) / 255. x.set_shape([None, None, None, 1]) for i in range(model_hparams.audio_compression): x = xnet_resblock(x, 2**(i + 1), True, "compress_block_%d" % i) return xnet_resblock(x, model_hparams.hidden_size, False, "compress_block_final")
[ "def", "audio_bottom", "(", "x", ",", "model_hparams", ",", "vocab_size", ")", ":", "del", "vocab_size", "# unused arg", "inputs", "=", "x", "with", "tf", ".", "variable_scope", "(", "\"audio_modality\"", ")", ":", "# TODO(aidangomez): Will need to sort out a better audio pipeline", "def", "xnet_resblock", "(", "x", ",", "filters", ",", "res_relu", ",", "name", ")", ":", "\"\"\"Xception block.\"\"\"", "with", "tf", ".", "variable_scope", "(", "name", ")", ":", "# Typically audio samples are >100k samples in length and have a width", "# of 2 or 4. Mono audio has a single channel while stereo has 2.", "y", "=", "common_layers", ".", "separable_conv_block", "(", "x", ",", "filters", ",", "[", "(", "(", "1", ",", "1", ")", ",", "(", "3", ",", "3", ")", ")", ",", "(", "(", "1", ",", "1", ")", ",", "(", "3", ",", "3", ")", ")", "]", ",", "first_relu", "=", "True", ",", "padding", "=", "\"SAME\"", ",", "force2d", "=", "True", ",", "name", "=", "\"sep_conv_block\"", ")", "y", "=", "common_layers", ".", "pool", "(", "y", ",", "(", "3", ",", "3", ")", ",", "\"MAX\"", ",", "\"SAME\"", ",", "strides", "=", "(", "2", ",", "2", ")", ")", "return", "y", "+", "common_layers", ".", "conv_block", "(", "x", ",", "filters", ",", "[", "(", "(", "1", ",", "1", ")", ",", "(", "1", ",", "1", ")", ")", "]", ",", "padding", "=", "\"SAME\"", ",", "strides", "=", "(", "2", ",", "2", ")", ",", "first_relu", "=", "res_relu", ",", "force2d", "=", "True", ",", "name", "=", "\"res_conv0\"", ")", "x", "=", "tf", ".", "to_float", "(", "inputs", ")", "/", "255.", "x", ".", "set_shape", "(", "[", "None", ",", "None", ",", "None", ",", "1", "]", ")", "for", "i", "in", "range", "(", "model_hparams", ".", "audio_compression", ")", ":", "x", "=", "xnet_resblock", "(", "x", ",", "2", "**", "(", "i", "+", "1", ")", ",", "True", ",", "\"compress_block_%d\"", "%", "i", ")", "return", "xnet_resblock", "(", "x", ",", "model_hparams", ".", "hidden_size", ",", "False", ",", "\"compress_block_final\"", ")" ]
Transform input from data space to model space. Args: x: A Tensor with shape [batch, ...] model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: body_input: A Tensor with shape [batch, ?, ?, model_hparams.hidden_size].
[ "Transform", "input", "from", "data", "space", "to", "model", "space", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L128-L173
22,304
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
image_targets_bottom
def image_targets_bottom(x, model_hparams, vocab_size): """Bottom transformation for target images.""" pixel_embedding_size = 64 inputs = x with tf.variable_scope("image_modality"): if not tf.executing_eagerly(): tf.summary.image( "targets_bottom", common_layers.tpu_safe_image_summary(inputs), max_outputs=1) inputs_shape = common_layers.shape_list(inputs) if len(inputs_shape) != 4: raise ValueError("Assuming images given as int tensors in the format " "[batch, height, width, channels] (256 values).") # We embed each of 256=vocab_size possible pixel values. embedding_var = tf.get_variable( "pixel_embedding", [vocab_size, pixel_embedding_size]) hot_inputs = tf.one_hot(tf.to_int32(inputs), vocab_size) hot_inputs = tf.reshape(hot_inputs, [-1, vocab_size]) embedded = tf.matmul(hot_inputs, embedding_var) # Let's now merge all channels that were embedded into a single vector. merged_size = pixel_embedding_size * inputs_shape[3] embedded = tf.reshape(embedded, inputs_shape[:3] + [merged_size]) merged = tf.layers.dense( embedded, model_hparams.hidden_size, name="merge_pixel_embedded_channels") return merged
python
def image_targets_bottom(x, model_hparams, vocab_size): """Bottom transformation for target images.""" pixel_embedding_size = 64 inputs = x with tf.variable_scope("image_modality"): if not tf.executing_eagerly(): tf.summary.image( "targets_bottom", common_layers.tpu_safe_image_summary(inputs), max_outputs=1) inputs_shape = common_layers.shape_list(inputs) if len(inputs_shape) != 4: raise ValueError("Assuming images given as int tensors in the format " "[batch, height, width, channels] (256 values).") # We embed each of 256=vocab_size possible pixel values. embedding_var = tf.get_variable( "pixel_embedding", [vocab_size, pixel_embedding_size]) hot_inputs = tf.one_hot(tf.to_int32(inputs), vocab_size) hot_inputs = tf.reshape(hot_inputs, [-1, vocab_size]) embedded = tf.matmul(hot_inputs, embedding_var) # Let's now merge all channels that were embedded into a single vector. merged_size = pixel_embedding_size * inputs_shape[3] embedded = tf.reshape(embedded, inputs_shape[:3] + [merged_size]) merged = tf.layers.dense( embedded, model_hparams.hidden_size, name="merge_pixel_embedded_channels") return merged
[ "def", "image_targets_bottom", "(", "x", ",", "model_hparams", ",", "vocab_size", ")", ":", "pixel_embedding_size", "=", "64", "inputs", "=", "x", "with", "tf", ".", "variable_scope", "(", "\"image_modality\"", ")", ":", "if", "not", "tf", ".", "executing_eagerly", "(", ")", ":", "tf", ".", "summary", ".", "image", "(", "\"targets_bottom\"", ",", "common_layers", ".", "tpu_safe_image_summary", "(", "inputs", ")", ",", "max_outputs", "=", "1", ")", "inputs_shape", "=", "common_layers", ".", "shape_list", "(", "inputs", ")", "if", "len", "(", "inputs_shape", ")", "!=", "4", ":", "raise", "ValueError", "(", "\"Assuming images given as int tensors in the format \"", "\"[batch, height, width, channels] (256 values).\"", ")", "# We embed each of 256=vocab_size possible pixel values.", "embedding_var", "=", "tf", ".", "get_variable", "(", "\"pixel_embedding\"", ",", "[", "vocab_size", ",", "pixel_embedding_size", "]", ")", "hot_inputs", "=", "tf", ".", "one_hot", "(", "tf", ".", "to_int32", "(", "inputs", ")", ",", "vocab_size", ")", "hot_inputs", "=", "tf", ".", "reshape", "(", "hot_inputs", ",", "[", "-", "1", ",", "vocab_size", "]", ")", "embedded", "=", "tf", ".", "matmul", "(", "hot_inputs", ",", "embedding_var", ")", "# Let's now merge all channels that were embedded into a single vector.", "merged_size", "=", "pixel_embedding_size", "*", "inputs_shape", "[", "3", "]", "embedded", "=", "tf", ".", "reshape", "(", "embedded", ",", "inputs_shape", "[", ":", "3", "]", "+", "[", "merged_size", "]", ")", "merged", "=", "tf", ".", "layers", ".", "dense", "(", "embedded", ",", "model_hparams", ".", "hidden_size", ",", "name", "=", "\"merge_pixel_embedded_channels\"", ")", "return", "merged" ]
Bottom transformation for target images.
[ "Bottom", "transformation", "for", "target", "images", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L260-L288
22,305
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
_image_channel_compress_bottom
def _image_channel_compress_bottom(inputs, model_hparams, name="bottom"): """Compresses channel-wise input pixels into whole pixel representions. Perform conversion of RGB pixel values to a real number in the range -1 to 1. This combines pixel channels to form a representation of shape [img_len, img_len]. Args: inputs: Tensor representing RGB pixel intensities as integers, of shape [batch, img_len, img_len, channels]. model_hparams: HParams, model hyperparmeters. name: string, scope. Returns: body_input: Tensor of shape [batch, img_len, img_len, model_hparams.hidden_size]. """ num_channels = 3 with tf.variable_scope(name): inputs = tf.to_float(inputs) hp = model_hparams if hp.mode != tf.estimator.ModeKeys.PREDICT: tf.summary.image( "inputs", common_layers.tpu_safe_image_summary(inputs), max_outputs=2) inputs = common_layers.convert_rgb_to_symmetric_real(inputs) # Reshape inputs to apply convolutions across [img_len, img_len*channels]. inputs_shape = common_layers.shape_list(inputs) inputs = tf.reshape( inputs, [-1, inputs_shape[1], inputs_shape[2] * inputs_shape[3], 1]) # Compress RGB intensities for each pixel using a convolution. outputs = tf.layers.conv2d( inputs, model_hparams.hidden_size, kernel_size=(1, num_channels), padding="VALID", strides=(1, num_channels), activation=tf.nn.relu, name="conv_input") return outputs
python
def _image_channel_compress_bottom(inputs, model_hparams, name="bottom"): """Compresses channel-wise input pixels into whole pixel representions. Perform conversion of RGB pixel values to a real number in the range -1 to 1. This combines pixel channels to form a representation of shape [img_len, img_len]. Args: inputs: Tensor representing RGB pixel intensities as integers, of shape [batch, img_len, img_len, channels]. model_hparams: HParams, model hyperparmeters. name: string, scope. Returns: body_input: Tensor of shape [batch, img_len, img_len, model_hparams.hidden_size]. """ num_channels = 3 with tf.variable_scope(name): inputs = tf.to_float(inputs) hp = model_hparams if hp.mode != tf.estimator.ModeKeys.PREDICT: tf.summary.image( "inputs", common_layers.tpu_safe_image_summary(inputs), max_outputs=2) inputs = common_layers.convert_rgb_to_symmetric_real(inputs) # Reshape inputs to apply convolutions across [img_len, img_len*channels]. inputs_shape = common_layers.shape_list(inputs) inputs = tf.reshape( inputs, [-1, inputs_shape[1], inputs_shape[2] * inputs_shape[3], 1]) # Compress RGB intensities for each pixel using a convolution. outputs = tf.layers.conv2d( inputs, model_hparams.hidden_size, kernel_size=(1, num_channels), padding="VALID", strides=(1, num_channels), activation=tf.nn.relu, name="conv_input") return outputs
[ "def", "_image_channel_compress_bottom", "(", "inputs", ",", "model_hparams", ",", "name", "=", "\"bottom\"", ")", ":", "num_channels", "=", "3", "with", "tf", ".", "variable_scope", "(", "name", ")", ":", "inputs", "=", "tf", ".", "to_float", "(", "inputs", ")", "hp", "=", "model_hparams", "if", "hp", ".", "mode", "!=", "tf", ".", "estimator", ".", "ModeKeys", ".", "PREDICT", ":", "tf", ".", "summary", ".", "image", "(", "\"inputs\"", ",", "common_layers", ".", "tpu_safe_image_summary", "(", "inputs", ")", ",", "max_outputs", "=", "2", ")", "inputs", "=", "common_layers", ".", "convert_rgb_to_symmetric_real", "(", "inputs", ")", "# Reshape inputs to apply convolutions across [img_len, img_len*channels].", "inputs_shape", "=", "common_layers", ".", "shape_list", "(", "inputs", ")", "inputs", "=", "tf", ".", "reshape", "(", "inputs", ",", "[", "-", "1", ",", "inputs_shape", "[", "1", "]", ",", "inputs_shape", "[", "2", "]", "*", "inputs_shape", "[", "3", "]", ",", "1", "]", ")", "# Compress RGB intensities for each pixel using a convolution.", "outputs", "=", "tf", ".", "layers", ".", "conv2d", "(", "inputs", ",", "model_hparams", ".", "hidden_size", ",", "kernel_size", "=", "(", "1", ",", "num_channels", ")", ",", "padding", "=", "\"VALID\"", ",", "strides", "=", "(", "1", ",", "num_channels", ")", ",", "activation", "=", "tf", ".", "nn", ".", "relu", ",", "name", "=", "\"conv_input\"", ")", "return", "outputs" ]
Compresses channel-wise input pixels into whole pixel representions. Perform conversion of RGB pixel values to a real number in the range -1 to 1. This combines pixel channels to form a representation of shape [img_len, img_len]. Args: inputs: Tensor representing RGB pixel intensities as integers, of shape [batch, img_len, img_len, channels]. model_hparams: HParams, model hyperparmeters. name: string, scope. Returns: body_input: Tensor of shape [batch, img_len, img_len, model_hparams.hidden_size].
[ "Compresses", "channel", "-", "wise", "input", "pixels", "into", "whole", "pixel", "representions", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L291-L333
22,306
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
image_channel_embeddings_bottom
def image_channel_embeddings_bottom(x, model_hparams, vocab_size): """Bottom transformation for image targets.""" del vocab_size # unused arg inputs = tf.to_int32(x) io_depth = model_hparams.num_channels tshape = common_layers.shape_list(inputs) hidden_size = model_hparams.hidden_size target_embeddings = cia.get_channel_embeddings( io_depth, inputs, hidden_size, "input_bottom") return tf.reshape(target_embeddings, [tshape[0], tshape[1], tshape[2] * io_depth, hidden_size])
python
def image_channel_embeddings_bottom(x, model_hparams, vocab_size): """Bottom transformation for image targets.""" del vocab_size # unused arg inputs = tf.to_int32(x) io_depth = model_hparams.num_channels tshape = common_layers.shape_list(inputs) hidden_size = model_hparams.hidden_size target_embeddings = cia.get_channel_embeddings( io_depth, inputs, hidden_size, "input_bottom") return tf.reshape(target_embeddings, [tshape[0], tshape[1], tshape[2] * io_depth, hidden_size])
[ "def", "image_channel_embeddings_bottom", "(", "x", ",", "model_hparams", ",", "vocab_size", ")", ":", "del", "vocab_size", "# unused arg", "inputs", "=", "tf", ".", "to_int32", "(", "x", ")", "io_depth", "=", "model_hparams", ".", "num_channels", "tshape", "=", "common_layers", ".", "shape_list", "(", "inputs", ")", "hidden_size", "=", "model_hparams", ".", "hidden_size", "target_embeddings", "=", "cia", ".", "get_channel_embeddings", "(", "io_depth", ",", "inputs", ",", "hidden_size", ",", "\"input_bottom\"", ")", "return", "tf", ".", "reshape", "(", "target_embeddings", ",", "[", "tshape", "[", "0", "]", ",", "tshape", "[", "1", "]", ",", "tshape", "[", "2", "]", "*", "io_depth", ",", "hidden_size", "]", ")" ]
Bottom transformation for image targets.
[ "Bottom", "transformation", "for", "image", "targets", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L346-L356
22,307
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
speech_recognition_bottom
def speech_recognition_bottom(x, model_hparams, vocab_size): """Use batchnorm instead of CMVN and shorten the stft with strided convs. Args: x: float32 tensor with shape [batch_size, len, 1, freqs * channels] model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: float32 tensor with shape [batch_size, shorter_len, 1, hidden_size] """ del vocab_size # unused arg inputs = x p = model_hparams num_mel_bins = p.audio_num_mel_bins num_channels = 3 if p.audio_add_delta_deltas else 1 with tf.variable_scope("speech_recognition_modality"): if p.audio_preproc_in_bottom: # Compute filterbanks with tf.variable_scope("fbanks"): waveforms = tf.squeeze(inputs, [2, 3]) mel_fbanks = common_audio.compute_mel_filterbank_features( waveforms, sample_rate=p.audio_sample_rate, dither=p.audio_dither, preemphasis=p.audio_preemphasis, frame_length=p.audio_frame_length, frame_step=p.audio_frame_step, lower_edge_hertz=p.audio_lower_edge_hertz, upper_edge_hertz=p.audio_upper_edge_hertz, num_mel_bins=p.audio_num_mel_bins, apply_mask=True) if p.audio_add_delta_deltas: mel_fbanks = common_audio.add_delta_deltas(mel_fbanks) x = tf.reshape(mel_fbanks, common_layers.shape_list(mel_fbanks)[:2] + [num_mel_bins, num_channels]) nonpadding_mask = 1. - common_attention.embedding_to_padding(x) num_of_nonpadding_elements = tf.reduce_sum( nonpadding_mask) * num_mel_bins * num_channels # This replaces CMVN estimation on data var_epsilon = 1e-09 mean = tf.reduce_sum( x, axis=[1], keepdims=True) / num_of_nonpadding_elements variance = (num_of_nonpadding_elements * mean**2. - 2. * mean * tf.reduce_sum(x, axis=[1], keepdims=True) + tf.reduce_sum(x**2, axis=[1], keepdims=True) ) / num_of_nonpadding_elements x = (x - mean) * tf.rsqrt(variance + var_epsilon) * tf.expand_dims( nonpadding_mask, -1) else: x = inputs # The convention is that the models are flattened along the spatial, # dimensions, thus the speech preprocessor treats frequencies and # channels as image colors (last axis) x.set_shape([None, None, num_mel_bins, num_channels]) # TODO(chorowski): how to specify bottom's hparams and avoid hardcoding? x = tf.pad(x, [[0, 0], [0, 8], [0, 0], [0, 0]]) for _ in range(2): x = tf.layers.conv2d( x, 128, (3, 3), (2, 2), use_bias=False) x = common_layers.layer_norm(x) x = tf.nn.relu(x) xshape = common_layers.shape_list(x) # apply a conv that will remove all frequencies and at the same time # project the output into desired hidden_size x = tf.pad(x, [[0, 0], [0, 2], [0, 0], [0, 0]]) x = tf.layers.conv2d(x, p.hidden_size, (3, xshape[2]), use_bias=False) assert common_layers.shape_list(x)[2] == 1 x = common_layers.layer_norm(x) x = tf.nn.relu(x) return x
python
def speech_recognition_bottom(x, model_hparams, vocab_size): """Use batchnorm instead of CMVN and shorten the stft with strided convs. Args: x: float32 tensor with shape [batch_size, len, 1, freqs * channels] model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: float32 tensor with shape [batch_size, shorter_len, 1, hidden_size] """ del vocab_size # unused arg inputs = x p = model_hparams num_mel_bins = p.audio_num_mel_bins num_channels = 3 if p.audio_add_delta_deltas else 1 with tf.variable_scope("speech_recognition_modality"): if p.audio_preproc_in_bottom: # Compute filterbanks with tf.variable_scope("fbanks"): waveforms = tf.squeeze(inputs, [2, 3]) mel_fbanks = common_audio.compute_mel_filterbank_features( waveforms, sample_rate=p.audio_sample_rate, dither=p.audio_dither, preemphasis=p.audio_preemphasis, frame_length=p.audio_frame_length, frame_step=p.audio_frame_step, lower_edge_hertz=p.audio_lower_edge_hertz, upper_edge_hertz=p.audio_upper_edge_hertz, num_mel_bins=p.audio_num_mel_bins, apply_mask=True) if p.audio_add_delta_deltas: mel_fbanks = common_audio.add_delta_deltas(mel_fbanks) x = tf.reshape(mel_fbanks, common_layers.shape_list(mel_fbanks)[:2] + [num_mel_bins, num_channels]) nonpadding_mask = 1. - common_attention.embedding_to_padding(x) num_of_nonpadding_elements = tf.reduce_sum( nonpadding_mask) * num_mel_bins * num_channels # This replaces CMVN estimation on data var_epsilon = 1e-09 mean = tf.reduce_sum( x, axis=[1], keepdims=True) / num_of_nonpadding_elements variance = (num_of_nonpadding_elements * mean**2. - 2. * mean * tf.reduce_sum(x, axis=[1], keepdims=True) + tf.reduce_sum(x**2, axis=[1], keepdims=True) ) / num_of_nonpadding_elements x = (x - mean) * tf.rsqrt(variance + var_epsilon) * tf.expand_dims( nonpadding_mask, -1) else: x = inputs # The convention is that the models are flattened along the spatial, # dimensions, thus the speech preprocessor treats frequencies and # channels as image colors (last axis) x.set_shape([None, None, num_mel_bins, num_channels]) # TODO(chorowski): how to specify bottom's hparams and avoid hardcoding? x = tf.pad(x, [[0, 0], [0, 8], [0, 0], [0, 0]]) for _ in range(2): x = tf.layers.conv2d( x, 128, (3, 3), (2, 2), use_bias=False) x = common_layers.layer_norm(x) x = tf.nn.relu(x) xshape = common_layers.shape_list(x) # apply a conv that will remove all frequencies and at the same time # project the output into desired hidden_size x = tf.pad(x, [[0, 0], [0, 2], [0, 0], [0, 0]]) x = tf.layers.conv2d(x, p.hidden_size, (3, xshape[2]), use_bias=False) assert common_layers.shape_list(x)[2] == 1 x = common_layers.layer_norm(x) x = tf.nn.relu(x) return x
[ "def", "speech_recognition_bottom", "(", "x", ",", "model_hparams", ",", "vocab_size", ")", ":", "del", "vocab_size", "# unused arg", "inputs", "=", "x", "p", "=", "model_hparams", "num_mel_bins", "=", "p", ".", "audio_num_mel_bins", "num_channels", "=", "3", "if", "p", ".", "audio_add_delta_deltas", "else", "1", "with", "tf", ".", "variable_scope", "(", "\"speech_recognition_modality\"", ")", ":", "if", "p", ".", "audio_preproc_in_bottom", ":", "# Compute filterbanks", "with", "tf", ".", "variable_scope", "(", "\"fbanks\"", ")", ":", "waveforms", "=", "tf", ".", "squeeze", "(", "inputs", ",", "[", "2", ",", "3", "]", ")", "mel_fbanks", "=", "common_audio", ".", "compute_mel_filterbank_features", "(", "waveforms", ",", "sample_rate", "=", "p", ".", "audio_sample_rate", ",", "dither", "=", "p", ".", "audio_dither", ",", "preemphasis", "=", "p", ".", "audio_preemphasis", ",", "frame_length", "=", "p", ".", "audio_frame_length", ",", "frame_step", "=", "p", ".", "audio_frame_step", ",", "lower_edge_hertz", "=", "p", ".", "audio_lower_edge_hertz", ",", "upper_edge_hertz", "=", "p", ".", "audio_upper_edge_hertz", ",", "num_mel_bins", "=", "p", ".", "audio_num_mel_bins", ",", "apply_mask", "=", "True", ")", "if", "p", ".", "audio_add_delta_deltas", ":", "mel_fbanks", "=", "common_audio", ".", "add_delta_deltas", "(", "mel_fbanks", ")", "x", "=", "tf", ".", "reshape", "(", "mel_fbanks", ",", "common_layers", ".", "shape_list", "(", "mel_fbanks", ")", "[", ":", "2", "]", "+", "[", "num_mel_bins", ",", "num_channels", "]", ")", "nonpadding_mask", "=", "1.", "-", "common_attention", ".", "embedding_to_padding", "(", "x", ")", "num_of_nonpadding_elements", "=", "tf", ".", "reduce_sum", "(", "nonpadding_mask", ")", "*", "num_mel_bins", "*", "num_channels", "# This replaces CMVN estimation on data", "var_epsilon", "=", "1e-09", "mean", "=", "tf", ".", "reduce_sum", "(", "x", ",", "axis", "=", "[", "1", "]", ",", "keepdims", "=", "True", ")", "/", "num_of_nonpadding_elements", "variance", "=", "(", "num_of_nonpadding_elements", "*", "mean", "**", "2.", "-", "2.", "*", "mean", "*", "tf", ".", "reduce_sum", "(", "x", ",", "axis", "=", "[", "1", "]", ",", "keepdims", "=", "True", ")", "+", "tf", ".", "reduce_sum", "(", "x", "**", "2", ",", "axis", "=", "[", "1", "]", ",", "keepdims", "=", "True", ")", ")", "/", "num_of_nonpadding_elements", "x", "=", "(", "x", "-", "mean", ")", "*", "tf", ".", "rsqrt", "(", "variance", "+", "var_epsilon", ")", "*", "tf", ".", "expand_dims", "(", "nonpadding_mask", ",", "-", "1", ")", "else", ":", "x", "=", "inputs", "# The convention is that the models are flattened along the spatial,", "# dimensions, thus the speech preprocessor treats frequencies and", "# channels as image colors (last axis)", "x", ".", "set_shape", "(", "[", "None", ",", "None", ",", "num_mel_bins", ",", "num_channels", "]", ")", "# TODO(chorowski): how to specify bottom's hparams and avoid hardcoding?", "x", "=", "tf", ".", "pad", "(", "x", ",", "[", "[", "0", ",", "0", "]", ",", "[", "0", ",", "8", "]", ",", "[", "0", ",", "0", "]", ",", "[", "0", ",", "0", "]", "]", ")", "for", "_", "in", "range", "(", "2", ")", ":", "x", "=", "tf", ".", "layers", ".", "conv2d", "(", "x", ",", "128", ",", "(", "3", ",", "3", ")", ",", "(", "2", ",", "2", ")", ",", "use_bias", "=", "False", ")", "x", "=", "common_layers", ".", "layer_norm", "(", "x", ")", "x", "=", "tf", ".", "nn", ".", "relu", "(", "x", ")", "xshape", "=", "common_layers", ".", "shape_list", "(", "x", ")", "# apply a conv that will remove all frequencies and at the same time", "# project the output into desired hidden_size", "x", "=", "tf", ".", "pad", "(", "x", ",", "[", "[", "0", ",", "0", "]", ",", "[", "0", ",", "2", "]", ",", "[", "0", ",", "0", "]", ",", "[", "0", ",", "0", "]", "]", ")", "x", "=", "tf", ".", "layers", ".", "conv2d", "(", "x", ",", "p", ".", "hidden_size", ",", "(", "3", ",", "xshape", "[", "2", "]", ")", ",", "use_bias", "=", "False", ")", "assert", "common_layers", ".", "shape_list", "(", "x", ")", "[", "2", "]", "==", "1", "x", "=", "common_layers", ".", "layer_norm", "(", "x", ")", "x", "=", "tf", ".", "nn", ".", "relu", "(", "x", ")", "return", "x" ]
Use batchnorm instead of CMVN and shorten the stft with strided convs. Args: x: float32 tensor with shape [batch_size, len, 1, freqs * channels] model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: float32 tensor with shape [batch_size, shorter_len, 1, hidden_size]
[ "Use", "batchnorm", "instead", "of", "CMVN", "and", "shorten", "the", "stft", "with", "strided", "convs", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L373-L452
22,308
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
get_weights
def get_weights(model_hparams, vocab_size, hidden_dim=None): """Create or get concatenated embedding or softmax variable. Args: model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. hidden_dim: dim of the variable. Defaults to _model_hparams' hidden_size Returns: a list of num_shards Tensors. """ if hidden_dim is None: hidden_dim = model_hparams.hidden_size num_shards = model_hparams.symbol_modality_num_shards shards = [] for i in range(num_shards): shard_size = (vocab_size // num_shards) + ( 1 if i < vocab_size % num_shards else 0) var_name = "weights_%d" % i shards.append( tf.get_variable( var_name, [shard_size, hidden_dim], initializer=tf.random_normal_initializer(0.0, hidden_dim**-0.5))) if num_shards == 1: ret = shards[0] else: ret = tf.concat(shards, 0) # Convert ret to tensor. if not tf.executing_eagerly(): ret = common_layers.convert_gradient_to_tensor(ret) return ret
python
def get_weights(model_hparams, vocab_size, hidden_dim=None): """Create or get concatenated embedding or softmax variable. Args: model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. hidden_dim: dim of the variable. Defaults to _model_hparams' hidden_size Returns: a list of num_shards Tensors. """ if hidden_dim is None: hidden_dim = model_hparams.hidden_size num_shards = model_hparams.symbol_modality_num_shards shards = [] for i in range(num_shards): shard_size = (vocab_size // num_shards) + ( 1 if i < vocab_size % num_shards else 0) var_name = "weights_%d" % i shards.append( tf.get_variable( var_name, [shard_size, hidden_dim], initializer=tf.random_normal_initializer(0.0, hidden_dim**-0.5))) if num_shards == 1: ret = shards[0] else: ret = tf.concat(shards, 0) # Convert ret to tensor. if not tf.executing_eagerly(): ret = common_layers.convert_gradient_to_tensor(ret) return ret
[ "def", "get_weights", "(", "model_hparams", ",", "vocab_size", ",", "hidden_dim", "=", "None", ")", ":", "if", "hidden_dim", "is", "None", ":", "hidden_dim", "=", "model_hparams", ".", "hidden_size", "num_shards", "=", "model_hparams", ".", "symbol_modality_num_shards", "shards", "=", "[", "]", "for", "i", "in", "range", "(", "num_shards", ")", ":", "shard_size", "=", "(", "vocab_size", "//", "num_shards", ")", "+", "(", "1", "if", "i", "<", "vocab_size", "%", "num_shards", "else", "0", ")", "var_name", "=", "\"weights_%d\"", "%", "i", "shards", ".", "append", "(", "tf", ".", "get_variable", "(", "var_name", ",", "[", "shard_size", ",", "hidden_dim", "]", ",", "initializer", "=", "tf", ".", "random_normal_initializer", "(", "0.0", ",", "hidden_dim", "**", "-", "0.5", ")", ")", ")", "if", "num_shards", "==", "1", ":", "ret", "=", "shards", "[", "0", "]", "else", ":", "ret", "=", "tf", ".", "concat", "(", "shards", ",", "0", ")", "# Convert ret to tensor.", "if", "not", "tf", ".", "executing_eagerly", "(", ")", ":", "ret", "=", "common_layers", ".", "convert_gradient_to_tensor", "(", "ret", ")", "return", "ret" ]
Create or get concatenated embedding or softmax variable. Args: model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. hidden_dim: dim of the variable. Defaults to _model_hparams' hidden_size Returns: a list of num_shards Tensors.
[ "Create", "or", "get", "concatenated", "embedding", "or", "softmax", "variable", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L455-L485
22,309
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
_symbol_bottom_simple
def _symbol_bottom_simple(x, model_hparams, vocab_size, name, reuse): """Bottom transformation for symbols.""" with tf.variable_scope(name, reuse=reuse): # Ensure the inputs are 3-D if len(x.get_shape()) == 4: x = tf.squeeze(x, axis=3) while len(x.get_shape()) < 3: x = tf.expand_dims(x, axis=-1) var = get_weights(model_hparams, vocab_size) x = common_layers.dropout_no_scaling( x, 1.0 - model_hparams.symbol_dropout) ret = common_layers.gather(var, x) if model_hparams.multiply_embedding_mode == "sqrt_depth": ret *= model_hparams.hidden_size**0.5 ret *= tf.expand_dims( common_layers.cast_like(tf.not_equal(x, 0), ret), -1) return ret
python
def _symbol_bottom_simple(x, model_hparams, vocab_size, name, reuse): """Bottom transformation for symbols.""" with tf.variable_scope(name, reuse=reuse): # Ensure the inputs are 3-D if len(x.get_shape()) == 4: x = tf.squeeze(x, axis=3) while len(x.get_shape()) < 3: x = tf.expand_dims(x, axis=-1) var = get_weights(model_hparams, vocab_size) x = common_layers.dropout_no_scaling( x, 1.0 - model_hparams.symbol_dropout) ret = common_layers.gather(var, x) if model_hparams.multiply_embedding_mode == "sqrt_depth": ret *= model_hparams.hidden_size**0.5 ret *= tf.expand_dims( common_layers.cast_like(tf.not_equal(x, 0), ret), -1) return ret
[ "def", "_symbol_bottom_simple", "(", "x", ",", "model_hparams", ",", "vocab_size", ",", "name", ",", "reuse", ")", ":", "with", "tf", ".", "variable_scope", "(", "name", ",", "reuse", "=", "reuse", ")", ":", "# Ensure the inputs are 3-D", "if", "len", "(", "x", ".", "get_shape", "(", ")", ")", "==", "4", ":", "x", "=", "tf", ".", "squeeze", "(", "x", ",", "axis", "=", "3", ")", "while", "len", "(", "x", ".", "get_shape", "(", ")", ")", "<", "3", ":", "x", "=", "tf", ".", "expand_dims", "(", "x", ",", "axis", "=", "-", "1", ")", "var", "=", "get_weights", "(", "model_hparams", ",", "vocab_size", ")", "x", "=", "common_layers", ".", "dropout_no_scaling", "(", "x", ",", "1.0", "-", "model_hparams", ".", "symbol_dropout", ")", "ret", "=", "common_layers", ".", "gather", "(", "var", ",", "x", ")", "if", "model_hparams", ".", "multiply_embedding_mode", "==", "\"sqrt_depth\"", ":", "ret", "*=", "model_hparams", ".", "hidden_size", "**", "0.5", "ret", "*=", "tf", ".", "expand_dims", "(", "common_layers", ".", "cast_like", "(", "tf", ".", "not_equal", "(", "x", ",", "0", ")", ",", "ret", ")", ",", "-", "1", ")", "return", "ret" ]
Bottom transformation for symbols.
[ "Bottom", "transformation", "for", "symbols", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L488-L505
22,310
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
symbol_targets_bottom
def symbol_targets_bottom(x, model_hparams, vocab_size): """Bottom transformation for target symbols.""" if (model_hparams.shared_embedding_and_softmax_weights or model_hparams.get("shared_embedding")): try: return _symbol_bottom_simple( x, model_hparams, vocab_size, "shared", reuse=True) except ValueError: # perhaps there were no inputs, and this is a new variable. return _symbol_bottom_simple( x, model_hparams, vocab_size, "shared", reuse=None) else: return _symbol_bottom_simple( x, model_hparams, vocab_size, "target_emb", reuse=None)
python
def symbol_targets_bottom(x, model_hparams, vocab_size): """Bottom transformation for target symbols.""" if (model_hparams.shared_embedding_and_softmax_weights or model_hparams.get("shared_embedding")): try: return _symbol_bottom_simple( x, model_hparams, vocab_size, "shared", reuse=True) except ValueError: # perhaps there were no inputs, and this is a new variable. return _symbol_bottom_simple( x, model_hparams, vocab_size, "shared", reuse=None) else: return _symbol_bottom_simple( x, model_hparams, vocab_size, "target_emb", reuse=None)
[ "def", "symbol_targets_bottom", "(", "x", ",", "model_hparams", ",", "vocab_size", ")", ":", "if", "(", "model_hparams", ".", "shared_embedding_and_softmax_weights", "or", "model_hparams", ".", "get", "(", "\"shared_embedding\"", ")", ")", ":", "try", ":", "return", "_symbol_bottom_simple", "(", "x", ",", "model_hparams", ",", "vocab_size", ",", "\"shared\"", ",", "reuse", "=", "True", ")", "except", "ValueError", ":", "# perhaps there were no inputs, and this is a new variable.", "return", "_symbol_bottom_simple", "(", "x", ",", "model_hparams", ",", "vocab_size", ",", "\"shared\"", ",", "reuse", "=", "None", ")", "else", ":", "return", "_symbol_bottom_simple", "(", "x", ",", "model_hparams", ",", "vocab_size", ",", "\"target_emb\"", ",", "reuse", "=", "None", ")" ]
Bottom transformation for target symbols.
[ "Bottom", "transformation", "for", "target", "symbols", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L517-L530
22,311
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
video_bitwise_bottom
def video_bitwise_bottom(x, model_hparams, vocab_size): """Bottom transformation for embedding video bitwise.""" pixel_embedding_size = 64 inputs = x with tf.variable_scope("video_modality_bitwise", reuse=tf.AUTO_REUSE): common_layers.summarize_video(inputs, "bottom") # Embed bitwise. assert vocab_size == 256 embedded = discretization.int_to_bit_embed(inputs, 8, pixel_embedding_size) # Project. return tf.layers.dense( embedded, model_hparams.hidden_size, name="merge_pixel_embedded_frames")
python
def video_bitwise_bottom(x, model_hparams, vocab_size): """Bottom transformation for embedding video bitwise.""" pixel_embedding_size = 64 inputs = x with tf.variable_scope("video_modality_bitwise", reuse=tf.AUTO_REUSE): common_layers.summarize_video(inputs, "bottom") # Embed bitwise. assert vocab_size == 256 embedded = discretization.int_to_bit_embed(inputs, 8, pixel_embedding_size) # Project. return tf.layers.dense( embedded, model_hparams.hidden_size, name="merge_pixel_embedded_frames")
[ "def", "video_bitwise_bottom", "(", "x", ",", "model_hparams", ",", "vocab_size", ")", ":", "pixel_embedding_size", "=", "64", "inputs", "=", "x", "with", "tf", ".", "variable_scope", "(", "\"video_modality_bitwise\"", ",", "reuse", "=", "tf", ".", "AUTO_REUSE", ")", ":", "common_layers", ".", "summarize_video", "(", "inputs", ",", "\"bottom\"", ")", "# Embed bitwise.", "assert", "vocab_size", "==", "256", "embedded", "=", "discretization", ".", "int_to_bit_embed", "(", "inputs", ",", "8", ",", "pixel_embedding_size", ")", "# Project.", "return", "tf", ".", "layers", ".", "dense", "(", "embedded", ",", "model_hparams", ".", "hidden_size", ",", "name", "=", "\"merge_pixel_embedded_frames\"", ")" ]
Bottom transformation for embedding video bitwise.
[ "Bottom", "transformation", "for", "embedding", "video", "bitwise", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L552-L566
22,312
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
video_pixel_noise_bottom
def video_pixel_noise_bottom(x, model_hparams, vocab_size): """Bottom transformation for video.""" input_noise = getattr(model_hparams, "video_modality_input_noise", 0.25) inputs = x if model_hparams.mode == tf.estimator.ModeKeys.TRAIN: background = tfp.stats.percentile(inputs, 50., axis=[0, 1, 2, 3]) input_shape = common_layers.shape_list(inputs) input_size = tf.reduce_prod(input_shape[:-1]) input_mask = tf.multinomial( tf.log([[input_noise, 1.-input_noise]]), input_size) input_mask = tf.reshape(tf.cast(input_mask, tf.int32), input_shape[:-1]+[1]) inputs = inputs * input_mask + background * (1 - input_mask) return video_bottom(inputs, model_hparams, vocab_size)
python
def video_pixel_noise_bottom(x, model_hparams, vocab_size): """Bottom transformation for video.""" input_noise = getattr(model_hparams, "video_modality_input_noise", 0.25) inputs = x if model_hparams.mode == tf.estimator.ModeKeys.TRAIN: background = tfp.stats.percentile(inputs, 50., axis=[0, 1, 2, 3]) input_shape = common_layers.shape_list(inputs) input_size = tf.reduce_prod(input_shape[:-1]) input_mask = tf.multinomial( tf.log([[input_noise, 1.-input_noise]]), input_size) input_mask = tf.reshape(tf.cast(input_mask, tf.int32), input_shape[:-1]+[1]) inputs = inputs * input_mask + background * (1 - input_mask) return video_bottom(inputs, model_hparams, vocab_size)
[ "def", "video_pixel_noise_bottom", "(", "x", ",", "model_hparams", ",", "vocab_size", ")", ":", "input_noise", "=", "getattr", "(", "model_hparams", ",", "\"video_modality_input_noise\"", ",", "0.25", ")", "inputs", "=", "x", "if", "model_hparams", ".", "mode", "==", "tf", ".", "estimator", ".", "ModeKeys", ".", "TRAIN", ":", "background", "=", "tfp", ".", "stats", ".", "percentile", "(", "inputs", ",", "50.", ",", "axis", "=", "[", "0", ",", "1", ",", "2", ",", "3", "]", ")", "input_shape", "=", "common_layers", ".", "shape_list", "(", "inputs", ")", "input_size", "=", "tf", ".", "reduce_prod", "(", "input_shape", "[", ":", "-", "1", "]", ")", "input_mask", "=", "tf", ".", "multinomial", "(", "tf", ".", "log", "(", "[", "[", "input_noise", ",", "1.", "-", "input_noise", "]", "]", ")", ",", "input_size", ")", "input_mask", "=", "tf", ".", "reshape", "(", "tf", ".", "cast", "(", "input_mask", ",", "tf", ".", "int32", ")", ",", "input_shape", "[", ":", "-", "1", "]", "+", "[", "1", "]", ")", "inputs", "=", "inputs", "*", "input_mask", "+", "background", "*", "(", "1", "-", "input_mask", ")", "return", "video_bottom", "(", "inputs", ",", "model_hparams", ",", "vocab_size", ")" ]
Bottom transformation for video.
[ "Bottom", "transformation", "for", "video", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L599-L612
22,313
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
convert_rgb_to_real
def convert_rgb_to_real(prediction, targets): """Convert prediction and target from rgb to real.""" prediction = tf.squeeze(prediction, axis=-1) prediction = common_layers.convert_rgb_to_real(prediction) targets = common_layers.convert_rgb_to_real(targets) return prediction, targets
python
def convert_rgb_to_real(prediction, targets): """Convert prediction and target from rgb to real.""" prediction = tf.squeeze(prediction, axis=-1) prediction = common_layers.convert_rgb_to_real(prediction) targets = common_layers.convert_rgb_to_real(targets) return prediction, targets
[ "def", "convert_rgb_to_real", "(", "prediction", ",", "targets", ")", ":", "prediction", "=", "tf", ".", "squeeze", "(", "prediction", ",", "axis", "=", "-", "1", ")", "prediction", "=", "common_layers", ".", "convert_rgb_to_real", "(", "prediction", ")", "targets", "=", "common_layers", ".", "convert_rgb_to_real", "(", "targets", ")", "return", "prediction", ",", "targets" ]
Convert prediction and target from rgb to real.
[ "Convert", "prediction", "and", "target", "from", "rgb", "to", "real", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L615-L620
22,314
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
ctc_symbol_loss
def ctc_symbol_loss(top_out, targets, model_hparams, vocab_size, weight_fn): """Compute the CTC loss.""" del model_hparams, vocab_size # unused arg logits = top_out with tf.name_scope("ctc_loss", values=[logits, targets]): # For CTC we assume targets are 1d, [batch, length, 1, 1] here. targets_shape = targets.get_shape().as_list() assert len(targets_shape) == 4 assert targets_shape[2] == 1 assert targets_shape[3] == 1 targets = tf.squeeze(targets, axis=[2, 3]) logits = tf.squeeze(logits, axis=[2, 3]) targets_mask = 1 - tf.to_int32(tf.equal(targets, 0)) targets_lengths = tf.reduce_sum(targets_mask, axis=1) sparse_targets = tf.keras.backend.ctc_label_dense_to_sparse( targets, targets_lengths) xent = tf.nn.ctc_loss( sparse_targets, logits, targets_lengths, time_major=False, preprocess_collapse_repeated=False, ctc_merge_repeated=False) weights = weight_fn(targets) return tf.reduce_sum(xent), tf.reduce_sum(weights)
python
def ctc_symbol_loss(top_out, targets, model_hparams, vocab_size, weight_fn): """Compute the CTC loss.""" del model_hparams, vocab_size # unused arg logits = top_out with tf.name_scope("ctc_loss", values=[logits, targets]): # For CTC we assume targets are 1d, [batch, length, 1, 1] here. targets_shape = targets.get_shape().as_list() assert len(targets_shape) == 4 assert targets_shape[2] == 1 assert targets_shape[3] == 1 targets = tf.squeeze(targets, axis=[2, 3]) logits = tf.squeeze(logits, axis=[2, 3]) targets_mask = 1 - tf.to_int32(tf.equal(targets, 0)) targets_lengths = tf.reduce_sum(targets_mask, axis=1) sparse_targets = tf.keras.backend.ctc_label_dense_to_sparse( targets, targets_lengths) xent = tf.nn.ctc_loss( sparse_targets, logits, targets_lengths, time_major=False, preprocess_collapse_repeated=False, ctc_merge_repeated=False) weights = weight_fn(targets) return tf.reduce_sum(xent), tf.reduce_sum(weights)
[ "def", "ctc_symbol_loss", "(", "top_out", ",", "targets", ",", "model_hparams", ",", "vocab_size", ",", "weight_fn", ")", ":", "del", "model_hparams", ",", "vocab_size", "# unused arg", "logits", "=", "top_out", "with", "tf", ".", "name_scope", "(", "\"ctc_loss\"", ",", "values", "=", "[", "logits", ",", "targets", "]", ")", ":", "# For CTC we assume targets are 1d, [batch, length, 1, 1] here.", "targets_shape", "=", "targets", ".", "get_shape", "(", ")", ".", "as_list", "(", ")", "assert", "len", "(", "targets_shape", ")", "==", "4", "assert", "targets_shape", "[", "2", "]", "==", "1", "assert", "targets_shape", "[", "3", "]", "==", "1", "targets", "=", "tf", ".", "squeeze", "(", "targets", ",", "axis", "=", "[", "2", ",", "3", "]", ")", "logits", "=", "tf", ".", "squeeze", "(", "logits", ",", "axis", "=", "[", "2", ",", "3", "]", ")", "targets_mask", "=", "1", "-", "tf", ".", "to_int32", "(", "tf", ".", "equal", "(", "targets", ",", "0", ")", ")", "targets_lengths", "=", "tf", ".", "reduce_sum", "(", "targets_mask", ",", "axis", "=", "1", ")", "sparse_targets", "=", "tf", ".", "keras", ".", "backend", ".", "ctc_label_dense_to_sparse", "(", "targets", ",", "targets_lengths", ")", "xent", "=", "tf", ".", "nn", ".", "ctc_loss", "(", "sparse_targets", ",", "logits", ",", "targets_lengths", ",", "time_major", "=", "False", ",", "preprocess_collapse_repeated", "=", "False", ",", "ctc_merge_repeated", "=", "False", ")", "weights", "=", "weight_fn", "(", "targets", ")", "return", "tf", ".", "reduce_sum", "(", "xent", ")", ",", "tf", ".", "reduce_sum", "(", "weights", ")" ]
Compute the CTC loss.
[ "Compute", "the", "CTC", "loss", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L638-L662
22,315
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
multi_label_loss
def multi_label_loss(top_out, targets, model_hparams, vocab_size, weights_fn): """Average loss over the labels.""" del vocab_size # unused arg logits = top_out num_labels = tf.shape(targets)[1] logits = tf.tile(logits, [1, num_labels, 1, 1, 1]) xent, weights = common_layers.padded_cross_entropy( logits, targets, model_hparams.label_smoothing, weights_fn=weights_fn, reduce_sum=False, ) xent = tf.squeeze(xent, [2, 3]) weights = tf.squeeze(weights, [2, 3]) # average loss over all labels loss = tf.reduce_sum(xent, axis=1) weights = tf.reduce_sum(weights, axis=1) loss /= (weights + 1e-8) weights = tf.to_float(tf.greater(weights, 0.)) return tf.reduce_sum(loss*weights), tf.reduce_sum(weights)
python
def multi_label_loss(top_out, targets, model_hparams, vocab_size, weights_fn): """Average loss over the labels.""" del vocab_size # unused arg logits = top_out num_labels = tf.shape(targets)[1] logits = tf.tile(logits, [1, num_labels, 1, 1, 1]) xent, weights = common_layers.padded_cross_entropy( logits, targets, model_hparams.label_smoothing, weights_fn=weights_fn, reduce_sum=False, ) xent = tf.squeeze(xent, [2, 3]) weights = tf.squeeze(weights, [2, 3]) # average loss over all labels loss = tf.reduce_sum(xent, axis=1) weights = tf.reduce_sum(weights, axis=1) loss /= (weights + 1e-8) weights = tf.to_float(tf.greater(weights, 0.)) return tf.reduce_sum(loss*weights), tf.reduce_sum(weights)
[ "def", "multi_label_loss", "(", "top_out", ",", "targets", ",", "model_hparams", ",", "vocab_size", ",", "weights_fn", ")", ":", "del", "vocab_size", "# unused arg", "logits", "=", "top_out", "num_labels", "=", "tf", ".", "shape", "(", "targets", ")", "[", "1", "]", "logits", "=", "tf", ".", "tile", "(", "logits", ",", "[", "1", ",", "num_labels", ",", "1", ",", "1", ",", "1", "]", ")", "xent", ",", "weights", "=", "common_layers", ".", "padded_cross_entropy", "(", "logits", ",", "targets", ",", "model_hparams", ".", "label_smoothing", ",", "weights_fn", "=", "weights_fn", ",", "reduce_sum", "=", "False", ",", ")", "xent", "=", "tf", ".", "squeeze", "(", "xent", ",", "[", "2", ",", "3", "]", ")", "weights", "=", "tf", ".", "squeeze", "(", "weights", ",", "[", "2", ",", "3", "]", ")", "# average loss over all labels", "loss", "=", "tf", ".", "reduce_sum", "(", "xent", ",", "axis", "=", "1", ")", "weights", "=", "tf", ".", "reduce_sum", "(", "weights", ",", "axis", "=", "1", ")", "loss", "/=", "(", "weights", "+", "1e-8", ")", "weights", "=", "tf", ".", "to_float", "(", "tf", ".", "greater", "(", "weights", ",", "0.", ")", ")", "return", "tf", ".", "reduce_sum", "(", "loss", "*", "weights", ")", ",", "tf", ".", "reduce_sum", "(", "weights", ")" ]
Average loss over the labels.
[ "Average", "loss", "over", "the", "labels", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L689-L711
22,316
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
one_hot_class_label_loss
def one_hot_class_label_loss(top_out, targets, model_hparams, vocab_size, weights_fn): """Apply softmax cross-entropy between outputs and targets. Args: top_out: logits Tensor with shape [batch, ?, ?, num_classes] targets: one-hot encoding Tensor with shape [batch, ?, ?, num_classes] model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. weights_fn: Returns: loss_scale (cross-entropy), loss_denom """ del model_hparams, vocab_size # unused arg loss_scale = tf.losses.softmax_cross_entropy( onehot_labels=targets, logits=top_out) weights = weights_fn(targets) loss_denom = tf.reduce_sum(weights) return loss_scale, loss_denom
python
def one_hot_class_label_loss(top_out, targets, model_hparams, vocab_size, weights_fn): """Apply softmax cross-entropy between outputs and targets. Args: top_out: logits Tensor with shape [batch, ?, ?, num_classes] targets: one-hot encoding Tensor with shape [batch, ?, ?, num_classes] model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. weights_fn: Returns: loss_scale (cross-entropy), loss_denom """ del model_hparams, vocab_size # unused arg loss_scale = tf.losses.softmax_cross_entropy( onehot_labels=targets, logits=top_out) weights = weights_fn(targets) loss_denom = tf.reduce_sum(weights) return loss_scale, loss_denom
[ "def", "one_hot_class_label_loss", "(", "top_out", ",", "targets", ",", "model_hparams", ",", "vocab_size", ",", "weights_fn", ")", ":", "del", "model_hparams", ",", "vocab_size", "# unused arg", "loss_scale", "=", "tf", ".", "losses", ".", "softmax_cross_entropy", "(", "onehot_labels", "=", "targets", ",", "logits", "=", "top_out", ")", "weights", "=", "weights_fn", "(", "targets", ")", "loss_denom", "=", "tf", ".", "reduce_sum", "(", "weights", ")", "return", "loss_scale", ",", "loss_denom" ]
Apply softmax cross-entropy between outputs and targets. Args: top_out: logits Tensor with shape [batch, ?, ?, num_classes] targets: one-hot encoding Tensor with shape [batch, ?, ?, num_classes] model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. weights_fn: Returns: loss_scale (cross-entropy), loss_denom
[ "Apply", "softmax", "cross", "-", "entropy", "between", "outputs", "and", "targets", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L714-L736
22,317
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
real_log_poisson_loss
def real_log_poisson_loss(top_out, targets, model_hparams, vocab_size, weights_fn): """Poisson loss for real.""" del model_hparams, vocab_size # unused arg predictions = top_out if (len(common_layers.shape_list(top_out)) != len( common_layers.shape_list(targets))): predictions = tf.squeeze(top_out, axis=[-1]) with tf.name_scope("log_possion"): weights = weights_fn(targets) lp_loss = tf.nn.log_poisson_loss(targets, predictions) return tf.reduce_sum(lp_loss * weights), tf.reduce_sum(weights)
python
def real_log_poisson_loss(top_out, targets, model_hparams, vocab_size, weights_fn): """Poisson loss for real.""" del model_hparams, vocab_size # unused arg predictions = top_out if (len(common_layers.shape_list(top_out)) != len( common_layers.shape_list(targets))): predictions = tf.squeeze(top_out, axis=[-1]) with tf.name_scope("log_possion"): weights = weights_fn(targets) lp_loss = tf.nn.log_poisson_loss(targets, predictions) return tf.reduce_sum(lp_loss * weights), tf.reduce_sum(weights)
[ "def", "real_log_poisson_loss", "(", "top_out", ",", "targets", ",", "model_hparams", ",", "vocab_size", ",", "weights_fn", ")", ":", "del", "model_hparams", ",", "vocab_size", "# unused arg", "predictions", "=", "top_out", "if", "(", "len", "(", "common_layers", ".", "shape_list", "(", "top_out", ")", ")", "!=", "len", "(", "common_layers", ".", "shape_list", "(", "targets", ")", ")", ")", ":", "predictions", "=", "tf", ".", "squeeze", "(", "top_out", ",", "axis", "=", "[", "-", "1", "]", ")", "with", "tf", ".", "name_scope", "(", "\"log_possion\"", ")", ":", "weights", "=", "weights_fn", "(", "targets", ")", "lp_loss", "=", "tf", ".", "nn", ".", "log_poisson_loss", "(", "targets", ",", "predictions", ")", "return", "tf", ".", "reduce_sum", "(", "lp_loss", "*", "weights", ")", ",", "tf", ".", "reduce_sum", "(", "weights", ")" ]
Poisson loss for real.
[ "Poisson", "loss", "for", "real", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L751-L765
22,318
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
class_label_top
def class_label_top(body_output, targets, model_hparams, vocab_size): """Transform inputs from model space to target space. Average over inner dims and a linear layer to logits. Args: body_output: A Tensor with shape [batch, ?, ?, body_output_size]. targets: model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: a Tensors, each with shape [batch_size, 1, 1, 1, vocab_size] """ del targets # unused arg with tf.variable_scope("class_label_modality_%d_%d" % ( vocab_size, model_hparams.hidden_size)): x = body_output x = tf.reduce_mean(x, axis=[1, 2], keepdims=True) res = tf.layers.dense(x, vocab_size) return tf.expand_dims(res, 3)
python
def class_label_top(body_output, targets, model_hparams, vocab_size): """Transform inputs from model space to target space. Average over inner dims and a linear layer to logits. Args: body_output: A Tensor with shape [batch, ?, ?, body_output_size]. targets: model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: a Tensors, each with shape [batch_size, 1, 1, 1, vocab_size] """ del targets # unused arg with tf.variable_scope("class_label_modality_%d_%d" % ( vocab_size, model_hparams.hidden_size)): x = body_output x = tf.reduce_mean(x, axis=[1, 2], keepdims=True) res = tf.layers.dense(x, vocab_size) return tf.expand_dims(res, 3)
[ "def", "class_label_top", "(", "body_output", ",", "targets", ",", "model_hparams", ",", "vocab_size", ")", ":", "del", "targets", "# unused arg", "with", "tf", ".", "variable_scope", "(", "\"class_label_modality_%d_%d\"", "%", "(", "vocab_size", ",", "model_hparams", ".", "hidden_size", ")", ")", ":", "x", "=", "body_output", "x", "=", "tf", ".", "reduce_mean", "(", "x", ",", "axis", "=", "[", "1", ",", "2", "]", ",", "keepdims", "=", "True", ")", "res", "=", "tf", ".", "layers", ".", "dense", "(", "x", ",", "vocab_size", ")", "return", "tf", ".", "expand_dims", "(", "res", ",", "3", ")" ]
Transform inputs from model space to target space. Average over inner dims and a linear layer to logits. Args: body_output: A Tensor with shape [batch, ?, ?, body_output_size]. targets: model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: a Tensors, each with shape [batch_size, 1, 1, 1, vocab_size]
[ "Transform", "inputs", "from", "model", "space", "to", "target", "space", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L927-L947
22,319
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
image_channel_compress_top
def image_channel_compress_top(body_output, targets, model_hparams, vocab_size): """Transforms body output to return logits. Args: body_output: Tensor of shape [batch, img_len, img_len, depth]. targets: model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: Tensor of shape [batch, img_len, img_len, channels, vocab_size]. """ del targets # unused arg with tf.variable_scope("image_channel_compress_modality"): hidden_size = model_hparams.hidden_size img_len = model_hparams.img_len channels = 3 # RGB batch = common_layers.shape_list(body_output)[0] x = tf.layers.conv2d( body_output, hidden_size * channels, kernel_size=(1, 1), strides=(1, 1), padding="VALID", activation=tf.nn.relu, name="decompress_conv") x = tf.reshape(x, [batch, img_len, img_len * channels, hidden_size]) x = common_layers.layer_preprocess(x, model_hparams) x = tf.layers.dense(x, vocab_size, use_bias=True, activation=None, name="output_conv") x = tf.reshape( x, [batch, img_len, img_len, channels, vocab_size]) return x
python
def image_channel_compress_top(body_output, targets, model_hparams, vocab_size): """Transforms body output to return logits. Args: body_output: Tensor of shape [batch, img_len, img_len, depth]. targets: model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: Tensor of shape [batch, img_len, img_len, channels, vocab_size]. """ del targets # unused arg with tf.variable_scope("image_channel_compress_modality"): hidden_size = model_hparams.hidden_size img_len = model_hparams.img_len channels = 3 # RGB batch = common_layers.shape_list(body_output)[0] x = tf.layers.conv2d( body_output, hidden_size * channels, kernel_size=(1, 1), strides=(1, 1), padding="VALID", activation=tf.nn.relu, name="decompress_conv") x = tf.reshape(x, [batch, img_len, img_len * channels, hidden_size]) x = common_layers.layer_preprocess(x, model_hparams) x = tf.layers.dense(x, vocab_size, use_bias=True, activation=None, name="output_conv") x = tf.reshape( x, [batch, img_len, img_len, channels, vocab_size]) return x
[ "def", "image_channel_compress_top", "(", "body_output", ",", "targets", ",", "model_hparams", ",", "vocab_size", ")", ":", "del", "targets", "# unused arg", "with", "tf", ".", "variable_scope", "(", "\"image_channel_compress_modality\"", ")", ":", "hidden_size", "=", "model_hparams", ".", "hidden_size", "img_len", "=", "model_hparams", ".", "img_len", "channels", "=", "3", "# RGB", "batch", "=", "common_layers", ".", "shape_list", "(", "body_output", ")", "[", "0", "]", "x", "=", "tf", ".", "layers", ".", "conv2d", "(", "body_output", ",", "hidden_size", "*", "channels", ",", "kernel_size", "=", "(", "1", ",", "1", ")", ",", "strides", "=", "(", "1", ",", "1", ")", ",", "padding", "=", "\"VALID\"", ",", "activation", "=", "tf", ".", "nn", ".", "relu", ",", "name", "=", "\"decompress_conv\"", ")", "x", "=", "tf", ".", "reshape", "(", "x", ",", "[", "batch", ",", "img_len", ",", "img_len", "*", "channels", ",", "hidden_size", "]", ")", "x", "=", "common_layers", ".", "layer_preprocess", "(", "x", ",", "model_hparams", ")", "x", "=", "tf", ".", "layers", ".", "dense", "(", "x", ",", "vocab_size", ",", "use_bias", "=", "True", ",", "activation", "=", "None", ",", "name", "=", "\"output_conv\"", ")", "x", "=", "tf", ".", "reshape", "(", "x", ",", "[", "batch", ",", "img_len", ",", "img_len", ",", "channels", ",", "vocab_size", "]", ")", "return", "x" ]
Transforms body output to return logits. Args: body_output: Tensor of shape [batch, img_len, img_len, depth]. targets: model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: Tensor of shape [batch, img_len, img_len, channels, vocab_size].
[ "Transforms", "body", "output", "to", "return", "logits", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L975-L1010
22,320
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
symbol_top
def symbol_top(body_output, targets, model_hparams, vocab_size): """Generate logits. Args: body_output: A Tensor with shape [batch, p0, p1, model_hparams.hidden_size]. targets: Unused. model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: logits: A Tensor with shape [batch, p0, p1, ?, vocab_size]. """ del targets # unused arg if model_hparams.shared_embedding_and_softmax_weights: scope_name = "shared" reuse = tf.AUTO_REUSE else: scope_name = "softmax" reuse = False with tf.variable_scope(scope_name, reuse=reuse): body_output_shape = common_layers.shape_list(body_output) var = get_weights(model_hparams, vocab_size, body_output_shape[-1]) if (model_hparams.factored_logits and model_hparams.mode == tf.estimator.ModeKeys.TRAIN): # insert channels dimension body_output = tf.expand_dims(body_output, 3) return common_layers.FactoredTensor(body_output, var) else: body_output = tf.reshape(body_output, [-1, body_output_shape[-1]]) logits = tf.matmul(body_output, var, transpose_b=True) return tf.reshape(logits, body_output_shape[:-1] + [1, vocab_size])
python
def symbol_top(body_output, targets, model_hparams, vocab_size): """Generate logits. Args: body_output: A Tensor with shape [batch, p0, p1, model_hparams.hidden_size]. targets: Unused. model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: logits: A Tensor with shape [batch, p0, p1, ?, vocab_size]. """ del targets # unused arg if model_hparams.shared_embedding_and_softmax_weights: scope_name = "shared" reuse = tf.AUTO_REUSE else: scope_name = "softmax" reuse = False with tf.variable_scope(scope_name, reuse=reuse): body_output_shape = common_layers.shape_list(body_output) var = get_weights(model_hparams, vocab_size, body_output_shape[-1]) if (model_hparams.factored_logits and model_hparams.mode == tf.estimator.ModeKeys.TRAIN): # insert channels dimension body_output = tf.expand_dims(body_output, 3) return common_layers.FactoredTensor(body_output, var) else: body_output = tf.reshape(body_output, [-1, body_output_shape[-1]]) logits = tf.matmul(body_output, var, transpose_b=True) return tf.reshape(logits, body_output_shape[:-1] + [1, vocab_size])
[ "def", "symbol_top", "(", "body_output", ",", "targets", ",", "model_hparams", ",", "vocab_size", ")", ":", "del", "targets", "# unused arg", "if", "model_hparams", ".", "shared_embedding_and_softmax_weights", ":", "scope_name", "=", "\"shared\"", "reuse", "=", "tf", ".", "AUTO_REUSE", "else", ":", "scope_name", "=", "\"softmax\"", "reuse", "=", "False", "with", "tf", ".", "variable_scope", "(", "scope_name", ",", "reuse", "=", "reuse", ")", ":", "body_output_shape", "=", "common_layers", ".", "shape_list", "(", "body_output", ")", "var", "=", "get_weights", "(", "model_hparams", ",", "vocab_size", ",", "body_output_shape", "[", "-", "1", "]", ")", "if", "(", "model_hparams", ".", "factored_logits", "and", "model_hparams", ".", "mode", "==", "tf", ".", "estimator", ".", "ModeKeys", ".", "TRAIN", ")", ":", "# insert channels dimension", "body_output", "=", "tf", ".", "expand_dims", "(", "body_output", ",", "3", ")", "return", "common_layers", ".", "FactoredTensor", "(", "body_output", ",", "var", ")", "else", ":", "body_output", "=", "tf", ".", "reshape", "(", "body_output", ",", "[", "-", "1", ",", "body_output_shape", "[", "-", "1", "]", "]", ")", "logits", "=", "tf", ".", "matmul", "(", "body_output", ",", "var", ",", "transpose_b", "=", "True", ")", "return", "tf", ".", "reshape", "(", "logits", ",", "body_output_shape", "[", ":", "-", "1", "]", "+", "[", "1", ",", "vocab_size", "]", ")" ]
Generate logits. Args: body_output: A Tensor with shape [batch, p0, p1, model_hparams.hidden_size]. targets: Unused. model_hparams: HParams, model hyperparmeters. vocab_size: int, vocabulary size. Returns: logits: A Tensor with shape [batch, p0, p1, ?, vocab_size].
[ "Generate", "logits", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L1105-L1137
22,321
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
get_bottom
def get_bottom(modality_type, value=None): """Gets default bottom transformation; if none available, return value.""" if modality_type == ModalityType.AUDIO: return audio_bottom elif modality_type == ModalityType.AUDIO_SPECTRAL: return audio_spectral_bottom elif modality_type in (ModalityType.CLASS_LABEL, ModalityType.MULTI_LABEL, ModalityType.ONE_HOT_CLASS_LABEL, ModalityType.SIGMOID_CLASS_LABEL, ModalityType.SIGMOID_MAX_POOLING_CLASS_LABEL, ModalityType.SOFTMAX_AVERAGE_POOLING_CLASS_LABEL, ModalityType.SOFTMAX_LAST_TIMESTEP_CLASS_LABEL, ModalityType.SOFTMAX_MAX_POOLING_CLASS_LABEL): return class_label_bottom elif modality_type in (ModalityType.CTC_SYMBOL, ModalityType.SYMBOL, ModalityType.SYMBOL_WEIGHTS_ALL): return symbol_bottom elif modality_type in (ModalityType.GENERIC_L2_LOSS, ModalityType.IDENTITY, ModalityType.IDENTITY_SYMBOL, ModalityType.IMAGE_CHANNEL_EMBEDDINGS_BOTTOM): return identity_bottom elif modality_type == ModalityType.IMAGE: return image_bottom elif modality_type in (ModalityType.IMAGE_CHANNEL_BOTTOM_IDENTITY, ModalityType.IMAGE_CHANNEL_COMPRESS): return image_channel_compress_bottom elif modality_type in (ModalityType.REAL, ModalityType.REAL_L2_LOSS, ModalityType.REAL_LOG_POISSON_LOSS): return real_bottom elif modality_type == ModalityType.SPEECH_RECOGNITION: return speech_recognition_bottom elif modality_type == ModalityType.SYMBOL_ONE_HOT: return symbol_one_hot_bottom elif modality_type in (ModalityType.VIDEO, ModalityType.VIDEO_L1, ModalityType.VIDEO_L2): return video_bottom elif modality_type == ModalityType.VIDEO_BITWISE: return video_bitwise_bottom elif modality_type == ModalityType.VIDEO_IDENTITY: return video_identity_bottom elif modality_type in (ModalityType.VIDEO_L1_RAW, ModalityType.VIDEO_L2_RAW): return video_raw_bottom elif modality_type == ModalityType.VIDEO_PIXEL_NOISE: return video_pixel_noise_bottom return value
python
def get_bottom(modality_type, value=None): """Gets default bottom transformation; if none available, return value.""" if modality_type == ModalityType.AUDIO: return audio_bottom elif modality_type == ModalityType.AUDIO_SPECTRAL: return audio_spectral_bottom elif modality_type in (ModalityType.CLASS_LABEL, ModalityType.MULTI_LABEL, ModalityType.ONE_HOT_CLASS_LABEL, ModalityType.SIGMOID_CLASS_LABEL, ModalityType.SIGMOID_MAX_POOLING_CLASS_LABEL, ModalityType.SOFTMAX_AVERAGE_POOLING_CLASS_LABEL, ModalityType.SOFTMAX_LAST_TIMESTEP_CLASS_LABEL, ModalityType.SOFTMAX_MAX_POOLING_CLASS_LABEL): return class_label_bottom elif modality_type in (ModalityType.CTC_SYMBOL, ModalityType.SYMBOL, ModalityType.SYMBOL_WEIGHTS_ALL): return symbol_bottom elif modality_type in (ModalityType.GENERIC_L2_LOSS, ModalityType.IDENTITY, ModalityType.IDENTITY_SYMBOL, ModalityType.IMAGE_CHANNEL_EMBEDDINGS_BOTTOM): return identity_bottom elif modality_type == ModalityType.IMAGE: return image_bottom elif modality_type in (ModalityType.IMAGE_CHANNEL_BOTTOM_IDENTITY, ModalityType.IMAGE_CHANNEL_COMPRESS): return image_channel_compress_bottom elif modality_type in (ModalityType.REAL, ModalityType.REAL_L2_LOSS, ModalityType.REAL_LOG_POISSON_LOSS): return real_bottom elif modality_type == ModalityType.SPEECH_RECOGNITION: return speech_recognition_bottom elif modality_type == ModalityType.SYMBOL_ONE_HOT: return symbol_one_hot_bottom elif modality_type in (ModalityType.VIDEO, ModalityType.VIDEO_L1, ModalityType.VIDEO_L2): return video_bottom elif modality_type == ModalityType.VIDEO_BITWISE: return video_bitwise_bottom elif modality_type == ModalityType.VIDEO_IDENTITY: return video_identity_bottom elif modality_type in (ModalityType.VIDEO_L1_RAW, ModalityType.VIDEO_L2_RAW): return video_raw_bottom elif modality_type == ModalityType.VIDEO_PIXEL_NOISE: return video_pixel_noise_bottom return value
[ "def", "get_bottom", "(", "modality_type", ",", "value", "=", "None", ")", ":", "if", "modality_type", "==", "ModalityType", ".", "AUDIO", ":", "return", "audio_bottom", "elif", "modality_type", "==", "ModalityType", ".", "AUDIO_SPECTRAL", ":", "return", "audio_spectral_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "CLASS_LABEL", ",", "ModalityType", ".", "MULTI_LABEL", ",", "ModalityType", ".", "ONE_HOT_CLASS_LABEL", ",", "ModalityType", ".", "SIGMOID_CLASS_LABEL", ",", "ModalityType", ".", "SIGMOID_MAX_POOLING_CLASS_LABEL", ",", "ModalityType", ".", "SOFTMAX_AVERAGE_POOLING_CLASS_LABEL", ",", "ModalityType", ".", "SOFTMAX_LAST_TIMESTEP_CLASS_LABEL", ",", "ModalityType", ".", "SOFTMAX_MAX_POOLING_CLASS_LABEL", ")", ":", "return", "class_label_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "CTC_SYMBOL", ",", "ModalityType", ".", "SYMBOL", ",", "ModalityType", ".", "SYMBOL_WEIGHTS_ALL", ")", ":", "return", "symbol_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "GENERIC_L2_LOSS", ",", "ModalityType", ".", "IDENTITY", ",", "ModalityType", ".", "IDENTITY_SYMBOL", ",", "ModalityType", ".", "IMAGE_CHANNEL_EMBEDDINGS_BOTTOM", ")", ":", "return", "identity_bottom", "elif", "modality_type", "==", "ModalityType", ".", "IMAGE", ":", "return", "image_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "IMAGE_CHANNEL_BOTTOM_IDENTITY", ",", "ModalityType", ".", "IMAGE_CHANNEL_COMPRESS", ")", ":", "return", "image_channel_compress_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "REAL", ",", "ModalityType", ".", "REAL_L2_LOSS", ",", "ModalityType", ".", "REAL_LOG_POISSON_LOSS", ")", ":", "return", "real_bottom", "elif", "modality_type", "==", "ModalityType", ".", "SPEECH_RECOGNITION", ":", "return", "speech_recognition_bottom", "elif", "modality_type", "==", "ModalityType", ".", "SYMBOL_ONE_HOT", ":", "return", "symbol_one_hot_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "VIDEO", ",", "ModalityType", ".", "VIDEO_L1", ",", "ModalityType", ".", "VIDEO_L2", ")", ":", "return", "video_bottom", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_BITWISE", ":", "return", "video_bitwise_bottom", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_IDENTITY", ":", "return", "video_identity_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "VIDEO_L1_RAW", ",", "ModalityType", ".", "VIDEO_L2_RAW", ")", ":", "return", "video_raw_bottom", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_PIXEL_NOISE", ":", "return", "video_pixel_noise_bottom", "return", "value" ]
Gets default bottom transformation; if none available, return value.
[ "Gets", "default", "bottom", "transformation", ";", "if", "none", "available", "return", "value", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L1192-L1242
22,322
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
get_loss
def get_loss(modality_type, value=None): """Gets default loss transformation; if none available, return value.""" if modality_type in (ModalityType.AUDIO, ModalityType.AUDIO_SPECTRAL, ModalityType.CLASS_LABEL, ModalityType.IDENTITY, ModalityType.IDENTITY_SYMBOL, ModalityType.IMAGE, ModalityType.IMAGE_CHANNEL_BOTTOM_IDENTITY, ModalityType.IMAGE_CHANNEL_COMPRESS, ModalityType.IMAGE_CHANNEL_EMBEDDINGS_BOTTOM, ModalityType.REAL, ModalityType.SPEECH_RECOGNITION, ModalityType.SYMBOL, ModalityType.SYMBOL_WEIGHTS_ALL): return generic_loss elif modality_type == ModalityType.CTC_SYMBOL: return ctc_symbol_loss elif modality_type == ModalityType.GENERIC_L2_LOSS: return generic_l2_loss elif modality_type == ModalityType.MULTI_LABEL: return multi_label_loss elif modality_type in (ModalityType.ONE_HOT_CLASS_LABEL, ModalityType.SOFTMAX_AVERAGE_POOLING_CLASS_LABEL, ModalityType.SOFTMAX_LAST_TIMESTEP_CLASS_LABEL, ModalityType.SOFTMAX_MAX_POOLING_CLASS_LABEL): return one_hot_class_label_loss elif modality_type == ModalityType.REAL_L2_LOSS: return real_l2_loss elif modality_type == ModalityType.REAL_LOG_POISSON_LOSS: return real_log_poisson_loss elif modality_type == ModalityType.SIGMOID_CLASS_LABEL: return sigmoid_class_label_loss elif modality_type == ModalityType.SIGMOID_MAX_POOLING_CLASS_LABEL: return sigmoid_max_pooling_class_label_loss elif modality_type == ModalityType.SYMBOL_ONE_HOT: return symbol_one_hot_loss elif modality_type in (ModalityType.VIDEO, ModalityType.VIDEO_BITWISE, ModalityType.VIDEO_PIXEL_NOISE): return video_loss elif modality_type == ModalityType.VIDEO_IDENTITY: return video_identity_loss elif modality_type == ModalityType.VIDEO_L1: return video_l1_loss elif modality_type == ModalityType.VIDEO_L1_RAW: return video_l1_raw_loss elif modality_type == ModalityType.VIDEO_L2: return video_l2_loss elif modality_type == ModalityType.VIDEO_L2_RAW: return video_l2_raw_loss return value
python
def get_loss(modality_type, value=None): """Gets default loss transformation; if none available, return value.""" if modality_type in (ModalityType.AUDIO, ModalityType.AUDIO_SPECTRAL, ModalityType.CLASS_LABEL, ModalityType.IDENTITY, ModalityType.IDENTITY_SYMBOL, ModalityType.IMAGE, ModalityType.IMAGE_CHANNEL_BOTTOM_IDENTITY, ModalityType.IMAGE_CHANNEL_COMPRESS, ModalityType.IMAGE_CHANNEL_EMBEDDINGS_BOTTOM, ModalityType.REAL, ModalityType.SPEECH_RECOGNITION, ModalityType.SYMBOL, ModalityType.SYMBOL_WEIGHTS_ALL): return generic_loss elif modality_type == ModalityType.CTC_SYMBOL: return ctc_symbol_loss elif modality_type == ModalityType.GENERIC_L2_LOSS: return generic_l2_loss elif modality_type == ModalityType.MULTI_LABEL: return multi_label_loss elif modality_type in (ModalityType.ONE_HOT_CLASS_LABEL, ModalityType.SOFTMAX_AVERAGE_POOLING_CLASS_LABEL, ModalityType.SOFTMAX_LAST_TIMESTEP_CLASS_LABEL, ModalityType.SOFTMAX_MAX_POOLING_CLASS_LABEL): return one_hot_class_label_loss elif modality_type == ModalityType.REAL_L2_LOSS: return real_l2_loss elif modality_type == ModalityType.REAL_LOG_POISSON_LOSS: return real_log_poisson_loss elif modality_type == ModalityType.SIGMOID_CLASS_LABEL: return sigmoid_class_label_loss elif modality_type == ModalityType.SIGMOID_MAX_POOLING_CLASS_LABEL: return sigmoid_max_pooling_class_label_loss elif modality_type == ModalityType.SYMBOL_ONE_HOT: return symbol_one_hot_loss elif modality_type in (ModalityType.VIDEO, ModalityType.VIDEO_BITWISE, ModalityType.VIDEO_PIXEL_NOISE): return video_loss elif modality_type == ModalityType.VIDEO_IDENTITY: return video_identity_loss elif modality_type == ModalityType.VIDEO_L1: return video_l1_loss elif modality_type == ModalityType.VIDEO_L1_RAW: return video_l1_raw_loss elif modality_type == ModalityType.VIDEO_L2: return video_l2_loss elif modality_type == ModalityType.VIDEO_L2_RAW: return video_l2_raw_loss return value
[ "def", "get_loss", "(", "modality_type", ",", "value", "=", "None", ")", ":", "if", "modality_type", "in", "(", "ModalityType", ".", "AUDIO", ",", "ModalityType", ".", "AUDIO_SPECTRAL", ",", "ModalityType", ".", "CLASS_LABEL", ",", "ModalityType", ".", "IDENTITY", ",", "ModalityType", ".", "IDENTITY_SYMBOL", ",", "ModalityType", ".", "IMAGE", ",", "ModalityType", ".", "IMAGE_CHANNEL_BOTTOM_IDENTITY", ",", "ModalityType", ".", "IMAGE_CHANNEL_COMPRESS", ",", "ModalityType", ".", "IMAGE_CHANNEL_EMBEDDINGS_BOTTOM", ",", "ModalityType", ".", "REAL", ",", "ModalityType", ".", "SPEECH_RECOGNITION", ",", "ModalityType", ".", "SYMBOL", ",", "ModalityType", ".", "SYMBOL_WEIGHTS_ALL", ")", ":", "return", "generic_loss", "elif", "modality_type", "==", "ModalityType", ".", "CTC_SYMBOL", ":", "return", "ctc_symbol_loss", "elif", "modality_type", "==", "ModalityType", ".", "GENERIC_L2_LOSS", ":", "return", "generic_l2_loss", "elif", "modality_type", "==", "ModalityType", ".", "MULTI_LABEL", ":", "return", "multi_label_loss", "elif", "modality_type", "in", "(", "ModalityType", ".", "ONE_HOT_CLASS_LABEL", ",", "ModalityType", ".", "SOFTMAX_AVERAGE_POOLING_CLASS_LABEL", ",", "ModalityType", ".", "SOFTMAX_LAST_TIMESTEP_CLASS_LABEL", ",", "ModalityType", ".", "SOFTMAX_MAX_POOLING_CLASS_LABEL", ")", ":", "return", "one_hot_class_label_loss", "elif", "modality_type", "==", "ModalityType", ".", "REAL_L2_LOSS", ":", "return", "real_l2_loss", "elif", "modality_type", "==", "ModalityType", ".", "REAL_LOG_POISSON_LOSS", ":", "return", "real_log_poisson_loss", "elif", "modality_type", "==", "ModalityType", ".", "SIGMOID_CLASS_LABEL", ":", "return", "sigmoid_class_label_loss", "elif", "modality_type", "==", "ModalityType", ".", "SIGMOID_MAX_POOLING_CLASS_LABEL", ":", "return", "sigmoid_max_pooling_class_label_loss", "elif", "modality_type", "==", "ModalityType", ".", "SYMBOL_ONE_HOT", ":", "return", "symbol_one_hot_loss", "elif", "modality_type", "in", "(", "ModalityType", ".", "VIDEO", ",", "ModalityType", ".", "VIDEO_BITWISE", ",", "ModalityType", ".", "VIDEO_PIXEL_NOISE", ")", ":", "return", "video_loss", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_IDENTITY", ":", "return", "video_identity_loss", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_L1", ":", "return", "video_l1_loss", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_L1_RAW", ":", "return", "video_l1_raw_loss", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_L2", ":", "return", "video_l2_loss", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_L2_RAW", ":", "return", "video_l2_raw_loss", "return", "value" ]
Gets default loss transformation; if none available, return value.
[ "Gets", "default", "loss", "transformation", ";", "if", "none", "available", "return", "value", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L1245-L1296
22,323
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
get_name
def get_name(modality_type, value=None): """Gets default name for transformations; if none available, return value.""" # For legacy reasons, modalities vary in their naming scheme. Future plans are # to remove any need for get_name. We do not recommend using it. if modality_type == ModalityType.AUDIO: return lambda model_hparams, vocab_size: "audio_modality" elif modality_type == ModalityType.AUDIO_SPECTRAL: return lambda model_hparams, vocab_size: "audio_spectral_modality" elif modality_type == ModalityType.GENERIC_L2_LOSS: return lambda model_hparams, vocab_size: "generic_l2_loss_modality" elif modality_type == ModalityType.IDENTITY: return lambda model_hparams, vocab_size: "identity_modality" elif modality_type == ModalityType.IMAGE: return lambda model_hparams, vocab_size: "image_modality" elif modality_type == ModalityType.IMAGE_CHANNEL_BOTTOM_IDENTITY: return (lambda model_hparams, vocab_size: # pylint: disable=g-long-lambda "image_channel_bottom_identity_modality") elif modality_type == ModalityType.IMAGE_CHANNEL_COMPRESS: return lambda model_hparams, vocab_size: "image_channel_compress_modality" elif modality_type == ModalityType.IMAGE_CHANNEL_EMBEDDINGS_BOTTOM: return lambda model_hparams, vocab_size: "image_channel_embeddings_bottom" elif modality_type == ModalityType.REAL: return lambda model_hparams, vocab_size: "real_modality" elif modality_type == ModalityType.REAL_L2_LOSS: return lambda model_hparams, vocab_size: "real_l2_loss_modality" elif modality_type == ModalityType.REAL_LOG_POISSON_LOSS: return lambda model_hparams, vocab_size: "real_log_poisson_loss_modality" elif modality_type == ModalityType.SPEECH_RECOGNITION: return lambda model_hparams, vocab_size: "speech_recognition_modality" elif modality_type == ModalityType.VIDEO: return lambda model_hparams, vocab_size: "video_modality" elif modality_type == ModalityType.VIDEO_BITWISE: return lambda model_hparams, vocab_size: "video_modality_bitwise" elif modality_type == ModalityType.VIDEO_IDENTITY: return lambda model_hparams, vocab_size: "video_modality_identity" elif modality_type == ModalityType.VIDEO_L1: return lambda model_hparams, vocab_size: "video_modality_l1" elif modality_type == ModalityType.VIDEO_L1_RAW: return lambda model_hparams, vocab_size: "video_modality_l1_raw" elif modality_type == ModalityType.VIDEO_L2: return lambda model_hparams, vocab_size: "video_modality_l2" elif modality_type == ModalityType.VIDEO_L2_RAW: return lambda model_hparams, vocab_size: "video_modality_l2_raw" elif modality_type == ModalityType.VIDEO_PIXEL_NOISE: return lambda model_hparams, vocab_size: "video_modality_pixel_noise" elif modality_type in (ModalityType.CLASS_LABEL, ModalityType.MULTI_LABEL, ModalityType.ONE_HOT_CLASS_LABEL): def name(model_hparams, vocab_size): return "class_label_modality_%d_%d" % (vocab_size, model_hparams.hidden_size) return name elif modality_type in (ModalityType.CTC_SYMBOL, ModalityType.IDENTITY_SYMBOL, ModalityType.SYMBOL, ModalityType.SYMBOL_WEIGHTS_ALL, ModalityType.SYMBOL_ONE_HOT): def name(model_hparams, vocab_size): return "symbol_modality_%d_%d" % (vocab_size, model_hparams.hidden_size) return name elif modality_type == ModalityType.SIGMOID_CLASS_LABEL: def name(model_hparams, vocab_size): return "sigmoid_class_symbol_modality_%d_%d" % (vocab_size, model_hparams.hidden_size) return name elif modality_type == ModalityType.SIGMOID_MAX_POOLING_CLASS_LABEL: def name(model_hparams, vocab_size): return "sigmoid_max_pooling_class_symbol_modality_%d_%d" % ( vocab_size, model_hparams.hidden_size) return name elif modality_type == ModalityType.SOFTMAX_AVERAGE_POOLING_CLASS_LABEL: def name(model_hparams, vocab_size): return "softmax_average_pooling_onehot_class_label_modality_%d_%d" % ( vocab_size, model_hparams.hidden_size) return name elif modality_type == ModalityType.SOFTMAX_LAST_TIMESTEP_CLASS_LABEL: def name(model_hparams, vocab_size): return "softmax_last_timestep_onehot_class_label_modality_%d_%d" % ( vocab_size, model_hparams.hidden_size) return name elif modality_type == ModalityType.SOFTMAX_MAX_POOLING_CLASS_LABEL: def name(model_hparams, vocab_size): return "softmax_max_pooling_onehot_class_label_modality_%d_%d" % ( vocab_size, model_hparams.hidden_size) return name return value
python
def get_name(modality_type, value=None): """Gets default name for transformations; if none available, return value.""" # For legacy reasons, modalities vary in their naming scheme. Future plans are # to remove any need for get_name. We do not recommend using it. if modality_type == ModalityType.AUDIO: return lambda model_hparams, vocab_size: "audio_modality" elif modality_type == ModalityType.AUDIO_SPECTRAL: return lambda model_hparams, vocab_size: "audio_spectral_modality" elif modality_type == ModalityType.GENERIC_L2_LOSS: return lambda model_hparams, vocab_size: "generic_l2_loss_modality" elif modality_type == ModalityType.IDENTITY: return lambda model_hparams, vocab_size: "identity_modality" elif modality_type == ModalityType.IMAGE: return lambda model_hparams, vocab_size: "image_modality" elif modality_type == ModalityType.IMAGE_CHANNEL_BOTTOM_IDENTITY: return (lambda model_hparams, vocab_size: # pylint: disable=g-long-lambda "image_channel_bottom_identity_modality") elif modality_type == ModalityType.IMAGE_CHANNEL_COMPRESS: return lambda model_hparams, vocab_size: "image_channel_compress_modality" elif modality_type == ModalityType.IMAGE_CHANNEL_EMBEDDINGS_BOTTOM: return lambda model_hparams, vocab_size: "image_channel_embeddings_bottom" elif modality_type == ModalityType.REAL: return lambda model_hparams, vocab_size: "real_modality" elif modality_type == ModalityType.REAL_L2_LOSS: return lambda model_hparams, vocab_size: "real_l2_loss_modality" elif modality_type == ModalityType.REAL_LOG_POISSON_LOSS: return lambda model_hparams, vocab_size: "real_log_poisson_loss_modality" elif modality_type == ModalityType.SPEECH_RECOGNITION: return lambda model_hparams, vocab_size: "speech_recognition_modality" elif modality_type == ModalityType.VIDEO: return lambda model_hparams, vocab_size: "video_modality" elif modality_type == ModalityType.VIDEO_BITWISE: return lambda model_hparams, vocab_size: "video_modality_bitwise" elif modality_type == ModalityType.VIDEO_IDENTITY: return lambda model_hparams, vocab_size: "video_modality_identity" elif modality_type == ModalityType.VIDEO_L1: return lambda model_hparams, vocab_size: "video_modality_l1" elif modality_type == ModalityType.VIDEO_L1_RAW: return lambda model_hparams, vocab_size: "video_modality_l1_raw" elif modality_type == ModalityType.VIDEO_L2: return lambda model_hparams, vocab_size: "video_modality_l2" elif modality_type == ModalityType.VIDEO_L2_RAW: return lambda model_hparams, vocab_size: "video_modality_l2_raw" elif modality_type == ModalityType.VIDEO_PIXEL_NOISE: return lambda model_hparams, vocab_size: "video_modality_pixel_noise" elif modality_type in (ModalityType.CLASS_LABEL, ModalityType.MULTI_LABEL, ModalityType.ONE_HOT_CLASS_LABEL): def name(model_hparams, vocab_size): return "class_label_modality_%d_%d" % (vocab_size, model_hparams.hidden_size) return name elif modality_type in (ModalityType.CTC_SYMBOL, ModalityType.IDENTITY_SYMBOL, ModalityType.SYMBOL, ModalityType.SYMBOL_WEIGHTS_ALL, ModalityType.SYMBOL_ONE_HOT): def name(model_hparams, vocab_size): return "symbol_modality_%d_%d" % (vocab_size, model_hparams.hidden_size) return name elif modality_type == ModalityType.SIGMOID_CLASS_LABEL: def name(model_hparams, vocab_size): return "sigmoid_class_symbol_modality_%d_%d" % (vocab_size, model_hparams.hidden_size) return name elif modality_type == ModalityType.SIGMOID_MAX_POOLING_CLASS_LABEL: def name(model_hparams, vocab_size): return "sigmoid_max_pooling_class_symbol_modality_%d_%d" % ( vocab_size, model_hparams.hidden_size) return name elif modality_type == ModalityType.SOFTMAX_AVERAGE_POOLING_CLASS_LABEL: def name(model_hparams, vocab_size): return "softmax_average_pooling_onehot_class_label_modality_%d_%d" % ( vocab_size, model_hparams.hidden_size) return name elif modality_type == ModalityType.SOFTMAX_LAST_TIMESTEP_CLASS_LABEL: def name(model_hparams, vocab_size): return "softmax_last_timestep_onehot_class_label_modality_%d_%d" % ( vocab_size, model_hparams.hidden_size) return name elif modality_type == ModalityType.SOFTMAX_MAX_POOLING_CLASS_LABEL: def name(model_hparams, vocab_size): return "softmax_max_pooling_onehot_class_label_modality_%d_%d" % ( vocab_size, model_hparams.hidden_size) return name return value
[ "def", "get_name", "(", "modality_type", ",", "value", "=", "None", ")", ":", "# For legacy reasons, modalities vary in their naming scheme. Future plans are", "# to remove any need for get_name. We do not recommend using it.", "if", "modality_type", "==", "ModalityType", ".", "AUDIO", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"audio_modality\"", "elif", "modality_type", "==", "ModalityType", ".", "AUDIO_SPECTRAL", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"audio_spectral_modality\"", "elif", "modality_type", "==", "ModalityType", ".", "GENERIC_L2_LOSS", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"generic_l2_loss_modality\"", "elif", "modality_type", "==", "ModalityType", ".", "IDENTITY", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"identity_modality\"", "elif", "modality_type", "==", "ModalityType", ".", "IMAGE", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"image_modality\"", "elif", "modality_type", "==", "ModalityType", ".", "IMAGE_CHANNEL_BOTTOM_IDENTITY", ":", "return", "(", "lambda", "model_hparams", ",", "vocab_size", ":", "# pylint: disable=g-long-lambda", "\"image_channel_bottom_identity_modality\"", ")", "elif", "modality_type", "==", "ModalityType", ".", "IMAGE_CHANNEL_COMPRESS", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"image_channel_compress_modality\"", "elif", "modality_type", "==", "ModalityType", ".", "IMAGE_CHANNEL_EMBEDDINGS_BOTTOM", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"image_channel_embeddings_bottom\"", "elif", "modality_type", "==", "ModalityType", ".", "REAL", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"real_modality\"", "elif", "modality_type", "==", "ModalityType", ".", "REAL_L2_LOSS", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"real_l2_loss_modality\"", "elif", "modality_type", "==", "ModalityType", ".", "REAL_LOG_POISSON_LOSS", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"real_log_poisson_loss_modality\"", "elif", "modality_type", "==", "ModalityType", ".", "SPEECH_RECOGNITION", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"speech_recognition_modality\"", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"video_modality\"", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_BITWISE", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"video_modality_bitwise\"", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_IDENTITY", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"video_modality_identity\"", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_L1", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"video_modality_l1\"", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_L1_RAW", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"video_modality_l1_raw\"", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_L2", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"video_modality_l2\"", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_L2_RAW", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"video_modality_l2_raw\"", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_PIXEL_NOISE", ":", "return", "lambda", "model_hparams", ",", "vocab_size", ":", "\"video_modality_pixel_noise\"", "elif", "modality_type", "in", "(", "ModalityType", ".", "CLASS_LABEL", ",", "ModalityType", ".", "MULTI_LABEL", ",", "ModalityType", ".", "ONE_HOT_CLASS_LABEL", ")", ":", "def", "name", "(", "model_hparams", ",", "vocab_size", ")", ":", "return", "\"class_label_modality_%d_%d\"", "%", "(", "vocab_size", ",", "model_hparams", ".", "hidden_size", ")", "return", "name", "elif", "modality_type", "in", "(", "ModalityType", ".", "CTC_SYMBOL", ",", "ModalityType", ".", "IDENTITY_SYMBOL", ",", "ModalityType", ".", "SYMBOL", ",", "ModalityType", ".", "SYMBOL_WEIGHTS_ALL", ",", "ModalityType", ".", "SYMBOL_ONE_HOT", ")", ":", "def", "name", "(", "model_hparams", ",", "vocab_size", ")", ":", "return", "\"symbol_modality_%d_%d\"", "%", "(", "vocab_size", ",", "model_hparams", ".", "hidden_size", ")", "return", "name", "elif", "modality_type", "==", "ModalityType", ".", "SIGMOID_CLASS_LABEL", ":", "def", "name", "(", "model_hparams", ",", "vocab_size", ")", ":", "return", "\"sigmoid_class_symbol_modality_%d_%d\"", "%", "(", "vocab_size", ",", "model_hparams", ".", "hidden_size", ")", "return", "name", "elif", "modality_type", "==", "ModalityType", ".", "SIGMOID_MAX_POOLING_CLASS_LABEL", ":", "def", "name", "(", "model_hparams", ",", "vocab_size", ")", ":", "return", "\"sigmoid_max_pooling_class_symbol_modality_%d_%d\"", "%", "(", "vocab_size", ",", "model_hparams", ".", "hidden_size", ")", "return", "name", "elif", "modality_type", "==", "ModalityType", ".", "SOFTMAX_AVERAGE_POOLING_CLASS_LABEL", ":", "def", "name", "(", "model_hparams", ",", "vocab_size", ")", ":", "return", "\"softmax_average_pooling_onehot_class_label_modality_%d_%d\"", "%", "(", "vocab_size", ",", "model_hparams", ".", "hidden_size", ")", "return", "name", "elif", "modality_type", "==", "ModalityType", ".", "SOFTMAX_LAST_TIMESTEP_CLASS_LABEL", ":", "def", "name", "(", "model_hparams", ",", "vocab_size", ")", ":", "return", "\"softmax_last_timestep_onehot_class_label_modality_%d_%d\"", "%", "(", "vocab_size", ",", "model_hparams", ".", "hidden_size", ")", "return", "name", "elif", "modality_type", "==", "ModalityType", ".", "SOFTMAX_MAX_POOLING_CLASS_LABEL", ":", "def", "name", "(", "model_hparams", ",", "vocab_size", ")", ":", "return", "\"softmax_max_pooling_onehot_class_label_modality_%d_%d\"", "%", "(", "vocab_size", ",", "model_hparams", ".", "hidden_size", ")", "return", "name", "return", "value" ]
Gets default name for transformations; if none available, return value.
[ "Gets", "default", "name", "for", "transformations", ";", "if", "none", "available", "return", "value", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L1299-L1384
22,324
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
get_targets_bottom
def get_targets_bottom(modality_type, value=None): """Gets default bottom transformation for targets; if none, return value.""" if modality_type == ModalityType.AUDIO: return make_targets_bottom(audio_bottom) elif modality_type == ModalityType.AUDIO_SPECTRAL: return make_targets_bottom(audio_spectral_bottom) elif modality_type in (ModalityType.CLASS_LABEL, ModalityType.MULTI_LABEL, ModalityType.ONE_HOT_CLASS_LABEL, ModalityType.SIGMOID_CLASS_LABEL, ModalityType.SIGMOID_MAX_POOLING_CLASS_LABEL, ModalityType.SOFTMAX_AVERAGE_POOLING_CLASS_LABEL, ModalityType.SOFTMAX_LAST_TIMESTEP_CLASS_LABEL, ModalityType.SOFTMAX_MAX_POOLING_CLASS_LABEL): return class_label_targets_bottom elif modality_type in (ModalityType.CTC_SYMBOL, ModalityType.SYMBOL, ModalityType.SYMBOL_WEIGHTS_ALL): return symbol_targets_bottom elif modality_type in (ModalityType.GENERIC_L2_LOSS, ModalityType.IDENTITY_SYMBOL): return identity_bottom elif modality_type == ModalityType.IDENTITY: return make_targets_bottom(identity_bottom) elif modality_type == ModalityType.IMAGE: return image_targets_bottom elif modality_type in (ModalityType.IMAGE_CHANNEL_BOTTOM_IDENTITY, ModalityType.IMAGE_CHANNEL_COMPRESS): return image_channel_compress_targets_bottom elif modality_type == ModalityType.IMAGE_CHANNEL_EMBEDDINGS_BOTTOM: return image_channel_embeddings_bottom elif modality_type in (ModalityType.REAL, ModalityType.REAL_L2_LOSS, ModalityType.REAL_LOG_POISSON_LOSS): return make_targets_bottom(real_bottom) elif modality_type == ModalityType.SPEECH_RECOGNITION: return make_targets_bottom(speech_recognition_bottom) elif modality_type == ModalityType.SYMBOL_ONE_HOT: return symbol_one_hot_bottom elif modality_type in (ModalityType.VIDEO, ModalityType.VIDEO_L1, ModalityType.VIDEO_L2): return video_targets_bottom elif modality_type == ModalityType.VIDEO_BITWISE: return video_bitwise_targets_bottom elif modality_type == ModalityType.VIDEO_IDENTITY: return video_identity_targets_bottom elif modality_type in (ModalityType.VIDEO_L1_RAW, ModalityType.VIDEO_L2_RAW): return video_raw_targets_bottom elif modality_type == ModalityType.VIDEO_PIXEL_NOISE: return make_targets_bottom(video_pixel_noise_bottom) return value
python
def get_targets_bottom(modality_type, value=None): """Gets default bottom transformation for targets; if none, return value.""" if modality_type == ModalityType.AUDIO: return make_targets_bottom(audio_bottom) elif modality_type == ModalityType.AUDIO_SPECTRAL: return make_targets_bottom(audio_spectral_bottom) elif modality_type in (ModalityType.CLASS_LABEL, ModalityType.MULTI_LABEL, ModalityType.ONE_HOT_CLASS_LABEL, ModalityType.SIGMOID_CLASS_LABEL, ModalityType.SIGMOID_MAX_POOLING_CLASS_LABEL, ModalityType.SOFTMAX_AVERAGE_POOLING_CLASS_LABEL, ModalityType.SOFTMAX_LAST_TIMESTEP_CLASS_LABEL, ModalityType.SOFTMAX_MAX_POOLING_CLASS_LABEL): return class_label_targets_bottom elif modality_type in (ModalityType.CTC_SYMBOL, ModalityType.SYMBOL, ModalityType.SYMBOL_WEIGHTS_ALL): return symbol_targets_bottom elif modality_type in (ModalityType.GENERIC_L2_LOSS, ModalityType.IDENTITY_SYMBOL): return identity_bottom elif modality_type == ModalityType.IDENTITY: return make_targets_bottom(identity_bottom) elif modality_type == ModalityType.IMAGE: return image_targets_bottom elif modality_type in (ModalityType.IMAGE_CHANNEL_BOTTOM_IDENTITY, ModalityType.IMAGE_CHANNEL_COMPRESS): return image_channel_compress_targets_bottom elif modality_type == ModalityType.IMAGE_CHANNEL_EMBEDDINGS_BOTTOM: return image_channel_embeddings_bottom elif modality_type in (ModalityType.REAL, ModalityType.REAL_L2_LOSS, ModalityType.REAL_LOG_POISSON_LOSS): return make_targets_bottom(real_bottom) elif modality_type == ModalityType.SPEECH_RECOGNITION: return make_targets_bottom(speech_recognition_bottom) elif modality_type == ModalityType.SYMBOL_ONE_HOT: return symbol_one_hot_bottom elif modality_type in (ModalityType.VIDEO, ModalityType.VIDEO_L1, ModalityType.VIDEO_L2): return video_targets_bottom elif modality_type == ModalityType.VIDEO_BITWISE: return video_bitwise_targets_bottom elif modality_type == ModalityType.VIDEO_IDENTITY: return video_identity_targets_bottom elif modality_type in (ModalityType.VIDEO_L1_RAW, ModalityType.VIDEO_L2_RAW): return video_raw_targets_bottom elif modality_type == ModalityType.VIDEO_PIXEL_NOISE: return make_targets_bottom(video_pixel_noise_bottom) return value
[ "def", "get_targets_bottom", "(", "modality_type", ",", "value", "=", "None", ")", ":", "if", "modality_type", "==", "ModalityType", ".", "AUDIO", ":", "return", "make_targets_bottom", "(", "audio_bottom", ")", "elif", "modality_type", "==", "ModalityType", ".", "AUDIO_SPECTRAL", ":", "return", "make_targets_bottom", "(", "audio_spectral_bottom", ")", "elif", "modality_type", "in", "(", "ModalityType", ".", "CLASS_LABEL", ",", "ModalityType", ".", "MULTI_LABEL", ",", "ModalityType", ".", "ONE_HOT_CLASS_LABEL", ",", "ModalityType", ".", "SIGMOID_CLASS_LABEL", ",", "ModalityType", ".", "SIGMOID_MAX_POOLING_CLASS_LABEL", ",", "ModalityType", ".", "SOFTMAX_AVERAGE_POOLING_CLASS_LABEL", ",", "ModalityType", ".", "SOFTMAX_LAST_TIMESTEP_CLASS_LABEL", ",", "ModalityType", ".", "SOFTMAX_MAX_POOLING_CLASS_LABEL", ")", ":", "return", "class_label_targets_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "CTC_SYMBOL", ",", "ModalityType", ".", "SYMBOL", ",", "ModalityType", ".", "SYMBOL_WEIGHTS_ALL", ")", ":", "return", "symbol_targets_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "GENERIC_L2_LOSS", ",", "ModalityType", ".", "IDENTITY_SYMBOL", ")", ":", "return", "identity_bottom", "elif", "modality_type", "==", "ModalityType", ".", "IDENTITY", ":", "return", "make_targets_bottom", "(", "identity_bottom", ")", "elif", "modality_type", "==", "ModalityType", ".", "IMAGE", ":", "return", "image_targets_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "IMAGE_CHANNEL_BOTTOM_IDENTITY", ",", "ModalityType", ".", "IMAGE_CHANNEL_COMPRESS", ")", ":", "return", "image_channel_compress_targets_bottom", "elif", "modality_type", "==", "ModalityType", ".", "IMAGE_CHANNEL_EMBEDDINGS_BOTTOM", ":", "return", "image_channel_embeddings_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "REAL", ",", "ModalityType", ".", "REAL_L2_LOSS", ",", "ModalityType", ".", "REAL_LOG_POISSON_LOSS", ")", ":", "return", "make_targets_bottom", "(", "real_bottom", ")", "elif", "modality_type", "==", "ModalityType", ".", "SPEECH_RECOGNITION", ":", "return", "make_targets_bottom", "(", "speech_recognition_bottom", ")", "elif", "modality_type", "==", "ModalityType", ".", "SYMBOL_ONE_HOT", ":", "return", "symbol_one_hot_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "VIDEO", ",", "ModalityType", ".", "VIDEO_L1", ",", "ModalityType", ".", "VIDEO_L2", ")", ":", "return", "video_targets_bottom", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_BITWISE", ":", "return", "video_bitwise_targets_bottom", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_IDENTITY", ":", "return", "video_identity_targets_bottom", "elif", "modality_type", "in", "(", "ModalityType", ".", "VIDEO_L1_RAW", ",", "ModalityType", ".", "VIDEO_L2_RAW", ")", ":", "return", "video_raw_targets_bottom", "elif", "modality_type", "==", "ModalityType", ".", "VIDEO_PIXEL_NOISE", ":", "return", "make_targets_bottom", "(", "video_pixel_noise_bottom", ")", "return", "value" ]
Gets default bottom transformation for targets; if none, return value.
[ "Gets", "default", "bottom", "transformation", "for", "targets", ";", "if", "none", "return", "value", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L1387-L1439
22,325
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
get_top
def get_top(modality_type, value=None): """Gets default top transformation; if none available, return value.""" if modality_type in (ModalityType.AUDIO, ModalityType.AUDIO_SPECTRAL, ModalityType.GENERIC_L2_LOSS, ModalityType.IDENTITY, ModalityType.IDENTITY_SYMBOL, ModalityType.IMAGE_CHANNEL_BOTTOM_IDENTITY, ModalityType.SPEECH_RECOGNITION, ModalityType.VIDEO_IDENTITY): return identity_top elif modality_type in (ModalityType.CLASS_LABEL, ModalityType.MULTI_LABEL, ModalityType.ONE_HOT_CLASS_LABEL, ModalityType.SIGMOID_CLASS_LABEL): return class_label_top elif modality_type in (ModalityType.CTC_SYMBOL, ModalityType.SYMBOL, ModalityType.SYMBOL_WEIGHTS_ALL): return symbol_top elif modality_type == ModalityType.IMAGE: return image_top elif modality_type == ModalityType.IMAGE_CHANNEL_COMPRESS: return image_channel_compress_top elif modality_type == ModalityType.IMAGE_CHANNEL_EMBEDDINGS_BOTTOM: return image_channel_embeddings_top elif modality_type in (ModalityType.REAL, ModalityType.REAL_L2_LOSS, ModalityType.REAL_LOG_POISSON_LOSS): return real_top elif modality_type == ModalityType.SIGMOID_MAX_POOLING_CLASS_LABEL: return sigmoid_max_pooling_class_label_top elif modality_type == ModalityType.SOFTMAX_AVERAGE_POOLING_CLASS_LABEL: return softmax_average_pooling_class_label_top elif modality_type == ModalityType.SOFTMAX_LAST_TIMESTEP_CLASS_LABEL: return softmax_last_timestep_class_label_top elif modality_type == ModalityType.SOFTMAX_MAX_POOLING_CLASS_LABEL: return softmax_max_pooling_class_label_top elif modality_type == ModalityType.SYMBOL_ONE_HOT: return symbol_one_hot_top elif modality_type in (ModalityType.VIDEO, ModalityType.VIDEO_BITWISE, ModalityType.VIDEO_PIXEL_NOISE): return video_top elif modality_type in (ModalityType.VIDEO_L1, ModalityType.VIDEO_L2): return video_l1_top elif modality_type in (ModalityType.VIDEO_L1_RAW, ModalityType.VIDEO_L2_RAW): return video_raw_top return value
python
def get_top(modality_type, value=None): """Gets default top transformation; if none available, return value.""" if modality_type in (ModalityType.AUDIO, ModalityType.AUDIO_SPECTRAL, ModalityType.GENERIC_L2_LOSS, ModalityType.IDENTITY, ModalityType.IDENTITY_SYMBOL, ModalityType.IMAGE_CHANNEL_BOTTOM_IDENTITY, ModalityType.SPEECH_RECOGNITION, ModalityType.VIDEO_IDENTITY): return identity_top elif modality_type in (ModalityType.CLASS_LABEL, ModalityType.MULTI_LABEL, ModalityType.ONE_HOT_CLASS_LABEL, ModalityType.SIGMOID_CLASS_LABEL): return class_label_top elif modality_type in (ModalityType.CTC_SYMBOL, ModalityType.SYMBOL, ModalityType.SYMBOL_WEIGHTS_ALL): return symbol_top elif modality_type == ModalityType.IMAGE: return image_top elif modality_type == ModalityType.IMAGE_CHANNEL_COMPRESS: return image_channel_compress_top elif modality_type == ModalityType.IMAGE_CHANNEL_EMBEDDINGS_BOTTOM: return image_channel_embeddings_top elif modality_type in (ModalityType.REAL, ModalityType.REAL_L2_LOSS, ModalityType.REAL_LOG_POISSON_LOSS): return real_top elif modality_type == ModalityType.SIGMOID_MAX_POOLING_CLASS_LABEL: return sigmoid_max_pooling_class_label_top elif modality_type == ModalityType.SOFTMAX_AVERAGE_POOLING_CLASS_LABEL: return softmax_average_pooling_class_label_top elif modality_type == ModalityType.SOFTMAX_LAST_TIMESTEP_CLASS_LABEL: return softmax_last_timestep_class_label_top elif modality_type == ModalityType.SOFTMAX_MAX_POOLING_CLASS_LABEL: return softmax_max_pooling_class_label_top elif modality_type == ModalityType.SYMBOL_ONE_HOT: return symbol_one_hot_top elif modality_type in (ModalityType.VIDEO, ModalityType.VIDEO_BITWISE, ModalityType.VIDEO_PIXEL_NOISE): return video_top elif modality_type in (ModalityType.VIDEO_L1, ModalityType.VIDEO_L2): return video_l1_top elif modality_type in (ModalityType.VIDEO_L1_RAW, ModalityType.VIDEO_L2_RAW): return video_raw_top return value
[ "def", "get_top", "(", "modality_type", ",", "value", "=", "None", ")", ":", "if", "modality_type", "in", "(", "ModalityType", ".", "AUDIO", ",", "ModalityType", ".", "AUDIO_SPECTRAL", ",", "ModalityType", ".", "GENERIC_L2_LOSS", ",", "ModalityType", ".", "IDENTITY", ",", "ModalityType", ".", "IDENTITY_SYMBOL", ",", "ModalityType", ".", "IMAGE_CHANNEL_BOTTOM_IDENTITY", ",", "ModalityType", ".", "SPEECH_RECOGNITION", ",", "ModalityType", ".", "VIDEO_IDENTITY", ")", ":", "return", "identity_top", "elif", "modality_type", "in", "(", "ModalityType", ".", "CLASS_LABEL", ",", "ModalityType", ".", "MULTI_LABEL", ",", "ModalityType", ".", "ONE_HOT_CLASS_LABEL", ",", "ModalityType", ".", "SIGMOID_CLASS_LABEL", ")", ":", "return", "class_label_top", "elif", "modality_type", "in", "(", "ModalityType", ".", "CTC_SYMBOL", ",", "ModalityType", ".", "SYMBOL", ",", "ModalityType", ".", "SYMBOL_WEIGHTS_ALL", ")", ":", "return", "symbol_top", "elif", "modality_type", "==", "ModalityType", ".", "IMAGE", ":", "return", "image_top", "elif", "modality_type", "==", "ModalityType", ".", "IMAGE_CHANNEL_COMPRESS", ":", "return", "image_channel_compress_top", "elif", "modality_type", "==", "ModalityType", ".", "IMAGE_CHANNEL_EMBEDDINGS_BOTTOM", ":", "return", "image_channel_embeddings_top", "elif", "modality_type", "in", "(", "ModalityType", ".", "REAL", ",", "ModalityType", ".", "REAL_L2_LOSS", ",", "ModalityType", ".", "REAL_LOG_POISSON_LOSS", ")", ":", "return", "real_top", "elif", "modality_type", "==", "ModalityType", ".", "SIGMOID_MAX_POOLING_CLASS_LABEL", ":", "return", "sigmoid_max_pooling_class_label_top", "elif", "modality_type", "==", "ModalityType", ".", "SOFTMAX_AVERAGE_POOLING_CLASS_LABEL", ":", "return", "softmax_average_pooling_class_label_top", "elif", "modality_type", "==", "ModalityType", ".", "SOFTMAX_LAST_TIMESTEP_CLASS_LABEL", ":", "return", "softmax_last_timestep_class_label_top", "elif", "modality_type", "==", "ModalityType", ".", "SOFTMAX_MAX_POOLING_CLASS_LABEL", ":", "return", "softmax_max_pooling_class_label_top", "elif", "modality_type", "==", "ModalityType", ".", "SYMBOL_ONE_HOT", ":", "return", "symbol_one_hot_top", "elif", "modality_type", "in", "(", "ModalityType", ".", "VIDEO", ",", "ModalityType", ".", "VIDEO_BITWISE", ",", "ModalityType", ".", "VIDEO_PIXEL_NOISE", ")", ":", "return", "video_top", "elif", "modality_type", "in", "(", "ModalityType", ".", "VIDEO_L1", ",", "ModalityType", ".", "VIDEO_L2", ")", ":", "return", "video_l1_top", "elif", "modality_type", "in", "(", "ModalityType", ".", "VIDEO_L1_RAW", ",", "ModalityType", ".", "VIDEO_L2_RAW", ")", ":", "return", "video_raw_top", "return", "value" ]
Gets default top transformation; if none available, return value.
[ "Gets", "default", "top", "transformation", ";", "if", "none", "available", "return", "value", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L1442-L1492
22,326
tensorflow/tensor2tensor
tensor2tensor/layers/modalities.py
get_weights_fn
def get_weights_fn(modality_type, value=None): """Gets default weights function; if none available, return value.""" if modality_type in (ModalityType.CTC_SYMBOL, ModalityType.IDENTITY_SYMBOL, ModalityType.MULTI_LABEL, ModalityType.SYMBOL, ModalityType.SYMBOL_ONE_HOT): return common_layers.weights_nonzero elif modality_type in ModalityType.get_choices(): return common_layers.weights_all return value
python
def get_weights_fn(modality_type, value=None): """Gets default weights function; if none available, return value.""" if modality_type in (ModalityType.CTC_SYMBOL, ModalityType.IDENTITY_SYMBOL, ModalityType.MULTI_LABEL, ModalityType.SYMBOL, ModalityType.SYMBOL_ONE_HOT): return common_layers.weights_nonzero elif modality_type in ModalityType.get_choices(): return common_layers.weights_all return value
[ "def", "get_weights_fn", "(", "modality_type", ",", "value", "=", "None", ")", ":", "if", "modality_type", "in", "(", "ModalityType", ".", "CTC_SYMBOL", ",", "ModalityType", ".", "IDENTITY_SYMBOL", ",", "ModalityType", ".", "MULTI_LABEL", ",", "ModalityType", ".", "SYMBOL", ",", "ModalityType", ".", "SYMBOL_ONE_HOT", ")", ":", "return", "common_layers", ".", "weights_nonzero", "elif", "modality_type", "in", "ModalityType", ".", "get_choices", "(", ")", ":", "return", "common_layers", ".", "weights_all", "return", "value" ]
Gets default weights function; if none available, return value.
[ "Gets", "default", "weights", "function", ";", "if", "none", "available", "return", "value", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/modalities.py#L1495-L1505
22,327
tensorflow/tensor2tensor
tensor2tensor/data_generators/paraphrase_ms_coco.py
create_combination
def create_combination(list_of_sentences): """Generates all possible pair combinations for the input list of sentences. For example: input = ["paraphrase1", "paraphrase2", "paraphrase3"] output = [("paraphrase1", "paraphrase2"), ("paraphrase1", "paraphrase3"), ("paraphrase2", "paraphrase3")] Args: list_of_sentences: the list of input sentences. Returns: the list of all possible sentence pairs. """ num_sentences = len(list_of_sentences) - 1 combinations = [] for i, _ in enumerate(list_of_sentences): if i == num_sentences: break num_pairs = num_sentences - i populated = num_pairs * [list_of_sentences[i]] zipped = list(zip(populated, list_of_sentences[i + 1:])) combinations += zipped return combinations
python
def create_combination(list_of_sentences): """Generates all possible pair combinations for the input list of sentences. For example: input = ["paraphrase1", "paraphrase2", "paraphrase3"] output = [("paraphrase1", "paraphrase2"), ("paraphrase1", "paraphrase3"), ("paraphrase2", "paraphrase3")] Args: list_of_sentences: the list of input sentences. Returns: the list of all possible sentence pairs. """ num_sentences = len(list_of_sentences) - 1 combinations = [] for i, _ in enumerate(list_of_sentences): if i == num_sentences: break num_pairs = num_sentences - i populated = num_pairs * [list_of_sentences[i]] zipped = list(zip(populated, list_of_sentences[i + 1:])) combinations += zipped return combinations
[ "def", "create_combination", "(", "list_of_sentences", ")", ":", "num_sentences", "=", "len", "(", "list_of_sentences", ")", "-", "1", "combinations", "=", "[", "]", "for", "i", ",", "_", "in", "enumerate", "(", "list_of_sentences", ")", ":", "if", "i", "==", "num_sentences", ":", "break", "num_pairs", "=", "num_sentences", "-", "i", "populated", "=", "num_pairs", "*", "[", "list_of_sentences", "[", "i", "]", "]", "zipped", "=", "list", "(", "zip", "(", "populated", ",", "list_of_sentences", "[", "i", "+", "1", ":", "]", ")", ")", "combinations", "+=", "zipped", "return", "combinations" ]
Generates all possible pair combinations for the input list of sentences. For example: input = ["paraphrase1", "paraphrase2", "paraphrase3"] output = [("paraphrase1", "paraphrase2"), ("paraphrase1", "paraphrase3"), ("paraphrase2", "paraphrase3")] Args: list_of_sentences: the list of input sentences. Returns: the list of all possible sentence pairs.
[ "Generates", "all", "possible", "pair", "combinations", "for", "the", "input", "list", "of", "sentences", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/paraphrase_ms_coco.py#L42-L67
22,328
tensorflow/tensor2tensor
tensor2tensor/models/image_transformer_2d.py
imagetransformer2d_base_8l_8_32_big
def imagetransformer2d_base_8l_8_32_big(): """hparams fo 8 layer big 2d model for cifar 10.""" hparams = image_transformer2d_base() hparams.num_heads = 16 hparams.hidden_size = 1024 hparams.filter_size = 2048 hparams.num_decoder_layers = 8 hparams.batch_size = 1 hparams.layer_prepostprocess_dropout = 0.3 hparams.query_shape = (8, 16) hparams.memory_flange = (0, 32) hparams.unconditional = int(False) return hparams
python
def imagetransformer2d_base_8l_8_32_big(): """hparams fo 8 layer big 2d model for cifar 10.""" hparams = image_transformer2d_base() hparams.num_heads = 16 hparams.hidden_size = 1024 hparams.filter_size = 2048 hparams.num_decoder_layers = 8 hparams.batch_size = 1 hparams.layer_prepostprocess_dropout = 0.3 hparams.query_shape = (8, 16) hparams.memory_flange = (0, 32) hparams.unconditional = int(False) return hparams
[ "def", "imagetransformer2d_base_8l_8_32_big", "(", ")", ":", "hparams", "=", "image_transformer2d_base", "(", ")", "hparams", ".", "num_heads", "=", "16", "hparams", ".", "hidden_size", "=", "1024", "hparams", ".", "filter_size", "=", "2048", "hparams", ".", "num_decoder_layers", "=", "8", "hparams", ".", "batch_size", "=", "1", "hparams", ".", "layer_prepostprocess_dropout", "=", "0.3", "hparams", ".", "query_shape", "=", "(", "8", ",", "16", ")", "hparams", ".", "memory_flange", "=", "(", "0", ",", "32", ")", "hparams", ".", "unconditional", "=", "int", "(", "False", ")", "return", "hparams" ]
hparams fo 8 layer big 2d model for cifar 10.
[ "hparams", "fo", "8", "layer", "big", "2d", "model", "for", "cifar", "10", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer_2d.py#L485-L497
22,329
tensorflow/tensor2tensor
tensor2tensor/models/image_transformer_2d.py
img2img_transformer2d_base
def img2img_transformer2d_base(): """Base params for img2img 2d attention.""" hparams = image_transformer2d_base() # learning related flags hparams.layer_preprocess_sequence = "n" hparams.layer_postprocess_sequence = "da" # This version seems to benefit from a higher learning rate. hparams.learning_rate = 0.2 hparams.layer_prepostprocess_dropout = 0.1 hparams.learning_rate_warmup_steps = 12000 hparams.filter_size = 2048 hparams.num_encoder_layers = 4 hparams.num_decoder_layers = 8 hparams.bottom["inputs"] = modalities.image_channel_embeddings_bottom hparams.dec_attention_type = cia.AttentionType.LOCAL_2D hparams.block_raster_scan = True return hparams
python
def img2img_transformer2d_base(): """Base params for img2img 2d attention.""" hparams = image_transformer2d_base() # learning related flags hparams.layer_preprocess_sequence = "n" hparams.layer_postprocess_sequence = "da" # This version seems to benefit from a higher learning rate. hparams.learning_rate = 0.2 hparams.layer_prepostprocess_dropout = 0.1 hparams.learning_rate_warmup_steps = 12000 hparams.filter_size = 2048 hparams.num_encoder_layers = 4 hparams.num_decoder_layers = 8 hparams.bottom["inputs"] = modalities.image_channel_embeddings_bottom hparams.dec_attention_type = cia.AttentionType.LOCAL_2D hparams.block_raster_scan = True return hparams
[ "def", "img2img_transformer2d_base", "(", ")", ":", "hparams", "=", "image_transformer2d_base", "(", ")", "# learning related flags", "hparams", ".", "layer_preprocess_sequence", "=", "\"n\"", "hparams", ".", "layer_postprocess_sequence", "=", "\"da\"", "# This version seems to benefit from a higher learning rate.", "hparams", ".", "learning_rate", "=", "0.2", "hparams", ".", "layer_prepostprocess_dropout", "=", "0.1", "hparams", ".", "learning_rate_warmup_steps", "=", "12000", "hparams", ".", "filter_size", "=", "2048", "hparams", ".", "num_encoder_layers", "=", "4", "hparams", ".", "num_decoder_layers", "=", "8", "hparams", ".", "bottom", "[", "\"inputs\"", "]", "=", "modalities", ".", "image_channel_embeddings_bottom", "hparams", ".", "dec_attention_type", "=", "cia", ".", "AttentionType", ".", "LOCAL_2D", "hparams", ".", "block_raster_scan", "=", "True", "return", "hparams" ]
Base params for img2img 2d attention.
[ "Base", "params", "for", "img2img", "2d", "attention", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer_2d.py#L585-L601
22,330
tensorflow/tensor2tensor
tensor2tensor/models/image_transformer_2d.py
img2img_transformer2d_q3
def img2img_transformer2d_q3(): """Current best hparams for local 2d.""" hparams = img2img_transformer2d_q1() hparams.batch_size = 2 hparams.query_shape = (8, 16) hparams.memory_flange = (8, 32) return hparams
python
def img2img_transformer2d_q3(): """Current best hparams for local 2d.""" hparams = img2img_transformer2d_q1() hparams.batch_size = 2 hparams.query_shape = (8, 16) hparams.memory_flange = (8, 32) return hparams
[ "def", "img2img_transformer2d_q3", "(", ")", ":", "hparams", "=", "img2img_transformer2d_q1", "(", ")", "hparams", ".", "batch_size", "=", "2", "hparams", ".", "query_shape", "=", "(", "8", ",", "16", ")", "hparams", ".", "memory_flange", "=", "(", "8", ",", "32", ")", "return", "hparams" ]
Current best hparams for local 2d.
[ "Current", "best", "hparams", "for", "local", "2d", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer_2d.py#L627-L633
22,331
tensorflow/tensor2tensor
tensor2tensor/models/image_transformer_2d.py
img2img_transformer_base
def img2img_transformer_base(): """Base params for local1d attention.""" hparams = image_transformer2d_base() # learning related flags hparams.layer_preprocess_sequence = "n" hparams.layer_postprocess_sequence = "da" # This version seems to benefit from a higher learning rate. hparams.learning_rate = 0.2 hparams.layer_prepostprocess_dropout = 0.1 hparams.learning_rate_warmup_steps = 12000 hparams.filter_size = 2048 hparams.num_encoder_layers = 4 hparams.num_decoder_layers = 8 hparams.block_length = 256 hparams.block_width = 256 hparams.dec_attention_type = cia.AttentionType.LOCAL_1D hparams.block_raster_scan = False return hparams
python
def img2img_transformer_base(): """Base params for local1d attention.""" hparams = image_transformer2d_base() # learning related flags hparams.layer_preprocess_sequence = "n" hparams.layer_postprocess_sequence = "da" # This version seems to benefit from a higher learning rate. hparams.learning_rate = 0.2 hparams.layer_prepostprocess_dropout = 0.1 hparams.learning_rate_warmup_steps = 12000 hparams.filter_size = 2048 hparams.num_encoder_layers = 4 hparams.num_decoder_layers = 8 hparams.block_length = 256 hparams.block_width = 256 hparams.dec_attention_type = cia.AttentionType.LOCAL_1D hparams.block_raster_scan = False return hparams
[ "def", "img2img_transformer_base", "(", ")", ":", "hparams", "=", "image_transformer2d_base", "(", ")", "# learning related flags", "hparams", ".", "layer_preprocess_sequence", "=", "\"n\"", "hparams", ".", "layer_postprocess_sequence", "=", "\"da\"", "# This version seems to benefit from a higher learning rate.", "hparams", ".", "learning_rate", "=", "0.2", "hparams", ".", "layer_prepostprocess_dropout", "=", "0.1", "hparams", ".", "learning_rate_warmup_steps", "=", "12000", "hparams", ".", "filter_size", "=", "2048", "hparams", ".", "num_encoder_layers", "=", "4", "hparams", ".", "num_decoder_layers", "=", "8", "hparams", ".", "block_length", "=", "256", "hparams", ".", "block_width", "=", "256", "hparams", ".", "dec_attention_type", "=", "cia", ".", "AttentionType", ".", "LOCAL_1D", "hparams", ".", "block_raster_scan", "=", "False", "return", "hparams" ]
Base params for local1d attention.
[ "Base", "params", "for", "local1d", "attention", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer_2d.py#L637-L654
22,332
tensorflow/tensor2tensor
tensor2tensor/models/image_transformer_2d.py
img2img_transformer_b3
def img2img_transformer_b3(): """Current best hparams for local 1d.""" hparams = img2img_transformer_base() hparams.batch_size = 2 hparams.layer_preprocess_sequence = "none" hparams.layer_postprocess_sequence = "dan" hparams.block_length = 128 hparams.sampling_temp = 0.9 return hparams
python
def img2img_transformer_b3(): """Current best hparams for local 1d.""" hparams = img2img_transformer_base() hparams.batch_size = 2 hparams.layer_preprocess_sequence = "none" hparams.layer_postprocess_sequence = "dan" hparams.block_length = 128 hparams.sampling_temp = 0.9 return hparams
[ "def", "img2img_transformer_b3", "(", ")", ":", "hparams", "=", "img2img_transformer_base", "(", ")", "hparams", ".", "batch_size", "=", "2", "hparams", ".", "layer_preprocess_sequence", "=", "\"none\"", "hparams", ".", "layer_postprocess_sequence", "=", "\"dan\"", "hparams", ".", "block_length", "=", "128", "hparams", ".", "sampling_temp", "=", "0.9", "return", "hparams" ]
Current best hparams for local 1d.
[ "Current", "best", "hparams", "for", "local", "1d", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer_2d.py#L678-L686
22,333
tensorflow/tensor2tensor
tensor2tensor/models/image_transformer_2d.py
img2img_transformer_dilated
def img2img_transformer_dilated(): """Try dilated.""" hparams = img2img_transformer_base() hparams.add_hparam("num_memory_blocks", 1) hparams.num_heads = 8 hparams.attention_key_channels = hparams.attention_value_channels = 0 hparams.hidden_size = 512 hparams.filter_size = 2048 hparams.num_decoder_layers = 8 hparams.sampling_method = "random" hparams.gap_sizes = [0, 16, 64, 0, 16, 64, 128, 0] hparams.dec_attention_type = cia.AttentionType.DILATED hparams.img_len = 64 hparams.block_length = 128 hparams.block_width = 128 return hparams
python
def img2img_transformer_dilated(): """Try dilated.""" hparams = img2img_transformer_base() hparams.add_hparam("num_memory_blocks", 1) hparams.num_heads = 8 hparams.attention_key_channels = hparams.attention_value_channels = 0 hparams.hidden_size = 512 hparams.filter_size = 2048 hparams.num_decoder_layers = 8 hparams.sampling_method = "random" hparams.gap_sizes = [0, 16, 64, 0, 16, 64, 128, 0] hparams.dec_attention_type = cia.AttentionType.DILATED hparams.img_len = 64 hparams.block_length = 128 hparams.block_width = 128 return hparams
[ "def", "img2img_transformer_dilated", "(", ")", ":", "hparams", "=", "img2img_transformer_base", "(", ")", "hparams", ".", "add_hparam", "(", "\"num_memory_blocks\"", ",", "1", ")", "hparams", ".", "num_heads", "=", "8", "hparams", ".", "attention_key_channels", "=", "hparams", ".", "attention_value_channels", "=", "0", "hparams", ".", "hidden_size", "=", "512", "hparams", ".", "filter_size", "=", "2048", "hparams", ".", "num_decoder_layers", "=", "8", "hparams", ".", "sampling_method", "=", "\"random\"", "hparams", ".", "gap_sizes", "=", "[", "0", ",", "16", ",", "64", ",", "0", ",", "16", ",", "64", ",", "128", ",", "0", "]", "hparams", ".", "dec_attention_type", "=", "cia", ".", "AttentionType", ".", "DILATED", "hparams", ".", "img_len", "=", "64", "hparams", ".", "block_length", "=", "128", "hparams", ".", "block_width", "=", "128", "return", "hparams" ]
Try dilated.
[ "Try", "dilated", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer_2d.py#L760-L775
22,334
tensorflow/tensor2tensor
tensor2tensor/models/image_transformer_2d.py
img2img_transformer_base_tpu
def img2img_transformer_base_tpu(): """Hparams for training img2img_transformer on tpu.""" hparams = img2img_transformer_base() update_hparams_for_tpu(hparams) hparams.batch_size = 2 hparams.num_heads = 4 # heads are expensive on tpu hparams.num_decoder_layers = 8 hparams.num_encoder_layers = 4 hparams.shared_embedding_and_softmax_weights = False return hparams
python
def img2img_transformer_base_tpu(): """Hparams for training img2img_transformer on tpu.""" hparams = img2img_transformer_base() update_hparams_for_tpu(hparams) hparams.batch_size = 2 hparams.num_heads = 4 # heads are expensive on tpu hparams.num_decoder_layers = 8 hparams.num_encoder_layers = 4 hparams.shared_embedding_and_softmax_weights = False return hparams
[ "def", "img2img_transformer_base_tpu", "(", ")", ":", "hparams", "=", "img2img_transformer_base", "(", ")", "update_hparams_for_tpu", "(", "hparams", ")", "hparams", ".", "batch_size", "=", "2", "hparams", ".", "num_heads", "=", "4", "# heads are expensive on tpu", "hparams", ".", "num_decoder_layers", "=", "8", "hparams", ".", "num_encoder_layers", "=", "4", "hparams", ".", "shared_embedding_and_softmax_weights", "=", "False", "return", "hparams" ]
Hparams for training img2img_transformer on tpu.
[ "Hparams", "for", "training", "img2img_transformer", "on", "tpu", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer_2d.py#L794-L803
22,335
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
ResidualFeedForward
def ResidualFeedForward(feature_depth, feedforward_depth, dropout, mode): """Residual feed-forward layer with normalization at start.""" return layers.Residual( layers.LayerNorm(), layers.Dense(feedforward_depth), layers.Relu(), layers.Dropout(rate=dropout, mode=mode), layers.Dense(feature_depth), layers.Dropout(rate=dropout, mode=mode) )
python
def ResidualFeedForward(feature_depth, feedforward_depth, dropout, mode): """Residual feed-forward layer with normalization at start.""" return layers.Residual( layers.LayerNorm(), layers.Dense(feedforward_depth), layers.Relu(), layers.Dropout(rate=dropout, mode=mode), layers.Dense(feature_depth), layers.Dropout(rate=dropout, mode=mode) )
[ "def", "ResidualFeedForward", "(", "feature_depth", ",", "feedforward_depth", ",", "dropout", ",", "mode", ")", ":", "return", "layers", ".", "Residual", "(", "layers", ".", "LayerNorm", "(", ")", ",", "layers", ".", "Dense", "(", "feedforward_depth", ")", ",", "layers", ".", "Relu", "(", ")", ",", "layers", ".", "Dropout", "(", "rate", "=", "dropout", ",", "mode", "=", "mode", ")", ",", "layers", ".", "Dense", "(", "feature_depth", ")", ",", "layers", ".", "Dropout", "(", "rate", "=", "dropout", ",", "mode", "=", "mode", ")", ")" ]
Residual feed-forward layer with normalization at start.
[ "Residual", "feed", "-", "forward", "layer", "with", "normalization", "at", "start", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L24-L36
22,336
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
EncoderLayer
def EncoderLayer(feature_depth, feedforward_depth, num_heads, dropout, mode): """Transformer encoder layer. The input to the encoder is a pair (embedded source, mask) where the mask is created from the original source to prevent attending to the padding part of the input. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) mode: str: 'train' or 'eval' Returns: the layer, returning a pair (actiavtions, mask). """ # The encoder block expects (activation, mask) as input and returns # the new activations only, we add the mask back to output next. encoder_block = layers.Serial( layers.Residual( # Attention block here. layers.Parallel(layers.LayerNorm(), layers.Identity()), layers.MultiHeadedAttention(feature_depth, num_heads=num_heads, dropout=dropout, mode=mode), layers.Dropout(rate=dropout, mode=mode), shortcut=layers.FirstBranch() ), ResidualFeedForward(feature_depth, feedforward_depth, dropout, mode=mode) ) # Now we add the mask back. return layers.Serial( layers.Reorder(output=((0, 1), 1)), # (x, mask) --> ((x, mask), mask) layers.Parallel(encoder_block, layers.Identity()) )
python
def EncoderLayer(feature_depth, feedforward_depth, num_heads, dropout, mode): """Transformer encoder layer. The input to the encoder is a pair (embedded source, mask) where the mask is created from the original source to prevent attending to the padding part of the input. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) mode: str: 'train' or 'eval' Returns: the layer, returning a pair (actiavtions, mask). """ # The encoder block expects (activation, mask) as input and returns # the new activations only, we add the mask back to output next. encoder_block = layers.Serial( layers.Residual( # Attention block here. layers.Parallel(layers.LayerNorm(), layers.Identity()), layers.MultiHeadedAttention(feature_depth, num_heads=num_heads, dropout=dropout, mode=mode), layers.Dropout(rate=dropout, mode=mode), shortcut=layers.FirstBranch() ), ResidualFeedForward(feature_depth, feedforward_depth, dropout, mode=mode) ) # Now we add the mask back. return layers.Serial( layers.Reorder(output=((0, 1), 1)), # (x, mask) --> ((x, mask), mask) layers.Parallel(encoder_block, layers.Identity()) )
[ "def", "EncoderLayer", "(", "feature_depth", ",", "feedforward_depth", ",", "num_heads", ",", "dropout", ",", "mode", ")", ":", "# The encoder block expects (activation, mask) as input and returns", "# the new activations only, we add the mask back to output next.", "encoder_block", "=", "layers", ".", "Serial", "(", "layers", ".", "Residual", "(", "# Attention block here.", "layers", ".", "Parallel", "(", "layers", ".", "LayerNorm", "(", ")", ",", "layers", ".", "Identity", "(", ")", ")", ",", "layers", ".", "MultiHeadedAttention", "(", "feature_depth", ",", "num_heads", "=", "num_heads", ",", "dropout", "=", "dropout", ",", "mode", "=", "mode", ")", ",", "layers", ".", "Dropout", "(", "rate", "=", "dropout", ",", "mode", "=", "mode", ")", ",", "shortcut", "=", "layers", ".", "FirstBranch", "(", ")", ")", ",", "ResidualFeedForward", "(", "feature_depth", ",", "feedforward_depth", ",", "dropout", ",", "mode", "=", "mode", ")", ")", "# Now we add the mask back.", "return", "layers", ".", "Serial", "(", "layers", ".", "Reorder", "(", "output", "=", "(", "(", "0", ",", "1", ")", ",", "1", ")", ")", ",", "# (x, mask) --> ((x, mask), mask)", "layers", ".", "Parallel", "(", "encoder_block", ",", "layers", ".", "Identity", "(", ")", ")", ")" ]
Transformer encoder layer. The input to the encoder is a pair (embedded source, mask) where the mask is created from the original source to prevent attending to the padding part of the input. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) mode: str: 'train' or 'eval' Returns: the layer, returning a pair (actiavtions, mask).
[ "Transformer", "encoder", "layer", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L39-L76
22,337
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
TransformerEncoder
def TransformerEncoder(vocab_size, num_classes=10, feature_depth=512, feedforward_depth=2048, num_layers=6, num_heads=8, dropout=0.1, max_len=2048, mode='train'): """Transformer encoder. Args: vocab_size: int: vocab size num_classes: how many classes on output feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_layers: int: number of encoder/decoder layers num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) max_len: int: maximum symbol length for positional encoding mode: str: 'train' or 'eval' Returns: the Transformer encoder layer. """ input_embedding = layers.Serial( layers.Embedding(feature_depth, vocab_size), layers.Dropout(rate=dropout, mode=mode), layers.PositionalEncoding(max_len=max_len) ) return layers.Serial( layers.Branch(), # Branch input to create embedding and mask. layers.Parallel(input_embedding, layers.PaddingMask()), layers.Serial(*[EncoderLayer(feature_depth, feedforward_depth, num_heads, dropout, mode) for _ in range(num_layers)]), layers.FirstBranch(), # Drop the mask. layers.LayerNorm(), layers.Mean(axis=1), # Average on length. layers.Dense(num_classes), layers.LogSoftmax() )
python
def TransformerEncoder(vocab_size, num_classes=10, feature_depth=512, feedforward_depth=2048, num_layers=6, num_heads=8, dropout=0.1, max_len=2048, mode='train'): """Transformer encoder. Args: vocab_size: int: vocab size num_classes: how many classes on output feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_layers: int: number of encoder/decoder layers num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) max_len: int: maximum symbol length for positional encoding mode: str: 'train' or 'eval' Returns: the Transformer encoder layer. """ input_embedding = layers.Serial( layers.Embedding(feature_depth, vocab_size), layers.Dropout(rate=dropout, mode=mode), layers.PositionalEncoding(max_len=max_len) ) return layers.Serial( layers.Branch(), # Branch input to create embedding and mask. layers.Parallel(input_embedding, layers.PaddingMask()), layers.Serial(*[EncoderLayer(feature_depth, feedforward_depth, num_heads, dropout, mode) for _ in range(num_layers)]), layers.FirstBranch(), # Drop the mask. layers.LayerNorm(), layers.Mean(axis=1), # Average on length. layers.Dense(num_classes), layers.LogSoftmax() )
[ "def", "TransformerEncoder", "(", "vocab_size", ",", "num_classes", "=", "10", ",", "feature_depth", "=", "512", ",", "feedforward_depth", "=", "2048", ",", "num_layers", "=", "6", ",", "num_heads", "=", "8", ",", "dropout", "=", "0.1", ",", "max_len", "=", "2048", ",", "mode", "=", "'train'", ")", ":", "input_embedding", "=", "layers", ".", "Serial", "(", "layers", ".", "Embedding", "(", "feature_depth", ",", "vocab_size", ")", ",", "layers", ".", "Dropout", "(", "rate", "=", "dropout", ",", "mode", "=", "mode", ")", ",", "layers", ".", "PositionalEncoding", "(", "max_len", "=", "max_len", ")", ")", "return", "layers", ".", "Serial", "(", "layers", ".", "Branch", "(", ")", ",", "# Branch input to create embedding and mask.", "layers", ".", "Parallel", "(", "input_embedding", ",", "layers", ".", "PaddingMask", "(", ")", ")", ",", "layers", ".", "Serial", "(", "*", "[", "EncoderLayer", "(", "feature_depth", ",", "feedforward_depth", ",", "num_heads", ",", "dropout", ",", "mode", ")", "for", "_", "in", "range", "(", "num_layers", ")", "]", ")", ",", "layers", ".", "FirstBranch", "(", ")", ",", "# Drop the mask.", "layers", ".", "LayerNorm", "(", ")", ",", "layers", ".", "Mean", "(", "axis", "=", "1", ")", ",", "# Average on length.", "layers", ".", "Dense", "(", "num_classes", ")", ",", "layers", ".", "LogSoftmax", "(", ")", ")" ]
Transformer encoder. Args: vocab_size: int: vocab size num_classes: how many classes on output feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_layers: int: number of encoder/decoder layers num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) max_len: int: maximum symbol length for positional encoding mode: str: 'train' or 'eval' Returns: the Transformer encoder layer.
[ "Transformer", "encoder", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L79-L120
22,338
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
DecoderLayer
def DecoderLayer(feature_depth, feedforward_depth, num_heads, dropout, mode): """Transformer decoder layer. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) mode: str: 'train' or 'eval' Returns: the layer. """ return layers.Serial( layers.Residual( # Self-attention block. layers.LayerNorm(), layers.Branch(), layers.Parallel(layers.Identity(), # activation for (q, k, v) layers.CausalMask(axis=-2)), # attention mask layers.MultiHeadedAttention(feature_depth, num_heads=num_heads, dropout=dropout, mode=mode), layers.Dropout(rate=dropout, mode=mode) ), ResidualFeedForward(feature_depth, feedforward_depth, dropout, mode=mode) )
python
def DecoderLayer(feature_depth, feedforward_depth, num_heads, dropout, mode): """Transformer decoder layer. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) mode: str: 'train' or 'eval' Returns: the layer. """ return layers.Serial( layers.Residual( # Self-attention block. layers.LayerNorm(), layers.Branch(), layers.Parallel(layers.Identity(), # activation for (q, k, v) layers.CausalMask(axis=-2)), # attention mask layers.MultiHeadedAttention(feature_depth, num_heads=num_heads, dropout=dropout, mode=mode), layers.Dropout(rate=dropout, mode=mode) ), ResidualFeedForward(feature_depth, feedforward_depth, dropout, mode=mode) )
[ "def", "DecoderLayer", "(", "feature_depth", ",", "feedforward_depth", ",", "num_heads", ",", "dropout", ",", "mode", ")", ":", "return", "layers", ".", "Serial", "(", "layers", ".", "Residual", "(", "# Self-attention block.", "layers", ".", "LayerNorm", "(", ")", ",", "layers", ".", "Branch", "(", ")", ",", "layers", ".", "Parallel", "(", "layers", ".", "Identity", "(", ")", ",", "# activation for (q, k, v)", "layers", ".", "CausalMask", "(", "axis", "=", "-", "2", ")", ")", ",", "# attention mask", "layers", ".", "MultiHeadedAttention", "(", "feature_depth", ",", "num_heads", "=", "num_heads", ",", "dropout", "=", "dropout", ",", "mode", "=", "mode", ")", ",", "layers", ".", "Dropout", "(", "rate", "=", "dropout", ",", "mode", "=", "mode", ")", ")", ",", "ResidualFeedForward", "(", "feature_depth", ",", "feedforward_depth", ",", "dropout", ",", "mode", "=", "mode", ")", ")" ]
Transformer decoder layer. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) mode: str: 'train' or 'eval' Returns: the layer.
[ "Transformer", "decoder", "layer", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L123-L151
22,339
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
ChunkedDecoderLayer
def ChunkedDecoderLayer(feature_depth, feedforward_depth, num_heads, dropout, chunk_selector, mode): """Transformer decoder layer operating on chunks. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) chunk_selector: a function from chunk number to list of chunks to attend. mode: str: 'train' or 'eval' Returns: the layer. """ return layers.Serial( layers.Residual( # Self-attention block. layers.Map(layers.LayerNorm()), layers.ChunkedCausalMultiHeadedAttention( feature_depth, num_heads=num_heads, dropout=dropout, chunk_selector=chunk_selector, mode=mode), layers.Map(layers.Dropout(rate=dropout, mode=mode)), ), layers.Map(ResidualFeedForward( feature_depth, feedforward_depth, dropout, mode=mode)) )
python
def ChunkedDecoderLayer(feature_depth, feedforward_depth, num_heads, dropout, chunk_selector, mode): """Transformer decoder layer operating on chunks. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) chunk_selector: a function from chunk number to list of chunks to attend. mode: str: 'train' or 'eval' Returns: the layer. """ return layers.Serial( layers.Residual( # Self-attention block. layers.Map(layers.LayerNorm()), layers.ChunkedCausalMultiHeadedAttention( feature_depth, num_heads=num_heads, dropout=dropout, chunk_selector=chunk_selector, mode=mode), layers.Map(layers.Dropout(rate=dropout, mode=mode)), ), layers.Map(ResidualFeedForward( feature_depth, feedforward_depth, dropout, mode=mode)) )
[ "def", "ChunkedDecoderLayer", "(", "feature_depth", ",", "feedforward_depth", ",", "num_heads", ",", "dropout", ",", "chunk_selector", ",", "mode", ")", ":", "return", "layers", ".", "Serial", "(", "layers", ".", "Residual", "(", "# Self-attention block.", "layers", ".", "Map", "(", "layers", ".", "LayerNorm", "(", ")", ")", ",", "layers", ".", "ChunkedCausalMultiHeadedAttention", "(", "feature_depth", ",", "num_heads", "=", "num_heads", ",", "dropout", "=", "dropout", ",", "chunk_selector", "=", "chunk_selector", ",", "mode", "=", "mode", ")", ",", "layers", ".", "Map", "(", "layers", ".", "Dropout", "(", "rate", "=", "dropout", ",", "mode", "=", "mode", ")", ")", ",", ")", ",", "layers", ".", "Map", "(", "ResidualFeedForward", "(", "feature_depth", ",", "feedforward_depth", ",", "dropout", ",", "mode", "=", "mode", ")", ")", ")" ]
Transformer decoder layer operating on chunks. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) chunk_selector: a function from chunk number to list of chunks to attend. mode: str: 'train' or 'eval' Returns: the layer.
[ "Transformer", "decoder", "layer", "operating", "on", "chunks", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L191-L220
22,340
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
ChunkedTransformerLM
def ChunkedTransformerLM(vocab_size, feature_depth=512, feedforward_depth=2048, num_layers=6, num_heads=8, dropout=0.1, chunk_selector=None, max_len=2048, mode='train'): """Transformer language model operating on chunks. The input to this model is a sequence presented as a list or tuple of chunks: (chunk1, chunk2, chunks3, ..., chunkN). Each chunk should have the same shape (batch, chunk-length) and together they represent a long sequence that's a concatenation chunk1,chunk2,...,chunkN. Chunked Transformer emulates the operation of a Transformer on this long sequence except for the chunked attention layer, which may attend to only a subset of the chunks to reduce memory use. Args: vocab_size: int: vocab size feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_layers: int: number of encoder/decoder layers num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) chunk_selector: a function from chunk number to list of chunks to attend (if None, attends to the previous chunks which is equivalent to setting chunk_selector(x) = [] if x < 1 else [x-1] (TransformerXL); we attend to the current chunk with a causal mask too, selected chunks unmasked). max_len: int: maximum symbol length for positional encoding mode: str: 'train' or 'eval' Returns: the layer. """ stack = [ChunkedDecoderLayer(feature_depth, feedforward_depth, num_heads, dropout, chunk_selector, mode) for _ in range(num_layers)] # Below each Map(L) applies the layer L to each chunk independently. return layers.Serial( layers.ShiftRight(), layers.Map(layers.Embedding(feature_depth, vocab_size)), layers.Map(layers.Dropout(rate=dropout, mode=mode)), layers.PositionalEncoding(max_len=max_len), layers.Serial(*stack), layers.Map(layers.LayerNorm()), layers.Map(layers.Dense(vocab_size)), layers.Map(layers.LogSoftmax()), )
python
def ChunkedTransformerLM(vocab_size, feature_depth=512, feedforward_depth=2048, num_layers=6, num_heads=8, dropout=0.1, chunk_selector=None, max_len=2048, mode='train'): """Transformer language model operating on chunks. The input to this model is a sequence presented as a list or tuple of chunks: (chunk1, chunk2, chunks3, ..., chunkN). Each chunk should have the same shape (batch, chunk-length) and together they represent a long sequence that's a concatenation chunk1,chunk2,...,chunkN. Chunked Transformer emulates the operation of a Transformer on this long sequence except for the chunked attention layer, which may attend to only a subset of the chunks to reduce memory use. Args: vocab_size: int: vocab size feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_layers: int: number of encoder/decoder layers num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) chunk_selector: a function from chunk number to list of chunks to attend (if None, attends to the previous chunks which is equivalent to setting chunk_selector(x) = [] if x < 1 else [x-1] (TransformerXL); we attend to the current chunk with a causal mask too, selected chunks unmasked). max_len: int: maximum symbol length for positional encoding mode: str: 'train' or 'eval' Returns: the layer. """ stack = [ChunkedDecoderLayer(feature_depth, feedforward_depth, num_heads, dropout, chunk_selector, mode) for _ in range(num_layers)] # Below each Map(L) applies the layer L to each chunk independently. return layers.Serial( layers.ShiftRight(), layers.Map(layers.Embedding(feature_depth, vocab_size)), layers.Map(layers.Dropout(rate=dropout, mode=mode)), layers.PositionalEncoding(max_len=max_len), layers.Serial(*stack), layers.Map(layers.LayerNorm()), layers.Map(layers.Dense(vocab_size)), layers.Map(layers.LogSoftmax()), )
[ "def", "ChunkedTransformerLM", "(", "vocab_size", ",", "feature_depth", "=", "512", ",", "feedforward_depth", "=", "2048", ",", "num_layers", "=", "6", ",", "num_heads", "=", "8", ",", "dropout", "=", "0.1", ",", "chunk_selector", "=", "None", ",", "max_len", "=", "2048", ",", "mode", "=", "'train'", ")", ":", "stack", "=", "[", "ChunkedDecoderLayer", "(", "feature_depth", ",", "feedforward_depth", ",", "num_heads", ",", "dropout", ",", "chunk_selector", ",", "mode", ")", "for", "_", "in", "range", "(", "num_layers", ")", "]", "# Below each Map(L) applies the layer L to each chunk independently.", "return", "layers", ".", "Serial", "(", "layers", ".", "ShiftRight", "(", ")", ",", "layers", ".", "Map", "(", "layers", ".", "Embedding", "(", "feature_depth", ",", "vocab_size", ")", ")", ",", "layers", ".", "Map", "(", "layers", ".", "Dropout", "(", "rate", "=", "dropout", ",", "mode", "=", "mode", ")", ")", ",", "layers", ".", "PositionalEncoding", "(", "max_len", "=", "max_len", ")", ",", "layers", ".", "Serial", "(", "*", "stack", ")", ",", "layers", ".", "Map", "(", "layers", ".", "LayerNorm", "(", ")", ")", ",", "layers", ".", "Map", "(", "layers", ".", "Dense", "(", "vocab_size", ")", ")", ",", "layers", ".", "Map", "(", "layers", ".", "LogSoftmax", "(", ")", ")", ",", ")" ]
Transformer language model operating on chunks. The input to this model is a sequence presented as a list or tuple of chunks: (chunk1, chunk2, chunks3, ..., chunkN). Each chunk should have the same shape (batch, chunk-length) and together they represent a long sequence that's a concatenation chunk1,chunk2,...,chunkN. Chunked Transformer emulates the operation of a Transformer on this long sequence except for the chunked attention layer, which may attend to only a subset of the chunks to reduce memory use. Args: vocab_size: int: vocab size feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_layers: int: number of encoder/decoder layers num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) chunk_selector: a function from chunk number to list of chunks to attend (if None, attends to the previous chunks which is equivalent to setting chunk_selector(x) = [] if x < 1 else [x-1] (TransformerXL); we attend to the current chunk with a causal mask too, selected chunks unmasked). max_len: int: maximum symbol length for positional encoding mode: str: 'train' or 'eval' Returns: the layer.
[ "Transformer", "language", "model", "operating", "on", "chunks", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L223-L273
22,341
tensorflow/tensor2tensor
tensor2tensor/models/mtf_transformer.py
mtf_transformer_paper_lm
def mtf_transformer_paper_lm(size): """Config for language-model experiments. Train these on languagemodel_lm1b32k_packed for 136000 steps (10 epochs) The size parameter is an integer that controls the number of heads and the size of the size of the feedforward hidden layers. Increasing size by 1 doubles each of these. Results: size params/10^9 log-ppl(per-token) -1 0.14 3.209 0 0.22 3.119 1 0.37 3.037 2 0.67 2.969 3 1.28 2.912 4 2.48 2.874 5 4.90 2.871 (to get word-level log-ppl, multiply by 1.1078) Args: size: an integer Returns: a hparams object """ n = 2 ** size hparams = mtf_transformer_base_lm() hparams.batch_size = 256 hparams.d_model = 1024 hparams.d_ff = int(8192 * n) hparams.d_kv = 256 hparams.num_heads = int(8 * n) hparams.shared_embedding_and_softmax_weights = False # one epoch for languagemodel_lm1b32k_packed = 13600 steps hparams.learning_rate_decay_steps = 13600 return hparams
python
def mtf_transformer_paper_lm(size): """Config for language-model experiments. Train these on languagemodel_lm1b32k_packed for 136000 steps (10 epochs) The size parameter is an integer that controls the number of heads and the size of the size of the feedforward hidden layers. Increasing size by 1 doubles each of these. Results: size params/10^9 log-ppl(per-token) -1 0.14 3.209 0 0.22 3.119 1 0.37 3.037 2 0.67 2.969 3 1.28 2.912 4 2.48 2.874 5 4.90 2.871 (to get word-level log-ppl, multiply by 1.1078) Args: size: an integer Returns: a hparams object """ n = 2 ** size hparams = mtf_transformer_base_lm() hparams.batch_size = 256 hparams.d_model = 1024 hparams.d_ff = int(8192 * n) hparams.d_kv = 256 hparams.num_heads = int(8 * n) hparams.shared_embedding_and_softmax_weights = False # one epoch for languagemodel_lm1b32k_packed = 13600 steps hparams.learning_rate_decay_steps = 13600 return hparams
[ "def", "mtf_transformer_paper_lm", "(", "size", ")", ":", "n", "=", "2", "**", "size", "hparams", "=", "mtf_transformer_base_lm", "(", ")", "hparams", ".", "batch_size", "=", "256", "hparams", ".", "d_model", "=", "1024", "hparams", ".", "d_ff", "=", "int", "(", "8192", "*", "n", ")", "hparams", ".", "d_kv", "=", "256", "hparams", ".", "num_heads", "=", "int", "(", "8", "*", "n", ")", "hparams", ".", "shared_embedding_and_softmax_weights", "=", "False", "# one epoch for languagemodel_lm1b32k_packed = 13600 steps", "hparams", ".", "learning_rate_decay_steps", "=", "13600", "return", "hparams" ]
Config for language-model experiments. Train these on languagemodel_lm1b32k_packed for 136000 steps (10 epochs) The size parameter is an integer that controls the number of heads and the size of the size of the feedforward hidden layers. Increasing size by 1 doubles each of these. Results: size params/10^9 log-ppl(per-token) -1 0.14 3.209 0 0.22 3.119 1 0.37 3.037 2 0.67 2.969 3 1.28 2.912 4 2.48 2.874 5 4.90 2.871 (to get word-level log-ppl, multiply by 1.1078) Args: size: an integer Returns: a hparams object
[ "Config", "for", "language", "-", "model", "experiments", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L953-L989
22,342
tensorflow/tensor2tensor
tensor2tensor/models/mtf_transformer.py
mtf_transformer_paper_tr
def mtf_transformer_paper_tr(size): """Config for translation experiments. Train these on translate_enfr_wmt32k_packed for 154000 steps (3 epochs) The size parameter is an integer that controls the number of heads and the size of the size of the feedforward hidden layers. Increasing size by 1 doubles each of these. Args: size: an integer Returns: a hparams object """ n = 2 ** size hparams = mtf_transformer_base() hparams.label_smoothing = 0.1 hparams.batch_size = 128 hparams.d_model = 1024 hparams.d_ff = int(4096 * n) hparams.num_heads = int(8 * n) hparams.shared_embedding_and_softmax_weights = False # one epoch for translate_enfr_wmt32k_packed = 51400 steps hparams.learning_rate_decay_steps = 51400 return hparams
python
def mtf_transformer_paper_tr(size): """Config for translation experiments. Train these on translate_enfr_wmt32k_packed for 154000 steps (3 epochs) The size parameter is an integer that controls the number of heads and the size of the size of the feedforward hidden layers. Increasing size by 1 doubles each of these. Args: size: an integer Returns: a hparams object """ n = 2 ** size hparams = mtf_transformer_base() hparams.label_smoothing = 0.1 hparams.batch_size = 128 hparams.d_model = 1024 hparams.d_ff = int(4096 * n) hparams.num_heads = int(8 * n) hparams.shared_embedding_and_softmax_weights = False # one epoch for translate_enfr_wmt32k_packed = 51400 steps hparams.learning_rate_decay_steps = 51400 return hparams
[ "def", "mtf_transformer_paper_tr", "(", "size", ")", ":", "n", "=", "2", "**", "size", "hparams", "=", "mtf_transformer_base", "(", ")", "hparams", ".", "label_smoothing", "=", "0.1", "hparams", ".", "batch_size", "=", "128", "hparams", ".", "d_model", "=", "1024", "hparams", ".", "d_ff", "=", "int", "(", "4096", "*", "n", ")", "hparams", ".", "num_heads", "=", "int", "(", "8", "*", "n", ")", "hparams", ".", "shared_embedding_and_softmax_weights", "=", "False", "# one epoch for translate_enfr_wmt32k_packed = 51400 steps", "hparams", ".", "learning_rate_decay_steps", "=", "51400", "return", "hparams" ]
Config for translation experiments. Train these on translate_enfr_wmt32k_packed for 154000 steps (3 epochs) The size parameter is an integer that controls the number of heads and the size of the size of the feedforward hidden layers. Increasing size by 1 doubles each of these. Args: size: an integer Returns: a hparams object
[ "Config", "for", "translation", "experiments", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L1041-L1065
22,343
tensorflow/tensor2tensor
tensor2tensor/models/mtf_transformer.py
mtf_transformer_lm_baseline
def mtf_transformer_lm_baseline(): """Small language model to run on 1 TPU. Run this on 2x2 on languagemodel_lm1b32k_packed for 272000 steps (10 epochs) Results: params/10^9 log-ppl(per-token) 0.14 3.202 Returns: a hparams """ hparams = mtf_transformer_paper_lm(-1) hparams.batch_size = 128 hparams.learning_rate_decay_steps = 27200 # one epoch on lm1b hparams.mesh_shape = "batch:8" return hparams
python
def mtf_transformer_lm_baseline(): """Small language model to run on 1 TPU. Run this on 2x2 on languagemodel_lm1b32k_packed for 272000 steps (10 epochs) Results: params/10^9 log-ppl(per-token) 0.14 3.202 Returns: a hparams """ hparams = mtf_transformer_paper_lm(-1) hparams.batch_size = 128 hparams.learning_rate_decay_steps = 27200 # one epoch on lm1b hparams.mesh_shape = "batch:8" return hparams
[ "def", "mtf_transformer_lm_baseline", "(", ")", ":", "hparams", "=", "mtf_transformer_paper_lm", "(", "-", "1", ")", "hparams", ".", "batch_size", "=", "128", "hparams", ".", "learning_rate_decay_steps", "=", "27200", "# one epoch on lm1b", "hparams", ".", "mesh_shape", "=", "\"batch:8\"", "return", "hparams" ]
Small language model to run on 1 TPU. Run this on 2x2 on languagemodel_lm1b32k_packed for 272000 steps (10 epochs) Results: params/10^9 log-ppl(per-token) 0.14 3.202 Returns: a hparams
[ "Small", "language", "model", "to", "run", "on", "1", "TPU", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L1171-L1186
22,344
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
graph_attention
def graph_attention(q, k, v, bias, dropout_rate=0.0, image_shapes=None, name=None, make_image_summary=True, save_weights_to=None, dropout_broadcast_dims=None, adjacency_matrix=None, num_edge_types=5): """graph attention. Args: q: a Tensor with shape [batch, heads, length_q, depth_k] k: a Tensor with shape [batch, heads, length_kv, depth_k] v: a Tensor with shape [batch, heads, length_kv, depth_v] bias: bias Tensor (see attention_bias()) dropout_rate: a floating point number image_shapes: optional tuple of integer scalars. see comments for attention_image_summary() name: an optional string make_image_summary: True if you want an image summary. save_weights_to: an optional dictionary to capture attention weights for vizualization; the weights tensor will be appended there under a string key created from the variable scope (including name). dropout_broadcast_dims: an optional list of integers less than 4 specifying in which dimensions to broadcast the dropout decisions. saves memory. adjacency_matrix: optional matrix of [batch, length, length] ids indicating edge type num_edge_types: an int indicating number of edge types Returns: A Tensor of shape [batch, length, depth(q)] """ with tf.variable_scope( name, default_name="dot_product_attention", values=[q, k, v]) as scope: # [batch, num_heads, query_length, memory_length] logits = tf.matmul(q, k, transpose_b=True) if adjacency_matrix is not None: key_head_depth = common_layers.shape_list(q)[-1] adjacency_vectors = make_edge_vectors( adjacency_matrix, num_edge_types, key_head_depth, name=name) # transposing q to be [batch, length_q, heads, depth_k] # to allow for matmul with [batch, length_q, length_q, depth_k] q_t = tf.transpose(q, [0, 2, 1, 3]) adj_logits = tf.matmul(q_t, adjacency_vectors, transpose_b=True) logits += tf.transpose(adj_logits, [0, 2, 1, 3]) # [batch, depth, num_nodes, num_nodes] if bias is not None: logits += bias weights = tf.nn.softmax(logits, name="attention_weights") if save_weights_to is not None: save_weights_to[scope.name] = weights # dropping out the attention links for each of the heads weights = common_layers.dropout_with_broadcast_dims( weights, 1.0 - dropout_rate, broadcast_dims=dropout_broadcast_dims) if common_layers.should_generate_summaries() and make_image_summary: common_attention.attention_image_summary(weights, image_shapes) return tf.matmul(weights, v)
python
def graph_attention(q, k, v, bias, dropout_rate=0.0, image_shapes=None, name=None, make_image_summary=True, save_weights_to=None, dropout_broadcast_dims=None, adjacency_matrix=None, num_edge_types=5): """graph attention. Args: q: a Tensor with shape [batch, heads, length_q, depth_k] k: a Tensor with shape [batch, heads, length_kv, depth_k] v: a Tensor with shape [batch, heads, length_kv, depth_v] bias: bias Tensor (see attention_bias()) dropout_rate: a floating point number image_shapes: optional tuple of integer scalars. see comments for attention_image_summary() name: an optional string make_image_summary: True if you want an image summary. save_weights_to: an optional dictionary to capture attention weights for vizualization; the weights tensor will be appended there under a string key created from the variable scope (including name). dropout_broadcast_dims: an optional list of integers less than 4 specifying in which dimensions to broadcast the dropout decisions. saves memory. adjacency_matrix: optional matrix of [batch, length, length] ids indicating edge type num_edge_types: an int indicating number of edge types Returns: A Tensor of shape [batch, length, depth(q)] """ with tf.variable_scope( name, default_name="dot_product_attention", values=[q, k, v]) as scope: # [batch, num_heads, query_length, memory_length] logits = tf.matmul(q, k, transpose_b=True) if adjacency_matrix is not None: key_head_depth = common_layers.shape_list(q)[-1] adjacency_vectors = make_edge_vectors( adjacency_matrix, num_edge_types, key_head_depth, name=name) # transposing q to be [batch, length_q, heads, depth_k] # to allow for matmul with [batch, length_q, length_q, depth_k] q_t = tf.transpose(q, [0, 2, 1, 3]) adj_logits = tf.matmul(q_t, adjacency_vectors, transpose_b=True) logits += tf.transpose(adj_logits, [0, 2, 1, 3]) # [batch, depth, num_nodes, num_nodes] if bias is not None: logits += bias weights = tf.nn.softmax(logits, name="attention_weights") if save_weights_to is not None: save_weights_to[scope.name] = weights # dropping out the attention links for each of the heads weights = common_layers.dropout_with_broadcast_dims( weights, 1.0 - dropout_rate, broadcast_dims=dropout_broadcast_dims) if common_layers.should_generate_summaries() and make_image_summary: common_attention.attention_image_summary(weights, image_shapes) return tf.matmul(weights, v)
[ "def", "graph_attention", "(", "q", ",", "k", ",", "v", ",", "bias", ",", "dropout_rate", "=", "0.0", ",", "image_shapes", "=", "None", ",", "name", "=", "None", ",", "make_image_summary", "=", "True", ",", "save_weights_to", "=", "None", ",", "dropout_broadcast_dims", "=", "None", ",", "adjacency_matrix", "=", "None", ",", "num_edge_types", "=", "5", ")", ":", "with", "tf", ".", "variable_scope", "(", "name", ",", "default_name", "=", "\"dot_product_attention\"", ",", "values", "=", "[", "q", ",", "k", ",", "v", "]", ")", "as", "scope", ":", "# [batch, num_heads, query_length, memory_length]", "logits", "=", "tf", ".", "matmul", "(", "q", ",", "k", ",", "transpose_b", "=", "True", ")", "if", "adjacency_matrix", "is", "not", "None", ":", "key_head_depth", "=", "common_layers", ".", "shape_list", "(", "q", ")", "[", "-", "1", "]", "adjacency_vectors", "=", "make_edge_vectors", "(", "adjacency_matrix", ",", "num_edge_types", ",", "key_head_depth", ",", "name", "=", "name", ")", "# transposing q to be [batch, length_q, heads, depth_k]", "# to allow for matmul with [batch, length_q, length_q, depth_k]", "q_t", "=", "tf", ".", "transpose", "(", "q", ",", "[", "0", ",", "2", ",", "1", ",", "3", "]", ")", "adj_logits", "=", "tf", ".", "matmul", "(", "q_t", ",", "adjacency_vectors", ",", "transpose_b", "=", "True", ")", "logits", "+=", "tf", ".", "transpose", "(", "adj_logits", ",", "[", "0", ",", "2", ",", "1", ",", "3", "]", ")", "# [batch, depth, num_nodes, num_nodes]", "if", "bias", "is", "not", "None", ":", "logits", "+=", "bias", "weights", "=", "tf", ".", "nn", ".", "softmax", "(", "logits", ",", "name", "=", "\"attention_weights\"", ")", "if", "save_weights_to", "is", "not", "None", ":", "save_weights_to", "[", "scope", ".", "name", "]", "=", "weights", "# dropping out the attention links for each of the heads", "weights", "=", "common_layers", ".", "dropout_with_broadcast_dims", "(", "weights", ",", "1.0", "-", "dropout_rate", ",", "broadcast_dims", "=", "dropout_broadcast_dims", ")", "if", "common_layers", ".", "should_generate_summaries", "(", ")", "and", "make_image_summary", ":", "common_attention", ".", "attention_image_summary", "(", "weights", ",", "image_shapes", ")", "return", "tf", ".", "matmul", "(", "weights", ",", "v", ")" ]
graph attention. Args: q: a Tensor with shape [batch, heads, length_q, depth_k] k: a Tensor with shape [batch, heads, length_kv, depth_k] v: a Tensor with shape [batch, heads, length_kv, depth_v] bias: bias Tensor (see attention_bias()) dropout_rate: a floating point number image_shapes: optional tuple of integer scalars. see comments for attention_image_summary() name: an optional string make_image_summary: True if you want an image summary. save_weights_to: an optional dictionary to capture attention weights for vizualization; the weights tensor will be appended there under a string key created from the variable scope (including name). dropout_broadcast_dims: an optional list of integers less than 4 specifying in which dimensions to broadcast the dropout decisions. saves memory. adjacency_matrix: optional matrix of [batch, length, length] ids indicating edge type num_edge_types: an int indicating number of edge types Returns: A Tensor of shape [batch, length, depth(q)]
[ "graph", "attention", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L185-L248
22,345
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
_compute_edge_transforms
def _compute_edge_transforms(node_states, depth, num_transforms, name="transform"): """Helper function that computes transformation for keys and values. Let B be the number of batches. Let N be the number of nodes in the graph. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries (total_key_depth). Let V be the size of the attention values (total_value_depth). Let T be the total number of transforms (num_transforms). Computes the transforms for keys or values for attention. * For each node N_j and edge type t, a key K_jt of size K is computed. When an edge of type t goes from node N_j to any other node, K_jt is the key that is in the attention process. * For each node N_j and edge type t, a value V_jt of size V is computed. When an edge of type t goes from node N_j to node N_i, Attention(Q_i, K_jt) produces a weight w_ijt. The message sent along this edge is w_ijt * V_jt. Args: node_states: A tensor of shape [B, L, D] depth: An integer (K or V) num_transforms: An integer (T), name: A name for the function Returns: x: A The attention keys or values for each node and edge type (shape [B, N*T, K or V]) """ node_shapes = common_layers.shape_list(node_states) x = common_layers.dense( node_states, depth * num_transforms, use_bias=False, name=name) batch = node_shapes[0] # B. length = node_shapes[1] # N. # Making the fourth dimension explicit by separating the vectors of size # K*T (in k) and V*T (in v) into two-dimensional matrices with shape [K, T] # (in k) and [V, T] in v. # x = tf.reshape(x, [batch, length, num_transforms, depth]) # Flatten out the fourth dimension. x = tf.reshape(x, [batch, length * num_transforms, depth]) return x
python
def _compute_edge_transforms(node_states, depth, num_transforms, name="transform"): """Helper function that computes transformation for keys and values. Let B be the number of batches. Let N be the number of nodes in the graph. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries (total_key_depth). Let V be the size of the attention values (total_value_depth). Let T be the total number of transforms (num_transforms). Computes the transforms for keys or values for attention. * For each node N_j and edge type t, a key K_jt of size K is computed. When an edge of type t goes from node N_j to any other node, K_jt is the key that is in the attention process. * For each node N_j and edge type t, a value V_jt of size V is computed. When an edge of type t goes from node N_j to node N_i, Attention(Q_i, K_jt) produces a weight w_ijt. The message sent along this edge is w_ijt * V_jt. Args: node_states: A tensor of shape [B, L, D] depth: An integer (K or V) num_transforms: An integer (T), name: A name for the function Returns: x: A The attention keys or values for each node and edge type (shape [B, N*T, K or V]) """ node_shapes = common_layers.shape_list(node_states) x = common_layers.dense( node_states, depth * num_transforms, use_bias=False, name=name) batch = node_shapes[0] # B. length = node_shapes[1] # N. # Making the fourth dimension explicit by separating the vectors of size # K*T (in k) and V*T (in v) into two-dimensional matrices with shape [K, T] # (in k) and [V, T] in v. # x = tf.reshape(x, [batch, length, num_transforms, depth]) # Flatten out the fourth dimension. x = tf.reshape(x, [batch, length * num_transforms, depth]) return x
[ "def", "_compute_edge_transforms", "(", "node_states", ",", "depth", ",", "num_transforms", ",", "name", "=", "\"transform\"", ")", ":", "node_shapes", "=", "common_layers", ".", "shape_list", "(", "node_states", ")", "x", "=", "common_layers", ".", "dense", "(", "node_states", ",", "depth", "*", "num_transforms", ",", "use_bias", "=", "False", ",", "name", "=", "name", ")", "batch", "=", "node_shapes", "[", "0", "]", "# B.", "length", "=", "node_shapes", "[", "1", "]", "# N.", "# Making the fourth dimension explicit by separating the vectors of size", "# K*T (in k) and V*T (in v) into two-dimensional matrices with shape [K, T]", "# (in k) and [V, T] in v.", "#", "x", "=", "tf", ".", "reshape", "(", "x", ",", "[", "batch", ",", "length", ",", "num_transforms", ",", "depth", "]", ")", "# Flatten out the fourth dimension.", "x", "=", "tf", ".", "reshape", "(", "x", ",", "[", "batch", ",", "length", "*", "num_transforms", ",", "depth", "]", ")", "return", "x" ]
Helper function that computes transformation for keys and values. Let B be the number of batches. Let N be the number of nodes in the graph. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries (total_key_depth). Let V be the size of the attention values (total_value_depth). Let T be the total number of transforms (num_transforms). Computes the transforms for keys or values for attention. * For each node N_j and edge type t, a key K_jt of size K is computed. When an edge of type t goes from node N_j to any other node, K_jt is the key that is in the attention process. * For each node N_j and edge type t, a value V_jt of size V is computed. When an edge of type t goes from node N_j to node N_i, Attention(Q_i, K_jt) produces a weight w_ijt. The message sent along this edge is w_ijt * V_jt. Args: node_states: A tensor of shape [B, L, D] depth: An integer (K or V) num_transforms: An integer (T), name: A name for the function Returns: x: A The attention keys or values for each node and edge type (shape [B, N*T, K or V])
[ "Helper", "function", "that", "computes", "transformation", "for", "keys", "and", "values", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L251-L301
22,346
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
compute_mpnn_qkv
def compute_mpnn_qkv(node_states, total_key_depth, total_value_depth, num_transforms): """Computes query, key and value for edge matrices. Let B be the number of batches. Let N be the number of nodes in the graph. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries (total_key_depth). Let V be the size of the attention values (total_value_depth). Let T be the total number of transforms (num_transforms). Computes the queries, keys, and values for attention. * For each node N_i in the graph, a query Q_i of size K is computed. This query is used to determine the relative weights to give to each of the node's incoming edges. * For each node N_j and edge type t, a key K_jt of size K is computed. When an edge of type t goes from node N_j to any other node, K_jt is the key that is in the attention process. * For each node N_j and edge type t, a value V_jt of size V is computed. When an edge of type t goes from node N_j to node N_i, Attention(Q_i, K_jt) produces a weight w_ijt. The message sent along this edge is w_ijt * V_jt. Args: node_states: A Tensor with shape [B, N, D]. total_key_depth: an integer (K). total_value_depth: an integer (V). num_transforms: a integer specifying number of transforms (T). This is typically the number of edge types. Returns: q: The attention queries for each destination node (shape [B, N, K]). k: The attention keys for each node and edge type (shape [B, N*T, K]). v: The attention values for each node and edge type (shape [B, N*T, V]). """ # node_states is initially a tensor with shape [B, N, D]. The call to dense # creates a D x K kernel that serves as a fully-connected layer. # # For each possible batch b and node n in the first two dimensions of # node_states, the corresponding size-D vector (the third dimension of # node_states) is the hidden state for node n in batch b. Each of these size-D # vectors is multiplied by the kernel to produce an attention query of size K. # The result is a tensor of size [B, N, K] containing the attention queries # for each node in each batch. q = common_layers.dense( node_states, total_key_depth, use_bias=False, name="q_mpnn") # Creates the attention keys in a manner similar to the process of creating # the attention queries. One key is created for each type of outgoing edge the # corresponding node might have, meaning k will have shape [B, N, K*T]. k = _compute_edge_transforms(node_states, total_key_depth, num_transforms, name="k_mpnn") v = _compute_edge_transforms(node_states, total_value_depth, num_transforms, name="v_mpnn") return q, k, v
python
def compute_mpnn_qkv(node_states, total_key_depth, total_value_depth, num_transforms): """Computes query, key and value for edge matrices. Let B be the number of batches. Let N be the number of nodes in the graph. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries (total_key_depth). Let V be the size of the attention values (total_value_depth). Let T be the total number of transforms (num_transforms). Computes the queries, keys, and values for attention. * For each node N_i in the graph, a query Q_i of size K is computed. This query is used to determine the relative weights to give to each of the node's incoming edges. * For each node N_j and edge type t, a key K_jt of size K is computed. When an edge of type t goes from node N_j to any other node, K_jt is the key that is in the attention process. * For each node N_j and edge type t, a value V_jt of size V is computed. When an edge of type t goes from node N_j to node N_i, Attention(Q_i, K_jt) produces a weight w_ijt. The message sent along this edge is w_ijt * V_jt. Args: node_states: A Tensor with shape [B, N, D]. total_key_depth: an integer (K). total_value_depth: an integer (V). num_transforms: a integer specifying number of transforms (T). This is typically the number of edge types. Returns: q: The attention queries for each destination node (shape [B, N, K]). k: The attention keys for each node and edge type (shape [B, N*T, K]). v: The attention values for each node and edge type (shape [B, N*T, V]). """ # node_states is initially a tensor with shape [B, N, D]. The call to dense # creates a D x K kernel that serves as a fully-connected layer. # # For each possible batch b and node n in the first two dimensions of # node_states, the corresponding size-D vector (the third dimension of # node_states) is the hidden state for node n in batch b. Each of these size-D # vectors is multiplied by the kernel to produce an attention query of size K. # The result is a tensor of size [B, N, K] containing the attention queries # for each node in each batch. q = common_layers.dense( node_states, total_key_depth, use_bias=False, name="q_mpnn") # Creates the attention keys in a manner similar to the process of creating # the attention queries. One key is created for each type of outgoing edge the # corresponding node might have, meaning k will have shape [B, N, K*T]. k = _compute_edge_transforms(node_states, total_key_depth, num_transforms, name="k_mpnn") v = _compute_edge_transforms(node_states, total_value_depth, num_transforms, name="v_mpnn") return q, k, v
[ "def", "compute_mpnn_qkv", "(", "node_states", ",", "total_key_depth", ",", "total_value_depth", ",", "num_transforms", ")", ":", "# node_states is initially a tensor with shape [B, N, D]. The call to dense", "# creates a D x K kernel that serves as a fully-connected layer.", "#", "# For each possible batch b and node n in the first two dimensions of", "# node_states, the corresponding size-D vector (the third dimension of", "# node_states) is the hidden state for node n in batch b. Each of these size-D", "# vectors is multiplied by the kernel to produce an attention query of size K.", "# The result is a tensor of size [B, N, K] containing the attention queries", "# for each node in each batch.", "q", "=", "common_layers", ".", "dense", "(", "node_states", ",", "total_key_depth", ",", "use_bias", "=", "False", ",", "name", "=", "\"q_mpnn\"", ")", "# Creates the attention keys in a manner similar to the process of creating", "# the attention queries. One key is created for each type of outgoing edge the", "# corresponding node might have, meaning k will have shape [B, N, K*T].", "k", "=", "_compute_edge_transforms", "(", "node_states", ",", "total_key_depth", ",", "num_transforms", ",", "name", "=", "\"k_mpnn\"", ")", "v", "=", "_compute_edge_transforms", "(", "node_states", ",", "total_value_depth", ",", "num_transforms", ",", "name", "=", "\"v_mpnn\"", ")", "return", "q", ",", "k", ",", "v" ]
Computes query, key and value for edge matrices. Let B be the number of batches. Let N be the number of nodes in the graph. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries (total_key_depth). Let V be the size of the attention values (total_value_depth). Let T be the total number of transforms (num_transforms). Computes the queries, keys, and values for attention. * For each node N_i in the graph, a query Q_i of size K is computed. This query is used to determine the relative weights to give to each of the node's incoming edges. * For each node N_j and edge type t, a key K_jt of size K is computed. When an edge of type t goes from node N_j to any other node, K_jt is the key that is in the attention process. * For each node N_j and edge type t, a value V_jt of size V is computed. When an edge of type t goes from node N_j to node N_i, Attention(Q_i, K_jt) produces a weight w_ijt. The message sent along this edge is w_ijt * V_jt. Args: node_states: A Tensor with shape [B, N, D]. total_key_depth: an integer (K). total_value_depth: an integer (V). num_transforms: a integer specifying number of transforms (T). This is typically the number of edge types. Returns: q: The attention queries for each destination node (shape [B, N, K]). k: The attention keys for each node and edge type (shape [B, N*T, K]). v: The attention values for each node and edge type (shape [B, N*T, V]).
[ "Computes", "query", "key", "and", "value", "for", "edge", "matrices", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L304-L364
22,347
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
sparse_message_pass_batched
def sparse_message_pass_batched(node_states, adjacency_matrices, num_edge_types, hidden_size, use_bias=True, average_aggregation=False, name="sparse_ggnn_batched"): """Identical to sparse_ggnn except that each input has a batch dimension. B = The batch size. N = The number of nodes in each batch. H = The size of the hidden states. T = The number of edge types. Args: node_states: Initial states of each node in the graph. Shape: [B, N, H] adjacency_matrices: Adjacency matrices of directed edges for each edge type and batch. Shape: [B, N, N, T] (sparse). num_edge_types: The number of edge types. T. hidden_size: The size of the hidden layer. H. use_bias: Whether to use bias in the hidden layer. average_aggregation: How to aggregate the incoming node messages. If average_aggregation is true, the messages are averaged. If it is false, they are summed. name: (optional) The scope within which tf variables should be created. Returns: The result of one round of message-passing of shape [B, N, H]. """ b, n = tf.shape(node_states)[0], tf.shape(node_states)[1] # Flatten the batch dimension of the node states. node_states = tf.reshape(node_states, [b*n, hidden_size]) # Flatten the batch dimension of the adjacency matrices. indices = adjacency_matrices.indices new_index2 = indices[:, 3] # The edge type dimension. # Offset N x N adjacency matrix by the batch number in which it appears. new_index0 = indices[:, 1] + indices[:, 0] * tf.cast(n, tf.int64) new_index1 = indices[:, 2] + indices[:, 0] * tf.cast(n, tf.int64) # Combine these indices as triples. new_indices = tf.stack([new_index0, new_index1, new_index2], axis=1) # Build the new sparse matrix. new_shape = [tf.cast(b*n, tf.int64), tf.cast(b*n, tf.int64), num_edge_types] adjacency_matrices = tf.SparseTensor(indices=new_indices, values=adjacency_matrices.values, dense_shape=new_shape) # Run a message-passing step and return the result with the batch dimension. node_states = sparse_message_pass( node_states, adjacency_matrices, num_edge_types, hidden_size, use_bias=use_bias, average_aggregation=average_aggregation, name=name) return tf.reshape(node_states, [b, n, hidden_size])
python
def sparse_message_pass_batched(node_states, adjacency_matrices, num_edge_types, hidden_size, use_bias=True, average_aggregation=False, name="sparse_ggnn_batched"): """Identical to sparse_ggnn except that each input has a batch dimension. B = The batch size. N = The number of nodes in each batch. H = The size of the hidden states. T = The number of edge types. Args: node_states: Initial states of each node in the graph. Shape: [B, N, H] adjacency_matrices: Adjacency matrices of directed edges for each edge type and batch. Shape: [B, N, N, T] (sparse). num_edge_types: The number of edge types. T. hidden_size: The size of the hidden layer. H. use_bias: Whether to use bias in the hidden layer. average_aggregation: How to aggregate the incoming node messages. If average_aggregation is true, the messages are averaged. If it is false, they are summed. name: (optional) The scope within which tf variables should be created. Returns: The result of one round of message-passing of shape [B, N, H]. """ b, n = tf.shape(node_states)[0], tf.shape(node_states)[1] # Flatten the batch dimension of the node states. node_states = tf.reshape(node_states, [b*n, hidden_size]) # Flatten the batch dimension of the adjacency matrices. indices = adjacency_matrices.indices new_index2 = indices[:, 3] # The edge type dimension. # Offset N x N adjacency matrix by the batch number in which it appears. new_index0 = indices[:, 1] + indices[:, 0] * tf.cast(n, tf.int64) new_index1 = indices[:, 2] + indices[:, 0] * tf.cast(n, tf.int64) # Combine these indices as triples. new_indices = tf.stack([new_index0, new_index1, new_index2], axis=1) # Build the new sparse matrix. new_shape = [tf.cast(b*n, tf.int64), tf.cast(b*n, tf.int64), num_edge_types] adjacency_matrices = tf.SparseTensor(indices=new_indices, values=adjacency_matrices.values, dense_shape=new_shape) # Run a message-passing step and return the result with the batch dimension. node_states = sparse_message_pass( node_states, adjacency_matrices, num_edge_types, hidden_size, use_bias=use_bias, average_aggregation=average_aggregation, name=name) return tf.reshape(node_states, [b, n, hidden_size])
[ "def", "sparse_message_pass_batched", "(", "node_states", ",", "adjacency_matrices", ",", "num_edge_types", ",", "hidden_size", ",", "use_bias", "=", "True", ",", "average_aggregation", "=", "False", ",", "name", "=", "\"sparse_ggnn_batched\"", ")", ":", "b", ",", "n", "=", "tf", ".", "shape", "(", "node_states", ")", "[", "0", "]", ",", "tf", ".", "shape", "(", "node_states", ")", "[", "1", "]", "# Flatten the batch dimension of the node states.", "node_states", "=", "tf", ".", "reshape", "(", "node_states", ",", "[", "b", "*", "n", ",", "hidden_size", "]", ")", "# Flatten the batch dimension of the adjacency matrices.", "indices", "=", "adjacency_matrices", ".", "indices", "new_index2", "=", "indices", "[", ":", ",", "3", "]", "# The edge type dimension.", "# Offset N x N adjacency matrix by the batch number in which it appears.", "new_index0", "=", "indices", "[", ":", ",", "1", "]", "+", "indices", "[", ":", ",", "0", "]", "*", "tf", ".", "cast", "(", "n", ",", "tf", ".", "int64", ")", "new_index1", "=", "indices", "[", ":", ",", "2", "]", "+", "indices", "[", ":", ",", "0", "]", "*", "tf", ".", "cast", "(", "n", ",", "tf", ".", "int64", ")", "# Combine these indices as triples.", "new_indices", "=", "tf", ".", "stack", "(", "[", "new_index0", ",", "new_index1", ",", "new_index2", "]", ",", "axis", "=", "1", ")", "# Build the new sparse matrix.", "new_shape", "=", "[", "tf", ".", "cast", "(", "b", "*", "n", ",", "tf", ".", "int64", ")", ",", "tf", ".", "cast", "(", "b", "*", "n", ",", "tf", ".", "int64", ")", ",", "num_edge_types", "]", "adjacency_matrices", "=", "tf", ".", "SparseTensor", "(", "indices", "=", "new_indices", ",", "values", "=", "adjacency_matrices", ".", "values", ",", "dense_shape", "=", "new_shape", ")", "# Run a message-passing step and return the result with the batch dimension.", "node_states", "=", "sparse_message_pass", "(", "node_states", ",", "adjacency_matrices", ",", "num_edge_types", ",", "hidden_size", ",", "use_bias", "=", "use_bias", ",", "average_aggregation", "=", "average_aggregation", ",", "name", "=", "name", ")", "return", "tf", ".", "reshape", "(", "node_states", ",", "[", "b", ",", "n", ",", "hidden_size", "]", ")" ]
Identical to sparse_ggnn except that each input has a batch dimension. B = The batch size. N = The number of nodes in each batch. H = The size of the hidden states. T = The number of edge types. Args: node_states: Initial states of each node in the graph. Shape: [B, N, H] adjacency_matrices: Adjacency matrices of directed edges for each edge type and batch. Shape: [B, N, N, T] (sparse). num_edge_types: The number of edge types. T. hidden_size: The size of the hidden layer. H. use_bias: Whether to use bias in the hidden layer. average_aggregation: How to aggregate the incoming node messages. If average_aggregation is true, the messages are averaged. If it is false, they are summed. name: (optional) The scope within which tf variables should be created. Returns: The result of one round of message-passing of shape [B, N, H].
[ "Identical", "to", "sparse_ggnn", "except", "that", "each", "input", "has", "a", "batch", "dimension", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L367-L428
22,348
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
sparse_message_pass
def sparse_message_pass(node_states, adjacency_matrices, num_edge_types, hidden_size, use_bias=True, average_aggregation=False, name="sparse_ggnn"): """One message-passing step for a GNN with a sparse adjacency matrix. Implements equation 2 (the message passing step) in [Li et al. 2015](https://arxiv.org/abs/1511.05493). N = The number of nodes in each batch. H = The size of the hidden states. T = The number of edge types. Args: node_states: Initial states of each node in the graph. Shape is [N, H]. adjacency_matrices: Adjacency matrix of directed edges for each edge type. Shape is [N, N, T] (sparse tensor). num_edge_types: The number of edge types. T. hidden_size: The size of the hidden state. H. use_bias: Whether to use bias in the hidden layer. average_aggregation: How to aggregate the incoming node messages. If average_aggregation is true, the messages are averaged. If it is false, they are summed. name: (optional) The scope within which tf variables should be created. Returns: The result of one step of Gated Graph Neural Network (GGNN) message passing. Shape: [N, H] """ n = tf.shape(node_states)[0] t = num_edge_types incoming_edges_per_type = tf.sparse_reduce_sum(adjacency_matrices, axis=1) # Convert the adjacency matrix into shape [T, N, N] - one [N, N] adjacency # matrix for each edge type. Since sparse tensor multiplication only supports # two-dimensional tensors, we actually convert the adjacency matrix into a # [T * N, N] tensor. adjacency_matrices = tf.sparse_transpose(adjacency_matrices, [2, 0, 1]) adjacency_matrices = tf.sparse_reshape(adjacency_matrices, [t * n, n]) # Multiply the adjacency matrix by the node states, producing a [T * N, H] # tensor. For each (edge type, node) pair, this tensor stores the sum of # the hidden states of the node's neighbors over incoming edges of that type. messages = tf.sparse_tensor_dense_matmul(adjacency_matrices, node_states) # Rearrange this tensor to have shape [N, T * H]. The incoming states of each # nodes neighbors are summed by edge type and then concatenated together into # a single T * H vector. messages = tf.reshape(messages, [t, n, hidden_size]) messages = tf.transpose(messages, [1, 0, 2]) messages = tf.reshape(messages, [n, t * hidden_size]) # Run each of those T * H vectors through a linear layer that produces # a vector of size H. This process is equivalent to running each H-sized # vector through a separate linear layer for each edge type and then adding # the results together. # # Note that, earlier on, we added together all of the states of neighbors # that were connected by edges of the same edge type. Since addition and # multiplying by a linear layer are commutative, this process was equivalent # to running each incoming edge through a linear layer separately and then # adding everything at the end. with tf.variable_scope(name, default_name="sparse_ggnn"): final_node_states = common_layers.dense( messages, hidden_size, use_bias=False) # Multiply the bias by for each edge type by the number of incoming nodes # of that edge type. if use_bias: bias = tf.get_variable("bias", initializer=tf.zeros([t, hidden_size])) final_node_states += tf.matmul(incoming_edges_per_type, bias) if average_aggregation: incoming_edges = tf.reduce_sum(incoming_edges_per_type, -1, keepdims=True) incoming_edges = tf.tile(incoming_edges, [1, hidden_size]) final_node_states /= incoming_edges + 1e-7 return tf.reshape(final_node_states, [n, hidden_size])
python
def sparse_message_pass(node_states, adjacency_matrices, num_edge_types, hidden_size, use_bias=True, average_aggregation=False, name="sparse_ggnn"): """One message-passing step for a GNN with a sparse adjacency matrix. Implements equation 2 (the message passing step) in [Li et al. 2015](https://arxiv.org/abs/1511.05493). N = The number of nodes in each batch. H = The size of the hidden states. T = The number of edge types. Args: node_states: Initial states of each node in the graph. Shape is [N, H]. adjacency_matrices: Adjacency matrix of directed edges for each edge type. Shape is [N, N, T] (sparse tensor). num_edge_types: The number of edge types. T. hidden_size: The size of the hidden state. H. use_bias: Whether to use bias in the hidden layer. average_aggregation: How to aggregate the incoming node messages. If average_aggregation is true, the messages are averaged. If it is false, they are summed. name: (optional) The scope within which tf variables should be created. Returns: The result of one step of Gated Graph Neural Network (GGNN) message passing. Shape: [N, H] """ n = tf.shape(node_states)[0] t = num_edge_types incoming_edges_per_type = tf.sparse_reduce_sum(adjacency_matrices, axis=1) # Convert the adjacency matrix into shape [T, N, N] - one [N, N] adjacency # matrix for each edge type. Since sparse tensor multiplication only supports # two-dimensional tensors, we actually convert the adjacency matrix into a # [T * N, N] tensor. adjacency_matrices = tf.sparse_transpose(adjacency_matrices, [2, 0, 1]) adjacency_matrices = tf.sparse_reshape(adjacency_matrices, [t * n, n]) # Multiply the adjacency matrix by the node states, producing a [T * N, H] # tensor. For each (edge type, node) pair, this tensor stores the sum of # the hidden states of the node's neighbors over incoming edges of that type. messages = tf.sparse_tensor_dense_matmul(adjacency_matrices, node_states) # Rearrange this tensor to have shape [N, T * H]. The incoming states of each # nodes neighbors are summed by edge type and then concatenated together into # a single T * H vector. messages = tf.reshape(messages, [t, n, hidden_size]) messages = tf.transpose(messages, [1, 0, 2]) messages = tf.reshape(messages, [n, t * hidden_size]) # Run each of those T * H vectors through a linear layer that produces # a vector of size H. This process is equivalent to running each H-sized # vector through a separate linear layer for each edge type and then adding # the results together. # # Note that, earlier on, we added together all of the states of neighbors # that were connected by edges of the same edge type. Since addition and # multiplying by a linear layer are commutative, this process was equivalent # to running each incoming edge through a linear layer separately and then # adding everything at the end. with tf.variable_scope(name, default_name="sparse_ggnn"): final_node_states = common_layers.dense( messages, hidden_size, use_bias=False) # Multiply the bias by for each edge type by the number of incoming nodes # of that edge type. if use_bias: bias = tf.get_variable("bias", initializer=tf.zeros([t, hidden_size])) final_node_states += tf.matmul(incoming_edges_per_type, bias) if average_aggregation: incoming_edges = tf.reduce_sum(incoming_edges_per_type, -1, keepdims=True) incoming_edges = tf.tile(incoming_edges, [1, hidden_size]) final_node_states /= incoming_edges + 1e-7 return tf.reshape(final_node_states, [n, hidden_size])
[ "def", "sparse_message_pass", "(", "node_states", ",", "adjacency_matrices", ",", "num_edge_types", ",", "hidden_size", ",", "use_bias", "=", "True", ",", "average_aggregation", "=", "False", ",", "name", "=", "\"sparse_ggnn\"", ")", ":", "n", "=", "tf", ".", "shape", "(", "node_states", ")", "[", "0", "]", "t", "=", "num_edge_types", "incoming_edges_per_type", "=", "tf", ".", "sparse_reduce_sum", "(", "adjacency_matrices", ",", "axis", "=", "1", ")", "# Convert the adjacency matrix into shape [T, N, N] - one [N, N] adjacency", "# matrix for each edge type. Since sparse tensor multiplication only supports", "# two-dimensional tensors, we actually convert the adjacency matrix into a", "# [T * N, N] tensor.", "adjacency_matrices", "=", "tf", ".", "sparse_transpose", "(", "adjacency_matrices", ",", "[", "2", ",", "0", ",", "1", "]", ")", "adjacency_matrices", "=", "tf", ".", "sparse_reshape", "(", "adjacency_matrices", ",", "[", "t", "*", "n", ",", "n", "]", ")", "# Multiply the adjacency matrix by the node states, producing a [T * N, H]", "# tensor. For each (edge type, node) pair, this tensor stores the sum of", "# the hidden states of the node's neighbors over incoming edges of that type.", "messages", "=", "tf", ".", "sparse_tensor_dense_matmul", "(", "adjacency_matrices", ",", "node_states", ")", "# Rearrange this tensor to have shape [N, T * H]. The incoming states of each", "# nodes neighbors are summed by edge type and then concatenated together into", "# a single T * H vector.", "messages", "=", "tf", ".", "reshape", "(", "messages", ",", "[", "t", ",", "n", ",", "hidden_size", "]", ")", "messages", "=", "tf", ".", "transpose", "(", "messages", ",", "[", "1", ",", "0", ",", "2", "]", ")", "messages", "=", "tf", ".", "reshape", "(", "messages", ",", "[", "n", ",", "t", "*", "hidden_size", "]", ")", "# Run each of those T * H vectors through a linear layer that produces", "# a vector of size H. This process is equivalent to running each H-sized", "# vector through a separate linear layer for each edge type and then adding", "# the results together.", "#", "# Note that, earlier on, we added together all of the states of neighbors", "# that were connected by edges of the same edge type. Since addition and", "# multiplying by a linear layer are commutative, this process was equivalent", "# to running each incoming edge through a linear layer separately and then", "# adding everything at the end.", "with", "tf", ".", "variable_scope", "(", "name", ",", "default_name", "=", "\"sparse_ggnn\"", ")", ":", "final_node_states", "=", "common_layers", ".", "dense", "(", "messages", ",", "hidden_size", ",", "use_bias", "=", "False", ")", "# Multiply the bias by for each edge type by the number of incoming nodes", "# of that edge type.", "if", "use_bias", ":", "bias", "=", "tf", ".", "get_variable", "(", "\"bias\"", ",", "initializer", "=", "tf", ".", "zeros", "(", "[", "t", ",", "hidden_size", "]", ")", ")", "final_node_states", "+=", "tf", ".", "matmul", "(", "incoming_edges_per_type", ",", "bias", ")", "if", "average_aggregation", ":", "incoming_edges", "=", "tf", ".", "reduce_sum", "(", "incoming_edges_per_type", ",", "-", "1", ",", "keepdims", "=", "True", ")", "incoming_edges", "=", "tf", ".", "tile", "(", "incoming_edges", ",", "[", "1", ",", "hidden_size", "]", ")", "final_node_states", "/=", "incoming_edges", "+", "1e-7", "return", "tf", ".", "reshape", "(", "final_node_states", ",", "[", "n", ",", "hidden_size", "]", ")" ]
One message-passing step for a GNN with a sparse adjacency matrix. Implements equation 2 (the message passing step) in [Li et al. 2015](https://arxiv.org/abs/1511.05493). N = The number of nodes in each batch. H = The size of the hidden states. T = The number of edge types. Args: node_states: Initial states of each node in the graph. Shape is [N, H]. adjacency_matrices: Adjacency matrix of directed edges for each edge type. Shape is [N, N, T] (sparse tensor). num_edge_types: The number of edge types. T. hidden_size: The size of the hidden state. H. use_bias: Whether to use bias in the hidden layer. average_aggregation: How to aggregate the incoming node messages. If average_aggregation is true, the messages are averaged. If it is false, they are summed. name: (optional) The scope within which tf variables should be created. Returns: The result of one step of Gated Graph Neural Network (GGNN) message passing. Shape: [N, H]
[ "One", "message", "-", "passing", "step", "for", "a", "GNN", "with", "a", "sparse", "adjacency", "matrix", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L431-L511
22,349
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
dot_product_mpnn_attention
def dot_product_mpnn_attention(q, k, v, adjacency_matrix, num_edge_types, num_transforms=None, use_weighted_sum=False, name=None): """Dot product attention with edge vectors. Let B be the number of batches. Let N be the number of nodes in the graph. Let K be the size of the attention keys/queries. Let V be the size of the attention values. Let T be the total number of transforms (num_transforms). Args: q: The query Tensor of shape [B, N, K]. k: The key Tensor of shape [B, T, N, K]. v: The value Tensor of shape [B, T, N, V]. adjacency_matrix: A Tensor of shape [B, N, N, T]. An entry at indices b, i, j, k is the indicator of the edge from node j to node i in batch b. A standard adjacency matrix will only have one edge type while a mutigraph will have multiple edge types. num_edge_types: An integer specifying number of edge types. num_transforms: An integer indicating number of transforms (T). If None, then num_transforms will be equal to num_edge_types. use_weighted_sum: If False, will only use a single transform per edge type. Otherwise, use a learned weighted sum of transforms per edge type. name: A string. Returns: A Tensor of shape [B, N, V] storing the result of computing attention weights using the queries and keys and combining the values according to those weights. Raises: ValueError: if num_transforms doesn't equal num_edge_types and not using weighted sum. """ with tf.variable_scope( name, default_name="dot_product_mpnn_attention", values=[q, k, v, adjacency_matrix, num_edge_types]): # If not explicitly set, use num_transforms set to num_edge_types. num_transforms = ( num_edge_types if num_transforms is None else num_transforms) if not use_weighted_sum and num_transforms != num_edge_types: raise ValueError("num_transforms must equal num_edge_types unless " "use_weighted_sum is True") # Computes the raw dot-product attention values between each query and # the corresponding keys it needs to consider. # # This operation takes the dot product of (the query for # each node) and (the key for each node for each possible edge type), # creating an N x N matrix for each edge type. The entry at index (i, j) # is the dot-product for the edge from node i to node j of the appropriate # type. These dot products will eventually become attention weights # specifying how much node i weights an edge of that type coming from node # j. all_edge_logits = tf.matmul( tf.tile(tf.expand_dims(q, axis=1), [1, num_edge_types, 1, 1]), k, transpose_b=True) # The adjacency matrix assumes there is only one directed edge (i <- j) for # each pair of nodes. If such an edge exists, it contains the integer # type of that edge at position (i, j) of the adjacency matrix. # # Construct edge_vectors of shape [B, N, N, T]. if use_weighted_sum: # Use dense representation for edge vectors. edge_vectors = make_edge_vectors( adjacency_matrix, num_edge_types, num_transforms) else: # Generate one-hot vectors based on edge types. # If there is an edge from node j to node i of type t, then index t of the # last dimension is 1 for entry (i, j) of the second and third dimensions. edge_vectors = tf.one_hot(adjacency_matrix, num_transforms) # Rearranging the dimensions to match the shape of all_edge_logits. edge_vectors = tf.transpose(edge_vectors, [0, 3, 1, 2]) # Element-wise multiplies all_edge_logits and edge_vectors. # # In other words: all_edge_logits contains N x N matrices of query-key # products. This element-wise multiplication zeroes out entries that do not # correspond to actual edges in the graph of the appropriate edge type. # all_edge_logits retains shape [B, T, N, N]. all_edge_logits *= edge_vectors # Since there can only be one edge from node A to node B, we can collapse # the T different adjacency matrices containing key-query pairs into one # adjacency matrix. logits is [B, N, N]. # TODO(dbieber): Use a reshape instead of reduce sum to attend over all # edges instead of over all neighboring nodes to handle the multigraph case. logits = tf.reduce_sum(all_edge_logits, axis=1) # For pairs of nodes with no edges between them, add a large negative bias # to each location without an edge so that the softmax of entries with the # value 0 become a small negative number instead. bias = 0 bias = tf.to_float(tf.equal( tf.reduce_sum(adjacency_matrix, axis=-1), 0)) * -1e9 logits += bias # Turn the raw key-query products into a probability distribution (or, # in terms of attention, weights). The softmax is computed across the # last dimension of logits. compatibility = tf.nn.softmax(logits) # Shape [B, N, N]. # Computes a summary showing the attention matrix as an image. Does not do # any work toward actually performing attention. common_attention.attention_image_summary( tf.expand_dims(compatibility, axis=1), None) # Repeats the attention matrix T times for each batch, producing # a tensor with shape [B, T, N, N] where the [N, N] component is T # repeats of the values found in compatibility. edge_compatibility = tf.tile( tf.expand_dims(compatibility, axis=1), [1, num_edge_types, 1, 1]) # Zeroes out the entries in edge_compatibility that do not correspond to # actual edges. edge_compatibility *= edge_vectors # Shape [B, T, N, N]. output = compute_values(edge_compatibility, v) return output
python
def dot_product_mpnn_attention(q, k, v, adjacency_matrix, num_edge_types, num_transforms=None, use_weighted_sum=False, name=None): """Dot product attention with edge vectors. Let B be the number of batches. Let N be the number of nodes in the graph. Let K be the size of the attention keys/queries. Let V be the size of the attention values. Let T be the total number of transforms (num_transforms). Args: q: The query Tensor of shape [B, N, K]. k: The key Tensor of shape [B, T, N, K]. v: The value Tensor of shape [B, T, N, V]. adjacency_matrix: A Tensor of shape [B, N, N, T]. An entry at indices b, i, j, k is the indicator of the edge from node j to node i in batch b. A standard adjacency matrix will only have one edge type while a mutigraph will have multiple edge types. num_edge_types: An integer specifying number of edge types. num_transforms: An integer indicating number of transforms (T). If None, then num_transforms will be equal to num_edge_types. use_weighted_sum: If False, will only use a single transform per edge type. Otherwise, use a learned weighted sum of transforms per edge type. name: A string. Returns: A Tensor of shape [B, N, V] storing the result of computing attention weights using the queries and keys and combining the values according to those weights. Raises: ValueError: if num_transforms doesn't equal num_edge_types and not using weighted sum. """ with tf.variable_scope( name, default_name="dot_product_mpnn_attention", values=[q, k, v, adjacency_matrix, num_edge_types]): # If not explicitly set, use num_transforms set to num_edge_types. num_transforms = ( num_edge_types if num_transforms is None else num_transforms) if not use_weighted_sum and num_transforms != num_edge_types: raise ValueError("num_transforms must equal num_edge_types unless " "use_weighted_sum is True") # Computes the raw dot-product attention values between each query and # the corresponding keys it needs to consider. # # This operation takes the dot product of (the query for # each node) and (the key for each node for each possible edge type), # creating an N x N matrix for each edge type. The entry at index (i, j) # is the dot-product for the edge from node i to node j of the appropriate # type. These dot products will eventually become attention weights # specifying how much node i weights an edge of that type coming from node # j. all_edge_logits = tf.matmul( tf.tile(tf.expand_dims(q, axis=1), [1, num_edge_types, 1, 1]), k, transpose_b=True) # The adjacency matrix assumes there is only one directed edge (i <- j) for # each pair of nodes. If such an edge exists, it contains the integer # type of that edge at position (i, j) of the adjacency matrix. # # Construct edge_vectors of shape [B, N, N, T]. if use_weighted_sum: # Use dense representation for edge vectors. edge_vectors = make_edge_vectors( adjacency_matrix, num_edge_types, num_transforms) else: # Generate one-hot vectors based on edge types. # If there is an edge from node j to node i of type t, then index t of the # last dimension is 1 for entry (i, j) of the second and third dimensions. edge_vectors = tf.one_hot(adjacency_matrix, num_transforms) # Rearranging the dimensions to match the shape of all_edge_logits. edge_vectors = tf.transpose(edge_vectors, [0, 3, 1, 2]) # Element-wise multiplies all_edge_logits and edge_vectors. # # In other words: all_edge_logits contains N x N matrices of query-key # products. This element-wise multiplication zeroes out entries that do not # correspond to actual edges in the graph of the appropriate edge type. # all_edge_logits retains shape [B, T, N, N]. all_edge_logits *= edge_vectors # Since there can only be one edge from node A to node B, we can collapse # the T different adjacency matrices containing key-query pairs into one # adjacency matrix. logits is [B, N, N]. # TODO(dbieber): Use a reshape instead of reduce sum to attend over all # edges instead of over all neighboring nodes to handle the multigraph case. logits = tf.reduce_sum(all_edge_logits, axis=1) # For pairs of nodes with no edges between them, add a large negative bias # to each location without an edge so that the softmax of entries with the # value 0 become a small negative number instead. bias = 0 bias = tf.to_float(tf.equal( tf.reduce_sum(adjacency_matrix, axis=-1), 0)) * -1e9 logits += bias # Turn the raw key-query products into a probability distribution (or, # in terms of attention, weights). The softmax is computed across the # last dimension of logits. compatibility = tf.nn.softmax(logits) # Shape [B, N, N]. # Computes a summary showing the attention matrix as an image. Does not do # any work toward actually performing attention. common_attention.attention_image_summary( tf.expand_dims(compatibility, axis=1), None) # Repeats the attention matrix T times for each batch, producing # a tensor with shape [B, T, N, N] where the [N, N] component is T # repeats of the values found in compatibility. edge_compatibility = tf.tile( tf.expand_dims(compatibility, axis=1), [1, num_edge_types, 1, 1]) # Zeroes out the entries in edge_compatibility that do not correspond to # actual edges. edge_compatibility *= edge_vectors # Shape [B, T, N, N]. output = compute_values(edge_compatibility, v) return output
[ "def", "dot_product_mpnn_attention", "(", "q", ",", "k", ",", "v", ",", "adjacency_matrix", ",", "num_edge_types", ",", "num_transforms", "=", "None", ",", "use_weighted_sum", "=", "False", ",", "name", "=", "None", ")", ":", "with", "tf", ".", "variable_scope", "(", "name", ",", "default_name", "=", "\"dot_product_mpnn_attention\"", ",", "values", "=", "[", "q", ",", "k", ",", "v", ",", "adjacency_matrix", ",", "num_edge_types", "]", ")", ":", "# If not explicitly set, use num_transforms set to num_edge_types.", "num_transforms", "=", "(", "num_edge_types", "if", "num_transforms", "is", "None", "else", "num_transforms", ")", "if", "not", "use_weighted_sum", "and", "num_transforms", "!=", "num_edge_types", ":", "raise", "ValueError", "(", "\"num_transforms must equal num_edge_types unless \"", "\"use_weighted_sum is True\"", ")", "# Computes the raw dot-product attention values between each query and", "# the corresponding keys it needs to consider.", "#", "# This operation takes the dot product of (the query for", "# each node) and (the key for each node for each possible edge type),", "# creating an N x N matrix for each edge type. The entry at index (i, j)", "# is the dot-product for the edge from node i to node j of the appropriate", "# type. These dot products will eventually become attention weights", "# specifying how much node i weights an edge of that type coming from node", "# j.", "all_edge_logits", "=", "tf", ".", "matmul", "(", "tf", ".", "tile", "(", "tf", ".", "expand_dims", "(", "q", ",", "axis", "=", "1", ")", ",", "[", "1", ",", "num_edge_types", ",", "1", ",", "1", "]", ")", ",", "k", ",", "transpose_b", "=", "True", ")", "# The adjacency matrix assumes there is only one directed edge (i <- j) for", "# each pair of nodes. If such an edge exists, it contains the integer", "# type of that edge at position (i, j) of the adjacency matrix.", "#", "# Construct edge_vectors of shape [B, N, N, T].", "if", "use_weighted_sum", ":", "# Use dense representation for edge vectors.", "edge_vectors", "=", "make_edge_vectors", "(", "adjacency_matrix", ",", "num_edge_types", ",", "num_transforms", ")", "else", ":", "# Generate one-hot vectors based on edge types.", "# If there is an edge from node j to node i of type t, then index t of the", "# last dimension is 1 for entry (i, j) of the second and third dimensions.", "edge_vectors", "=", "tf", ".", "one_hot", "(", "adjacency_matrix", ",", "num_transforms", ")", "# Rearranging the dimensions to match the shape of all_edge_logits.", "edge_vectors", "=", "tf", ".", "transpose", "(", "edge_vectors", ",", "[", "0", ",", "3", ",", "1", ",", "2", "]", ")", "# Element-wise multiplies all_edge_logits and edge_vectors.", "#", "# In other words: all_edge_logits contains N x N matrices of query-key", "# products. This element-wise multiplication zeroes out entries that do not", "# correspond to actual edges in the graph of the appropriate edge type.", "# all_edge_logits retains shape [B, T, N, N].", "all_edge_logits", "*=", "edge_vectors", "# Since there can only be one edge from node A to node B, we can collapse", "# the T different adjacency matrices containing key-query pairs into one", "# adjacency matrix. logits is [B, N, N].", "# TODO(dbieber): Use a reshape instead of reduce sum to attend over all", "# edges instead of over all neighboring nodes to handle the multigraph case.", "logits", "=", "tf", ".", "reduce_sum", "(", "all_edge_logits", ",", "axis", "=", "1", ")", "# For pairs of nodes with no edges between them, add a large negative bias", "# to each location without an edge so that the softmax of entries with the", "# value 0 become a small negative number instead.", "bias", "=", "0", "bias", "=", "tf", ".", "to_float", "(", "tf", ".", "equal", "(", "tf", ".", "reduce_sum", "(", "adjacency_matrix", ",", "axis", "=", "-", "1", ")", ",", "0", ")", ")", "*", "-", "1e9", "logits", "+=", "bias", "# Turn the raw key-query products into a probability distribution (or,", "# in terms of attention, weights). The softmax is computed across the", "# last dimension of logits.", "compatibility", "=", "tf", ".", "nn", ".", "softmax", "(", "logits", ")", "# Shape [B, N, N].", "# Computes a summary showing the attention matrix as an image. Does not do", "# any work toward actually performing attention.", "common_attention", ".", "attention_image_summary", "(", "tf", ".", "expand_dims", "(", "compatibility", ",", "axis", "=", "1", ")", ",", "None", ")", "# Repeats the attention matrix T times for each batch, producing", "# a tensor with shape [B, T, N, N] where the [N, N] component is T", "# repeats of the values found in compatibility.", "edge_compatibility", "=", "tf", ".", "tile", "(", "tf", ".", "expand_dims", "(", "compatibility", ",", "axis", "=", "1", ")", ",", "[", "1", ",", "num_edge_types", ",", "1", ",", "1", "]", ")", "# Zeroes out the entries in edge_compatibility that do not correspond to", "# actual edges.", "edge_compatibility", "*=", "edge_vectors", "# Shape [B, T, N, N].", "output", "=", "compute_values", "(", "edge_compatibility", ",", "v", ")", "return", "output" ]
Dot product attention with edge vectors. Let B be the number of batches. Let N be the number of nodes in the graph. Let K be the size of the attention keys/queries. Let V be the size of the attention values. Let T be the total number of transforms (num_transforms). Args: q: The query Tensor of shape [B, N, K]. k: The key Tensor of shape [B, T, N, K]. v: The value Tensor of shape [B, T, N, V]. adjacency_matrix: A Tensor of shape [B, N, N, T]. An entry at indices b, i, j, k is the indicator of the edge from node j to node i in batch b. A standard adjacency matrix will only have one edge type while a mutigraph will have multiple edge types. num_edge_types: An integer specifying number of edge types. num_transforms: An integer indicating number of transforms (T). If None, then num_transforms will be equal to num_edge_types. use_weighted_sum: If False, will only use a single transform per edge type. Otherwise, use a learned weighted sum of transforms per edge type. name: A string. Returns: A Tensor of shape [B, N, V] storing the result of computing attention weights using the queries and keys and combining the values according to those weights. Raises: ValueError: if num_transforms doesn't equal num_edge_types and not using weighted sum.
[ "Dot", "product", "attention", "with", "edge", "vectors", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L652-L783
22,350
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
ggnn_fast_dense
def ggnn_fast_dense(node_states, adjacency_matrix, num_edge_types, total_value_depth, name=None): """ggnn version of the MPNN from Gilmer et al. Let B be the number of batches. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries. Let V be the size of the output of the ggnn. Let T be the number of transforms / edge types. Args: node_states: The value Tensor of shape [B, T, N, D]. adjacency_matrix: A Tensor of shape [B, N, N, T]. An entry at indices b, i, j, k is the indicator of the edge from node j to node i in batch b. A standard adjacency matrix will only have values of one, while a mutigraph may have larger integer values. num_edge_types: An integer specifying number of edge types. total_value_depth: An integer (V) name: A string. Returns: A Tensor of shape [B, N, V] storing the result of computing attention weights using the queries and keys and combining the values according to those weights. Raises: ValueError: if num_transforms doesn't equal num_edge_types and not using weighted sum. """ # between the same nodes (with only one edge of each type. adjacency_matrix # will need to be converted to shape [B, T, N, N]. with tf.variable_scope( name, default_name="ggnn_fast_dense", values=[node_states, adjacency_matrix, num_edge_types]): nodes_shape = common_layers.shape_list(node_states) v = _compute_edge_transforms(node_states, total_value_depth, num_edge_types, name="v_mpnn") v = tf.reshape(v, [nodes_shape[0], nodes_shape[1], num_edge_types, total_value_depth ]) # Shape [B, N, T, V]. v = tf.transpose(v, [0, 2, 1, 3]) # Shape [B, T, N, V]. # Rearranging the dimensions to match the shape of all_edge_logits. edge_vectors = tf.transpose(adjacency_matrix, [0, 3, 1, 2]) output = compute_values(edge_vectors, v) return output
python
def ggnn_fast_dense(node_states, adjacency_matrix, num_edge_types, total_value_depth, name=None): """ggnn version of the MPNN from Gilmer et al. Let B be the number of batches. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries. Let V be the size of the output of the ggnn. Let T be the number of transforms / edge types. Args: node_states: The value Tensor of shape [B, T, N, D]. adjacency_matrix: A Tensor of shape [B, N, N, T]. An entry at indices b, i, j, k is the indicator of the edge from node j to node i in batch b. A standard adjacency matrix will only have values of one, while a mutigraph may have larger integer values. num_edge_types: An integer specifying number of edge types. total_value_depth: An integer (V) name: A string. Returns: A Tensor of shape [B, N, V] storing the result of computing attention weights using the queries and keys and combining the values according to those weights. Raises: ValueError: if num_transforms doesn't equal num_edge_types and not using weighted sum. """ # between the same nodes (with only one edge of each type. adjacency_matrix # will need to be converted to shape [B, T, N, N]. with tf.variable_scope( name, default_name="ggnn_fast_dense", values=[node_states, adjacency_matrix, num_edge_types]): nodes_shape = common_layers.shape_list(node_states) v = _compute_edge_transforms(node_states, total_value_depth, num_edge_types, name="v_mpnn") v = tf.reshape(v, [nodes_shape[0], nodes_shape[1], num_edge_types, total_value_depth ]) # Shape [B, N, T, V]. v = tf.transpose(v, [0, 2, 1, 3]) # Shape [B, T, N, V]. # Rearranging the dimensions to match the shape of all_edge_logits. edge_vectors = tf.transpose(adjacency_matrix, [0, 3, 1, 2]) output = compute_values(edge_vectors, v) return output
[ "def", "ggnn_fast_dense", "(", "node_states", ",", "adjacency_matrix", ",", "num_edge_types", ",", "total_value_depth", ",", "name", "=", "None", ")", ":", "# between the same nodes (with only one edge of each type. adjacency_matrix", "# will need to be converted to shape [B, T, N, N].", "with", "tf", ".", "variable_scope", "(", "name", ",", "default_name", "=", "\"ggnn_fast_dense\"", ",", "values", "=", "[", "node_states", ",", "adjacency_matrix", ",", "num_edge_types", "]", ")", ":", "nodes_shape", "=", "common_layers", ".", "shape_list", "(", "node_states", ")", "v", "=", "_compute_edge_transforms", "(", "node_states", ",", "total_value_depth", ",", "num_edge_types", ",", "name", "=", "\"v_mpnn\"", ")", "v", "=", "tf", ".", "reshape", "(", "v", ",", "[", "nodes_shape", "[", "0", "]", ",", "nodes_shape", "[", "1", "]", ",", "num_edge_types", ",", "total_value_depth", "]", ")", "# Shape [B, N, T, V].", "v", "=", "tf", ".", "transpose", "(", "v", ",", "[", "0", ",", "2", ",", "1", ",", "3", "]", ")", "# Shape [B, T, N, V].", "# Rearranging the dimensions to match the shape of all_edge_logits.", "edge_vectors", "=", "tf", ".", "transpose", "(", "adjacency_matrix", ",", "[", "0", ",", "3", ",", "1", ",", "2", "]", ")", "output", "=", "compute_values", "(", "edge_vectors", ",", "v", ")", "return", "output" ]
ggnn version of the MPNN from Gilmer et al. Let B be the number of batches. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries. Let V be the size of the output of the ggnn. Let T be the number of transforms / edge types. Args: node_states: The value Tensor of shape [B, T, N, D]. adjacency_matrix: A Tensor of shape [B, N, N, T]. An entry at indices b, i, j, k is the indicator of the edge from node j to node i in batch b. A standard adjacency matrix will only have values of one, while a mutigraph may have larger integer values. num_edge_types: An integer specifying number of edge types. total_value_depth: An integer (V) name: A string. Returns: A Tensor of shape [B, N, V] storing the result of computing attention weights using the queries and keys and combining the values according to those weights. Raises: ValueError: if num_transforms doesn't equal num_edge_types and not using weighted sum.
[ "ggnn", "version", "of", "the", "MPNN", "from", "Gilmer", "et", "al", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L786-L837
22,351
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
compute_values
def compute_values(edge_compatibility, v): """Compute values. If edge compatibilities is just adjacency, we get ggnn. Args: edge_compatibility: A tensor of shape [batch, num_transforms, length, depth] v: A tensor of shape [batch, num_transforms, length, depth] Returns: output: A [batch, length, depth] tensor """ # Computes the incoming value vectors for each node by weighting them # according to the attention weights. These values are still segregated by # edge type. # Shape = [B, T, N, V]. all_edge_values = tf.matmul(tf.to_float(edge_compatibility), v) # Combines the weighted value vectors together across edge types into a # single N x V matrix for each batch. output = tf.reduce_sum(all_edge_values, axis=1) # Shape [B, N, V]. return output
python
def compute_values(edge_compatibility, v): """Compute values. If edge compatibilities is just adjacency, we get ggnn. Args: edge_compatibility: A tensor of shape [batch, num_transforms, length, depth] v: A tensor of shape [batch, num_transforms, length, depth] Returns: output: A [batch, length, depth] tensor """ # Computes the incoming value vectors for each node by weighting them # according to the attention weights. These values are still segregated by # edge type. # Shape = [B, T, N, V]. all_edge_values = tf.matmul(tf.to_float(edge_compatibility), v) # Combines the weighted value vectors together across edge types into a # single N x V matrix for each batch. output = tf.reduce_sum(all_edge_values, axis=1) # Shape [B, N, V]. return output
[ "def", "compute_values", "(", "edge_compatibility", ",", "v", ")", ":", "# Computes the incoming value vectors for each node by weighting them", "# according to the attention weights. These values are still segregated by", "# edge type.", "# Shape = [B, T, N, V].", "all_edge_values", "=", "tf", ".", "matmul", "(", "tf", ".", "to_float", "(", "edge_compatibility", ")", ",", "v", ")", "# Combines the weighted value vectors together across edge types into a", "# single N x V matrix for each batch.", "output", "=", "tf", ".", "reduce_sum", "(", "all_edge_values", ",", "axis", "=", "1", ")", "# Shape [B, N, V].", "return", "output" ]
Compute values. If edge compatibilities is just adjacency, we get ggnn. Args: edge_compatibility: A tensor of shape [batch, num_transforms, length, depth] v: A tensor of shape [batch, num_transforms, length, depth] Returns: output: A [batch, length, depth] tensor
[ "Compute", "values", ".", "If", "edge", "compatibilities", "is", "just", "adjacency", "we", "get", "ggnn", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L840-L860
22,352
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
precompute_edge_matrices
def precompute_edge_matrices(adjacency, hparams): """Precompute the a_in and a_out tensors. (we don't want to add to the graph everytime _fprop is called) Args: adjacency: placeholder of real valued vectors of shape [B, L, L, E] hparams: HParams object Returns: edge_matrices: [batch, L * D, L * D] the dense matrix for message passing viewed as a block matrix (L,L) blocks of size (D,D). Each plot is a function of the edge vector of the adjacency matrix at that spot. """ batch_size, num_nodes, _, edge_dim = common_layers.shape_list(adjacency) # build the edge_network for incoming edges with tf.variable_scope("edge_network"): x = tf.reshape( adjacency, [batch_size * num_nodes * num_nodes, edge_dim], name="adj_reshape_in") for ip_layer in range(hparams.edge_network_layers): name = "edge_network_layer_%d"%ip_layer x = tf.layers.dense(common_layers.layer_preprocess(x, hparams), hparams.edge_network_hidden_size, activation=tf.nn.relu, name=name) x = tf.layers.dense(common_layers.layer_preprocess(x, hparams), hparams.hidden_size**2, activation=None, name="edge_network_output") # x = [batch * l * l, d *d] edge_matrices_flat = tf.reshape(x, [batch_size, num_nodes, num_nodes, hparams.hidden_size, hparams.hidden_size]) # reshape to [batch, l * d, l *d] edge_matrices = tf.reshape( tf.transpose(edge_matrices_flat, [0, 1, 3, 2, 4]), [ -1, num_nodes * hparams.hidden_size, num_nodes * hparams.hidden_size ], name="edge_matrices") return edge_matrices
python
def precompute_edge_matrices(adjacency, hparams): """Precompute the a_in and a_out tensors. (we don't want to add to the graph everytime _fprop is called) Args: adjacency: placeholder of real valued vectors of shape [B, L, L, E] hparams: HParams object Returns: edge_matrices: [batch, L * D, L * D] the dense matrix for message passing viewed as a block matrix (L,L) blocks of size (D,D). Each plot is a function of the edge vector of the adjacency matrix at that spot. """ batch_size, num_nodes, _, edge_dim = common_layers.shape_list(adjacency) # build the edge_network for incoming edges with tf.variable_scope("edge_network"): x = tf.reshape( adjacency, [batch_size * num_nodes * num_nodes, edge_dim], name="adj_reshape_in") for ip_layer in range(hparams.edge_network_layers): name = "edge_network_layer_%d"%ip_layer x = tf.layers.dense(common_layers.layer_preprocess(x, hparams), hparams.edge_network_hidden_size, activation=tf.nn.relu, name=name) x = tf.layers.dense(common_layers.layer_preprocess(x, hparams), hparams.hidden_size**2, activation=None, name="edge_network_output") # x = [batch * l * l, d *d] edge_matrices_flat = tf.reshape(x, [batch_size, num_nodes, num_nodes, hparams.hidden_size, hparams.hidden_size]) # reshape to [batch, l * d, l *d] edge_matrices = tf.reshape( tf.transpose(edge_matrices_flat, [0, 1, 3, 2, 4]), [ -1, num_nodes * hparams.hidden_size, num_nodes * hparams.hidden_size ], name="edge_matrices") return edge_matrices
[ "def", "precompute_edge_matrices", "(", "adjacency", ",", "hparams", ")", ":", "batch_size", ",", "num_nodes", ",", "_", ",", "edge_dim", "=", "common_layers", ".", "shape_list", "(", "adjacency", ")", "# build the edge_network for incoming edges", "with", "tf", ".", "variable_scope", "(", "\"edge_network\"", ")", ":", "x", "=", "tf", ".", "reshape", "(", "adjacency", ",", "[", "batch_size", "*", "num_nodes", "*", "num_nodes", ",", "edge_dim", "]", ",", "name", "=", "\"adj_reshape_in\"", ")", "for", "ip_layer", "in", "range", "(", "hparams", ".", "edge_network_layers", ")", ":", "name", "=", "\"edge_network_layer_%d\"", "%", "ip_layer", "x", "=", "tf", ".", "layers", ".", "dense", "(", "common_layers", ".", "layer_preprocess", "(", "x", ",", "hparams", ")", ",", "hparams", ".", "edge_network_hidden_size", ",", "activation", "=", "tf", ".", "nn", ".", "relu", ",", "name", "=", "name", ")", "x", "=", "tf", ".", "layers", ".", "dense", "(", "common_layers", ".", "layer_preprocess", "(", "x", ",", "hparams", ")", ",", "hparams", ".", "hidden_size", "**", "2", ",", "activation", "=", "None", ",", "name", "=", "\"edge_network_output\"", ")", "# x = [batch * l * l, d *d]", "edge_matrices_flat", "=", "tf", ".", "reshape", "(", "x", ",", "[", "batch_size", ",", "num_nodes", ",", "num_nodes", ",", "hparams", ".", "hidden_size", ",", "hparams", ".", "hidden_size", "]", ")", "# reshape to [batch, l * d, l *d]", "edge_matrices", "=", "tf", ".", "reshape", "(", "tf", ".", "transpose", "(", "edge_matrices_flat", ",", "[", "0", ",", "1", ",", "3", ",", "2", ",", "4", "]", ")", ",", "[", "-", "1", ",", "num_nodes", "*", "hparams", ".", "hidden_size", ",", "num_nodes", "*", "hparams", ".", "hidden_size", "]", ",", "name", "=", "\"edge_matrices\"", ")", "return", "edge_matrices" ]
Precompute the a_in and a_out tensors. (we don't want to add to the graph everytime _fprop is called) Args: adjacency: placeholder of real valued vectors of shape [B, L, L, E] hparams: HParams object Returns: edge_matrices: [batch, L * D, L * D] the dense matrix for message passing viewed as a block matrix (L,L) blocks of size (D,D). Each plot is a function of the edge vector of the adjacency matrix at that spot.
[ "Precompute", "the", "a_in", "and", "a_out", "tensors", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L863-L907
22,353
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
generate_files_distributed
def generate_files_distributed(generator, output_name, output_dir, num_shards=1, max_cases=None, task_id=0): """generate_files but with a single writer writing to shard task_id.""" assert task_id < num_shards output_filename = sharded_name(output_name, task_id, num_shards) output_file = os.path.join(output_dir, output_filename) tf.logging.info("Writing to file %s", output_file) writer = tf.python_io.TFRecordWriter(output_file) counter = 0 for case in generator: if counter % 100000 == 0: tf.logging.info("Generating case %d for %s." % (counter, output_name)) counter += 1 if max_cases and counter > max_cases: break example = to_example(case) writer.write(example.SerializeToString()) writer.close() return output_file
python
def generate_files_distributed(generator, output_name, output_dir, num_shards=1, max_cases=None, task_id=0): """generate_files but with a single writer writing to shard task_id.""" assert task_id < num_shards output_filename = sharded_name(output_name, task_id, num_shards) output_file = os.path.join(output_dir, output_filename) tf.logging.info("Writing to file %s", output_file) writer = tf.python_io.TFRecordWriter(output_file) counter = 0 for case in generator: if counter % 100000 == 0: tf.logging.info("Generating case %d for %s." % (counter, output_name)) counter += 1 if max_cases and counter > max_cases: break example = to_example(case) writer.write(example.SerializeToString()) writer.close() return output_file
[ "def", "generate_files_distributed", "(", "generator", ",", "output_name", ",", "output_dir", ",", "num_shards", "=", "1", ",", "max_cases", "=", "None", ",", "task_id", "=", "0", ")", ":", "assert", "task_id", "<", "num_shards", "output_filename", "=", "sharded_name", "(", "output_name", ",", "task_id", ",", "num_shards", ")", "output_file", "=", "os", ".", "path", ".", "join", "(", "output_dir", ",", "output_filename", ")", "tf", ".", "logging", ".", "info", "(", "\"Writing to file %s\"", ",", "output_file", ")", "writer", "=", "tf", ".", "python_io", ".", "TFRecordWriter", "(", "output_file", ")", "counter", "=", "0", "for", "case", "in", "generator", ":", "if", "counter", "%", "100000", "==", "0", ":", "tf", ".", "logging", ".", "info", "(", "\"Generating case %d for %s.\"", "%", "(", "counter", ",", "output_name", ")", ")", "counter", "+=", "1", "if", "max_cases", "and", "counter", ">", "max_cases", ":", "break", "example", "=", "to_example", "(", "case", ")", "writer", ".", "write", "(", "example", ".", "SerializeToString", "(", ")", ")", "writer", ".", "close", "(", ")", "return", "output_file" ]
generate_files but with a single writer writing to shard task_id.
[ "generate_files", "but", "with", "a", "single", "writer", "writing", "to", "shard", "task_id", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L65-L89
22,354
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
generate_files
def generate_files(generator, output_filenames, max_cases=None, cycle_every_n=1): """Generate cases from a generator and save as TFRecord files. Generated cases are transformed to tf.Example protos and saved as TFRecords in sharded files named output_dir/output_name-00..N-of-00..M=num_shards. Args: generator: a generator yielding (string -> int/float/str list) dictionaries. output_filenames: List of output file paths. max_cases: maximum number of cases to get from the generator; if None (default), we use the generator until StopIteration is raised. cycle_every_n: how many cases from the generator to take before switching to the next shard; by default set to 1, switch every case. """ if outputs_exist(output_filenames): tf.logging.info("Skipping generator because outputs files exists at {}" .format(output_filenames)) return tmp_filenames = [fname + ".incomplete" for fname in output_filenames] num_shards = len(output_filenames) # Check if is training or eval, ref: train_data_filenames(). if num_shards > 0: if "-train" in output_filenames[0]: tag = "train" elif "-dev" in output_filenames[0]: tag = "eval" else: tag = "other" writers = [tf.python_io.TFRecordWriter(fname) for fname in tmp_filenames] counter, shard = 0, 0 for case in generator: if case is None: continue if counter % 100000 == 0: tf.logging.info("Generating case %d." % counter) counter += 1 if max_cases and counter > max_cases: break example = to_example(case) writers[shard].write(example.SerializeToString()) if counter % cycle_every_n == 0: shard = (shard + 1) % num_shards for writer in writers: writer.close() for tmp_name, final_name in zip(tmp_filenames, output_filenames): tf.gfile.Rename(tmp_name, final_name) if num_shards > 0: if tag == "train": mlperf_log.transformer_print( key=mlperf_log.PREPROC_NUM_TRAIN_EXAMPLES, value=counter) elif tag == "eval": mlperf_log.transformer_print( key=mlperf_log.PREPROC_NUM_EVAL_EXAMPLES, value=counter) tf.logging.info("Generated %s Examples", counter)
python
def generate_files(generator, output_filenames, max_cases=None, cycle_every_n=1): """Generate cases from a generator and save as TFRecord files. Generated cases are transformed to tf.Example protos and saved as TFRecords in sharded files named output_dir/output_name-00..N-of-00..M=num_shards. Args: generator: a generator yielding (string -> int/float/str list) dictionaries. output_filenames: List of output file paths. max_cases: maximum number of cases to get from the generator; if None (default), we use the generator until StopIteration is raised. cycle_every_n: how many cases from the generator to take before switching to the next shard; by default set to 1, switch every case. """ if outputs_exist(output_filenames): tf.logging.info("Skipping generator because outputs files exists at {}" .format(output_filenames)) return tmp_filenames = [fname + ".incomplete" for fname in output_filenames] num_shards = len(output_filenames) # Check if is training or eval, ref: train_data_filenames(). if num_shards > 0: if "-train" in output_filenames[0]: tag = "train" elif "-dev" in output_filenames[0]: tag = "eval" else: tag = "other" writers = [tf.python_io.TFRecordWriter(fname) for fname in tmp_filenames] counter, shard = 0, 0 for case in generator: if case is None: continue if counter % 100000 == 0: tf.logging.info("Generating case %d." % counter) counter += 1 if max_cases and counter > max_cases: break example = to_example(case) writers[shard].write(example.SerializeToString()) if counter % cycle_every_n == 0: shard = (shard + 1) % num_shards for writer in writers: writer.close() for tmp_name, final_name in zip(tmp_filenames, output_filenames): tf.gfile.Rename(tmp_name, final_name) if num_shards > 0: if tag == "train": mlperf_log.transformer_print( key=mlperf_log.PREPROC_NUM_TRAIN_EXAMPLES, value=counter) elif tag == "eval": mlperf_log.transformer_print( key=mlperf_log.PREPROC_NUM_EVAL_EXAMPLES, value=counter) tf.logging.info("Generated %s Examples", counter)
[ "def", "generate_files", "(", "generator", ",", "output_filenames", ",", "max_cases", "=", "None", ",", "cycle_every_n", "=", "1", ")", ":", "if", "outputs_exist", "(", "output_filenames", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Skipping generator because outputs files exists at {}\"", ".", "format", "(", "output_filenames", ")", ")", "return", "tmp_filenames", "=", "[", "fname", "+", "\".incomplete\"", "for", "fname", "in", "output_filenames", "]", "num_shards", "=", "len", "(", "output_filenames", ")", "# Check if is training or eval, ref: train_data_filenames().", "if", "num_shards", ">", "0", ":", "if", "\"-train\"", "in", "output_filenames", "[", "0", "]", ":", "tag", "=", "\"train\"", "elif", "\"-dev\"", "in", "output_filenames", "[", "0", "]", ":", "tag", "=", "\"eval\"", "else", ":", "tag", "=", "\"other\"", "writers", "=", "[", "tf", ".", "python_io", ".", "TFRecordWriter", "(", "fname", ")", "for", "fname", "in", "tmp_filenames", "]", "counter", ",", "shard", "=", "0", ",", "0", "for", "case", "in", "generator", ":", "if", "case", "is", "None", ":", "continue", "if", "counter", "%", "100000", "==", "0", ":", "tf", ".", "logging", ".", "info", "(", "\"Generating case %d.\"", "%", "counter", ")", "counter", "+=", "1", "if", "max_cases", "and", "counter", ">", "max_cases", ":", "break", "example", "=", "to_example", "(", "case", ")", "writers", "[", "shard", "]", ".", "write", "(", "example", ".", "SerializeToString", "(", ")", ")", "if", "counter", "%", "cycle_every_n", "==", "0", ":", "shard", "=", "(", "shard", "+", "1", ")", "%", "num_shards", "for", "writer", "in", "writers", ":", "writer", ".", "close", "(", ")", "for", "tmp_name", ",", "final_name", "in", "zip", "(", "tmp_filenames", ",", "output_filenames", ")", ":", "tf", ".", "gfile", ".", "Rename", "(", "tmp_name", ",", "final_name", ")", "if", "num_shards", ">", "0", ":", "if", "tag", "==", "\"train\"", ":", "mlperf_log", ".", "transformer_print", "(", "key", "=", "mlperf_log", ".", "PREPROC_NUM_TRAIN_EXAMPLES", ",", "value", "=", "counter", ")", "elif", "tag", "==", "\"eval\"", ":", "mlperf_log", ".", "transformer_print", "(", "key", "=", "mlperf_log", ".", "PREPROC_NUM_EVAL_EXAMPLES", ",", "value", "=", "counter", ")", "tf", ".", "logging", ".", "info", "(", "\"Generated %s Examples\"", ",", "counter", ")" ]
Generate cases from a generator and save as TFRecord files. Generated cases are transformed to tf.Example protos and saved as TFRecords in sharded files named output_dir/output_name-00..N-of-00..M=num_shards. Args: generator: a generator yielding (string -> int/float/str list) dictionaries. output_filenames: List of output file paths. max_cases: maximum number of cases to get from the generator; if None (default), we use the generator until StopIteration is raised. cycle_every_n: how many cases from the generator to take before switching to the next shard; by default set to 1, switch every case.
[ "Generate", "cases", "from", "a", "generator", "and", "save", "as", "TFRecord", "files", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L134-L193
22,355
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
download_report_hook
def download_report_hook(count, block_size, total_size): """Report hook for download progress. Args: count: current block number block_size: block size total_size: total size """ percent = int(count * block_size * 100 / total_size) print("\r%d%%" % percent + " completed", end="\r")
python
def download_report_hook(count, block_size, total_size): """Report hook for download progress. Args: count: current block number block_size: block size total_size: total size """ percent = int(count * block_size * 100 / total_size) print("\r%d%%" % percent + " completed", end="\r")
[ "def", "download_report_hook", "(", "count", ",", "block_size", ",", "total_size", ")", ":", "percent", "=", "int", "(", "count", "*", "block_size", "*", "100", "/", "total_size", ")", "print", "(", "\"\\r%d%%\"", "%", "percent", "+", "\" completed\"", ",", "end", "=", "\"\\r\"", ")" ]
Report hook for download progress. Args: count: current block number block_size: block size total_size: total size
[ "Report", "hook", "for", "download", "progress", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L196-L205
22,356
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
maybe_download
def maybe_download(directory, filename, uri): """Download filename from uri unless it's already in directory. Copies a remote file to local if that local file does not already exist. If the local file pre-exists this function call, it does not check that the local file is a copy of the remote. Remote filenames can be filepaths, any URI readable by tensorflow.gfile, or a URL. Args: directory: path to the directory that will be used. filename: name of the file to download to (do nothing if it already exists). uri: URI to copy (or download) from. Returns: The path to the downloaded file. """ tf.gfile.MakeDirs(directory) filepath = os.path.join(directory, filename) if tf.gfile.Exists(filepath): tf.logging.info("Not downloading, file already found: %s" % filepath) return filepath tf.logging.info("Downloading %s to %s" % (uri, filepath)) try: tf.gfile.Copy(uri, filepath) except tf.errors.UnimplementedError: if uri.startswith("http"): inprogress_filepath = filepath + ".incomplete" inprogress_filepath, _ = urllib.urlretrieve( uri, inprogress_filepath, reporthook=download_report_hook) # Print newline to clear the carriage return from the download progress print() tf.gfile.Rename(inprogress_filepath, filepath) else: raise ValueError("Unrecognized URI: " + filepath) statinfo = os.stat(filepath) tf.logging.info("Successfully downloaded %s, %s bytes." % (filename, statinfo.st_size)) return filepath
python
def maybe_download(directory, filename, uri): """Download filename from uri unless it's already in directory. Copies a remote file to local if that local file does not already exist. If the local file pre-exists this function call, it does not check that the local file is a copy of the remote. Remote filenames can be filepaths, any URI readable by tensorflow.gfile, or a URL. Args: directory: path to the directory that will be used. filename: name of the file to download to (do nothing if it already exists). uri: URI to copy (or download) from. Returns: The path to the downloaded file. """ tf.gfile.MakeDirs(directory) filepath = os.path.join(directory, filename) if tf.gfile.Exists(filepath): tf.logging.info("Not downloading, file already found: %s" % filepath) return filepath tf.logging.info("Downloading %s to %s" % (uri, filepath)) try: tf.gfile.Copy(uri, filepath) except tf.errors.UnimplementedError: if uri.startswith("http"): inprogress_filepath = filepath + ".incomplete" inprogress_filepath, _ = urllib.urlretrieve( uri, inprogress_filepath, reporthook=download_report_hook) # Print newline to clear the carriage return from the download progress print() tf.gfile.Rename(inprogress_filepath, filepath) else: raise ValueError("Unrecognized URI: " + filepath) statinfo = os.stat(filepath) tf.logging.info("Successfully downloaded %s, %s bytes." % (filename, statinfo.st_size)) return filepath
[ "def", "maybe_download", "(", "directory", ",", "filename", ",", "uri", ")", ":", "tf", ".", "gfile", ".", "MakeDirs", "(", "directory", ")", "filepath", "=", "os", ".", "path", ".", "join", "(", "directory", ",", "filename", ")", "if", "tf", ".", "gfile", ".", "Exists", "(", "filepath", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Not downloading, file already found: %s\"", "%", "filepath", ")", "return", "filepath", "tf", ".", "logging", ".", "info", "(", "\"Downloading %s to %s\"", "%", "(", "uri", ",", "filepath", ")", ")", "try", ":", "tf", ".", "gfile", ".", "Copy", "(", "uri", ",", "filepath", ")", "except", "tf", ".", "errors", ".", "UnimplementedError", ":", "if", "uri", ".", "startswith", "(", "\"http\"", ")", ":", "inprogress_filepath", "=", "filepath", "+", "\".incomplete\"", "inprogress_filepath", ",", "_", "=", "urllib", ".", "urlretrieve", "(", "uri", ",", "inprogress_filepath", ",", "reporthook", "=", "download_report_hook", ")", "# Print newline to clear the carriage return from the download progress", "print", "(", ")", "tf", ".", "gfile", ".", "Rename", "(", "inprogress_filepath", ",", "filepath", ")", "else", ":", "raise", "ValueError", "(", "\"Unrecognized URI: \"", "+", "filepath", ")", "statinfo", "=", "os", ".", "stat", "(", "filepath", ")", "tf", ".", "logging", ".", "info", "(", "\"Successfully downloaded %s, %s bytes.\"", "%", "(", "filename", ",", "statinfo", ".", "st_size", ")", ")", "return", "filepath" ]
Download filename from uri unless it's already in directory. Copies a remote file to local if that local file does not already exist. If the local file pre-exists this function call, it does not check that the local file is a copy of the remote. Remote filenames can be filepaths, any URI readable by tensorflow.gfile, or a URL. Args: directory: path to the directory that will be used. filename: name of the file to download to (do nothing if it already exists). uri: URI to copy (or download) from. Returns: The path to the downloaded file.
[ "Download", "filename", "from", "uri", "unless", "it", "s", "already", "in", "directory", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L208-L248
22,357
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
maybe_download_from_drive
def maybe_download_from_drive(directory, filename, url): """Download filename from Google drive unless it's already in directory. Args: directory: path to the directory that will be used. filename: name of the file to download to (do nothing if it already exists). url: URL to download from. Returns: The path to the downloaded file. """ if not tf.gfile.Exists(directory): tf.logging.info("Creating directory %s" % directory) tf.gfile.MakeDirs(directory) filepath = os.path.join(directory, filename) confirm_token = None if tf.gfile.Exists(filepath): tf.logging.info("Not downloading, file already found: %s" % filepath) return filepath # Since the file is big, drive will scan it for virus and take it to a # warning page. We find the confirm token on this page and append it to the # URL to start the download process. confirm_token = None session = requests.Session() response = session.get(url, stream=True) for k, v in response.cookies.items(): if k.startswith("download_warning"): confirm_token = v if confirm_token: url = url + "&confirm=" + confirm_token tf.logging.info("Downloading %s to %s" % (url, filepath)) response = session.get(url, stream=True) # Now begin the download. chunk_size = 16 * 1024 with open(filepath, "wb") as f: for chunk in response.iter_content(chunk_size): if chunk: f.write(chunk) # Print newline to clear the carriage return from the download progress print() statinfo = os.stat(filepath) tf.logging.info("Successfully downloaded %s, %s bytes." % (filename, statinfo.st_size)) return filepath
python
def maybe_download_from_drive(directory, filename, url): """Download filename from Google drive unless it's already in directory. Args: directory: path to the directory that will be used. filename: name of the file to download to (do nothing if it already exists). url: URL to download from. Returns: The path to the downloaded file. """ if not tf.gfile.Exists(directory): tf.logging.info("Creating directory %s" % directory) tf.gfile.MakeDirs(directory) filepath = os.path.join(directory, filename) confirm_token = None if tf.gfile.Exists(filepath): tf.logging.info("Not downloading, file already found: %s" % filepath) return filepath # Since the file is big, drive will scan it for virus and take it to a # warning page. We find the confirm token on this page and append it to the # URL to start the download process. confirm_token = None session = requests.Session() response = session.get(url, stream=True) for k, v in response.cookies.items(): if k.startswith("download_warning"): confirm_token = v if confirm_token: url = url + "&confirm=" + confirm_token tf.logging.info("Downloading %s to %s" % (url, filepath)) response = session.get(url, stream=True) # Now begin the download. chunk_size = 16 * 1024 with open(filepath, "wb") as f: for chunk in response.iter_content(chunk_size): if chunk: f.write(chunk) # Print newline to clear the carriage return from the download progress print() statinfo = os.stat(filepath) tf.logging.info("Successfully downloaded %s, %s bytes." % (filename, statinfo.st_size)) return filepath
[ "def", "maybe_download_from_drive", "(", "directory", ",", "filename", ",", "url", ")", ":", "if", "not", "tf", ".", "gfile", ".", "Exists", "(", "directory", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Creating directory %s\"", "%", "directory", ")", "tf", ".", "gfile", ".", "MakeDirs", "(", "directory", ")", "filepath", "=", "os", ".", "path", ".", "join", "(", "directory", ",", "filename", ")", "confirm_token", "=", "None", "if", "tf", ".", "gfile", ".", "Exists", "(", "filepath", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Not downloading, file already found: %s\"", "%", "filepath", ")", "return", "filepath", "# Since the file is big, drive will scan it for virus and take it to a", "# warning page. We find the confirm token on this page and append it to the", "# URL to start the download process.", "confirm_token", "=", "None", "session", "=", "requests", ".", "Session", "(", ")", "response", "=", "session", ".", "get", "(", "url", ",", "stream", "=", "True", ")", "for", "k", ",", "v", "in", "response", ".", "cookies", ".", "items", "(", ")", ":", "if", "k", ".", "startswith", "(", "\"download_warning\"", ")", ":", "confirm_token", "=", "v", "if", "confirm_token", ":", "url", "=", "url", "+", "\"&confirm=\"", "+", "confirm_token", "tf", ".", "logging", ".", "info", "(", "\"Downloading %s to %s\"", "%", "(", "url", ",", "filepath", ")", ")", "response", "=", "session", ".", "get", "(", "url", ",", "stream", "=", "True", ")", "# Now begin the download.", "chunk_size", "=", "16", "*", "1024", "with", "open", "(", "filepath", ",", "\"wb\"", ")", "as", "f", ":", "for", "chunk", "in", "response", ".", "iter_content", "(", "chunk_size", ")", ":", "if", "chunk", ":", "f", ".", "write", "(", "chunk", ")", "# Print newline to clear the carriage return from the download progress", "print", "(", ")", "statinfo", "=", "os", ".", "stat", "(", "filepath", ")", "tf", ".", "logging", ".", "info", "(", "\"Successfully downloaded %s, %s bytes.\"", "%", "(", "filename", ",", "statinfo", ".", "st_size", ")", ")", "return", "filepath" ]
Download filename from Google drive unless it's already in directory. Args: directory: path to the directory that will be used. filename: name of the file to download to (do nothing if it already exists). url: URL to download from. Returns: The path to the downloaded file.
[ "Download", "filename", "from", "Google", "drive", "unless", "it", "s", "already", "in", "directory", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L251-L298
22,358
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
gunzip_file
def gunzip_file(gz_path, new_path): """Unzips from gz_path into new_path. Args: gz_path: path to the zipped file. new_path: path to where the file will be unzipped. """ if tf.gfile.Exists(new_path): tf.logging.info("File %s already exists, skipping unpacking" % new_path) return tf.logging.info("Unpacking %s to %s" % (gz_path, new_path)) # We may be unpacking into a newly created directory, add write mode. mode = stat.S_IRWXU or stat.S_IXGRP or stat.S_IRGRP or stat.S_IROTH os.chmod(os.path.dirname(new_path), mode) with gzip.open(gz_path, "rb") as gz_file: with tf.gfile.GFile(new_path, mode="wb") as new_file: for line in gz_file: new_file.write(line)
python
def gunzip_file(gz_path, new_path): """Unzips from gz_path into new_path. Args: gz_path: path to the zipped file. new_path: path to where the file will be unzipped. """ if tf.gfile.Exists(new_path): tf.logging.info("File %s already exists, skipping unpacking" % new_path) return tf.logging.info("Unpacking %s to %s" % (gz_path, new_path)) # We may be unpacking into a newly created directory, add write mode. mode = stat.S_IRWXU or stat.S_IXGRP or stat.S_IRGRP or stat.S_IROTH os.chmod(os.path.dirname(new_path), mode) with gzip.open(gz_path, "rb") as gz_file: with tf.gfile.GFile(new_path, mode="wb") as new_file: for line in gz_file: new_file.write(line)
[ "def", "gunzip_file", "(", "gz_path", ",", "new_path", ")", ":", "if", "tf", ".", "gfile", ".", "Exists", "(", "new_path", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"File %s already exists, skipping unpacking\"", "%", "new_path", ")", "return", "tf", ".", "logging", ".", "info", "(", "\"Unpacking %s to %s\"", "%", "(", "gz_path", ",", "new_path", ")", ")", "# We may be unpacking into a newly created directory, add write mode.", "mode", "=", "stat", ".", "S_IRWXU", "or", "stat", ".", "S_IXGRP", "or", "stat", ".", "S_IRGRP", "or", "stat", ".", "S_IROTH", "os", ".", "chmod", "(", "os", ".", "path", ".", "dirname", "(", "new_path", ")", ",", "mode", ")", "with", "gzip", ".", "open", "(", "gz_path", ",", "\"rb\"", ")", "as", "gz_file", ":", "with", "tf", ".", "gfile", ".", "GFile", "(", "new_path", ",", "mode", "=", "\"wb\"", ")", "as", "new_file", ":", "for", "line", "in", "gz_file", ":", "new_file", ".", "write", "(", "line", ")" ]
Unzips from gz_path into new_path. Args: gz_path: path to the zipped file. new_path: path to where the file will be unzipped.
[ "Unzips", "from", "gz_path", "into", "new_path", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L301-L318
22,359
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
get_or_generate_vocab_inner
def get_or_generate_vocab_inner(data_dir, vocab_filename, vocab_size, generator, max_subtoken_length=None, reserved_tokens=None): """Inner implementation for vocab generators. Args: data_dir: The base directory where data and vocab files are stored. If None, then do not save the vocab even if it doesn't exist. vocab_filename: relative filename where vocab file is stored vocab_size: target size of the vocabulary constructed by SubwordTextEncoder generator: a generator that produces tokens from the vocabulary max_subtoken_length: an optional integer. Set this to a finite value to avoid quadratic costs during vocab building. reserved_tokens: List of reserved tokens. `text_encoder.RESERVED_TOKENS` should be a prefix of `reserved_tokens`. If `None`, defaults to `RESERVED_TOKENS`. Returns: A SubwordTextEncoder vocabulary object. """ if data_dir and vocab_filename: vocab_filepath = os.path.join(data_dir, vocab_filename) if tf.gfile.Exists(vocab_filepath): tf.logging.info("Found vocab file: %s", vocab_filepath) return text_encoder.SubwordTextEncoder(vocab_filepath) else: vocab_filepath = None tf.logging.info("Generating vocab file: %s", vocab_filepath) vocab = text_encoder.SubwordTextEncoder.build_from_generator( generator, vocab_size, max_subtoken_length=max_subtoken_length, reserved_tokens=reserved_tokens) if vocab_filepath: tf.gfile.MakeDirs(data_dir) vocab.store_to_file(vocab_filepath) return vocab
python
def get_or_generate_vocab_inner(data_dir, vocab_filename, vocab_size, generator, max_subtoken_length=None, reserved_tokens=None): """Inner implementation for vocab generators. Args: data_dir: The base directory where data and vocab files are stored. If None, then do not save the vocab even if it doesn't exist. vocab_filename: relative filename where vocab file is stored vocab_size: target size of the vocabulary constructed by SubwordTextEncoder generator: a generator that produces tokens from the vocabulary max_subtoken_length: an optional integer. Set this to a finite value to avoid quadratic costs during vocab building. reserved_tokens: List of reserved tokens. `text_encoder.RESERVED_TOKENS` should be a prefix of `reserved_tokens`. If `None`, defaults to `RESERVED_TOKENS`. Returns: A SubwordTextEncoder vocabulary object. """ if data_dir and vocab_filename: vocab_filepath = os.path.join(data_dir, vocab_filename) if tf.gfile.Exists(vocab_filepath): tf.logging.info("Found vocab file: %s", vocab_filepath) return text_encoder.SubwordTextEncoder(vocab_filepath) else: vocab_filepath = None tf.logging.info("Generating vocab file: %s", vocab_filepath) vocab = text_encoder.SubwordTextEncoder.build_from_generator( generator, vocab_size, max_subtoken_length=max_subtoken_length, reserved_tokens=reserved_tokens) if vocab_filepath: tf.gfile.MakeDirs(data_dir) vocab.store_to_file(vocab_filepath) return vocab
[ "def", "get_or_generate_vocab_inner", "(", "data_dir", ",", "vocab_filename", ",", "vocab_size", ",", "generator", ",", "max_subtoken_length", "=", "None", ",", "reserved_tokens", "=", "None", ")", ":", "if", "data_dir", "and", "vocab_filename", ":", "vocab_filepath", "=", "os", ".", "path", ".", "join", "(", "data_dir", ",", "vocab_filename", ")", "if", "tf", ".", "gfile", ".", "Exists", "(", "vocab_filepath", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Found vocab file: %s\"", ",", "vocab_filepath", ")", "return", "text_encoder", ".", "SubwordTextEncoder", "(", "vocab_filepath", ")", "else", ":", "vocab_filepath", "=", "None", "tf", ".", "logging", ".", "info", "(", "\"Generating vocab file: %s\"", ",", "vocab_filepath", ")", "vocab", "=", "text_encoder", ".", "SubwordTextEncoder", ".", "build_from_generator", "(", "generator", ",", "vocab_size", ",", "max_subtoken_length", "=", "max_subtoken_length", ",", "reserved_tokens", "=", "reserved_tokens", ")", "if", "vocab_filepath", ":", "tf", ".", "gfile", ".", "MakeDirs", "(", "data_dir", ")", "vocab", ".", "store_to_file", "(", "vocab_filepath", ")", "return", "vocab" ]
Inner implementation for vocab generators. Args: data_dir: The base directory where data and vocab files are stored. If None, then do not save the vocab even if it doesn't exist. vocab_filename: relative filename where vocab file is stored vocab_size: target size of the vocabulary constructed by SubwordTextEncoder generator: a generator that produces tokens from the vocabulary max_subtoken_length: an optional integer. Set this to a finite value to avoid quadratic costs during vocab building. reserved_tokens: List of reserved tokens. `text_encoder.RESERVED_TOKENS` should be a prefix of `reserved_tokens`. If `None`, defaults to `RESERVED_TOKENS`. Returns: A SubwordTextEncoder vocabulary object.
[ "Inner", "implementation", "for", "vocab", "generators", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L321-L358
22,360
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
get_or_generate_vocab
def get_or_generate_vocab(data_dir, tmp_dir, vocab_filename, vocab_size, sources, file_byte_budget=1e6, max_subtoken_length=None): """Generate a vocabulary from the datasets in sources.""" vocab_generator = generate_lines_for_vocab(tmp_dir, sources, file_byte_budget) return get_or_generate_vocab_inner(data_dir, vocab_filename, vocab_size, vocab_generator, max_subtoken_length)
python
def get_or_generate_vocab(data_dir, tmp_dir, vocab_filename, vocab_size, sources, file_byte_budget=1e6, max_subtoken_length=None): """Generate a vocabulary from the datasets in sources.""" vocab_generator = generate_lines_for_vocab(tmp_dir, sources, file_byte_budget) return get_or_generate_vocab_inner(data_dir, vocab_filename, vocab_size, vocab_generator, max_subtoken_length)
[ "def", "get_or_generate_vocab", "(", "data_dir", ",", "tmp_dir", ",", "vocab_filename", ",", "vocab_size", ",", "sources", ",", "file_byte_budget", "=", "1e6", ",", "max_subtoken_length", "=", "None", ")", ":", "vocab_generator", "=", "generate_lines_for_vocab", "(", "tmp_dir", ",", "sources", ",", "file_byte_budget", ")", "return", "get_or_generate_vocab_inner", "(", "data_dir", ",", "vocab_filename", ",", "vocab_size", ",", "vocab_generator", ",", "max_subtoken_length", ")" ]
Generate a vocabulary from the datasets in sources.
[ "Generate", "a", "vocabulary", "from", "the", "datasets", "in", "sources", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L361-L368
22,361
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
generate_lines_for_vocab
def generate_lines_for_vocab(tmp_dir, sources, file_byte_budget=1e6): """Generate lines for vocabulary generation.""" tf.logging.info("Generating vocab from: %s", str(sources)) for source in sources: url = source[0] filename = os.path.basename(url) compressed_file = maybe_download(tmp_dir, filename, url) for lang_file in source[1]: tf.logging.info("Reading file: %s" % lang_file) filepath = os.path.join(tmp_dir, lang_file) # Extract from tar if needed. if not tf.gfile.Exists(filepath): read_type = "r:gz" if filename.endswith("tgz") else "r" with tarfile.open(compressed_file, read_type) as corpus_tar: corpus_tar.extractall(tmp_dir) # For some datasets a second extraction is necessary. if lang_file.endswith(".gz"): new_filepath = os.path.join(tmp_dir, lang_file[:-3]) if tf.gfile.Exists(new_filepath): tf.logging.info( "Subdirectory %s already exists, skipping unpacking" % filepath) else: tf.logging.info("Unpacking subdirectory %s" % filepath) gunzip_file(filepath, new_filepath) filepath = new_filepath with tf.gfile.GFile(filepath, mode="r") as source_file: file_byte_budget_ = file_byte_budget counter = 0 countermax = int(source_file.size() / file_byte_budget_ / 2) for line in source_file: if counter < countermax: counter += 1 else: if file_byte_budget_ <= 0: break line = line.strip() file_byte_budget_ -= len(line) counter = 0 yield line
python
def generate_lines_for_vocab(tmp_dir, sources, file_byte_budget=1e6): """Generate lines for vocabulary generation.""" tf.logging.info("Generating vocab from: %s", str(sources)) for source in sources: url = source[0] filename = os.path.basename(url) compressed_file = maybe_download(tmp_dir, filename, url) for lang_file in source[1]: tf.logging.info("Reading file: %s" % lang_file) filepath = os.path.join(tmp_dir, lang_file) # Extract from tar if needed. if not tf.gfile.Exists(filepath): read_type = "r:gz" if filename.endswith("tgz") else "r" with tarfile.open(compressed_file, read_type) as corpus_tar: corpus_tar.extractall(tmp_dir) # For some datasets a second extraction is necessary. if lang_file.endswith(".gz"): new_filepath = os.path.join(tmp_dir, lang_file[:-3]) if tf.gfile.Exists(new_filepath): tf.logging.info( "Subdirectory %s already exists, skipping unpacking" % filepath) else: tf.logging.info("Unpacking subdirectory %s" % filepath) gunzip_file(filepath, new_filepath) filepath = new_filepath with tf.gfile.GFile(filepath, mode="r") as source_file: file_byte_budget_ = file_byte_budget counter = 0 countermax = int(source_file.size() / file_byte_budget_ / 2) for line in source_file: if counter < countermax: counter += 1 else: if file_byte_budget_ <= 0: break line = line.strip() file_byte_budget_ -= len(line) counter = 0 yield line
[ "def", "generate_lines_for_vocab", "(", "tmp_dir", ",", "sources", ",", "file_byte_budget", "=", "1e6", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Generating vocab from: %s\"", ",", "str", "(", "sources", ")", ")", "for", "source", "in", "sources", ":", "url", "=", "source", "[", "0", "]", "filename", "=", "os", ".", "path", ".", "basename", "(", "url", ")", "compressed_file", "=", "maybe_download", "(", "tmp_dir", ",", "filename", ",", "url", ")", "for", "lang_file", "in", "source", "[", "1", "]", ":", "tf", ".", "logging", ".", "info", "(", "\"Reading file: %s\"", "%", "lang_file", ")", "filepath", "=", "os", ".", "path", ".", "join", "(", "tmp_dir", ",", "lang_file", ")", "# Extract from tar if needed.", "if", "not", "tf", ".", "gfile", ".", "Exists", "(", "filepath", ")", ":", "read_type", "=", "\"r:gz\"", "if", "filename", ".", "endswith", "(", "\"tgz\"", ")", "else", "\"r\"", "with", "tarfile", ".", "open", "(", "compressed_file", ",", "read_type", ")", "as", "corpus_tar", ":", "corpus_tar", ".", "extractall", "(", "tmp_dir", ")", "# For some datasets a second extraction is necessary.", "if", "lang_file", ".", "endswith", "(", "\".gz\"", ")", ":", "new_filepath", "=", "os", ".", "path", ".", "join", "(", "tmp_dir", ",", "lang_file", "[", ":", "-", "3", "]", ")", "if", "tf", ".", "gfile", ".", "Exists", "(", "new_filepath", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Subdirectory %s already exists, skipping unpacking\"", "%", "filepath", ")", "else", ":", "tf", ".", "logging", ".", "info", "(", "\"Unpacking subdirectory %s\"", "%", "filepath", ")", "gunzip_file", "(", "filepath", ",", "new_filepath", ")", "filepath", "=", "new_filepath", "with", "tf", ".", "gfile", ".", "GFile", "(", "filepath", ",", "mode", "=", "\"r\"", ")", "as", "source_file", ":", "file_byte_budget_", "=", "file_byte_budget", "counter", "=", "0", "countermax", "=", "int", "(", "source_file", ".", "size", "(", ")", "/", "file_byte_budget_", "/", "2", ")", "for", "line", "in", "source_file", ":", "if", "counter", "<", "countermax", ":", "counter", "+=", "1", "else", ":", "if", "file_byte_budget_", "<=", "0", ":", "break", "line", "=", "line", ".", "strip", "(", ")", "file_byte_budget_", "-=", "len", "(", "line", ")", "counter", "=", "0", "yield", "line" ]
Generate lines for vocabulary generation.
[ "Generate", "lines", "for", "vocabulary", "generation", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L371-L413
22,362
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
get_or_generate_tabbed_vocab
def get_or_generate_tabbed_vocab(data_dir, tmp_dir, source_filename, index, vocab_filename, vocab_size): r"""Generate a vocabulary from a tabbed source file. The source is a file of source, target pairs, where each line contains a source string and a target string, separated by a tab ('\t') character. The index parameter specifies 0 for the source or 1 for the target. Args: data_dir: path to the data directory. tmp_dir: path to the temporary directory. source_filename: the name of the tab-separated source file. index: index. vocab_filename: the name of the vocabulary file. vocab_size: vocabulary size. Returns: The vocabulary. """ def generate(): filepath = os.path.join(tmp_dir, source_filename) tf.logging.info("Generating vocab from %s", filepath) with tf.gfile.GFile(filepath, mode="r") as source_file: for line in source_file: line = line.strip() if line and "\t" in line: parts = line.split("\t", 1) part = parts[index].strip() yield part return get_or_generate_vocab_inner(data_dir, vocab_filename, vocab_size, generate())
python
def get_or_generate_tabbed_vocab(data_dir, tmp_dir, source_filename, index, vocab_filename, vocab_size): r"""Generate a vocabulary from a tabbed source file. The source is a file of source, target pairs, where each line contains a source string and a target string, separated by a tab ('\t') character. The index parameter specifies 0 for the source or 1 for the target. Args: data_dir: path to the data directory. tmp_dir: path to the temporary directory. source_filename: the name of the tab-separated source file. index: index. vocab_filename: the name of the vocabulary file. vocab_size: vocabulary size. Returns: The vocabulary. """ def generate(): filepath = os.path.join(tmp_dir, source_filename) tf.logging.info("Generating vocab from %s", filepath) with tf.gfile.GFile(filepath, mode="r") as source_file: for line in source_file: line = line.strip() if line and "\t" in line: parts = line.split("\t", 1) part = parts[index].strip() yield part return get_or_generate_vocab_inner(data_dir, vocab_filename, vocab_size, generate())
[ "def", "get_or_generate_tabbed_vocab", "(", "data_dir", ",", "tmp_dir", ",", "source_filename", ",", "index", ",", "vocab_filename", ",", "vocab_size", ")", ":", "def", "generate", "(", ")", ":", "filepath", "=", "os", ".", "path", ".", "join", "(", "tmp_dir", ",", "source_filename", ")", "tf", ".", "logging", ".", "info", "(", "\"Generating vocab from %s\"", ",", "filepath", ")", "with", "tf", ".", "gfile", ".", "GFile", "(", "filepath", ",", "mode", "=", "\"r\"", ")", "as", "source_file", ":", "for", "line", "in", "source_file", ":", "line", "=", "line", ".", "strip", "(", ")", "if", "line", "and", "\"\\t\"", "in", "line", ":", "parts", "=", "line", ".", "split", "(", "\"\\t\"", ",", "1", ")", "part", "=", "parts", "[", "index", "]", ".", "strip", "(", ")", "yield", "part", "return", "get_or_generate_vocab_inner", "(", "data_dir", ",", "vocab_filename", ",", "vocab_size", ",", "generate", "(", ")", ")" ]
r"""Generate a vocabulary from a tabbed source file. The source is a file of source, target pairs, where each line contains a source string and a target string, separated by a tab ('\t') character. The index parameter specifies 0 for the source or 1 for the target. Args: data_dir: path to the data directory. tmp_dir: path to the temporary directory. source_filename: the name of the tab-separated source file. index: index. vocab_filename: the name of the vocabulary file. vocab_size: vocabulary size. Returns: The vocabulary.
[ "r", "Generate", "a", "vocabulary", "from", "a", "tabbed", "source", "file", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L416-L447
22,363
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
get_or_generate_txt_vocab
def get_or_generate_txt_vocab(data_dir, vocab_filename, vocab_size, filepatterns): """Generate a vocabulary from txt files with example-per-line.""" if isinstance(filepatterns, str): filepatterns = [filepatterns] def generate(): tf.logging.info("Generating vocab from %s", filepatterns) for filepattern in filepatterns: for filename in tf.gfile.Glob(filepattern): with tf.gfile.GFile(filename, mode="r") as source_file: for line in source_file: yield line.strip() return get_or_generate_vocab_inner(data_dir, vocab_filename, vocab_size, generate())
python
def get_or_generate_txt_vocab(data_dir, vocab_filename, vocab_size, filepatterns): """Generate a vocabulary from txt files with example-per-line.""" if isinstance(filepatterns, str): filepatterns = [filepatterns] def generate(): tf.logging.info("Generating vocab from %s", filepatterns) for filepattern in filepatterns: for filename in tf.gfile.Glob(filepattern): with tf.gfile.GFile(filename, mode="r") as source_file: for line in source_file: yield line.strip() return get_or_generate_vocab_inner(data_dir, vocab_filename, vocab_size, generate())
[ "def", "get_or_generate_txt_vocab", "(", "data_dir", ",", "vocab_filename", ",", "vocab_size", ",", "filepatterns", ")", ":", "if", "isinstance", "(", "filepatterns", ",", "str", ")", ":", "filepatterns", "=", "[", "filepatterns", "]", "def", "generate", "(", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Generating vocab from %s\"", ",", "filepatterns", ")", "for", "filepattern", "in", "filepatterns", ":", "for", "filename", "in", "tf", ".", "gfile", ".", "Glob", "(", "filepattern", ")", ":", "with", "tf", ".", "gfile", ".", "GFile", "(", "filename", ",", "mode", "=", "\"r\"", ")", "as", "source_file", ":", "for", "line", "in", "source_file", ":", "yield", "line", ".", "strip", "(", ")", "return", "get_or_generate_vocab_inner", "(", "data_dir", ",", "vocab_filename", ",", "vocab_size", ",", "generate", "(", ")", ")" ]
Generate a vocabulary from txt files with example-per-line.
[ "Generate", "a", "vocabulary", "from", "txt", "files", "with", "example", "-", "per", "-", "line", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L450-L465
22,364
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
_shuffle_single
def _shuffle_single(fname, extra_fn=None): """Shuffle a single file of records. Args: fname: a string extra_fn: an optional function from list of TFRecords to list of TFRecords to be called after shuffling. """ records = read_records(fname) random.shuffle(records) if extra_fn is not None: records = extra_fn(records) out_fname = fname.replace(UNSHUFFLED_SUFFIX, "") write_records(records, out_fname) tf.gfile.Remove(fname)
python
def _shuffle_single(fname, extra_fn=None): """Shuffle a single file of records. Args: fname: a string extra_fn: an optional function from list of TFRecords to list of TFRecords to be called after shuffling. """ records = read_records(fname) random.shuffle(records) if extra_fn is not None: records = extra_fn(records) out_fname = fname.replace(UNSHUFFLED_SUFFIX, "") write_records(records, out_fname) tf.gfile.Remove(fname)
[ "def", "_shuffle_single", "(", "fname", ",", "extra_fn", "=", "None", ")", ":", "records", "=", "read_records", "(", "fname", ")", "random", ".", "shuffle", "(", "records", ")", "if", "extra_fn", "is", "not", "None", ":", "records", "=", "extra_fn", "(", "records", ")", "out_fname", "=", "fname", ".", "replace", "(", "UNSHUFFLED_SUFFIX", ",", "\"\"", ")", "write_records", "(", "records", ",", "out_fname", ")", "tf", ".", "gfile", ".", "Remove", "(", "fname", ")" ]
Shuffle a single file of records. Args: fname: a string extra_fn: an optional function from list of TFRecords to list of TFRecords to be called after shuffling.
[ "Shuffle", "a", "single", "file", "of", "records", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L499-L513
22,365
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
shuffle_dataset
def shuffle_dataset(filenames, extra_fn=None): """Shuffles the dataset. Args: filenames: a list of strings extra_fn: an optional function from list of records to list of records to be called after shuffling a file. """ if outputs_exist(filenames): tf.logging.info("Skipping shuffle because output files exist") return tf.logging.info("Shuffling data...") for filename in filenames: _shuffle_single(filename, extra_fn=extra_fn) tf.logging.info("Data shuffled.")
python
def shuffle_dataset(filenames, extra_fn=None): """Shuffles the dataset. Args: filenames: a list of strings extra_fn: an optional function from list of records to list of records to be called after shuffling a file. """ if outputs_exist(filenames): tf.logging.info("Skipping shuffle because output files exist") return tf.logging.info("Shuffling data...") for filename in filenames: _shuffle_single(filename, extra_fn=extra_fn) tf.logging.info("Data shuffled.")
[ "def", "shuffle_dataset", "(", "filenames", ",", "extra_fn", "=", "None", ")", ":", "if", "outputs_exist", "(", "filenames", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Skipping shuffle because output files exist\"", ")", "return", "tf", ".", "logging", ".", "info", "(", "\"Shuffling data...\"", ")", "for", "filename", "in", "filenames", ":", "_shuffle_single", "(", "filename", ",", "extra_fn", "=", "extra_fn", ")", "tf", ".", "logging", ".", "info", "(", "\"Data shuffled.\"", ")" ]
Shuffles the dataset. Args: filenames: a list of strings extra_fn: an optional function from list of records to list of records to be called after shuffling a file.
[ "Shuffles", "the", "dataset", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L516-L530
22,366
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
pack_examples
def pack_examples(examples, has_inputs, packed_length=256, spacing=2, queue_size=10, chop_long_sequences=False): """Pack examples into longer examples. If has_inputs=False, we are packing single-sequence examples with targets only and no inputs. In this case, we concatenate the targets from several examples to form each new example. We insert a number of zeros for spacing between the original sequences. This is to help the sequences stay separate under convolutions. If chop_long_sequences is set, then any input sequence longer than packed_length gets chopped up into multiple examples. Otherwise, long sequences are emitted as singletons. If has_inputs=True, then we are packing sequence-to-sequence examples. We combine several examples by concatenating the inputs (as above) and concatenating the targets (as above). Chopping of long sequences is not supported. The packed examples are represented as dictionaries containing: "inputs", "targets": the packed sequences described above "inputs_segmentation", "targets_segmentation": Sequences aligned with "inputs", "targets" specifying to which original sequence each position belongs. Numbering starts from 1, and 0 is used for spacing. This information is useful for preventing attention across segments. e.g. [1 1 1 1 1 1 0 0 2 2 2 0 0 3 3 3 3 3 0 0 4 4 4] "inputs_position", "targets_position": Sequences aligned with "inputs", "targets" specifying position within the original sequence. This is useful for positional encodings. e.g. [0 1 2 3 4 5 0 0 0 1 2 0 0 0 1 2 3 4 0 0 0 1 2] Args: examples: a generator returning feature dictionaries. has_inputs: a boolean packed_length: an integer spacing: an integer queue_size: an integer chop_long_sequences: a boolean Yields: feature dictionaries. """ packer = SequencePairPacker if has_inputs else SequencePacker combined = [] for example in examples: x = ((example["inputs"], example["targets"]) if has_inputs else example["targets"]) if chop_long_sequences and len(x) > packed_length: assert not has_inputs num_fragments = len(x) // packed_length for i in range(num_fragments): yield packer( x[packed_length * i:packed_length * (i + 1)], spacing).to_dict() x = x[packed_length * num_fragments:] added = False for c in combined: if c.can_fit(x, packed_length): c.add(x) added = True break if not added: if len(combined) == queue_size: yield combined[0].to_dict() combined = combined[1:] combined.append(packer(x, spacing)) for c in combined: yield c.to_dict()
python
def pack_examples(examples, has_inputs, packed_length=256, spacing=2, queue_size=10, chop_long_sequences=False): """Pack examples into longer examples. If has_inputs=False, we are packing single-sequence examples with targets only and no inputs. In this case, we concatenate the targets from several examples to form each new example. We insert a number of zeros for spacing between the original sequences. This is to help the sequences stay separate under convolutions. If chop_long_sequences is set, then any input sequence longer than packed_length gets chopped up into multiple examples. Otherwise, long sequences are emitted as singletons. If has_inputs=True, then we are packing sequence-to-sequence examples. We combine several examples by concatenating the inputs (as above) and concatenating the targets (as above). Chopping of long sequences is not supported. The packed examples are represented as dictionaries containing: "inputs", "targets": the packed sequences described above "inputs_segmentation", "targets_segmentation": Sequences aligned with "inputs", "targets" specifying to which original sequence each position belongs. Numbering starts from 1, and 0 is used for spacing. This information is useful for preventing attention across segments. e.g. [1 1 1 1 1 1 0 0 2 2 2 0 0 3 3 3 3 3 0 0 4 4 4] "inputs_position", "targets_position": Sequences aligned with "inputs", "targets" specifying position within the original sequence. This is useful for positional encodings. e.g. [0 1 2 3 4 5 0 0 0 1 2 0 0 0 1 2 3 4 0 0 0 1 2] Args: examples: a generator returning feature dictionaries. has_inputs: a boolean packed_length: an integer spacing: an integer queue_size: an integer chop_long_sequences: a boolean Yields: feature dictionaries. """ packer = SequencePairPacker if has_inputs else SequencePacker combined = [] for example in examples: x = ((example["inputs"], example["targets"]) if has_inputs else example["targets"]) if chop_long_sequences and len(x) > packed_length: assert not has_inputs num_fragments = len(x) // packed_length for i in range(num_fragments): yield packer( x[packed_length * i:packed_length * (i + 1)], spacing).to_dict() x = x[packed_length * num_fragments:] added = False for c in combined: if c.can_fit(x, packed_length): c.add(x) added = True break if not added: if len(combined) == queue_size: yield combined[0].to_dict() combined = combined[1:] combined.append(packer(x, spacing)) for c in combined: yield c.to_dict()
[ "def", "pack_examples", "(", "examples", ",", "has_inputs", ",", "packed_length", "=", "256", ",", "spacing", "=", "2", ",", "queue_size", "=", "10", ",", "chop_long_sequences", "=", "False", ")", ":", "packer", "=", "SequencePairPacker", "if", "has_inputs", "else", "SequencePacker", "combined", "=", "[", "]", "for", "example", "in", "examples", ":", "x", "=", "(", "(", "example", "[", "\"inputs\"", "]", ",", "example", "[", "\"targets\"", "]", ")", "if", "has_inputs", "else", "example", "[", "\"targets\"", "]", ")", "if", "chop_long_sequences", "and", "len", "(", "x", ")", ">", "packed_length", ":", "assert", "not", "has_inputs", "num_fragments", "=", "len", "(", "x", ")", "//", "packed_length", "for", "i", "in", "range", "(", "num_fragments", ")", ":", "yield", "packer", "(", "x", "[", "packed_length", "*", "i", ":", "packed_length", "*", "(", "i", "+", "1", ")", "]", ",", "spacing", ")", ".", "to_dict", "(", ")", "x", "=", "x", "[", "packed_length", "*", "num_fragments", ":", "]", "added", "=", "False", "for", "c", "in", "combined", ":", "if", "c", ".", "can_fit", "(", "x", ",", "packed_length", ")", ":", "c", ".", "add", "(", "x", ")", "added", "=", "True", "break", "if", "not", "added", ":", "if", "len", "(", "combined", ")", "==", "queue_size", ":", "yield", "combined", "[", "0", "]", ".", "to_dict", "(", ")", "combined", "=", "combined", "[", "1", ":", "]", "combined", ".", "append", "(", "packer", "(", "x", ",", "spacing", ")", ")", "for", "c", "in", "combined", ":", "yield", "c", ".", "to_dict", "(", ")" ]
Pack examples into longer examples. If has_inputs=False, we are packing single-sequence examples with targets only and no inputs. In this case, we concatenate the targets from several examples to form each new example. We insert a number of zeros for spacing between the original sequences. This is to help the sequences stay separate under convolutions. If chop_long_sequences is set, then any input sequence longer than packed_length gets chopped up into multiple examples. Otherwise, long sequences are emitted as singletons. If has_inputs=True, then we are packing sequence-to-sequence examples. We combine several examples by concatenating the inputs (as above) and concatenating the targets (as above). Chopping of long sequences is not supported. The packed examples are represented as dictionaries containing: "inputs", "targets": the packed sequences described above "inputs_segmentation", "targets_segmentation": Sequences aligned with "inputs", "targets" specifying to which original sequence each position belongs. Numbering starts from 1, and 0 is used for spacing. This information is useful for preventing attention across segments. e.g. [1 1 1 1 1 1 0 0 2 2 2 0 0 3 3 3 3 3 0 0 4 4 4] "inputs_position", "targets_position": Sequences aligned with "inputs", "targets" specifying position within the original sequence. This is useful for positional encodings. e.g. [0 1 2 3 4 5 0 0 0 1 2 0 0 0 1 2 3 4 0 0 0 1 2] Args: examples: a generator returning feature dictionaries. has_inputs: a boolean packed_length: an integer spacing: an integer queue_size: an integer chop_long_sequences: a boolean Yields: feature dictionaries.
[ "Pack", "examples", "into", "longer", "examples", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L589-L660
22,367
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
make_tmp_dir
def make_tmp_dir(suffix="", prefix="tmp", dir=None): # pylint: disable=redefined-builtin """Make a temporary directory.""" if dir is None: return tempfile.mkdtemp(suffix, prefix, dir) else: while True: rand_term = random.randint(1, 9999) tmp_dir = os.path.join(dir, "%s%d%s" % (prefix, rand_term, suffix)) if tf.gfile.Exists(tmp_dir): continue tf.gfile.MakeDirs(tmp_dir) break return tmp_dir
python
def make_tmp_dir(suffix="", prefix="tmp", dir=None): # pylint: disable=redefined-builtin """Make a temporary directory.""" if dir is None: return tempfile.mkdtemp(suffix, prefix, dir) else: while True: rand_term = random.randint(1, 9999) tmp_dir = os.path.join(dir, "%s%d%s" % (prefix, rand_term, suffix)) if tf.gfile.Exists(tmp_dir): continue tf.gfile.MakeDirs(tmp_dir) break return tmp_dir
[ "def", "make_tmp_dir", "(", "suffix", "=", "\"\"", ",", "prefix", "=", "\"tmp\"", ",", "dir", "=", "None", ")", ":", "# pylint: disable=redefined-builtin", "if", "dir", "is", "None", ":", "return", "tempfile", ".", "mkdtemp", "(", "suffix", ",", "prefix", ",", "dir", ")", "else", ":", "while", "True", ":", "rand_term", "=", "random", ".", "randint", "(", "1", ",", "9999", ")", "tmp_dir", "=", "os", ".", "path", ".", "join", "(", "dir", ",", "\"%s%d%s\"", "%", "(", "prefix", ",", "rand_term", ",", "suffix", ")", ")", "if", "tf", ".", "gfile", ".", "Exists", "(", "tmp_dir", ")", ":", "continue", "tf", ".", "gfile", ".", "MakeDirs", "(", "tmp_dir", ")", "break", "return", "tmp_dir" ]
Make a temporary directory.
[ "Make", "a", "temporary", "directory", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L883-L895
22,368
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
tfrecord_iterator_for_problem
def tfrecord_iterator_for_problem(problem, data_dir, dataset_split=tf.estimator.ModeKeys.TRAIN): """Iterate over the records on disk for the Problem.""" filenames = tf.gfile.Glob(problem.filepattern(data_dir, mode=dataset_split)) example_spec = problem.example_reading_spec()[0] return tfrecord_iterator(filenames, example_spec=example_spec)
python
def tfrecord_iterator_for_problem(problem, data_dir, dataset_split=tf.estimator.ModeKeys.TRAIN): """Iterate over the records on disk for the Problem.""" filenames = tf.gfile.Glob(problem.filepattern(data_dir, mode=dataset_split)) example_spec = problem.example_reading_spec()[0] return tfrecord_iterator(filenames, example_spec=example_spec)
[ "def", "tfrecord_iterator_for_problem", "(", "problem", ",", "data_dir", ",", "dataset_split", "=", "tf", ".", "estimator", ".", "ModeKeys", ".", "TRAIN", ")", ":", "filenames", "=", "tf", ".", "gfile", ".", "Glob", "(", "problem", ".", "filepattern", "(", "data_dir", ",", "mode", "=", "dataset_split", ")", ")", "example_spec", "=", "problem", ".", "example_reading_spec", "(", ")", "[", "0", "]", "return", "tfrecord_iterator", "(", "filenames", ",", "example_spec", "=", "example_spec", ")" ]
Iterate over the records on disk for the Problem.
[ "Iterate", "over", "the", "records", "on", "disk", "for", "the", "Problem", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L898-L903
22,369
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
tfrecord_iterator
def tfrecord_iterator(filenames, gzipped=False, example_spec=None): """Yields records from TFRecord files. Args: filenames: list<str>, list of TFRecord filenames to read from. gzipped: bool, whether the TFRecord files are gzip-encoded. example_spec: dict<str feature name, tf.VarLenFeature/tf.FixedLenFeature>, if provided, will parse each record as a tensorflow.Example proto. Yields: Records (or parsed Examples, if example_spec is provided) from files. """ with tf.Graph().as_default(): dataset = tf.data.Dataset.from_tensor_slices(filenames) def _load_records(filename): return tf.data.TFRecordDataset( filename, compression_type=tf.constant("GZIP") if gzipped else None, buffer_size=16 * 1000 * 1000) dataset = dataset.flat_map(_load_records) def _parse_example(ex_ser): return tf.parse_single_example(ex_ser, example_spec) if example_spec: dataset = dataset.map(_parse_example, num_parallel_calls=32) dataset = dataset.prefetch(100) record_it = dataset.make_one_shot_iterator().get_next() with tf.Session() as sess: while True: try: ex = sess.run(record_it) yield ex except tf.errors.OutOfRangeError: break
python
def tfrecord_iterator(filenames, gzipped=False, example_spec=None): """Yields records from TFRecord files. Args: filenames: list<str>, list of TFRecord filenames to read from. gzipped: bool, whether the TFRecord files are gzip-encoded. example_spec: dict<str feature name, tf.VarLenFeature/tf.FixedLenFeature>, if provided, will parse each record as a tensorflow.Example proto. Yields: Records (or parsed Examples, if example_spec is provided) from files. """ with tf.Graph().as_default(): dataset = tf.data.Dataset.from_tensor_slices(filenames) def _load_records(filename): return tf.data.TFRecordDataset( filename, compression_type=tf.constant("GZIP") if gzipped else None, buffer_size=16 * 1000 * 1000) dataset = dataset.flat_map(_load_records) def _parse_example(ex_ser): return tf.parse_single_example(ex_ser, example_spec) if example_spec: dataset = dataset.map(_parse_example, num_parallel_calls=32) dataset = dataset.prefetch(100) record_it = dataset.make_one_shot_iterator().get_next() with tf.Session() as sess: while True: try: ex = sess.run(record_it) yield ex except tf.errors.OutOfRangeError: break
[ "def", "tfrecord_iterator", "(", "filenames", ",", "gzipped", "=", "False", ",", "example_spec", "=", "None", ")", ":", "with", "tf", ".", "Graph", "(", ")", ".", "as_default", "(", ")", ":", "dataset", "=", "tf", ".", "data", ".", "Dataset", ".", "from_tensor_slices", "(", "filenames", ")", "def", "_load_records", "(", "filename", ")", ":", "return", "tf", ".", "data", ".", "TFRecordDataset", "(", "filename", ",", "compression_type", "=", "tf", ".", "constant", "(", "\"GZIP\"", ")", "if", "gzipped", "else", "None", ",", "buffer_size", "=", "16", "*", "1000", "*", "1000", ")", "dataset", "=", "dataset", ".", "flat_map", "(", "_load_records", ")", "def", "_parse_example", "(", "ex_ser", ")", ":", "return", "tf", ".", "parse_single_example", "(", "ex_ser", ",", "example_spec", ")", "if", "example_spec", ":", "dataset", "=", "dataset", ".", "map", "(", "_parse_example", ",", "num_parallel_calls", "=", "32", ")", "dataset", "=", "dataset", ".", "prefetch", "(", "100", ")", "record_it", "=", "dataset", ".", "make_one_shot_iterator", "(", ")", ".", "get_next", "(", ")", "with", "tf", ".", "Session", "(", ")", "as", "sess", ":", "while", "True", ":", "try", ":", "ex", "=", "sess", ".", "run", "(", "record_it", ")", "yield", "ex", "except", "tf", ".", "errors", ".", "OutOfRangeError", ":", "break" ]
Yields records from TFRecord files. Args: filenames: list<str>, list of TFRecord filenames to read from. gzipped: bool, whether the TFRecord files are gzip-encoded. example_spec: dict<str feature name, tf.VarLenFeature/tf.FixedLenFeature>, if provided, will parse each record as a tensorflow.Example proto. Yields: Records (or parsed Examples, if example_spec is provided) from files.
[ "Yields", "records", "from", "TFRecord", "files", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L906-L943
22,370
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
random_deinterleave
def random_deinterleave(text, separator_symbol="X"): """Create a fill-in-the-blanks training example from text. Split on spaces, then cut into segments at random points. Alternate segments are assigned to the two output strings. separator_symbol separates segments within each of the outputs. example: text="The quick brown fox jumps over the lazy dog." returns: ("X quick brown X the lazy X", "The X fox jumps over X dog.") The two outputs can also be reversed to yield an instance of the same problem. Args: text: a string separator_symbol: a string Returns: a pair of strings """ words = text.strip().split(" ") n = len(words) if n <= 1: return text, "" cut = [False] * n cut[0] = True num_cuts = int(math.exp(random.uniform(0, math.log(n)))) for _ in range(num_cuts): cut[random.randint(1, n -1)] = True out = [[], []] part = random.randint(0, 1) for i in range(n): if cut[i]: out[part].append(separator_symbol) part = 1 - part out[part].append(words[i]) return " ".join(out[0]), " ".join(out[1])
python
def random_deinterleave(text, separator_symbol="X"): """Create a fill-in-the-blanks training example from text. Split on spaces, then cut into segments at random points. Alternate segments are assigned to the two output strings. separator_symbol separates segments within each of the outputs. example: text="The quick brown fox jumps over the lazy dog." returns: ("X quick brown X the lazy X", "The X fox jumps over X dog.") The two outputs can also be reversed to yield an instance of the same problem. Args: text: a string separator_symbol: a string Returns: a pair of strings """ words = text.strip().split(" ") n = len(words) if n <= 1: return text, "" cut = [False] * n cut[0] = True num_cuts = int(math.exp(random.uniform(0, math.log(n)))) for _ in range(num_cuts): cut[random.randint(1, n -1)] = True out = [[], []] part = random.randint(0, 1) for i in range(n): if cut[i]: out[part].append(separator_symbol) part = 1 - part out[part].append(words[i]) return " ".join(out[0]), " ".join(out[1])
[ "def", "random_deinterleave", "(", "text", ",", "separator_symbol", "=", "\"X\"", ")", ":", "words", "=", "text", ".", "strip", "(", ")", ".", "split", "(", "\" \"", ")", "n", "=", "len", "(", "words", ")", "if", "n", "<=", "1", ":", "return", "text", ",", "\"\"", "cut", "=", "[", "False", "]", "*", "n", "cut", "[", "0", "]", "=", "True", "num_cuts", "=", "int", "(", "math", ".", "exp", "(", "random", ".", "uniform", "(", "0", ",", "math", ".", "log", "(", "n", ")", ")", ")", ")", "for", "_", "in", "range", "(", "num_cuts", ")", ":", "cut", "[", "random", ".", "randint", "(", "1", ",", "n", "-", "1", ")", "]", "=", "True", "out", "=", "[", "[", "]", ",", "[", "]", "]", "part", "=", "random", ".", "randint", "(", "0", ",", "1", ")", "for", "i", "in", "range", "(", "n", ")", ":", "if", "cut", "[", "i", "]", ":", "out", "[", "part", "]", ".", "append", "(", "separator_symbol", ")", "part", "=", "1", "-", "part", "out", "[", "part", "]", ".", "append", "(", "words", "[", "i", "]", ")", "return", "\" \"", ".", "join", "(", "out", "[", "0", "]", ")", ",", "\" \"", ".", "join", "(", "out", "[", "1", "]", ")" ]
Create a fill-in-the-blanks training example from text. Split on spaces, then cut into segments at random points. Alternate segments are assigned to the two output strings. separator_symbol separates segments within each of the outputs. example: text="The quick brown fox jumps over the lazy dog." returns: ("X quick brown X the lazy X", "The X fox jumps over X dog.") The two outputs can also be reversed to yield an instance of the same problem. Args: text: a string separator_symbol: a string Returns: a pair of strings
[ "Create", "a", "fill", "-", "in", "-", "the", "-", "blanks", "training", "example", "from", "text", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L946-L981
22,371
tensorflow/tensor2tensor
tensor2tensor/models/neural_gpu.py
neural_gpu_body
def neural_gpu_body(inputs, hparams, name=None): """The core Neural GPU.""" with tf.variable_scope(name, "neural_gpu"): def step(state, inp): # pylint: disable=missing-docstring x = tf.nn.dropout(state, 1.0 - hparams.dropout) for layer in range(hparams.num_hidden_layers): x = common_layers.conv_gru( x, (hparams.kernel_height, hparams.kernel_width), hparams.hidden_size, name="cgru_%d" % layer) # Padding input is zeroed-out in the modality, we check this by summing. padding_inp = tf.less(tf.reduce_sum(tf.abs(inp), axis=[1, 2]), 0.00001) new_state = tf.where(padding_inp, state, x) # No-op where inp is padding. return new_state return tf.foldl( step, tf.transpose(inputs, [1, 0, 2, 3]), initializer=inputs, parallel_iterations=1, swap_memory=True)
python
def neural_gpu_body(inputs, hparams, name=None): """The core Neural GPU.""" with tf.variable_scope(name, "neural_gpu"): def step(state, inp): # pylint: disable=missing-docstring x = tf.nn.dropout(state, 1.0 - hparams.dropout) for layer in range(hparams.num_hidden_layers): x = common_layers.conv_gru( x, (hparams.kernel_height, hparams.kernel_width), hparams.hidden_size, name="cgru_%d" % layer) # Padding input is zeroed-out in the modality, we check this by summing. padding_inp = tf.less(tf.reduce_sum(tf.abs(inp), axis=[1, 2]), 0.00001) new_state = tf.where(padding_inp, state, x) # No-op where inp is padding. return new_state return tf.foldl( step, tf.transpose(inputs, [1, 0, 2, 3]), initializer=inputs, parallel_iterations=1, swap_memory=True)
[ "def", "neural_gpu_body", "(", "inputs", ",", "hparams", ",", "name", "=", "None", ")", ":", "with", "tf", ".", "variable_scope", "(", "name", ",", "\"neural_gpu\"", ")", ":", "def", "step", "(", "state", ",", "inp", ")", ":", "# pylint: disable=missing-docstring", "x", "=", "tf", ".", "nn", ".", "dropout", "(", "state", ",", "1.0", "-", "hparams", ".", "dropout", ")", "for", "layer", "in", "range", "(", "hparams", ".", "num_hidden_layers", ")", ":", "x", "=", "common_layers", ".", "conv_gru", "(", "x", ",", "(", "hparams", ".", "kernel_height", ",", "hparams", ".", "kernel_width", ")", ",", "hparams", ".", "hidden_size", ",", "name", "=", "\"cgru_%d\"", "%", "layer", ")", "# Padding input is zeroed-out in the modality, we check this by summing.", "padding_inp", "=", "tf", ".", "less", "(", "tf", ".", "reduce_sum", "(", "tf", ".", "abs", "(", "inp", ")", ",", "axis", "=", "[", "1", ",", "2", "]", ")", ",", "0.00001", ")", "new_state", "=", "tf", ".", "where", "(", "padding_inp", ",", "state", ",", "x", ")", "# No-op where inp is padding.", "return", "new_state", "return", "tf", ".", "foldl", "(", "step", ",", "tf", ".", "transpose", "(", "inputs", ",", "[", "1", ",", "0", ",", "2", ",", "3", "]", ")", ",", "initializer", "=", "inputs", ",", "parallel_iterations", "=", "1", ",", "swap_memory", "=", "True", ")" ]
The core Neural GPU.
[ "The", "core", "Neural", "GPU", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/neural_gpu.py#L31-L52
22,372
tensorflow/tensor2tensor
tensor2tensor/trax/layers/combinators.py
_reorder_shape
def _reorder_shape(input_shape, output=None): # pylint: disable=invalid-name """Helper to determine the shape of reorder output.""" if output is None: return input_shape return base.nested_map(output, lambda i: input_shape[i])
python
def _reorder_shape(input_shape, output=None): # pylint: disable=invalid-name """Helper to determine the shape of reorder output.""" if output is None: return input_shape return base.nested_map(output, lambda i: input_shape[i])
[ "def", "_reorder_shape", "(", "input_shape", ",", "output", "=", "None", ")", ":", "# pylint: disable=invalid-name", "if", "output", "is", "None", ":", "return", "input_shape", "return", "base", ".", "nested_map", "(", "output", ",", "lambda", "i", ":", "input_shape", "[", "i", "]", ")" ]
Helper to determine the shape of reorder output.
[ "Helper", "to", "determine", "the", "shape", "of", "reorder", "output", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L77-L81
22,373
tensorflow/tensor2tensor
tensor2tensor/trax/layers/combinators.py
Reorder
def Reorder(x, params, output=None, **kwargs): """Reorder a tuple into another tuple. For example, we can re-order (x, y) into (y, x) or even (y, (x, y), y). The output argument specifies how to re-order, using integers that refer to indices in the input tuple. For example, if input = (x, y, z) then Reorder(input, output=(1, 0, 2)) = (y, x, z) Reorder(input, output=(0, 0)) = (x, x) Reorder(input, output=(0, (1, 1))) = (x, (y, y)) Reorder(input, output=((2, 0), (1, 1))) = ((z, x), (y, y)) By default (if no output is given) Reorder does nothing (Identity). Args: x: the input tuple to re-order. params: layer parameters (unused). output: the specification of the output tuple: a nested tuple of ints. **kwargs: other arguments (unused). Returns: The re-ordered tuple with the same shape as output. """ del params, kwargs if output is None: return x return base.nested_map(output, lambda i: x[i])
python
def Reorder(x, params, output=None, **kwargs): """Reorder a tuple into another tuple. For example, we can re-order (x, y) into (y, x) or even (y, (x, y), y). The output argument specifies how to re-order, using integers that refer to indices in the input tuple. For example, if input = (x, y, z) then Reorder(input, output=(1, 0, 2)) = (y, x, z) Reorder(input, output=(0, 0)) = (x, x) Reorder(input, output=(0, (1, 1))) = (x, (y, y)) Reorder(input, output=((2, 0), (1, 1))) = ((z, x), (y, y)) By default (if no output is given) Reorder does nothing (Identity). Args: x: the input tuple to re-order. params: layer parameters (unused). output: the specification of the output tuple: a nested tuple of ints. **kwargs: other arguments (unused). Returns: The re-ordered tuple with the same shape as output. """ del params, kwargs if output is None: return x return base.nested_map(output, lambda i: x[i])
[ "def", "Reorder", "(", "x", ",", "params", ",", "output", "=", "None", ",", "*", "*", "kwargs", ")", ":", "del", "params", ",", "kwargs", "if", "output", "is", "None", ":", "return", "x", "return", "base", ".", "nested_map", "(", "output", ",", "lambda", "i", ":", "x", "[", "i", "]", ")" ]
Reorder a tuple into another tuple. For example, we can re-order (x, y) into (y, x) or even (y, (x, y), y). The output argument specifies how to re-order, using integers that refer to indices in the input tuple. For example, if input = (x, y, z) then Reorder(input, output=(1, 0, 2)) = (y, x, z) Reorder(input, output=(0, 0)) = (x, x) Reorder(input, output=(0, (1, 1))) = (x, (y, y)) Reorder(input, output=((2, 0), (1, 1))) = ((z, x), (y, y)) By default (if no output is given) Reorder does nothing (Identity). Args: x: the input tuple to re-order. params: layer parameters (unused). output: the specification of the output tuple: a nested tuple of ints. **kwargs: other arguments (unused). Returns: The re-ordered tuple with the same shape as output.
[ "Reorder", "a", "tuple", "into", "another", "tuple", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L85-L115
22,374
tensorflow/tensor2tensor
tensor2tensor/trax/layers/combinators.py
_concatenate_shape
def _concatenate_shape(input_shape, axis=-1): # pylint: disable=invalid-name """Helper to determine the shape of Concatenate output.""" ax = axis % len(input_shape[0]) concat_size = sum(shape[ax] for shape in input_shape) out_shape = input_shape[0][:ax] + (concat_size,) + input_shape[0][ax+1:] return out_shape
python
def _concatenate_shape(input_shape, axis=-1): # pylint: disable=invalid-name """Helper to determine the shape of Concatenate output.""" ax = axis % len(input_shape[0]) concat_size = sum(shape[ax] for shape in input_shape) out_shape = input_shape[0][:ax] + (concat_size,) + input_shape[0][ax+1:] return out_shape
[ "def", "_concatenate_shape", "(", "input_shape", ",", "axis", "=", "-", "1", ")", ":", "# pylint: disable=invalid-name", "ax", "=", "axis", "%", "len", "(", "input_shape", "[", "0", "]", ")", "concat_size", "=", "sum", "(", "shape", "[", "ax", "]", "for", "shape", "in", "input_shape", ")", "out_shape", "=", "input_shape", "[", "0", "]", "[", ":", "ax", "]", "+", "(", "concat_size", ",", ")", "+", "input_shape", "[", "0", "]", "[", "ax", "+", "1", ":", "]", "return", "out_shape" ]
Helper to determine the shape of Concatenate output.
[ "Helper", "to", "determine", "the", "shape", "of", "Concatenate", "output", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L189-L194
22,375
tensorflow/tensor2tensor
tensor2tensor/trax/layers/combinators.py
Residual
def Residual(*layers, **kwargs): """Constructs a residual version of layers, summing input to layers output.""" shortcut = kwargs.get('shortcut', Identity()) # pylint: disable=no-value-for-parameter if len(layers) > 1: return Serial( Branch(), # pylint: disable=no-value-for-parameter Parallel(Serial(*layers), shortcut), SumBranches() # pylint: disable=no-value-for-parameter ) elif len(layers) == 1: return Serial( Branch(), # pylint: disable=no-value-for-parameter Parallel(layers[0], shortcut), SumBranches() # pylint: disable=no-value-for-parameter ) else: raise ValueError('Empty residual combinator.')
python
def Residual(*layers, **kwargs): """Constructs a residual version of layers, summing input to layers output.""" shortcut = kwargs.get('shortcut', Identity()) # pylint: disable=no-value-for-parameter if len(layers) > 1: return Serial( Branch(), # pylint: disable=no-value-for-parameter Parallel(Serial(*layers), shortcut), SumBranches() # pylint: disable=no-value-for-parameter ) elif len(layers) == 1: return Serial( Branch(), # pylint: disable=no-value-for-parameter Parallel(layers[0], shortcut), SumBranches() # pylint: disable=no-value-for-parameter ) else: raise ValueError('Empty residual combinator.')
[ "def", "Residual", "(", "*", "layers", ",", "*", "*", "kwargs", ")", ":", "shortcut", "=", "kwargs", ".", "get", "(", "'shortcut'", ",", "Identity", "(", ")", ")", "# pylint: disable=no-value-for-parameter", "if", "len", "(", "layers", ")", ">", "1", ":", "return", "Serial", "(", "Branch", "(", ")", ",", "# pylint: disable=no-value-for-parameter", "Parallel", "(", "Serial", "(", "*", "layers", ")", ",", "shortcut", ")", ",", "SumBranches", "(", ")", "# pylint: disable=no-value-for-parameter", ")", "elif", "len", "(", "layers", ")", "==", "1", ":", "return", "Serial", "(", "Branch", "(", ")", ",", "# pylint: disable=no-value-for-parameter", "Parallel", "(", "layers", "[", "0", "]", ",", "shortcut", ")", ",", "SumBranches", "(", ")", "# pylint: disable=no-value-for-parameter", ")", "else", ":", "raise", "ValueError", "(", "'Empty residual combinator.'", ")" ]
Constructs a residual version of layers, summing input to layers output.
[ "Constructs", "a", "residual", "version", "of", "layers", "summing", "input", "to", "layers", "output", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L240-L256
22,376
tensorflow/tensor2tensor
tensor2tensor/models/research/universal_transformer.py
update_hparams_for_universal_transformer
def update_hparams_for_universal_transformer(hparams): """Adds default hparams for all of the variants of the Universal Transformer. Args: hparams: default hparams (usually one of the standard hparams from transformer model (like "transformer_base") Returns: hparams with default values for Universal Transformers hyper-parameters """ hparams.daisy_chain_variables = False # Breaks multi-gpu in while loops. # If not None, mixes vanilla transformer with Universal Transformer. # Options: None, "before_ut", and "after_ut". hparams.add_hparam("mix_with_transformer", None) # Number of vanilla transformer layers used to be mixed with u-transofmer. hparams.add_hparam("num_mixedin_layers", 2) # Number of transformer layers within the recurrent block (default is 1). hparams.add_hparam("num_inrecurrence_layers", 1) # Type of recurrency: # basic, highway, skip, dwa, act, rnn, gru, lstm. hparams.add_hparam("recurrence_type", "basic") # Number of steps (which is equivalent to num layer in transformer). hparams.add_hparam("num_rec_steps", hparams.num_hidden_layers) # Add the positional mebedding at each step(horisontal timing) hparams.add_hparam("add_position_timing_signal", True) if hparams.add_position_timing_signal: hparams.pos = None # Logic of position shifting when using timing signal: # None, "random", "step" hparams.add_hparam("position_start_index", None) # Add an step embedding at each step (vertical timing) hparams.add_hparam("add_step_timing_signal", True) # Either "learned" or "sinusoid" hparams.add_hparam("step_timing_signal_type", "learned") # Add or concat the timing signal (applied both on position and step timing). # Options: "add" and "concat". hparams.add_hparam("add_or_concat_timing_signal", "add") # Add SRU at the beginning of each Universal Transformer step. # This can be considered as a position timing signal hparams.add_hparam("add_sru", False) # Default ffn layer is separable convolution. # Options: "fc" and "sepconv". hparams.add_hparam("transformer_ffn_type", "fc") # Transform bias (in models with highway or skip connection). hparams.add_hparam("transform_bias_init", -1.0) hparams.add_hparam("couple_carry_transform_gates", True) # Depth-wise attention (grid-transformer!) hparams: # Adds depth embedding, if true. hparams.add_hparam("depth_embedding", True) # Learns attention weights for elements (instead of positions), if true. hparams.add_hparam("dwa_elements", True) # Type of ffn_layer used for gate in skip, highway, etc. # "dense" or "dense_dropconnect". # With dense_relu_dense, the bias/kernel initializations will not be applied. hparams.add_hparam("gate_ffn_layer", "dense") # LSTM forget bias for lstm style recurrence. hparams.add_hparam("lstm_forget_bias", 1.0) # Uses the memory at the last step as the final output, if true. hparams.add_hparam("use_memory_as_final_state", False) # if also add a ffn unit to the transition function when using gru/lstm hparams.add_hparam("add_ffn_unit_to_the_transition_function", False) # Type of act: basic/accumulated/global (instead of position-wise!)/random. hparams.add_hparam("act_type", "basic") # Max number of steps (forces halting at this step). hparams.add_hparam("act_max_steps", 2 * hparams.num_hidden_layers) hparams.add_hparam("act_halting_bias_init", 1.0) hparams.add_hparam("act_epsilon", 0.01) hparams.add_hparam("act_loss_weight", 0.01) return hparams
python
def update_hparams_for_universal_transformer(hparams): """Adds default hparams for all of the variants of the Universal Transformer. Args: hparams: default hparams (usually one of the standard hparams from transformer model (like "transformer_base") Returns: hparams with default values for Universal Transformers hyper-parameters """ hparams.daisy_chain_variables = False # Breaks multi-gpu in while loops. # If not None, mixes vanilla transformer with Universal Transformer. # Options: None, "before_ut", and "after_ut". hparams.add_hparam("mix_with_transformer", None) # Number of vanilla transformer layers used to be mixed with u-transofmer. hparams.add_hparam("num_mixedin_layers", 2) # Number of transformer layers within the recurrent block (default is 1). hparams.add_hparam("num_inrecurrence_layers", 1) # Type of recurrency: # basic, highway, skip, dwa, act, rnn, gru, lstm. hparams.add_hparam("recurrence_type", "basic") # Number of steps (which is equivalent to num layer in transformer). hparams.add_hparam("num_rec_steps", hparams.num_hidden_layers) # Add the positional mebedding at each step(horisontal timing) hparams.add_hparam("add_position_timing_signal", True) if hparams.add_position_timing_signal: hparams.pos = None # Logic of position shifting when using timing signal: # None, "random", "step" hparams.add_hparam("position_start_index", None) # Add an step embedding at each step (vertical timing) hparams.add_hparam("add_step_timing_signal", True) # Either "learned" or "sinusoid" hparams.add_hparam("step_timing_signal_type", "learned") # Add or concat the timing signal (applied both on position and step timing). # Options: "add" and "concat". hparams.add_hparam("add_or_concat_timing_signal", "add") # Add SRU at the beginning of each Universal Transformer step. # This can be considered as a position timing signal hparams.add_hparam("add_sru", False) # Default ffn layer is separable convolution. # Options: "fc" and "sepconv". hparams.add_hparam("transformer_ffn_type", "fc") # Transform bias (in models with highway or skip connection). hparams.add_hparam("transform_bias_init", -1.0) hparams.add_hparam("couple_carry_transform_gates", True) # Depth-wise attention (grid-transformer!) hparams: # Adds depth embedding, if true. hparams.add_hparam("depth_embedding", True) # Learns attention weights for elements (instead of positions), if true. hparams.add_hparam("dwa_elements", True) # Type of ffn_layer used for gate in skip, highway, etc. # "dense" or "dense_dropconnect". # With dense_relu_dense, the bias/kernel initializations will not be applied. hparams.add_hparam("gate_ffn_layer", "dense") # LSTM forget bias for lstm style recurrence. hparams.add_hparam("lstm_forget_bias", 1.0) # Uses the memory at the last step as the final output, if true. hparams.add_hparam("use_memory_as_final_state", False) # if also add a ffn unit to the transition function when using gru/lstm hparams.add_hparam("add_ffn_unit_to_the_transition_function", False) # Type of act: basic/accumulated/global (instead of position-wise!)/random. hparams.add_hparam("act_type", "basic") # Max number of steps (forces halting at this step). hparams.add_hparam("act_max_steps", 2 * hparams.num_hidden_layers) hparams.add_hparam("act_halting_bias_init", 1.0) hparams.add_hparam("act_epsilon", 0.01) hparams.add_hparam("act_loss_weight", 0.01) return hparams
[ "def", "update_hparams_for_universal_transformer", "(", "hparams", ")", ":", "hparams", ".", "daisy_chain_variables", "=", "False", "# Breaks multi-gpu in while loops.", "# If not None, mixes vanilla transformer with Universal Transformer.", "# Options: None, \"before_ut\", and \"after_ut\".", "hparams", ".", "add_hparam", "(", "\"mix_with_transformer\"", ",", "None", ")", "# Number of vanilla transformer layers used to be mixed with u-transofmer.", "hparams", ".", "add_hparam", "(", "\"num_mixedin_layers\"", ",", "2", ")", "# Number of transformer layers within the recurrent block (default is 1).", "hparams", ".", "add_hparam", "(", "\"num_inrecurrence_layers\"", ",", "1", ")", "# Type of recurrency:", "# basic, highway, skip, dwa, act, rnn, gru, lstm.", "hparams", ".", "add_hparam", "(", "\"recurrence_type\"", ",", "\"basic\"", ")", "# Number of steps (which is equivalent to num layer in transformer).", "hparams", ".", "add_hparam", "(", "\"num_rec_steps\"", ",", "hparams", ".", "num_hidden_layers", ")", "# Add the positional mebedding at each step(horisontal timing)", "hparams", ".", "add_hparam", "(", "\"add_position_timing_signal\"", ",", "True", ")", "if", "hparams", ".", "add_position_timing_signal", ":", "hparams", ".", "pos", "=", "None", "# Logic of position shifting when using timing signal:", "# None, \"random\", \"step\"", "hparams", ".", "add_hparam", "(", "\"position_start_index\"", ",", "None", ")", "# Add an step embedding at each step (vertical timing)", "hparams", ".", "add_hparam", "(", "\"add_step_timing_signal\"", ",", "True", ")", "# Either \"learned\" or \"sinusoid\"", "hparams", ".", "add_hparam", "(", "\"step_timing_signal_type\"", ",", "\"learned\"", ")", "# Add or concat the timing signal (applied both on position and step timing).", "# Options: \"add\" and \"concat\".", "hparams", ".", "add_hparam", "(", "\"add_or_concat_timing_signal\"", ",", "\"add\"", ")", "# Add SRU at the beginning of each Universal Transformer step.", "# This can be considered as a position timing signal", "hparams", ".", "add_hparam", "(", "\"add_sru\"", ",", "False", ")", "# Default ffn layer is separable convolution.", "# Options: \"fc\" and \"sepconv\".", "hparams", ".", "add_hparam", "(", "\"transformer_ffn_type\"", ",", "\"fc\"", ")", "# Transform bias (in models with highway or skip connection).", "hparams", ".", "add_hparam", "(", "\"transform_bias_init\"", ",", "-", "1.0", ")", "hparams", ".", "add_hparam", "(", "\"couple_carry_transform_gates\"", ",", "True", ")", "# Depth-wise attention (grid-transformer!) hparams:", "# Adds depth embedding, if true.", "hparams", ".", "add_hparam", "(", "\"depth_embedding\"", ",", "True", ")", "# Learns attention weights for elements (instead of positions), if true.", "hparams", ".", "add_hparam", "(", "\"dwa_elements\"", ",", "True", ")", "# Type of ffn_layer used for gate in skip, highway, etc.", "# \"dense\" or \"dense_dropconnect\".", "# With dense_relu_dense, the bias/kernel initializations will not be applied.", "hparams", ".", "add_hparam", "(", "\"gate_ffn_layer\"", ",", "\"dense\"", ")", "# LSTM forget bias for lstm style recurrence.", "hparams", ".", "add_hparam", "(", "\"lstm_forget_bias\"", ",", "1.0", ")", "# Uses the memory at the last step as the final output, if true.", "hparams", ".", "add_hparam", "(", "\"use_memory_as_final_state\"", ",", "False", ")", "# if also add a ffn unit to the transition function when using gru/lstm", "hparams", ".", "add_hparam", "(", "\"add_ffn_unit_to_the_transition_function\"", ",", "False", ")", "# Type of act: basic/accumulated/global (instead of position-wise!)/random.", "hparams", ".", "add_hparam", "(", "\"act_type\"", ",", "\"basic\"", ")", "# Max number of steps (forces halting at this step).", "hparams", ".", "add_hparam", "(", "\"act_max_steps\"", ",", "2", "*", "hparams", ".", "num_hidden_layers", ")", "hparams", ".", "add_hparam", "(", "\"act_halting_bias_init\"", ",", "1.0", ")", "hparams", ".", "add_hparam", "(", "\"act_epsilon\"", ",", "0.01", ")", "hparams", ".", "add_hparam", "(", "\"act_loss_weight\"", ",", "0.01", ")", "return", "hparams" ]
Adds default hparams for all of the variants of the Universal Transformer. Args: hparams: default hparams (usually one of the standard hparams from transformer model (like "transformer_base") Returns: hparams with default values for Universal Transformers hyper-parameters
[ "Adds", "default", "hparams", "for", "all", "of", "the", "variants", "of", "the", "Universal", "Transformer", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L352-L436
22,377
tensorflow/tensor2tensor
tensor2tensor/models/research/universal_transformer.py
universal_transformer_base
def universal_transformer_base(): """Base parameters for Universal Transformer.""" hparams = transformer.transformer_base() # To have a similar capacity to the transformer_base with 6 layers, # we need to increase the size of the UT's layer # since, in fact, UT has a single layer repeating multiple times. hparams.hidden_size = 1024 hparams.filter_size = 4096 hparams.num_heads = 16 hparams.layer_prepostprocess_dropout = 0.3 hparams = update_hparams_for_universal_transformer(hparams) return hparams
python
def universal_transformer_base(): """Base parameters for Universal Transformer.""" hparams = transformer.transformer_base() # To have a similar capacity to the transformer_base with 6 layers, # we need to increase the size of the UT's layer # since, in fact, UT has a single layer repeating multiple times. hparams.hidden_size = 1024 hparams.filter_size = 4096 hparams.num_heads = 16 hparams.layer_prepostprocess_dropout = 0.3 hparams = update_hparams_for_universal_transformer(hparams) return hparams
[ "def", "universal_transformer_base", "(", ")", ":", "hparams", "=", "transformer", ".", "transformer_base", "(", ")", "# To have a similar capacity to the transformer_base with 6 layers,", "# we need to increase the size of the UT's layer", "# since, in fact, UT has a single layer repeating multiple times.", "hparams", ".", "hidden_size", "=", "1024", "hparams", ".", "filter_size", "=", "4096", "hparams", ".", "num_heads", "=", "16", "hparams", ".", "layer_prepostprocess_dropout", "=", "0.3", "hparams", "=", "update_hparams_for_universal_transformer", "(", "hparams", ")", "return", "hparams" ]
Base parameters for Universal Transformer.
[ "Base", "parameters", "for", "Universal", "Transformer", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L440-L451
22,378
tensorflow/tensor2tensor
tensor2tensor/models/research/universal_transformer.py
adaptive_universal_transformer_multilayer_tpu
def adaptive_universal_transformer_multilayer_tpu(): """Multi-layer config for adaptive Transformer on TPU.""" hparams = adaptive_universal_transformer_base_tpu() hparams.num_inrecurrence_layers = 2 hparams.mix_with_transformer = "before_ut,after_ut" hparams.num_mixedin_layers = 1 hparams.transformer_ffn_type = "sepconv" # TODO(lukaszkaiser): the options below don't work on TPU yet, make them work. # hparams.add_step_timing_signal = True # hparams.add_sru = True # hparams.self_attention_type = "dot_product_relative_v2" # hparams.max_relative_position = 256 return hparams
python
def adaptive_universal_transformer_multilayer_tpu(): """Multi-layer config for adaptive Transformer on TPU.""" hparams = adaptive_universal_transformer_base_tpu() hparams.num_inrecurrence_layers = 2 hparams.mix_with_transformer = "before_ut,after_ut" hparams.num_mixedin_layers = 1 hparams.transformer_ffn_type = "sepconv" # TODO(lukaszkaiser): the options below don't work on TPU yet, make them work. # hparams.add_step_timing_signal = True # hparams.add_sru = True # hparams.self_attention_type = "dot_product_relative_v2" # hparams.max_relative_position = 256 return hparams
[ "def", "adaptive_universal_transformer_multilayer_tpu", "(", ")", ":", "hparams", "=", "adaptive_universal_transformer_base_tpu", "(", ")", "hparams", ".", "num_inrecurrence_layers", "=", "2", "hparams", ".", "mix_with_transformer", "=", "\"before_ut,after_ut\"", "hparams", ".", "num_mixedin_layers", "=", "1", "hparams", ".", "transformer_ffn_type", "=", "\"sepconv\"", "# TODO(lukaszkaiser): the options below don't work on TPU yet, make them work.", "# hparams.add_step_timing_signal = True", "# hparams.add_sru = True", "# hparams.self_attention_type = \"dot_product_relative_v2\"", "# hparams.max_relative_position = 256", "return", "hparams" ]
Multi-layer config for adaptive Transformer on TPU.
[ "Multi", "-", "layer", "config", "for", "adaptive", "Transformer", "on", "TPU", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L543-L555
22,379
tensorflow/tensor2tensor
tensor2tensor/models/research/universal_transformer.py
adaptive_universal_transformer_multilayer_hard
def adaptive_universal_transformer_multilayer_hard(): """Multi-layer config for adaptive Transformer with hard attention.""" hparams = adaptive_universal_transformer_multilayer_tpu() hparams.batch_size = 256 hparams.hard_attention_k = 8 hparams.add_step_timing_signal = True # hparams.add_sru = True # This is very slow on GPUs, does it help? hparams.self_attention_type = "dot_product_relative_v2" hparams.max_relative_position = 256 return hparams
python
def adaptive_universal_transformer_multilayer_hard(): """Multi-layer config for adaptive Transformer with hard attention.""" hparams = adaptive_universal_transformer_multilayer_tpu() hparams.batch_size = 256 hparams.hard_attention_k = 8 hparams.add_step_timing_signal = True # hparams.add_sru = True # This is very slow on GPUs, does it help? hparams.self_attention_type = "dot_product_relative_v2" hparams.max_relative_position = 256 return hparams
[ "def", "adaptive_universal_transformer_multilayer_hard", "(", ")", ":", "hparams", "=", "adaptive_universal_transformer_multilayer_tpu", "(", ")", "hparams", ".", "batch_size", "=", "256", "hparams", ".", "hard_attention_k", "=", "8", "hparams", ".", "add_step_timing_signal", "=", "True", "# hparams.add_sru = True # This is very slow on GPUs, does it help?", "hparams", ".", "self_attention_type", "=", "\"dot_product_relative_v2\"", "hparams", ".", "max_relative_position", "=", "256", "return", "hparams" ]
Multi-layer config for adaptive Transformer with hard attention.
[ "Multi", "-", "layer", "config", "for", "adaptive", "Transformer", "with", "hard", "attention", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L559-L568
22,380
tensorflow/tensor2tensor
tensor2tensor/trax/models/neural_gpu.py
ConvDiagonalGRU
def ConvDiagonalGRU(units, kernel_size=(3, 3)): """Build convolutional GRU with diagonal gating as in ImprovedNGPU.""" def BuildConv(): return layers.Conv(filters=units, kernel_size=kernel_size, padding='SAME') return layers.GeneralGRUCell( candidate_transform=BuildConv, memory_transform=DiagonalGate, gate_nonlinearity=layers.HardSigmoid, candidate_nonlinearity=layers.HardTanh)
python
def ConvDiagonalGRU(units, kernel_size=(3, 3)): """Build convolutional GRU with diagonal gating as in ImprovedNGPU.""" def BuildConv(): return layers.Conv(filters=units, kernel_size=kernel_size, padding='SAME') return layers.GeneralGRUCell( candidate_transform=BuildConv, memory_transform=DiagonalGate, gate_nonlinearity=layers.HardSigmoid, candidate_nonlinearity=layers.HardTanh)
[ "def", "ConvDiagonalGRU", "(", "units", ",", "kernel_size", "=", "(", "3", ",", "3", ")", ")", ":", "def", "BuildConv", "(", ")", ":", "return", "layers", ".", "Conv", "(", "filters", "=", "units", ",", "kernel_size", "=", "kernel_size", ",", "padding", "=", "'SAME'", ")", "return", "layers", ".", "GeneralGRUCell", "(", "candidate_transform", "=", "BuildConv", ",", "memory_transform", "=", "DiagonalGate", ",", "gate_nonlinearity", "=", "layers", ".", "HardSigmoid", ",", "candidate_nonlinearity", "=", "layers", ".", "HardTanh", ")" ]
Build convolutional GRU with diagonal gating as in ImprovedNGPU.
[ "Build", "convolutional", "GRU", "with", "diagonal", "gating", "as", "in", "ImprovedNGPU", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/neural_gpu.py#L50-L60
22,381
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
strip_ids
def strip_ids(ids, ids_to_strip): """Strip ids_to_strip from the end ids.""" ids = list(ids) while ids and ids[-1] in ids_to_strip: ids.pop() return ids
python
def strip_ids(ids, ids_to_strip): """Strip ids_to_strip from the end ids.""" ids = list(ids) while ids and ids[-1] in ids_to_strip: ids.pop() return ids
[ "def", "strip_ids", "(", "ids", ",", "ids_to_strip", ")", ":", "ids", "=", "list", "(", "ids", ")", "while", "ids", "and", "ids", "[", "-", "1", "]", "in", "ids_to_strip", ":", "ids", ".", "pop", "(", ")", "return", "ids" ]
Strip ids_to_strip from the end ids.
[ "Strip", "ids_to_strip", "from", "the", "end", "ids", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L99-L104
22,382
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
_escape_token
def _escape_token(token, alphabet): """Escape away underscores and OOV characters and append '_'. This allows the token to be expressed as the concatenation of a list of subtokens from the vocabulary. The underscore acts as a sentinel which allows us to invertibly concatenate multiple such lists. Args: token: A unicode string to be escaped. alphabet: A set of all characters in the vocabulary's alphabet. Returns: escaped_token: An escaped unicode string. Raises: ValueError: If the provided token is not unicode. """ if not isinstance(token, six.text_type): raise ValueError("Expected string type for token, got %s" % type(token)) token = token.replace(u"\\", u"\\\\").replace(u"_", u"\\u") ret = [c if c in alphabet and c != u"\n" else r"\%d;" % ord(c) for c in token] return u"".join(ret) + "_"
python
def _escape_token(token, alphabet): """Escape away underscores and OOV characters and append '_'. This allows the token to be expressed as the concatenation of a list of subtokens from the vocabulary. The underscore acts as a sentinel which allows us to invertibly concatenate multiple such lists. Args: token: A unicode string to be escaped. alphabet: A set of all characters in the vocabulary's alphabet. Returns: escaped_token: An escaped unicode string. Raises: ValueError: If the provided token is not unicode. """ if not isinstance(token, six.text_type): raise ValueError("Expected string type for token, got %s" % type(token)) token = token.replace(u"\\", u"\\\\").replace(u"_", u"\\u") ret = [c if c in alphabet and c != u"\n" else r"\%d;" % ord(c) for c in token] return u"".join(ret) + "_"
[ "def", "_escape_token", "(", "token", ",", "alphabet", ")", ":", "if", "not", "isinstance", "(", "token", ",", "six", ".", "text_type", ")", ":", "raise", "ValueError", "(", "\"Expected string type for token, got %s\"", "%", "type", "(", "token", ")", ")", "token", "=", "token", ".", "replace", "(", "u\"\\\\\"", ",", "u\"\\\\\\\\\"", ")", ".", "replace", "(", "u\"_\"", ",", "u\"\\\\u\"", ")", "ret", "=", "[", "c", "if", "c", "in", "alphabet", "and", "c", "!=", "u\"\\n\"", "else", "r\"\\%d;\"", "%", "ord", "(", "c", ")", "for", "c", "in", "token", "]", "return", "u\"\"", ".", "join", "(", "ret", ")", "+", "\"_\"" ]
Escape away underscores and OOV characters and append '_'. This allows the token to be expressed as the concatenation of a list of subtokens from the vocabulary. The underscore acts as a sentinel which allows us to invertibly concatenate multiple such lists. Args: token: A unicode string to be escaped. alphabet: A set of all characters in the vocabulary's alphabet. Returns: escaped_token: An escaped unicode string. Raises: ValueError: If the provided token is not unicode.
[ "Escape", "away", "underscores", "and", "OOV", "characters", "and", "append", "_", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L400-L422
22,383
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TextEncoder.encode
def encode(self, s): """Transform a human-readable string into a sequence of int ids. The ids should be in the range [num_reserved_ids, vocab_size). Ids [0, num_reserved_ids) are reserved. EOS is not appended. Args: s: human-readable string to be converted. Returns: ids: list of integers """ return [int(w) + self._num_reserved_ids for w in s.split()]
python
def encode(self, s): """Transform a human-readable string into a sequence of int ids. The ids should be in the range [num_reserved_ids, vocab_size). Ids [0, num_reserved_ids) are reserved. EOS is not appended. Args: s: human-readable string to be converted. Returns: ids: list of integers """ return [int(w) + self._num_reserved_ids for w in s.split()]
[ "def", "encode", "(", "self", ",", "s", ")", ":", "return", "[", "int", "(", "w", ")", "+", "self", ".", "_num_reserved_ids", "for", "w", "in", "s", ".", "split", "(", ")", "]" ]
Transform a human-readable string into a sequence of int ids. The ids should be in the range [num_reserved_ids, vocab_size). Ids [0, num_reserved_ids) are reserved. EOS is not appended. Args: s: human-readable string to be converted. Returns: ids: list of integers
[ "Transform", "a", "human", "-", "readable", "string", "into", "a", "sequence", "of", "int", "ids", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L117-L131
22,384
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TextEncoder.decode
def decode(self, ids, strip_extraneous=False): """Transform a sequence of int ids into a human-readable string. EOS is not expected in ids. Args: ids: list of integers to be converted. strip_extraneous: bool, whether to strip off extraneous tokens (EOS and PAD). Returns: s: human-readable string. """ if strip_extraneous: ids = strip_ids(ids, list(range(self._num_reserved_ids or 0))) return " ".join(self.decode_list(ids))
python
def decode(self, ids, strip_extraneous=False): """Transform a sequence of int ids into a human-readable string. EOS is not expected in ids. Args: ids: list of integers to be converted. strip_extraneous: bool, whether to strip off extraneous tokens (EOS and PAD). Returns: s: human-readable string. """ if strip_extraneous: ids = strip_ids(ids, list(range(self._num_reserved_ids or 0))) return " ".join(self.decode_list(ids))
[ "def", "decode", "(", "self", ",", "ids", ",", "strip_extraneous", "=", "False", ")", ":", "if", "strip_extraneous", ":", "ids", "=", "strip_ids", "(", "ids", ",", "list", "(", "range", "(", "self", ".", "_num_reserved_ids", "or", "0", ")", ")", ")", "return", "\" \"", ".", "join", "(", "self", ".", "decode_list", "(", "ids", ")", ")" ]
Transform a sequence of int ids into a human-readable string. EOS is not expected in ids. Args: ids: list of integers to be converted. strip_extraneous: bool, whether to strip off extraneous tokens (EOS and PAD). Returns: s: human-readable string.
[ "Transform", "a", "sequence", "of", "int", "ids", "into", "a", "human", "-", "readable", "string", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L133-L148
22,385
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TextEncoder.decode_list
def decode_list(self, ids): """Transform a sequence of int ids into a their string versions. This method supports transforming individual input/output ids to their string versions so that sequence to/from text conversions can be visualized in a human readable format. Args: ids: list of integers to be converted. Returns: strs: list of human-readable string. """ decoded_ids = [] for id_ in ids: if 0 <= id_ < self._num_reserved_ids: decoded_ids.append(RESERVED_TOKENS[int(id_)]) else: decoded_ids.append(id_ - self._num_reserved_ids) return [str(d) for d in decoded_ids]
python
def decode_list(self, ids): """Transform a sequence of int ids into a their string versions. This method supports transforming individual input/output ids to their string versions so that sequence to/from text conversions can be visualized in a human readable format. Args: ids: list of integers to be converted. Returns: strs: list of human-readable string. """ decoded_ids = [] for id_ in ids: if 0 <= id_ < self._num_reserved_ids: decoded_ids.append(RESERVED_TOKENS[int(id_)]) else: decoded_ids.append(id_ - self._num_reserved_ids) return [str(d) for d in decoded_ids]
[ "def", "decode_list", "(", "self", ",", "ids", ")", ":", "decoded_ids", "=", "[", "]", "for", "id_", "in", "ids", ":", "if", "0", "<=", "id_", "<", "self", ".", "_num_reserved_ids", ":", "decoded_ids", ".", "append", "(", "RESERVED_TOKENS", "[", "int", "(", "id_", ")", "]", ")", "else", ":", "decoded_ids", ".", "append", "(", "id_", "-", "self", ".", "_num_reserved_ids", ")", "return", "[", "str", "(", "d", ")", "for", "d", "in", "decoded_ids", "]" ]
Transform a sequence of int ids into a their string versions. This method supports transforming individual input/output ids to their string versions so that sequence to/from text conversions can be visualized in a human readable format. Args: ids: list of integers to be converted. Returns: strs: list of human-readable string.
[ "Transform", "a", "sequence", "of", "int", "ids", "into", "a", "their", "string", "versions", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L150-L169
22,386
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TokenTextEncoder.encode
def encode(self, s): """Converts a space-separated string of tokens to a list of ids.""" sentence = s tokens = sentence.strip().split() if self._replace_oov is not None: tokens = [t if t in self._token_to_id else self._replace_oov for t in tokens] ret = [self._token_to_id[tok] for tok in tokens] return ret[::-1] if self._reverse else ret
python
def encode(self, s): """Converts a space-separated string of tokens to a list of ids.""" sentence = s tokens = sentence.strip().split() if self._replace_oov is not None: tokens = [t if t in self._token_to_id else self._replace_oov for t in tokens] ret = [self._token_to_id[tok] for tok in tokens] return ret[::-1] if self._reverse else ret
[ "def", "encode", "(", "self", ",", "s", ")", ":", "sentence", "=", "s", "tokens", "=", "sentence", ".", "strip", "(", ")", ".", "split", "(", ")", "if", "self", ".", "_replace_oov", "is", "not", "None", ":", "tokens", "=", "[", "t", "if", "t", "in", "self", ".", "_token_to_id", "else", "self", ".", "_replace_oov", "for", "t", "in", "tokens", "]", "ret", "=", "[", "self", ".", "_token_to_id", "[", "tok", "]", "for", "tok", "in", "tokens", "]", "return", "ret", "[", ":", ":", "-", "1", "]", "if", "self", ".", "_reverse", "else", "ret" ]
Converts a space-separated string of tokens to a list of ids.
[ "Converts", "a", "space", "-", "separated", "string", "of", "tokens", "to", "a", "list", "of", "ids", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L314-L322
22,387
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TokenTextEncoder._init_vocab_from_list
def _init_vocab_from_list(self, vocab_list): """Initialize tokens from a list of tokens. It is ok if reserved tokens appear in the vocab list. They will be removed. The set of tokens in vocab_list should be unique. Args: vocab_list: A list of tokens. """ def token_gen(): for token in vocab_list: if token not in RESERVED_TOKENS: yield token self._init_vocab(token_gen())
python
def _init_vocab_from_list(self, vocab_list): """Initialize tokens from a list of tokens. It is ok if reserved tokens appear in the vocab list. They will be removed. The set of tokens in vocab_list should be unique. Args: vocab_list: A list of tokens. """ def token_gen(): for token in vocab_list: if token not in RESERVED_TOKENS: yield token self._init_vocab(token_gen())
[ "def", "_init_vocab_from_list", "(", "self", ",", "vocab_list", ")", ":", "def", "token_gen", "(", ")", ":", "for", "token", "in", "vocab_list", ":", "if", "token", "not", "in", "RESERVED_TOKENS", ":", "yield", "token", "self", ".", "_init_vocab", "(", "token_gen", "(", ")", ")" ]
Initialize tokens from a list of tokens. It is ok if reserved tokens appear in the vocab list. They will be removed. The set of tokens in vocab_list should be unique. Args: vocab_list: A list of tokens.
[ "Initialize", "tokens", "from", "a", "list", "of", "tokens", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L353-L367
22,388
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TokenTextEncoder._init_vocab
def _init_vocab(self, token_generator, add_reserved_tokens=True): """Initialize vocabulary with tokens from token_generator.""" self._id_to_token = {} non_reserved_start_index = 0 if add_reserved_tokens: self._id_to_token.update(enumerate(RESERVED_TOKENS)) non_reserved_start_index = len(RESERVED_TOKENS) self._id_to_token.update( enumerate(token_generator, start=non_reserved_start_index)) # _token_to_id is the reverse of _id_to_token self._token_to_id = dict((v, k) for k, v in six.iteritems(self._id_to_token))
python
def _init_vocab(self, token_generator, add_reserved_tokens=True): """Initialize vocabulary with tokens from token_generator.""" self._id_to_token = {} non_reserved_start_index = 0 if add_reserved_tokens: self._id_to_token.update(enumerate(RESERVED_TOKENS)) non_reserved_start_index = len(RESERVED_TOKENS) self._id_to_token.update( enumerate(token_generator, start=non_reserved_start_index)) # _token_to_id is the reverse of _id_to_token self._token_to_id = dict((v, k) for k, v in six.iteritems(self._id_to_token))
[ "def", "_init_vocab", "(", "self", ",", "token_generator", ",", "add_reserved_tokens", "=", "True", ")", ":", "self", ".", "_id_to_token", "=", "{", "}", "non_reserved_start_index", "=", "0", "if", "add_reserved_tokens", ":", "self", ".", "_id_to_token", ".", "update", "(", "enumerate", "(", "RESERVED_TOKENS", ")", ")", "non_reserved_start_index", "=", "len", "(", "RESERVED_TOKENS", ")", "self", ".", "_id_to_token", ".", "update", "(", "enumerate", "(", "token_generator", ",", "start", "=", "non_reserved_start_index", ")", ")", "# _token_to_id is the reverse of _id_to_token", "self", ".", "_token_to_id", "=", "dict", "(", "(", "v", ",", "k", ")", "for", "k", ",", "v", "in", "six", ".", "iteritems", "(", "self", ".", "_id_to_token", ")", ")" ]
Initialize vocabulary with tokens from token_generator.
[ "Initialize", "vocabulary", "with", "tokens", "from", "token_generator", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L369-L384
22,389
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TokenTextEncoder.store_to_file
def store_to_file(self, filename): """Write vocab file to disk. Vocab files have one token per line. The file ends in a newline. Reserved tokens are written to the vocab file as well. Args: filename: Full path of the file to store the vocab to. """ with tf.gfile.Open(filename, "w") as f: for i in range(len(self._id_to_token)): f.write(self._id_to_token[i] + "\n")
python
def store_to_file(self, filename): """Write vocab file to disk. Vocab files have one token per line. The file ends in a newline. Reserved tokens are written to the vocab file as well. Args: filename: Full path of the file to store the vocab to. """ with tf.gfile.Open(filename, "w") as f: for i in range(len(self._id_to_token)): f.write(self._id_to_token[i] + "\n")
[ "def", "store_to_file", "(", "self", ",", "filename", ")", ":", "with", "tf", ".", "gfile", ".", "Open", "(", "filename", ",", "\"w\"", ")", "as", "f", ":", "for", "i", "in", "range", "(", "len", "(", "self", ".", "_id_to_token", ")", ")", ":", "f", ".", "write", "(", "self", ".", "_id_to_token", "[", "i", "]", "+", "\"\\n\"", ")" ]
Write vocab file to disk. Vocab files have one token per line. The file ends in a newline. Reserved tokens are written to the vocab file as well. Args: filename: Full path of the file to store the vocab to.
[ "Write", "vocab", "file", "to", "disk", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L386-L397
22,390
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder.decode
def decode(self, ids, strip_extraneous=False): """Converts a sequence of subtoken ids to a native string. Args: ids: a list of integers in the range [0, vocab_size) strip_extraneous: bool, whether to strip off extraneous tokens (EOS and PAD). Returns: a native string """ if strip_extraneous: ids = strip_ids(ids, list(range(self._num_reserved_ids or 0))) return unicode_to_native( tokenizer.decode(self._subtoken_ids_to_tokens(ids)))
python
def decode(self, ids, strip_extraneous=False): """Converts a sequence of subtoken ids to a native string. Args: ids: a list of integers in the range [0, vocab_size) strip_extraneous: bool, whether to strip off extraneous tokens (EOS and PAD). Returns: a native string """ if strip_extraneous: ids = strip_ids(ids, list(range(self._num_reserved_ids or 0))) return unicode_to_native( tokenizer.decode(self._subtoken_ids_to_tokens(ids)))
[ "def", "decode", "(", "self", ",", "ids", ",", "strip_extraneous", "=", "False", ")", ":", "if", "strip_extraneous", ":", "ids", "=", "strip_ids", "(", "ids", ",", "list", "(", "range", "(", "self", ".", "_num_reserved_ids", "or", "0", ")", ")", ")", "return", "unicode_to_native", "(", "tokenizer", ".", "decode", "(", "self", ".", "_subtoken_ids_to_tokens", "(", "ids", ")", ")", ")" ]
Converts a sequence of subtoken ids to a native string. Args: ids: a list of integers in the range [0, vocab_size) strip_extraneous: bool, whether to strip off extraneous tokens (EOS and PAD). Returns: a native string
[ "Converts", "a", "sequence", "of", "subtoken", "ids", "to", "a", "native", "string", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L522-L536
22,391
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._tokens_to_subtoken_ids
def _tokens_to_subtoken_ids(self, tokens): """Converts a list of tokens to a list of subtoken ids. Args: tokens: a list of strings. Returns: a list of integers in the range [0, vocab_size) """ ret = [] for token in tokens: ret.extend(self._token_to_subtoken_ids(token)) return ret
python
def _tokens_to_subtoken_ids(self, tokens): """Converts a list of tokens to a list of subtoken ids. Args: tokens: a list of strings. Returns: a list of integers in the range [0, vocab_size) """ ret = [] for token in tokens: ret.extend(self._token_to_subtoken_ids(token)) return ret
[ "def", "_tokens_to_subtoken_ids", "(", "self", ",", "tokens", ")", ":", "ret", "=", "[", "]", "for", "token", "in", "tokens", ":", "ret", ".", "extend", "(", "self", ".", "_token_to_subtoken_ids", "(", "token", ")", ")", "return", "ret" ]
Converts a list of tokens to a list of subtoken ids. Args: tokens: a list of strings. Returns: a list of integers in the range [0, vocab_size)
[ "Converts", "a", "list", "of", "tokens", "to", "a", "list", "of", "subtoken", "ids", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L546-L557
22,392
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._token_to_subtoken_ids
def _token_to_subtoken_ids(self, token): """Converts token to a list of subtoken ids. Args: token: a string. Returns: a list of integers in the range [0, vocab_size) """ cache_location = hash(token) % self._cache_size cache_key, cache_value = self._cache[cache_location] if cache_key == token: return cache_value ret = self._escaped_token_to_subtoken_ids( _escape_token(token, self._alphabet)) self._cache[cache_location] = (token, ret) return ret
python
def _token_to_subtoken_ids(self, token): """Converts token to a list of subtoken ids. Args: token: a string. Returns: a list of integers in the range [0, vocab_size) """ cache_location = hash(token) % self._cache_size cache_key, cache_value = self._cache[cache_location] if cache_key == token: return cache_value ret = self._escaped_token_to_subtoken_ids( _escape_token(token, self._alphabet)) self._cache[cache_location] = (token, ret) return ret
[ "def", "_token_to_subtoken_ids", "(", "self", ",", "token", ")", ":", "cache_location", "=", "hash", "(", "token", ")", "%", "self", ".", "_cache_size", "cache_key", ",", "cache_value", "=", "self", ".", "_cache", "[", "cache_location", "]", "if", "cache_key", "==", "token", ":", "return", "cache_value", "ret", "=", "self", ".", "_escaped_token_to_subtoken_ids", "(", "_escape_token", "(", "token", ",", "self", ".", "_alphabet", ")", ")", "self", ".", "_cache", "[", "cache_location", "]", "=", "(", "token", ",", "ret", ")", "return", "ret" ]
Converts token to a list of subtoken ids. Args: token: a string. Returns: a list of integers in the range [0, vocab_size)
[ "Converts", "token", "to", "a", "list", "of", "subtoken", "ids", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L559-L574
22,393
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._subtoken_ids_to_tokens
def _subtoken_ids_to_tokens(self, subtokens): """Converts a list of subtoken ids to a list of tokens. Args: subtokens: a list of integers in the range [0, vocab_size) Returns: a list of strings. """ concatenated = "".join( [self._subtoken_id_to_subtoken_string(s) for s in subtokens]) split = concatenated.split("_") ret = [] for t in split: if t: unescaped = _unescape_token(t + "_") if unescaped: ret.append(unescaped) return ret
python
def _subtoken_ids_to_tokens(self, subtokens): """Converts a list of subtoken ids to a list of tokens. Args: subtokens: a list of integers in the range [0, vocab_size) Returns: a list of strings. """ concatenated = "".join( [self._subtoken_id_to_subtoken_string(s) for s in subtokens]) split = concatenated.split("_") ret = [] for t in split: if t: unescaped = _unescape_token(t + "_") if unescaped: ret.append(unescaped) return ret
[ "def", "_subtoken_ids_to_tokens", "(", "self", ",", "subtokens", ")", ":", "concatenated", "=", "\"\"", ".", "join", "(", "[", "self", ".", "_subtoken_id_to_subtoken_string", "(", "s", ")", "for", "s", "in", "subtokens", "]", ")", "split", "=", "concatenated", ".", "split", "(", "\"_\"", ")", "ret", "=", "[", "]", "for", "t", "in", "split", ":", "if", "t", ":", "unescaped", "=", "_unescape_token", "(", "t", "+", "\"_\"", ")", "if", "unescaped", ":", "ret", ".", "append", "(", "unescaped", ")", "return", "ret" ]
Converts a list of subtoken ids to a list of tokens. Args: subtokens: a list of integers in the range [0, vocab_size) Returns: a list of strings.
[ "Converts", "a", "list", "of", "subtoken", "ids", "to", "a", "list", "of", "tokens", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L576-L593
22,394
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._subtoken_id_to_subtoken_string
def _subtoken_id_to_subtoken_string(self, subtoken): """Converts a subtoken integer ID to a subtoken string.""" if 0 <= subtoken < self.vocab_size: return self._all_subtoken_strings[subtoken] return u""
python
def _subtoken_id_to_subtoken_string(self, subtoken): """Converts a subtoken integer ID to a subtoken string.""" if 0 <= subtoken < self.vocab_size: return self._all_subtoken_strings[subtoken] return u""
[ "def", "_subtoken_id_to_subtoken_string", "(", "self", ",", "subtoken", ")", ":", "if", "0", "<=", "subtoken", "<", "self", ".", "vocab_size", ":", "return", "self", ".", "_all_subtoken_strings", "[", "subtoken", "]", "return", "u\"\"" ]
Converts a subtoken integer ID to a subtoken string.
[ "Converts", "a", "subtoken", "integer", "ID", "to", "a", "subtoken", "string", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L595-L599
22,395
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._escaped_token_to_subtoken_ids
def _escaped_token_to_subtoken_ids(self, escaped_token): """Converts an escaped token string to a list of subtoken IDs. Args: escaped_token: An escaped token as a unicode string. Returns: A list of subtoken IDs as integers. """ return [ self._subtoken_string_to_id[subtoken] for subtoken in self._escaped_token_to_subtoken_strings(escaped_token) ]
python
def _escaped_token_to_subtoken_ids(self, escaped_token): """Converts an escaped token string to a list of subtoken IDs. Args: escaped_token: An escaped token as a unicode string. Returns: A list of subtoken IDs as integers. """ return [ self._subtoken_string_to_id[subtoken] for subtoken in self._escaped_token_to_subtoken_strings(escaped_token) ]
[ "def", "_escaped_token_to_subtoken_ids", "(", "self", ",", "escaped_token", ")", ":", "return", "[", "self", ".", "_subtoken_string_to_id", "[", "subtoken", "]", "for", "subtoken", "in", "self", ".", "_escaped_token_to_subtoken_strings", "(", "escaped_token", ")", "]" ]
Converts an escaped token string to a list of subtoken IDs. Args: escaped_token: An escaped token as a unicode string. Returns: A list of subtoken IDs as integers.
[ "Converts", "an", "escaped", "token", "string", "to", "a", "list", "of", "subtoken", "IDs", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L631-L642
22,396
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder.build_from_generator
def build_from_generator(cls, generator, target_size, max_subtoken_length=None, reserved_tokens=None): """Builds a SubwordTextEncoder from the generated text. Args: generator: yields text. target_size: int, approximate vocabulary size to create. max_subtoken_length: Maximum length of a subtoken. If this is not set, then the runtime and memory use of creating the vocab is quadratic in the length of the longest token. If this is set, then it is instead O(max_subtoken_length * length of longest token). reserved_tokens: List of reserved tokens. The global variable `RESERVED_TOKENS` must be a prefix of `reserved_tokens`. If this argument is `None`, it will use `RESERVED_TOKENS`. Returns: SubwordTextEncoder with `vocab_size` approximately `target_size`. """ token_counts = collections.defaultdict(int) for item in generator: for tok in tokenizer.encode(native_to_unicode(item)): token_counts[tok] += 1 encoder = cls.build_to_target_size( target_size, token_counts, 1, 1e3, max_subtoken_length=max_subtoken_length, reserved_tokens=reserved_tokens) return encoder
python
def build_from_generator(cls, generator, target_size, max_subtoken_length=None, reserved_tokens=None): """Builds a SubwordTextEncoder from the generated text. Args: generator: yields text. target_size: int, approximate vocabulary size to create. max_subtoken_length: Maximum length of a subtoken. If this is not set, then the runtime and memory use of creating the vocab is quadratic in the length of the longest token. If this is set, then it is instead O(max_subtoken_length * length of longest token). reserved_tokens: List of reserved tokens. The global variable `RESERVED_TOKENS` must be a prefix of `reserved_tokens`. If this argument is `None`, it will use `RESERVED_TOKENS`. Returns: SubwordTextEncoder with `vocab_size` approximately `target_size`. """ token_counts = collections.defaultdict(int) for item in generator: for tok in tokenizer.encode(native_to_unicode(item)): token_counts[tok] += 1 encoder = cls.build_to_target_size( target_size, token_counts, 1, 1e3, max_subtoken_length=max_subtoken_length, reserved_tokens=reserved_tokens) return encoder
[ "def", "build_from_generator", "(", "cls", ",", "generator", ",", "target_size", ",", "max_subtoken_length", "=", "None", ",", "reserved_tokens", "=", "None", ")", ":", "token_counts", "=", "collections", ".", "defaultdict", "(", "int", ")", "for", "item", "in", "generator", ":", "for", "tok", "in", "tokenizer", ".", "encode", "(", "native_to_unicode", "(", "item", ")", ")", ":", "token_counts", "[", "tok", "]", "+=", "1", "encoder", "=", "cls", ".", "build_to_target_size", "(", "target_size", ",", "token_counts", ",", "1", ",", "1e3", ",", "max_subtoken_length", "=", "max_subtoken_length", ",", "reserved_tokens", "=", "reserved_tokens", ")", "return", "encoder" ]
Builds a SubwordTextEncoder from the generated text. Args: generator: yields text. target_size: int, approximate vocabulary size to create. max_subtoken_length: Maximum length of a subtoken. If this is not set, then the runtime and memory use of creating the vocab is quadratic in the length of the longest token. If this is set, then it is instead O(max_subtoken_length * length of longest token). reserved_tokens: List of reserved tokens. The global variable `RESERVED_TOKENS` must be a prefix of `reserved_tokens`. If this argument is `None`, it will use `RESERVED_TOKENS`. Returns: SubwordTextEncoder with `vocab_size` approximately `target_size`.
[ "Builds", "a", "SubwordTextEncoder", "from", "the", "generated", "text", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L645-L674
22,397
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder.build_to_target_size
def build_to_target_size(cls, target_size, token_counts, min_val, max_val, max_subtoken_length=None, reserved_tokens=None, num_iterations=4): """Builds a SubwordTextEncoder that has `vocab_size` near `target_size`. Uses simple recursive binary search to find a minimum token count that most closely matches the `target_size`. Args: target_size: Desired vocab_size to approximate. token_counts: A dictionary of token counts, mapping string to int. min_val: An integer; lower bound for the minimum token count. max_val: An integer; upper bound for the minimum token count. max_subtoken_length: Maximum length of a subtoken. If this is not set, then the runtime and memory use of creating the vocab is quadratic in the length of the longest token. If this is set, then it is instead O(max_subtoken_length * length of longest token). reserved_tokens: List of reserved tokens. The global variable `RESERVED_TOKENS` must be a prefix of `reserved_tokens`. If this argument is `None`, it will use `RESERVED_TOKENS`. num_iterations: An integer; how many iterations of refinement. Returns: A SubwordTextEncoder instance. Raises: ValueError: If `min_val` is greater than `max_val`. """ if min_val > max_val: raise ValueError("Lower bound for the minimum token count " "is greater than the upper bound.") if target_size < 1: raise ValueError("Target size must be positive.") if reserved_tokens is None: reserved_tokens = RESERVED_TOKENS def bisect(min_val, max_val): """Bisection to find the right size.""" present_count = (max_val + min_val) // 2 tf.logging.info("Trying min_count %d" % present_count) subtokenizer = cls() subtokenizer.build_from_token_counts( token_counts, present_count, num_iterations, max_subtoken_length=max_subtoken_length, reserved_tokens=reserved_tokens) # Being within 1% of the target size is ok. is_ok = abs(subtokenizer.vocab_size - target_size) * 100 < target_size # If min_val == max_val, we can't do any better than this. if is_ok or min_val >= max_val or present_count < 2: return subtokenizer if subtokenizer.vocab_size > target_size: other_subtokenizer = bisect(present_count + 1, max_val) else: other_subtokenizer = bisect(min_val, present_count - 1) if other_subtokenizer is None: return subtokenizer if (abs(other_subtokenizer.vocab_size - target_size) < abs(subtokenizer.vocab_size - target_size)): return other_subtokenizer return subtokenizer return bisect(min_val, max_val)
python
def build_to_target_size(cls, target_size, token_counts, min_val, max_val, max_subtoken_length=None, reserved_tokens=None, num_iterations=4): """Builds a SubwordTextEncoder that has `vocab_size` near `target_size`. Uses simple recursive binary search to find a minimum token count that most closely matches the `target_size`. Args: target_size: Desired vocab_size to approximate. token_counts: A dictionary of token counts, mapping string to int. min_val: An integer; lower bound for the minimum token count. max_val: An integer; upper bound for the minimum token count. max_subtoken_length: Maximum length of a subtoken. If this is not set, then the runtime and memory use of creating the vocab is quadratic in the length of the longest token. If this is set, then it is instead O(max_subtoken_length * length of longest token). reserved_tokens: List of reserved tokens. The global variable `RESERVED_TOKENS` must be a prefix of `reserved_tokens`. If this argument is `None`, it will use `RESERVED_TOKENS`. num_iterations: An integer; how many iterations of refinement. Returns: A SubwordTextEncoder instance. Raises: ValueError: If `min_val` is greater than `max_val`. """ if min_val > max_val: raise ValueError("Lower bound for the minimum token count " "is greater than the upper bound.") if target_size < 1: raise ValueError("Target size must be positive.") if reserved_tokens is None: reserved_tokens = RESERVED_TOKENS def bisect(min_val, max_val): """Bisection to find the right size.""" present_count = (max_val + min_val) // 2 tf.logging.info("Trying min_count %d" % present_count) subtokenizer = cls() subtokenizer.build_from_token_counts( token_counts, present_count, num_iterations, max_subtoken_length=max_subtoken_length, reserved_tokens=reserved_tokens) # Being within 1% of the target size is ok. is_ok = abs(subtokenizer.vocab_size - target_size) * 100 < target_size # If min_val == max_val, we can't do any better than this. if is_ok or min_val >= max_val or present_count < 2: return subtokenizer if subtokenizer.vocab_size > target_size: other_subtokenizer = bisect(present_count + 1, max_val) else: other_subtokenizer = bisect(min_val, present_count - 1) if other_subtokenizer is None: return subtokenizer if (abs(other_subtokenizer.vocab_size - target_size) < abs(subtokenizer.vocab_size - target_size)): return other_subtokenizer return subtokenizer return bisect(min_val, max_val)
[ "def", "build_to_target_size", "(", "cls", ",", "target_size", ",", "token_counts", ",", "min_val", ",", "max_val", ",", "max_subtoken_length", "=", "None", ",", "reserved_tokens", "=", "None", ",", "num_iterations", "=", "4", ")", ":", "if", "min_val", ">", "max_val", ":", "raise", "ValueError", "(", "\"Lower bound for the minimum token count \"", "\"is greater than the upper bound.\"", ")", "if", "target_size", "<", "1", ":", "raise", "ValueError", "(", "\"Target size must be positive.\"", ")", "if", "reserved_tokens", "is", "None", ":", "reserved_tokens", "=", "RESERVED_TOKENS", "def", "bisect", "(", "min_val", ",", "max_val", ")", ":", "\"\"\"Bisection to find the right size.\"\"\"", "present_count", "=", "(", "max_val", "+", "min_val", ")", "//", "2", "tf", ".", "logging", ".", "info", "(", "\"Trying min_count %d\"", "%", "present_count", ")", "subtokenizer", "=", "cls", "(", ")", "subtokenizer", ".", "build_from_token_counts", "(", "token_counts", ",", "present_count", ",", "num_iterations", ",", "max_subtoken_length", "=", "max_subtoken_length", ",", "reserved_tokens", "=", "reserved_tokens", ")", "# Being within 1% of the target size is ok.", "is_ok", "=", "abs", "(", "subtokenizer", ".", "vocab_size", "-", "target_size", ")", "*", "100", "<", "target_size", "# If min_val == max_val, we can't do any better than this.", "if", "is_ok", "or", "min_val", ">=", "max_val", "or", "present_count", "<", "2", ":", "return", "subtokenizer", "if", "subtokenizer", ".", "vocab_size", ">", "target_size", ":", "other_subtokenizer", "=", "bisect", "(", "present_count", "+", "1", ",", "max_val", ")", "else", ":", "other_subtokenizer", "=", "bisect", "(", "min_val", ",", "present_count", "-", "1", ")", "if", "other_subtokenizer", "is", "None", ":", "return", "subtokenizer", "if", "(", "abs", "(", "other_subtokenizer", ".", "vocab_size", "-", "target_size", ")", "<", "abs", "(", "subtokenizer", ".", "vocab_size", "-", "target_size", ")", ")", ":", "return", "other_subtokenizer", "return", "subtokenizer", "return", "bisect", "(", "min_val", ",", "max_val", ")" ]
Builds a SubwordTextEncoder that has `vocab_size` near `target_size`. Uses simple recursive binary search to find a minimum token count that most closely matches the `target_size`. Args: target_size: Desired vocab_size to approximate. token_counts: A dictionary of token counts, mapping string to int. min_val: An integer; lower bound for the minimum token count. max_val: An integer; upper bound for the minimum token count. max_subtoken_length: Maximum length of a subtoken. If this is not set, then the runtime and memory use of creating the vocab is quadratic in the length of the longest token. If this is set, then it is instead O(max_subtoken_length * length of longest token). reserved_tokens: List of reserved tokens. The global variable `RESERVED_TOKENS` must be a prefix of `reserved_tokens`. If this argument is `None`, it will use `RESERVED_TOKENS`. num_iterations: An integer; how many iterations of refinement. Returns: A SubwordTextEncoder instance. Raises: ValueError: If `min_val` is greater than `max_val`.
[ "Builds", "a", "SubwordTextEncoder", "that", "has", "vocab_size", "near", "target_size", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L677-L748
22,398
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder.build_from_token_counts
def build_from_token_counts(self, token_counts, min_count, num_iterations=4, reserved_tokens=None, max_subtoken_length=None): """Train a SubwordTextEncoder based on a dictionary of word counts. Args: token_counts: a dictionary of Unicode strings to int. min_count: an integer - discard subtokens with lower counts. num_iterations: an integer. how many iterations of refinement. reserved_tokens: List of reserved tokens. The global variable `RESERVED_TOKENS` must be a prefix of `reserved_tokens`. If this argument is `None`, it will use `RESERVED_TOKENS`. max_subtoken_length: Maximum length of a subtoken. If this is not set, then the runtime and memory use of creating the vocab is quadratic in the length of the longest token. If this is set, then it is instead O(max_subtoken_length * length of longest token). Raises: ValueError: if reserved is not 0 or len(RESERVED_TOKENS). In this case, it is not clear what the space is being reserved for, or when it will be filled in. """ if reserved_tokens is None: reserved_tokens = RESERVED_TOKENS else: # There is not complete freedom in replacing RESERVED_TOKENS. for default, proposed in zip(RESERVED_TOKENS, reserved_tokens): if default != proposed: raise ValueError("RESERVED_TOKENS must be a prefix of " "reserved_tokens.") # Initialize the alphabet. Note, this must include reserved tokens or it can # result in encoding failures. alphabet_tokens = chain(six.iterkeys(token_counts), [native_to_unicode(t) for t in reserved_tokens]) self._init_alphabet_from_tokens(alphabet_tokens) # Bootstrap the initial list of subtokens with the characters from the # alphabet plus the escaping characters. self._init_subtokens_from_list(list(self._alphabet), reserved_tokens=reserved_tokens) # We build iteratively. On each iteration, we segment all the words, # then count the resulting potential subtokens, keeping the ones # with high enough counts for our new vocabulary. if min_count < 1: min_count = 1 for i in range(num_iterations): tf.logging.info("Iteration {0}".format(i)) # Collect all substrings of the encoded token that break along current # subtoken boundaries. subtoken_counts = collections.defaultdict(int) for token, count in six.iteritems(token_counts): iter_start_time = time.time() escaped_token = _escape_token(token, self._alphabet) subtokens = self._escaped_token_to_subtoken_strings(escaped_token) start = 0 for subtoken in subtokens: last_position = len(escaped_token) + 1 if max_subtoken_length is not None: last_position = min(last_position, start + max_subtoken_length) for end in range(start + 1, last_position): new_subtoken = escaped_token[start:end] subtoken_counts[new_subtoken] += count start += len(subtoken) iter_time_secs = time.time() - iter_start_time if iter_time_secs > 0.1: tf.logging.info(u"Processing token [{0}] took {1} seconds, consider " "setting Text2TextProblem.max_subtoken_length to a " "smaller value.".format(token, iter_time_secs)) # Array of sets of candidate subtoken strings, by length. len_to_subtoken_strings = [] for subtoken_string, count in six.iteritems(subtoken_counts): lsub = len(subtoken_string) if count >= min_count: while len(len_to_subtoken_strings) <= lsub: len_to_subtoken_strings.append(set()) len_to_subtoken_strings[lsub].add(subtoken_string) # Consider the candidates longest to shortest, so that if we accept # a longer subtoken string, we can decrement the counts of its prefixes. new_subtoken_strings = [] for lsub in range(len(len_to_subtoken_strings) - 1, 0, -1): subtoken_strings = len_to_subtoken_strings[lsub] for subtoken_string in subtoken_strings: count = subtoken_counts[subtoken_string] if count >= min_count: # Exclude alphabet tokens here, as they must be included later, # explicitly, regardless of count. if subtoken_string not in self._alphabet: new_subtoken_strings.append((count, subtoken_string)) for l in range(1, lsub): subtoken_counts[subtoken_string[:l]] -= count # Include the alphabet explicitly to guarantee all strings are encodable. new_subtoken_strings.extend((subtoken_counts.get(a, 0), a) for a in self._alphabet) new_subtoken_strings.sort(reverse=True) # Reinitialize to the candidate vocabulary. new_subtoken_strings = [subtoken for _, subtoken in new_subtoken_strings] if reserved_tokens: escaped_reserved_tokens = [ _escape_token(native_to_unicode(t), self._alphabet) for t in reserved_tokens ] new_subtoken_strings = escaped_reserved_tokens + new_subtoken_strings self._init_subtokens_from_list(new_subtoken_strings) tf.logging.info("vocab_size = %d" % self.vocab_size)
python
def build_from_token_counts(self, token_counts, min_count, num_iterations=4, reserved_tokens=None, max_subtoken_length=None): """Train a SubwordTextEncoder based on a dictionary of word counts. Args: token_counts: a dictionary of Unicode strings to int. min_count: an integer - discard subtokens with lower counts. num_iterations: an integer. how many iterations of refinement. reserved_tokens: List of reserved tokens. The global variable `RESERVED_TOKENS` must be a prefix of `reserved_tokens`. If this argument is `None`, it will use `RESERVED_TOKENS`. max_subtoken_length: Maximum length of a subtoken. If this is not set, then the runtime and memory use of creating the vocab is quadratic in the length of the longest token. If this is set, then it is instead O(max_subtoken_length * length of longest token). Raises: ValueError: if reserved is not 0 or len(RESERVED_TOKENS). In this case, it is not clear what the space is being reserved for, or when it will be filled in. """ if reserved_tokens is None: reserved_tokens = RESERVED_TOKENS else: # There is not complete freedom in replacing RESERVED_TOKENS. for default, proposed in zip(RESERVED_TOKENS, reserved_tokens): if default != proposed: raise ValueError("RESERVED_TOKENS must be a prefix of " "reserved_tokens.") # Initialize the alphabet. Note, this must include reserved tokens or it can # result in encoding failures. alphabet_tokens = chain(six.iterkeys(token_counts), [native_to_unicode(t) for t in reserved_tokens]) self._init_alphabet_from_tokens(alphabet_tokens) # Bootstrap the initial list of subtokens with the characters from the # alphabet plus the escaping characters. self._init_subtokens_from_list(list(self._alphabet), reserved_tokens=reserved_tokens) # We build iteratively. On each iteration, we segment all the words, # then count the resulting potential subtokens, keeping the ones # with high enough counts for our new vocabulary. if min_count < 1: min_count = 1 for i in range(num_iterations): tf.logging.info("Iteration {0}".format(i)) # Collect all substrings of the encoded token that break along current # subtoken boundaries. subtoken_counts = collections.defaultdict(int) for token, count in six.iteritems(token_counts): iter_start_time = time.time() escaped_token = _escape_token(token, self._alphabet) subtokens = self._escaped_token_to_subtoken_strings(escaped_token) start = 0 for subtoken in subtokens: last_position = len(escaped_token) + 1 if max_subtoken_length is not None: last_position = min(last_position, start + max_subtoken_length) for end in range(start + 1, last_position): new_subtoken = escaped_token[start:end] subtoken_counts[new_subtoken] += count start += len(subtoken) iter_time_secs = time.time() - iter_start_time if iter_time_secs > 0.1: tf.logging.info(u"Processing token [{0}] took {1} seconds, consider " "setting Text2TextProblem.max_subtoken_length to a " "smaller value.".format(token, iter_time_secs)) # Array of sets of candidate subtoken strings, by length. len_to_subtoken_strings = [] for subtoken_string, count in six.iteritems(subtoken_counts): lsub = len(subtoken_string) if count >= min_count: while len(len_to_subtoken_strings) <= lsub: len_to_subtoken_strings.append(set()) len_to_subtoken_strings[lsub].add(subtoken_string) # Consider the candidates longest to shortest, so that if we accept # a longer subtoken string, we can decrement the counts of its prefixes. new_subtoken_strings = [] for lsub in range(len(len_to_subtoken_strings) - 1, 0, -1): subtoken_strings = len_to_subtoken_strings[lsub] for subtoken_string in subtoken_strings: count = subtoken_counts[subtoken_string] if count >= min_count: # Exclude alphabet tokens here, as they must be included later, # explicitly, regardless of count. if subtoken_string not in self._alphabet: new_subtoken_strings.append((count, subtoken_string)) for l in range(1, lsub): subtoken_counts[subtoken_string[:l]] -= count # Include the alphabet explicitly to guarantee all strings are encodable. new_subtoken_strings.extend((subtoken_counts.get(a, 0), a) for a in self._alphabet) new_subtoken_strings.sort(reverse=True) # Reinitialize to the candidate vocabulary. new_subtoken_strings = [subtoken for _, subtoken in new_subtoken_strings] if reserved_tokens: escaped_reserved_tokens = [ _escape_token(native_to_unicode(t), self._alphabet) for t in reserved_tokens ] new_subtoken_strings = escaped_reserved_tokens + new_subtoken_strings self._init_subtokens_from_list(new_subtoken_strings) tf.logging.info("vocab_size = %d" % self.vocab_size)
[ "def", "build_from_token_counts", "(", "self", ",", "token_counts", ",", "min_count", ",", "num_iterations", "=", "4", ",", "reserved_tokens", "=", "None", ",", "max_subtoken_length", "=", "None", ")", ":", "if", "reserved_tokens", "is", "None", ":", "reserved_tokens", "=", "RESERVED_TOKENS", "else", ":", "# There is not complete freedom in replacing RESERVED_TOKENS.", "for", "default", ",", "proposed", "in", "zip", "(", "RESERVED_TOKENS", ",", "reserved_tokens", ")", ":", "if", "default", "!=", "proposed", ":", "raise", "ValueError", "(", "\"RESERVED_TOKENS must be a prefix of \"", "\"reserved_tokens.\"", ")", "# Initialize the alphabet. Note, this must include reserved tokens or it can", "# result in encoding failures.", "alphabet_tokens", "=", "chain", "(", "six", ".", "iterkeys", "(", "token_counts", ")", ",", "[", "native_to_unicode", "(", "t", ")", "for", "t", "in", "reserved_tokens", "]", ")", "self", ".", "_init_alphabet_from_tokens", "(", "alphabet_tokens", ")", "# Bootstrap the initial list of subtokens with the characters from the", "# alphabet plus the escaping characters.", "self", ".", "_init_subtokens_from_list", "(", "list", "(", "self", ".", "_alphabet", ")", ",", "reserved_tokens", "=", "reserved_tokens", ")", "# We build iteratively. On each iteration, we segment all the words,", "# then count the resulting potential subtokens, keeping the ones", "# with high enough counts for our new vocabulary.", "if", "min_count", "<", "1", ":", "min_count", "=", "1", "for", "i", "in", "range", "(", "num_iterations", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Iteration {0}\"", ".", "format", "(", "i", ")", ")", "# Collect all substrings of the encoded token that break along current", "# subtoken boundaries.", "subtoken_counts", "=", "collections", ".", "defaultdict", "(", "int", ")", "for", "token", ",", "count", "in", "six", ".", "iteritems", "(", "token_counts", ")", ":", "iter_start_time", "=", "time", ".", "time", "(", ")", "escaped_token", "=", "_escape_token", "(", "token", ",", "self", ".", "_alphabet", ")", "subtokens", "=", "self", ".", "_escaped_token_to_subtoken_strings", "(", "escaped_token", ")", "start", "=", "0", "for", "subtoken", "in", "subtokens", ":", "last_position", "=", "len", "(", "escaped_token", ")", "+", "1", "if", "max_subtoken_length", "is", "not", "None", ":", "last_position", "=", "min", "(", "last_position", ",", "start", "+", "max_subtoken_length", ")", "for", "end", "in", "range", "(", "start", "+", "1", ",", "last_position", ")", ":", "new_subtoken", "=", "escaped_token", "[", "start", ":", "end", "]", "subtoken_counts", "[", "new_subtoken", "]", "+=", "count", "start", "+=", "len", "(", "subtoken", ")", "iter_time_secs", "=", "time", ".", "time", "(", ")", "-", "iter_start_time", "if", "iter_time_secs", ">", "0.1", ":", "tf", ".", "logging", ".", "info", "(", "u\"Processing token [{0}] took {1} seconds, consider \"", "\"setting Text2TextProblem.max_subtoken_length to a \"", "\"smaller value.\"", ".", "format", "(", "token", ",", "iter_time_secs", ")", ")", "# Array of sets of candidate subtoken strings, by length.", "len_to_subtoken_strings", "=", "[", "]", "for", "subtoken_string", ",", "count", "in", "six", ".", "iteritems", "(", "subtoken_counts", ")", ":", "lsub", "=", "len", "(", "subtoken_string", ")", "if", "count", ">=", "min_count", ":", "while", "len", "(", "len_to_subtoken_strings", ")", "<=", "lsub", ":", "len_to_subtoken_strings", ".", "append", "(", "set", "(", ")", ")", "len_to_subtoken_strings", "[", "lsub", "]", ".", "add", "(", "subtoken_string", ")", "# Consider the candidates longest to shortest, so that if we accept", "# a longer subtoken string, we can decrement the counts of its prefixes.", "new_subtoken_strings", "=", "[", "]", "for", "lsub", "in", "range", "(", "len", "(", "len_to_subtoken_strings", ")", "-", "1", ",", "0", ",", "-", "1", ")", ":", "subtoken_strings", "=", "len_to_subtoken_strings", "[", "lsub", "]", "for", "subtoken_string", "in", "subtoken_strings", ":", "count", "=", "subtoken_counts", "[", "subtoken_string", "]", "if", "count", ">=", "min_count", ":", "# Exclude alphabet tokens here, as they must be included later,", "# explicitly, regardless of count.", "if", "subtoken_string", "not", "in", "self", ".", "_alphabet", ":", "new_subtoken_strings", ".", "append", "(", "(", "count", ",", "subtoken_string", ")", ")", "for", "l", "in", "range", "(", "1", ",", "lsub", ")", ":", "subtoken_counts", "[", "subtoken_string", "[", ":", "l", "]", "]", "-=", "count", "# Include the alphabet explicitly to guarantee all strings are encodable.", "new_subtoken_strings", ".", "extend", "(", "(", "subtoken_counts", ".", "get", "(", "a", ",", "0", ")", ",", "a", ")", "for", "a", "in", "self", ".", "_alphabet", ")", "new_subtoken_strings", ".", "sort", "(", "reverse", "=", "True", ")", "# Reinitialize to the candidate vocabulary.", "new_subtoken_strings", "=", "[", "subtoken", "for", "_", ",", "subtoken", "in", "new_subtoken_strings", "]", "if", "reserved_tokens", ":", "escaped_reserved_tokens", "=", "[", "_escape_token", "(", "native_to_unicode", "(", "t", ")", ",", "self", ".", "_alphabet", ")", "for", "t", "in", "reserved_tokens", "]", "new_subtoken_strings", "=", "escaped_reserved_tokens", "+", "new_subtoken_strings", "self", ".", "_init_subtokens_from_list", "(", "new_subtoken_strings", ")", "tf", ".", "logging", ".", "info", "(", "\"vocab_size = %d\"", "%", "self", ".", "vocab_size", ")" ]
Train a SubwordTextEncoder based on a dictionary of word counts. Args: token_counts: a dictionary of Unicode strings to int. min_count: an integer - discard subtokens with lower counts. num_iterations: an integer. how many iterations of refinement. reserved_tokens: List of reserved tokens. The global variable `RESERVED_TOKENS` must be a prefix of `reserved_tokens`. If this argument is `None`, it will use `RESERVED_TOKENS`. max_subtoken_length: Maximum length of a subtoken. If this is not set, then the runtime and memory use of creating the vocab is quadratic in the length of the longest token. If this is set, then it is instead O(max_subtoken_length * length of longest token). Raises: ValueError: if reserved is not 0 or len(RESERVED_TOKENS). In this case, it is not clear what the space is being reserved for, or when it will be filled in.
[ "Train", "a", "SubwordTextEncoder", "based", "on", "a", "dictionary", "of", "word", "counts", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L750-L866
22,399
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder.dump
def dump(self): """Debugging dump of the current subtoken vocabulary.""" subtoken_strings = [(i, s) for s, i in six.iteritems(self._subtoken_string_to_id)] print(u", ".join(u"{0} : '{1}'".format(i, s) for i, s in sorted(subtoken_strings)))
python
def dump(self): """Debugging dump of the current subtoken vocabulary.""" subtoken_strings = [(i, s) for s, i in six.iteritems(self._subtoken_string_to_id)] print(u", ".join(u"{0} : '{1}'".format(i, s) for i, s in sorted(subtoken_strings)))
[ "def", "dump", "(", "self", ")", ":", "subtoken_strings", "=", "[", "(", "i", ",", "s", ")", "for", "s", ",", "i", "in", "six", ".", "iteritems", "(", "self", ".", "_subtoken_string_to_id", ")", "]", "print", "(", "u\", \"", ".", "join", "(", "u\"{0} : '{1}'\"", ".", "format", "(", "i", ",", "s", ")", "for", "i", ",", "s", "in", "sorted", "(", "subtoken_strings", ")", ")", ")" ]
Debugging dump of the current subtoken vocabulary.
[ "Debugging", "dump", "of", "the", "current", "subtoken", "vocabulary", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L872-L877