diff stringlengths 65 26.7k | message stringlengths 7 9.92k |
|---|---|
diff --git a/go/test/endtoend/sharded/shared_keyspace_test.go b/go/test/endtoend/sharded/shared_keyspace_test.go
index <HASH>..<HASH> 100644
--- a/go/test/endtoend/sharded/shared_keyspace_test.go
+++ b/go/test/endtoend/sharded/shared_keyspace_test.go
@@ -259,4 +259,4 @@ func initCluster(shardNames []string, totalTabletsRequired int) {
keyspace.Shards = append(keyspace.Shards, *shard)
}
clusterInstance.Keyspaces = append(clusterInstance.Keyspaces, keyspace)
-}
+}
\ No newline at end of file | Converted sharded test from py to go (#<I>)
* Converted sharded test from py to go |
diff --git a/ui/js/controllers/perf/compare.js b/ui/js/controllers/perf/compare.js
index <HASH>..<HASH> 100644
--- a/ui/js/controllers/perf/compare.js
+++ b/ui/js/controllers/perf/compare.js
@@ -142,6 +142,12 @@ perf.controller('CompareChooserCtrl', [
}
);
};
+
+ // if we have a try push prepopulated, automatically offer a new revision
+ if ($scope.newRevision.length === 12) {
+ $scope.updateNewRevisionTips();
+ $scope.getPreviousRevision();
+ }
});
}]); | Bug <I> - Automatically offer to set base revision when preloading perf compare chooser |
diff --git a/lib/etl/control/destination/database_destination.rb b/lib/etl/control/destination/database_destination.rb
index <HASH>..<HASH> 100644
--- a/lib/etl/control/destination/database_destination.rb
+++ b/lib/etl/control/destination/database_destination.rb
@@ -59,10 +59,10 @@ module ETL #:nodoc:
names = []
values = []
order.each do |name|
- names << name
+ names << "`#{name}`"
values << conn.quote(row[name]) # TODO: this is probably not database agnostic
end
- q = "INSERT INTO #{table_name} (#{names.join(',')}) VALUES (#{values.join(',')})"
+ q = "INSERT INTO `#{table_name}` (#{names.join(',')}) VALUES (#{values.join(',')})"
ETL::Engine.logger.debug("Executing insert: #{q}")
conn.insert(q, "Insert row #{current_row}")
@current_row += 1 | Added backticking to table and column names for database destinations |
diff --git a/tests/test_simplenote.py b/tests/test_simplenote.py
index <HASH>..<HASH> 100644
--- a/tests/test_simplenote.py
+++ b/tests/test_simplenote.py
@@ -8,10 +8,10 @@ from simplenote import Simplenote
class TestSimplenote(unittest.TestCase):
def setUp(self):
- res, status = Simplenote(self.user, self.password).get_note_list()
- [Simplenote(self.user, self.password).delete_note(n["key"]) for n in res]
self.user = "simplenote-test@lordofhosts.de"
self.password = "foobar"
+ res, status = Simplenote(self.user, self.password).get_note_list()
+ [Simplenote(self.user, self.password).delete_note(n["key"]) for n in res]
self.unicode_note = "∮ E⋅da = Q, n → ∞, ∑ f(i) = ∏ g(i), ⎧⎡⎛┌─────┐⎞⎤⎫"
Simplenote(self.user, self.password).add_note("First Note.")
Simplenote(self.user, self.password).add_note("Second Note.") | assign before usage, d'oh |
diff --git a/pyemma/coordinates/data/sparsifier.py b/pyemma/coordinates/data/sparsifier.py
index <HASH>..<HASH> 100644
--- a/pyemma/coordinates/data/sparsifier.py
+++ b/pyemma/coordinates/data/sparsifier.py
@@ -58,5 +58,5 @@ class Sparsifier(Transformer):
% (self.data_producer.dimension() - self.dimension()))
self._varying_indices = np.array(self._varying_indices, dtype=int)
- def _map_array(self, X):
+ def _transform_array(self, X):
return X[:, self._varying_indices] | [coor/sparsifier] adopted to map method refactoring |
diff --git a/lib/instance/login_user_manager.rb b/lib/instance/login_user_manager.rb
index <HASH>..<HASH> 100644
--- a/lib/instance/login_user_manager.rb
+++ b/lib/instance/login_user_manager.rb
@@ -86,8 +86,11 @@ module RightScale
# nil
def add_user(username, uid)
uid = Integer(uid)
-
- %x(sudo useradd -s /bin/bash -u #{uid} -m #{Shellwords.escape(username)})
+
+ useradd = ['usr/bin/useradd', 'usr/sbin/useradd', 'bin/useradd', 'sbin/useradd'].collect { |key| key if File.executable? key }.first
+ raise SystemConflict, "Failed to find a suitable implementation of 'useradd'." unless useradd
+
+ %x(sudo #{useradd} -s /bin/bash -u #{uid} -m #{Shellwords.escape(username)})
case $?.exitstatus
when 0 | Fix: When logging in as the Rightscale user, 'useradd' is not on the path |
diff --git a/src/Sulu/Bundle/MediaBundle/Media/Manager/DefaultMediaManager.php b/src/Sulu/Bundle/MediaBundle/Media/Manager/DefaultMediaManager.php
index <HASH>..<HASH> 100644
--- a/src/Sulu/Bundle/MediaBundle/Media/Manager/DefaultMediaManager.php
+++ b/src/Sulu/Bundle/MediaBundle/Media/Manager/DefaultMediaManager.php
@@ -823,7 +823,7 @@ class DefaultMediaManager implements MediaManagerInterface
* @param Media $media
* @return Media
*/
- protected function addFormatsAndUrl(Media $media)
+ public function addFormatsAndUrl(Media $media)
{
$media->setFormats(
$this->formatManager->getFormats( | changed addFormatsAndUrl function from private to protected |
diff --git a/src/main/java/io/reactivex/Completable.java b/src/main/java/io/reactivex/Completable.java
index <HASH>..<HASH> 100644
--- a/src/main/java/io/reactivex/Completable.java
+++ b/src/main/java/io/reactivex/Completable.java
@@ -1227,6 +1227,7 @@ public abstract class Completable implements CompletableSource {
* @return the throwable if this terminated with an error, null otherwise
* @throws RuntimeException that wraps an InterruptedException if the wait is interrupted
*/
+ @Nullable
@CheckReturnValue
@SchedulerSupport(SchedulerSupport.NONE)
public final Throwable blockingGet() {
@@ -1250,6 +1251,7 @@ public abstract class Completable implements CompletableSource {
* @throws RuntimeException that wraps an InterruptedException if the wait is interrupted or
* TimeoutException if the specified timeout elapsed before it
*/
+ @Nullable
@CheckReturnValue
@SchedulerSupport(SchedulerSupport.NONE)
public final Throwable blockingGet(long timeout, TimeUnit unit) { | Add Nullable annotations for blocking methods in Completable (#<I>) |
diff --git a/core/azure-core/src/main/java/com/azure/core/http/policy/HttpLoggingPolicy.java b/core/azure-core/src/main/java/com/azure/core/http/policy/HttpLoggingPolicy.java
index <HASH>..<HASH> 100644
--- a/core/azure-core/src/main/java/com/azure/core/http/policy/HttpLoggingPolicy.java
+++ b/core/azure-core/src/main/java/com/azure/core/http/policy/HttpLoggingPolicy.java
@@ -145,7 +145,7 @@ public class HttpLoggingPolicy implements HttpPipelinePolicy {
logger.asInfo().log("Response body:\n{}", bodyStr);
logger.asInfo().log("<-- END HTTP");
return bufferedResponse;
- });
+ }).switchIfEmpty(Mono.defer(() -> Mono.just(bufferedResponse)));
} else {
logger.asInfo().log("(body content not logged)");
logger.asInfo().log("<-- END HTTP"); | If body is empty then return publisher emitting response instead of propagating empty publisher (#<I>) |
diff --git a/lib/sprockets/railtie.rb b/lib/sprockets/railtie.rb
index <HASH>..<HASH> 100644
--- a/lib/sprockets/railtie.rb
+++ b/lib/sprockets/railtie.rb
@@ -4,6 +4,7 @@ require 'action_controller/railtie'
require 'active_support/core_ext/module/remove_method'
require 'sprockets'
require 'sprockets/rails/helper'
+require 'sprockets/rails/version'
module Rails
class Application | railtie.rb gets required directly, and needs VERSION.
Fixes #<I>. |
diff --git a/src/Controller/Component/PaginatorComponent.php b/src/Controller/Component/PaginatorComponent.php
index <HASH>..<HASH> 100644
--- a/src/Controller/Component/PaginatorComponent.php
+++ b/src/Controller/Component/PaginatorComponent.php
@@ -35,7 +35,7 @@ use UnexpectedValueException;
*
* @link https://book.cakephp.org/4/en/controllers/components/pagination.html
* @mixin \Cake\Datasource\Paginator
- * @deprecated 4.4.0 Use Cake\Datasource\Paginator directly.
+ * @deprecated 4.4.0 Use Cake\Datasource\Paginator directly. Will be removed in 6.0.
*/
class PaginatorComponent extends Component
{ | Update deprecated tag with <I> removal |
diff --git a/stubilous/server.py b/stubilous/server.py
index <HASH>..<HASH> 100644
--- a/stubilous/server.py
+++ b/stubilous/server.py
@@ -1,3 +1,4 @@
+import logging
from flask import Flask, make_response
@@ -31,6 +32,10 @@ def init_routes(app, routes):
route.status,
route.headers)))
+ @app.errorhandler(404)
+ def incorrect_route(ex):
+ logging.error(ex)
+
def run(config):
app = create_app() | Notify about not found urls |
diff --git a/elasticsearch_dsl/filter.py b/elasticsearch_dsl/filter.py
index <HASH>..<HASH> 100644
--- a/elasticsearch_dsl/filter.py
+++ b/elasticsearch_dsl/filter.py
@@ -53,4 +53,14 @@ class Bool(BoolMixin, Filter):
# register this as Bool for Filter
Filter._bool = Bool
-EMPTY_FILTER = Bool()
+class MatchAll(Filter):
+ name = 'match_all'
+ def __add__(self, other):
+ return other._clone()
+ __and__ = __rand__ = __radd__ = __add__
+
+ def __or__(self, other):
+ return self
+ __ror__ = __or__
+
+EMPTY_FILTER = MatchAll()
diff --git a/test_elasticsearch_dsl/test_search.py b/test_elasticsearch_dsl/test_search.py
index <HASH>..<HASH> 100644
--- a/test_elasticsearch_dsl/test_search.py
+++ b/test_elasticsearch_dsl/test_search.py
@@ -181,7 +181,7 @@ def test_complex_example():
}
},
'post_filter': {
- 'bool': {'must': [{'terms': {'tags': ['prague', 'czech']}}]}
+ 'terms': {'tags': ['prague', 'czech']}
},
'aggs': {
'per_country': { | Don't wrap all filters in Bool, introduce MatchAll instead |
diff --git a/system/Router/Router.php b/system/Router/Router.php
index <HASH>..<HASH> 100644
--- a/system/Router/Router.php
+++ b/system/Router/Router.php
@@ -389,9 +389,9 @@ class Router implements RouterInterface
foreach ($routes as $key => $value)
{
- $key = $key === '/' ? $key : ltrim($key, '/ ');
- $priority = $this->collection->getRoutesOptions($key)['order'] ?? 0;
- $order[$priority][$key] = $value;
+ $key = $key === '/' ? $key : ltrim($key, '/ ');
+ $priority = $this->collection->getRoutesOptions($key)['order'] ?? 0;
+ $order[abs((int)$priority)][$key] = $value;
}
ksort($order);
$routes = array_merge(...$order); | Added casting to integer and absolute value |
diff --git a/javascript/libjoynr-js/src/main/js/joynr/proxy/ProxyOperation.js b/javascript/libjoynr-js/src/main/js/joynr/proxy/ProxyOperation.js
index <HASH>..<HASH> 100644
--- a/javascript/libjoynr-js/src/main/js/joynr/proxy/ProxyOperation.js
+++ b/javascript/libjoynr-js/src/main/js/joynr/proxy/ProxyOperation.js
@@ -95,7 +95,7 @@ function checkArguments(operationArguments) {
try {
if (Constructor && Constructor.checkMembers) {
- Constructor.checkMembers(argumentValue, Typing.checkPropertyIfDefined);
+ Constructor.checkMembers(argumentValue, Typing.checkProperty);
}
} catch (error) {
errors.push(error.message);
@@ -350,4 +350,4 @@ function ProxyOperation(parent, settings, operationName, operationSignatures) {
return Object.freeze(this);
}
-module.exports = ProxyOperation;
\ No newline at end of file
+module.exports = ProxyOperation; | [JS] Use Typing.checkProperty for parameter checks in ProxyOperation
* if a parameter of a method call is a struct, use checkProperty
instead of checkPropertyIfDefined to make sure missing members
are detected.
Change-Id: Idb8a<I>fa9e<I>adc<I>ddf<I>a<I>d<I>b9d<I> |
diff --git a/Net/Gearman/Job.php b/Net/Gearman/Job.php
index <HASH>..<HASH> 100644
--- a/Net/Gearman/Job.php
+++ b/Net/Gearman/Job.php
@@ -30,6 +30,11 @@ if (!defined('NET_GEARMAN_JOB_PATH')) {
define('NET_GEARMAN_JOB_PATH', 'Net/Gearman/Job');
}
+// Define this if you want your Jobs to have a prefix requirement
+if (!defined('NET_GEARMAN_JOB_CLASS_PREFIX')) {
+ define('NET_GEARMAN_JOB_CLASS_PREFIX', 'Net_Gearman_Job_');
+}
+
/**
* Job creation class
*
@@ -63,7 +68,7 @@ abstract class Net_Gearman_Job
{
$file = NET_GEARMAN_JOB_PATH . '/' . $job . '.php';
include_once $file;
- $class = 'Net_Gearman_Job_' . $job;
+ $class = NET_GEARMAN_JOB_CLASS_PREFIX . $job;
if (!class_exists($class)) {
throw new Net_Gearman_Job_Exception('Invalid Job class');
} | Added a define for job class prefix |
diff --git a/lib/strong_migrations/migration.rb b/lib/strong_migrations/migration.rb
index <HASH>..<HASH> 100644
--- a/lib/strong_migrations/migration.rb
+++ b/lib/strong_migrations/migration.rb
@@ -222,7 +222,7 @@ end"
def backfill_code(table, column, default)
model = table.to_s.classify
- "#{model}.in_batches.update_all #{column}: #{default.inspect}"
+ "#{model}.in_batches do |relation| \n relation.update_all #{column}: #{default.inspect}\n sleep(0.1)\n end"
end
def stop!(message, header: "Custom check") | Added throttling to backfill instructions |
diff --git a/src/extensions/reorder-columns/bootstrap-table-reorder-columns.js b/src/extensions/reorder-columns/bootstrap-table-reorder-columns.js
index <HASH>..<HASH> 100644
--- a/src/extensions/reorder-columns/bootstrap-table-reorder-columns.js
+++ b/src/extensions/reorder-columns/bootstrap-table-reorder-columns.js
@@ -129,7 +129,9 @@ $.BootstrapTable = class extends $.BootstrapTable {
})
this.columnsSortOrder = sortOrder
- this.persistReorderColumnsState(this)
+ if (this.options.cookie) {
+ this.persistReorderColumnsState(this)
+ }
const ths = []
const formatters = [] | only execute that function if the cookie plugin is enabled (#<I>) |
diff --git a/pysat/_files.py b/pysat/_files.py
index <HASH>..<HASH> 100644
--- a/pysat/_files.py
+++ b/pysat/_files.py
@@ -152,7 +152,7 @@ class Files(object):
# store write to disk preference
self.write_to_disk = write_to_disk
if self.write_to_disk is False:
- # using blank memory rather than loading from diisk
+ # using blank memory rather than loading from disk
self._previous_file_list = pds.Series([], dtype='a')
self._current_file_list = pds.Series([], dtype='a')
@@ -720,11 +720,11 @@ def parse_delimited_filenames(files, format_str, delimiter):
import collections
# create storage for data to be parsed from filenames
- stored = collections.OrderedDict()
- stored['year'] = []; stored['month'] = []; stored['day'] = [];
- stored['hour'] = []; stored['min'] = []; stored['sec'] = [];
- stored['version'] = []; stored['revision'] = [];
-
+ ordered_keys = ['year', 'month', 'day', 'hour', 'min', 'sec',
+ 'version', 'revision']
+ stored = collections.OrderedDict({kk:list() for kk in ordered_keys})
+
+ # exit early if there are no files
if len(files) == 0:
stored['files'] = []
# include format string as convenience for later functions | Fixed comment typo. Cleaned up generation of OrderedDict |
diff --git a/src/Repositories/PdoRepository.php b/src/Repositories/PdoRepository.php
index <HASH>..<HASH> 100644
--- a/src/Repositories/PdoRepository.php
+++ b/src/Repositories/PdoRepository.php
@@ -61,18 +61,15 @@ abstract class PdoRepository extends Repository
*/
public static function getPdoParamName($columnName)
{
- $alias = $columnName;
- $count = 1;
+ if (isset(self::$pdoParamAliasesUsed[$columnName])) {
+ self::$pdoParamAliasesUsed[$columnName]++;
- while(in_array($alias, self::$pdoParamAliasesUsed)){
- $count++;
+ return $columnName . self::$pdoParamAliasesUsed[$columnName];
+ } else {
+ self::$pdoParamAliasesUsed[$columnName] = 1;
- $alias = $columnName.$count;
+ return $columnName;
}
-
- self::$pdoParamAliasesUsed[] = $alias;
-
- return $alias;
}
/** | Speed improvement by using isset instead of in_array |
diff --git a/src/Http/Validation/SeatSettings.php b/src/Http/Validation/SeatSettings.php
index <HASH>..<HASH> 100644
--- a/src/Http/Validation/SeatSettings.php
+++ b/src/Http/Validation/SeatSettings.php
@@ -54,6 +54,7 @@ class SeatSettings extends FormRequest
$allowed_force_min_mask = implode(',', Seat::$options['force_min_mask']);
$allowed_sso = implode(',', Seat::$options['allow_sso']);
$allowed_tracking = implode(',', Seat::$options['allow_tracking']);
+ $require_activation = implode(',', Seat::$options['require_activation']);
return [
'registration' => 'required|in:' . $allowed_registration,
@@ -63,6 +64,7 @@ class SeatSettings extends FormRequest
'min_corporation_access_mask' => 'required|numeric',
'allow_sso' => 'required|in:' . $allowed_sso,
'allow_tracking' => 'required|in:' . $allowed_tracking,
+ 'require_activation' => 'required|in:' . $require_activation,
];
}
} | Ensure require_activation is validated. |
diff --git a/neurom/core/__init__.py b/neurom/core/__init__.py
index <HASH>..<HASH> 100644
--- a/neurom/core/__init__.py
+++ b/neurom/core/__init__.py
@@ -30,7 +30,6 @@
from .tree import i_chain2 as _chain_neurites
from .tree import Tree as _Tree
-from .types import NeuriteType
def iter_neurites(obj, mapfun=None, filt=None):
diff --git a/neurom/core/section_neuron.py b/neurom/core/section_neuron.py
index <HASH>..<HASH> 100644
--- a/neurom/core/section_neuron.py
+++ b/neurom/core/section_neuron.py
@@ -33,7 +33,7 @@ from collections import defaultdict
from collections import namedtuple
import numpy as np
from neurom.io.hdf5 import H5
-from neurom.core import NeuriteType
+from neurom.core.types import NeuriteType
from neurom.core.tree import Tree, ipreorder, ibifurcation_point
from neurom.core.types import tree_type_checker as is_type
from neurom.core.dataformat import POINT_TYPE | Remove import of core.types.NeuriteTypes in core.__init__.py
This import was causing problems in setup.py, which is run before
the Enum<I> package has been installed. |
diff --git a/src/CachePlugin.php b/src/CachePlugin.php
index <HASH>..<HASH> 100644
--- a/src/CachePlugin.php
+++ b/src/CachePlugin.php
@@ -76,7 +76,6 @@ final class CachePlugin implements Plugin
throw new \InvalidArgumentException('You can\'t provide config option "respect_cache_headers" and "respect_response_cache_directives". '.'Use "respect_response_cache_directives" instead.');
}
-
$optionsResolver = new OptionsResolver();
$this->configureOptions($optionsResolver);
$this->config = $optionsResolver->resolve($config); | Fix CI # 3
- that has been a fine restriction |
diff --git a/orianna/src/main/java/com/merakianalytics/orianna/datapipeline/ImageDownloader.java b/orianna/src/main/java/com/merakianalytics/orianna/datapipeline/ImageDownloader.java
index <HASH>..<HASH> 100644
--- a/orianna/src/main/java/com/merakianalytics/orianna/datapipeline/ImageDownloader.java
+++ b/orianna/src/main/java/com/merakianalytics/orianna/datapipeline/ImageDownloader.java
@@ -15,6 +15,7 @@ import com.merakianalytics.datapipelines.sources.AbstractDataSource;
import com.merakianalytics.datapipelines.sources.Get;
import com.merakianalytics.datapipelines.sources.GetMany;
import com.merakianalytics.orianna.datapipeline.common.HTTPClient;
+import com.merakianalytics.orianna.datapipeline.common.HTTPClient.Configuration;
import com.merakianalytics.orianna.datapipeline.common.HTTPClient.Response;
import com.merakianalytics.orianna.datapipeline.common.Utilities;
import com.merakianalytics.orianna.types.common.OriannaException;
@@ -23,7 +24,9 @@ public class ImageDownloader extends AbstractDataSource {
private final HTTPClient client;
public ImageDownloader() {
- client = new HTTPClient();
+ final Configuration config = new Configuration();
+ config.setHttps(false); // TODO: Make this configurable
+ client = new HTTPClient(config);
}
@Get(BufferedImage.class) | Don't try to use SSL for ddrgon images |
diff --git a/gson/src/main/java/com/google/gson/JsonSerializationVisitor.java b/gson/src/main/java/com/google/gson/JsonSerializationVisitor.java
index <HASH>..<HASH> 100644
--- a/gson/src/main/java/com/google/gson/JsonSerializationVisitor.java
+++ b/gson/src/main/java/com/google/gson/JsonSerializationVisitor.java
@@ -135,10 +135,8 @@ final class JsonSerializationVisitor implements ObjectNavigator.Visitor {
}
public void visitPrimitive(Object obj) {
- if (obj != null) {
- JsonElement json = new JsonPrimitive(obj);
- assignToRoot(json);
- }
+ JsonElement json = obj == null ? JsonNull.createJsonNull() : new JsonPrimitive(obj);
+ assignToRoot(json);
}
private void addAsChildOfObject(Field f, Type fieldType, Object fieldValue) { | Implemented suggestions from code review r<I> by adding a JsonNull for primitives if the value is null. |
diff --git a/cohorts/load.py b/cohorts/load.py
index <HASH>..<HASH> 100644
--- a/cohorts/load.py
+++ b/cohorts/load.py
@@ -273,7 +273,7 @@ class Cohort(object):
def plot_benefit(self, on, col=None, col_equals=None):
plot_col, df = self.plot_init(on, col, col_equals)
original_len = len(df)
- df = df[~df[self.benefit_col].isnull()]
+ df = df[df[self.benefit_col].notnull()]
updated_len = len(df)
df[self.benefit_col] = df[self.benefit_col].apply(bool)
if updated_len < original_len: | not isnull to notnull |
diff --git a/helios-client/src/main/java/com/spotify/helios/client/AuthenticatingHttpConnector.java b/helios-client/src/main/java/com/spotify/helios/client/AuthenticatingHttpConnector.java
index <HASH>..<HASH> 100644
--- a/helios-client/src/main/java/com/spotify/helios/client/AuthenticatingHttpConnector.java
+++ b/helios-client/src/main/java/com/spotify/helios/client/AuthenticatingHttpConnector.java
@@ -210,7 +210,9 @@ public class AuthenticatingHttpConnector implements HttpConnector {
} catch (Exception e) {
// We catch everything because right now the masters do not require authentication.
// So delay reporting errors to the user until the servers return 401 Unauthorized.
- log.debug("Couldn't get identities from ssh-agent", e);
+ log.debug("Unable to get identities from ssh-agent. Note that this might not indicate"
+ + " an actual problem unless your Helios cluster requires authentication"
+ + " for all requests.", e);
}
} | less-scary error message when SSH_AUTH_SOCK is not set
Although we only output the previous message at DEBUG, this log message
can be confusing for users if they have failed tests or another problem
with their client talking to Helios that leads them to think that the
SSH_AUTH_SOCK is the real issue.
So try to wordsmith the message a little bit to indicate it probably
isn't a real issue in all cases. |
diff --git a/lib/Rails/ActiveRecord/Base.php b/lib/Rails/ActiveRecord/Base.php
index <HASH>..<HASH> 100755
--- a/lib/Rails/ActiveRecord/Base.php
+++ b/lib/Rails/ActiveRecord/Base.php
@@ -878,7 +878,7 @@ abstract class Base
static::connection()->executeSql($d);
if (ActiveRecord::lastError()) {
- $this->errors()->addToBase(ActiveRecord::lastError());
+ # The error is logged by Connection.
return false;
}
}
diff --git a/lib/Rails/ActiveRecord/Connection.php b/lib/Rails/ActiveRecord/Connection.php
index <HASH>..<HASH> 100755
--- a/lib/Rails/ActiveRecord/Connection.php
+++ b/lib/Rails/ActiveRecord/Connection.php
@@ -133,7 +133,7 @@ class Connection
$msg = "Error on database query execution\n";
$msg .= $e->getMessage();
- Rails::log()->message($msg);
+ Rails::log()->warning($msg);
}
return $stmt;
} | Changed error handling upon save.
If an error occured in AR\Base when saving a model, the whole array would be added to the errors object of the model, instead of only adding the message of the error.
This would cause an error when trying to implode the error messages with fullMessages(). Furthermore, storing database errors in the model could be dangerous as they could be shown to the users.
Now if an error occurs, it's ignored, because AR\Connection already logs them as warnings. |
diff --git a/panphon/_panphon.py b/panphon/_panphon.py
index <HASH>..<HASH> 100644
--- a/panphon/_panphon.py
+++ b/panphon/_panphon.py
@@ -388,3 +388,7 @@ class FeatureTable(object):
pattern = u''.join(sequence)
regex = re.compile(pattern)
return regex
+
+ def segment_to_vector(self, seg):
+ ft_dict = dict(self.seg_dict[seg])
+ return [ft_dict[name] for name in self.names] | Added segment_to_vector |
diff --git a/mod/wiki/lib.php b/mod/wiki/lib.php
index <HASH>..<HASH> 100644
--- a/mod/wiki/lib.php
+++ b/mod/wiki/lib.php
@@ -438,7 +438,9 @@ function wiki_extend_navigation(navigation_node $navref, $course, $module, $cm)
return false;
}
- $gid = groups_get_activity_group($cm);
+ if (!$gid = groups_get_activity_group($cm)){
+ $gid = 0;
+ }
if (!$subwiki = wiki_get_subwiki_by_group($cm->instance, $gid, $userid)){
return null;
} else { | [MDL-<I>]
Fixing this problem. |
diff --git a/rinoh/float.py b/rinoh/float.py
index <HASH>..<HASH> 100644
--- a/rinoh/float.py
+++ b/rinoh/float.py
@@ -55,6 +55,11 @@ class ImageBase(Flowable):
self.dpi = dpi
self.rotate = rotate
+ @property
+ def filename(self):
+ if isinstance(self.filename_or_file, str):
+ return self.filename_or_file
+
def initial_state(self, container):
return ImageState(self) | Image.filename: for use in selectors |
diff --git a/src/server/pfs/cmds/cmds.go b/src/server/pfs/cmds/cmds.go
index <HASH>..<HASH> 100644
--- a/src/server/pfs/cmds/cmds.go
+++ b/src/server/pfs/cmds/cmds.go
@@ -804,6 +804,7 @@ func putFileHelper(client *client.APIClient, repo, commit, path, source string,
if source == "-" {
limiter.Acquire()
defer limiter.Release()
+ fmt.Println("Reading from STDIN:")
return putFile(os.Stdin)
}
// try parsing the filename as a url, if it is one do a PutFileURL | Print a message if you are putting file through STDIN |
diff --git a/test/tabletmanager.py b/test/tabletmanager.py
index <HASH>..<HASH> 100755
--- a/test/tabletmanager.py
+++ b/test/tabletmanager.py
@@ -7,10 +7,6 @@ warnings.simplefilter("ignore")
import json
import logging
-import os
-import signal
-from subprocess import PIPE
-import threading
import time
import unittest
import urllib
@@ -21,7 +17,6 @@ import utils
import tablet
from mysql_flavor import mysql_flavor
from protocols_flavor import protocols_flavor
-from vtdb import dbexceptions
tablet_62344 = tablet.Tablet(62344)
tablet_62044 = tablet.Tablet(62044) | test: Remove unused imports from tabletmanager.py. |
diff --git a/eventkit/models.py b/eventkit/models.py
index <HASH>..<HASH> 100644
--- a/eventkit/models.py
+++ b/eventkit/models.py
@@ -393,7 +393,7 @@ class AbstractEvent(PolymorphicMPTTModel, AbstractBaseModel):
# `starts` and `end_repeat` are exclusive and will not be included in
# the occurrences.
- if starts and end_repeat:
+ if starts:
missing = rruleset.between(starts, end_repeat)
return missing
return [] | Remove check for `end_repeat` existance as it should always be there. |
diff --git a/lib/modules/crafity.Synchronizer.js b/lib/modules/crafity.Synchronizer.js
index <HASH>..<HASH> 100755
--- a/lib/modules/crafity.Synchronizer.js
+++ b/lib/modules/crafity.Synchronizer.js
@@ -88,15 +88,11 @@ function Synchronizer(finish) {
if (self.listeners("finished").length) {
onfinishCalled = true;
self.emit("finished", err);
- onfinishHandlers.forEach(function (onfinish) {
- onfinish.call(onfinish, err);
- });
}
} else {
try {
callback.apply(callback, arguments);
} catch (err) {
- if (finished) { return; }
finished = true;
lastError = err;
if (self.listeners("finished").length) {
@@ -118,7 +114,7 @@ function Synchronizer(finish) {
this.onfinish = function (finish) {
self.on("finished", finish);
-
+
if (lastError) {
finish(lastError, null);
} else if (finished && !handlerCount) { | Fixed an issue when the synchronizer finishes it stopped to early |
diff --git a/internal/model/queue_test.go b/internal/model/queue_test.go
index <HASH>..<HASH> 100644
--- a/internal/model/queue_test.go
+++ b/internal/model/queue_test.go
@@ -191,7 +191,7 @@ func BenchmarkJobQueuePushPopDone10k(b *testing.B) {
for _, f := range files {
q.Push(f.Name)
}
- for range files {
+ for _ = range files {
n, _ := q.Pop()
q.Done(n)
} | Don't use Go <I> range syntax in queue_test.go, since the listed requirement is Go <I>. |
diff --git a/lxd/firewall/drivers/drivers_xtables.go b/lxd/firewall/drivers/drivers_xtables.go
index <HASH>..<HASH> 100644
--- a/lxd/firewall/drivers/drivers_xtables.go
+++ b/lxd/firewall/drivers/drivers_xtables.go
@@ -418,7 +418,7 @@ func (d Xtables) InstanceClearProxyNAT(projectName string, instanceName string,
}
if len(errs) > 0 {
- return err
+ return fmt.Errorf("Failed to remove proxy NAT rules for %q: %v", deviceName, errs)
}
return nil | lxd/firewall/drivers/drivers/xtables: Improves proxy NAT rule removal errors |
diff --git a/great_expectations/data_context/data_context.py b/great_expectations/data_context/data_context.py
index <HASH>..<HASH> 100644
--- a/great_expectations/data_context/data_context.py
+++ b/great_expectations/data_context/data_context.py
@@ -682,6 +682,7 @@ class BaseDataContext(object):
raise ValueError(
"Unable to load datasource `%s` -- no configuration found or invalid configuration." % datasource_name
)
+ datasource_config = datasourceConfigSchema.load(datasource_config)
datasource = self._build_datasource_from_config(datasource_name, datasource_config)
self._datasources[datasource_name] = datasource
return datasource | Load datasource config before building datasource from config |
diff --git a/lib/binding/collection.js b/lib/binding/collection.js
index <HASH>..<HASH> 100644
--- a/lib/binding/collection.js
+++ b/lib/binding/collection.js
@@ -19,6 +19,10 @@ function Collection() {
var removed = source.splice(idx, count);
this.length = source.length;
api.fire('changed', { removed: removed, removeIdx: idx });
+ },
+
+ get : function (idx) {
+ return source[idx];
}
};
diff --git a/lib/controls/itemsControl.js b/lib/controls/itemsControl.js
index <HASH>..<HASH> 100644
--- a/lib/controls/itemsControl.js
+++ b/lib/controls/itemsControl.js
@@ -55,8 +55,15 @@ function appendChildren(itemsControl) {
ensureCanAppendChildren(itemsControl);
var itemSource = itemsControl._itemSource;
- for (var i = 0; i < itemSource.length; ++i) {
- itemsControl._addItem(itemSource[i]);
+ var i;
+ if (typeof itemSource.get === 'function') {
+ for (i = 0; i < itemSource.length; ++i) {
+ itemsControl._addItem(itemSource.get(i));
+ }
+ } else {
+ for (i = 0; i < itemSource.length; ++i) {
+ itemsControl._addItem(itemSource[i]);
+ }
}
} | Dirty iteration over observable collection
Ideally it should be array like, but I don't want to kill garbage
collection |
diff --git a/main/core/Manager/WorkspaceManager.php b/main/core/Manager/WorkspaceManager.php
index <HASH>..<HASH> 100644
--- a/main/core/Manager/WorkspaceManager.php
+++ b/main/core/Manager/WorkspaceManager.php
@@ -1487,6 +1487,15 @@ class WorkspaceManager
}
}
}
+
+ foreach ($copy->getChildren() as $child) {
+ foreach ($resourceNode->getChildren() as $sourceChild) {
+ if ($child->getName() === $sourceChild->getName()) {
+ $this->duplicateRights($sourceChild, $child, $workspaceRoles);
+ }
+ }
+ }
+
$this->om->flush();
} | [CoreBundle] Right handling for children in creation from model. (#<I>) |
diff --git a/webpack.config.js b/webpack.config.js
index <HASH>..<HASH> 100644
--- a/webpack.config.js
+++ b/webpack.config.js
@@ -30,4 +30,7 @@ module.exports = {
}),
new ForkTsCheckerWebpackPlugin(),
],
-};
+ node: {
+ fs: 'empty'
+ }
+} | Do not try to include Node's fs module
This was caused by rdflib now `require`ing solid-auth-cli, which
Webpack then tried to bundle. Since that code path should never be
hit in production, we can safely exclude it from the bundle. |
diff --git a/lib/dredd_hooks/server.rb b/lib/dredd_hooks/server.rb
index <HASH>..<HASH> 100644
--- a/lib/dredd_hooks/server.rb
+++ b/lib/dredd_hooks/server.rb
@@ -49,10 +49,14 @@ module DreddHooks
transaction = events_handler.handle(event, transaction)
- response = {
- "uuid" => message['uuid'],
- "event" => event,
- "data" => transaction
+ response(message['uuid'], event, transaction)
+ end
+
+ def response(message_uuid, event, transaction)
+ {
+ uuid: message_uuid,
+ event: event,
+ data: transaction,
}.to_json
end | Minor refactor extract method Server#response |
diff --git a/src/Saber.php b/src/Saber.php
index <HASH>..<HASH> 100644
--- a/src/Saber.php
+++ b/src/Saber.php
@@ -7,7 +7,6 @@
namespace Swlib;
-use Swlib\Http\BufferStream;
use Swlib\Http\ContentType;
use Swlib\Http\Exception\HttpExceptionMask;
use Swlib\Http\SwUploadFile;
@@ -19,6 +18,7 @@ use Swlib\Saber\ResponseMap;
use Swlib\Saber\WebSocket;
use Swlib\Util\DataParser;
use Swlib\Util\TypeDetector;
+use function Swlib\Http\stream_for;
class Saber
{
@@ -598,7 +598,7 @@ class Saber
} else {
$options['data'] = null;
}
- $buffer = $options['data'] ? new BufferStream((string)$options['data']) : null;
+ $buffer = $options['data'] ? stream_for((string)$options['data']) : null;
if (isset($buffer)) {
$request->withBody($buffer);
} | use stream_for instead of BufferStream |
diff --git a/master/buildbot/steps/transfer.py b/master/buildbot/steps/transfer.py
index <HASH>..<HASH> 100644
--- a/master/buildbot/steps/transfer.py
+++ b/master/buildbot/steps/transfer.py
@@ -95,6 +95,7 @@ class FileUpload(_TransferBuildStep, WorkerAPICompatMixin):
renderables = [
'masterdest',
'url',
+ 'urlText',
'workersrc',
] | FileUpload: make the urlText renderable
This change allows for making urlText to depend on properties of the
build. |
diff --git a/Gruntfile.js b/Gruntfile.js
index <HASH>..<HASH> 100644
--- a/Gruntfile.js
+++ b/Gruntfile.js
@@ -271,7 +271,7 @@ module.exports = function(grunt) {
// http://twitter.github.com/bower/
grunt.registerTask('component', 'Build the FullCalendar component for the Bower package manager', [
- 'clean:build',
+ 'clean:component',
'submodules',
'uglify', // we want the minified JS in there
'copy:component',
@@ -306,11 +306,8 @@ module.exports = function(grunt) {
/* Clean Up Files
----------------------------------------------------------------------------------------------------*/
- config.clean.build = [
- 'build/out/*',
- 'build/component/*'
- ];
-
+ config.clean.build = 'build/out/*';
+ config.clean.component = 'build/component/*';
config.clean.dist = 'dist/*'; | when cleaning build files, don't clear component by default |
diff --git a/src/User/User.php b/src/User/User.php
index <HASH>..<HASH> 100644
--- a/src/User/User.php
+++ b/src/User/User.php
@@ -540,6 +540,9 @@ class User extends \Hubzero\Database\Relational
// know what user is logged in
App::get('session')->set('user', App::get('user')->getInstance());
$this->guest = false;
+
+ $data = App::get('user')->getInstance()->toArray();
+ \Event::trigger('user.onUserLogin', array($data));
}
}
catch (Exception $e) | [feat] Adding login event trigger after JWT login |
diff --git a/src/components/LoginForm.js b/src/components/LoginForm.js
index <HASH>..<HASH> 100644
--- a/src/components/LoginForm.js
+++ b/src/components/LoginForm.js
@@ -41,7 +41,9 @@ class DefaultLoginForm extends React.Component {
var fields = null;
var socialProviders = null;
- if (data && data.form) {
+ if (err) {
+ fields = defaultFields;
+ } else if (data && data.form) {
fields = data.form.fields;
if (!this.props.hideSocial) {
data.accountStores.forEach((accountStore) => {
diff --git a/src/components/RegistrationForm.js b/src/components/RegistrationForm.js
index <HASH>..<HASH> 100644
--- a/src/components/RegistrationForm.js
+++ b/src/components/RegistrationForm.js
@@ -53,13 +53,13 @@ class DefaultRegistrationForm extends React.Component {
}
];
-
-
UserStore.getRegisterViewData((err, data) => {
var fields = null;
var socialProviders = null;
- if (data && data.form) {
+ if (err) {
+ fields = defaultFields;
+ } else if (data && data.form) {
fields = data.form.fields;
if (!this.props.hideSocial) {
data.accountStores.forEach((accountStore) => { | fix defaultFields fallback for the Login and Registration form |
diff --git a/Lib/fontbakery/checkrunner.py b/Lib/fontbakery/checkrunner.py
index <HASH>..<HASH> 100644
--- a/Lib/fontbakery/checkrunner.py
+++ b/Lib/fontbakery/checkrunner.py
@@ -962,6 +962,21 @@ class Spec:
raise SetupError('Spec fails expected checks test:\n{}'.format(
'\n'.join(message)))
+ def is_numerical_id(checkid):
+ try:
+ int(checkid.split('/')[-1])
+ return True
+ except:
+ return False
+ numerical_check_ids = [c for c in registered_checks if is_numerical_id(c)]
+ if numerical_check_ids:
+ num = len(numerical_check_ids)
+ percentage = 100.0 * num / len(registered_checks)
+ print(f'\nThere still are {num} numerical check IDs. ({percentage:.2f}%)\n'
+ 'They should all be renamed to keyword-based IDs.\n'
+ 'See: https://github.com/googlefonts/fontbakery/issues/2238\n\n')
+ # assert(percentage == 0)
+
def resolve_alias(self, original_name):
name = original_name
seen = set() | computing how many numerical check IDs still remain to...
... be renamed into keyword-based IDs.
This will in the future become a hard-enforcement runtime self-test.
(issue #<I>) |
diff --git a/src/UserData/Extractor/ExtractorInterface.php b/src/UserData/Extractor/ExtractorInterface.php
index <HASH>..<HASH> 100644
--- a/src/UserData/Extractor/ExtractorInterface.php
+++ b/src/UserData/Extractor/ExtractorInterface.php
@@ -56,7 +56,7 @@ interface ExtractorInterface {
/**
* Get service id
*
- * @return string
+ * @return string String, like "google" or "facebook"
* @throws Exception
*/
public function getServiceId(); | Updated phpdoc for ExtractorInterface::getServiceId() |
diff --git a/tests/test_ellipsoid.py b/tests/test_ellipsoid.py
index <HASH>..<HASH> 100644
--- a/tests/test_ellipsoid.py
+++ b/tests/test_ellipsoid.py
@@ -22,6 +22,7 @@ def test_sample():
for i in range(nsim):
R.append(mu.sample()[0])
R = np.array(R)
+ assert (all([mu.contains(_) for _ in R]))
# here I'm checking that all the points are uniformly distributed
# within each ellipsoid
@@ -60,7 +61,7 @@ def test_sample_q():
R.append(x)
break
R = np.array(R)
-
+ assert (all([mu.contains(_) for _ in R]))
# here I'm checking that all the points are uniformly distributed
# within each ellipsoid
for curc in [cen1, cen2]: | add additional ellipsoidal test |
diff --git a/lark/lexer.py b/lark/lexer.py
index <HASH>..<HASH> 100644
--- a/lark/lexer.py
+++ b/lark/lexer.py
@@ -157,12 +157,7 @@ class Token(str):
end_pos: int
def __new__(cls, type_, value, start_pos=None, line=None, column=None, end_line=None, end_column=None, end_pos=None):
- try:
- inst = super(Token, cls).__new__(cls, value)
- except UnicodeDecodeError:
- value = value.decode('latin1')
- inst = super(Token, cls).__new__(cls, value)
-
+ inst = super(Token, cls).__new__(cls, value)
inst.type = type_
inst.start_pos = start_pos
inst.value = value | Remove Py2-related unicode patch |
diff --git a/lib/action_kit_api/event.rb b/lib/action_kit_api/event.rb
index <HASH>..<HASH> 100644
--- a/lib/action_kit_api/event.rb
+++ b/lib/action_kit_api/event.rb
@@ -20,9 +20,6 @@ module ActionKitApi
@required_attrs = [:campaign_id, :creator_id]
@read_only_attrs = [:attendee_count, :host_is_confirmed, :is_full, :is_in_past, :is_open_for_signup, :status_summary]
- # Set us up some defaults
- args[:is_approved] ||= true
-
super
end
diff --git a/lib/action_kit_api/event_campaign.rb b/lib/action_kit_api/event_campaign.rb
index <HASH>..<HASH> 100644
--- a/lib/action_kit_api/event_campaign.rb
+++ b/lib/action_kit_api/event_campaign.rb
@@ -33,11 +33,6 @@ module ActionKitApi
(args[0]).merge!(:campaign_id => self.id)
event = ActionKitApi::Event.new(*args)
- event.save
-
- result = ActionKitApi.connection.call("EventSignup.create", {:event_id => event.id, :user_id => event.creator_id, :status => 'host'})
-
- event
end
# Uses public_search so is subject those limitations | Cleaned up event and event_campaign |
diff --git a/mode/rst/rst.js b/mode/rst/rst.js
index <HASH>..<HASH> 100644
--- a/mode/rst/rst.js
+++ b/mode/rst/rst.js
@@ -501,9 +501,9 @@ CodeMirror.defineMode('rst', function (config, options) {
rx_uri_protocol + rx_uri_domain + rx_uri_path
);
- var rx_strong = /^\*\*[^\*\s](?:[^\*]*[^\*\s])?\*\*/;
- var rx_emphasis = /^\*[^\*\s](?:[^\*]*[^\*\s])?\*/;
- var rx_literal = /^``[^`\s](?:[^`]*[^`\s])``/;
+ var rx_strong = /^\*\*[^\*\s](?:[^\*]*[^\*\s])?\*\*(\s+|$)/;
+ var rx_emphasis = /^[^\*]\*[^\*\s](?:[^\*]*[^\*\s])?\*(\s+|$)/;
+ var rx_literal = /^``[^`\s](?:[^`]*[^`\s])``(\s+|$)/;
var rx_number = /^(?:[\d]+(?:[\.,]\d+)*)/;
var rx_positive = /^(?:\s\+[\d]+(?:[\.,]\d+)*)/; | [rst-mode] Fixed strong, emphasis & literal rST
rST does not allow mixing strong with simple emphasis, further it
requires the stars to have a preceding and trailing whitespaces. Same
for literal text; e.g.
**strong** is correct, but **strong* is not, or
*emphasis* is correct, but *emphasis** is not. |
diff --git a/src/core/propertyTypes.js b/src/core/propertyTypes.js
index <HASH>..<HASH> 100644
--- a/src/core/propertyTypes.js
+++ b/src/core/propertyTypes.js
@@ -41,10 +41,9 @@ module.exports.registerPropertyType = registerPropertyType;
function arrayParse (value) {
if (Array.isArray(value)) { return value; }
- if (value === null || value.length === 0) { return []; }
- return value.split(',').map(function (str) {
- return str.trim();
- });
+ if (!value || typeof value !== 'string') { return []; }
+ return value.split(',').map(trim);
+ function trim (str) { return str.trim(); }
}
function arrayStringify (value) { | For you my friend? This one time. Next time cousin Borat pays a visit.. Borat Devops. |
diff --git a/pandas/tseries/tests/test_timeseries.py b/pandas/tseries/tests/test_timeseries.py
index <HASH>..<HASH> 100644
--- a/pandas/tseries/tests/test_timeseries.py
+++ b/pandas/tseries/tests/test_timeseries.py
@@ -1014,13 +1014,13 @@ class TestTimeSeries(unittest.TestCase):
def test_constructor_int64_nocopy(self):
# #1624
- arr = np.arange(1000)
+ arr = np.arange(1000, dtype=np.int64)
index = DatetimeIndex(arr)
arr[50:100] = -1
self.assert_((index.asi8[50:100] == -1).all())
- arr = np.arange(1000)
+ arr = np.arange(1000, dtype=np.int64)
index = DatetimeIndex(arr, copy=True)
arr[50:100] = -1 | BUG: fix windows/<I>-bit builds |
diff --git a/lib/gdriveWrapper.js b/lib/gdriveWrapper.js
index <HASH>..<HASH> 100644
--- a/lib/gdriveWrapper.js
+++ b/lib/gdriveWrapper.js
@@ -211,7 +211,7 @@ gdriveWrapper.prototype.downloadNewFiles = function(gdriveDirectory, targetDirec
filesBeingDownloaded--;
checkGetNextFiles();
} catch (err) {
- complete(new Error(5, "Failed to append to existing file list"));
+ complete(new Error('Failed to append to existing file list'));
return;
}
} else { | Remove numbers from error creation
Looks like passing a number for the error is not supported
by default, remove |
diff --git a/src/Core42/Model/DefaultModel.php b/src/Core42/Model/DefaultModel.php
index <HASH>..<HASH> 100644
--- a/src/Core42/Model/DefaultModel.php
+++ b/src/Core42/Model/DefaultModel.php
@@ -22,10 +22,11 @@ class DefaultModel extends AbstractModel
$variableName = lcfirst(substr($method, 3));
if (strncasecmp($method, "get", 3) === 0) {
- $return = $this->properties[$variableName];
+ $return = $this->get($variableName);
} elseif (strncasecmp($method, "set", 3) === 0) {
- $return = $this;
- $this->properties[$variableName] = $params[0];
+ $this->properties[$variableName] = $variableName;
+
+ $return = $this->set($variableName, $params[0]);
} else {
throw new \Exception("Method {$method} not found");
}
@@ -44,4 +45,25 @@ class DefaultModel extends AbstractModel
$this->$setter($value);
}
}
+
+ /**
+ * @param array $data
+ */
+ public function hydrate(array $data)
+ {
+ $this->exchangeArray($data);
+ }
+
+ /**
+ * @return array
+ */
+ public function extract()
+ {
+ $array = array();
+ foreach ($this->properties as $variableName) {
+ $array[$variableName] = $this->get($variableName);
+ }
+
+ return $array;
+ }
} | hydrate/extract feature for default model, added support for diff() |
diff --git a/setup.py b/setup.py
index <HASH>..<HASH> 100644
--- a/setup.py
+++ b/setup.py
@@ -120,6 +120,7 @@ setup(
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
+ "Programming Language :: Python :: 3.7",
"Programming Language :: Python :: Implementation :: CPython",
"Topic :: System :: Systems Administration :: Authentication/Directory"],
ext_modules = [ | Claim support for Python <I> |
diff --git a/tests/ignite/engine/test_create_supervised.py b/tests/ignite/engine/test_create_supervised.py
index <HASH>..<HASH> 100644
--- a/tests/ignite/engine/test_create_supervised.py
+++ b/tests/ignite/engine/test_create_supervised.py
@@ -391,6 +391,7 @@ def test_create_supervised_trainer_amp_error(mock_torch_cuda_amp_module):
_test_create_supervised_trainer(amp_mode="amp", scaler=True)
+@pytest.mark.skipif(LooseVersion(torch.__version__) < LooseVersion("1.5.0"), reason="Skip if < 1.5.0")
def test_create_supervised_trainer_scaler_not_amp():
scaler = torch.cuda.amp.GradScaler(enabled=torch.cuda.is_available()) | Fix test_create_supervised on old PyTorch without amp (#<I>)
* Fix test_create_supervised on old PyTorch without amp
* move to torch <I> as minimal version |
diff --git a/wakeonlan.py b/wakeonlan.py
index <HASH>..<HASH> 100644
--- a/wakeonlan.py
+++ b/wakeonlan.py
@@ -108,8 +108,8 @@ if __name__ == "__main__":
elif i is len(args) - 1:
mac_address = args[i]
else:
- sys.exit(help_message)
+ sys.exit(help_message % args[0])
success = send_magic_packet(mac_address, ip_address=ip_address, port=port)
- print("Magic packet sent succesfully." if success else help_message)
+ print("Magic packet sent succesfully." if success else help_message % args[0]) | help message %s is now shown as args[0] |
diff --git a/log.go b/log.go
index <HASH>..<HASH> 100644
--- a/log.go
+++ b/log.go
@@ -70,6 +70,7 @@ func (j *JSONFormatter) Format(e *log.Entry) ([]byte, error) {
FileField: t.Loc(),
FunctionField: t.FuncName(),
})
+ new.Time = e.Time
new.Level = e.Level
new.Message = e.Message
e = new | JSONFormatter time fix
JSONFormatter was always showing "time":"<I>-<I>-<I>T<I>:<I>:<I>Z" |
diff --git a/spec/github/api/callbacks_spec.rb b/spec/github/api/callbacks_spec.rb
index <HASH>..<HASH> 100644
--- a/spec/github/api/callbacks_spec.rb
+++ b/spec/github/api/callbacks_spec.rb
@@ -28,14 +28,14 @@ end
describe Github::API, '#callbacks' do
it "retrieves only public api methods" do
- expect(ApiTest.request_methods.to_a).to eq([
+ expect(ApiTest.request_methods.to_a - [
'list',
'list_with_callback_apitest',
'list_without_callback_apitest',
'get',
'get_with_callback_apitest',
'get_without_callback_apitest'
- ])
+ ]).to be_empty
end
it "execute before callback" do | Fix spec result dependency for jruby. |
diff --git a/salt/utils/iam.py b/salt/utils/iam.py
index <HASH>..<HASH> 100644
--- a/salt/utils/iam.py
+++ b/salt/utils/iam.py
@@ -13,7 +13,7 @@ import time
import requests
import pprint
from six.moves import range
-import six
+import salt.utils.six as six
log = logging.getLogger(__name__) | Replaced import six in file /salt/utils/iam.py |
diff --git a/lib/veewee/provider/core/box/issh.rb b/lib/veewee/provider/core/box/issh.rb
index <HASH>..<HASH> 100644
--- a/lib/veewee/provider/core/box/issh.rb
+++ b/lib/veewee/provider/core/box/issh.rb
@@ -25,7 +25,7 @@ module Veewee
def ssh_commandline_options(options)
command_options = [
- "-q", #Suppress warning messages
+ #"-q", #Suppress warning messages
# "-T", #Pseudo-terminal will not be allocated because stdin is not a terminal.
"-p #{ssh_options[:port]}",
"-o UserKnownHostsFile=/dev/null", | remove the -q for ssh , to see warning and errors |
diff --git a/test/grpc_test.go b/test/grpc_test.go
index <HASH>..<HASH> 100644
--- a/test/grpc_test.go
+++ b/test/grpc_test.go
@@ -22,7 +22,8 @@ func TestGrpc(t *testing.T) {
}
defer g.Stop()
- ctx, _ := context.WithTimeout(context.Background(), 5*time.Second)
+ ctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)
+ defer cancel()
conn, err := grpc.DialContext(ctx, tcp, grpc.WithInsecure(), grpc.WithBlock())
if err != nil {
t.Fatalf("Expected no error but got: %s", err) | Fix grpc test vet warning (#<I>)
This fixes the vet warning: the cancel function returned by
context.WithTimeout should be called, not discarded, to avoid a context
leak. |
diff --git a/bin/firenze.js b/bin/firenze.js
index <HASH>..<HASH> 100755
--- a/bin/firenze.js
+++ b/bin/firenze.js
@@ -16,6 +16,7 @@ if (command === 'migration') {
.command('list', 'List migrations')
.command('current', 'Show current migration version')
.command('run [name]', 'Run specific migration by name')
+ .command('runAll', 'Run all pending migrations')
.command('rollback [name]', 'Roll back specific migration by name')
.describe('db', 'Path to database file')
.describe('db-name', 'Name of database instance (if path exports multiple named instances)')
@@ -102,6 +103,17 @@ if (command === 'migration') {
.finally(function () {
db.close();
});
+ } else if (subCommand === 'runAll') {
+ migration.runAll()
+ .then(function () {
+ console.log('Successfully ran all migrations.');
+ })
+ .catch(function (error) {
+ console.log(chalk.bgRed('Error:') + ' ' + error);
+ })
+ .finally(function () {
+ db.close();
+ });
} else if (subCommand === 'rollback') {
if (typeof argv._[2] === 'undefined') {
console.log(chalk.bgRed('Error:') + ' No name given'); | migrations: run all pending migrations via CLI. |
diff --git a/cmd/syncthing/memsize_solaris.go b/cmd/syncthing/memsize_solaris.go
index <HASH>..<HASH> 100644
--- a/cmd/syncthing/memsize_solaris.go
+++ b/cmd/syncthing/memsize_solaris.go
@@ -1,3 +1,5 @@
+// +build solaris
+
package main
import ( | Avoid build error in Go<I> |
diff --git a/src/draw/handler/Draw.Circle.js b/src/draw/handler/Draw.Circle.js
index <HASH>..<HASH> 100644
--- a/src/draw/handler/Draw.Circle.js
+++ b/src/draw/handler/Draw.Circle.js
@@ -15,7 +15,7 @@ L.Draw.Circle = L.Draw.SimpleShape.extend({
clickable: true
},
showRadius: true,
- metric: true, // Whether to use the metric meaurement system or imperial
+ metric: true, // Whether to use the metric measurement system or imperial
feet: true // When not metric, use feet instead of yards for display
},
diff --git a/src/draw/handler/Draw.Polyline.js b/src/draw/handler/Draw.Polyline.js
index <HASH>..<HASH> 100644
--- a/src/draw/handler/Draw.Polyline.js
+++ b/src/draw/handler/Draw.Polyline.js
@@ -30,7 +30,7 @@ L.Draw.Polyline = L.Draw.Feature.extend({
fill: false,
clickable: true
},
- metric: true, // Whether to use the metric meaurement system or imperial
+ metric: true, // Whether to use the metric measurement system or imperial
feet: true, // When not metric, to use feet instead of yards for display.
showLength: true, // Whether to display distance in the tooltip
zIndexOffset: 2000 // This should be > than the highest z-index any map layers | [Typo] s/meaurement/measurement |
diff --git a/elasticsearch/connection/base.py b/elasticsearch/connection/base.py
index <HASH>..<HASH> 100644
--- a/elasticsearch/connection/base.py
+++ b/elasticsearch/connection/base.py
@@ -42,7 +42,7 @@ class Connection(object):
return json.dumps(json.loads(data), sort_keys=True, indent=2, separators=(',', ': '))
except (ValueError, TypeError):
# non-json data or a bulk request
- return repr(data)
+ return data
logger.info(
'%s %s [status:%s request:%.3fs]', method, full_url, | Make sure bulk requests are logged correctly |
diff --git a/lib/index.js b/lib/index.js
index <HASH>..<HASH> 100644
--- a/lib/index.js
+++ b/lib/index.js
@@ -43,7 +43,7 @@ app.conditional = function*(next) {
app.subdomainrouter=function * (next){
let origin = this.request.get("origin").replace("http://","").split(":")[0].split(".");
if(origin.length >=3 ){
- let reqo = thi.request.get("origin").replace(origin[0]+".","");
+ let reqo = this.request.get("origin").replace(origin[0]+".","");
if(app.routers[origin[0]]!==undefined && reqo===server.hostname){
let port = this.request.host.split(":")[1];
port = port ? ":"+port:""; | [ci skip] Fixed bug with production server |
diff --git a/packages/ember-runtime/lib/mixins/enumerable.js b/packages/ember-runtime/lib/mixins/enumerable.js
index <HASH>..<HASH> 100644
--- a/packages/ember-runtime/lib/mixins/enumerable.js
+++ b/packages/ember-runtime/lib/mixins/enumerable.js
@@ -285,7 +285,7 @@ export default Mixin.create({
@method getEach
@param {String} key name of the property
@return {Array} The mapped array.
- @private
+ @public
*/
getEach: aliasMethod('mapBy'), | [DOC Release] Mark Enumerable.getEach() as public
[ci skip] |
diff --git a/lib/ffprobe.js b/lib/ffprobe.js
index <HASH>..<HASH> 100644
--- a/lib/ffprobe.js
+++ b/lib/ffprobe.js
@@ -11,7 +11,8 @@ function parseFfprobeOutput(out) {
var lines = out.split(/\r\n|\r|\n/);
var data = {
streams: [],
- format: {}
+ format: {},
+ chapters: []
};
function parseBlock(name) {
@@ -46,6 +47,9 @@ function parseFfprobeOutput(out) {
if (line.match(/^\[stream/i)) {
var stream = parseBlock('stream');
data.streams.push(stream);
+ } else if (line.match(/^\[chapter/i)) {
+ var chapter = parseBlock('chapter');
+ data.chapters.push(chapter);
} else if (line.toLowerCase() === '[format]') {
data.format = parseBlock('format');
}
@@ -105,6 +109,7 @@ module.exports = function(proto) {
break;
}
+
if (index === null) {
if (!this._currentInput) {
return callback(new Error('No input specified')); | ffprobe - parsing of chapters metadata |
diff --git a/wsgidav/http_authenticator.py b/wsgidav/http_authenticator.py
index <HASH>..<HASH> 100644
--- a/wsgidav/http_authenticator.py
+++ b/wsgidav/http_authenticator.py
@@ -217,6 +217,12 @@ class HTTPAuthenticator(BaseMiddleware):
elif authmethod == "basic" and self._acceptbasic:
return self.authBasicAuthRequest(environ, start_response)
+ # The requested auth method is not supported.
+ elif self._defaultdigest and self._acceptdigest:
+ return self.sendDigestAuthResponse(environ, start_response)
+ elif self._acceptbasic:
+ return self.sendBasicAuthResponse(environ, start_response)
+
util.log(
"HTTPAuthenticator: respond with 400 Bad request; Auth-Method: %s" % authmethod) | ISSUE <I>: Return <I> when auth method is not supported (#<I>)
Thanks. |
diff --git a/server/irc/channel.js b/server/irc/channel.js
index <HASH>..<HASH> 100644
--- a/server/irc/channel.js
+++ b/server/irc/channel.js
@@ -4,7 +4,9 @@ var util = require('util'),
var IrcChannel = function(irc_connection, name) {
this.irc_connection = irc_connection;
- this.name = name;
+
+ // Lowercase the channel name so we don't run into case-sensitive issues
+ this.name = name.toLowerCase();
this.members = [];
this.ban_list_buffer = []; | Channel case-sensitivity issues. Main cause of "Cannot call method 'clientEvent' of undefined" |
diff --git a/fusesoc/capi2/core.py b/fusesoc/capi2/core.py
index <HASH>..<HASH> 100644
--- a/fusesoc/capi2/core.py
+++ b/fusesoc/capi2/core.py
@@ -48,7 +48,7 @@ class String(str):
return t.expr
else:
return []
- word = Word(alphanums+':>.[]_-,=~/')
+ word = Word(alphanums+':<>.[]_-,=~/')
conditional = Forward()
conditional << (Optional("!")("negate") + word("cond") + Suppress('?') + Suppress('(') + OneOrMore(conditional ^ word)("expr") + Suppress(')')).setParseAction(cb_conditional)
#string = (function ^ word) | Allow < character in Word parsing.
This allows dependencies to include restrictions on less than a specific version |
diff --git a/test/getConfig.spec.js b/test/getConfig.spec.js
index <HASH>..<HASH> 100644
--- a/test/getConfig.spec.js
+++ b/test/getConfig.spec.js
@@ -74,8 +74,10 @@ afterAll(done => {
beforeEach(done => {
rimraf(outputFolder, () => {
- mkdirSync(outputFolder);
- done();
+ setTimeout(() => {
+ mkdirSync(outputFolder);
+ done();
+ }, 50);
});
}); | Odd test result on Travis, I suspect rimraf is calling the callback before it's really finished, it happened before |
diff --git a/lib/OpenLayers/Layer/WMS/Untiled.js b/lib/OpenLayers/Layer/WMS/Untiled.js
index <HASH>..<HASH> 100644
--- a/lib/OpenLayers/Layer/WMS/Untiled.js
+++ b/lib/OpenLayers/Layer/WMS/Untiled.js
@@ -107,8 +107,6 @@ OpenLayers.Layer.WMS.Untiled.prototype =
*/
moveTo:function(bounds, zoomChanged, minor) {
-
-
if (!minor) {
if (bounds == null) {
@@ -132,7 +130,7 @@ OpenLayers.Layer.WMS.Untiled.prototype =
// update div
var img = this.imgDiv;
- if (this.transparent) {
+ if (this.params.TRANSPARENT == 'true') {
OpenLayers.Util.modifyAlphaImageDiv(this.imgDiv,
null,
pos,
@@ -206,7 +204,7 @@ OpenLayers.Layer.WMS.Untiled.prototype =
var pos = this.map.getLayerPxFromViewPortPx(tl);
//create div
- if (this.transparent) {
+ if (this.params.TRANSPARENT == 'true') {
this.imgDiv = OpenLayers.Util.createAlphaImageDiv(null,
pos,
size, | check for transparency was not correct
git-svn-id: <URL> |
diff --git a/src/Models/MetadataTrait.php b/src/Models/MetadataTrait.php
index <HASH>..<HASH> 100644
--- a/src/Models/MetadataTrait.php
+++ b/src/Models/MetadataTrait.php
@@ -16,6 +16,7 @@ trait MetadataTrait
'integer' => EdmPrimitiveType::INT32,
'string' => EdmPrimitiveType::STRING,
'datetime' => EdmPrimitiveType::DATETIME,
+ 'float' => EdmPrimitiveType::SINGLE,
'decimal' => EdmPrimitiveType::DECIMAL,
'text' => EdmPrimitiveType::STRING,
'boolean' => EdmPrimitiveType::BOOLEAN, | added float to the type mapping (#<I>) |
diff --git a/test/Labrador/Test/Unit/RendererTest.php b/test/Labrador/Test/Unit/RendererTest.php
index <HASH>..<HASH> 100644
--- a/test/Labrador/Test/Unit/RendererTest.php
+++ b/test/Labrador/Test/Unit/RendererTest.php
@@ -29,9 +29,8 @@ class RendererTest extends UnitTestCase {
$renderer = $this->getRenderer();
$expected = <<<TEXT
partial
-
TEXT;
- $actual = $renderer->renderPartial('partial');
+ $actual = trim($renderer->renderPartial('partial'), PHP_EOL);
$this->assertSame($expected, $actual);
}
@@ -42,7 +41,7 @@ layout
foobar
TEXT;
- $actual = $renderer->renderPartial('layout', ['_content' => 'foobar']);
+ $actual = trim($renderer->renderPartial('layout', ['_content' => 'foobar']), PHP_EOL);
$this->assertSame($expected, $actual);
}
@@ -54,7 +53,7 @@ layout
partial
TEXT;
- $actual = $renderer->render('partial');
+ $actual = trim($renderer->render('partial'), PHP_EOL);
$this->assertSame($expected, $actual);
} | trimming up content on tests
some editors can be setup to automagically include blank lines and some
don’t. ultimately the presence of a new line at the end of the content
is of little consequence and shouldn’t impact the confidence of the
tests. |
diff --git a/test/resources/6_setup.js b/test/resources/6_setup.js
index <HASH>..<HASH> 100644
--- a/test/resources/6_setup.js
+++ b/test/resources/6_setup.js
@@ -27,7 +27,7 @@ describe('Resource:', function() {
});
});
- it.only('should generate the settings-file', function(done) {
+ it('should generate the settings-file', function(done) {
client.post('/setup/commit', function(err, req, res) {
assert.ifError(err);
assert.equal(res.statusCode, 201); | Oups! Run all the tests |
diff --git a/crispy_forms/tests/test_layout.py b/crispy_forms/tests/test_layout.py
index <HASH>..<HASH> 100644
--- a/crispy_forms/tests/test_layout.py
+++ b/crispy_forms/tests/test_layout.py
@@ -270,7 +270,6 @@ def test_column_has_css_classes(settings):
c = Context({'form': form, 'form_helper': form_helper})
html = template.render(c)
- print(html)
if settings.CRISPY_TEMPLATE_PACK == 'uni_form':
assert html.count('formColumn') == 1 | Remove print statement from test file. (#<I>) |
diff --git a/server/workflow/workflow_server.go b/server/workflow/workflow_server.go
index <HASH>..<HASH> 100644
--- a/server/workflow/workflow_server.go
+++ b/server/workflow/workflow_server.go
@@ -299,9 +299,6 @@ func (s *workflowServer) DeleteWorkflow(ctx context.Context, req *workflowpkg.Wo
if err != nil {
return nil, err
}
- if wf.Finalizers != nil && !req.Force {
- return nil, fmt.Errorf("%s has finalizer. Use argo delete --force to delete this workflow", wf.Name)
- }
if req.Force {
_, err := auth.GetWfClient(ctx).ArgoprojV1alpha1().Workflows(wf.Namespace).Patch(ctx, wf.Name, types.MergePatchType, []byte("{\"metadata\":{\"finalizers\":null}}"), metav1.PatchOptions{})
if err != nil { | fix: removed error check which prevented deleting successful artGC wfs. (#<I>)
fix: removed error check which prevented deleting successful artGC workflows |
diff --git a/lib/semantic_logger/appender/syslog.rb b/lib/semantic_logger/appender/syslog.rb
index <HASH>..<HASH> 100644
--- a/lib/semantic_logger/appender/syslog.rb
+++ b/lib/semantic_logger/appender/syslog.rb
@@ -164,7 +164,8 @@ module SemanticLogger
def reopen
case @protocol
when :syslog
- ::Syslog.open(application, options, facility)
+ method = ::Syslog.opened? ? :reopen : :open
+ ::Syslog.send(method, application, options, facility)
when :tcp
# Use the local logger for @remote_syslog so errors with the remote logger can be recorded locally.
@tcp_client_options[:logger] = logger | Now syslog appender calls ::Syslog.reopen if it is already opened. |
diff --git a/test/Formatter/SignatureFormatterTest.php b/test/Formatter/SignatureFormatterTest.php
index <HASH>..<HASH> 100644
--- a/test/Formatter/SignatureFormatterTest.php
+++ b/test/Formatter/SignatureFormatterTest.php
@@ -13,6 +13,7 @@ namespace Psy\Test\Formatter;
use Psy\Formatter\SignatureFormatter;
use Psy\Reflection\ReflectionClassConstant;
+use Psy\Reflection\ReflectionConstant_;
class SignatureFormatterTest extends \PHPUnit\Framework\TestCase
{
@@ -68,6 +69,18 @@ class SignatureFormatterTest extends \PHPUnit\Framework\TestCase
new \ReflectionMethod('Psy\Test\Formatter\Fixtures\BoringTrait', 'boringMethod'),
'public function boringMethod($one = 1)',
],
+ [
+ new ReflectionConstant_('E_ERROR'),
+ 'define("E_ERROR", 1)',
+ ],
+ [
+ new ReflectionConstant_('PHP_VERSION'),
+ 'define("PHP_VERSION", "' . PHP_VERSION . '")',
+ ],
+ [
+ new ReflectionConstant_('__LINE__'),
+ 'define("__LINE__", null)', // @todo show this as `unknown` in red or something?
+ ],
];
} | Add test coverage for constant signature formatting. |
diff --git a/tags_include.go b/tags_include.go
index <HASH>..<HASH> 100644
--- a/tags_include.go
+++ b/tags_include.go
@@ -1,6 +1,7 @@
package pongo2
import (
+ "errors"
"path/filepath"
)
@@ -40,6 +41,11 @@ func (node *tagIncludeNode) Execute(ctx *ExecutionContext) (string, error) {
if err != nil {
return "", err
}
+
+ if filename.String() == "" {
+ return "", errors.New("Filename for 'include'-tag evaluated to an empty string.")
+ }
+
// Get include-filename relative to the including-template directory
including_dir := filepath.Dir(ctx.template.name)
included_filename := filepath.Join(including_dir, filename.String()) | include-tag lazy filename validation added. |
diff --git a/contribs/src/main/java/com/netflix/conductor/contribs/AMQPModule.java b/contribs/src/main/java/com/netflix/conductor/contribs/AMQPModule.java
index <HASH>..<HASH> 100644
--- a/contribs/src/main/java/com/netflix/conductor/contribs/AMQPModule.java
+++ b/contribs/src/main/java/com/netflix/conductor/contribs/AMQPModule.java
@@ -55,7 +55,7 @@ public class AMQPModule extends AbstractModule {
}
@ProvidesIntoMap
- @StringMapKey("amqp")
+ @StringMapKey("amqp_queue")
@Singleton
@Named(EVENT_QUEUE_PROVIDERS_QUALIFIER)
public EventQueueProvider getAMQQueueEventQueueProvider(Configuration config) { | Fix issue in which a queue in an event-handler is defined as "amqp:..." and no events are consumed. (#<I>)
Rename StringMapKey for amqp queue from "amqp" to amqp_queue"
Issue: <URL> |
diff --git a/commands.go b/commands.go
index <HASH>..<HASH> 100644
--- a/commands.go
+++ b/commands.go
@@ -730,11 +730,11 @@ func cmdEnv(c *cli.Context) {
switch userShell {
case "fish":
- fmt.Printf("%s\n\nset -x DOCKER_TLS_VERIFY 1;\nset -x DOCKER_CERT_PATH %q;\nset -x DOCKER_HOST %s;\n",
- usageHint, cfg.machineDir, dockerHost)
+ fmt.Printf("set -x DOCKER_TLS_VERIFY 1;\nset -x DOCKER_CERT_PATH %q;\nset -x DOCKER_HOST %s;\n\n%s\n",
+ cfg.machineDir, dockerHost, usageHint)
default:
- fmt.Printf("%s\nexport DOCKER_TLS_VERIFY=1\nexport DOCKER_CERT_PATH=%q\nexport DOCKER_HOST=%s\n",
- usageHint, cfg.machineDir, dockerHost)
+ fmt.Printf("export DOCKER_TLS_VERIFY=1\nexport DOCKER_CERT_PATH=%q\nexport DOCKER_HOST=%s\n\n%s\n",
+ cfg.machineDir, dockerHost, usageHint)
}
} | Moved env usage hint to end of output to avoid shell expansion issue |
diff --git a/core/src/main/java/net/kuujo/vertigo/cluster/data/MapEvent.java b/core/src/main/java/net/kuujo/vertigo/cluster/data/MapEvent.java
index <HASH>..<HASH> 100644
--- a/core/src/main/java/net/kuujo/vertigo/cluster/data/MapEvent.java
+++ b/core/src/main/java/net/kuujo/vertigo/cluster/data/MapEvent.java
@@ -122,4 +122,27 @@ public class MapEvent<K, V> implements JsonSerializable {
return (V) value;
}
+ @Override
+ public int hashCode() {
+ int hashCode = 23;
+ hashCode = 37 * hashCode + type.hashCode();
+ hashCode = 37 * hashCode + key.hashCode();
+ hashCode = 37 * hashCode + value.hashCode();
+ return hashCode;
+ }
+
+ @Override
+ public String toString() {
+ return String.format("MapEvent(%s)[%s:%s]", type, key, value);
+ }
+
+ @Override
+ public boolean equals(Object object) {
+ if (!(object instanceof MapEvent)) {
+ return false;
+ }
+ MapEvent<?, ?> event = (MapEvent<?, ?>) object;
+ return event.type.equals(type) && event.key.equals(key) && event.value.equals(value);
+ }
+
} | Add hash support to MapEvent. |
diff --git a/URLNormalizer.php b/URLNormalizer.php
index <HASH>..<HASH> 100644
--- a/URLNormalizer.php
+++ b/URLNormalizer.php
@@ -105,7 +105,12 @@ class URLNormalizer {
$fragment = '#' . $this->fragment;
}
- return $this->scheme . $this->host . $this->port . $this->user . $this->pass . $this->path . $query . $fragment;
+ $port = '';
+ if ( $this->port ) {
+ $port = ':' . $this->port;
+ }
+
+ return $this->scheme . $this->host . $port . $this->user . $this->pass . $this->path . $query . $fragment;
}
/**
diff --git a/URLNormalizerTest.php b/URLNormalizerTest.php
index <HASH>..<HASH> 100644
--- a/URLNormalizerTest.php
+++ b/URLNormalizerTest.php
@@ -130,4 +130,11 @@ class URLNormalizerTest extends PHPUnit_Framework_TestCase
$this->fixture->setUrl( $url );
$this->assertEquals( $url, $this->fixture->normalize() );
}
+
+ public function testPortNumbersArePreserved() {
+ $url = 'http://example.com:81/index.html';
+
+ $this->fixture->setUrl( $url );
+ $this->assertEquals( $url, $this->fixture->normalize() );
+ }
} | Port numbers are now correctly preserved after normalisation |
diff --git a/lib/sass/script/string.rb b/lib/sass/script/string.rb
index <HASH>..<HASH> 100644
--- a/lib/sass/script/string.rb
+++ b/lib/sass/script/string.rb
@@ -46,6 +46,8 @@ module Sass::Script
if type == :identifier
if context == :equals && Sass::SCSS::RX.escape_ident(self.value).include?(?\\)
return "unquote(#{Sass::Script::String.new(self.value, :string).to_sass})"
+ elsif context == :equals && self.value.size == 0
+ return "\"\""
end
return self.value
end | [SCSS] Handle empty strings in variable assigment when converting to scss. |
diff --git a/src/Drupal/Driver/Cores/Drupal6.php b/src/Drupal/Driver/Cores/Drupal6.php
index <HASH>..<HASH> 100644
--- a/src/Drupal/Driver/Cores/Drupal6.php
+++ b/src/Drupal/Driver/Cores/Drupal6.php
@@ -29,8 +29,10 @@ class Drupal6 implements CoreInterface {
}
// Bootstrap Drupal.
+ $current_path = getcwd();
chdir(DRUPAL_ROOT);
drupal_bootstrap(DRUPAL_BOOTSTRAP_FULL);
+ chdir($current_path);
}
/**
diff --git a/src/Drupal/Driver/Cores/Drupal8.php b/src/Drupal/Driver/Cores/Drupal8.php
index <HASH>..<HASH> 100644
--- a/src/Drupal/Driver/Cores/Drupal8.php
+++ b/src/Drupal/Driver/Cores/Drupal8.php
@@ -29,8 +29,10 @@ class Drupal8 implements CoreInterface {
}
// Bootstrap Drupal.
+ $current_path = getcwd();
chdir(DRUPAL_ROOT);
drupal_bootstrap(DRUPAL_BOOTSTRAP_FULL);
+ chdir($current_path);
}
/** | Follow-up to issue #<I>. |
diff --git a/src/components/radioButton/demoMultiColumn/script.js b/src/components/radioButton/demoMultiColumn/script.js
index <HASH>..<HASH> 100644
--- a/src/components/radioButton/demoMultiColumn/script.js
+++ b/src/components/radioButton/demoMultiColumn/script.js
@@ -5,13 +5,13 @@ angular
self.contacts = [{
'id': 1,
- 'fullName': 'Maria Guadalupe',
+ 'fullName': 'María Guadalupe',
'lastName': 'Guadalupe',
'title': "CEO, Found"
}, {
'id': 2,
- 'fullName': 'Gabriel García Marquéz',
- 'lastName': 'Marquéz',
+ 'fullName': 'Gabriel García Márquez',
+ 'lastName': 'Márquez',
'title': "VP Sales & Marketing"
}, {
'id': 3, | docs(radiobutton): correct two misspellings on demo (#<I>) |
diff --git a/lib/travis/cli/console.rb b/lib/travis/cli/console.rb
index <HASH>..<HASH> 100644
--- a/lib/travis/cli/console.rb
+++ b/lib/travis/cli/console.rb
@@ -23,7 +23,9 @@ module Travis
require 'pry'
true
rescue LoadError
- $stderr.puts 'You need to install pry to use Travis CLI console.'
+ $stderr.puts 'You need to install pry to use Travis CLI console. Try'
+ $stderr.puts
+ $stderr.puts '$ (sudo) gem install pry'
false
end
end | Add message about how to install Pry (#<I>) |
diff --git a/setup.py b/setup.py
index <HASH>..<HASH> 100644
--- a/setup.py
+++ b/setup.py
@@ -12,7 +12,7 @@ def read(fname):
setup(
name="qds_sdk",
- version="1.9.0",
+ version="unreleased",
author="Qubole",
author_email="dev@qubole.com",
description=("Python SDK for coding to the Qubole Data Service API"), | fix: usr: setup.py should indicate unreleased branch as version==unreleased
when using qds-sdk as a dependency - pip and setuptools look at the
version in setup() to find the stable version. if this does not match
with the one specified by install_requires then pip/setuptools fail and
install the stable version instead |
diff --git a/docs/source/conf.py b/docs/source/conf.py
index <HASH>..<HASH> 100644
--- a/docs/source/conf.py
+++ b/docs/source/conf.py
@@ -166,11 +166,11 @@ napoleon_use_rtype = False
add_module_names = False
-html_context = {
- 'css_files': [
- '_static/rtd_overrides.css', # overrides for wide tables in RTD theme
- ],
-}
+# html_context = {
+# 'css_files': [
+# '_static/rtd_overrides.css', # overrides for wide tables in RTD theme
+# ],
+# }
# Theme options are theme-specific and customize the look and feel of a theme
# further. For a list of options available for each theme, see the | DOC hm, this seems to break rtfd |
diff --git a/activerecord/lib/active_record/connection_adapters/abstract/database_statements.rb b/activerecord/lib/active_record/connection_adapters/abstract/database_statements.rb
index <HASH>..<HASH> 100644
--- a/activerecord/lib/active_record/connection_adapters/abstract/database_statements.rb
+++ b/activerecord/lib/active_record/connection_adapters/abstract/database_statements.rb
@@ -174,14 +174,9 @@ module ActiveRecord
options.assert_valid_keys :requires_new, :joinable
last_transaction_joinable = @transaction_joinable
- if options.has_key?(:joinable)
- @transaction_joinable = options[:joinable]
- else
- @transaction_joinable = true
- end
- requires_new = options[:requires_new] || !last_transaction_joinable
-
- transaction_open = false
+ @transaction_joinable = options.fetch(:joinable, true)
+ requires_new = options[:requires_new] || !last_transaction_joinable
+ transaction_open = false
begin
if requires_new || open_transactions == 0 | use Hash#fetch to eliminate conditional |
diff --git a/src/foremast/elb/format_listeners.py b/src/foremast/elb/format_listeners.py
index <HASH>..<HASH> 100644
--- a/src/foremast/elb/format_listeners.py
+++ b/src/foremast/elb/format_listeners.py
@@ -134,6 +134,7 @@ def format_cert_name(env='', account='', region='', certificate=None):
Args:
env (str): Account environment name
account (str): Account number for ARN
+ region (str): AWS Region.
certificate (str): Name of SSL certificate
Returns:
@@ -165,6 +166,7 @@ def generate_custom_cert_name(env='', region='', account='', certificate=None):
Args:
env (str): Account environment name
+ region (str): AWS Region.
account (str): Account number for ARN.
certificate (str): Name of SSL certificate.
diff --git a/tests/elb/test_elb.py b/tests/elb/test_elb.py
index <HASH>..<HASH> 100644
--- a/tests/elb/test_elb.py
+++ b/tests/elb/test_elb.py
@@ -17,8 +17,6 @@
import json
from unittest import mock
-import io
-
from foremast.elb import SpinnakerELB
from foremast.elb.format_listeners import format_cert_name, format_listeners
from foremast.elb.splay_health import splay_health | Fixed docstrings and unneeded imports |
diff --git a/lib/bakeware/app_builder.rb b/lib/bakeware/app_builder.rb
index <HASH>..<HASH> 100755
--- a/lib/bakeware/app_builder.rb
+++ b/lib/bakeware/app_builder.rb
@@ -99,6 +99,7 @@ module Bakeware
copy_file 'asset_sync', 'config/initializers/asset_sync.rb'
copy_file 'timeout', 'config/initializers/timeout.rb'
copy_file 'Procfile', 'Procfile'
+ concat_file 'import_scss_styles', 'app/assets/stylesheets/application.css.scss'
inject_into_file 'Procfile', "worker: env QUEUE=* bundle exec rake resque:work",
:after => "\n"
replace_in_file 'Procfile',
@@ -136,7 +137,6 @@ module Bakeware
copy_file 'app/assets/stylesheets/application.css',
'app/assets/stylesheets/application.css.scss'
remove_file 'app/assets/stylesheets/application.css'
- concat_file 'import_scss_styles', 'app/assets/stylesheets/application.css.scss'
create_file 'app/assets/stylesheets/_screen.scss'
end | compass import no longer gets included unless meaty is switched on |
diff --git a/liquibase-core/src/main/java/liquibase/integration/spring/SpringLiquibase.java b/liquibase-core/src/main/java/liquibase/integration/spring/SpringLiquibase.java
index <HASH>..<HASH> 100644
--- a/liquibase-core/src/main/java/liquibase/integration/spring/SpringLiquibase.java
+++ b/liquibase-core/src/main/java/liquibase/integration/spring/SpringLiquibase.java
@@ -72,14 +72,26 @@ import org.springframework.core.io.ResourceLoader;
* @author Rob Schoening
*/
public class SpringLiquibase implements InitializingBean, BeanNameAware, ResourceLoaderAware {
+
+ private boolean dropFirst = false;
+
+ public boolean isDropFirst() {
+ return dropFirst;
+ }
+
+ public void setDropFirst(boolean dropFirst) {
+ this.dropFirst = dropFirst;
+ }
+
+
public class SpringResourceOpener implements ResourceAccessor {
private String parentFile;
+
public SpringResourceOpener(String parentFile) {
this.parentFile = parentFile;
}
-
public InputStream getResourceAsStream(String file) throws IOException {
try {
Resource resource = getResource(file);
@@ -252,6 +264,10 @@ public class SpringLiquibase implements InitializingBean, BeanNameAware, Resourc
}
}
+ if (isDropFirst()) {
+ liquibase.dropAll();
+ }
+
return liquibase;
} | CORE-<I>
Add "dropFirst" property to the SpringLiquibase |
diff --git a/lib/registry.js b/lib/registry.js
index <HASH>..<HASH> 100644
--- a/lib/registry.js
+++ b/lib/registry.js
@@ -41,7 +41,7 @@ exports.value = function (hive, key, valueName, callback) {
key = Array.isArray(key) ? key.join('\\') : key
- const result = registry.enumerateValues(hive, key)
+ const result = registry.enumerateValuesSafe(hive, key)
const expectedName = valueName || ''
const fqk = [shortHive || hive, key].join('\\')
@@ -63,7 +63,7 @@ exports.values = function (hive, key, callback) {
hive = SHORT_HIVES.get(hive)
}
- const result = registry.enumerateValues(hive, key)
+ const result = registry.enumerateValuesSafe(hive, key)
const obj = {}
for (const item of result) { | Avoid errors if the registry is not readable (#<I>)
Previously this could happen if the user did not have access
to the given hive or key. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.