content_type
stringclasses
8 values
main_lang
stringclasses
7 values
message
stringlengths
1
50
sha
stringlengths
40
40
patch
stringlengths
52
962k
file_count
int64
1
300
Javascript
Javascript
fix word spelling
f6210923c012e64738859be4b92700f6937d103b
<ide><path>.eslintrc.js <ide> module.exports = { <ide> ...["implements", "const", "memberof", "readonly", "yields"].reduce( <ide> (acc, tag) => { <ide> acc[tag] = { <del> message: `@${tag} currently not supported in Typescript` <add> message: `@${tag} currently not supported in TypeScript` <ide> }; <ide> return acc; <ide> },
1
Java
Java
use a surface switch
bc06d1cc9c8683115a85207c7f461c5ff82b5ef1
<ide><path>ReactAndroid/src/main/java/com/facebook/react/ReactInstanceManager.java <ide> private void attachRootViewToInstance(final ReactRootView rootView) { <ide> <ide> @Nullable Bundle initialProperties = rootView.getAppProperties(); <ide> final int rootTag = uiManagerModule.addRootView( <del> rootView, <add> rootView.getView(), <ide> initialProperties == null ? <ide> new WritableNativeMap() : Arguments.fromBundle(initialProperties), <ide> rootView.getInitialUITemplate()); <ide><path>ReactAndroid/src/main/java/com/facebook/react/ReactRootView.java <ide> public interface ReactRootViewEventListener { <ide> private int mLastWidth = 0; <ide> private int mLastHeight = 0; <ide> private @UIManagerType int mUIManagerType = DEFAULT; <add> private final boolean mUseSurface; <ide> <ide> public ReactRootView(Context context) { <ide> super(context); <add> mUseSurface = false; <add> init(); <add> } <add> <add> public ReactRootView(Context context, boolean useSurface) { <add> super(context); <add> mUseSurface = useSurface; <ide> init(); <ide> } <ide> <ide> public ReactRootView(Context context, AttributeSet attrs) { <ide> super(context, attrs); <add> mUseSurface = false; <ide> init(); <ide> } <ide> <ide> public ReactRootView(Context context, AttributeSet attrs, int defStyle) { <ide> super(context, attrs, defStyle); <add> mUseSurface = false; <ide> init(); <ide> } <ide> <ide> private void init() { <ide> setClipChildren(false); <ide> } <ide> <add> public View getView() { <add> // TODO add mUseSurface to return surface here <add> return this; <add> } <add> <ide> @Override <ide> protected void onMeasure(int widthMeasureSpec, int heightMeasureSpec) { <add> if (mUseSurface) { <add> super.onMeasure(widthMeasureSpec, heightMeasureSpec); <add> return; <add> } <add> <ide> Systrace.beginSection(TRACE_TAG_REACT_JAVA_BRIDGE, "ReactRootView.onMeasure"); <ide> try { <ide> boolean measureSpecsUpdated = widthMeasureSpec != mWidthMeasureSpec || <ide> public void requestDisallowInterceptTouchEvent(boolean disallowIntercept) { <ide> <ide> @Override <ide> protected void onLayout(boolean changed, int left, int top, int right, int bottom) { <add> if (mUseSurface) { <add> super.onLayout(changed, left, top, right, bottom); <add> } <ide> // No-op since UIManagerModule handles actually laying out children. <ide> } <ide> <ide> public void startReactApplication( <ide> mAppProperties = initialProperties; <ide> mInitialUITemplate = initialUITemplate; <ide> <add> if (mUseSurface) { <add> // TODO initialize surface here <add> } <add> <ide> if (!mReactInstanceManager.hasStartedCreatingInitialContext()) { <ide> mReactInstanceManager.createReactContextInBackground(); <ide> } <ide> public void setAppProperties(@Nullable Bundle appProperties) { <ide> return; <ide> } <ide> <del> if (mWasMeasured) { <del> updateRootLayoutSpecs(mWidthMeasureSpec, mHeightMeasureSpec); <del> } <ide> CatalystInstance catalystInstance = reactContext.getCatalystInstance(); <add> String jsAppModuleName = getJSModuleName(); <ide> <del> WritableNativeMap appParams = new WritableNativeMap(); <del> appParams.putDouble("rootTag", getRootViewTag()); <del> @Nullable Bundle appProperties = getAppProperties(); <del> if (appProperties != null) { <del> appParams.putMap("initialProps", Arguments.fromBundle(appProperties)); <del> } <del> if (getUIManagerType() == FABRIC) { <del> appParams.putBoolean("fabric", true); <del> } <add> if (mUseSurface) { <add> // TODO call surface's runApplication <add> } else { <ide> <del> mShouldLogContentAppeared = true; <add> if (mWasMeasured) { <add> updateRootLayoutSpecs(mWidthMeasureSpec, mHeightMeasureSpec); <add> } <ide> <del> String jsAppModuleName = getJSModuleName(); <del> catalystInstance.getJSModule(AppRegistry.class).runApplication(jsAppModuleName, appParams); <add> WritableNativeMap appParams = new WritableNativeMap(); <add> appParams.putDouble("rootTag", getRootViewTag()); <add> @Nullable Bundle appProperties = getAppProperties(); <add> if (appProperties != null) { <add> appParams.putMap("initialProps", Arguments.fromBundle(appProperties)); <add> } <add> if (getUIManagerType() == FABRIC) { <add> appParams.putBoolean("fabric", true); <add> } <add> <add> mShouldLogContentAppeared = true; <add> <add> catalystInstance.getJSModule(AppRegistry.class).runApplication(jsAppModuleName, appParams); <add> } <ide> } finally { <ide> Systrace.endSection(TRACE_TAG_REACT_JAVA_BRIDGE); <ide> }
2
Javascript
Javascript
fix strictequal assertion arguments
90e72daad5b9020e90b11c9f125cbe996354d679
<ide><path>test/parallel/test-buffer-alloc.js <ide> const SlowBuffer = require('buffer').SlowBuffer; <ide> <ide> <ide> const b = Buffer.allocUnsafe(1024); <del>assert.strictEqual(1024, b.length); <add>assert.strictEqual(b.length, 1024); <ide> <ide> b[0] = -1; <ide> assert.strictEqual(b[0], 255); <ide> for (let i = 0; i < 1024; i++) { <ide> } <ide> <ide> const c = Buffer.allocUnsafe(512); <del>assert.strictEqual(512, c.length); <add>assert.strictEqual(c.length, 512); <ide> <ide> const d = Buffer.from([]); <del>assert.strictEqual(0, d.length); <add>assert.strictEqual(d.length, 0); <ide> <ide> // Test offset properties <ide> { <ide> const b = Buffer.alloc(128); <del> assert.strictEqual(128, b.length); <del> assert.strictEqual(0, b.byteOffset); <del> assert.strictEqual(0, b.offset); <add> assert.strictEqual(b.length, 128); <add> assert.strictEqual(b.byteOffset, 0); <add> assert.strictEqual(b.offset, 0); <ide> } <ide> <ide> // Test creating a Buffer from a Uint32Array <ide> Buffer.alloc(1).write('', 1, 0); <ide> <ide> { <ide> const slice = b.slice(100, 150); <del> assert.strictEqual(50, slice.length); <add> assert.strictEqual(slice.length, 50); <ide> for (let i = 0; i < 50; i++) { <ide> assert.strictEqual(b[100 + i], slice[i]); <ide> } <ide> Buffer.alloc(1).write('', 1, 0); <ide> const a = Buffer.allocUnsafe(8); <ide> for (let i = 0; i < 8; i++) a[i] = i; <ide> const b = a.slice(4, 8); <del> assert.strictEqual(4, b[0]); <del> assert.strictEqual(5, b[1]); <del> assert.strictEqual(6, b[2]); <del> assert.strictEqual(7, b[3]); <add> assert.strictEqual(b[0], 4); <add> assert.strictEqual(b[1], 5); <add> assert.strictEqual(b[2], 6); <add> assert.strictEqual(b[3], 7); <ide> const c = b.slice(2, 4); <del> assert.strictEqual(6, c[0]); <del> assert.strictEqual(7, c[1]); <add> assert.strictEqual(c[0], 6); <add> assert.strictEqual(c[1], 7); <ide> } <ide> <ide> { <ide> Buffer.alloc(1).write('', 1, 0); <ide> // <ide> // Test toString('base64') <ide> // <del>assert.strictEqual('TWFu', (Buffer.from('Man')).toString('base64')); <add>assert.strictEqual((Buffer.from('Man')).toString('base64'), 'TWFu'); <ide> <ide> { <ide> // test that regular and URL-safe base64 both work <ide> assert.deepStrictEqual(Buffer.from(' YWJvcnVtLg', 'base64'), <ide> const b = Buffer.from(s); <ide> <ide> for (let i = 0; i < l; i++) { <del> assert.strictEqual('h'.charCodeAt(0), b[i]); <add> assert.strictEqual(b[i], 'h'.charCodeAt(0)); <ide> } <ide> <ide> const sb = b.toString(); <ide> function buildBuffer(data) { <ide> <ide> const x = buildBuffer([0x81, 0xa3, 0x66, 0x6f, 0x6f, 0xa3, 0x62, 0x61, 0x72]); <ide> <del>assert.strictEqual('<Buffer 81 a3 66 6f 6f a3 62 61 72>', x.inspect()); <add>assert.strictEqual(x.inspect(), '<Buffer 81 a3 66 6f 6f a3 62 61 72>'); <ide> <ide> { <ide> const z = x.slice(4); <del> assert.strictEqual(5, z.length); <del> assert.strictEqual(0x6f, z[0]); <del> assert.strictEqual(0xa3, z[1]); <del> assert.strictEqual(0x62, z[2]); <del> assert.strictEqual(0x61, z[3]); <del> assert.strictEqual(0x72, z[4]); <add> assert.strictEqual(z.length, 5); <add> assert.strictEqual(z[0], 0x6f); <add> assert.strictEqual(z[1], 0xa3); <add> assert.strictEqual(z[2], 0x62); <add> assert.strictEqual(z[3], 0x61); <add> assert.strictEqual(z[4], 0x72); <ide> } <ide> <ide> { <ide> assert.strictEqual('<Buffer 81 a3 66 6f 6f a3 62 61 72>', x.inspect()); <ide> <ide> { <ide> const z = x.slice(0, 4); <del> assert.strictEqual(4, z.length); <del> assert.strictEqual(0x81, z[0]); <del> assert.strictEqual(0xa3, z[1]); <add> assert.strictEqual(z.length, 4); <add> assert.strictEqual(z[0], 0x81); <add> assert.strictEqual(z[1], 0xa3); <ide> } <ide> <ide> { <ide> const z = x.slice(0, 9); <del> assert.strictEqual(9, z.length); <add> assert.strictEqual(z.length, 9); <ide> } <ide> <ide> { <ide> const z = x.slice(1, 4); <del> assert.strictEqual(3, z.length); <del> assert.strictEqual(0xa3, z[0]); <add> assert.strictEqual(z.length, 3); <add> assert.strictEqual(z[0], 0xa3); <ide> } <ide> <ide> { <ide> const z = x.slice(2, 4); <del> assert.strictEqual(2, z.length); <del> assert.strictEqual(0x66, z[0]); <del> assert.strictEqual(0x6f, z[1]); <add> assert.strictEqual(z.length, 2); <add> assert.strictEqual(z[0], 0x66); <add> assert.strictEqual(z[1], 0x6f); <ide> } <ide> <ide> ['ucs2', 'ucs-2', 'utf16le', 'utf-16le'].forEach((encoding) => { <ide> assert.strictEqual('<Buffer 81 a3 66 6f 6f a3 62 61 72>', x.inspect()); <ide> const b = Buffer.from([0xde, 0xad, 0xbe, 0xef]); <ide> let s = String.fromCharCode(0xffff); <ide> b.write(s, 0, 'latin1'); <del> assert.strictEqual(0xff, b[0]); <del> assert.strictEqual(0xad, b[1]); <del> assert.strictEqual(0xbe, b[2]); <del> assert.strictEqual(0xef, b[3]); <add> assert.strictEqual(b[0], 0xff); <add> assert.strictEqual(b[1], 0xad); <add> assert.strictEqual(b[2], 0xbe); <add> assert.strictEqual(b[3], 0xef); <ide> s = String.fromCharCode(0xaaee); <ide> b.write(s, 0, 'latin1'); <del> assert.strictEqual(0xee, b[0]); <del> assert.strictEqual(0xad, b[1]); <del> assert.strictEqual(0xbe, b[2]); <del> assert.strictEqual(0xef, b[3]); <add> assert.strictEqual(b[0], 0xee); <add> assert.strictEqual(b[1], 0xad); <add> assert.strictEqual(b[2], 0xbe); <add> assert.strictEqual(b[3], 0xef); <ide> } <ide> <ide> { <ide> // Binary encoding should write only one byte per character. <ide> const b = Buffer.from([0xde, 0xad, 0xbe, 0xef]); <ide> let s = String.fromCharCode(0xffff); <ide> b.write(s, 0, 'latin1'); <del> assert.strictEqual(0xff, b[0]); <del> assert.strictEqual(0xad, b[1]); <del> assert.strictEqual(0xbe, b[2]); <del> assert.strictEqual(0xef, b[3]); <add> assert.strictEqual(b[0], 0xff); <add> assert.strictEqual(b[1], 0xad); <add> assert.strictEqual(b[2], 0xbe); <add> assert.strictEqual(b[3], 0xef); <ide> s = String.fromCharCode(0xaaee); <ide> b.write(s, 0, 'latin1'); <del> assert.strictEqual(0xee, b[0]); <del> assert.strictEqual(0xad, b[1]); <del> assert.strictEqual(0xbe, b[2]); <del> assert.strictEqual(0xef, b[3]); <add> assert.strictEqual(b[0], 0xee); <add> assert.strictEqual(b[1], 0xad); <add> assert.strictEqual(b[2], 0xbe); <add> assert.strictEqual(b[3], 0xef); <ide> } <ide> <ide> { <ide> assert.strictEqual('<Buffer 81 a3 66 6f 6f a3 62 61 72>', x.inspect()); <ide> { <ide> // test offset returns are correct <ide> const b = Buffer.allocUnsafe(16); <del> assert.strictEqual(4, b.writeUInt32LE(0, 0)); <del> assert.strictEqual(6, b.writeUInt16LE(0, 4)); <del> assert.strictEqual(7, b.writeUInt8(0, 6)); <del> assert.strictEqual(8, b.writeInt8(0, 7)); <del> assert.strictEqual(16, b.writeDoubleLE(0, 8)); <add> assert.strictEqual(b.writeUInt32LE(0, 0), 4); <add> assert.strictEqual(b.writeUInt16LE(0, 4), 6); <add> assert.strictEqual(b.writeUInt8(0, 6), 7); <add> assert.strictEqual(b.writeInt8(0, 7), 8); <add> assert.strictEqual(b.writeDoubleLE(0, 8), 16); <ide> } <ide> <ide> {
1
PHP
PHP
apply fixes from styleci
22cdcf7cd02a6e35146b4737321a7ce18a075d43
<ide><path>tests/Integration/Console/ConsoleApplicationTest.php <ide> use Illuminate\Console\Command; <ide> use Orchestra\Testbench\TestCase; <ide> use Illuminate\Contracts\Console\Kernel; <del>use Illuminate\Foundation\Testing\PendingCommand; <ide> <ide> class ConsoleApplicationTest extends TestCase <ide> {
1
Go
Go
address feedback from tonis
0cba7740d41369eee33b671f26276325580bc07b
<ide><path>builder/dockerfile/builder.go <ide> func addNodesForLabelOption(dockerfile *parser.Node, labels map[string]string) { <ide> // <ide> // TODO: Remove? <ide> func BuildFromConfig(config *container.Config, changes []string, os string) (*container.Config, error) { <add> if !system.IsOSSupported(os) { <add> return nil, errdefs.InvalidParameter(system.ErrNotSupportedOperatingSystem) <add> } <ide> if len(changes) == 0 { <ide> return config, nil <ide> } <ide><path>builder/dockerfile/dispatchers.go <ide> func initializeStage(d dispatchRequest, cmd *instructions.Stage) error { <ide> return err <ide> } <ide> state := d.state <del> state.beginStage(cmd.Name, image) <add> if err := state.beginStage(cmd.Name, image); err != nil { <add> return err <add> } <ide> if len(state.runConfig.OnBuild) > 0 { <ide> triggers := state.runConfig.OnBuild <ide> state.runConfig.OnBuild = nil <ide> func resolveCmdLine(cmd instructions.ShellDependantCmdLine, runConfig *container <ide> // RUN [ "echo", "hi" ] # echo hi <ide> // <ide> func dispatchRun(d dispatchRequest, c *instructions.RunCommand) error { <del> <add> if !system.IsOSSupported(d.state.operatingSystem) { <add> return system.ErrNotSupportedOperatingSystem <add> } <ide> stateRunConfig := d.state.runConfig <ide> cmdFromArgs := resolveCmdLine(c.ShellDependantCmdLine, stateRunConfig, d.state.operatingSystem) <ide> buildArgs := d.state.buildArgs.FilterAllowed(stateRunConfig.Env) <ide><path>builder/dockerfile/evaluator.go <ide> package dockerfile <ide> <ide> import ( <ide> "reflect" <add> "runtime" <ide> "strconv" <ide> "strings" <ide> <ide> func (s *dispatchState) hasFromImage() bool { <ide> return s.imageID != "" || (s.baseImage != nil && s.baseImage.ImageID() == "") <ide> } <ide> <del>func (s *dispatchState) beginStage(stageName string, image builder.Image) { <add>func (s *dispatchState) beginStage(stageName string, image builder.Image) error { <ide> s.stageName = stageName <ide> s.imageID = image.ImageID() <ide> s.operatingSystem = image.OperatingSystem() <add> if s.operatingSystem == "" { // In case it isn't set <add> s.operatingSystem = runtime.GOOS <add> } <add> if !system.IsOSSupported(s.operatingSystem) { <add> return system.ErrNotSupportedOperatingSystem <add> } <ide> <ide> if image.RunConfig() != nil { <ide> // copy avoids referencing the same instance when 2 stages have the same base <ide> func (s *dispatchState) beginStage(stageName string, image builder.Image) { <ide> s.setDefaultPath() <ide> s.runConfig.OpenStdin = false <ide> s.runConfig.StdinOnce = false <add> return nil <ide> } <ide> <ide> // Add the default PATH to runConfig.ENV if one exists for the operating system and there <ide><path>daemon/build.go <ide> import ( <ide> "github.com/docker/docker/pkg/containerfs" <ide> "github.com/docker/docker/pkg/idtools" <ide> "github.com/docker/docker/pkg/stringid" <add> "github.com/docker/docker/pkg/system" <ide> "github.com/docker/docker/registry" <ide> "github.com/pkg/errors" <ide> "github.com/sirupsen/logrus" <ide> func (rl *releaseableLayer) Commit() (builder.ReleaseableLayer, error) { <ide> if err != nil { <ide> return nil, err <ide> } <del> // TODO: An optimization woudld be to handle empty layers before returning <add> // TODO: An optimization would be to handle empty layers before returning <ide> return &releaseableLayer{layerStore: rl.layerStore, roLayer: newLayer}, nil <ide> } <ide> <ide> func (daemon *Daemon) pullForBuilder(ctx context.Context, name string, authConfi <ide> // leaking of layers. <ide> func (daemon *Daemon) GetImageAndReleasableLayer(ctx context.Context, refOrID string, opts backend.GetImageAndLayerOptions) (builder.Image, builder.ReleaseableLayer, error) { <ide> if refOrID == "" { <add> if !system.IsOSSupported(opts.OS) { <add> return nil, nil, system.ErrNotSupportedOperatingSystem <add> } <ide> layer, err := newReleasableLayerForImage(nil, daemon.layerStores[opts.OS]) <ide> return nil, layer, err <ide> } <ide> func (daemon *Daemon) GetImageAndReleasableLayer(ctx context.Context, refOrID st <ide> } <ide> // TODO: shouldn't we error out if error is different from "not found" ? <ide> if image != nil { <add> if !system.IsOSSupported(image.OperatingSystem()) { <add> return nil, nil, system.ErrNotSupportedOperatingSystem <add> } <ide> layer, err := newReleasableLayerForImage(image, daemon.layerStores[image.OperatingSystem()]) <ide> return image, layer, err <ide> } <ide> func (daemon *Daemon) GetImageAndReleasableLayer(ctx context.Context, refOrID st <ide> if err != nil { <ide> return nil, nil, err <ide> } <add> if !system.IsOSSupported(image.OperatingSystem()) { <add> return nil, nil, system.ErrNotSupportedOperatingSystem <add> } <ide> layer, err := newReleasableLayerForImage(image, daemon.layerStores[image.OperatingSystem()]) <ide> return image, layer, err <ide> } <ide><path>daemon/commit.go <ide> import ( <ide> "github.com/docker/docker/image" <ide> "github.com/docker/docker/layer" <ide> "github.com/docker/docker/pkg/ioutils" <add> "github.com/docker/docker/pkg/system" <ide> "github.com/pkg/errors" <ide> ) <ide> <ide> func (daemon *Daemon) Commit(name string, c *backend.ContainerCommitConfig) (str <ide> daemon.containerPause(container) <ide> defer daemon.containerUnpause(container) <ide> } <add> if !system.IsOSSupported(container.OS) { <add> return "", system.ErrNotSupportedOperatingSystem <add> } <ide> <ide> if c.MergeConfigs && c.Config == nil { <ide> c.Config = container.Config <ide> func (daemon *Daemon) Commit(name string, c *backend.ContainerCommitConfig) (str <ide> } <ide> <ide> func (daemon *Daemon) exportContainerRw(container *container.Container) (arch io.ReadCloser, err error) { <add> // Note: Indexing by OS is safe as only called from `Commit` which has already performed validation <ide> rwlayer, err := daemon.layerStores[container.OS].GetRWLayer(container.ID) <ide> if err != nil { <ide> return nil, err <ide><path>daemon/create.go <ide> func (daemon *Daemon) setRWLayer(container *container.Container) error { <ide> StorageOpt: container.HostConfig.StorageOpt, <ide> } <ide> <add> // Indexing by OS is safe here as validation of OS has already been performed in create() (the only <add> // caller), and guaranteed non-nil <ide> rwLayer, err := daemon.layerStores[container.OS].CreateRWLayer(container.ID, layerID, rwLayerOpts) <ide> if err != nil { <ide> return err <ide><path>daemon/daemon.go <ide> func (daemon *Daemon) restore() error { <ide> logrus.Errorf("Failed to load container %v: %v", id, err) <ide> continue <ide> } <del> <add> if !system.IsOSSupported(container.OS) { <add> logrus.Errorf("Failed to load container %v: %s (%q)", id, system.ErrNotSupportedOperatingSystem, container.OS) <add> continue <add> } <ide> // Ignore the container if it does not support the current driver being used by the graph <ide> currentDriverForContainerOS := daemon.graphDrivers[container.OS] <ide> if (container.Driver == "" && currentDriverForContainerOS == "aufs") || container.Driver == currentDriverForContainerOS { <ide><path>daemon/delete.go <ide> func (daemon *Daemon) cleanupContainer(container *container.Container, forceRemo <ide> return fmt.Errorf("Could not kill running container %s, cannot remove - %v", container.ID, err) <ide> } <ide> } <add> if !system.IsOSSupported(container.OS) { <add> return fmt.Errorf("cannot remove %s: %s ", container.ID, system.ErrNotSupportedOperatingSystem) <add> } <ide> <ide> // stop collection of stats for the container regardless <ide> // if stats are currently getting collected. <ide><path>daemon/export.go <ide> import ( <ide> "github.com/docker/docker/errdefs" <ide> "github.com/docker/docker/pkg/archive" <ide> "github.com/docker/docker/pkg/ioutils" <add> "github.com/docker/docker/pkg/system" <ide> ) <ide> <ide> // ContainerExport writes the contents of the container to the given <ide> func (daemon *Daemon) ContainerExport(name string, out io.Writer) error { <ide> } <ide> <ide> func (daemon *Daemon) containerExport(container *container.Container) (arch io.ReadCloser, err error) { <add> if !system.IsOSSupported(container.OS) { <add> return nil, fmt.Errorf("cannot export %s: %s ", container.ID, system.ErrNotSupportedOperatingSystem) <add> } <ide> rwlayer, err := daemon.layerStores[container.OS].GetRWLayer(container.ID) <ide> if err != nil { <ide> return nil, err <ide><path>daemon/image_delete.go <ide> import ( <ide> "github.com/docker/docker/errdefs" <ide> "github.com/docker/docker/image" <ide> "github.com/docker/docker/pkg/stringid" <add> "github.com/docker/docker/pkg/system" <ide> "github.com/pkg/errors" <ide> ) <ide> <ide> func (daemon *Daemon) ImageDelete(imageRef string, force, prune bool) ([]types.I <ide> start := time.Now() <ide> records := []types.ImageDeleteResponseItem{} <ide> <del> imgID, _, err := daemon.GetImageIDAndOS(imageRef) <add> imgID, operatingSystem, err := daemon.GetImageIDAndOS(imageRef) <ide> if err != nil { <ide> return nil, err <ide> } <add> if !system.IsOSSupported(operatingSystem) { <add> return nil, errors.Errorf("unable to delete image: %q", system.ErrNotSupportedOperatingSystem) <add> } <ide> <ide> repoRefs := daemon.referenceStore.References(imgID.Digest()) <ide> <ide><path>daemon/image_inspect.go <ide> import ( <ide> "github.com/docker/distribution/reference" <ide> "github.com/docker/docker/api/types" <ide> "github.com/docker/docker/layer" <add> "github.com/docker/docker/pkg/system" <ide> "github.com/pkg/errors" <ide> ) <ide> <ide> func (daemon *Daemon) LookupImage(name string) (*types.ImageInspect, error) { <ide> if err != nil { <ide> return nil, errors.Wrapf(err, "no such image: %s", name) <ide> } <del> <add> if !system.IsOSSupported(img.OperatingSystem()) { <add> return nil, system.ErrNotSupportedOperatingSystem <add> } <ide> refs := daemon.referenceStore.References(img.ID().Digest()) <ide> repoTags := []string{} <ide> repoDigests := []string{} <ide><path>daemon/images.go <ide> import ( <ide> "github.com/docker/docker/container" <ide> "github.com/docker/docker/image" <ide> "github.com/docker/docker/layer" <add> "github.com/docker/docker/pkg/system" <ide> ) <ide> <ide> var acceptedImageFilterTags = map[string]bool{ <ide> func (daemon *Daemon) Images(imageFilters filters.Args, all bool, withExtraAttrs <ide> } <ide> } <ide> <add> // Skip any images with an unsupported operating system to avoid a potential <add> // panic when indexing through the layerstore. Don't error as we want to list <add> // the other images. This should never happen, but here as a safety precaution. <add> if !system.IsOSSupported(img.OperatingSystem()) { <add> continue <add> } <add> <ide> layerID := img.RootFS.ChainID() <ide> var size int64 <ide> if layerID != "" { <ide><path>daemon/oci_windows.go <ide> func (daemon *Daemon) createSpec(c *container.Container) (*specs.Spec, error) { <ide> max := len(img.RootFS.DiffIDs) <ide> for i := 1; i <= max; i++ { <ide> img.RootFS.DiffIDs = img.RootFS.DiffIDs[:i] <add> if !system.IsOSSupported(img.OperatingSystem()) { <add> return nil, fmt.Errorf("cannot get layerpath for ImageID %s: %s ", img.RootFS.ChainID(), system.ErrNotSupportedOperatingSystem) <add> } <ide> layerPath, err := layer.GetLayerPath(daemon.layerStores[img.OperatingSystem()], img.RootFS.ChainID()) <ide> if err != nil { <ide> return nil, fmt.Errorf("failed to get layer path from graphdriver %s for ImageID %s - %s", daemon.layerStores[img.OperatingSystem()], img.RootFS.ChainID(), err) <ide><path>distribution/config.go <ide> func (s *imageConfigStore) RootFSAndOSFromConfig(c []byte) (*image.RootFS, strin <ide> if os == "" { <ide> os = runtime.GOOS <ide> } <add> if !system.IsOSSupported(os) { <add> return nil, "", system.ErrNotSupportedOperatingSystem <add> } <ide> return unmarshalledConfig.RootFS, os, nil <ide> } <ide> <ide><path>distribution/push_v1.go <ide> import ( <ide> "github.com/docker/docker/pkg/ioutils" <ide> "github.com/docker/docker/pkg/progress" <ide> "github.com/docker/docker/pkg/stringid" <add> "github.com/docker/docker/pkg/system" <ide> "github.com/docker/docker/registry" <ide> "github.com/opencontainers/go-digest" <ide> "github.com/sirupsen/logrus" <ide> func (p *v1Pusher) imageListForTag(imgID image.ID, dependenciesSeen map[layer.Ch <ide> <ide> topLayerID := img.RootFS.ChainID() <ide> <add> if !system.IsOSSupported(img.OperatingSystem()) { <add> return nil, system.ErrNotSupportedOperatingSystem <add> } <ide> pl, err := p.config.LayerStores[img.OperatingSystem()].Get(topLayerID) <ide> *referencedLayers = append(*referencedLayers, pl) <ide> if err != nil { <ide><path>distribution/xfer/download.go <ide> import ( <ide> "github.com/docker/docker/pkg/archive" <ide> "github.com/docker/docker/pkg/ioutils" <ide> "github.com/docker/docker/pkg/progress" <add> "github.com/docker/docker/pkg/system" <ide> "github.com/sirupsen/logrus" <ide> "golang.org/x/net/context" <ide> ) <ide> func (ldm *LayerDownloadManager) Download(ctx context.Context, initialRootFS ima <ide> downloadsByKey = make(map[string]*downloadTransfer) <ide> ) <ide> <del> // Assume that the operating system is the host OS if blank <add> // Assume that the operating system is the host OS if blank, and validate it <add> // to ensure we don't cause a panic by an invalid index into the layerstores. <ide> if os == "" { <ide> os = runtime.GOOS <ide> } <add> if !system.IsOSSupported(os) { <add> return image.RootFS{}, nil, system.ErrNotSupportedOperatingSystem <add> } <ide> <ide> rootFS := initialRootFS <ide> for _, descriptor := range layers { <ide><path>image/store.go <ide> import ( <ide> <ide> "github.com/docker/distribution/digestset" <ide> "github.com/docker/docker/layer" <add> "github.com/docker/docker/pkg/system" <ide> "github.com/opencontainers/go-digest" <ide> "github.com/pkg/errors" <ide> "github.com/sirupsen/logrus" <ide> func (is *store) restore() error { <ide> } <ide> var l layer.Layer <ide> if chainID := img.RootFS.ChainID(); chainID != "" { <add> if !system.IsOSSupported(img.OperatingSystem()) { <add> return system.ErrNotSupportedOperatingSystem <add> } <ide> l, err = is.lss[img.OperatingSystem()].Get(chainID) <ide> if err != nil { <ide> return err <ide> func (is *store) Create(config []byte) (ID, error) { <ide> <ide> var l layer.Layer <ide> if layerID != "" { <add> if !system.IsOSSupported(img.OperatingSystem()) { <add> return "", system.ErrNotSupportedOperatingSystem <add> } <ide> l, err = is.lss[img.OperatingSystem()].Get(layerID) <ide> if err != nil { <ide> return "", errors.Wrapf(err, "failed to get layer %s", layerID) <ide><path>layer/filestore_unix.go <ide> package layer <ide> <ide> import "runtime" <ide> <del>// SetOS writes the "os" file to the layer filestore <del>func (fm *fileMetadataTransaction) SetOS(os string) error { <add>// setOS writes the "os" file to the layer filestore <add>func (fm *fileMetadataTransaction) setOS(os string) error { <ide> return nil <ide> } <ide> <del>// GetOS reads the "os" file from the layer filestore <del>func (fms *fileMetadataStore) GetOS(layer ChainID) (string, error) { <add>// getOS reads the "os" file from the layer filestore <add>func (fms *fileMetadataStore) getOS(layer ChainID) (string, error) { <ide> return runtime.GOOS, nil <ide> } <ide><path>layer/filestore_windows.go <ide> import ( <ide> "strings" <ide> ) <ide> <del>// SetOS writes the "os" file to the layer filestore <del>func (fm *fileMetadataTransaction) SetOS(os string) error { <add>// setOS writes the "os" file to the layer filestore <add>func (fm *fileMetadataTransaction) setOS(os string) error { <ide> if os == "" { <ide> return nil <ide> } <ide> return fm.ws.WriteFile("os", []byte(os), 0644) <ide> } <ide> <del>// GetOS reads the "os" file from the layer filestore <del>func (fms *fileMetadataStore) GetOS(layer ChainID) (string, error) { <add>// getOS reads the "os" file from the layer filestore <add>func (fms *fileMetadataStore) getOS(layer ChainID) (string, error) { <ide> contentBytes, err := ioutil.ReadFile(fms.getLayerFilename(layer, "os")) <ide> if err != nil { <ide> // For backwards compatibility, the os file may not exist. Default to "windows" if missing. <ide><path>layer/layer.go <ide> type MetadataTransaction interface { <ide> SetDiffID(DiffID) error <ide> SetCacheID(string) error <ide> SetDescriptor(distribution.Descriptor) error <del> SetOS(string) error <add> setOS(string) error <ide> TarSplitWriter(compressInput bool) (io.WriteCloser, error) <ide> <ide> Commit(ChainID) error <ide> type MetadataStore interface { <ide> GetDiffID(ChainID) (DiffID, error) <ide> GetCacheID(ChainID) (string, error) <ide> GetDescriptor(ChainID) (distribution.Descriptor, error) <del> GetOS(ChainID) (string, error) <add> getOS(ChainID) (string, error) <ide> TarSplitReader(ChainID) (io.ReadCloser, error) <ide> <ide> SetMountID(string, string) error <ide><path>layer/layer_store.go <ide> func (ls *layerStore) loadLayer(layer ChainID) (*roLayer, error) { <ide> return nil, fmt.Errorf("failed to get descriptor for %s: %s", layer, err) <ide> } <ide> <del> os, err := ls.store.GetOS(layer) <add> os, err := ls.store.getOS(layer) <ide> if err != nil { <ide> return nil, fmt.Errorf("failed to get operating system for %s: %s", layer, err) <ide> } <ide><path>layer/ro_layer.go <ide> func storeLayer(tx MetadataTransaction, layer *roLayer) error { <ide> return err <ide> } <ide> } <del> if err := tx.SetOS(layer.layerStore.os); err != nil { <add> if err := tx.setOS(layer.layerStore.os); err != nil { <ide> return err <ide> } <ide> <ide><path>pkg/system/errors.go <ide> import ( <ide> var ( <ide> // ErrNotSupportedPlatform means the platform is not supported. <ide> ErrNotSupportedPlatform = errors.New("platform and architecture is not supported") <add> <add> // ErrNotSupportedOperatingSystem means the operating system is not supported. <add> ErrNotSupportedOperatingSystem = errors.New("operating system is not supported") <ide> ) <ide><path>pkg/system/lcow.go <ide> func IsOSSupported(os string) bool { <ide> if runtime.GOOS == os { <ide> return true <ide> } <del> if LCOWSupported() && runtime.GOOS == "windows" && os == "linux" { <add> if LCOWSupported() && os == "linux" { <ide> return true <ide> } <ide> return false
24
PHP
PHP
add accessor for reading individual files
96630e240a7b0133c443b3899c7636efb1102855
<ide><path>src/Network/Request.php <ide> public function getAttributes() <ide> return $this->attributes + $emulated; <ide> } <ide> <add> /** <add> * Get the uploaded file from a dotted path. <add> * <add> * @param string $path The dot separated path to the file you want. <add> * @return null|Psr\Http\Message\UploadedFileInterface <add> */ <add> public function getUploadedFile($path) <add> { <add> $file = Hash::get($this->uploadedFiles, $path); <add> if (!$file instanceof UploadedFile) { <add> return null; <add> } <add> <add> return $file; <add> } <add> <ide> /** <ide> * Get the array of uploaded files from the request. <ide> * <ide><path>tests/TestCase/Network/RequestTest.php <ide> public function testWithUploadedFiles() <ide> $this->assertSame(['picture' => $file], $new->getUploadedFiles()); <ide> } <ide> <add> /** <add> * Test getting a single file <add> * <add> * @return void <add> */ <add> public function testGetUploadedFile() <add> { <add> $file = new UploadedFile( <add> __FILE__, <add> 123, <add> UPLOAD_ERR_OK, <add> 'test.php', <add> 'text/plain' <add> ); <add> $request = new Request(); <add> $new = $request->withUploadedFiles(['picture' => $file]); <add> $this->assertNull($new->getUploadedFile('')); <add> $this->assertSame($file, $new->getUploadedFile('picture')); <add> <add> $new = $request->withUploadedFiles([ <add> 'pictures' => [ <add> [ <add> 'image' => $file <add> ] <add> ] <add> ]); <add> $this->assertNull($new->getUploadedFile('pictures')); <add> $this->assertNull($new->getUploadedFile('pictures.0')); <add> $this->assertNull($new->getUploadedFile('pictures.1')); <add> $this->assertSame($file, $new->getUploadedFile('pictures.0.image')); <add> } <add> <ide> /** <ide> * Test replacing files with an invalid file <ide> *
2
Text
Text
update seed code for d3.js challenge
f8aa6c04e20cf363d727d93c366f34894625110a
<ide><path>curriculum/challenges/english/04-data-visualization/data-visualization-with-d3/change-the-presentation-of-a-bar-chart.md <ide> assert( <ide> .bar { <ide> width: 25px; <ide> height: 100px; <del> /* Only change code below this line */ <add> /* Add your code below this line */ <ide> <ide> <del> /* Only change code above this line */ <add> /* Add your code above this line */ <ide> display: inline-block; <ide> background-color: blue; <ide> } <ide> assert( <ide> .enter() <ide> .append("div") <ide> .attr("class", "bar") <del> .style("height", (d) => (d + "px")) <add> .style("height", (d) => (d + "px")) // Change this line <ide> </script> <ide> </body> <ide> ```
1
Javascript
Javascript
add info circle to claim cert steps
1805631e4a03e7c97950d22090e06650ed9a62a7
<ide><path>client/src/templates/Introduction/components/ClaimCertSteps.js <ide> import { withTranslation, useTranslation } from 'react-i18next'; <ide> import { connect } from 'react-redux'; <ide> import { createSelector } from 'reselect'; <ide> <add>import IntroInformation from '../../../assets/icons/IntroInformation'; <ide> import GreenPass from '../../../assets/icons/GreenPass'; <ide> import GreenNotCompleted from '../../../assets/icons/GreenNotCompleted'; <ide> import { userSelector } from '../../../redux'; <ide> const ClaimCertSteps = ({ certSlug, i18nCertText, superBlock, user }) => { <ide> {t('certification-card.complete-project', { <ide> i18nCertText <ide> })} <add> <span className='badge map-badge map-project-checkmark'> <add> <IntroInformation style={mapIconStyle} /> <add> </span> <ide> </a> <ide> </li> <ide> <li className='map-challenge-title map-project-wrap'> <del> <Link to={certClaimLink}>{t('certification-card.set-claim')}</Link> <add> <Link to={certClaimLink}> <add> {t('certification-card.set-claim')} <add> <span className='badge map-badge map-project-checkmark'> <add> <IntroInformation style={mapIconStyle} /> <add> </span> <add> </Link> <ide> </li> <ide> </ul> <ide> );
1
PHP
PHP
add doc blocks to callbackstatement
a79b614bef74503ac2c9bb18a33658b305d40fe7
<ide><path>Cake/Database/Statement/CallbackStatement.php <ide> */ <ide> namespace Cake\Database\Statement; <ide> <add>/** <add> * Wraps a statement in a callback that allows row results <add> * to be modified when being fetched. <add> * <add> * This is used by CakePHP to eagerly load association data. <add> */ <ide> class CallbackStatement extends StatementDecorator { <ide> <add>/** <add> * A callback function to be applied to results. <add> * <add> * @var callable <add> */ <ide> protected $_callback; <ide> <add>/** <add> * Constructor <add> * <add> * @param Cake\Database\StatementInterface $statement The statement to decorate. <add> * @param Cake\Database\Driver $driver The driver instance used by the statement. <add> * @param callable $callback The callback to apply to results before they are returned. <add> */ <ide> public function __construct($statement, $driver, $callback) { <ide> parent::__construct($statement, $driver); <ide> $this->_callback = $callback; <ide> } <ide> <add>/** <add> * Fetch a row from the statement. <add> * <add> * The result will be procssed by the callback when it is not `false. <add> * <add> * @param string $type Either 'num' or 'assoc' to indicate the result format you would like. <add> * @return mixed <add> */ <ide> public function fetch($type = 'num') { <ide> $callback = $this->_callback; <ide> $row = $this->_statement->fetch($type); <ide> return $row === false ? $row : $callback($row); <ide> } <ide> <add>/** <add> * Fetch all rows from the statement. <add> * <add> * Each row in the result will be procssed by the callback when it is not `false. <add> * <add> * @param string $type Either 'num' or 'assoc' to indicate the result format you would like. <add> * @return mixed <add> */ <ide> public function fetchAll($type = 'num') { <ide> return array_map($this->_callback, $this->_statement->fetchAll($type)); <ide> }
1
Ruby
Ruby
use rails() instead of system()
75afb43a25eacaefd7487dfa1f95be03c06c62ee
<ide><path>railties/test/application/assets_test.rb <ide> def teardown <ide> def precompile!(env = nil) <ide> with_env env.to_h do <ide> quietly do <del> precompile_task = "bin/rails assets:precompile --trace 2>&1" <del> output = Dir.chdir(app_path) { %x[ #{precompile_task} ] } <del> assert $?.success?, output <del> output <add> rails ["assets:precompile", "--trace"] <ide> end <ide> end <ide> end <ide> def with_env(env) <ide> <ide> def clean_assets! <ide> quietly do <del> assert Dir.chdir(app_path) { system("bin/rails assets:clobber") } <add> rails ["assets:clobber"] <ide> end <ide> end <ide>
1
Javascript
Javascript
add tests for jwt authorization
36c473799839bd4b31d29a0aa418f8c5f527d1f5
<ide><path>api-server/server/middlewares/jwt-authorizaion.test.js <del>/* global describe xdescribe it expect */ <del>import { isWhiteListedPath } from './jwt-authorization'; <del> <del>describe('jwt-authorization', () => { <del> describe('isWhiteListedPath', () => { <del> const whiteList = [/^\/is-ok\//, /^\/this-is\/also\/ok\//]; <del> <del> it('returns a boolean', () => { <del> const result = isWhiteListedPath(); <del> <del> expect(typeof result).toBe('boolean'); <del> }); <del> <del> it('returns true for a white listed path', () => { <del> expect.assertions(2); <del> <del> const resultA = isWhiteListedPath('/is-ok/should-be/good', whiteList); <del> const resultB = isWhiteListedPath('/this-is/also/ok/surely', whiteList); <del> expect(resultA).toBe(true); <del> expect(resultB).toBe(true); <del> }); <del> <del> it('returns false for a non-white-listed path', () => { <del> const result = isWhiteListedPath('/hax0r-42/no-go', whiteList); <del> expect(result).toBe(false); <del> }); <del> }); <del> <del> xdescribe('authorizeByJWT'); <del>}); <ide><path>api-server/server/middlewares/request-authorizaion.test.js <add>/* global describe it expect */ <add>import sinon from 'sinon'; <add>import { mockReq, mockRes } from 'sinon-express-mock'; <add>import jwt from 'jsonwebtoken'; <add> <add>import createRequestAuthorization, { <add> isWhiteListedPath <add>} from './request-authorization'; <add> <add>const validJWTSecret = 'this is a super secret string'; <add>const invalidJWTSecret = 'This is not correct secret'; <add>const now = new Date(Date.now()); <add>const theBeginningOfTime = new Date(0); <add>const accessToken = { <add> id: '123abc', <add> userId: '456def', <add> ttl: 60000, <add> created: now <add>}; <add>const users = { <add> '456def': { <add> username: 'camperbot', <add> progressTimestamps: [1, 2, 3, 4] <add> } <add>}; <add>const mockGetUserById = id => <add> id in users ? Promise.resolve(users[id]) : Promise.reject('No user found'); <add> <add>describe('request-authorization', () => { <add> describe('isWhiteListedPath', () => { <add> const whiteList = [/^\/is-ok\//, /^\/this-is\/also\/ok\//]; <add> <add> it('returns a boolean', () => { <add> const result = isWhiteListedPath(); <add> <add> expect(typeof result).toBe('boolean'); <add> }); <add> <add> it('returns true for a white listed path', () => { <add> expect.assertions(2); <add> <add> const resultA = isWhiteListedPath('/is-ok/should-be/good', whiteList); <add> const resultB = isWhiteListedPath('/this-is/also/ok/surely', whiteList); <add> expect(resultA).toBe(true); <add> expect(resultB).toBe(true); <add> }); <add> <add> it('returns false for a non-white-listed path', () => { <add> const result = isWhiteListedPath('/hax0r-42/no-go', whiteList); <add> expect(result).toBe(false); <add> }); <add> }); <add> <add> describe('createRequestAuthorization', () => { <add> const requestAuthorization = createRequestAuthorization({ <add> _jwtSecret: validJWTSecret, <add> getUserById: mockGetUserById <add> }); <add> <add> it('is a function', () => { <add> expect(typeof requestAuthorization).toEqual('function'); <add> }); <add> <add> it('throws when no access token is present', () => { <add> expect.assertions(2); <add> const req = mockReq({ path: '/internal/some-path/that-needs/auth' }); <add> const res = mockRes(); <add> const next = sinon.spy(); <add> expect(() => requestAuthorization(req, res, next)).toThrowError( <add> 'Access token is required for this request' <add> ); <add> expect(next.called).toBe(false); <add> }); <add> <add> it('throws when the access token is invalid', () => { <add> expect.assertions(2); <add> const invalidJWT = jwt.sign({ accessToken }, invalidJWTSecret); <add> const req = mockReq({ <add> path: '/internal/some-path/that-needs/auth', <add> // eslint-disable-next-line camelcase <add> cookie: { jwt_access_token: invalidJWT } <add> }); <add> const res = mockRes(); <add> const next = sinon.spy(); <add> <add> expect(() => requestAuthorization(req, res, next)).toThrowError( <add> 'invalid signature' <add> ); <add> expect(next.called).toBe(false); <add> }); <add> <add> it('throws when the access token has expired', () => { <add> expect.assertions(2); <add> const invalidJWT = jwt.sign( <add> { accessToken: { ...accessToken, created: theBeginningOfTime } }, <add> validJWTSecret <add> ); <add> const req = mockReq({ <add> path: '/internal/some-path/that-needs/auth', <add> // eslint-disable-next-line camelcase <add> cookie: { jwt_access_token: invalidJWT } <add> }); <add> const res = mockRes(); <add> const next = sinon.spy(); <add> <add> expect(() => requestAuthorization(req, res, next)).toThrowError( <add> 'Access token is no longer vaild' <add> ); <add> expect(next.called).toBe(false); <add> }); <add> <add> it('adds the user to the request object', async done => { <add> expect.assertions(5); <add> const validJWT = jwt.sign({ accessToken }, validJWTSecret); <add> const req = mockReq({ <add> path: '/internal/some-path/that-needs/auth', <add> // eslint-disable-next-line camelcase <add> cookie: { jwt_access_token: validJWT } <add> }); <add> const res = mockRes(); <add> const next = sinon.spy(); <add> await requestAuthorization(req, res, next); <add> expect(next.called).toBe(true); <add> expect(req).toHaveProperty('user'); <add> expect(req.user).toEqual(users['456def']); <add> expect(req.user).toHaveProperty('points'); <add> expect(req.user.points).toEqual(4); <add> return done(); <add> }); <add> <add> it('calls next if request does not require authorization', async () => { <add> const req = mockReq({ path: '/unauthenticated/another/route' }); <add> const res = mockRes(); <add> const next = sinon.spy(); <add> await requestAuthorization(req, res, next); <add> expect(next.called).toBe(true); <add> }); <add> }); <add>}); <add><path>api-server/server/middlewares/request-authorization.js <del><path>api-server/server/middlewares/jwt-authorization.js <ide> import loopback from 'loopback'; <ide> import jwt from 'jsonwebtoken'; <ide> import { isBefore } from 'date-fns'; <add>import { isEmpty } from 'lodash'; <ide> <ide> import { homeLocation } from '../../../config/env'; <ide> <ide> export function isWhiteListedPath(path, whiteListREs = _whiteListREs) { <ide> return whiteListREs.some(re => re.test(path)); <ide> } <ide> <del>export default () => <del> function authorizeByJWT(req, res, next) { <add>export default ({ <add> _jwtSecret = process.env.JWT_SECRET, <add> getUserById = _getUserById <add>} = {}) => <add> function requestAuthorisation(req, res, next) { <ide> const { path } = req; <ide> if (apiProxyRE.test(path) && !isWhiteListedPath(path)) { <ide> const cookie = <ide> export default () => <ide> } <ide> let token; <ide> try { <del> token = jwt.verify(cookie, process.env.JWT_SECRET); <add> token = jwt.verify(cookie, _jwtSecret); <ide> } catch (err) { <ide> throw wrapHandledError(new Error(err.message), { <ide> type: 'info', <ide> export default () => <ide> status: 403 <ide> }); <ide> } <del> if (!req.user) { <del> const User = loopback.getModelByType('User'); <del> return User.findById(userId) <add> <add> if (isEmpty(req.user)) { <add> return getUserById(userId) <ide> .then(user => { <ide> if (user) { <ide> user.points = user.progressTimestamps.length; <ide> export default () => <ide> .then(next) <ide> .catch(next); <ide> } else { <del> return next(); <add> return Promise.resolve(next()); <ide> } <ide> } <del> return next(); <add> return Promise.resolve(next()); <ide> }; <add> <add>export function _getUserById(id) { <add> const User = loopback.getModelByType('User'); <add> return new Promise((resolve, reject) => <add> User.findById(id, (err, instance) => { <add> if (err) { <add> return reject(err); <add> } <add> return resolve(instance); <add> }) <add> ); <add>}
3
Java
Java
fix related classes for issue
6bbd921cd4e0da242b2fa97a80fdf488ebf70e44
<ide><path>rxjava-core/src/main/java/rx/internal/operators/OnSubscribeFromIterable.java <ide> public void request(long n) { <ide> } <ide> o.onNext(it.next()); <ide> } <del> o.onCompleted(); <add> if (!o.isUnsubscribed()) { <add> o.onCompleted(); <add> } <ide> } else if(n > 0) { <ide> // backpressure is requested <ide> long _c = REQUESTED_UPDATER.getAndAdd(this, n); <ide><path>rxjava-core/src/main/java/rx/internal/operators/OnSubscribeRange.java <ide> public void request(long n) { <ide> } <ide> o.onNext((int) i); <ide> } <del> o.onCompleted(); <add> if (!o.isUnsubscribed()) { <add> o.onCompleted(); <add> } <ide> } else if (n > 0) { <ide> // backpressure is requested <ide> long _c = REQUESTED_UPDATER.getAndAdd(this, n); <ide><path>rxjava-core/src/main/java/rx/internal/operators/OperatorDoOnEach.java <ide> public void onCompleted() { <ide> onError(e); <ide> return; <ide> } <del> observer.onCompleted(); <ide> // Set `done` here so that the error in `doOnEachObserver.onCompleted()` can be noticed by observer <ide> done = true; <add> observer.onCompleted(); <ide> } <ide> <ide> @Override <ide> public void onError(Throwable e) { <ide> if (done) { <ide> return; <ide> } <add> done = true; <ide> try { <ide> doOnEachObserver.onError(e); <ide> } catch (Throwable e2) { <ide> observer.onError(e2); <ide> return; <ide> } <ide> observer.onError(e); <del> done = true; <ide> } <ide> <ide> @Override <ide><path>rxjava-core/src/main/java/rx/internal/operators/OperatorTakeWhile.java <ide> public Subscriber<? super T> call(final Subscriber<? super T> subscriber) { <ide> <ide> private int counter = 0; <ide> <add> private boolean done = false; <add> <ide> @Override <ide> public void onNext(T args) { <ide> boolean isSelected; <ide> try { <ide> isSelected = predicate.call(args, counter++); <ide> } catch (Throwable e) { <add> done = true; <ide> subscriber.onError(e); <ide> unsubscribe(); <ide> return; <ide> } <ide> if (isSelected) { <ide> subscriber.onNext(args); <ide> } else { <add> done = true; <ide> subscriber.onCompleted(); <ide> unsubscribe(); <ide> } <ide> } <ide> <ide> @Override <ide> public void onCompleted() { <del> subscriber.onCompleted(); <add> if (!done) { <add> subscriber.onCompleted(); <add> } <ide> } <ide> <ide> @Override <ide> public void onError(Throwable e) { <del> subscriber.onError(e); <add> if (!done) { <add> subscriber.onError(e); <add> } <ide> } <ide> <ide> };
4
Ruby
Ruby
use short formula name in bottle commit
656c713d6c549a1ae1dc25cdc1f9bb887dfd8440
<ide><path>Library/Homebrew/cmd/bottle.rb <ide> def merge <ide> end <ide> <ide> unless ARGV.include? "--no-commit" <add> short_name = formula_name.split("/", -1).last <ide> pkg_version = bottle_hash["formula"]["pkg_version"] <ide> <ide> path.parent.cd do <ide> safe_system "git", "commit", "--no-edit", "--verbose", <del> "--message=#{formula_name}: #{update_or_add} #{pkg_version} bottle.", <add> "--message=#{short_name}: #{update_or_add} #{pkg_version} bottle.", <ide> "--", path <ide> end <ide> end
1
Ruby
Ruby
ignore processing instructions when parsing html
303d379dadaf8c5b0b4518173b2689b003c213a8
<ide><path>actionpack/lib/action_controller/vendor/html-scanner/html/tokenizer.rb <ide> def scan_tag <ide> elsif @scanner.scan(/!/) # doctype <ide> tag << @scanner.matched <ide> tag << consume_quoted_regions <add> elsif @scanner.scan(/\?/) # processing instructions, drop them <add> tag = '' <add> @scanner.scan_until(/\/>/) <ide> else <ide> tag << consume_quoted_regions <ide> end
1
PHP
PHP
update tmp constant in tests
f80a10a3b0cf385b690238f5cc5913d3e6ef3137
<ide><path>lib/Cake/Test/init.php <ide> define('APP', ROOT . DS . APP_DIR . DS); <ide> define('WWW_ROOT', APP . WEBROOT_DIR . DS); <ide> define('TESTS', APP . 'Test' . DS); <del>define('TMP', APP . 'tmp' . DS); <add>define('TMP', sys_get_temp_dir() . DS); <ide> define('LOGS', TMP . 'logs' . DS); <ide> define('CACHE', TMP . 'cache' . DS); <ide> define('CAKE_CORE_INCLUDE_PATH', ROOT . DS . 'lib'); <ide> define('CORE_TEST_CASES', CAKE . 'Test' . DS . 'TestCase'); <ide> define('LOG_ERROR', LOG_ERR); <ide> <add> <ide> @mkdir(LOGS); <ide> @mkdir(CACHE); <ide>
1
Go
Go
simplify the kill "sig" prefix stripping code
4bf70317ed3c2467f444eac9b7865b170da6366c
<ide><path>server/server.go <ide> func (srv *Server) ContainerKill(job *engine.Job) engine.Status { <ide> // The largest legal signal is 31, so let's parse on 5 bits <ide> sig, err = strconv.ParseUint(job.Args[1], 10, 5) <ide> if err != nil { <del> // The signal is not a number, treat it as a string <del> sig = uint64(signal.SignalMap[job.Args[1]]) <del> if sig == 0 && strings.HasPrefix(job.Args[1], "SIG") { <del> // If signal is prefixed with SIG, try with it stripped (ie, "SIGKILL", etc) <del> sig = uint64(signal.SignalMap[job.Args[1][3:]]) <del> } <add> // The signal is not a number, treat it as a string (either like "KILL" or like "SIGKILL") <add> sig = uint64(signal.SignalMap[strings.TrimPrefix(job.Args[1], "SIG")]) <ide> if sig == 0 { <ide> return job.Errorf("Invalid signal: %s", job.Args[1]) <ide> }
1
Text
Text
update styleguide keybinding
be561f0e5e1090cdcd0b2fae5de8b4cd29cbc21d
<ide><path>docs/creating-a-package.md <ide> Ideally, you won't need much in the way of styling. We've provided a standard <ide> set of components which define both the colors and UI elements for any package <ide> that fits into Atom seamlessly. You can view all of Atom's UI components by opening <ide> the styleguide: open the command palette (`cmd-shift-P`) and search for _styleguide_, <del>or just type `cmd-ctrl-G`. <add>or just type `cmd-ctrl-shift-G`. <ide> <ide> If you _do_ need special styling, try to keep only structural styles in the package <ide> stylesheets. If you _must_ specify colors and sizing, these should be taken from
1
Go
Go
add tests for loadmanifest digest verification
74528be903d2f6c9e07ab9a3931b050d5dae6562
<ide><path>graph/manifest_test.go <ide> import ( <ide> "os" <ide> "testing" <ide> <add> "github.com/docker/distribution/digest" <ide> "github.com/docker/docker/image" <ide> "github.com/docker/docker/pkg/tarsum" <ide> "github.com/docker/docker/registry" <ide> "github.com/docker/docker/runconfig" <ide> "github.com/docker/docker/utils" <add> "github.com/docker/libtrust" <ide> ) <ide> <ide> const ( <ide> func TestManifestTarsumCache(t *testing.T) { <ide> t.Fatalf("Unexpected json value\nExpected:\n%s\nActual:\n%s", v1compat, manifest.History[0].V1Compatibility) <ide> } <ide> } <add> <add>// TestManifestDigestCheck ensures that loadManifest properly verifies the <add>// remote and local digest. <add>func TestManifestDigestCheck(t *testing.T) { <add> tmp, err := utils.TestDirectory("") <add> if err != nil { <add> t.Fatal(err) <add> } <add> defer os.RemoveAll(tmp) <add> store := mkTestTagStore(tmp, t) <add> defer store.graph.driver.Cleanup() <add> <add> archive, err := fakeTar() <add> if err != nil { <add> t.Fatal(err) <add> } <add> img := &image.Image{ID: testManifestImageID} <add> if err := store.graph.Register(img, archive); err != nil { <add> t.Fatal(err) <add> } <add> if err := store.Tag(testManifestImageName, testManifestTag, testManifestImageID, false); err != nil { <add> t.Fatal(err) <add> } <add> <add> if cs, err := img.GetCheckSum(store.graph.ImageRoot(testManifestImageID)); err != nil { <add> t.Fatal(err) <add> } else if cs != "" { <add> t.Fatalf("Non-empty checksum file after register") <add> } <add> <add> // Generate manifest <add> payload, err := store.newManifest(testManifestImageName, testManifestImageName, testManifestTag) <add> if err != nil { <add> t.Fatalf("unexpected error generating test manifest: %v", err) <add> } <add> <add> pk, err := libtrust.GenerateECP256PrivateKey() <add> if err != nil { <add> t.Fatalf("unexpected error generating private key: %v", err) <add> } <add> <add> sig, err := libtrust.NewJSONSignature(payload) <add> if err != nil { <add> t.Fatalf("error creating signature: %v", err) <add> } <add> <add> if err := sig.Sign(pk); err != nil { <add> t.Fatalf("error signing manifest bytes: %v", err) <add> } <add> <add> signedBytes, err := sig.PrettySignature("signatures") <add> if err != nil { <add> t.Fatalf("error getting signed bytes: %v", err) <add> } <add> <add> dgst, err := digest.FromBytes(payload) <add> if err != nil { <add> t.Fatalf("error getting digest of manifest: %v", err) <add> } <add> <add> // use this as the "bad" digest <add> zeroDigest, err := digest.FromBytes([]byte{}) <add> if err != nil { <add> t.Fatalf("error making zero digest: %v", err) <add> } <add> <add> // Remote and local match, everything should look good <add> local, _, _, err := store.loadManifest(signedBytes, dgst.String(), dgst) <add> if err != nil { <add> t.Fatalf("unexpected error verifying local and remote digest: %v", err) <add> } <add> <add> if local != dgst { <add> t.Fatalf("local digest not correctly calculated: %v", err) <add> } <add> <add> // remote and no local, since pulling by tag <add> local, _, _, err = store.loadManifest(signedBytes, "tag", dgst) <add> if err != nil { <add> t.Fatalf("unexpected error verifying tag pull and remote digest: %v", err) <add> } <add> <add> if local != dgst { <add> t.Fatalf("local digest not correctly calculated: %v", err) <add> } <add> <add> // remote and differing local, this is the most important to fail <add> local, _, _, err = store.loadManifest(signedBytes, zeroDigest.String(), dgst) <add> if err == nil { <add> t.Fatalf("error expected when verifying with differing local digest") <add> } <add> <add> // no remote, no local (by tag) <add> local, _, _, err = store.loadManifest(signedBytes, "tag", "") <add> if err != nil { <add> t.Fatalf("unexpected error verifying manifest without remote digest: %v", err) <add> } <add> <add> if local != dgst { <add> t.Fatalf("local digest not correctly calculated: %v", err) <add> } <add> <add> // no remote, with local <add> local, _, _, err = store.loadManifest(signedBytes, dgst.String(), "") <add> if err != nil { <add> t.Fatalf("unexpected error verifying manifest without remote digest: %v", err) <add> } <add> <add> if local != dgst { <add> t.Fatalf("local digest not correctly calculated: %v", err) <add> } <add> <add> // bad remote, we fail the check. <add> local, _, _, err = store.loadManifest(signedBytes, dgst.String(), zeroDigest) <add> if err == nil { <add> t.Fatalf("error expected when verifying with differing remote digest") <add> } <add>} <ide><path>graph/tags_unit_test.go <ide> import ( <ide> "github.com/docker/docker/daemon/graphdriver" <ide> _ "github.com/docker/docker/daemon/graphdriver/vfs" // import the vfs driver so it is used in the tests <ide> "github.com/docker/docker/image" <add> "github.com/docker/docker/trust" <ide> "github.com/docker/docker/utils" <ide> ) <ide> <ide> func mkTestTagStore(root string, t *testing.T) *TagStore { <ide> if err != nil { <ide> t.Fatal(err) <ide> } <add> <add> trust, err := trust.NewTrustStore(root + "/trust") <add> if err != nil { <add> t.Fatal(err) <add> } <add> <ide> tagCfg := &TagStoreConfig{ <ide> Graph: graph, <ide> Events: events.New(), <add> Trust: trust, <ide> } <ide> store, err := NewTagStore(path.Join(root, "tags"), tagCfg) <ide> if err != nil {
2
Ruby
Ruby
make `in_batches` queries to preparable
7d5399379cb9ac2d3ffafa28fdc844d7b6c18ab8
<ide><path>activerecord/lib/active_record/relation/batches.rb <ide> def in_batches(of: 1000, start: nil, finish: nil, load: false, error_on_ignore: <ide> end <ide> end <ide> <del> batch_relation = relation.where(arel_attribute(primary_key).gt(primary_key_offset)) <add> attr = Relation::QueryAttribute.new(primary_key, primary_key_offset, klass.type_for_attribute(primary_key)) <add> batch_relation = relation.where(arel_attribute(primary_key).gt(Arel::Nodes::BindParam.new(attr))) <ide> end <ide> end <ide> <ide> private <ide> <ide> def apply_limits(relation, start, finish) <del> relation = relation.where(arel_attribute(primary_key).gteq(start)) if start <del> relation = relation.where(arel_attribute(primary_key).lteq(finish)) if finish <add> if start <add> attr = Relation::QueryAttribute.new(primary_key, start, klass.type_for_attribute(primary_key)) <add> relation = relation.where(arel_attribute(primary_key).gteq(Arel::Nodes::BindParam.new(attr))) <add> end <add> if finish <add> attr = Relation::QueryAttribute.new(primary_key, finish, klass.type_for_attribute(primary_key)) <add> relation = relation.where(arel_attribute(primary_key).lteq(Arel::Nodes::BindParam.new(attr))) <add> end <ide> relation <ide> end <ide>
1
Text
Text
add docker commands to local setup
1d6c43c612fbb8cc4d7f76d0ab07f63eb59a9960
<ide><path>docs/how-to-setup-freecodecamp-locally.md <ide> A quick reference to the commands that you will need when working locally. <ide> <ide> **Docker Build:** <ide> <del>> ### Todo: Add docker quick command list <add>| command | description | <add>| ------- | ----------- | <add>| `npm run docker:init` | Prepare containers for installation of dependencies. | <add>| `npm run docker:install` | Install / re-install all dependencies and bootstraps the different services. | <add>| `npm run docker:seed` | Parse all the challenge markdown files and inserts them into MongoDB. | <add>| `npm run docker:develop` | Start the freeCodeCamp API Server and Client Applications. | <add>| `npm run docker:test:init` | Bootstrap the test container, necessary for testing in docker. | <add>| `npm run docker:test -- -c "npm run test` | Run all JS tests in the system, including client, server, lint and challenge tests. | <add>| `npm run docker:test -- -c "npm run test:curriculum` | Run the curriculum test suite. | <add>| `npm run docker:test -- -c "npm run test:client` | Run the client test suite. | <add>| `npm run docker:test -- -c "npm run test:server` | Run the server test suite. | <add>| `npm run docker:clean` | Uninstall all dependencies and cleans up caches. | <ide> <ide> **Local Build:** <ide>
1
Text
Text
fix typo in actionpack changelog.md
c3d5c9447a49fae7a4507b3512d65cd52f789d99
<ide><path>actionpack/CHANGELOG.md <ide> <ide> *Gustavo Gutierrez* <ide> <del>* Calling `ActionController::Parameters#transform_keys/!` without a block now returns <add>* Calling `ActionController::Parameters#transform_keys!` without a block now returns <ide> an enumerator for the parameters instead of the underlying hash. <ide> <ide> *Eugene Kenny*
1
Javascript
Javascript
add placeholder text to forum posts
90fe00937431b68f1af13f31a4a0af6a935ae130
<ide><path>client/src/templates/Challenges/redux/create-question-epic.js <ide> function createQuestionEpic(action$, state$, { window }) { <ide> <ide> let textMessage = dedent( <ide> `**Tell us what's happening:** <add> Describe your issue in detail here. <ide> <ide> ${ <ide> projectFormValues.length
1
Text
Text
prepare providers release 0.0.2a1
5a439e84eb6c0544dc6c3d6a9f4ceeb2172cd5d0
<ide><path>airflow/providers/amazon/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [8afdb6ac6](https://github.com/apache/airflow/commit/8afdb6ac6a7997cb14806bc2734c81c00ed8da97) | 2020-10-26 | Fix spellings (#11825) | <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [6ce855af1](https://github.com/apache/airflow/commit/6ce855af118daeaa4c249669079ab9d9aad23945) | 2020-10-24 | Fix spelling (#11821) | <add>| [3934ef224](https://github.com/apache/airflow/commit/3934ef22494db6d9613c229aaa82ea6a366b7c2f) | 2020-10-24 | Remove redundant builtins imports (#11809) | <add>| [4c8e033c0](https://github.com/apache/airflow/commit/4c8e033c0ee7d28963d504a9216205155f20f58f) | 2020-10-24 | Fix spelling and grammar (#11814) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [0df60b773](https://github.com/apache/airflow/commit/0df60b773671ecf8d4e5f582ac2be200cf2a2edd) | 2020-10-23 | Add reattach flag to ECSOperator (#10643) | <add>| [b9d677cdd](https://github.com/apache/airflow/commit/b9d677cdd660e0be8278a64658e73359276a9682) | 2020-10-22 | Add type hints to aws provider (#11531) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [674368f66](https://github.com/apache/airflow/commit/674368f66cf61b2a105f326f23868ac3aee08807) | 2020-10-19 | Fixes MySQLToS3 float to int conversion (#10437) | <add>| [0823d46a7](https://github.com/apache/airflow/commit/0823d46a7f267f2e45195a175021825367938add) | 2020-10-16 | Add type annotations for AWS operators and hooks (#11434) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/amazon/README.md <ide> <ide> # Package apache-airflow-providers-amazon <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Secrets](#secrets) <ide> - [Moved secrets](#moved-secrets) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [8afdb6ac6](https://github.com/apache/airflow/commit/8afdb6ac6a7997cb14806bc2734c81c00ed8da97) | 2020-10-26 | Fix spellings (#11825) | <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [6ce855af1](https://github.com/apache/airflow/commit/6ce855af118daeaa4c249669079ab9d9aad23945) | 2020-10-24 | Fix spelling (#11821) | <add>| [3934ef224](https://github.com/apache/airflow/commit/3934ef22494db6d9613c229aaa82ea6a366b7c2f) | 2020-10-24 | Remove redundant builtins imports (#11809) | <add>| [4c8e033c0](https://github.com/apache/airflow/commit/4c8e033c0ee7d28963d504a9216205155f20f58f) | 2020-10-24 | Fix spelling and grammar (#11814) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [0df60b773](https://github.com/apache/airflow/commit/0df60b773671ecf8d4e5f582ac2be200cf2a2edd) | 2020-10-23 | Add reattach flag to ECSOperator (#10643) | <add>| [b9d677cdd](https://github.com/apache/airflow/commit/b9d677cdd660e0be8278a64658e73359276a9682) | 2020-10-22 | Add type hints to aws provider (#11531) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [674368f66](https://github.com/apache/airflow/commit/674368f66cf61b2a105f326f23868ac3aee08807) | 2020-10-19 | Fixes MySQLToS3 float to int conversion (#10437) | <add>| [0823d46a7](https://github.com/apache/airflow/commit/0823d46a7f267f2e45195a175021825367938add) | 2020-10-16 | Add type annotations for AWS operators and hooks (#11434) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/apache/cassandra/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [0646849e3](https://github.com/apache/airflow/commit/0646849e3dacdc2bc62705ae136f3ad3b16232e9) | 2020-10-14 | Add protocol_version to conn_config for Cassandrahook (#11036) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/apache/cassandra/README.md <ide> <ide> # Package apache-airflow-providers-apache-cassandra <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [0646849e3](https://github.com/apache/airflow/commit/0646849e3dacdc2bc62705ae136f3ad3b16232e9) | 2020-10-14 | Add protocol_version to conn_config for Cassandrahook (#11036) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/apache/druid/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/apache/druid/README.md <ide> <ide> # Package apache-airflow-providers-apache-druid <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/apache/hdfs/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/apache/hdfs/README.md <ide> <ide> # Package apache-airflow-providers-apache-hdfs <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/apache/hive/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/apache/hive/README.md <ide> <ide> # Package apache-airflow-providers-apache-hive <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/apache/kylin/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/apache/kylin/README.md <ide> <ide> # Package apache-airflow-providers-apache-kylin <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [New hooks](#new-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/apache/livy/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/apache/livy/README.md <ide> <ide> # Package apache-airflow-providers-apache-livy <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [New hooks](#new-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/apache/pig/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/apache/pig/README.md <ide> <ide> # Package apache-airflow-providers-apache-pig <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/apache/pinot/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/apache/pinot/README.md <ide> <ide> # Package apache-airflow-providers-apache-pinot <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/apache/spark/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/apache/spark/README.md <ide> <ide> # Package apache-airflow-providers-apache-spark <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/apache/sqoop/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/apache/sqoop/README.md <ide> <ide> # Package apache-airflow-providers-apache-sqoop <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/celery/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/celery/README.md <ide> <ide> # Package apache-airflow-providers-celery <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Sensors](#sensors) <ide> - [Moved sensors](#moved-sensors) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/cloudant/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/cloudant/README.md <ide> <ide> # Package apache-airflow-providers-cloudant <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/cncf/kubernetes/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [53e606210](https://github.com/apache/airflow/commit/53e6062105be0ae1761a354e2055eb0779d12e73) | 2020-10-21 | Enforce strict rules for yamllint (#11709) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [eee4e30f2](https://github.com/apache/airflow/commit/eee4e30f2caf02e16088ff5d1af1ea380a73e982) | 2020-10-15 | Add better debug logging to K8sexec and K8sPodOp (#11502) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/cncf/kubernetes/README.md <ide> <ide> # Package apache-airflow-providers-cncf-kubernetes <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [New hooks](#new-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [53e606210](https://github.com/apache/airflow/commit/53e6062105be0ae1761a354e2055eb0779d12e73) | 2020-10-21 | Enforce strict rules for yamllint (#11709) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [eee4e30f2](https://github.com/apache/airflow/commit/eee4e30f2caf02e16088ff5d1af1ea380a73e982) | 2020-10-15 | Add better debug logging to K8sexec and K8sPodOp (#11502) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/databricks/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/databricks/README.md <ide> <ide> # Package apache-airflow-providers-databricks <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/datadog/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/datadog/README.md <ide> <ide> # Package apache-airflow-providers-datadog <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/dingding/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [172820db4](https://github.com/apache/airflow/commit/172820db4d2009dd26fa8aef4a864fb8a3d7e78d) | 2020-10-21 | Fix case of GitHub (#11398) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/dingding/README.md <ide> <ide> # Package apache-airflow-providers-dingding <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [172820db4](https://github.com/apache/airflow/commit/172820db4d2009dd26fa8aef4a864fb8a3d7e78d) | 2020-10-21 | Fix case of GitHub (#11398) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/discord/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/discord/README.md <ide> <ide> # Package apache-airflow-providers-discord <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/docker/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/docker/README.md <ide> <ide> # Package apache-airflow-providers-docker <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/elasticsearch/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/elasticsearch/README.md <ide> <ide> # Package apache-airflow-providers-elasticsearch <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [New hooks](#new-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/exasol/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/exasol/README.md <ide> <ide> # Package apache-airflow-providers-exasol <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [New hooks](#new-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/facebook/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/facebook/README.md <ide> <ide> # Package apache-airflow-providers-facebook <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [New hooks](#new-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/ftp/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/ftp/README.md <ide> <ide> # Package apache-airflow-providers-ftp <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/google/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------| <add>| [240c7d4d7](https://github.com/apache/airflow/commit/240c7d4d72aac8f6aab98f5913e8f54c4f1372ff) | 2020-10-26 | Google Memcached hooks - improve protobuf messages handling (#11743) | <add>| [8afdb6ac6](https://github.com/apache/airflow/commit/8afdb6ac6a7997cb14806bc2734c81c00ed8da97) | 2020-10-26 | Fix spellings (#11825) | <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [6ce855af1](https://github.com/apache/airflow/commit/6ce855af118daeaa4c249669079ab9d9aad23945) | 2020-10-24 | Fix spelling (#11821) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [727c739af](https://github.com/apache/airflow/commit/727c739afb565d4d394a8faedc969334cb8e738e) | 2020-10-22 | Improve Cloud Memorystore for Redis example (#11735) | <add>| [1da8379c9](https://github.com/apache/airflow/commit/1da8379c913843834353b44861c62f332a461bdf) | 2020-10-22 | Fix static checks after merging #10121 (#11737) | <add>| [91503308c](https://github.com/apache/airflow/commit/91503308c723b186ce6f4026f2a3e2c21030f6e5) | 2020-10-22 | Add Google Cloud Memorystore Memcached Operators (#10121) | <add>| [950c16d0b](https://github.com/apache/airflow/commit/950c16d0b0ab67bb7af11909de751029faf0313a) | 2020-10-21 | Retry requests in case of error in Google ML Engine Hook (#11712) | <add>| [2bfc53b5e](https://github.com/apache/airflow/commit/2bfc53b5eb67406d418371b74dc9bc5a07be238e) | 2020-10-21 | Fix doc errors in google provider files. (#11713) | <add>| [53e606210](https://github.com/apache/airflow/commit/53e6062105be0ae1761a354e2055eb0779d12e73) | 2020-10-21 | Enforce strict rules for yamllint (#11709) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [2d854c350](https://github.com/apache/airflow/commit/2d854c3505ccad66e9a7d94267e51bed800433c2) | 2020-10-19 | Add service_account to Google ML Engine operator (#11619) | <add>| [46a121fb7](https://github.com/apache/airflow/commit/46a121fb7b77c0964e053b58750e2d8bc2bd0b2a) | 2020-10-18 | docs: Update Bigquery clustering docstrings (#11232) | <add>| [49c58147f](https://github.com/apache/airflow/commit/49c58147fed8a52869d0b0ecc00c102c11972ad0) | 2020-10-18 | Strict type checking for provider Google (#11609) | <add>| [0823d46a7](https://github.com/apache/airflow/commit/0823d46a7f267f2e45195a175021825367938add) | 2020-10-16 | Add type annotations for AWS operators and hooks (#11434) | <add>| [3c10ca650](https://github.com/apache/airflow/commit/3c10ca6504be37fabff9a10caefea3fe4df31a02) | 2020-10-16 | Add DataflowStartFlexTemplateOperator (#8550) | <add>| [8865d14df](https://github.com/apache/airflow/commit/8865d14df4d58dd5f1a4d2ff81c77469959f175a) | 2020-10-16 | Strict type checking for provider google cloud (#11548) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/google/README.md <ide> <ide> # Package apache-airflow-providers-google <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Secrets](#secrets) <ide> - [Moved secrets](#moved-secrets) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> You can install this package on top of an existing airflow 2.* installation via <ide> | google-cloud-kms | &gt;=1.2.1,&lt;2.0.0 | <ide> | google-cloud-language | &gt;=1.1.1,&lt;2.0.0 | <ide> | google-cloud-logging | &gt;=1.14.0,&lt;2.0.0 | <add>| google-cloud-memcache | &gt;=0.2.0 | <ide> | google-cloud-monitoring | &gt;=0.34.0,&lt;2.0.0 | <ide> | google-cloud-pubsub | &gt;=1.0.0,&lt;2.0.0 | <ide> | google-cloud-redis | &gt;=0.3.0,&lt;2.0.0 | <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> | [cloud.operators.cloud_memorystore.CloudMemorystoreGetInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) | <ide> | [cloud.operators.cloud_memorystore.CloudMemorystoreImportOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) | <ide> | [cloud.operators.cloud_memorystore.CloudMemorystoreListInstancesOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) | <add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedApplyParametersOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) | <add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedCreateInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) | <add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedDeleteInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) | <add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedGetInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) | <add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedListInstancesOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) | <add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedUpdateInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) | <add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedUpdateParametersOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) | <ide> | [cloud.operators.cloud_memorystore.CloudMemorystoreScaleInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) | <ide> | [cloud.operators.cloud_memorystore.CloudMemorystoreUpdateInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) | <ide> | [cloud.operators.cloud_storage_transfer_service.CloudDataTransferServiceGCSToGCSOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_storage_transfer_service.py) | <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> | [cloud.operators.datacatalog.CloudDataCatalogUpdateTagOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/datacatalog.py) | <ide> | [cloud.operators.datacatalog.CloudDataCatalogUpdateTagTemplateFieldOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/datacatalog.py) | <ide> | [cloud.operators.datacatalog.CloudDataCatalogUpdateTagTemplateOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/datacatalog.py) | <add>| [cloud.operators.dataflow.DataflowStartFlexTemplateOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/dataflow.py) | <ide> | [cloud.operators.datafusion.CloudDataFusionCreateInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/datafusion.py) | <ide> | [cloud.operators.datafusion.CloudDataFusionCreatePipelineOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/datafusion.py) | <ide> | [cloud.operators.datafusion.CloudDataFusionDeleteInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/datafusion.py) | <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> | [cloud.hooks.automl.CloudAutoMLHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/automl.py) | <ide> | [cloud.hooks.bigquery_dts.BiqQueryDataTransferServiceHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/bigquery_dts.py) | <ide> | [cloud.hooks.cloud_memorystore.CloudMemorystoreHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/cloud_memorystore.py) | <add>| [cloud.hooks.cloud_memorystore.CloudMemorystoreMemcachedHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/cloud_memorystore.py) | <ide> | [cloud.hooks.datacatalog.CloudDataCatalogHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/datacatalog.py) | <ide> | [cloud.hooks.datafusion.DataFusionHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/datafusion.py) | <ide> | [cloud.hooks.dataprep.GoogleDataprepHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/dataprep.py) | <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------| <add>| [240c7d4d7](https://github.com/apache/airflow/commit/240c7d4d72aac8f6aab98f5913e8f54c4f1372ff) | 2020-10-26 | Google Memcached hooks - improve protobuf messages handling (#11743) | <add>| [8afdb6ac6](https://github.com/apache/airflow/commit/8afdb6ac6a7997cb14806bc2734c81c00ed8da97) | 2020-10-26 | Fix spellings (#11825) | <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [6ce855af1](https://github.com/apache/airflow/commit/6ce855af118daeaa4c249669079ab9d9aad23945) | 2020-10-24 | Fix spelling (#11821) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [727c739af](https://github.com/apache/airflow/commit/727c739afb565d4d394a8faedc969334cb8e738e) | 2020-10-22 | Improve Cloud Memorystore for Redis example (#11735) | <add>| [1da8379c9](https://github.com/apache/airflow/commit/1da8379c913843834353b44861c62f332a461bdf) | 2020-10-22 | Fix static checks after merging #10121 (#11737) | <add>| [91503308c](https://github.com/apache/airflow/commit/91503308c723b186ce6f4026f2a3e2c21030f6e5) | 2020-10-22 | Add Google Cloud Memorystore Memcached Operators (#10121) | <add>| [950c16d0b](https://github.com/apache/airflow/commit/950c16d0b0ab67bb7af11909de751029faf0313a) | 2020-10-21 | Retry requests in case of error in Google ML Engine Hook (#11712) | <add>| [2bfc53b5e](https://github.com/apache/airflow/commit/2bfc53b5eb67406d418371b74dc9bc5a07be238e) | 2020-10-21 | Fix doc errors in google provider files. (#11713) | <add>| [53e606210](https://github.com/apache/airflow/commit/53e6062105be0ae1761a354e2055eb0779d12e73) | 2020-10-21 | Enforce strict rules for yamllint (#11709) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [2d854c350](https://github.com/apache/airflow/commit/2d854c3505ccad66e9a7d94267e51bed800433c2) | 2020-10-19 | Add service_account to Google ML Engine operator (#11619) | <add>| [46a121fb7](https://github.com/apache/airflow/commit/46a121fb7b77c0964e053b58750e2d8bc2bd0b2a) | 2020-10-18 | docs: Update Bigquery clustering docstrings (#11232) | <add>| [49c58147f](https://github.com/apache/airflow/commit/49c58147fed8a52869d0b0ecc00c102c11972ad0) | 2020-10-18 | Strict type checking for provider Google (#11609) | <add>| [0823d46a7](https://github.com/apache/airflow/commit/0823d46a7f267f2e45195a175021825367938add) | 2020-10-16 | Add type annotations for AWS operators and hooks (#11434) | <add>| [3c10ca650](https://github.com/apache/airflow/commit/3c10ca6504be37fabff9a10caefea3fe4df31a02) | 2020-10-16 | Add DataflowStartFlexTemplateOperator (#8550) | <add>| [8865d14df](https://github.com/apache/airflow/commit/8865d14df4d58dd5f1a4d2ff81c77469959f175a) | 2020-10-16 | Strict type checking for provider google cloud (#11548) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/grpc/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/grpc/README.md <ide> <ide> # Package apache-airflow-providers-grpc <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [New hooks](#new-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/hashicorp/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/hashicorp/README.md <ide> <ide> # Package apache-airflow-providers-hashicorp <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Secrets](#secrets) <ide> - [Moved secrets](#moved-secrets) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/http/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [3cddc1182](https://github.com/apache/airflow/commit/3cddc11821ff8f9ed0811384c0643f756a2b3dfa) | 2020-10-16 | Updated template_fields_rendereds for PostgresOperator and SimpleHttpOperator (#11555) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/http/README.md <ide> <ide> # Package apache-airflow-providers-http <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [3cddc1182](https://github.com/apache/airflow/commit/3cddc11821ff8f9ed0811384c0643f756a2b3dfa) | 2020-10-16 | Updated template_fields_rendereds for PostgresOperator and SimpleHttpOperator (#11555) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/imap/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/imap/README.md <ide> <ide> # Package apache-airflow-providers-imap <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/jdbc/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/jdbc/README.md <ide> <ide> # Package apache-airflow-providers-jdbc <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/jenkins/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [8afdb6ac6](https://github.com/apache/airflow/commit/8afdb6ac6a7997cb14806bc2734c81c00ed8da97) | 2020-10-26 | Fix spellings (#11825) | <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/jenkins/README.md <ide> <ide> # Package apache-airflow-providers-jenkins <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [8afdb6ac6](https://github.com/apache/airflow/commit/8afdb6ac6a7997cb14806bc2734c81c00ed8da97) | 2020-10-26 | Fix spellings (#11825) | <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/jira/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/jira/README.md <ide> <ide> # Package apache-airflow-providers-jira <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/microsoft/azure/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [6ce855af1](https://github.com/apache/airflow/commit/6ce855af118daeaa4c249669079ab9d9aad23945) | 2020-10-24 | Fix spelling (#11821) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [f8ff217e2](https://github.com/apache/airflow/commit/f8ff217e2f2152bbb9fc701ff4c0b6eb447ad65c) | 2020-10-18 | Fix incorrect typing and move config args out of extra connection config to operator args (#11635) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/microsoft/azure/README.md <ide> <ide> # Package apache-airflow-providers-microsoft-azure <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Secrets](#secrets) <ide> - [New secrets](#new-secrets) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [6ce855af1](https://github.com/apache/airflow/commit/6ce855af118daeaa4c249669079ab9d9aad23945) | 2020-10-24 | Fix spelling (#11821) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [f8ff217e2](https://github.com/apache/airflow/commit/f8ff217e2f2152bbb9fc701ff4c0b6eb447ad65c) | 2020-10-18 | Fix incorrect typing and move config args out of extra connection config to operator args (#11635) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/microsoft/mssql/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [765d29ecc](https://github.com/apache/airflow/commit/765d29ecc9fd6a3220efa0a6c4ce10848f5cbf82) | 2020-10-15 | Pymssql is maintained again (#11537) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/microsoft/mssql/README.md <ide> <ide> # Package apache-airflow-providers-microsoft-mssql <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> You can install this package on top of an existing airflow 2.* installation via <ide> <ide> | PIP package | Version required | <ide> |:--------------|:-------------------| <del>| pymssql | ~=2.1.1 | <add>| pymssql | ~=2.1,&gt;=2.1.5 | <ide> <ide> ## Cross provider package dependencies <ide> <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [765d29ecc](https://github.com/apache/airflow/commit/765d29ecc9fd6a3220efa0a6c4ce10848f5cbf82) | 2020-10-15 | Pymssql is maintained again (#11537) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/microsoft/winrm/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/microsoft/winrm/README.md <ide> <ide> # Package apache-airflow-providers-microsoft-winrm <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/mongo/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/mongo/README.md <ide> <ide> # Package apache-airflow-providers-mongo <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/mysql/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/mysql/README.md <ide> <ide> # Package apache-airflow-providers-mysql <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/odbc/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/odbc/README.md <ide> <ide> # Package apache-airflow-providers-odbc <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [New hooks](#new-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/openfaas/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/openfaas/README.md <ide> <ide> # Package apache-airflow-providers-openfaas <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/opsgenie/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/opsgenie/README.md <ide> <ide> # Package apache-airflow-providers-opsgenie <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/oracle/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/oracle/README.md <ide> <ide> # Package apache-airflow-providers-oracle <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/pagerduty/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [3ee618623](https://github.com/apache/airflow/commit/3ee618623be6079ed177da793b490cb7436d5cb6) | 2020-10-20 | Switch PagerdutyHook from pypd to use pdpyras instead (#11151) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/pagerduty/README.md <ide> <ide> # Package apache-airflow-providers-pagerduty <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [3ee618623](https://github.com/apache/airflow/commit/3ee618623be6079ed177da793b490cb7436d5cb6) | 2020-10-20 | Switch PagerdutyHook from pypd to use pdpyras instead (#11151) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/papermill/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [cb7c67dea](https://github.com/apache/airflow/commit/cb7c67dea9cd9b9c5de10e355b63039446003149) | 2020-10-20 | Fix example DAGs in pip packages (#11687) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/papermill/README.md <ide> <ide> # Package apache-airflow-providers-papermill <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Operators](#operators) <ide> - [Moved operators](#moved-operators) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [cb7c67dea](https://github.com/apache/airflow/commit/cb7c67dea9cd9b9c5de10e355b63039446003149) | 2020-10-20 | Fix example DAGs in pip packages (#11687) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/plexus/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [56d72e3ff](https://github.com/apache/airflow/commit/56d72e3ff8798a2662847355d1b73b2c1f57b31f) | 2020-10-24 | Replace non-empty sets with set literals (#11810) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/plexus/README.md <ide> <ide> # Package apache-airflow-providers-plexus <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [New hooks](#new-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [56d72e3ff](https://github.com/apache/airflow/commit/56d72e3ff8798a2662847355d1b73b2c1f57b31f) | 2020-10-24 | Replace non-empty sets with set literals (#11810) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/postgres/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [a4dc11fae](https://github.com/apache/airflow/commit/a4dc11fae63d56bc6cbb029525113948862fd45d) | 2020-10-19 | Change to pass all extra connection paramaters to psycopg2 (#11019) | <add>| [3cddc1182](https://github.com/apache/airflow/commit/3cddc11821ff8f9ed0811384c0643f756a2b3dfa) | 2020-10-16 | Updated template_fields_rendereds for PostgresOperator and SimpleHttpOperator (#11555) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/postgres/README.md <ide> <ide> # Package apache-airflow-providers-postgres <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [a4dc11fae](https://github.com/apache/airflow/commit/a4dc11fae63d56bc6cbb029525113948862fd45d) | 2020-10-19 | Change to pass all extra connection paramaters to psycopg2 (#11019) | <add>| [3cddc1182](https://github.com/apache/airflow/commit/3cddc11821ff8f9ed0811384c0643f756a2b3dfa) | 2020-10-16 | Updated template_fields_rendereds for PostgresOperator and SimpleHttpOperator (#11555) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/presto/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [1543923c1](https://github.com/apache/airflow/commit/1543923c197f658533ca0a0bb259b59a002cce43) | 2020-10-20 | Add Kerberos Auth for PrestoHook (#10488) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/presto/README.md <ide> <ide> # Package apache-airflow-providers-presto <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [1543923c1](https://github.com/apache/airflow/commit/1543923c197f658533ca0a0bb259b59a002cce43) | 2020-10-20 | Add Kerberos Auth for PrestoHook (#10488) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/qubole/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/qubole/README.md <ide> <ide> # Package apache-airflow-providers-qubole <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/redis/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/redis/README.md <ide> <ide> # Package apache-airflow-providers-redis <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/salesforce/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/salesforce/README.md <ide> <ide> # Package apache-airflow-providers-salesforce <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [New hooks](#new-hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/samba/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/samba/README.md <ide> <ide> # Package apache-airflow-providers-samba <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/segment/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/segment/README.md <ide> <ide> # Package apache-airflow-providers-segment <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/sftp/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [ae791e191](https://github.com/apache/airflow/commit/ae791e19163b4ee6e84940d2567adbf0a2626fb4) | 2020-10-21 | Fix formatting errors introduced in #11720 (#11733) | <add>| [1fb3c28e1](https://github.com/apache/airflow/commit/1fb3c28e1a4fd54c9d83dccd413659a7a87c7315) | 2020-10-21 | Add support for setting ciphers for SFTPHook (#11720) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/sftp/README.md <ide> <ide> # Package apache-airflow-providers-sftp <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [ae791e191](https://github.com/apache/airflow/commit/ae791e19163b4ee6e84940d2567adbf0a2626fb4) | 2020-10-21 | Fix formatting errors introduced in #11720 (#11733) | <add>| [1fb3c28e1](https://github.com/apache/airflow/commit/1fb3c28e1a4fd54c9d83dccd413659a7a87c7315) | 2020-10-21 | Add support for setting ciphers for SFTPHook (#11720) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/singularity/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/singularity/README.md <ide> <ide> # Package apache-airflow-providers-singularity <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Operators](#operators) <ide> - [New operators](#new-operators) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/slack/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [4fb5c017f](https://github.com/apache/airflow/commit/4fb5c017fe5ca41ed95547a857c9c39efc4f1476) | 2020-10-21 | Check response status in slack webhook hook. (#11620) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/slack/README.md <ide> <ide> # Package apache-airflow-providers-slack <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [4fb5c017f](https://github.com/apache/airflow/commit/4fb5c017fe5ca41ed95547a857c9c39efc4f1476) | 2020-10-21 | Check response status in slack webhook hook. (#11620) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/snowflake/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/snowflake/README.md <ide> <ide> # Package apache-airflow-providers-snowflake <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [New hooks](#new-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/sqlite/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/sqlite/README.md <ide> <ide> # Package apache-airflow-providers-sqlite <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/ssh/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/ssh/README.md <ide> <ide> # Package apache-airflow-providers-ssh <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/vertica/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/vertica/README.md <ide> <ide> # Package apache-airflow-providers-vertica <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/yandex/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/yandex/README.md <ide> <ide> # Package apache-airflow-providers-yandex <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [New hooks](#new-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject | <ide><path>airflow/providers/zendesk/PROVIDER_CHANGES_0.0.2a1.md <add> <add> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <ide><path>airflow/providers/zendesk/README.md <ide> <ide> # Package apache-airflow-providers-zendesk <ide> <del>Release: 0.0.1 <add>Release: 0.0.2a1 <ide> <ide> **Table of contents** <ide> <ide> Release: 0.0.1 <ide> - [Hooks](#hooks) <ide> - [Moved hooks](#moved-hooks) <ide> - [Releases](#releases) <add> - [Release 0.0.2a1](#release-002a1) <ide> - [Release 0.0.1](#release-001) <ide> <ide> ## Provider package <ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/ <ide> <ide> ## Releases <ide> <add>### Release 0.0.2a1 <add> <add>| Commit | Committed | Subject | <add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------| <add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) | <add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) | <add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) | <add> <add> <ide> ### Release 0.0.1 <ide> <ide> | Commit | Committed | Subject |
118
Ruby
Ruby
remove unused column types override
ea6f28c8a114396376bafe26e1af5d2ca3df6fdb
<ide><path>activerecord/lib/active_record/core.rb <ide> def initialize(attributes = nil, options = {}) <ide> end <ide> <ide> @attributes = defaults <del> @column_types_override = nil <ide> @column_types = self.class.column_types <ide> <ide> init_internals <ide> def initialize(attributes = nil, options = {}) <ide> # post.title # => 'hello world' <ide> def init_with(coder) <ide> @attributes = coder['attributes'] <del> @column_types_override = coder['column_types'] <ide> @column_types = self.class.column_types <ide> <ide> init_internals <ide> def encode_with(coder) <ide> # FIXME: Remove this when we better serialize attributes <ide> coder['raw_attributes'] = attributes_before_type_cast <ide> coder['attributes'] = @attributes <del> coder['column_types'] = @column_types_override <ide> coder['new_record'] = new_record? <ide> end <ide> <ide><path>activerecord/lib/active_record/persistence.rb <ide> def instantiate(attributes, column_types = {}) <ide> type = column_types.fetch(name) { klass.type_for_attribute(name) } <ide> h[name] = Attribute.from_database(value, type) <ide> end <del> <del> klass.allocate.init_with( <del> 'attributes' => attributes, <del> 'column_types' => column_types, <del> 'new_record' => false, <del> ) <add> klass.allocate.init_with('attributes' => attributes, 'new_record' => false) <ide> end <ide> <ide> private <ide> def reload(options = nil) <ide> <ide> @attributes.update(fresh_object.instance_variable_get('@attributes')) <ide> <del> @column_types = self.class.column_types <del> @column_types_override = fresh_object.instance_variable_get('@column_types_override') <add> @column_types = self.class.column_types <ide> self <ide> end <ide>
2
Python
Python
move trainers to core/
cbbba228150140a14eea2361868bf66700feff4b
<ide><path>official/core/base_task_test.py <add># Lint as: python3 <add># Copyright 2020 The TensorFlow Authors. All Rights Reserved. <add># <add># Licensed under the Apache License, Version 2.0 (the "License"); <add># you may not use this file except in compliance with the License. <add># You may obtain a copy of the License at <add># <add># http://www.apache.org/licenses/LICENSE-2.0 <add># <add># Unless required by applicable law or agreed to in writing, software <add># distributed under the License is distributed on an "AS IS" BASIS, <add># WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. <add># See the License for the specific language governing permissions and <add># limitations under the License. <add># ============================================================================== <add>"""Tests for tensorflow_models.core.base_task.""" <add> <add>import functools <add> <add>from absl.testing import parameterized <add>import tensorflow as tf <add> <add>from tensorflow.python.distribute import combinations <add>from tensorflow.python.distribute import strategy_combinations <add>from official.utils.testing import mock_task <add> <add> <add>def all_strategy_combinations(): <add> return combinations.combine( <add> distribution=[ <add> strategy_combinations.default_strategy, <add> strategy_combinations.tpu_strategy, <add> strategy_combinations.one_device_strategy_gpu, <add> ], <add> mode='eager', <add> ) <add> <add> <add>class TaskKerasTest(tf.test.TestCase, parameterized.TestCase): <add> <add> @combinations.generate(all_strategy_combinations()) <add> def test_task_with_step_override(self, distribution): <add> with distribution.scope(): <add> task = mock_task.MockTask() <add> model = task.build_model() <add> model = task.compile_model( <add> model, <add> optimizer=tf.keras.optimizers.SGD(learning_rate=1e-3), <add> metrics=task.build_metrics(), <add> train_step=task.train_step, <add> validation_step=task.validation_step) <add> <add> dataset = task.build_inputs(params=None) <add> logs = model.fit(dataset, epochs=1, steps_per_epoch=2) <add> self.assertIn('loss', logs.history) <add> self.assertIn('acc', logs.history) <add> <add> # Without specifying metrics through compile. <add> with distribution.scope(): <add> train_metrics = task.build_metrics(training=True) <add> val_metrics = task.build_metrics(training=False) <add> model = task.build_model() <add> model = task.compile_model( <add> model, <add> optimizer=tf.keras.optimizers.SGD(learning_rate=1e-3), <add> train_step=functools.partial(task.train_step, metrics=train_metrics), <add> validation_step=functools.partial( <add> task.validation_step, metrics=val_metrics)) <add> logs = model.fit(dataset, epochs=1, steps_per_epoch=2) <add> self.assertIn('loss', logs.history) <add> self.assertIn('acc', logs.history) <add> <add> def test_task_with_fit(self): <add> task = mock_task.MockTask() <add> model = task.build_model() <add> model = task.compile_model( <add> model, <add> optimizer=tf.keras.optimizers.SGD(learning_rate=1e-3), <add> loss=tf.keras.losses.CategoricalCrossentropy(), <add> metrics=task.build_metrics()) <add> dataset = task.build_inputs(params=None) <add> logs = model.fit(dataset, epochs=1, steps_per_epoch=2) <add> self.assertIn('loss', logs.history) <add> self.assertIn('acc', logs.history) <add> self.assertLen(model.evaluate(dataset, steps=1), 2) <add> <add> def test_task_invalid_compile(self): <add> task = mock_task.MockTask() <add> model = task.build_model() <add> with self.assertRaises(ValueError): <add> _ = task.compile_model( <add> model, <add> optimizer=tf.keras.optimizers.SGD(learning_rate=1e-3), <add> loss=tf.keras.losses.CategoricalCrossentropy(), <add> metrics=task.build_metrics(), <add> train_step=task.train_step) <add> <add> <add>if __name__ == '__main__': <add> tf.test.main() <ide><path>official/core/base_trainer.py <add># Lint as: python3 <add># Copyright 2020 The TensorFlow Authors. All Rights Reserved. <add># <add># Licensed under the Apache License, Version 2.0 (the "License"); <add># you may not use this file except in compliance with the License. <add># You may obtain a copy of the License at <add># <add># http://www.apache.org/licenses/LICENSE-2.0 <add># <add># Unless required by applicable law or agreed to in writing, software <add># distributed under the License is distributed on an "AS IS" BASIS, <add># WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. <add># See the License for the specific language governing permissions and <add># limitations under the License. <add># ============================================================================== <add>"""Standard Trainer implementation. <add> <add>The base trainer implements the Orbit `StandardTrainable` and <add>`StandardEvaluable` interfaces. Trainers inside this project should be <add>interchangable and independent on model architectures and tasks. <add>""" <add>import gin <add>import orbit <add>import tensorflow as tf <add> <add>from official.core import base_task <add>from official.modeling import optimization <add>from official.modeling import performance <add>from official.modeling.hyperparams import config_definitions <add> <add> <add>ExperimentConfig = config_definitions.ExperimentConfig <add> <add> <add>@gin.configurable <add>class Trainer(orbit.StandardTrainer, orbit.StandardEvaluator): <add> """Implements the common trainer shared for TensorFlow models.""" <add> <add> def __init__(self, <add> config: ExperimentConfig, <add> task: base_task.Task, <add> train: bool = True, <add> evaluate: bool = True, <add> model=None, <add> optimizer=None): <add> """Initialize common trainer for TensorFlow models. <add> <add> Args: <add> config: An `ExperimentConfig` instance specifying experiment config. <add> task: A base_task.Task instance. <add> train: bool, whether or not this trainer will be used for training. <add> default to True. <add> evaluate: bool, whether or not this trainer will be used for evaluation. <add> default to True. <add> model: tf.keras.Model instance. If provided, it will be used instead <add> of building model using task.build_model(). Default to None. <add> optimizer: tf.keras.optimizers.Optimizer instance. If provided, it will <add> used instead of the optimizer from config. Default to None. <add> """ <add> # Gets the current distribution strategy. If not inside any strategy scope, <add> # it gets a single-replica no-op strategy. <add> self._strategy = tf.distribute.get_strategy() <add> self._config = config <add> self._task = task <add> <add> self._model = model or task.build_model() <add> <add> if optimizer is None: <add> opt_factory = optimization.OptimizerFactory( <add> config.trainer.optimizer_config) <add> self._optimizer = opt_factory.build_optimizer( <add> opt_factory.build_learning_rate()) <add> else: <add> self._optimizer = optimizer <add> <add> # Configuring optimizer when loss_scale is set in runtime config. This helps <add> # avoiding overflow/underflow for float16 computations. <add> if config.runtime.loss_scale: <add> self._optimizer = performance.configure_optimizer( <add> self._optimizer, <add> use_float16=config.runtime.mixed_precision_dtype == 'float16', <add> loss_scale=config.runtime.loss_scale) <add> <add> # global_step increases by 1 after each training iteration. <add> # We should have global_step.numpy() == self.optimizer.iterations.numpy() <add> # when there is only 1 optimizer. <add> self._global_step = orbit.utils.create_global_step() <add> if hasattr(self.model, 'checkpoint_items'): <add> checkpoint_items = self.model.checkpoint_items <add> else: <add> checkpoint_items = {} <add> self._checkpoint = tf.train.Checkpoint( <add> global_step=self.global_step, model=self.model, <add> optimizer=self.optimizer, **checkpoint_items) <add> <add> self._train_loss = tf.keras.metrics.Mean('training_loss', dtype=tf.float32) <add> self._validation_loss = tf.keras.metrics.Mean( <add> 'validation_loss', dtype=tf.float32) <add> self._train_metrics = self.task.build_metrics( <add> training=True) + self.model.metrics <add> self._validation_metrics = self.task.build_metrics( <add> training=False) + self.model.metrics <add> <add> if train: <add> train_dataset = orbit.utils.make_distributed_dataset( <add> self.strategy, self.task.build_inputs, self.config.task.train_data) <add> orbit.StandardTrainer.__init__( <add> self, <add> train_dataset, <add> options=orbit.StandardTrainerOptions( <add> use_tf_while_loop=config.trainer.train_tf_while_loop, <add> use_tf_function=config.trainer.train_tf_function, <add> use_tpu_summary_optimization=config.trainer.allow_tpu_summary)) <add> <add> if evaluate: <add> eval_dataset = orbit.utils.make_distributed_dataset( <add> self.strategy, self.task.build_inputs, <add> self.config.task.validation_data) <add> orbit.StandardEvaluator.__init__( <add> self, <add> eval_dataset, <add> options=orbit.StandardEvaluatorOptions( <add> use_tf_function=config.trainer.eval_tf_function)) <add> <add> @property <add> def strategy(self): <add> return self._strategy <add> <add> @property <add> def config(self): <add> return self._config <add> <add> @property <add> def task(self): <add> return self._task <add> <add> @property <add> def model(self): <add> return self._model <add> <add> @property <add> def optimizer(self): <add> return self._optimizer <add> <add> @property <add> def global_step(self): <add> return self._global_step <add> <add> @property <add> def train_loss(self): <add> """Accesses the training loss metric object.""" <add> return self._train_loss <add> <add> @property <add> def validation_loss(self): <add> """Accesses the validation loss metric object.""" <add> return self._validation_loss <add> <add> @property <add> def train_metrics(self): <add> """Accesses all training metric objects.""" <add> return self._train_metrics <add> <add> @property <add> def validation_metrics(self): <add> """Accesses all validation metric metric objects.""" <add> return self._validation_metrics <add> <add> def initialize(self): <add> """A callback function. <add> <add> This function will be called when no checkpoint found for the model. <add> If there is a checkpoint, the checkpoint will be loaded and this function <add> will not be called. Tasks may use this callback function to load a <add> pretrained checkpoint, saved under a directory other than the model_dir. <add> """ <add> self.task.initialize(self.model) <add> <add> @property <add> def checkpoint(self): <add> """Accesses the training checkpoint.""" <add> return self._checkpoint <add> <add> def train_loop_end(self): <add> """See base class.""" <add> logs = {} <add> for metric in self.train_metrics + [self.train_loss]: <add> logs[metric.name] = metric.result() <add> metric.reset_states() <add> if callable(self.optimizer.learning_rate): <add> logs['learning_rate'] = self.optimizer.learning_rate(self.global_step) <add> else: <add> logs['learning_rate'] = self.optimizer.learning_rate <add> return logs <add> <add> def train_step(self, iterator): <add> """See base class.""" <add> <add> def step_fn(inputs): <add> logs = self.task.train_step( <add> inputs, <add> model=self.model, <add> optimizer=self.optimizer, <add> metrics=self.train_metrics) <add> self._train_loss.update_state(logs[self.task.loss]) <add> self.global_step.assign_add(1) <add> <add> self.strategy.run(step_fn, args=(next(iterator),)) <add> <add> def eval_begin(self): <add> """Sets up metrics.""" <add> for metric in self.validation_metrics + [self.validation_loss]: <add> metric.reset_states() <add> <add> def eval_step(self, iterator): <add> """See base class.""" <add> <add> def step_fn(inputs): <add> logs = self.task.validation_step( <add> inputs, model=self.model, metrics=self.validation_metrics) <add> self._validation_loss.update_state(logs[self.task.loss]) <add> return logs <add> <add> distributed_outputs = self.strategy.run(step_fn, args=(next(iterator),)) <add> return tf.nest.map_structure(self.strategy.experimental_local_results, <add> distributed_outputs) <add> <add> def eval_end(self, aggregated_logs=None): <add> """Processes evaluation results.""" <add> logs = {} <add> for metric in self.validation_metrics + [self.validation_loss]: <add> logs[metric.name] = metric.result() <add> if aggregated_logs: <add> metrics = self.task.reduce_aggregated_logs(aggregated_logs) <add> logs.update(metrics) <add> return logs <add> <add> def eval_reduce(self, state=None, step_outputs=None): <add> return self.task.aggregate_logs(state, step_outputs) <ide><path>official/core/base_trainer_test.py <add># Lint as: python3 <add># Copyright 2020 The TensorFlow Authors. All Rights Reserved. <add># <add># Licensed under the Apache License, Version 2.0 (the "License"); <add># you may not use this file except in compliance with the License. <add># You may obtain a copy of the License at <add># <add># http://www.apache.org/licenses/LICENSE-2.0 <add># <add># Unless required by applicable law or agreed to in writing, software <add># distributed under the License is distributed on an "AS IS" BASIS, <add># WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. <add># See the License for the specific language governing permissions and <add># limitations under the License. <add># ============================================================================== <add>"""Tests for tensorflow_models.core.trainers.trainer.""" <add># pylint: disable=g-direct-tensorflow-import <add>from absl.testing import parameterized <add>import tensorflow as tf <add> <add>from tensorflow.python.distribute import combinations <add>from tensorflow.python.distribute import strategy_combinations <add>from official.core import base_trainer as trainer_lib <add>from official.modeling.hyperparams import config_definitions as cfg <add>from official.utils.testing import mock_task <add> <add> <add>def all_strategy_combinations(): <add> return combinations.combine( <add> distribution=[ <add> strategy_combinations.default_strategy, <add> strategy_combinations.tpu_strategy, <add> strategy_combinations.one_device_strategy_gpu, <add> ], <add> mode='eager', <add> ) <add> <add> <add>class TrainerTest(tf.test.TestCase, parameterized.TestCase): <add> <add> def setUp(self): <add> super().setUp() <add> self._config = cfg.ExperimentConfig( <add> trainer=cfg.TrainerConfig( <add> optimizer_config=cfg.OptimizationConfig( <add> {'optimizer': { <add> 'type': 'sgd' <add> }, <add> 'learning_rate': { <add> 'type': 'constant' <add> }}))) <add> <add> def create_test_trainer(self): <add> task = mock_task.MockTask() <add> trainer = trainer_lib.Trainer(self._config, task) <add> return trainer <add> <add> @combinations.generate(all_strategy_combinations()) <add> def test_trainer_train(self, distribution): <add> with distribution.scope(): <add> trainer = self.create_test_trainer() <add> logs = trainer.train(tf.convert_to_tensor(5, dtype=tf.int32)) <add> self.assertIn('training_loss', logs) <add> self.assertIn('learning_rate', logs) <add> <add> @combinations.generate(all_strategy_combinations()) <add> def test_trainer_validate(self, distribution): <add> with distribution.scope(): <add> trainer = self.create_test_trainer() <add> logs = trainer.evaluate(tf.convert_to_tensor(5, dtype=tf.int32)) <add> self.assertIn('validation_loss', logs) <add> self.assertEqual(logs['acc'], 5. * distribution.num_replicas_in_sync) <add> <add> @combinations.generate( <add> combinations.combine( <add> mixed_precision_dtype=['float32', 'bfloat16', 'float16'], <add> loss_scale=[None, 'dynamic', 128, 256], <add> )) <add> def test_configure_optimizer(self, mixed_precision_dtype, loss_scale): <add> config = cfg.ExperimentConfig( <add> runtime=cfg.RuntimeConfig( <add> mixed_precision_dtype=mixed_precision_dtype, loss_scale=loss_scale), <add> trainer=cfg.TrainerConfig( <add> optimizer_config=cfg.OptimizationConfig( <add> {'optimizer': { <add> 'type': 'sgd' <add> }, <add> 'learning_rate': { <add> 'type': 'constant' <add> }}))) <add> task = mock_task.MockTask() <add> trainer = trainer_lib.Trainer(config, task) <add> if mixed_precision_dtype != 'float16': <add> self.assertIsInstance(trainer.optimizer, tf.keras.optimizers.SGD) <add> elif mixed_precision_dtype == 'float16' and loss_scale is None: <add> self.assertIsInstance(trainer.optimizer, tf.keras.optimizers.SGD) <add> else: <add> self.assertIsInstance( <add> trainer.optimizer, <add> tf.keras.mixed_precision.experimental.LossScaleOptimizer) <add> <add> metrics = trainer.train(tf.convert_to_tensor(5, dtype=tf.int32)) <add> self.assertIn('training_loss', metrics) <add> <add> <add>if __name__ == '__main__': <add> tf.test.main() <ide><path>official/utils/testing/mock_task.py <add># Lint as: python3 <add># Copyright 2020 The TensorFlow Authors. All Rights Reserved. <add># <add># Licensed under the Apache License, Version 2.0 (the "License"); <add># you may not use this file except in compliance with the License. <add># You may obtain a copy of the License at <add># <add># http://www.apache.org/licenses/LICENSE-2.0 <add># <add># Unless required by applicable law or agreed to in writing, software <add># distributed under the License is distributed on an "AS IS" BASIS, <add># WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. <add># See the License for the specific language governing permissions and <add># limitations under the License. <add># ============================================================================== <add>"""Mock task for testing.""" <add> <add>import dataclasses <add>import numpy as np <add>import tensorflow as tf <add> <add>from official.core import base_task <add>from official.core import exp_factory <add>from official.core import task_factory <add>from official.modeling.hyperparams import config_definitions as cfg <add> <add> <add>class MockModel(tf.keras.Model): <add> <add> def __init__(self, network): <add> super().__init__() <add> self.network = network <add> <add> def call(self, inputs): <add> outputs = self.network(inputs) <add> self.add_loss(tf.reduce_mean(outputs)) <add> return outputs <add> <add> <add>@dataclasses.dataclass <add>class MockTaskConfig(cfg.TaskConfig): <add> pass <add> <add> <add>@task_factory.register_task_cls(MockTaskConfig) <add>class MockTask(base_task.Task): <add> """Mock task object for testing.""" <add> <add> def __init__(self, params=None, logging_dir=None): <add> super().__init__(params=params, logging_dir=logging_dir) <add> <add> def build_model(self, *arg, **kwargs): <add> inputs = tf.keras.layers.Input(shape=(2,), name="random", dtype=tf.float32) <add> outputs = tf.keras.layers.Dense(1)(inputs) <add> network = tf.keras.Model(inputs=inputs, outputs=outputs) <add> return MockModel(network) <add> <add> def build_metrics(self, training: bool = True): <add> del training <add> return [tf.keras.metrics.Accuracy(name="acc")] <add> <add> def build_inputs(self, params): <add> <add> def generate_data(_): <add> x = tf.zeros(shape=(2,), dtype=tf.float32) <add> label = tf.zeros([1], dtype=tf.int32) <add> return x, label <add> <add> dataset = tf.data.Dataset.range(1) <add> dataset = dataset.repeat() <add> dataset = dataset.map( <add> generate_data, num_parallel_calls=tf.data.experimental.AUTOTUNE) <add> return dataset.prefetch(buffer_size=1).batch(2, drop_remainder=True) <add> <add> def aggregate_logs(self, state, step_outputs): <add> if state is None: <add> state = {} <add> for key, value in step_outputs.items(): <add> if key not in state: <add> state[key] = [] <add> state[key].append( <add> np.concatenate([np.expand_dims(v.numpy(), axis=0) for v in value])) <add> return state <add> <add> def reduce_aggregated_logs(self, aggregated_logs): <add> for k, v in aggregated_logs.items(): <add> aggregated_logs[k] = np.sum(np.stack(v, axis=0)) <add> return aggregated_logs <add> <add> <add>@exp_factory.register_config_factory("mock") <add>def mock_experiment() -> cfg.ExperimentConfig: <add> config = cfg.ExperimentConfig( <add> task=MockTaskConfig(), trainer=cfg.TrainerConfig()) <add> return config
4
Python
Python
fix assertion error
7e8f1053bb07f4f73914b5784a87de6dcaea6bcb
<ide><path>tests/test_appctx.py <ide> def index(): <ide> c = app.test_client() <ide> res = c.get('/') <ide> assert res.status_code == 200 <del> assert res.data == u'' <add> assert res.data == b'' <ide> assert called == ['request', 'app'] <ide> <ide>
1
Go
Go
fix internal ipvlan network to work in swarm
c7f8bfa0010f6f92c875e0c11df7ecd397eca35c
<ide><path>libnetwork/drivers/ipvlan/ipvlan_network.go <ide> func parseNetworkOptions(id string, option options.Generic) (*configuration, err <ide> return nil, err <ide> } <ide> } <del> // setting the parent to "" will trigger an isolated network dummy parent link <ide> if val, ok := option[netlabel.Internal]; ok { <ide> if internal, ok := val.(bool); ok && internal { <ide> config.Internal = true <del> // empty --parent= and --internal are handled the same. <del> config.Parent = "" <ide> } <ide> } <ide> return config, nil
1
PHP
PHP
fix bug in url generator
3ef7bf2bf4fc7c7b240ea1f0f1bb2e364603f4ca
<ide><path>src/Illuminate/Routing/UrlGenerator.php <ide> public function secureAsset($path) <ide> */ <ide> protected function getScheme($secure) <ide> { <del> if (is_null($secure)) <add> if ( ! $secure) <ide> { <ide> return $this->request->getScheme().'://'; <ide> } <ide><path>tests/Routing/RoutingUrlGeneratorTest.php <ide> public function testBasicGeneration() <ide> <ide> $this->assertEquals('http://www.foo.com/foo/bar', $url->asset('foo/bar')); <ide> $this->assertEquals('https://www.foo.com/foo/bar', $url->asset('foo/bar', true)); <del> } <add> } <ide> <ide> <ide> public function testBasicRouteGeneration() <ide> public function testBasicRouteGeneration() <ide> } <ide> <ide> <add> public function testRoutesMaintainRequestScheme() <add> { <add> $url = new UrlGenerator( <add> $routes = new Illuminate\Routing\RouteCollection, <add> $request = Illuminate\Http\Request::create('https://www.foo.com/') <add> ); <add> <add> /** <add> * Named Routes <add> */ <add> $route = new Illuminate\Routing\Route(array('GET'), 'foo/bar', array('as' => 'foo')); <add> $routes->add($route); <add> <add> $this->assertEquals('https://www.foo.com/foo/bar', $url->route('foo')); <add> } <add> <add> <ide> public function testRoutesWithDomains() <ide> { <del> $url = new UrlGenerator( <add> $url = new UrlGenerator( <ide> $routes = new Illuminate\Routing\RouteCollection, <ide> $request = Illuminate\Http\Request::create('http://www.foo.com/') <ide> );
2
Javascript
Javascript
add uniquename for css module classes
bdd5d3d9d391faebf826cb951aa5a95d7f96ff07
<ide><path>lib/css/CssLoadingRuntimeModule.js <ide> class CssLoadingRuntimeModule extends RuntimeModule { <ide> const code = Template.asString([ <ide> "link = document.createElement('link');", <ide> uniqueName <del> ? 'link.setAttribute("data-webpack", dataWebpackPrefix + key);' <add> ? 'link.setAttribute("data-webpack", uniqueName + ":" + key);' <ide> : "", <ide> "link.setAttribute(loadingAttribute, 1);", <ide> 'link.rel = "stylesheet";', <ide> class CssLoadingRuntimeModule extends RuntimeModule { <ide> id => `${JSON.stringify(id)}:0` <ide> ).join(",")}};`, <ide> "", <add> uniqueName <add> ? `var uniqueName = ${JSON.stringify( <add> runtimeTemplate.outputOptions.uniqueName <add> )};` <add> : "// data-webpack is not used as build has no uniqueName", <ide> `var loadCssChunkData = ${runtimeTemplate.basicFunction("chunkId, link", [ <ide> 'var data, token = "", token2, exports = {}, exportsWithId = [], i = 0, cc = 1;', <ide> "try { if(!link) link = loadStylesheet(chunkId); data = link.sheet.cssRules; data = data[data.length - 1].style; } catch(e) { data = getComputedStyle(document.head); }", <del> 'data = data.getPropertyValue("--webpack-" + chunkId);', <add> `data = data.getPropertyValue(${ <add> uniqueName <add> ? runtimeTemplate.concatenation( <add> "--webpack-", <add> { expr: "uniqueName" }, <add> "-", <add> { expr: "chunkId" } <add> ) <add> : runtimeTemplate.concatenation("--webpack-", { expr: "chunkId" }) <add> });`, <ide> "if(!data) return;", <ide> "for(; cc; i++) {", <ide> Template.indent([ <ide> class CssLoadingRuntimeModule extends RuntimeModule { <ide> )}) { exports[token2.replace(/^_/, "")] = token.replace(/^_/, ""); token = ""; }`, <ide> `else if(cc == ${cc( <ide> "/" <del> )}) { token = token.replace(/^_/, ""); exports[token] = token + "_"; exportsWithId.push(token); token = ""; }`, <add> )}) { token = token.replace(/^_/, ""); exports[token] = token; exportsWithId.push(token); token = ""; }`, <ide> `else if(!cc || cc == ${cc( <ide> "," <ide> )}) { token = token.replace(/^_/, ""); exportsWithId.forEach(${runtimeTemplate.expressionFunction( <del> "exports[x] += token", <add> `exports[x] = ${ <add> uniqueName <add> ? runtimeTemplate.concatenation( <add> { expr: "uniqueName" }, <add> "-", <add> { expr: "token" }, <add> "-", <add> { expr: "exports[x]" } <add> ) <add> : runtimeTemplate.concatenation({ expr: "token" }, "-", { <add> expr: "exports[x]" <add> }) <add> }`, <ide> "x" <ide> )}); ${RuntimeGlobals.makeNamespaceObject}(exports); ${ <ide> RuntimeGlobals.moduleFactories <ide> class CssLoadingRuntimeModule extends RuntimeModule { <ide> "installedChunks[chunkId] = 0;" <ide> ])}`, <ide> 'var loadingAttribute = "data-webpack-loading";', <del> uniqueName <del> ? `var dataWebpackPrefix = ${JSON.stringify(uniqueName + ":")};` <del> : "// data-webpack is not used as build has no uniqueName", <ide> `var loadStylesheet = ${runtimeTemplate.basicFunction( <ide> "chunkId, url, done" + (withHmr ? ", hmr" : ""), <ide> [ <ide> class CssLoadingRuntimeModule extends RuntimeModule { <ide> "var l = links[i];", <ide> `if(l.getAttribute("href") == url${ <ide> uniqueName <del> ? ' || l.getAttribute("data-webpack") == dataWebpackPrefix + key' <add> ? ' || l.getAttribute("data-webpack") == uniqueName + ":" + key' <ide> : "" <ide> }) { link = l; break; }` <ide> ]), <ide><path>lib/css/CssModulesPlugin.js <ide> class CssModulesPlugin { <ide> this.renderChunk({ <ide> chunk, <ide> chunkGraph, <del> codeGenerationResults <add> codeGenerationResults, <add> uniqueName: compilation.outputOptions.uniqueName <ide> }), <ide> filenameTemplate: CssModulesPlugin.getChunkFilenameTemplate( <ide> chunk, <ide> class CssModulesPlugin { <ide> return [...(cssImports || []), ...(cssContent || [])]; <ide> } <ide> <del> renderChunk({ chunk, chunkGraph, codeGenerationResults }) { <add> renderChunk({ uniqueName, chunk, chunkGraph, codeGenerationResults }) { <ide> const modules = this.getOrderedChunkCssModules(chunk, chunkGraph); <ide> const source = new ConcatSource(); <ide> const metaData = []; <ide> class CssModulesPlugin { <ide> `${ <ide> exports <ide> ? Array.from(exports, ([n, v]) => <del> v === `${n}_${moduleId}` <add> v === `${uniqueName ? uniqueName + "-" : ""}${moduleId}-${n}` <ide> ? `${escapeCss(n)}/` <ide> : `${escapeCss(n)}(${escapeCss(v)})` <ide> ).join("") <ide> class CssModulesPlugin { <ide> } <ide> } <ide> source.add( <del> `head{--webpack-${escapeCss(chunk.id, true)}:${metaData.join(",")};}` <add> `head{--webpack-${escapeCss( <add> (uniqueName ? uniqueName + "-" : "") + chunk.id, <add> true <add> )}:${metaData.join(",")};}` <ide> ); <ide> return source; <ide> } <ide><path>lib/dependencies/CssLocalIdentifierDependency.js <ide> CssLocalIdentifierDependency.Template = class CssLocalIdentifierDependencyTempla <ide> apply( <ide> dependency, <ide> source, <del> { module, moduleGraph, chunkGraph, runtime, cssExports } <add> { module, moduleGraph, chunkGraph, runtime, runtimeTemplate, cssExports } <ide> ) { <ide> const dep = /** @type {CssLocalIdentifierDependency} */ (dependency); <ide> const used = moduleGraph <ide> .getExportInfo(module, dep.name) <ide> .getUsedName(dep.name, runtime); <ide> const moduleId = chunkGraph.getModuleId(module); <del> const identifier = (used || "_") + "_" + moduleId; <add> const identifier = <add> (runtimeTemplate.outputOptions.uniqueName <add> ? runtimeTemplate.outputOptions.uniqueName + "-" <add> : "") + <add> moduleId + <add> "-" + <add> (used || ""); <ide> source.replace( <ide> dep.range[0], <ide> dep.range[1] - 1, <ide><path>test/configCases/css/css-modules/index.js <ide> it("should allow to create css modules", done => { <ide> try { <ide> expect(x).toEqual({ <ide> global: undefined, <del> class: prod ? "S_491" : "class_./style.module.css", <add> class: prod ? "my-app-491-S" : "./style.module.css-class", <ide> local: prod <del> ? "Zw_491 yl_491 J__491 gc_491" <del> : "local1_./style.module.css local2_./style.module.css local3_./style.module.css local4_./style.module.css", <add> ? "my-app-491-Zw my-app-491-yl my-app-491-J_ my-app-491-gc" <add> : "./style.module.css-local1 ./style.module.css-local2 ./style.module.css-local3 ./style.module.css-local4", <ide> local2: prod <del> ? "Xg_491 AY_491" <del> : "local5_./style.module.css local6_./style.module.css", <add> ? "my-app-491-Xg my-app-491-AY" <add> : "./style.module.css-local5 ./style.module.css-local6", <ide> nested: prod <del> ? "RX_491 undefined X2_491" <del> : "nested1_./style.module.css undefined nested3_./style.module.css" <add> ? "my-app-491-RX undefined my-app-491-X2" <add> : "./style.module.css-nested1 undefined ./style.module.css-nested3" <ide> }); <ide> } catch (e) { <ide> return done(e); <ide><path>test/configCases/css/css-modules/webpack.config.js <ide> module.exports = [ <ide> { <ide> target: "web", <ide> mode: "production", <add> output: { <add> uniqueName: "my-app" <add> }, <ide> experiments: { <ide> css: true <ide> }
5
Javascript
Javascript
fix dependencies when snapshot is missing
8ee39ac3ac651a0d2093292ad16e1a215b8816f1
<ide><path>lib/NormalModule.js <ide> class NormalModule extends Module { <ide> contextDependencies: contextDeps, <ide> missingDependencies: missingDeps <ide> } = this.buildInfo; <del> fileDependencies.addAll(fileDeps); <del> contextDependencies.addAll(contextDeps); <del> missingDependencies.addAll(missingDeps); <add> if (fileDeps !== undefined) fileDependencies.addAll(fileDeps); <add> if (contextDeps !== undefined) contextDependencies.addAll(contextDeps); <add> if (missingDeps !== undefined) missingDependencies.addAll(missingDeps); <ide> } <del> if (buildDeps) { <add> if (buildDeps !== undefined) { <ide> buildDependencies.addAll(buildDeps); <ide> } <ide> }
1
Javascript
Javascript
create name/dashname at seed
b1e34f8cc8554458200009fb2b2f2c82b05823f1
<ide><path>seed/index.js <ide> require('babel/register'); <ide> require('dotenv').load(); <ide> var fs = require('fs'), <add> _ = require('lodash'), <ide> path = require('path'), <ide> app = require('../server/server'), <ide> nonprofits = require('./nonprofits.json'), <ide> jobs = require('./jobs.json'); <ide> <del>var challangesRegex = /^(bonfire:|waypoint:|zipline:|basejump:|hike:)/i; <del> <ide> function getFilesFor(dir) { <ide> return fs.readdirSync(path.join(__dirname, '/' + dir)); <ide> } <ide> Challenge.destroyAll(function(err, info) { <ide> console.log('Deleted ', info); <ide> } <ide> challenges.forEach(function(file) { <del> var challenges = require('./challenges/' + file).challenges <add> var challengeSpec = require('./challenges/' + file); <add> var challenges = challengeSpec.challenges <ide> .map(function(challenge) { <ide> // NOTE(berks): add title for displaying in views <del> //challenge.title = challenge.name.replace(challangesRegex, '').trim(); <del> challenge.name = challenge.title.replace(/[^a-zA-Z0-9 ]/g, ''); // Remove non-alphanumwhitespace chars <del> challenge.dashedName = challenge.name.replace(/\s/g, '-'); // Replace with dasherize(); <add> challenge.name = <add> _.capitalize(challenge.type) + <add> ': ' + <add> challenge.title.replace(/[^a-zA-Z0-9\s]/g, ''); <add> challenge.dashedName = challenge.name.toLowerCase().replace(/\s/g, '-'); <ide> return challenge; <ide> }); <ide>
1
Javascript
Javascript
remove redundant code
416499c872ba3db45a4dd4b88e015b679931c15f
<ide><path>lib/timers.js <ide> exports.enroll = function(item, msecs) { <ide> // it will reset its timeout. <ide> exports.active = function(item) { <ide> var msecs = item._idleTimeout; <del> if (msecs >= 0) { <del> var list = lists[msecs]; <del> if (!list || L.isEmpty(list)) { <del> insert(item, msecs); <del> } else { <del> item._idleStart = Timer.now(); <del> L.append(list, item); <del> } <del> } <add> if (msecs >= 0) <add> insert(item, msecs); <ide> }; <ide> <ide>
1
Ruby
Ruby
use ||= instead of "return ... if ... "
4f119ad85ff77386312b4a5c3289ec53a6f29769
<ide><path>Library/Homebrew/linkage_checker.rb <ide> def unexpected_broken_dylibs <ide> end <ide> <ide> def unexpected_present_dylibs <del> return @unexpected_present_dylibs if @unexpected_present_dylibs <del> <del> @unexpected_present_dylibs = @formula.class.allowed_missing_libraries.reject do |allowed_missing_lib| <add> @unexpected_present_dylibs ||= @formula.class.allowed_missing_libraries.reject do |allowed_missing_lib| <ide> @broken_dylibs.any? do |broken_lib| <ide> case allowed_missing_lib <ide> when Regexp
1
Go
Go
use mock for search
9a0d7fe0182da541cc99eab9a4930616792e95c3
<ide><path>registry/registry.go <ide> func (r *Registry) PushImageJSONIndex(remote string, imgList []*ImgData, validat <ide> <ide> func (r *Registry) SearchRepositories(term string) (*SearchResults, error) { <ide> utils.Debugf("Index server: %s", r.indexEndpoint) <del> u := IndexServerAddress() + "search?q=" + url.QueryEscape(term) <add> u := r.indexEndpoint + "search?q=" + url.QueryEscape(term) <ide> req, err := r.reqFactory.NewRequest("GET", u, nil) <ide> if err != nil { <ide> return nil, err <ide><path>registry/registry_mock_test.go <ide> func handlerAuth(w http.ResponseWriter, r *http.Request) { <ide> } <ide> <ide> func handlerSearch(w http.ResponseWriter, r *http.Request) { <del> writeResponse(w, "{}", 200) <add> result := &SearchResults{ <add> Query: "fakequery", <add> NumResults: 1, <add> Results: []SearchResult{{Name: "fakeimage", StarCount: 42}}, <add> } <add> writeResponse(w, result, 200) <ide> } <ide> <ide> func TestPing(t *testing.T) { <ide><path>registry/registry_test.go <ide> func TestPushImageJSONIndex(t *testing.T) { <ide> <ide> func TestSearchRepositories(t *testing.T) { <ide> r := spawnTestRegistry(t) <del> results, err := r.SearchRepositories("supercalifragilisticepsialidocious") <add> results, err := r.SearchRepositories("fakequery") <ide> if err != nil { <ide> t.Fatal(err) <ide> } <ide> if results == nil { <ide> t.Fatal("Expected non-nil SearchResults object") <ide> } <del> assertEqual(t, results.NumResults, 0, "Expected 0 search results") <add> assertEqual(t, results.NumResults, 1, "Expected 1 search results") <add> assertEqual(t, results.Query, "fakequery", "Expected 'fakequery' as query") <add> assertEqual(t, results.Results[0].StarCount, 42, "Expected 'fakeimage' a ot hae 42 stars") <ide> } <ide> <ide> func TestValidRepositoryName(t *testing.T) {
3
Javascript
Javascript
fix lint error
a582a1a188b981f8d72d25d26827c837c0ad8be9
<ide><path>src/get-release-channel.js <ide> module.exports = function(version) { <ide> // This matches stable, dev (with or without commit hash) and any other <ide> // release channel following the pattern '1.00.0-channel0' <del> const match = version.match( <del> /\d+\.\d+\.\d+(-([a-z]+)(\d+|-\w{4,})?)?$/ <del> ); <add> const match = version.match(/\d+\.\d+\.\d+(-([a-z]+)(\d+|-\w{4,})?)?$/); <ide> if (!match) { <ide> return 'unrecognized'; <ide> } else if (match[2]) { <ide> return match[2]; <ide> } <ide> <ide> return 'stable'; <del>} <add>};
1
Mixed
Python
add (noun chunks) syntax iterators for danish
e3222fdec9efcc9a4a5d0d9b2f92099fbf36ed7b
<ide><path>.github/contributors/ophelielacroix.md <add># spaCy contributor agreement <add> <add>This spaCy Contributor Agreement (**"SCA"**) is based on the <add>[Oracle Contributor Agreement](http://www.oracle.com/technetwork/oca-405177.pdf). <add>The SCA applies to any contribution that you make to any product or project <add>managed by us (the **"project"**), and sets out the intellectual property rights <add>you grant to us in the contributed materials. The term **"us"** shall mean <add>[ExplosionAI GmbH](https://explosion.ai/legal). The term <add>**"you"** shall mean the person or entity identified below. <add> <add>If you agree to be bound by these terms, fill in the information requested <add>below and include the filled-in version with your first pull request, under the <add>folder [`.github/contributors/`](/.github/contributors/). The name of the file <add>should be your GitHub username, with the extension `.md`. For example, the user <add>example_user would create the file `.github/contributors/example_user.md`. <add> <add>Read this agreement carefully before signing. These terms and conditions <add>constitute a binding legal agreement. <add> <add>## Contributor Agreement <add> <add>1. The term "contribution" or "contributed materials" means any source code, <add>object code, patch, tool, sample, graphic, specification, manual, <add>documentation, or any other material posted or submitted by you to the project. <add> <add>2. With respect to any worldwide copyrights, or copyright applications and <add>registrations, in your contribution: <add> <add> * you hereby assign to us joint ownership, and to the extent that such <add> assignment is or becomes invalid, ineffective or unenforceable, you hereby <add> grant to us a perpetual, irrevocable, non-exclusive, worldwide, no-charge, <add> royalty-free, unrestricted license to exercise all rights under those <add> copyrights. This includes, at our option, the right to sublicense these same <add> rights to third parties through multiple levels of sublicensees or other <add> licensing arrangements; <add> <add> * you agree that each of us can do all things in relation to your <add> contribution as if each of us were the sole owners, and if one of us makes <add> a derivative work of your contribution, the one who makes the derivative <add> work (or has it made will be the sole owner of that derivative work; <add> <add> * you agree that you will not assert any moral rights in your contribution <add> against us, our licensees or transferees; <add> <add> * you agree that we may register a copyright in your contribution and <add> exercise all ownership rights associated with it; and <add> <add> * you agree that neither of us has any duty to consult with, obtain the <add> consent of, pay or render an accounting to the other for any use or <add> distribution of your contribution. <add> <add>3. With respect to any patents you own, or that you can license without payment <add>to any third party, you hereby grant to us a perpetual, irrevocable, <add>non-exclusive, worldwide, no-charge, royalty-free license to: <add> <add> * make, have made, use, sell, offer to sell, import, and otherwise transfer <add> your contribution in whole or in part, alone or in combination with or <add> included in any product, work or materials arising out of the project to <add> which your contribution was submitted, and <add> <add> * at our option, to sublicense these same rights to third parties through <add> multiple levels of sublicensees or other licensing arrangements. <add> <add>4. Except as set out above, you keep all right, title, and interest in your <add>contribution. The rights that you grant to us under these terms are effective <add>on the date you first submitted a contribution to us, even if your submission <add>took place before the date you sign these terms. <add> <add>5. You covenant, represent, warrant and agree that: <add> <add> * Each contribution that you submit is and shall be an original work of <add> authorship and you can legally grant the rights set out in this SCA; <add> <add> * to the best of your knowledge, each contribution will not violate any <add> third party's copyrights, trademarks, patents, or other intellectual <add> property rights; and <add> <add> * each contribution shall be in compliance with U.S. export control laws and <add> other applicable export and import laws. You agree to notify us if you <add> become aware of any circumstance which would make any of the foregoing <add> representations inaccurate in any respect. We may publicly disclose your <add> participation in the project, including the fact that you have signed the SCA. <add> <add>6. This SCA is governed by the laws of the State of California and applicable <add>U.S. Federal law. Any choice of law rules will not apply. <add> <add>7. Please place an “x” on one of the applicable statement below. Please do NOT <add>mark both statements: <add> <add> * [X] I am signing on behalf of myself as an individual and no other person <add> or entity, including my employer, has or will have rights with respect to my <add> contributions. <add> <add> * [ ] I am signing on behalf of my employer or a legal entity and I have the <add> actual authority to contractually bind that entity. <add> <add>## Contributor Details <add> <add>| Field | Entry | <add>|-------------------------------|-----------------| <add>| Name | Ophélie Lacroix | <add>| Company name (if applicable) | | <add>| Title or role (if applicable) | | <add>| Date | | <add>| GitHub username | ophelielacroix | <add>| Website (optional) | | <ide><path>spacy/lang/da/__init__.py <ide> from .lex_attrs import LEX_ATTRS <ide> from .morph_rules import MORPH_RULES <ide> from ..tag_map import TAG_MAP <add>from .syntax_iterators import SYNTAX_ITERATORS <ide> <ide> from ..tokenizer_exceptions import BASE_EXCEPTIONS <ide> from ...language import Language <ide> class DanishDefaults(Language.Defaults): <ide> suffixes = TOKENIZER_SUFFIXES <ide> tag_map = TAG_MAP <ide> stop_words = STOP_WORDS <add> syntax_iterators = SYNTAX_ITERATORS <ide> <ide> <ide> class Danish(Language): <ide><path>spacy/lang/da/syntax_iterators.py <add># coding: utf8 <add>from __future__ import unicode_literals <add> <add>from ...symbols import NOUN, PROPN, PRON, VERB, AUX <add>from ...errors import Errors <add> <add> <add>def noun_chunks(doclike): <add> def is_verb_token(tok): <add> return tok.pos in [VERB, AUX] <add> <add> def next_token(tok): <add> try: <add> return tok.nbor() <add> except IndexError: <add> return None <add> <add> def get_left_bound(doc, root): <add> left_bound = root <add> for tok in reversed(list(root.lefts)): <add> if tok.dep in np_left_deps: <add> left_bound = tok <add> return left_bound <add> <add> def get_right_bound(doc, root): <add> right_bound = root <add> for tok in root.rights: <add> if tok.dep in np_right_deps: <add> right = get_right_bound(doc, tok) <add> if list( <add> filter( <add> lambda t: is_verb_token(t) or t.dep in stop_deps, <add> doc[root.i : right.i], <add> ) <add> ): <add> break <add> else: <add> right_bound = right <add> return right_bound <add> <add> def get_bounds(doc, root): <add> return get_left_bound(doc, root), get_right_bound(doc, root) <add> <add> doc = doclike.doc <add> <add> if not doc.is_parsed: <add> raise ValueError(Errors.E029) <add> <add> if not len(doc): <add> return <add> <add> left_labels = [ <add> "det", <add> "fixed", <add> "nmod:poss", <add> "amod", <add> "flat", <add> "goeswith", <add> "nummod", <add> "appos", <add> ] <add> right_labels = ["fixed", "nmod:poss", "amod", "flat", "goeswith", "nummod", "appos"] <add> stop_labels = ["punct"] <add> <add> np_label = doc.vocab.strings.add("NP") <add> np_left_deps = [doc.vocab.strings.add(label) for label in left_labels] <add> np_right_deps = [doc.vocab.strings.add(label) for label in right_labels] <add> stop_deps = [doc.vocab.strings.add(label) for label in stop_labels] <add> <add> chunks = [] <add> prev_right = -1 <add> for token in doclike: <add> if token.pos in [PROPN, NOUN, PRON]: <add> left, right = get_bounds(doc, token) <add> if left.i <= prev_right: <add> continue <add> yield left.i, right.i + 1, np_label <add> prev_right = right.i <add> <add> <add>SYNTAX_ITERATORS = {"noun_chunks": noun_chunks} <ide><path>spacy/tests/lang/da/test_noun_chunks.py <add># coding: utf-8 <add>from __future__ import unicode_literals <add> <add>import pytest <add> <add>from ...util import get_doc <add> <add> <add>def test_noun_chunks_is_parsed(da_tokenizer): <add> """Test that noun_chunks raises Value Error for 'da' language if Doc is not parsed. <add> To check this test, we're constructing a Doc <add> with a new Vocab here and forcing is_parsed to 'False' <add> to make sure the noun chunks don't run. <add> """ <add> doc = da_tokenizer("Det er en sætning") <add> doc.is_parsed = False <add> with pytest.raises(ValueError): <add> list(doc.noun_chunks) <add> <add> <add> <add>DA_NP_TEST_EXAMPLES = [ <add> ( <add> "Hun elsker at plukker frugt.", <add> ['PRON', 'VERB', 'PART', 'VERB', 'NOUN', 'PUNCT'], <add> ['nsubj', 'ROOT', 'mark', 'obj', 'obj', 'punct'], <add> [1, 0, 1, -2, -1, -4], <add> ["Hun", "frugt"], <add> ), <add> ( <add> "Påfugle er de smukkeste fugle.", <add> ['NOUN', 'AUX', 'DET', 'ADJ', 'NOUN', 'PUNCT'], <add> ['nsubj', 'cop', 'det', 'amod', 'ROOT', 'punct'], <add> [4, 3, 2, 1, 0, -1], <add> ["Påfugle", "de smukkeste fugle"], <add> ), <add> ( <add> "Rikke og Jacob Jensen glæder sig til en hyggelig skovtur", <add> ['PROPN', 'CCONJ', 'PROPN', 'PROPN', 'VERB', 'PRON', 'ADP', 'DET', 'ADJ', 'NOUN'], <add> ['nsubj', 'cc', 'conj', 'flat', 'ROOT', 'obj', 'case', 'det', 'amod', 'obl'], <add> [4, 1, -2, -1, 0, -1, 3, 2, 1, -5], <add> ["Rikke", "Jacob Jensen", "sig", "en hyggelig skovtur"], <add> ), <add>] <add> <add> <add>@pytest.mark.parametrize( <add> "text,pos,deps,heads,expected_noun_chunks", DA_NP_TEST_EXAMPLES <add>) <add>def test_da_noun_chunks(da_tokenizer, text, pos, deps, heads, expected_noun_chunks): <add> tokens = da_tokenizer(text) <add> <add> assert len(heads) == len(pos) <add> doc = get_doc( <add> tokens.vocab, words=[t.text for t in tokens], heads=heads, deps=deps, pos=pos <add> ) <add> <add> noun_chunks = list(doc.noun_chunks) <add> assert len(noun_chunks) == len(expected_noun_chunks) <add> for i, np in enumerate(noun_chunks): <add> assert np.text == expected_noun_chunks[i]
4
Javascript
Javascript
add array.flat polyfill to nomodule-polyfills
20c546710d758a17ecb7abac64208d2fa09186e4
<ide><path>packages/next-polyfill-nomodule/src/index.js <ide> import 'core-js/features/array/fill' <ide> import 'core-js/features/array/find' <ide> import 'core-js/features/array/find-index' <ide> import 'core-js/features/array/flat-map' <add>import 'core-js/features/array/flat' <ide> import 'core-js/features/array/from' <ide> import 'core-js/features/array/includes' <ide> import 'core-js/features/array/iterator'
1
PHP
PHP
apply fixes from styleci
89d9cd46e6cf764171cae2c762872aacf87d65dd
<ide><path>src/Illuminate/Database/Eloquent/Relations/MorphPivot.php <ide> <ide> namespace Illuminate\Database\Eloquent\Relations; <ide> <del>use Illuminate\Database\Eloquent\Builder; <ide> use Illuminate\Support\Str; <ide> <ide> class MorphPivot extends Pivot
1
Mixed
Javascript
create temp dir in common.js
a6b8ee19b85bbd798510191f0aee596f36b909d2
<ide><path>test/common.js <ide> exports.tmpDirName = 'tmp'; <ide> exports.PORT = +process.env.NODE_COMMON_PORT || 12346; <ide> exports.isWindows = process.platform === 'win32'; <ide> <add>function rimrafSync(p) { <add> try { <add> var st = fs.lstatSync(p); <add> } catch (e) { <add> if (e.code === 'ENOENT') <add> return; <add> } <add> <add> try { <add> if (st && st.isDirectory()) <add> rmdirSync(p, null); <add> else <add> fs.unlinkSync(p); <add> } catch (e) { <add> if (e.code === 'ENOENT') <add> return; <add> if (e.code === 'EPERM') <add> return rmdirSync(p, er); <add> if (e.code !== 'EISDIR') <add> throw e; <add> rmdirSync(p, e); <add> } <add>} <add> <add>function rmdirSync(p, originalEr) { <add> try { <add> fs.rmdirSync(p); <add> } catch (e) { <add> if (e.code === 'ENOTDIR') <add> throw originalEr; <add> if (e.code === 'ENOTEMPTY' || e.code === 'EEXIST' || e.code === 'EPERM') { <add> fs.readdirSync(p).forEach(function(f) { <add> rimrafSync(path.join(p, f)); <add> }); <add> fs.rmdirSync(p); <add> } <add> } <add>} <add> <add>function refreshTmpDir() { <add> if (!process.send) { // Not a child process <add> try { <add> rimrafSync(exports.tmpDir); <add> } catch (e) { <add> } <add> <add> try { <add> fs.mkdirSync(exports.tmpDir); <add> } catch (e) { <add> } <add> } <add>} <add> <ide> if (process.env.TEST_THREAD_ID) { <ide> // Distribute ports in parallel tests <ide> if (!process.env.NODE_COMMON_PORT) <ide> if (process.env.TEST_THREAD_ID) { <ide> } <ide> exports.tmpDir = path.join(exports.testDir, exports.tmpDirName); <ide> <add>refreshTmpDir(); <add> <ide> var opensslCli = null; <ide> var inFreeBSDJail = null; <ide> var localhostIPv4 = null; <ide><path>test/fixtures/print-chars-from-buffer.js <del>var common = require('../common'); <ide> var assert = require('assert'); <ide> <ide> var n = parseInt(process.argv[2]); <ide><path>test/fixtures/print-chars.js <del>var common = require('../common'); <ide> var assert = require('assert'); <ide> <ide> var n = parseInt(process.argv[2]); <ide><path>test/sequential/test-fs-watch-recursive.js <ide> if (process.platform === 'darwin') { <ide> watcher.on('change', function(event, filename) { <ide> assert.ok('change' === event || 'rename' === event); <ide> <del> // Ignore stale events generated by mkdir <del> if (filename === testsubdirName) <add> // Ignore stale events generated by mkdir and other tests <add> if (filename !== relativePathOne) <ide> return; <ide> <del> assert.equal(relativePathOne, filename); <del> <ide> watcher.close(); <ide> cleanup(); <ide> ++watchSeenOne; <ide><path>test/testpy/__init__.py <ide> import test <ide> import os <ide> import shutil <del>from shutil import rmtree <ide> from os import mkdir <ide> from glob import glob <ide> from os.path import join, dirname, exists <ide> def __init__(self, path, file, arch, mode, context, config, additional=[]): <ide> self.tmpdir = join(dirname(self.config.root), 'tmp') <ide> self.additional_flags = additional <ide> <del> def GetTmpDir(self): <del> return "%s.%d" % (self.tmpdir, self.thread_id) <del> <del> <del> def AfterRun(self, result): <del> # delete the whole tmp dir <del> try: <del> rmtree(self.GetTmpDir()) <del> except: <del> pass <del> # make it again. <del> try: <del> mkdir(self.GetTmpDir()) <del> except: <del> pass <del> <del> def BeforeRun(self): <del> # delete the whole tmp dir <del> try: <del> rmtree(self.GetTmpDir()) <del> except: <del> pass <del> # make it again. <del> # intermittently fails on win32, so keep trying <del> while not os.path.exists(self.GetTmpDir()): <del> try: <del> mkdir(self.GetTmpDir()) <del> except: <del> pass <ide> <ide> def GetLabel(self): <ide> return "%s %s" % (self.mode, self.GetName())
5
Go
Go
fix av in build due to userns
6d71f277608b2a655df0942d00607f47dbcaa37b
<ide><path>pkg/archive/diff.go <ide> func UnpackLayer(dest string, layer Reader, options *TarOptions) (size int64, er <ide> defer pools.BufioReader32KPool.Put(trBuf) <ide> <ide> var dirs []*tar.Header <add> <add> if options == nil { <add> options = &TarOptions{} <add> } <add> if options.ExcludePatterns == nil { <add> options.ExcludePatterns = []string{} <add> } <ide> remappedRootUID, remappedRootGID, err := idtools.GetRootUIDGID(options.UIDMaps, options.GIDMaps) <ide> if err != nil { <ide> return 0, err
1
Go
Go
resolve connection reset by peer regression
dc0ee98805c1e9282c729a79cdf10e59bad3cb09
<ide><path>daemon/attach.go <ide> func (daemon *Daemon) ContainerAttach(prefixOrName string, c *backend.ContainerA <ide> } <ide> <ide> cfg := stream.AttachConfig{ <del> UseStdin: c.UseStdin && container.Config.OpenStdin, <add> UseStdin: c.UseStdin, <ide> UseStdout: c.UseStdout, <ide> UseStderr: c.UseStderr, <ide> TTY: container.Config.Tty, <ide> func (daemon *Daemon) ContainerAttachRaw(prefixOrName string, stdin io.ReadClose <ide> return err <ide> } <ide> cfg := stream.AttachConfig{ <del> UseStdin: stdin != nil && container.Config.OpenStdin, <add> UseStdin: stdin != nil, <ide> UseStdout: stdout != nil, <ide> UseStderr: stderr != nil, <ide> TTY: container.Config.Tty, <ide> func (daemon *Daemon) containerAttach(c *container.Container, cfg *stream.Attach <ide> cfg.Stdin = r <ide> } <ide> <add> if !c.Config.OpenStdin { <add> cfg.Stdin = nil <add> } <add> <ide> waitChan := make(chan struct{}) <ide> if c.Config.StdinOnce && !c.Config.Tty { <ide> defer func() {
1
Javascript
Javascript
upgrade stats to es6
b7c1b5c6724944a92de54590a5171f2805a48cd0
<ide><path>lib/Stats.js <ide> MIT License http://www.opensource.org/licenses/mit-license.php <ide> Author Tobias Koppers @sokra <ide> */ <del>var RequestShortener = require("./RequestShortener"); <del>var SizeFormatHelpers = require("./SizeFormatHelpers"); <add>"use strict"; <ide> <del>function Stats(compilation) { <del> this.compilation = compilation; <del> this.hash = compilation.hash; <del>} <del>module.exports = Stats; <add>const RequestShortener = require("./RequestShortener"); <add>const SizeFormatHelpers = require("./SizeFormatHelpers"); <ide> <del>Stats.prototype.hasWarnings = function() { <del> return this.compilation.warnings.length > 0; <del>}; <add>const d = (v, def) => v === undefined ? def : v; <ide> <del>Stats.prototype.hasErrors = function() { <del> return this.compilation.errors.length > 0; <del>}; <add>class Stats { <add> constructor(compilation) { <add> this.compilation = compilation; <add> this.hash = compilation.hash; <add> } <ide> <del>Stats.prototype.toJson = function toJson(options, forToString) { <del> if(typeof options === "boolean" || typeof options === "string") { <del> options = Stats.presetToOptions(options); <del> } else if(!options) { <del> options = {}; <add> hasWarnings() { <add> return this.compilation.warnings.length > 0; <ide> } <ide> <del> function d(v, def) { <del> return v === undefined ? def : v; <add> hasErrors() { <add> return this.compilation.errors.length > 0; <ide> } <del> var compilation = this.compilation; <del> var requestShortener = new RequestShortener(d(options.context, process.cwd())); <del> var showPerformance = d(options.performance, true); <del> var showHash = d(options.hash, true); <del> var showVersion = d(options.version, true); <del> var showTimings = d(options.timings, true); <del> var showAssets = d(options.assets, true); <del> var showEntrypoints = d(options.entrypoints, !forToString); <del> var showChunks = d(options.chunks, true); <del> var showChunkModules = d(options.chunkModules, !!forToString); <del> var showChunkOrigins = d(options.chunkOrigins, !forToString); <del> var showModules = d(options.modules, !forToString); <del> var showDepth = d(options.depth, !forToString); <del> var showCachedModules = d(options.cached, true); <del> var showCachedAssets = d(options.cachedAssets, true); <del> var showReasons = d(options.reasons, !forToString); <del> var showUsedExports = d(options.usedExports, !forToString); <del> var showProvidedExports = d(options.providedExports, !forToString); <del> var showChildren = d(options.children, true); <del> var showSource = d(options.source, !forToString); <del> var showErrors = d(options.errors, true); <del> var showErrorDetails = d(options.errorDetails, !forToString); <del> var showWarnings = d(options.warnings, true); <del> var showPublicPath = d(options.publicPath, !forToString); <del> var excludeModules = [].concat(d(options.exclude, [])).map(function(str) { <del> if(typeof str !== "string") return str; <del> return new RegExp("[\\\\/]" + str.replace(/[\-\[\]\/\{\}\(\)\*\+\?\.\\\^\$\|]/g, "\\$&") + "([\\\\/]|$|!|\\?)"); <del> }); <del> var maxModules = d(options.maxModules, forToString ? 15 : Infinity); <del> var sortModules = d(options.modulesSort, "id"); <del> var sortChunks = d(options.chunksSort, "id"); <del> var sortAssets = d(options.assetsSort, ""); <del> <del> function createModuleFilter() { <del> var i = 0; <del> return function(module) { <del> if(!showCachedModules && !module.built) { <del> return false; <del> } <del> if(excludeModules.length > 0) { <del> var ident = requestShortener.shorten(module.resource); <del> var excluded = excludeModules.some(function(regExp) { <del> return regExp.test(ident); <del> }); <del> if(excluded) <add> <add> toJson(options, forToString) { <add> if(typeof options === "boolean" || typeof options === "string") { <add> options = Stats.presetToOptions(options); <add> } else if(!options) { <add> options = {}; <add> } <add> <add> const compilation = this.compilation; <add> const requestShortener = new RequestShortener(d(options.context, process.cwd())); <add> const showPerformance = d(options.performance, true); <add> const showHash = d(options.hash, true); <add> const showVersion = d(options.version, true); <add> const showTimings = d(options.timings, true); <add> const showAssets = d(options.assets, true); <add> const showEntrypoints = d(options.entrypoints, !forToString); <add> const showChunks = d(options.chunks, true); <add> const showChunkModules = d(options.chunkModules, !!forToString); <add> const showChunkOrigins = d(options.chunkOrigins, !forToString); <add> const showModules = d(options.modules, !forToString); <add> const showDepth = d(options.depth, !forToString); <add> const showCachedModules = d(options.cached, true); <add> const showCachedAssets = d(options.cachedAssets, true); <add> const showReasons = d(options.reasons, !forToString); <add> const showUsedExports = d(options.usedExports, !forToString); <add> const showProvidedExports = d(options.providedExports, !forToString); <add> const showChildren = d(options.children, true); <add> const showSource = d(options.source, !forToString); <add> const showErrors = d(options.errors, true); <add> const showErrorDetails = d(options.errorDetails, !forToString); <add> const showWarnings = d(options.warnings, true); <add> const showPublicPath = d(options.publicPath, !forToString); <add> const excludeModules = [].concat(d(options.exclude, [])).map(str => { <add> if(typeof str !== "string") return str; <add> return new RegExp(`[\\\\/]${str.replace(/[\-\[\]\/\{\}\(\)\*\+\?\.\\\^\$\|]/g, "\\$&")}([\\\\/]|$|!|\\?)`); <add> }); <add> const maxModules = d(options.maxModules, forToString ? 15 : Infinity); <add> const sortModules = d(options.modulesSort, "id"); <add> const sortChunks = d(options.chunksSort, "id"); <add> const sortAssets = d(options.assetsSort, ""); <add> <add> const createModuleFilter = () => { <add> let i = 0; <add> return module => { <add> if(!showCachedModules && !module.built) { <ide> return false; <add> } <add> if(excludeModules.length > 0) { <add> const ident = requestShortener.shorten(module.resource); <add> const excluded = excludeModules.some(regExp => regExp.test(ident)); <add> if(excluded) <add> return false; <add> } <add> return i++ < maxModules; <add> }; <add> } <add> <add> const sortByField = (field) => { <add> let callback; <add> if(!field) { <add> callback = () => 0; <add> return callback; <ide> } <del> return i++ < maxModules; <del> }; <del> } <ide> <del> function sortByField(field) { <del> if(!field) return function() { <del> return 0; <del> }; <del> if(field[0] === "!") { <del> field = field.substr(1); <del> return function(a, b) { <add> if(field[0] === "!") { <add> field = field.substr(1); <add> callback = (a, b) => { <add> if(a[field] === null && b[field] === null) return 0; <add> if(a[field] === null) return 1; <add> if(b[field] === null) return -1; <add> if(a[field] === b[field]) return 0; <add> return a[field] < b[field] ? 1 : -1; <add> }; <add> } <add> callback = (a, b) => { <ide> if(a[field] === null && b[field] === null) return 0; <ide> if(a[field] === null) return 1; <ide> if(b[field] === null) return -1; <ide> if(a[field] === b[field]) return 0; <del> return a[field] < b[field] ? 1 : -1; <add> return a[field] < b[field] ? -1 : 1; <ide> }; <add> return callback; <ide> } <del> return function(a, b) { <del> if(a[field] === null && b[field] === null) return 0; <del> if(a[field] === null) return 1; <del> if(b[field] === null) return -1; <del> if(a[field] === b[field]) return 0; <del> return a[field] < b[field] ? -1 : 1; <del> }; <del> } <ide> <del> function formatError(e) { <del> var text = ""; <del> if(typeof e === "string") <del> e = { <del> message: e <del> }; <del> if(e.chunk) { <del> text += "chunk " + (e.chunk.name || e.chunk.id) + <del> (e.chunk.hasRuntime() ? " [entry]" : e.chunk.isInitial() ? " [initial]" : "") + "\n"; <del> } <del> if(e.file) { <del> text += e.file + "\n"; <del> } <del> if(e.module && e.module.readableIdentifier && typeof e.module.readableIdentifier === "function") { <del> text += e.module.readableIdentifier(requestShortener) + "\n"; <del> } <del> text += e.message; <del> if(showErrorDetails && e.details) text += "\n" + e.details; <del> if(showErrorDetails && e.missing) text += e.missing.map(function(item) { <del> return "\n[" + item + "]"; <del> }).join(""); <del> if(e.dependencies && e.origin) { <del> text += "\n @ " + e.origin.readableIdentifier(requestShortener); <del> e.dependencies.forEach(function(dep) { <del> if(!dep.loc) return; <del> if(typeof dep.loc === "string") return; <del> if(!dep.loc.start) return; <del> if(!dep.loc.end) return; <del> text += " " + dep.loc.start.line + ":" + dep.loc.start.column + "-" + <del> (dep.loc.start.line !== dep.loc.end.line ? dep.loc.end.line + ":" : "") + dep.loc.end.column; <del> }); <del> var current = e.origin; <del> while(current.issuer) { <del> current = current.issuer; <del> text += "\n @ " + current.readableIdentifier(requestShortener); <add> const formatError = (e) => { <add> let text = ""; <add> if(typeof e === "string") <add> e = { <add> message: e <add> }; <add> if(e.chunk) { <add> text += `chunk ${e.chunk.name || e.chunk.id}${e.chunk.hasRuntime() ? " [entry]" : e.chunk.isInitial() ? " [initial]" : ""}\n`; <add> } <add> if(e.file) { <add> text += `${e.file}\n`; <add> } <add> if(e.module && e.module.readableIdentifier && typeof e.module.readableIdentifier === "function") { <add> text += `${e.module.readableIdentifier(requestShortener)}\n`; <add> } <add> text += e.message; <add> if(showErrorDetails && e.details) text += `\n${e.details}`; <add> if(showErrorDetails && e.missing) text += e.missing.map(item => `\n[${item}]`).join(""); <add> if(e.dependencies && e.origin) { <add> text += `\n @ ${e.origin.readableIdentifier(requestShortener)}`; <add> e.dependencies.forEach(dep => { <add> if(!dep.loc) return; <add> if(typeof dep.loc === "string") return; <add> if(!dep.loc.start) return; <add> if(!dep.loc.end) return; <add> text += ` ${dep.loc.start.line}:${dep.loc.start.column}-${dep.loc.start.line !== dep.loc.end.line ? dep.loc.end.line + ":" : ""}${dep.loc.end.column}`; <add> }); <add> let current = e.origin; <add> while(current.issuer) { <add> current = current.issuer; <add> text += `\n @ ${current.readableIdentifier(requestShortener)}`; <add> } <ide> } <add> return text; <ide> } <del> return text; <del> } <ide> <del> var obj = { <del> errors: compilation.errors.map(formatError), <del> warnings: compilation.warnings.map(formatError) <del> }; <del> <del> //We just hint other renderers since actually omitting <del> //errors/warnings from the JSON would be kind of weird. <del> Object.defineProperty(obj, "_showWarnings", { <del> value: showWarnings, <del> enumerable: false <del> }); <del> Object.defineProperty(obj, "_showErrors", { <del> value: showErrors, <del> enumerable: false <del> }); <del> <del> if(showVersion) { <del> obj.version = require("../package.json").version; <del> } <add> const obj = { <add> errors: compilation.errors.map(formatError), <add> warnings: compilation.warnings.map(formatError) <add> }; <ide> <del> if(showHash) obj.hash = this.hash; <del> if(showTimings && this.startTime && this.endTime) { <del> obj.time = this.endTime - this.startTime; <del> } <del> if(compilation.needAdditionalPass) { <del> obj.needAdditionalPass = true; <del> } <del> if(showPublicPath) { <del> obj.publicPath = this.compilation.mainTemplate.getPublicPath({ <del> hash: this.compilation.hash <add> //We just hint other renderers since actually omitting <add> //errors/warnings from the JSON would be kind of weird. <add> Object.defineProperty(obj, "_showWarnings", { <add> value: showWarnings, <add> enumerable: false <add> }); <add> Object.defineProperty(obj, "_showErrors", { <add> value: showErrors, <add> enumerable: false <ide> }); <del> } <del> if(showAssets) { <del> var assetsByFile = {}; <del> obj.assetsByChunkName = {}; <del> obj.assets = Object.keys(compilation.assets).map(function(asset) { <del> var obj = { <del> name: asset, <del> size: compilation.assets[asset].size(), <del> chunks: [], <del> chunkNames: [], <del> emitted: compilation.assets[asset].emitted <del> }; <ide> <del> if(showPerformance) { <del> obj.isOverSizeLimit = compilation.assets[asset].isOverSizeLimit; <del> } <add> if(showVersion) { <add> obj.version = require("../package.json").version; <add> } <ide> <del> assetsByFile[asset] = obj; <del> return obj; <del> }).filter(function(asset) { <del> return showCachedAssets || asset.emitted; <del> }); <del> compilation.chunks.forEach(function(chunk) { <del> chunk.files.forEach(function(asset) { <del> if(assetsByFile[asset]) { <del> chunk.ids.forEach(function(id) { <del> assetsByFile[asset].chunks.push(id); <del> }); <del> if(chunk.name) { <del> assetsByFile[asset].chunkNames.push(chunk.name); <del> if(obj.assetsByChunkName[chunk.name]) <del> obj.assetsByChunkName[chunk.name] = [].concat(obj.assetsByChunkName[chunk.name]).concat([asset]); <del> else <del> obj.assetsByChunkName[chunk.name] = asset; <del> } <del> } <add> if(showHash) obj.hash = this.hash; <add> if(showTimings && this.startTime && this.endTime) { <add> obj.time = this.endTime - this.startTime; <add> } <add> if(compilation.needAdditionalPass) { <add> obj.needAdditionalPass = true; <add> } <add> if(showPublicPath) { <add> obj.publicPath = this.compilation.mainTemplate.getPublicPath({ <add> hash: this.compilation.hash <ide> }); <del> }); <del> obj.assets.sort(sortByField(sortAssets)); <del> } <add> } <add> if(showAssets) { <add> const assetsByFile = {}; <add> obj.assetsByChunkName = {}; <add> obj.assets = Object.keys(compilation.assets).map(asset => { <add> const obj = { <add> name: asset, <add> size: compilation.assets[asset].size(), <add> chunks: [], <add> chunkNames: [], <add> emitted: compilation.assets[asset].emitted <add> }; <ide> <del> if(showEntrypoints) { <del> obj.entrypoints = {}; <del> Object.keys(compilation.entrypoints).forEach(function(name) { <del> var ep = compilation.entrypoints[name]; <del> obj.entrypoints[name] = { <del> chunks: ep.chunks.map(function(c) { <del> return c.id; <del> }), <del> assets: ep.chunks.reduce(function(array, c) { <del> return array.concat(c.files || []); <del> }, []) <del> }; <del> if(showPerformance) { <del> obj.entrypoints[name].isOverSizeLimit = ep.isOverSizeLimit; <del> } <del> }); <del> } <add> if(showPerformance) { <add> obj.isOverSizeLimit = compilation.assets[asset].isOverSizeLimit; <add> } <ide> <del> function fnModule(module) { <del> var obj = { <del> id: module.id, <del> identifier: module.identifier(), <del> name: module.readableIdentifier(requestShortener), <del> index: module.index, <del> index2: module.index2, <del> size: module.size(), <del> cacheable: !!module.cacheable, <del> built: !!module.built, <del> optional: !!module.optional, <del> prefetched: !!module.prefetched, <del> chunks: module.chunks.map(function(chunk) { <del> return chunk.id; <del> }), <del> assets: Object.keys(module.assets || {}), <del> issuer: module.issuer && module.issuer.identifier(), <del> issuerId: module.issuer && module.issuer.id, <del> issuerName: module.issuer && module.issuer.readableIdentifier(requestShortener), <del> profile: module.profile, <del> failed: !!module.error, <del> errors: module.errors && module.dependenciesErrors && (module.errors.length + module.dependenciesErrors.length), <del> warnings: module.errors && module.dependenciesErrors && (module.warnings.length + module.dependenciesWarnings.length) <del> }; <del> if(showReasons) { <del> obj.reasons = module.reasons.filter(function(reason) { <del> return reason.dependency && reason.module; <del> }).map(function(reason) { <del> var obj = { <del> moduleId: reason.module.id, <del> moduleIdentifier: reason.module.identifier(), <del> module: reason.module.readableIdentifier(requestShortener), <del> moduleName: reason.module.readableIdentifier(requestShortener), <del> type: reason.dependency.type, <del> userRequest: reason.dependency.userRequest <del> }; <del> var dep = reason.dependency; <del> if(dep.templateModules) obj.templateModules = dep.templateModules.map(function(module) { <del> return module.id; <del> }); <del> if(typeof dep.loc === "object") obj.loc = dep.loc.start.line + ":" + dep.loc.start.column + "-" + <del> (dep.loc.start.line !== dep.loc.end.line ? dep.loc.end.line + ":" : "") + dep.loc.end.column; <add> assetsByFile[asset] = obj; <ide> return obj; <del> }).sort(function(a, b) { <del> return a.moduleId - b.moduleId; <add> }).filter(asset => showCachedAssets || asset.emitted); <add> <add> compilation.chunks.forEach(chunk => { <add> chunk.files.forEach(asset => { <add> if(assetsByFile[asset]) { <add> chunk.ids.forEach(id => { <add> assetsByFile[asset].chunks.push(id); <add> }); <add> if(chunk.name) { <add> assetsByFile[asset].chunkNames.push(chunk.name); <add> if(obj.assetsByChunkName[chunk.name]) <add> obj.assetsByChunkName[chunk.name] = [].concat(obj.assetsByChunkName[chunk.name]).concat([asset]); <add> else <add> obj.assetsByChunkName[chunk.name] = asset; <add> } <add> } <add> }); <ide> }); <add> obj.assets.sort(sortByField(sortAssets)); <ide> } <del> if(showUsedExports) { <del> obj.usedExports = module.used ? module.usedExports : false; <del> } <del> if(showProvidedExports) { <del> obj.providedExports = Array.isArray(module.providedExports) ? module.providedExports : null; <del> } <del> if(showDepth) { <del> obj.depth = module.depth; <del> } <del> if(showSource && module._source) { <del> obj.source = module._source.source(); <add> <add> if(showEntrypoints) { <add> obj.entrypoints = {}; <add> Object.keys(compilation.entrypoints).forEach(name => { <add> const ep = compilation.entrypoints[name]; <add> obj.entrypoints[name] = { <add> chunks: ep.chunks.map(c => c.id), <add> assets: ep.chunks.reduce((array, c) => array.concat(c.files || []), []) <add> }; <add> if(showPerformance) { <add> obj.entrypoints[name].isOverSizeLimit = ep.isOverSizeLimit; <add> } <add> }); <ide> } <del> return obj; <del> } <del> if(showChunks) { <del> obj.chunks = compilation.chunks.map(function(chunk) { <del> var obj = { <del> id: chunk.id, <del> rendered: chunk.rendered, <del> initial: chunk.isInitial(), <del> entry: chunk.hasRuntime(), <del> recorded: chunk.recorded, <del> extraAsync: !!chunk.extraAsync, <del> size: chunk.modules.reduce(function(size, module) { <del> return size + module.size(); <del> }, 0), <del> names: chunk.name ? [chunk.name] : [], <del> files: chunk.files.slice(), <del> hash: chunk.renderedHash, <del> parents: chunk.parents.map(function(c) { <del> return c.id; <del> }) <add> <add> function fnModule(module) { <add> const obj = { <add> id: module.id, <add> identifier: module.identifier(), <add> name: module.readableIdentifier(requestShortener), <add> index: module.index, <add> index2: module.index2, <add> size: module.size(), <add> cacheable: !!module.cacheable, <add> built: !!module.built, <add> optional: !!module.optional, <add> prefetched: !!module.prefetched, <add> chunks: module.chunks.map(chunk => chunk.id), <add> assets: Object.keys(module.assets || {}), <add> issuer: module.issuer && module.issuer.identifier(), <add> issuerId: module.issuer && module.issuer.id, <add> issuerName: module.issuer && module.issuer.readableIdentifier(requestShortener), <add> profile: module.profile, <add> failed: !!module.error, <add> errors: module.errors && module.dependenciesErrors && (module.errors.length + module.dependenciesErrors.length), <add> warnings: module.errors && module.dependenciesErrors && (module.warnings.length + module.dependenciesWarnings.length) <ide> }; <del> if(showChunkModules) { <del> obj.modules = chunk.modules <del> .slice() <del> .sort(sortByField("depth")) <del> .filter(createModuleFilter()) <del> .map(fnModule); <del> obj.filteredModules = chunk.modules.length - obj.modules.length; <del> obj.modules.sort(sortByField(sortModules)); <add> if(showReasons) { <add> obj.reasons = module.reasons.filter(reason => reason.dependency && reason.module).map(reason => { <add> const obj = { <add> moduleId: reason.module.id, <add> moduleIdentifier: reason.module.identifier(), <add> module: reason.module.readableIdentifier(requestShortener), <add> moduleName: reason.module.readableIdentifier(requestShortener), <add> type: reason.dependency.type, <add> userRequest: reason.dependency.userRequest <add> }; <add> const dep = reason.dependency; <add> if(dep.templateModules) obj.templateModules = dep.templateModules.map(module => module.id); <add> if(typeof dep.loc === "object") obj.loc = `${dep.loc.start.line}:${dep.loc.start.column}-${dep.loc.start.line !== dep.loc.end.line ? dep.loc.end.line + ":" : ""}${dep.loc.end.column}`; <add> return obj; <add> }).sort((a, b) => a.moduleId - b.moduleId); <ide> } <del> if(showChunkOrigins) { <del> obj.origins = chunk.origins.map(function(origin) { <del> return { <add> if(showUsedExports) { <add> obj.usedExports = module.used ? module.usedExports : false; <add> } <add> if(showProvidedExports) { <add> obj.providedExports = Array.isArray(module.providedExports) ? module.providedExports : null; <add> } <add> if(showDepth) { <add> obj.depth = module.depth; <add> } <add> if(showSource && module._source) { <add> obj.source = module._source.source(); <add> } <add> return obj; <add> } <add> if(showChunks) { <add> obj.chunks = compilation.chunks.map(chunk => { <add> const obj = { <add> id: chunk.id, <add> rendered: chunk.rendered, <add> initial: chunk.isInitial(), <add> entry: chunk.hasRuntime(), <add> recorded: chunk.recorded, <add> extraAsync: !!chunk.extraAsync, <add> size: chunk.modules.reduce((size, module) => size + module.size(), 0), <add> names: chunk.name ? [chunk.name] : [], <add> files: chunk.files.slice(), <add> hash: chunk.renderedHash, <add> parents: chunk.parents.map(c => c.id) <add> }; <add> if(showChunkModules) { <add> obj.modules = chunk.modules <add> .slice() <add> .sort(sortByField("depth")) <add> .filter(createModuleFilter()) <add> .map(fnModule); <add> obj.filteredModules = chunk.modules.length - obj.modules.length; <add> obj.modules.sort(sortByField(sortModules)); <add> } <add> if(showChunkOrigins) { <add> obj.origins = chunk.origins.map(origin => ({ <ide> moduleId: origin.module ? origin.module.id : undefined, <ide> module: origin.module ? origin.module.identifier() : "", <ide> moduleIdentifier: origin.module ? origin.module.identifier() : "", <ide> moduleName: origin.module ? origin.module.readableIdentifier(requestShortener) : "", <del> loc: typeof origin.loc === "object" ? obj.loc = origin.loc.start.line + ":" + origin.loc.start.column + "-" + <del> (origin.loc.start.line !== origin.loc.end.line ? origin.loc.end.line + ":" : "") + origin.loc.end.column : "", <add> loc: typeof origin.loc === "object" ? obj.loc = `${origin.loc.start.line}:${origin.loc.start.column}-${origin.loc.start.line !== origin.loc.end.line ? origin.loc.end.line + ":" : ""}${origin.loc.end.column}` : "", <ide> name: origin.name, <ide> reasons: origin.reasons || [] <del> }; <del> }); <del> } <del> return obj; <del> }); <del> obj.chunks.sort(sortByField(sortChunks)); <del> } <del> if(showModules) { <del> obj.modules = compilation.modules <del> .slice() <del> .sort(sortByField("depth")) <del> .filter(createModuleFilter()) <del> .map(fnModule); <del> obj.filteredModules = compilation.modules.length - obj.modules.length; <del> obj.modules.sort(sortByField(sortModules)); <del> } <del> if(showChildren) { <del> obj.children = compilation.children.map(function(child) { <del> var obj = new Stats(child).toJson(options, forToString); <del> delete obj.hash; <del> delete obj.version; <del> obj.name = child.name; <del> return obj; <del> }); <add> })); <add> } <add> return obj; <add> }); <add> obj.chunks.sort(sortByField(sortChunks)); <add> } <add> if(showModules) { <add> obj.modules = compilation.modules <add> .slice() <add> .sort(sortByField("depth")) <add> .filter(createModuleFilter()) <add> .map(fnModule); <add> obj.filteredModules = compilation.modules.length - obj.modules.length; <add> obj.modules.sort(sortByField(sortModules)); <add> } <add> if(showChildren) { <add> obj.children = compilation.children.map(child => { <add> const obj = new Stats(child).toJson(options, forToString); <add> delete obj.hash; <add> delete obj.version; <add> obj.name = child.name; <add> return obj; <add> }); <add> } <add> <add> return obj; <ide> } <ide> <del> return obj; <del>}; <add> toString(options) { <add> if(typeof options === "boolean" || typeof options === "string") { <add> options = Stats.presetToOptions(options); <add> } else if(!options) { <add> options = {}; <add> } <ide> <del>Stats.prototype.toString = function toString(options) { <del> if(typeof options === "boolean" || typeof options === "string") { <del> options = Stats.presetToOptions(options); <del> } else if(!options) options = {}; <add> const useColors = d(options.colors, false); <ide> <del> function d(v, def) { <del> return v === undefined ? def : v; <del> } <del> var useColors = d(options.colors, false); <del> <del> var obj = this.toJson(options, true); <del> <del> return Stats.jsonToString(obj, useColors); <del>}; <del> <del>Stats.jsonToString = function jsonToString(obj, useColors) { <del> var buf = []; <del> <del> var defaultColors = { <del> bold: "\u001b[1m", <del> yellow: "\u001b[1m\u001b[33m", <del> red: "\u001b[1m\u001b[31m", <del> green: "\u001b[1m\u001b[32m", <del> cyan: "\u001b[1m\u001b[36m", <del> magenta: "\u001b[1m\u001b[35m" <del> }; <del> <del> var colors = Object.keys(defaultColors).reduce(function(obj, color) { <del> obj[color] = function(str) { <del> if(useColors) { <del> buf.push( <del> (useColors === true || useColors[color] === undefined) ? <del> defaultColors[color] : useColors[color] <del> ); <del> } <del> buf.push(str); <del> if(useColors) { <del> buf.push("\u001b[39m\u001b[22m"); <del> } <del> }; <del> return obj; <del> }, { <del> normal: function(str) { <del> buf.push(str); <del> } <del> }); <del> <del> function coloredTime(time) { <del> var times = [800, 400, 200, 100]; <del> if(obj.time) { <del> times = [obj.time / 2, obj.time / 4, obj.time / 8, obj.time / 16]; <del> } <del> if(time < times[3]) <del> colors.normal(time + "ms"); <del> else if(time < times[2]) <del> colors.bold(time + "ms"); <del> else if(time < times[1]) <del> colors.green(time + "ms"); <del> else if(time < times[0]) <del> colors.yellow(time + "ms"); <del> else <del> colors.red(time + "ms"); <del> } <add> const obj = this.toJson(options, true); <ide> <del> function newline() { <del> buf.push("\n"); <add> return Stats.jsonToString(obj, useColors); <ide> } <ide> <del> function getText(arr, row, col) { <del> return arr[row][col].value; <del> } <add> static jsonToString(obj, useColors) { <add> const buf = []; <ide> <del> function table(array, align, splitter) { <del> var row; <del> var rows = array.length; <del> var col; <del> var cols = array[0].length; <del> var colSizes = new Array(cols); <del> var value; <del> for(col = 0; col < cols; col++) <del> colSizes[col] = 0; <del> for(row = 0; row < rows; row++) { <del> for(col = 0; col < cols; col++) { <del> value = getText(array, row, col) + ""; <del> if(value.length > colSizes[col]) { <del> colSizes[col] = value.length; <add> const defaultColors = { <add> bold: "\u001b[1m", <add> yellow: "\u001b[1m\u001b[33m", <add> red: "\u001b[1m\u001b[31m", <add> green: "\u001b[1m\u001b[32m", <add> cyan: "\u001b[1m\u001b[36m", <add> magenta: "\u001b[1m\u001b[35m" <add> }; <add> <add> const colors = Object.keys(defaultColors).reduce((obj, color) => { <add> obj[color] = str => { <add> if(useColors) { <add> buf.push( <add> (useColors === true || useColors[color] === undefined) ? <add> defaultColors[color] : useColors[color] <add> ); <ide> } <del> } <del> } <del> for(row = 0; row < rows; row++) { <del> for(col = 0; col < cols; col++) { <del> var format = array[row][col].color; <del> value = getText(array, row, col) + ""; <del> var l = value.length; <del> if(align[col] === "l") <del> format(value); <del> for(; l < colSizes[col] && col !== cols - 1; l++) <del> colors.normal(" "); <del> if(align[col] === "r") <del> format(value); <del> if(col + 1 < cols && colSizes[col] !== 0) <del> colors.normal(splitter || " "); <del> } <del> newline(); <del> } <del> } <add> buf.push(str); <add> if(useColors) { <add> buf.push("\u001b[39m\u001b[22m"); <add> } <add> }; <add> return obj; <add> }, { <add> normal: (str) => buf.push(str) <add> }); <ide> <del> function getAssetColor(asset, defaultColor) { <del> if(asset.isOverSizeLimit) { <del> return colors.yellow; <add> const coloredTime = (time) => { <add> let times = [800, 400, 200, 100]; <add> if(obj.time) { <add> times = [obj.time / 2, obj.time / 4, obj.time / 8, obj.time / 16]; <add> } <add> if(time < times[3]) <add> colors.normal(`${time}ms`); <add> else if(time < times[2]) <add> colors.bold(`${time}ms`); <add> else if(time < times[1]) <add> colors.green(`${time}ms`); <add> else if(time < times[0]) <add> colors.yellow(`${time}ms`); <add> else <add> colors.red(`${time}ms`); <ide> } <ide> <del> return defaultColor; <del> } <add> const newline = () => buf.push("\n"); <ide> <del> if(obj.hash) { <del> colors.normal("Hash: "); <del> colors.bold(obj.hash); <del> newline(); <del> } <del> if(obj.version) { <del> colors.normal("Version: webpack "); <del> colors.bold(obj.version); <del> newline(); <del> } <del> if(typeof obj.time === "number") { <del> colors.normal("Time: "); <del> colors.bold(obj.time); <del> colors.normal("ms"); <del> newline(); <del> } <del> if(obj.publicPath) { <del> colors.normal("PublicPath: "); <del> colors.bold(obj.publicPath); <del> newline(); <del> } <add> const getText = (arr, row, col) => { <add> return arr[row][col].value; <add> } <ide> <del> if(obj.assets && obj.assets.length > 0) { <del> var t = [ <del> [{ <del> value: "Asset", <del> color: colors.bold <del> }, { <del> value: "Size", <del> color: colors.bold <del> }, { <del> value: "Chunks", <del> color: colors.bold <del> }, { <del> value: "", <del> color: colors.bold <del> }, { <del> value: "", <del> color: colors.bold <del> }, { <del> value: "Chunk Names", <del> color: colors.bold <del> }] <del> ]; <del> obj.assets.forEach(function(asset) { <del> t.push([{ <del> value: asset.name, <del> color: getAssetColor(asset, colors.green) <del> }, { <del> value: SizeFormatHelpers.formatSize(asset.size), <del> color: getAssetColor(asset, colors.normal) <del> }, { <del> value: asset.chunks.join(", "), <del> color: colors.bold <del> }, { <del> value: asset.emitted ? "[emitted]" : "", <del> color: colors.green <del> }, { <del> value: asset.isOverSizeLimit ? "[big]" : "", <del> color: getAssetColor(asset, colors.normal) <del> }, { <del> value: asset.chunkNames.join(", "), <del> color: colors.normal <del> }]); <del> }); <del> table(t, "rrrlll"); <del> } <del> if(obj.entrypoints) { <del> Object.keys(obj.entrypoints).forEach(function(name) { <del> var ep = obj.entrypoints[name]; <del> colors.normal("Entrypoint "); <del> colors.bold(name); <del> if(ep.isOverSizeLimit) { <del> colors.normal(" "); <del> colors.yellow("[big]"); <add> const table = (array, align, splitter) => { <add> let row; <add> const rows = array.length; <add> let col; <add> const cols = array[0].length; <add> const colSizes = new Array(cols); <add> let value; <add> for(col = 0; col < cols; col++) <add> colSizes[col] = 0; <add> for(row = 0; row < rows; row++) { <add> for(col = 0; col < cols; col++) { <add> value = `${getText(array, row, col)}`; <add> if(value.length > colSizes[col]) { <add> colSizes[col] = value.length; <add> } <add> } <ide> } <del> colors.normal(" ="); <del> ep.assets.forEach(function(asset) { <del> colors.normal(" "); <del> colors.green(asset); <del> }); <del> newline(); <del> }); <del> } <del> var modulesByIdentifier = {}; <del> if(obj.modules) { <del> obj.modules.forEach(function(module) { <del> modulesByIdentifier["$" + module.identifier] = module; <del> }); <del> } else if(obj.chunks) { <del> obj.chunks.forEach(function(chunk) { <del> if(chunk.modules) { <del> chunk.modules.forEach(function(module) { <del> modulesByIdentifier["$" + module.identifier] = module; <del> }); <add> for(row = 0; row < rows; row++) { <add> for(col = 0; col < cols; col++) { <add> const format = array[row][col].color; <add> value = `${getText(array, row, col)}`; <add> let l = value.length; <add> if(align[col] === "l") <add> format(value); <add> for(; l < colSizes[col] && col !== cols - 1; l++) <add> colors.normal(" "); <add> if(align[col] === "r") <add> format(value); <add> if(col + 1 < cols && colSizes[col] !== 0) <add> colors.normal(splitter || " "); <add> } <add> newline(); <ide> } <del> }); <del> } <del> <del> function processModuleAttributes(module) { <del> colors.normal(" "); <del> colors.normal(SizeFormatHelpers.formatSize(module.size)); <del> if(module.chunks) { <del> module.chunks.forEach(function(chunk) { <del> colors.normal(" {"); <del> colors.yellow(chunk); <del> colors.normal("}"); <del> }); <ide> } <del> if(typeof module.depth === "number") { <del> colors.normal(" [depth " + module.depth + "]"); <del> } <del> if(!module.cacheable) { <del> colors.red(" [not cacheable]"); <add> <add> const getAssetColor = (asset, defaultColor) => { <add> if(asset.isOverSizeLimit) { <add> return colors.yellow; <add> } <add> <add> return defaultColor; <ide> } <del> if(module.optional) { <del> colors.yellow(" [optional]"); <add> <add> if(obj.hash) { <add> colors.normal("Hash: "); <add> colors.bold(obj.hash); <add> newline(); <ide> } <del> if(module.built) { <del> colors.green(" [built]"); <add> if(obj.version) { <add> colors.normal("Version: webpack "); <add> colors.bold(obj.version); <add> newline(); <ide> } <del> if(module.prefetched) { <del> colors.magenta(" [prefetched]"); <add> if(typeof obj.time === "number") { <add> colors.normal("Time: "); <add> colors.bold(obj.time); <add> colors.normal("ms"); <add> newline(); <ide> } <del> if(module.failed) <del> colors.red(" [failed]"); <del> if(module.warnings) <del> colors.yellow(" [" + module.warnings + " warning" + (module.warnings === 1 ? "" : "s") + "]"); <del> if(module.errors) <del> colors.red(" [" + module.errors + " error" + (module.errors === 1 ? "" : "s") + "]"); <del> } <del> <del> function processModuleContent(module, prefix) { <del> if(Array.isArray(module.providedExports)) { <del> colors.normal(prefix); <del> colors.cyan("[exports: " + module.providedExports.join(", ") + "]"); <add> if(obj.publicPath) { <add> colors.normal("PublicPath: "); <add> colors.bold(obj.publicPath); <ide> newline(); <ide> } <del> if(module.usedExports !== undefined) { <del> if(module.usedExports !== true) { <del> colors.normal(prefix); <del> if(module.usedExports === false) <del> colors.cyan("[no exports used]"); <del> else <del> colors.cyan("[only some exports used: " + module.usedExports.join(", ") + "]"); <del> newline(); <del> } <add> <add> if(obj.assets && obj.assets.length > 0) { <add> const t = [ <add> [{ <add> value: "Asset", <add> color: colors.bold <add> }, { <add> value: "Size", <add> color: colors.bold <add> }, { <add> value: "Chunks", <add> color: colors.bold <add> }, { <add> value: "", <add> color: colors.bold <add> }, { <add> value: "", <add> color: colors.bold <add> }, { <add> value: "Chunk Names", <add> color: colors.bold <add> }] <add> ]; <add> obj.assets.forEach(asset => { <add> t.push([{ <add> value: asset.name, <add> color: getAssetColor(asset, colors.green) <add> }, { <add> value: SizeFormatHelpers.formatSize(asset.size), <add> color: getAssetColor(asset, colors.normal) <add> }, { <add> value: asset.chunks.join(", "), <add> color: colors.bold <add> }, { <add> value: asset.emitted ? "[emitted]" : "", <add> color: colors.green <add> }, { <add> value: asset.isOverSizeLimit ? "[big]" : "", <add> color: getAssetColor(asset, colors.normal) <add> }, { <add> value: asset.chunkNames.join(", "), <add> color: colors.normal <add> }]); <add> }); <add> table(t, "rrrlll"); <ide> } <del> if(module.reasons) { <del> module.reasons.forEach(function(reason) { <del> colors.normal(prefix); <del> colors.normal(reason.type); <del> colors.normal(" "); <del> colors.cyan(reason.userRequest); <del> if(reason.templateModules) colors.cyan(reason.templateModules.join(" ")); <del> colors.normal(" ["); <del> colors.normal(reason.moduleId); <del> colors.normal("] "); <del> colors.magenta(reason.module); <del> if(reason.loc) { <add> if(obj.entrypoints) { <add> Object.keys(obj.entrypoints).forEach(name => { <add> const ep = obj.entrypoints[name]; <add> colors.normal("Entrypoint "); <add> colors.bold(name); <add> if(ep.isOverSizeLimit) { <ide> colors.normal(" "); <del> colors.normal(reason.loc); <add> colors.yellow("[big]"); <ide> } <add> colors.normal(" ="); <add> ep.assets.forEach(asset => { <add> colors.normal(" "); <add> colors.green(asset); <add> }); <ide> newline(); <ide> }); <ide> } <del> if(module.profile) { <del> colors.normal(prefix); <del> var sum = 0; <del> var path = []; <del> var current = module; <del> while(current.issuer) { <del> path.unshift(current = current.issuer); <del> } <del> path.forEach(function(module) { <del> colors.normal("["); <del> colors.normal(module.id); <del> colors.normal("] "); <del> if(module.profile) { <del> var time = (module.profile.factory || 0) + (module.profile.building || 0); <del> coloredTime(time); <del> sum += time; <del> colors.normal(" "); <del> } <del> colors.normal("->"); <add> const modulesByIdentifier = {}; <add> if(obj.modules) { <add> obj.modules.forEach(module => { <add> modulesByIdentifier[`$${module.identifier}`] = module; <ide> }); <del> Object.keys(module.profile).forEach(function(key) { <del> colors.normal(" " + key + ":"); <del> var time = module.profile[key]; <del> coloredTime(time); <del> sum += time; <add> } else if(obj.chunks) { <add> obj.chunks.forEach(chunk => { <add> if(chunk.modules) { <add> chunk.modules.forEach(module => { <add> modulesByIdentifier[`$${module.identifier}`] = module; <add> }); <add> } <ide> }); <del> colors.normal(" = "); <del> coloredTime(sum); <del> newline(); <ide> } <del> } <ide> <del> if(obj.chunks) { <del> obj.chunks.forEach(function(chunk) { <del> colors.normal("chunk "); <del> if(chunk.id < 1000) colors.normal(" "); <del> if(chunk.id < 100) colors.normal(" "); <del> if(chunk.id < 10) colors.normal(" "); <del> colors.normal("{"); <del> colors.yellow(chunk.id); <del> colors.normal("} "); <del> colors.green(chunk.files.join(", ")); <del> if(chunk.names && chunk.names.length > 0) { <del> colors.normal(" ("); <del> colors.normal(chunk.names.join(", ")); <del> colors.normal(")"); <del> } <add> const processModuleAttributes = (module) => { <ide> colors.normal(" "); <del> colors.normal(SizeFormatHelpers.formatSize(chunk.size)); <del> chunk.parents.forEach(function(id) { <del> colors.normal(" {"); <del> colors.yellow(id); <del> colors.normal("}"); <del> }); <del> if(chunk.entry) { <del> colors.yellow(" [entry]"); <del> } else if(chunk.initial) { <del> colors.yellow(" [initial]"); <add> colors.normal(SizeFormatHelpers.formatSize(module.size)); <add> if(module.chunks) { <add> module.chunks.forEach(chunk => { <add> colors.normal(" {"); <add> colors.yellow(chunk); <add> colors.normal("}"); <add> }); <ide> } <del> if(chunk.rendered) { <del> colors.green(" [rendered]"); <add> if(typeof module.depth === "number") { <add> colors.normal(` [depth ${module.depth}]`); <ide> } <del> if(chunk.recorded) { <del> colors.green(" [recorded]"); <add> if(!module.cacheable) { <add> colors.red(" [not cacheable]"); <ide> } <del> newline(); <del> if(chunk.origins) { <del> chunk.origins.forEach(function(origin) { <del> colors.normal(" > "); <del> if(origin.reasons && origin.reasons.length) { <del> colors.yellow(origin.reasons.join(" ")); <del> colors.normal(" "); <del> } <del> if(origin.name) { <del> colors.normal(origin.name); <add> if(module.optional) { <add> colors.yellow(" [optional]"); <add> } <add> if(module.built) { <add> colors.green(" [built]"); <add> } <add> if(module.prefetched) { <add> colors.magenta(" [prefetched]"); <add> } <add> if(module.failed) <add> colors.red(" [failed]"); <add> if(module.warnings) <add> colors.yellow(` [${module.warnings} warning${module.warnings === 1 ? "" : "s"}]`); <add> if(module.errors) <add> colors.red(` [${module.errors} error${module.errors === 1 ? "" : "s"}]`); <add> } <add> <add> const processModuleContent = (module, prefix) => { <add> if(Array.isArray(module.providedExports)) { <add> colors.normal(prefix); <add> colors.cyan(`[exports: ${module.providedExports.join(", ")}]`); <add> newline(); <add> } <add> if(module.usedExports !== undefined) { <add> if(module.usedExports !== true) { <add> colors.normal(prefix); <add> if(module.usedExports === false) <add> colors.cyan("[no exports used]"); <add> else <add> colors.cyan(`[only some exports used: ${module.usedExports.join(", ")}]`); <add> newline(); <add> } <add> } <add> if(module.reasons) { <add> module.reasons.forEach(reason => { <add> colors.normal(prefix); <add> colors.normal(reason.type); <add> colors.normal(" "); <add> colors.cyan(reason.userRequest); <add> if(reason.templateModules) colors.cyan(reason.templateModules.join(" ")); <add> colors.normal(" ["); <add> colors.normal(reason.moduleId); <add> colors.normal("] "); <add> colors.magenta(reason.module); <add> if(reason.loc) { <ide> colors.normal(" "); <del> } <del> if(origin.module) { <del> colors.normal("["); <del> colors.normal(origin.moduleId); <del> colors.normal("] "); <del> var module = modulesByIdentifier["$" + origin.module]; <del> if(module) { <del> colors.bold(module.name); <del> colors.normal(" "); <del> } <del> if(origin.loc) { <del> colors.normal(origin.loc); <del> } <add> colors.normal(reason.loc); <ide> } <ide> newline(); <ide> }); <ide> } <del> if(chunk.modules) { <del> chunk.modules.forEach(function(module) { <del> colors.normal(" "); <del> if(module.id < 1000) colors.normal(" "); <del> if(module.id < 100) colors.normal(" "); <del> if(module.id < 10) colors.normal(" "); <add> if(module.profile) { <add> colors.normal(prefix); <add> let sum = 0; <add> const path = []; <add> let current = module; <add> while(current.issuer) { <add> path.unshift(current = current.issuer); <add> } <add> path.forEach(module => { <ide> colors.normal("["); <ide> colors.normal(module.id); <ide> colors.normal("] "); <del> colors.bold(module.name); <del> processModuleAttributes(module); <del> newline(); <del> processModuleContent(module, " "); <add> if(module.profile) { <add> const time = (module.profile.factory || 0) + (module.profile.building || 0); <add> coloredTime(time); <add> sum += time; <add> colors.normal(" "); <add> } <add> colors.normal("->"); <ide> }); <del> if(chunk.filteredModules > 0) { <del> colors.normal(" + " + chunk.filteredModules + " hidden modules"); <del> newline(); <del> } <add> Object.keys(module.profile).forEach(key => { <add> colors.normal(` ${key}:`); <add> const time = module.profile[key]; <add> coloredTime(time); <add> sum += time; <add> }); <add> colors.normal(" = "); <add> coloredTime(sum); <add> newline(); <ide> } <del> }); <del> } <del> if(obj.modules) { <del> obj.modules.forEach(function(module) { <del> if(module.id < 1000) colors.normal(" "); <del> if(module.id < 100) colors.normal(" "); <del> if(module.id < 10) colors.normal(" "); <del> colors.normal("["); <del> colors.normal(module.id); <del> colors.normal("] "); <del> colors.bold(module.name || module.identifier); <del> processModuleAttributes(module); <del> newline(); <del> processModuleContent(module, " "); <del> }); <del> if(obj.filteredModules > 0) { <del> colors.normal(" + " + obj.filteredModules + " hidden modules"); <del> newline(); <ide> } <del> } <del> if(obj._showWarnings && obj.warnings) { <del> obj.warnings.forEach(function(warning) { <del> newline(); <del> colors.yellow("WARNING in " + warning); <del> newline(); <del> }); <del> } <del> if(obj._showErrors && obj.errors) { <del> obj.errors.forEach(function(error) { <del> newline(); <del> colors.red("ERROR in " + error); <del> newline(); <del> }); <del> } <del> if(obj.children) { <del> obj.children.forEach(function(child) { <del> if(child.name) { <del> colors.normal("Child "); <del> colors.bold(child.name); <del> colors.normal(":"); <del> } else { <del> colors.normal("Child"); <add> <add> if(obj.chunks) { <add> obj.chunks.forEach(chunk => { <add> colors.normal("chunk "); <add> if(chunk.id < 1000) colors.normal(" "); <add> if(chunk.id < 100) colors.normal(" "); <add> if(chunk.id < 10) colors.normal(" "); <add> colors.normal("{"); <add> colors.yellow(chunk.id); <add> colors.normal("} "); <add> colors.green(chunk.files.join(", ")); <add> if(chunk.names && chunk.names.length > 0) { <add> colors.normal(" ("); <add> colors.normal(chunk.names.join(", ")); <add> colors.normal(")"); <add> } <add> colors.normal(" "); <add> colors.normal(SizeFormatHelpers.formatSize(chunk.size)); <add> chunk.parents.forEach(id => { <add> colors.normal(" {"); <add> colors.yellow(id); <add> colors.normal("}"); <add> }); <add> if(chunk.entry) { <add> colors.yellow(" [entry]"); <add> } else if(chunk.initial) { <add> colors.yellow(" [initial]"); <add> } <add> if(chunk.rendered) { <add> colors.green(" [rendered]"); <add> } <add> if(chunk.recorded) { <add> colors.green(" [recorded]"); <add> } <add> newline(); <add> if(chunk.origins) { <add> chunk.origins.forEach(origin => { <add> colors.normal(" > "); <add> if(origin.reasons && origin.reasons.length) { <add> colors.yellow(origin.reasons.join(" ")); <add> colors.normal(" "); <add> } <add> if(origin.name) { <add> colors.normal(origin.name); <add> colors.normal(" "); <add> } <add> if(origin.module) { <add> colors.normal("["); <add> colors.normal(origin.moduleId); <add> colors.normal("] "); <add> const module = modulesByIdentifier[`$${origin.module}`]; <add> if(module) { <add> colors.bold(module.name); <add> colors.normal(" "); <add> } <add> if(origin.loc) { <add> colors.normal(origin.loc); <add> } <add> } <add> newline(); <add> }); <add> } <add> if(chunk.modules) { <add> chunk.modules.forEach(module => { <add> colors.normal(" "); <add> if(module.id < 1000) colors.normal(" "); <add> if(module.id < 100) colors.normal(" "); <add> if(module.id < 10) colors.normal(" "); <add> colors.normal("["); <add> colors.normal(module.id); <add> colors.normal("] "); <add> colors.bold(module.name); <add> processModuleAttributes(module); <add> newline(); <add> processModuleContent(module, " "); <add> }); <add> if(chunk.filteredModules > 0) { <add> colors.normal(` + ${chunk.filteredModules} hidden modules`); <add> newline(); <add> } <add> } <add> }); <add> } <add> if(obj.modules) { <add> obj.modules.forEach(module => { <add> if(module.id < 1000) colors.normal(" "); <add> if(module.id < 100) colors.normal(" "); <add> if(module.id < 10) colors.normal(" "); <add> colors.normal("["); <add> colors.normal(module.id); <add> colors.normal("] "); <add> colors.bold(module.name || module.identifier); <add> processModuleAttributes(module); <add> newline(); <add> processModuleContent(module, " "); <add> }); <add> if(obj.filteredModules > 0) { <add> colors.normal(` + ${obj.filteredModules} hidden modules`); <add> newline(); <ide> } <del> newline(); <del> buf.push(" "); <del> buf.push(Stats.jsonToString(child, useColors).replace(/\n/g, "\n ")); <del> newline(); <del> }); <del> } <del> if(obj.needAdditionalPass) { <del> colors.yellow("Compilation needs an additional pass and will compile again."); <del> } <add> } <add> if(obj._showWarnings && obj.warnings) { <add> obj.warnings.forEach(warning => { <add> newline(); <add> colors.yellow(`WARNING in ${warning}`); <add> newline(); <add> }); <add> } <add> if(obj._showErrors && obj.errors) { <add> obj.errors.forEach(error => { <add> newline(); <add> colors.red(`ERROR in ${error}`); <add> newline(); <add> }); <add> } <add> if(obj.children) { <add> obj.children.forEach(child => { <add> if(child.name) { <add> colors.normal("Child "); <add> colors.bold(child.name); <add> colors.normal(":"); <add> } else { <add> colors.normal("Child"); <add> } <add> newline(); <add> buf.push(" "); <add> buf.push(Stats.jsonToString(child, useColors).replace(/\n/g, "\n ")); <add> newline(); <add> }); <add> } <add> if(obj.needAdditionalPass) { <add> colors.yellow("Compilation needs an additional pass and will compile again."); <add> } <ide> <del> while(buf[buf.length - 1] === "\n") buf.pop(); <del> return buf.join(""); <del>}; <del> <del>Stats.presetToOptions = function(name) { <del> //Accepted values: none, errors-only, minimal, normal, verbose <del> //Any other falsy value will behave as 'none', truthy values as 'normal' <del> var pn = (typeof name === "string") && name.toLowerCase() || name; <del> if(pn === "none" || !pn) { <del> return { <del> hash: false, <del> version: false, <del> timings: false, <del> assets: false, <del> entrypoints: false, <del> chunks: false, <del> chunkModules: false, <del> modules: false, <del> reasons: false, <del> depth: false, <del> usedExports: false, <del> providedExports: false, <del> children: false, <del> source: false, <del> errors: false, <del> errorDetails: false, <del> warnings: false, <del> publicPath: false <del> }; <del> } else { <del> return { <del> hash: pn !== "errors-only" && pn !== "minimal", <del> version: pn === "verbose", <del> timings: pn !== "errors-only" && pn !== "minimal", <del> assets: pn === "verbose", <del> entrypoints: pn === "verbose", <del> chunks: pn !== "errors-only", <del> chunkModules: pn === "verbose", <del> //warnings: pn !== "errors-only", <del> errorDetails: pn !== "errors-only" && pn !== "minimal", <del> reasons: pn === "verbose", <del> depth: pn === "verbose", <del> usedExports: pn === "verbose", <del> providedExports: pn === "verbose", <del> colors: true <del> }; <add> while(buf[buf.length - 1] === "\n") buf.pop(); <add> return buf.join(""); <add> } <add> <add> static presetToOptions(name) { <add> //Accepted values: none, errors-only, minimal, normal, verbose <add> //Any other falsy value will behave as 'none', truthy values as 'normal' <add> const pn = (typeof name === "string") && name.toLowerCase() || name; <add> if(pn === "none" || !pn) { <add> return { <add> hash: false, <add> version: false, <add> timings: false, <add> assets: false, <add> entrypoints: false, <add> chunks: false, <add> chunkModules: false, <add> modules: false, <add> reasons: false, <add> depth: false, <add> usedExports: false, <add> providedExports: false, <add> children: false, <add> source: false, <add> errors: false, <add> errorDetails: false, <add> warnings: false, <add> publicPath: false <add> }; <add> } else { <add> return { <add> hash: pn !== "errors-only" && pn !== "minimal", <add> version: pn === "verbose", <add> timings: pn !== "errors-only" && pn !== "minimal", <add> assets: pn === "verbose", <add> entrypoints: pn === "verbose", <add> chunks: pn !== "errors-only", <add> chunkModules: pn === "verbose", <add> //warnings: pn !== "errors-only", <add> errorDetails: pn !== "errors-only" && pn !== "minimal", <add> reasons: pn === "verbose", <add> depth: pn === "verbose", <add> usedExports: pn === "verbose", <add> providedExports: pn === "verbose", <add> colors: true <add> }; <add> } <ide> } <del>}; <add>} <add> <add>module.exports = Stats;
1
Mixed
Ruby
dump warns when `mysqldump` is not in path
0a5fdcd5ae2f7928ce9ce492c9815ad71f858a9f
<ide><path>activerecord/CHANGELOG.md <ide> ## Rails 4.0.0 (unreleased) ## <ide> <add>* Warn when `rake db:structure:dump` with a mysl database and <add> `mysqldump` is not in the PATH or fails. <add> Fixes #9518. <add> <add> *Yves Senn* <add> <ide> * Remove `connection#structure_dump`, which is no longer used. *Yves Senn* <ide> <ide> * Make it possible to execute migrations without a transaction even <ide><path>activerecord/lib/active_record/tasks/mysql_database_tasks.rb <ide> def structure_dump(filename) <ide> args.concat(["--result-file", "#{filename}"]) <ide> args.concat(["--no-data"]) <ide> args.concat(["#{configuration['database']}"]) <del> Kernel.system(*args) <add> unless Kernel.system(*args) <add> $stderr.puts "Could not dump the database structure. "\ <add> "Make sure `mysqldump` is in your PATH and check the command output for warnings." <add> end <ide> end <ide> <ide> def structure_load(filename) <ide><path>activerecord/test/cases/tasks/mysql_rake_test.rb <ide> def setup <ide> <ide> def test_structure_dump <ide> filename = "awesome-file.sql" <del> Kernel.expects(:system).with("mysqldump", "--result-file", filename, "--no-data", "test-db") <add> Kernel.expects(:system).with("mysqldump", "--result-file", filename, "--no-data", "test-db").returns(true) <ide> <ide> ActiveRecord::Tasks::DatabaseTasks.structure_dump(@configuration, filename) <ide> end <add> <add> def test_warn_when_external_structure_dump_fails <add> filename = "awesome-file.sql" <add> Kernel.expects(:system).with("mysqldump", "--result-file", filename, "--no-data", "test-db").returns(false) <add> <add> warnings = capture(:stderr) do <add> ActiveRecord::Tasks::DatabaseTasks.structure_dump(@configuration, filename) <add> end <add> <add> assert_match(/Could not dump the database structure/, warnings) <add> end <ide> end <ide> <ide> class MySQLStructureLoadTest < ActiveRecord::TestCase
3
PHP
PHP
add regression test for
264054d9ef36fb6b355417fb56a3cb1147441a6d
<ide><path>tests/TestCase/ORM/EntityTest.php <ide> public function testErrors() { <ide> $this->assertSame($entity, $entity->errors('foo', 'bar')); <ide> $this->assertEquals(['bar'], $entity->errors('foo')); <ide> <add> $this->assertEquals([], $entity->errors('boo')); <add> $entity['boo'] = [ <add> 'someting' => 'stupid', <add> 'and' => false <add> ]; <add> $this->assertEquals([], $entity->errors('boo')); <add> <ide> $entity->errors('foo', 'other error'); <ide> $this->assertEquals(['other error'], $entity->errors('foo')); <ide>
1
Ruby
Ruby
remove unneeded requires
a227359c6289613c870c8bb1481754ab00325c12
<ide><path>actionpack/lib/action_view/digestor.rb <del>require 'active_support/core_ext' <del>require 'logger' <del> <ide> module ActionView <ide> class Digestor <ide> EXPLICIT_DEPENDENCY = /# Template Dependency: ([^ ]+)/ <ide> def nested_dependencies <ide> def logical_name <ide> name.gsub(%r|/_|, "/") <ide> end <del> <add> <ide> def directory <ide> name.split("/").first <ide> end <ide> def source <ide> @source ||= finder.find(logical_name, [], partial?, formats: [ format ]).source <ide> end <ide> <del> <ide> def dependency_digest <ide> dependencies.collect do |template_name| <ide> Digestor.digest(template_name, format, finder, partial: true) <ide> def render_dependencies <ide> <ide> # render("headline") => render("message/headline") <ide> collect { |name| name.include?("/") ? name : "#{directory}/#{name}" }. <del> <add> <ide> # replace quotes from string renders <ide> collect { |name| name.gsub(/["']/, "") } <ide> end <ide> def explicit_dependencies <ide> source.scan(EXPLICIT_DEPENDENCY).flatten.uniq <ide> end <ide> end <del>end <ide>\ No newline at end of file <add>end
1
Ruby
Ruby
add nuclear to github_prerelease_allowlist
a176c3f75ba60ecca3701ceb219de6592dbe66d6
<ide><path>Library/Homebrew/utils/shared_audits.rb <ide> def github_release_data(user, repo, tag) <ide> "gitless" => "0.8.8", <ide> "home-assistant" => :all, <ide> "infrakit" => "0.5", <add> "nuclear" => :all, <ide> "pock" => :all, <ide> "riff" => "0.5.0", <ide> "telegram-cli" => "1.3.1",
1
PHP
PHP
add methods to get/set middleware on a route
9d9e566d8552450202a2d742f0024857d8f6d01b
<ide><path>src/Routing/Route/Route.php <ide> class Route <ide> */ <ide> protected $_extensions = []; <ide> <add> /** <add> * List of middleware that should be applied. <add> * <add> * @var array <add> */ <add> protected $middleware = []; <add> <ide> /** <ide> * Valid HTTP methods. <ide> * <ide> class Route <ide> * ### Options <ide> * <ide> * - `_ext` - Defines the extensions used for this route. <add> * - `_middleware` - Define the middleware names for this route. <ide> * - `pass` - Copies the listed parameters into params['pass']. <ide> * - `_host` - Define the host name pattern if you want this route to only match <ide> * specific host names. You can use `.*` and to create wildcard subdomains/hosts <ide> public function __construct($template, $defaults = [], array $options = []) <ide> unset($defaults['[method]']); <ide> } <ide> $this->defaults = (array)$defaults; <del> $this->options = $options + ['_ext' => []]; <add> $this->options = $options + ['_ext' => [], '_middleware' => []]; <ide> $this->setExtensions((array)$this->options['_ext']); <add> $this->setMiddleware((array)$this->options['_middleware']); <add> unset($this->options['_middleware']); <ide> } <ide> <ide> /** <ide> public function staticPath() <ide> return $this->template; <ide> } <ide> <add> /** <add> * Set the names of the middleware that should be applied to this route. <add> * <add> * @param array $middleware The list of middleware names to apply to this route. <add> * Middleware names will not be checked until the route is matched. <add> * @return $this <add> */ <add> public function setMiddleware(array $middleware) <add> { <add> $this->middleware = $middleware; <add> <add> return $this; <add> } <add> <add> /** <add> * Get the names of the middleware that should be applied to this route. <add> * <add> * @return array <add> */ <add> public function getMiddleware() <add> { <add> return $this->middleware; <add> } <add> <ide> /** <ide> * Set state magic method to support var_export <ide> * <ide><path>tests/TestCase/Routing/Route/RouteTest.php <ide> public function testConstruction() <ide> $this->assertFalse($route->compiled()); <ide> } <ide> <add> /** <add> * Test set middleware in the constructor <add> * <add> * @return void <add> */ <add> public function testConstructorSetMiddleware() <add> { <add> $route = new Route('/:controller/:action/*', [], ['_middleware' => ['auth', 'cookie']]); <add> $this->assertSame(['auth', 'cookie'], $route->getMiddleware()); <add> } <add> <ide> /** <ide> * Test Route compiling. <ide> * <ide> public function testSetPersist() <ide> $this->assertSame($result, $route, 'Should return this'); <ide> $this->assertEquals(['date'], $route->options['persist']); <ide> } <add> <add> /** <add> * Test setting/getting middleware. <add> * <add> * @return void <add> */ <add> public function testSetMiddleware() <add> { <add> $route = new Route('/reviews/:date/:id', ['controller' => 'Reviews', 'action' => 'view']); <add> $result = $route->setMiddleware(['auth', 'cookie']); <add> $this->assertSame($result, $route); <add> $this->assertSame(['auth', 'cookie'], $route->getMiddleware()); <add> } <ide> }
2
Mixed
Javascript
support abortsignal in createsocket
21c8c7e5114c56c8326aa15f11e69d58326553da
<ide><path>doc/api/dgram.md <ide> chained. <ide> <!-- YAML <ide> added: v0.11.13 <ide> changes: <add> - version: REPLACEME <add> pr-url: https://github.com/nodejs/node/pull/37026 <add> description: AbortSignal support was added. <ide> - version: v11.4.0 <ide> pr-url: https://github.com/nodejs/node/pull/23798 <ide> description: The `ipv6Only` option is supported. <ide> changes: <ide> * `recvBufferSize` {number} Sets the `SO_RCVBUF` socket value. <ide> * `sendBufferSize` {number} Sets the `SO_SNDBUF` socket value. <ide> * `lookup` {Function} Custom lookup function. **Default:** [`dns.lookup()`][]. <add> * `signal` {AbortSignal} An AbortSignal that may be used to close a socket. <ide> * `callback` {Function} Attached as a listener for `'message'` events. Optional. <ide> * Returns: {dgram.Socket} <ide> <ide> method will bind the socket to the "all interfaces" address on a random port <ide> and port can be retrieved using [`socket.address().address`][] and <ide> [`socket.address().port`][]. <ide> <add>If the `signal` option is enabled, calling `.abort()` on the corresponding <add>`AbortController` is similar to calling `.close()` on the socket: <add> <add>```js <add>const controller = new AbortController(); <add>const { signal } = controller; <add>const server = dgram.createSocket({ type: 'udp4', signal }); <add>server.on('message', (msg, rinfo) => { <add> console.log(`server got: ${msg} from ${rinfo.address}:${rinfo.port}`); <add>}); <add>// Later, when you want to close the server. <add>controller.abort(); <add>``` <add> <ide> ### `dgram.createSocket(type[, callback])` <ide> <!-- YAML <ide> added: v0.1.99 <ide><path>lib/dgram.js <ide> const { <ide> } = errors.codes; <ide> const { <ide> isInt32, <add> validateAbortSignal, <ide> validateString, <ide> validateNumber, <ide> validatePort, <ide> function Socket(type, listener) { <ide> recvBufferSize, <ide> sendBufferSize <ide> }; <add> <add> if (options?.signal !== undefined) { <add> const { signal } = options; <add> validateAbortSignal(signal, 'options.signal'); <add> const onAborted = () => { <add> this.close(); <add> }; <add> if (signal.aborted) { <add> onAborted(); <add> } else { <add> signal.addEventListener('abort', onAborted); <add> this.once('close', () => signal.removeEventListener('abort', onAborted)); <add> } <add> } <ide> } <ide> ObjectSetPrototypeOf(Socket.prototype, EventEmitter.prototype); <ide> ObjectSetPrototypeOf(Socket, EventEmitter); <ide><path>test/parallel/test-dgram-close-signal.js <add>'use strict'; <add> <add>const common = require('../common'); <add>const assert = require('assert'); <add>const dgram = require('dgram'); <add> <add>{ <add> // Test bad signal. <add> assert.throws( <add> () => dgram.createSocket({ type: 'udp4', signal: {} }), <add> { <add> code: 'ERR_INVALID_ARG_TYPE', <add> name: 'TypeError' <add> }); <add>} <add> <add>{ <add> // Test close. <add> const controller = new AbortController(); <add> const { signal } = controller; <add> const server = dgram.createSocket({ type: 'udp4', signal }); <add> server.on('close', common.mustCall()); <add> controller.abort(); <add>} <add> <add>{ <add> // Test close with pre-aborted signal. <add> const controller = new AbortController(); <add> controller.abort(); <add> const { signal } = controller; <add> const server = dgram.createSocket({ type: 'udp4', signal }); <add> server.on('close', common.mustCall()); <add>}
3
Python
Python
set version to v3.4.1
5c2a00cef04b8c6e93e81cd1ca1d752f320c6e5d
<ide><path>spacy/about.py <ide> # fmt: off <ide> __title__ = "spacy" <del>__version__ = "3.4.0" <add>__version__ = "3.4.1" <ide> __download_url__ = "https://github.com/explosion/spacy-models/releases/download" <ide> __compatibility__ = "https://raw.githubusercontent.com/explosion/spacy-models/master/compatibility.json" <ide> __projects__ = "https://github.com/explosion/projects"
1
Javascript
Javascript
add highlight decorations, but no tests yet
a4224922a3933c93a2827bc914bfb7ad7a95ba3e
<ide><path>src/text-editor-component.js <ide> class TextEditorComponent { <ide> this.cursorsToRender = [] <ide> this.decorationsToRender = { <ide> lineNumbers: new Map(), <del> lines: new Map() <add> lines: new Map(), <add> highlights: new Map() <add> } <add> this.decorationsToMeasure = { <add> highlights: new Map() <ide> } <ide> <ide> if (this.props.model) this.observeModel() <ide> class TextEditorComponent { <ide> if (longestLineToMeasure) this.measureLongestLineWidth(longestLineToMeasure) <ide> if (this.pendingAutoscroll) this.finalizeAutoscroll() <ide> this.positionCursorsToRender() <add> this.updateHighlightsToRender() <ide> <ide> etch.updateSync(this) <ide> <ide> class TextEditorComponent { <ide> for (let row = tileStartRow; row < tileEndRow; row++) { <ide> lineDecorations[row - tileStartRow] = this.decorationsToRender.lines.get(row) <ide> } <add> const highlightDecorations = this.decorationsToRender.highlights.get(tileStartRow) <ide> <ide> tileNodes[tileIndex] = $(LinesTileComponent, { <ide> key: tileIndex, <ide> height: tileHeight, <ide> width: tileWidth, <ide> top: this.topPixelPositionForRow(tileStartRow), <add> lineHeight: this.measurements.lineHeight, <ide> screenLines: screenLines.slice(tileStartRow - startRow, tileEndRow - startRow), <ide> lineDecorations, <add> highlightDecorations, <ide> displayLayer, <ide> lineNodesByScreenLineId, <ide> textNodesByScreenLineId <ide> class TextEditorComponent { <ide> }) <ide> const lastCursorMarker = model.getLastCursor().getMarker() <ide> <del> this.cursorsToRender.length = cursorMarkers.length <add> this.cursorsToRender.length = 0 <ide> this.lastCursorIndex = -1 <ide> for (let i = 0; i < cursorMarkers.length; i++) { <ide> const cursorMarker = cursorMarkers[i] <ide> if (cursorMarker === lastCursorMarker) this.lastCursorIndex = i <ide> const screenPosition = cursorMarker.getHeadScreenPosition() <ide> const {row, column} = screenPosition <add> <add> if (row < this.getRenderedStartRow() || row >= this.getRenderedEndRow()) { <add> continue <add> } <add> <ide> this.requestHorizontalMeasurement(row, column) <ide> let columnWidth = 0 <ide> if (model.lineLengthForScreenRow(row) > column) { <ide> columnWidth = 1 <ide> this.requestHorizontalMeasurement(row, column + 1) <ide> } <del> this.cursorsToRender[i] = { <add> this.cursorsToRender.push({ <ide> screenPosition, columnWidth, <ide> pixelTop: 0, pixelLeft: 0, pixelWidth: 0 <del> } <add> }) <ide> } <ide> } <ide> <ide> queryDecorationsToRender () { <ide> this.decorationsToRender.lineNumbers.clear() <ide> this.decorationsToRender.lines.clear() <add> this.decorationsToMeasure.highlights.clear() <ide> <ide> const decorationsByMarker = <ide> this.getModel().decorationManager.decorationPropertiesByMarkerForScreenRowRange( <ide> class TextEditorComponent { <ide> switch (type) { <ide> case 'line': <ide> case 'line-number': <del> const decorationsByRow = (type === 'line') ? this.decorationsToRender.lines : this.decorationsToRender.lineNumbers <add> this.addLineDecorationToRender(type, decoration, screenRange, reversed) <add> break <add> case 'highlight': <add> this.addHighlightDecorationToMeasure(decoration, screenRange) <add> break <add> } <add> } <add> } <ide> <del> let omitLastRow = false <del> if (screenRange.isEmpty()) { <del> if (decoration.onlyNonEmpty) return <del> } else { <del> if (decoration.onlyEmpty) return <del> if (decoration.omitEmptyLastRow !== false) { <del> omitLastRow = screenRange.end.column === 0 <del> } <del> } <add> addLineDecorationToRender (type, decoration, screenRange, reversed) { <add> const decorationsByRow = (type === 'line') ? this.decorationsToRender.lines : this.decorationsToRender.lineNumbers <add> <add> let omitLastRow = false <add> if (screenRange.isEmpty()) { <add> if (decoration.onlyNonEmpty) return <add> } else { <add> if (decoration.onlyEmpty) return <add> if (decoration.omitEmptyLastRow !== false) { <add> omitLastRow = screenRange.end.column === 0 <add> } <add> } <ide> <del> let startRow = screenRange.start.row <del> let endRow = screenRange.end.row <add> let startRow = screenRange.start.row <add> let endRow = screenRange.end.row <ide> <del> if (decoration.onlyHead) { <del> if (reversed) { <del> endRow = startRow <del> } else { <del> startRow = endRow <del> } <del> } <add> if (decoration.onlyHead) { <add> if (reversed) { <add> endRow = startRow <add> } else { <add> startRow = endRow <add> } <add> } <ide> <del> for (let row = startRow; row <= endRow; row++) { <del> if (omitLastRow && row === endRow) break <del> const currentClassName = decorationsByRow.get(row) <del> const newClassName = currentClassName ? currentClassName + ' ' + decoration.class : decoration.class <del> decorationsByRow.set(row, newClassName) <del> } <del> break <add> for (let row = startRow; row <= endRow; row++) { <add> if (omitLastRow && row === endRow) break <add> const currentClassName = decorationsByRow.get(row) <add> const newClassName = currentClassName ? currentClassName + ' ' + decoration.class : decoration.class <add> decorationsByRow.set(row, newClassName) <add> } <add> } <add> <add> addHighlightDecorationToMeasure(decoration, screenRange) { <add> screenRange = constrainRangeToRows(screenRange, this.getRenderedStartRow(), this.getRenderedEndRow()) <add> if (screenRange.isEmpty()) return <add> let tileStartRow = this.tileStartRowForRow(screenRange.start.row) <add> const rowsPerTile = this.getRowsPerTile() <add> <add> while (tileStartRow <= screenRange.end.row) { <add> const tileEndRow = tileStartRow + rowsPerTile <add> const screenRangeInTile = constrainRangeToRows(screenRange, tileStartRow, tileEndRow) <add> <add> let tileHighlights = this.decorationsToMeasure.highlights.get(tileStartRow) <add> if (!tileHighlights) { <add> tileHighlights = [] <add> this.decorationsToMeasure.highlights.set(tileStartRow, tileHighlights) <ide> } <add> tileHighlights.push({decoration, screenRange: screenRangeInTile}) <add> <add> this.requestHorizontalMeasurement(screenRangeInTile.start.row, screenRangeInTile.start.column) <add> this.requestHorizontalMeasurement(screenRangeInTile.end.row, screenRangeInTile.end.column) <add> <add> tileStartRow += rowsPerTile <ide> } <ide> } <ide> <add> updateHighlightsToRender () { <add> this.decorationsToRender.highlights.clear() <add> this.decorationsToMeasure.highlights.forEach((highlights, tileRow) => { <add> for (let i = 0, length = highlights.length; i < length; i++) { <add> const highlight = highlights[i] <add> const {start, end} = highlight.screenRange <add> highlight.startPixelTop = this.pixelTopForScreenRow(start.row) <add> highlight.startPixelLeft = this.pixelLeftForScreenRowAndColumn(start.row, start.column) <add> highlight.endPixelTop = this.pixelTopForScreenRow(end.row + 1) <add> highlight.endPixelLeft = this.pixelLeftForScreenRowAndColumn(end.row, end.column) <add> } <add> this.decorationsToRender.highlights.set(tileRow, highlights) <add> }) <add> } <ide> <ide> positionCursorsToRender () { <ide> const height = this.measurements.lineHeight + 'px' <ide> class TextEditorComponent { <ide> } <ide> <ide> requestHorizontalMeasurement (row, column) { <add> if (column === 0) return <ide> let columns = this.horizontalPositionsToMeasure.get(row) <ide> if (columns == null) { <ide> columns = [] <ide> class TextEditorComponent { <ide> <ide> const screenLine = this.getModel().displayLayer.getScreenLine(row) <ide> const lineNode = this.lineNodesByScreenLineId.get(screenLine.id) <add> <add> if (!lineNode) { <add> const error = new Error('Requested measurement of a line that is not currently rendered') <add> error.metadata = {row, columnsToMeasure} <add> throw error <add> } <add> <ide> const textNodes = this.textNodesByScreenLineId.get(screenLine.id) <ide> let positionsForLine = this.horizontalPixelPositionsByScreenLineId.get(screenLine.id) <ide> if (positionsForLine == null) { <ide> class LinesTileComponent { <ide> } <ide> <ide> render () { <add> const {height, width, top} = this.props <add> <add> return $.div( <add> { <add> style: { <add> contain: 'strict', <add> position: 'absolute', <add> height: height + 'px', <add> width: width + 'px', <add> willChange: 'transform', <add> transform: `translateY(${top}px)`, <add> backgroundColor: 'inherit' <add> } <add> }, <add> this.renderHighlights(), <add> this.renderLines() <add> ) <add> <add> } <add> <add> renderHighlights () { <add> const {top, height, width, lineHeight, highlightDecorations} = this.props <add> <add> let children = null <add> if (highlightDecorations) { <add> const decorationCount = highlightDecorations.length <add> children = new Array(decorationCount) <add> for (let i = 0; i < decorationCount; i++) { <add> const highlightProps = Object.assign( <add> {parentTileTop: top, lineHeight}, <add> highlightDecorations[i] <add> ) <add> children[i] = $(HighlightComponent, highlightProps) <add> } <add> } <add> <add> return $.div( <add> { <add> style: { <add> position: 'absolute', <add> contain: 'strict', <add> height: height + 'px', <add> width: width + 'px' <add> }, <add> }, children <add> ) <add> } <add> <add> renderLines () { <ide> const { <ide> height, width, top, <ide> screenLines, lineDecorations, displayLayer, <ide> class LinesTileComponent { <ide> <ide> return $.div({ <ide> style: { <del> contain: 'strict', <ide> position: 'absolute', <add> contain: 'strict', <ide> height: height + 'px', <ide> width: width + 'px', <del> willChange: 'transform', <del> transform: `translateY(${top}px)`, <del> backgroundColor: 'inherit' <ide> } <ide> }, children) <ide> } <ide> class LinesTileComponent { <ide> if (oldProps.width !== newProps.width) return true <ide> if (!arraysEqual(oldProps.screenLines, newProps.screenLines)) return true <ide> if (!arraysEqual(oldProps.lineDecorations, newProps.lineDecorations)) return true <add> <add> if (!oldProps.highlightDecorations && newProps.highlightDecorations) return true <add> if (oldProps.highlightDecorations && !newProps.highlightDecorations) return true <add> <add> if (oldProps.highlightDecorations && newProps.highlightDecorations) { <add> if (oldProps.highlightDecorations.length !== newProps.highlightDecorations.length) return true <add> <add> for (let i = 0, length = oldProps.highlightDecorations.length; i < length; i++) { <add> const oldHighlight = oldProps.highlightDecorations[i] <add> const newHighlight = newProps.highlightDecorations[i] <add> if (oldHighlight.decoration.class !== newHighlight.decoration.class) return true <add> if (oldHighlight.startPixelLeft !== newHighlight.startPixelLeft) return true <add> if (oldHighlight.endPixelLeft !== newHighlight.endPixelLeft) return true <add> if (!oldHighlight.screenRange.isEqual(newHighlight.screenRange)) return true <add> } <add> } <add> <ide> return false <ide> } <ide> } <ide> class LineComponent { <ide> } <ide> } <ide> <add>class HighlightComponent { <add> constructor (props) { <add> this.props = props <add> etch.initialize(this) <add> } <add> <add> update (props) { <add> this.props = props <add> etch.updateSync(this) <add> } <add> <add> render () { <add> let {startPixelTop, endPixelTop} = this.props <add> const { <add> decoration, screenRange, parentTileTop, lineHeight, <add> startPixelLeft, endPixelLeft, <add> } = this.props <add> startPixelTop -= parentTileTop <add> endPixelTop -= parentTileTop <add> <add> let children <add> if (screenRange.start.row === screenRange.end.row) { <add> children = $.div({ <add> className: 'region', <add> style: { <add> position: 'absolute', <add> boxSizing: 'border-box', <add> top: startPixelTop + 'px', <add> left: startPixelLeft + 'px', <add> width: endPixelLeft - startPixelLeft + 'px', <add> height: lineHeight + 'px' <add> } <add> }) <add> } else { <add> children = [] <add> children.push($.div({ <add> className: 'region', <add> style: { <add> position: 'absolute', <add> boxSizing: 'border-box', <add> top: startPixelTop + 'px', <add> left: startPixelLeft + 'px', <add> right: 0, <add> height: lineHeight + 'px' <add> } <add> })) <add> <add> if (screenRange.end.row - screenRange.start.row > 1) { <add> children.push($.div({ <add> className: 'region', <add> style: { <add> position: 'absolute', <add> boxSizing: 'border-box', <add> top: startPixelTop + lineHeight + 'px', <add> left: 0, <add> right: 0, <add> height: endPixelTop - startPixelTop - (lineHeight * 2) + 'px' <add> } <add> })) <add> } <add> <add> children.push($.div({ <add> className: 'region', <add> style: { <add> position: 'absolute', <add> boxSizing: 'border-box', <add> top: endPixelTop - lineHeight + 'px', <add> left: 0, <add> width: endPixelLeft + 'px', <add> height: lineHeight + 'px' <add> } <add> })) <add> } <add> <add> const className = 'highlight ' + decoration.class <add> return $.div({className}, children) <add> } <add>} <add> <ide> const classNamesByScopeName = new Map() <ide> function classNameForScopeName (scopeName) { <ide> let classString = classNamesByScopeName.get(scopeName) <ide> function arraysEqual(a, b) { <ide> } <ide> return true <ide> } <add> <add>function constrainRangeToRows (range, startRow, endRow) { <add> if (range.start.row < startRow || range.end.row >= endRow) { <add> range = range.copy() <add> if (range.start.row < startRow) { <add> range.start.row = startRow <add> range.start.column = 0 <add> } <add> if (range.end.row >= endRow) { <add> range.end.row = endRow <add> range.end.column = 0 <add> } <add> } <add> return range <add>}
1
Javascript
Javascript
move reserved names to top of module
d5479d8fa7c52c79912aba45e3c86e2db32738dd
<ide><path>lib/optimize/ConcatenatedModule.js <ide> const createHash = require("../util/createHash"); <ide> /** @typedef {import("../WebpackError")} WebpackError */ <ide> /** @typedef {import("../util/createHash").Hash} Hash */ <ide> <add>const RESERVED_NAMES = [ <add> // internal name <add> "__WEBPACK_MODULE_DEFAULT_EXPORT__", <add> <add> // keywords <add> "abstract,arguments,async,await,boolean,break,byte,case,catch,char,class,const,continue", <add> "debugger,default,delete,do,double,else,enum,eval,export,extends,false,final,finally,float", <add> "for,function,goto,if,implements,import,in,instanceof,int,interface,let,long,native,new,null", <add> "package,private,protected,public,return,short,static,super,switch,synchronized,this,throw", <add> "throws,transient,true,try,typeof,var,void,volatile,while,with,yield", <add> <add> // commonjs <add> "module,__dirname,__filename,exports", <add> <add> // js globals <add> "Array,Date,eval,function,hasOwnProperty,Infinity,isFinite,isNaN,isPrototypeOf,length,Math", <add> "NaN,name,Number,Object,prototype,String,toString,undefined,valueOf", <add> <add> // browser globals <add> "alert,all,anchor,anchors,area,assign,blur,button,checkbox,clearInterval,clearTimeout", <add> "clientInformation,close,closed,confirm,constructor,crypto,decodeURI,decodeURIComponent", <add> "defaultStatus,document,element,elements,embed,embeds,encodeURI,encodeURIComponent,escape", <add> "event,fileUpload,focus,form,forms,frame,innerHeight,innerWidth,layer,layers,link,location", <add> "mimeTypes,navigate,navigator,frames,frameRate,hidden,history,image,images,offscreenBuffering", <add> "open,opener,option,outerHeight,outerWidth,packages,pageXOffset,pageYOffset,parent,parseFloat", <add> "parseInt,password,pkcs11,plugin,prompt,propertyIsEnum,radio,reset,screenX,screenY,scroll", <add> "secure,select,self,setInterval,setTimeout,status,submit,taint,text,textarea,top,unescape", <add> "untaint,window", <add> <add> // window events <add> "onblur,onclick,onerror,onfocus,onkeydown,onkeypress,onkeyup,onmouseover,onload,onmouseup,onmousedown,onsubmit" <add>] <add> .join(",") <add> .split(","); <add> <ide> /** <ide> * @typedef {Object} ConcatenationEntry <ide> * @property {"concatenated" | "external"} type <ide> class ConcatenatedModule extends Module { <ide> } <ide> <ide> // List of all used names to avoid conflicts <del> const allUsedNames = new Set([ <del> "__WEBPACK_MODULE_DEFAULT_EXPORT__", // avoid using this internal name <del> <del> "abstract", <del> "arguments", <del> "async", <del> "await", <del> "boolean", <del> "break", <del> "byte", <del> "case", <del> "catch", <del> "char", <del> "class", <del> "const", <del> "continue", <del> "debugger", <del> "default", <del> "delete", <del> "do", <del> "double", <del> "else", <del> "enum", <del> "eval", <del> "export", <del> "extends", <del> "false", <del> "final", <del> "finally", <del> "float", <del> "for", <del> "function", <del> "goto", <del> "if", <del> "implements", <del> "import", <del> "in", <del> "instanceof", <del> "int", <del> "interface", <del> "let", <del> "long", <del> "native", <del> "new", <del> "null", <del> "package", <del> "private", <del> "protected", <del> "public", <del> "return", <del> "short", <del> "static", <del> "super", <del> "switch", <del> "synchronized", <del> "this", <del> "throw", <del> "throws", <del> "transient", <del> "true", <del> "try", <del> "typeof", <del> "var", <del> "void", <del> "volatile", <del> "while", <del> "with", <del> "yield", <del> <del> "module", <del> "__dirname", <del> "__filename", <del> "exports", <del> <del> "Array", <del> "Date", <del> "eval", <del> "function", <del> "hasOwnProperty", <del> "Infinity", <del> "isFinite", <del> "isNaN", <del> "isPrototypeOf", <del> "length", <del> "Math", <del> "NaN", <del> "name", <del> "Number", <del> "Object", <del> "prototype", <del> "String", <del> "toString", <del> "undefined", <del> "valueOf", <del> <del> "alert", <del> "all", <del> "anchor", <del> "anchors", <del> "area", <del> "assign", <del> "blur", <del> "button", <del> "checkbox", <del> "clearInterval", <del> "clearTimeout", <del> "clientInformation", <del> "close", <del> "closed", <del> "confirm", <del> "constructor", <del> "crypto", <del> "decodeURI", <del> "decodeURIComponent", <del> "defaultStatus", <del> "document", <del> "element", <del> "elements", <del> "embed", <del> "embeds", <del> "encodeURI", <del> "encodeURIComponent", <del> "escape", <del> "event", <del> "fileUpload", <del> "focus", <del> "form", <del> "forms", <del> "frame", <del> "innerHeight", <del> "innerWidth", <del> "layer", <del> "layers", <del> "link", <del> "location", <del> "mimeTypes", <del> "navigate", <del> "navigator", <del> "frames", <del> "frameRate", <del> "hidden", <del> "history", <del> "image", <del> "images", <del> "offscreenBuffering", <del> "open", <del> "opener", <del> "option", <del> "outerHeight", <del> "outerWidth", <del> "packages", <del> "pageXOffset", <del> "pageYOffset", <del> "parent", <del> "parseFloat", <del> "parseInt", <del> "password", <del> "pkcs11", <del> "plugin", <del> "prompt", <del> "propertyIsEnum", <del> "radio", <del> "reset", <del> "screenX", <del> "screenY", <del> "scroll", <del> "secure", <del> "select", <del> "self", <del> "setInterval", <del> "setTimeout", <del> "status", <del> "submit", <del> "taint", <del> "text", <del> "textarea", <del> "top", <del> "unescape", <del> "untaint", <del> "window", <del> <del> "onblur", <del> "onclick", <del> "onerror", <del> "onfocus", <del> "onkeydown", <del> "onkeypress", <del> "onkeyup", <del> "onmouseover", <del> "onload", <del> "onmouseup", <del> "onmousedown", <del> "onsubmit" <del> ]); <add> const allUsedNames = new Set(RESERVED_NAMES); <ide> <ide> // Set of already checked scopes <ide> const alreadyCheckedScopes = new Set();
1
Java
Java
fix javadoc in dependencydescriptor
acae174f8f0a9013b10b10a65e63c0caf2a28337
<ide><path>spring-beans/src/main/java/org/springframework/beans/factory/config/DependencyDescriptor.java <ide> public void initParameterNameDiscovery(@Nullable ParameterNameDiscoverer paramet <ide> <ide> /** <ide> * Determine the name of the wrapped parameter/field. <del> * @return the declared name (never {@code null}) <add> * @return the declared name (may be {@code null}) <ide> */ <ide> @Nullable <ide> public String getDependencyName() {
1
Javascript
Javascript
move hydration code out of normal suspense path
b2763d3eaa4a56ce9c973945783aba7cac63478f
<ide><path>packages/react-reconciler/src/ReactFiberBeginWork.new.js <ide> function updateSuspenseComponent(current, workInProgress, renderLanes) { <ide> // a stack. <ide> if (current === null) { <ide> // Initial mount <add> <add> // Special path for hydration <ide> // If we're currently hydrating, try to hydrate this boundary. <ide> tryToClaimNextHydratableInstance(workInProgress); <ide> // This could've been a dehydrated suspense component. <ide> function updateSuspenseComponent(current, workInProgress, renderLanes) { <ide> } else { <ide> // This is an update. <ide> <del> // If the current fiber has a SuspenseState, that means it's already showing <del> // a fallback. <add> // Special path for hydration <ide> const prevState: null | SuspenseState = current.memoizedState; <ide> if (prevState !== null) { <del> // The current tree is already showing a fallback <del> <del> // Special path for hydration <ide> const dehydrated = prevState.dehydrated; <ide> if (dehydrated !== null) { <del> if (!didSuspend) { <del> return updateDehydratedSuspenseComponent( <del> current, <del> workInProgress, <del> dehydrated, <del> prevState, <del> renderLanes, <del> ); <del> } else if (workInProgress.flags & ForceClientRender) { <del> // Something errored during hydration. Try again without hydrating. <del> workInProgress.flags &= ~ForceClientRender; <del> return retrySuspenseComponentWithoutHydrating( <del> current, <del> workInProgress, <del> renderLanes, <del> new Error( <del> 'There was an error while hydrating this Suspense boundary. ' + <del> 'Switched to client rendering.', <del> ), <del> ); <del> } else if ( <del> (workInProgress.memoizedState: null | SuspenseState) !== null <del> ) { <del> // Something suspended and we should still be in dehydrated mode. <del> // Leave the existing child in place. <del> workInProgress.child = current.child; <del> // The dehydrated completion pass expects this flag to be there <del> // but the normal suspense pass doesn't. <del> workInProgress.flags |= DidCapture; <del> return null; <del> } else { <del> // Suspended but we should no longer be in dehydrated mode. <del> // Therefore we now have to render the fallback. <del> const nextPrimaryChildren = nextProps.children; <del> const nextFallbackChildren = nextProps.fallback; <del> const fallbackChildFragment = mountSuspenseFallbackAfterRetryWithoutHydrating( <del> current, <del> workInProgress, <del> nextPrimaryChildren, <del> nextFallbackChildren, <del> renderLanes, <del> ); <del> const primaryChildFragment: Fiber = (workInProgress.child: any); <del> primaryChildFragment.memoizedState = mountSuspenseOffscreenState( <del> renderLanes, <del> ); <del> workInProgress.memoizedState = SUSPENDED_MARKER; <del> return fallbackChildFragment; <del> } <del> } <del> <del> if (showFallback) { <del> const nextFallbackChildren = nextProps.fallback; <del> const nextPrimaryChildren = nextProps.children; <del> const fallbackChildFragment = updateSuspenseFallbackChildren( <add> return updateDehydratedSuspenseComponent( <ide> current, <ide> workInProgress, <del> nextPrimaryChildren, <del> nextFallbackChildren, <del> renderLanes, <del> ); <del> const primaryChildFragment: Fiber = (workInProgress.child: any); <del> const prevOffscreenState: OffscreenState | null = (current.child: any) <del> .memoizedState; <del> primaryChildFragment.memoizedState = <del> prevOffscreenState === null <del> ? mountSuspenseOffscreenState(renderLanes) <del> : updateSuspenseOffscreenState(prevOffscreenState, renderLanes); <del> if (enableTransitionTracing) { <del> const currentTransitions = getSuspendedTransitions(); <del> if (currentTransitions !== null) { <del> const primaryChildUpdateQueue: OffscreenQueue = { <del> transitions: currentTransitions, <del> }; <del> primaryChildFragment.updateQueue = primaryChildUpdateQueue; <del> } <del> } <del> primaryChildFragment.childLanes = getRemainingWorkInPrimaryTree( <del> current, <del> renderLanes, <del> ); <del> workInProgress.memoizedState = SUSPENDED_MARKER; <del> return fallbackChildFragment; <del> } else { <del> const nextPrimaryChildren = nextProps.children; <del> const primaryChildFragment = updateSuspensePrimaryChildren( <del> current, <del> workInProgress, <del> nextPrimaryChildren, <add> didSuspend, <add> nextProps, <add> dehydrated, <add> prevState, <ide> renderLanes, <ide> ); <del> workInProgress.memoizedState = null; <del> return primaryChildFragment; <ide> } <del> } else { <del> // The current tree is not already showing a fallback. <del> if (showFallback) { <del> // Timed out. <del> const nextFallbackChildren = nextProps.fallback; <del> const nextPrimaryChildren = nextProps.children; <del> const fallbackChildFragment = updateSuspenseFallbackChildren( <del> current, <del> workInProgress, <del> nextPrimaryChildren, <del> nextFallbackChildren, <del> renderLanes, <del> ); <del> const primaryChildFragment: Fiber = (workInProgress.child: any); <del> const prevOffscreenState: OffscreenState | null = (current.child: any) <del> .memoizedState; <del> primaryChildFragment.memoizedState = <del> prevOffscreenState === null <del> ? mountSuspenseOffscreenState(renderLanes) <del> : updateSuspenseOffscreenState(prevOffscreenState, renderLanes); <del> primaryChildFragment.childLanes = getRemainingWorkInPrimaryTree( <del> current, <del> renderLanes, <del> ); <add> } <ide> <del> if (enableTransitionTracing) { <del> const currentTransitions = getSuspendedTransitions(); <del> if (currentTransitions !== null) { <del> const primaryChildUpdateQueue: OffscreenQueue = { <del> transitions: currentTransitions, <del> }; <del> primaryChildFragment.updateQueue = primaryChildUpdateQueue; <del> } <add> if (showFallback) { <add> const nextFallbackChildren = nextProps.fallback; <add> const nextPrimaryChildren = nextProps.children; <add> const fallbackChildFragment = updateSuspenseFallbackChildren( <add> current, <add> workInProgress, <add> nextPrimaryChildren, <add> nextFallbackChildren, <add> renderLanes, <add> ); <add> const primaryChildFragment: Fiber = (workInProgress.child: any); <add> const prevOffscreenState: OffscreenState | null = (current.child: any) <add> .memoizedState; <add> primaryChildFragment.memoizedState = <add> prevOffscreenState === null <add> ? mountSuspenseOffscreenState(renderLanes) <add> : updateSuspenseOffscreenState(prevOffscreenState, renderLanes); <add> if (enableTransitionTracing) { <add> const currentTransitions = getSuspendedTransitions(); <add> if (currentTransitions !== null) { <add> const primaryChildUpdateQueue: OffscreenQueue = { <add> transitions: currentTransitions, <add> }; <add> primaryChildFragment.updateQueue = primaryChildUpdateQueue; <ide> } <del> <del> // Skip the primary children, and continue working on the <del> // fallback children. <del> workInProgress.memoizedState = SUSPENDED_MARKER; <del> return fallbackChildFragment; <del> } else { <del> // Still haven't timed out. Continue rendering the children, like we <del> // normally do. <del> const nextPrimaryChildren = nextProps.children; <del> const primaryChildFragment = updateSuspensePrimaryChildren( <del> current, <del> workInProgress, <del> nextPrimaryChildren, <del> renderLanes, <del> ); <del> workInProgress.memoizedState = null; <del> return primaryChildFragment; <ide> } <add> primaryChildFragment.childLanes = getRemainingWorkInPrimaryTree( <add> current, <add> renderLanes, <add> ); <add> workInProgress.memoizedState = SUSPENDED_MARKER; <add> return fallbackChildFragment; <add> } else { <add> const nextPrimaryChildren = nextProps.children; <add> const primaryChildFragment = updateSuspensePrimaryChildren( <add> current, <add> workInProgress, <add> nextPrimaryChildren, <add> renderLanes, <add> ); <add> workInProgress.memoizedState = null; <add> return primaryChildFragment; <ide> } <ide> } <ide> } <ide> function mountDehydratedSuspenseComponent( <ide> function updateDehydratedSuspenseComponent( <ide> current: Fiber, <ide> workInProgress: Fiber, <add> didSuspend: boolean, <add> nextProps: any, <ide> suspenseInstance: SuspenseInstance, <ide> suspenseState: SuspenseState, <ide> renderLanes: Lanes, <ide> ): null | Fiber { <del> // We should never be hydrating at this point because it is the first pass, <del> // but after we've already committed once. <del> warnIfHydrating(); <add> if (!didSuspend) { <add> // This is the first render pass. Attempt to hydrate. <ide> <del> if ((workInProgress.mode & ConcurrentMode) === NoMode) { <del> return retrySuspenseComponentWithoutHydrating( <del> current, <del> workInProgress, <del> renderLanes, <del> // TODO: When we delete legacy mode, we should make this error argument <del> // required — every concurrent mode path that causes hydration to <del> // de-opt to client rendering should have an error message. <del> null, <del> ); <del> } <add> // We should never be hydrating at this point because it is the first pass, <add> // but after we've already committed once. <add> warnIfHydrating(); <ide> <del> if (isSuspenseInstanceFallback(suspenseInstance)) { <del> // This boundary is in a permanent fallback state. In this case, we'll never <del> // get an update and we'll never be able to hydrate the final content. Let's just try the <del> // client side render instead. <del> return retrySuspenseComponentWithoutHydrating( <del> current, <del> workInProgress, <del> renderLanes, <del> // TODO: The server should serialize the error message so we can log it <del> // here on the client. Or, in production, a hash/id that corresponds to <del> // the error. <del> new Error( <del> 'The server could not finish this Suspense boundary, likely ' + <del> 'due to an error during server rendering. Switched to ' + <del> 'client rendering.', <del> ), <del> ); <del> } <add> if ((workInProgress.mode & ConcurrentMode) === NoMode) { <add> return retrySuspenseComponentWithoutHydrating( <add> current, <add> workInProgress, <add> renderLanes, <add> // TODO: When we delete legacy mode, we should make this error argument <add> // required — every concurrent mode path that causes hydration to <add> // de-opt to client rendering should have an error message. <add> null, <add> ); <add> } <ide> <del> if ( <del> enableLazyContextPropagation && <del> // TODO: Factoring is a little weird, since we check this right below, too. <del> // But don't want to re-arrange the if-else chain until/unless this <del> // feature lands. <del> !didReceiveUpdate <del> ) { <del> // We need to check if any children have context before we decide to bail <del> // out, so propagate the changes now. <del> lazilyPropagateParentContextChanges(current, workInProgress, renderLanes); <del> } <del> <del> // We use lanes to indicate that a child might depend on context, so if <del> // any context has changed, we need to treat is as if the input might have changed. <del> const hasContextChanged = includesSomeLane(renderLanes, current.childLanes); <del> if (didReceiveUpdate || hasContextChanged) { <del> // This boundary has changed since the first render. This means that we are now unable to <del> // hydrate it. We might still be able to hydrate it using a higher priority lane. <del> const root = getWorkInProgressRoot(); <del> if (root !== null) { <del> const attemptHydrationAtLane = getBumpedLaneForHydration( <del> root, <add> if (isSuspenseInstanceFallback(suspenseInstance)) { <add> // This boundary is in a permanent fallback state. In this case, we'll never <add> // get an update and we'll never be able to hydrate the final content. Let's just try the <add> // client side render instead. <add> return retrySuspenseComponentWithoutHydrating( <add> current, <add> workInProgress, <ide> renderLanes, <add> // TODO: The server should serialize the error message so we can log it <add> // here on the client. Or, in production, a hash/id that corresponds to <add> // the error. <add> new Error( <add> 'The server could not finish this Suspense boundary, likely ' + <add> 'due to an error during server rendering. Switched to ' + <add> 'client rendering.', <add> ), <ide> ); <del> if ( <del> attemptHydrationAtLane !== NoLane && <del> attemptHydrationAtLane !== suspenseState.retryLane <del> ) { <del> // Intentionally mutating since this render will get interrupted. This <del> // is one of the very rare times where we mutate the current tree <del> // during the render phase. <del> suspenseState.retryLane = attemptHydrationAtLane; <del> // TODO: Ideally this would inherit the event time of the current render <del> const eventTime = NoTimestamp; <del> scheduleUpdateOnFiber(current, attemptHydrationAtLane, eventTime); <del> } else { <del> // We have already tried to ping at a higher priority than we're rendering with <del> // so if we got here, we must have failed to hydrate at those levels. We must <del> // now give up. Instead, we're going to delete the whole subtree and instead inject <del> // a new real Suspense boundary to take its place, which may render content <del> // or fallback. This might suspend for a while and if it does we might still have <del> // an opportunity to hydrate before this pass commits. <del> } <del> } <del> <del> // If we have scheduled higher pri work above, this will probably just abort the render <del> // since we now have higher priority work, but in case it doesn't, we need to prepare to <del> // render something, if we time out. Even if that requires us to delete everything and <del> // skip hydration. <del> // Delay having to do this as long as the suspense timeout allows us. <del> renderDidSuspendDelayIfPossible(); <del> return retrySuspenseComponentWithoutHydrating( <del> current, <del> workInProgress, <del> renderLanes, <del> new Error( <del> 'This Suspense boundary received an update before it finished ' + <del> 'hydrating. This caused the boundary to switch to client rendering. ' + <del> 'The usual way to fix this is to wrap the original update ' + <del> 'in startTransition.', <del> ), <del> ); <del> } else if (isSuspenseInstancePending(suspenseInstance)) { <del> // This component is still pending more data from the server, so we can't hydrate its <del> // content. We treat it as if this component suspended itself. It might seem as if <del> // we could just try to render it client-side instead. However, this will perform a <del> // lot of unnecessary work and is unlikely to complete since it often will suspend <del> // on missing data anyway. Additionally, the server might be able to render more <del> // than we can on the client yet. In that case we'd end up with more fallback states <del> // on the client than if we just leave it alone. If the server times out or errors <del> // these should update this boundary to the permanent Fallback state instead. <del> // Mark it as having captured (i.e. suspended). <del> workInProgress.flags |= DidCapture; <del> // Leave the child in place. I.e. the dehydrated fragment. <del> workInProgress.child = current.child; <del> // Register a callback to retry this boundary once the server has sent the result. <del> const retry = retryDehydratedSuspenseBoundary.bind(null, current); <del> registerSuspenseInstanceRetry(suspenseInstance, retry); <del> return null; <add> } <add> <add> if ( <add> enableLazyContextPropagation && <add> // TODO: Factoring is a little weird, since we check this right below, too. <add> // But don't want to re-arrange the if-else chain until/unless this <add> // feature lands. <add> !didReceiveUpdate <add> ) { <add> // We need to check if any children have context before we decide to bail <add> // out, so propagate the changes now. <add> lazilyPropagateParentContextChanges(current, workInProgress, renderLanes); <add> } <add> <add> // We use lanes to indicate that a child might depend on context, so if <add> // any context has changed, we need to treat is as if the input might have changed. <add> const hasContextChanged = includesSomeLane(renderLanes, current.childLanes); <add> if (didReceiveUpdate || hasContextChanged) { <add> // This boundary has changed since the first render. This means that we are now unable to <add> // hydrate it. We might still be able to hydrate it using a higher priority lane. <add> const root = getWorkInProgressRoot(); <add> if (root !== null) { <add> const attemptHydrationAtLane = getBumpedLaneForHydration( <add> root, <add> renderLanes, <add> ); <add> if ( <add> attemptHydrationAtLane !== NoLane && <add> attemptHydrationAtLane !== suspenseState.retryLane <add> ) { <add> // Intentionally mutating since this render will get interrupted. This <add> // is one of the very rare times where we mutate the current tree <add> // during the render phase. <add> suspenseState.retryLane = attemptHydrationAtLane; <add> // TODO: Ideally this would inherit the event time of the current render <add> const eventTime = NoTimestamp; <add> scheduleUpdateOnFiber(current, attemptHydrationAtLane, eventTime); <add> } else { <add> // We have already tried to ping at a higher priority than we're rendering with <add> // so if we got here, we must have failed to hydrate at those levels. We must <add> // now give up. Instead, we're going to delete the whole subtree and instead inject <add> // a new real Suspense boundary to take its place, which may render content <add> // or fallback. This might suspend for a while and if it does we might still have <add> // an opportunity to hydrate before this pass commits. <add> } <add> } <add> <add> // If we have scheduled higher pri work above, this will probably just abort the render <add> // since we now have higher priority work, but in case it doesn't, we need to prepare to <add> // render something, if we time out. Even if that requires us to delete everything and <add> // skip hydration. <add> // Delay having to do this as long as the suspense timeout allows us. <add> renderDidSuspendDelayIfPossible(); <add> return retrySuspenseComponentWithoutHydrating( <add> current, <add> workInProgress, <add> renderLanes, <add> new Error( <add> 'This Suspense boundary received an update before it finished ' + <add> 'hydrating. This caused the boundary to switch to client rendering. ' + <add> 'The usual way to fix this is to wrap the original update ' + <add> 'in startTransition.', <add> ), <add> ); <add> } else if (isSuspenseInstancePending(suspenseInstance)) { <add> // This component is still pending more data from the server, so we can't hydrate its <add> // content. We treat it as if this component suspended itself. It might seem as if <add> // we could just try to render it client-side instead. However, this will perform a <add> // lot of unnecessary work and is unlikely to complete since it often will suspend <add> // on missing data anyway. Additionally, the server might be able to render more <add> // than we can on the client yet. In that case we'd end up with more fallback states <add> // on the client than if we just leave it alone. If the server times out or errors <add> // these should update this boundary to the permanent Fallback state instead. <add> // Mark it as having captured (i.e. suspended). <add> workInProgress.flags |= DidCapture; <add> // Leave the child in place. I.e. the dehydrated fragment. <add> workInProgress.child = current.child; <add> // Register a callback to retry this boundary once the server has sent the result. <add> const retry = retryDehydratedSuspenseBoundary.bind(null, current); <add> registerSuspenseInstanceRetry(suspenseInstance, retry); <add> return null; <add> } else { <add> // This is the first attempt. <add> reenterHydrationStateFromDehydratedSuspenseInstance( <add> workInProgress, <add> suspenseInstance, <add> suspenseState.treeContext, <add> ); <add> const primaryChildren = nextProps.children; <add> const primaryChildFragment = mountSuspensePrimaryChildren( <add> workInProgress, <add> primaryChildren, <add> renderLanes, <add> ); <add> // Mark the children as hydrating. This is a fast path to know whether this <add> // tree is part of a hydrating tree. This is used to determine if a child <add> // node has fully mounted yet, and for scheduling event replaying. <add> // Conceptually this is similar to Placement in that a new subtree is <add> // inserted into the React tree here. It just happens to not need DOM <add> // mutations because it already exists. <add> primaryChildFragment.flags |= Hydrating; <add> return primaryChildFragment; <add> } <ide> } else { <del> // This is the first attempt. <del> reenterHydrationStateFromDehydratedSuspenseInstance( <del> workInProgress, <del> suspenseInstance, <del> suspenseState.treeContext, <del> ); <del> const nextProps = workInProgress.pendingProps; <del> const primaryChildren = nextProps.children; <del> const primaryChildFragment = mountSuspensePrimaryChildren( <del> workInProgress, <del> primaryChildren, <del> renderLanes, <del> ); <del> // Mark the children as hydrating. This is a fast path to know whether this <del> // tree is part of a hydrating tree. This is used to determine if a child <del> // node has fully mounted yet, and for scheduling event replaying. <del> // Conceptually this is similar to Placement in that a new subtree is <del> // inserted into the React tree here. It just happens to not need DOM <del> // mutations because it already exists. <del> primaryChildFragment.flags |= Hydrating; <del> return primaryChildFragment; <add> // This is the second render pass. We already attempted to hydrated, but <add> // something either suspended or errored. <add> <add> if (workInProgress.flags & ForceClientRender) { <add> // Something errored during hydration. Try again without hydrating. <add> workInProgress.flags &= ~ForceClientRender; <add> return retrySuspenseComponentWithoutHydrating( <add> current, <add> workInProgress, <add> renderLanes, <add> new Error( <add> 'There was an error while hydrating this Suspense boundary. ' + <add> 'Switched to client rendering.', <add> ), <add> ); <add> } else if ((workInProgress.memoizedState: null | SuspenseState) !== null) { <add> // Something suspended and we should still be in dehydrated mode. <add> // Leave the existing child in place. <add> workInProgress.child = current.child; <add> // The dehydrated completion pass expects this flag to be there <add> // but the normal suspense pass doesn't. <add> workInProgress.flags |= DidCapture; <add> return null; <add> } else { <add> // Suspended but we should no longer be in dehydrated mode. <add> // Therefore we now have to render the fallback. <add> const nextPrimaryChildren = nextProps.children; <add> const nextFallbackChildren = nextProps.fallback; <add> const fallbackChildFragment = mountSuspenseFallbackAfterRetryWithoutHydrating( <add> current, <add> workInProgress, <add> nextPrimaryChildren, <add> nextFallbackChildren, <add> renderLanes, <add> ); <add> const primaryChildFragment: Fiber = (workInProgress.child: any); <add> primaryChildFragment.memoizedState = mountSuspenseOffscreenState( <add> renderLanes, <add> ); <add> workInProgress.memoizedState = SUSPENDED_MARKER; <add> return fallbackChildFragment; <add> } <ide> } <ide> } <ide> <ide><path>packages/react-reconciler/src/ReactFiberBeginWork.old.js <ide> function updateSuspenseComponent(current, workInProgress, renderLanes) { <ide> // a stack. <ide> if (current === null) { <ide> // Initial mount <add> <add> // Special path for hydration <ide> // If we're currently hydrating, try to hydrate this boundary. <ide> tryToClaimNextHydratableInstance(workInProgress); <ide> // This could've been a dehydrated suspense component. <ide> function updateSuspenseComponent(current, workInProgress, renderLanes) { <ide> } else { <ide> // This is an update. <ide> <del> // If the current fiber has a SuspenseState, that means it's already showing <del> // a fallback. <add> // Special path for hydration <ide> const prevState: null | SuspenseState = current.memoizedState; <ide> if (prevState !== null) { <del> // The current tree is already showing a fallback <del> <del> // Special path for hydration <ide> const dehydrated = prevState.dehydrated; <ide> if (dehydrated !== null) { <del> if (!didSuspend) { <del> return updateDehydratedSuspenseComponent( <del> current, <del> workInProgress, <del> dehydrated, <del> prevState, <del> renderLanes, <del> ); <del> } else if (workInProgress.flags & ForceClientRender) { <del> // Something errored during hydration. Try again without hydrating. <del> workInProgress.flags &= ~ForceClientRender; <del> return retrySuspenseComponentWithoutHydrating( <del> current, <del> workInProgress, <del> renderLanes, <del> new Error( <del> 'There was an error while hydrating this Suspense boundary. ' + <del> 'Switched to client rendering.', <del> ), <del> ); <del> } else if ( <del> (workInProgress.memoizedState: null | SuspenseState) !== null <del> ) { <del> // Something suspended and we should still be in dehydrated mode. <del> // Leave the existing child in place. <del> workInProgress.child = current.child; <del> // The dehydrated completion pass expects this flag to be there <del> // but the normal suspense pass doesn't. <del> workInProgress.flags |= DidCapture; <del> return null; <del> } else { <del> // Suspended but we should no longer be in dehydrated mode. <del> // Therefore we now have to render the fallback. <del> const nextPrimaryChildren = nextProps.children; <del> const nextFallbackChildren = nextProps.fallback; <del> const fallbackChildFragment = mountSuspenseFallbackAfterRetryWithoutHydrating( <del> current, <del> workInProgress, <del> nextPrimaryChildren, <del> nextFallbackChildren, <del> renderLanes, <del> ); <del> const primaryChildFragment: Fiber = (workInProgress.child: any); <del> primaryChildFragment.memoizedState = mountSuspenseOffscreenState( <del> renderLanes, <del> ); <del> workInProgress.memoizedState = SUSPENDED_MARKER; <del> return fallbackChildFragment; <del> } <del> } <del> <del> if (showFallback) { <del> const nextFallbackChildren = nextProps.fallback; <del> const nextPrimaryChildren = nextProps.children; <del> const fallbackChildFragment = updateSuspenseFallbackChildren( <add> return updateDehydratedSuspenseComponent( <ide> current, <ide> workInProgress, <del> nextPrimaryChildren, <del> nextFallbackChildren, <del> renderLanes, <del> ); <del> const primaryChildFragment: Fiber = (workInProgress.child: any); <del> const prevOffscreenState: OffscreenState | null = (current.child: any) <del> .memoizedState; <del> primaryChildFragment.memoizedState = <del> prevOffscreenState === null <del> ? mountSuspenseOffscreenState(renderLanes) <del> : updateSuspenseOffscreenState(prevOffscreenState, renderLanes); <del> if (enableTransitionTracing) { <del> const currentTransitions = getSuspendedTransitions(); <del> if (currentTransitions !== null) { <del> const primaryChildUpdateQueue: OffscreenQueue = { <del> transitions: currentTransitions, <del> }; <del> primaryChildFragment.updateQueue = primaryChildUpdateQueue; <del> } <del> } <del> primaryChildFragment.childLanes = getRemainingWorkInPrimaryTree( <del> current, <del> renderLanes, <del> ); <del> workInProgress.memoizedState = SUSPENDED_MARKER; <del> return fallbackChildFragment; <del> } else { <del> const nextPrimaryChildren = nextProps.children; <del> const primaryChildFragment = updateSuspensePrimaryChildren( <del> current, <del> workInProgress, <del> nextPrimaryChildren, <add> didSuspend, <add> nextProps, <add> dehydrated, <add> prevState, <ide> renderLanes, <ide> ); <del> workInProgress.memoizedState = null; <del> return primaryChildFragment; <ide> } <del> } else { <del> // The current tree is not already showing a fallback. <del> if (showFallback) { <del> // Timed out. <del> const nextFallbackChildren = nextProps.fallback; <del> const nextPrimaryChildren = nextProps.children; <del> const fallbackChildFragment = updateSuspenseFallbackChildren( <del> current, <del> workInProgress, <del> nextPrimaryChildren, <del> nextFallbackChildren, <del> renderLanes, <del> ); <del> const primaryChildFragment: Fiber = (workInProgress.child: any); <del> const prevOffscreenState: OffscreenState | null = (current.child: any) <del> .memoizedState; <del> primaryChildFragment.memoizedState = <del> prevOffscreenState === null <del> ? mountSuspenseOffscreenState(renderLanes) <del> : updateSuspenseOffscreenState(prevOffscreenState, renderLanes); <del> primaryChildFragment.childLanes = getRemainingWorkInPrimaryTree( <del> current, <del> renderLanes, <del> ); <add> } <ide> <del> if (enableTransitionTracing) { <del> const currentTransitions = getSuspendedTransitions(); <del> if (currentTransitions !== null) { <del> const primaryChildUpdateQueue: OffscreenQueue = { <del> transitions: currentTransitions, <del> }; <del> primaryChildFragment.updateQueue = primaryChildUpdateQueue; <del> } <add> if (showFallback) { <add> const nextFallbackChildren = nextProps.fallback; <add> const nextPrimaryChildren = nextProps.children; <add> const fallbackChildFragment = updateSuspenseFallbackChildren( <add> current, <add> workInProgress, <add> nextPrimaryChildren, <add> nextFallbackChildren, <add> renderLanes, <add> ); <add> const primaryChildFragment: Fiber = (workInProgress.child: any); <add> const prevOffscreenState: OffscreenState | null = (current.child: any) <add> .memoizedState; <add> primaryChildFragment.memoizedState = <add> prevOffscreenState === null <add> ? mountSuspenseOffscreenState(renderLanes) <add> : updateSuspenseOffscreenState(prevOffscreenState, renderLanes); <add> if (enableTransitionTracing) { <add> const currentTransitions = getSuspendedTransitions(); <add> if (currentTransitions !== null) { <add> const primaryChildUpdateQueue: OffscreenQueue = { <add> transitions: currentTransitions, <add> }; <add> primaryChildFragment.updateQueue = primaryChildUpdateQueue; <ide> } <del> <del> // Skip the primary children, and continue working on the <del> // fallback children. <del> workInProgress.memoizedState = SUSPENDED_MARKER; <del> return fallbackChildFragment; <del> } else { <del> // Still haven't timed out. Continue rendering the children, like we <del> // normally do. <del> const nextPrimaryChildren = nextProps.children; <del> const primaryChildFragment = updateSuspensePrimaryChildren( <del> current, <del> workInProgress, <del> nextPrimaryChildren, <del> renderLanes, <del> ); <del> workInProgress.memoizedState = null; <del> return primaryChildFragment; <ide> } <add> primaryChildFragment.childLanes = getRemainingWorkInPrimaryTree( <add> current, <add> renderLanes, <add> ); <add> workInProgress.memoizedState = SUSPENDED_MARKER; <add> return fallbackChildFragment; <add> } else { <add> const nextPrimaryChildren = nextProps.children; <add> const primaryChildFragment = updateSuspensePrimaryChildren( <add> current, <add> workInProgress, <add> nextPrimaryChildren, <add> renderLanes, <add> ); <add> workInProgress.memoizedState = null; <add> return primaryChildFragment; <ide> } <ide> } <ide> } <ide> function mountDehydratedSuspenseComponent( <ide> function updateDehydratedSuspenseComponent( <ide> current: Fiber, <ide> workInProgress: Fiber, <add> didSuspend: boolean, <add> nextProps: any, <ide> suspenseInstance: SuspenseInstance, <ide> suspenseState: SuspenseState, <ide> renderLanes: Lanes, <ide> ): null | Fiber { <del> // We should never be hydrating at this point because it is the first pass, <del> // but after we've already committed once. <del> warnIfHydrating(); <add> if (!didSuspend) { <add> // This is the first render pass. Attempt to hydrate. <ide> <del> if ((workInProgress.mode & ConcurrentMode) === NoMode) { <del> return retrySuspenseComponentWithoutHydrating( <del> current, <del> workInProgress, <del> renderLanes, <del> // TODO: When we delete legacy mode, we should make this error argument <del> // required — every concurrent mode path that causes hydration to <del> // de-opt to client rendering should have an error message. <del> null, <del> ); <del> } <add> // We should never be hydrating at this point because it is the first pass, <add> // but after we've already committed once. <add> warnIfHydrating(); <ide> <del> if (isSuspenseInstanceFallback(suspenseInstance)) { <del> // This boundary is in a permanent fallback state. In this case, we'll never <del> // get an update and we'll never be able to hydrate the final content. Let's just try the <del> // client side render instead. <del> return retrySuspenseComponentWithoutHydrating( <del> current, <del> workInProgress, <del> renderLanes, <del> // TODO: The server should serialize the error message so we can log it <del> // here on the client. Or, in production, a hash/id that corresponds to <del> // the error. <del> new Error( <del> 'The server could not finish this Suspense boundary, likely ' + <del> 'due to an error during server rendering. Switched to ' + <del> 'client rendering.', <del> ), <del> ); <del> } <add> if ((workInProgress.mode & ConcurrentMode) === NoMode) { <add> return retrySuspenseComponentWithoutHydrating( <add> current, <add> workInProgress, <add> renderLanes, <add> // TODO: When we delete legacy mode, we should make this error argument <add> // required — every concurrent mode path that causes hydration to <add> // de-opt to client rendering should have an error message. <add> null, <add> ); <add> } <ide> <del> if ( <del> enableLazyContextPropagation && <del> // TODO: Factoring is a little weird, since we check this right below, too. <del> // But don't want to re-arrange the if-else chain until/unless this <del> // feature lands. <del> !didReceiveUpdate <del> ) { <del> // We need to check if any children have context before we decide to bail <del> // out, so propagate the changes now. <del> lazilyPropagateParentContextChanges(current, workInProgress, renderLanes); <del> } <del> <del> // We use lanes to indicate that a child might depend on context, so if <del> // any context has changed, we need to treat is as if the input might have changed. <del> const hasContextChanged = includesSomeLane(renderLanes, current.childLanes); <del> if (didReceiveUpdate || hasContextChanged) { <del> // This boundary has changed since the first render. This means that we are now unable to <del> // hydrate it. We might still be able to hydrate it using a higher priority lane. <del> const root = getWorkInProgressRoot(); <del> if (root !== null) { <del> const attemptHydrationAtLane = getBumpedLaneForHydration( <del> root, <add> if (isSuspenseInstanceFallback(suspenseInstance)) { <add> // This boundary is in a permanent fallback state. In this case, we'll never <add> // get an update and we'll never be able to hydrate the final content. Let's just try the <add> // client side render instead. <add> return retrySuspenseComponentWithoutHydrating( <add> current, <add> workInProgress, <ide> renderLanes, <add> // TODO: The server should serialize the error message so we can log it <add> // here on the client. Or, in production, a hash/id that corresponds to <add> // the error. <add> new Error( <add> 'The server could not finish this Suspense boundary, likely ' + <add> 'due to an error during server rendering. Switched to ' + <add> 'client rendering.', <add> ), <ide> ); <del> if ( <del> attemptHydrationAtLane !== NoLane && <del> attemptHydrationAtLane !== suspenseState.retryLane <del> ) { <del> // Intentionally mutating since this render will get interrupted. This <del> // is one of the very rare times where we mutate the current tree <del> // during the render phase. <del> suspenseState.retryLane = attemptHydrationAtLane; <del> // TODO: Ideally this would inherit the event time of the current render <del> const eventTime = NoTimestamp; <del> scheduleUpdateOnFiber(current, attemptHydrationAtLane, eventTime); <del> } else { <del> // We have already tried to ping at a higher priority than we're rendering with <del> // so if we got here, we must have failed to hydrate at those levels. We must <del> // now give up. Instead, we're going to delete the whole subtree and instead inject <del> // a new real Suspense boundary to take its place, which may render content <del> // or fallback. This might suspend for a while and if it does we might still have <del> // an opportunity to hydrate before this pass commits. <del> } <del> } <del> <del> // If we have scheduled higher pri work above, this will probably just abort the render <del> // since we now have higher priority work, but in case it doesn't, we need to prepare to <del> // render something, if we time out. Even if that requires us to delete everything and <del> // skip hydration. <del> // Delay having to do this as long as the suspense timeout allows us. <del> renderDidSuspendDelayIfPossible(); <del> return retrySuspenseComponentWithoutHydrating( <del> current, <del> workInProgress, <del> renderLanes, <del> new Error( <del> 'This Suspense boundary received an update before it finished ' + <del> 'hydrating. This caused the boundary to switch to client rendering. ' + <del> 'The usual way to fix this is to wrap the original update ' + <del> 'in startTransition.', <del> ), <del> ); <del> } else if (isSuspenseInstancePending(suspenseInstance)) { <del> // This component is still pending more data from the server, so we can't hydrate its <del> // content. We treat it as if this component suspended itself. It might seem as if <del> // we could just try to render it client-side instead. However, this will perform a <del> // lot of unnecessary work and is unlikely to complete since it often will suspend <del> // on missing data anyway. Additionally, the server might be able to render more <del> // than we can on the client yet. In that case we'd end up with more fallback states <del> // on the client than if we just leave it alone. If the server times out or errors <del> // these should update this boundary to the permanent Fallback state instead. <del> // Mark it as having captured (i.e. suspended). <del> workInProgress.flags |= DidCapture; <del> // Leave the child in place. I.e. the dehydrated fragment. <del> workInProgress.child = current.child; <del> // Register a callback to retry this boundary once the server has sent the result. <del> const retry = retryDehydratedSuspenseBoundary.bind(null, current); <del> registerSuspenseInstanceRetry(suspenseInstance, retry); <del> return null; <add> } <add> <add> if ( <add> enableLazyContextPropagation && <add> // TODO: Factoring is a little weird, since we check this right below, too. <add> // But don't want to re-arrange the if-else chain until/unless this <add> // feature lands. <add> !didReceiveUpdate <add> ) { <add> // We need to check if any children have context before we decide to bail <add> // out, so propagate the changes now. <add> lazilyPropagateParentContextChanges(current, workInProgress, renderLanes); <add> } <add> <add> // We use lanes to indicate that a child might depend on context, so if <add> // any context has changed, we need to treat is as if the input might have changed. <add> const hasContextChanged = includesSomeLane(renderLanes, current.childLanes); <add> if (didReceiveUpdate || hasContextChanged) { <add> // This boundary has changed since the first render. This means that we are now unable to <add> // hydrate it. We might still be able to hydrate it using a higher priority lane. <add> const root = getWorkInProgressRoot(); <add> if (root !== null) { <add> const attemptHydrationAtLane = getBumpedLaneForHydration( <add> root, <add> renderLanes, <add> ); <add> if ( <add> attemptHydrationAtLane !== NoLane && <add> attemptHydrationAtLane !== suspenseState.retryLane <add> ) { <add> // Intentionally mutating since this render will get interrupted. This <add> // is one of the very rare times where we mutate the current tree <add> // during the render phase. <add> suspenseState.retryLane = attemptHydrationAtLane; <add> // TODO: Ideally this would inherit the event time of the current render <add> const eventTime = NoTimestamp; <add> scheduleUpdateOnFiber(current, attemptHydrationAtLane, eventTime); <add> } else { <add> // We have already tried to ping at a higher priority than we're rendering with <add> // so if we got here, we must have failed to hydrate at those levels. We must <add> // now give up. Instead, we're going to delete the whole subtree and instead inject <add> // a new real Suspense boundary to take its place, which may render content <add> // or fallback. This might suspend for a while and if it does we might still have <add> // an opportunity to hydrate before this pass commits. <add> } <add> } <add> <add> // If we have scheduled higher pri work above, this will probably just abort the render <add> // since we now have higher priority work, but in case it doesn't, we need to prepare to <add> // render something, if we time out. Even if that requires us to delete everything and <add> // skip hydration. <add> // Delay having to do this as long as the suspense timeout allows us. <add> renderDidSuspendDelayIfPossible(); <add> return retrySuspenseComponentWithoutHydrating( <add> current, <add> workInProgress, <add> renderLanes, <add> new Error( <add> 'This Suspense boundary received an update before it finished ' + <add> 'hydrating. This caused the boundary to switch to client rendering. ' + <add> 'The usual way to fix this is to wrap the original update ' + <add> 'in startTransition.', <add> ), <add> ); <add> } else if (isSuspenseInstancePending(suspenseInstance)) { <add> // This component is still pending more data from the server, so we can't hydrate its <add> // content. We treat it as if this component suspended itself. It might seem as if <add> // we could just try to render it client-side instead. However, this will perform a <add> // lot of unnecessary work and is unlikely to complete since it often will suspend <add> // on missing data anyway. Additionally, the server might be able to render more <add> // than we can on the client yet. In that case we'd end up with more fallback states <add> // on the client than if we just leave it alone. If the server times out or errors <add> // these should update this boundary to the permanent Fallback state instead. <add> // Mark it as having captured (i.e. suspended). <add> workInProgress.flags |= DidCapture; <add> // Leave the child in place. I.e. the dehydrated fragment. <add> workInProgress.child = current.child; <add> // Register a callback to retry this boundary once the server has sent the result. <add> const retry = retryDehydratedSuspenseBoundary.bind(null, current); <add> registerSuspenseInstanceRetry(suspenseInstance, retry); <add> return null; <add> } else { <add> // This is the first attempt. <add> reenterHydrationStateFromDehydratedSuspenseInstance( <add> workInProgress, <add> suspenseInstance, <add> suspenseState.treeContext, <add> ); <add> const primaryChildren = nextProps.children; <add> const primaryChildFragment = mountSuspensePrimaryChildren( <add> workInProgress, <add> primaryChildren, <add> renderLanes, <add> ); <add> // Mark the children as hydrating. This is a fast path to know whether this <add> // tree is part of a hydrating tree. This is used to determine if a child <add> // node has fully mounted yet, and for scheduling event replaying. <add> // Conceptually this is similar to Placement in that a new subtree is <add> // inserted into the React tree here. It just happens to not need DOM <add> // mutations because it already exists. <add> primaryChildFragment.flags |= Hydrating; <add> return primaryChildFragment; <add> } <ide> } else { <del> // This is the first attempt. <del> reenterHydrationStateFromDehydratedSuspenseInstance( <del> workInProgress, <del> suspenseInstance, <del> suspenseState.treeContext, <del> ); <del> const nextProps = workInProgress.pendingProps; <del> const primaryChildren = nextProps.children; <del> const primaryChildFragment = mountSuspensePrimaryChildren( <del> workInProgress, <del> primaryChildren, <del> renderLanes, <del> ); <del> // Mark the children as hydrating. This is a fast path to know whether this <del> // tree is part of a hydrating tree. This is used to determine if a child <del> // node has fully mounted yet, and for scheduling event replaying. <del> // Conceptually this is similar to Placement in that a new subtree is <del> // inserted into the React tree here. It just happens to not need DOM <del> // mutations because it already exists. <del> primaryChildFragment.flags |= Hydrating; <del> return primaryChildFragment; <add> // This is the second render pass. We already attempted to hydrated, but <add> // something either suspended or errored. <add> <add> if (workInProgress.flags & ForceClientRender) { <add> // Something errored during hydration. Try again without hydrating. <add> workInProgress.flags &= ~ForceClientRender; <add> return retrySuspenseComponentWithoutHydrating( <add> current, <add> workInProgress, <add> renderLanes, <add> new Error( <add> 'There was an error while hydrating this Suspense boundary. ' + <add> 'Switched to client rendering.', <add> ), <add> ); <add> } else if ((workInProgress.memoizedState: null | SuspenseState) !== null) { <add> // Something suspended and we should still be in dehydrated mode. <add> // Leave the existing child in place. <add> workInProgress.child = current.child; <add> // The dehydrated completion pass expects this flag to be there <add> // but the normal suspense pass doesn't. <add> workInProgress.flags |= DidCapture; <add> return null; <add> } else { <add> // Suspended but we should no longer be in dehydrated mode. <add> // Therefore we now have to render the fallback. <add> const nextPrimaryChildren = nextProps.children; <add> const nextFallbackChildren = nextProps.fallback; <add> const fallbackChildFragment = mountSuspenseFallbackAfterRetryWithoutHydrating( <add> current, <add> workInProgress, <add> nextPrimaryChildren, <add> nextFallbackChildren, <add> renderLanes, <add> ); <add> const primaryChildFragment: Fiber = (workInProgress.child: any); <add> primaryChildFragment.memoizedState = mountSuspenseOffscreenState( <add> renderLanes, <add> ); <add> workInProgress.memoizedState = SUSPENDED_MARKER; <add> return fallbackChildFragment; <add> } <ide> } <ide> } <ide> <ide><path>packages/react-reconciler/src/ReactFiberCompleteWork.new.js <ide> function bubbleProperties(completedWork: Fiber) { <ide> return didBailout; <ide> } <ide> <add>function completeDehydratedSuspenseBoundary( <add> current: Fiber | null, <add> workInProgress: Fiber, <add> nextState: SuspenseState | null, <add>): boolean { <add> if ( <add> hasUnhydratedTailNodes() && <add> (workInProgress.mode & ConcurrentMode) !== NoMode && <add> (workInProgress.flags & DidCapture) === NoFlags <add> ) { <add> warnIfUnhydratedTailNodes(workInProgress); <add> resetHydrationState(); <add> workInProgress.flags |= ForceClientRender | Incomplete | ShouldCapture; <add> <add> return false; <add> } <add> <add> const wasHydrated = popHydrationState(workInProgress); <add> <add> if (nextState !== null && nextState.dehydrated !== null) { <add> // We might be inside a hydration state the first time we're picking up this <add> // Suspense boundary, and also after we've reentered it for further hydration. <add> if (current === null) { <add> if (!wasHydrated) { <add> throw new Error( <add> 'A dehydrated suspense component was completed without a hydrated node. ' + <add> 'This is probably a bug in React.', <add> ); <add> } <add> prepareToHydrateHostSuspenseInstance(workInProgress); <add> bubbleProperties(workInProgress); <add> if (enableProfilerTimer) { <add> if ((workInProgress.mode & ProfileMode) !== NoMode) { <add> const isTimedOutSuspense = nextState !== null; <add> if (isTimedOutSuspense) { <add> // Don't count time spent in a timed out Suspense subtree as part of the base duration. <add> const primaryChildFragment = workInProgress.child; <add> if (primaryChildFragment !== null) { <add> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator <add> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number); <add> } <add> } <add> } <add> } <add> return false; <add> } else { <add> // We might have reentered this boundary to hydrate it. If so, we need to reset the hydration <add> // state since we're now exiting out of it. popHydrationState doesn't do that for us. <add> resetHydrationState(); <add> if ((workInProgress.flags & DidCapture) === NoFlags) { <add> // This boundary did not suspend so it's now hydrated and unsuspended. <add> workInProgress.memoizedState = null; <add> } <add> // If nothing suspended, we need to schedule an effect to mark this boundary <add> // as having hydrated so events know that they're free to be invoked. <add> // It's also a signal to replay events and the suspense callback. <add> // If something suspended, schedule an effect to attach retry listeners. <add> // So we might as well always mark this. <add> workInProgress.flags |= Update; <add> bubbleProperties(workInProgress); <add> if (enableProfilerTimer) { <add> if ((workInProgress.mode & ProfileMode) !== NoMode) { <add> const isTimedOutSuspense = nextState !== null; <add> if (isTimedOutSuspense) { <add> // Don't count time spent in a timed out Suspense subtree as part of the base duration. <add> const primaryChildFragment = workInProgress.child; <add> if (primaryChildFragment !== null) { <add> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator <add> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number); <add> } <add> } <add> } <add> } <add> return false; <add> } <add> } else { <add> // Successfully completed this tree. If this was a forced client render, <add> // there may have been recoverable errors during first hydration <add> // attempt. If so, add them to a queue so we can log them in the <add> // commit phase. <add> upgradeHydrationErrorsToRecoverable(); <add> <add> // Fall through to normal Suspense path <add> return true; <add> } <add>} <add> <ide> function completeWork( <ide> current: Fiber | null, <ide> workInProgress: Fiber, <ide> function completeWork( <ide> popSuspenseContext(workInProgress); <ide> const nextState: null | SuspenseState = workInProgress.memoizedState; <ide> <add> // Special path for dehydrated boundaries. We may eventually move this <add> // to its own fiber type so that we can add other kinds of hydration <add> // boundaries that aren't associated with a Suspense tree. In anticipation <add> // of such a refactor, all the hydration logic is contained in <add> // this branch. <ide> if ( <del> hasUnhydratedTailNodes() && <del> (workInProgress.mode & ConcurrentMode) !== NoMode && <del> (workInProgress.flags & DidCapture) === NoFlags <add> current === null || <add> (current.memoizedState !== null && <add> current.memoizedState.dehydrated !== null) <ide> ) { <del> warnIfUnhydratedTailNodes(workInProgress); <del> resetHydrationState(); <del> workInProgress.flags |= ForceClientRender | Incomplete | ShouldCapture; <del> return workInProgress; <del> } <del> if (nextState !== null && nextState.dehydrated !== null) { <del> // We might be inside a hydration state the first time we're picking up this <del> // Suspense boundary, and also after we've reentered it for further hydration. <del> const wasHydrated = popHydrationState(workInProgress); <del> if (current === null) { <del> if (!wasHydrated) { <del> throw new Error( <del> 'A dehydrated suspense component was completed without a hydrated node. ' + <del> 'This is probably a bug in React.', <del> ); <del> } <del> prepareToHydrateHostSuspenseInstance(workInProgress); <del> bubbleProperties(workInProgress); <del> if (enableProfilerTimer) { <del> if ((workInProgress.mode & ProfileMode) !== NoMode) { <del> const isTimedOutSuspense = nextState !== null; <del> if (isTimedOutSuspense) { <del> // Don't count time spent in a timed out Suspense subtree as part of the base duration. <del> const primaryChildFragment = workInProgress.child; <del> if (primaryChildFragment !== null) { <del> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator <del> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number); <del> } <del> } <del> } <del> } <del> return null; <del> } else { <del> // We might have reentered this boundary to hydrate it. If so, we need to reset the hydration <del> // state since we're now exiting out of it. popHydrationState doesn't do that for us. <del> resetHydrationState(); <del> if ((workInProgress.flags & DidCapture) === NoFlags) { <del> // This boundary did not suspend so it's now hydrated and unsuspended. <del> workInProgress.memoizedState = null; <del> } <del> // If nothing suspended, we need to schedule an effect to mark this boundary <del> // as having hydrated so events know that they're free to be invoked. <del> // It's also a signal to replay events and the suspense callback. <del> // If something suspended, schedule an effect to attach retry listeners. <del> // So we might as well always mark this. <del> workInProgress.flags |= Update; <del> bubbleProperties(workInProgress); <del> if (enableProfilerTimer) { <del> if ((workInProgress.mode & ProfileMode) !== NoMode) { <del> const isTimedOutSuspense = nextState !== null; <del> if (isTimedOutSuspense) { <del> // Don't count time spent in a timed out Suspense subtree as part of the base duration. <del> const primaryChildFragment = workInProgress.child; <del> if (primaryChildFragment !== null) { <del> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator <del> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number); <del> } <del> } <del> } <add> const fallthroughToNormalSuspensePath = completeDehydratedSuspenseBoundary( <add> current, <add> workInProgress, <add> nextState, <add> ); <add> if (!fallthroughToNormalSuspensePath) { <add> if (workInProgress.flags & ShouldCapture) { <add> // Special case. There were remaining unhydrated nodes. We treat <add> // this as a mismatch. Revert to client rendering. <add> return workInProgress; <add> } else { <add> // Did not finish hydrating, either because this is the initial <add> // render or because something suspended. <add> return null; <ide> } <del> return null; <ide> } <del> } <ide> <del> // Successfully completed this tree. If this was a forced client render, <del> // there may have been recoverable errors during first hydration <del> // attempt. If so, add them to a queue so we can log them in the <del> // commit phase. <del> upgradeHydrationErrorsToRecoverable(); <add> // Continue with the normal Suspense path. <add> } <ide> <ide> if ((workInProgress.flags & DidCapture) !== NoFlags) { <ide> // Something suspended. Re-render with the fallback children. <ide> function completeWork( <ide> } <ide> <ide> const nextDidTimeout = nextState !== null; <del> let prevDidTimeout = false; <del> if (current === null) { <del> popHydrationState(workInProgress); <del> } else { <del> const prevState: null | SuspenseState = current.memoizedState; <del> prevDidTimeout = prevState !== null; <del> } <add> const prevDidTimeout = <add> current !== null && <add> (current.memoizedState: null | SuspenseState) !== null; <ide> <ide> if (enableCache && nextDidTimeout) { <ide> const offscreenFiber: Fiber = (workInProgress.child: any); <ide><path>packages/react-reconciler/src/ReactFiberCompleteWork.old.js <ide> function bubbleProperties(completedWork: Fiber) { <ide> return didBailout; <ide> } <ide> <add>function completeDehydratedSuspenseBoundary( <add> current: Fiber | null, <add> workInProgress: Fiber, <add> nextState: SuspenseState | null, <add>): boolean { <add> if ( <add> hasUnhydratedTailNodes() && <add> (workInProgress.mode & ConcurrentMode) !== NoMode && <add> (workInProgress.flags & DidCapture) === NoFlags <add> ) { <add> warnIfUnhydratedTailNodes(workInProgress); <add> resetHydrationState(); <add> workInProgress.flags |= ForceClientRender | Incomplete | ShouldCapture; <add> <add> return false; <add> } <add> <add> const wasHydrated = popHydrationState(workInProgress); <add> <add> if (nextState !== null && nextState.dehydrated !== null) { <add> // We might be inside a hydration state the first time we're picking up this <add> // Suspense boundary, and also after we've reentered it for further hydration. <add> if (current === null) { <add> if (!wasHydrated) { <add> throw new Error( <add> 'A dehydrated suspense component was completed without a hydrated node. ' + <add> 'This is probably a bug in React.', <add> ); <add> } <add> prepareToHydrateHostSuspenseInstance(workInProgress); <add> bubbleProperties(workInProgress); <add> if (enableProfilerTimer) { <add> if ((workInProgress.mode & ProfileMode) !== NoMode) { <add> const isTimedOutSuspense = nextState !== null; <add> if (isTimedOutSuspense) { <add> // Don't count time spent in a timed out Suspense subtree as part of the base duration. <add> const primaryChildFragment = workInProgress.child; <add> if (primaryChildFragment !== null) { <add> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator <add> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number); <add> } <add> } <add> } <add> } <add> return false; <add> } else { <add> // We might have reentered this boundary to hydrate it. If so, we need to reset the hydration <add> // state since we're now exiting out of it. popHydrationState doesn't do that for us. <add> resetHydrationState(); <add> if ((workInProgress.flags & DidCapture) === NoFlags) { <add> // This boundary did not suspend so it's now hydrated and unsuspended. <add> workInProgress.memoizedState = null; <add> } <add> // If nothing suspended, we need to schedule an effect to mark this boundary <add> // as having hydrated so events know that they're free to be invoked. <add> // It's also a signal to replay events and the suspense callback. <add> // If something suspended, schedule an effect to attach retry listeners. <add> // So we might as well always mark this. <add> workInProgress.flags |= Update; <add> bubbleProperties(workInProgress); <add> if (enableProfilerTimer) { <add> if ((workInProgress.mode & ProfileMode) !== NoMode) { <add> const isTimedOutSuspense = nextState !== null; <add> if (isTimedOutSuspense) { <add> // Don't count time spent in a timed out Suspense subtree as part of the base duration. <add> const primaryChildFragment = workInProgress.child; <add> if (primaryChildFragment !== null) { <add> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator <add> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number); <add> } <add> } <add> } <add> } <add> return false; <add> } <add> } else { <add> // Successfully completed this tree. If this was a forced client render, <add> // there may have been recoverable errors during first hydration <add> // attempt. If so, add them to a queue so we can log them in the <add> // commit phase. <add> upgradeHydrationErrorsToRecoverable(); <add> <add> // Fall through to normal Suspense path <add> return true; <add> } <add>} <add> <ide> function completeWork( <ide> current: Fiber | null, <ide> workInProgress: Fiber, <ide> function completeWork( <ide> popSuspenseContext(workInProgress); <ide> const nextState: null | SuspenseState = workInProgress.memoizedState; <ide> <add> // Special path for dehydrated boundaries. We may eventually move this <add> // to its own fiber type so that we can add other kinds of hydration <add> // boundaries that aren't associated with a Suspense tree. In anticipation <add> // of such a refactor, all the hydration logic is contained in <add> // this branch. <ide> if ( <del> hasUnhydratedTailNodes() && <del> (workInProgress.mode & ConcurrentMode) !== NoMode && <del> (workInProgress.flags & DidCapture) === NoFlags <add> current === null || <add> (current.memoizedState !== null && <add> current.memoizedState.dehydrated !== null) <ide> ) { <del> warnIfUnhydratedTailNodes(workInProgress); <del> resetHydrationState(); <del> workInProgress.flags |= ForceClientRender | Incomplete | ShouldCapture; <del> return workInProgress; <del> } <del> if (nextState !== null && nextState.dehydrated !== null) { <del> // We might be inside a hydration state the first time we're picking up this <del> // Suspense boundary, and also after we've reentered it for further hydration. <del> const wasHydrated = popHydrationState(workInProgress); <del> if (current === null) { <del> if (!wasHydrated) { <del> throw new Error( <del> 'A dehydrated suspense component was completed without a hydrated node. ' + <del> 'This is probably a bug in React.', <del> ); <del> } <del> prepareToHydrateHostSuspenseInstance(workInProgress); <del> bubbleProperties(workInProgress); <del> if (enableProfilerTimer) { <del> if ((workInProgress.mode & ProfileMode) !== NoMode) { <del> const isTimedOutSuspense = nextState !== null; <del> if (isTimedOutSuspense) { <del> // Don't count time spent in a timed out Suspense subtree as part of the base duration. <del> const primaryChildFragment = workInProgress.child; <del> if (primaryChildFragment !== null) { <del> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator <del> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number); <del> } <del> } <del> } <del> } <del> return null; <del> } else { <del> // We might have reentered this boundary to hydrate it. If so, we need to reset the hydration <del> // state since we're now exiting out of it. popHydrationState doesn't do that for us. <del> resetHydrationState(); <del> if ((workInProgress.flags & DidCapture) === NoFlags) { <del> // This boundary did not suspend so it's now hydrated and unsuspended. <del> workInProgress.memoizedState = null; <del> } <del> // If nothing suspended, we need to schedule an effect to mark this boundary <del> // as having hydrated so events know that they're free to be invoked. <del> // It's also a signal to replay events and the suspense callback. <del> // If something suspended, schedule an effect to attach retry listeners. <del> // So we might as well always mark this. <del> workInProgress.flags |= Update; <del> bubbleProperties(workInProgress); <del> if (enableProfilerTimer) { <del> if ((workInProgress.mode & ProfileMode) !== NoMode) { <del> const isTimedOutSuspense = nextState !== null; <del> if (isTimedOutSuspense) { <del> // Don't count time spent in a timed out Suspense subtree as part of the base duration. <del> const primaryChildFragment = workInProgress.child; <del> if (primaryChildFragment !== null) { <del> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator <del> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number); <del> } <del> } <del> } <add> const fallthroughToNormalSuspensePath = completeDehydratedSuspenseBoundary( <add> current, <add> workInProgress, <add> nextState, <add> ); <add> if (!fallthroughToNormalSuspensePath) { <add> if (workInProgress.flags & ShouldCapture) { <add> // Special case. There were remaining unhydrated nodes. We treat <add> // this as a mismatch. Revert to client rendering. <add> return workInProgress; <add> } else { <add> // Did not finish hydrating, either because this is the initial <add> // render or because something suspended. <add> return null; <ide> } <del> return null; <ide> } <del> } <ide> <del> // Successfully completed this tree. If this was a forced client render, <del> // there may have been recoverable errors during first hydration <del> // attempt. If so, add them to a queue so we can log them in the <del> // commit phase. <del> upgradeHydrationErrorsToRecoverable(); <add> // Continue with the normal Suspense path. <add> } <ide> <ide> if ((workInProgress.flags & DidCapture) !== NoFlags) { <ide> // Something suspended. Re-render with the fallback children. <ide> function completeWork( <ide> } <ide> <ide> const nextDidTimeout = nextState !== null; <del> let prevDidTimeout = false; <del> if (current === null) { <del> popHydrationState(workInProgress); <del> } else { <del> const prevState: null | SuspenseState = current.memoizedState; <del> prevDidTimeout = prevState !== null; <del> } <add> const prevDidTimeout = <add> current !== null && <add> (current.memoizedState: null | SuspenseState) !== null; <ide> <ide> if (enableCache && nextDidTimeout) { <ide> const offscreenFiber: Fiber = (workInProgress.child: any);
4
Javascript
Javascript
remove obsolete sharedarraybuffer eslint config
82ef450e0e22ef3d5cff03d26aed7d708fa10537
<ide><path>.eslintrc.js <ide> module.exports = { <ide> ], <ide> <ide> globals: { <del> SharedArrayBuffer: true, <del> <ide> spyOnDev: true, <ide> spyOnDevAndProd: true, <ide> spyOnProd: true, <ide><path>scripts/rollup/validate/eslintrc.cjs.js <ide> module.exports = { <ide> trustedTypes: true, <ide> <ide> // Scheduler profiling <del> SharedArrayBuffer: true, <ide> Int32Array: true, <ide> ArrayBuffer: true, <ide> <ide><path>scripts/rollup/validate/eslintrc.cjs2015.js <ide> module.exports = { <ide> trustedTypes: true, <ide> <ide> // Scheduler profiling <del> SharedArrayBuffer: true, <ide> Int32Array: true, <ide> ArrayBuffer: true, <ide> <ide><path>scripts/rollup/validate/eslintrc.esm.js <ide> module.exports = { <ide> trustedTypes: true, <ide> <ide> // Scheduler profiling <del> SharedArrayBuffer: true, <ide> Int32Array: true, <ide> ArrayBuffer: true, <ide> <ide><path>scripts/rollup/validate/eslintrc.fb.js <ide> module.exports = { <ide> trustedTypes: true, <ide> <ide> // Scheduler profiling <del> SharedArrayBuffer: true, <ide> Int32Array: true, <ide> ArrayBuffer: true, <ide> <ide><path>scripts/rollup/validate/eslintrc.rn.js <ide> module.exports = { <ide> trustedTypes: true, <ide> <ide> // Scheduler profiling <del> SharedArrayBuffer: true, <ide> Int32Array: true, <ide> ArrayBuffer: true, <ide> <ide><path>scripts/rollup/validate/eslintrc.umd.js <ide> module.exports = { <ide> trustedTypes: true, <ide> <ide> // Scheduler profiling <del> SharedArrayBuffer: true, <ide> Int32Array: true, <ide> ArrayBuffer: true, <ide>
7
Ruby
Ruby
relax hub requirement
6dc72f2679208f34d9d09f547dda5a2ba0714a9f
<ide><path>Library/Homebrew/dev-cmd/boneyard-formula-pr.rb <ide> def boneyard_formula_pr <ide> tap_migrations = tap_migrations.sort.inject({}) { |a, e| a.merge!(e[0] => e[1]) } <ide> tap_migrations_path.atomic_write(JSON.pretty_generate(tap_migrations) + "\n") <ide> end <del> unless Formula["hub"].any_version_installed? || local_only <add> unless which("hub") || local_only <ide> if ARGV.dry_run? <ide> puts "brew install hub" <ide> else
1
Python
Python
remove unused imports.
51a2b441c65e5df4fa7b540c2c1211e3adff53cb
<ide><path>official/mnist/mnist.py <ide> from __future__ import division <ide> from __future__ import print_function <ide> <del>import argparse <del>import sys <del> <ide> from absl import app as absl_app <ide> from absl import flags <ide> import tensorflow as tf # pylint: disable=g-bad-import-order <ide><path>official/mnist/mnist_eager.py <ide> from __future__ import print_function <ide> <ide> import os <del>import sys <ide> import time <ide> <ide> # pylint: disable=g-bad-import-order <ide><path>official/resnet/cifar10_main.py <ide> from __future__ import print_function <ide> <ide> import os <del>import sys <ide> <ide> from absl import app as absl_app <ide> from absl import flags <ide><path>official/resnet/imagenet_main.py <ide> from __future__ import print_function <ide> <ide> import os <del>import sys <ide> <ide> from absl import app as absl_app <ide> from absl import flags <ide><path>official/resnet/resnet_run_loop.py <ide> from __future__ import division <ide> from __future__ import print_function <ide> <del>import argparse <ide> import os <ide> <ide> from absl import flags <ide><path>official/wide_deep/wide_deep.py <ide> from __future__ import division <ide> from __future__ import print_function <ide> <del>import argparse <ide> import os <ide> import shutil <del>import sys <ide> <ide> from absl import app as absl_app <ide> from absl import flags
6
Ruby
Ruby
remove a comment that is not fact in tests
41758964e9515ae879fce0be687da2f5953bc950
<ide><path>activerecord/test/cases/calculations_test.rb <ide> def test_pluck_with_selection_clause <ide> assert_equal [50, 53, 55, 60], Account.pluck(Arel.sql("DISTINCT credit_limit")).sort <ide> assert_equal [50, 53, 55, 60], Account.pluck(Arel.sql("DISTINCT accounts.credit_limit")).sort <ide> assert_equal [50, 53, 55, 60], Account.pluck(Arel.sql("DISTINCT(credit_limit)")).sort <del> <del> # MySQL returns "SUM(DISTINCT(credit_limit))" as the column name unless <del> # an alias is provided. Without the alias, the column cannot be found <del> # and properly typecast. <del> assert_equal [50 + 53 + 55 + 60], Account.pluck(Arel.sql("SUM(DISTINCT(credit_limit)) as credit_limit")) <add> assert_equal [50 + 53 + 55 + 60], Account.pluck(Arel.sql("SUM(DISTINCT(credit_limit))")) <ide> end <ide> <ide> def test_plucks_with_ids
1
PHP
PHP
fix missing use of binding key in belongstomany
6fcff4875f07951294c1185fc4e1769929639ad5
<ide><path>src/ORM/Association/BelongsToMany.php <ide> protected function _generateJunctionAssociations(Table $junction, Table $source, <ide> <ide> if (!$junction->hasAssociation($sAlias)) { <ide> $junction->belongsTo($sAlias, [ <add> 'bindingKey' => $this->getBindingKey(), <ide> 'foreignKey' => $this->getForeignKey(), <ide> 'targetTable' => $source, <ide> ]); <ide><path>tests/Fixture/ArticlesTagsBindingKeysFixture.php <add><?php <add>/** <add> * CakePHP(tm) : Rapid Development Framework (https://cakephp.org) <add> * Copyright (c) Cake Software Foundation, Inc. (https://cakefoundation.org) <add> * <add> * Licensed under The MIT License <add> * For full copyright and license information, please see the LICENSE.txt <add> * Redistributions of files must retain the above copyright notice <add> * <add> * @copyright Copyright (c) Cake Software Foundation, Inc. (https://cakefoundation.org) <add> * @link https://cakephp.org CakePHP(tm) Project <add> * @since 4.3.7 <add> * @license https://opensource.org/licenses/mit-license.php MIT License <add> */ <add>namespace Cake\Test\Fixture; <add> <add>use Cake\TestSuite\Fixture\TestFixture; <add> <add>/** <add> * Fixture for testing bindingKey in belongstomany associations. <add> */ <add>class ArticlesTagsBindingKeysFixture extends TestFixture <add>{ <add> /** <add> * records property <add> * <add> * @var array <add> */ <add> public $records = [ <add> ['article_id' => 1, 'tagname' => 'tag1'], <add> ['article_id' => 1, 'tagname' => 'tag2'], <add> ['article_id' => 2, 'tagname' => 'tag1'], <add> ['article_id' => 2, 'tagname' => 'tag3'], <add> ]; <add>} <ide><path>tests/TestCase/ORM/Association/BelongsToManyTest.php <ide> class BelongsToManyTest extends TestCase <ide> 'core.Tags', <ide> 'core.SpecialTags', <ide> 'core.ArticlesTags', <add> 'core.ArticlesTagsBindingKeys', <ide> 'core.BinaryUuidItems', <ide> 'core.BinaryUuidTags', <ide> 'core.BinaryUuidItemsBinaryUuidTags', <ide> public function testCustomTargetBindingKeyLink(): void <ide> $this->assertCount(1, $results); <ide> $this->assertCount(3, $results[0]->special_tags); <ide> } <add> <add> /** <add> * Test custom binding key for target table association <add> */ <add> public function testBindingKeyMatching(): void <add> { <add> $table = $this->getTableLocator()->get('Tags'); <add> $table->belongsToMany('Articles', [ <add> 'through' => 'ArticlesTagsBindingKeys', <add> 'foreignKey' => 'tagname', <add> 'targetForeignKey' => 'article_id', <add> 'bindingKey' => 'name', <add> ]); <add> $query = $table->find() <add> ->matching('Articles', function ($q) { <add> return $q->where('Articles.id > 0'); <add> }); <add> $results = $query->all(); <add> <add> // 4 records in the junction table. <add> $this->assertCount(4, $results); <add> } <ide> } <ide><path>tests/schema.php <ide> ], <ide> ], <ide> ], <add> [ <add> 'table' => 'articles_tags_binding_keys', <add> 'columns' => [ <add> 'article_id' => [ <add> 'type' => 'integer', <add> 'null' => false, <add> ], <add> 'tagname' => [ <add> 'type' => 'string', <add> 'null' => false, <add> ], <add> ], <add> 'constraints' => [ <add> 'unique_tag' => [ <add> 'type' => 'primary', <add> 'columns' => [ <add> 'article_id', <add> 'tagname', <add> ], <add> ], <add> ], <add> ], <ide> [ <ide> 'table' => 'profiles', <ide> 'columns' => [
4
Ruby
Ruby
add newline to failure message
c3b309e7957d38952a85b4d320ac73112d2733af
<ide><path>Library/Homebrew/requirements/java_requirement.rb <ide> def initialize(tags) <ide> def message <ide> version_string = " #{@version}" if @version <ide> <del> s = "Java#{version_string} is required to install this formula." <add> s = "Java#{version_string} is required to install this formula.\n" <ide> s += super <ide> s <ide> end
1
Javascript
Javascript
remove extra scene parameter from jsm version
447d9dd95897527b504ace0b173056a30f3fe59e
<ide><path>examples/jsm/renderers/CSS2DRenderer.js <ide> var CSS2DRenderer = function () { <ide> <ide> for ( var i = 0, l = object.children.length; i < l; i ++ ) { <ide> <del> renderObject( object.children[ i ], scene, camera ); <add> renderObject( object.children[ i ], camera ); <ide> <ide> } <ide>
1
Javascript
Javascript
handle dashes in keys in tohavedata matcher
2ef6d1be965bb7e579403b809331370849709b6c
<ide><path>vendor/jasmine-jquery.js <ide> var jQueryMatchers = { <ide> <ide> toHaveData: function(key, expectedValue) { <ide> if (this.actual instanceof HTMLElement) { <del> return hasProperty(this.actual.dataset[key], expectedValue) <add> var camelCaseKey <add> for (var part of key.split('-')) { <add> if (camelCaseKey) { <add> camelCaseKey += part[0].toUpperCase() + part.substring(1) <add> } else { <add> camelCaseKey = part <add> } <add> } <add> return hasProperty(this.actual.dataset[camelCaseKey], expectedValue) <ide> } else { <ide> return hasProperty(this.actual.data(key), expectedValue) <ide> }
1
Ruby
Ruby
replace capture2e with utils.popen
e8f78878cb81c4dec14f7e1498ead71892502072
<ide><path>Library/Homebrew/utils/livecheck.rb <ide> # frozen_string_literal: true <ide> <del>require "open3" <del> <ide> module Livecheck <del> def livecheck_formula_response(formula_name) <del> ohai "Checking livecheck formula : #{formula_name}" <del> command_args = ["brew", "livecheck", formula_name, "--quiet"] <add> def livecheck_formula_response(name) <add> ohai "Checking livecheck formula : #{name}" <ide> <del> response = Open3.capture2e(*command_args) <add> response = Utils.popen_read("brew", "livecheck", name, "--quiet").chomp <ide> parse_livecheck_response(response) <ide> end <ide> <ide><path>Library/Homebrew/utils/versions.rb <ide> # frozen_string_literal: true <ide> <del>require "open3" <ide> require "formula" <ide> <ide> module Versions <ide> def current_formula_version(formula_name) <ide> end <ide> <ide> def bump_formula_pr(formula_name, url) <del> command_args = ["brew", "bump-formula-pr", "--no-browse", <del> "--dry-run", formula_name, "--url=#{url}"] <add> response = Utils.popen_read("brew", "bump-formula-pr", "--no-browse", <add> "--dry-run", formula_name, "--url=#{url}").chomp <ide> <del> response = Open3.capture2e(*command_args) <ide> parse_formula_bump_response(response) <ide> end <ide>
2
Ruby
Ruby
add tests for credentials command
8a331566bfe05ccc85f25805693aa892cec449a5
<ide><path>railties/test/commands/credentials_test.rb <ide> class Rails::Command::CredentialsCommandTest < ActiveSupport::TestCase <ide> <ide> teardown { teardown_app } <ide> <add> test "edit without editor gives hint" do <add> assert_match "No $EDITOR to open credentials in", run_edit_command(editor: "") <add> end <add> <add> test "edit credentials" do <add> # Run twice to ensure credentials can be reread after first edit pass. <add> 2.times do <add> assert_match(/access_key_id: 123/, run_edit_command) <add> end <add> end <add> <add> test "show credentials" do <add> assert_match(/access_key_id: 123/, run_show_command) <add> end <add> <ide> test "edit command does not add master key to gitignore when already exist" do <ide> run_edit_command <ide> <ide> def run_edit_command(editor: "cat") <ide> rails "credentials:edit" <ide> end <ide> end <add> <add> def run_show_command <add> rails "credentials:show" <add> end <ide> end
1
Java
Java
fix checkstyle violations
385292dc41c62e96a083c41c9c686825db251d63
<ide><path>spring-core/src/main/java/org/springframework/core/io/support/PathMatchingResourcePatternResolver.java <ide> protected Set<Resource> doFindPathMatchingFileResources(Resource rootDirResource <ide> FileSystem fileSystem; <ide> try { <ide> fileSystem = FileSystems.getFileSystem(rootDirResource.getURI().resolve("/")); <del> } catch (Exception e) { <add> } <add> catch (Exception ex) { <ide> fileSystem = FileSystems.newFileSystem(rootDirResource.getURI().resolve("/"), Map.of(), <ide> ClassUtils.getDefaultClassLoader()); <ide> } <ide> protected Set<Resource> doFindPathMatchingFileResources(Resource rootDirResource <ide> if (getPathMatcher().match(patternPath.toString(), file.toString())) { <ide> try { <ide> result.add(new UrlResource(file.toUri())); <del> } catch (MalformedURLException e) { <add> } <add> catch (MalformedURLException ex) { <ide> // ignore <ide> } <ide> } <ide> }); <del> } catch (NoSuchFileException e) { <add> } <add> catch (NoSuchFileException ex) { <ide> // ignore <ide> } <ide> try { <ide> fileSystem.close(); <del> } catch (UnsupportedOperationException e) { <add> } <add> catch (UnsupportedOperationException ex) { <ide> // ignore <ide> } <ide> return result;
1
Python
Python
add a method to convert str to ascii in python 3
54b61fd5a090061b1016e1fee115c6808bbb06ed
<ide><path>glances/compat.py <ide> def mean(numbers): <ide> def to_ascii(s): <ide> """Convert the bytes string to a ASCII string <ide> Usefull to remove accent (diacritics)""" <del> return str(s, 'utf-8') <add> if isinstance(s, binary_type): <add> return s <add> return s.encode('ascii', 'ignore') <ide> <ide> def listitems(d): <ide> return list(d.items()) <ide> def to_ascii(s): <ide> Usefull to remove accent (diacritics)""" <ide> if isinstance(s, binary_type): <ide> return s <del> return unicodedata.normalize('NFKD', s).encode('ASCII', 'ignore') <add> return unicodedata.normalize('NFKD', s).encode('ascii', 'ignore') <ide> <ide> def listitems(d): <ide> return d.items()
1
Ruby
Ruby
add test for build.without --without-foo
02a1406a2e4c2d0506861248eb767d99f5ce4515
<ide><path>Library/Homebrew/rubocops/lines_cop.rb <ide> def audit_formula(_node, _class_node, _parent_class_node, body_node) <ide> problem "Don't negate 'build.without?': use 'build.with?'" <ide> end <ide> <del> # find_instance_method_call(body_node, :build, :without?) do |m| <del> # arg = parameters(m).first <del> # next unless match = regex_match_group(arg, %r{-?-?without-(.*)}) <del> # problem "Don't duplicate 'without': Use `build.without? \"#{match[1]}\"` to check for \"--without-#{match[1]}\"" <del> # end <del> # <add> find_instance_method_call(body_node, :build, :without?) do |m| <add> arg = parameters(m).first <add> next unless match = regex_match_group(arg, %r{-?-?without-(.*)}) <add> problem "Don't duplicate 'without': Use `build.without? \"#{match[1]}\"` to check for \"--without-#{match[1]}\"" <add> end <add> <ide> # find_instance_method_call(body_node, :build, :with?) do |m| <ide> # arg = parameters(m).first <ide> # next unless match = regex_match_group(arg, %r{-?-?with-(.*)}) <ide><path>Library/Homebrew/test/rubocops/lines_cop_spec.rb <ide> def post_install <ide> expect_offense(expected, actual) <ide> end <ide> end <add> <add> it "with negated build.with?" do <add> source = <<-EOS.undent <add> class Foo < Formula <add> desc "foo" <add> url 'http://example.com/foo-1.0.tgz' <add> def post_install <add> return if build.without? "--without-bar" <add> end <add> end <add> EOS <add> <add> expected_offenses = [{ message: "Don't duplicate 'without': Use `build.without? \"bar\"` to check for \"--without-bar\"", <add> severity: :convention, <add> line: 5, <add> column: 30, <add> source: source }] <add> <add> inspect_source(cop, source) <add> <add> expected_offenses.zip(cop.offenses).each do |expected, actual| <add> expect_offense(expected, actual) <add> end <add> end <ide> end <ide> def expect_offense(expected, actual) <ide> expect(actual.message).to eq(expected[:message])
2
Text
Text
add links to inline html table
dd023df135207086c86129aa1683ea4688f97f53
<ide><path>doc/api/crypto.md <ide> The following constants exported by `crypto.constants` apply to various uses of <ide> the `crypto`, `tls`, and `https` modules and are generally specific to OpenSSL. <ide> <ide> ### OpenSSL Options <del> <add><!--lint disable maximum-line-length--> <ide> <table> <ide> <tr> <ide> <th>Constant</th> <ide> the `crypto`, `tls`, and `https` modules and are generally specific to OpenSSL. <ide> <tr> <ide> <td><code>SSL_OP_ALL</code></td> <ide> <td>Applies multiple bug workarounds within OpenSSL. See <del> https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html for <del> detail.</td> <add> <a href="https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html">https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html</a> <add> for detail.</td> <ide> </tr> <ide> <tr> <ide> <td><code>SSL_OP_ALLOW_UNSAFE_LEGACY_RENEGOTIATION</code></td> <ide> <td>Allows legacy insecure renegotiation between OpenSSL and unpatched <ide> clients or servers. See <del> https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html.</td> <add> <a href="https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html">https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html</a>.</td> <ide> </tr> <ide> <tr> <ide> <td><code>SSL_OP_CIPHER_SERVER_PREFERENCE</code></td> <ide> <td>Attempts to use the server's preferences instead of the client's when <ide> selecting a cipher. Behavior depends on protocol version. See <del> https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html.</td> <add> <a href="https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html">https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html</a>.</td> <ide> </tr> <ide> <tr> <ide> <td><code>SSL_OP_CISCO_ANYCONNECT</code></td> <ide> the `crypto`, `tls`, and `https` modules and are generally specific to OpenSSL. <ide> </table> <ide> <ide> ### OpenSSL Engine Constants <add><!--lint enable maximum-line-length remark-lint--> <ide> <ide> <table> <ide> <tr>
1
Text
Text
add optional sudo to make install in readme
ab04a434761cf66d107481d58798f36d3cb49d46
<ide><path>README.md <ide> Prerequisites: <ide> ```text <ide> $ ./configure <ide> $ make <del>$ make install <add>$ [sudo] make install <ide> ``` <ide> <ide> If your Python binary is in a non-standard location or has a <ide> non-standard name, run the following instead: <ide> $ export PYTHON=/path/to/python <ide> $ $PYTHON ./configure <ide> $ make <del>$ make install <add>$ [sudo] make install <ide> ``` <ide> <ide> To run the tests:
1
Python
Python
update 2 mnist examples
ca23406974546c5214a3ec73eb903952c14ab2f0
<ide><path>examples/mnist_cnn.py <ide> import numpy as np <ide> np.random.seed(1337) # for reproducibility <ide> <add>import keras <ide> from keras.datasets import mnist <ide> from keras.models import Sequential <del>from keras.layers import Dense, Dropout, Activation, Flatten <del>from keras.layers import Convolution2D, MaxPooling2D <del>from keras.utils import np_utils <add>from keras.layers import Dense, Dropout, Flatten <add>from keras.layers import Conv2D, MaxPooling2D <ide> from keras import backend as K <ide> <ide> batch_size = 128 <del>nb_classes = 10 <del>nb_epoch = 12 <add>num_classes = 10 <add>num_epoch = 12 <ide> <ide> # input image dimensions <ide> img_rows, img_cols = 28, 28 <del># number of convolutional filters to use <del>nb_filters = 32 <del># size of pooling area for max pooling <del>pool_size = (2, 2) <del># convolution kernel size <del>kernel_size = (3, 3) <ide> <ide> # the data, shuffled and split between train and test sets <del>(X_train, y_train), (X_test, y_test) = mnist.load_data() <add>(x_train, y_train), (x_test, y_test) = mnist.load_data() <ide> <ide> if K.image_data_format() == 'channels_first': <del> X_train = X_train.reshape(X_train.shape[0], 1, img_rows, img_cols) <del> X_test = X_test.reshape(X_test.shape[0], 1, img_rows, img_cols) <add> x_train = x_train.reshape(x_train.shape[0], 1, img_rows, img_cols) <add> x_test = x_test.reshape(x_test.shape[0], 1, img_rows, img_cols) <ide> input_shape = (1, img_rows, img_cols) <ide> else: <del> X_train = X_train.reshape(X_train.shape[0], img_rows, img_cols, 1) <del> X_test = X_test.reshape(X_test.shape[0], img_rows, img_cols, 1) <add> x_train = x_train.reshape(x_train.shape[0], img_rows, img_cols, 1) <add> x_test = x_test.reshape(x_test.shape[0], img_rows, img_cols, 1) <ide> input_shape = (img_rows, img_cols, 1) <ide> <del>X_train = X_train.astype('float32') <del>X_test = X_test.astype('float32') <del>X_train /= 255 <del>X_test /= 255 <del>print('X_train shape:', X_train.shape) <del>print(X_train.shape[0], 'train samples') <del>print(X_test.shape[0], 'test samples') <add>x_train = x_train.astype('float32') <add>x_test = x_test.astype('float32') <add>x_train /= 255 <add>x_test /= 255 <add>print('x_train shape:', x_train.shape) <add>print(x_train.shape[0], 'train samples') <add>print(x_test.shape[0], 'test samples') <ide> <ide> # convert class vectors to binary class matrices <del>Y_train = np_utils.to_categorical(y_train, nb_classes) <del>Y_test = np_utils.to_categorical(y_test, nb_classes) <add>y_train = keras.utils.np_utils.to_categorical(y_train, num_classes) <add>y_test = keras.utils.np_utils.to_categorical(y_test, num_classes) <ide> <ide> model = Sequential() <del> <del>model.add(Convolution2D(nb_filters, kernel_size[0], kernel_size[1], <del> border_mode='valid', <del> input_shape=input_shape)) <del>model.add(Activation('relu')) <del>model.add(Convolution2D(nb_filters, kernel_size[0], kernel_size[1])) <del>model.add(Activation('relu')) <del>model.add(MaxPooling2D(pool_size=pool_size)) <add>model.add(Conv2D(32, kernel_size=(3, 3), <add> activation='relu', <add> input_shape=input_shape)) <add>model.add(Conv2D(64, (3, 3), activation='relu')) <add>model.add(MaxPooling2D(pool_size=(2, 2))) <ide> model.add(Dropout(0.25)) <del> <ide> model.add(Flatten()) <del>model.add(Dense(128)) <del>model.add(Activation('relu')) <add>model.add(Dense(128, activation='relu')) <ide> model.add(Dropout(0.5)) <del>model.add(Dense(nb_classes)) <del>model.add(Activation('softmax')) <add>model.add(Dense(num_classes, activation='softmax')) <ide> <del>model.compile(loss='categorical_crossentropy', <del> optimizer='adadelta', <add>model.compile(loss=keras.losses.categorical_crossentropy, <add> optimizer=keras.optimizers.Adadelta(), <ide> metrics=['accuracy']) <ide> <del>model.fit(X_train, Y_train, batch_size=batch_size, nb_epoch=nb_epoch, <del> verbose=1, validation_data=(X_test, Y_test)) <del>score = model.evaluate(X_test, Y_test, verbose=0) <del>print('Test score:', score[0]) <add>model.fit(x_train, y_train, batch_size=batch_size, num_epoch=num_epoch, <add> verbose=1, validation_data=(x_test, y_test)) <add>score = model.evaluate(x_test, y_test, verbose=0) <add>print('Test loss:', score[0]) <ide> print('Test accuracy:', score[1]) <ide><path>examples/mnist_mlp.py <ide> <ide> from keras.datasets import mnist <ide> from keras.models import Sequential <del>from keras.layers.core import Dense, Dropout, Activation <add>from keras.layers import Dense, Dropout, Activation <ide> from keras.optimizers import RMSprop <ide> from keras.utils import np_utils <ide> <ide> <ide> batch_size = 128 <del>nb_classes = 10 <del>nb_epoch = 20 <add>num_classes = 10 <add>num_epoch = 20 <ide> <ide> # the data, shuffled and split between train and test sets <del>(X_train, y_train), (X_test, y_test) = mnist.load_data() <add>(x_train, y_train), (x_test, y_test) = mnist.load_data() <ide> <del>X_train = X_train.reshape(60000, 784) <del>X_test = X_test.reshape(10000, 784) <del>X_train = X_train.astype('float32') <del>X_test = X_test.astype('float32') <del>X_train /= 255 <del>X_test /= 255 <del>print(X_train.shape[0], 'train samples') <del>print(X_test.shape[0], 'test samples') <add>x_train = x_train.reshape(60000, 784) <add>x_test = x_test.reshape(10000, 784) <add>x_train = x_train.astype('float32') <add>x_test = x_test.astype('float32') <add>x_train /= 255 <add>x_test /= 255 <add>print(x_train.shape[0], 'train samples') <add>print(x_test.shape[0], 'test samples') <ide> <ide> # convert class vectors to binary class matrices <del>Y_train = np_utils.to_categorical(y_train, nb_classes) <del>Y_test = np_utils.to_categorical(y_test, nb_classes) <add>y_train = np_utils.to_categorical(y_train, num_classes) <add>y_test = np_utils.to_categorical(y_test, num_classes) <ide> <ide> model = Sequential() <del>model.add(Dense(512, input_shape=(784,))) <del>model.add(Activation('relu')) <add>model.add(Dense(512, activation='relu', input_shape=(784,))) <ide> model.add(Dropout(0.2)) <del>model.add(Dense(512)) <del>model.add(Activation('relu')) <add>model.add(Dense(512, activation='relu')) <ide> model.add(Dropout(0.2)) <del>model.add(Dense(10)) <del>model.add(Activation('softmax')) <add>model.add(Dense(10, activation='softmax')) <ide> <ide> model.summary() <ide> <ide> model.compile(loss='categorical_crossentropy', <ide> optimizer=RMSprop(), <ide> metrics=['accuracy']) <ide> <del>history = model.fit(X_train, Y_train, <del> batch_size=batch_size, nb_epoch=nb_epoch, <del> verbose=1, validation_data=(X_test, Y_test)) <del>score = model.evaluate(X_test, Y_test, verbose=0) <del>print('Test score:', score[0]) <add>history = model.fit(x_train, y_train, <add> batch_size=batch_size, num_epoch=num_epoch, <add> verbose=1, validation_data=(x_test, y_test)) <add>score = model.evaluate(x_test, y_test, verbose=0) <add>print('Test loss:', score[0]) <ide> print('Test accuracy:', score[1])
2
PHP
PHP
add tests for cachemanager
fdcd963158587e9c69980bbf53ad8e1c34d70d9f
<ide><path>src/Illuminate/Cache/CacheManager.php <ide> protected function getPrefix(array $config) <ide> * Get the cache connection configuration. <ide> * <ide> * @param string $name <del> * @return array <add> * @return array|null <ide> */ <ide> protected function getConfig($name) <ide> { <ide><path>tests/Cache/CacheManagerTest.php <ide> <ide> use Illuminate\Cache\ArrayStore; <ide> use Illuminate\Cache\CacheManager; <add>use Illuminate\Config\Repository; <add>use Illuminate\Container\Container; <add>use InvalidArgumentException; <ide> use Mockery as m; <ide> use PHPUnit\Framework\TestCase; <ide> <ide> public function testForgetDriverForgets() <ide> $cacheManager->forgetDriver('forget'); <ide> $this->assertNull($cacheManager->store('forget')->get('foo')); <ide> } <add> <add> public function testThrowExceptionWhenUnknownDriverIsUsed() <add> { <add> $this->expectException(InvalidArgumentException::class); <add> $this->expectExceptionMessage('Driver [unknown_taxi_driver] is not supported.'); <add> <add> $userConfig = [ <add> 'cache' => [ <add> 'stores' => [ <add> 'my_store' => [ <add> 'driver' => 'unknown_taxi_driver', <add> ], <add> ], <add> ], <add> ]; <add> <add> $app = Container::getInstance(); <add> $app->bind('config', function () use ($userConfig) { <add> return new Repository($userConfig); <add> }); <add> <add> $cacheManager = new CacheManager($app); <add> <add> $cacheManager->store('my_store'); <add> } <add> <add> public function testThrowExceptionWhenUnknownStoreIsUsed() <add> { <add> $this->expectException(InvalidArgumentException::class); <add> $this->expectExceptionMessage('Cache store [alien_store] is not defined.'); <add> <add> $userConfig = [ <add> 'cache' => [ <add> 'stores' => [ <add> 'my_store' => [ <add> 'driver' => 'array', <add> ], <add> ], <add> ], <add> ]; <add> <add> $app = Container::getInstance(); <add> $app->bind('config', function () use ($userConfig) { <add> return new Repository($userConfig); <add> }); <add> <add> $cacheManager = new CacheManager($app); <add> <add> $cacheManager->store('alien_store'); <add> } <ide> }
2
Javascript
Javascript
remove unused require
d6e2abfd69301eb2bada951920db779c773dbea8
<ide><path>lib/dependencies/CommonJsRequireDependencyParserPlugin.js <ide> */ <ide> "use strict"; <ide> <del>const ConstDependency = require("./ConstDependency"); <ide> const CommonJsRequireDependency = require("./CommonJsRequireDependency"); <ide> const CommonJsRequireContextDependency = require("./CommonJsRequireContextDependency"); <ide> const RequireHeaderDependency = require("./RequireHeaderDependency");
1
PHP
PHP
fix code styling
7e7013222f98ff31db89fb732d047ce910c8753a
<ide><path>src/Illuminate/View/Compilers/BladeCompiler.php <ide> protected function compileOverwrite($expression) <ide> */ <ide> protected function compileUnless($expression) <ide> { <del> return "<?php if ( ! $expression): ?>"; <add> return "<?php if (! $expression): ?>"; <ide> } <ide> <ide> /** <ide><path>tests/View/ViewBladeCompilerTest.php <ide> public function testUnlessStatementsAreCompiled() <ide> $string = '@unless (name(foo(bar))) <ide> breeze <ide> @endunless'; <del> $expected = '<?php if ( ! (name(foo(bar)))): ?> <add> $expected = '<?php if (! (name(foo(bar)))): ?> <ide> breeze <ide> <?php endif; ?>'; <ide> $this->assertEquals($expected, $compiler->compileString($string));
2
Ruby
Ruby
prevent negative ids in output of #inspect
5bbce6fccc541b9941628cda4eda1f84c5a909ad
<ide><path>actionview/lib/action_view/template/resolver.rb <ide> def initialize <ide> end <ide> <ide> def inspect <del> "#<#{self.class.name}:0x#{(object_id << 1).to_s(16)} keys=#{@data.size} queries=#{@query_cache.size}>" <add> "#{to_s[0..-2]} keys=#{@data.size} queries=#{@query_cache.size}>" <ide> end <ide> <ide> # Cache the templates returned by the block
1
Text
Text
add quotes around the brew invocation
953a1bf20b6944ef14bd9f6e5f5bc37ed191eba8
<ide><path>docs/Shell-Completion.md <ide> To make Homebrew's completions available in `bash`, you must source the definiti <ide> <ide> ```sh <ide> if type brew &>/dev/null; then <del> HOMEBREW_PREFIX=$(brew --prefix) <add> HOMEBREW_PREFIX="$(brew --prefix)" <ide> if [[ -r "${HOMEBREW_PREFIX}/etc/profile.d/bash_completion.sh" ]]; then <ide> source "${HOMEBREW_PREFIX}/etc/profile.d/bash_completion.sh" <ide> else
1
Ruby
Ruby
allow methods outside the model
9f0f7ec2223ae60ee6e2eab5bc9f036a480e9f81
<ide><path>actionview/lib/action_view/helpers/form_helper.rb <ide> def apply_form_for_options!(record, object, options) #:nodoc: <ide> # <input type="text" name="post[title]" value="<the title of the post>"> <ide> # </form> <ide> # <add> # # Though the fields don't have to correspond to model attributes: <add> # <%= form_with model: Cat.new do |form| %> <add> # <%= form.text_field :cats_dont_have_gills %> <add> # <%= form.text_field :but_in_forms_they_can %> <add> # <% end %> <add> # # => <add> # <form action="/cats" method="post" data-remote="true"> <add> # <input type="text" name="cat[cats_dont_have_gills]"> <add> # <input type="text" name="cat[but_in_forms_they_can]"> <add> # </form> <add> # <ide> # The parameters in the forms are accessible in controllers according to <ide> # their name nesting. So inputs named +title+ and <tt>post[title]</tt> are <ide> # accessible as <tt>params[:title]</tt> and <tt>params[:post][:title]</tt> <ide> def apply_form_for_options!(record, object, options) #:nodoc: <ide> # form_with(**options.merge(builder: LabellingFormBuilder), &block) <ide> # end <ide> def form_with(model: nil, scope: nil, url: nil, format: nil, **options) <add> options[:allow_method_names_outside_object] = true <ide> options[:skip_default_ids] = true <ide> <ide> if model <ide> def to_model <ide> def initialize(object_name, object, template, options) <ide> @nested_child_index = {} <ide> @object_name, @object, @template, @options = object_name, object, template, options <del> @default_options = @options ? @options.slice(:index, :namespace, :skip_default_ids) : {} <add> @default_options = @options ? @options.slice(:index, :namespace, :skip_default_ids, :allow_method_names_outside_object) : {} <ide> <ide> convert_to_legacy_options(@options) <ide> <ide><path>actionview/lib/action_view/helpers/tags/base.rb <ide> def initialize(object_name, method_name, template_object, options = {}) <ide> @object_name.sub!(/\[\]$/, "") || @object_name.sub!(/\[\]\]$/, "]") <ide> @object = retrieve_object(options.delete(:object)) <ide> @skip_default_ids = options.delete(:skip_default_ids) <add> @allow_method_names_outside_object = options.delete(:allow_method_names_outside_object) <ide> @options = options <ide> @auto_index = Regexp.last_match ? retrieve_autoindex(Regexp.last_match.pre_match) : nil <ide> end <ide> def render <ide> private <ide> <ide> def value(object) <del> object.public_send @method_name if object <add> if @allow_method_names_outside_object <add> object.public_send @method_name if object && object.respond_to?(@method_name) <add> else <add> object.public_send @method_name if object <add> end <ide> end <ide> <ide> def value_before_type_cast(object) <ide> def add_default_name_and_id(options) <ide> <ide> def tag_name(multiple = false, index = nil) <ide> # a little duplication to construct less strings <del> if index <add> case <add> when @object_name.empty? <add> "#{sanitized_method_name}#{"[]" if multiple}" <add> when index <ide> "#{@object_name}[#{index}][#{sanitized_method_name}]#{"[]" if multiple}" <ide> else <ide> "#{@object_name}[#{sanitized_method_name}]#{"[]" if multiple}" <ide> def tag_name(multiple = false, index = nil) <ide> <ide> def tag_id(index = nil) <ide> # a little duplication to construct less strings <del> if index <add> case <add> when @object_name.empty? <add> sanitized_method_name.dup <add> when index <ide> "#{sanitized_object_name}_#{index}_#{sanitized_method_name}" <ide> else <ide> "#{sanitized_object_name}_#{sanitized_method_name}" <ide><path>actionview/test/template/form_helper/form_with_test.rb <ide> def test_form_with <ide> assert_dom_equal expected, output_buffer <ide> end <ide> <add> def test_form_with_only_url_on_create <add> form_with(url: "/posts") do |f| <add> concat f.label :title, "Label me" <add> concat f.text_field :title <add> end <add> <add> expected = whole_form("/posts") do <add> '<label for="title">Label me</label>' + <add> '<input type="text" name="title">' <add> end <add> <add> assert_dom_equal expected, output_buffer <add> end <add> <add> def test_form_with_only_url_on_update <add> form_with(url: "/posts/123") do |f| <add> concat f.label :title, 'Label me' <add> concat f.text_field :title <add> end <add> <add> expected = whole_form("/posts/123") do <add> '<label for="title">Label me</label>' + <add> '<input type="text" name="title">' <add> end <add> <add> assert_dom_equal expected, output_buffer <add> end <add> <add> def test_form_with_general_attributes <add> form_with(url: "/posts/123") do |f| <add> concat f.text_field :no_model_to_back_this_badboy <add> end <add> <add> expected = whole_form("/posts/123") do <add> '<input type="text" name="no_model_to_back_this_badboy">' <add> end <add> <add> assert_dom_equal expected, output_buffer <add> end <add> <add> def test_form_with_attribute_not_on_model <add> form_with(model: @post) do |f| <add> concat f.text_field :this_dont_exist_on_post <add> end <add> <add> expected = whole_form("/posts/123", method: :patch) do <add> '<input type="text" name="post[this_dont_exist_on_post]">' <add> end <add> <add> assert_dom_equal expected, output_buffer <add> end <add> <add> def test_form_with_doesnt_call_private_or_protected_properties_on_form_object_skipping_value <add> obj = Class.new do <add> private def private_property <add> "That would be great." <add> end <add> <add> protected def protected_property <add> "I believe you have my stapler." <add> end <add> end.new <add> <add> form_with(model: obj, scope: "other_name", url: "/", id: "edit-other-name") do |f| <add> assert_dom_equal '<input type="hidden" name="other_name[private_property]">', f.hidden_field(:private_property) <add> assert_dom_equal '<input type="hidden" name="other_name[protected_property]">', f.hidden_field(:protected_property) <add> end <add> end <add> <ide> def test_form_with_with_collection_radio_buttons <ide> post = Post.new <ide> def post.active; false; end <ide> def test_form_with_with_symbol_scope <ide> assert_dom_equal expected, output_buffer <ide> end <ide> <del> def test_form_tags_do_not_call_private_properties_on_form_object <del> obj = Class.new do <del> private def private_property <del> raise "This method should not be called." <del> end <del> end.new <del> <del> form_with(model: obj, scope: "other_name", url: "/", id: "edit-other-name") do |f| <del> assert_raise(NoMethodError) { f.hidden_field(:private_property) } <del> end <del> end <del> <ide> def test_form_with_with_method_as_part_of_html_options <ide> form_with(model: @post, url: "/", id: "create-post", html: { method: :delete }) do |f| <ide> concat f.text_field(:title)
3
Python
Python
attempt a port from
1eb7cc3017a6def34fb448781578888764d1e659
<ide><path>spacy/lang/ga/__init__.py <add># coding: utf8 <add>from __future__ import unicode_literals <add> <add>from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS <add>from .stop_words import STOP_WORDS <add> <add>from ..tokenizer_exceptions import BASE_EXCEPTIONS <add>from ...language import Language <add>from ...attrs import LANG <add>from ...util import update_exc <add> <add> <add>class Irish(Language): <add> lang = 'nb' <add> <add> class Defaults(Language.Defaults): <add> lex_attr_getters = dict(Language.Defaults.lex_attr_getters) <add> lex_attr_getters[LANG] = lambda text: 'ga' <add> <add> tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS) <add> stop_words = set(STOP_WORDS) <add> <add> <add>__all__ = ['Irish'] <ide><path>spacy/lang/ga/stop_words.py <add># encoding: utf8 <add>from __future__ import unicode_literals <add> <add> <add>STOP_WORDS = set(""" <add>a ach ag agus an aon ar arna as <add> <add>ba beirt bhúr <add> <add>caoga ceathair ceathrar chomh chuig chun cois céad cúig cúigear <add> <add>daichead dar de deich deichniúr den dhá do don dtí dá dár dó <add> <add>faoi faoin faoina faoinár fara fiche <add> <add>gach gan go gur <add> <add>haon hocht <add> <add>i iad idir in ina ins inár is <add> <add>le leis lena lenár <add> <add>mar mo muid mé <add> <add>na nach naoi naonúr ná ní níor nó nócha <add> <add>ocht ochtar ochtó os <add> <add>roimh <add> <add>sa seacht seachtar seachtó seasca seisear siad sibh sinn sna sé sí <add> <add>tar thar thú triúr trí trína trínár tríocha tú <add> <add>um <add> <add>ár <add> <add>é éis <add> <add>í <add> <add>ó ón óna ónár <add>""".split()) <ide><path>spacy/lang/ga/tokenizer_exceptions.py <add># encoding: utf8 <add>from __future__ import unicode_literals <add> <add>from ..symbols import ORTH, LEMMA, NORM <add> <add> <add>_exc = { <add> "'acha'n": [ <add> {ORTH: "'ach", LEMMA: "gach", NORM: "gach"}, <add> {ORTH: "a'n", LEMMA: "aon", NORM: "aon"}], <add> <add> "dem'": [ <add> {ORTH: "de", LEMMA: "de", NORM: "de"}, <add> {ORTH: "m'", LEMMA: "mo", NORM: "mo"}], <add> <add> "ded'": [ <add> {ORTH: "de", LEMMA: "de", NORM: "de"}, <add> {ORTH: "d'", LEMMA: "do", NORM: "do"}], <add> <add> "lem'": [ <add> {ORTH: "le", LEMMA: "le", NORM: "le"}, <add> {ORTH: "m'", LEMMA: "mo", NORM: "mo"}], <add> <add> "led'": [ <add> {ORTH: "le", LEMMA: "le", NORM: "le"}, <add> {ORTH: "d'", LEMMA: "mo", NORM: "do"}], <add> <add> "a.C.n.": [ <add> {ORTH: "a.", LEMMA: "ante"}, <add> {ORTH: "C.", LEMMA: "Christum"}, <add> {ORTH: "n.", LEMMA: "natum"}], <add> <add> "m.sh.": [ <add> {ORTH: "m.", LEMMA: "mar"}, <add> {ORTH: "sh.", LEMMA: "sampla"}], <add> <add> "M.F.": [ <add> {ORTH: "M.", LEMMA: "Meán"}, <add> {ORTH: "F.", LEMMA: "Fómhar"}], <add> <add> "M.Fómh.": [ <add> {ORTH: "M.", LEMMA: "Meán"}, <add> {ORTH: "Fómh.", LEMMA: "Fómhar"}], <add> <add> "R.C.": [ <add> {ORTH: "Rr.", LEMMA: "roimh"}, <add> {ORTH: "C.", LEMMA: "Críost"}], <add> <add> "r.Ch.": [ <add> {ORTH: "r.", LEMMA: "roimh"}, <add> {ORTH: "Ch.", LEMMA: "Críost"}], <add> <add> "r.Chr.": [ <add> {ORTH: "r.", LEMMA: "roimh"}, <add> {ORTH: "Chr.", LEMMA: "Críost"}], <add> <add> "R.Ch.": [ <add> {ORTH: "R.", LEMMA: "roimh"}, <add> {ORTH: "Ch.", LEMMA: "Críost"}], <add> <add> "R.Chr.": [ <add> {ORTH: "R.", LEMMA: "roimh"}, <add> {ORTH: "Chr.", LEMMA: "Críost"}], <add> <add> "⁊rl.": [ <add> {ORTH: "⁊", LEMMA: "agus"}, <add> {ORTH: "rl.", LEMMA: "araile"}], <add> <add> "srl.": [ <add> {ORTH: "s", LEMMA: "agus"}, <add> {ORTH: "rl.", LEMMA: "araile"}], <add> <add>} <add> <add>for exc_data in [ <add> {ORTH: "'gus", LEMMA: "agus", NORM: "agus"}, <add> {ORTH: "'ach", LEMMA: "gach", NORM: "gach"}, <add> {ORTH: "ao'", LEMMA: "aon", NORM: "aon"}, <add> {ORTH: "'niar", LEMMA: "aniar", NORM: "aniar"}, <add> {ORTH: "'níos", LEMMA: "aníos", NORM: "aníos"}, <add> {ORTH: "'ndiu", LEMMA: "inniu", NORM: "inniu"}, <add> {ORTH: "'nocht", LEMMA: "anocht", NORM: "anocht"}, <add> {ORTH: "m'", LEMMA: "mo"},, <add> {ORTH: "Aib.", LEMMA: "Aibreán"}, <add> {ORTH: "Ath.", LEMMA: "athair"}, <add> {ORTH: "Beal.", LEMMA: "Bealtaine"}, <add> {ORTH: "Co.", LEMMA: "contae"}, <add> {ORTH: "Ean.", LEMMA: "Eanáir"}, <add> {ORTH: "Feab.", LEMMA: "Feabhra"}, <add> {ORTH: "gCo.", LEMMA: "contae"}, <add> {ORTH: ".i.", LEMMA: "eadhon"}, <add> {ORTH: "lch.", LEMMA: "leathanach"}, <add> {ORTH: "Lch.", LEMMA: "leathanach"}, <add> {ORTH: "lgh.", LEMMA: "leathanach"}, <add> {ORTH: "Lgh.", LEMMA: "leathanach"}, <add> {ORTH: "Lún.", LEMMA: "Lúnasa"}, <add> {ORTH: "Már.", LEMMA: "Márta"}, <add> {ORTH: "Meith.", LEMMA: "Meitheamh"}, <add> {ORTH: "Noll.", LEMMA: "Nollaig"}, <add> {ORTH: "Samh.", LEMMA: "Samhain"}, <add> {ORTH: "tAth.", LEMMA: "athair"}, <add> {ORTH: "tUas.", LEMMA: "Uasal"}, <add> {ORTH: "teo.", LEMMA: "teoranta"}, <add> {ORTH: "Teo.", LEMMA: "teoranta"}, <add> {ORTH: "Uas.", LEMMA: "Uasal"}, <add> {ORTH: "uimh.", LEMMA: "uimhir"}, <add> {ORTH: "Uimh.", LEMMA: "uimhir"}]: <add> _exc[exc_data[ORTH]] = [dict(exc_data)], <add> <add>for orth in [ <add> "d'"]: <add> _exc[orth] = [{ORTH: orth}] <add> <add> <add>TOKENIZER_EXCEPTIONS = dict(_exc)
3
Python
Python
change duration_ms to duration"
1d4c7993a9c6dcacaca5074a80b1043e977c43fb
<ide><path>tools/test.py <ide> def HasRun(self, output): <ide> (duration.seconds + duration.days * 24 * 3600) * 10**6) / 10**6 <ide> <ide> logger.info(' ---') <del> logger.info(' duration: %d.%ds' % (total_seconds, duration.microseconds / 1000)) <add> logger.info(' duration_ms: %d.%d' % (total_seconds, duration.microseconds / 1000)) <ide> logger.info(' ...') <ide> <ide> def Done(self):
1
Text
Text
release notes for 0.10.0 chicken-hands
30e5f6274a7148c139aeafdb8098f63cbdbd5d33
<ide><path>CHANGELOG.md <add>- The Latest Stable Release: <a href="#0.9.19">0.9.19 canine-psychokinesis</a> <add>- The Latest Unstable Release: <a href="#0.10.0">0.10.0 chicken-hands</a> <add> <add> <ide> <a name="0.10.0"><a/> <del># 0.10.0 chicken-hands (in-progress) # <add># 0.10.0 chicken-hands (2011-09-02) # <add> <add>## Features: <add> <add>- complete rewrite of the Scope implementation with several API and semantic changes. Please see: <add> - [angular.scope API docs](http://docs-next.angularjs.org/#!/api/angular.scope) <add> - [scopes dev guide article](http://docs-next.angularjs.org/#!/guide/dev_guide.scopes) <add> - [scope.js source file](https://github.com/angular/angular.js/blob/master/src/Scope.js) <add> - breaking changes section of this changelog <add>- added event system to scopes (see [$on], [$emit] and [$broadcast]) <add>- added i18n and l10n support for date, currency and number filters see [i18n] docs for more info <add>- added localizable [ng:pluralize] widget <add>- added [ng:cloak] directive for hiding uncompiled templates <add> <add> <add>## Bug Fix: <add> <add>- make [ng:class] friendly towards other code adding/removing classes <add> ([commit](https://github.com/angular/angular.js/commit/2a8fe56997fddbad673748ce02abf649a709c4ca)) <add>- several [jqLite] bugfixes and improvements <add>- [ng:href], [ng:src] and friends now work properly when no expression is present in the attribute <add> value. <add> (Issue [#534](https://github.com/angular/angular.js/issues/534)) <add>- expose missing [lowercase], [uppercase] and [isDate] APIs. <add> <add> <add>## Docs: <add> <add>- many (but not all just yet) api docs were proof-read and improved <add> <add> <add>## Breaking Changes: <add> <add>- many scope related changes: <add> - $onEval is no more (use $watch with a fn as the only param if you really miss it) <add> - $eval without params doesn't trigger model mutation observations (use $apply/$digest instead) <add> - $digest propagates through the scope tree automatically (this is the desired behavior anyway) <add> - $watch various API changes <add> - scope is now the first argument passed into the $watch listener <add> - `this` in the $watch listener is undefined instead of current scope <add> - objects and arrays are watched and compared by equality and not just identity <add> - the initial execution of the $watch listener now executes asynchronously with respect to the <add> code registering it via $watch <add> - exceptionHandler argument is no more <add> - initRun argument is no more <add> - angular.scope does not create child scopes by taking parent as the first argument - use $new <add> instead <add> - scope.$set and scope.$get were removed, use direct property assignment instead or $eval <add>- $route.onChange was removed and replaced with $beforeRouteChange, $afterRouteChange and <add> $routeUpdate events that can be used together with the new $routeParams service <add>- `angular.equals()` now uses `===` instead of `==` when comparing primitives <ide> <ide> <ide> <ide> with the `$route` service <ide> <ide> <ide> <add>[lowercase]: http://docs.angularjs.org/#!/api/angular.lowercase <add>[uppercase]: http://docs.angularjs.org/#!/api/angular.uppercase <add>[isDate]: http://docs.angularjs.org/#!/api/angular.isDate <ide> [scope]: http://docs.angularjs.org/#!/api/angular.scope <ide> [compile]: http://docs.angularjs.org/#!/api/angular.compile <ide> [element]: http://docs.angularjs.org/#!/api/angular.element <ide> with the `$route` service <ide> [ng:readonly]: http://docs.angularjs.org/#!/api/angular.directive.ng:readonly <ide> [ng:show]: http://docs.angularjs.org/#!/api/angular.directive.ng:show <ide> [ng:hide]: http://docs.angularjs.org/#!/api/angular.directive.ng:hide <add>[ng:class]: http://docs.angularjs.org/#!/api/angular.directive.ng:class <add>[ng:src]: http://docs.angularjs.org/#!/api/angular.directive.ng:src <add>[ng:href]: http://docs.angularjs.org/#!/api/angular.directive.ng:href <ide> [$defer]: http://docs.angularjs.org/#!/api/angular.service.$defer <ide> [$cookies]: http://docs.angularjs.org/#!/api/angular.service.$cookies <ide> [$xhr]: http://docs.angularjs.org/#!/api/angular.service.$xhr <ide> with the `$route` service <ide> [jqLite]: http://docs.angularjs.org/#!/api/angular.element <ide> [angular.version]: http://docs.angularjs.org/#!/api/angular.version <ide> [Jstd Scenario Adapter]: https://github.com/angular/angular.js/blob/master/src/jstd-scenario-adapter/Adapter.js <add>[i18n]: http://docs-next.angularjs.org/#!/guide/dev_guide.i18n <add>[ng:pluralize]: http://docs-next.angularjs.org/#!/api/angular.widget.ng:pluralize <add>[ng:cloak]: http://docs-next.angularjs.org/#!/api/angular.directive.ng:cloak <add>[$on]: http://docs-next.angularjs.org/#!/api/angular.scope.$on <add>[$emit]: http://docs-next.angularjs.org/#!/api/angular.scope.$emit <add>[$broadcast]: http://docs-next.angularjs.org/#!/api/angular.scope.$broadcast
1
Javascript
Javascript
add support for title in schema
c76f7ec2f4bcca5e3361ff77fc022075368dad91
<ide><path>test/Schemas.lint.js <ide> describe("Schemas", () => { <ide> const allowedProperties = [ <ide> "definitions", <ide> "$ref", <del> "id", <add> "$id", <add> "title", <ide> "items", <ide> "properties", <ide> "additionalProperties", <ide><path>tooling/format-schemas.js <ide> const PROPERTIES = [ <ide> "$ref", <ide> "definitions", <ide> <add> "$id", <ide> "id", <add> "title", <ide> "description", <ide> "type", <ide>
2
Text
Text
update devtools post to note beta 2
3ed9581adbdc2bcf1bca0800050e7907624a83aa
<ide><path>docs/_posts/2015-08-03-new-react-devtools-beta.md <ide> Let us know what issues you run into <ide> [on GitHub](https://github.com/facebook/react-devtools/issues), and check out <ide> [the README](https://github.com/facebook/react-devtools/tree/devtools-next) <ide> for more info. <add> <add>## Update <add>*August 12, 2015* <add> <add>A second beta is out, with a number of bugfixes. It is also listed on the <add>[releases page](https://github.com/facebook/react-devtools/releases).
1
Text
Text
add missing link references
6c76de13c5849fb8273272490c55460e33c0c556
<ide><path>doc/api/assert.md <ide> Due to the confusing notation, it is recommended not to use a string as the <ide> second argument. This might lead to difficult-to-spot errors. <ide> <ide> [`Error.captureStackTrace`]: errors.html#errors_error_capturestacktrace_targetobject_constructoropt <add>[`Error`]: errors.html#errors_class_error <ide> [`Map`]: https://developer.mozilla.org/en/docs/Web/JavaScript/Reference/Global_Objects/Map <ide> [`Object.is()`]: https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Object/is <ide> [`RegExp`]: https://developer.mozilla.org/en-US/docs/Web/JavaScript/Guide/Regular_Expressions <ide><path>doc/api/errors.md <ide> Creation of a [`zlib`][] object failed due to incorrect configuration. <ide> [`hash.digest()`]: crypto.html#crypto_hash_digest_encoding <ide> [`hash.update()`]: crypto.html#crypto_hash_update_data_inputencoding <ide> [`readable._read()`]: stream.html#stream_readable_read_size_1 <add>[`server.close()`]: net.html#net_server_close_callback <ide> [`sign.sign()`]: crypto.html#crypto_sign_sign_privatekey_outputformat <ide> [`stream.pipe()`]: stream.html#stream_readable_pipe_destination_options <ide> [`stream.push()`]: stream.html#stream_readable_push_chunk_encoding
2
Javascript
Javascript
reduce the number of db calls for getsessionuser
892e6862ed4ffe4a9d01fec0045475b24e9be1da
<ide><path>api-server/common/models/user.js <ide> export default function(User) { <ide> }); <ide> <ide> User.prototype.getPoints$ = function getPoints$() { <add> if ( <add> Array.isArray(this.progressTimestamps) && <add> this.progressTimestamps.length <add> ) { <add> return Observable.of(this.progressTimestamps); <add> } <ide> const id = this.getId(); <ide> const filter = { <ide> where: { id }, <ide> export default function(User) { <ide> }); <ide> }; <ide> User.prototype.getCompletedChallenges$ = function getCompletedChallenges$() { <add> if ( <add> Array.isArray(this.completedChallenges) && <add> this.completedChallenges.length <add> ) { <add> return Observable.of(this.completedChallenges); <add> } <ide> const id = this.getId(); <ide> const filter = { <ide> where: { id }, <ide><path>api-server/server/boot_tests/challenge.test.js <ide> import { <ide> } from './fixtures'; <ide> <ide> describe('boot/challenge', () => { <del> xdescribe('backendChallengeCompleted'); <add> xdescribe('backendChallengeCompleted', () => {}); <ide> <ide> describe('buildUserUpdate', () => { <ide> it('returns an Object with a nested "completedChallenges" property', () => { <ide> describe('boot/challenge', () => { <ide> }); <ide> }); <ide> <del> xdescribe('modernChallengeCompleted'); <add> xdescribe('modernChallengeCompleted', () => {}); <ide> <del> xdescribe('projectCompleted'); <add> xdescribe('projectCompleted', () => {}); <ide> <ide> describe('redirectToCurrentChallenge', () => { <ide> const mockHomeLocation = 'https://www.example.com'; <ide><path>api-server/server/utils/user-stats.js <ide> export function getUserById(id, User = loopback.getModelByType('User')) { <ide> if (err || isEmpty(instance)) { <ide> return reject(err || 'No user instance found'); <ide> } <del> instance.points = <del> (instance.progressTimestamps && instance.progressTimestamps.length) || <del> 1; <add> <ide> let completedChallengeCount = 0; <ide> let completedProjectCount = 0; <ide> if ('completedChallenges' in instance) { <ide> export function getUserById(id, User = loopback.getModelByType('User')) { <ide> } <ide> }); <ide> } <add> <ide> instance.completedChallengeCount = completedChallengeCount; <ide> instance.completedProjectCount = completedProjectCount; <ide> instance.completedCertCount = getCompletedCertCount(instance); <ide> instance.completedLegacyCertCount = getLegacyCertCount(instance); <del> instance.completedChallenges = []; <del> delete instance.progressTimestamps; <add> instance.points = <add> (instance.progressTimestamps && instance.progressTimestamps.length) || <add> 1; <ide> return resolve(instance); <ide> }) <ide> ); <ide><path>api-server/server/utils/user-stats.test.js <ide> describe('user stats', () => { <ide> .then(user => { <ide> expect(user).toEqual(mockUser); <ide> <del> // fields added or removed by getUserById <del> expect(user).not.toHaveProperty('progressTimestamps'); <add> expect(user).toHaveProperty('progressTimestamps'); <ide> expect(user).toHaveProperty('completedChallengeCount'); <ide> expect(user).toHaveProperty('completedProjectCount'); <ide> expect(user).toHaveProperty('completedCertCount'); <ide> expect(user).toHaveProperty('completedLegacyCertCount'); <del> expect(user.completedChallenges).toEqual([]); <add> expect(user).toHaveProperty('completedChallenges'); <ide> }) <ide> .then(done) <ide> .catch(done);
4
Javascript
Javascript
add error logs in __debug__
96a7adf49e20197f802813224362470c66a38d20
<ide><path>src/devtools/store.js <ide> export default class Store extends EventEmitter { <ide> <ide> i = i + 3; <ide> <add> if (__DEBUG__) { <add> if (this._idToElement.has(id)) { <add> console.error( <add> 'Store already contains fiber ' + id + '. This is a bug.' <add> ); <add> } <add> } <add> <ide> if (type === ElementTypeRoot) { <ide> if (__DEBUG__) { <ide> debug('Add', `new root fiber ${id}`); <ide> export default class Store extends EventEmitter { <ide> case TREE_OPERATION_REMOVE: <ide> id = ((operations[i + 1]: any): number); <ide> <add> if (__DEBUG__) { <add> if (!this._idToElement.has(id)) { <add> console.error( <add> 'Store does not contain fiber ' + id + '. This is a bug.' <add> ); <add> } <add> } <add> <ide> i = i + 2; <ide> <ide> element = ((this._idToElement.get(id): any): Element);
1
Java
Java
update javadoc on uricomponentsbuilder.query
dacbf4cb34d3b2cf0747daa16903c91d696655a4
<ide><path>spring-web/src/main/java/org/springframework/web/util/UriComponentsBuilder.java <ide> public UriComponentsBuilder pathSegment(String... pathSegments) throws IllegalAr <ide> * be parsed unambiguously. Such values should be substituted for URI <ide> * variables to enable correct parsing: <ide> * <pre class="code"> <del> * String uriString = &quot;/hotels/42?filter={value}&quot;; <del> * UriComponentsBuilder.fromUriString(uriString).buildAndExpand(&quot;hot&amp;cold&quot;); <add> * UriComponentsBuilder.fromUriString(&quot;/hotels/42&quot;) <add> * .query(&quot;filter={value}&quot;) <add> * .buildAndExpand(&quot;hot&amp;cold&quot;); <ide> * </pre> <ide> * @param query the query string <ide> * @return this UriComponentsBuilder
1
Ruby
Ruby
remove unused method
3c100cfe8ed5a875b0bbdc8fa4e8f2b0cbf78676
<ide><path>activerecord/test/models/developer.rb <ide> def find_least_recent <ide> def log=(message) <ide> audit_logs.build :message => message <ide> end <del> <del> def self.all_johns <del> self.with_exclusive_scope :find => where(:name => 'John') do <del> self.all <del> end <del> end <ide> end <ide> <ide> class AuditLog < ActiveRecord::Base
1
Ruby
Ruby
apply suggestions from code review
bdb64aa178c1531661d74f58cd2435c30e029272
<ide><path>Library/Homebrew/formula_installer.rb <ide> def puts_requirement_messages <ide> end <ide> <ide> def forbidden_license_check <del> forbidden_licenses = (Homebrew::EnvConfig.forbidden_licenses || "").split(" ") <add> forbidden_licenses = Homebrew::EnvConfig.forbidden_licenses.to_s.split(" ") <ide> return if forbidden_licenses.blank? <ide> <ide> if forbidden_licenses.include? formula.license
1
Ruby
Ruby
stop test resetting global state
e2b6919223cd24bc8e64b4b5230733c0d8f8e6fc
<ide><path>activerecord/test/cases/yaml_serialization_test.rb <ide> class YamlSerializationTest < ActiveRecord::TestCase <ide> fixtures :topics <ide> <ide> def test_to_yaml_with_time_with_zone_should_not_raise_exception <add> tz = Time.zone <ide> Time.zone = ActiveSupport::TimeZone["Pacific Time (US & Canada)"] <ide> ActiveRecord::Base.time_zone_aware_attributes = true <add> <ide> topic = Topic.new(:written_on => DateTime.now) <ide> assert_nothing_raised { topic.to_yaml } <add> <add> ensure <add> Time.zone = tz <add> ActiveRecord::Base.time_zone_aware_attributes = false <ide> end <ide> <ide> def test_roundtrip
1
Javascript
Javascript
use a domain to catch async errors safely
c0721bcd66829356950b58cc532d6e3d8bbfc641
<ide><path>lib/repl.js <ide> var fs = require('fs'); <ide> var rl = require('readline'); <ide> var Console = require('console').Console; <ide> var EventEmitter = require('events').EventEmitter; <add>var domain = require('domain'); <ide> <ide> // If obj.hasOwnProperty has been overridden, then calling <ide> // obj.hasOwnProperty(prop) will break. <ide> function REPLServer(prompt, stream, eval_, useGlobal, ignoreUndefined) { <ide> <ide> EventEmitter.call(this); <ide> <del> var options, input, output; <add> var options, input, output, dom; <ide> if (typeof prompt == 'object') { <ide> // an options object was given <ide> options = prompt; <ide> function REPLServer(prompt, stream, eval_, useGlobal, ignoreUndefined) { <ide> useGlobal = options.useGlobal; <ide> ignoreUndefined = options.ignoreUndefined; <ide> prompt = options.prompt; <add> dom = options.domain; <ide> } else if (typeof prompt != 'string') { <ide> throw new Error('An options Object, or a prompt String are required'); <ide> } else { <ide> function REPLServer(prompt, stream, eval_, useGlobal, ignoreUndefined) { <ide> <ide> var self = this; <ide> <add> self._domain = dom || domain.create(); <add> <ide> self.useGlobal = !!useGlobal; <ide> self.ignoreUndefined = !!ignoreUndefined; <ide> <del> self.eval = eval_ || function(code, context, file, cb) { <add> eval_ = eval_ || defaultEval; <add> <add> function defaultEval(code, context, file, cb) { <ide> var err, result; <ide> try { <ide> if (self.useGlobal) { <ide> function REPLServer(prompt, stream, eval_, useGlobal, ignoreUndefined) { <ide> } catch (e) { <ide> err = e; <ide> } <del> if (err && process.domain) { <add> if (err && process.domain && !isSyntaxError(err)) { <ide> process.domain.emit('error', err); <ide> process.domain.exit(); <ide> } <ide> else { <ide> cb(err, result); <ide> } <del> }; <add> } <add> <add> self.eval = self._domain.bind(eval_); <add> <add> self._domain.on('error', function(e) { <add> self.outputStream.write((e.stack || e) + '\n'); <add> self.bufferedCommand = ''; <add> self.displayPrompt(); <add> }); <ide> <ide> if (!input && !output) { <ide> // legacy API, passing a 'stream'/'socket' option <ide> function REPLServer(prompt, stream, eval_, useGlobal, ignoreUndefined) { <ide> self.displayPrompt(); <ide> return; <ide> } else if (e) { <del> self.outputStream.write((e.stack || e) + '\n'); <add> self._domain.emit('error', e); <ide> } <ide> <ide> // Clear buffer if no SyntaxErrors <ide><path>test/simple/test-repl-domain.js <ide> // OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE <ide> // USE OR OTHER DEALINGS IN THE SOFTWARE. <ide> <add>var assert = require('assert'); <add>var common = require('../common.js'); <add> <ide> var util = require('util'); <ide> var repl = require('repl'); <ide> <ide><path>test/simple/test-repl-options.js <ide> assert.equal(r2.rli.terminal, false); <ide> assert.equal(r2.useColors, true); <ide> assert.equal(r2.useGlobal, true); <ide> assert.equal(r2.ignoreUndefined, true); <del>assert.equal(r2.eval, evaler); <ide> assert.equal(r2.writer, writer); <ide><path>test/simple/test-repl-timeout-throw.js <add>// Copyright Joyent, Inc. and other Node contributors. <add>// <add>// Permission is hereby granted, free of charge, to any person obtaining a <add>// copy of this software and associated documentation files (the <add>// "Software"), to deal in the Software without restriction, including <add>// without limitation the rights to use, copy, modify, merge, publish, <add>// distribute, sublicense, and/or sell copies of the Software, and to permit <add>// persons to whom the Software is furnished to do so, subject to the <add>// following conditions: <add>// <add>// The above copyright notice and this permission notice shall be included <add>// in all copies or substantial portions of the Software. <add>// <add>// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS <add>// OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF <add>// MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN <add>// NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, <add>// DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR <add>// OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE <add>// USE OR OTHER DEALINGS IN THE SOFTWARE. <add> <add>var assert = require('assert'); <add>var common = require('../common.js'); <add> <add>var spawn = require('child_process').spawn; <add> <add>var child = spawn(process.execPath, [ '-i' ], { <add> stdio: [null, null, 2] <add>}); <add> <add>var stdout = ''; <add>child.stdout.setEncoding('utf8'); <add>child.stdout.on('data', function(c) { <add> process.stdout.write(c); <add> stdout += c; <add>}); <add> <add>child.stdin.write = function(original) { return function(c) { <add> process.stderr.write(c); <add> return original.call(child.stdin, c); <add>}}(child.stdin.write); <add> <add>child.stdout.once('data', function() { <add> child.stdin.write('var throws = 0;'); <add> child.stdin.write('process.on("exit",function(){console.log(throws)});'); <add> child.stdin.write('function thrower(){console.log("THROW",throws++);XXX};'); <add> child.stdin.write('setTimeout(thrower);""\n'); <add> <add> setTimeout(fsTest, 50); <add> function fsTest() { <add> var f = JSON.stringify(__filename); <add> child.stdin.write('fs.readFile(' + f + ', thrower);\n'); <add> setTimeout(eeTest, 50); <add> } <add> <add> function eeTest() { <add> child.stdin.write('setTimeout(function() {\n' + <add> ' var events = require("events");\n' + <add> ' var e = new events.EventEmitter;\n' + <add> ' process.nextTick(function() {\n' + <add> ' e.on("x", thrower);\n' + <add> ' setTimeout(function() {\n' + <add> ' e.emit("x");\n' + <add> ' });\n' + <add> ' });\n' + <add> '});"";\n'); <add> <add> setTimeout(child.stdin.end.bind(child.stdin), 50); <add> } <add>}); <add> <add>child.on('exit', function(c) { <add> assert(!c); <add> // make sure we got 3 throws, in the end. <add> var lastLine = stdout.trim().split(/\r?\n/).pop(); <add> assert.equal(lastLine, '> 3'); <add>});
4
Javascript
Javascript
add missing semicolon
debf96c742934543e210c48c7c11199aa614e381
<ide><path>d3.layout.js <ide> d3.layout.chord = function() { <ide> if (source.value || target.value) { <ide> chords.push(source.value < target.value <ide> ? {source: target, target: source} <del> : {source: source, target: target}) <add> : {source: source, target: target}); <ide> } <ide> } <ide> } <ide><path>src/layout/chord.js <ide> d3.layout.chord = function() { <ide> if (source.value || target.value) { <ide> chords.push(source.value < target.value <ide> ? {source: target, target: source} <del> : {source: source, target: target}) <add> : {source: source, target: target}); <ide> } <ide> } <ide> }
2
Javascript
Javascript
clarify the comparator parameter
b9d2b30808db367af1c31fa0fa6cef8f1bc086d3
<ide><path>src/ng/filter/filter.js <ide> * The final result is an array of those elements that the predicate returned true for. <ide> * <ide> * @param {function(actual, expected)|true|false} [comparator] Comparator which is used in <del> * determining if the expected value (from the filter expression) and actual value (from <del> * the object in the array) should be considered a match. <add> * determining if values retrieved using `expression` (when it is not a function) should be <add> * considered a match based on the the expected value (from the filter expression) and actual <add> * value (from the object in the array). <ide> * <ide> * Can be one of: <ide> *
1
PHP
PHP
fix cookie sessions
a6c638b91c250a38f42536ad7123a45a2d8a9529
<ide><path>src/Illuminate/Session/Middleware/StartSession.php <ide> use Illuminate\Session\SessionManager; <ide> use Illuminate\Session\SessionInterface; <ide> use Symfony\Component\HttpFoundation\Cookie; <add>use Illuminate\Session\CookieSessionHandler; <ide> use Symfony\Component\HttpFoundation\Request; <ide> use Symfony\Component\HttpFoundation\Response; <del>use Illuminate\Contracts\Routing\Middleware as MiddlewareContract; <ide> use Illuminate\Contracts\Routing\TerminableMiddleware; <add>use Illuminate\Contracts\Routing\Middleware as MiddlewareContract; <ide> <ide> class StartSession implements MiddlewareContract, TerminableMiddleware { <ide> <ide> public function handle($request, Closure $next) <ide> */ <ide> public function terminate($request, $response) <ide> { <del> if ($this->sessionConfigured()) <add> if ($this->sessionConfigured() && ! $this->usingCookieSessions()) <ide> { <ide> $this->manager->driver()->save(); <ide> } <ide> protected function configHitsLottery(array $config) <ide> */ <ide> protected function addCookieToResponse(Response $response, SessionInterface $session) <ide> { <add> if ($this->usingCookieSessions()) <add> { <add> $this->manager->driver()->save(); <add> } <add> <ide> if ($this->sessionIsPersistent($config = $this->manager->getSessionConfig())) <ide> { <ide> $response->headers->setCookie(new Cookie( <ide> protected function sessionIsPersistent(array $config = null) <ide> return ! in_array($config['driver'], array(null, 'array')); <ide> } <ide> <add> /** <add> * Determine if the session is using cookie sessions. <add> * <add> * @return bool <add> */ <add> protected function usingCookieSessions() <add> { <add> if ( ! $this->sessionConfigured()) return false; <add> <add> return $this->manager->driver()->getHandler() instanceof CookieSessionHandler; <add> } <add> <ide> }
1
PHP
PHP
fix cs error
c3508f3b9c3c601df2fc63e3e6cf25a1b57a3f42
<ide><path>src/Error/BaseErrorHandler.php <ide> namespace Cake\Error; <ide> <ide> use Cake\Core\Configure; <add>use Cake\Error\PHP7ErrorException; <ide> use Cake\Log\Log; <ide> use Cake\Routing\Router; <del>use Exception; <ide> use Error; <del>use Cake\Error\PHP7ErrorException; <add>use Exception; <ide> <ide> /** <ide> * Base error handler that provides logic common to the CLI + web
1
Javascript
Javascript
keep stdio after exit
9e84a26cb3d390a6ce051b75c892094fff233f92
<ide><path>lib/internal/worker.js <ide> class Worker extends EventEmitter { <ide> this[kPublicPort] = null; <ide> <ide> const { stdout, stderr } = this[kParentSideStdio]; <del> this[kParentSideStdio] = null; <ide> <ide> if (!stdout._readableState.ended) { <ide> debug(`[${threadId}] explicitly closes stdout for ${this.threadId}`); <ide> class Worker extends EventEmitter { <ide> } <ide> <ide> get stdin() { <del> if (this[kParentSideStdio] === null) return null; <del> <ide> return this[kParentSideStdio].stdin; <ide> } <ide> <ide> get stdout() { <del> if (this[kParentSideStdio] === null) return null; <del> <ide> return this[kParentSideStdio].stdout; <ide> } <ide> <ide> get stderr() { <del> if (this[kParentSideStdio] === null) return null; <del> <ide> return this[kParentSideStdio].stderr; <ide> } <ide> } <ide><path>test/parallel/test-worker-safe-getters.js <ide> if (isMainThread) { <ide> stderr: true <ide> }); <ide> <add> const { stdin, stdout, stderr } = w; <add> <ide> w.on('exit', common.mustCall((code) => { <ide> assert.strictEqual(code, 0); <ide> <ide> if (isMainThread) { <ide> w.ref(); <ide> w.unref(); <ide> <del> // Although not browser specific, probably wise to <del> // make sure the stream getters don't throw either. <del> w.stdin; <del> w.stdout; <del> w.stderr; <del> <ide> // Sanity check. <ide> assert.strictEqual(w.threadId, -1); <del> assert.strictEqual(w.stdin, null); <del> assert.strictEqual(w.stdout, null); <del> assert.strictEqual(w.stderr, null); <add> assert.strictEqual(w.stdin, stdin); <add> assert.strictEqual(w.stdout, stdout); <add> assert.strictEqual(w.stderr, stderr); <ide> })); <ide> } else { <ide> process.exit(0);
2
Python
Python
compress target assigner
4f135c700383fa41c239c00ac1255b5b4eae4b27
<ide><path>research/object_detection/core/target_assigner.py <ide> def assign(self, <ide> reg_targets = self._create_regression_targets(anchors, <ide> groundtruth_boxes, <ide> match) <del> cls_targets = self._create_classification_targets(groundtruth_labels, <del> unmatched_class_label, <del> match) <del> reg_weights = self._create_regression_weights(match, groundtruth_weights) <add> cls_targets = match.gather_based_on_match( <add> groundtruth_labels, <add> unmatched_value=unmatched_class_label, <add> ignored_value=unmatched_class_label) <add> reg_weights = match.gather_based_on_match(groundtruth_weights, <add> ignored_value=0., <add> unmatched_value=0.) <add> cls_weights = match.gather_based_on_match( <add> groundtruth_weights, <add> ignored_value=0., <add> unmatched_value=self._negative_class_weight) <ide> <del> cls_weights = self._create_classification_weights(match, <del> groundtruth_weights) <ide> # convert cls_weights from per-anchor to per-class. <ide> class_label_shape = tf.shape(cls_targets)[1:] <ide> weights_shape = tf.shape(cls_weights) <ide> def _create_regression_targets(self, anchors, groundtruth_boxes, match): <ide> <ide> # Zero out the unmatched and ignored regression targets. <ide> unmatched_ignored_reg_targets = tf.tile( <del> self._default_regression_target(), [match_results_shape[0], 1]) <add> tf.constant([4 * [0]], tf.float32), [match_results_shape[0], 1]) <ide> matched_anchors_mask = match.matched_column_indicator() <ide> reg_targets = tf.where(matched_anchors_mask, <ide> matched_reg_targets, <ide> unmatched_ignored_reg_targets) <ide> return reg_targets <del> <del> def _default_regression_target(self): <del> """Returns the default target for anchors to regress to. <del> <del> Default regression targets are set to zero (though in <del> this implementation what these targets are set to should <del> not matter as the regression weight of any box set to <del> regress to the default target is zero). <del> <del> Returns: <del> default_target: a float32 tensor with shape [1, box_code_dimension] <del> """ <del> return tf.constant([4 * [0]], tf.float32) <del> <del> def _create_classification_targets(self, groundtruth_labels, <del> unmatched_class_label, match): <del> """Create classification targets for each anchor. <del> <del> Assign a classification target of for each anchor to the matching <del> groundtruth label that is provided by match. Anchors that are not matched <del> to anything are given the target self._unmatched_cls_target <del> <del> Args: <del> groundtruth_labels: a tensor of shape [num_gt_boxes, d_1, ... d_k] <del> with labels for each of the ground_truth boxes. The subshape <del> [d_1, ... d_k] can be empty (corresponding to scalar labels). <del> unmatched_class_label: a float32 tensor with shape [d_1, d_2, ..., d_k] <del> which is consistent with the classification target for each <del> anchor (and can be empty for scalar targets). This shape must thus be <del> compatible with the groundtruth labels that are passed to the "assign" <del> function (which have shape [num_gt_boxes, d_1, d_2, ..., d_k]). <del> match: a matcher.Match object that provides a matching between anchors <del> and groundtruth boxes. <del> <del> Returns: <del> a float32 tensor with shape [num_anchors, d_1, d_2 ... d_k], where the <del> subshape [d_1, ..., d_k] is compatible with groundtruth_labels which has <del> shape [num_gt_boxes, d_1, d_2, ... d_k]. <del> """ <del> return match.gather_based_on_match( <del> groundtruth_labels, <del> unmatched_value=unmatched_class_label, <del> ignored_value=unmatched_class_label) <del> <del> def _create_regression_weights(self, match, groundtruth_weights): <del> """Set regression weight for each anchor. <del> <del> Only positive anchors are set to contribute to the regression loss, so this <del> method returns a weight of 1 for every positive anchor and 0 for every <del> negative anchor. <del> <del> Args: <del> match: a matcher.Match object that provides a matching between anchors <del> and groundtruth boxes. <del> groundtruth_weights: a float tensor of shape [M] indicating the weight to <del> assign to all anchors match to a particular groundtruth box. <del> <del> Returns: <del> a float32 tensor with shape [num_anchors] representing regression weights. <del> """ <del> return match.gather_based_on_match( <del> groundtruth_weights, ignored_value=0., unmatched_value=0.) <del> <del> def _create_classification_weights(self, <del> match, <del> groundtruth_weights): <del> """Create classification weights for each anchor. <del> <del> Positive (matched) anchors are associated with a weight of <del> positive_class_weight and negative (unmatched) anchors are associated with <del> a weight of negative_class_weight. When anchors are ignored, weights are set <del> to zero. By default, both positive/negative weights are set to 1.0, <del> but they can be adjusted to handle class imbalance (which is almost always <del> the case in object detection). <del> <del> Args: <del> match: a matcher.Match object that provides a matching between anchors <del> and groundtruth boxes. <del> groundtruth_weights: a float tensor of shape [M] indicating the weight to <del> assign to all anchors match to a particular groundtruth box. <del> <del> Returns: <del> a float32 tensor with shape [num_anchors] representing classification <del> weights. <del> """ <del> return match.gather_based_on_match( <del> groundtruth_weights, <del> ignored_value=0., <del> unmatched_value=self._negative_class_weight)
1
PHP
PHP
remove proxy for money temporarily
1f394c1fa52439e44a18d59414eff009cb682a27
<ide><path>src/Validation/Validator.php <ide> public function integer($field, $message = null, $when = null) <ide> ]); <ide> } <ide> <del> /** <del> * Add a validation rule to ensure a field is a money value. <del> * <del> * @param string $field The field you want to apply the rule to. <del> * @param string $symbolPosition The position of the currency symbol <del> * @param string|null $message The error message when the rule fails. <del> * @param string|callable|null $when Either 'create' or 'update' or a callable that returns <del> * true when the validation rule should be applied. <del> * @see \Cake\Validation\Validation::money() <del> * @return $this <del> */ <del> public function money($field, $symbolPosition = 'left', $message = null, $when = null) <del> { <del> $extra = array_filter(['on' => $when, 'message' => $message]); <del> return $this->add($field, 'money', $extra + [ <del> 'rule' => ['money', $symbolPosition] <del> ]); <del> } <del> <ide> /** <ide> * Add a validation rule for a multiple select. Comparison is case sensitive by default. <ide> * <ide><path>tests/TestCase/Validation/ValidatorTest.php <ide> public function testInteger() <ide> $this->assertNotEmpty($validator->errors(['username' => 'not integer'])); <ide> } <ide> <del> /** <del> * Tests the money proxy method <del> * <del> * @return void <del> */ <del> public function testMoney() <del> { <del> $validator = new Validator(); <del> $this->assertProxyMethod($validator, 'money', 'left', ['left']); <del> $this->assertNotEmpty($validator->errors(['username' => 'not money'])); <del> } <del> <ide> /** <ide> * Tests the multiple proxy method <ide> *
2
Text
Text
add changelog entry for [ci skip]
c087cfc6717335c1cd4b8e438004cedac2dfa492
<ide><path>activerecord/CHANGELOG.md <add>* Use `UPDATE` rather than `SET` when enabling the `standard_conforming_strings` <add> setting as this allows us to avoid disabling errors on the PostgreSQL connection. <add> The former behavior would cause problems when using a connection pooling tool like <add> PgBouncer because it's not guaranteed to have the same connection between calls to <add> `execute` and it could leave the connection with errors disabled. <add> <add> Fixes #22101. <add> <add> *Harry Marr* <add> <ide> * Set `scope.reordering_value` to `true` if :reordering values are specified. <ide> <ide> Fixes #21886.
1
Text
Text
add ayase-252 to collaborators
3e8eda25319e8dc682de4f8028fae35da916d996
<ide><path>README.md <ide> For information about the governance of the Node.js project, see <ide> **Anatoli Papirovski** &lt;apapirovski@mac.com&gt; (he/him) <ide> * [AshCripps](https://github.com/AshCripps) - <ide> **Ash Cripps** &lt;acripps@redhat.com&gt; <add>* [Ayase-252](https://github.com/Ayase-252) - <add>**Qingyu Deng** &lt;i@ayase-lab.com&gt; <ide> * [bcoe](https://github.com/bcoe) - <ide> **Ben Coe** &lt;bencoe@gmail.com&gt; (he/him) <ide> * [bengl](https://github.com/bengl) -
1
Ruby
Ruby
recommend casks where available
7afe1ed67aa2509791705cf6d54ecfe771c61862
<ide><path>Library/Homebrew/requirements.rb <ide> def message; <ide> <<-EOS.undent <ide> A LaTeX distribution is required for Homebrew to install this formula. <ide> <del> You can install MacTeX distribution from: <add> You can install MacTeX distribution with: <add> brew cask install mactex <add> <add> Or from: <ide> http://www.tug.org/mactex/ <ide> <ide> Make sure that "/usr/texbin", or the location you installed it to, is in <ide> def message <ide> <<-EOS.undent <ide> Java#{version_string} is required to install this formula. <ide> <del> You can install Java from: <del> http://www.oracle.com/technetwork/java/javase/downloads/index.html <add> You can install the Java Development Kit (JDK) with: <add> brew cask install java <ide> <del> Make sure you install both the JRE and JDK. <add> Or from: <add> http://www.oracle.com/technetwork/java/javase/downloads/index.html <ide> EOS <ide> end <ide> end <ide><path>Library/Homebrew/requirements/x11_dependency.rb <ide> def initialize(name="x11", tags=[]) <ide> @name = name <ide> if /(\d\.)+\d/ === tags.first <ide> @min_version = Version.new(tags.shift) <add> @min_version_string = " #{@min_version}" <ide> else <ide> @min_version = Version.new("0.0.0") <add> @min_version_string = "" <ide> end <ide> super(tags) <ide> end <ide> def initialize(name="x11", tags=[]) <ide> end <ide> <ide> def message; <<-EOS.undent <del> Unsatisfied dependency: XQuartz #{@min_version} <del> Homebrew does not package XQuartz. Installers may be found at: <add> You can install XQuartz#{@min_version_string} with: <add> brew cask install xquartz <add> <add> Or from: <ide> https://xquartz.macosforge.org <ide> EOS <ide> end
2
PHP
PHP
use argument unpacking
f80d6f89a3b4b35933892b29857d7cad1cda29bd
<ide><path>src/Illuminate/Support/Facades/Facade.php <ide> public static function __callStatic($method, $args) <ide> throw new RuntimeException('A facade root has not been set.'); <ide> } <ide> <del> switch (count($args)) { <del> case 0: <del> return $instance->$method(); <del> <del> case 1: <del> return $instance->$method($args[0]); <del> <del> case 2: <del> return $instance->$method($args[0], $args[1]); <del> <del> case 3: <del> return $instance->$method($args[0], $args[1], $args[2]); <del> <del> case 4: <del> return $instance->$method($args[0], $args[1], $args[2], $args[3]); <del> <del> default: <del> return call_user_func_array([$instance, $method], $args); <del> } <add> return $instance->$method(...$args); <ide> } <ide> }
1
Ruby
Ruby
require conversions to use string#ord
f802eb2f0039512105d6ed985e1ed3245f7316d5
<ide><path>activesupport/lib/active_support/cache/file_store.rb <ide> require 'active_support/core_ext/file/atomic' <add>require 'active_support/core_ext/string/conversions' <ide> <ide> module ActiveSupport <ide> module Cache
1