content_type
stringclasses 8
values | main_lang
stringclasses 7
values | message
stringlengths 1
50
| sha
stringlengths 40
40
| patch
stringlengths 52
962k
| file_count
int64 1
300
|
|---|---|---|---|---|---|
Javascript
|
Javascript
|
fix word spelling
|
f6210923c012e64738859be4b92700f6937d103b
|
<ide><path>.eslintrc.js
<ide> module.exports = {
<ide> ...["implements", "const", "memberof", "readonly", "yields"].reduce(
<ide> (acc, tag) => {
<ide> acc[tag] = {
<del> message: `@${tag} currently not supported in Typescript`
<add> message: `@${tag} currently not supported in TypeScript`
<ide> };
<ide> return acc;
<ide> },
| 1
|
Java
|
Java
|
use a surface switch
|
bc06d1cc9c8683115a85207c7f461c5ff82b5ef1
|
<ide><path>ReactAndroid/src/main/java/com/facebook/react/ReactInstanceManager.java
<ide> private void attachRootViewToInstance(final ReactRootView rootView) {
<ide>
<ide> @Nullable Bundle initialProperties = rootView.getAppProperties();
<ide> final int rootTag = uiManagerModule.addRootView(
<del> rootView,
<add> rootView.getView(),
<ide> initialProperties == null ?
<ide> new WritableNativeMap() : Arguments.fromBundle(initialProperties),
<ide> rootView.getInitialUITemplate());
<ide><path>ReactAndroid/src/main/java/com/facebook/react/ReactRootView.java
<ide> public interface ReactRootViewEventListener {
<ide> private int mLastWidth = 0;
<ide> private int mLastHeight = 0;
<ide> private @UIManagerType int mUIManagerType = DEFAULT;
<add> private final boolean mUseSurface;
<ide>
<ide> public ReactRootView(Context context) {
<ide> super(context);
<add> mUseSurface = false;
<add> init();
<add> }
<add>
<add> public ReactRootView(Context context, boolean useSurface) {
<add> super(context);
<add> mUseSurface = useSurface;
<ide> init();
<ide> }
<ide>
<ide> public ReactRootView(Context context, AttributeSet attrs) {
<ide> super(context, attrs);
<add> mUseSurface = false;
<ide> init();
<ide> }
<ide>
<ide> public ReactRootView(Context context, AttributeSet attrs, int defStyle) {
<ide> super(context, attrs, defStyle);
<add> mUseSurface = false;
<ide> init();
<ide> }
<ide>
<ide> private void init() {
<ide> setClipChildren(false);
<ide> }
<ide>
<add> public View getView() {
<add> // TODO add mUseSurface to return surface here
<add> return this;
<add> }
<add>
<ide> @Override
<ide> protected void onMeasure(int widthMeasureSpec, int heightMeasureSpec) {
<add> if (mUseSurface) {
<add> super.onMeasure(widthMeasureSpec, heightMeasureSpec);
<add> return;
<add> }
<add>
<ide> Systrace.beginSection(TRACE_TAG_REACT_JAVA_BRIDGE, "ReactRootView.onMeasure");
<ide> try {
<ide> boolean measureSpecsUpdated = widthMeasureSpec != mWidthMeasureSpec ||
<ide> public void requestDisallowInterceptTouchEvent(boolean disallowIntercept) {
<ide>
<ide> @Override
<ide> protected void onLayout(boolean changed, int left, int top, int right, int bottom) {
<add> if (mUseSurface) {
<add> super.onLayout(changed, left, top, right, bottom);
<add> }
<ide> // No-op since UIManagerModule handles actually laying out children.
<ide> }
<ide>
<ide> public void startReactApplication(
<ide> mAppProperties = initialProperties;
<ide> mInitialUITemplate = initialUITemplate;
<ide>
<add> if (mUseSurface) {
<add> // TODO initialize surface here
<add> }
<add>
<ide> if (!mReactInstanceManager.hasStartedCreatingInitialContext()) {
<ide> mReactInstanceManager.createReactContextInBackground();
<ide> }
<ide> public void setAppProperties(@Nullable Bundle appProperties) {
<ide> return;
<ide> }
<ide>
<del> if (mWasMeasured) {
<del> updateRootLayoutSpecs(mWidthMeasureSpec, mHeightMeasureSpec);
<del> }
<ide> CatalystInstance catalystInstance = reactContext.getCatalystInstance();
<add> String jsAppModuleName = getJSModuleName();
<ide>
<del> WritableNativeMap appParams = new WritableNativeMap();
<del> appParams.putDouble("rootTag", getRootViewTag());
<del> @Nullable Bundle appProperties = getAppProperties();
<del> if (appProperties != null) {
<del> appParams.putMap("initialProps", Arguments.fromBundle(appProperties));
<del> }
<del> if (getUIManagerType() == FABRIC) {
<del> appParams.putBoolean("fabric", true);
<del> }
<add> if (mUseSurface) {
<add> // TODO call surface's runApplication
<add> } else {
<ide>
<del> mShouldLogContentAppeared = true;
<add> if (mWasMeasured) {
<add> updateRootLayoutSpecs(mWidthMeasureSpec, mHeightMeasureSpec);
<add> }
<ide>
<del> String jsAppModuleName = getJSModuleName();
<del> catalystInstance.getJSModule(AppRegistry.class).runApplication(jsAppModuleName, appParams);
<add> WritableNativeMap appParams = new WritableNativeMap();
<add> appParams.putDouble("rootTag", getRootViewTag());
<add> @Nullable Bundle appProperties = getAppProperties();
<add> if (appProperties != null) {
<add> appParams.putMap("initialProps", Arguments.fromBundle(appProperties));
<add> }
<add> if (getUIManagerType() == FABRIC) {
<add> appParams.putBoolean("fabric", true);
<add> }
<add>
<add> mShouldLogContentAppeared = true;
<add>
<add> catalystInstance.getJSModule(AppRegistry.class).runApplication(jsAppModuleName, appParams);
<add> }
<ide> } finally {
<ide> Systrace.endSection(TRACE_TAG_REACT_JAVA_BRIDGE);
<ide> }
| 2
|
Javascript
|
Javascript
|
fix strictequal assertion arguments
|
90e72daad5b9020e90b11c9f125cbe996354d679
|
<ide><path>test/parallel/test-buffer-alloc.js
<ide> const SlowBuffer = require('buffer').SlowBuffer;
<ide>
<ide>
<ide> const b = Buffer.allocUnsafe(1024);
<del>assert.strictEqual(1024, b.length);
<add>assert.strictEqual(b.length, 1024);
<ide>
<ide> b[0] = -1;
<ide> assert.strictEqual(b[0], 255);
<ide> for (let i = 0; i < 1024; i++) {
<ide> }
<ide>
<ide> const c = Buffer.allocUnsafe(512);
<del>assert.strictEqual(512, c.length);
<add>assert.strictEqual(c.length, 512);
<ide>
<ide> const d = Buffer.from([]);
<del>assert.strictEqual(0, d.length);
<add>assert.strictEqual(d.length, 0);
<ide>
<ide> // Test offset properties
<ide> {
<ide> const b = Buffer.alloc(128);
<del> assert.strictEqual(128, b.length);
<del> assert.strictEqual(0, b.byteOffset);
<del> assert.strictEqual(0, b.offset);
<add> assert.strictEqual(b.length, 128);
<add> assert.strictEqual(b.byteOffset, 0);
<add> assert.strictEqual(b.offset, 0);
<ide> }
<ide>
<ide> // Test creating a Buffer from a Uint32Array
<ide> Buffer.alloc(1).write('', 1, 0);
<ide>
<ide> {
<ide> const slice = b.slice(100, 150);
<del> assert.strictEqual(50, slice.length);
<add> assert.strictEqual(slice.length, 50);
<ide> for (let i = 0; i < 50; i++) {
<ide> assert.strictEqual(b[100 + i], slice[i]);
<ide> }
<ide> Buffer.alloc(1).write('', 1, 0);
<ide> const a = Buffer.allocUnsafe(8);
<ide> for (let i = 0; i < 8; i++) a[i] = i;
<ide> const b = a.slice(4, 8);
<del> assert.strictEqual(4, b[0]);
<del> assert.strictEqual(5, b[1]);
<del> assert.strictEqual(6, b[2]);
<del> assert.strictEqual(7, b[3]);
<add> assert.strictEqual(b[0], 4);
<add> assert.strictEqual(b[1], 5);
<add> assert.strictEqual(b[2], 6);
<add> assert.strictEqual(b[3], 7);
<ide> const c = b.slice(2, 4);
<del> assert.strictEqual(6, c[0]);
<del> assert.strictEqual(7, c[1]);
<add> assert.strictEqual(c[0], 6);
<add> assert.strictEqual(c[1], 7);
<ide> }
<ide>
<ide> {
<ide> Buffer.alloc(1).write('', 1, 0);
<ide> //
<ide> // Test toString('base64')
<ide> //
<del>assert.strictEqual('TWFu', (Buffer.from('Man')).toString('base64'));
<add>assert.strictEqual((Buffer.from('Man')).toString('base64'), 'TWFu');
<ide>
<ide> {
<ide> // test that regular and URL-safe base64 both work
<ide> assert.deepStrictEqual(Buffer.from(' YWJvcnVtLg', 'base64'),
<ide> const b = Buffer.from(s);
<ide>
<ide> for (let i = 0; i < l; i++) {
<del> assert.strictEqual('h'.charCodeAt(0), b[i]);
<add> assert.strictEqual(b[i], 'h'.charCodeAt(0));
<ide> }
<ide>
<ide> const sb = b.toString();
<ide> function buildBuffer(data) {
<ide>
<ide> const x = buildBuffer([0x81, 0xa3, 0x66, 0x6f, 0x6f, 0xa3, 0x62, 0x61, 0x72]);
<ide>
<del>assert.strictEqual('<Buffer 81 a3 66 6f 6f a3 62 61 72>', x.inspect());
<add>assert.strictEqual(x.inspect(), '<Buffer 81 a3 66 6f 6f a3 62 61 72>');
<ide>
<ide> {
<ide> const z = x.slice(4);
<del> assert.strictEqual(5, z.length);
<del> assert.strictEqual(0x6f, z[0]);
<del> assert.strictEqual(0xa3, z[1]);
<del> assert.strictEqual(0x62, z[2]);
<del> assert.strictEqual(0x61, z[3]);
<del> assert.strictEqual(0x72, z[4]);
<add> assert.strictEqual(z.length, 5);
<add> assert.strictEqual(z[0], 0x6f);
<add> assert.strictEqual(z[1], 0xa3);
<add> assert.strictEqual(z[2], 0x62);
<add> assert.strictEqual(z[3], 0x61);
<add> assert.strictEqual(z[4], 0x72);
<ide> }
<ide>
<ide> {
<ide> assert.strictEqual('<Buffer 81 a3 66 6f 6f a3 62 61 72>', x.inspect());
<ide>
<ide> {
<ide> const z = x.slice(0, 4);
<del> assert.strictEqual(4, z.length);
<del> assert.strictEqual(0x81, z[0]);
<del> assert.strictEqual(0xa3, z[1]);
<add> assert.strictEqual(z.length, 4);
<add> assert.strictEqual(z[0], 0x81);
<add> assert.strictEqual(z[1], 0xa3);
<ide> }
<ide>
<ide> {
<ide> const z = x.slice(0, 9);
<del> assert.strictEqual(9, z.length);
<add> assert.strictEqual(z.length, 9);
<ide> }
<ide>
<ide> {
<ide> const z = x.slice(1, 4);
<del> assert.strictEqual(3, z.length);
<del> assert.strictEqual(0xa3, z[0]);
<add> assert.strictEqual(z.length, 3);
<add> assert.strictEqual(z[0], 0xa3);
<ide> }
<ide>
<ide> {
<ide> const z = x.slice(2, 4);
<del> assert.strictEqual(2, z.length);
<del> assert.strictEqual(0x66, z[0]);
<del> assert.strictEqual(0x6f, z[1]);
<add> assert.strictEqual(z.length, 2);
<add> assert.strictEqual(z[0], 0x66);
<add> assert.strictEqual(z[1], 0x6f);
<ide> }
<ide>
<ide> ['ucs2', 'ucs-2', 'utf16le', 'utf-16le'].forEach((encoding) => {
<ide> assert.strictEqual('<Buffer 81 a3 66 6f 6f a3 62 61 72>', x.inspect());
<ide> const b = Buffer.from([0xde, 0xad, 0xbe, 0xef]);
<ide> let s = String.fromCharCode(0xffff);
<ide> b.write(s, 0, 'latin1');
<del> assert.strictEqual(0xff, b[0]);
<del> assert.strictEqual(0xad, b[1]);
<del> assert.strictEqual(0xbe, b[2]);
<del> assert.strictEqual(0xef, b[3]);
<add> assert.strictEqual(b[0], 0xff);
<add> assert.strictEqual(b[1], 0xad);
<add> assert.strictEqual(b[2], 0xbe);
<add> assert.strictEqual(b[3], 0xef);
<ide> s = String.fromCharCode(0xaaee);
<ide> b.write(s, 0, 'latin1');
<del> assert.strictEqual(0xee, b[0]);
<del> assert.strictEqual(0xad, b[1]);
<del> assert.strictEqual(0xbe, b[2]);
<del> assert.strictEqual(0xef, b[3]);
<add> assert.strictEqual(b[0], 0xee);
<add> assert.strictEqual(b[1], 0xad);
<add> assert.strictEqual(b[2], 0xbe);
<add> assert.strictEqual(b[3], 0xef);
<ide> }
<ide>
<ide> {
<ide> // Binary encoding should write only one byte per character.
<ide> const b = Buffer.from([0xde, 0xad, 0xbe, 0xef]);
<ide> let s = String.fromCharCode(0xffff);
<ide> b.write(s, 0, 'latin1');
<del> assert.strictEqual(0xff, b[0]);
<del> assert.strictEqual(0xad, b[1]);
<del> assert.strictEqual(0xbe, b[2]);
<del> assert.strictEqual(0xef, b[3]);
<add> assert.strictEqual(b[0], 0xff);
<add> assert.strictEqual(b[1], 0xad);
<add> assert.strictEqual(b[2], 0xbe);
<add> assert.strictEqual(b[3], 0xef);
<ide> s = String.fromCharCode(0xaaee);
<ide> b.write(s, 0, 'latin1');
<del> assert.strictEqual(0xee, b[0]);
<del> assert.strictEqual(0xad, b[1]);
<del> assert.strictEqual(0xbe, b[2]);
<del> assert.strictEqual(0xef, b[3]);
<add> assert.strictEqual(b[0], 0xee);
<add> assert.strictEqual(b[1], 0xad);
<add> assert.strictEqual(b[2], 0xbe);
<add> assert.strictEqual(b[3], 0xef);
<ide> }
<ide>
<ide> {
<ide> assert.strictEqual('<Buffer 81 a3 66 6f 6f a3 62 61 72>', x.inspect());
<ide> {
<ide> // test offset returns are correct
<ide> const b = Buffer.allocUnsafe(16);
<del> assert.strictEqual(4, b.writeUInt32LE(0, 0));
<del> assert.strictEqual(6, b.writeUInt16LE(0, 4));
<del> assert.strictEqual(7, b.writeUInt8(0, 6));
<del> assert.strictEqual(8, b.writeInt8(0, 7));
<del> assert.strictEqual(16, b.writeDoubleLE(0, 8));
<add> assert.strictEqual(b.writeUInt32LE(0, 0), 4);
<add> assert.strictEqual(b.writeUInt16LE(0, 4), 6);
<add> assert.strictEqual(b.writeUInt8(0, 6), 7);
<add> assert.strictEqual(b.writeInt8(0, 7), 8);
<add> assert.strictEqual(b.writeDoubleLE(0, 8), 16);
<ide> }
<ide>
<ide> {
| 1
|
PHP
|
PHP
|
apply fixes from styleci
|
22cdcf7cd02a6e35146b4737321a7ce18a075d43
|
<ide><path>tests/Integration/Console/ConsoleApplicationTest.php
<ide> use Illuminate\Console\Command;
<ide> use Orchestra\Testbench\TestCase;
<ide> use Illuminate\Contracts\Console\Kernel;
<del>use Illuminate\Foundation\Testing\PendingCommand;
<ide>
<ide> class ConsoleApplicationTest extends TestCase
<ide> {
| 1
|
Go
|
Go
|
address feedback from tonis
|
0cba7740d41369eee33b671f26276325580bc07b
|
<ide><path>builder/dockerfile/builder.go
<ide> func addNodesForLabelOption(dockerfile *parser.Node, labels map[string]string) {
<ide> //
<ide> // TODO: Remove?
<ide> func BuildFromConfig(config *container.Config, changes []string, os string) (*container.Config, error) {
<add> if !system.IsOSSupported(os) {
<add> return nil, errdefs.InvalidParameter(system.ErrNotSupportedOperatingSystem)
<add> }
<ide> if len(changes) == 0 {
<ide> return config, nil
<ide> }
<ide><path>builder/dockerfile/dispatchers.go
<ide> func initializeStage(d dispatchRequest, cmd *instructions.Stage) error {
<ide> return err
<ide> }
<ide> state := d.state
<del> state.beginStage(cmd.Name, image)
<add> if err := state.beginStage(cmd.Name, image); err != nil {
<add> return err
<add> }
<ide> if len(state.runConfig.OnBuild) > 0 {
<ide> triggers := state.runConfig.OnBuild
<ide> state.runConfig.OnBuild = nil
<ide> func resolveCmdLine(cmd instructions.ShellDependantCmdLine, runConfig *container
<ide> // RUN [ "echo", "hi" ] # echo hi
<ide> //
<ide> func dispatchRun(d dispatchRequest, c *instructions.RunCommand) error {
<del>
<add> if !system.IsOSSupported(d.state.operatingSystem) {
<add> return system.ErrNotSupportedOperatingSystem
<add> }
<ide> stateRunConfig := d.state.runConfig
<ide> cmdFromArgs := resolveCmdLine(c.ShellDependantCmdLine, stateRunConfig, d.state.operatingSystem)
<ide> buildArgs := d.state.buildArgs.FilterAllowed(stateRunConfig.Env)
<ide><path>builder/dockerfile/evaluator.go
<ide> package dockerfile
<ide>
<ide> import (
<ide> "reflect"
<add> "runtime"
<ide> "strconv"
<ide> "strings"
<ide>
<ide> func (s *dispatchState) hasFromImage() bool {
<ide> return s.imageID != "" || (s.baseImage != nil && s.baseImage.ImageID() == "")
<ide> }
<ide>
<del>func (s *dispatchState) beginStage(stageName string, image builder.Image) {
<add>func (s *dispatchState) beginStage(stageName string, image builder.Image) error {
<ide> s.stageName = stageName
<ide> s.imageID = image.ImageID()
<ide> s.operatingSystem = image.OperatingSystem()
<add> if s.operatingSystem == "" { // In case it isn't set
<add> s.operatingSystem = runtime.GOOS
<add> }
<add> if !system.IsOSSupported(s.operatingSystem) {
<add> return system.ErrNotSupportedOperatingSystem
<add> }
<ide>
<ide> if image.RunConfig() != nil {
<ide> // copy avoids referencing the same instance when 2 stages have the same base
<ide> func (s *dispatchState) beginStage(stageName string, image builder.Image) {
<ide> s.setDefaultPath()
<ide> s.runConfig.OpenStdin = false
<ide> s.runConfig.StdinOnce = false
<add> return nil
<ide> }
<ide>
<ide> // Add the default PATH to runConfig.ENV if one exists for the operating system and there
<ide><path>daemon/build.go
<ide> import (
<ide> "github.com/docker/docker/pkg/containerfs"
<ide> "github.com/docker/docker/pkg/idtools"
<ide> "github.com/docker/docker/pkg/stringid"
<add> "github.com/docker/docker/pkg/system"
<ide> "github.com/docker/docker/registry"
<ide> "github.com/pkg/errors"
<ide> "github.com/sirupsen/logrus"
<ide> func (rl *releaseableLayer) Commit() (builder.ReleaseableLayer, error) {
<ide> if err != nil {
<ide> return nil, err
<ide> }
<del> // TODO: An optimization woudld be to handle empty layers before returning
<add> // TODO: An optimization would be to handle empty layers before returning
<ide> return &releaseableLayer{layerStore: rl.layerStore, roLayer: newLayer}, nil
<ide> }
<ide>
<ide> func (daemon *Daemon) pullForBuilder(ctx context.Context, name string, authConfi
<ide> // leaking of layers.
<ide> func (daemon *Daemon) GetImageAndReleasableLayer(ctx context.Context, refOrID string, opts backend.GetImageAndLayerOptions) (builder.Image, builder.ReleaseableLayer, error) {
<ide> if refOrID == "" {
<add> if !system.IsOSSupported(opts.OS) {
<add> return nil, nil, system.ErrNotSupportedOperatingSystem
<add> }
<ide> layer, err := newReleasableLayerForImage(nil, daemon.layerStores[opts.OS])
<ide> return nil, layer, err
<ide> }
<ide> func (daemon *Daemon) GetImageAndReleasableLayer(ctx context.Context, refOrID st
<ide> }
<ide> // TODO: shouldn't we error out if error is different from "not found" ?
<ide> if image != nil {
<add> if !system.IsOSSupported(image.OperatingSystem()) {
<add> return nil, nil, system.ErrNotSupportedOperatingSystem
<add> }
<ide> layer, err := newReleasableLayerForImage(image, daemon.layerStores[image.OperatingSystem()])
<ide> return image, layer, err
<ide> }
<ide> func (daemon *Daemon) GetImageAndReleasableLayer(ctx context.Context, refOrID st
<ide> if err != nil {
<ide> return nil, nil, err
<ide> }
<add> if !system.IsOSSupported(image.OperatingSystem()) {
<add> return nil, nil, system.ErrNotSupportedOperatingSystem
<add> }
<ide> layer, err := newReleasableLayerForImage(image, daemon.layerStores[image.OperatingSystem()])
<ide> return image, layer, err
<ide> }
<ide><path>daemon/commit.go
<ide> import (
<ide> "github.com/docker/docker/image"
<ide> "github.com/docker/docker/layer"
<ide> "github.com/docker/docker/pkg/ioutils"
<add> "github.com/docker/docker/pkg/system"
<ide> "github.com/pkg/errors"
<ide> )
<ide>
<ide> func (daemon *Daemon) Commit(name string, c *backend.ContainerCommitConfig) (str
<ide> daemon.containerPause(container)
<ide> defer daemon.containerUnpause(container)
<ide> }
<add> if !system.IsOSSupported(container.OS) {
<add> return "", system.ErrNotSupportedOperatingSystem
<add> }
<ide>
<ide> if c.MergeConfigs && c.Config == nil {
<ide> c.Config = container.Config
<ide> func (daemon *Daemon) Commit(name string, c *backend.ContainerCommitConfig) (str
<ide> }
<ide>
<ide> func (daemon *Daemon) exportContainerRw(container *container.Container) (arch io.ReadCloser, err error) {
<add> // Note: Indexing by OS is safe as only called from `Commit` which has already performed validation
<ide> rwlayer, err := daemon.layerStores[container.OS].GetRWLayer(container.ID)
<ide> if err != nil {
<ide> return nil, err
<ide><path>daemon/create.go
<ide> func (daemon *Daemon) setRWLayer(container *container.Container) error {
<ide> StorageOpt: container.HostConfig.StorageOpt,
<ide> }
<ide>
<add> // Indexing by OS is safe here as validation of OS has already been performed in create() (the only
<add> // caller), and guaranteed non-nil
<ide> rwLayer, err := daemon.layerStores[container.OS].CreateRWLayer(container.ID, layerID, rwLayerOpts)
<ide> if err != nil {
<ide> return err
<ide><path>daemon/daemon.go
<ide> func (daemon *Daemon) restore() error {
<ide> logrus.Errorf("Failed to load container %v: %v", id, err)
<ide> continue
<ide> }
<del>
<add> if !system.IsOSSupported(container.OS) {
<add> logrus.Errorf("Failed to load container %v: %s (%q)", id, system.ErrNotSupportedOperatingSystem, container.OS)
<add> continue
<add> }
<ide> // Ignore the container if it does not support the current driver being used by the graph
<ide> currentDriverForContainerOS := daemon.graphDrivers[container.OS]
<ide> if (container.Driver == "" && currentDriverForContainerOS == "aufs") || container.Driver == currentDriverForContainerOS {
<ide><path>daemon/delete.go
<ide> func (daemon *Daemon) cleanupContainer(container *container.Container, forceRemo
<ide> return fmt.Errorf("Could not kill running container %s, cannot remove - %v", container.ID, err)
<ide> }
<ide> }
<add> if !system.IsOSSupported(container.OS) {
<add> return fmt.Errorf("cannot remove %s: %s ", container.ID, system.ErrNotSupportedOperatingSystem)
<add> }
<ide>
<ide> // stop collection of stats for the container regardless
<ide> // if stats are currently getting collected.
<ide><path>daemon/export.go
<ide> import (
<ide> "github.com/docker/docker/errdefs"
<ide> "github.com/docker/docker/pkg/archive"
<ide> "github.com/docker/docker/pkg/ioutils"
<add> "github.com/docker/docker/pkg/system"
<ide> )
<ide>
<ide> // ContainerExport writes the contents of the container to the given
<ide> func (daemon *Daemon) ContainerExport(name string, out io.Writer) error {
<ide> }
<ide>
<ide> func (daemon *Daemon) containerExport(container *container.Container) (arch io.ReadCloser, err error) {
<add> if !system.IsOSSupported(container.OS) {
<add> return nil, fmt.Errorf("cannot export %s: %s ", container.ID, system.ErrNotSupportedOperatingSystem)
<add> }
<ide> rwlayer, err := daemon.layerStores[container.OS].GetRWLayer(container.ID)
<ide> if err != nil {
<ide> return nil, err
<ide><path>daemon/image_delete.go
<ide> import (
<ide> "github.com/docker/docker/errdefs"
<ide> "github.com/docker/docker/image"
<ide> "github.com/docker/docker/pkg/stringid"
<add> "github.com/docker/docker/pkg/system"
<ide> "github.com/pkg/errors"
<ide> )
<ide>
<ide> func (daemon *Daemon) ImageDelete(imageRef string, force, prune bool) ([]types.I
<ide> start := time.Now()
<ide> records := []types.ImageDeleteResponseItem{}
<ide>
<del> imgID, _, err := daemon.GetImageIDAndOS(imageRef)
<add> imgID, operatingSystem, err := daemon.GetImageIDAndOS(imageRef)
<ide> if err != nil {
<ide> return nil, err
<ide> }
<add> if !system.IsOSSupported(operatingSystem) {
<add> return nil, errors.Errorf("unable to delete image: %q", system.ErrNotSupportedOperatingSystem)
<add> }
<ide>
<ide> repoRefs := daemon.referenceStore.References(imgID.Digest())
<ide>
<ide><path>daemon/image_inspect.go
<ide> import (
<ide> "github.com/docker/distribution/reference"
<ide> "github.com/docker/docker/api/types"
<ide> "github.com/docker/docker/layer"
<add> "github.com/docker/docker/pkg/system"
<ide> "github.com/pkg/errors"
<ide> )
<ide>
<ide> func (daemon *Daemon) LookupImage(name string) (*types.ImageInspect, error) {
<ide> if err != nil {
<ide> return nil, errors.Wrapf(err, "no such image: %s", name)
<ide> }
<del>
<add> if !system.IsOSSupported(img.OperatingSystem()) {
<add> return nil, system.ErrNotSupportedOperatingSystem
<add> }
<ide> refs := daemon.referenceStore.References(img.ID().Digest())
<ide> repoTags := []string{}
<ide> repoDigests := []string{}
<ide><path>daemon/images.go
<ide> import (
<ide> "github.com/docker/docker/container"
<ide> "github.com/docker/docker/image"
<ide> "github.com/docker/docker/layer"
<add> "github.com/docker/docker/pkg/system"
<ide> )
<ide>
<ide> var acceptedImageFilterTags = map[string]bool{
<ide> func (daemon *Daemon) Images(imageFilters filters.Args, all bool, withExtraAttrs
<ide> }
<ide> }
<ide>
<add> // Skip any images with an unsupported operating system to avoid a potential
<add> // panic when indexing through the layerstore. Don't error as we want to list
<add> // the other images. This should never happen, but here as a safety precaution.
<add> if !system.IsOSSupported(img.OperatingSystem()) {
<add> continue
<add> }
<add>
<ide> layerID := img.RootFS.ChainID()
<ide> var size int64
<ide> if layerID != "" {
<ide><path>daemon/oci_windows.go
<ide> func (daemon *Daemon) createSpec(c *container.Container) (*specs.Spec, error) {
<ide> max := len(img.RootFS.DiffIDs)
<ide> for i := 1; i <= max; i++ {
<ide> img.RootFS.DiffIDs = img.RootFS.DiffIDs[:i]
<add> if !system.IsOSSupported(img.OperatingSystem()) {
<add> return nil, fmt.Errorf("cannot get layerpath for ImageID %s: %s ", img.RootFS.ChainID(), system.ErrNotSupportedOperatingSystem)
<add> }
<ide> layerPath, err := layer.GetLayerPath(daemon.layerStores[img.OperatingSystem()], img.RootFS.ChainID())
<ide> if err != nil {
<ide> return nil, fmt.Errorf("failed to get layer path from graphdriver %s for ImageID %s - %s", daemon.layerStores[img.OperatingSystem()], img.RootFS.ChainID(), err)
<ide><path>distribution/config.go
<ide> func (s *imageConfigStore) RootFSAndOSFromConfig(c []byte) (*image.RootFS, strin
<ide> if os == "" {
<ide> os = runtime.GOOS
<ide> }
<add> if !system.IsOSSupported(os) {
<add> return nil, "", system.ErrNotSupportedOperatingSystem
<add> }
<ide> return unmarshalledConfig.RootFS, os, nil
<ide> }
<ide>
<ide><path>distribution/push_v1.go
<ide> import (
<ide> "github.com/docker/docker/pkg/ioutils"
<ide> "github.com/docker/docker/pkg/progress"
<ide> "github.com/docker/docker/pkg/stringid"
<add> "github.com/docker/docker/pkg/system"
<ide> "github.com/docker/docker/registry"
<ide> "github.com/opencontainers/go-digest"
<ide> "github.com/sirupsen/logrus"
<ide> func (p *v1Pusher) imageListForTag(imgID image.ID, dependenciesSeen map[layer.Ch
<ide>
<ide> topLayerID := img.RootFS.ChainID()
<ide>
<add> if !system.IsOSSupported(img.OperatingSystem()) {
<add> return nil, system.ErrNotSupportedOperatingSystem
<add> }
<ide> pl, err := p.config.LayerStores[img.OperatingSystem()].Get(topLayerID)
<ide> *referencedLayers = append(*referencedLayers, pl)
<ide> if err != nil {
<ide><path>distribution/xfer/download.go
<ide> import (
<ide> "github.com/docker/docker/pkg/archive"
<ide> "github.com/docker/docker/pkg/ioutils"
<ide> "github.com/docker/docker/pkg/progress"
<add> "github.com/docker/docker/pkg/system"
<ide> "github.com/sirupsen/logrus"
<ide> "golang.org/x/net/context"
<ide> )
<ide> func (ldm *LayerDownloadManager) Download(ctx context.Context, initialRootFS ima
<ide> downloadsByKey = make(map[string]*downloadTransfer)
<ide> )
<ide>
<del> // Assume that the operating system is the host OS if blank
<add> // Assume that the operating system is the host OS if blank, and validate it
<add> // to ensure we don't cause a panic by an invalid index into the layerstores.
<ide> if os == "" {
<ide> os = runtime.GOOS
<ide> }
<add> if !system.IsOSSupported(os) {
<add> return image.RootFS{}, nil, system.ErrNotSupportedOperatingSystem
<add> }
<ide>
<ide> rootFS := initialRootFS
<ide> for _, descriptor := range layers {
<ide><path>image/store.go
<ide> import (
<ide>
<ide> "github.com/docker/distribution/digestset"
<ide> "github.com/docker/docker/layer"
<add> "github.com/docker/docker/pkg/system"
<ide> "github.com/opencontainers/go-digest"
<ide> "github.com/pkg/errors"
<ide> "github.com/sirupsen/logrus"
<ide> func (is *store) restore() error {
<ide> }
<ide> var l layer.Layer
<ide> if chainID := img.RootFS.ChainID(); chainID != "" {
<add> if !system.IsOSSupported(img.OperatingSystem()) {
<add> return system.ErrNotSupportedOperatingSystem
<add> }
<ide> l, err = is.lss[img.OperatingSystem()].Get(chainID)
<ide> if err != nil {
<ide> return err
<ide> func (is *store) Create(config []byte) (ID, error) {
<ide>
<ide> var l layer.Layer
<ide> if layerID != "" {
<add> if !system.IsOSSupported(img.OperatingSystem()) {
<add> return "", system.ErrNotSupportedOperatingSystem
<add> }
<ide> l, err = is.lss[img.OperatingSystem()].Get(layerID)
<ide> if err != nil {
<ide> return "", errors.Wrapf(err, "failed to get layer %s", layerID)
<ide><path>layer/filestore_unix.go
<ide> package layer
<ide>
<ide> import "runtime"
<ide>
<del>// SetOS writes the "os" file to the layer filestore
<del>func (fm *fileMetadataTransaction) SetOS(os string) error {
<add>// setOS writes the "os" file to the layer filestore
<add>func (fm *fileMetadataTransaction) setOS(os string) error {
<ide> return nil
<ide> }
<ide>
<del>// GetOS reads the "os" file from the layer filestore
<del>func (fms *fileMetadataStore) GetOS(layer ChainID) (string, error) {
<add>// getOS reads the "os" file from the layer filestore
<add>func (fms *fileMetadataStore) getOS(layer ChainID) (string, error) {
<ide> return runtime.GOOS, nil
<ide> }
<ide><path>layer/filestore_windows.go
<ide> import (
<ide> "strings"
<ide> )
<ide>
<del>// SetOS writes the "os" file to the layer filestore
<del>func (fm *fileMetadataTransaction) SetOS(os string) error {
<add>// setOS writes the "os" file to the layer filestore
<add>func (fm *fileMetadataTransaction) setOS(os string) error {
<ide> if os == "" {
<ide> return nil
<ide> }
<ide> return fm.ws.WriteFile("os", []byte(os), 0644)
<ide> }
<ide>
<del>// GetOS reads the "os" file from the layer filestore
<del>func (fms *fileMetadataStore) GetOS(layer ChainID) (string, error) {
<add>// getOS reads the "os" file from the layer filestore
<add>func (fms *fileMetadataStore) getOS(layer ChainID) (string, error) {
<ide> contentBytes, err := ioutil.ReadFile(fms.getLayerFilename(layer, "os"))
<ide> if err != nil {
<ide> // For backwards compatibility, the os file may not exist. Default to "windows" if missing.
<ide><path>layer/layer.go
<ide> type MetadataTransaction interface {
<ide> SetDiffID(DiffID) error
<ide> SetCacheID(string) error
<ide> SetDescriptor(distribution.Descriptor) error
<del> SetOS(string) error
<add> setOS(string) error
<ide> TarSplitWriter(compressInput bool) (io.WriteCloser, error)
<ide>
<ide> Commit(ChainID) error
<ide> type MetadataStore interface {
<ide> GetDiffID(ChainID) (DiffID, error)
<ide> GetCacheID(ChainID) (string, error)
<ide> GetDescriptor(ChainID) (distribution.Descriptor, error)
<del> GetOS(ChainID) (string, error)
<add> getOS(ChainID) (string, error)
<ide> TarSplitReader(ChainID) (io.ReadCloser, error)
<ide>
<ide> SetMountID(string, string) error
<ide><path>layer/layer_store.go
<ide> func (ls *layerStore) loadLayer(layer ChainID) (*roLayer, error) {
<ide> return nil, fmt.Errorf("failed to get descriptor for %s: %s", layer, err)
<ide> }
<ide>
<del> os, err := ls.store.GetOS(layer)
<add> os, err := ls.store.getOS(layer)
<ide> if err != nil {
<ide> return nil, fmt.Errorf("failed to get operating system for %s: %s", layer, err)
<ide> }
<ide><path>layer/ro_layer.go
<ide> func storeLayer(tx MetadataTransaction, layer *roLayer) error {
<ide> return err
<ide> }
<ide> }
<del> if err := tx.SetOS(layer.layerStore.os); err != nil {
<add> if err := tx.setOS(layer.layerStore.os); err != nil {
<ide> return err
<ide> }
<ide>
<ide><path>pkg/system/errors.go
<ide> import (
<ide> var (
<ide> // ErrNotSupportedPlatform means the platform is not supported.
<ide> ErrNotSupportedPlatform = errors.New("platform and architecture is not supported")
<add>
<add> // ErrNotSupportedOperatingSystem means the operating system is not supported.
<add> ErrNotSupportedOperatingSystem = errors.New("operating system is not supported")
<ide> )
<ide><path>pkg/system/lcow.go
<ide> func IsOSSupported(os string) bool {
<ide> if runtime.GOOS == os {
<ide> return true
<ide> }
<del> if LCOWSupported() && runtime.GOOS == "windows" && os == "linux" {
<add> if LCOWSupported() && os == "linux" {
<ide> return true
<ide> }
<ide> return false
| 24
|
PHP
|
PHP
|
add accessor for reading individual files
|
96630e240a7b0133c443b3899c7636efb1102855
|
<ide><path>src/Network/Request.php
<ide> public function getAttributes()
<ide> return $this->attributes + $emulated;
<ide> }
<ide>
<add> /**
<add> * Get the uploaded file from a dotted path.
<add> *
<add> * @param string $path The dot separated path to the file you want.
<add> * @return null|Psr\Http\Message\UploadedFileInterface
<add> */
<add> public function getUploadedFile($path)
<add> {
<add> $file = Hash::get($this->uploadedFiles, $path);
<add> if (!$file instanceof UploadedFile) {
<add> return null;
<add> }
<add>
<add> return $file;
<add> }
<add>
<ide> /**
<ide> * Get the array of uploaded files from the request.
<ide> *
<ide><path>tests/TestCase/Network/RequestTest.php
<ide> public function testWithUploadedFiles()
<ide> $this->assertSame(['picture' => $file], $new->getUploadedFiles());
<ide> }
<ide>
<add> /**
<add> * Test getting a single file
<add> *
<add> * @return void
<add> */
<add> public function testGetUploadedFile()
<add> {
<add> $file = new UploadedFile(
<add> __FILE__,
<add> 123,
<add> UPLOAD_ERR_OK,
<add> 'test.php',
<add> 'text/plain'
<add> );
<add> $request = new Request();
<add> $new = $request->withUploadedFiles(['picture' => $file]);
<add> $this->assertNull($new->getUploadedFile(''));
<add> $this->assertSame($file, $new->getUploadedFile('picture'));
<add>
<add> $new = $request->withUploadedFiles([
<add> 'pictures' => [
<add> [
<add> 'image' => $file
<add> ]
<add> ]
<add> ]);
<add> $this->assertNull($new->getUploadedFile('pictures'));
<add> $this->assertNull($new->getUploadedFile('pictures.0'));
<add> $this->assertNull($new->getUploadedFile('pictures.1'));
<add> $this->assertSame($file, $new->getUploadedFile('pictures.0.image'));
<add> }
<add>
<ide> /**
<ide> * Test replacing files with an invalid file
<ide> *
| 2
|
Text
|
Text
|
update seed code for d3.js challenge
|
f8aa6c04e20cf363d727d93c366f34894625110a
|
<ide><path>curriculum/challenges/english/04-data-visualization/data-visualization-with-d3/change-the-presentation-of-a-bar-chart.md
<ide> assert(
<ide> .bar {
<ide> width: 25px;
<ide> height: 100px;
<del> /* Only change code below this line */
<add> /* Add your code below this line */
<ide>
<ide>
<del> /* Only change code above this line */
<add> /* Add your code above this line */
<ide> display: inline-block;
<ide> background-color: blue;
<ide> }
<ide> assert(
<ide> .enter()
<ide> .append("div")
<ide> .attr("class", "bar")
<del> .style("height", (d) => (d + "px"))
<add> .style("height", (d) => (d + "px")) // Change this line
<ide> </script>
<ide> </body>
<ide> ```
| 1
|
Javascript
|
Javascript
|
add info circle to claim cert steps
|
1805631e4a03e7c97950d22090e06650ed9a62a7
|
<ide><path>client/src/templates/Introduction/components/ClaimCertSteps.js
<ide> import { withTranslation, useTranslation } from 'react-i18next';
<ide> import { connect } from 'react-redux';
<ide> import { createSelector } from 'reselect';
<ide>
<add>import IntroInformation from '../../../assets/icons/IntroInformation';
<ide> import GreenPass from '../../../assets/icons/GreenPass';
<ide> import GreenNotCompleted from '../../../assets/icons/GreenNotCompleted';
<ide> import { userSelector } from '../../../redux';
<ide> const ClaimCertSteps = ({ certSlug, i18nCertText, superBlock, user }) => {
<ide> {t('certification-card.complete-project', {
<ide> i18nCertText
<ide> })}
<add> <span className='badge map-badge map-project-checkmark'>
<add> <IntroInformation style={mapIconStyle} />
<add> </span>
<ide> </a>
<ide> </li>
<ide> <li className='map-challenge-title map-project-wrap'>
<del> <Link to={certClaimLink}>{t('certification-card.set-claim')}</Link>
<add> <Link to={certClaimLink}>
<add> {t('certification-card.set-claim')}
<add> <span className='badge map-badge map-project-checkmark'>
<add> <IntroInformation style={mapIconStyle} />
<add> </span>
<add> </Link>
<ide> </li>
<ide> </ul>
<ide> );
| 1
|
PHP
|
PHP
|
add doc blocks to callbackstatement
|
a79b614bef74503ac2c9bb18a33658b305d40fe7
|
<ide><path>Cake/Database/Statement/CallbackStatement.php
<ide> */
<ide> namespace Cake\Database\Statement;
<ide>
<add>/**
<add> * Wraps a statement in a callback that allows row results
<add> * to be modified when being fetched.
<add> *
<add> * This is used by CakePHP to eagerly load association data.
<add> */
<ide> class CallbackStatement extends StatementDecorator {
<ide>
<add>/**
<add> * A callback function to be applied to results.
<add> *
<add> * @var callable
<add> */
<ide> protected $_callback;
<ide>
<add>/**
<add> * Constructor
<add> *
<add> * @param Cake\Database\StatementInterface $statement The statement to decorate.
<add> * @param Cake\Database\Driver $driver The driver instance used by the statement.
<add> * @param callable $callback The callback to apply to results before they are returned.
<add> */
<ide> public function __construct($statement, $driver, $callback) {
<ide> parent::__construct($statement, $driver);
<ide> $this->_callback = $callback;
<ide> }
<ide>
<add>/**
<add> * Fetch a row from the statement.
<add> *
<add> * The result will be procssed by the callback when it is not `false.
<add> *
<add> * @param string $type Either 'num' or 'assoc' to indicate the result format you would like.
<add> * @return mixed
<add> */
<ide> public function fetch($type = 'num') {
<ide> $callback = $this->_callback;
<ide> $row = $this->_statement->fetch($type);
<ide> return $row === false ? $row : $callback($row);
<ide> }
<ide>
<add>/**
<add> * Fetch all rows from the statement.
<add> *
<add> * Each row in the result will be procssed by the callback when it is not `false.
<add> *
<add> * @param string $type Either 'num' or 'assoc' to indicate the result format you would like.
<add> * @return mixed
<add> */
<ide> public function fetchAll($type = 'num') {
<ide> return array_map($this->_callback, $this->_statement->fetchAll($type));
<ide> }
| 1
|
Ruby
|
Ruby
|
use rails() instead of system()
|
75afb43a25eacaefd7487dfa1f95be03c06c62ee
|
<ide><path>railties/test/application/assets_test.rb
<ide> def teardown
<ide> def precompile!(env = nil)
<ide> with_env env.to_h do
<ide> quietly do
<del> precompile_task = "bin/rails assets:precompile --trace 2>&1"
<del> output = Dir.chdir(app_path) { %x[ #{precompile_task} ] }
<del> assert $?.success?, output
<del> output
<add> rails ["assets:precompile", "--trace"]
<ide> end
<ide> end
<ide> end
<ide> def with_env(env)
<ide>
<ide> def clean_assets!
<ide> quietly do
<del> assert Dir.chdir(app_path) { system("bin/rails assets:clobber") }
<add> rails ["assets:clobber"]
<ide> end
<ide> end
<ide>
| 1
|
Javascript
|
Javascript
|
add tests for jwt authorization
|
36c473799839bd4b31d29a0aa418f8c5f527d1f5
|
<ide><path>api-server/server/middlewares/jwt-authorizaion.test.js
<del>/* global describe xdescribe it expect */
<del>import { isWhiteListedPath } from './jwt-authorization';
<del>
<del>describe('jwt-authorization', () => {
<del> describe('isWhiteListedPath', () => {
<del> const whiteList = [/^\/is-ok\//, /^\/this-is\/also\/ok\//];
<del>
<del> it('returns a boolean', () => {
<del> const result = isWhiteListedPath();
<del>
<del> expect(typeof result).toBe('boolean');
<del> });
<del>
<del> it('returns true for a white listed path', () => {
<del> expect.assertions(2);
<del>
<del> const resultA = isWhiteListedPath('/is-ok/should-be/good', whiteList);
<del> const resultB = isWhiteListedPath('/this-is/also/ok/surely', whiteList);
<del> expect(resultA).toBe(true);
<del> expect(resultB).toBe(true);
<del> });
<del>
<del> it('returns false for a non-white-listed path', () => {
<del> const result = isWhiteListedPath('/hax0r-42/no-go', whiteList);
<del> expect(result).toBe(false);
<del> });
<del> });
<del>
<del> xdescribe('authorizeByJWT');
<del>});
<ide><path>api-server/server/middlewares/request-authorizaion.test.js
<add>/* global describe it expect */
<add>import sinon from 'sinon';
<add>import { mockReq, mockRes } from 'sinon-express-mock';
<add>import jwt from 'jsonwebtoken';
<add>
<add>import createRequestAuthorization, {
<add> isWhiteListedPath
<add>} from './request-authorization';
<add>
<add>const validJWTSecret = 'this is a super secret string';
<add>const invalidJWTSecret = 'This is not correct secret';
<add>const now = new Date(Date.now());
<add>const theBeginningOfTime = new Date(0);
<add>const accessToken = {
<add> id: '123abc',
<add> userId: '456def',
<add> ttl: 60000,
<add> created: now
<add>};
<add>const users = {
<add> '456def': {
<add> username: 'camperbot',
<add> progressTimestamps: [1, 2, 3, 4]
<add> }
<add>};
<add>const mockGetUserById = id =>
<add> id in users ? Promise.resolve(users[id]) : Promise.reject('No user found');
<add>
<add>describe('request-authorization', () => {
<add> describe('isWhiteListedPath', () => {
<add> const whiteList = [/^\/is-ok\//, /^\/this-is\/also\/ok\//];
<add>
<add> it('returns a boolean', () => {
<add> const result = isWhiteListedPath();
<add>
<add> expect(typeof result).toBe('boolean');
<add> });
<add>
<add> it('returns true for a white listed path', () => {
<add> expect.assertions(2);
<add>
<add> const resultA = isWhiteListedPath('/is-ok/should-be/good', whiteList);
<add> const resultB = isWhiteListedPath('/this-is/also/ok/surely', whiteList);
<add> expect(resultA).toBe(true);
<add> expect(resultB).toBe(true);
<add> });
<add>
<add> it('returns false for a non-white-listed path', () => {
<add> const result = isWhiteListedPath('/hax0r-42/no-go', whiteList);
<add> expect(result).toBe(false);
<add> });
<add> });
<add>
<add> describe('createRequestAuthorization', () => {
<add> const requestAuthorization = createRequestAuthorization({
<add> _jwtSecret: validJWTSecret,
<add> getUserById: mockGetUserById
<add> });
<add>
<add> it('is a function', () => {
<add> expect(typeof requestAuthorization).toEqual('function');
<add> });
<add>
<add> it('throws when no access token is present', () => {
<add> expect.assertions(2);
<add> const req = mockReq({ path: '/internal/some-path/that-needs/auth' });
<add> const res = mockRes();
<add> const next = sinon.spy();
<add> expect(() => requestAuthorization(req, res, next)).toThrowError(
<add> 'Access token is required for this request'
<add> );
<add> expect(next.called).toBe(false);
<add> });
<add>
<add> it('throws when the access token is invalid', () => {
<add> expect.assertions(2);
<add> const invalidJWT = jwt.sign({ accessToken }, invalidJWTSecret);
<add> const req = mockReq({
<add> path: '/internal/some-path/that-needs/auth',
<add> // eslint-disable-next-line camelcase
<add> cookie: { jwt_access_token: invalidJWT }
<add> });
<add> const res = mockRes();
<add> const next = sinon.spy();
<add>
<add> expect(() => requestAuthorization(req, res, next)).toThrowError(
<add> 'invalid signature'
<add> );
<add> expect(next.called).toBe(false);
<add> });
<add>
<add> it('throws when the access token has expired', () => {
<add> expect.assertions(2);
<add> const invalidJWT = jwt.sign(
<add> { accessToken: { ...accessToken, created: theBeginningOfTime } },
<add> validJWTSecret
<add> );
<add> const req = mockReq({
<add> path: '/internal/some-path/that-needs/auth',
<add> // eslint-disable-next-line camelcase
<add> cookie: { jwt_access_token: invalidJWT }
<add> });
<add> const res = mockRes();
<add> const next = sinon.spy();
<add>
<add> expect(() => requestAuthorization(req, res, next)).toThrowError(
<add> 'Access token is no longer vaild'
<add> );
<add> expect(next.called).toBe(false);
<add> });
<add>
<add> it('adds the user to the request object', async done => {
<add> expect.assertions(5);
<add> const validJWT = jwt.sign({ accessToken }, validJWTSecret);
<add> const req = mockReq({
<add> path: '/internal/some-path/that-needs/auth',
<add> // eslint-disable-next-line camelcase
<add> cookie: { jwt_access_token: validJWT }
<add> });
<add> const res = mockRes();
<add> const next = sinon.spy();
<add> await requestAuthorization(req, res, next);
<add> expect(next.called).toBe(true);
<add> expect(req).toHaveProperty('user');
<add> expect(req.user).toEqual(users['456def']);
<add> expect(req.user).toHaveProperty('points');
<add> expect(req.user.points).toEqual(4);
<add> return done();
<add> });
<add>
<add> it('calls next if request does not require authorization', async () => {
<add> const req = mockReq({ path: '/unauthenticated/another/route' });
<add> const res = mockRes();
<add> const next = sinon.spy();
<add> await requestAuthorization(req, res, next);
<add> expect(next.called).toBe(true);
<add> });
<add> });
<add>});
<add><path>api-server/server/middlewares/request-authorization.js
<del><path>api-server/server/middlewares/jwt-authorization.js
<ide> import loopback from 'loopback';
<ide> import jwt from 'jsonwebtoken';
<ide> import { isBefore } from 'date-fns';
<add>import { isEmpty } from 'lodash';
<ide>
<ide> import { homeLocation } from '../../../config/env';
<ide>
<ide> export function isWhiteListedPath(path, whiteListREs = _whiteListREs) {
<ide> return whiteListREs.some(re => re.test(path));
<ide> }
<ide>
<del>export default () =>
<del> function authorizeByJWT(req, res, next) {
<add>export default ({
<add> _jwtSecret = process.env.JWT_SECRET,
<add> getUserById = _getUserById
<add>} = {}) =>
<add> function requestAuthorisation(req, res, next) {
<ide> const { path } = req;
<ide> if (apiProxyRE.test(path) && !isWhiteListedPath(path)) {
<ide> const cookie =
<ide> export default () =>
<ide> }
<ide> let token;
<ide> try {
<del> token = jwt.verify(cookie, process.env.JWT_SECRET);
<add> token = jwt.verify(cookie, _jwtSecret);
<ide> } catch (err) {
<ide> throw wrapHandledError(new Error(err.message), {
<ide> type: 'info',
<ide> export default () =>
<ide> status: 403
<ide> });
<ide> }
<del> if (!req.user) {
<del> const User = loopback.getModelByType('User');
<del> return User.findById(userId)
<add>
<add> if (isEmpty(req.user)) {
<add> return getUserById(userId)
<ide> .then(user => {
<ide> if (user) {
<ide> user.points = user.progressTimestamps.length;
<ide> export default () =>
<ide> .then(next)
<ide> .catch(next);
<ide> } else {
<del> return next();
<add> return Promise.resolve(next());
<ide> }
<ide> }
<del> return next();
<add> return Promise.resolve(next());
<ide> };
<add>
<add>export function _getUserById(id) {
<add> const User = loopback.getModelByType('User');
<add> return new Promise((resolve, reject) =>
<add> User.findById(id, (err, instance) => {
<add> if (err) {
<add> return reject(err);
<add> }
<add> return resolve(instance);
<add> })
<add> );
<add>}
| 3
|
Java
|
Java
|
fix related classes for issue
|
6bbd921cd4e0da242b2fa97a80fdf488ebf70e44
|
<ide><path>rxjava-core/src/main/java/rx/internal/operators/OnSubscribeFromIterable.java
<ide> public void request(long n) {
<ide> }
<ide> o.onNext(it.next());
<ide> }
<del> o.onCompleted();
<add> if (!o.isUnsubscribed()) {
<add> o.onCompleted();
<add> }
<ide> } else if(n > 0) {
<ide> // backpressure is requested
<ide> long _c = REQUESTED_UPDATER.getAndAdd(this, n);
<ide><path>rxjava-core/src/main/java/rx/internal/operators/OnSubscribeRange.java
<ide> public void request(long n) {
<ide> }
<ide> o.onNext((int) i);
<ide> }
<del> o.onCompleted();
<add> if (!o.isUnsubscribed()) {
<add> o.onCompleted();
<add> }
<ide> } else if (n > 0) {
<ide> // backpressure is requested
<ide> long _c = REQUESTED_UPDATER.getAndAdd(this, n);
<ide><path>rxjava-core/src/main/java/rx/internal/operators/OperatorDoOnEach.java
<ide> public void onCompleted() {
<ide> onError(e);
<ide> return;
<ide> }
<del> observer.onCompleted();
<ide> // Set `done` here so that the error in `doOnEachObserver.onCompleted()` can be noticed by observer
<ide> done = true;
<add> observer.onCompleted();
<ide> }
<ide>
<ide> @Override
<ide> public void onError(Throwable e) {
<ide> if (done) {
<ide> return;
<ide> }
<add> done = true;
<ide> try {
<ide> doOnEachObserver.onError(e);
<ide> } catch (Throwable e2) {
<ide> observer.onError(e2);
<ide> return;
<ide> }
<ide> observer.onError(e);
<del> done = true;
<ide> }
<ide>
<ide> @Override
<ide><path>rxjava-core/src/main/java/rx/internal/operators/OperatorTakeWhile.java
<ide> public Subscriber<? super T> call(final Subscriber<? super T> subscriber) {
<ide>
<ide> private int counter = 0;
<ide>
<add> private boolean done = false;
<add>
<ide> @Override
<ide> public void onNext(T args) {
<ide> boolean isSelected;
<ide> try {
<ide> isSelected = predicate.call(args, counter++);
<ide> } catch (Throwable e) {
<add> done = true;
<ide> subscriber.onError(e);
<ide> unsubscribe();
<ide> return;
<ide> }
<ide> if (isSelected) {
<ide> subscriber.onNext(args);
<ide> } else {
<add> done = true;
<ide> subscriber.onCompleted();
<ide> unsubscribe();
<ide> }
<ide> }
<ide>
<ide> @Override
<ide> public void onCompleted() {
<del> subscriber.onCompleted();
<add> if (!done) {
<add> subscriber.onCompleted();
<add> }
<ide> }
<ide>
<ide> @Override
<ide> public void onError(Throwable e) {
<del> subscriber.onError(e);
<add> if (!done) {
<add> subscriber.onError(e);
<add> }
<ide> }
<ide>
<ide> };
| 4
|
Ruby
|
Ruby
|
use short formula name in bottle commit
|
656c713d6c549a1ae1dc25cdc1f9bb887dfd8440
|
<ide><path>Library/Homebrew/cmd/bottle.rb
<ide> def merge
<ide> end
<ide>
<ide> unless ARGV.include? "--no-commit"
<add> short_name = formula_name.split("/", -1).last
<ide> pkg_version = bottle_hash["formula"]["pkg_version"]
<ide>
<ide> path.parent.cd do
<ide> safe_system "git", "commit", "--no-edit", "--verbose",
<del> "--message=#{formula_name}: #{update_or_add} #{pkg_version} bottle.",
<add> "--message=#{short_name}: #{update_or_add} #{pkg_version} bottle.",
<ide> "--", path
<ide> end
<ide> end
| 1
|
Ruby
|
Ruby
|
ignore processing instructions when parsing html
|
303d379dadaf8c5b0b4518173b2689b003c213a8
|
<ide><path>actionpack/lib/action_controller/vendor/html-scanner/html/tokenizer.rb
<ide> def scan_tag
<ide> elsif @scanner.scan(/!/) # doctype
<ide> tag << @scanner.matched
<ide> tag << consume_quoted_regions
<add> elsif @scanner.scan(/\?/) # processing instructions, drop them
<add> tag = ''
<add> @scanner.scan_until(/\/>/)
<ide> else
<ide> tag << consume_quoted_regions
<ide> end
| 1
|
PHP
|
PHP
|
update tmp constant in tests
|
f80a10a3b0cf385b690238f5cc5913d3e6ef3137
|
<ide><path>lib/Cake/Test/init.php
<ide> define('APP', ROOT . DS . APP_DIR . DS);
<ide> define('WWW_ROOT', APP . WEBROOT_DIR . DS);
<ide> define('TESTS', APP . 'Test' . DS);
<del>define('TMP', APP . 'tmp' . DS);
<add>define('TMP', sys_get_temp_dir() . DS);
<ide> define('LOGS', TMP . 'logs' . DS);
<ide> define('CACHE', TMP . 'cache' . DS);
<ide> define('CAKE_CORE_INCLUDE_PATH', ROOT . DS . 'lib');
<ide> define('CORE_TEST_CASES', CAKE . 'Test' . DS . 'TestCase');
<ide> define('LOG_ERROR', LOG_ERR);
<ide>
<add>
<ide> @mkdir(LOGS);
<ide> @mkdir(CACHE);
<ide>
| 1
|
Go
|
Go
|
simplify the kill "sig" prefix stripping code
|
4bf70317ed3c2467f444eac9b7865b170da6366c
|
<ide><path>server/server.go
<ide> func (srv *Server) ContainerKill(job *engine.Job) engine.Status {
<ide> // The largest legal signal is 31, so let's parse on 5 bits
<ide> sig, err = strconv.ParseUint(job.Args[1], 10, 5)
<ide> if err != nil {
<del> // The signal is not a number, treat it as a string
<del> sig = uint64(signal.SignalMap[job.Args[1]])
<del> if sig == 0 && strings.HasPrefix(job.Args[1], "SIG") {
<del> // If signal is prefixed with SIG, try with it stripped (ie, "SIGKILL", etc)
<del> sig = uint64(signal.SignalMap[job.Args[1][3:]])
<del> }
<add> // The signal is not a number, treat it as a string (either like "KILL" or like "SIGKILL")
<add> sig = uint64(signal.SignalMap[strings.TrimPrefix(job.Args[1], "SIG")])
<ide> if sig == 0 {
<ide> return job.Errorf("Invalid signal: %s", job.Args[1])
<ide> }
| 1
|
Text
|
Text
|
update styleguide keybinding
|
be561f0e5e1090cdcd0b2fae5de8b4cd29cbc21d
|
<ide><path>docs/creating-a-package.md
<ide> Ideally, you won't need much in the way of styling. We've provided a standard
<ide> set of components which define both the colors and UI elements for any package
<ide> that fits into Atom seamlessly. You can view all of Atom's UI components by opening
<ide> the styleguide: open the command palette (`cmd-shift-P`) and search for _styleguide_,
<del>or just type `cmd-ctrl-G`.
<add>or just type `cmd-ctrl-shift-G`.
<ide>
<ide> If you _do_ need special styling, try to keep only structural styles in the package
<ide> stylesheets. If you _must_ specify colors and sizing, these should be taken from
| 1
|
Go
|
Go
|
add tests for loadmanifest digest verification
|
74528be903d2f6c9e07ab9a3931b050d5dae6562
|
<ide><path>graph/manifest_test.go
<ide> import (
<ide> "os"
<ide> "testing"
<ide>
<add> "github.com/docker/distribution/digest"
<ide> "github.com/docker/docker/image"
<ide> "github.com/docker/docker/pkg/tarsum"
<ide> "github.com/docker/docker/registry"
<ide> "github.com/docker/docker/runconfig"
<ide> "github.com/docker/docker/utils"
<add> "github.com/docker/libtrust"
<ide> )
<ide>
<ide> const (
<ide> func TestManifestTarsumCache(t *testing.T) {
<ide> t.Fatalf("Unexpected json value\nExpected:\n%s\nActual:\n%s", v1compat, manifest.History[0].V1Compatibility)
<ide> }
<ide> }
<add>
<add>// TestManifestDigestCheck ensures that loadManifest properly verifies the
<add>// remote and local digest.
<add>func TestManifestDigestCheck(t *testing.T) {
<add> tmp, err := utils.TestDirectory("")
<add> if err != nil {
<add> t.Fatal(err)
<add> }
<add> defer os.RemoveAll(tmp)
<add> store := mkTestTagStore(tmp, t)
<add> defer store.graph.driver.Cleanup()
<add>
<add> archive, err := fakeTar()
<add> if err != nil {
<add> t.Fatal(err)
<add> }
<add> img := &image.Image{ID: testManifestImageID}
<add> if err := store.graph.Register(img, archive); err != nil {
<add> t.Fatal(err)
<add> }
<add> if err := store.Tag(testManifestImageName, testManifestTag, testManifestImageID, false); err != nil {
<add> t.Fatal(err)
<add> }
<add>
<add> if cs, err := img.GetCheckSum(store.graph.ImageRoot(testManifestImageID)); err != nil {
<add> t.Fatal(err)
<add> } else if cs != "" {
<add> t.Fatalf("Non-empty checksum file after register")
<add> }
<add>
<add> // Generate manifest
<add> payload, err := store.newManifest(testManifestImageName, testManifestImageName, testManifestTag)
<add> if err != nil {
<add> t.Fatalf("unexpected error generating test manifest: %v", err)
<add> }
<add>
<add> pk, err := libtrust.GenerateECP256PrivateKey()
<add> if err != nil {
<add> t.Fatalf("unexpected error generating private key: %v", err)
<add> }
<add>
<add> sig, err := libtrust.NewJSONSignature(payload)
<add> if err != nil {
<add> t.Fatalf("error creating signature: %v", err)
<add> }
<add>
<add> if err := sig.Sign(pk); err != nil {
<add> t.Fatalf("error signing manifest bytes: %v", err)
<add> }
<add>
<add> signedBytes, err := sig.PrettySignature("signatures")
<add> if err != nil {
<add> t.Fatalf("error getting signed bytes: %v", err)
<add> }
<add>
<add> dgst, err := digest.FromBytes(payload)
<add> if err != nil {
<add> t.Fatalf("error getting digest of manifest: %v", err)
<add> }
<add>
<add> // use this as the "bad" digest
<add> zeroDigest, err := digest.FromBytes([]byte{})
<add> if err != nil {
<add> t.Fatalf("error making zero digest: %v", err)
<add> }
<add>
<add> // Remote and local match, everything should look good
<add> local, _, _, err := store.loadManifest(signedBytes, dgst.String(), dgst)
<add> if err != nil {
<add> t.Fatalf("unexpected error verifying local and remote digest: %v", err)
<add> }
<add>
<add> if local != dgst {
<add> t.Fatalf("local digest not correctly calculated: %v", err)
<add> }
<add>
<add> // remote and no local, since pulling by tag
<add> local, _, _, err = store.loadManifest(signedBytes, "tag", dgst)
<add> if err != nil {
<add> t.Fatalf("unexpected error verifying tag pull and remote digest: %v", err)
<add> }
<add>
<add> if local != dgst {
<add> t.Fatalf("local digest not correctly calculated: %v", err)
<add> }
<add>
<add> // remote and differing local, this is the most important to fail
<add> local, _, _, err = store.loadManifest(signedBytes, zeroDigest.String(), dgst)
<add> if err == nil {
<add> t.Fatalf("error expected when verifying with differing local digest")
<add> }
<add>
<add> // no remote, no local (by tag)
<add> local, _, _, err = store.loadManifest(signedBytes, "tag", "")
<add> if err != nil {
<add> t.Fatalf("unexpected error verifying manifest without remote digest: %v", err)
<add> }
<add>
<add> if local != dgst {
<add> t.Fatalf("local digest not correctly calculated: %v", err)
<add> }
<add>
<add> // no remote, with local
<add> local, _, _, err = store.loadManifest(signedBytes, dgst.String(), "")
<add> if err != nil {
<add> t.Fatalf("unexpected error verifying manifest without remote digest: %v", err)
<add> }
<add>
<add> if local != dgst {
<add> t.Fatalf("local digest not correctly calculated: %v", err)
<add> }
<add>
<add> // bad remote, we fail the check.
<add> local, _, _, err = store.loadManifest(signedBytes, dgst.String(), zeroDigest)
<add> if err == nil {
<add> t.Fatalf("error expected when verifying with differing remote digest")
<add> }
<add>}
<ide><path>graph/tags_unit_test.go
<ide> import (
<ide> "github.com/docker/docker/daemon/graphdriver"
<ide> _ "github.com/docker/docker/daemon/graphdriver/vfs" // import the vfs driver so it is used in the tests
<ide> "github.com/docker/docker/image"
<add> "github.com/docker/docker/trust"
<ide> "github.com/docker/docker/utils"
<ide> )
<ide>
<ide> func mkTestTagStore(root string, t *testing.T) *TagStore {
<ide> if err != nil {
<ide> t.Fatal(err)
<ide> }
<add>
<add> trust, err := trust.NewTrustStore(root + "/trust")
<add> if err != nil {
<add> t.Fatal(err)
<add> }
<add>
<ide> tagCfg := &TagStoreConfig{
<ide> Graph: graph,
<ide> Events: events.New(),
<add> Trust: trust,
<ide> }
<ide> store, err := NewTagStore(path.Join(root, "tags"), tagCfg)
<ide> if err != nil {
| 2
|
Ruby
|
Ruby
|
make `in_batches` queries to preparable
|
7d5399379cb9ac2d3ffafa28fdc844d7b6c18ab8
|
<ide><path>activerecord/lib/active_record/relation/batches.rb
<ide> def in_batches(of: 1000, start: nil, finish: nil, load: false, error_on_ignore:
<ide> end
<ide> end
<ide>
<del> batch_relation = relation.where(arel_attribute(primary_key).gt(primary_key_offset))
<add> attr = Relation::QueryAttribute.new(primary_key, primary_key_offset, klass.type_for_attribute(primary_key))
<add> batch_relation = relation.where(arel_attribute(primary_key).gt(Arel::Nodes::BindParam.new(attr)))
<ide> end
<ide> end
<ide>
<ide> private
<ide>
<ide> def apply_limits(relation, start, finish)
<del> relation = relation.where(arel_attribute(primary_key).gteq(start)) if start
<del> relation = relation.where(arel_attribute(primary_key).lteq(finish)) if finish
<add> if start
<add> attr = Relation::QueryAttribute.new(primary_key, start, klass.type_for_attribute(primary_key))
<add> relation = relation.where(arel_attribute(primary_key).gteq(Arel::Nodes::BindParam.new(attr)))
<add> end
<add> if finish
<add> attr = Relation::QueryAttribute.new(primary_key, finish, klass.type_for_attribute(primary_key))
<add> relation = relation.where(arel_attribute(primary_key).lteq(Arel::Nodes::BindParam.new(attr)))
<add> end
<ide> relation
<ide> end
<ide>
| 1
|
Text
|
Text
|
add docker commands to local setup
|
1d6c43c612fbb8cc4d7f76d0ab07f63eb59a9960
|
<ide><path>docs/how-to-setup-freecodecamp-locally.md
<ide> A quick reference to the commands that you will need when working locally.
<ide>
<ide> **Docker Build:**
<ide>
<del>> ### Todo: Add docker quick command list
<add>| command | description |
<add>| ------- | ----------- |
<add>| `npm run docker:init` | Prepare containers for installation of dependencies. |
<add>| `npm run docker:install` | Install / re-install all dependencies and bootstraps the different services. |
<add>| `npm run docker:seed` | Parse all the challenge markdown files and inserts them into MongoDB. |
<add>| `npm run docker:develop` | Start the freeCodeCamp API Server and Client Applications. |
<add>| `npm run docker:test:init` | Bootstrap the test container, necessary for testing in docker. |
<add>| `npm run docker:test -- -c "npm run test` | Run all JS tests in the system, including client, server, lint and challenge tests. |
<add>| `npm run docker:test -- -c "npm run test:curriculum` | Run the curriculum test suite. |
<add>| `npm run docker:test -- -c "npm run test:client` | Run the client test suite. |
<add>| `npm run docker:test -- -c "npm run test:server` | Run the server test suite. |
<add>| `npm run docker:clean` | Uninstall all dependencies and cleans up caches. |
<ide>
<ide> **Local Build:**
<ide>
| 1
|
Text
|
Text
|
fix typo in actionpack changelog.md
|
c3d5c9447a49fae7a4507b3512d65cd52f789d99
|
<ide><path>actionpack/CHANGELOG.md
<ide>
<ide> *Gustavo Gutierrez*
<ide>
<del>* Calling `ActionController::Parameters#transform_keys/!` without a block now returns
<add>* Calling `ActionController::Parameters#transform_keys!` without a block now returns
<ide> an enumerator for the parameters instead of the underlying hash.
<ide>
<ide> *Eugene Kenny*
| 1
|
Javascript
|
Javascript
|
add placeholder text to forum posts
|
90fe00937431b68f1af13f31a4a0af6a935ae130
|
<ide><path>client/src/templates/Challenges/redux/create-question-epic.js
<ide> function createQuestionEpic(action$, state$, { window }) {
<ide>
<ide> let textMessage = dedent(
<ide> `**Tell us what's happening:**
<add> Describe your issue in detail here.
<ide>
<ide> ${
<ide> projectFormValues.length
| 1
|
Text
|
Text
|
prepare providers release 0.0.2a1
|
5a439e84eb6c0544dc6c3d6a9f4ceeb2172cd5d0
|
<ide><path>airflow/providers/amazon/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [8afdb6ac6](https://github.com/apache/airflow/commit/8afdb6ac6a7997cb14806bc2734c81c00ed8da97) | 2020-10-26 | Fix spellings (#11825) |
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [6ce855af1](https://github.com/apache/airflow/commit/6ce855af118daeaa4c249669079ab9d9aad23945) | 2020-10-24 | Fix spelling (#11821) |
<add>| [3934ef224](https://github.com/apache/airflow/commit/3934ef22494db6d9613c229aaa82ea6a366b7c2f) | 2020-10-24 | Remove redundant builtins imports (#11809) |
<add>| [4c8e033c0](https://github.com/apache/airflow/commit/4c8e033c0ee7d28963d504a9216205155f20f58f) | 2020-10-24 | Fix spelling and grammar (#11814) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [0df60b773](https://github.com/apache/airflow/commit/0df60b773671ecf8d4e5f582ac2be200cf2a2edd) | 2020-10-23 | Add reattach flag to ECSOperator (#10643) |
<add>| [b9d677cdd](https://github.com/apache/airflow/commit/b9d677cdd660e0be8278a64658e73359276a9682) | 2020-10-22 | Add type hints to aws provider (#11531) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [674368f66](https://github.com/apache/airflow/commit/674368f66cf61b2a105f326f23868ac3aee08807) | 2020-10-19 | Fixes MySQLToS3 float to int conversion (#10437) |
<add>| [0823d46a7](https://github.com/apache/airflow/commit/0823d46a7f267f2e45195a175021825367938add) | 2020-10-16 | Add type annotations for AWS operators and hooks (#11434) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/amazon/README.md
<ide>
<ide> # Package apache-airflow-providers-amazon
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Secrets](#secrets)
<ide> - [Moved secrets](#moved-secrets)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [8afdb6ac6](https://github.com/apache/airflow/commit/8afdb6ac6a7997cb14806bc2734c81c00ed8da97) | 2020-10-26 | Fix spellings (#11825) |
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [6ce855af1](https://github.com/apache/airflow/commit/6ce855af118daeaa4c249669079ab9d9aad23945) | 2020-10-24 | Fix spelling (#11821) |
<add>| [3934ef224](https://github.com/apache/airflow/commit/3934ef22494db6d9613c229aaa82ea6a366b7c2f) | 2020-10-24 | Remove redundant builtins imports (#11809) |
<add>| [4c8e033c0](https://github.com/apache/airflow/commit/4c8e033c0ee7d28963d504a9216205155f20f58f) | 2020-10-24 | Fix spelling and grammar (#11814) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [0df60b773](https://github.com/apache/airflow/commit/0df60b773671ecf8d4e5f582ac2be200cf2a2edd) | 2020-10-23 | Add reattach flag to ECSOperator (#10643) |
<add>| [b9d677cdd](https://github.com/apache/airflow/commit/b9d677cdd660e0be8278a64658e73359276a9682) | 2020-10-22 | Add type hints to aws provider (#11531) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [674368f66](https://github.com/apache/airflow/commit/674368f66cf61b2a105f326f23868ac3aee08807) | 2020-10-19 | Fixes MySQLToS3 float to int conversion (#10437) |
<add>| [0823d46a7](https://github.com/apache/airflow/commit/0823d46a7f267f2e45195a175021825367938add) | 2020-10-16 | Add type annotations for AWS operators and hooks (#11434) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/apache/cassandra/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [0646849e3](https://github.com/apache/airflow/commit/0646849e3dacdc2bc62705ae136f3ad3b16232e9) | 2020-10-14 | Add protocol_version to conn_config for Cassandrahook (#11036) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/apache/cassandra/README.md
<ide>
<ide> # Package apache-airflow-providers-apache-cassandra
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [0646849e3](https://github.com/apache/airflow/commit/0646849e3dacdc2bc62705ae136f3ad3b16232e9) | 2020-10-14 | Add protocol_version to conn_config for Cassandrahook (#11036) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/apache/druid/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/apache/druid/README.md
<ide>
<ide> # Package apache-airflow-providers-apache-druid
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/apache/hdfs/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/apache/hdfs/README.md
<ide>
<ide> # Package apache-airflow-providers-apache-hdfs
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/apache/hive/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/apache/hive/README.md
<ide>
<ide> # Package apache-airflow-providers-apache-hive
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/apache/kylin/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/apache/kylin/README.md
<ide>
<ide> # Package apache-airflow-providers-apache-kylin
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [New hooks](#new-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/apache/livy/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/apache/livy/README.md
<ide>
<ide> # Package apache-airflow-providers-apache-livy
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [New hooks](#new-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/apache/pig/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/apache/pig/README.md
<ide>
<ide> # Package apache-airflow-providers-apache-pig
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/apache/pinot/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/apache/pinot/README.md
<ide>
<ide> # Package apache-airflow-providers-apache-pinot
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/apache/spark/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/apache/spark/README.md
<ide>
<ide> # Package apache-airflow-providers-apache-spark
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/apache/sqoop/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/apache/sqoop/README.md
<ide>
<ide> # Package apache-airflow-providers-apache-sqoop
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/celery/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/celery/README.md
<ide>
<ide> # Package apache-airflow-providers-celery
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Sensors](#sensors)
<ide> - [Moved sensors](#moved-sensors)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/cloudant/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/cloudant/README.md
<ide>
<ide> # Package apache-airflow-providers-cloudant
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/cncf/kubernetes/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [53e606210](https://github.com/apache/airflow/commit/53e6062105be0ae1761a354e2055eb0779d12e73) | 2020-10-21 | Enforce strict rules for yamllint (#11709) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [eee4e30f2](https://github.com/apache/airflow/commit/eee4e30f2caf02e16088ff5d1af1ea380a73e982) | 2020-10-15 | Add better debug logging to K8sexec and K8sPodOp (#11502) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/cncf/kubernetes/README.md
<ide>
<ide> # Package apache-airflow-providers-cncf-kubernetes
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [New hooks](#new-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [53e606210](https://github.com/apache/airflow/commit/53e6062105be0ae1761a354e2055eb0779d12e73) | 2020-10-21 | Enforce strict rules for yamllint (#11709) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [eee4e30f2](https://github.com/apache/airflow/commit/eee4e30f2caf02e16088ff5d1af1ea380a73e982) | 2020-10-15 | Add better debug logging to K8sexec and K8sPodOp (#11502) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/databricks/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/databricks/README.md
<ide>
<ide> # Package apache-airflow-providers-databricks
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/datadog/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/datadog/README.md
<ide>
<ide> # Package apache-airflow-providers-datadog
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/dingding/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [172820db4](https://github.com/apache/airflow/commit/172820db4d2009dd26fa8aef4a864fb8a3d7e78d) | 2020-10-21 | Fix case of GitHub (#11398) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/dingding/README.md
<ide>
<ide> # Package apache-airflow-providers-dingding
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [172820db4](https://github.com/apache/airflow/commit/172820db4d2009dd26fa8aef4a864fb8a3d7e78d) | 2020-10-21 | Fix case of GitHub (#11398) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/discord/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/discord/README.md
<ide>
<ide> # Package apache-airflow-providers-discord
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/docker/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/docker/README.md
<ide>
<ide> # Package apache-airflow-providers-docker
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/elasticsearch/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/elasticsearch/README.md
<ide>
<ide> # Package apache-airflow-providers-elasticsearch
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [New hooks](#new-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/exasol/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/exasol/README.md
<ide>
<ide> # Package apache-airflow-providers-exasol
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [New hooks](#new-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/facebook/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/facebook/README.md
<ide>
<ide> # Package apache-airflow-providers-facebook
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [New hooks](#new-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/ftp/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/ftp/README.md
<ide>
<ide> # Package apache-airflow-providers-ftp
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/google/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------|
<add>| [240c7d4d7](https://github.com/apache/airflow/commit/240c7d4d72aac8f6aab98f5913e8f54c4f1372ff) | 2020-10-26 | Google Memcached hooks - improve protobuf messages handling (#11743) |
<add>| [8afdb6ac6](https://github.com/apache/airflow/commit/8afdb6ac6a7997cb14806bc2734c81c00ed8da97) | 2020-10-26 | Fix spellings (#11825) |
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [6ce855af1](https://github.com/apache/airflow/commit/6ce855af118daeaa4c249669079ab9d9aad23945) | 2020-10-24 | Fix spelling (#11821) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [727c739af](https://github.com/apache/airflow/commit/727c739afb565d4d394a8faedc969334cb8e738e) | 2020-10-22 | Improve Cloud Memorystore for Redis example (#11735) |
<add>| [1da8379c9](https://github.com/apache/airflow/commit/1da8379c913843834353b44861c62f332a461bdf) | 2020-10-22 | Fix static checks after merging #10121 (#11737) |
<add>| [91503308c](https://github.com/apache/airflow/commit/91503308c723b186ce6f4026f2a3e2c21030f6e5) | 2020-10-22 | Add Google Cloud Memorystore Memcached Operators (#10121) |
<add>| [950c16d0b](https://github.com/apache/airflow/commit/950c16d0b0ab67bb7af11909de751029faf0313a) | 2020-10-21 | Retry requests in case of error in Google ML Engine Hook (#11712) |
<add>| [2bfc53b5e](https://github.com/apache/airflow/commit/2bfc53b5eb67406d418371b74dc9bc5a07be238e) | 2020-10-21 | Fix doc errors in google provider files. (#11713) |
<add>| [53e606210](https://github.com/apache/airflow/commit/53e6062105be0ae1761a354e2055eb0779d12e73) | 2020-10-21 | Enforce strict rules for yamllint (#11709) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [2d854c350](https://github.com/apache/airflow/commit/2d854c3505ccad66e9a7d94267e51bed800433c2) | 2020-10-19 | Add service_account to Google ML Engine operator (#11619) |
<add>| [46a121fb7](https://github.com/apache/airflow/commit/46a121fb7b77c0964e053b58750e2d8bc2bd0b2a) | 2020-10-18 | docs: Update Bigquery clustering docstrings (#11232) |
<add>| [49c58147f](https://github.com/apache/airflow/commit/49c58147fed8a52869d0b0ecc00c102c11972ad0) | 2020-10-18 | Strict type checking for provider Google (#11609) |
<add>| [0823d46a7](https://github.com/apache/airflow/commit/0823d46a7f267f2e45195a175021825367938add) | 2020-10-16 | Add type annotations for AWS operators and hooks (#11434) |
<add>| [3c10ca650](https://github.com/apache/airflow/commit/3c10ca6504be37fabff9a10caefea3fe4df31a02) | 2020-10-16 | Add DataflowStartFlexTemplateOperator (#8550) |
<add>| [8865d14df](https://github.com/apache/airflow/commit/8865d14df4d58dd5f1a4d2ff81c77469959f175a) | 2020-10-16 | Strict type checking for provider google cloud (#11548) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/google/README.md
<ide>
<ide> # Package apache-airflow-providers-google
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Secrets](#secrets)
<ide> - [Moved secrets](#moved-secrets)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> You can install this package on top of an existing airflow 2.* installation via
<ide> | google-cloud-kms | >=1.2.1,<2.0.0 |
<ide> | google-cloud-language | >=1.1.1,<2.0.0 |
<ide> | google-cloud-logging | >=1.14.0,<2.0.0 |
<add>| google-cloud-memcache | >=0.2.0 |
<ide> | google-cloud-monitoring | >=0.34.0,<2.0.0 |
<ide> | google-cloud-pubsub | >=1.0.0,<2.0.0 |
<ide> | google-cloud-redis | >=0.3.0,<2.0.0 |
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide> | [cloud.operators.cloud_memorystore.CloudMemorystoreGetInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) |
<ide> | [cloud.operators.cloud_memorystore.CloudMemorystoreImportOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) |
<ide> | [cloud.operators.cloud_memorystore.CloudMemorystoreListInstancesOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) |
<add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedApplyParametersOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) |
<add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedCreateInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) |
<add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedDeleteInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) |
<add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedGetInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) |
<add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedListInstancesOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) |
<add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedUpdateInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) |
<add>| [cloud.operators.cloud_memorystore.CloudMemorystoreMemcachedUpdateParametersOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) |
<ide> | [cloud.operators.cloud_memorystore.CloudMemorystoreScaleInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) |
<ide> | [cloud.operators.cloud_memorystore.CloudMemorystoreUpdateInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_memorystore.py) |
<ide> | [cloud.operators.cloud_storage_transfer_service.CloudDataTransferServiceGCSToGCSOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/cloud_storage_transfer_service.py) |
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide> | [cloud.operators.datacatalog.CloudDataCatalogUpdateTagOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/datacatalog.py) |
<ide> | [cloud.operators.datacatalog.CloudDataCatalogUpdateTagTemplateFieldOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/datacatalog.py) |
<ide> | [cloud.operators.datacatalog.CloudDataCatalogUpdateTagTemplateOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/datacatalog.py) |
<add>| [cloud.operators.dataflow.DataflowStartFlexTemplateOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/dataflow.py) |
<ide> | [cloud.operators.datafusion.CloudDataFusionCreateInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/datafusion.py) |
<ide> | [cloud.operators.datafusion.CloudDataFusionCreatePipelineOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/datafusion.py) |
<ide> | [cloud.operators.datafusion.CloudDataFusionDeleteInstanceOperator](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/datafusion.py) |
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide> | [cloud.hooks.automl.CloudAutoMLHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/automl.py) |
<ide> | [cloud.hooks.bigquery_dts.BiqQueryDataTransferServiceHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/bigquery_dts.py) |
<ide> | [cloud.hooks.cloud_memorystore.CloudMemorystoreHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/cloud_memorystore.py) |
<add>| [cloud.hooks.cloud_memorystore.CloudMemorystoreMemcachedHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/cloud_memorystore.py) |
<ide> | [cloud.hooks.datacatalog.CloudDataCatalogHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/datacatalog.py) |
<ide> | [cloud.hooks.datafusion.DataFusionHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/datafusion.py) |
<ide> | [cloud.hooks.dataprep.GoogleDataprepHook](https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/hooks/dataprep.py) |
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------|
<add>| [240c7d4d7](https://github.com/apache/airflow/commit/240c7d4d72aac8f6aab98f5913e8f54c4f1372ff) | 2020-10-26 | Google Memcached hooks - improve protobuf messages handling (#11743) |
<add>| [8afdb6ac6](https://github.com/apache/airflow/commit/8afdb6ac6a7997cb14806bc2734c81c00ed8da97) | 2020-10-26 | Fix spellings (#11825) |
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [6ce855af1](https://github.com/apache/airflow/commit/6ce855af118daeaa4c249669079ab9d9aad23945) | 2020-10-24 | Fix spelling (#11821) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [727c739af](https://github.com/apache/airflow/commit/727c739afb565d4d394a8faedc969334cb8e738e) | 2020-10-22 | Improve Cloud Memorystore for Redis example (#11735) |
<add>| [1da8379c9](https://github.com/apache/airflow/commit/1da8379c913843834353b44861c62f332a461bdf) | 2020-10-22 | Fix static checks after merging #10121 (#11737) |
<add>| [91503308c](https://github.com/apache/airflow/commit/91503308c723b186ce6f4026f2a3e2c21030f6e5) | 2020-10-22 | Add Google Cloud Memorystore Memcached Operators (#10121) |
<add>| [950c16d0b](https://github.com/apache/airflow/commit/950c16d0b0ab67bb7af11909de751029faf0313a) | 2020-10-21 | Retry requests in case of error in Google ML Engine Hook (#11712) |
<add>| [2bfc53b5e](https://github.com/apache/airflow/commit/2bfc53b5eb67406d418371b74dc9bc5a07be238e) | 2020-10-21 | Fix doc errors in google provider files. (#11713) |
<add>| [53e606210](https://github.com/apache/airflow/commit/53e6062105be0ae1761a354e2055eb0779d12e73) | 2020-10-21 | Enforce strict rules for yamllint (#11709) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [2d854c350](https://github.com/apache/airflow/commit/2d854c3505ccad66e9a7d94267e51bed800433c2) | 2020-10-19 | Add service_account to Google ML Engine operator (#11619) |
<add>| [46a121fb7](https://github.com/apache/airflow/commit/46a121fb7b77c0964e053b58750e2d8bc2bd0b2a) | 2020-10-18 | docs: Update Bigquery clustering docstrings (#11232) |
<add>| [49c58147f](https://github.com/apache/airflow/commit/49c58147fed8a52869d0b0ecc00c102c11972ad0) | 2020-10-18 | Strict type checking for provider Google (#11609) |
<add>| [0823d46a7](https://github.com/apache/airflow/commit/0823d46a7f267f2e45195a175021825367938add) | 2020-10-16 | Add type annotations for AWS operators and hooks (#11434) |
<add>| [3c10ca650](https://github.com/apache/airflow/commit/3c10ca6504be37fabff9a10caefea3fe4df31a02) | 2020-10-16 | Add DataflowStartFlexTemplateOperator (#8550) |
<add>| [8865d14df](https://github.com/apache/airflow/commit/8865d14df4d58dd5f1a4d2ff81c77469959f175a) | 2020-10-16 | Strict type checking for provider google cloud (#11548) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/grpc/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/grpc/README.md
<ide>
<ide> # Package apache-airflow-providers-grpc
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [New hooks](#new-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/hashicorp/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/hashicorp/README.md
<ide>
<ide> # Package apache-airflow-providers-hashicorp
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Secrets](#secrets)
<ide> - [Moved secrets](#moved-secrets)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/http/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [3cddc1182](https://github.com/apache/airflow/commit/3cddc11821ff8f9ed0811384c0643f756a2b3dfa) | 2020-10-16 | Updated template_fields_rendereds for PostgresOperator and SimpleHttpOperator (#11555) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/http/README.md
<ide>
<ide> # Package apache-airflow-providers-http
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [3cddc1182](https://github.com/apache/airflow/commit/3cddc11821ff8f9ed0811384c0643f756a2b3dfa) | 2020-10-16 | Updated template_fields_rendereds for PostgresOperator and SimpleHttpOperator (#11555) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/imap/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/imap/README.md
<ide>
<ide> # Package apache-airflow-providers-imap
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/jdbc/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/jdbc/README.md
<ide>
<ide> # Package apache-airflow-providers-jdbc
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/jenkins/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [8afdb6ac6](https://github.com/apache/airflow/commit/8afdb6ac6a7997cb14806bc2734c81c00ed8da97) | 2020-10-26 | Fix spellings (#11825) |
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/jenkins/README.md
<ide>
<ide> # Package apache-airflow-providers-jenkins
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [8afdb6ac6](https://github.com/apache/airflow/commit/8afdb6ac6a7997cb14806bc2734c81c00ed8da97) | 2020-10-26 | Fix spellings (#11825) |
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/jira/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/jira/README.md
<ide>
<ide> # Package apache-airflow-providers-jira
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/microsoft/azure/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [6ce855af1](https://github.com/apache/airflow/commit/6ce855af118daeaa4c249669079ab9d9aad23945) | 2020-10-24 | Fix spelling (#11821) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [f8ff217e2](https://github.com/apache/airflow/commit/f8ff217e2f2152bbb9fc701ff4c0b6eb447ad65c) | 2020-10-18 | Fix incorrect typing and move config args out of extra connection config to operator args (#11635) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/microsoft/azure/README.md
<ide>
<ide> # Package apache-airflow-providers-microsoft-azure
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Secrets](#secrets)
<ide> - [New secrets](#new-secrets)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [6ce855af1](https://github.com/apache/airflow/commit/6ce855af118daeaa4c249669079ab9d9aad23945) | 2020-10-24 | Fix spelling (#11821) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [f8ff217e2](https://github.com/apache/airflow/commit/f8ff217e2f2152bbb9fc701ff4c0b6eb447ad65c) | 2020-10-18 | Fix incorrect typing and move config args out of extra connection config to operator args (#11635) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/microsoft/mssql/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [765d29ecc](https://github.com/apache/airflow/commit/765d29ecc9fd6a3220efa0a6c4ce10848f5cbf82) | 2020-10-15 | Pymssql is maintained again (#11537) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/microsoft/mssql/README.md
<ide>
<ide> # Package apache-airflow-providers-microsoft-mssql
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> You can install this package on top of an existing airflow 2.* installation via
<ide>
<ide> | PIP package | Version required |
<ide> |:--------------|:-------------------|
<del>| pymssql | ~=2.1.1 |
<add>| pymssql | ~=2.1,>=2.1.5 |
<ide>
<ide> ## Cross provider package dependencies
<ide>
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [765d29ecc](https://github.com/apache/airflow/commit/765d29ecc9fd6a3220efa0a6c4ce10848f5cbf82) | 2020-10-15 | Pymssql is maintained again (#11537) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/microsoft/winrm/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/microsoft/winrm/README.md
<ide>
<ide> # Package apache-airflow-providers-microsoft-winrm
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/mongo/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/mongo/README.md
<ide>
<ide> # Package apache-airflow-providers-mongo
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/mysql/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/mysql/README.md
<ide>
<ide> # Package apache-airflow-providers-mysql
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/odbc/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/odbc/README.md
<ide>
<ide> # Package apache-airflow-providers-odbc
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [New hooks](#new-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/openfaas/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/openfaas/README.md
<ide>
<ide> # Package apache-airflow-providers-openfaas
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/opsgenie/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/opsgenie/README.md
<ide>
<ide> # Package apache-airflow-providers-opsgenie
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/oracle/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/oracle/README.md
<ide>
<ide> # Package apache-airflow-providers-oracle
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/pagerduty/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [3ee618623](https://github.com/apache/airflow/commit/3ee618623be6079ed177da793b490cb7436d5cb6) | 2020-10-20 | Switch PagerdutyHook from pypd to use pdpyras instead (#11151) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/pagerduty/README.md
<ide>
<ide> # Package apache-airflow-providers-pagerduty
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [3ee618623](https://github.com/apache/airflow/commit/3ee618623be6079ed177da793b490cb7436d5cb6) | 2020-10-20 | Switch PagerdutyHook from pypd to use pdpyras instead (#11151) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/papermill/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [cb7c67dea](https://github.com/apache/airflow/commit/cb7c67dea9cd9b9c5de10e355b63039446003149) | 2020-10-20 | Fix example DAGs in pip packages (#11687) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/papermill/README.md
<ide>
<ide> # Package apache-airflow-providers-papermill
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Operators](#operators)
<ide> - [Moved operators](#moved-operators)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [cb7c67dea](https://github.com/apache/airflow/commit/cb7c67dea9cd9b9c5de10e355b63039446003149) | 2020-10-20 | Fix example DAGs in pip packages (#11687) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/plexus/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [56d72e3ff](https://github.com/apache/airflow/commit/56d72e3ff8798a2662847355d1b73b2c1f57b31f) | 2020-10-24 | Replace non-empty sets with set literals (#11810) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/plexus/README.md
<ide>
<ide> # Package apache-airflow-providers-plexus
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [New hooks](#new-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [56d72e3ff](https://github.com/apache/airflow/commit/56d72e3ff8798a2662847355d1b73b2c1f57b31f) | 2020-10-24 | Replace non-empty sets with set literals (#11810) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/postgres/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [a4dc11fae](https://github.com/apache/airflow/commit/a4dc11fae63d56bc6cbb029525113948862fd45d) | 2020-10-19 | Change to pass all extra connection paramaters to psycopg2 (#11019) |
<add>| [3cddc1182](https://github.com/apache/airflow/commit/3cddc11821ff8f9ed0811384c0643f756a2b3dfa) | 2020-10-16 | Updated template_fields_rendereds for PostgresOperator and SimpleHttpOperator (#11555) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/postgres/README.md
<ide>
<ide> # Package apache-airflow-providers-postgres
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:---------------------------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [a4dc11fae](https://github.com/apache/airflow/commit/a4dc11fae63d56bc6cbb029525113948862fd45d) | 2020-10-19 | Change to pass all extra connection paramaters to psycopg2 (#11019) |
<add>| [3cddc1182](https://github.com/apache/airflow/commit/3cddc11821ff8f9ed0811384c0643f756a2b3dfa) | 2020-10-16 | Updated template_fields_rendereds for PostgresOperator and SimpleHttpOperator (#11555) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/presto/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [1543923c1](https://github.com/apache/airflow/commit/1543923c197f658533ca0a0bb259b59a002cce43) | 2020-10-20 | Add Kerberos Auth for PrestoHook (#10488) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/presto/README.md
<ide>
<ide> # Package apache-airflow-providers-presto
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [1543923c1](https://github.com/apache/airflow/commit/1543923c197f658533ca0a0bb259b59a002cce43) | 2020-10-20 | Add Kerberos Auth for PrestoHook (#10488) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/qubole/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/qubole/README.md
<ide>
<ide> # Package apache-airflow-providers-qubole
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/redis/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/redis/README.md
<ide>
<ide> # Package apache-airflow-providers-redis
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/salesforce/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/salesforce/README.md
<ide>
<ide> # Package apache-airflow-providers-salesforce
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [New hooks](#new-hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/samba/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/samba/README.md
<ide>
<ide> # Package apache-airflow-providers-samba
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/segment/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/segment/README.md
<ide>
<ide> # Package apache-airflow-providers-segment
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/sftp/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [ae791e191](https://github.com/apache/airflow/commit/ae791e19163b4ee6e84940d2567adbf0a2626fb4) | 2020-10-21 | Fix formatting errors introduced in #11720 (#11733) |
<add>| [1fb3c28e1](https://github.com/apache/airflow/commit/1fb3c28e1a4fd54c9d83dccd413659a7a87c7315) | 2020-10-21 | Add support for setting ciphers for SFTPHook (#11720) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/sftp/README.md
<ide>
<ide> # Package apache-airflow-providers-sftp
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [ae791e191](https://github.com/apache/airflow/commit/ae791e19163b4ee6e84940d2567adbf0a2626fb4) | 2020-10-21 | Fix formatting errors introduced in #11720 (#11733) |
<add>| [1fb3c28e1](https://github.com/apache/airflow/commit/1fb3c28e1a4fd54c9d83dccd413659a7a87c7315) | 2020-10-21 | Add support for setting ciphers for SFTPHook (#11720) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/singularity/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/singularity/README.md
<ide>
<ide> # Package apache-airflow-providers-singularity
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Operators](#operators)
<ide> - [New operators](#new-operators)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/slack/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [4fb5c017f](https://github.com/apache/airflow/commit/4fb5c017fe5ca41ed95547a857c9c39efc4f1476) | 2020-10-21 | Check response status in slack webhook hook. (#11620) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/slack/README.md
<ide>
<ide> # Package apache-airflow-providers-slack
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [4fb5c017f](https://github.com/apache/airflow/commit/4fb5c017fe5ca41ed95547a857c9c39efc4f1476) | 2020-10-21 | Check response status in slack webhook hook. (#11620) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/snowflake/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/snowflake/README.md
<ide>
<ide> # Package apache-airflow-providers-snowflake
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [New hooks](#new-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [483068745](https://github.com/apache/airflow/commit/48306874538eea7cfd42358d5ebb59705204bfc4) | 2020-10-24 | Use Python 3 style super classes (#11806) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/sqlite/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/sqlite/README.md
<ide>
<ide> # Package apache-airflow-providers-sqlite
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/ssh/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/ssh/README.md
<ide>
<ide> # Package apache-airflow-providers-ssh
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/vertica/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/vertica/README.md
<ide>
<ide> # Package apache-airflow-providers-vertica
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/yandex/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/yandex/README.md
<ide>
<ide> # Package apache-airflow-providers-yandex
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [New hooks](#new-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
<ide><path>airflow/providers/zendesk/PROVIDER_CHANGES_0.0.2a1.md
<add>
<add>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<ide><path>airflow/providers/zendesk/README.md
<ide>
<ide> # Package apache-airflow-providers-zendesk
<ide>
<del>Release: 0.0.1
<add>Release: 0.0.2a1
<ide>
<ide> **Table of contents**
<ide>
<ide> Release: 0.0.1
<ide> - [Hooks](#hooks)
<ide> - [Moved hooks](#moved-hooks)
<ide> - [Releases](#releases)
<add> - [Release 0.0.2a1](#release-002a1)
<ide> - [Release 0.0.1](#release-001)
<ide>
<ide> ## Provider package
<ide> in [Naming conventions for provider packages](https://github.com/apache/airflow/
<ide>
<ide> ## Releases
<ide>
<add>### Release 0.0.2a1
<add>
<add>| Commit | Committed | Subject |
<add>|:-----------------------------------------------------------------------------------------------|:------------|:-------------------------------------------------------------------|
<add>| [872b1566a](https://github.com/apache/airflow/commit/872b1566a11cb73297e657ff325161721b296574) | 2020-10-25 | Generated backport providers readmes/setup for 2020.10.29 (#11826) |
<add>| [349b0811c](https://github.com/apache/airflow/commit/349b0811c3022605426ba57d30936240a7c2848a) | 2020-10-20 | Add D200 pydocstyle check (#11688) |
<add>| [16e712971](https://github.com/apache/airflow/commit/16e7129719f1c0940aef2a93bed81368e997a746) | 2020-10-13 | Added support for provider packages for Airflow 2.0 (#11487) |
<add>
<add>
<ide> ### Release 0.0.1
<ide>
<ide> | Commit | Committed | Subject |
| 118
|
Ruby
|
Ruby
|
remove unused column types override
|
ea6f28c8a114396376bafe26e1af5d2ca3df6fdb
|
<ide><path>activerecord/lib/active_record/core.rb
<ide> def initialize(attributes = nil, options = {})
<ide> end
<ide>
<ide> @attributes = defaults
<del> @column_types_override = nil
<ide> @column_types = self.class.column_types
<ide>
<ide> init_internals
<ide> def initialize(attributes = nil, options = {})
<ide> # post.title # => 'hello world'
<ide> def init_with(coder)
<ide> @attributes = coder['attributes']
<del> @column_types_override = coder['column_types']
<ide> @column_types = self.class.column_types
<ide>
<ide> init_internals
<ide> def encode_with(coder)
<ide> # FIXME: Remove this when we better serialize attributes
<ide> coder['raw_attributes'] = attributes_before_type_cast
<ide> coder['attributes'] = @attributes
<del> coder['column_types'] = @column_types_override
<ide> coder['new_record'] = new_record?
<ide> end
<ide>
<ide><path>activerecord/lib/active_record/persistence.rb
<ide> def instantiate(attributes, column_types = {})
<ide> type = column_types.fetch(name) { klass.type_for_attribute(name) }
<ide> h[name] = Attribute.from_database(value, type)
<ide> end
<del>
<del> klass.allocate.init_with(
<del> 'attributes' => attributes,
<del> 'column_types' => column_types,
<del> 'new_record' => false,
<del> )
<add> klass.allocate.init_with('attributes' => attributes, 'new_record' => false)
<ide> end
<ide>
<ide> private
<ide> def reload(options = nil)
<ide>
<ide> @attributes.update(fresh_object.instance_variable_get('@attributes'))
<ide>
<del> @column_types = self.class.column_types
<del> @column_types_override = fresh_object.instance_variable_get('@column_types_override')
<add> @column_types = self.class.column_types
<ide> self
<ide> end
<ide>
| 2
|
Python
|
Python
|
move trainers to core/
|
cbbba228150140a14eea2361868bf66700feff4b
|
<ide><path>official/core/base_task_test.py
<add># Lint as: python3
<add># Copyright 2020 The TensorFlow Authors. All Rights Reserved.
<add>#
<add># Licensed under the Apache License, Version 2.0 (the "License");
<add># you may not use this file except in compliance with the License.
<add># You may obtain a copy of the License at
<add>#
<add># http://www.apache.org/licenses/LICENSE-2.0
<add>#
<add># Unless required by applicable law or agreed to in writing, software
<add># distributed under the License is distributed on an "AS IS" BASIS,
<add># WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
<add># See the License for the specific language governing permissions and
<add># limitations under the License.
<add># ==============================================================================
<add>"""Tests for tensorflow_models.core.base_task."""
<add>
<add>import functools
<add>
<add>from absl.testing import parameterized
<add>import tensorflow as tf
<add>
<add>from tensorflow.python.distribute import combinations
<add>from tensorflow.python.distribute import strategy_combinations
<add>from official.utils.testing import mock_task
<add>
<add>
<add>def all_strategy_combinations():
<add> return combinations.combine(
<add> distribution=[
<add> strategy_combinations.default_strategy,
<add> strategy_combinations.tpu_strategy,
<add> strategy_combinations.one_device_strategy_gpu,
<add> ],
<add> mode='eager',
<add> )
<add>
<add>
<add>class TaskKerasTest(tf.test.TestCase, parameterized.TestCase):
<add>
<add> @combinations.generate(all_strategy_combinations())
<add> def test_task_with_step_override(self, distribution):
<add> with distribution.scope():
<add> task = mock_task.MockTask()
<add> model = task.build_model()
<add> model = task.compile_model(
<add> model,
<add> optimizer=tf.keras.optimizers.SGD(learning_rate=1e-3),
<add> metrics=task.build_metrics(),
<add> train_step=task.train_step,
<add> validation_step=task.validation_step)
<add>
<add> dataset = task.build_inputs(params=None)
<add> logs = model.fit(dataset, epochs=1, steps_per_epoch=2)
<add> self.assertIn('loss', logs.history)
<add> self.assertIn('acc', logs.history)
<add>
<add> # Without specifying metrics through compile.
<add> with distribution.scope():
<add> train_metrics = task.build_metrics(training=True)
<add> val_metrics = task.build_metrics(training=False)
<add> model = task.build_model()
<add> model = task.compile_model(
<add> model,
<add> optimizer=tf.keras.optimizers.SGD(learning_rate=1e-3),
<add> train_step=functools.partial(task.train_step, metrics=train_metrics),
<add> validation_step=functools.partial(
<add> task.validation_step, metrics=val_metrics))
<add> logs = model.fit(dataset, epochs=1, steps_per_epoch=2)
<add> self.assertIn('loss', logs.history)
<add> self.assertIn('acc', logs.history)
<add>
<add> def test_task_with_fit(self):
<add> task = mock_task.MockTask()
<add> model = task.build_model()
<add> model = task.compile_model(
<add> model,
<add> optimizer=tf.keras.optimizers.SGD(learning_rate=1e-3),
<add> loss=tf.keras.losses.CategoricalCrossentropy(),
<add> metrics=task.build_metrics())
<add> dataset = task.build_inputs(params=None)
<add> logs = model.fit(dataset, epochs=1, steps_per_epoch=2)
<add> self.assertIn('loss', logs.history)
<add> self.assertIn('acc', logs.history)
<add> self.assertLen(model.evaluate(dataset, steps=1), 2)
<add>
<add> def test_task_invalid_compile(self):
<add> task = mock_task.MockTask()
<add> model = task.build_model()
<add> with self.assertRaises(ValueError):
<add> _ = task.compile_model(
<add> model,
<add> optimizer=tf.keras.optimizers.SGD(learning_rate=1e-3),
<add> loss=tf.keras.losses.CategoricalCrossentropy(),
<add> metrics=task.build_metrics(),
<add> train_step=task.train_step)
<add>
<add>
<add>if __name__ == '__main__':
<add> tf.test.main()
<ide><path>official/core/base_trainer.py
<add># Lint as: python3
<add># Copyright 2020 The TensorFlow Authors. All Rights Reserved.
<add>#
<add># Licensed under the Apache License, Version 2.0 (the "License");
<add># you may not use this file except in compliance with the License.
<add># You may obtain a copy of the License at
<add>#
<add># http://www.apache.org/licenses/LICENSE-2.0
<add>#
<add># Unless required by applicable law or agreed to in writing, software
<add># distributed under the License is distributed on an "AS IS" BASIS,
<add># WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
<add># See the License for the specific language governing permissions and
<add># limitations under the License.
<add># ==============================================================================
<add>"""Standard Trainer implementation.
<add>
<add>The base trainer implements the Orbit `StandardTrainable` and
<add>`StandardEvaluable` interfaces. Trainers inside this project should be
<add>interchangable and independent on model architectures and tasks.
<add>"""
<add>import gin
<add>import orbit
<add>import tensorflow as tf
<add>
<add>from official.core import base_task
<add>from official.modeling import optimization
<add>from official.modeling import performance
<add>from official.modeling.hyperparams import config_definitions
<add>
<add>
<add>ExperimentConfig = config_definitions.ExperimentConfig
<add>
<add>
<add>@gin.configurable
<add>class Trainer(orbit.StandardTrainer, orbit.StandardEvaluator):
<add> """Implements the common trainer shared for TensorFlow models."""
<add>
<add> def __init__(self,
<add> config: ExperimentConfig,
<add> task: base_task.Task,
<add> train: bool = True,
<add> evaluate: bool = True,
<add> model=None,
<add> optimizer=None):
<add> """Initialize common trainer for TensorFlow models.
<add>
<add> Args:
<add> config: An `ExperimentConfig` instance specifying experiment config.
<add> task: A base_task.Task instance.
<add> train: bool, whether or not this trainer will be used for training.
<add> default to True.
<add> evaluate: bool, whether or not this trainer will be used for evaluation.
<add> default to True.
<add> model: tf.keras.Model instance. If provided, it will be used instead
<add> of building model using task.build_model(). Default to None.
<add> optimizer: tf.keras.optimizers.Optimizer instance. If provided, it will
<add> used instead of the optimizer from config. Default to None.
<add> """
<add> # Gets the current distribution strategy. If not inside any strategy scope,
<add> # it gets a single-replica no-op strategy.
<add> self._strategy = tf.distribute.get_strategy()
<add> self._config = config
<add> self._task = task
<add>
<add> self._model = model or task.build_model()
<add>
<add> if optimizer is None:
<add> opt_factory = optimization.OptimizerFactory(
<add> config.trainer.optimizer_config)
<add> self._optimizer = opt_factory.build_optimizer(
<add> opt_factory.build_learning_rate())
<add> else:
<add> self._optimizer = optimizer
<add>
<add> # Configuring optimizer when loss_scale is set in runtime config. This helps
<add> # avoiding overflow/underflow for float16 computations.
<add> if config.runtime.loss_scale:
<add> self._optimizer = performance.configure_optimizer(
<add> self._optimizer,
<add> use_float16=config.runtime.mixed_precision_dtype == 'float16',
<add> loss_scale=config.runtime.loss_scale)
<add>
<add> # global_step increases by 1 after each training iteration.
<add> # We should have global_step.numpy() == self.optimizer.iterations.numpy()
<add> # when there is only 1 optimizer.
<add> self._global_step = orbit.utils.create_global_step()
<add> if hasattr(self.model, 'checkpoint_items'):
<add> checkpoint_items = self.model.checkpoint_items
<add> else:
<add> checkpoint_items = {}
<add> self._checkpoint = tf.train.Checkpoint(
<add> global_step=self.global_step, model=self.model,
<add> optimizer=self.optimizer, **checkpoint_items)
<add>
<add> self._train_loss = tf.keras.metrics.Mean('training_loss', dtype=tf.float32)
<add> self._validation_loss = tf.keras.metrics.Mean(
<add> 'validation_loss', dtype=tf.float32)
<add> self._train_metrics = self.task.build_metrics(
<add> training=True) + self.model.metrics
<add> self._validation_metrics = self.task.build_metrics(
<add> training=False) + self.model.metrics
<add>
<add> if train:
<add> train_dataset = orbit.utils.make_distributed_dataset(
<add> self.strategy, self.task.build_inputs, self.config.task.train_data)
<add> orbit.StandardTrainer.__init__(
<add> self,
<add> train_dataset,
<add> options=orbit.StandardTrainerOptions(
<add> use_tf_while_loop=config.trainer.train_tf_while_loop,
<add> use_tf_function=config.trainer.train_tf_function,
<add> use_tpu_summary_optimization=config.trainer.allow_tpu_summary))
<add>
<add> if evaluate:
<add> eval_dataset = orbit.utils.make_distributed_dataset(
<add> self.strategy, self.task.build_inputs,
<add> self.config.task.validation_data)
<add> orbit.StandardEvaluator.__init__(
<add> self,
<add> eval_dataset,
<add> options=orbit.StandardEvaluatorOptions(
<add> use_tf_function=config.trainer.eval_tf_function))
<add>
<add> @property
<add> def strategy(self):
<add> return self._strategy
<add>
<add> @property
<add> def config(self):
<add> return self._config
<add>
<add> @property
<add> def task(self):
<add> return self._task
<add>
<add> @property
<add> def model(self):
<add> return self._model
<add>
<add> @property
<add> def optimizer(self):
<add> return self._optimizer
<add>
<add> @property
<add> def global_step(self):
<add> return self._global_step
<add>
<add> @property
<add> def train_loss(self):
<add> """Accesses the training loss metric object."""
<add> return self._train_loss
<add>
<add> @property
<add> def validation_loss(self):
<add> """Accesses the validation loss metric object."""
<add> return self._validation_loss
<add>
<add> @property
<add> def train_metrics(self):
<add> """Accesses all training metric objects."""
<add> return self._train_metrics
<add>
<add> @property
<add> def validation_metrics(self):
<add> """Accesses all validation metric metric objects."""
<add> return self._validation_metrics
<add>
<add> def initialize(self):
<add> """A callback function.
<add>
<add> This function will be called when no checkpoint found for the model.
<add> If there is a checkpoint, the checkpoint will be loaded and this function
<add> will not be called. Tasks may use this callback function to load a
<add> pretrained checkpoint, saved under a directory other than the model_dir.
<add> """
<add> self.task.initialize(self.model)
<add>
<add> @property
<add> def checkpoint(self):
<add> """Accesses the training checkpoint."""
<add> return self._checkpoint
<add>
<add> def train_loop_end(self):
<add> """See base class."""
<add> logs = {}
<add> for metric in self.train_metrics + [self.train_loss]:
<add> logs[metric.name] = metric.result()
<add> metric.reset_states()
<add> if callable(self.optimizer.learning_rate):
<add> logs['learning_rate'] = self.optimizer.learning_rate(self.global_step)
<add> else:
<add> logs['learning_rate'] = self.optimizer.learning_rate
<add> return logs
<add>
<add> def train_step(self, iterator):
<add> """See base class."""
<add>
<add> def step_fn(inputs):
<add> logs = self.task.train_step(
<add> inputs,
<add> model=self.model,
<add> optimizer=self.optimizer,
<add> metrics=self.train_metrics)
<add> self._train_loss.update_state(logs[self.task.loss])
<add> self.global_step.assign_add(1)
<add>
<add> self.strategy.run(step_fn, args=(next(iterator),))
<add>
<add> def eval_begin(self):
<add> """Sets up metrics."""
<add> for metric in self.validation_metrics + [self.validation_loss]:
<add> metric.reset_states()
<add>
<add> def eval_step(self, iterator):
<add> """See base class."""
<add>
<add> def step_fn(inputs):
<add> logs = self.task.validation_step(
<add> inputs, model=self.model, metrics=self.validation_metrics)
<add> self._validation_loss.update_state(logs[self.task.loss])
<add> return logs
<add>
<add> distributed_outputs = self.strategy.run(step_fn, args=(next(iterator),))
<add> return tf.nest.map_structure(self.strategy.experimental_local_results,
<add> distributed_outputs)
<add>
<add> def eval_end(self, aggregated_logs=None):
<add> """Processes evaluation results."""
<add> logs = {}
<add> for metric in self.validation_metrics + [self.validation_loss]:
<add> logs[metric.name] = metric.result()
<add> if aggregated_logs:
<add> metrics = self.task.reduce_aggregated_logs(aggregated_logs)
<add> logs.update(metrics)
<add> return logs
<add>
<add> def eval_reduce(self, state=None, step_outputs=None):
<add> return self.task.aggregate_logs(state, step_outputs)
<ide><path>official/core/base_trainer_test.py
<add># Lint as: python3
<add># Copyright 2020 The TensorFlow Authors. All Rights Reserved.
<add>#
<add># Licensed under the Apache License, Version 2.0 (the "License");
<add># you may not use this file except in compliance with the License.
<add># You may obtain a copy of the License at
<add>#
<add># http://www.apache.org/licenses/LICENSE-2.0
<add>#
<add># Unless required by applicable law or agreed to in writing, software
<add># distributed under the License is distributed on an "AS IS" BASIS,
<add># WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
<add># See the License for the specific language governing permissions and
<add># limitations under the License.
<add># ==============================================================================
<add>"""Tests for tensorflow_models.core.trainers.trainer."""
<add># pylint: disable=g-direct-tensorflow-import
<add>from absl.testing import parameterized
<add>import tensorflow as tf
<add>
<add>from tensorflow.python.distribute import combinations
<add>from tensorflow.python.distribute import strategy_combinations
<add>from official.core import base_trainer as trainer_lib
<add>from official.modeling.hyperparams import config_definitions as cfg
<add>from official.utils.testing import mock_task
<add>
<add>
<add>def all_strategy_combinations():
<add> return combinations.combine(
<add> distribution=[
<add> strategy_combinations.default_strategy,
<add> strategy_combinations.tpu_strategy,
<add> strategy_combinations.one_device_strategy_gpu,
<add> ],
<add> mode='eager',
<add> )
<add>
<add>
<add>class TrainerTest(tf.test.TestCase, parameterized.TestCase):
<add>
<add> def setUp(self):
<add> super().setUp()
<add> self._config = cfg.ExperimentConfig(
<add> trainer=cfg.TrainerConfig(
<add> optimizer_config=cfg.OptimizationConfig(
<add> {'optimizer': {
<add> 'type': 'sgd'
<add> },
<add> 'learning_rate': {
<add> 'type': 'constant'
<add> }})))
<add>
<add> def create_test_trainer(self):
<add> task = mock_task.MockTask()
<add> trainer = trainer_lib.Trainer(self._config, task)
<add> return trainer
<add>
<add> @combinations.generate(all_strategy_combinations())
<add> def test_trainer_train(self, distribution):
<add> with distribution.scope():
<add> trainer = self.create_test_trainer()
<add> logs = trainer.train(tf.convert_to_tensor(5, dtype=tf.int32))
<add> self.assertIn('training_loss', logs)
<add> self.assertIn('learning_rate', logs)
<add>
<add> @combinations.generate(all_strategy_combinations())
<add> def test_trainer_validate(self, distribution):
<add> with distribution.scope():
<add> trainer = self.create_test_trainer()
<add> logs = trainer.evaluate(tf.convert_to_tensor(5, dtype=tf.int32))
<add> self.assertIn('validation_loss', logs)
<add> self.assertEqual(logs['acc'], 5. * distribution.num_replicas_in_sync)
<add>
<add> @combinations.generate(
<add> combinations.combine(
<add> mixed_precision_dtype=['float32', 'bfloat16', 'float16'],
<add> loss_scale=[None, 'dynamic', 128, 256],
<add> ))
<add> def test_configure_optimizer(self, mixed_precision_dtype, loss_scale):
<add> config = cfg.ExperimentConfig(
<add> runtime=cfg.RuntimeConfig(
<add> mixed_precision_dtype=mixed_precision_dtype, loss_scale=loss_scale),
<add> trainer=cfg.TrainerConfig(
<add> optimizer_config=cfg.OptimizationConfig(
<add> {'optimizer': {
<add> 'type': 'sgd'
<add> },
<add> 'learning_rate': {
<add> 'type': 'constant'
<add> }})))
<add> task = mock_task.MockTask()
<add> trainer = trainer_lib.Trainer(config, task)
<add> if mixed_precision_dtype != 'float16':
<add> self.assertIsInstance(trainer.optimizer, tf.keras.optimizers.SGD)
<add> elif mixed_precision_dtype == 'float16' and loss_scale is None:
<add> self.assertIsInstance(trainer.optimizer, tf.keras.optimizers.SGD)
<add> else:
<add> self.assertIsInstance(
<add> trainer.optimizer,
<add> tf.keras.mixed_precision.experimental.LossScaleOptimizer)
<add>
<add> metrics = trainer.train(tf.convert_to_tensor(5, dtype=tf.int32))
<add> self.assertIn('training_loss', metrics)
<add>
<add>
<add>if __name__ == '__main__':
<add> tf.test.main()
<ide><path>official/utils/testing/mock_task.py
<add># Lint as: python3
<add># Copyright 2020 The TensorFlow Authors. All Rights Reserved.
<add>#
<add># Licensed under the Apache License, Version 2.0 (the "License");
<add># you may not use this file except in compliance with the License.
<add># You may obtain a copy of the License at
<add>#
<add># http://www.apache.org/licenses/LICENSE-2.0
<add>#
<add># Unless required by applicable law or agreed to in writing, software
<add># distributed under the License is distributed on an "AS IS" BASIS,
<add># WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
<add># See the License for the specific language governing permissions and
<add># limitations under the License.
<add># ==============================================================================
<add>"""Mock task for testing."""
<add>
<add>import dataclasses
<add>import numpy as np
<add>import tensorflow as tf
<add>
<add>from official.core import base_task
<add>from official.core import exp_factory
<add>from official.core import task_factory
<add>from official.modeling.hyperparams import config_definitions as cfg
<add>
<add>
<add>class MockModel(tf.keras.Model):
<add>
<add> def __init__(self, network):
<add> super().__init__()
<add> self.network = network
<add>
<add> def call(self, inputs):
<add> outputs = self.network(inputs)
<add> self.add_loss(tf.reduce_mean(outputs))
<add> return outputs
<add>
<add>
<add>@dataclasses.dataclass
<add>class MockTaskConfig(cfg.TaskConfig):
<add> pass
<add>
<add>
<add>@task_factory.register_task_cls(MockTaskConfig)
<add>class MockTask(base_task.Task):
<add> """Mock task object for testing."""
<add>
<add> def __init__(self, params=None, logging_dir=None):
<add> super().__init__(params=params, logging_dir=logging_dir)
<add>
<add> def build_model(self, *arg, **kwargs):
<add> inputs = tf.keras.layers.Input(shape=(2,), name="random", dtype=tf.float32)
<add> outputs = tf.keras.layers.Dense(1)(inputs)
<add> network = tf.keras.Model(inputs=inputs, outputs=outputs)
<add> return MockModel(network)
<add>
<add> def build_metrics(self, training: bool = True):
<add> del training
<add> return [tf.keras.metrics.Accuracy(name="acc")]
<add>
<add> def build_inputs(self, params):
<add>
<add> def generate_data(_):
<add> x = tf.zeros(shape=(2,), dtype=tf.float32)
<add> label = tf.zeros([1], dtype=tf.int32)
<add> return x, label
<add>
<add> dataset = tf.data.Dataset.range(1)
<add> dataset = dataset.repeat()
<add> dataset = dataset.map(
<add> generate_data, num_parallel_calls=tf.data.experimental.AUTOTUNE)
<add> return dataset.prefetch(buffer_size=1).batch(2, drop_remainder=True)
<add>
<add> def aggregate_logs(self, state, step_outputs):
<add> if state is None:
<add> state = {}
<add> for key, value in step_outputs.items():
<add> if key not in state:
<add> state[key] = []
<add> state[key].append(
<add> np.concatenate([np.expand_dims(v.numpy(), axis=0) for v in value]))
<add> return state
<add>
<add> def reduce_aggregated_logs(self, aggregated_logs):
<add> for k, v in aggregated_logs.items():
<add> aggregated_logs[k] = np.sum(np.stack(v, axis=0))
<add> return aggregated_logs
<add>
<add>
<add>@exp_factory.register_config_factory("mock")
<add>def mock_experiment() -> cfg.ExperimentConfig:
<add> config = cfg.ExperimentConfig(
<add> task=MockTaskConfig(), trainer=cfg.TrainerConfig())
<add> return config
| 4
|
Python
|
Python
|
fix assertion error
|
7e8f1053bb07f4f73914b5784a87de6dcaea6bcb
|
<ide><path>tests/test_appctx.py
<ide> def index():
<ide> c = app.test_client()
<ide> res = c.get('/')
<ide> assert res.status_code == 200
<del> assert res.data == u''
<add> assert res.data == b''
<ide> assert called == ['request', 'app']
<ide>
<ide>
| 1
|
Go
|
Go
|
fix internal ipvlan network to work in swarm
|
c7f8bfa0010f6f92c875e0c11df7ecd397eca35c
|
<ide><path>libnetwork/drivers/ipvlan/ipvlan_network.go
<ide> func parseNetworkOptions(id string, option options.Generic) (*configuration, err
<ide> return nil, err
<ide> }
<ide> }
<del> // setting the parent to "" will trigger an isolated network dummy parent link
<ide> if val, ok := option[netlabel.Internal]; ok {
<ide> if internal, ok := val.(bool); ok && internal {
<ide> config.Internal = true
<del> // empty --parent= and --internal are handled the same.
<del> config.Parent = ""
<ide> }
<ide> }
<ide> return config, nil
| 1
|
PHP
|
PHP
|
fix bug in url generator
|
3ef7bf2bf4fc7c7b240ea1f0f1bb2e364603f4ca
|
<ide><path>src/Illuminate/Routing/UrlGenerator.php
<ide> public function secureAsset($path)
<ide> */
<ide> protected function getScheme($secure)
<ide> {
<del> if (is_null($secure))
<add> if ( ! $secure)
<ide> {
<ide> return $this->request->getScheme().'://';
<ide> }
<ide><path>tests/Routing/RoutingUrlGeneratorTest.php
<ide> public function testBasicGeneration()
<ide>
<ide> $this->assertEquals('http://www.foo.com/foo/bar', $url->asset('foo/bar'));
<ide> $this->assertEquals('https://www.foo.com/foo/bar', $url->asset('foo/bar', true));
<del> }
<add> }
<ide>
<ide>
<ide> public function testBasicRouteGeneration()
<ide> public function testBasicRouteGeneration()
<ide> }
<ide>
<ide>
<add> public function testRoutesMaintainRequestScheme()
<add> {
<add> $url = new UrlGenerator(
<add> $routes = new Illuminate\Routing\RouteCollection,
<add> $request = Illuminate\Http\Request::create('https://www.foo.com/')
<add> );
<add>
<add> /**
<add> * Named Routes
<add> */
<add> $route = new Illuminate\Routing\Route(array('GET'), 'foo/bar', array('as' => 'foo'));
<add> $routes->add($route);
<add>
<add> $this->assertEquals('https://www.foo.com/foo/bar', $url->route('foo'));
<add> }
<add>
<add>
<ide> public function testRoutesWithDomains()
<ide> {
<del> $url = new UrlGenerator(
<add> $url = new UrlGenerator(
<ide> $routes = new Illuminate\Routing\RouteCollection,
<ide> $request = Illuminate\Http\Request::create('http://www.foo.com/')
<ide> );
| 2
|
Javascript
|
Javascript
|
add uniquename for css module classes
|
bdd5d3d9d391faebf826cb951aa5a95d7f96ff07
|
<ide><path>lib/css/CssLoadingRuntimeModule.js
<ide> class CssLoadingRuntimeModule extends RuntimeModule {
<ide> const code = Template.asString([
<ide> "link = document.createElement('link');",
<ide> uniqueName
<del> ? 'link.setAttribute("data-webpack", dataWebpackPrefix + key);'
<add> ? 'link.setAttribute("data-webpack", uniqueName + ":" + key);'
<ide> : "",
<ide> "link.setAttribute(loadingAttribute, 1);",
<ide> 'link.rel = "stylesheet";',
<ide> class CssLoadingRuntimeModule extends RuntimeModule {
<ide> id => `${JSON.stringify(id)}:0`
<ide> ).join(",")}};`,
<ide> "",
<add> uniqueName
<add> ? `var uniqueName = ${JSON.stringify(
<add> runtimeTemplate.outputOptions.uniqueName
<add> )};`
<add> : "// data-webpack is not used as build has no uniqueName",
<ide> `var loadCssChunkData = ${runtimeTemplate.basicFunction("chunkId, link", [
<ide> 'var data, token = "", token2, exports = {}, exportsWithId = [], i = 0, cc = 1;',
<ide> "try { if(!link) link = loadStylesheet(chunkId); data = link.sheet.cssRules; data = data[data.length - 1].style; } catch(e) { data = getComputedStyle(document.head); }",
<del> 'data = data.getPropertyValue("--webpack-" + chunkId);',
<add> `data = data.getPropertyValue(${
<add> uniqueName
<add> ? runtimeTemplate.concatenation(
<add> "--webpack-",
<add> { expr: "uniqueName" },
<add> "-",
<add> { expr: "chunkId" }
<add> )
<add> : runtimeTemplate.concatenation("--webpack-", { expr: "chunkId" })
<add> });`,
<ide> "if(!data) return;",
<ide> "for(; cc; i++) {",
<ide> Template.indent([
<ide> class CssLoadingRuntimeModule extends RuntimeModule {
<ide> )}) { exports[token2.replace(/^_/, "")] = token.replace(/^_/, ""); token = ""; }`,
<ide> `else if(cc == ${cc(
<ide> "/"
<del> )}) { token = token.replace(/^_/, ""); exports[token] = token + "_"; exportsWithId.push(token); token = ""; }`,
<add> )}) { token = token.replace(/^_/, ""); exports[token] = token; exportsWithId.push(token); token = ""; }`,
<ide> `else if(!cc || cc == ${cc(
<ide> ","
<ide> )}) { token = token.replace(/^_/, ""); exportsWithId.forEach(${runtimeTemplate.expressionFunction(
<del> "exports[x] += token",
<add> `exports[x] = ${
<add> uniqueName
<add> ? runtimeTemplate.concatenation(
<add> { expr: "uniqueName" },
<add> "-",
<add> { expr: "token" },
<add> "-",
<add> { expr: "exports[x]" }
<add> )
<add> : runtimeTemplate.concatenation({ expr: "token" }, "-", {
<add> expr: "exports[x]"
<add> })
<add> }`,
<ide> "x"
<ide> )}); ${RuntimeGlobals.makeNamespaceObject}(exports); ${
<ide> RuntimeGlobals.moduleFactories
<ide> class CssLoadingRuntimeModule extends RuntimeModule {
<ide> "installedChunks[chunkId] = 0;"
<ide> ])}`,
<ide> 'var loadingAttribute = "data-webpack-loading";',
<del> uniqueName
<del> ? `var dataWebpackPrefix = ${JSON.stringify(uniqueName + ":")};`
<del> : "// data-webpack is not used as build has no uniqueName",
<ide> `var loadStylesheet = ${runtimeTemplate.basicFunction(
<ide> "chunkId, url, done" + (withHmr ? ", hmr" : ""),
<ide> [
<ide> class CssLoadingRuntimeModule extends RuntimeModule {
<ide> "var l = links[i];",
<ide> `if(l.getAttribute("href") == url${
<ide> uniqueName
<del> ? ' || l.getAttribute("data-webpack") == dataWebpackPrefix + key'
<add> ? ' || l.getAttribute("data-webpack") == uniqueName + ":" + key'
<ide> : ""
<ide> }) { link = l; break; }`
<ide> ]),
<ide><path>lib/css/CssModulesPlugin.js
<ide> class CssModulesPlugin {
<ide> this.renderChunk({
<ide> chunk,
<ide> chunkGraph,
<del> codeGenerationResults
<add> codeGenerationResults,
<add> uniqueName: compilation.outputOptions.uniqueName
<ide> }),
<ide> filenameTemplate: CssModulesPlugin.getChunkFilenameTemplate(
<ide> chunk,
<ide> class CssModulesPlugin {
<ide> return [...(cssImports || []), ...(cssContent || [])];
<ide> }
<ide>
<del> renderChunk({ chunk, chunkGraph, codeGenerationResults }) {
<add> renderChunk({ uniqueName, chunk, chunkGraph, codeGenerationResults }) {
<ide> const modules = this.getOrderedChunkCssModules(chunk, chunkGraph);
<ide> const source = new ConcatSource();
<ide> const metaData = [];
<ide> class CssModulesPlugin {
<ide> `${
<ide> exports
<ide> ? Array.from(exports, ([n, v]) =>
<del> v === `${n}_${moduleId}`
<add> v === `${uniqueName ? uniqueName + "-" : ""}${moduleId}-${n}`
<ide> ? `${escapeCss(n)}/`
<ide> : `${escapeCss(n)}(${escapeCss(v)})`
<ide> ).join("")
<ide> class CssModulesPlugin {
<ide> }
<ide> }
<ide> source.add(
<del> `head{--webpack-${escapeCss(chunk.id, true)}:${metaData.join(",")};}`
<add> `head{--webpack-${escapeCss(
<add> (uniqueName ? uniqueName + "-" : "") + chunk.id,
<add> true
<add> )}:${metaData.join(",")};}`
<ide> );
<ide> return source;
<ide> }
<ide><path>lib/dependencies/CssLocalIdentifierDependency.js
<ide> CssLocalIdentifierDependency.Template = class CssLocalIdentifierDependencyTempla
<ide> apply(
<ide> dependency,
<ide> source,
<del> { module, moduleGraph, chunkGraph, runtime, cssExports }
<add> { module, moduleGraph, chunkGraph, runtime, runtimeTemplate, cssExports }
<ide> ) {
<ide> const dep = /** @type {CssLocalIdentifierDependency} */ (dependency);
<ide> const used = moduleGraph
<ide> .getExportInfo(module, dep.name)
<ide> .getUsedName(dep.name, runtime);
<ide> const moduleId = chunkGraph.getModuleId(module);
<del> const identifier = (used || "_") + "_" + moduleId;
<add> const identifier =
<add> (runtimeTemplate.outputOptions.uniqueName
<add> ? runtimeTemplate.outputOptions.uniqueName + "-"
<add> : "") +
<add> moduleId +
<add> "-" +
<add> (used || "");
<ide> source.replace(
<ide> dep.range[0],
<ide> dep.range[1] - 1,
<ide><path>test/configCases/css/css-modules/index.js
<ide> it("should allow to create css modules", done => {
<ide> try {
<ide> expect(x).toEqual({
<ide> global: undefined,
<del> class: prod ? "S_491" : "class_./style.module.css",
<add> class: prod ? "my-app-491-S" : "./style.module.css-class",
<ide> local: prod
<del> ? "Zw_491 yl_491 J__491 gc_491"
<del> : "local1_./style.module.css local2_./style.module.css local3_./style.module.css local4_./style.module.css",
<add> ? "my-app-491-Zw my-app-491-yl my-app-491-J_ my-app-491-gc"
<add> : "./style.module.css-local1 ./style.module.css-local2 ./style.module.css-local3 ./style.module.css-local4",
<ide> local2: prod
<del> ? "Xg_491 AY_491"
<del> : "local5_./style.module.css local6_./style.module.css",
<add> ? "my-app-491-Xg my-app-491-AY"
<add> : "./style.module.css-local5 ./style.module.css-local6",
<ide> nested: prod
<del> ? "RX_491 undefined X2_491"
<del> : "nested1_./style.module.css undefined nested3_./style.module.css"
<add> ? "my-app-491-RX undefined my-app-491-X2"
<add> : "./style.module.css-nested1 undefined ./style.module.css-nested3"
<ide> });
<ide> } catch (e) {
<ide> return done(e);
<ide><path>test/configCases/css/css-modules/webpack.config.js
<ide> module.exports = [
<ide> {
<ide> target: "web",
<ide> mode: "production",
<add> output: {
<add> uniqueName: "my-app"
<add> },
<ide> experiments: {
<ide> css: true
<ide> }
| 5
|
Javascript
|
Javascript
|
fix dependencies when snapshot is missing
|
8ee39ac3ac651a0d2093292ad16e1a215b8816f1
|
<ide><path>lib/NormalModule.js
<ide> class NormalModule extends Module {
<ide> contextDependencies: contextDeps,
<ide> missingDependencies: missingDeps
<ide> } = this.buildInfo;
<del> fileDependencies.addAll(fileDeps);
<del> contextDependencies.addAll(contextDeps);
<del> missingDependencies.addAll(missingDeps);
<add> if (fileDeps !== undefined) fileDependencies.addAll(fileDeps);
<add> if (contextDeps !== undefined) contextDependencies.addAll(contextDeps);
<add> if (missingDeps !== undefined) missingDependencies.addAll(missingDeps);
<ide> }
<del> if (buildDeps) {
<add> if (buildDeps !== undefined) {
<ide> buildDependencies.addAll(buildDeps);
<ide> }
<ide> }
| 1
|
Javascript
|
Javascript
|
create name/dashname at seed
|
b1e34f8cc8554458200009fb2b2f2c82b05823f1
|
<ide><path>seed/index.js
<ide> require('babel/register');
<ide> require('dotenv').load();
<ide> var fs = require('fs'),
<add> _ = require('lodash'),
<ide> path = require('path'),
<ide> app = require('../server/server'),
<ide> nonprofits = require('./nonprofits.json'),
<ide> jobs = require('./jobs.json');
<ide>
<del>var challangesRegex = /^(bonfire:|waypoint:|zipline:|basejump:|hike:)/i;
<del>
<ide> function getFilesFor(dir) {
<ide> return fs.readdirSync(path.join(__dirname, '/' + dir));
<ide> }
<ide> Challenge.destroyAll(function(err, info) {
<ide> console.log('Deleted ', info);
<ide> }
<ide> challenges.forEach(function(file) {
<del> var challenges = require('./challenges/' + file).challenges
<add> var challengeSpec = require('./challenges/' + file);
<add> var challenges = challengeSpec.challenges
<ide> .map(function(challenge) {
<ide> // NOTE(berks): add title for displaying in views
<del> //challenge.title = challenge.name.replace(challangesRegex, '').trim();
<del> challenge.name = challenge.title.replace(/[^a-zA-Z0-9 ]/g, ''); // Remove non-alphanumwhitespace chars
<del> challenge.dashedName = challenge.name.replace(/\s/g, '-'); // Replace with dasherize();
<add> challenge.name =
<add> _.capitalize(challenge.type) +
<add> ': ' +
<add> challenge.title.replace(/[^a-zA-Z0-9\s]/g, '');
<add> challenge.dashedName = challenge.name.toLowerCase().replace(/\s/g, '-');
<ide> return challenge;
<ide> });
<ide>
| 1
|
Javascript
|
Javascript
|
remove redundant code
|
416499c872ba3db45a4dd4b88e015b679931c15f
|
<ide><path>lib/timers.js
<ide> exports.enroll = function(item, msecs) {
<ide> // it will reset its timeout.
<ide> exports.active = function(item) {
<ide> var msecs = item._idleTimeout;
<del> if (msecs >= 0) {
<del> var list = lists[msecs];
<del> if (!list || L.isEmpty(list)) {
<del> insert(item, msecs);
<del> } else {
<del> item._idleStart = Timer.now();
<del> L.append(list, item);
<del> }
<del> }
<add> if (msecs >= 0)
<add> insert(item, msecs);
<ide> };
<ide>
<ide>
| 1
|
Ruby
|
Ruby
|
use ||= instead of "return ... if ... "
|
4f119ad85ff77386312b4a5c3289ec53a6f29769
|
<ide><path>Library/Homebrew/linkage_checker.rb
<ide> def unexpected_broken_dylibs
<ide> end
<ide>
<ide> def unexpected_present_dylibs
<del> return @unexpected_present_dylibs if @unexpected_present_dylibs
<del>
<del> @unexpected_present_dylibs = @formula.class.allowed_missing_libraries.reject do |allowed_missing_lib|
<add> @unexpected_present_dylibs ||= @formula.class.allowed_missing_libraries.reject do |allowed_missing_lib|
<ide> @broken_dylibs.any? do |broken_lib|
<ide> case allowed_missing_lib
<ide> when Regexp
| 1
|
Go
|
Go
|
use mock for search
|
9a0d7fe0182da541cc99eab9a4930616792e95c3
|
<ide><path>registry/registry.go
<ide> func (r *Registry) PushImageJSONIndex(remote string, imgList []*ImgData, validat
<ide>
<ide> func (r *Registry) SearchRepositories(term string) (*SearchResults, error) {
<ide> utils.Debugf("Index server: %s", r.indexEndpoint)
<del> u := IndexServerAddress() + "search?q=" + url.QueryEscape(term)
<add> u := r.indexEndpoint + "search?q=" + url.QueryEscape(term)
<ide> req, err := r.reqFactory.NewRequest("GET", u, nil)
<ide> if err != nil {
<ide> return nil, err
<ide><path>registry/registry_mock_test.go
<ide> func handlerAuth(w http.ResponseWriter, r *http.Request) {
<ide> }
<ide>
<ide> func handlerSearch(w http.ResponseWriter, r *http.Request) {
<del> writeResponse(w, "{}", 200)
<add> result := &SearchResults{
<add> Query: "fakequery",
<add> NumResults: 1,
<add> Results: []SearchResult{{Name: "fakeimage", StarCount: 42}},
<add> }
<add> writeResponse(w, result, 200)
<ide> }
<ide>
<ide> func TestPing(t *testing.T) {
<ide><path>registry/registry_test.go
<ide> func TestPushImageJSONIndex(t *testing.T) {
<ide>
<ide> func TestSearchRepositories(t *testing.T) {
<ide> r := spawnTestRegistry(t)
<del> results, err := r.SearchRepositories("supercalifragilisticepsialidocious")
<add> results, err := r.SearchRepositories("fakequery")
<ide> if err != nil {
<ide> t.Fatal(err)
<ide> }
<ide> if results == nil {
<ide> t.Fatal("Expected non-nil SearchResults object")
<ide> }
<del> assertEqual(t, results.NumResults, 0, "Expected 0 search results")
<add> assertEqual(t, results.NumResults, 1, "Expected 1 search results")
<add> assertEqual(t, results.Query, "fakequery", "Expected 'fakequery' as query")
<add> assertEqual(t, results.Results[0].StarCount, 42, "Expected 'fakeimage' a ot hae 42 stars")
<ide> }
<ide>
<ide> func TestValidRepositoryName(t *testing.T) {
| 3
|
Javascript
|
Javascript
|
fix lint error
|
a582a1a188b981f8d72d25d26827c837c0ad8be9
|
<ide><path>src/get-release-channel.js
<ide> module.exports = function(version) {
<ide> // This matches stable, dev (with or without commit hash) and any other
<ide> // release channel following the pattern '1.00.0-channel0'
<del> const match = version.match(
<del> /\d+\.\d+\.\d+(-([a-z]+)(\d+|-\w{4,})?)?$/
<del> );
<add> const match = version.match(/\d+\.\d+\.\d+(-([a-z]+)(\d+|-\w{4,})?)?$/);
<ide> if (!match) {
<ide> return 'unrecognized';
<ide> } else if (match[2]) {
<ide> return match[2];
<ide> }
<ide>
<ide> return 'stable';
<del>}
<add>};
| 1
|
Mixed
|
Python
|
add (noun chunks) syntax iterators for danish
|
e3222fdec9efcc9a4a5d0d9b2f92099fbf36ed7b
|
<ide><path>.github/contributors/ophelielacroix.md
<add># spaCy contributor agreement
<add>
<add>This spaCy Contributor Agreement (**"SCA"**) is based on the
<add>[Oracle Contributor Agreement](http://www.oracle.com/technetwork/oca-405177.pdf).
<add>The SCA applies to any contribution that you make to any product or project
<add>managed by us (the **"project"**), and sets out the intellectual property rights
<add>you grant to us in the contributed materials. The term **"us"** shall mean
<add>[ExplosionAI GmbH](https://explosion.ai/legal). The term
<add>**"you"** shall mean the person or entity identified below.
<add>
<add>If you agree to be bound by these terms, fill in the information requested
<add>below and include the filled-in version with your first pull request, under the
<add>folder [`.github/contributors/`](/.github/contributors/). The name of the file
<add>should be your GitHub username, with the extension `.md`. For example, the user
<add>example_user would create the file `.github/contributors/example_user.md`.
<add>
<add>Read this agreement carefully before signing. These terms and conditions
<add>constitute a binding legal agreement.
<add>
<add>## Contributor Agreement
<add>
<add>1. The term "contribution" or "contributed materials" means any source code,
<add>object code, patch, tool, sample, graphic, specification, manual,
<add>documentation, or any other material posted or submitted by you to the project.
<add>
<add>2. With respect to any worldwide copyrights, or copyright applications and
<add>registrations, in your contribution:
<add>
<add> * you hereby assign to us joint ownership, and to the extent that such
<add> assignment is or becomes invalid, ineffective or unenforceable, you hereby
<add> grant to us a perpetual, irrevocable, non-exclusive, worldwide, no-charge,
<add> royalty-free, unrestricted license to exercise all rights under those
<add> copyrights. This includes, at our option, the right to sublicense these same
<add> rights to third parties through multiple levels of sublicensees or other
<add> licensing arrangements;
<add>
<add> * you agree that each of us can do all things in relation to your
<add> contribution as if each of us were the sole owners, and if one of us makes
<add> a derivative work of your contribution, the one who makes the derivative
<add> work (or has it made will be the sole owner of that derivative work;
<add>
<add> * you agree that you will not assert any moral rights in your contribution
<add> against us, our licensees or transferees;
<add>
<add> * you agree that we may register a copyright in your contribution and
<add> exercise all ownership rights associated with it; and
<add>
<add> * you agree that neither of us has any duty to consult with, obtain the
<add> consent of, pay or render an accounting to the other for any use or
<add> distribution of your contribution.
<add>
<add>3. With respect to any patents you own, or that you can license without payment
<add>to any third party, you hereby grant to us a perpetual, irrevocable,
<add>non-exclusive, worldwide, no-charge, royalty-free license to:
<add>
<add> * make, have made, use, sell, offer to sell, import, and otherwise transfer
<add> your contribution in whole or in part, alone or in combination with or
<add> included in any product, work or materials arising out of the project to
<add> which your contribution was submitted, and
<add>
<add> * at our option, to sublicense these same rights to third parties through
<add> multiple levels of sublicensees or other licensing arrangements.
<add>
<add>4. Except as set out above, you keep all right, title, and interest in your
<add>contribution. The rights that you grant to us under these terms are effective
<add>on the date you first submitted a contribution to us, even if your submission
<add>took place before the date you sign these terms.
<add>
<add>5. You covenant, represent, warrant and agree that:
<add>
<add> * Each contribution that you submit is and shall be an original work of
<add> authorship and you can legally grant the rights set out in this SCA;
<add>
<add> * to the best of your knowledge, each contribution will not violate any
<add> third party's copyrights, trademarks, patents, or other intellectual
<add> property rights; and
<add>
<add> * each contribution shall be in compliance with U.S. export control laws and
<add> other applicable export and import laws. You agree to notify us if you
<add> become aware of any circumstance which would make any of the foregoing
<add> representations inaccurate in any respect. We may publicly disclose your
<add> participation in the project, including the fact that you have signed the SCA.
<add>
<add>6. This SCA is governed by the laws of the State of California and applicable
<add>U.S. Federal law. Any choice of law rules will not apply.
<add>
<add>7. Please place an “x” on one of the applicable statement below. Please do NOT
<add>mark both statements:
<add>
<add> * [X] I am signing on behalf of myself as an individual and no other person
<add> or entity, including my employer, has or will have rights with respect to my
<add> contributions.
<add>
<add> * [ ] I am signing on behalf of my employer or a legal entity and I have the
<add> actual authority to contractually bind that entity.
<add>
<add>## Contributor Details
<add>
<add>| Field | Entry |
<add>|-------------------------------|-----------------|
<add>| Name | Ophélie Lacroix |
<add>| Company name (if applicable) | |
<add>| Title or role (if applicable) | |
<add>| Date | |
<add>| GitHub username | ophelielacroix |
<add>| Website (optional) | |
<ide><path>spacy/lang/da/__init__.py
<ide> from .lex_attrs import LEX_ATTRS
<ide> from .morph_rules import MORPH_RULES
<ide> from ..tag_map import TAG_MAP
<add>from .syntax_iterators import SYNTAX_ITERATORS
<ide>
<ide> from ..tokenizer_exceptions import BASE_EXCEPTIONS
<ide> from ...language import Language
<ide> class DanishDefaults(Language.Defaults):
<ide> suffixes = TOKENIZER_SUFFIXES
<ide> tag_map = TAG_MAP
<ide> stop_words = STOP_WORDS
<add> syntax_iterators = SYNTAX_ITERATORS
<ide>
<ide>
<ide> class Danish(Language):
<ide><path>spacy/lang/da/syntax_iterators.py
<add># coding: utf8
<add>from __future__ import unicode_literals
<add>
<add>from ...symbols import NOUN, PROPN, PRON, VERB, AUX
<add>from ...errors import Errors
<add>
<add>
<add>def noun_chunks(doclike):
<add> def is_verb_token(tok):
<add> return tok.pos in [VERB, AUX]
<add>
<add> def next_token(tok):
<add> try:
<add> return tok.nbor()
<add> except IndexError:
<add> return None
<add>
<add> def get_left_bound(doc, root):
<add> left_bound = root
<add> for tok in reversed(list(root.lefts)):
<add> if tok.dep in np_left_deps:
<add> left_bound = tok
<add> return left_bound
<add>
<add> def get_right_bound(doc, root):
<add> right_bound = root
<add> for tok in root.rights:
<add> if tok.dep in np_right_deps:
<add> right = get_right_bound(doc, tok)
<add> if list(
<add> filter(
<add> lambda t: is_verb_token(t) or t.dep in stop_deps,
<add> doc[root.i : right.i],
<add> )
<add> ):
<add> break
<add> else:
<add> right_bound = right
<add> return right_bound
<add>
<add> def get_bounds(doc, root):
<add> return get_left_bound(doc, root), get_right_bound(doc, root)
<add>
<add> doc = doclike.doc
<add>
<add> if not doc.is_parsed:
<add> raise ValueError(Errors.E029)
<add>
<add> if not len(doc):
<add> return
<add>
<add> left_labels = [
<add> "det",
<add> "fixed",
<add> "nmod:poss",
<add> "amod",
<add> "flat",
<add> "goeswith",
<add> "nummod",
<add> "appos",
<add> ]
<add> right_labels = ["fixed", "nmod:poss", "amod", "flat", "goeswith", "nummod", "appos"]
<add> stop_labels = ["punct"]
<add>
<add> np_label = doc.vocab.strings.add("NP")
<add> np_left_deps = [doc.vocab.strings.add(label) for label in left_labels]
<add> np_right_deps = [doc.vocab.strings.add(label) for label in right_labels]
<add> stop_deps = [doc.vocab.strings.add(label) for label in stop_labels]
<add>
<add> chunks = []
<add> prev_right = -1
<add> for token in doclike:
<add> if token.pos in [PROPN, NOUN, PRON]:
<add> left, right = get_bounds(doc, token)
<add> if left.i <= prev_right:
<add> continue
<add> yield left.i, right.i + 1, np_label
<add> prev_right = right.i
<add>
<add>
<add>SYNTAX_ITERATORS = {"noun_chunks": noun_chunks}
<ide><path>spacy/tests/lang/da/test_noun_chunks.py
<add># coding: utf-8
<add>from __future__ import unicode_literals
<add>
<add>import pytest
<add>
<add>from ...util import get_doc
<add>
<add>
<add>def test_noun_chunks_is_parsed(da_tokenizer):
<add> """Test that noun_chunks raises Value Error for 'da' language if Doc is not parsed.
<add> To check this test, we're constructing a Doc
<add> with a new Vocab here and forcing is_parsed to 'False'
<add> to make sure the noun chunks don't run.
<add> """
<add> doc = da_tokenizer("Det er en sætning")
<add> doc.is_parsed = False
<add> with pytest.raises(ValueError):
<add> list(doc.noun_chunks)
<add>
<add>
<add>
<add>DA_NP_TEST_EXAMPLES = [
<add> (
<add> "Hun elsker at plukker frugt.",
<add> ['PRON', 'VERB', 'PART', 'VERB', 'NOUN', 'PUNCT'],
<add> ['nsubj', 'ROOT', 'mark', 'obj', 'obj', 'punct'],
<add> [1, 0, 1, -2, -1, -4],
<add> ["Hun", "frugt"],
<add> ),
<add> (
<add> "Påfugle er de smukkeste fugle.",
<add> ['NOUN', 'AUX', 'DET', 'ADJ', 'NOUN', 'PUNCT'],
<add> ['nsubj', 'cop', 'det', 'amod', 'ROOT', 'punct'],
<add> [4, 3, 2, 1, 0, -1],
<add> ["Påfugle", "de smukkeste fugle"],
<add> ),
<add> (
<add> "Rikke og Jacob Jensen glæder sig til en hyggelig skovtur",
<add> ['PROPN', 'CCONJ', 'PROPN', 'PROPN', 'VERB', 'PRON', 'ADP', 'DET', 'ADJ', 'NOUN'],
<add> ['nsubj', 'cc', 'conj', 'flat', 'ROOT', 'obj', 'case', 'det', 'amod', 'obl'],
<add> [4, 1, -2, -1, 0, -1, 3, 2, 1, -5],
<add> ["Rikke", "Jacob Jensen", "sig", "en hyggelig skovtur"],
<add> ),
<add>]
<add>
<add>
<add>@pytest.mark.parametrize(
<add> "text,pos,deps,heads,expected_noun_chunks", DA_NP_TEST_EXAMPLES
<add>)
<add>def test_da_noun_chunks(da_tokenizer, text, pos, deps, heads, expected_noun_chunks):
<add> tokens = da_tokenizer(text)
<add>
<add> assert len(heads) == len(pos)
<add> doc = get_doc(
<add> tokens.vocab, words=[t.text for t in tokens], heads=heads, deps=deps, pos=pos
<add> )
<add>
<add> noun_chunks = list(doc.noun_chunks)
<add> assert len(noun_chunks) == len(expected_noun_chunks)
<add> for i, np in enumerate(noun_chunks):
<add> assert np.text == expected_noun_chunks[i]
| 4
|
Javascript
|
Javascript
|
add array.flat polyfill to nomodule-polyfills
|
20c546710d758a17ecb7abac64208d2fa09186e4
|
<ide><path>packages/next-polyfill-nomodule/src/index.js
<ide> import 'core-js/features/array/fill'
<ide> import 'core-js/features/array/find'
<ide> import 'core-js/features/array/find-index'
<ide> import 'core-js/features/array/flat-map'
<add>import 'core-js/features/array/flat'
<ide> import 'core-js/features/array/from'
<ide> import 'core-js/features/array/includes'
<ide> import 'core-js/features/array/iterator'
| 1
|
PHP
|
PHP
|
apply fixes from styleci
|
89d9cd46e6cf764171cae2c762872aacf87d65dd
|
<ide><path>src/Illuminate/Database/Eloquent/Relations/MorphPivot.php
<ide>
<ide> namespace Illuminate\Database\Eloquent\Relations;
<ide>
<del>use Illuminate\Database\Eloquent\Builder;
<ide> use Illuminate\Support\Str;
<ide>
<ide> class MorphPivot extends Pivot
| 1
|
Mixed
|
Javascript
|
create temp dir in common.js
|
a6b8ee19b85bbd798510191f0aee596f36b909d2
|
<ide><path>test/common.js
<ide> exports.tmpDirName = 'tmp';
<ide> exports.PORT = +process.env.NODE_COMMON_PORT || 12346;
<ide> exports.isWindows = process.platform === 'win32';
<ide>
<add>function rimrafSync(p) {
<add> try {
<add> var st = fs.lstatSync(p);
<add> } catch (e) {
<add> if (e.code === 'ENOENT')
<add> return;
<add> }
<add>
<add> try {
<add> if (st && st.isDirectory())
<add> rmdirSync(p, null);
<add> else
<add> fs.unlinkSync(p);
<add> } catch (e) {
<add> if (e.code === 'ENOENT')
<add> return;
<add> if (e.code === 'EPERM')
<add> return rmdirSync(p, er);
<add> if (e.code !== 'EISDIR')
<add> throw e;
<add> rmdirSync(p, e);
<add> }
<add>}
<add>
<add>function rmdirSync(p, originalEr) {
<add> try {
<add> fs.rmdirSync(p);
<add> } catch (e) {
<add> if (e.code === 'ENOTDIR')
<add> throw originalEr;
<add> if (e.code === 'ENOTEMPTY' || e.code === 'EEXIST' || e.code === 'EPERM') {
<add> fs.readdirSync(p).forEach(function(f) {
<add> rimrafSync(path.join(p, f));
<add> });
<add> fs.rmdirSync(p);
<add> }
<add> }
<add>}
<add>
<add>function refreshTmpDir() {
<add> if (!process.send) { // Not a child process
<add> try {
<add> rimrafSync(exports.tmpDir);
<add> } catch (e) {
<add> }
<add>
<add> try {
<add> fs.mkdirSync(exports.tmpDir);
<add> } catch (e) {
<add> }
<add> }
<add>}
<add>
<ide> if (process.env.TEST_THREAD_ID) {
<ide> // Distribute ports in parallel tests
<ide> if (!process.env.NODE_COMMON_PORT)
<ide> if (process.env.TEST_THREAD_ID) {
<ide> }
<ide> exports.tmpDir = path.join(exports.testDir, exports.tmpDirName);
<ide>
<add>refreshTmpDir();
<add>
<ide> var opensslCli = null;
<ide> var inFreeBSDJail = null;
<ide> var localhostIPv4 = null;
<ide><path>test/fixtures/print-chars-from-buffer.js
<del>var common = require('../common');
<ide> var assert = require('assert');
<ide>
<ide> var n = parseInt(process.argv[2]);
<ide><path>test/fixtures/print-chars.js
<del>var common = require('../common');
<ide> var assert = require('assert');
<ide>
<ide> var n = parseInt(process.argv[2]);
<ide><path>test/sequential/test-fs-watch-recursive.js
<ide> if (process.platform === 'darwin') {
<ide> watcher.on('change', function(event, filename) {
<ide> assert.ok('change' === event || 'rename' === event);
<ide>
<del> // Ignore stale events generated by mkdir
<del> if (filename === testsubdirName)
<add> // Ignore stale events generated by mkdir and other tests
<add> if (filename !== relativePathOne)
<ide> return;
<ide>
<del> assert.equal(relativePathOne, filename);
<del>
<ide> watcher.close();
<ide> cleanup();
<ide> ++watchSeenOne;
<ide><path>test/testpy/__init__.py
<ide> import test
<ide> import os
<ide> import shutil
<del>from shutil import rmtree
<ide> from os import mkdir
<ide> from glob import glob
<ide> from os.path import join, dirname, exists
<ide> def __init__(self, path, file, arch, mode, context, config, additional=[]):
<ide> self.tmpdir = join(dirname(self.config.root), 'tmp')
<ide> self.additional_flags = additional
<ide>
<del> def GetTmpDir(self):
<del> return "%s.%d" % (self.tmpdir, self.thread_id)
<del>
<del>
<del> def AfterRun(self, result):
<del> # delete the whole tmp dir
<del> try:
<del> rmtree(self.GetTmpDir())
<del> except:
<del> pass
<del> # make it again.
<del> try:
<del> mkdir(self.GetTmpDir())
<del> except:
<del> pass
<del>
<del> def BeforeRun(self):
<del> # delete the whole tmp dir
<del> try:
<del> rmtree(self.GetTmpDir())
<del> except:
<del> pass
<del> # make it again.
<del> # intermittently fails on win32, so keep trying
<del> while not os.path.exists(self.GetTmpDir()):
<del> try:
<del> mkdir(self.GetTmpDir())
<del> except:
<del> pass
<ide>
<ide> def GetLabel(self):
<ide> return "%s %s" % (self.mode, self.GetName())
| 5
|
Go
|
Go
|
fix av in build due to userns
|
6d71f277608b2a655df0942d00607f47dbcaa37b
|
<ide><path>pkg/archive/diff.go
<ide> func UnpackLayer(dest string, layer Reader, options *TarOptions) (size int64, er
<ide> defer pools.BufioReader32KPool.Put(trBuf)
<ide>
<ide> var dirs []*tar.Header
<add>
<add> if options == nil {
<add> options = &TarOptions{}
<add> }
<add> if options.ExcludePatterns == nil {
<add> options.ExcludePatterns = []string{}
<add> }
<ide> remappedRootUID, remappedRootGID, err := idtools.GetRootUIDGID(options.UIDMaps, options.GIDMaps)
<ide> if err != nil {
<ide> return 0, err
| 1
|
Go
|
Go
|
resolve connection reset by peer regression
|
dc0ee98805c1e9282c729a79cdf10e59bad3cb09
|
<ide><path>daemon/attach.go
<ide> func (daemon *Daemon) ContainerAttach(prefixOrName string, c *backend.ContainerA
<ide> }
<ide>
<ide> cfg := stream.AttachConfig{
<del> UseStdin: c.UseStdin && container.Config.OpenStdin,
<add> UseStdin: c.UseStdin,
<ide> UseStdout: c.UseStdout,
<ide> UseStderr: c.UseStderr,
<ide> TTY: container.Config.Tty,
<ide> func (daemon *Daemon) ContainerAttachRaw(prefixOrName string, stdin io.ReadClose
<ide> return err
<ide> }
<ide> cfg := stream.AttachConfig{
<del> UseStdin: stdin != nil && container.Config.OpenStdin,
<add> UseStdin: stdin != nil,
<ide> UseStdout: stdout != nil,
<ide> UseStderr: stderr != nil,
<ide> TTY: container.Config.Tty,
<ide> func (daemon *Daemon) containerAttach(c *container.Container, cfg *stream.Attach
<ide> cfg.Stdin = r
<ide> }
<ide>
<add> if !c.Config.OpenStdin {
<add> cfg.Stdin = nil
<add> }
<add>
<ide> waitChan := make(chan struct{})
<ide> if c.Config.StdinOnce && !c.Config.Tty {
<ide> defer func() {
| 1
|
Javascript
|
Javascript
|
upgrade stats to es6
|
b7c1b5c6724944a92de54590a5171f2805a48cd0
|
<ide><path>lib/Stats.js
<ide> MIT License http://www.opensource.org/licenses/mit-license.php
<ide> Author Tobias Koppers @sokra
<ide> */
<del>var RequestShortener = require("./RequestShortener");
<del>var SizeFormatHelpers = require("./SizeFormatHelpers");
<add>"use strict";
<ide>
<del>function Stats(compilation) {
<del> this.compilation = compilation;
<del> this.hash = compilation.hash;
<del>}
<del>module.exports = Stats;
<add>const RequestShortener = require("./RequestShortener");
<add>const SizeFormatHelpers = require("./SizeFormatHelpers");
<ide>
<del>Stats.prototype.hasWarnings = function() {
<del> return this.compilation.warnings.length > 0;
<del>};
<add>const d = (v, def) => v === undefined ? def : v;
<ide>
<del>Stats.prototype.hasErrors = function() {
<del> return this.compilation.errors.length > 0;
<del>};
<add>class Stats {
<add> constructor(compilation) {
<add> this.compilation = compilation;
<add> this.hash = compilation.hash;
<add> }
<ide>
<del>Stats.prototype.toJson = function toJson(options, forToString) {
<del> if(typeof options === "boolean" || typeof options === "string") {
<del> options = Stats.presetToOptions(options);
<del> } else if(!options) {
<del> options = {};
<add> hasWarnings() {
<add> return this.compilation.warnings.length > 0;
<ide> }
<ide>
<del> function d(v, def) {
<del> return v === undefined ? def : v;
<add> hasErrors() {
<add> return this.compilation.errors.length > 0;
<ide> }
<del> var compilation = this.compilation;
<del> var requestShortener = new RequestShortener(d(options.context, process.cwd()));
<del> var showPerformance = d(options.performance, true);
<del> var showHash = d(options.hash, true);
<del> var showVersion = d(options.version, true);
<del> var showTimings = d(options.timings, true);
<del> var showAssets = d(options.assets, true);
<del> var showEntrypoints = d(options.entrypoints, !forToString);
<del> var showChunks = d(options.chunks, true);
<del> var showChunkModules = d(options.chunkModules, !!forToString);
<del> var showChunkOrigins = d(options.chunkOrigins, !forToString);
<del> var showModules = d(options.modules, !forToString);
<del> var showDepth = d(options.depth, !forToString);
<del> var showCachedModules = d(options.cached, true);
<del> var showCachedAssets = d(options.cachedAssets, true);
<del> var showReasons = d(options.reasons, !forToString);
<del> var showUsedExports = d(options.usedExports, !forToString);
<del> var showProvidedExports = d(options.providedExports, !forToString);
<del> var showChildren = d(options.children, true);
<del> var showSource = d(options.source, !forToString);
<del> var showErrors = d(options.errors, true);
<del> var showErrorDetails = d(options.errorDetails, !forToString);
<del> var showWarnings = d(options.warnings, true);
<del> var showPublicPath = d(options.publicPath, !forToString);
<del> var excludeModules = [].concat(d(options.exclude, [])).map(function(str) {
<del> if(typeof str !== "string") return str;
<del> return new RegExp("[\\\\/]" + str.replace(/[\-\[\]\/\{\}\(\)\*\+\?\.\\\^\$\|]/g, "\\$&") + "([\\\\/]|$|!|\\?)");
<del> });
<del> var maxModules = d(options.maxModules, forToString ? 15 : Infinity);
<del> var sortModules = d(options.modulesSort, "id");
<del> var sortChunks = d(options.chunksSort, "id");
<del> var sortAssets = d(options.assetsSort, "");
<del>
<del> function createModuleFilter() {
<del> var i = 0;
<del> return function(module) {
<del> if(!showCachedModules && !module.built) {
<del> return false;
<del> }
<del> if(excludeModules.length > 0) {
<del> var ident = requestShortener.shorten(module.resource);
<del> var excluded = excludeModules.some(function(regExp) {
<del> return regExp.test(ident);
<del> });
<del> if(excluded)
<add>
<add> toJson(options, forToString) {
<add> if(typeof options === "boolean" || typeof options === "string") {
<add> options = Stats.presetToOptions(options);
<add> } else if(!options) {
<add> options = {};
<add> }
<add>
<add> const compilation = this.compilation;
<add> const requestShortener = new RequestShortener(d(options.context, process.cwd()));
<add> const showPerformance = d(options.performance, true);
<add> const showHash = d(options.hash, true);
<add> const showVersion = d(options.version, true);
<add> const showTimings = d(options.timings, true);
<add> const showAssets = d(options.assets, true);
<add> const showEntrypoints = d(options.entrypoints, !forToString);
<add> const showChunks = d(options.chunks, true);
<add> const showChunkModules = d(options.chunkModules, !!forToString);
<add> const showChunkOrigins = d(options.chunkOrigins, !forToString);
<add> const showModules = d(options.modules, !forToString);
<add> const showDepth = d(options.depth, !forToString);
<add> const showCachedModules = d(options.cached, true);
<add> const showCachedAssets = d(options.cachedAssets, true);
<add> const showReasons = d(options.reasons, !forToString);
<add> const showUsedExports = d(options.usedExports, !forToString);
<add> const showProvidedExports = d(options.providedExports, !forToString);
<add> const showChildren = d(options.children, true);
<add> const showSource = d(options.source, !forToString);
<add> const showErrors = d(options.errors, true);
<add> const showErrorDetails = d(options.errorDetails, !forToString);
<add> const showWarnings = d(options.warnings, true);
<add> const showPublicPath = d(options.publicPath, !forToString);
<add> const excludeModules = [].concat(d(options.exclude, [])).map(str => {
<add> if(typeof str !== "string") return str;
<add> return new RegExp(`[\\\\/]${str.replace(/[\-\[\]\/\{\}\(\)\*\+\?\.\\\^\$\|]/g, "\\$&")}([\\\\/]|$|!|\\?)`);
<add> });
<add> const maxModules = d(options.maxModules, forToString ? 15 : Infinity);
<add> const sortModules = d(options.modulesSort, "id");
<add> const sortChunks = d(options.chunksSort, "id");
<add> const sortAssets = d(options.assetsSort, "");
<add>
<add> const createModuleFilter = () => {
<add> let i = 0;
<add> return module => {
<add> if(!showCachedModules && !module.built) {
<ide> return false;
<add> }
<add> if(excludeModules.length > 0) {
<add> const ident = requestShortener.shorten(module.resource);
<add> const excluded = excludeModules.some(regExp => regExp.test(ident));
<add> if(excluded)
<add> return false;
<add> }
<add> return i++ < maxModules;
<add> };
<add> }
<add>
<add> const sortByField = (field) => {
<add> let callback;
<add> if(!field) {
<add> callback = () => 0;
<add> return callback;
<ide> }
<del> return i++ < maxModules;
<del> };
<del> }
<ide>
<del> function sortByField(field) {
<del> if(!field) return function() {
<del> return 0;
<del> };
<del> if(field[0] === "!") {
<del> field = field.substr(1);
<del> return function(a, b) {
<add> if(field[0] === "!") {
<add> field = field.substr(1);
<add> callback = (a, b) => {
<add> if(a[field] === null && b[field] === null) return 0;
<add> if(a[field] === null) return 1;
<add> if(b[field] === null) return -1;
<add> if(a[field] === b[field]) return 0;
<add> return a[field] < b[field] ? 1 : -1;
<add> };
<add> }
<add> callback = (a, b) => {
<ide> if(a[field] === null && b[field] === null) return 0;
<ide> if(a[field] === null) return 1;
<ide> if(b[field] === null) return -1;
<ide> if(a[field] === b[field]) return 0;
<del> return a[field] < b[field] ? 1 : -1;
<add> return a[field] < b[field] ? -1 : 1;
<ide> };
<add> return callback;
<ide> }
<del> return function(a, b) {
<del> if(a[field] === null && b[field] === null) return 0;
<del> if(a[field] === null) return 1;
<del> if(b[field] === null) return -1;
<del> if(a[field] === b[field]) return 0;
<del> return a[field] < b[field] ? -1 : 1;
<del> };
<del> }
<ide>
<del> function formatError(e) {
<del> var text = "";
<del> if(typeof e === "string")
<del> e = {
<del> message: e
<del> };
<del> if(e.chunk) {
<del> text += "chunk " + (e.chunk.name || e.chunk.id) +
<del> (e.chunk.hasRuntime() ? " [entry]" : e.chunk.isInitial() ? " [initial]" : "") + "\n";
<del> }
<del> if(e.file) {
<del> text += e.file + "\n";
<del> }
<del> if(e.module && e.module.readableIdentifier && typeof e.module.readableIdentifier === "function") {
<del> text += e.module.readableIdentifier(requestShortener) + "\n";
<del> }
<del> text += e.message;
<del> if(showErrorDetails && e.details) text += "\n" + e.details;
<del> if(showErrorDetails && e.missing) text += e.missing.map(function(item) {
<del> return "\n[" + item + "]";
<del> }).join("");
<del> if(e.dependencies && e.origin) {
<del> text += "\n @ " + e.origin.readableIdentifier(requestShortener);
<del> e.dependencies.forEach(function(dep) {
<del> if(!dep.loc) return;
<del> if(typeof dep.loc === "string") return;
<del> if(!dep.loc.start) return;
<del> if(!dep.loc.end) return;
<del> text += " " + dep.loc.start.line + ":" + dep.loc.start.column + "-" +
<del> (dep.loc.start.line !== dep.loc.end.line ? dep.loc.end.line + ":" : "") + dep.loc.end.column;
<del> });
<del> var current = e.origin;
<del> while(current.issuer) {
<del> current = current.issuer;
<del> text += "\n @ " + current.readableIdentifier(requestShortener);
<add> const formatError = (e) => {
<add> let text = "";
<add> if(typeof e === "string")
<add> e = {
<add> message: e
<add> };
<add> if(e.chunk) {
<add> text += `chunk ${e.chunk.name || e.chunk.id}${e.chunk.hasRuntime() ? " [entry]" : e.chunk.isInitial() ? " [initial]" : ""}\n`;
<add> }
<add> if(e.file) {
<add> text += `${e.file}\n`;
<add> }
<add> if(e.module && e.module.readableIdentifier && typeof e.module.readableIdentifier === "function") {
<add> text += `${e.module.readableIdentifier(requestShortener)}\n`;
<add> }
<add> text += e.message;
<add> if(showErrorDetails && e.details) text += `\n${e.details}`;
<add> if(showErrorDetails && e.missing) text += e.missing.map(item => `\n[${item}]`).join("");
<add> if(e.dependencies && e.origin) {
<add> text += `\n @ ${e.origin.readableIdentifier(requestShortener)}`;
<add> e.dependencies.forEach(dep => {
<add> if(!dep.loc) return;
<add> if(typeof dep.loc === "string") return;
<add> if(!dep.loc.start) return;
<add> if(!dep.loc.end) return;
<add> text += ` ${dep.loc.start.line}:${dep.loc.start.column}-${dep.loc.start.line !== dep.loc.end.line ? dep.loc.end.line + ":" : ""}${dep.loc.end.column}`;
<add> });
<add> let current = e.origin;
<add> while(current.issuer) {
<add> current = current.issuer;
<add> text += `\n @ ${current.readableIdentifier(requestShortener)}`;
<add> }
<ide> }
<add> return text;
<ide> }
<del> return text;
<del> }
<ide>
<del> var obj = {
<del> errors: compilation.errors.map(formatError),
<del> warnings: compilation.warnings.map(formatError)
<del> };
<del>
<del> //We just hint other renderers since actually omitting
<del> //errors/warnings from the JSON would be kind of weird.
<del> Object.defineProperty(obj, "_showWarnings", {
<del> value: showWarnings,
<del> enumerable: false
<del> });
<del> Object.defineProperty(obj, "_showErrors", {
<del> value: showErrors,
<del> enumerable: false
<del> });
<del>
<del> if(showVersion) {
<del> obj.version = require("../package.json").version;
<del> }
<add> const obj = {
<add> errors: compilation.errors.map(formatError),
<add> warnings: compilation.warnings.map(formatError)
<add> };
<ide>
<del> if(showHash) obj.hash = this.hash;
<del> if(showTimings && this.startTime && this.endTime) {
<del> obj.time = this.endTime - this.startTime;
<del> }
<del> if(compilation.needAdditionalPass) {
<del> obj.needAdditionalPass = true;
<del> }
<del> if(showPublicPath) {
<del> obj.publicPath = this.compilation.mainTemplate.getPublicPath({
<del> hash: this.compilation.hash
<add> //We just hint other renderers since actually omitting
<add> //errors/warnings from the JSON would be kind of weird.
<add> Object.defineProperty(obj, "_showWarnings", {
<add> value: showWarnings,
<add> enumerable: false
<add> });
<add> Object.defineProperty(obj, "_showErrors", {
<add> value: showErrors,
<add> enumerable: false
<ide> });
<del> }
<del> if(showAssets) {
<del> var assetsByFile = {};
<del> obj.assetsByChunkName = {};
<del> obj.assets = Object.keys(compilation.assets).map(function(asset) {
<del> var obj = {
<del> name: asset,
<del> size: compilation.assets[asset].size(),
<del> chunks: [],
<del> chunkNames: [],
<del> emitted: compilation.assets[asset].emitted
<del> };
<ide>
<del> if(showPerformance) {
<del> obj.isOverSizeLimit = compilation.assets[asset].isOverSizeLimit;
<del> }
<add> if(showVersion) {
<add> obj.version = require("../package.json").version;
<add> }
<ide>
<del> assetsByFile[asset] = obj;
<del> return obj;
<del> }).filter(function(asset) {
<del> return showCachedAssets || asset.emitted;
<del> });
<del> compilation.chunks.forEach(function(chunk) {
<del> chunk.files.forEach(function(asset) {
<del> if(assetsByFile[asset]) {
<del> chunk.ids.forEach(function(id) {
<del> assetsByFile[asset].chunks.push(id);
<del> });
<del> if(chunk.name) {
<del> assetsByFile[asset].chunkNames.push(chunk.name);
<del> if(obj.assetsByChunkName[chunk.name])
<del> obj.assetsByChunkName[chunk.name] = [].concat(obj.assetsByChunkName[chunk.name]).concat([asset]);
<del> else
<del> obj.assetsByChunkName[chunk.name] = asset;
<del> }
<del> }
<add> if(showHash) obj.hash = this.hash;
<add> if(showTimings && this.startTime && this.endTime) {
<add> obj.time = this.endTime - this.startTime;
<add> }
<add> if(compilation.needAdditionalPass) {
<add> obj.needAdditionalPass = true;
<add> }
<add> if(showPublicPath) {
<add> obj.publicPath = this.compilation.mainTemplate.getPublicPath({
<add> hash: this.compilation.hash
<ide> });
<del> });
<del> obj.assets.sort(sortByField(sortAssets));
<del> }
<add> }
<add> if(showAssets) {
<add> const assetsByFile = {};
<add> obj.assetsByChunkName = {};
<add> obj.assets = Object.keys(compilation.assets).map(asset => {
<add> const obj = {
<add> name: asset,
<add> size: compilation.assets[asset].size(),
<add> chunks: [],
<add> chunkNames: [],
<add> emitted: compilation.assets[asset].emitted
<add> };
<ide>
<del> if(showEntrypoints) {
<del> obj.entrypoints = {};
<del> Object.keys(compilation.entrypoints).forEach(function(name) {
<del> var ep = compilation.entrypoints[name];
<del> obj.entrypoints[name] = {
<del> chunks: ep.chunks.map(function(c) {
<del> return c.id;
<del> }),
<del> assets: ep.chunks.reduce(function(array, c) {
<del> return array.concat(c.files || []);
<del> }, [])
<del> };
<del> if(showPerformance) {
<del> obj.entrypoints[name].isOverSizeLimit = ep.isOverSizeLimit;
<del> }
<del> });
<del> }
<add> if(showPerformance) {
<add> obj.isOverSizeLimit = compilation.assets[asset].isOverSizeLimit;
<add> }
<ide>
<del> function fnModule(module) {
<del> var obj = {
<del> id: module.id,
<del> identifier: module.identifier(),
<del> name: module.readableIdentifier(requestShortener),
<del> index: module.index,
<del> index2: module.index2,
<del> size: module.size(),
<del> cacheable: !!module.cacheable,
<del> built: !!module.built,
<del> optional: !!module.optional,
<del> prefetched: !!module.prefetched,
<del> chunks: module.chunks.map(function(chunk) {
<del> return chunk.id;
<del> }),
<del> assets: Object.keys(module.assets || {}),
<del> issuer: module.issuer && module.issuer.identifier(),
<del> issuerId: module.issuer && module.issuer.id,
<del> issuerName: module.issuer && module.issuer.readableIdentifier(requestShortener),
<del> profile: module.profile,
<del> failed: !!module.error,
<del> errors: module.errors && module.dependenciesErrors && (module.errors.length + module.dependenciesErrors.length),
<del> warnings: module.errors && module.dependenciesErrors && (module.warnings.length + module.dependenciesWarnings.length)
<del> };
<del> if(showReasons) {
<del> obj.reasons = module.reasons.filter(function(reason) {
<del> return reason.dependency && reason.module;
<del> }).map(function(reason) {
<del> var obj = {
<del> moduleId: reason.module.id,
<del> moduleIdentifier: reason.module.identifier(),
<del> module: reason.module.readableIdentifier(requestShortener),
<del> moduleName: reason.module.readableIdentifier(requestShortener),
<del> type: reason.dependency.type,
<del> userRequest: reason.dependency.userRequest
<del> };
<del> var dep = reason.dependency;
<del> if(dep.templateModules) obj.templateModules = dep.templateModules.map(function(module) {
<del> return module.id;
<del> });
<del> if(typeof dep.loc === "object") obj.loc = dep.loc.start.line + ":" + dep.loc.start.column + "-" +
<del> (dep.loc.start.line !== dep.loc.end.line ? dep.loc.end.line + ":" : "") + dep.loc.end.column;
<add> assetsByFile[asset] = obj;
<ide> return obj;
<del> }).sort(function(a, b) {
<del> return a.moduleId - b.moduleId;
<add> }).filter(asset => showCachedAssets || asset.emitted);
<add>
<add> compilation.chunks.forEach(chunk => {
<add> chunk.files.forEach(asset => {
<add> if(assetsByFile[asset]) {
<add> chunk.ids.forEach(id => {
<add> assetsByFile[asset].chunks.push(id);
<add> });
<add> if(chunk.name) {
<add> assetsByFile[asset].chunkNames.push(chunk.name);
<add> if(obj.assetsByChunkName[chunk.name])
<add> obj.assetsByChunkName[chunk.name] = [].concat(obj.assetsByChunkName[chunk.name]).concat([asset]);
<add> else
<add> obj.assetsByChunkName[chunk.name] = asset;
<add> }
<add> }
<add> });
<ide> });
<add> obj.assets.sort(sortByField(sortAssets));
<ide> }
<del> if(showUsedExports) {
<del> obj.usedExports = module.used ? module.usedExports : false;
<del> }
<del> if(showProvidedExports) {
<del> obj.providedExports = Array.isArray(module.providedExports) ? module.providedExports : null;
<del> }
<del> if(showDepth) {
<del> obj.depth = module.depth;
<del> }
<del> if(showSource && module._source) {
<del> obj.source = module._source.source();
<add>
<add> if(showEntrypoints) {
<add> obj.entrypoints = {};
<add> Object.keys(compilation.entrypoints).forEach(name => {
<add> const ep = compilation.entrypoints[name];
<add> obj.entrypoints[name] = {
<add> chunks: ep.chunks.map(c => c.id),
<add> assets: ep.chunks.reduce((array, c) => array.concat(c.files || []), [])
<add> };
<add> if(showPerformance) {
<add> obj.entrypoints[name].isOverSizeLimit = ep.isOverSizeLimit;
<add> }
<add> });
<ide> }
<del> return obj;
<del> }
<del> if(showChunks) {
<del> obj.chunks = compilation.chunks.map(function(chunk) {
<del> var obj = {
<del> id: chunk.id,
<del> rendered: chunk.rendered,
<del> initial: chunk.isInitial(),
<del> entry: chunk.hasRuntime(),
<del> recorded: chunk.recorded,
<del> extraAsync: !!chunk.extraAsync,
<del> size: chunk.modules.reduce(function(size, module) {
<del> return size + module.size();
<del> }, 0),
<del> names: chunk.name ? [chunk.name] : [],
<del> files: chunk.files.slice(),
<del> hash: chunk.renderedHash,
<del> parents: chunk.parents.map(function(c) {
<del> return c.id;
<del> })
<add>
<add> function fnModule(module) {
<add> const obj = {
<add> id: module.id,
<add> identifier: module.identifier(),
<add> name: module.readableIdentifier(requestShortener),
<add> index: module.index,
<add> index2: module.index2,
<add> size: module.size(),
<add> cacheable: !!module.cacheable,
<add> built: !!module.built,
<add> optional: !!module.optional,
<add> prefetched: !!module.prefetched,
<add> chunks: module.chunks.map(chunk => chunk.id),
<add> assets: Object.keys(module.assets || {}),
<add> issuer: module.issuer && module.issuer.identifier(),
<add> issuerId: module.issuer && module.issuer.id,
<add> issuerName: module.issuer && module.issuer.readableIdentifier(requestShortener),
<add> profile: module.profile,
<add> failed: !!module.error,
<add> errors: module.errors && module.dependenciesErrors && (module.errors.length + module.dependenciesErrors.length),
<add> warnings: module.errors && module.dependenciesErrors && (module.warnings.length + module.dependenciesWarnings.length)
<ide> };
<del> if(showChunkModules) {
<del> obj.modules = chunk.modules
<del> .slice()
<del> .sort(sortByField("depth"))
<del> .filter(createModuleFilter())
<del> .map(fnModule);
<del> obj.filteredModules = chunk.modules.length - obj.modules.length;
<del> obj.modules.sort(sortByField(sortModules));
<add> if(showReasons) {
<add> obj.reasons = module.reasons.filter(reason => reason.dependency && reason.module).map(reason => {
<add> const obj = {
<add> moduleId: reason.module.id,
<add> moduleIdentifier: reason.module.identifier(),
<add> module: reason.module.readableIdentifier(requestShortener),
<add> moduleName: reason.module.readableIdentifier(requestShortener),
<add> type: reason.dependency.type,
<add> userRequest: reason.dependency.userRequest
<add> };
<add> const dep = reason.dependency;
<add> if(dep.templateModules) obj.templateModules = dep.templateModules.map(module => module.id);
<add> if(typeof dep.loc === "object") obj.loc = `${dep.loc.start.line}:${dep.loc.start.column}-${dep.loc.start.line !== dep.loc.end.line ? dep.loc.end.line + ":" : ""}${dep.loc.end.column}`;
<add> return obj;
<add> }).sort((a, b) => a.moduleId - b.moduleId);
<ide> }
<del> if(showChunkOrigins) {
<del> obj.origins = chunk.origins.map(function(origin) {
<del> return {
<add> if(showUsedExports) {
<add> obj.usedExports = module.used ? module.usedExports : false;
<add> }
<add> if(showProvidedExports) {
<add> obj.providedExports = Array.isArray(module.providedExports) ? module.providedExports : null;
<add> }
<add> if(showDepth) {
<add> obj.depth = module.depth;
<add> }
<add> if(showSource && module._source) {
<add> obj.source = module._source.source();
<add> }
<add> return obj;
<add> }
<add> if(showChunks) {
<add> obj.chunks = compilation.chunks.map(chunk => {
<add> const obj = {
<add> id: chunk.id,
<add> rendered: chunk.rendered,
<add> initial: chunk.isInitial(),
<add> entry: chunk.hasRuntime(),
<add> recorded: chunk.recorded,
<add> extraAsync: !!chunk.extraAsync,
<add> size: chunk.modules.reduce((size, module) => size + module.size(), 0),
<add> names: chunk.name ? [chunk.name] : [],
<add> files: chunk.files.slice(),
<add> hash: chunk.renderedHash,
<add> parents: chunk.parents.map(c => c.id)
<add> };
<add> if(showChunkModules) {
<add> obj.modules = chunk.modules
<add> .slice()
<add> .sort(sortByField("depth"))
<add> .filter(createModuleFilter())
<add> .map(fnModule);
<add> obj.filteredModules = chunk.modules.length - obj.modules.length;
<add> obj.modules.sort(sortByField(sortModules));
<add> }
<add> if(showChunkOrigins) {
<add> obj.origins = chunk.origins.map(origin => ({
<ide> moduleId: origin.module ? origin.module.id : undefined,
<ide> module: origin.module ? origin.module.identifier() : "",
<ide> moduleIdentifier: origin.module ? origin.module.identifier() : "",
<ide> moduleName: origin.module ? origin.module.readableIdentifier(requestShortener) : "",
<del> loc: typeof origin.loc === "object" ? obj.loc = origin.loc.start.line + ":" + origin.loc.start.column + "-" +
<del> (origin.loc.start.line !== origin.loc.end.line ? origin.loc.end.line + ":" : "") + origin.loc.end.column : "",
<add> loc: typeof origin.loc === "object" ? obj.loc = `${origin.loc.start.line}:${origin.loc.start.column}-${origin.loc.start.line !== origin.loc.end.line ? origin.loc.end.line + ":" : ""}${origin.loc.end.column}` : "",
<ide> name: origin.name,
<ide> reasons: origin.reasons || []
<del> };
<del> });
<del> }
<del> return obj;
<del> });
<del> obj.chunks.sort(sortByField(sortChunks));
<del> }
<del> if(showModules) {
<del> obj.modules = compilation.modules
<del> .slice()
<del> .sort(sortByField("depth"))
<del> .filter(createModuleFilter())
<del> .map(fnModule);
<del> obj.filteredModules = compilation.modules.length - obj.modules.length;
<del> obj.modules.sort(sortByField(sortModules));
<del> }
<del> if(showChildren) {
<del> obj.children = compilation.children.map(function(child) {
<del> var obj = new Stats(child).toJson(options, forToString);
<del> delete obj.hash;
<del> delete obj.version;
<del> obj.name = child.name;
<del> return obj;
<del> });
<add> }));
<add> }
<add> return obj;
<add> });
<add> obj.chunks.sort(sortByField(sortChunks));
<add> }
<add> if(showModules) {
<add> obj.modules = compilation.modules
<add> .slice()
<add> .sort(sortByField("depth"))
<add> .filter(createModuleFilter())
<add> .map(fnModule);
<add> obj.filteredModules = compilation.modules.length - obj.modules.length;
<add> obj.modules.sort(sortByField(sortModules));
<add> }
<add> if(showChildren) {
<add> obj.children = compilation.children.map(child => {
<add> const obj = new Stats(child).toJson(options, forToString);
<add> delete obj.hash;
<add> delete obj.version;
<add> obj.name = child.name;
<add> return obj;
<add> });
<add> }
<add>
<add> return obj;
<ide> }
<ide>
<del> return obj;
<del>};
<add> toString(options) {
<add> if(typeof options === "boolean" || typeof options === "string") {
<add> options = Stats.presetToOptions(options);
<add> } else if(!options) {
<add> options = {};
<add> }
<ide>
<del>Stats.prototype.toString = function toString(options) {
<del> if(typeof options === "boolean" || typeof options === "string") {
<del> options = Stats.presetToOptions(options);
<del> } else if(!options) options = {};
<add> const useColors = d(options.colors, false);
<ide>
<del> function d(v, def) {
<del> return v === undefined ? def : v;
<del> }
<del> var useColors = d(options.colors, false);
<del>
<del> var obj = this.toJson(options, true);
<del>
<del> return Stats.jsonToString(obj, useColors);
<del>};
<del>
<del>Stats.jsonToString = function jsonToString(obj, useColors) {
<del> var buf = [];
<del>
<del> var defaultColors = {
<del> bold: "\u001b[1m",
<del> yellow: "\u001b[1m\u001b[33m",
<del> red: "\u001b[1m\u001b[31m",
<del> green: "\u001b[1m\u001b[32m",
<del> cyan: "\u001b[1m\u001b[36m",
<del> magenta: "\u001b[1m\u001b[35m"
<del> };
<del>
<del> var colors = Object.keys(defaultColors).reduce(function(obj, color) {
<del> obj[color] = function(str) {
<del> if(useColors) {
<del> buf.push(
<del> (useColors === true || useColors[color] === undefined) ?
<del> defaultColors[color] : useColors[color]
<del> );
<del> }
<del> buf.push(str);
<del> if(useColors) {
<del> buf.push("\u001b[39m\u001b[22m");
<del> }
<del> };
<del> return obj;
<del> }, {
<del> normal: function(str) {
<del> buf.push(str);
<del> }
<del> });
<del>
<del> function coloredTime(time) {
<del> var times = [800, 400, 200, 100];
<del> if(obj.time) {
<del> times = [obj.time / 2, obj.time / 4, obj.time / 8, obj.time / 16];
<del> }
<del> if(time < times[3])
<del> colors.normal(time + "ms");
<del> else if(time < times[2])
<del> colors.bold(time + "ms");
<del> else if(time < times[1])
<del> colors.green(time + "ms");
<del> else if(time < times[0])
<del> colors.yellow(time + "ms");
<del> else
<del> colors.red(time + "ms");
<del> }
<add> const obj = this.toJson(options, true);
<ide>
<del> function newline() {
<del> buf.push("\n");
<add> return Stats.jsonToString(obj, useColors);
<ide> }
<ide>
<del> function getText(arr, row, col) {
<del> return arr[row][col].value;
<del> }
<add> static jsonToString(obj, useColors) {
<add> const buf = [];
<ide>
<del> function table(array, align, splitter) {
<del> var row;
<del> var rows = array.length;
<del> var col;
<del> var cols = array[0].length;
<del> var colSizes = new Array(cols);
<del> var value;
<del> for(col = 0; col < cols; col++)
<del> colSizes[col] = 0;
<del> for(row = 0; row < rows; row++) {
<del> for(col = 0; col < cols; col++) {
<del> value = getText(array, row, col) + "";
<del> if(value.length > colSizes[col]) {
<del> colSizes[col] = value.length;
<add> const defaultColors = {
<add> bold: "\u001b[1m",
<add> yellow: "\u001b[1m\u001b[33m",
<add> red: "\u001b[1m\u001b[31m",
<add> green: "\u001b[1m\u001b[32m",
<add> cyan: "\u001b[1m\u001b[36m",
<add> magenta: "\u001b[1m\u001b[35m"
<add> };
<add>
<add> const colors = Object.keys(defaultColors).reduce((obj, color) => {
<add> obj[color] = str => {
<add> if(useColors) {
<add> buf.push(
<add> (useColors === true || useColors[color] === undefined) ?
<add> defaultColors[color] : useColors[color]
<add> );
<ide> }
<del> }
<del> }
<del> for(row = 0; row < rows; row++) {
<del> for(col = 0; col < cols; col++) {
<del> var format = array[row][col].color;
<del> value = getText(array, row, col) + "";
<del> var l = value.length;
<del> if(align[col] === "l")
<del> format(value);
<del> for(; l < colSizes[col] && col !== cols - 1; l++)
<del> colors.normal(" ");
<del> if(align[col] === "r")
<del> format(value);
<del> if(col + 1 < cols && colSizes[col] !== 0)
<del> colors.normal(splitter || " ");
<del> }
<del> newline();
<del> }
<del> }
<add> buf.push(str);
<add> if(useColors) {
<add> buf.push("\u001b[39m\u001b[22m");
<add> }
<add> };
<add> return obj;
<add> }, {
<add> normal: (str) => buf.push(str)
<add> });
<ide>
<del> function getAssetColor(asset, defaultColor) {
<del> if(asset.isOverSizeLimit) {
<del> return colors.yellow;
<add> const coloredTime = (time) => {
<add> let times = [800, 400, 200, 100];
<add> if(obj.time) {
<add> times = [obj.time / 2, obj.time / 4, obj.time / 8, obj.time / 16];
<add> }
<add> if(time < times[3])
<add> colors.normal(`${time}ms`);
<add> else if(time < times[2])
<add> colors.bold(`${time}ms`);
<add> else if(time < times[1])
<add> colors.green(`${time}ms`);
<add> else if(time < times[0])
<add> colors.yellow(`${time}ms`);
<add> else
<add> colors.red(`${time}ms`);
<ide> }
<ide>
<del> return defaultColor;
<del> }
<add> const newline = () => buf.push("\n");
<ide>
<del> if(obj.hash) {
<del> colors.normal("Hash: ");
<del> colors.bold(obj.hash);
<del> newline();
<del> }
<del> if(obj.version) {
<del> colors.normal("Version: webpack ");
<del> colors.bold(obj.version);
<del> newline();
<del> }
<del> if(typeof obj.time === "number") {
<del> colors.normal("Time: ");
<del> colors.bold(obj.time);
<del> colors.normal("ms");
<del> newline();
<del> }
<del> if(obj.publicPath) {
<del> colors.normal("PublicPath: ");
<del> colors.bold(obj.publicPath);
<del> newline();
<del> }
<add> const getText = (arr, row, col) => {
<add> return arr[row][col].value;
<add> }
<ide>
<del> if(obj.assets && obj.assets.length > 0) {
<del> var t = [
<del> [{
<del> value: "Asset",
<del> color: colors.bold
<del> }, {
<del> value: "Size",
<del> color: colors.bold
<del> }, {
<del> value: "Chunks",
<del> color: colors.bold
<del> }, {
<del> value: "",
<del> color: colors.bold
<del> }, {
<del> value: "",
<del> color: colors.bold
<del> }, {
<del> value: "Chunk Names",
<del> color: colors.bold
<del> }]
<del> ];
<del> obj.assets.forEach(function(asset) {
<del> t.push([{
<del> value: asset.name,
<del> color: getAssetColor(asset, colors.green)
<del> }, {
<del> value: SizeFormatHelpers.formatSize(asset.size),
<del> color: getAssetColor(asset, colors.normal)
<del> }, {
<del> value: asset.chunks.join(", "),
<del> color: colors.bold
<del> }, {
<del> value: asset.emitted ? "[emitted]" : "",
<del> color: colors.green
<del> }, {
<del> value: asset.isOverSizeLimit ? "[big]" : "",
<del> color: getAssetColor(asset, colors.normal)
<del> }, {
<del> value: asset.chunkNames.join(", "),
<del> color: colors.normal
<del> }]);
<del> });
<del> table(t, "rrrlll");
<del> }
<del> if(obj.entrypoints) {
<del> Object.keys(obj.entrypoints).forEach(function(name) {
<del> var ep = obj.entrypoints[name];
<del> colors.normal("Entrypoint ");
<del> colors.bold(name);
<del> if(ep.isOverSizeLimit) {
<del> colors.normal(" ");
<del> colors.yellow("[big]");
<add> const table = (array, align, splitter) => {
<add> let row;
<add> const rows = array.length;
<add> let col;
<add> const cols = array[0].length;
<add> const colSizes = new Array(cols);
<add> let value;
<add> for(col = 0; col < cols; col++)
<add> colSizes[col] = 0;
<add> for(row = 0; row < rows; row++) {
<add> for(col = 0; col < cols; col++) {
<add> value = `${getText(array, row, col)}`;
<add> if(value.length > colSizes[col]) {
<add> colSizes[col] = value.length;
<add> }
<add> }
<ide> }
<del> colors.normal(" =");
<del> ep.assets.forEach(function(asset) {
<del> colors.normal(" ");
<del> colors.green(asset);
<del> });
<del> newline();
<del> });
<del> }
<del> var modulesByIdentifier = {};
<del> if(obj.modules) {
<del> obj.modules.forEach(function(module) {
<del> modulesByIdentifier["$" + module.identifier] = module;
<del> });
<del> } else if(obj.chunks) {
<del> obj.chunks.forEach(function(chunk) {
<del> if(chunk.modules) {
<del> chunk.modules.forEach(function(module) {
<del> modulesByIdentifier["$" + module.identifier] = module;
<del> });
<add> for(row = 0; row < rows; row++) {
<add> for(col = 0; col < cols; col++) {
<add> const format = array[row][col].color;
<add> value = `${getText(array, row, col)}`;
<add> let l = value.length;
<add> if(align[col] === "l")
<add> format(value);
<add> for(; l < colSizes[col] && col !== cols - 1; l++)
<add> colors.normal(" ");
<add> if(align[col] === "r")
<add> format(value);
<add> if(col + 1 < cols && colSizes[col] !== 0)
<add> colors.normal(splitter || " ");
<add> }
<add> newline();
<ide> }
<del> });
<del> }
<del>
<del> function processModuleAttributes(module) {
<del> colors.normal(" ");
<del> colors.normal(SizeFormatHelpers.formatSize(module.size));
<del> if(module.chunks) {
<del> module.chunks.forEach(function(chunk) {
<del> colors.normal(" {");
<del> colors.yellow(chunk);
<del> colors.normal("}");
<del> });
<ide> }
<del> if(typeof module.depth === "number") {
<del> colors.normal(" [depth " + module.depth + "]");
<del> }
<del> if(!module.cacheable) {
<del> colors.red(" [not cacheable]");
<add>
<add> const getAssetColor = (asset, defaultColor) => {
<add> if(asset.isOverSizeLimit) {
<add> return colors.yellow;
<add> }
<add>
<add> return defaultColor;
<ide> }
<del> if(module.optional) {
<del> colors.yellow(" [optional]");
<add>
<add> if(obj.hash) {
<add> colors.normal("Hash: ");
<add> colors.bold(obj.hash);
<add> newline();
<ide> }
<del> if(module.built) {
<del> colors.green(" [built]");
<add> if(obj.version) {
<add> colors.normal("Version: webpack ");
<add> colors.bold(obj.version);
<add> newline();
<ide> }
<del> if(module.prefetched) {
<del> colors.magenta(" [prefetched]");
<add> if(typeof obj.time === "number") {
<add> colors.normal("Time: ");
<add> colors.bold(obj.time);
<add> colors.normal("ms");
<add> newline();
<ide> }
<del> if(module.failed)
<del> colors.red(" [failed]");
<del> if(module.warnings)
<del> colors.yellow(" [" + module.warnings + " warning" + (module.warnings === 1 ? "" : "s") + "]");
<del> if(module.errors)
<del> colors.red(" [" + module.errors + " error" + (module.errors === 1 ? "" : "s") + "]");
<del> }
<del>
<del> function processModuleContent(module, prefix) {
<del> if(Array.isArray(module.providedExports)) {
<del> colors.normal(prefix);
<del> colors.cyan("[exports: " + module.providedExports.join(", ") + "]");
<add> if(obj.publicPath) {
<add> colors.normal("PublicPath: ");
<add> colors.bold(obj.publicPath);
<ide> newline();
<ide> }
<del> if(module.usedExports !== undefined) {
<del> if(module.usedExports !== true) {
<del> colors.normal(prefix);
<del> if(module.usedExports === false)
<del> colors.cyan("[no exports used]");
<del> else
<del> colors.cyan("[only some exports used: " + module.usedExports.join(", ") + "]");
<del> newline();
<del> }
<add>
<add> if(obj.assets && obj.assets.length > 0) {
<add> const t = [
<add> [{
<add> value: "Asset",
<add> color: colors.bold
<add> }, {
<add> value: "Size",
<add> color: colors.bold
<add> }, {
<add> value: "Chunks",
<add> color: colors.bold
<add> }, {
<add> value: "",
<add> color: colors.bold
<add> }, {
<add> value: "",
<add> color: colors.bold
<add> }, {
<add> value: "Chunk Names",
<add> color: colors.bold
<add> }]
<add> ];
<add> obj.assets.forEach(asset => {
<add> t.push([{
<add> value: asset.name,
<add> color: getAssetColor(asset, colors.green)
<add> }, {
<add> value: SizeFormatHelpers.formatSize(asset.size),
<add> color: getAssetColor(asset, colors.normal)
<add> }, {
<add> value: asset.chunks.join(", "),
<add> color: colors.bold
<add> }, {
<add> value: asset.emitted ? "[emitted]" : "",
<add> color: colors.green
<add> }, {
<add> value: asset.isOverSizeLimit ? "[big]" : "",
<add> color: getAssetColor(asset, colors.normal)
<add> }, {
<add> value: asset.chunkNames.join(", "),
<add> color: colors.normal
<add> }]);
<add> });
<add> table(t, "rrrlll");
<ide> }
<del> if(module.reasons) {
<del> module.reasons.forEach(function(reason) {
<del> colors.normal(prefix);
<del> colors.normal(reason.type);
<del> colors.normal(" ");
<del> colors.cyan(reason.userRequest);
<del> if(reason.templateModules) colors.cyan(reason.templateModules.join(" "));
<del> colors.normal(" [");
<del> colors.normal(reason.moduleId);
<del> colors.normal("] ");
<del> colors.magenta(reason.module);
<del> if(reason.loc) {
<add> if(obj.entrypoints) {
<add> Object.keys(obj.entrypoints).forEach(name => {
<add> const ep = obj.entrypoints[name];
<add> colors.normal("Entrypoint ");
<add> colors.bold(name);
<add> if(ep.isOverSizeLimit) {
<ide> colors.normal(" ");
<del> colors.normal(reason.loc);
<add> colors.yellow("[big]");
<ide> }
<add> colors.normal(" =");
<add> ep.assets.forEach(asset => {
<add> colors.normal(" ");
<add> colors.green(asset);
<add> });
<ide> newline();
<ide> });
<ide> }
<del> if(module.profile) {
<del> colors.normal(prefix);
<del> var sum = 0;
<del> var path = [];
<del> var current = module;
<del> while(current.issuer) {
<del> path.unshift(current = current.issuer);
<del> }
<del> path.forEach(function(module) {
<del> colors.normal("[");
<del> colors.normal(module.id);
<del> colors.normal("] ");
<del> if(module.profile) {
<del> var time = (module.profile.factory || 0) + (module.profile.building || 0);
<del> coloredTime(time);
<del> sum += time;
<del> colors.normal(" ");
<del> }
<del> colors.normal("->");
<add> const modulesByIdentifier = {};
<add> if(obj.modules) {
<add> obj.modules.forEach(module => {
<add> modulesByIdentifier[`$${module.identifier}`] = module;
<ide> });
<del> Object.keys(module.profile).forEach(function(key) {
<del> colors.normal(" " + key + ":");
<del> var time = module.profile[key];
<del> coloredTime(time);
<del> sum += time;
<add> } else if(obj.chunks) {
<add> obj.chunks.forEach(chunk => {
<add> if(chunk.modules) {
<add> chunk.modules.forEach(module => {
<add> modulesByIdentifier[`$${module.identifier}`] = module;
<add> });
<add> }
<ide> });
<del> colors.normal(" = ");
<del> coloredTime(sum);
<del> newline();
<ide> }
<del> }
<ide>
<del> if(obj.chunks) {
<del> obj.chunks.forEach(function(chunk) {
<del> colors.normal("chunk ");
<del> if(chunk.id < 1000) colors.normal(" ");
<del> if(chunk.id < 100) colors.normal(" ");
<del> if(chunk.id < 10) colors.normal(" ");
<del> colors.normal("{");
<del> colors.yellow(chunk.id);
<del> colors.normal("} ");
<del> colors.green(chunk.files.join(", "));
<del> if(chunk.names && chunk.names.length > 0) {
<del> colors.normal(" (");
<del> colors.normal(chunk.names.join(", "));
<del> colors.normal(")");
<del> }
<add> const processModuleAttributes = (module) => {
<ide> colors.normal(" ");
<del> colors.normal(SizeFormatHelpers.formatSize(chunk.size));
<del> chunk.parents.forEach(function(id) {
<del> colors.normal(" {");
<del> colors.yellow(id);
<del> colors.normal("}");
<del> });
<del> if(chunk.entry) {
<del> colors.yellow(" [entry]");
<del> } else if(chunk.initial) {
<del> colors.yellow(" [initial]");
<add> colors.normal(SizeFormatHelpers.formatSize(module.size));
<add> if(module.chunks) {
<add> module.chunks.forEach(chunk => {
<add> colors.normal(" {");
<add> colors.yellow(chunk);
<add> colors.normal("}");
<add> });
<ide> }
<del> if(chunk.rendered) {
<del> colors.green(" [rendered]");
<add> if(typeof module.depth === "number") {
<add> colors.normal(` [depth ${module.depth}]`);
<ide> }
<del> if(chunk.recorded) {
<del> colors.green(" [recorded]");
<add> if(!module.cacheable) {
<add> colors.red(" [not cacheable]");
<ide> }
<del> newline();
<del> if(chunk.origins) {
<del> chunk.origins.forEach(function(origin) {
<del> colors.normal(" > ");
<del> if(origin.reasons && origin.reasons.length) {
<del> colors.yellow(origin.reasons.join(" "));
<del> colors.normal(" ");
<del> }
<del> if(origin.name) {
<del> colors.normal(origin.name);
<add> if(module.optional) {
<add> colors.yellow(" [optional]");
<add> }
<add> if(module.built) {
<add> colors.green(" [built]");
<add> }
<add> if(module.prefetched) {
<add> colors.magenta(" [prefetched]");
<add> }
<add> if(module.failed)
<add> colors.red(" [failed]");
<add> if(module.warnings)
<add> colors.yellow(` [${module.warnings} warning${module.warnings === 1 ? "" : "s"}]`);
<add> if(module.errors)
<add> colors.red(` [${module.errors} error${module.errors === 1 ? "" : "s"}]`);
<add> }
<add>
<add> const processModuleContent = (module, prefix) => {
<add> if(Array.isArray(module.providedExports)) {
<add> colors.normal(prefix);
<add> colors.cyan(`[exports: ${module.providedExports.join(", ")}]`);
<add> newline();
<add> }
<add> if(module.usedExports !== undefined) {
<add> if(module.usedExports !== true) {
<add> colors.normal(prefix);
<add> if(module.usedExports === false)
<add> colors.cyan("[no exports used]");
<add> else
<add> colors.cyan(`[only some exports used: ${module.usedExports.join(", ")}]`);
<add> newline();
<add> }
<add> }
<add> if(module.reasons) {
<add> module.reasons.forEach(reason => {
<add> colors.normal(prefix);
<add> colors.normal(reason.type);
<add> colors.normal(" ");
<add> colors.cyan(reason.userRequest);
<add> if(reason.templateModules) colors.cyan(reason.templateModules.join(" "));
<add> colors.normal(" [");
<add> colors.normal(reason.moduleId);
<add> colors.normal("] ");
<add> colors.magenta(reason.module);
<add> if(reason.loc) {
<ide> colors.normal(" ");
<del> }
<del> if(origin.module) {
<del> colors.normal("[");
<del> colors.normal(origin.moduleId);
<del> colors.normal("] ");
<del> var module = modulesByIdentifier["$" + origin.module];
<del> if(module) {
<del> colors.bold(module.name);
<del> colors.normal(" ");
<del> }
<del> if(origin.loc) {
<del> colors.normal(origin.loc);
<del> }
<add> colors.normal(reason.loc);
<ide> }
<ide> newline();
<ide> });
<ide> }
<del> if(chunk.modules) {
<del> chunk.modules.forEach(function(module) {
<del> colors.normal(" ");
<del> if(module.id < 1000) colors.normal(" ");
<del> if(module.id < 100) colors.normal(" ");
<del> if(module.id < 10) colors.normal(" ");
<add> if(module.profile) {
<add> colors.normal(prefix);
<add> let sum = 0;
<add> const path = [];
<add> let current = module;
<add> while(current.issuer) {
<add> path.unshift(current = current.issuer);
<add> }
<add> path.forEach(module => {
<ide> colors.normal("[");
<ide> colors.normal(module.id);
<ide> colors.normal("] ");
<del> colors.bold(module.name);
<del> processModuleAttributes(module);
<del> newline();
<del> processModuleContent(module, " ");
<add> if(module.profile) {
<add> const time = (module.profile.factory || 0) + (module.profile.building || 0);
<add> coloredTime(time);
<add> sum += time;
<add> colors.normal(" ");
<add> }
<add> colors.normal("->");
<ide> });
<del> if(chunk.filteredModules > 0) {
<del> colors.normal(" + " + chunk.filteredModules + " hidden modules");
<del> newline();
<del> }
<add> Object.keys(module.profile).forEach(key => {
<add> colors.normal(` ${key}:`);
<add> const time = module.profile[key];
<add> coloredTime(time);
<add> sum += time;
<add> });
<add> colors.normal(" = ");
<add> coloredTime(sum);
<add> newline();
<ide> }
<del> });
<del> }
<del> if(obj.modules) {
<del> obj.modules.forEach(function(module) {
<del> if(module.id < 1000) colors.normal(" ");
<del> if(module.id < 100) colors.normal(" ");
<del> if(module.id < 10) colors.normal(" ");
<del> colors.normal("[");
<del> colors.normal(module.id);
<del> colors.normal("] ");
<del> colors.bold(module.name || module.identifier);
<del> processModuleAttributes(module);
<del> newline();
<del> processModuleContent(module, " ");
<del> });
<del> if(obj.filteredModules > 0) {
<del> colors.normal(" + " + obj.filteredModules + " hidden modules");
<del> newline();
<ide> }
<del> }
<del> if(obj._showWarnings && obj.warnings) {
<del> obj.warnings.forEach(function(warning) {
<del> newline();
<del> colors.yellow("WARNING in " + warning);
<del> newline();
<del> });
<del> }
<del> if(obj._showErrors && obj.errors) {
<del> obj.errors.forEach(function(error) {
<del> newline();
<del> colors.red("ERROR in " + error);
<del> newline();
<del> });
<del> }
<del> if(obj.children) {
<del> obj.children.forEach(function(child) {
<del> if(child.name) {
<del> colors.normal("Child ");
<del> colors.bold(child.name);
<del> colors.normal(":");
<del> } else {
<del> colors.normal("Child");
<add>
<add> if(obj.chunks) {
<add> obj.chunks.forEach(chunk => {
<add> colors.normal("chunk ");
<add> if(chunk.id < 1000) colors.normal(" ");
<add> if(chunk.id < 100) colors.normal(" ");
<add> if(chunk.id < 10) colors.normal(" ");
<add> colors.normal("{");
<add> colors.yellow(chunk.id);
<add> colors.normal("} ");
<add> colors.green(chunk.files.join(", "));
<add> if(chunk.names && chunk.names.length > 0) {
<add> colors.normal(" (");
<add> colors.normal(chunk.names.join(", "));
<add> colors.normal(")");
<add> }
<add> colors.normal(" ");
<add> colors.normal(SizeFormatHelpers.formatSize(chunk.size));
<add> chunk.parents.forEach(id => {
<add> colors.normal(" {");
<add> colors.yellow(id);
<add> colors.normal("}");
<add> });
<add> if(chunk.entry) {
<add> colors.yellow(" [entry]");
<add> } else if(chunk.initial) {
<add> colors.yellow(" [initial]");
<add> }
<add> if(chunk.rendered) {
<add> colors.green(" [rendered]");
<add> }
<add> if(chunk.recorded) {
<add> colors.green(" [recorded]");
<add> }
<add> newline();
<add> if(chunk.origins) {
<add> chunk.origins.forEach(origin => {
<add> colors.normal(" > ");
<add> if(origin.reasons && origin.reasons.length) {
<add> colors.yellow(origin.reasons.join(" "));
<add> colors.normal(" ");
<add> }
<add> if(origin.name) {
<add> colors.normal(origin.name);
<add> colors.normal(" ");
<add> }
<add> if(origin.module) {
<add> colors.normal("[");
<add> colors.normal(origin.moduleId);
<add> colors.normal("] ");
<add> const module = modulesByIdentifier[`$${origin.module}`];
<add> if(module) {
<add> colors.bold(module.name);
<add> colors.normal(" ");
<add> }
<add> if(origin.loc) {
<add> colors.normal(origin.loc);
<add> }
<add> }
<add> newline();
<add> });
<add> }
<add> if(chunk.modules) {
<add> chunk.modules.forEach(module => {
<add> colors.normal(" ");
<add> if(module.id < 1000) colors.normal(" ");
<add> if(module.id < 100) colors.normal(" ");
<add> if(module.id < 10) colors.normal(" ");
<add> colors.normal("[");
<add> colors.normal(module.id);
<add> colors.normal("] ");
<add> colors.bold(module.name);
<add> processModuleAttributes(module);
<add> newline();
<add> processModuleContent(module, " ");
<add> });
<add> if(chunk.filteredModules > 0) {
<add> colors.normal(` + ${chunk.filteredModules} hidden modules`);
<add> newline();
<add> }
<add> }
<add> });
<add> }
<add> if(obj.modules) {
<add> obj.modules.forEach(module => {
<add> if(module.id < 1000) colors.normal(" ");
<add> if(module.id < 100) colors.normal(" ");
<add> if(module.id < 10) colors.normal(" ");
<add> colors.normal("[");
<add> colors.normal(module.id);
<add> colors.normal("] ");
<add> colors.bold(module.name || module.identifier);
<add> processModuleAttributes(module);
<add> newline();
<add> processModuleContent(module, " ");
<add> });
<add> if(obj.filteredModules > 0) {
<add> colors.normal(` + ${obj.filteredModules} hidden modules`);
<add> newline();
<ide> }
<del> newline();
<del> buf.push(" ");
<del> buf.push(Stats.jsonToString(child, useColors).replace(/\n/g, "\n "));
<del> newline();
<del> });
<del> }
<del> if(obj.needAdditionalPass) {
<del> colors.yellow("Compilation needs an additional pass and will compile again.");
<del> }
<add> }
<add> if(obj._showWarnings && obj.warnings) {
<add> obj.warnings.forEach(warning => {
<add> newline();
<add> colors.yellow(`WARNING in ${warning}`);
<add> newline();
<add> });
<add> }
<add> if(obj._showErrors && obj.errors) {
<add> obj.errors.forEach(error => {
<add> newline();
<add> colors.red(`ERROR in ${error}`);
<add> newline();
<add> });
<add> }
<add> if(obj.children) {
<add> obj.children.forEach(child => {
<add> if(child.name) {
<add> colors.normal("Child ");
<add> colors.bold(child.name);
<add> colors.normal(":");
<add> } else {
<add> colors.normal("Child");
<add> }
<add> newline();
<add> buf.push(" ");
<add> buf.push(Stats.jsonToString(child, useColors).replace(/\n/g, "\n "));
<add> newline();
<add> });
<add> }
<add> if(obj.needAdditionalPass) {
<add> colors.yellow("Compilation needs an additional pass and will compile again.");
<add> }
<ide>
<del> while(buf[buf.length - 1] === "\n") buf.pop();
<del> return buf.join("");
<del>};
<del>
<del>Stats.presetToOptions = function(name) {
<del> //Accepted values: none, errors-only, minimal, normal, verbose
<del> //Any other falsy value will behave as 'none', truthy values as 'normal'
<del> var pn = (typeof name === "string") && name.toLowerCase() || name;
<del> if(pn === "none" || !pn) {
<del> return {
<del> hash: false,
<del> version: false,
<del> timings: false,
<del> assets: false,
<del> entrypoints: false,
<del> chunks: false,
<del> chunkModules: false,
<del> modules: false,
<del> reasons: false,
<del> depth: false,
<del> usedExports: false,
<del> providedExports: false,
<del> children: false,
<del> source: false,
<del> errors: false,
<del> errorDetails: false,
<del> warnings: false,
<del> publicPath: false
<del> };
<del> } else {
<del> return {
<del> hash: pn !== "errors-only" && pn !== "minimal",
<del> version: pn === "verbose",
<del> timings: pn !== "errors-only" && pn !== "minimal",
<del> assets: pn === "verbose",
<del> entrypoints: pn === "verbose",
<del> chunks: pn !== "errors-only",
<del> chunkModules: pn === "verbose",
<del> //warnings: pn !== "errors-only",
<del> errorDetails: pn !== "errors-only" && pn !== "minimal",
<del> reasons: pn === "verbose",
<del> depth: pn === "verbose",
<del> usedExports: pn === "verbose",
<del> providedExports: pn === "verbose",
<del> colors: true
<del> };
<add> while(buf[buf.length - 1] === "\n") buf.pop();
<add> return buf.join("");
<add> }
<add>
<add> static presetToOptions(name) {
<add> //Accepted values: none, errors-only, minimal, normal, verbose
<add> //Any other falsy value will behave as 'none', truthy values as 'normal'
<add> const pn = (typeof name === "string") && name.toLowerCase() || name;
<add> if(pn === "none" || !pn) {
<add> return {
<add> hash: false,
<add> version: false,
<add> timings: false,
<add> assets: false,
<add> entrypoints: false,
<add> chunks: false,
<add> chunkModules: false,
<add> modules: false,
<add> reasons: false,
<add> depth: false,
<add> usedExports: false,
<add> providedExports: false,
<add> children: false,
<add> source: false,
<add> errors: false,
<add> errorDetails: false,
<add> warnings: false,
<add> publicPath: false
<add> };
<add> } else {
<add> return {
<add> hash: pn !== "errors-only" && pn !== "minimal",
<add> version: pn === "verbose",
<add> timings: pn !== "errors-only" && pn !== "minimal",
<add> assets: pn === "verbose",
<add> entrypoints: pn === "verbose",
<add> chunks: pn !== "errors-only",
<add> chunkModules: pn === "verbose",
<add> //warnings: pn !== "errors-only",
<add> errorDetails: pn !== "errors-only" && pn !== "minimal",
<add> reasons: pn === "verbose",
<add> depth: pn === "verbose",
<add> usedExports: pn === "verbose",
<add> providedExports: pn === "verbose",
<add> colors: true
<add> };
<add> }
<ide> }
<del>};
<add>}
<add>
<add>module.exports = Stats;
| 1
|
Mixed
|
Ruby
|
dump warns when `mysqldump` is not in path
|
0a5fdcd5ae2f7928ce9ce492c9815ad71f858a9f
|
<ide><path>activerecord/CHANGELOG.md
<ide> ## Rails 4.0.0 (unreleased) ##
<ide>
<add>* Warn when `rake db:structure:dump` with a mysl database and
<add> `mysqldump` is not in the PATH or fails.
<add> Fixes #9518.
<add>
<add> *Yves Senn*
<add>
<ide> * Remove `connection#structure_dump`, which is no longer used. *Yves Senn*
<ide>
<ide> * Make it possible to execute migrations without a transaction even
<ide><path>activerecord/lib/active_record/tasks/mysql_database_tasks.rb
<ide> def structure_dump(filename)
<ide> args.concat(["--result-file", "#{filename}"])
<ide> args.concat(["--no-data"])
<ide> args.concat(["#{configuration['database']}"])
<del> Kernel.system(*args)
<add> unless Kernel.system(*args)
<add> $stderr.puts "Could not dump the database structure. "\
<add> "Make sure `mysqldump` is in your PATH and check the command output for warnings."
<add> end
<ide> end
<ide>
<ide> def structure_load(filename)
<ide><path>activerecord/test/cases/tasks/mysql_rake_test.rb
<ide> def setup
<ide>
<ide> def test_structure_dump
<ide> filename = "awesome-file.sql"
<del> Kernel.expects(:system).with("mysqldump", "--result-file", filename, "--no-data", "test-db")
<add> Kernel.expects(:system).with("mysqldump", "--result-file", filename, "--no-data", "test-db").returns(true)
<ide>
<ide> ActiveRecord::Tasks::DatabaseTasks.structure_dump(@configuration, filename)
<ide> end
<add>
<add> def test_warn_when_external_structure_dump_fails
<add> filename = "awesome-file.sql"
<add> Kernel.expects(:system).with("mysqldump", "--result-file", filename, "--no-data", "test-db").returns(false)
<add>
<add> warnings = capture(:stderr) do
<add> ActiveRecord::Tasks::DatabaseTasks.structure_dump(@configuration, filename)
<add> end
<add>
<add> assert_match(/Could not dump the database structure/, warnings)
<add> end
<ide> end
<ide>
<ide> class MySQLStructureLoadTest < ActiveRecord::TestCase
| 3
|
PHP
|
PHP
|
add regression test for
|
264054d9ef36fb6b355417fb56a3cb1147441a6d
|
<ide><path>tests/TestCase/ORM/EntityTest.php
<ide> public function testErrors() {
<ide> $this->assertSame($entity, $entity->errors('foo', 'bar'));
<ide> $this->assertEquals(['bar'], $entity->errors('foo'));
<ide>
<add> $this->assertEquals([], $entity->errors('boo'));
<add> $entity['boo'] = [
<add> 'someting' => 'stupid',
<add> 'and' => false
<add> ];
<add> $this->assertEquals([], $entity->errors('boo'));
<add>
<ide> $entity->errors('foo', 'other error');
<ide> $this->assertEquals(['other error'], $entity->errors('foo'));
<ide>
| 1
|
Ruby
|
Ruby
|
remove unneeded requires
|
a227359c6289613c870c8bb1481754ab00325c12
|
<ide><path>actionpack/lib/action_view/digestor.rb
<del>require 'active_support/core_ext'
<del>require 'logger'
<del>
<ide> module ActionView
<ide> class Digestor
<ide> EXPLICIT_DEPENDENCY = /# Template Dependency: ([^ ]+)/
<ide> def nested_dependencies
<ide> def logical_name
<ide> name.gsub(%r|/_|, "/")
<ide> end
<del>
<add>
<ide> def directory
<ide> name.split("/").first
<ide> end
<ide> def source
<ide> @source ||= finder.find(logical_name, [], partial?, formats: [ format ]).source
<ide> end
<ide>
<del>
<ide> def dependency_digest
<ide> dependencies.collect do |template_name|
<ide> Digestor.digest(template_name, format, finder, partial: true)
<ide> def render_dependencies
<ide>
<ide> # render("headline") => render("message/headline")
<ide> collect { |name| name.include?("/") ? name : "#{directory}/#{name}" }.
<del>
<add>
<ide> # replace quotes from string renders
<ide> collect { |name| name.gsub(/["']/, "") }
<ide> end
<ide> def explicit_dependencies
<ide> source.scan(EXPLICIT_DEPENDENCY).flatten.uniq
<ide> end
<ide> end
<del>end
<ide>\ No newline at end of file
<add>end
| 1
|
Ruby
|
Ruby
|
add nuclear to github_prerelease_allowlist
|
a176c3f75ba60ecca3701ceb219de6592dbe66d6
|
<ide><path>Library/Homebrew/utils/shared_audits.rb
<ide> def github_release_data(user, repo, tag)
<ide> "gitless" => "0.8.8",
<ide> "home-assistant" => :all,
<ide> "infrakit" => "0.5",
<add> "nuclear" => :all,
<ide> "pock" => :all,
<ide> "riff" => "0.5.0",
<ide> "telegram-cli" => "1.3.1",
| 1
|
PHP
|
PHP
|
add methods to get/set middleware on a route
|
9d9e566d8552450202a2d742f0024857d8f6d01b
|
<ide><path>src/Routing/Route/Route.php
<ide> class Route
<ide> */
<ide> protected $_extensions = [];
<ide>
<add> /**
<add> * List of middleware that should be applied.
<add> *
<add> * @var array
<add> */
<add> protected $middleware = [];
<add>
<ide> /**
<ide> * Valid HTTP methods.
<ide> *
<ide> class Route
<ide> * ### Options
<ide> *
<ide> * - `_ext` - Defines the extensions used for this route.
<add> * - `_middleware` - Define the middleware names for this route.
<ide> * - `pass` - Copies the listed parameters into params['pass'].
<ide> * - `_host` - Define the host name pattern if you want this route to only match
<ide> * specific host names. You can use `.*` and to create wildcard subdomains/hosts
<ide> public function __construct($template, $defaults = [], array $options = [])
<ide> unset($defaults['[method]']);
<ide> }
<ide> $this->defaults = (array)$defaults;
<del> $this->options = $options + ['_ext' => []];
<add> $this->options = $options + ['_ext' => [], '_middleware' => []];
<ide> $this->setExtensions((array)$this->options['_ext']);
<add> $this->setMiddleware((array)$this->options['_middleware']);
<add> unset($this->options['_middleware']);
<ide> }
<ide>
<ide> /**
<ide> public function staticPath()
<ide> return $this->template;
<ide> }
<ide>
<add> /**
<add> * Set the names of the middleware that should be applied to this route.
<add> *
<add> * @param array $middleware The list of middleware names to apply to this route.
<add> * Middleware names will not be checked until the route is matched.
<add> * @return $this
<add> */
<add> public function setMiddleware(array $middleware)
<add> {
<add> $this->middleware = $middleware;
<add>
<add> return $this;
<add> }
<add>
<add> /**
<add> * Get the names of the middleware that should be applied to this route.
<add> *
<add> * @return array
<add> */
<add> public function getMiddleware()
<add> {
<add> return $this->middleware;
<add> }
<add>
<ide> /**
<ide> * Set state magic method to support var_export
<ide> *
<ide><path>tests/TestCase/Routing/Route/RouteTest.php
<ide> public function testConstruction()
<ide> $this->assertFalse($route->compiled());
<ide> }
<ide>
<add> /**
<add> * Test set middleware in the constructor
<add> *
<add> * @return void
<add> */
<add> public function testConstructorSetMiddleware()
<add> {
<add> $route = new Route('/:controller/:action/*', [], ['_middleware' => ['auth', 'cookie']]);
<add> $this->assertSame(['auth', 'cookie'], $route->getMiddleware());
<add> }
<add>
<ide> /**
<ide> * Test Route compiling.
<ide> *
<ide> public function testSetPersist()
<ide> $this->assertSame($result, $route, 'Should return this');
<ide> $this->assertEquals(['date'], $route->options['persist']);
<ide> }
<add>
<add> /**
<add> * Test setting/getting middleware.
<add> *
<add> * @return void
<add> */
<add> public function testSetMiddleware()
<add> {
<add> $route = new Route('/reviews/:date/:id', ['controller' => 'Reviews', 'action' => 'view']);
<add> $result = $route->setMiddleware(['auth', 'cookie']);
<add> $this->assertSame($result, $route);
<add> $this->assertSame(['auth', 'cookie'], $route->getMiddleware());
<add> }
<ide> }
| 2
|
Mixed
|
Javascript
|
support abortsignal in createsocket
|
21c8c7e5114c56c8326aa15f11e69d58326553da
|
<ide><path>doc/api/dgram.md
<ide> chained.
<ide> <!-- YAML
<ide> added: v0.11.13
<ide> changes:
<add> - version: REPLACEME
<add> pr-url: https://github.com/nodejs/node/pull/37026
<add> description: AbortSignal support was added.
<ide> - version: v11.4.0
<ide> pr-url: https://github.com/nodejs/node/pull/23798
<ide> description: The `ipv6Only` option is supported.
<ide> changes:
<ide> * `recvBufferSize` {number} Sets the `SO_RCVBUF` socket value.
<ide> * `sendBufferSize` {number} Sets the `SO_SNDBUF` socket value.
<ide> * `lookup` {Function} Custom lookup function. **Default:** [`dns.lookup()`][].
<add> * `signal` {AbortSignal} An AbortSignal that may be used to close a socket.
<ide> * `callback` {Function} Attached as a listener for `'message'` events. Optional.
<ide> * Returns: {dgram.Socket}
<ide>
<ide> method will bind the socket to the "all interfaces" address on a random port
<ide> and port can be retrieved using [`socket.address().address`][] and
<ide> [`socket.address().port`][].
<ide>
<add>If the `signal` option is enabled, calling `.abort()` on the corresponding
<add>`AbortController` is similar to calling `.close()` on the socket:
<add>
<add>```js
<add>const controller = new AbortController();
<add>const { signal } = controller;
<add>const server = dgram.createSocket({ type: 'udp4', signal });
<add>server.on('message', (msg, rinfo) => {
<add> console.log(`server got: ${msg} from ${rinfo.address}:${rinfo.port}`);
<add>});
<add>// Later, when you want to close the server.
<add>controller.abort();
<add>```
<add>
<ide> ### `dgram.createSocket(type[, callback])`
<ide> <!-- YAML
<ide> added: v0.1.99
<ide><path>lib/dgram.js
<ide> const {
<ide> } = errors.codes;
<ide> const {
<ide> isInt32,
<add> validateAbortSignal,
<ide> validateString,
<ide> validateNumber,
<ide> validatePort,
<ide> function Socket(type, listener) {
<ide> recvBufferSize,
<ide> sendBufferSize
<ide> };
<add>
<add> if (options?.signal !== undefined) {
<add> const { signal } = options;
<add> validateAbortSignal(signal, 'options.signal');
<add> const onAborted = () => {
<add> this.close();
<add> };
<add> if (signal.aborted) {
<add> onAborted();
<add> } else {
<add> signal.addEventListener('abort', onAborted);
<add> this.once('close', () => signal.removeEventListener('abort', onAborted));
<add> }
<add> }
<ide> }
<ide> ObjectSetPrototypeOf(Socket.prototype, EventEmitter.prototype);
<ide> ObjectSetPrototypeOf(Socket, EventEmitter);
<ide><path>test/parallel/test-dgram-close-signal.js
<add>'use strict';
<add>
<add>const common = require('../common');
<add>const assert = require('assert');
<add>const dgram = require('dgram');
<add>
<add>{
<add> // Test bad signal.
<add> assert.throws(
<add> () => dgram.createSocket({ type: 'udp4', signal: {} }),
<add> {
<add> code: 'ERR_INVALID_ARG_TYPE',
<add> name: 'TypeError'
<add> });
<add>}
<add>
<add>{
<add> // Test close.
<add> const controller = new AbortController();
<add> const { signal } = controller;
<add> const server = dgram.createSocket({ type: 'udp4', signal });
<add> server.on('close', common.mustCall());
<add> controller.abort();
<add>}
<add>
<add>{
<add> // Test close with pre-aborted signal.
<add> const controller = new AbortController();
<add> controller.abort();
<add> const { signal } = controller;
<add> const server = dgram.createSocket({ type: 'udp4', signal });
<add> server.on('close', common.mustCall());
<add>}
| 3
|
Python
|
Python
|
set version to v3.4.1
|
5c2a00cef04b8c6e93e81cd1ca1d752f320c6e5d
|
<ide><path>spacy/about.py
<ide> # fmt: off
<ide> __title__ = "spacy"
<del>__version__ = "3.4.0"
<add>__version__ = "3.4.1"
<ide> __download_url__ = "https://github.com/explosion/spacy-models/releases/download"
<ide> __compatibility__ = "https://raw.githubusercontent.com/explosion/spacy-models/master/compatibility.json"
<ide> __projects__ = "https://github.com/explosion/projects"
| 1
|
Javascript
|
Javascript
|
add highlight decorations, but no tests yet
|
a4224922a3933c93a2827bc914bfb7ad7a95ba3e
|
<ide><path>src/text-editor-component.js
<ide> class TextEditorComponent {
<ide> this.cursorsToRender = []
<ide> this.decorationsToRender = {
<ide> lineNumbers: new Map(),
<del> lines: new Map()
<add> lines: new Map(),
<add> highlights: new Map()
<add> }
<add> this.decorationsToMeasure = {
<add> highlights: new Map()
<ide> }
<ide>
<ide> if (this.props.model) this.observeModel()
<ide> class TextEditorComponent {
<ide> if (longestLineToMeasure) this.measureLongestLineWidth(longestLineToMeasure)
<ide> if (this.pendingAutoscroll) this.finalizeAutoscroll()
<ide> this.positionCursorsToRender()
<add> this.updateHighlightsToRender()
<ide>
<ide> etch.updateSync(this)
<ide>
<ide> class TextEditorComponent {
<ide> for (let row = tileStartRow; row < tileEndRow; row++) {
<ide> lineDecorations[row - tileStartRow] = this.decorationsToRender.lines.get(row)
<ide> }
<add> const highlightDecorations = this.decorationsToRender.highlights.get(tileStartRow)
<ide>
<ide> tileNodes[tileIndex] = $(LinesTileComponent, {
<ide> key: tileIndex,
<ide> height: tileHeight,
<ide> width: tileWidth,
<ide> top: this.topPixelPositionForRow(tileStartRow),
<add> lineHeight: this.measurements.lineHeight,
<ide> screenLines: screenLines.slice(tileStartRow - startRow, tileEndRow - startRow),
<ide> lineDecorations,
<add> highlightDecorations,
<ide> displayLayer,
<ide> lineNodesByScreenLineId,
<ide> textNodesByScreenLineId
<ide> class TextEditorComponent {
<ide> })
<ide> const lastCursorMarker = model.getLastCursor().getMarker()
<ide>
<del> this.cursorsToRender.length = cursorMarkers.length
<add> this.cursorsToRender.length = 0
<ide> this.lastCursorIndex = -1
<ide> for (let i = 0; i < cursorMarkers.length; i++) {
<ide> const cursorMarker = cursorMarkers[i]
<ide> if (cursorMarker === lastCursorMarker) this.lastCursorIndex = i
<ide> const screenPosition = cursorMarker.getHeadScreenPosition()
<ide> const {row, column} = screenPosition
<add>
<add> if (row < this.getRenderedStartRow() || row >= this.getRenderedEndRow()) {
<add> continue
<add> }
<add>
<ide> this.requestHorizontalMeasurement(row, column)
<ide> let columnWidth = 0
<ide> if (model.lineLengthForScreenRow(row) > column) {
<ide> columnWidth = 1
<ide> this.requestHorizontalMeasurement(row, column + 1)
<ide> }
<del> this.cursorsToRender[i] = {
<add> this.cursorsToRender.push({
<ide> screenPosition, columnWidth,
<ide> pixelTop: 0, pixelLeft: 0, pixelWidth: 0
<del> }
<add> })
<ide> }
<ide> }
<ide>
<ide> queryDecorationsToRender () {
<ide> this.decorationsToRender.lineNumbers.clear()
<ide> this.decorationsToRender.lines.clear()
<add> this.decorationsToMeasure.highlights.clear()
<ide>
<ide> const decorationsByMarker =
<ide> this.getModel().decorationManager.decorationPropertiesByMarkerForScreenRowRange(
<ide> class TextEditorComponent {
<ide> switch (type) {
<ide> case 'line':
<ide> case 'line-number':
<del> const decorationsByRow = (type === 'line') ? this.decorationsToRender.lines : this.decorationsToRender.lineNumbers
<add> this.addLineDecorationToRender(type, decoration, screenRange, reversed)
<add> break
<add> case 'highlight':
<add> this.addHighlightDecorationToMeasure(decoration, screenRange)
<add> break
<add> }
<add> }
<add> }
<ide>
<del> let omitLastRow = false
<del> if (screenRange.isEmpty()) {
<del> if (decoration.onlyNonEmpty) return
<del> } else {
<del> if (decoration.onlyEmpty) return
<del> if (decoration.omitEmptyLastRow !== false) {
<del> omitLastRow = screenRange.end.column === 0
<del> }
<del> }
<add> addLineDecorationToRender (type, decoration, screenRange, reversed) {
<add> const decorationsByRow = (type === 'line') ? this.decorationsToRender.lines : this.decorationsToRender.lineNumbers
<add>
<add> let omitLastRow = false
<add> if (screenRange.isEmpty()) {
<add> if (decoration.onlyNonEmpty) return
<add> } else {
<add> if (decoration.onlyEmpty) return
<add> if (decoration.omitEmptyLastRow !== false) {
<add> omitLastRow = screenRange.end.column === 0
<add> }
<add> }
<ide>
<del> let startRow = screenRange.start.row
<del> let endRow = screenRange.end.row
<add> let startRow = screenRange.start.row
<add> let endRow = screenRange.end.row
<ide>
<del> if (decoration.onlyHead) {
<del> if (reversed) {
<del> endRow = startRow
<del> } else {
<del> startRow = endRow
<del> }
<del> }
<add> if (decoration.onlyHead) {
<add> if (reversed) {
<add> endRow = startRow
<add> } else {
<add> startRow = endRow
<add> }
<add> }
<ide>
<del> for (let row = startRow; row <= endRow; row++) {
<del> if (omitLastRow && row === endRow) break
<del> const currentClassName = decorationsByRow.get(row)
<del> const newClassName = currentClassName ? currentClassName + ' ' + decoration.class : decoration.class
<del> decorationsByRow.set(row, newClassName)
<del> }
<del> break
<add> for (let row = startRow; row <= endRow; row++) {
<add> if (omitLastRow && row === endRow) break
<add> const currentClassName = decorationsByRow.get(row)
<add> const newClassName = currentClassName ? currentClassName + ' ' + decoration.class : decoration.class
<add> decorationsByRow.set(row, newClassName)
<add> }
<add> }
<add>
<add> addHighlightDecorationToMeasure(decoration, screenRange) {
<add> screenRange = constrainRangeToRows(screenRange, this.getRenderedStartRow(), this.getRenderedEndRow())
<add> if (screenRange.isEmpty()) return
<add> let tileStartRow = this.tileStartRowForRow(screenRange.start.row)
<add> const rowsPerTile = this.getRowsPerTile()
<add>
<add> while (tileStartRow <= screenRange.end.row) {
<add> const tileEndRow = tileStartRow + rowsPerTile
<add> const screenRangeInTile = constrainRangeToRows(screenRange, tileStartRow, tileEndRow)
<add>
<add> let tileHighlights = this.decorationsToMeasure.highlights.get(tileStartRow)
<add> if (!tileHighlights) {
<add> tileHighlights = []
<add> this.decorationsToMeasure.highlights.set(tileStartRow, tileHighlights)
<ide> }
<add> tileHighlights.push({decoration, screenRange: screenRangeInTile})
<add>
<add> this.requestHorizontalMeasurement(screenRangeInTile.start.row, screenRangeInTile.start.column)
<add> this.requestHorizontalMeasurement(screenRangeInTile.end.row, screenRangeInTile.end.column)
<add>
<add> tileStartRow += rowsPerTile
<ide> }
<ide> }
<ide>
<add> updateHighlightsToRender () {
<add> this.decorationsToRender.highlights.clear()
<add> this.decorationsToMeasure.highlights.forEach((highlights, tileRow) => {
<add> for (let i = 0, length = highlights.length; i < length; i++) {
<add> const highlight = highlights[i]
<add> const {start, end} = highlight.screenRange
<add> highlight.startPixelTop = this.pixelTopForScreenRow(start.row)
<add> highlight.startPixelLeft = this.pixelLeftForScreenRowAndColumn(start.row, start.column)
<add> highlight.endPixelTop = this.pixelTopForScreenRow(end.row + 1)
<add> highlight.endPixelLeft = this.pixelLeftForScreenRowAndColumn(end.row, end.column)
<add> }
<add> this.decorationsToRender.highlights.set(tileRow, highlights)
<add> })
<add> }
<ide>
<ide> positionCursorsToRender () {
<ide> const height = this.measurements.lineHeight + 'px'
<ide> class TextEditorComponent {
<ide> }
<ide>
<ide> requestHorizontalMeasurement (row, column) {
<add> if (column === 0) return
<ide> let columns = this.horizontalPositionsToMeasure.get(row)
<ide> if (columns == null) {
<ide> columns = []
<ide> class TextEditorComponent {
<ide>
<ide> const screenLine = this.getModel().displayLayer.getScreenLine(row)
<ide> const lineNode = this.lineNodesByScreenLineId.get(screenLine.id)
<add>
<add> if (!lineNode) {
<add> const error = new Error('Requested measurement of a line that is not currently rendered')
<add> error.metadata = {row, columnsToMeasure}
<add> throw error
<add> }
<add>
<ide> const textNodes = this.textNodesByScreenLineId.get(screenLine.id)
<ide> let positionsForLine = this.horizontalPixelPositionsByScreenLineId.get(screenLine.id)
<ide> if (positionsForLine == null) {
<ide> class LinesTileComponent {
<ide> }
<ide>
<ide> render () {
<add> const {height, width, top} = this.props
<add>
<add> return $.div(
<add> {
<add> style: {
<add> contain: 'strict',
<add> position: 'absolute',
<add> height: height + 'px',
<add> width: width + 'px',
<add> willChange: 'transform',
<add> transform: `translateY(${top}px)`,
<add> backgroundColor: 'inherit'
<add> }
<add> },
<add> this.renderHighlights(),
<add> this.renderLines()
<add> )
<add>
<add> }
<add>
<add> renderHighlights () {
<add> const {top, height, width, lineHeight, highlightDecorations} = this.props
<add>
<add> let children = null
<add> if (highlightDecorations) {
<add> const decorationCount = highlightDecorations.length
<add> children = new Array(decorationCount)
<add> for (let i = 0; i < decorationCount; i++) {
<add> const highlightProps = Object.assign(
<add> {parentTileTop: top, lineHeight},
<add> highlightDecorations[i]
<add> )
<add> children[i] = $(HighlightComponent, highlightProps)
<add> }
<add> }
<add>
<add> return $.div(
<add> {
<add> style: {
<add> position: 'absolute',
<add> contain: 'strict',
<add> height: height + 'px',
<add> width: width + 'px'
<add> },
<add> }, children
<add> )
<add> }
<add>
<add> renderLines () {
<ide> const {
<ide> height, width, top,
<ide> screenLines, lineDecorations, displayLayer,
<ide> class LinesTileComponent {
<ide>
<ide> return $.div({
<ide> style: {
<del> contain: 'strict',
<ide> position: 'absolute',
<add> contain: 'strict',
<ide> height: height + 'px',
<ide> width: width + 'px',
<del> willChange: 'transform',
<del> transform: `translateY(${top}px)`,
<del> backgroundColor: 'inherit'
<ide> }
<ide> }, children)
<ide> }
<ide> class LinesTileComponent {
<ide> if (oldProps.width !== newProps.width) return true
<ide> if (!arraysEqual(oldProps.screenLines, newProps.screenLines)) return true
<ide> if (!arraysEqual(oldProps.lineDecorations, newProps.lineDecorations)) return true
<add>
<add> if (!oldProps.highlightDecorations && newProps.highlightDecorations) return true
<add> if (oldProps.highlightDecorations && !newProps.highlightDecorations) return true
<add>
<add> if (oldProps.highlightDecorations && newProps.highlightDecorations) {
<add> if (oldProps.highlightDecorations.length !== newProps.highlightDecorations.length) return true
<add>
<add> for (let i = 0, length = oldProps.highlightDecorations.length; i < length; i++) {
<add> const oldHighlight = oldProps.highlightDecorations[i]
<add> const newHighlight = newProps.highlightDecorations[i]
<add> if (oldHighlight.decoration.class !== newHighlight.decoration.class) return true
<add> if (oldHighlight.startPixelLeft !== newHighlight.startPixelLeft) return true
<add> if (oldHighlight.endPixelLeft !== newHighlight.endPixelLeft) return true
<add> if (!oldHighlight.screenRange.isEqual(newHighlight.screenRange)) return true
<add> }
<add> }
<add>
<ide> return false
<ide> }
<ide> }
<ide> class LineComponent {
<ide> }
<ide> }
<ide>
<add>class HighlightComponent {
<add> constructor (props) {
<add> this.props = props
<add> etch.initialize(this)
<add> }
<add>
<add> update (props) {
<add> this.props = props
<add> etch.updateSync(this)
<add> }
<add>
<add> render () {
<add> let {startPixelTop, endPixelTop} = this.props
<add> const {
<add> decoration, screenRange, parentTileTop, lineHeight,
<add> startPixelLeft, endPixelLeft,
<add> } = this.props
<add> startPixelTop -= parentTileTop
<add> endPixelTop -= parentTileTop
<add>
<add> let children
<add> if (screenRange.start.row === screenRange.end.row) {
<add> children = $.div({
<add> className: 'region',
<add> style: {
<add> position: 'absolute',
<add> boxSizing: 'border-box',
<add> top: startPixelTop + 'px',
<add> left: startPixelLeft + 'px',
<add> width: endPixelLeft - startPixelLeft + 'px',
<add> height: lineHeight + 'px'
<add> }
<add> })
<add> } else {
<add> children = []
<add> children.push($.div({
<add> className: 'region',
<add> style: {
<add> position: 'absolute',
<add> boxSizing: 'border-box',
<add> top: startPixelTop + 'px',
<add> left: startPixelLeft + 'px',
<add> right: 0,
<add> height: lineHeight + 'px'
<add> }
<add> }))
<add>
<add> if (screenRange.end.row - screenRange.start.row > 1) {
<add> children.push($.div({
<add> className: 'region',
<add> style: {
<add> position: 'absolute',
<add> boxSizing: 'border-box',
<add> top: startPixelTop + lineHeight + 'px',
<add> left: 0,
<add> right: 0,
<add> height: endPixelTop - startPixelTop - (lineHeight * 2) + 'px'
<add> }
<add> }))
<add> }
<add>
<add> children.push($.div({
<add> className: 'region',
<add> style: {
<add> position: 'absolute',
<add> boxSizing: 'border-box',
<add> top: endPixelTop - lineHeight + 'px',
<add> left: 0,
<add> width: endPixelLeft + 'px',
<add> height: lineHeight + 'px'
<add> }
<add> }))
<add> }
<add>
<add> const className = 'highlight ' + decoration.class
<add> return $.div({className}, children)
<add> }
<add>}
<add>
<ide> const classNamesByScopeName = new Map()
<ide> function classNameForScopeName (scopeName) {
<ide> let classString = classNamesByScopeName.get(scopeName)
<ide> function arraysEqual(a, b) {
<ide> }
<ide> return true
<ide> }
<add>
<add>function constrainRangeToRows (range, startRow, endRow) {
<add> if (range.start.row < startRow || range.end.row >= endRow) {
<add> range = range.copy()
<add> if (range.start.row < startRow) {
<add> range.start.row = startRow
<add> range.start.column = 0
<add> }
<add> if (range.end.row >= endRow) {
<add> range.end.row = endRow
<add> range.end.column = 0
<add> }
<add> }
<add> return range
<add>}
| 1
|
Javascript
|
Javascript
|
move reserved names to top of module
|
d5479d8fa7c52c79912aba45e3c86e2db32738dd
|
<ide><path>lib/optimize/ConcatenatedModule.js
<ide> const createHash = require("../util/createHash");
<ide> /** @typedef {import("../WebpackError")} WebpackError */
<ide> /** @typedef {import("../util/createHash").Hash} Hash */
<ide>
<add>const RESERVED_NAMES = [
<add> // internal name
<add> "__WEBPACK_MODULE_DEFAULT_EXPORT__",
<add>
<add> // keywords
<add> "abstract,arguments,async,await,boolean,break,byte,case,catch,char,class,const,continue",
<add> "debugger,default,delete,do,double,else,enum,eval,export,extends,false,final,finally,float",
<add> "for,function,goto,if,implements,import,in,instanceof,int,interface,let,long,native,new,null",
<add> "package,private,protected,public,return,short,static,super,switch,synchronized,this,throw",
<add> "throws,transient,true,try,typeof,var,void,volatile,while,with,yield",
<add>
<add> // commonjs
<add> "module,__dirname,__filename,exports",
<add>
<add> // js globals
<add> "Array,Date,eval,function,hasOwnProperty,Infinity,isFinite,isNaN,isPrototypeOf,length,Math",
<add> "NaN,name,Number,Object,prototype,String,toString,undefined,valueOf",
<add>
<add> // browser globals
<add> "alert,all,anchor,anchors,area,assign,blur,button,checkbox,clearInterval,clearTimeout",
<add> "clientInformation,close,closed,confirm,constructor,crypto,decodeURI,decodeURIComponent",
<add> "defaultStatus,document,element,elements,embed,embeds,encodeURI,encodeURIComponent,escape",
<add> "event,fileUpload,focus,form,forms,frame,innerHeight,innerWidth,layer,layers,link,location",
<add> "mimeTypes,navigate,navigator,frames,frameRate,hidden,history,image,images,offscreenBuffering",
<add> "open,opener,option,outerHeight,outerWidth,packages,pageXOffset,pageYOffset,parent,parseFloat",
<add> "parseInt,password,pkcs11,plugin,prompt,propertyIsEnum,radio,reset,screenX,screenY,scroll",
<add> "secure,select,self,setInterval,setTimeout,status,submit,taint,text,textarea,top,unescape",
<add> "untaint,window",
<add>
<add> // window events
<add> "onblur,onclick,onerror,onfocus,onkeydown,onkeypress,onkeyup,onmouseover,onload,onmouseup,onmousedown,onsubmit"
<add>]
<add> .join(",")
<add> .split(",");
<add>
<ide> /**
<ide> * @typedef {Object} ConcatenationEntry
<ide> * @property {"concatenated" | "external"} type
<ide> class ConcatenatedModule extends Module {
<ide> }
<ide>
<ide> // List of all used names to avoid conflicts
<del> const allUsedNames = new Set([
<del> "__WEBPACK_MODULE_DEFAULT_EXPORT__", // avoid using this internal name
<del>
<del> "abstract",
<del> "arguments",
<del> "async",
<del> "await",
<del> "boolean",
<del> "break",
<del> "byte",
<del> "case",
<del> "catch",
<del> "char",
<del> "class",
<del> "const",
<del> "continue",
<del> "debugger",
<del> "default",
<del> "delete",
<del> "do",
<del> "double",
<del> "else",
<del> "enum",
<del> "eval",
<del> "export",
<del> "extends",
<del> "false",
<del> "final",
<del> "finally",
<del> "float",
<del> "for",
<del> "function",
<del> "goto",
<del> "if",
<del> "implements",
<del> "import",
<del> "in",
<del> "instanceof",
<del> "int",
<del> "interface",
<del> "let",
<del> "long",
<del> "native",
<del> "new",
<del> "null",
<del> "package",
<del> "private",
<del> "protected",
<del> "public",
<del> "return",
<del> "short",
<del> "static",
<del> "super",
<del> "switch",
<del> "synchronized",
<del> "this",
<del> "throw",
<del> "throws",
<del> "transient",
<del> "true",
<del> "try",
<del> "typeof",
<del> "var",
<del> "void",
<del> "volatile",
<del> "while",
<del> "with",
<del> "yield",
<del>
<del> "module",
<del> "__dirname",
<del> "__filename",
<del> "exports",
<del>
<del> "Array",
<del> "Date",
<del> "eval",
<del> "function",
<del> "hasOwnProperty",
<del> "Infinity",
<del> "isFinite",
<del> "isNaN",
<del> "isPrototypeOf",
<del> "length",
<del> "Math",
<del> "NaN",
<del> "name",
<del> "Number",
<del> "Object",
<del> "prototype",
<del> "String",
<del> "toString",
<del> "undefined",
<del> "valueOf",
<del>
<del> "alert",
<del> "all",
<del> "anchor",
<del> "anchors",
<del> "area",
<del> "assign",
<del> "blur",
<del> "button",
<del> "checkbox",
<del> "clearInterval",
<del> "clearTimeout",
<del> "clientInformation",
<del> "close",
<del> "closed",
<del> "confirm",
<del> "constructor",
<del> "crypto",
<del> "decodeURI",
<del> "decodeURIComponent",
<del> "defaultStatus",
<del> "document",
<del> "element",
<del> "elements",
<del> "embed",
<del> "embeds",
<del> "encodeURI",
<del> "encodeURIComponent",
<del> "escape",
<del> "event",
<del> "fileUpload",
<del> "focus",
<del> "form",
<del> "forms",
<del> "frame",
<del> "innerHeight",
<del> "innerWidth",
<del> "layer",
<del> "layers",
<del> "link",
<del> "location",
<del> "mimeTypes",
<del> "navigate",
<del> "navigator",
<del> "frames",
<del> "frameRate",
<del> "hidden",
<del> "history",
<del> "image",
<del> "images",
<del> "offscreenBuffering",
<del> "open",
<del> "opener",
<del> "option",
<del> "outerHeight",
<del> "outerWidth",
<del> "packages",
<del> "pageXOffset",
<del> "pageYOffset",
<del> "parent",
<del> "parseFloat",
<del> "parseInt",
<del> "password",
<del> "pkcs11",
<del> "plugin",
<del> "prompt",
<del> "propertyIsEnum",
<del> "radio",
<del> "reset",
<del> "screenX",
<del> "screenY",
<del> "scroll",
<del> "secure",
<del> "select",
<del> "self",
<del> "setInterval",
<del> "setTimeout",
<del> "status",
<del> "submit",
<del> "taint",
<del> "text",
<del> "textarea",
<del> "top",
<del> "unescape",
<del> "untaint",
<del> "window",
<del>
<del> "onblur",
<del> "onclick",
<del> "onerror",
<del> "onfocus",
<del> "onkeydown",
<del> "onkeypress",
<del> "onkeyup",
<del> "onmouseover",
<del> "onload",
<del> "onmouseup",
<del> "onmousedown",
<del> "onsubmit"
<del> ]);
<add> const allUsedNames = new Set(RESERVED_NAMES);
<ide>
<ide> // Set of already checked scopes
<ide> const alreadyCheckedScopes = new Set();
| 1
|
Java
|
Java
|
fix javadoc in dependencydescriptor
|
acae174f8f0a9013b10b10a65e63c0caf2a28337
|
<ide><path>spring-beans/src/main/java/org/springframework/beans/factory/config/DependencyDescriptor.java
<ide> public void initParameterNameDiscovery(@Nullable ParameterNameDiscoverer paramet
<ide>
<ide> /**
<ide> * Determine the name of the wrapped parameter/field.
<del> * @return the declared name (never {@code null})
<add> * @return the declared name (may be {@code null})
<ide> */
<ide> @Nullable
<ide> public String getDependencyName() {
| 1
|
Javascript
|
Javascript
|
move hydration code out of normal suspense path
|
b2763d3eaa4a56ce9c973945783aba7cac63478f
|
<ide><path>packages/react-reconciler/src/ReactFiberBeginWork.new.js
<ide> function updateSuspenseComponent(current, workInProgress, renderLanes) {
<ide> // a stack.
<ide> if (current === null) {
<ide> // Initial mount
<add>
<add> // Special path for hydration
<ide> // If we're currently hydrating, try to hydrate this boundary.
<ide> tryToClaimNextHydratableInstance(workInProgress);
<ide> // This could've been a dehydrated suspense component.
<ide> function updateSuspenseComponent(current, workInProgress, renderLanes) {
<ide> } else {
<ide> // This is an update.
<ide>
<del> // If the current fiber has a SuspenseState, that means it's already showing
<del> // a fallback.
<add> // Special path for hydration
<ide> const prevState: null | SuspenseState = current.memoizedState;
<ide> if (prevState !== null) {
<del> // The current tree is already showing a fallback
<del>
<del> // Special path for hydration
<ide> const dehydrated = prevState.dehydrated;
<ide> if (dehydrated !== null) {
<del> if (!didSuspend) {
<del> return updateDehydratedSuspenseComponent(
<del> current,
<del> workInProgress,
<del> dehydrated,
<del> prevState,
<del> renderLanes,
<del> );
<del> } else if (workInProgress.flags & ForceClientRender) {
<del> // Something errored during hydration. Try again without hydrating.
<del> workInProgress.flags &= ~ForceClientRender;
<del> return retrySuspenseComponentWithoutHydrating(
<del> current,
<del> workInProgress,
<del> renderLanes,
<del> new Error(
<del> 'There was an error while hydrating this Suspense boundary. ' +
<del> 'Switched to client rendering.',
<del> ),
<del> );
<del> } else if (
<del> (workInProgress.memoizedState: null | SuspenseState) !== null
<del> ) {
<del> // Something suspended and we should still be in dehydrated mode.
<del> // Leave the existing child in place.
<del> workInProgress.child = current.child;
<del> // The dehydrated completion pass expects this flag to be there
<del> // but the normal suspense pass doesn't.
<del> workInProgress.flags |= DidCapture;
<del> return null;
<del> } else {
<del> // Suspended but we should no longer be in dehydrated mode.
<del> // Therefore we now have to render the fallback.
<del> const nextPrimaryChildren = nextProps.children;
<del> const nextFallbackChildren = nextProps.fallback;
<del> const fallbackChildFragment = mountSuspenseFallbackAfterRetryWithoutHydrating(
<del> current,
<del> workInProgress,
<del> nextPrimaryChildren,
<del> nextFallbackChildren,
<del> renderLanes,
<del> );
<del> const primaryChildFragment: Fiber = (workInProgress.child: any);
<del> primaryChildFragment.memoizedState = mountSuspenseOffscreenState(
<del> renderLanes,
<del> );
<del> workInProgress.memoizedState = SUSPENDED_MARKER;
<del> return fallbackChildFragment;
<del> }
<del> }
<del>
<del> if (showFallback) {
<del> const nextFallbackChildren = nextProps.fallback;
<del> const nextPrimaryChildren = nextProps.children;
<del> const fallbackChildFragment = updateSuspenseFallbackChildren(
<add> return updateDehydratedSuspenseComponent(
<ide> current,
<ide> workInProgress,
<del> nextPrimaryChildren,
<del> nextFallbackChildren,
<del> renderLanes,
<del> );
<del> const primaryChildFragment: Fiber = (workInProgress.child: any);
<del> const prevOffscreenState: OffscreenState | null = (current.child: any)
<del> .memoizedState;
<del> primaryChildFragment.memoizedState =
<del> prevOffscreenState === null
<del> ? mountSuspenseOffscreenState(renderLanes)
<del> : updateSuspenseOffscreenState(prevOffscreenState, renderLanes);
<del> if (enableTransitionTracing) {
<del> const currentTransitions = getSuspendedTransitions();
<del> if (currentTransitions !== null) {
<del> const primaryChildUpdateQueue: OffscreenQueue = {
<del> transitions: currentTransitions,
<del> };
<del> primaryChildFragment.updateQueue = primaryChildUpdateQueue;
<del> }
<del> }
<del> primaryChildFragment.childLanes = getRemainingWorkInPrimaryTree(
<del> current,
<del> renderLanes,
<del> );
<del> workInProgress.memoizedState = SUSPENDED_MARKER;
<del> return fallbackChildFragment;
<del> } else {
<del> const nextPrimaryChildren = nextProps.children;
<del> const primaryChildFragment = updateSuspensePrimaryChildren(
<del> current,
<del> workInProgress,
<del> nextPrimaryChildren,
<add> didSuspend,
<add> nextProps,
<add> dehydrated,
<add> prevState,
<ide> renderLanes,
<ide> );
<del> workInProgress.memoizedState = null;
<del> return primaryChildFragment;
<ide> }
<del> } else {
<del> // The current tree is not already showing a fallback.
<del> if (showFallback) {
<del> // Timed out.
<del> const nextFallbackChildren = nextProps.fallback;
<del> const nextPrimaryChildren = nextProps.children;
<del> const fallbackChildFragment = updateSuspenseFallbackChildren(
<del> current,
<del> workInProgress,
<del> nextPrimaryChildren,
<del> nextFallbackChildren,
<del> renderLanes,
<del> );
<del> const primaryChildFragment: Fiber = (workInProgress.child: any);
<del> const prevOffscreenState: OffscreenState | null = (current.child: any)
<del> .memoizedState;
<del> primaryChildFragment.memoizedState =
<del> prevOffscreenState === null
<del> ? mountSuspenseOffscreenState(renderLanes)
<del> : updateSuspenseOffscreenState(prevOffscreenState, renderLanes);
<del> primaryChildFragment.childLanes = getRemainingWorkInPrimaryTree(
<del> current,
<del> renderLanes,
<del> );
<add> }
<ide>
<del> if (enableTransitionTracing) {
<del> const currentTransitions = getSuspendedTransitions();
<del> if (currentTransitions !== null) {
<del> const primaryChildUpdateQueue: OffscreenQueue = {
<del> transitions: currentTransitions,
<del> };
<del> primaryChildFragment.updateQueue = primaryChildUpdateQueue;
<del> }
<add> if (showFallback) {
<add> const nextFallbackChildren = nextProps.fallback;
<add> const nextPrimaryChildren = nextProps.children;
<add> const fallbackChildFragment = updateSuspenseFallbackChildren(
<add> current,
<add> workInProgress,
<add> nextPrimaryChildren,
<add> nextFallbackChildren,
<add> renderLanes,
<add> );
<add> const primaryChildFragment: Fiber = (workInProgress.child: any);
<add> const prevOffscreenState: OffscreenState | null = (current.child: any)
<add> .memoizedState;
<add> primaryChildFragment.memoizedState =
<add> prevOffscreenState === null
<add> ? mountSuspenseOffscreenState(renderLanes)
<add> : updateSuspenseOffscreenState(prevOffscreenState, renderLanes);
<add> if (enableTransitionTracing) {
<add> const currentTransitions = getSuspendedTransitions();
<add> if (currentTransitions !== null) {
<add> const primaryChildUpdateQueue: OffscreenQueue = {
<add> transitions: currentTransitions,
<add> };
<add> primaryChildFragment.updateQueue = primaryChildUpdateQueue;
<ide> }
<del>
<del> // Skip the primary children, and continue working on the
<del> // fallback children.
<del> workInProgress.memoizedState = SUSPENDED_MARKER;
<del> return fallbackChildFragment;
<del> } else {
<del> // Still haven't timed out. Continue rendering the children, like we
<del> // normally do.
<del> const nextPrimaryChildren = nextProps.children;
<del> const primaryChildFragment = updateSuspensePrimaryChildren(
<del> current,
<del> workInProgress,
<del> nextPrimaryChildren,
<del> renderLanes,
<del> );
<del> workInProgress.memoizedState = null;
<del> return primaryChildFragment;
<ide> }
<add> primaryChildFragment.childLanes = getRemainingWorkInPrimaryTree(
<add> current,
<add> renderLanes,
<add> );
<add> workInProgress.memoizedState = SUSPENDED_MARKER;
<add> return fallbackChildFragment;
<add> } else {
<add> const nextPrimaryChildren = nextProps.children;
<add> const primaryChildFragment = updateSuspensePrimaryChildren(
<add> current,
<add> workInProgress,
<add> nextPrimaryChildren,
<add> renderLanes,
<add> );
<add> workInProgress.memoizedState = null;
<add> return primaryChildFragment;
<ide> }
<ide> }
<ide> }
<ide> function mountDehydratedSuspenseComponent(
<ide> function updateDehydratedSuspenseComponent(
<ide> current: Fiber,
<ide> workInProgress: Fiber,
<add> didSuspend: boolean,
<add> nextProps: any,
<ide> suspenseInstance: SuspenseInstance,
<ide> suspenseState: SuspenseState,
<ide> renderLanes: Lanes,
<ide> ): null | Fiber {
<del> // We should never be hydrating at this point because it is the first pass,
<del> // but after we've already committed once.
<del> warnIfHydrating();
<add> if (!didSuspend) {
<add> // This is the first render pass. Attempt to hydrate.
<ide>
<del> if ((workInProgress.mode & ConcurrentMode) === NoMode) {
<del> return retrySuspenseComponentWithoutHydrating(
<del> current,
<del> workInProgress,
<del> renderLanes,
<del> // TODO: When we delete legacy mode, we should make this error argument
<del> // required — every concurrent mode path that causes hydration to
<del> // de-opt to client rendering should have an error message.
<del> null,
<del> );
<del> }
<add> // We should never be hydrating at this point because it is the first pass,
<add> // but after we've already committed once.
<add> warnIfHydrating();
<ide>
<del> if (isSuspenseInstanceFallback(suspenseInstance)) {
<del> // This boundary is in a permanent fallback state. In this case, we'll never
<del> // get an update and we'll never be able to hydrate the final content. Let's just try the
<del> // client side render instead.
<del> return retrySuspenseComponentWithoutHydrating(
<del> current,
<del> workInProgress,
<del> renderLanes,
<del> // TODO: The server should serialize the error message so we can log it
<del> // here on the client. Or, in production, a hash/id that corresponds to
<del> // the error.
<del> new Error(
<del> 'The server could not finish this Suspense boundary, likely ' +
<del> 'due to an error during server rendering. Switched to ' +
<del> 'client rendering.',
<del> ),
<del> );
<del> }
<add> if ((workInProgress.mode & ConcurrentMode) === NoMode) {
<add> return retrySuspenseComponentWithoutHydrating(
<add> current,
<add> workInProgress,
<add> renderLanes,
<add> // TODO: When we delete legacy mode, we should make this error argument
<add> // required — every concurrent mode path that causes hydration to
<add> // de-opt to client rendering should have an error message.
<add> null,
<add> );
<add> }
<ide>
<del> if (
<del> enableLazyContextPropagation &&
<del> // TODO: Factoring is a little weird, since we check this right below, too.
<del> // But don't want to re-arrange the if-else chain until/unless this
<del> // feature lands.
<del> !didReceiveUpdate
<del> ) {
<del> // We need to check if any children have context before we decide to bail
<del> // out, so propagate the changes now.
<del> lazilyPropagateParentContextChanges(current, workInProgress, renderLanes);
<del> }
<del>
<del> // We use lanes to indicate that a child might depend on context, so if
<del> // any context has changed, we need to treat is as if the input might have changed.
<del> const hasContextChanged = includesSomeLane(renderLanes, current.childLanes);
<del> if (didReceiveUpdate || hasContextChanged) {
<del> // This boundary has changed since the first render. This means that we are now unable to
<del> // hydrate it. We might still be able to hydrate it using a higher priority lane.
<del> const root = getWorkInProgressRoot();
<del> if (root !== null) {
<del> const attemptHydrationAtLane = getBumpedLaneForHydration(
<del> root,
<add> if (isSuspenseInstanceFallback(suspenseInstance)) {
<add> // This boundary is in a permanent fallback state. In this case, we'll never
<add> // get an update and we'll never be able to hydrate the final content. Let's just try the
<add> // client side render instead.
<add> return retrySuspenseComponentWithoutHydrating(
<add> current,
<add> workInProgress,
<ide> renderLanes,
<add> // TODO: The server should serialize the error message so we can log it
<add> // here on the client. Or, in production, a hash/id that corresponds to
<add> // the error.
<add> new Error(
<add> 'The server could not finish this Suspense boundary, likely ' +
<add> 'due to an error during server rendering. Switched to ' +
<add> 'client rendering.',
<add> ),
<ide> );
<del> if (
<del> attemptHydrationAtLane !== NoLane &&
<del> attemptHydrationAtLane !== suspenseState.retryLane
<del> ) {
<del> // Intentionally mutating since this render will get interrupted. This
<del> // is one of the very rare times where we mutate the current tree
<del> // during the render phase.
<del> suspenseState.retryLane = attemptHydrationAtLane;
<del> // TODO: Ideally this would inherit the event time of the current render
<del> const eventTime = NoTimestamp;
<del> scheduleUpdateOnFiber(current, attemptHydrationAtLane, eventTime);
<del> } else {
<del> // We have already tried to ping at a higher priority than we're rendering with
<del> // so if we got here, we must have failed to hydrate at those levels. We must
<del> // now give up. Instead, we're going to delete the whole subtree and instead inject
<del> // a new real Suspense boundary to take its place, which may render content
<del> // or fallback. This might suspend for a while and if it does we might still have
<del> // an opportunity to hydrate before this pass commits.
<del> }
<del> }
<del>
<del> // If we have scheduled higher pri work above, this will probably just abort the render
<del> // since we now have higher priority work, but in case it doesn't, we need to prepare to
<del> // render something, if we time out. Even if that requires us to delete everything and
<del> // skip hydration.
<del> // Delay having to do this as long as the suspense timeout allows us.
<del> renderDidSuspendDelayIfPossible();
<del> return retrySuspenseComponentWithoutHydrating(
<del> current,
<del> workInProgress,
<del> renderLanes,
<del> new Error(
<del> 'This Suspense boundary received an update before it finished ' +
<del> 'hydrating. This caused the boundary to switch to client rendering. ' +
<del> 'The usual way to fix this is to wrap the original update ' +
<del> 'in startTransition.',
<del> ),
<del> );
<del> } else if (isSuspenseInstancePending(suspenseInstance)) {
<del> // This component is still pending more data from the server, so we can't hydrate its
<del> // content. We treat it as if this component suspended itself. It might seem as if
<del> // we could just try to render it client-side instead. However, this will perform a
<del> // lot of unnecessary work and is unlikely to complete since it often will suspend
<del> // on missing data anyway. Additionally, the server might be able to render more
<del> // than we can on the client yet. In that case we'd end up with more fallback states
<del> // on the client than if we just leave it alone. If the server times out or errors
<del> // these should update this boundary to the permanent Fallback state instead.
<del> // Mark it as having captured (i.e. suspended).
<del> workInProgress.flags |= DidCapture;
<del> // Leave the child in place. I.e. the dehydrated fragment.
<del> workInProgress.child = current.child;
<del> // Register a callback to retry this boundary once the server has sent the result.
<del> const retry = retryDehydratedSuspenseBoundary.bind(null, current);
<del> registerSuspenseInstanceRetry(suspenseInstance, retry);
<del> return null;
<add> }
<add>
<add> if (
<add> enableLazyContextPropagation &&
<add> // TODO: Factoring is a little weird, since we check this right below, too.
<add> // But don't want to re-arrange the if-else chain until/unless this
<add> // feature lands.
<add> !didReceiveUpdate
<add> ) {
<add> // We need to check if any children have context before we decide to bail
<add> // out, so propagate the changes now.
<add> lazilyPropagateParentContextChanges(current, workInProgress, renderLanes);
<add> }
<add>
<add> // We use lanes to indicate that a child might depend on context, so if
<add> // any context has changed, we need to treat is as if the input might have changed.
<add> const hasContextChanged = includesSomeLane(renderLanes, current.childLanes);
<add> if (didReceiveUpdate || hasContextChanged) {
<add> // This boundary has changed since the first render. This means that we are now unable to
<add> // hydrate it. We might still be able to hydrate it using a higher priority lane.
<add> const root = getWorkInProgressRoot();
<add> if (root !== null) {
<add> const attemptHydrationAtLane = getBumpedLaneForHydration(
<add> root,
<add> renderLanes,
<add> );
<add> if (
<add> attemptHydrationAtLane !== NoLane &&
<add> attemptHydrationAtLane !== suspenseState.retryLane
<add> ) {
<add> // Intentionally mutating since this render will get interrupted. This
<add> // is one of the very rare times where we mutate the current tree
<add> // during the render phase.
<add> suspenseState.retryLane = attemptHydrationAtLane;
<add> // TODO: Ideally this would inherit the event time of the current render
<add> const eventTime = NoTimestamp;
<add> scheduleUpdateOnFiber(current, attemptHydrationAtLane, eventTime);
<add> } else {
<add> // We have already tried to ping at a higher priority than we're rendering with
<add> // so if we got here, we must have failed to hydrate at those levels. We must
<add> // now give up. Instead, we're going to delete the whole subtree and instead inject
<add> // a new real Suspense boundary to take its place, which may render content
<add> // or fallback. This might suspend for a while and if it does we might still have
<add> // an opportunity to hydrate before this pass commits.
<add> }
<add> }
<add>
<add> // If we have scheduled higher pri work above, this will probably just abort the render
<add> // since we now have higher priority work, but in case it doesn't, we need to prepare to
<add> // render something, if we time out. Even if that requires us to delete everything and
<add> // skip hydration.
<add> // Delay having to do this as long as the suspense timeout allows us.
<add> renderDidSuspendDelayIfPossible();
<add> return retrySuspenseComponentWithoutHydrating(
<add> current,
<add> workInProgress,
<add> renderLanes,
<add> new Error(
<add> 'This Suspense boundary received an update before it finished ' +
<add> 'hydrating. This caused the boundary to switch to client rendering. ' +
<add> 'The usual way to fix this is to wrap the original update ' +
<add> 'in startTransition.',
<add> ),
<add> );
<add> } else if (isSuspenseInstancePending(suspenseInstance)) {
<add> // This component is still pending more data from the server, so we can't hydrate its
<add> // content. We treat it as if this component suspended itself. It might seem as if
<add> // we could just try to render it client-side instead. However, this will perform a
<add> // lot of unnecessary work and is unlikely to complete since it often will suspend
<add> // on missing data anyway. Additionally, the server might be able to render more
<add> // than we can on the client yet. In that case we'd end up with more fallback states
<add> // on the client than if we just leave it alone. If the server times out or errors
<add> // these should update this boundary to the permanent Fallback state instead.
<add> // Mark it as having captured (i.e. suspended).
<add> workInProgress.flags |= DidCapture;
<add> // Leave the child in place. I.e. the dehydrated fragment.
<add> workInProgress.child = current.child;
<add> // Register a callback to retry this boundary once the server has sent the result.
<add> const retry = retryDehydratedSuspenseBoundary.bind(null, current);
<add> registerSuspenseInstanceRetry(suspenseInstance, retry);
<add> return null;
<add> } else {
<add> // This is the first attempt.
<add> reenterHydrationStateFromDehydratedSuspenseInstance(
<add> workInProgress,
<add> suspenseInstance,
<add> suspenseState.treeContext,
<add> );
<add> const primaryChildren = nextProps.children;
<add> const primaryChildFragment = mountSuspensePrimaryChildren(
<add> workInProgress,
<add> primaryChildren,
<add> renderLanes,
<add> );
<add> // Mark the children as hydrating. This is a fast path to know whether this
<add> // tree is part of a hydrating tree. This is used to determine if a child
<add> // node has fully mounted yet, and for scheduling event replaying.
<add> // Conceptually this is similar to Placement in that a new subtree is
<add> // inserted into the React tree here. It just happens to not need DOM
<add> // mutations because it already exists.
<add> primaryChildFragment.flags |= Hydrating;
<add> return primaryChildFragment;
<add> }
<ide> } else {
<del> // This is the first attempt.
<del> reenterHydrationStateFromDehydratedSuspenseInstance(
<del> workInProgress,
<del> suspenseInstance,
<del> suspenseState.treeContext,
<del> );
<del> const nextProps = workInProgress.pendingProps;
<del> const primaryChildren = nextProps.children;
<del> const primaryChildFragment = mountSuspensePrimaryChildren(
<del> workInProgress,
<del> primaryChildren,
<del> renderLanes,
<del> );
<del> // Mark the children as hydrating. This is a fast path to know whether this
<del> // tree is part of a hydrating tree. This is used to determine if a child
<del> // node has fully mounted yet, and for scheduling event replaying.
<del> // Conceptually this is similar to Placement in that a new subtree is
<del> // inserted into the React tree here. It just happens to not need DOM
<del> // mutations because it already exists.
<del> primaryChildFragment.flags |= Hydrating;
<del> return primaryChildFragment;
<add> // This is the second render pass. We already attempted to hydrated, but
<add> // something either suspended or errored.
<add>
<add> if (workInProgress.flags & ForceClientRender) {
<add> // Something errored during hydration. Try again without hydrating.
<add> workInProgress.flags &= ~ForceClientRender;
<add> return retrySuspenseComponentWithoutHydrating(
<add> current,
<add> workInProgress,
<add> renderLanes,
<add> new Error(
<add> 'There was an error while hydrating this Suspense boundary. ' +
<add> 'Switched to client rendering.',
<add> ),
<add> );
<add> } else if ((workInProgress.memoizedState: null | SuspenseState) !== null) {
<add> // Something suspended and we should still be in dehydrated mode.
<add> // Leave the existing child in place.
<add> workInProgress.child = current.child;
<add> // The dehydrated completion pass expects this flag to be there
<add> // but the normal suspense pass doesn't.
<add> workInProgress.flags |= DidCapture;
<add> return null;
<add> } else {
<add> // Suspended but we should no longer be in dehydrated mode.
<add> // Therefore we now have to render the fallback.
<add> const nextPrimaryChildren = nextProps.children;
<add> const nextFallbackChildren = nextProps.fallback;
<add> const fallbackChildFragment = mountSuspenseFallbackAfterRetryWithoutHydrating(
<add> current,
<add> workInProgress,
<add> nextPrimaryChildren,
<add> nextFallbackChildren,
<add> renderLanes,
<add> );
<add> const primaryChildFragment: Fiber = (workInProgress.child: any);
<add> primaryChildFragment.memoizedState = mountSuspenseOffscreenState(
<add> renderLanes,
<add> );
<add> workInProgress.memoizedState = SUSPENDED_MARKER;
<add> return fallbackChildFragment;
<add> }
<ide> }
<ide> }
<ide>
<ide><path>packages/react-reconciler/src/ReactFiberBeginWork.old.js
<ide> function updateSuspenseComponent(current, workInProgress, renderLanes) {
<ide> // a stack.
<ide> if (current === null) {
<ide> // Initial mount
<add>
<add> // Special path for hydration
<ide> // If we're currently hydrating, try to hydrate this boundary.
<ide> tryToClaimNextHydratableInstance(workInProgress);
<ide> // This could've been a dehydrated suspense component.
<ide> function updateSuspenseComponent(current, workInProgress, renderLanes) {
<ide> } else {
<ide> // This is an update.
<ide>
<del> // If the current fiber has a SuspenseState, that means it's already showing
<del> // a fallback.
<add> // Special path for hydration
<ide> const prevState: null | SuspenseState = current.memoizedState;
<ide> if (prevState !== null) {
<del> // The current tree is already showing a fallback
<del>
<del> // Special path for hydration
<ide> const dehydrated = prevState.dehydrated;
<ide> if (dehydrated !== null) {
<del> if (!didSuspend) {
<del> return updateDehydratedSuspenseComponent(
<del> current,
<del> workInProgress,
<del> dehydrated,
<del> prevState,
<del> renderLanes,
<del> );
<del> } else if (workInProgress.flags & ForceClientRender) {
<del> // Something errored during hydration. Try again without hydrating.
<del> workInProgress.flags &= ~ForceClientRender;
<del> return retrySuspenseComponentWithoutHydrating(
<del> current,
<del> workInProgress,
<del> renderLanes,
<del> new Error(
<del> 'There was an error while hydrating this Suspense boundary. ' +
<del> 'Switched to client rendering.',
<del> ),
<del> );
<del> } else if (
<del> (workInProgress.memoizedState: null | SuspenseState) !== null
<del> ) {
<del> // Something suspended and we should still be in dehydrated mode.
<del> // Leave the existing child in place.
<del> workInProgress.child = current.child;
<del> // The dehydrated completion pass expects this flag to be there
<del> // but the normal suspense pass doesn't.
<del> workInProgress.flags |= DidCapture;
<del> return null;
<del> } else {
<del> // Suspended but we should no longer be in dehydrated mode.
<del> // Therefore we now have to render the fallback.
<del> const nextPrimaryChildren = nextProps.children;
<del> const nextFallbackChildren = nextProps.fallback;
<del> const fallbackChildFragment = mountSuspenseFallbackAfterRetryWithoutHydrating(
<del> current,
<del> workInProgress,
<del> nextPrimaryChildren,
<del> nextFallbackChildren,
<del> renderLanes,
<del> );
<del> const primaryChildFragment: Fiber = (workInProgress.child: any);
<del> primaryChildFragment.memoizedState = mountSuspenseOffscreenState(
<del> renderLanes,
<del> );
<del> workInProgress.memoizedState = SUSPENDED_MARKER;
<del> return fallbackChildFragment;
<del> }
<del> }
<del>
<del> if (showFallback) {
<del> const nextFallbackChildren = nextProps.fallback;
<del> const nextPrimaryChildren = nextProps.children;
<del> const fallbackChildFragment = updateSuspenseFallbackChildren(
<add> return updateDehydratedSuspenseComponent(
<ide> current,
<ide> workInProgress,
<del> nextPrimaryChildren,
<del> nextFallbackChildren,
<del> renderLanes,
<del> );
<del> const primaryChildFragment: Fiber = (workInProgress.child: any);
<del> const prevOffscreenState: OffscreenState | null = (current.child: any)
<del> .memoizedState;
<del> primaryChildFragment.memoizedState =
<del> prevOffscreenState === null
<del> ? mountSuspenseOffscreenState(renderLanes)
<del> : updateSuspenseOffscreenState(prevOffscreenState, renderLanes);
<del> if (enableTransitionTracing) {
<del> const currentTransitions = getSuspendedTransitions();
<del> if (currentTransitions !== null) {
<del> const primaryChildUpdateQueue: OffscreenQueue = {
<del> transitions: currentTransitions,
<del> };
<del> primaryChildFragment.updateQueue = primaryChildUpdateQueue;
<del> }
<del> }
<del> primaryChildFragment.childLanes = getRemainingWorkInPrimaryTree(
<del> current,
<del> renderLanes,
<del> );
<del> workInProgress.memoizedState = SUSPENDED_MARKER;
<del> return fallbackChildFragment;
<del> } else {
<del> const nextPrimaryChildren = nextProps.children;
<del> const primaryChildFragment = updateSuspensePrimaryChildren(
<del> current,
<del> workInProgress,
<del> nextPrimaryChildren,
<add> didSuspend,
<add> nextProps,
<add> dehydrated,
<add> prevState,
<ide> renderLanes,
<ide> );
<del> workInProgress.memoizedState = null;
<del> return primaryChildFragment;
<ide> }
<del> } else {
<del> // The current tree is not already showing a fallback.
<del> if (showFallback) {
<del> // Timed out.
<del> const nextFallbackChildren = nextProps.fallback;
<del> const nextPrimaryChildren = nextProps.children;
<del> const fallbackChildFragment = updateSuspenseFallbackChildren(
<del> current,
<del> workInProgress,
<del> nextPrimaryChildren,
<del> nextFallbackChildren,
<del> renderLanes,
<del> );
<del> const primaryChildFragment: Fiber = (workInProgress.child: any);
<del> const prevOffscreenState: OffscreenState | null = (current.child: any)
<del> .memoizedState;
<del> primaryChildFragment.memoizedState =
<del> prevOffscreenState === null
<del> ? mountSuspenseOffscreenState(renderLanes)
<del> : updateSuspenseOffscreenState(prevOffscreenState, renderLanes);
<del> primaryChildFragment.childLanes = getRemainingWorkInPrimaryTree(
<del> current,
<del> renderLanes,
<del> );
<add> }
<ide>
<del> if (enableTransitionTracing) {
<del> const currentTransitions = getSuspendedTransitions();
<del> if (currentTransitions !== null) {
<del> const primaryChildUpdateQueue: OffscreenQueue = {
<del> transitions: currentTransitions,
<del> };
<del> primaryChildFragment.updateQueue = primaryChildUpdateQueue;
<del> }
<add> if (showFallback) {
<add> const nextFallbackChildren = nextProps.fallback;
<add> const nextPrimaryChildren = nextProps.children;
<add> const fallbackChildFragment = updateSuspenseFallbackChildren(
<add> current,
<add> workInProgress,
<add> nextPrimaryChildren,
<add> nextFallbackChildren,
<add> renderLanes,
<add> );
<add> const primaryChildFragment: Fiber = (workInProgress.child: any);
<add> const prevOffscreenState: OffscreenState | null = (current.child: any)
<add> .memoizedState;
<add> primaryChildFragment.memoizedState =
<add> prevOffscreenState === null
<add> ? mountSuspenseOffscreenState(renderLanes)
<add> : updateSuspenseOffscreenState(prevOffscreenState, renderLanes);
<add> if (enableTransitionTracing) {
<add> const currentTransitions = getSuspendedTransitions();
<add> if (currentTransitions !== null) {
<add> const primaryChildUpdateQueue: OffscreenQueue = {
<add> transitions: currentTransitions,
<add> };
<add> primaryChildFragment.updateQueue = primaryChildUpdateQueue;
<ide> }
<del>
<del> // Skip the primary children, and continue working on the
<del> // fallback children.
<del> workInProgress.memoizedState = SUSPENDED_MARKER;
<del> return fallbackChildFragment;
<del> } else {
<del> // Still haven't timed out. Continue rendering the children, like we
<del> // normally do.
<del> const nextPrimaryChildren = nextProps.children;
<del> const primaryChildFragment = updateSuspensePrimaryChildren(
<del> current,
<del> workInProgress,
<del> nextPrimaryChildren,
<del> renderLanes,
<del> );
<del> workInProgress.memoizedState = null;
<del> return primaryChildFragment;
<ide> }
<add> primaryChildFragment.childLanes = getRemainingWorkInPrimaryTree(
<add> current,
<add> renderLanes,
<add> );
<add> workInProgress.memoizedState = SUSPENDED_MARKER;
<add> return fallbackChildFragment;
<add> } else {
<add> const nextPrimaryChildren = nextProps.children;
<add> const primaryChildFragment = updateSuspensePrimaryChildren(
<add> current,
<add> workInProgress,
<add> nextPrimaryChildren,
<add> renderLanes,
<add> );
<add> workInProgress.memoizedState = null;
<add> return primaryChildFragment;
<ide> }
<ide> }
<ide> }
<ide> function mountDehydratedSuspenseComponent(
<ide> function updateDehydratedSuspenseComponent(
<ide> current: Fiber,
<ide> workInProgress: Fiber,
<add> didSuspend: boolean,
<add> nextProps: any,
<ide> suspenseInstance: SuspenseInstance,
<ide> suspenseState: SuspenseState,
<ide> renderLanes: Lanes,
<ide> ): null | Fiber {
<del> // We should never be hydrating at this point because it is the first pass,
<del> // but after we've already committed once.
<del> warnIfHydrating();
<add> if (!didSuspend) {
<add> // This is the first render pass. Attempt to hydrate.
<ide>
<del> if ((workInProgress.mode & ConcurrentMode) === NoMode) {
<del> return retrySuspenseComponentWithoutHydrating(
<del> current,
<del> workInProgress,
<del> renderLanes,
<del> // TODO: When we delete legacy mode, we should make this error argument
<del> // required — every concurrent mode path that causes hydration to
<del> // de-opt to client rendering should have an error message.
<del> null,
<del> );
<del> }
<add> // We should never be hydrating at this point because it is the first pass,
<add> // but after we've already committed once.
<add> warnIfHydrating();
<ide>
<del> if (isSuspenseInstanceFallback(suspenseInstance)) {
<del> // This boundary is in a permanent fallback state. In this case, we'll never
<del> // get an update and we'll never be able to hydrate the final content. Let's just try the
<del> // client side render instead.
<del> return retrySuspenseComponentWithoutHydrating(
<del> current,
<del> workInProgress,
<del> renderLanes,
<del> // TODO: The server should serialize the error message so we can log it
<del> // here on the client. Or, in production, a hash/id that corresponds to
<del> // the error.
<del> new Error(
<del> 'The server could not finish this Suspense boundary, likely ' +
<del> 'due to an error during server rendering. Switched to ' +
<del> 'client rendering.',
<del> ),
<del> );
<del> }
<add> if ((workInProgress.mode & ConcurrentMode) === NoMode) {
<add> return retrySuspenseComponentWithoutHydrating(
<add> current,
<add> workInProgress,
<add> renderLanes,
<add> // TODO: When we delete legacy mode, we should make this error argument
<add> // required — every concurrent mode path that causes hydration to
<add> // de-opt to client rendering should have an error message.
<add> null,
<add> );
<add> }
<ide>
<del> if (
<del> enableLazyContextPropagation &&
<del> // TODO: Factoring is a little weird, since we check this right below, too.
<del> // But don't want to re-arrange the if-else chain until/unless this
<del> // feature lands.
<del> !didReceiveUpdate
<del> ) {
<del> // We need to check if any children have context before we decide to bail
<del> // out, so propagate the changes now.
<del> lazilyPropagateParentContextChanges(current, workInProgress, renderLanes);
<del> }
<del>
<del> // We use lanes to indicate that a child might depend on context, so if
<del> // any context has changed, we need to treat is as if the input might have changed.
<del> const hasContextChanged = includesSomeLane(renderLanes, current.childLanes);
<del> if (didReceiveUpdate || hasContextChanged) {
<del> // This boundary has changed since the first render. This means that we are now unable to
<del> // hydrate it. We might still be able to hydrate it using a higher priority lane.
<del> const root = getWorkInProgressRoot();
<del> if (root !== null) {
<del> const attemptHydrationAtLane = getBumpedLaneForHydration(
<del> root,
<add> if (isSuspenseInstanceFallback(suspenseInstance)) {
<add> // This boundary is in a permanent fallback state. In this case, we'll never
<add> // get an update and we'll never be able to hydrate the final content. Let's just try the
<add> // client side render instead.
<add> return retrySuspenseComponentWithoutHydrating(
<add> current,
<add> workInProgress,
<ide> renderLanes,
<add> // TODO: The server should serialize the error message so we can log it
<add> // here on the client. Or, in production, a hash/id that corresponds to
<add> // the error.
<add> new Error(
<add> 'The server could not finish this Suspense boundary, likely ' +
<add> 'due to an error during server rendering. Switched to ' +
<add> 'client rendering.',
<add> ),
<ide> );
<del> if (
<del> attemptHydrationAtLane !== NoLane &&
<del> attemptHydrationAtLane !== suspenseState.retryLane
<del> ) {
<del> // Intentionally mutating since this render will get interrupted. This
<del> // is one of the very rare times where we mutate the current tree
<del> // during the render phase.
<del> suspenseState.retryLane = attemptHydrationAtLane;
<del> // TODO: Ideally this would inherit the event time of the current render
<del> const eventTime = NoTimestamp;
<del> scheduleUpdateOnFiber(current, attemptHydrationAtLane, eventTime);
<del> } else {
<del> // We have already tried to ping at a higher priority than we're rendering with
<del> // so if we got here, we must have failed to hydrate at those levels. We must
<del> // now give up. Instead, we're going to delete the whole subtree and instead inject
<del> // a new real Suspense boundary to take its place, which may render content
<del> // or fallback. This might suspend for a while and if it does we might still have
<del> // an opportunity to hydrate before this pass commits.
<del> }
<del> }
<del>
<del> // If we have scheduled higher pri work above, this will probably just abort the render
<del> // since we now have higher priority work, but in case it doesn't, we need to prepare to
<del> // render something, if we time out. Even if that requires us to delete everything and
<del> // skip hydration.
<del> // Delay having to do this as long as the suspense timeout allows us.
<del> renderDidSuspendDelayIfPossible();
<del> return retrySuspenseComponentWithoutHydrating(
<del> current,
<del> workInProgress,
<del> renderLanes,
<del> new Error(
<del> 'This Suspense boundary received an update before it finished ' +
<del> 'hydrating. This caused the boundary to switch to client rendering. ' +
<del> 'The usual way to fix this is to wrap the original update ' +
<del> 'in startTransition.',
<del> ),
<del> );
<del> } else if (isSuspenseInstancePending(suspenseInstance)) {
<del> // This component is still pending more data from the server, so we can't hydrate its
<del> // content. We treat it as if this component suspended itself. It might seem as if
<del> // we could just try to render it client-side instead. However, this will perform a
<del> // lot of unnecessary work and is unlikely to complete since it often will suspend
<del> // on missing data anyway. Additionally, the server might be able to render more
<del> // than we can on the client yet. In that case we'd end up with more fallback states
<del> // on the client than if we just leave it alone. If the server times out or errors
<del> // these should update this boundary to the permanent Fallback state instead.
<del> // Mark it as having captured (i.e. suspended).
<del> workInProgress.flags |= DidCapture;
<del> // Leave the child in place. I.e. the dehydrated fragment.
<del> workInProgress.child = current.child;
<del> // Register a callback to retry this boundary once the server has sent the result.
<del> const retry = retryDehydratedSuspenseBoundary.bind(null, current);
<del> registerSuspenseInstanceRetry(suspenseInstance, retry);
<del> return null;
<add> }
<add>
<add> if (
<add> enableLazyContextPropagation &&
<add> // TODO: Factoring is a little weird, since we check this right below, too.
<add> // But don't want to re-arrange the if-else chain until/unless this
<add> // feature lands.
<add> !didReceiveUpdate
<add> ) {
<add> // We need to check if any children have context before we decide to bail
<add> // out, so propagate the changes now.
<add> lazilyPropagateParentContextChanges(current, workInProgress, renderLanes);
<add> }
<add>
<add> // We use lanes to indicate that a child might depend on context, so if
<add> // any context has changed, we need to treat is as if the input might have changed.
<add> const hasContextChanged = includesSomeLane(renderLanes, current.childLanes);
<add> if (didReceiveUpdate || hasContextChanged) {
<add> // This boundary has changed since the first render. This means that we are now unable to
<add> // hydrate it. We might still be able to hydrate it using a higher priority lane.
<add> const root = getWorkInProgressRoot();
<add> if (root !== null) {
<add> const attemptHydrationAtLane = getBumpedLaneForHydration(
<add> root,
<add> renderLanes,
<add> );
<add> if (
<add> attemptHydrationAtLane !== NoLane &&
<add> attemptHydrationAtLane !== suspenseState.retryLane
<add> ) {
<add> // Intentionally mutating since this render will get interrupted. This
<add> // is one of the very rare times where we mutate the current tree
<add> // during the render phase.
<add> suspenseState.retryLane = attemptHydrationAtLane;
<add> // TODO: Ideally this would inherit the event time of the current render
<add> const eventTime = NoTimestamp;
<add> scheduleUpdateOnFiber(current, attemptHydrationAtLane, eventTime);
<add> } else {
<add> // We have already tried to ping at a higher priority than we're rendering with
<add> // so if we got here, we must have failed to hydrate at those levels. We must
<add> // now give up. Instead, we're going to delete the whole subtree and instead inject
<add> // a new real Suspense boundary to take its place, which may render content
<add> // or fallback. This might suspend for a while and if it does we might still have
<add> // an opportunity to hydrate before this pass commits.
<add> }
<add> }
<add>
<add> // If we have scheduled higher pri work above, this will probably just abort the render
<add> // since we now have higher priority work, but in case it doesn't, we need to prepare to
<add> // render something, if we time out. Even if that requires us to delete everything and
<add> // skip hydration.
<add> // Delay having to do this as long as the suspense timeout allows us.
<add> renderDidSuspendDelayIfPossible();
<add> return retrySuspenseComponentWithoutHydrating(
<add> current,
<add> workInProgress,
<add> renderLanes,
<add> new Error(
<add> 'This Suspense boundary received an update before it finished ' +
<add> 'hydrating. This caused the boundary to switch to client rendering. ' +
<add> 'The usual way to fix this is to wrap the original update ' +
<add> 'in startTransition.',
<add> ),
<add> );
<add> } else if (isSuspenseInstancePending(suspenseInstance)) {
<add> // This component is still pending more data from the server, so we can't hydrate its
<add> // content. We treat it as if this component suspended itself. It might seem as if
<add> // we could just try to render it client-side instead. However, this will perform a
<add> // lot of unnecessary work and is unlikely to complete since it often will suspend
<add> // on missing data anyway. Additionally, the server might be able to render more
<add> // than we can on the client yet. In that case we'd end up with more fallback states
<add> // on the client than if we just leave it alone. If the server times out or errors
<add> // these should update this boundary to the permanent Fallback state instead.
<add> // Mark it as having captured (i.e. suspended).
<add> workInProgress.flags |= DidCapture;
<add> // Leave the child in place. I.e. the dehydrated fragment.
<add> workInProgress.child = current.child;
<add> // Register a callback to retry this boundary once the server has sent the result.
<add> const retry = retryDehydratedSuspenseBoundary.bind(null, current);
<add> registerSuspenseInstanceRetry(suspenseInstance, retry);
<add> return null;
<add> } else {
<add> // This is the first attempt.
<add> reenterHydrationStateFromDehydratedSuspenseInstance(
<add> workInProgress,
<add> suspenseInstance,
<add> suspenseState.treeContext,
<add> );
<add> const primaryChildren = nextProps.children;
<add> const primaryChildFragment = mountSuspensePrimaryChildren(
<add> workInProgress,
<add> primaryChildren,
<add> renderLanes,
<add> );
<add> // Mark the children as hydrating. This is a fast path to know whether this
<add> // tree is part of a hydrating tree. This is used to determine if a child
<add> // node has fully mounted yet, and for scheduling event replaying.
<add> // Conceptually this is similar to Placement in that a new subtree is
<add> // inserted into the React tree here. It just happens to not need DOM
<add> // mutations because it already exists.
<add> primaryChildFragment.flags |= Hydrating;
<add> return primaryChildFragment;
<add> }
<ide> } else {
<del> // This is the first attempt.
<del> reenterHydrationStateFromDehydratedSuspenseInstance(
<del> workInProgress,
<del> suspenseInstance,
<del> suspenseState.treeContext,
<del> );
<del> const nextProps = workInProgress.pendingProps;
<del> const primaryChildren = nextProps.children;
<del> const primaryChildFragment = mountSuspensePrimaryChildren(
<del> workInProgress,
<del> primaryChildren,
<del> renderLanes,
<del> );
<del> // Mark the children as hydrating. This is a fast path to know whether this
<del> // tree is part of a hydrating tree. This is used to determine if a child
<del> // node has fully mounted yet, and for scheduling event replaying.
<del> // Conceptually this is similar to Placement in that a new subtree is
<del> // inserted into the React tree here. It just happens to not need DOM
<del> // mutations because it already exists.
<del> primaryChildFragment.flags |= Hydrating;
<del> return primaryChildFragment;
<add> // This is the second render pass. We already attempted to hydrated, but
<add> // something either suspended or errored.
<add>
<add> if (workInProgress.flags & ForceClientRender) {
<add> // Something errored during hydration. Try again without hydrating.
<add> workInProgress.flags &= ~ForceClientRender;
<add> return retrySuspenseComponentWithoutHydrating(
<add> current,
<add> workInProgress,
<add> renderLanes,
<add> new Error(
<add> 'There was an error while hydrating this Suspense boundary. ' +
<add> 'Switched to client rendering.',
<add> ),
<add> );
<add> } else if ((workInProgress.memoizedState: null | SuspenseState) !== null) {
<add> // Something suspended and we should still be in dehydrated mode.
<add> // Leave the existing child in place.
<add> workInProgress.child = current.child;
<add> // The dehydrated completion pass expects this flag to be there
<add> // but the normal suspense pass doesn't.
<add> workInProgress.flags |= DidCapture;
<add> return null;
<add> } else {
<add> // Suspended but we should no longer be in dehydrated mode.
<add> // Therefore we now have to render the fallback.
<add> const nextPrimaryChildren = nextProps.children;
<add> const nextFallbackChildren = nextProps.fallback;
<add> const fallbackChildFragment = mountSuspenseFallbackAfterRetryWithoutHydrating(
<add> current,
<add> workInProgress,
<add> nextPrimaryChildren,
<add> nextFallbackChildren,
<add> renderLanes,
<add> );
<add> const primaryChildFragment: Fiber = (workInProgress.child: any);
<add> primaryChildFragment.memoizedState = mountSuspenseOffscreenState(
<add> renderLanes,
<add> );
<add> workInProgress.memoizedState = SUSPENDED_MARKER;
<add> return fallbackChildFragment;
<add> }
<ide> }
<ide> }
<ide>
<ide><path>packages/react-reconciler/src/ReactFiberCompleteWork.new.js
<ide> function bubbleProperties(completedWork: Fiber) {
<ide> return didBailout;
<ide> }
<ide>
<add>function completeDehydratedSuspenseBoundary(
<add> current: Fiber | null,
<add> workInProgress: Fiber,
<add> nextState: SuspenseState | null,
<add>): boolean {
<add> if (
<add> hasUnhydratedTailNodes() &&
<add> (workInProgress.mode & ConcurrentMode) !== NoMode &&
<add> (workInProgress.flags & DidCapture) === NoFlags
<add> ) {
<add> warnIfUnhydratedTailNodes(workInProgress);
<add> resetHydrationState();
<add> workInProgress.flags |= ForceClientRender | Incomplete | ShouldCapture;
<add>
<add> return false;
<add> }
<add>
<add> const wasHydrated = popHydrationState(workInProgress);
<add>
<add> if (nextState !== null && nextState.dehydrated !== null) {
<add> // We might be inside a hydration state the first time we're picking up this
<add> // Suspense boundary, and also after we've reentered it for further hydration.
<add> if (current === null) {
<add> if (!wasHydrated) {
<add> throw new Error(
<add> 'A dehydrated suspense component was completed without a hydrated node. ' +
<add> 'This is probably a bug in React.',
<add> );
<add> }
<add> prepareToHydrateHostSuspenseInstance(workInProgress);
<add> bubbleProperties(workInProgress);
<add> if (enableProfilerTimer) {
<add> if ((workInProgress.mode & ProfileMode) !== NoMode) {
<add> const isTimedOutSuspense = nextState !== null;
<add> if (isTimedOutSuspense) {
<add> // Don't count time spent in a timed out Suspense subtree as part of the base duration.
<add> const primaryChildFragment = workInProgress.child;
<add> if (primaryChildFragment !== null) {
<add> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator
<add> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number);
<add> }
<add> }
<add> }
<add> }
<add> return false;
<add> } else {
<add> // We might have reentered this boundary to hydrate it. If so, we need to reset the hydration
<add> // state since we're now exiting out of it. popHydrationState doesn't do that for us.
<add> resetHydrationState();
<add> if ((workInProgress.flags & DidCapture) === NoFlags) {
<add> // This boundary did not suspend so it's now hydrated and unsuspended.
<add> workInProgress.memoizedState = null;
<add> }
<add> // If nothing suspended, we need to schedule an effect to mark this boundary
<add> // as having hydrated so events know that they're free to be invoked.
<add> // It's also a signal to replay events and the suspense callback.
<add> // If something suspended, schedule an effect to attach retry listeners.
<add> // So we might as well always mark this.
<add> workInProgress.flags |= Update;
<add> bubbleProperties(workInProgress);
<add> if (enableProfilerTimer) {
<add> if ((workInProgress.mode & ProfileMode) !== NoMode) {
<add> const isTimedOutSuspense = nextState !== null;
<add> if (isTimedOutSuspense) {
<add> // Don't count time spent in a timed out Suspense subtree as part of the base duration.
<add> const primaryChildFragment = workInProgress.child;
<add> if (primaryChildFragment !== null) {
<add> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator
<add> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number);
<add> }
<add> }
<add> }
<add> }
<add> return false;
<add> }
<add> } else {
<add> // Successfully completed this tree. If this was a forced client render,
<add> // there may have been recoverable errors during first hydration
<add> // attempt. If so, add them to a queue so we can log them in the
<add> // commit phase.
<add> upgradeHydrationErrorsToRecoverable();
<add>
<add> // Fall through to normal Suspense path
<add> return true;
<add> }
<add>}
<add>
<ide> function completeWork(
<ide> current: Fiber | null,
<ide> workInProgress: Fiber,
<ide> function completeWork(
<ide> popSuspenseContext(workInProgress);
<ide> const nextState: null | SuspenseState = workInProgress.memoizedState;
<ide>
<add> // Special path for dehydrated boundaries. We may eventually move this
<add> // to its own fiber type so that we can add other kinds of hydration
<add> // boundaries that aren't associated with a Suspense tree. In anticipation
<add> // of such a refactor, all the hydration logic is contained in
<add> // this branch.
<ide> if (
<del> hasUnhydratedTailNodes() &&
<del> (workInProgress.mode & ConcurrentMode) !== NoMode &&
<del> (workInProgress.flags & DidCapture) === NoFlags
<add> current === null ||
<add> (current.memoizedState !== null &&
<add> current.memoizedState.dehydrated !== null)
<ide> ) {
<del> warnIfUnhydratedTailNodes(workInProgress);
<del> resetHydrationState();
<del> workInProgress.flags |= ForceClientRender | Incomplete | ShouldCapture;
<del> return workInProgress;
<del> }
<del> if (nextState !== null && nextState.dehydrated !== null) {
<del> // We might be inside a hydration state the first time we're picking up this
<del> // Suspense boundary, and also after we've reentered it for further hydration.
<del> const wasHydrated = popHydrationState(workInProgress);
<del> if (current === null) {
<del> if (!wasHydrated) {
<del> throw new Error(
<del> 'A dehydrated suspense component was completed without a hydrated node. ' +
<del> 'This is probably a bug in React.',
<del> );
<del> }
<del> prepareToHydrateHostSuspenseInstance(workInProgress);
<del> bubbleProperties(workInProgress);
<del> if (enableProfilerTimer) {
<del> if ((workInProgress.mode & ProfileMode) !== NoMode) {
<del> const isTimedOutSuspense = nextState !== null;
<del> if (isTimedOutSuspense) {
<del> // Don't count time spent in a timed out Suspense subtree as part of the base duration.
<del> const primaryChildFragment = workInProgress.child;
<del> if (primaryChildFragment !== null) {
<del> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator
<del> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number);
<del> }
<del> }
<del> }
<del> }
<del> return null;
<del> } else {
<del> // We might have reentered this boundary to hydrate it. If so, we need to reset the hydration
<del> // state since we're now exiting out of it. popHydrationState doesn't do that for us.
<del> resetHydrationState();
<del> if ((workInProgress.flags & DidCapture) === NoFlags) {
<del> // This boundary did not suspend so it's now hydrated and unsuspended.
<del> workInProgress.memoizedState = null;
<del> }
<del> // If nothing suspended, we need to schedule an effect to mark this boundary
<del> // as having hydrated so events know that they're free to be invoked.
<del> // It's also a signal to replay events and the suspense callback.
<del> // If something suspended, schedule an effect to attach retry listeners.
<del> // So we might as well always mark this.
<del> workInProgress.flags |= Update;
<del> bubbleProperties(workInProgress);
<del> if (enableProfilerTimer) {
<del> if ((workInProgress.mode & ProfileMode) !== NoMode) {
<del> const isTimedOutSuspense = nextState !== null;
<del> if (isTimedOutSuspense) {
<del> // Don't count time spent in a timed out Suspense subtree as part of the base duration.
<del> const primaryChildFragment = workInProgress.child;
<del> if (primaryChildFragment !== null) {
<del> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator
<del> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number);
<del> }
<del> }
<del> }
<add> const fallthroughToNormalSuspensePath = completeDehydratedSuspenseBoundary(
<add> current,
<add> workInProgress,
<add> nextState,
<add> );
<add> if (!fallthroughToNormalSuspensePath) {
<add> if (workInProgress.flags & ShouldCapture) {
<add> // Special case. There were remaining unhydrated nodes. We treat
<add> // this as a mismatch. Revert to client rendering.
<add> return workInProgress;
<add> } else {
<add> // Did not finish hydrating, either because this is the initial
<add> // render or because something suspended.
<add> return null;
<ide> }
<del> return null;
<ide> }
<del> }
<ide>
<del> // Successfully completed this tree. If this was a forced client render,
<del> // there may have been recoverable errors during first hydration
<del> // attempt. If so, add them to a queue so we can log them in the
<del> // commit phase.
<del> upgradeHydrationErrorsToRecoverable();
<add> // Continue with the normal Suspense path.
<add> }
<ide>
<ide> if ((workInProgress.flags & DidCapture) !== NoFlags) {
<ide> // Something suspended. Re-render with the fallback children.
<ide> function completeWork(
<ide> }
<ide>
<ide> const nextDidTimeout = nextState !== null;
<del> let prevDidTimeout = false;
<del> if (current === null) {
<del> popHydrationState(workInProgress);
<del> } else {
<del> const prevState: null | SuspenseState = current.memoizedState;
<del> prevDidTimeout = prevState !== null;
<del> }
<add> const prevDidTimeout =
<add> current !== null &&
<add> (current.memoizedState: null | SuspenseState) !== null;
<ide>
<ide> if (enableCache && nextDidTimeout) {
<ide> const offscreenFiber: Fiber = (workInProgress.child: any);
<ide><path>packages/react-reconciler/src/ReactFiberCompleteWork.old.js
<ide> function bubbleProperties(completedWork: Fiber) {
<ide> return didBailout;
<ide> }
<ide>
<add>function completeDehydratedSuspenseBoundary(
<add> current: Fiber | null,
<add> workInProgress: Fiber,
<add> nextState: SuspenseState | null,
<add>): boolean {
<add> if (
<add> hasUnhydratedTailNodes() &&
<add> (workInProgress.mode & ConcurrentMode) !== NoMode &&
<add> (workInProgress.flags & DidCapture) === NoFlags
<add> ) {
<add> warnIfUnhydratedTailNodes(workInProgress);
<add> resetHydrationState();
<add> workInProgress.flags |= ForceClientRender | Incomplete | ShouldCapture;
<add>
<add> return false;
<add> }
<add>
<add> const wasHydrated = popHydrationState(workInProgress);
<add>
<add> if (nextState !== null && nextState.dehydrated !== null) {
<add> // We might be inside a hydration state the first time we're picking up this
<add> // Suspense boundary, and also after we've reentered it for further hydration.
<add> if (current === null) {
<add> if (!wasHydrated) {
<add> throw new Error(
<add> 'A dehydrated suspense component was completed without a hydrated node. ' +
<add> 'This is probably a bug in React.',
<add> );
<add> }
<add> prepareToHydrateHostSuspenseInstance(workInProgress);
<add> bubbleProperties(workInProgress);
<add> if (enableProfilerTimer) {
<add> if ((workInProgress.mode & ProfileMode) !== NoMode) {
<add> const isTimedOutSuspense = nextState !== null;
<add> if (isTimedOutSuspense) {
<add> // Don't count time spent in a timed out Suspense subtree as part of the base duration.
<add> const primaryChildFragment = workInProgress.child;
<add> if (primaryChildFragment !== null) {
<add> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator
<add> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number);
<add> }
<add> }
<add> }
<add> }
<add> return false;
<add> } else {
<add> // We might have reentered this boundary to hydrate it. If so, we need to reset the hydration
<add> // state since we're now exiting out of it. popHydrationState doesn't do that for us.
<add> resetHydrationState();
<add> if ((workInProgress.flags & DidCapture) === NoFlags) {
<add> // This boundary did not suspend so it's now hydrated and unsuspended.
<add> workInProgress.memoizedState = null;
<add> }
<add> // If nothing suspended, we need to schedule an effect to mark this boundary
<add> // as having hydrated so events know that they're free to be invoked.
<add> // It's also a signal to replay events and the suspense callback.
<add> // If something suspended, schedule an effect to attach retry listeners.
<add> // So we might as well always mark this.
<add> workInProgress.flags |= Update;
<add> bubbleProperties(workInProgress);
<add> if (enableProfilerTimer) {
<add> if ((workInProgress.mode & ProfileMode) !== NoMode) {
<add> const isTimedOutSuspense = nextState !== null;
<add> if (isTimedOutSuspense) {
<add> // Don't count time spent in a timed out Suspense subtree as part of the base duration.
<add> const primaryChildFragment = workInProgress.child;
<add> if (primaryChildFragment !== null) {
<add> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator
<add> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number);
<add> }
<add> }
<add> }
<add> }
<add> return false;
<add> }
<add> } else {
<add> // Successfully completed this tree. If this was a forced client render,
<add> // there may have been recoverable errors during first hydration
<add> // attempt. If so, add them to a queue so we can log them in the
<add> // commit phase.
<add> upgradeHydrationErrorsToRecoverable();
<add>
<add> // Fall through to normal Suspense path
<add> return true;
<add> }
<add>}
<add>
<ide> function completeWork(
<ide> current: Fiber | null,
<ide> workInProgress: Fiber,
<ide> function completeWork(
<ide> popSuspenseContext(workInProgress);
<ide> const nextState: null | SuspenseState = workInProgress.memoizedState;
<ide>
<add> // Special path for dehydrated boundaries. We may eventually move this
<add> // to its own fiber type so that we can add other kinds of hydration
<add> // boundaries that aren't associated with a Suspense tree. In anticipation
<add> // of such a refactor, all the hydration logic is contained in
<add> // this branch.
<ide> if (
<del> hasUnhydratedTailNodes() &&
<del> (workInProgress.mode & ConcurrentMode) !== NoMode &&
<del> (workInProgress.flags & DidCapture) === NoFlags
<add> current === null ||
<add> (current.memoizedState !== null &&
<add> current.memoizedState.dehydrated !== null)
<ide> ) {
<del> warnIfUnhydratedTailNodes(workInProgress);
<del> resetHydrationState();
<del> workInProgress.flags |= ForceClientRender | Incomplete | ShouldCapture;
<del> return workInProgress;
<del> }
<del> if (nextState !== null && nextState.dehydrated !== null) {
<del> // We might be inside a hydration state the first time we're picking up this
<del> // Suspense boundary, and also after we've reentered it for further hydration.
<del> const wasHydrated = popHydrationState(workInProgress);
<del> if (current === null) {
<del> if (!wasHydrated) {
<del> throw new Error(
<del> 'A dehydrated suspense component was completed without a hydrated node. ' +
<del> 'This is probably a bug in React.',
<del> );
<del> }
<del> prepareToHydrateHostSuspenseInstance(workInProgress);
<del> bubbleProperties(workInProgress);
<del> if (enableProfilerTimer) {
<del> if ((workInProgress.mode & ProfileMode) !== NoMode) {
<del> const isTimedOutSuspense = nextState !== null;
<del> if (isTimedOutSuspense) {
<del> // Don't count time spent in a timed out Suspense subtree as part of the base duration.
<del> const primaryChildFragment = workInProgress.child;
<del> if (primaryChildFragment !== null) {
<del> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator
<del> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number);
<del> }
<del> }
<del> }
<del> }
<del> return null;
<del> } else {
<del> // We might have reentered this boundary to hydrate it. If so, we need to reset the hydration
<del> // state since we're now exiting out of it. popHydrationState doesn't do that for us.
<del> resetHydrationState();
<del> if ((workInProgress.flags & DidCapture) === NoFlags) {
<del> // This boundary did not suspend so it's now hydrated and unsuspended.
<del> workInProgress.memoizedState = null;
<del> }
<del> // If nothing suspended, we need to schedule an effect to mark this boundary
<del> // as having hydrated so events know that they're free to be invoked.
<del> // It's also a signal to replay events and the suspense callback.
<del> // If something suspended, schedule an effect to attach retry listeners.
<del> // So we might as well always mark this.
<del> workInProgress.flags |= Update;
<del> bubbleProperties(workInProgress);
<del> if (enableProfilerTimer) {
<del> if ((workInProgress.mode & ProfileMode) !== NoMode) {
<del> const isTimedOutSuspense = nextState !== null;
<del> if (isTimedOutSuspense) {
<del> // Don't count time spent in a timed out Suspense subtree as part of the base duration.
<del> const primaryChildFragment = workInProgress.child;
<del> if (primaryChildFragment !== null) {
<del> // $FlowFixMe Flow doesn't support type casting in combination with the -= operator
<del> workInProgress.treeBaseDuration -= ((primaryChildFragment.treeBaseDuration: any): number);
<del> }
<del> }
<del> }
<add> const fallthroughToNormalSuspensePath = completeDehydratedSuspenseBoundary(
<add> current,
<add> workInProgress,
<add> nextState,
<add> );
<add> if (!fallthroughToNormalSuspensePath) {
<add> if (workInProgress.flags & ShouldCapture) {
<add> // Special case. There were remaining unhydrated nodes. We treat
<add> // this as a mismatch. Revert to client rendering.
<add> return workInProgress;
<add> } else {
<add> // Did not finish hydrating, either because this is the initial
<add> // render or because something suspended.
<add> return null;
<ide> }
<del> return null;
<ide> }
<del> }
<ide>
<del> // Successfully completed this tree. If this was a forced client render,
<del> // there may have been recoverable errors during first hydration
<del> // attempt. If so, add them to a queue so we can log them in the
<del> // commit phase.
<del> upgradeHydrationErrorsToRecoverable();
<add> // Continue with the normal Suspense path.
<add> }
<ide>
<ide> if ((workInProgress.flags & DidCapture) !== NoFlags) {
<ide> // Something suspended. Re-render with the fallback children.
<ide> function completeWork(
<ide> }
<ide>
<ide> const nextDidTimeout = nextState !== null;
<del> let prevDidTimeout = false;
<del> if (current === null) {
<del> popHydrationState(workInProgress);
<del> } else {
<del> const prevState: null | SuspenseState = current.memoizedState;
<del> prevDidTimeout = prevState !== null;
<del> }
<add> const prevDidTimeout =
<add> current !== null &&
<add> (current.memoizedState: null | SuspenseState) !== null;
<ide>
<ide> if (enableCache && nextDidTimeout) {
<ide> const offscreenFiber: Fiber = (workInProgress.child: any);
| 4
|
Javascript
|
Javascript
|
remove obsolete sharedarraybuffer eslint config
|
82ef450e0e22ef3d5cff03d26aed7d708fa10537
|
<ide><path>.eslintrc.js
<ide> module.exports = {
<ide> ],
<ide>
<ide> globals: {
<del> SharedArrayBuffer: true,
<del>
<ide> spyOnDev: true,
<ide> spyOnDevAndProd: true,
<ide> spyOnProd: true,
<ide><path>scripts/rollup/validate/eslintrc.cjs.js
<ide> module.exports = {
<ide> trustedTypes: true,
<ide>
<ide> // Scheduler profiling
<del> SharedArrayBuffer: true,
<ide> Int32Array: true,
<ide> ArrayBuffer: true,
<ide>
<ide><path>scripts/rollup/validate/eslintrc.cjs2015.js
<ide> module.exports = {
<ide> trustedTypes: true,
<ide>
<ide> // Scheduler profiling
<del> SharedArrayBuffer: true,
<ide> Int32Array: true,
<ide> ArrayBuffer: true,
<ide>
<ide><path>scripts/rollup/validate/eslintrc.esm.js
<ide> module.exports = {
<ide> trustedTypes: true,
<ide>
<ide> // Scheduler profiling
<del> SharedArrayBuffer: true,
<ide> Int32Array: true,
<ide> ArrayBuffer: true,
<ide>
<ide><path>scripts/rollup/validate/eslintrc.fb.js
<ide> module.exports = {
<ide> trustedTypes: true,
<ide>
<ide> // Scheduler profiling
<del> SharedArrayBuffer: true,
<ide> Int32Array: true,
<ide> ArrayBuffer: true,
<ide>
<ide><path>scripts/rollup/validate/eslintrc.rn.js
<ide> module.exports = {
<ide> trustedTypes: true,
<ide>
<ide> // Scheduler profiling
<del> SharedArrayBuffer: true,
<ide> Int32Array: true,
<ide> ArrayBuffer: true,
<ide>
<ide><path>scripts/rollup/validate/eslintrc.umd.js
<ide> module.exports = {
<ide> trustedTypes: true,
<ide>
<ide> // Scheduler profiling
<del> SharedArrayBuffer: true,
<ide> Int32Array: true,
<ide> ArrayBuffer: true,
<ide>
| 7
|
Ruby
|
Ruby
|
relax hub requirement
|
6dc72f2679208f34d9d09f547dda5a2ba0714a9f
|
<ide><path>Library/Homebrew/dev-cmd/boneyard-formula-pr.rb
<ide> def boneyard_formula_pr
<ide> tap_migrations = tap_migrations.sort.inject({}) { |a, e| a.merge!(e[0] => e[1]) }
<ide> tap_migrations_path.atomic_write(JSON.pretty_generate(tap_migrations) + "\n")
<ide> end
<del> unless Formula["hub"].any_version_installed? || local_only
<add> unless which("hub") || local_only
<ide> if ARGV.dry_run?
<ide> puts "brew install hub"
<ide> else
| 1
|
Python
|
Python
|
remove unused imports.
|
51a2b441c65e5df4fa7b540c2c1211e3adff53cb
|
<ide><path>official/mnist/mnist.py
<ide> from __future__ import division
<ide> from __future__ import print_function
<ide>
<del>import argparse
<del>import sys
<del>
<ide> from absl import app as absl_app
<ide> from absl import flags
<ide> import tensorflow as tf # pylint: disable=g-bad-import-order
<ide><path>official/mnist/mnist_eager.py
<ide> from __future__ import print_function
<ide>
<ide> import os
<del>import sys
<ide> import time
<ide>
<ide> # pylint: disable=g-bad-import-order
<ide><path>official/resnet/cifar10_main.py
<ide> from __future__ import print_function
<ide>
<ide> import os
<del>import sys
<ide>
<ide> from absl import app as absl_app
<ide> from absl import flags
<ide><path>official/resnet/imagenet_main.py
<ide> from __future__ import print_function
<ide>
<ide> import os
<del>import sys
<ide>
<ide> from absl import app as absl_app
<ide> from absl import flags
<ide><path>official/resnet/resnet_run_loop.py
<ide> from __future__ import division
<ide> from __future__ import print_function
<ide>
<del>import argparse
<ide> import os
<ide>
<ide> from absl import flags
<ide><path>official/wide_deep/wide_deep.py
<ide> from __future__ import division
<ide> from __future__ import print_function
<ide>
<del>import argparse
<ide> import os
<ide> import shutil
<del>import sys
<ide>
<ide> from absl import app as absl_app
<ide> from absl import flags
| 6
|
Ruby
|
Ruby
|
remove a comment that is not fact in tests
|
41758964e9515ae879fce0be687da2f5953bc950
|
<ide><path>activerecord/test/cases/calculations_test.rb
<ide> def test_pluck_with_selection_clause
<ide> assert_equal [50, 53, 55, 60], Account.pluck(Arel.sql("DISTINCT credit_limit")).sort
<ide> assert_equal [50, 53, 55, 60], Account.pluck(Arel.sql("DISTINCT accounts.credit_limit")).sort
<ide> assert_equal [50, 53, 55, 60], Account.pluck(Arel.sql("DISTINCT(credit_limit)")).sort
<del>
<del> # MySQL returns "SUM(DISTINCT(credit_limit))" as the column name unless
<del> # an alias is provided. Without the alias, the column cannot be found
<del> # and properly typecast.
<del> assert_equal [50 + 53 + 55 + 60], Account.pluck(Arel.sql("SUM(DISTINCT(credit_limit)) as credit_limit"))
<add> assert_equal [50 + 53 + 55 + 60], Account.pluck(Arel.sql("SUM(DISTINCT(credit_limit))"))
<ide> end
<ide>
<ide> def test_plucks_with_ids
| 1
|
PHP
|
PHP
|
fix missing use of binding key in belongstomany
|
6fcff4875f07951294c1185fc4e1769929639ad5
|
<ide><path>src/ORM/Association/BelongsToMany.php
<ide> protected function _generateJunctionAssociations(Table $junction, Table $source,
<ide>
<ide> if (!$junction->hasAssociation($sAlias)) {
<ide> $junction->belongsTo($sAlias, [
<add> 'bindingKey' => $this->getBindingKey(),
<ide> 'foreignKey' => $this->getForeignKey(),
<ide> 'targetTable' => $source,
<ide> ]);
<ide><path>tests/Fixture/ArticlesTagsBindingKeysFixture.php
<add><?php
<add>/**
<add> * CakePHP(tm) : Rapid Development Framework (https://cakephp.org)
<add> * Copyright (c) Cake Software Foundation, Inc. (https://cakefoundation.org)
<add> *
<add> * Licensed under The MIT License
<add> * For full copyright and license information, please see the LICENSE.txt
<add> * Redistributions of files must retain the above copyright notice
<add> *
<add> * @copyright Copyright (c) Cake Software Foundation, Inc. (https://cakefoundation.org)
<add> * @link https://cakephp.org CakePHP(tm) Project
<add> * @since 4.3.7
<add> * @license https://opensource.org/licenses/mit-license.php MIT License
<add> */
<add>namespace Cake\Test\Fixture;
<add>
<add>use Cake\TestSuite\Fixture\TestFixture;
<add>
<add>/**
<add> * Fixture for testing bindingKey in belongstomany associations.
<add> */
<add>class ArticlesTagsBindingKeysFixture extends TestFixture
<add>{
<add> /**
<add> * records property
<add> *
<add> * @var array
<add> */
<add> public $records = [
<add> ['article_id' => 1, 'tagname' => 'tag1'],
<add> ['article_id' => 1, 'tagname' => 'tag2'],
<add> ['article_id' => 2, 'tagname' => 'tag1'],
<add> ['article_id' => 2, 'tagname' => 'tag3'],
<add> ];
<add>}
<ide><path>tests/TestCase/ORM/Association/BelongsToManyTest.php
<ide> class BelongsToManyTest extends TestCase
<ide> 'core.Tags',
<ide> 'core.SpecialTags',
<ide> 'core.ArticlesTags',
<add> 'core.ArticlesTagsBindingKeys',
<ide> 'core.BinaryUuidItems',
<ide> 'core.BinaryUuidTags',
<ide> 'core.BinaryUuidItemsBinaryUuidTags',
<ide> public function testCustomTargetBindingKeyLink(): void
<ide> $this->assertCount(1, $results);
<ide> $this->assertCount(3, $results[0]->special_tags);
<ide> }
<add>
<add> /**
<add> * Test custom binding key for target table association
<add> */
<add> public function testBindingKeyMatching(): void
<add> {
<add> $table = $this->getTableLocator()->get('Tags');
<add> $table->belongsToMany('Articles', [
<add> 'through' => 'ArticlesTagsBindingKeys',
<add> 'foreignKey' => 'tagname',
<add> 'targetForeignKey' => 'article_id',
<add> 'bindingKey' => 'name',
<add> ]);
<add> $query = $table->find()
<add> ->matching('Articles', function ($q) {
<add> return $q->where('Articles.id > 0');
<add> });
<add> $results = $query->all();
<add>
<add> // 4 records in the junction table.
<add> $this->assertCount(4, $results);
<add> }
<ide> }
<ide><path>tests/schema.php
<ide> ],
<ide> ],
<ide> ],
<add> [
<add> 'table' => 'articles_tags_binding_keys',
<add> 'columns' => [
<add> 'article_id' => [
<add> 'type' => 'integer',
<add> 'null' => false,
<add> ],
<add> 'tagname' => [
<add> 'type' => 'string',
<add> 'null' => false,
<add> ],
<add> ],
<add> 'constraints' => [
<add> 'unique_tag' => [
<add> 'type' => 'primary',
<add> 'columns' => [
<add> 'article_id',
<add> 'tagname',
<add> ],
<add> ],
<add> ],
<add> ],
<ide> [
<ide> 'table' => 'profiles',
<ide> 'columns' => [
| 4
|
Ruby
|
Ruby
|
add newline to failure message
|
c3b309e7957d38952a85b4d320ac73112d2733af
|
<ide><path>Library/Homebrew/requirements/java_requirement.rb
<ide> def initialize(tags)
<ide> def message
<ide> version_string = " #{@version}" if @version
<ide>
<del> s = "Java#{version_string} is required to install this formula."
<add> s = "Java#{version_string} is required to install this formula.\n"
<ide> s += super
<ide> s
<ide> end
| 1
|
Javascript
|
Javascript
|
remove extra scene parameter from jsm version
|
447d9dd95897527b504ace0b173056a30f3fe59e
|
<ide><path>examples/jsm/renderers/CSS2DRenderer.js
<ide> var CSS2DRenderer = function () {
<ide>
<ide> for ( var i = 0, l = object.children.length; i < l; i ++ ) {
<ide>
<del> renderObject( object.children[ i ], scene, camera );
<add> renderObject( object.children[ i ], camera );
<ide>
<ide> }
<ide>
| 1
|
Javascript
|
Javascript
|
handle dashes in keys in tohavedata matcher
|
2ef6d1be965bb7e579403b809331370849709b6c
|
<ide><path>vendor/jasmine-jquery.js
<ide> var jQueryMatchers = {
<ide>
<ide> toHaveData: function(key, expectedValue) {
<ide> if (this.actual instanceof HTMLElement) {
<del> return hasProperty(this.actual.dataset[key], expectedValue)
<add> var camelCaseKey
<add> for (var part of key.split('-')) {
<add> if (camelCaseKey) {
<add> camelCaseKey += part[0].toUpperCase() + part.substring(1)
<add> } else {
<add> camelCaseKey = part
<add> }
<add> }
<add> return hasProperty(this.actual.dataset[camelCaseKey], expectedValue)
<ide> } else {
<ide> return hasProperty(this.actual.data(key), expectedValue)
<ide> }
| 1
|
Ruby
|
Ruby
|
replace capture2e with utils.popen
|
e8f78878cb81c4dec14f7e1498ead71892502072
|
<ide><path>Library/Homebrew/utils/livecheck.rb
<ide> # frozen_string_literal: true
<ide>
<del>require "open3"
<del>
<ide> module Livecheck
<del> def livecheck_formula_response(formula_name)
<del> ohai "Checking livecheck formula : #{formula_name}"
<del> command_args = ["brew", "livecheck", formula_name, "--quiet"]
<add> def livecheck_formula_response(name)
<add> ohai "Checking livecheck formula : #{name}"
<ide>
<del> response = Open3.capture2e(*command_args)
<add> response = Utils.popen_read("brew", "livecheck", name, "--quiet").chomp
<ide> parse_livecheck_response(response)
<ide> end
<ide>
<ide><path>Library/Homebrew/utils/versions.rb
<ide> # frozen_string_literal: true
<ide>
<del>require "open3"
<ide> require "formula"
<ide>
<ide> module Versions
<ide> def current_formula_version(formula_name)
<ide> end
<ide>
<ide> def bump_formula_pr(formula_name, url)
<del> command_args = ["brew", "bump-formula-pr", "--no-browse",
<del> "--dry-run", formula_name, "--url=#{url}"]
<add> response = Utils.popen_read("brew", "bump-formula-pr", "--no-browse",
<add> "--dry-run", formula_name, "--url=#{url}").chomp
<ide>
<del> response = Open3.capture2e(*command_args)
<ide> parse_formula_bump_response(response)
<ide> end
<ide>
| 2
|
Ruby
|
Ruby
|
add tests for credentials command
|
8a331566bfe05ccc85f25805693aa892cec449a5
|
<ide><path>railties/test/commands/credentials_test.rb
<ide> class Rails::Command::CredentialsCommandTest < ActiveSupport::TestCase
<ide>
<ide> teardown { teardown_app }
<ide>
<add> test "edit without editor gives hint" do
<add> assert_match "No $EDITOR to open credentials in", run_edit_command(editor: "")
<add> end
<add>
<add> test "edit credentials" do
<add> # Run twice to ensure credentials can be reread after first edit pass.
<add> 2.times do
<add> assert_match(/access_key_id: 123/, run_edit_command)
<add> end
<add> end
<add>
<add> test "show credentials" do
<add> assert_match(/access_key_id: 123/, run_show_command)
<add> end
<add>
<ide> test "edit command does not add master key to gitignore when already exist" do
<ide> run_edit_command
<ide>
<ide> def run_edit_command(editor: "cat")
<ide> rails "credentials:edit"
<ide> end
<ide> end
<add>
<add> def run_show_command
<add> rails "credentials:show"
<add> end
<ide> end
| 1
|
Java
|
Java
|
fix checkstyle violations
|
385292dc41c62e96a083c41c9c686825db251d63
|
<ide><path>spring-core/src/main/java/org/springframework/core/io/support/PathMatchingResourcePatternResolver.java
<ide> protected Set<Resource> doFindPathMatchingFileResources(Resource rootDirResource
<ide> FileSystem fileSystem;
<ide> try {
<ide> fileSystem = FileSystems.getFileSystem(rootDirResource.getURI().resolve("/"));
<del> } catch (Exception e) {
<add> }
<add> catch (Exception ex) {
<ide> fileSystem = FileSystems.newFileSystem(rootDirResource.getURI().resolve("/"), Map.of(),
<ide> ClassUtils.getDefaultClassLoader());
<ide> }
<ide> protected Set<Resource> doFindPathMatchingFileResources(Resource rootDirResource
<ide> if (getPathMatcher().match(patternPath.toString(), file.toString())) {
<ide> try {
<ide> result.add(new UrlResource(file.toUri()));
<del> } catch (MalformedURLException e) {
<add> }
<add> catch (MalformedURLException ex) {
<ide> // ignore
<ide> }
<ide> }
<ide> });
<del> } catch (NoSuchFileException e) {
<add> }
<add> catch (NoSuchFileException ex) {
<ide> // ignore
<ide> }
<ide> try {
<ide> fileSystem.close();
<del> } catch (UnsupportedOperationException e) {
<add> }
<add> catch (UnsupportedOperationException ex) {
<ide> // ignore
<ide> }
<ide> return result;
| 1
|
Python
|
Python
|
add a method to convert str to ascii in python 3
|
54b61fd5a090061b1016e1fee115c6808bbb06ed
|
<ide><path>glances/compat.py
<ide> def mean(numbers):
<ide> def to_ascii(s):
<ide> """Convert the bytes string to a ASCII string
<ide> Usefull to remove accent (diacritics)"""
<del> return str(s, 'utf-8')
<add> if isinstance(s, binary_type):
<add> return s
<add> return s.encode('ascii', 'ignore')
<ide>
<ide> def listitems(d):
<ide> return list(d.items())
<ide> def to_ascii(s):
<ide> Usefull to remove accent (diacritics)"""
<ide> if isinstance(s, binary_type):
<ide> return s
<del> return unicodedata.normalize('NFKD', s).encode('ASCII', 'ignore')
<add> return unicodedata.normalize('NFKD', s).encode('ascii', 'ignore')
<ide>
<ide> def listitems(d):
<ide> return d.items()
| 1
|
Ruby
|
Ruby
|
add test for build.without --without-foo
|
02a1406a2e4c2d0506861248eb767d99f5ce4515
|
<ide><path>Library/Homebrew/rubocops/lines_cop.rb
<ide> def audit_formula(_node, _class_node, _parent_class_node, body_node)
<ide> problem "Don't negate 'build.without?': use 'build.with?'"
<ide> end
<ide>
<del> # find_instance_method_call(body_node, :build, :without?) do |m|
<del> # arg = parameters(m).first
<del> # next unless match = regex_match_group(arg, %r{-?-?without-(.*)})
<del> # problem "Don't duplicate 'without': Use `build.without? \"#{match[1]}\"` to check for \"--without-#{match[1]}\""
<del> # end
<del> #
<add> find_instance_method_call(body_node, :build, :without?) do |m|
<add> arg = parameters(m).first
<add> next unless match = regex_match_group(arg, %r{-?-?without-(.*)})
<add> problem "Don't duplicate 'without': Use `build.without? \"#{match[1]}\"` to check for \"--without-#{match[1]}\""
<add> end
<add>
<ide> # find_instance_method_call(body_node, :build, :with?) do |m|
<ide> # arg = parameters(m).first
<ide> # next unless match = regex_match_group(arg, %r{-?-?with-(.*)})
<ide><path>Library/Homebrew/test/rubocops/lines_cop_spec.rb
<ide> def post_install
<ide> expect_offense(expected, actual)
<ide> end
<ide> end
<add>
<add> it "with negated build.with?" do
<add> source = <<-EOS.undent
<add> class Foo < Formula
<add> desc "foo"
<add> url 'http://example.com/foo-1.0.tgz'
<add> def post_install
<add> return if build.without? "--without-bar"
<add> end
<add> end
<add> EOS
<add>
<add> expected_offenses = [{ message: "Don't duplicate 'without': Use `build.without? \"bar\"` to check for \"--without-bar\"",
<add> severity: :convention,
<add> line: 5,
<add> column: 30,
<add> source: source }]
<add>
<add> inspect_source(cop, source)
<add>
<add> expected_offenses.zip(cop.offenses).each do |expected, actual|
<add> expect_offense(expected, actual)
<add> end
<add> end
<ide> end
<ide> def expect_offense(expected, actual)
<ide> expect(actual.message).to eq(expected[:message])
| 2
|
Text
|
Text
|
add links to inline html table
|
dd023df135207086c86129aa1683ea4688f97f53
|
<ide><path>doc/api/crypto.md
<ide> The following constants exported by `crypto.constants` apply to various uses of
<ide> the `crypto`, `tls`, and `https` modules and are generally specific to OpenSSL.
<ide>
<ide> ### OpenSSL Options
<del>
<add><!--lint disable maximum-line-length-->
<ide> <table>
<ide> <tr>
<ide> <th>Constant</th>
<ide> the `crypto`, `tls`, and `https` modules and are generally specific to OpenSSL.
<ide> <tr>
<ide> <td><code>SSL_OP_ALL</code></td>
<ide> <td>Applies multiple bug workarounds within OpenSSL. See
<del> https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html for
<del> detail.</td>
<add> <a href="https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html">https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html</a>
<add> for detail.</td>
<ide> </tr>
<ide> <tr>
<ide> <td><code>SSL_OP_ALLOW_UNSAFE_LEGACY_RENEGOTIATION</code></td>
<ide> <td>Allows legacy insecure renegotiation between OpenSSL and unpatched
<ide> clients or servers. See
<del> https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html.</td>
<add> <a href="https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html">https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html</a>.</td>
<ide> </tr>
<ide> <tr>
<ide> <td><code>SSL_OP_CIPHER_SERVER_PREFERENCE</code></td>
<ide> <td>Attempts to use the server's preferences instead of the client's when
<ide> selecting a cipher. Behavior depends on protocol version. See
<del> https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html.</td>
<add> <a href="https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html">https://www.openssl.org/docs/man1.0.2/ssl/SSL_CTX_set_options.html</a>.</td>
<ide> </tr>
<ide> <tr>
<ide> <td><code>SSL_OP_CISCO_ANYCONNECT</code></td>
<ide> the `crypto`, `tls`, and `https` modules and are generally specific to OpenSSL.
<ide> </table>
<ide>
<ide> ### OpenSSL Engine Constants
<add><!--lint enable maximum-line-length remark-lint-->
<ide>
<ide> <table>
<ide> <tr>
| 1
|
Text
|
Text
|
add optional sudo to make install in readme
|
ab04a434761cf66d107481d58798f36d3cb49d46
|
<ide><path>README.md
<ide> Prerequisites:
<ide> ```text
<ide> $ ./configure
<ide> $ make
<del>$ make install
<add>$ [sudo] make install
<ide> ```
<ide>
<ide> If your Python binary is in a non-standard location or has a
<ide> non-standard name, run the following instead:
<ide> $ export PYTHON=/path/to/python
<ide> $ $PYTHON ./configure
<ide> $ make
<del>$ make install
<add>$ [sudo] make install
<ide> ```
<ide>
<ide> To run the tests:
| 1
|
Python
|
Python
|
update 2 mnist examples
|
ca23406974546c5214a3ec73eb903952c14ab2f0
|
<ide><path>examples/mnist_cnn.py
<ide> import numpy as np
<ide> np.random.seed(1337) # for reproducibility
<ide>
<add>import keras
<ide> from keras.datasets import mnist
<ide> from keras.models import Sequential
<del>from keras.layers import Dense, Dropout, Activation, Flatten
<del>from keras.layers import Convolution2D, MaxPooling2D
<del>from keras.utils import np_utils
<add>from keras.layers import Dense, Dropout, Flatten
<add>from keras.layers import Conv2D, MaxPooling2D
<ide> from keras import backend as K
<ide>
<ide> batch_size = 128
<del>nb_classes = 10
<del>nb_epoch = 12
<add>num_classes = 10
<add>num_epoch = 12
<ide>
<ide> # input image dimensions
<ide> img_rows, img_cols = 28, 28
<del># number of convolutional filters to use
<del>nb_filters = 32
<del># size of pooling area for max pooling
<del>pool_size = (2, 2)
<del># convolution kernel size
<del>kernel_size = (3, 3)
<ide>
<ide> # the data, shuffled and split between train and test sets
<del>(X_train, y_train), (X_test, y_test) = mnist.load_data()
<add>(x_train, y_train), (x_test, y_test) = mnist.load_data()
<ide>
<ide> if K.image_data_format() == 'channels_first':
<del> X_train = X_train.reshape(X_train.shape[0], 1, img_rows, img_cols)
<del> X_test = X_test.reshape(X_test.shape[0], 1, img_rows, img_cols)
<add> x_train = x_train.reshape(x_train.shape[0], 1, img_rows, img_cols)
<add> x_test = x_test.reshape(x_test.shape[0], 1, img_rows, img_cols)
<ide> input_shape = (1, img_rows, img_cols)
<ide> else:
<del> X_train = X_train.reshape(X_train.shape[0], img_rows, img_cols, 1)
<del> X_test = X_test.reshape(X_test.shape[0], img_rows, img_cols, 1)
<add> x_train = x_train.reshape(x_train.shape[0], img_rows, img_cols, 1)
<add> x_test = x_test.reshape(x_test.shape[0], img_rows, img_cols, 1)
<ide> input_shape = (img_rows, img_cols, 1)
<ide>
<del>X_train = X_train.astype('float32')
<del>X_test = X_test.astype('float32')
<del>X_train /= 255
<del>X_test /= 255
<del>print('X_train shape:', X_train.shape)
<del>print(X_train.shape[0], 'train samples')
<del>print(X_test.shape[0], 'test samples')
<add>x_train = x_train.astype('float32')
<add>x_test = x_test.astype('float32')
<add>x_train /= 255
<add>x_test /= 255
<add>print('x_train shape:', x_train.shape)
<add>print(x_train.shape[0], 'train samples')
<add>print(x_test.shape[0], 'test samples')
<ide>
<ide> # convert class vectors to binary class matrices
<del>Y_train = np_utils.to_categorical(y_train, nb_classes)
<del>Y_test = np_utils.to_categorical(y_test, nb_classes)
<add>y_train = keras.utils.np_utils.to_categorical(y_train, num_classes)
<add>y_test = keras.utils.np_utils.to_categorical(y_test, num_classes)
<ide>
<ide> model = Sequential()
<del>
<del>model.add(Convolution2D(nb_filters, kernel_size[0], kernel_size[1],
<del> border_mode='valid',
<del> input_shape=input_shape))
<del>model.add(Activation('relu'))
<del>model.add(Convolution2D(nb_filters, kernel_size[0], kernel_size[1]))
<del>model.add(Activation('relu'))
<del>model.add(MaxPooling2D(pool_size=pool_size))
<add>model.add(Conv2D(32, kernel_size=(3, 3),
<add> activation='relu',
<add> input_shape=input_shape))
<add>model.add(Conv2D(64, (3, 3), activation='relu'))
<add>model.add(MaxPooling2D(pool_size=(2, 2)))
<ide> model.add(Dropout(0.25))
<del>
<ide> model.add(Flatten())
<del>model.add(Dense(128))
<del>model.add(Activation('relu'))
<add>model.add(Dense(128, activation='relu'))
<ide> model.add(Dropout(0.5))
<del>model.add(Dense(nb_classes))
<del>model.add(Activation('softmax'))
<add>model.add(Dense(num_classes, activation='softmax'))
<ide>
<del>model.compile(loss='categorical_crossentropy',
<del> optimizer='adadelta',
<add>model.compile(loss=keras.losses.categorical_crossentropy,
<add> optimizer=keras.optimizers.Adadelta(),
<ide> metrics=['accuracy'])
<ide>
<del>model.fit(X_train, Y_train, batch_size=batch_size, nb_epoch=nb_epoch,
<del> verbose=1, validation_data=(X_test, Y_test))
<del>score = model.evaluate(X_test, Y_test, verbose=0)
<del>print('Test score:', score[0])
<add>model.fit(x_train, y_train, batch_size=batch_size, num_epoch=num_epoch,
<add> verbose=1, validation_data=(x_test, y_test))
<add>score = model.evaluate(x_test, y_test, verbose=0)
<add>print('Test loss:', score[0])
<ide> print('Test accuracy:', score[1])
<ide><path>examples/mnist_mlp.py
<ide>
<ide> from keras.datasets import mnist
<ide> from keras.models import Sequential
<del>from keras.layers.core import Dense, Dropout, Activation
<add>from keras.layers import Dense, Dropout, Activation
<ide> from keras.optimizers import RMSprop
<ide> from keras.utils import np_utils
<ide>
<ide>
<ide> batch_size = 128
<del>nb_classes = 10
<del>nb_epoch = 20
<add>num_classes = 10
<add>num_epoch = 20
<ide>
<ide> # the data, shuffled and split between train and test sets
<del>(X_train, y_train), (X_test, y_test) = mnist.load_data()
<add>(x_train, y_train), (x_test, y_test) = mnist.load_data()
<ide>
<del>X_train = X_train.reshape(60000, 784)
<del>X_test = X_test.reshape(10000, 784)
<del>X_train = X_train.astype('float32')
<del>X_test = X_test.astype('float32')
<del>X_train /= 255
<del>X_test /= 255
<del>print(X_train.shape[0], 'train samples')
<del>print(X_test.shape[0], 'test samples')
<add>x_train = x_train.reshape(60000, 784)
<add>x_test = x_test.reshape(10000, 784)
<add>x_train = x_train.astype('float32')
<add>x_test = x_test.astype('float32')
<add>x_train /= 255
<add>x_test /= 255
<add>print(x_train.shape[0], 'train samples')
<add>print(x_test.shape[0], 'test samples')
<ide>
<ide> # convert class vectors to binary class matrices
<del>Y_train = np_utils.to_categorical(y_train, nb_classes)
<del>Y_test = np_utils.to_categorical(y_test, nb_classes)
<add>y_train = np_utils.to_categorical(y_train, num_classes)
<add>y_test = np_utils.to_categorical(y_test, num_classes)
<ide>
<ide> model = Sequential()
<del>model.add(Dense(512, input_shape=(784,)))
<del>model.add(Activation('relu'))
<add>model.add(Dense(512, activation='relu', input_shape=(784,)))
<ide> model.add(Dropout(0.2))
<del>model.add(Dense(512))
<del>model.add(Activation('relu'))
<add>model.add(Dense(512, activation='relu'))
<ide> model.add(Dropout(0.2))
<del>model.add(Dense(10))
<del>model.add(Activation('softmax'))
<add>model.add(Dense(10, activation='softmax'))
<ide>
<ide> model.summary()
<ide>
<ide> model.compile(loss='categorical_crossentropy',
<ide> optimizer=RMSprop(),
<ide> metrics=['accuracy'])
<ide>
<del>history = model.fit(X_train, Y_train,
<del> batch_size=batch_size, nb_epoch=nb_epoch,
<del> verbose=1, validation_data=(X_test, Y_test))
<del>score = model.evaluate(X_test, Y_test, verbose=0)
<del>print('Test score:', score[0])
<add>history = model.fit(x_train, y_train,
<add> batch_size=batch_size, num_epoch=num_epoch,
<add> verbose=1, validation_data=(x_test, y_test))
<add>score = model.evaluate(x_test, y_test, verbose=0)
<add>print('Test loss:', score[0])
<ide> print('Test accuracy:', score[1])
| 2
|
PHP
|
PHP
|
add tests for cachemanager
|
fdcd963158587e9c69980bbf53ad8e1c34d70d9f
|
<ide><path>src/Illuminate/Cache/CacheManager.php
<ide> protected function getPrefix(array $config)
<ide> * Get the cache connection configuration.
<ide> *
<ide> * @param string $name
<del> * @return array
<add> * @return array|null
<ide> */
<ide> protected function getConfig($name)
<ide> {
<ide><path>tests/Cache/CacheManagerTest.php
<ide>
<ide> use Illuminate\Cache\ArrayStore;
<ide> use Illuminate\Cache\CacheManager;
<add>use Illuminate\Config\Repository;
<add>use Illuminate\Container\Container;
<add>use InvalidArgumentException;
<ide> use Mockery as m;
<ide> use PHPUnit\Framework\TestCase;
<ide>
<ide> public function testForgetDriverForgets()
<ide> $cacheManager->forgetDriver('forget');
<ide> $this->assertNull($cacheManager->store('forget')->get('foo'));
<ide> }
<add>
<add> public function testThrowExceptionWhenUnknownDriverIsUsed()
<add> {
<add> $this->expectException(InvalidArgumentException::class);
<add> $this->expectExceptionMessage('Driver [unknown_taxi_driver] is not supported.');
<add>
<add> $userConfig = [
<add> 'cache' => [
<add> 'stores' => [
<add> 'my_store' => [
<add> 'driver' => 'unknown_taxi_driver',
<add> ],
<add> ],
<add> ],
<add> ];
<add>
<add> $app = Container::getInstance();
<add> $app->bind('config', function () use ($userConfig) {
<add> return new Repository($userConfig);
<add> });
<add>
<add> $cacheManager = new CacheManager($app);
<add>
<add> $cacheManager->store('my_store');
<add> }
<add>
<add> public function testThrowExceptionWhenUnknownStoreIsUsed()
<add> {
<add> $this->expectException(InvalidArgumentException::class);
<add> $this->expectExceptionMessage('Cache store [alien_store] is not defined.');
<add>
<add> $userConfig = [
<add> 'cache' => [
<add> 'stores' => [
<add> 'my_store' => [
<add> 'driver' => 'array',
<add> ],
<add> ],
<add> ],
<add> ];
<add>
<add> $app = Container::getInstance();
<add> $app->bind('config', function () use ($userConfig) {
<add> return new Repository($userConfig);
<add> });
<add>
<add> $cacheManager = new CacheManager($app);
<add>
<add> $cacheManager->store('alien_store');
<add> }
<ide> }
| 2
|
Javascript
|
Javascript
|
remove unused require
|
d6e2abfd69301eb2bada951920db779c773dbea8
|
<ide><path>lib/dependencies/CommonJsRequireDependencyParserPlugin.js
<ide> */
<ide> "use strict";
<ide>
<del>const ConstDependency = require("./ConstDependency");
<ide> const CommonJsRequireDependency = require("./CommonJsRequireDependency");
<ide> const CommonJsRequireContextDependency = require("./CommonJsRequireContextDependency");
<ide> const RequireHeaderDependency = require("./RequireHeaderDependency");
| 1
|
PHP
|
PHP
|
fix code styling
|
7e7013222f98ff31db89fb732d047ce910c8753a
|
<ide><path>src/Illuminate/View/Compilers/BladeCompiler.php
<ide> protected function compileOverwrite($expression)
<ide> */
<ide> protected function compileUnless($expression)
<ide> {
<del> return "<?php if ( ! $expression): ?>";
<add> return "<?php if (! $expression): ?>";
<ide> }
<ide>
<ide> /**
<ide><path>tests/View/ViewBladeCompilerTest.php
<ide> public function testUnlessStatementsAreCompiled()
<ide> $string = '@unless (name(foo(bar)))
<ide> breeze
<ide> @endunless';
<del> $expected = '<?php if ( ! (name(foo(bar)))): ?>
<add> $expected = '<?php if (! (name(foo(bar)))): ?>
<ide> breeze
<ide> <?php endif; ?>';
<ide> $this->assertEquals($expected, $compiler->compileString($string));
| 2
|
Ruby
|
Ruby
|
prevent negative ids in output of #inspect
|
5bbce6fccc541b9941628cda4eda1f84c5a909ad
|
<ide><path>actionview/lib/action_view/template/resolver.rb
<ide> def initialize
<ide> end
<ide>
<ide> def inspect
<del> "#<#{self.class.name}:0x#{(object_id << 1).to_s(16)} keys=#{@data.size} queries=#{@query_cache.size}>"
<add> "#{to_s[0..-2]} keys=#{@data.size} queries=#{@query_cache.size}>"
<ide> end
<ide>
<ide> # Cache the templates returned by the block
| 1
|
Text
|
Text
|
add quotes around the brew invocation
|
953a1bf20b6944ef14bd9f6e5f5bc37ed191eba8
|
<ide><path>docs/Shell-Completion.md
<ide> To make Homebrew's completions available in `bash`, you must source the definiti
<ide>
<ide> ```sh
<ide> if type brew &>/dev/null; then
<del> HOMEBREW_PREFIX=$(brew --prefix)
<add> HOMEBREW_PREFIX="$(brew --prefix)"
<ide> if [[ -r "${HOMEBREW_PREFIX}/etc/profile.d/bash_completion.sh" ]]; then
<ide> source "${HOMEBREW_PREFIX}/etc/profile.d/bash_completion.sh"
<ide> else
| 1
|
Ruby
|
Ruby
|
allow methods outside the model
|
9f0f7ec2223ae60ee6e2eab5bc9f036a480e9f81
|
<ide><path>actionview/lib/action_view/helpers/form_helper.rb
<ide> def apply_form_for_options!(record, object, options) #:nodoc:
<ide> # <input type="text" name="post[title]" value="<the title of the post>">
<ide> # </form>
<ide> #
<add> # # Though the fields don't have to correspond to model attributes:
<add> # <%= form_with model: Cat.new do |form| %>
<add> # <%= form.text_field :cats_dont_have_gills %>
<add> # <%= form.text_field :but_in_forms_they_can %>
<add> # <% end %>
<add> # # =>
<add> # <form action="/cats" method="post" data-remote="true">
<add> # <input type="text" name="cat[cats_dont_have_gills]">
<add> # <input type="text" name="cat[but_in_forms_they_can]">
<add> # </form>
<add> #
<ide> # The parameters in the forms are accessible in controllers according to
<ide> # their name nesting. So inputs named +title+ and <tt>post[title]</tt> are
<ide> # accessible as <tt>params[:title]</tt> and <tt>params[:post][:title]</tt>
<ide> def apply_form_for_options!(record, object, options) #:nodoc:
<ide> # form_with(**options.merge(builder: LabellingFormBuilder), &block)
<ide> # end
<ide> def form_with(model: nil, scope: nil, url: nil, format: nil, **options)
<add> options[:allow_method_names_outside_object] = true
<ide> options[:skip_default_ids] = true
<ide>
<ide> if model
<ide> def to_model
<ide> def initialize(object_name, object, template, options)
<ide> @nested_child_index = {}
<ide> @object_name, @object, @template, @options = object_name, object, template, options
<del> @default_options = @options ? @options.slice(:index, :namespace, :skip_default_ids) : {}
<add> @default_options = @options ? @options.slice(:index, :namespace, :skip_default_ids, :allow_method_names_outside_object) : {}
<ide>
<ide> convert_to_legacy_options(@options)
<ide>
<ide><path>actionview/lib/action_view/helpers/tags/base.rb
<ide> def initialize(object_name, method_name, template_object, options = {})
<ide> @object_name.sub!(/\[\]$/, "") || @object_name.sub!(/\[\]\]$/, "]")
<ide> @object = retrieve_object(options.delete(:object))
<ide> @skip_default_ids = options.delete(:skip_default_ids)
<add> @allow_method_names_outside_object = options.delete(:allow_method_names_outside_object)
<ide> @options = options
<ide> @auto_index = Regexp.last_match ? retrieve_autoindex(Regexp.last_match.pre_match) : nil
<ide> end
<ide> def render
<ide> private
<ide>
<ide> def value(object)
<del> object.public_send @method_name if object
<add> if @allow_method_names_outside_object
<add> object.public_send @method_name if object && object.respond_to?(@method_name)
<add> else
<add> object.public_send @method_name if object
<add> end
<ide> end
<ide>
<ide> def value_before_type_cast(object)
<ide> def add_default_name_and_id(options)
<ide>
<ide> def tag_name(multiple = false, index = nil)
<ide> # a little duplication to construct less strings
<del> if index
<add> case
<add> when @object_name.empty?
<add> "#{sanitized_method_name}#{"[]" if multiple}"
<add> when index
<ide> "#{@object_name}[#{index}][#{sanitized_method_name}]#{"[]" if multiple}"
<ide> else
<ide> "#{@object_name}[#{sanitized_method_name}]#{"[]" if multiple}"
<ide> def tag_name(multiple = false, index = nil)
<ide>
<ide> def tag_id(index = nil)
<ide> # a little duplication to construct less strings
<del> if index
<add> case
<add> when @object_name.empty?
<add> sanitized_method_name.dup
<add> when index
<ide> "#{sanitized_object_name}_#{index}_#{sanitized_method_name}"
<ide> else
<ide> "#{sanitized_object_name}_#{sanitized_method_name}"
<ide><path>actionview/test/template/form_helper/form_with_test.rb
<ide> def test_form_with
<ide> assert_dom_equal expected, output_buffer
<ide> end
<ide>
<add> def test_form_with_only_url_on_create
<add> form_with(url: "/posts") do |f|
<add> concat f.label :title, "Label me"
<add> concat f.text_field :title
<add> end
<add>
<add> expected = whole_form("/posts") do
<add> '<label for="title">Label me</label>' +
<add> '<input type="text" name="title">'
<add> end
<add>
<add> assert_dom_equal expected, output_buffer
<add> end
<add>
<add> def test_form_with_only_url_on_update
<add> form_with(url: "/posts/123") do |f|
<add> concat f.label :title, 'Label me'
<add> concat f.text_field :title
<add> end
<add>
<add> expected = whole_form("/posts/123") do
<add> '<label for="title">Label me</label>' +
<add> '<input type="text" name="title">'
<add> end
<add>
<add> assert_dom_equal expected, output_buffer
<add> end
<add>
<add> def test_form_with_general_attributes
<add> form_with(url: "/posts/123") do |f|
<add> concat f.text_field :no_model_to_back_this_badboy
<add> end
<add>
<add> expected = whole_form("/posts/123") do
<add> '<input type="text" name="no_model_to_back_this_badboy">'
<add> end
<add>
<add> assert_dom_equal expected, output_buffer
<add> end
<add>
<add> def test_form_with_attribute_not_on_model
<add> form_with(model: @post) do |f|
<add> concat f.text_field :this_dont_exist_on_post
<add> end
<add>
<add> expected = whole_form("/posts/123", method: :patch) do
<add> '<input type="text" name="post[this_dont_exist_on_post]">'
<add> end
<add>
<add> assert_dom_equal expected, output_buffer
<add> end
<add>
<add> def test_form_with_doesnt_call_private_or_protected_properties_on_form_object_skipping_value
<add> obj = Class.new do
<add> private def private_property
<add> "That would be great."
<add> end
<add>
<add> protected def protected_property
<add> "I believe you have my stapler."
<add> end
<add> end.new
<add>
<add> form_with(model: obj, scope: "other_name", url: "/", id: "edit-other-name") do |f|
<add> assert_dom_equal '<input type="hidden" name="other_name[private_property]">', f.hidden_field(:private_property)
<add> assert_dom_equal '<input type="hidden" name="other_name[protected_property]">', f.hidden_field(:protected_property)
<add> end
<add> end
<add>
<ide> def test_form_with_with_collection_radio_buttons
<ide> post = Post.new
<ide> def post.active; false; end
<ide> def test_form_with_with_symbol_scope
<ide> assert_dom_equal expected, output_buffer
<ide> end
<ide>
<del> def test_form_tags_do_not_call_private_properties_on_form_object
<del> obj = Class.new do
<del> private def private_property
<del> raise "This method should not be called."
<del> end
<del> end.new
<del>
<del> form_with(model: obj, scope: "other_name", url: "/", id: "edit-other-name") do |f|
<del> assert_raise(NoMethodError) { f.hidden_field(:private_property) }
<del> end
<del> end
<del>
<ide> def test_form_with_with_method_as_part_of_html_options
<ide> form_with(model: @post, url: "/", id: "create-post", html: { method: :delete }) do |f|
<ide> concat f.text_field(:title)
| 3
|
Python
|
Python
|
attempt a port from
|
1eb7cc3017a6def34fb448781578888764d1e659
|
<ide><path>spacy/lang/ga/__init__.py
<add># coding: utf8
<add>from __future__ import unicode_literals
<add>
<add>from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS
<add>from .stop_words import STOP_WORDS
<add>
<add>from ..tokenizer_exceptions import BASE_EXCEPTIONS
<add>from ...language import Language
<add>from ...attrs import LANG
<add>from ...util import update_exc
<add>
<add>
<add>class Irish(Language):
<add> lang = 'nb'
<add>
<add> class Defaults(Language.Defaults):
<add> lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
<add> lex_attr_getters[LANG] = lambda text: 'ga'
<add>
<add> tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS)
<add> stop_words = set(STOP_WORDS)
<add>
<add>
<add>__all__ = ['Irish']
<ide><path>spacy/lang/ga/stop_words.py
<add># encoding: utf8
<add>from __future__ import unicode_literals
<add>
<add>
<add>STOP_WORDS = set("""
<add>a ach ag agus an aon ar arna as
<add>
<add>ba beirt bhúr
<add>
<add>caoga ceathair ceathrar chomh chuig chun cois céad cúig cúigear
<add>
<add>daichead dar de deich deichniúr den dhá do don dtí dá dár dó
<add>
<add>faoi faoin faoina faoinár fara fiche
<add>
<add>gach gan go gur
<add>
<add>haon hocht
<add>
<add>i iad idir in ina ins inár is
<add>
<add>le leis lena lenár
<add>
<add>mar mo muid mé
<add>
<add>na nach naoi naonúr ná ní níor nó nócha
<add>
<add>ocht ochtar ochtó os
<add>
<add>roimh
<add>
<add>sa seacht seachtar seachtó seasca seisear siad sibh sinn sna sé sí
<add>
<add>tar thar thú triúr trí trína trínár tríocha tú
<add>
<add>um
<add>
<add>ár
<add>
<add>é éis
<add>
<add>í
<add>
<add>ó ón óna ónár
<add>""".split())
<ide><path>spacy/lang/ga/tokenizer_exceptions.py
<add># encoding: utf8
<add>from __future__ import unicode_literals
<add>
<add>from ..symbols import ORTH, LEMMA, NORM
<add>
<add>
<add>_exc = {
<add> "'acha'n": [
<add> {ORTH: "'ach", LEMMA: "gach", NORM: "gach"},
<add> {ORTH: "a'n", LEMMA: "aon", NORM: "aon"}],
<add>
<add> "dem'": [
<add> {ORTH: "de", LEMMA: "de", NORM: "de"},
<add> {ORTH: "m'", LEMMA: "mo", NORM: "mo"}],
<add>
<add> "ded'": [
<add> {ORTH: "de", LEMMA: "de", NORM: "de"},
<add> {ORTH: "d'", LEMMA: "do", NORM: "do"}],
<add>
<add> "lem'": [
<add> {ORTH: "le", LEMMA: "le", NORM: "le"},
<add> {ORTH: "m'", LEMMA: "mo", NORM: "mo"}],
<add>
<add> "led'": [
<add> {ORTH: "le", LEMMA: "le", NORM: "le"},
<add> {ORTH: "d'", LEMMA: "mo", NORM: "do"}],
<add>
<add> "a.C.n.": [
<add> {ORTH: "a.", LEMMA: "ante"},
<add> {ORTH: "C.", LEMMA: "Christum"},
<add> {ORTH: "n.", LEMMA: "natum"}],
<add>
<add> "m.sh.": [
<add> {ORTH: "m.", LEMMA: "mar"},
<add> {ORTH: "sh.", LEMMA: "sampla"}],
<add>
<add> "M.F.": [
<add> {ORTH: "M.", LEMMA: "Meán"},
<add> {ORTH: "F.", LEMMA: "Fómhar"}],
<add>
<add> "M.Fómh.": [
<add> {ORTH: "M.", LEMMA: "Meán"},
<add> {ORTH: "Fómh.", LEMMA: "Fómhar"}],
<add>
<add> "R.C.": [
<add> {ORTH: "Rr.", LEMMA: "roimh"},
<add> {ORTH: "C.", LEMMA: "Críost"}],
<add>
<add> "r.Ch.": [
<add> {ORTH: "r.", LEMMA: "roimh"},
<add> {ORTH: "Ch.", LEMMA: "Críost"}],
<add>
<add> "r.Chr.": [
<add> {ORTH: "r.", LEMMA: "roimh"},
<add> {ORTH: "Chr.", LEMMA: "Críost"}],
<add>
<add> "R.Ch.": [
<add> {ORTH: "R.", LEMMA: "roimh"},
<add> {ORTH: "Ch.", LEMMA: "Críost"}],
<add>
<add> "R.Chr.": [
<add> {ORTH: "R.", LEMMA: "roimh"},
<add> {ORTH: "Chr.", LEMMA: "Críost"}],
<add>
<add> "⁊rl.": [
<add> {ORTH: "⁊", LEMMA: "agus"},
<add> {ORTH: "rl.", LEMMA: "araile"}],
<add>
<add> "srl.": [
<add> {ORTH: "s", LEMMA: "agus"},
<add> {ORTH: "rl.", LEMMA: "araile"}],
<add>
<add>}
<add>
<add>for exc_data in [
<add> {ORTH: "'gus", LEMMA: "agus", NORM: "agus"},
<add> {ORTH: "'ach", LEMMA: "gach", NORM: "gach"},
<add> {ORTH: "ao'", LEMMA: "aon", NORM: "aon"},
<add> {ORTH: "'niar", LEMMA: "aniar", NORM: "aniar"},
<add> {ORTH: "'níos", LEMMA: "aníos", NORM: "aníos"},
<add> {ORTH: "'ndiu", LEMMA: "inniu", NORM: "inniu"},
<add> {ORTH: "'nocht", LEMMA: "anocht", NORM: "anocht"},
<add> {ORTH: "m'", LEMMA: "mo"},,
<add> {ORTH: "Aib.", LEMMA: "Aibreán"},
<add> {ORTH: "Ath.", LEMMA: "athair"},
<add> {ORTH: "Beal.", LEMMA: "Bealtaine"},
<add> {ORTH: "Co.", LEMMA: "contae"},
<add> {ORTH: "Ean.", LEMMA: "Eanáir"},
<add> {ORTH: "Feab.", LEMMA: "Feabhra"},
<add> {ORTH: "gCo.", LEMMA: "contae"},
<add> {ORTH: ".i.", LEMMA: "eadhon"},
<add> {ORTH: "lch.", LEMMA: "leathanach"},
<add> {ORTH: "Lch.", LEMMA: "leathanach"},
<add> {ORTH: "lgh.", LEMMA: "leathanach"},
<add> {ORTH: "Lgh.", LEMMA: "leathanach"},
<add> {ORTH: "Lún.", LEMMA: "Lúnasa"},
<add> {ORTH: "Már.", LEMMA: "Márta"},
<add> {ORTH: "Meith.", LEMMA: "Meitheamh"},
<add> {ORTH: "Noll.", LEMMA: "Nollaig"},
<add> {ORTH: "Samh.", LEMMA: "Samhain"},
<add> {ORTH: "tAth.", LEMMA: "athair"},
<add> {ORTH: "tUas.", LEMMA: "Uasal"},
<add> {ORTH: "teo.", LEMMA: "teoranta"},
<add> {ORTH: "Teo.", LEMMA: "teoranta"},
<add> {ORTH: "Uas.", LEMMA: "Uasal"},
<add> {ORTH: "uimh.", LEMMA: "uimhir"},
<add> {ORTH: "Uimh.", LEMMA: "uimhir"}]:
<add> _exc[exc_data[ORTH]] = [dict(exc_data)],
<add>
<add>for orth in [
<add> "d'"]:
<add> _exc[orth] = [{ORTH: orth}]
<add>
<add>
<add>TOKENIZER_EXCEPTIONS = dict(_exc)
| 3
|
Python
|
Python
|
change duration_ms to duration"
|
1d4c7993a9c6dcacaca5074a80b1043e977c43fb
|
<ide><path>tools/test.py
<ide> def HasRun(self, output):
<ide> (duration.seconds + duration.days * 24 * 3600) * 10**6) / 10**6
<ide>
<ide> logger.info(' ---')
<del> logger.info(' duration: %d.%ds' % (total_seconds, duration.microseconds / 1000))
<add> logger.info(' duration_ms: %d.%d' % (total_seconds, duration.microseconds / 1000))
<ide> logger.info(' ...')
<ide>
<ide> def Done(self):
| 1
|
Text
|
Text
|
release notes for 0.10.0 chicken-hands
|
30e5f6274a7148c139aeafdb8098f63cbdbd5d33
|
<ide><path>CHANGELOG.md
<add>- The Latest Stable Release: <a href="#0.9.19">0.9.19 canine-psychokinesis</a>
<add>- The Latest Unstable Release: <a href="#0.10.0">0.10.0 chicken-hands</a>
<add>
<add>
<ide> <a name="0.10.0"><a/>
<del># 0.10.0 chicken-hands (in-progress) #
<add># 0.10.0 chicken-hands (2011-09-02) #
<add>
<add>## Features:
<add>
<add>- complete rewrite of the Scope implementation with several API and semantic changes. Please see:
<add> - [angular.scope API docs](http://docs-next.angularjs.org/#!/api/angular.scope)
<add> - [scopes dev guide article](http://docs-next.angularjs.org/#!/guide/dev_guide.scopes)
<add> - [scope.js source file](https://github.com/angular/angular.js/blob/master/src/Scope.js)
<add> - breaking changes section of this changelog
<add>- added event system to scopes (see [$on], [$emit] and [$broadcast])
<add>- added i18n and l10n support for date, currency and number filters see [i18n] docs for more info
<add>- added localizable [ng:pluralize] widget
<add>- added [ng:cloak] directive for hiding uncompiled templates
<add>
<add>
<add>## Bug Fix:
<add>
<add>- make [ng:class] friendly towards other code adding/removing classes
<add> ([commit](https://github.com/angular/angular.js/commit/2a8fe56997fddbad673748ce02abf649a709c4ca))
<add>- several [jqLite] bugfixes and improvements
<add>- [ng:href], [ng:src] and friends now work properly when no expression is present in the attribute
<add> value.
<add> (Issue [#534](https://github.com/angular/angular.js/issues/534))
<add>- expose missing [lowercase], [uppercase] and [isDate] APIs.
<add>
<add>
<add>## Docs:
<add>
<add>- many (but not all just yet) api docs were proof-read and improved
<add>
<add>
<add>## Breaking Changes:
<add>
<add>- many scope related changes:
<add> - $onEval is no more (use $watch with a fn as the only param if you really miss it)
<add> - $eval without params doesn't trigger model mutation observations (use $apply/$digest instead)
<add> - $digest propagates through the scope tree automatically (this is the desired behavior anyway)
<add> - $watch various API changes
<add> - scope is now the first argument passed into the $watch listener
<add> - `this` in the $watch listener is undefined instead of current scope
<add> - objects and arrays are watched and compared by equality and not just identity
<add> - the initial execution of the $watch listener now executes asynchronously with respect to the
<add> code registering it via $watch
<add> - exceptionHandler argument is no more
<add> - initRun argument is no more
<add> - angular.scope does not create child scopes by taking parent as the first argument - use $new
<add> instead
<add> - scope.$set and scope.$get were removed, use direct property assignment instead or $eval
<add>- $route.onChange was removed and replaced with $beforeRouteChange, $afterRouteChange and
<add> $routeUpdate events that can be used together with the new $routeParams service
<add>- `angular.equals()` now uses `===` instead of `==` when comparing primitives
<ide>
<ide>
<ide>
<ide> with the `$route` service
<ide>
<ide>
<ide>
<add>[lowercase]: http://docs.angularjs.org/#!/api/angular.lowercase
<add>[uppercase]: http://docs.angularjs.org/#!/api/angular.uppercase
<add>[isDate]: http://docs.angularjs.org/#!/api/angular.isDate
<ide> [scope]: http://docs.angularjs.org/#!/api/angular.scope
<ide> [compile]: http://docs.angularjs.org/#!/api/angular.compile
<ide> [element]: http://docs.angularjs.org/#!/api/angular.element
<ide> with the `$route` service
<ide> [ng:readonly]: http://docs.angularjs.org/#!/api/angular.directive.ng:readonly
<ide> [ng:show]: http://docs.angularjs.org/#!/api/angular.directive.ng:show
<ide> [ng:hide]: http://docs.angularjs.org/#!/api/angular.directive.ng:hide
<add>[ng:class]: http://docs.angularjs.org/#!/api/angular.directive.ng:class
<add>[ng:src]: http://docs.angularjs.org/#!/api/angular.directive.ng:src
<add>[ng:href]: http://docs.angularjs.org/#!/api/angular.directive.ng:href
<ide> [$defer]: http://docs.angularjs.org/#!/api/angular.service.$defer
<ide> [$cookies]: http://docs.angularjs.org/#!/api/angular.service.$cookies
<ide> [$xhr]: http://docs.angularjs.org/#!/api/angular.service.$xhr
<ide> with the `$route` service
<ide> [jqLite]: http://docs.angularjs.org/#!/api/angular.element
<ide> [angular.version]: http://docs.angularjs.org/#!/api/angular.version
<ide> [Jstd Scenario Adapter]: https://github.com/angular/angular.js/blob/master/src/jstd-scenario-adapter/Adapter.js
<add>[i18n]: http://docs-next.angularjs.org/#!/guide/dev_guide.i18n
<add>[ng:pluralize]: http://docs-next.angularjs.org/#!/api/angular.widget.ng:pluralize
<add>[ng:cloak]: http://docs-next.angularjs.org/#!/api/angular.directive.ng:cloak
<add>[$on]: http://docs-next.angularjs.org/#!/api/angular.scope.$on
<add>[$emit]: http://docs-next.angularjs.org/#!/api/angular.scope.$emit
<add>[$broadcast]: http://docs-next.angularjs.org/#!/api/angular.scope.$broadcast
| 1
|
Javascript
|
Javascript
|
add support for title in schema
|
c76f7ec2f4bcca5e3361ff77fc022075368dad91
|
<ide><path>test/Schemas.lint.js
<ide> describe("Schemas", () => {
<ide> const allowedProperties = [
<ide> "definitions",
<ide> "$ref",
<del> "id",
<add> "$id",
<add> "title",
<ide> "items",
<ide> "properties",
<ide> "additionalProperties",
<ide><path>tooling/format-schemas.js
<ide> const PROPERTIES = [
<ide> "$ref",
<ide> "definitions",
<ide>
<add> "$id",
<ide> "id",
<add> "title",
<ide> "description",
<ide> "type",
<ide>
| 2
|
Text
|
Text
|
update devtools post to note beta 2
|
3ed9581adbdc2bcf1bca0800050e7907624a83aa
|
<ide><path>docs/_posts/2015-08-03-new-react-devtools-beta.md
<ide> Let us know what issues you run into
<ide> [on GitHub](https://github.com/facebook/react-devtools/issues), and check out
<ide> [the README](https://github.com/facebook/react-devtools/tree/devtools-next)
<ide> for more info.
<add>
<add>## Update
<add>*August 12, 2015*
<add>
<add>A second beta is out, with a number of bugfixes. It is also listed on the
<add>[releases page](https://github.com/facebook/react-devtools/releases).
| 1
|
Text
|
Text
|
add missing link references
|
6c76de13c5849fb8273272490c55460e33c0c556
|
<ide><path>doc/api/assert.md
<ide> Due to the confusing notation, it is recommended not to use a string as the
<ide> second argument. This might lead to difficult-to-spot errors.
<ide>
<ide> [`Error.captureStackTrace`]: errors.html#errors_error_capturestacktrace_targetobject_constructoropt
<add>[`Error`]: errors.html#errors_class_error
<ide> [`Map`]: https://developer.mozilla.org/en/docs/Web/JavaScript/Reference/Global_Objects/Map
<ide> [`Object.is()`]: https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Object/is
<ide> [`RegExp`]: https://developer.mozilla.org/en-US/docs/Web/JavaScript/Guide/Regular_Expressions
<ide><path>doc/api/errors.md
<ide> Creation of a [`zlib`][] object failed due to incorrect configuration.
<ide> [`hash.digest()`]: crypto.html#crypto_hash_digest_encoding
<ide> [`hash.update()`]: crypto.html#crypto_hash_update_data_inputencoding
<ide> [`readable._read()`]: stream.html#stream_readable_read_size_1
<add>[`server.close()`]: net.html#net_server_close_callback
<ide> [`sign.sign()`]: crypto.html#crypto_sign_sign_privatekey_outputformat
<ide> [`stream.pipe()`]: stream.html#stream_readable_pipe_destination_options
<ide> [`stream.push()`]: stream.html#stream_readable_push_chunk_encoding
| 2
|
Javascript
|
Javascript
|
reduce the number of db calls for getsessionuser
|
892e6862ed4ffe4a9d01fec0045475b24e9be1da
|
<ide><path>api-server/common/models/user.js
<ide> export default function(User) {
<ide> });
<ide>
<ide> User.prototype.getPoints$ = function getPoints$() {
<add> if (
<add> Array.isArray(this.progressTimestamps) &&
<add> this.progressTimestamps.length
<add> ) {
<add> return Observable.of(this.progressTimestamps);
<add> }
<ide> const id = this.getId();
<ide> const filter = {
<ide> where: { id },
<ide> export default function(User) {
<ide> });
<ide> };
<ide> User.prototype.getCompletedChallenges$ = function getCompletedChallenges$() {
<add> if (
<add> Array.isArray(this.completedChallenges) &&
<add> this.completedChallenges.length
<add> ) {
<add> return Observable.of(this.completedChallenges);
<add> }
<ide> const id = this.getId();
<ide> const filter = {
<ide> where: { id },
<ide><path>api-server/server/boot_tests/challenge.test.js
<ide> import {
<ide> } from './fixtures';
<ide>
<ide> describe('boot/challenge', () => {
<del> xdescribe('backendChallengeCompleted');
<add> xdescribe('backendChallengeCompleted', () => {});
<ide>
<ide> describe('buildUserUpdate', () => {
<ide> it('returns an Object with a nested "completedChallenges" property', () => {
<ide> describe('boot/challenge', () => {
<ide> });
<ide> });
<ide>
<del> xdescribe('modernChallengeCompleted');
<add> xdescribe('modernChallengeCompleted', () => {});
<ide>
<del> xdescribe('projectCompleted');
<add> xdescribe('projectCompleted', () => {});
<ide>
<ide> describe('redirectToCurrentChallenge', () => {
<ide> const mockHomeLocation = 'https://www.example.com';
<ide><path>api-server/server/utils/user-stats.js
<ide> export function getUserById(id, User = loopback.getModelByType('User')) {
<ide> if (err || isEmpty(instance)) {
<ide> return reject(err || 'No user instance found');
<ide> }
<del> instance.points =
<del> (instance.progressTimestamps && instance.progressTimestamps.length) ||
<del> 1;
<add>
<ide> let completedChallengeCount = 0;
<ide> let completedProjectCount = 0;
<ide> if ('completedChallenges' in instance) {
<ide> export function getUserById(id, User = loopback.getModelByType('User')) {
<ide> }
<ide> });
<ide> }
<add>
<ide> instance.completedChallengeCount = completedChallengeCount;
<ide> instance.completedProjectCount = completedProjectCount;
<ide> instance.completedCertCount = getCompletedCertCount(instance);
<ide> instance.completedLegacyCertCount = getLegacyCertCount(instance);
<del> instance.completedChallenges = [];
<del> delete instance.progressTimestamps;
<add> instance.points =
<add> (instance.progressTimestamps && instance.progressTimestamps.length) ||
<add> 1;
<ide> return resolve(instance);
<ide> })
<ide> );
<ide><path>api-server/server/utils/user-stats.test.js
<ide> describe('user stats', () => {
<ide> .then(user => {
<ide> expect(user).toEqual(mockUser);
<ide>
<del> // fields added or removed by getUserById
<del> expect(user).not.toHaveProperty('progressTimestamps');
<add> expect(user).toHaveProperty('progressTimestamps');
<ide> expect(user).toHaveProperty('completedChallengeCount');
<ide> expect(user).toHaveProperty('completedProjectCount');
<ide> expect(user).toHaveProperty('completedCertCount');
<ide> expect(user).toHaveProperty('completedLegacyCertCount');
<del> expect(user.completedChallenges).toEqual([]);
<add> expect(user).toHaveProperty('completedChallenges');
<ide> })
<ide> .then(done)
<ide> .catch(done);
| 4
|
Javascript
|
Javascript
|
add error logs in __debug__
|
96a7adf49e20197f802813224362470c66a38d20
|
<ide><path>src/devtools/store.js
<ide> export default class Store extends EventEmitter {
<ide>
<ide> i = i + 3;
<ide>
<add> if (__DEBUG__) {
<add> if (this._idToElement.has(id)) {
<add> console.error(
<add> 'Store already contains fiber ' + id + '. This is a bug.'
<add> );
<add> }
<add> }
<add>
<ide> if (type === ElementTypeRoot) {
<ide> if (__DEBUG__) {
<ide> debug('Add', `new root fiber ${id}`);
<ide> export default class Store extends EventEmitter {
<ide> case TREE_OPERATION_REMOVE:
<ide> id = ((operations[i + 1]: any): number);
<ide>
<add> if (__DEBUG__) {
<add> if (!this._idToElement.has(id)) {
<add> console.error(
<add> 'Store does not contain fiber ' + id + '. This is a bug.'
<add> );
<add> }
<add> }
<add>
<ide> i = i + 2;
<ide>
<ide> element = ((this._idToElement.get(id): any): Element);
| 1
|
Java
|
Java
|
update javadoc on uricomponentsbuilder.query
|
dacbf4cb34d3b2cf0747daa16903c91d696655a4
|
<ide><path>spring-web/src/main/java/org/springframework/web/util/UriComponentsBuilder.java
<ide> public UriComponentsBuilder pathSegment(String... pathSegments) throws IllegalAr
<ide> * be parsed unambiguously. Such values should be substituted for URI
<ide> * variables to enable correct parsing:
<ide> * <pre class="code">
<del> * String uriString = "/hotels/42?filter={value}";
<del> * UriComponentsBuilder.fromUriString(uriString).buildAndExpand("hot&cold");
<add> * UriComponentsBuilder.fromUriString("/hotels/42")
<add> * .query("filter={value}")
<add> * .buildAndExpand("hot&cold");
<ide> * </pre>
<ide> * @param query the query string
<ide> * @return this UriComponentsBuilder
| 1
|
Ruby
|
Ruby
|
remove unused method
|
3c100cfe8ed5a875b0bbdc8fa4e8f2b0cbf78676
|
<ide><path>activerecord/test/models/developer.rb
<ide> def find_least_recent
<ide> def log=(message)
<ide> audit_logs.build :message => message
<ide> end
<del>
<del> def self.all_johns
<del> self.with_exclusive_scope :find => where(:name => 'John') do
<del> self.all
<del> end
<del> end
<ide> end
<ide>
<ide> class AuditLog < ActiveRecord::Base
| 1
|
Ruby
|
Ruby
|
apply suggestions from code review
|
bdb64aa178c1531661d74f58cd2435c30e029272
|
<ide><path>Library/Homebrew/formula_installer.rb
<ide> def puts_requirement_messages
<ide> end
<ide>
<ide> def forbidden_license_check
<del> forbidden_licenses = (Homebrew::EnvConfig.forbidden_licenses || "").split(" ")
<add> forbidden_licenses = Homebrew::EnvConfig.forbidden_licenses.to_s.split(" ")
<ide> return if forbidden_licenses.blank?
<ide>
<ide> if forbidden_licenses.include? formula.license
| 1
|
Ruby
|
Ruby
|
stop test resetting global state
|
e2b6919223cd24bc8e64b4b5230733c0d8f8e6fc
|
<ide><path>activerecord/test/cases/yaml_serialization_test.rb
<ide> class YamlSerializationTest < ActiveRecord::TestCase
<ide> fixtures :topics
<ide>
<ide> def test_to_yaml_with_time_with_zone_should_not_raise_exception
<add> tz = Time.zone
<ide> Time.zone = ActiveSupport::TimeZone["Pacific Time (US & Canada)"]
<ide> ActiveRecord::Base.time_zone_aware_attributes = true
<add>
<ide> topic = Topic.new(:written_on => DateTime.now)
<ide> assert_nothing_raised { topic.to_yaml }
<add>
<add> ensure
<add> Time.zone = tz
<add> ActiveRecord::Base.time_zone_aware_attributes = false
<ide> end
<ide>
<ide> def test_roundtrip
| 1
|
Javascript
|
Javascript
|
use a domain to catch async errors safely
|
c0721bcd66829356950b58cc532d6e3d8bbfc641
|
<ide><path>lib/repl.js
<ide> var fs = require('fs');
<ide> var rl = require('readline');
<ide> var Console = require('console').Console;
<ide> var EventEmitter = require('events').EventEmitter;
<add>var domain = require('domain');
<ide>
<ide> // If obj.hasOwnProperty has been overridden, then calling
<ide> // obj.hasOwnProperty(prop) will break.
<ide> function REPLServer(prompt, stream, eval_, useGlobal, ignoreUndefined) {
<ide>
<ide> EventEmitter.call(this);
<ide>
<del> var options, input, output;
<add> var options, input, output, dom;
<ide> if (typeof prompt == 'object') {
<ide> // an options object was given
<ide> options = prompt;
<ide> function REPLServer(prompt, stream, eval_, useGlobal, ignoreUndefined) {
<ide> useGlobal = options.useGlobal;
<ide> ignoreUndefined = options.ignoreUndefined;
<ide> prompt = options.prompt;
<add> dom = options.domain;
<ide> } else if (typeof prompt != 'string') {
<ide> throw new Error('An options Object, or a prompt String are required');
<ide> } else {
<ide> function REPLServer(prompt, stream, eval_, useGlobal, ignoreUndefined) {
<ide>
<ide> var self = this;
<ide>
<add> self._domain = dom || domain.create();
<add>
<ide> self.useGlobal = !!useGlobal;
<ide> self.ignoreUndefined = !!ignoreUndefined;
<ide>
<del> self.eval = eval_ || function(code, context, file, cb) {
<add> eval_ = eval_ || defaultEval;
<add>
<add> function defaultEval(code, context, file, cb) {
<ide> var err, result;
<ide> try {
<ide> if (self.useGlobal) {
<ide> function REPLServer(prompt, stream, eval_, useGlobal, ignoreUndefined) {
<ide> } catch (e) {
<ide> err = e;
<ide> }
<del> if (err && process.domain) {
<add> if (err && process.domain && !isSyntaxError(err)) {
<ide> process.domain.emit('error', err);
<ide> process.domain.exit();
<ide> }
<ide> else {
<ide> cb(err, result);
<ide> }
<del> };
<add> }
<add>
<add> self.eval = self._domain.bind(eval_);
<add>
<add> self._domain.on('error', function(e) {
<add> self.outputStream.write((e.stack || e) + '\n');
<add> self.bufferedCommand = '';
<add> self.displayPrompt();
<add> });
<ide>
<ide> if (!input && !output) {
<ide> // legacy API, passing a 'stream'/'socket' option
<ide> function REPLServer(prompt, stream, eval_, useGlobal, ignoreUndefined) {
<ide> self.displayPrompt();
<ide> return;
<ide> } else if (e) {
<del> self.outputStream.write((e.stack || e) + '\n');
<add> self._domain.emit('error', e);
<ide> }
<ide>
<ide> // Clear buffer if no SyntaxErrors
<ide><path>test/simple/test-repl-domain.js
<ide> // OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
<ide> // USE OR OTHER DEALINGS IN THE SOFTWARE.
<ide>
<add>var assert = require('assert');
<add>var common = require('../common.js');
<add>
<ide> var util = require('util');
<ide> var repl = require('repl');
<ide>
<ide><path>test/simple/test-repl-options.js
<ide> assert.equal(r2.rli.terminal, false);
<ide> assert.equal(r2.useColors, true);
<ide> assert.equal(r2.useGlobal, true);
<ide> assert.equal(r2.ignoreUndefined, true);
<del>assert.equal(r2.eval, evaler);
<ide> assert.equal(r2.writer, writer);
<ide><path>test/simple/test-repl-timeout-throw.js
<add>// Copyright Joyent, Inc. and other Node contributors.
<add>//
<add>// Permission is hereby granted, free of charge, to any person obtaining a
<add>// copy of this software and associated documentation files (the
<add>// "Software"), to deal in the Software without restriction, including
<add>// without limitation the rights to use, copy, modify, merge, publish,
<add>// distribute, sublicense, and/or sell copies of the Software, and to permit
<add>// persons to whom the Software is furnished to do so, subject to the
<add>// following conditions:
<add>//
<add>// The above copyright notice and this permission notice shall be included
<add>// in all copies or substantial portions of the Software.
<add>//
<add>// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
<add>// OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
<add>// MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN
<add>// NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM,
<add>// DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
<add>// OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
<add>// USE OR OTHER DEALINGS IN THE SOFTWARE.
<add>
<add>var assert = require('assert');
<add>var common = require('../common.js');
<add>
<add>var spawn = require('child_process').spawn;
<add>
<add>var child = spawn(process.execPath, [ '-i' ], {
<add> stdio: [null, null, 2]
<add>});
<add>
<add>var stdout = '';
<add>child.stdout.setEncoding('utf8');
<add>child.stdout.on('data', function(c) {
<add> process.stdout.write(c);
<add> stdout += c;
<add>});
<add>
<add>child.stdin.write = function(original) { return function(c) {
<add> process.stderr.write(c);
<add> return original.call(child.stdin, c);
<add>}}(child.stdin.write);
<add>
<add>child.stdout.once('data', function() {
<add> child.stdin.write('var throws = 0;');
<add> child.stdin.write('process.on("exit",function(){console.log(throws)});');
<add> child.stdin.write('function thrower(){console.log("THROW",throws++);XXX};');
<add> child.stdin.write('setTimeout(thrower);""\n');
<add>
<add> setTimeout(fsTest, 50);
<add> function fsTest() {
<add> var f = JSON.stringify(__filename);
<add> child.stdin.write('fs.readFile(' + f + ', thrower);\n');
<add> setTimeout(eeTest, 50);
<add> }
<add>
<add> function eeTest() {
<add> child.stdin.write('setTimeout(function() {\n' +
<add> ' var events = require("events");\n' +
<add> ' var e = new events.EventEmitter;\n' +
<add> ' process.nextTick(function() {\n' +
<add> ' e.on("x", thrower);\n' +
<add> ' setTimeout(function() {\n' +
<add> ' e.emit("x");\n' +
<add> ' });\n' +
<add> ' });\n' +
<add> '});"";\n');
<add>
<add> setTimeout(child.stdin.end.bind(child.stdin), 50);
<add> }
<add>});
<add>
<add>child.on('exit', function(c) {
<add> assert(!c);
<add> // make sure we got 3 throws, in the end.
<add> var lastLine = stdout.trim().split(/\r?\n/).pop();
<add> assert.equal(lastLine, '> 3');
<add>});
| 4
|
Javascript
|
Javascript
|
add missing semicolon
|
debf96c742934543e210c48c7c11199aa614e381
|
<ide><path>d3.layout.js
<ide> d3.layout.chord = function() {
<ide> if (source.value || target.value) {
<ide> chords.push(source.value < target.value
<ide> ? {source: target, target: source}
<del> : {source: source, target: target})
<add> : {source: source, target: target});
<ide> }
<ide> }
<ide> }
<ide><path>src/layout/chord.js
<ide> d3.layout.chord = function() {
<ide> if (source.value || target.value) {
<ide> chords.push(source.value < target.value
<ide> ? {source: target, target: source}
<del> : {source: source, target: target})
<add> : {source: source, target: target});
<ide> }
<ide> }
<ide> }
| 2
|
Javascript
|
Javascript
|
clarify the comparator parameter
|
b9d2b30808db367af1c31fa0fa6cef8f1bc086d3
|
<ide><path>src/ng/filter/filter.js
<ide> * The final result is an array of those elements that the predicate returned true for.
<ide> *
<ide> * @param {function(actual, expected)|true|false} [comparator] Comparator which is used in
<del> * determining if the expected value (from the filter expression) and actual value (from
<del> * the object in the array) should be considered a match.
<add> * determining if values retrieved using `expression` (when it is not a function) should be
<add> * considered a match based on the the expected value (from the filter expression) and actual
<add> * value (from the object in the array).
<ide> *
<ide> * Can be one of:
<ide> *
| 1
|
PHP
|
PHP
|
fix cookie sessions
|
a6c638b91c250a38f42536ad7123a45a2d8a9529
|
<ide><path>src/Illuminate/Session/Middleware/StartSession.php
<ide> use Illuminate\Session\SessionManager;
<ide> use Illuminate\Session\SessionInterface;
<ide> use Symfony\Component\HttpFoundation\Cookie;
<add>use Illuminate\Session\CookieSessionHandler;
<ide> use Symfony\Component\HttpFoundation\Request;
<ide> use Symfony\Component\HttpFoundation\Response;
<del>use Illuminate\Contracts\Routing\Middleware as MiddlewareContract;
<ide> use Illuminate\Contracts\Routing\TerminableMiddleware;
<add>use Illuminate\Contracts\Routing\Middleware as MiddlewareContract;
<ide>
<ide> class StartSession implements MiddlewareContract, TerminableMiddleware {
<ide>
<ide> public function handle($request, Closure $next)
<ide> */
<ide> public function terminate($request, $response)
<ide> {
<del> if ($this->sessionConfigured())
<add> if ($this->sessionConfigured() && ! $this->usingCookieSessions())
<ide> {
<ide> $this->manager->driver()->save();
<ide> }
<ide> protected function configHitsLottery(array $config)
<ide> */
<ide> protected function addCookieToResponse(Response $response, SessionInterface $session)
<ide> {
<add> if ($this->usingCookieSessions())
<add> {
<add> $this->manager->driver()->save();
<add> }
<add>
<ide> if ($this->sessionIsPersistent($config = $this->manager->getSessionConfig()))
<ide> {
<ide> $response->headers->setCookie(new Cookie(
<ide> protected function sessionIsPersistent(array $config = null)
<ide> return ! in_array($config['driver'], array(null, 'array'));
<ide> }
<ide>
<add> /**
<add> * Determine if the session is using cookie sessions.
<add> *
<add> * @return bool
<add> */
<add> protected function usingCookieSessions()
<add> {
<add> if ( ! $this->sessionConfigured()) return false;
<add>
<add> return $this->manager->driver()->getHandler() instanceof CookieSessionHandler;
<add> }
<add>
<ide> }
| 1
|
PHP
|
PHP
|
fix cs error
|
c3508f3b9c3c601df2fc63e3e6cf25a1b57a3f42
|
<ide><path>src/Error/BaseErrorHandler.php
<ide> namespace Cake\Error;
<ide>
<ide> use Cake\Core\Configure;
<add>use Cake\Error\PHP7ErrorException;
<ide> use Cake\Log\Log;
<ide> use Cake\Routing\Router;
<del>use Exception;
<ide> use Error;
<del>use Cake\Error\PHP7ErrorException;
<add>use Exception;
<ide>
<ide> /**
<ide> * Base error handler that provides logic common to the CLI + web
| 1
|
Javascript
|
Javascript
|
keep stdio after exit
|
9e84a26cb3d390a6ce051b75c892094fff233f92
|
<ide><path>lib/internal/worker.js
<ide> class Worker extends EventEmitter {
<ide> this[kPublicPort] = null;
<ide>
<ide> const { stdout, stderr } = this[kParentSideStdio];
<del> this[kParentSideStdio] = null;
<ide>
<ide> if (!stdout._readableState.ended) {
<ide> debug(`[${threadId}] explicitly closes stdout for ${this.threadId}`);
<ide> class Worker extends EventEmitter {
<ide> }
<ide>
<ide> get stdin() {
<del> if (this[kParentSideStdio] === null) return null;
<del>
<ide> return this[kParentSideStdio].stdin;
<ide> }
<ide>
<ide> get stdout() {
<del> if (this[kParentSideStdio] === null) return null;
<del>
<ide> return this[kParentSideStdio].stdout;
<ide> }
<ide>
<ide> get stderr() {
<del> if (this[kParentSideStdio] === null) return null;
<del>
<ide> return this[kParentSideStdio].stderr;
<ide> }
<ide> }
<ide><path>test/parallel/test-worker-safe-getters.js
<ide> if (isMainThread) {
<ide> stderr: true
<ide> });
<ide>
<add> const { stdin, stdout, stderr } = w;
<add>
<ide> w.on('exit', common.mustCall((code) => {
<ide> assert.strictEqual(code, 0);
<ide>
<ide> if (isMainThread) {
<ide> w.ref();
<ide> w.unref();
<ide>
<del> // Although not browser specific, probably wise to
<del> // make sure the stream getters don't throw either.
<del> w.stdin;
<del> w.stdout;
<del> w.stderr;
<del>
<ide> // Sanity check.
<ide> assert.strictEqual(w.threadId, -1);
<del> assert.strictEqual(w.stdin, null);
<del> assert.strictEqual(w.stdout, null);
<del> assert.strictEqual(w.stderr, null);
<add> assert.strictEqual(w.stdin, stdin);
<add> assert.strictEqual(w.stdout, stdout);
<add> assert.strictEqual(w.stderr, stderr);
<ide> }));
<ide> } else {
<ide> process.exit(0);
| 2
|
Python
|
Python
|
compress target assigner
|
4f135c700383fa41c239c00ac1255b5b4eae4b27
|
<ide><path>research/object_detection/core/target_assigner.py
<ide> def assign(self,
<ide> reg_targets = self._create_regression_targets(anchors,
<ide> groundtruth_boxes,
<ide> match)
<del> cls_targets = self._create_classification_targets(groundtruth_labels,
<del> unmatched_class_label,
<del> match)
<del> reg_weights = self._create_regression_weights(match, groundtruth_weights)
<add> cls_targets = match.gather_based_on_match(
<add> groundtruth_labels,
<add> unmatched_value=unmatched_class_label,
<add> ignored_value=unmatched_class_label)
<add> reg_weights = match.gather_based_on_match(groundtruth_weights,
<add> ignored_value=0.,
<add> unmatched_value=0.)
<add> cls_weights = match.gather_based_on_match(
<add> groundtruth_weights,
<add> ignored_value=0.,
<add> unmatched_value=self._negative_class_weight)
<ide>
<del> cls_weights = self._create_classification_weights(match,
<del> groundtruth_weights)
<ide> # convert cls_weights from per-anchor to per-class.
<ide> class_label_shape = tf.shape(cls_targets)[1:]
<ide> weights_shape = tf.shape(cls_weights)
<ide> def _create_regression_targets(self, anchors, groundtruth_boxes, match):
<ide>
<ide> # Zero out the unmatched and ignored regression targets.
<ide> unmatched_ignored_reg_targets = tf.tile(
<del> self._default_regression_target(), [match_results_shape[0], 1])
<add> tf.constant([4 * [0]], tf.float32), [match_results_shape[0], 1])
<ide> matched_anchors_mask = match.matched_column_indicator()
<ide> reg_targets = tf.where(matched_anchors_mask,
<ide> matched_reg_targets,
<ide> unmatched_ignored_reg_targets)
<ide> return reg_targets
<del>
<del> def _default_regression_target(self):
<del> """Returns the default target for anchors to regress to.
<del>
<del> Default regression targets are set to zero (though in
<del> this implementation what these targets are set to should
<del> not matter as the regression weight of any box set to
<del> regress to the default target is zero).
<del>
<del> Returns:
<del> default_target: a float32 tensor with shape [1, box_code_dimension]
<del> """
<del> return tf.constant([4 * [0]], tf.float32)
<del>
<del> def _create_classification_targets(self, groundtruth_labels,
<del> unmatched_class_label, match):
<del> """Create classification targets for each anchor.
<del>
<del> Assign a classification target of for each anchor to the matching
<del> groundtruth label that is provided by match. Anchors that are not matched
<del> to anything are given the target self._unmatched_cls_target
<del>
<del> Args:
<del> groundtruth_labels: a tensor of shape [num_gt_boxes, d_1, ... d_k]
<del> with labels for each of the ground_truth boxes. The subshape
<del> [d_1, ... d_k] can be empty (corresponding to scalar labels).
<del> unmatched_class_label: a float32 tensor with shape [d_1, d_2, ..., d_k]
<del> which is consistent with the classification target for each
<del> anchor (and can be empty for scalar targets). This shape must thus be
<del> compatible with the groundtruth labels that are passed to the "assign"
<del> function (which have shape [num_gt_boxes, d_1, d_2, ..., d_k]).
<del> match: a matcher.Match object that provides a matching between anchors
<del> and groundtruth boxes.
<del>
<del> Returns:
<del> a float32 tensor with shape [num_anchors, d_1, d_2 ... d_k], where the
<del> subshape [d_1, ..., d_k] is compatible with groundtruth_labels which has
<del> shape [num_gt_boxes, d_1, d_2, ... d_k].
<del> """
<del> return match.gather_based_on_match(
<del> groundtruth_labels,
<del> unmatched_value=unmatched_class_label,
<del> ignored_value=unmatched_class_label)
<del>
<del> def _create_regression_weights(self, match, groundtruth_weights):
<del> """Set regression weight for each anchor.
<del>
<del> Only positive anchors are set to contribute to the regression loss, so this
<del> method returns a weight of 1 for every positive anchor and 0 for every
<del> negative anchor.
<del>
<del> Args:
<del> match: a matcher.Match object that provides a matching between anchors
<del> and groundtruth boxes.
<del> groundtruth_weights: a float tensor of shape [M] indicating the weight to
<del> assign to all anchors match to a particular groundtruth box.
<del>
<del> Returns:
<del> a float32 tensor with shape [num_anchors] representing regression weights.
<del> """
<del> return match.gather_based_on_match(
<del> groundtruth_weights, ignored_value=0., unmatched_value=0.)
<del>
<del> def _create_classification_weights(self,
<del> match,
<del> groundtruth_weights):
<del> """Create classification weights for each anchor.
<del>
<del> Positive (matched) anchors are associated with a weight of
<del> positive_class_weight and negative (unmatched) anchors are associated with
<del> a weight of negative_class_weight. When anchors are ignored, weights are set
<del> to zero. By default, both positive/negative weights are set to 1.0,
<del> but they can be adjusted to handle class imbalance (which is almost always
<del> the case in object detection).
<del>
<del> Args:
<del> match: a matcher.Match object that provides a matching between anchors
<del> and groundtruth boxes.
<del> groundtruth_weights: a float tensor of shape [M] indicating the weight to
<del> assign to all anchors match to a particular groundtruth box.
<del>
<del> Returns:
<del> a float32 tensor with shape [num_anchors] representing classification
<del> weights.
<del> """
<del> return match.gather_based_on_match(
<del> groundtruth_weights,
<del> ignored_value=0.,
<del> unmatched_value=self._negative_class_weight)
| 1
|
PHP
|
PHP
|
remove proxy for money temporarily
|
1f394c1fa52439e44a18d59414eff009cb682a27
|
<ide><path>src/Validation/Validator.php
<ide> public function integer($field, $message = null, $when = null)
<ide> ]);
<ide> }
<ide>
<del> /**
<del> * Add a validation rule to ensure a field is a money value.
<del> *
<del> * @param string $field The field you want to apply the rule to.
<del> * @param string $symbolPosition The position of the currency symbol
<del> * @param string|null $message The error message when the rule fails.
<del> * @param string|callable|null $when Either 'create' or 'update' or a callable that returns
<del> * true when the validation rule should be applied.
<del> * @see \Cake\Validation\Validation::money()
<del> * @return $this
<del> */
<del> public function money($field, $symbolPosition = 'left', $message = null, $when = null)
<del> {
<del> $extra = array_filter(['on' => $when, 'message' => $message]);
<del> return $this->add($field, 'money', $extra + [
<del> 'rule' => ['money', $symbolPosition]
<del> ]);
<del> }
<del>
<ide> /**
<ide> * Add a validation rule for a multiple select. Comparison is case sensitive by default.
<ide> *
<ide><path>tests/TestCase/Validation/ValidatorTest.php
<ide> public function testInteger()
<ide> $this->assertNotEmpty($validator->errors(['username' => 'not integer']));
<ide> }
<ide>
<del> /**
<del> * Tests the money proxy method
<del> *
<del> * @return void
<del> */
<del> public function testMoney()
<del> {
<del> $validator = new Validator();
<del> $this->assertProxyMethod($validator, 'money', 'left', ['left']);
<del> $this->assertNotEmpty($validator->errors(['username' => 'not money']));
<del> }
<del>
<ide> /**
<ide> * Tests the multiple proxy method
<ide> *
| 2
|
Text
|
Text
|
add changelog entry for [ci skip]
|
c087cfc6717335c1cd4b8e438004cedac2dfa492
|
<ide><path>activerecord/CHANGELOG.md
<add>* Use `UPDATE` rather than `SET` when enabling the `standard_conforming_strings`
<add> setting as this allows us to avoid disabling errors on the PostgreSQL connection.
<add> The former behavior would cause problems when using a connection pooling tool like
<add> PgBouncer because it's not guaranteed to have the same connection between calls to
<add> `execute` and it could leave the connection with errors disabled.
<add>
<add> Fixes #22101.
<add>
<add> *Harry Marr*
<add>
<ide> * Set `scope.reordering_value` to `true` if :reordering values are specified.
<ide>
<ide> Fixes #21886.
| 1
|
Text
|
Text
|
add ayase-252 to collaborators
|
3e8eda25319e8dc682de4f8028fae35da916d996
|
<ide><path>README.md
<ide> For information about the governance of the Node.js project, see
<ide> **Anatoli Papirovski** <apapirovski@mac.com> (he/him)
<ide> * [AshCripps](https://github.com/AshCripps) -
<ide> **Ash Cripps** <acripps@redhat.com>
<add>* [Ayase-252](https://github.com/Ayase-252) -
<add>**Qingyu Deng** <i@ayase-lab.com>
<ide> * [bcoe](https://github.com/bcoe) -
<ide> **Ben Coe** <bencoe@gmail.com> (he/him)
<ide> * [bengl](https://github.com/bengl) -
| 1
|
Ruby
|
Ruby
|
recommend casks where available
|
7afe1ed67aa2509791705cf6d54ecfe771c61862
|
<ide><path>Library/Homebrew/requirements.rb
<ide> def message;
<ide> <<-EOS.undent
<ide> A LaTeX distribution is required for Homebrew to install this formula.
<ide>
<del> You can install MacTeX distribution from:
<add> You can install MacTeX distribution with:
<add> brew cask install mactex
<add>
<add> Or from:
<ide> http://www.tug.org/mactex/
<ide>
<ide> Make sure that "/usr/texbin", or the location you installed it to, is in
<ide> def message
<ide> <<-EOS.undent
<ide> Java#{version_string} is required to install this formula.
<ide>
<del> You can install Java from:
<del> http://www.oracle.com/technetwork/java/javase/downloads/index.html
<add> You can install the Java Development Kit (JDK) with:
<add> brew cask install java
<ide>
<del> Make sure you install both the JRE and JDK.
<add> Or from:
<add> http://www.oracle.com/technetwork/java/javase/downloads/index.html
<ide> EOS
<ide> end
<ide> end
<ide><path>Library/Homebrew/requirements/x11_dependency.rb
<ide> def initialize(name="x11", tags=[])
<ide> @name = name
<ide> if /(\d\.)+\d/ === tags.first
<ide> @min_version = Version.new(tags.shift)
<add> @min_version_string = " #{@min_version}"
<ide> else
<ide> @min_version = Version.new("0.0.0")
<add> @min_version_string = ""
<ide> end
<ide> super(tags)
<ide> end
<ide> def initialize(name="x11", tags=[])
<ide> end
<ide>
<ide> def message; <<-EOS.undent
<del> Unsatisfied dependency: XQuartz #{@min_version}
<del> Homebrew does not package XQuartz. Installers may be found at:
<add> You can install XQuartz#{@min_version_string} with:
<add> brew cask install xquartz
<add>
<add> Or from:
<ide> https://xquartz.macosforge.org
<ide> EOS
<ide> end
| 2
|
PHP
|
PHP
|
use argument unpacking
|
f80d6f89a3b4b35933892b29857d7cad1cda29bd
|
<ide><path>src/Illuminate/Support/Facades/Facade.php
<ide> public static function __callStatic($method, $args)
<ide> throw new RuntimeException('A facade root has not been set.');
<ide> }
<ide>
<del> switch (count($args)) {
<del> case 0:
<del> return $instance->$method();
<del>
<del> case 1:
<del> return $instance->$method($args[0]);
<del>
<del> case 2:
<del> return $instance->$method($args[0], $args[1]);
<del>
<del> case 3:
<del> return $instance->$method($args[0], $args[1], $args[2]);
<del>
<del> case 4:
<del> return $instance->$method($args[0], $args[1], $args[2], $args[3]);
<del>
<del> default:
<del> return call_user_func_array([$instance, $method], $args);
<del> }
<add> return $instance->$method(...$args);
<ide> }
<ide> }
| 1
|
Ruby
|
Ruby
|
require conversions to use string#ord
|
f802eb2f0039512105d6ed985e1ed3245f7316d5
|
<ide><path>activesupport/lib/active_support/cache/file_store.rb
<ide> require 'active_support/core_ext/file/atomic'
<add>require 'active_support/core_ext/string/conversions'
<ide>
<ide> module ActiveSupport
<ide> module Cache
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.