Dataset Viewer
Auto-converted to Parquet Duplicate
query
stringlengths
10
3.85k
ru_query
stringlengths
9
3.76k
document
stringlengths
17
430k
metadata
dict
negatives
listlengths
97
100
negative_scores
listlengths
97
100
document_score
stringlengths
5
10
document_rank
stringclasses
2 values
Returns the value of the 'go_package' option of the first .proto file found in the same directory as projectFile
Возвращает значение опции 'go_package' первого файла .proto, найденного в той же директории, что и projectFile
func detectGoPackageForProject(projectFile string) (string, error) { var goPkg string projectDir := filepath.Dir(projectFile) if err := filepath.Walk(projectDir, func(protoFile string, info os.FileInfo, err error) error { // already set if goPkg != "" { return nil } if !strings.HasSuffix(protoFile, ".proto") { return nil } // search for go_package on protos in the same dir as the project.json if projectDir != filepath.Dir(protoFile) { return nil } content, err := ioutil.ReadFile(protoFile) if err != nil { return err } lines := strings.Split(string(content), "\n") for _, line := range lines { goPackage := goPackageStatementRegex.FindStringSubmatch(line) if len(goPackage) == 0 { continue } if len(goPackage) != 2 { return errors.Errorf("parsing go_package error: from %v found %v", line, goPackage) } goPkg = goPackage[1] break } return nil }); err != nil { return "", err } if goPkg == "" { return "", errors.Errorf("no go_package statement found in root dir of project %v", projectFile) } return goPkg, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (g *Generator) GoFilePackage(depfile *fdep.DepFile) string {\n\treturn fproto_wrap.BaseName(g.GoWrapPackage(depfile))\n}", "func (g *Generator) GoPackage(depfile *fdep.DepFile) string {\n\tfor _, o := range depfile.ProtoFile.Options {\n\t\tif o.Name == \"go_package\" {\n\t\t\treturn o.Value.String()\n\t\t}\n\t}\n\treturn path.Dir(depfile.FilePath)\n}", "func (c *common) GetPackage() string { return c.file.GetPackage() }", "func (pkg *goPackage) firstGoFile() string {\n\tgoSrcs := []platformStringsBuilder{\n\t\tpkg.library.sources,\n\t\tpkg.binary.sources,\n\t\tpkg.test.sources,\n\t}\n\tfor _, sb := range goSrcs {\n\t\tif sb.strs != nil {\n\t\t\tfor s := range sb.strs {\n\t\t\t\tif strings.HasSuffix(s, \".go\") {\n\t\t\t\t\treturn s\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn \"\"\n}", "func (fd *File) GoPackagePath() string {\n\treturn fd.builder.GoPackagePath\n}", "func GoPackage(packageName string) string {\n\tif packageName == \"\" {\n\t\treturn \"\"\n\t}\n\tsplit := strings.Split(packageName, \".\")\n\treturn split[len(split)-1] + \"pb\"\n}", "func goPackageName(d *descriptor.FileDescriptorProto) (name string, explicit bool) {\n\t// Does the file have a \"go_package\" option?\n\tif _, pkg, ok := goPackageOption(d); ok {\n\t\treturn pkg, true\n\t}\n\n\t// Does the file have a package clause?\n\tif pkg := d.GetPackage(); pkg != \"\" {\n\t\treturn pkg, false\n\t}\n\t// Use the file base name.\n\treturn baseName(d.GetName()), false\n}", "func (d *FileDescriptor) goPackageName() (name string, explicit bool) {\n\t// Does the file have a \"go_package\" option?\n\tif _, pkg, ok := d.goPackageOption(); ok {\n\t\treturn pkg, true\n\t}\n\n\t// Does the file have a package clause?\n\tif pkg := d.GetPackage(); pkg != \"\" {\n\t\treturn pkg, false\n\t}\n\t// Use the file base name.\n\treturn baseName(d.GetName()), false\n}", "func (d *FileDescriptor) PackageName() string { return uniquePackageOf(d.FileDescriptorProto) }", "func (c *common) PackageName() string { return uniquePackageOf(c.file) }", "func (d *FileDescriptor) goFileName(pathType pathType) string {\n\tname := *d.Name\n\tif ext := path.Ext(name); ext == \".proto\" || ext == \".protodevel\" {\n\t\tname = name[:len(name)-len(ext)]\n\t}\n\tname += \".cobra.pb.go\"\n\n\tif pathType == pathTypeSourceRelative {\n\t\treturn name\n\t}\n\n\t// Does the file have a \"go_package\" option?\n\t// If it does, it may override the filename.\n\tif impPath, _, ok := d.goPackageOption(); ok && impPath != \"\" {\n\t\t// Replace the existing dirname with the declared import path.\n\t\t_, name = path.Split(name)\n\t\tname = path.Join(impPath, name)\n\t\treturn name\n\t}\n\n\treturn name\n}", "func (*GetProjectRequest) Descriptor() ([]byte, []int) {\n\treturn file_user_proto_rawDescGZIP(), []int{25}\n}", "func goFileName(d *descriptor.FileDescriptorProto) string {\n\tname := *d.Name\n\tif ext := path.Ext(name); ext == \".proto\" || ext == \".protodevel\" {\n\t\tname = name[:len(name)-len(ext)]\n\t}\n\tname += \".nrpc.go\"\n\n\t// Does the file have a \"go_package\" option?\n\t// If it does, it may override the filename.\n\tif impPath, _, ok := goPackageOption(d); ok && impPath != \"\" {\n\t\t// Replace the existing dirname with the declared import path.\n\t\t_, name = path.Split(name)\n\t\tname = path.Join(impPath, name)\n\t\treturn name\n\t}\n\n\treturn name\n}", "func (pp *protoPackage) pkgPath() string {\n\treturn strings.Replace(pp.Pkg, \".\", \"/\", -1)\n}", "func goPkg(fileName string) (string, error) {\n\tcontent, err := os.ReadFile(fileName)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tvar pkgName string\n\tif match := goPkgOptRe.FindSubmatch(content); len(match) > 0 {\n\t\tpn, err := strconv.Unquote(string(match[1]))\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tpkgName = pn\n\t}\n\tif p := strings.IndexRune(pkgName, ';'); p > 0 {\n\t\tpkgName = pkgName[:p]\n\t}\n\treturn pkgName, nil\n}", "func (project Project) Package() (string, error) {\n\n\tif project.packageName != \"\" {\n\t\treturn project.packageName, nil\n\t}\n\n\tgoModPath := project.RelPath(GoModFileName)\n\tif !project.FileExists(goModPath) {\n\t\treturn \"\", errors.New(\"Failed to determine the package name for this project\")\n\t}\n\n\tb, err := ioutil.ReadFile(goModPath)\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"Failed to read the go.mod file\")\n\t}\n\n\tmod, err := gomod.Parse(goModPath, b)\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"Failed to parse the go.mod file\")\n\t}\n\n\tproject.packageName = strings.TrimSuffix(mod.Name, \"/\")\n\n\treturn project.packageName, nil\n\n}", "func (g *Generator) GoWrapFilePackage(depfile *fdep.DepFile) string {\n\tif g.PkgSource != nil {\n\t\tif p, ok := g.PkgSource.GetFilePkg(g, depfile); ok {\n\t\t\treturn p\n\t\t}\n\t}\n\n\treturn \"fw\" + fproto_wrap.BaseName(g.GoWrapPackage(depfile))\n}", "func (f *FileStruct) GetPersistPackageOption() string {\n\tif f.Desc == nil || f.Desc.GetOptions() == nil {\n\t\treturn \"\"\n\t}\n\tif proto.HasExtension(f.Desc.GetOptions(), persist.E_Package) {\n\t\tpkg, err := proto.GetExtension(f.Desc.GetOptions(), persist.E_Package)\n\t\tif err != nil {\n\t\t\tlogrus.WithError(err).Debug(\"Error\")\n\t\t\treturn \"\"\n\t\t}\n\t\t//logrus.WithField(\"pkg\", *pkg.(*string)).Info(\"Package\")\n\t\treturn *pkg.(*string)\n\t}\n\tlogrus.WithField(\"File Options\", f.Desc.GetOptions()).Debug(\"file options\")\n\treturn \"\"\n}", "func Which(s protoreflect.FullName) ProtoFile {\r\n\treturn wellKnownTypes[s]\r\n}", "func GetPackageName(source string) string {\n\tfileNode, err := parser.ParseFile(\"\", source, nil, parser.ImportsOnly)\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\n\treturn fileNode.Name.Name()\n}", "func (*GetProjectResponse) Descriptor() ([]byte, []int) {\n\treturn file_user_proto_rawDescGZIP(), []int{26}\n}", "func (*Project) Descriptor() ([]byte, []int) {\n\treturn file_proto_carbon_proto_rawDescGZIP(), []int{0}\n}", "func ProtoFromFileDescriptor(d protoreflect.FileDescriptor) *descriptorpb.FileDescriptorProto {\n\tif imp, ok := d.(protoreflect.FileImport); ok {\n\t\td = imp.FileDescriptor\n\t}\n\ttype canProto interface {\n\t\tFileDescriptorProto() *descriptorpb.FileDescriptorProto\n\t}\n\tif res, ok := d.(canProto); ok {\n\t\treturn res.FileDescriptorProto()\n\t}\n\tif res, ok := d.(DescriptorProtoWrapper); ok {\n\t\tif fd, ok := res.AsProto().(*descriptorpb.FileDescriptorProto); ok {\n\t\t\treturn fd\n\t\t}\n\t}\n\treturn protodesc.ToFileDescriptorProto(d)\n}", "func goPackageOption(d *descriptor.FileDescriptorProto) (impPath, pkg string, ok bool) {\n\tpkg = d.GetOptions().GetGoPackage()\n\tif pkg == \"\" {\n\t\treturn\n\t}\n\tok = true\n\t// The presence of a slash implies there's an import path.\n\tslash := strings.LastIndex(pkg, \"/\")\n\tif slash < 0 {\n\t\treturn\n\t}\n\timpPath, pkg = pkg, pkg[slash+1:]\n\t// A semicolon-delimited suffix overrides the package name.\n\tsc := strings.IndexByte(impPath, ';')\n\tif sc < 0 {\n\t\treturn\n\t}\n\timpPath, pkg = impPath[:sc], impPath[sc+1:]\n\treturn\n}", "func (*CodeGeneratorResponse_File) Descriptor() ([]byte, []int) {\n\treturn file_google_protobuf_compiler_plugin_proto_rawDescGZIP(), []int{2, 0}\n}", "func ToFileDescriptorProto(file protoreflect.FileDescriptor) *descriptorpb.FileDescriptorProto {\n\tp := &descriptorpb.FileDescriptorProto{\n\t\tName: proto.String(file.Path()),\n\t\tOptions: proto.Clone(file.Options()).(*descriptorpb.FileOptions),\n\t}\n\tif file.Package() != \"\" {\n\t\tp.Package = proto.String(string(file.Package()))\n\t}\n\tfor i, imports := 0, file.Imports(); i < imports.Len(); i++ {\n\t\timp := imports.Get(i)\n\t\tp.Dependency = append(p.Dependency, imp.Path())\n\t\tif imp.IsPublic {\n\t\t\tp.PublicDependency = append(p.PublicDependency, int32(i))\n\t\t}\n\t\tif imp.IsWeak {\n\t\t\tp.WeakDependency = append(p.WeakDependency, int32(i))\n\t\t}\n\t}\n\tfor i, locs := 0, file.SourceLocations(); i < locs.Len(); i++ {\n\t\tloc := locs.Get(i)\n\t\tl := &descriptorpb.SourceCodeInfo_Location{}\n\t\tl.Path = append(l.Path, loc.Path...)\n\t\tif loc.StartLine == loc.EndLine {\n\t\t\tl.Span = []int32{int32(loc.StartLine), int32(loc.StartColumn), int32(loc.EndColumn)}\n\t\t} else {\n\t\t\tl.Span = []int32{int32(loc.StartLine), int32(loc.StartColumn), int32(loc.EndLine), int32(loc.EndColumn)}\n\t\t}\n\t\tl.LeadingDetachedComments = append([]string(nil), loc.LeadingDetachedComments...)\n\t\tif loc.LeadingComments != \"\" {\n\t\t\tl.LeadingComments = proto.String(loc.LeadingComments)\n\t\t}\n\t\tif loc.TrailingComments != \"\" {\n\t\t\tl.TrailingComments = proto.String(loc.TrailingComments)\n\t\t}\n\t\tif p.SourceCodeInfo == nil {\n\t\t\tp.SourceCodeInfo = &descriptorpb.SourceCodeInfo{}\n\t\t}\n\t\tp.SourceCodeInfo.Location = append(p.SourceCodeInfo.Location, l)\n\n\t}\n\tfor i, messages := 0, file.Messages(); i < messages.Len(); i++ {\n\t\tp.MessageType = append(p.MessageType, ToDescriptorProto(messages.Get(i)))\n\t}\n\tfor i, enums := 0, file.Enums(); i < enums.Len(); i++ {\n\t\tp.EnumType = append(p.EnumType, ToEnumDescriptorProto(enums.Get(i)))\n\t}\n\tfor i, services := 0, file.Services(); i < services.Len(); i++ {\n\t\tp.Service = append(p.Service, ToServiceDescriptorProto(services.Get(i)))\n\t}\n\tfor i, exts := 0, file.Extensions(); i < exts.Len(); i++ {\n\t\tp.Extension = append(p.Extension, ToFieldDescriptorProto(exts.Get(i)))\n\t}\n\tif syntax := file.Syntax(); syntax != protoreflect.Proto2 {\n\t\tp.Syntax = proto.String(file.Syntax().String())\n\t}\n\treturn p\n}", "func (*PatchProject) Descriptor() ([]byte, []int) {\n\treturn file_determined_project_v1_project_proto_rawDescGZIP(), []int{4}\n}", "func (*Project) Descriptor() ([]byte, []int) {\n\treturn file_determined_project_v1_project_proto_rawDescGZIP(), []int{2}\n}", "func (d *FileDescriptor) goPackageOption() (impPath, pkg string, ok bool) {\n\tpkg = d.GetOptions().GetGoPackage()\n\tif pkg == \"\" {\n\t\treturn\n\t}\n\tok = true\n\t// The presence of a slash implies there's an import path.\n\tslash := strings.LastIndex(pkg, \"/\")\n\tif slash < 0 {\n\t\treturn\n\t}\n\timpPath, pkg = pkg, pkg[slash+1:]\n\t// A semicolon-delimited suffix overrides the package name.\n\tsc := strings.IndexByte(impPath, ';')\n\tif sc < 0 {\n\t\treturn\n\t}\n\timpPath, pkg = impPath[:sc], impPath[sc+1:]\n\treturn\n}", "func (*ProjectSimple) Descriptor() ([]byte, []int) {\n\treturn file_user_proto_rawDescGZIP(), []int{29}\n}", "func (*GoPackageInfo) Descriptor() ([]byte, []int) {\n\treturn file_kythe_proto_go_proto_rawDescGZIP(), []int{1}\n}", "func deduceGenPkgName(genFiles []*descriptor.FileDescriptorProto) (string, error) {\n\tvar genPkgName string\n\tfor _, f := range genFiles {\n\t\tname, explicit := goPackageName(f)\n\t\tif explicit {\n\t\t\tname = stringutils.CleanIdentifier(name)\n\t\t\tif genPkgName != \"\" && genPkgName != name {\n\t\t\t\t// Make sure they're all set consistently.\n\t\t\t\treturn \"\", errors.Errorf(\"files have conflicting go_package settings, must be the same: %q and %q\", genPkgName, name)\n\t\t\t}\n\t\t\tgenPkgName = name\n\t\t}\n\t}\n\tif genPkgName != \"\" {\n\t\treturn genPkgName, nil\n\t}\n\n\t// If there is no explicit setting, then check the implicit package name\n\t// (derived from the protobuf package name) of the files and make sure it's\n\t// consistent.\n\tfor _, f := range genFiles {\n\t\tname, _ := goPackageName(f)\n\t\tname = stringutils.CleanIdentifier(name)\n\t\tif genPkgName != \"\" && genPkgName != name {\n\t\t\treturn \"\", errors.Errorf(\"files have conflicting package names, must be the same or overridden with go_package: %q and %q\", genPkgName, name)\n\t\t}\n\t\tgenPkgName = name\n\t}\n\n\t// All the files have the same name, so we're good.\n\treturn genPkgName, nil\n}", "func GetFirstGoPath() string {\n\treturn strings.Split(os.Getenv(\"GOPATH\"), \":\")[0]\n}", "func lookupProjPath(protoAbs string) (result string) {\n\tlastIndex := len(protoAbs)\n\tcurPath := protoAbs\n\n\tfor lastIndex > 0 {\n\t\tif fileExist(curPath+\"/cmd\") && fileExist(curPath+\"/api\") {\n\t\t\tresult = curPath\n\t\t\treturn\n\t\t}\n\t\tlastIndex = strings.LastIndex(curPath, string(os.PathSeparator))\n\t\tcurPath = protoAbs[:lastIndex]\n\t}\n\tresult = \"\"\n\treturn\n}", "func Namespace(file *descriptor.FileDescriptorProto) string {\n\toptions := file.GetOptions()\n\n\t// When there is a namespace option defined we use it\n\tif options.PhpNamespace != nil {\n\t\treturn options.GetPhpNamespace()\n\t}\n\n\treturn Name(file.GetPackage())\n}", "func protobufName(f *ast.Field) string {\n\tfor _, attr := range f.Attrs {\n\t\tif strings.HasPrefix(attr.Text, \"@protobuf\") {\n\t\t\tfor _, str := range strings.Split(attr.Text[10:len(attr.Text)-1], \",\") {\n\t\t\t\tif strings.HasPrefix(str, \"name=\") {\n\t\t\t\t\treturn str[5:]\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn \"\"\n}", "func (*Project) Descriptor() ([]byte, []int) {\n\treturn file_management_proto_rawDescGZIP(), []int{79}\n}", "func (*Project) Descriptor() ([]byte, []int) {\n\treturn file_management_proto_rawDescGZIP(), []int{76}\n}", "func (*ExistingFile) Descriptor() ([]byte, []int) {\n\treturn file_protoconfig_go_kingpinv2_v1_extensions_proto_rawDescGZIP(), []int{1}\n}", "func getTmplFileDesc(fds []*descriptor.FileDescriptorProto) (string, *descriptor.FileDescriptorProto, error) {\n\tvar templateDescriptorProto *descriptor.FileDescriptorProto\n\tfor _, fd := range fds {\n\t\tif fd.GetOptions() == nil || !proto.HasExtension(fd.GetOptions(), tmpl.E_TemplateVariety) {\n\t\t\tcontinue\n\t\t}\n\t\tif templateDescriptorProto != nil {\n\t\t\treturn \"\", nil, fmt.Errorf(\n\t\t\t\t\"proto files %s and %s, both have the option %s. Only one proto file is allowed with this options\",\n\t\t\t\tfd.GetName(), templateDescriptorProto.GetName(), tmpl.E_TemplateVariety.Name)\n\t\t}\n\t\ttemplateDescriptorProto = fd\n\t}\n\n\tif templateDescriptorProto == nil {\n\t\treturn \"\", nil, fmt.Errorf(\"there has to be one proto file that has the extension %s\", tmpl.E_TemplateVariety.Name)\n\t}\n\n\tvar tmplName string\n\tif nameExt, err := proto.GetExtension(templateDescriptorProto.GetOptions(), tmpl.E_TemplateName); err != nil {\n\t\treturn \"\", nil, fmt.Errorf(\n\t\t\t\"proto files %s is missing required template_name option\", templateDescriptorProto.GetName())\n\t} else if err := validateTmplName(*(nameExt.(*string))); err != nil {\n\t\treturn \"\", nil, err\n\t} else {\n\t\ttmplName = *(nameExt.(*string))\n\t}\n\n\treturn tmplName, templateDescriptorProto, nil\n}", "func generateFile(gen *protogen.Plugin, file *protogen.File) {\n\tfilename := file.GeneratedFilenamePrefix + \"_message.pb.go\"\n\tg := gen.NewGeneratedFile(filename, file.GoImportPath)\n\n\tg.P(\"// Code generated by protoc-gen-message-validator. DO NOT EDIT.\")\n\tg.P()\n\tg.P(\"package \", file.GoPackageName)\n\tg.P()\n\n\tfor _, message := range file.Messages {\n\t\tstructName := string(message.Desc.Name())\n\t\tprefix := strings.ToLower(string(structName[0]))\n\n\t\tfor _, subMessage := range message.Messages {\n\t\t\tif subMessage.Desc.IsMapEntry() {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tsubStructName := string(subMessage.Desc.Name())\n\t\t\tgenerateMessage(fmt.Sprintf(\"%s_%s\", structName, subStructName), prefix, subMessage, g)\n\t\t}\n\n\t\tgenerateMessage(structName, prefix, message, g)\n\t}\n}", "func (*AppGroup) Descriptor() ([]byte, []int) {\n\treturn file_common_proto_rawDescGZIP(), []int{1}\n}", "func (*UpdateProjectRequest) Descriptor() ([]byte, []int) {\n\treturn file_user_proto_rawDescGZIP(), []int{21}\n}", "func (*ProjectModel) Descriptor() ([]byte, []int) {\n\treturn file_determined_project_v1_project_proto_rawDescGZIP(), []int{3}\n}", "func (*Project) Descriptor() ([]byte, []int) {\n\treturn file_api_ocp_project_api_ocp_project_api_proto_rawDescGZIP(), []int{12}\n}", "func (s *Stub) GetProject() string {\n\treturn \"\"\n}", "func (*UpdateProjectResponse) Descriptor() ([]byte, []int) {\n\treturn file_user_proto_rawDescGZIP(), []int{22}\n}", "func guessPackageName(b *util.BuildCtxt, base string) string {\n\tcwd, err := os.Getwd()\n\tif err != nil {\n\t\treturn \"main\"\n\t}\n\n\tpkg, err := b.Import(base, cwd, 0)\n\tif err != nil {\n\t\t// There may not be any top level Go source files but the project may\n\t\t// still be within the GOPATH.\n\t\tif strings.HasPrefix(base, b.GOPATH) {\n\t\t\tp := strings.TrimPrefix(base, b.GOPATH)\n\t\t\treturn strings.Trim(p, string(os.PathSeparator))\n\t\t}\n\t}\n\n\treturn pkg.ImportPath\n}", "func GetGoPackage(url string) string {\n\tswitch {\n\tcase strings.Contains(url, \";\"):\n\t\tidx := strings.LastIndex(url, \";\")\n\t\treturn url[idx+1:]\n\tcase strings.Contains(url, \"/\"):\n\t\tidx := strings.LastIndex(url, \"/\")\n\t\treturn url[idx+1:]\n\tdefault:\n\t\treturn url\n\t}\n}", "func (g *Generator) GoWrapPackage(depfile *fdep.DepFile) string {\n\tif g.PkgSource != nil {\n\t\tif p, ok := g.PkgSource.GetPkg(g, depfile); ok {\n\t\t\treturn p\n\t\t}\n\t}\n\n\tfor _, o := range depfile.ProtoFile.Options {\n\t\tif o.Name == \"gowrap_package\" {\n\t\t\treturn o.Value.String()\n\t\t}\n\t}\n\n\t// prepend \"fpwrap\"\n\tfor _, o := range depfile.ProtoFile.Options {\n\t\tif o.Name == \"go_package\" {\n\t\t\treturn path.Join(\"fpwrap\", o.Value.String())\n\t\t}\n\t}\n\treturn path.Join(\"fpwrap\", path.Dir(depfile.FilePath))\n}", "func getFileExtensionBySdk(precompiledObjectPath string) (string, error) {\n\tsdk := strings.Split(precompiledObjectPath, string(os.PathSeparator))[0]\n\tvar extension string\n\tswitch sdk {\n\tcase pb.Sdk_SDK_JAVA.String():\n\t\textension = javaExtension\n\tcase pb.Sdk_SDK_PYTHON.String():\n\t\textension = pyExtension\n\tcase pb.Sdk_SDK_GO.String():\n\t\textension = goExtension\n\tcase pb.Sdk_SDK_SCIO.String():\n\t\textension = scioExtension\n\tdefault:\n\t\treturn \"\", fmt.Errorf(\"\")\n\t}\n\treturn extension, nil\n}", "func (p *Provider) GetProject() string {\n\to := p.opts\n\tif len(o.projects) > 1 {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"multiple projects not supported (%d specified)\", len(o.projects)))\n\t}\n\treturn o.projects[0]\n}", "func (*ListProjectsResponse_Project) Descriptor() ([]byte, []int) {\n\treturn file_web_proto_rawDescGZIP(), []int{1, 0}\n}", "func (*ProjectMember) Descriptor() ([]byte, []int) {\n\treturn file_management_proto_rawDescGZIP(), []int{78}\n}", "func (*GetUserProjectsRequest) Descriptor() ([]byte, []int) {\n\treturn file_user_proto_rawDescGZIP(), []int{27}\n}", "func (t *Type) fullGoPackageName() string {\n\tif t.qname.namespace != t.Namespace {\n\t\treturn t.qname.namespace.fullGoPackageName\n\t}\n\treturn \"\"\n}", "func (i Import) Package() string {\n\tif v := i.Alias; len(v) != 0 {\n\t\treturn v\n\t}\n\n\tif v := i.Path; len(v) != 0 {\n\t\tparts := strings.Split(v, \"/\")\n\t\tpkg := parts[len(parts)-1]\n\t\treturn pkg\n\t}\n\n\treturn \"\"\n}", "func (*CreateProjectResponse) Descriptor() ([]byte, []int) {\n\treturn file_user_proto_rawDescGZIP(), []int{18}\n}", "func (*CreateProjectRequest) Descriptor() ([]byte, []int) {\n\treturn file_user_proto_rawDescGZIP(), []int{17}\n}", "func (*Projects) Descriptor() ([]byte, []int) {\n\treturn file_management_proto_rawDescGZIP(), []int{78}\n}", "func (b *Buffer) FileType() string {\n\treturn b.Settings[\"filetype\"].(string)\n}", "func (*ProjectMember) Descriptor() ([]byte, []int) {\n\treturn file_management_proto_rawDescGZIP(), []int{81}\n}", "func (x *fastReflection_ModuleOptions) Get(descriptor protoreflect.FieldDescriptor) protoreflect.Value {\n\tswitch descriptor.FullName() {\n\tcase \"cosmos.autocli.v1.ModuleOptions.tx\":\n\t\tvalue := x.Tx\n\t\treturn protoreflect.ValueOfMessage(value.ProtoReflect())\n\tcase \"cosmos.autocli.v1.ModuleOptions.query\":\n\t\tvalue := x.Query\n\t\treturn protoreflect.ValueOfMessage(value.ProtoReflect())\n\tdefault:\n\t\tif descriptor.IsExtension() {\n\t\t\tpanic(fmt.Errorf(\"proto3 declared messages do not support extensions: cosmos.autocli.v1.ModuleOptions\"))\n\t\t}\n\t\tpanic(fmt.Errorf(\"message cosmos.autocli.v1.ModuleOptions does not contain field %s\", descriptor.FullName()))\n\t}\n}", "func (*DescribeProjectRequest) Descriptor() ([]byte, []int) {\n\treturn file_api_ocp_project_api_ocp_project_api_proto_rawDescGZIP(), []int{8}\n}", "func GetProto(src string) (string, error) {\n\tparsed, err := url.Parse(src)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif len(parsed.Scheme) > 0 {\n\t\treturn parsed.Scheme, nil\n\t}\n\n\treturn \"\", nil\n}", "func (*NewProject) Descriptor() ([]byte, []int) {\n\treturn file_api_ocp_project_api_ocp_project_api_proto_rawDescGZIP(), []int{13}\n}", "func goPath() string {\n\tgpDefault := build.Default.GOPATH\n\tgps := filepath.SplitList(gpDefault)\n\n\treturn gps[0]\n}", "func (p *plugin) analyzeFile(f *descriptor.FileDescriptorProto) error {\n\tif f.GetSyntax() != \"proto3\" {\n\t\treturn fmt.Errorf(\"unsupported syntax '%s', must be 'proto3'\", f.GetSyntax())\n\t}\n\n\tfile := goFile{structs: map[string]goStruct{}}\n\n\tfor _, m := range f.GetMessageType() {\n\t\tif err := p.analyzeMessageType(file, []string{}, m); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to analyze message type '%s': %s\", m.GetName(), err.Error())\n\t\t}\n\t}\n\n\tif len(file.structs) > 0 {\n\t\tn := filepath.Base(f.GetName())\n\t\tn = strings.TrimSuffix(n, filepath.Ext(n))\n\t\tp.targetFiles[n+\".pb.go\"] = file\n\t}\n\n\treturn nil\n}", "func (resolver *NpmResolver) ParsePkgFile(pkgFile string) (*Package, error) {\n\tcontent, err := ioutil.ReadFile(pkgFile)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar packageInfo Package\n\tif err := json.Unmarshal(content, &packageInfo); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &packageInfo, nil\n}", "func getPackageName(datatypeName string) string {\n\tparts := strings.Split(datatypeName, \".\")\n\tif len(parts) == 1 {\n\t\treturn \"\" // no package name\n\t}\n\n\toffset := 0\n\tfor i, p := range parts {\n\t\tif unicode.IsUpper(rune(p[0])) {\n\t\t\tbreak\n\t\t}\n\n\t\toffset += len(p)\n\t\tif i > 0 {\n\t\t\toffset += 1 // also account for the '.'\n\t\t}\n\t}\n\n\treturn datatypeName[:offset]\n}", "func (pp *protoPackage) absPath() string {\n\treturn path.Join(pp.Path, pp.pkgPath())\n}", "func (f *tmplFuncs) resolvePkgPath(pkg string) string {\n\t// Test this proto file itself:\n\tif stripExt(filepath.Base(*f.f.Name)) == pkg {\n\t\treturn *f.f.Name\n\t}\n\n\t// Test each dependency:\n\tfor _, p := range f.f.Dependency {\n\t\tif stripExt(filepath.Base(p)) == pkg {\n\t\t\treturn p\n\t\t}\n\t}\n\treturn \"\"\n}", "func (*MyCompany) Descriptor() ([]byte, []int) {\n\treturn file_parser_company_proto_rawDescGZIP(), []int{21}\n}", "func IsCommonProto(f *desc.FileDescriptor) bool {\n\tp := f.GetPackage()\n\tfor _, prefix := range []string{\"google.api\", \"google.protobuf\", \"google.rpc\", \"google.longrunning\"} {\n\t\tif strings.HasPrefix(p, prefix) {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (i *Import) GetTFVCProject() string {\n\tif i == nil || i.TFVCProject == nil {\n\t\treturn \"\"\n\t}\n\treturn *i.TFVCProject\n}", "func (*Program) Descriptor() ([]byte, []int) {\n\treturn file_proto_common_proto_rawDescGZIP(), []int{1}\n}", "func (*DescribeProjectResponse) Descriptor() ([]byte, []int) {\n\treturn file_api_ocp_project_api_ocp_project_api_proto_rawDescGZIP(), []int{9}\n}", "func (*DcsProject) Descriptor() ([]byte, []int) {\n\treturn file_dcs_model_proto_rawDescGZIP(), []int{2}\n}", "func (o *ProformaArray) GetProject() string {\n\tif o == nil || o.Project == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Project\n}", "func (*CheckProjectTokenResponse) Descriptor() ([]byte, []int) {\n\treturn file_user_proto_rawDescGZIP(), []int{33}\n}", "func (*ProjectColumn) Descriptor() ([]byte, []int) {\n\treturn file_determined_project_v1_project_proto_rawDescGZIP(), []int{0}\n}", "func getCallerPackage() string {\n\tconst replayModeRecordModeCaller = 3\n\tc := caller.Get(replayModeRecordModeCaller)\n\tpkg := strings.SplitN(c, \".\", 2)[0]\n\treturn path.Base(pkg)\n}", "func (pr *PkgDecoder) PkgPath() string { return pr.pkgPath }", "func (b *baseBuilder) GetFile() *FileBuilder {\n\tp := b.parent\n\tfor p != nil {\n\t\tif fb, ok := p.(*FileBuilder); ok {\n\t\t\treturn fb\n\t\t}\n\t\tp = p.GetParent()\n\t}\n\treturn nil\n}", "func GoServicePackagePath(name string) string {\n\treturn filepath.Join(PaceBase, ServiceBase, name)\n}", "func (*UpdateProjectRequest) Descriptor() ([]byte, []int) {\n\treturn file_api_ocp_project_api_ocp_project_api_proto_rawDescGZIP(), []int{10}\n}", "func GetPkgName() string {\n\t_, filePath, _, _ := runtime.Caller(0)\n\tfile, _ := os.Open(filePath)\n\tr := bufio.NewReader(file)\n\tline, _, _ := r.ReadLine()\n\tpkgName := bytes.TrimPrefix(line, []byte(\"package \"))\n\n\treturn string(pkgName)\n}", "func (p *Parser) Package() string {\n\treturn p.asTree.Name.Name\n}", "func ProtoServiceName(fullname string) Option {\n\treturn func(a *appOptions) {\n\t\ta.protoService = fullname\n\t}\n}", "func (*Projects) Descriptor() ([]byte, []int) {\n\treturn file_management_proto_rawDescGZIP(), []int{75}\n}", "func packageFilename(pwd, relativePath string) string {\n\tfullPath := filepath.Join(pwd, relativePath)\n\treturn strings.TrimPrefix(strings.TrimPrefix(fullPath, filepath.Join(gopath(), \"src\")), \"/\")\n}", "func (*CheckProjectTokenRequest) Descriptor() ([]byte, []int) {\n\treturn file_user_proto_rawDescGZIP(), []int{32}\n}", "func (*ProjectID) Descriptor() ([]byte, []int) {\n\treturn file_management_proto_rawDescGZIP(), []int{6}\n}", "func (*ProjectID) Descriptor() ([]byte, []int) {\n\treturn file_management_proto_rawDescGZIP(), []int{6}\n}", "func (o *ProformaArray) GetProjectOk() (*string, bool) {\n\tif o == nil || o.Project == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Project, true\n}", "func getPackageName(f string) string {\n\tfor {\n\t\tlastPeriod := strings.LastIndex(f, \".\")\n\t\tlastSlash := strings.LastIndex(f, \"/\")\n\t\tif lastPeriod > lastSlash {\n\t\t\tf = f[:lastPeriod]\n\t\t} else {\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn f\n}", "func Project(ctx context.Context, project string) (*configpb.ProjectConfig, error) {\n\tconfigs, err := Projects(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif c, ok := configs[project]; ok {\n\t\treturn c, nil\n\t}\n\treturn nil, ErrNotFoundProjectConfig\n}", "func (*GetMyRequest) Descriptor() ([]byte, []int) {\n\treturn file_parser_company_proto_rawDescGZIP(), []int{6}\n}", "func (*GlobalOptions) Descriptor() ([]byte, []int) {\n\treturn file_github_com_google_cloudprober_targets_gce_proto_config_proto_rawDescGZIP(), []int{3}\n}", "func (pi *PackageInfo) FileVName(file *ast.File) *spb.VName {\n\tif v := pi.fileVName[file]; v != nil {\n\t\treturn v\n\t}\n\tv := proto.Clone(pi.VName).(*spb.VName)\n\tv.Language = \"\"\n\tv.Signature = \"\"\n\tv.Path = pi.FileSet.Position(file.Pos()).Filename\n\treturn v\n}" ]
[ "0.6514599", "0.6418956", "0.62568474", "0.6100058", "0.6059389", "0.6033528", "0.5760218", "0.571393", "0.56766814", "0.56747204", "0.55886865", "0.55810106", "0.55710924", "0.5528127", "0.5502348", "0.55021805", "0.5456921", "0.5454185", "0.54367715", "0.54307157", "0.54025465", "0.53424656", "0.5329483", "0.5327733", "0.53020066", "0.52819157", "0.52795607", "0.52257407", "0.52240294", "0.51456493", "0.5137637", "0.5134006", "0.5133966", "0.51120484", "0.5107064", "0.50847185", "0.5082025", "0.507457", "0.50681937", "0.5033462", "0.50315183", "0.50245786", "0.5019538", "0.5017327", "0.50053775", "0.50039554", "0.49943617", "0.49909803", "0.49890676", "0.49804884", "0.49707755", "0.4931849", "0.49237567", "0.4921064", "0.4905122", "0.49045667", "0.49015555", "0.49012348", "0.4892581", "0.4889273", "0.48874435", "0.48817107", "0.48771524", "0.4876309", "0.48685285", "0.48641378", "0.4861054", "0.48610055", "0.48546317", "0.4849436", "0.48461854", "0.48250479", "0.4822479", "0.48205945", "0.48192737", "0.48135585", "0.48060772", "0.4804189", "0.48035768", "0.48003185", "0.47990838", "0.4787936", "0.47851935", "0.47817233", "0.4781146", "0.47775137", "0.47767115", "0.4776245", "0.47737917", "0.47712553", "0.476954", "0.47676048", "0.47668436", "0.47668436", "0.47661626", "0.47600132", "0.47533566", "0.47468054", "0.47438252", "0.4743572" ]
0.7273096
0
NewQueueManagerWithDefaults instantiates a new QueueManager object This constructor will only assign default values to properties that have it defined, but it doesn't guarantee that properties required by API are set
NewQueueManagerWithDefaults создает новый объект QueueManager. Этот конструктор присваивает значения по умолчанию только тем свойствам, которые определены, но он не гарантирует, что свойства, требуемые API, будут заданы.
func NewQueueManagerWithDefaults() *QueueManager { this := QueueManager{} return &this }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewQueueManager(logger log.Logger, cfg config.QueueConfig, externalLabels labels.Labels, relabelConfigs []*relabel.Config, client StorageClient, flushDeadline time.Duration) *QueueManager {\n\tif logger == nil {\n\t\tlogger = log.NewNopLogger()\n\t} else {\n\t\tlogger = log.With(logger, \"queue\", client.Name())\n\t}\n\tt := &QueueManager{\n\t\tlogger: logger,\n\t\tflushDeadline: flushDeadline,\n\t\tcfg: cfg,\n\t\texternalLabels: externalLabels,\n\t\trelabelConfigs: relabelConfigs,\n\t\tclient: client,\n\t\tqueueName: client.Name(),\n\n\t\tlogLimiter: rate.NewLimiter(logRateLimit, logBurst),\n\t\tnumShards: cfg.MinShards,\n\t\treshardChan: make(chan int),\n\t\tquit: make(chan struct{}),\n\n\t\tsamplesIn: newEWMARate(ewmaWeight, shardUpdateDuration),\n\t\tsamplesOut: newEWMARate(ewmaWeight, shardUpdateDuration),\n\t\tsamplesOutDuration: newEWMARate(ewmaWeight, shardUpdateDuration),\n\t}\n\tt.shards = t.newShards(t.numShards)\n\tnumShards.WithLabelValues(t.queueName).Set(float64(t.numShards))\n\tshardCapacity.WithLabelValues(t.queueName).Set(float64(t.cfg.Capacity))\n\n\t// Initialize counter labels to zero.\n\tsentBatchDuration.WithLabelValues(t.queueName)\n\tsucceededSamplesTotal.WithLabelValues(t.queueName)\n\tfailedSamplesTotal.WithLabelValues(t.queueName)\n\tdroppedSamplesTotal.WithLabelValues(t.queueName)\n\n\treturn t\n}", "func NewDefaultClient() QueueClient {\n\treturn &inMemoryQueue{queues: make(map[string][]string)}\n}", "func DefaultQueue(queue string) func(*Locker) error {\n\treturn func(l *Locker) error {\n\t\tl.DefaultQueue = queue\n\t\treturn nil\n\t}\n}", "func (m ManagedConsumerConfig) setDefaults() ManagedConsumerConfig {\n\tif m.NewConsumerTimeout <= 0 {\n\t\tm.NewConsumerTimeout = 5 * time.Second\n\t}\n\tif m.InitialReconnectDelay <= 0 {\n\t\tm.InitialReconnectDelay = 1 * time.Second\n\t}\n\tif m.MaxReconnectDelay <= 0 {\n\t\tm.MaxReconnectDelay = 5 * time.Minute\n\t}\n\t// unbuffered queue not allowed\n\tif m.QueueSize <= 0 {\n\t\tm.QueueSize = 128\n\t}\n\n\treturn m\n}", "func Default() *JobManager {\n\tif _default == nil {\n\t\t_defaultLock.Lock()\n\t\tdefer _defaultLock.Unlock()\n\n\t\tif _default == nil {\n\t\t\t_default = New()\n\t\t}\n\t}\n\treturn _default\n}", "func NewDefault(db *bolt.DB) (q queue.WaitQueue, err error) {\n\treturn New(db, DefaultBucket, DefaultMemQueueSize, DefaultBufSize)\n}", "func DefaultQueue(queue string) func(*Config) error {\n\treturn func(c *Config) error {\n\t\tc.DefaultQueue = queue\n\t\treturn nil\n\t}\n}", "func (obj *RabbitQueue) Default() {\n\trabbitQueueLog.Info(\"default\", \"name\", obj.Name, \"namespace\", obj.Namespace)\n\n\tif obj.Spec.QueueName == \"\" {\n\t\tobj.Spec.QueueName = obj.Name\n\t}\n}", "func NewQueueManager(name string, clusters []string, aliasQueues []AliasQueue, remoteQueues []RemoteQueue, clusterQueues []ClusterQueue, ) *QueueManager {\n\tthis := QueueManager{}\n\tthis.Name = name\n\tthis.Clusters = clusters\n\tthis.AliasQueues = aliasQueues\n\tthis.RemoteQueues = remoteQueues\n\tthis.ClusterQueues = clusterQueues\n\treturn &this\n}", "func DefaultQueueSettings() QueueSettings {\n\treturn QueueSettings{\n\t\tEnabled: true,\n\t\tNumConsumers: 10,\n\t\t// For 5000 queue elements at 100 requests/sec gives about 50 sec of survival of destination outage.\n\t\t// This is a pretty decent value for production.\n\t\t// User should calculate this from the perspective of how many seconds to buffer in case of a backend outage,\n\t\t// multiply that by the number of requests per seconds.\n\t\tQueueSize: 5000,\n\t\tPersistentStorageEnabled: false,\n\t}\n}", "func NewQueueManager(q amboy.Queue) Manager {\n\treturn &queueManager{\n\t\tqueue: q,\n\t}\n}", "func NewDefault(m map[string]interface{}) (share.Manager, error) {\n\tc := &config{}\n\tif err := mapstructure.Decode(m, c); err != nil {\n\t\terr = errors.Wrap(err, \"error creating a new manager\")\n\t\treturn nil, err\n\t}\n\n\ts, err := metadata.NewCS3Storage(c.GatewayAddr, c.ProviderAddr, c.ServiceUserID, c.ServiceUserIdp, c.MachineAuthAPIKey)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tindexer := indexer.CreateIndexer(s)\n\n\tclient, err := pool.GetGatewayServiceClient(c.GatewayAddr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn New(client, s, indexer)\n}", "func NewQueue(maxQueueSize, maxFlowSize uint64, helper Interface) *Queue {\n\tif maxFlowSize > maxQueueSize {\n\t\tpanic(\"MaxFlowSize > MaxQueueSize\")\n\t}\n\n\tif helper == nil {\n\t\tpanic(\"helper is nil\")\n\t}\n\n\tq := new(Queue)\n\tq.cond.L = &q.lock\n\tq.maxQueueSize = maxQueueSize\n\tq.maxFlowSize = maxFlowSize\n\tq.helper = helper\n\tq.flows = make(map[uint64]*flowInfo)\n\n\treturn q\n}", "func New(mqURL string) (models.MessageQueue, error) {\n\tmq, err := newmq(mqURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &metricMQ{mq}, nil\n}", "func NewDefaultMQService() *mqServiceImpl {\n\treturn &mqServiceImpl{}\n}", "func New() *queue {\n\treturn &queue{\n\t\titems: make([]item, DefaultCapacity),\n\t\tcapacity: DefaultCapacity,\n\t}\n}", "func newDefaultPodManager() *podManager {\n\treturn &podManager{\n\t\trunningPods: make(map[string]*runningPod),\n\t\trequests: make(chan *cniserver.PodRequest, 20),\n\t\treattachPods: make(map[string]*corev1.Pod),\n\t}\n}", "func NewDefaultManager() *Manager {\n\tm := NewManager()\n\n\t// default config\n\tm.SetAuthorizeCodeExp(time.Minute * 10)\n\tm.SetImplicitTokenCfg(&Config{AccessTokenExp: time.Hour * 1})\n\tm.SetClientTokenCfg(&Config{AccessTokenExp: time.Hour * 2})\n\tm.SetAuthorizeCodeTokenCfg(&Config{IsGenerateRefresh: true, AccessTokenExp: time.Hour * 2, RefreshTokenExp: time.Hour * 24 * 3})\n\tm.SetPasswordTokenCfg(&Config{IsGenerateRefresh: true, AccessTokenExp: time.Hour * 2, RefreshTokenExp: time.Hour * 24 * 7})\n\n\tm.MapTokenModel(models.NewToken())\n\tm.MapAuthorizeGenerate(generates.NewAuthorizeGenerate())\n\tm.MapAccessGenerate(generates.NewAccessGenerate())\n\n\treturn m\n}", "func NewQueue(maximumCapacity int, initialCapacity int, factory TokenFactory) *Queue {\n\tq := &Queue{\n\t\tmaxCapacity: maximumCapacity,\n\t\tavailableTokens: make(chan (Token), maximumCapacity),\n\t\tcommittedTokens: make(chan (Token), maximumCapacity),\n\t\tdiscardTokens: make(chan (Token), maximumCapacity),\n\t\tcloseTokens: make(chan (Token)),\n\t}\n\n\tfor i := 0; i < maximumCapacity; i++ {\n\t\ttoken := factory()\n\t\tif token == nil {\n\t\t\treturn nil\n\t\t}\n\t\tq.discardTokens <- token\n\t\tq.validTokens = append(q.validTokens, token)\n\t}\n\n\tq.EnableDisableTokens(initialCapacity)\n\n\treturn q\n}", "func NewBasicMockMessageQueue() lanternmq.MessageQueue {\n\tmq := BasicMockMessageQueue{}\n\tmq.Queue = make(chan []byte, 20)\n\n\tmq.ConnectFn = func(username string, password string, host string, port string) error {\n\t\treturn nil\n\t}\n\n\tmq.CreateChannelFn = func() (lanternmq.ChannelID, error) {\n\t\treturn 1, nil\n\t}\n\n\tmq.NumConcurrentMsgsFn = func(chID lanternmq.ChannelID, num int) error {\n\t\treturn nil\n\t}\n\n\tmq.QueueExistsFn = func(chId lanternmq.ChannelID, qName string) (bool, error) {\n\t\treturn true, nil\n\t}\n\n\tmq.DeclareQueueFn = func(chID lanternmq.ChannelID, name string) error {\n\t\treturn nil\n\t}\n\n\tmq.PublishToQueueFn = func(chID lanternmq.ChannelID, qName string, message string) error {\n\t\tif len(mq.Queue) < 20 {\n\t\t\tmq.Queue <- []byte(message)\n\t\t} else {\n\t\t\treturn errors.New(\"queue full - unable to add new message\")\n\t\t}\n\t\treturn nil\n\t}\n\n\tmq.ConsumeFromQueueFn = func(chID lanternmq.ChannelID, qName string) (lanternmq.Messages, error) {\n\t\treturn nil, nil\n\t}\n\n\tmq.ProcessMessagesFn = func(ctx context.Context, msgs lanternmq.Messages, handler lanternmq.MessageHandler, args *map[string]interface{}, errs chan<- error) {\n\t\tfor msg := range mq.Queue {\n\t\t\terr := handler(msg, args)\n\t\t\tif err != nil {\n\t\t\t\terrs <- err\n\t\t\t}\n\t\t}\n\t}\n\n\tmq.CloseFn = func() {}\n\treturn &mq\n}", "func New() *Queue {\r\n\treturn &Queue{nil,nil,0}\r\n}", "func New(mqURL string) (models.MessageQueue, error) {\n\t// Play with URL schemes here: https://play.golang.org/p/xWAf9SpCBW\n\tu, err := url.Parse(mqURL)\n\tif err != nil {\n\t\tlogrus.WithError(err).WithFields(logrus.Fields{\"url\": mqURL}).Fatal(\"bad MQ URL\")\n\t}\n\tlogrus.WithFields(logrus.Fields{\"mq\": u.Scheme}).Debug(\"selecting MQ\")\n\tswitch u.Scheme {\n\tcase \"memory\":\n\t\treturn NewMemoryMQ(), nil\n\tcase \"redis\":\n\t\treturn NewRedisMQ(u)\n\tcase \"bolt\":\n\t\treturn NewBoltMQ(u)\n\t}\n\tif strings.HasPrefix(u.Scheme, \"ironmq\") {\n\t\treturn NewIronMQ(u), nil\n\t}\n\n\treturn nil, fmt.Errorf(\"mq type not supported %v\", u.Scheme)\n}", "func NewQueue() *Queue {\n return &Queue{member: make([]interface{}, 0)}\n}", "func New() Manager {\n\treturn Manager{\n\t\tState: make(map[string]string),\n\t\tClientHolder: make(map[string]utils.Set),\n\t\tClientQueue: make(map[string]utils.Queue),\n\t}\n}", "func NewManager(logger logging.Logger) SessionManager {\n\treturn &defaultSessionManager{\n\t\tlogger: logger,\n\t\ttasks: make(map[string]exec.Execer),\n\n\t\tquit: make(chan struct{}),\n\t}\n}", "func (queue *Queue) Init() (err error) {\n\tclient := queue.GetClient()\n\n\tparams := &sqs.CreateQueueInput{\n\t\tQueueName: aws.String(queue.Name + deadLetterQueueSuffix),\n\t\tAttributes: map[string]*string{\n\t\t\t\"MessageRetentionPeriod\": aws.String(\"1209600\"),\n\t\t},\n\t}\n\tresp, err := client.CreateQueue(params)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"queueName\": queue.Name,\n\t\t\t\"error\": err,\n\t\t}).Error(\"Createing the dead letter queue\")\n\t\treturn\n\t}\n\n\tqueue.DeadLetterQueueURL = *resp.QueueUrl\n\tlog.WithFields(log.Fields{\n\t\t\"QueueUrl\": queue.DeadLetterQueueURL,\n\t}).Info(\"Dead Letter Queue initialized\")\n\n\tqueueArnAttributeName := \"QueueArn\"\n\tdeadLetterQueueAttributes, err := queue.GetAttributesByQueueURL(queue.DeadLetterQueueURL, []*string{&queueArnAttributeName})\n\tif err != nil {\n\t\treturn\n\t}\n\tredrivePolicy := &RedrivePolicy{\n\t\tMaxReceiveCount: MaxReceiveCountBeforeDead,\n\t\tDeadLetterTargetArn: *deadLetterQueueAttributes.Attributes[queueArnAttributeName],\n\t}\n\tredrivePolicyString, err := redrivePolicy.GetAsAWSString()\n\tif err != nil {\n\t\treturn\n\t}\n\tparams = &sqs.CreateQueueInput{\n\t\tQueueName: aws.String(queue.Name),\n\t\tAttributes: map[string]*string{\n\t\t\t\"RedrivePolicy\": redrivePolicyString,\n\t\t\t\"MessageRetentionPeriod\": aws.String(\"1209600\"),\n\t\t},\n\t}\n\tresp, err = client.CreateQueue(params)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"queueName\": queue.Name,\n\t\t\t\"error\": err,\n\t\t}).Error(\"Createing the queue\")\n\t\treturn\n\t}\n\n\tqueue.URL = *resp.QueueUrl\n\tlog.WithFields(log.Fields{\n\t\t\"QueueUrl\": queue.URL,\n\t}).Info(\"Queue initialized\")\n\n\treturn\n}", "func New() *Queue {\n\treturn &Queue{nil, nil, 0}\n}", "func New() *Queue {\n\treturn &Queue{nil, nil, 0}\n}", "func MyQueueConstructor() MyQueue {\n\treturn MyQueue{}\n}", "func New() *JobManager {\n\tjm := JobManager{\n\t\theartbeatInterval: DefaultHeartbeatInterval,\n\t\tjobs: map[string]*JobMeta{},\n\t\ttasks: map[string]*TaskMeta{},\n\t}\n\tjm.schedulerWorker = async.NewInterval(jm.runDueJobs, DefaultHeartbeatInterval)\n\tjm.killHangingTasksWorker = async.NewInterval(jm.killHangingTasks, DefaultHeartbeatInterval)\n\treturn &jm\n}", "func WithDefaultMaxTries(n int) Option {\n\treturn func(q *Queue) {\n\t\tq.maxTries = n\n\t}\n}", "func NewDefaultRestrictionManager(maxValueLength int) *DefaultRestrictionManager {\n\tif maxValueLength == 0 {\n\t\tmaxValueLength = defaultMaxValueLength\n\t}\n\treturn &DefaultRestrictionManager{\n\t\tdefaultRestriction: &Restriction{keyAllowed: true, maxValueLength: maxValueLength},\n\t}\n}", "func (manager *Manager) SetDefaults() {\n\tmanager.viperConfig.SetDefault(workspace, \"No name\")\n\n\tmanager.viperConfig.SetDefault(tcpAddress, \"localhost:8888\")\n\tmanager.viperConfig.SetDefault(tcpConnectionType, \"tcp\")\n\n\tmanager.viperConfig.SetDefault(httpAddress, \":8080\")\n\n\tmanager.viperConfig.SetDefault(shutdownTimeout, 15*time.Second)\n\tmanager.viperConfig.SetDefault(readTimeout, 10*time.Second)\n\tmanager.viperConfig.SetDefault(writeTimeout, 10*time.Second)\n\n\tmanager.viperConfig.SetDefault(websocketReadBufferSize, 1024)\n\tmanager.viperConfig.SetDefault(websocketWriteBufferSize, 1024)\n\tmanager.viperConfig.SetDefault(websocketMaxMessageSize, 512)\n\tmanager.viperConfig.SetDefault(websocketWriteWait, 10*time.Second)\n\tmanager.viperConfig.SetDefault(websocketPongWait, 60*time.Second)\n\tmanager.viperConfig.SetDefault(websocketPingPeriod, 60*0.9*time.Second)\n\n\tmanager.viperConfig.SetDefault(httpTimeout, 1*time.Second)\n\n\tmanager.viperConfig.SetDefault(logLevel, \"debug\")\n}", "func NewQueue() Queue {\n\treturn Queue{}\n}", "func NewQueue(name string) *Queue {\n\tredisClient := GetRedisClientFromConfig()\n\tqueue := &Queue{Name: name, RedisClient: redisClient}\n\treturn queue\n}", "func NewQueue(name string, itemType reflect.Type, maxQueueSize uint32) Queue {\n\tq := queue{\n\t\tname: name,\n\t\titemType: itemType,\n\t\tchannel: make(chan interface{}, maxQueueSize),\n\t}\n\treturn &q\n}", "func setupManager(username string, password string, brokerIp string, brokerPort int, manager *Manager, exchange string, queueName string) error {\n\tamqpURI := getAmqpUri(username, password, brokerIp, brokerPort)\n\tmanager.logger.Debugf(\"dialing %s\", amqpURI)\n\tvar err error\n\tmanager.Connection, err = amqp.Dial(amqpURI)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tmanager.logger.Debugf(\"got Connection, getting Channel\")\n\tmanager.Channel, err = manager.Connection.Channel()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tmanager.logger.Debugf(\"got Channel, declaring Exchange (%q)\", exchange)\n\n\tmanager.logger.Debugf(\"declared Exchange, declaring Queue %q\", queueName)\n\tqueue, err := manager.Channel.QueueDeclare(\n\t\tqueueName,\n\t\ttrue,\n\t\ttrue,\n\t\tfalse,\n\t\tfalse,\n\t\tnil,\n\t)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tmanager.logger.Debugf(\"declared Queue (%q, %d messages, %d consumers), binding to Exchange\",\n\t\tqueue.Name, queue.Messages, queue.Consumers)\n\n\tif err = manager.Channel.QueueBind(\n\t\tqueue.Name, // name of the queue\n\t\tqueue.Name, // bindingKey\n\t\texchange, // sourceExchange\n\t\tfalse, // noWait\n\t\tnil, // arguments\n\t); err != nil {\n\t\treturn err\n\t}\n\n\tmanager.logger.Debug(\"Queue bound to Exchange, starting Consume\")\n\treturn nil\n}", "func NewQueue(maxWorkers int, maxQueue int) *Queue {\n\tq := make(chan Job, maxQueue)\n\treturn &Queue{\n\t\tq,\n\t\ttrue,\n\t\t&Dispatcher{\n\t\t\tjobQueue: q,\n\t\t\tworkerPool: make(chan chan Job, maxWorkers),\n\t\t\tMaxWorkers: maxWorkers,\n\t\t},\n\t}\n}", "func LocalQueueFactory(size int) Option {\n\treturn func(env Environment) {\n\t\tif size < 5 {\n\t\t\tsize = 5\n\t\t}\n\t\tQueueFactory(makeLocalEventQueueFactory(size))\n\t}\n}", "func NewFakeQueueDispatcher() (dispatcher *FakeQueueDispatcher) {\n\tdispatcher = &FakeQueueDispatcher{}\n\tdispatcher.Messages = make([]interface{}, 0)\n\treturn\n}", "func setupDefaults() {\n\tclient.DefaultClient = grpcCli.NewClient()\n\tserver.DefaultServer = grpcSvr.NewServer()\n\tnetwork.DefaultNetwork = mucpNet.NewNetwork()\n\tmetrics.DefaultMetricsReporter = noopMet.New()\n\n\t// setup rpc implementations after the client is configured\n\tauth.DefaultAuth = authSrv.NewAuth()\n\tbroker.DefaultBroker = brokerSrv.NewBroker()\n\tevents.DefaultStream = eventsSrv.NewStream()\n\tevents.DefaultStore = eventsSrv.NewStore()\n\tregistry.DefaultRegistry = registrySrv.NewRegistry()\n\trouter.DefaultRouter = routerSrv.NewRouter()\n\tstore.DefaultStore = storeSrv.NewStore()\n\tstore.DefaultBlobStore = storeSrv.NewBlobStore()\n\truntime.DefaultRuntime = runtimeSrv.NewRuntime()\n}", "func SetDefault(jm *JobManager) {\n\t_defaultLock.Lock()\n\t_default = jm\n\t_defaultLock.Unlock()\n}", "func NewDefaultPriorityQueue() *PriorityQueue {\n\treturn NewPriorityQueue(func(interface{}) bool { return false })\n}", "func NewManager(h Handler,\n\tusername string,\n\tpassword string,\n\tbrokerIp string,\n\tbrokerPort int,\n\texchange string,\n\tqueueName string,\n\tworkers int,\n\tallocate bool,\n\tmanagerName string,\n\thandleFunction handlerFunction,\n\tlogLevel string,\n\tnet catalogue.BaseNetworkInt,\n\timg catalogue.BaseImageInt) (*Manager, error) {\n\n\tmanager := &Manager{\n\t\tConnection: nil,\n\t\tChannel: nil,\n\t\tallocate: allocate,\n\t\tworkers: workers,\n\t\terrorChan: make(chan error),\n\t\tlogger: GetLogger(managerName, logLevel),\n\t\thandlerFunction: handleFunction,\n\t\thandler: h,\n\t\timage: img,\n\t\tnetwork: net,\n\t}\n\n\terr := setupManager(username, password, brokerIp, brokerPort, manager, exchange, queueName)\n\tif err != nil {\n\t\tmanager.logger.Errorf(\"Error while setup the amqp thing: %v\", err)\n\t\treturn nil, err\n\t}\n\tmanager.queueName = queueName\n\treturn manager, nil\n}", "func NewQueue(ctx context.Context, queueID string, db *sql.DB, conf QueueConfig) (*Queue, error) {\n\tq := &Queue{ID: queueID}\n\tq.repo = repo.NewRepository(db)\n\tq.PollRate = 100 * time.Millisecond // Default\n\tq.queueSize = 10000 // Default\n\tq.retries = 3 // Default\n\tq.IsMultiQueue = conf.IsMultiQueue\n\tq.baseDelay = 3 * time.Second // Default\n\n\tif conf.PollingRate > 0 {\n\t\tq.PollRate = conf.PollingRate\n\t}\n\tif conf.Qsize > 0 {\n\t\tq.queueSize = conf.Qsize\n\t}\n\tif conf.BaseDelay > 0 {\n\t\tq.baseDelay = conf.BaseDelay\n\t}\n\tif conf.Retries >= 0 {\n\t\tq.retries = conf.Retries\n\t}\n\t// Multilevel Queue/channel created\n\ttemp := mlQueue{}\n\ttemp.notifier = make([]chan JobChan, 1)\n\ttemp.notifier[0] = make(chan JobChan, q.queueSize)\n\ttemp.total = 1\n\tq.mq = temp\n\n\tm := make(map[string][]worker.Worker)\n\tq.workers = m\n\tvar wg sync.WaitGroup\n\tq.wg = &wg\n\n\t// resume stopped jobs\n\terr := q.ResumePendingJobs(ctx)\n\tif err != nil {\n\t\tlogger.Log.Error(\"Unable to resume jobs from bucket: %s\", zap.Error(err))\n\t\t// Don't fail out, this isn't really fatal. But maybe it should be?\n\t}\n\treturn q, nil\n}", "func New(name string, c config.Config) *Queue {\n\treturn &Queue{\n\t\tname: name,\n\t\tconf: c,\n\t}\n}", "func New() Queue {\n\treturn Queue{list: linkedlist.New()}\n}", "func NewDefaultManager(\n\tdemands *cache.SafeDemandCache,\n\tbinpacker *binpacker.Binpacker,\n\tinstanceGroupLabel string) Manager {\n\treturn &defaultManager{\n\t\tdemands: demands,\n\t\tbinpacker: binpacker,\n\t\tinstanceGroupLabel: instanceGroupLabel,\n\t}\n}", "func NewQueue(action WorkAction, options ...QueueOption) *Queue {\n\tq := Queue{\n\t\tLatch: NewLatch(),\n\t\tAction: action,\n\t\tContext: context.Background(),\n\t\tMaxWork: DefaultQueueMaxWork,\n\t\tParallelism: runtime.NumCPU(),\n\t}\n\tfor _, option := range options {\n\t\toption(&q)\n\t}\n\treturn &q\n}", "func NewDefaults() map[string]interface{} {\n\tdefaults := make(map[string]interface{})\n\n\tdefaults[authPostgresURI] = \"postgresql://postgres:postgres@localhost:5432/test?sslmode=disable\"\n\tdefaults[authMigrationVersion] = 0\n\n\tdefaults[gatewayAddr] = \":10000\"\n\tdefaults[gatewayEndpoint] = \"/graphql\"\n\tdefaults[gatewayServePlayground] = true\n\tdefaults[gatewayPlaygroundEndpoint] = \"/playground\"\n\tdefaults[gatewayEnableIntrospection] = true\n\n\tdefaults[seedUserLogin] = \"root\"\n\tdefaults[seedUserPassword] = \"root\"\n\tdefaults[seedRoleTitle] = \"ROOT\"\n\tdefaults[seedRoleSuper] = true\n\n\tdefaults[sessionAccessTokenTTL] = 1000000\n\tdefaults[sessionRefreshTokenTTl] = 5000000\n\n\treturn defaults\n}", "func New(opt *Options) *Queue {\n\tif client == nil {\n\t\tredisOpt := &redis.Options{\n\t\t\tAddr: opt.Connection.Addr,\n\t\t\tPassword: opt.Connection.Password,\n\t\t\tDB: opt.Connection.DB,\n\t\t\tMaxRetries: opt.Connection.MaxRetries,\n\t\t\tDialTimeout: opt.Connection.DialTimeout,\n\t\t\tReadTimeout: opt.Connection.ReadTimeout,\n\t\t\tWriteTimeout: opt.Connection.WriteTimeout,\n\t\t\tPoolSize: opt.Connection.PoolSize,\n\t\t\tPoolTimeout: opt.Connection.PoolTimeout,\n\t\t\tIdleTimeout: opt.Connection.IdleTimeout,\n\t\t}\n\t\tclient = redis.NewClient(redisOpt)\n\t}\n\n\treturn &Queue{\n\t\tjobChannel: make(chan string, 1000),\n\t\tconcurrency: opt.Concurrency,\n\t\tqueueName: opt.QueueName,\n\t\tprocessor: opt.Processor,\n\t\terrorHandler: opt.ErrorHandler,\n\t}\n}", "func (o *SendJobCommandParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func New() *Queue {\n\tq := new(Queue)\n\tq.length = 0\n\tq.s1 = stack.New()\n\tq.s2 = stack.New()\n\n\treturn q\n}", "func NewQueue() *Queue {\n\treturn &Queue{nil, nil, 0}\n}", "func defaultConsumerOptions() *consumerOptions {\n\treturn &consumerOptions{\n\t\tqueueDepth: 10000,\n\t\tconcurrency: 10,\n\t\tStats: &NilConsumerStatsCollector{},\n\t}\n}", "func NewQueue(action func(interface{}) error) *QueueWorker {\n\treturn &QueueWorker{\n\t\taction: action,\n\t\tlatch: &Latch{},\n\t\tmaxWork: DefaultQueueWorkerMaxWork,\n\t}\n}", "func (o *GetGPUArchitectureParams) SetDefaults() {\n\tvar (\n\t\tallowUnstableDefault = bool(false)\n\t)\n\n\tval := GetGPUArchitectureParams{\n\t\tAllowUnstable: &allowUnstableDefault,\n\t}\n\n\tval.timeout = o.timeout\n\tval.Context = o.Context\n\tval.HTTPClient = o.HTTPClient\n\t*o = val\n}", "func NewQueue(l int) *Queue {\n\tif l == -1 {\n\t\treturn &Queue{\n\t\t\tQueue: make([]types.Event, 0),\n\t\t\tL: int(^uint(0) >> 1), // max integer value, architecture independent\n\t\t}\n\t}\n\tq := &Queue{\n\t\tQueue: make([]types.Event, 0, l),\n\t\tL: l,\n\t}\n\tlog.WithFields(log.Fields{\"Capacity\": q.L}).Debugf(\"Creating queue\")\n\treturn q\n}", "func NewDefaults() *Client {\n\treturn &Client{\n\t\tsigkil: make(chan os.Signal, 1),\n\t\tsighup: make(chan os.Signal, 1),\n\t\tmenu: make(map[string]ui.MenuItem),\n\t\tplex: &logs.Timer{},\n\t\talert: &logs.Cooler{},\n\t\tLogger: logs.New(),\n\t\tConfig: &configfile.Config{\n\t\t\tApps: &apps.Apps{\n\t\t\t\tURLBase: \"/\",\n\t\t\t},\n\t\t\tServices: &services.Config{\n\t\t\t\tInterval: cnfg.Duration{Duration: services.DefaultSendInterval},\n\t\t\t\tParallel: 1,\n\t\t\t},\n\t\t\tBindAddr: configfile.DefaultBindAddr,\n\t\t\tSnapshot: &snapshot.Config{\n\t\t\t\tTimeout: cnfg.Duration{Duration: snapshot.DefaultTimeout},\n\t\t\t},\n\t\t\tLogs: &logs.Logs{\n\t\t\t\tLogFiles: DefaultLogFiles,\n\t\t\t\tLogFileMb: DefaultLogFileMb,\n\t\t\t},\n\t\t\tTimeout: cnfg.Duration{Duration: configfile.DefaultTimeout},\n\t\t}, Flags: &Flags{\n\t\t\tFlagSet: flag.NewFlagSet(DefaultName, flag.ExitOnError),\n\t\t\tConfigFile: os.Getenv(DefaultEnvPrefix + \"_CONFIG_FILE\"),\n\t\t\tEnvPrefix: DefaultEnvPrefix,\n\t\t},\n\t}\n}", "func New(hint int) *Queue {\n\treturn &Queue{\n\t\titems: make([]interface{}, 0, hint),\n\t}\n}", "func initMailQueueProducer() (err error) {\n\tnsqCfg := nsq.NewConfig()\n\tnsqCfg.UserAgent = \"tmail.queue\"\n\tNsqQueueProducer, err = nsq.NewProducer(\"127.0.0.1:4150\", nsqCfg)\n\tif Cfg.GetDebugEnabled() {\n\t\tNsqQueueProducer.SetLogger(Log, 0)\n\t} else {\n\t\tNsqQueueProducer.SetLogger(Log, 4)\n\t}\n\treturn err\n}", "func New(name string) (*Queue, error) {\n\tqueue := Queue{Name: name}\n\terr := queue.Init()\n\n\treturn &queue, err\n}", "func New(maxSize int, dropBehavior DropBehavior) *Queue {\n\treturn &Queue{\n\t\tmaxSize: maxSize,\n\t\tdropBehavior: dropBehavior,\n\t}\n}", "func New(delegate Delegate, settings Settings) (*Queue, error) {\n\tconst op = \"pq/new\"\n\n\tif delegate == nil {\n\t\treturn nil, errOp(op).of(InvalidParam).report(\"delegate must not be nil\")\n\t}\n\n\taccessor, errKind := makeAccess(delegate)\n\tif errKind != NoError {\n\t\treturn nil, errOp(op).of(errKind)\n\t}\n\n\tpageSize := delegate.PageSize()\n\n\tq := &Queue{\n\t\taccessor: accessor,\n\t\tsettings: settings,\n\t\tpagePool: newPagePool(pageSize),\n\t}\n\n\t// use pointer address as ID for correlating error messages\n\tq.id = queueID(uintptr(unsafe.Pointer(q)))\n\taccessor.quID = q.id\n\n\trootBuf, err := q.accessor.ReadRoot()\n\tif err != nil {\n\t\treturn nil, wrapErr(op, err).of(InitFailed).\n\t\t\treport(\"failed to read queue header\")\n\t}\n\n\troot := castQueueRootPage(rootBuf[:])\n\tif root.version.Get() != queueVersion {\n\t\tcause := &Error{\n\t\t\tkind: InitFailed,\n\t\t\tmsg: fmt.Sprintf(\"queue version %v\", root.version.Get()),\n\t\t}\n\t\treturn nil, wrapErr(op, cause).of(InitFailed)\n\t}\n\n\ttracef(\"open queue: %p (pageSize: %v)\\n\", q, pageSize)\n\ttraceQueueHeader(root)\n\n\tq.version = root.version.Get()\n\tq.hdrOffset = q.accessor.RootFileOffset()\n\tq.onInit()\n\treturn q, nil\n}", "func NewPrinterDefaults()(*PrinterDefaults) {\n m := &PrinterDefaults{\n }\n m.backingStore = ie8677ce2c7e1b4c22e9c3827ecd078d41185424dd9eeb92b7d971ed2d49a392e.BackingStoreFactoryInstance();\n m.SetAdditionalData(make(map[string]any))\n return m\n}", "func (r *yandexMessageQueueReconciler) SetupWithManager(mgr ctrl.Manager) error {\n\treturn ctrl.NewControllerManagedBy(mgr).\n\t\tFor(&connectorsv1.YandexMessageQueue{}).\n\t\tComplete(r)\n}", "func (o *RTRCheckAdminCommandStatusParams) SetDefaults() {\n\tvar (\n\t\tsequenceIDDefault = int64(0)\n\t)\n\n\tval := RTRCheckAdminCommandStatusParams{\n\t\tSequenceID: sequenceIDDefault,\n\t}\n\n\tval.timeout = o.timeout\n\tval.Context = o.Context\n\tval.HTTPClient = o.HTTPClient\n\t*o = val\n}", "func New(cfg Config, pubSub pubSub, metrics metricsProvider) (*Queue, error) {\n\tmsgChan, err := pubSub.SubscribeWithOpts(context.Background(), topic, spi.WithPool(cfg.PoolSize))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"subscribe to topic [%s]: %w\", topic, err)\n\t}\n\n\tq := &Queue{\n\t\tpubSub: pubSub,\n\t\tmsgChan: msgChan,\n\t\tjsonMarshal: json.Marshal,\n\t\tjsonUnmarshal: json.Unmarshal,\n\t\tmetrics: metrics,\n\t}\n\n\tq.Lifecycle = lifecycle.New(\"operation-queue\",\n\t\tlifecycle.WithStart(q.start),\n\t\tlifecycle.WithStop(q.stop),\n\t)\n\n\tq.Start()\n\n\treturn q, nil\n}", "func NewQueue() Queue {\r\n\tvar empty []int\r\n\treturn Queue{empty, len(empty)}\r\n}", "func New() *Queue {\n\titems := []*item.Item{}\n\tlock := &sync.Mutex{}\n\treturn &Queue{items, lock}\n}", "func NewQueue(\n\tservers []string,\n\topts QueueOptions,\n) (Queue, error) {\n\tq, err := newQueue(servers, opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tq.initConnections(servers)\n\tgo q.reportMetrics()\n\n\treturn q, nil\n}", "func NewQueue(storage Storage, reQueueTimeout time.Duration) Queue {\n\tif reQueueTimeout < 1 {\n\t\treQueueTimeout = time.Minute * 30\n\t}\n\n\tname := \"gocelery\"\n\tq := &queue{\n\t\tstorage: storage,\n\t\thead: 0,\n\t\ttail: 0,\n\t\trequeueTimeout: reQueueTimeout,\n\t\tqueuePrefix: fmt.Sprintf(\"%s-queue-\", name),\n\t\tqueueAckPrefix: fmt.Sprintf(\"%s-ack-\", name),\n\t}\n\n\t// restore the old state from the DB\n\tq.loadHeadTail()\n\treturn q\n}", "func (s *Store) CreateQueue(name string, overriddenSettings ...QueueSetting) (QueueMeta, QueueSettings, error) {\n\tif !isValidQueueName(name) {\n\t\treturn QueueMeta{}, QueueSettings{}, ErrInvalidQueueName\n\t}\n\n\tmeta := QueueMeta{Name: name, Created: time.Now()}\n\tsettings := defaultQueueSettings()\n\n\tfor _, setting := range overriddenSettings {\n\t\tif err := setting(&settings); err != nil {\n\t\t\treturn QueueMeta{}, QueueSettings{}, err\n\t\t}\n\t}\n\n\treturn meta, settings, s.db.Update(func(tx *bolt.Tx) error {\n\t\tqueues := tx.Bucket([]byte(\"Queues\"))\n\n\t\tbucket, err := queues.CreateBucket([]byte(name))\n\t\tif err != nil {\n\t\t\tif err == bolt.ErrBucketExists {\n\t\t\t\treturn ErrQueueExists\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\n\t\t// Meta\n\n\t\tmetaBucket, err := bucket.CreateBucketIfNotExists([]byte(\"Meta\"))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err = metaBucket.Put([]byte(\"Name\"), []byte(name)); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err = metaBucket.Put([]byte(\"Created\"), encodeTime(meta.Created)); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Settings\n\n\t\tsettingsBucket, err := bucket.CreateBucketIfNotExists([]byte(\"Settings\"))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err = settingsBucket.Put([]byte(\"LeaseDuration\"), encodeInt(settings.LeaseDuration)); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err = settingsBucket.Put([]byte(\"MessageRetentionPeriod\"), encodeInt(settings.MessageRetentionPeriod)); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err = settingsBucket.Put([]byte(\"DelaySeconds\"), encodeInt(settings.DelaySeconds)); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Message Buckets\n\n\t\tmessages, err := bucket.CreateBucketIfNotExists([]byte(\"Messages\"))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif _, err := messages.CreateBucketIfNotExists([]byte(\"Visible\")); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif _, err := messages.CreateBucketIfNotExists([]byte(\"Leased\")); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif _, err := messages.CreateBucketIfNotExists([]byte(\"Delayed\")); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t})\n}", "func DefaultQueueKeysFunc(_ runtime.Object) []string {\n\treturn []string{DefaultQueueKey}\n}", "func New(cfg config.Queue, n notifier) *Queue {\n\tq := &Queue{\n\t\taddCh: make(chan struct{}, cfg.QueueSize),\n\t\tpopCh: make(chan struct{}, cfg.GoRoutinesSize),\n\t\taddMessage: make(chan entity.NotifierMessage, 1),\n\t\tpopMessage: make(chan entity.NotifierMessage, 1),\n\t\tnotifier: n,\n\t}\n\n\tgo q.pop()\n\tgo q.add()\n\n\treturn q\n}", "func New(cb Done, transport http.RoundTripper) *Manager {\n\treturn &Manager{\n\t\tkeys: sets.NewString(),\n\t\tcb: cb,\n\t\ttransport: transport,\n\t}\n}", "func newDefaultContainerConfig() ContainerConfig {\n\treturn ContainerConfig{\n\t\tCPU: newMinMaxAllocation(),\n\t\tMemory: newMinMaxAllocation(),\n\t\tBlockRead: newMinMaxAllocation(),\n\t\tBlockWrite: newMinMaxAllocation(),\n\t\tNetworkRx: newMinMaxAllocation(),\n\t\tNetworkTx: newMinMaxAllocation(),\n\t}\n}", "func NewQueue(ctx *pulumi.Context,\n\tname string, args *QueueArgs, opts ...pulumi.ResourceOption) (*Queue, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.HoursOfOperationArn == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'HoursOfOperationArn'\")\n\t}\n\tif args.InstanceArn == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'InstanceArn'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Queue\n\terr := ctx.RegisterResource(\"aws-native:connect:Queue\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (o *ListHetznerSizesParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (t *OpenconfigQos_Qos_Queues) NewQueue(Name string) (*OpenconfigQos_Qos_Queues_Queue, error){\n\n\t// Initialise the list within the receiver struct if it has not already been\n\t// created.\n\tif t.Queue == nil {\n\t\tt.Queue = make(map[string]*OpenconfigQos_Qos_Queues_Queue)\n\t}\n\n\tkey := Name\n\n\t// Ensure that this key has not already been used in the\n\t// list. Keyed YANG lists do not allow duplicate keys to\n\t// be created.\n\tif _, ok := t.Queue[key]; ok {\n\t\treturn nil, fmt.Errorf(\"duplicate key %v for list Queue\", key)\n\t}\n\n\tt.Queue[key] = &OpenconfigQos_Qos_Queues_Queue{\n\t\tName: &Name,\n\t}\n\n\treturn t.Queue[key], nil\n}", "func GetDefaultManager() *Manager {\n\treturn defaultManager\n}", "func (o *GetFqdnCacheParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func NewAPIRequestManager() *APIRequestManager {\n\treturn &APIRequestManager{\n\t\tqueue: make(chan *WorkerItem, 10),\n\t}\n}", "func New() *Queue {\r\n\treturn &Queue{\r\n\t\tdata: []int{},\r\n\t}\r\n}", "func setDefault(c *Config) {\n\tc.Token = \"\"\n\tc.GuildID = \"\"\n}", "func New(ctx context.Context, cfg models.Config) (*Queue, error) {\n\tconn, err := connect(ctx, cfg)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to connect to RabbitMQ \")\n\t}\n\n\tch, err := conn.Channel()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to open a channel \")\n\t}\n\n\t_, err = ch.QueueDeclare(\"ItemQueue\", false, false, false, false, nil)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to declare a queue \")\n\t}\n\n\treturn &Queue{ch, conn}, nil\n}", "func (c *InitManagementClusterInput) Defaults(ctx context.Context) {\n\tc.Config.Defaults()\n\tif c.Scheme == nil {\n\t\tc.Scheme = runtime.NewScheme()\n\t}\n\tif c.NewManagementClusterFn == nil {\n\t\tc.NewManagementClusterFn = func() (ManagementCluster, error) {\n\t\t\treturn kind.NewCluster(ctx, c.ManagementClusterName, c.Scheme)\n\t\t}\n\t}\n}", "func (c *InitManagementClusterInput) Defaults(ctx context.Context) {\n\tc.Config.Defaults()\n\tif c.Scheme == nil {\n\t\tc.Scheme = runtime.NewScheme()\n\t}\n\tif c.NewManagementClusterFn == nil {\n\t\tc.NewManagementClusterFn = func() (ManagementCluster, error) {\n\t\t\treturn kind.NewCluster(ctx, c.ManagementClusterName, c.Scheme)\n\t\t}\n\t}\n}", "func NewSmsTrackingWithDefaults() *SmsTracking {\n\tthis := SmsTracking{}\n\treturn &this\n}", "func (cc *ConstructionCreate) defaults() {\n\tif _, ok := cc.mutation.RawProduction(); !ok {\n\t\tv := construction.DefaultRawProduction\n\t\tcc.mutation.SetRawProduction(v)\n\t}\n\tif _, ok := cc.mutation.Production(); !ok {\n\t\tv := construction.DefaultProduction\n\t\tcc.mutation.SetProduction(v)\n\t}\n\tif _, ok := cc.mutation.GetType(); !ok {\n\t\tv := construction.DefaultType\n\t\tcc.mutation.SetType(v)\n\t}\n\tif _, ok := cc.mutation.Level(); !ok {\n\t\tv := construction.DefaultLevel\n\t\tcc.mutation.SetLevel(v)\n\t}\n\tif _, ok := cc.mutation.Modifier(); !ok {\n\t\tv := construction.DefaultModifier\n\t\tcc.mutation.SetModifier(v)\n\t}\n\tif _, ok := cc.mutation.LastUpdated(); !ok {\n\t\tv := construction.DefaultLastUpdated()\n\t\tcc.mutation.SetLastUpdated(v)\n\t}\n\tif _, ok := cc.mutation.NeedRefresh(); !ok {\n\t\tv := construction.DefaultNeedRefresh\n\t\tcc.mutation.SetNeedRefresh(v)\n\t}\n}", "func (t *OpenconfigQos_Qos_Interfaces_Interface_Input_Queues) NewQueue(Name string) (*OpenconfigQos_Qos_Interfaces_Interface_Input_Queues_Queue, error){\n\n\t// Initialise the list within the receiver struct if it has not already been\n\t// created.\n\tif t.Queue == nil {\n\t\tt.Queue = make(map[string]*OpenconfigQos_Qos_Interfaces_Interface_Input_Queues_Queue)\n\t}\n\n\tkey := Name\n\n\t// Ensure that this key has not already been used in the\n\t// list. Keyed YANG lists do not allow duplicate keys to\n\t// be created.\n\tif _, ok := t.Queue[key]; ok {\n\t\treturn nil, fmt.Errorf(\"duplicate key %v for list Queue\", key)\n\t}\n\n\tt.Queue[key] = &OpenconfigQos_Qos_Interfaces_Interface_Input_Queues_Queue{\n\t\tName: &Name,\n\t}\n\n\treturn t.Queue[key], nil\n}", "func (o *StorageServiceMetricsHintsInProgressGetParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func NewQueue(args []func(http.ResponseWriter, *http.Request) (http.ResponseWriter, *http.Request)) *Queue {\n\tq := &Queue{}\n\tfor _, f := range args {\n\t\tq.list = append(q.list, f)\n\t}\n\treturn q\n}", "func Constructor() MyQueue {\n\treturn Myqueue{list: listNew()}\n}", "func NewQueue() *Queue {\n\treturn &Queue{}\n}", "func NewQueue() *Queue {\n\treturn &Queue{}\n}", "func (o *GetBundleByKeyParams) SetDefaults() {\n\tvar (\n\t\tauditDefault = string(\"NONE\")\n\n\t\tincludedDeletedDefault = bool(false)\n\t)\n\n\tval := GetBundleByKeyParams{\n\t\tAudit: &auditDefault,\n\t\tIncludedDeleted: &includedDeletedDefault,\n\t}\n\n\tval.timeout = o.timeout\n\tval.Context = o.Context\n\tval.HTTPClient = o.HTTPClient\n\t*o = val\n}", "func New(\n\tlogger *zap.SugaredLogger,\n\tflushFunc, closeFunc func(),\n\topts Options,\n) *Queue {\n\tif flushFunc == nil {\n\t\tflushFunc = func() {}\n\t}\n\tif closeFunc == nil {\n\t\tcloseFunc = func() {}\n\t}\n\tif opts.Rate == 0 {\n\t\topts.Rate = 5 * time.Second\n\t}\n\n\tvar counter = int32(0)\n\treturn &Queue{\n\t\tl: logger,\n\n\t\tcloseFunc: closeFunc,\n\t\tflushFunc: flushFunc,\n\n\t\tpendingC: make(chan func(), 3*opts.BatchSize),\n\t\tpending: &counter,\n\t\trate: opts.Rate,\n\t\tbatchSize: opts.BatchSize,\n\n\t\tstopC: make(chan bool, 1),\n\t\tstopped: false,\n\t}\n}", "func (o *QtreeCollectionGetParams) SetDefaults() {\n\tvar (\n\t\treturnRecordsDefault = bool(true)\n\n\t\treturnTimeoutDefault = int64(15)\n\t)\n\n\tval := QtreeCollectionGetParams{\n\t\tReturnRecords: &returnRecordsDefault,\n\t\tReturnTimeout: &returnTimeoutDefault,\n\t}\n\n\tval.timeout = o.timeout\n\tval.Context = o.Context\n\tval.HTTPClient = o.HTTPClient\n\t*o = val\n}", "func newQueue() *Queue {\n\tl := list.New()\n\treturn &Queue{Elements: l}\n}" ]
[ "0.6368561", "0.5929313", "0.5928396", "0.59103316", "0.5819523", "0.58130354", "0.5753316", "0.5739753", "0.568538", "0.5661701", "0.5606484", "0.5487868", "0.54643965", "0.5463153", "0.54630005", "0.5449085", "0.544472", "0.54301214", "0.540289", "0.5370558", "0.535047", "0.5341004", "0.5334277", "0.5295785", "0.52576846", "0.52551335", "0.52515256", "0.52515256", "0.52440447", "0.5240541", "0.52186775", "0.52153325", "0.5206354", "0.51808494", "0.5172658", "0.5166101", "0.5157915", "0.5146065", "0.5141454", "0.5139676", "0.5139047", "0.51379216", "0.51322454", "0.5129915", "0.5124572", "0.51156694", "0.5109505", "0.5105227", "0.5103686", "0.51033306", "0.5093908", "0.50899905", "0.5077126", "0.50655735", "0.5062785", "0.5044816", "0.50357693", "0.50288504", "0.5013021", "0.50087065", "0.50034255", "0.49959683", "0.49950668", "0.49800548", "0.4976274", "0.49653354", "0.49580646", "0.49579832", "0.49559307", "0.49503115", "0.49397472", "0.4929721", "0.49250063", "0.4920364", "0.49179143", "0.49174854", "0.49174672", "0.49137354", "0.4901583", "0.48928303", "0.48868117", "0.4878816", "0.48771858", "0.48752692", "0.48736435", "0.48721433", "0.48712355", "0.48712355", "0.48708367", "0.48632345", "0.4862766", "0.48562503", "0.4856162", "0.48510158", "0.4849206", "0.4849206", "0.48466817", "0.48451757", "0.48441622", "0.48435262" ]
0.7690877
0
GetClusters returns the Clusters field value
GetClusters возвращает значение поля Clusters
func (o *QueueManager) GetClusters() []string { if o == nil { var ret []string return ret } return o.Clusters }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *Config) GetClusters(ctx context.Context, quiet bool, filterMap map[string]string, clustersName ...string) (string, error) {\n\tc.Logger.Debugf(\"Sending parameters to server to get the clusters %q\", strings.Join(clustersName, \", \"))\n\n\tfilter := MapToSlice(filterMap)\n\n\treturn c.RunGRPCnRESTFunc(\"get\", true,\n\t\tfunc() (string, error) {\n\t\t\treturn c.getClustersGRPC(ctx, quiet, filter, clustersName...)\n\t\t},\n\t\tfunc() (string, error) {\n\t\t\treturn c.getClustersHTTP(quiet, filter, clustersName...)\n\t\t})\n}", "func (c *ClientImpl) GetClusters(ctx context.Context, hcpHostURL string) (models.ClusterResp, error) {\n\tspan, _ := opentracing.StartSpanFromContext(ctx, \"Get Clusters\")\n\tdefer span.Finish()\n\n\tsession, err := c.getSession(ctx, hcpHostURL, hcpUserName, hcpPassword)\n\tif err != nil {\n\t\treturn models.ClusterResp{}, err\n\t}\n\n\tstatus = Failure\n\tmonitor := metrics.StartExternalCall(externalSvcName, \"Fetch Clusters\")\n\tdefer func() { monitor.RecordWithStatus(status) }()\n\n\tresp, err := mlopsHttp.ExecuteHTTPRequest(\n\t\tctx,\n\t\tc.client,\n\t\thcpHostURL+clusterPathV2,\n\t\thttp.MethodGet,\n\t\tmap[string]string{sessionHeader: session},\n\t\tbytes.NewReader(nil),\n\t)\n\tif err != nil {\n\t\treturn models.ClusterResp{}, errors.Wrapf(err, \"while fetching clusters in MLOps controller platform.\")\n\t}\n\tresp.Body.Close()\n\n\tstatus = Success\n\n\terr = c.deleteSession(ctx, hcpHostURL, session)\n\tif err != nil {\n\t\treturn models.ClusterResp{}, err\n\t}\n\n\tclustersResp := models.ClusterResp{}\n\tjson.NewDecoder(resp.Body).Decode(&clustersResp)\n\n\treturn clustersResp, nil\n}", "func (a *Client) GetClusters(params *GetClustersParams, opts ...ClientOption) (*GetClustersOK, *GetClustersMultiStatus, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetClustersParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"GetClusters\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/kubernetes-protection/entities/kubernetes/clusters/v1\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\", \"application/octet-stream\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &GetClustersReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tswitch value := result.(type) {\n\tcase *GetClustersOK:\n\t\treturn value, nil, nil\n\tcase *GetClustersMultiStatus:\n\t\treturn nil, value, nil\n\t}\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for kubernetes_protection: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (s *RaftDatabase) Clusters() int {\n\treturn GetArg(s.name, \"clusters\").Int(s.clusters)\n}", "func (o AppProjectSpecSyncWindowsOutput) Clusters() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v AppProjectSpecSyncWindows) []string { return v.Clusters }).(pulumi.StringArrayOutput)\n}", "func Clusters() (clusters map[string][]string) {\n\tclusters = make(map[string][]string)\n\tif addr := AccessConsulAddr(); addr != \"\" && Region() != \"\" {\n\t\treturn getClustersFromConsul(addr, Region())\n\t}\n\tcs := Get(\"Key-ClusterMgrCluster\").(map[string]string)\n\tfor key, value := range cs {\n\t\tclusters[key] = strings.Split(value, \" \")\n\t}\n\treturn\n}", "func (a ClustersAPI) Get(clusterID string) (httpmodels.GetResp, error) {\n\tvar clusterInfo httpmodels.GetResp\n\n\tdata := struct {\n\t\tClusterID string `json:\"cluster_id,omitempty\" url:\"cluster_id,omitempty\"`\n\t}{\n\t\tclusterID,\n\t}\n\tresp, err := a.Client.performQuery(http.MethodGet, \"/clusters/get\", data, nil)\n\tif err != nil {\n\t\treturn clusterInfo, err\n\t}\n\n\terr = json.Unmarshal(resp, &clusterInfo)\n\treturn clusterInfo, err\n}", "func (c *cloud) Clusters() (cloudprovider.Clusters, bool) {\n\treturn nil, false\n}", "func (e *ECS) ListClusters(req *ListClustersReq) (\n\t*ListClustersResp, error) {\n\tif req == nil {\n\t\treturn nil, fmt.Errorf(\"The req params cannot be nil\")\n\t}\n\n\tparams := makeParams(\"ListClusters\")\n\tif req.MaxResults > 0 {\n\t\tparams[\"maxResults\"] = strconv.Itoa(int(req.MaxResults))\n\t}\n\tif req.NextToken != \"\" {\n\t\tparams[\"nextToken\"] = req.NextToken\n\t}\n\n\tresp := new(ListClustersResp)\n\tif err := e.query(params, resp); err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func (h *httpCloud) Clusters() (cloudprovider.Clusters, bool) {\n\treturn nil, false\n}", "func (c *cloud) Clusters() (cloudprovider.Clusters, bool) {\n\tklog.V(4).Infof(\"Clusters called\")\n\treturn nil, false\n}", "func (a ClustersAPI) List() ([]httpmodels.GetResp, error) {\n\tvar clusterList = struct {\n\t\tClusters []httpmodels.GetResp `json:\"clusters,omitempty\" url:\"clusters,omitempty\"`\n\t}{}\n\n\tresp, err := a.Client.performQuery(http.MethodGet, \"/clusters/list\", nil, nil)\n\tif err != nil {\n\t\treturn clusterList.Clusters, err\n\t}\n\n\terr = json.Unmarshal(resp, &clusterList)\n\treturn clusterList.Clusters, err\n}", "func (ch *ClusterHandler) GetClusters() app.Adapter {\n\treturn func(h http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tcontext := app.GetRequestContext(r)\n\n\t\t\tlogger := log.WithFields(log.Fields{\"package\": \"handlers\", \"event\": \"get_clusters\", \"request\": context.RequestId()})\n\n\t\t\tclusters, err := ch.service.GetClusters(context.RequestId())\n\t\t\tif err != nil {\n\t\t\t\tresponse := ErrorResponseAttributes{Title: \"get_clusters_error\", Detail: err.Error()}\n\t\t\t\tlogger.Error(err.Error())\n\t\t\t\trespondWithJson(w, newErrorResponse(&response, context.RequestId()), http.StatusInternalServerError)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\trespondWithJson(w, newClustersResponse(clusters, context.RequestId()), http.StatusOK)\n\t\t})\n\t}\n}", "func (c *Client) GetClusters(ctx context.Context) <-chan GetClusterResult {\n\t// TODO Make the concurrency configurable\n\tconcurrency := int(math.Min(5, float64(runtime.NumCPU())))\n\tresults := make(chan GetClusterResult, concurrency)\n\n\tclusterNames, err := c.GetClusterNames(ctx)\n\tif err != nil {\n\t\tclose(results)\n\t\treturn results\n\t}\n\n\tvar wg sync.WaitGroup\n\n\tgo func() {\n\t\tdefer close(results)\n\t\tfor _, clusterName := range clusterNames {\n\t\t\twg.Add(1)\n\t\t\tgo func(name string) {\n\t\t\t\tdefer wg.Done()\n\t\t\t\tcluster, err := c.GetCluster(ctx, name)\n\t\t\t\tresult := GetClusterResult{Cluster: cluster, Error: err}\n\t\t\t\tresults <- result\n\t\t\t}(clusterName)\n\t\t}\n\t\twg.Wait()\n\t}()\n\n\treturn results\n}", "func (cloud *Cloud) Clusters() (cloudprovider.Clusters, bool) {\n\treturn nil, false\n}", "func (o *QueueManager) GetClustersOk() (*[]string, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn &o.Clusters, true\n}", "func (client OpenShiftManagedClustersClient) Get(ctx context.Context, resourceGroupName string, resourceName string) (result v20180930preview.OpenShiftManagedCluster, err error) {\n\treq, err := client.GetPreparer(ctx, resourceGroupName, resourceName)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"containerservice.OpenShiftManagedClustersClient\", \"Get\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.GetSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"containerservice.OpenShiftManagedClustersClient\", \"Get\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.GetResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"containerservice.OpenShiftManagedClustersClient\", \"Get\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func FetchClusters(c *gin.Context) {\n\n\tbanzaiUtils.LogInfo(banzaiConstants.TagListClusters, \"Start listing clusters\")\n\n\tvar clusters []banzaiSimpleTypes.ClusterSimple\n\tvar response []*cloud.ClusterRepresentation\n\tdatabase.Find(&clusters)\n\n\tif len(clusters) <= 0 {\n\t\tbanzaiUtils.LogInfo(banzaiConstants.TagListClusters, \"No clusters found\")\n\t\tcloud.SetResponseBodyJson(c, http.StatusNotFound, gin.H{\n\t\t\tcloud.JsonKeyStatus: http.StatusNotFound,\n\t\t\tcloud.JsonKeyMessage: \"No clusters found!\",\n\t\t})\n\t\treturn\n\t}\n\n\tfor _, cl := range clusters {\n\t\tclust := cloud.GetClusterRepresentation(&cl)\n\t\tif clust != nil {\n\t\t\tbanzaiUtils.LogInfo(banzaiConstants.TagListClusters, fmt.Sprintf(\"Append %#v cluster representation to response\", clust))\n\t\t\tresponse = append(response, clust)\n\t\t}\n\n\t}\n\tcloud.SetResponseBodyJson(c, http.StatusOK, gin.H{\n\t\tcloud.JsonKeyStatus: http.StatusOK,\n\t\tcloud.JsonKeyData: response,\n\t})\n}", "func (a *DefaultApiService) ListClusters(ctx _context.Context, localVarOptionals *ListClustersOpts) (Clusters, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue Clusters\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/clusters\"\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tif localVarOptionals != nil && localVarOptionals.Id.IsSet() {\n\t\tt:=localVarOptionals.Id.Value()\n\t\tif reflect.TypeOf(t).Kind() == reflect.Slice {\n\t\t\ts := reflect.ValueOf(t)\n\t\t\tfor i := 0; i < s.Len(); i++ {\n\t\t\t\tlocalVarQueryParams.Add(\"id[]\", parameterToString(s.Index(i), \"multi\"))\n\t\t\t}\n\t\t} else {\n\t\t\tlocalVarQueryParams.Add(\"id[]\", parameterToString(t, \"multi\"))\n\t\t}\n\t}\n\tif localVarOptionals != nil && localVarOptionals.NotId.IsSet() {\n\t\tt:=localVarOptionals.NotId.Value()\n\t\tif reflect.TypeOf(t).Kind() == reflect.Slice {\n\t\t\ts := reflect.ValueOf(t)\n\t\t\tfor i := 0; i < s.Len(); i++ {\n\t\t\t\tlocalVarQueryParams.Add(\"!id[]\", parameterToString(s.Index(i), \"multi\"))\n\t\t\t}\n\t\t} else {\n\t\t\tlocalVarQueryParams.Add(\"!id[]\", parameterToString(t, \"multi\"))\n\t\t}\n\t}\n\tif localVarOptionals != nil && localVarOptionals.StoryCountMin.IsSet() {\n\t\tlocalVarQueryParams.Add(\"story_count.min\", parameterToString(localVarOptionals.StoryCountMin.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.StoryCountMax.IsSet() {\n\t\tlocalVarQueryParams.Add(\"story_count.max\", parameterToString(localVarOptionals.StoryCountMax.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.TimeStart.IsSet() {\n\t\tlocalVarQueryParams.Add(\"time.start\", parameterToString(localVarOptionals.TimeStart.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.TimeEnd.IsSet() {\n\t\tlocalVarQueryParams.Add(\"time.end\", parameterToString(localVarOptionals.TimeEnd.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.EarliestStoryStart.IsSet() {\n\t\tlocalVarQueryParams.Add(\"earliest_story.start\", parameterToString(localVarOptionals.EarliestStoryStart.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.EarliestStoryEnd.IsSet() {\n\t\tlocalVarQueryParams.Add(\"earliest_story.end\", parameterToString(localVarOptionals.EarliestStoryEnd.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.LatestStoryStart.IsSet() {\n\t\tlocalVarQueryParams.Add(\"latest_story.start\", parameterToString(localVarOptionals.LatestStoryStart.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.LatestStoryEnd.IsSet() {\n\t\tlocalVarQueryParams.Add(\"latest_story.end\", parameterToString(localVarOptionals.LatestStoryEnd.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.LocationCountry.IsSet() {\n\t\tt:=localVarOptionals.LocationCountry.Value()\n\t\tif reflect.TypeOf(t).Kind() == reflect.Slice {\n\t\t\ts := reflect.ValueOf(t)\n\t\t\tfor i := 0; i < s.Len(); i++ {\n\t\t\t\tlocalVarQueryParams.Add(\"location.country\", parameterToString(s.Index(i), \"multi\"))\n\t\t\t}\n\t\t} else {\n\t\t\tlocalVarQueryParams.Add(\"location.country\", parameterToString(t, \"multi\"))\n\t\t}\n\t}\n\tif localVarOptionals != nil && localVarOptionals.NotLocationCountry.IsSet() {\n\t\tt:=localVarOptionals.NotLocationCountry.Value()\n\t\tif reflect.TypeOf(t).Kind() == reflect.Slice {\n\t\t\ts := reflect.ValueOf(t)\n\t\t\tfor i := 0; i < s.Len(); i++ {\n\t\t\t\tlocalVarQueryParams.Add(\"!location.country\", parameterToString(s.Index(i), \"multi\"))\n\t\t\t}\n\t\t} else {\n\t\t\tlocalVarQueryParams.Add(\"!location.country\", parameterToString(t, \"multi\"))\n\t\t}\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Return_.IsSet() {\n\t\tt:=localVarOptionals.Return_.Value()\n\t\tif reflect.TypeOf(t).Kind() == reflect.Slice {\n\t\t\ts := reflect.ValueOf(t)\n\t\t\tfor i := 0; i < s.Len(); i++ {\n\t\t\t\tlocalVarQueryParams.Add(\"return[]\", parameterToString(s.Index(i), \"multi\"))\n\t\t\t}\n\t\t} else {\n\t\t\tlocalVarQueryParams.Add(\"return[]\", parameterToString(t, \"multi\"))\n\t\t}\n\t}\n\tif localVarOptionals != nil && localVarOptionals.SortBy.IsSet() {\n\t\tlocalVarQueryParams.Add(\"sort_by\", parameterToString(localVarOptionals.SortBy.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.SortDirection.IsSet() {\n\t\tlocalVarQueryParams.Add(\"sort_direction\", parameterToString(localVarOptionals.SortDirection.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Cursor.IsSet() {\n\t\tlocalVarQueryParams.Add(\"cursor\", parameterToString(localVarOptionals.Cursor.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.PerPage.IsSet() {\n\t\tlocalVarQueryParams.Add(\"per_page\", parameterToString(localVarOptionals.PerPage.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\", \"text/xml\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"X-AYLIEN-NewsAPI-Application-ID\"] = key\n\t\t}\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"X-AYLIEN-NewsAPI-Application-Key\"] = key\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 401 {\n\t\t\tvar v Errors\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v Errors\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 422 {\n\t\t\tvar v Errors\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 429 {\n\t\t\tvar v Errors\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 500 {\n\t\t\tvar v Errors\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (az *Cloud) Clusters() (cloudprovider.Clusters, bool) {\n\treturn nil, false\n}", "func ExampleClustersClient_Get() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armservicefabric.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := clientFactory.NewClustersClient().Get(ctx, \"resRg\", \"myCluster\", nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// You could use response here. We use blank identifier for just demo purposes.\n\t_ = res\n\t// If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t// res.Cluster = armservicefabric.Cluster{\n\t// \tName: to.Ptr(\"myCluster\"),\n\t// \tType: to.Ptr(\"Microsoft.ServiceFabric/clusters\"),\n\t// \tEtag: to.Ptr(\"W/\\\"636462502169240745\\\"\"),\n\t// \tID: to.Ptr(\"/subscriptions/00000000-0000-0000-0000-000000000000/resourcegroups/resRg/providers/Microsoft.ServiceFabric/clusters/myCluster\"),\n\t// \tLocation: to.Ptr(\"eastus\"),\n\t// \tTags: map[string]*string{\n\t// \t},\n\t// \tProperties: &armservicefabric.ClusterProperties{\n\t// \t\tAddOnFeatures: []*armservicefabric.AddOnFeatures{\n\t// \t\t\tto.Ptr(armservicefabric.AddOnFeaturesRepairManager),\n\t// \t\t\tto.Ptr(armservicefabric.AddOnFeaturesDNSService),\n\t// \t\t\tto.Ptr(armservicefabric.AddOnFeaturesBackupRestoreService),\n\t// \t\t\tto.Ptr(armservicefabric.AddOnFeaturesResourceMonitorService)},\n\t// \t\t\tAvailableClusterVersions: []*armservicefabric.ClusterVersionDetails{\n\t// \t\t\t\t{\n\t// \t\t\t\t\tCodeVersion: to.Ptr(\"6.1.480.9494\"),\n\t// \t\t\t\t\tEnvironment: to.Ptr(armservicefabric.ClusterEnvironmentWindows),\n\t// \t\t\t\t\tSupportExpiryUTC: to.Ptr(\"2018-06-15T23:59:59.9999999\"),\n\t// \t\t\t}},\n\t// \t\t\tAzureActiveDirectory: &armservicefabric.AzureActiveDirectory{\n\t// \t\t\t\tClientApplication: to.Ptr(\"d151ad89-4bce-4ae8-b3d1-1dc79679fa75\"),\n\t// \t\t\t\tClusterApplication: to.Ptr(\"5886372e-7bf4-4878-a497-8098aba608ae\"),\n\t// \t\t\t\tTenantID: to.Ptr(\"6abcc6a0-8666-43f1-87b8-172cf86a9f9c\"),\n\t// \t\t\t},\n\t// \t\t\tCertificateCommonNames: &armservicefabric.ServerCertificateCommonNames{\n\t// \t\t\t\tCommonNames: []*armservicefabric.ServerCertificateCommonName{\n\t// \t\t\t\t\t{\n\t// \t\t\t\t\t\tCertificateCommonName: to.Ptr(\"abc.com\"),\n\t// \t\t\t\t\t\tCertificateIssuerThumbprint: to.Ptr(\"12599211F8F14C90AFA9532AD79A6F2CA1C00622\"),\n\t// \t\t\t\t}},\n\t// \t\t\t\tX509StoreName: to.Ptr(armservicefabric.StoreNameMy),\n\t// \t\t\t},\n\t// \t\t\tClientCertificateCommonNames: []*armservicefabric.ClientCertificateCommonName{\n\t// \t\t\t\t{\n\t// \t\t\t\t\tCertificateCommonName: to.Ptr(\"abc.com\"),\n\t// \t\t\t\t\tCertificateIssuerThumbprint: to.Ptr(\"5F3660C715EBBDA31DB1FFDCF508302348DE8E7A\"),\n\t// \t\t\t\t\tIsAdmin: to.Ptr(true),\n\t// \t\t\t}},\n\t// \t\t\tClientCertificateThumbprints: []*armservicefabric.ClientCertificateThumbprint{\n\t// \t\t\t\t{\n\t// \t\t\t\t\tCertificateThumbprint: to.Ptr(\"5F3660C715EBBDA31DB1FFDCF508302348DE8E7A\"),\n\t// \t\t\t\t\tIsAdmin: to.Ptr(true),\n\t// \t\t\t}},\n\t// \t\t\tClusterCodeVersion: to.Ptr(\"6.1.480.9494\"),\n\t// \t\t\tClusterEndpoint: to.Ptr(\"https://eastus.servicefabric.azure.com\"),\n\t// \t\t\tClusterID: to.Ptr(\"92584666-9889-4ae8-8d02-91902923d37f\"),\n\t// \t\t\tClusterState: to.Ptr(armservicefabric.ClusterStateWaitingForNodes),\n\t// \t\t\tDiagnosticsStorageAccountConfig: &armservicefabric.DiagnosticsStorageAccountConfig{\n\t// \t\t\t\tBlobEndpoint: to.Ptr(\"https://diag.blob.core.windows.net/\"),\n\t// \t\t\t\tProtectedAccountKeyName: to.Ptr(\"StorageAccountKey1\"),\n\t// \t\t\t\tQueueEndpoint: to.Ptr(\"https://diag.queue.core.windows.net/\"),\n\t// \t\t\t\tStorageAccountName: to.Ptr(\"diag\"),\n\t// \t\t\t\tTableEndpoint: to.Ptr(\"https://diag.table.core.windows.net/\"),\n\t// \t\t\t},\n\t// \t\t\tFabricSettings: []*armservicefabric.SettingsSectionDescription{\n\t// \t\t\t\t{\n\t// \t\t\t\t\tName: to.Ptr(\"UpgradeService\"),\n\t// \t\t\t\t\tParameters: []*armservicefabric.SettingsParameterDescription{\n\t// \t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\tName: to.Ptr(\"AppPollIntervalInSeconds\"),\n\t// \t\t\t\t\t\t\tValue: to.Ptr(\"60\"),\n\t// \t\t\t\t\t}},\n\t// \t\t\t}},\n\t// \t\t\tManagementEndpoint: to.Ptr(\"https://myCluster.eastus.cloudapp.azure.com:19080\"),\n\t// \t\t\tNodeTypes: []*armservicefabric.NodeTypeDescription{\n\t// \t\t\t\t{\n\t// \t\t\t\t\tName: to.Ptr(\"nt1vm\"),\n\t// \t\t\t\t\tApplicationPorts: &armservicefabric.EndpointRangeDescription{\n\t// \t\t\t\t\t\tEndPort: to.Ptr[int32](30000),\n\t// \t\t\t\t\t\tStartPort: to.Ptr[int32](20000),\n\t// \t\t\t\t\t},\n\t// \t\t\t\t\tClientConnectionEndpointPort: to.Ptr[int32](19000),\n\t// \t\t\t\t\tDurabilityLevel: to.Ptr(armservicefabric.DurabilityLevelBronze),\n\t// \t\t\t\t\tEphemeralPorts: &armservicefabric.EndpointRangeDescription{\n\t// \t\t\t\t\t\tEndPort: to.Ptr[int32](64000),\n\t// \t\t\t\t\t\tStartPort: to.Ptr[int32](49000),\n\t// \t\t\t\t\t},\n\t// \t\t\t\t\tHTTPGatewayEndpointPort: to.Ptr[int32](19007),\n\t// \t\t\t\t\tIsPrimary: to.Ptr(true),\n\t// \t\t\t\t\tVMInstanceCount: to.Ptr[int32](5),\n\t// \t\t\t}},\n\t// \t\t\tProvisioningState: to.Ptr(armservicefabric.ProvisioningStateSucceeded),\n\t// \t\t\tReliabilityLevel: to.Ptr(armservicefabric.ReliabilityLevelSilver),\n\t// \t\t\tReverseProxyCertificateCommonNames: &armservicefabric.ServerCertificateCommonNames{\n\t// \t\t\t\tCommonNames: []*armservicefabric.ServerCertificateCommonName{\n\t// \t\t\t\t\t{\n\t// \t\t\t\t\t\tCertificateCommonName: to.Ptr(\"abc.com\"),\n\t// \t\t\t\t\t\tCertificateIssuerThumbprint: to.Ptr(\"12599211F8F14C90AFA9532AD79A6F2CA1C00622\"),\n\t// \t\t\t\t}},\n\t// \t\t\t\tX509StoreName: to.Ptr(armservicefabric.StoreNameMy),\n\t// \t\t\t},\n\t// \t\t\tUpgradeDescription: &armservicefabric.ClusterUpgradePolicy{\n\t// \t\t\t\tDeltaHealthPolicy: &armservicefabric.ClusterUpgradeDeltaHealthPolicy{\n\t// \t\t\t\t\tApplicationDeltaHealthPolicies: map[string]*armservicefabric.ApplicationDeltaHealthPolicy{\n\t// \t\t\t\t\t\t\"fabric:/myApp1\": &armservicefabric.ApplicationDeltaHealthPolicy{\n\t// \t\t\t\t\t\t\tDefaultServiceTypeDeltaHealthPolicy: &armservicefabric.ServiceTypeDeltaHealthPolicy{\n\t// \t\t\t\t\t\t\t\tMaxPercentDeltaUnhealthyServices: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tServiceTypeDeltaHealthPolicies: map[string]*armservicefabric.ServiceTypeDeltaHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\"myServiceType1\": &armservicefabric.ServiceTypeDeltaHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\tMaxPercentDeltaUnhealthyServices: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t},\n\t// \t\t\t\t\t},\n\t// \t\t\t\t\tMaxPercentDeltaUnhealthyApplications: to.Ptr[int32](0),\n\t// \t\t\t\t\tMaxPercentDeltaUnhealthyNodes: to.Ptr[int32](0),\n\t// \t\t\t\t\tMaxPercentUpgradeDomainDeltaUnhealthyNodes: to.Ptr[int32](0),\n\t// \t\t\t\t},\n\t// \t\t\t\tForceRestart: to.Ptr(false),\n\t// \t\t\t\tHealthCheckRetryTimeout: to.Ptr(\"00:05:00\"),\n\t// \t\t\t\tHealthCheckStableDuration: to.Ptr(\"00:00:30\"),\n\t// \t\t\t\tHealthCheckWaitDuration: to.Ptr(\"00:00:30\"),\n\t// \t\t\t\tHealthPolicy: &armservicefabric.ClusterHealthPolicy{\n\t// \t\t\t\t\tApplicationHealthPolicies: map[string]*armservicefabric.ApplicationHealthPolicy{\n\t// \t\t\t\t\t\t\"fabric:/myApp1\": &armservicefabric.ApplicationHealthPolicy{\n\t// \t\t\t\t\t\t\tDefaultServiceTypeHealthPolicy: &armservicefabric.ServiceTypeHealthPolicy{\n\t// \t\t\t\t\t\t\t\tMaxPercentUnhealthyServices: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tServiceTypeHealthPolicies: map[string]*armservicefabric.ServiceTypeHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\"myServiceType1\": &armservicefabric.ServiceTypeHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\tMaxPercentUnhealthyServices: to.Ptr[int32](100),\n\t// \t\t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t},\n\t// \t\t\t\t\t},\n\t// \t\t\t\t\tMaxPercentUnhealthyApplications: to.Ptr[int32](0),\n\t// \t\t\t\t\tMaxPercentUnhealthyNodes: to.Ptr[int32](0),\n\t// \t\t\t\t},\n\t// \t\t\t\tUpgradeDomainTimeout: to.Ptr(\"00:15:00\"),\n\t// \t\t\t\tUpgradeReplicaSetCheckTimeout: to.Ptr(\"00:10:00\"),\n\t// \t\t\t\tUpgradeTimeout: to.Ptr(\"01:00:00\"),\n\t// \t\t\t},\n\t// \t\t\tUpgradeMode: to.Ptr(armservicefabric.UpgradeModeManual),\n\t// \t\t\tVMImage: to.Ptr(\"Windows\"),\n\t// \t\t},\n\t// \t}\n}", "func (a *Client) ListClusters(params *ListClustersParams, authInfo runtime.ClientAuthInfoWriter) (*ListClustersOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListClustersParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"ListClusters\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/v1/clusters\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &ListClustersReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*ListClustersOK), nil\n\n}", "func (s *Server) GetClusters() []*api.Cluster {\n\tinstances := s.doGetClusters()\n\tclusters := make([]*api.Cluster, len(instances))\n\tfor i, instance := range instances {\n\t\tclusters[i] = convertClusterToAPI(instance)\n\t}\n\treturn clusters\n}", "func (bc *Baiducloud) Clusters() (cloudprovider.Clusters, bool) {\n\treturn nil, false\n}", "func (adm Admin) ListClusters() (string, error) {\n\tconn := newConnection(adm.ZkSvr)\n\terr := conn.Connect()\n\tif err != nil {\n\t\tfmt.Println(\"Failed to connect to zookeeper.\")\n\t\treturn \"\", err\n\t}\n\tdefer conn.Disconnect()\n\n\tvar clusters []string\n\n\tchildren, err := conn.Children(\"/\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tfor _, cluster := range children {\n\t\tif ok, err := conn.IsClusterSetup(cluster); ok && err == nil {\n\t\t\tclusters = append(clusters, cluster)\n\t\t}\n\t}\n\n\tvar buffer bytes.Buffer\n\tbuffer.WriteString(\"Existing clusters: \\n\")\n\n\tfor _, cluster := range clusters {\n\t\tbuffer.WriteString(\" \" + cluster + \"\\n\")\n\t}\n\treturn buffer.String(), nil\n}", "func (q *QueryResolver) Clusters(ctx context.Context) ([]*ClusterInfoResolver, error) {\n\tgrpcAPI := q.Env.VizierClusterInfo\n\tresp, err := grpcAPI.GetClusterInfo(ctx, &cloudpb.GetClusterInfoRequest{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar res []*ClusterInfoResolver\n\tfor _, cluster := range resp.Clusters {\n\t\tresolver, err := clusterInfoToResolver(cluster)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tres = append(res, resolver)\n\t}\n\treturn res, nil\n}", "func (r *ProjectsInstancesClustersService) Get(name string) *ProjectsInstancesClustersGetCall {\n\tc := &ProjectsInstancesClustersGetCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\treturn c\n}", "func (a *Client) GetCombinedCloudClusters(params *GetCombinedCloudClustersParams, opts ...ClientOption) (*GetCombinedCloudClustersOK, *GetCombinedCloudClustersMultiStatus, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetCombinedCloudClustersParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"GetCombinedCloudClusters\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/kubernetes-protection/entities/cloud_cluster/v1\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\", \"application/octet-stream\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &GetCombinedCloudClustersReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tswitch value := result.(type) {\n\tcase *GetCombinedCloudClustersOK:\n\t\treturn value, nil, nil\n\tcase *GetCombinedCloudClustersMultiStatus:\n\t\treturn nil, value, nil\n\t}\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for kubernetes_protection: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (c *Client) GetClustersSync(ctx context.Context) ([]*Cluster, error) {\n\tclusters := make([]*Cluster, 0)\n\n\tfor result := range c.GetClusters(ctx) {\n\t\tif result.Error != nil {\n\t\t\treturn nil, result.Error\n\t\t}\n\t\tclusters = append(clusters, result.Cluster)\n\t}\n\n\treturn clusters, nil\n}", "func (adm Admin) ListClusters() (string, error) {\n\tvar clusters []string\n\n\tchildren, err := adm.zkClient.Children(\"/\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tfor _, cluster := range children {\n\t\tif ok, err := adm.isClusterSetup(cluster); ok && err == nil {\n\t\t\tclusters = append(clusters, cluster)\n\t\t}\n\t}\n\n\tvar buffer bytes.Buffer\n\tbuffer.WriteString(\"Existing clusters: \\n\")\n\n\tfor _, cluster := range clusters {\n\t\tbuffer.WriteString(\" \" + cluster + \"\\n\")\n\t}\n\treturn buffer.String(), nil\n}", "func handleGetClusters(c *Context, w http.ResponseWriter, r *http.Request) {\n\tpaging, err := parsePaging(r.URL)\n\tif err != nil {\n\t\tc.Logger.WithError(err).Error(\"failed to parse paging parameters\")\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tfilter := &model.ClusterFilter{\n\t\tPaging: paging,\n\t}\n\n\tclusters, err := c.Store.GetClusterDTOs(filter)\n\tif err != nil {\n\t\tc.Logger.WithError(err).Error(\"failed to query clusters\")\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\tif clusters == nil {\n\t\tclusters = []*model.ClusterDTO{}\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\toutputJSON(c, w, clusters)\n}", "func (s *clusterService) Clusters(ctx context.Context, options ...rest.HTTPClientOption) ([]cluster.Cluster, error) {\n\t_, err := Start(ctx, s.Factories().ClusterCacheFactory(), options...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tclusterCache.RLock()\n\tdefer clusterCache.RUnlock()\n\n\treturn Clusters(clusterCache.Clusters()), nil\n}", "func (c Client) ListClusters() (ClusterList, error) {\n\tbody, err := c.watsonClient.MakeRequest(\"GET\", c.version+\"/solr_clusters\", nil, nil)\n\tif err != nil {\n\t\treturn ClusterList{}, err\n\t}\n\tvar response ClusterList\n\terr = json.Unmarshal(body, &response)\n\treturn response, err\n}", "func (c *krakenClusters) Get(name string, options v1.GetOptions) (result *v1alpha1.KrakenCluster, err error) {\n\tresult = &v1alpha1.KrakenCluster{}\n\terr = c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"krakenclusters\").\n\t\tName(name).\n\t\tVersionedParams(&options, scheme.ParameterCodec).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (c starterClusterServiceOp) List(ctx context.Context) (*[]models.Cluster, *Response, error) {\n\tvar clusterList []models.Cluster\n\tgraphqlRequest := models.GraphqlRequest{\n\t\tName: \"clusters\",\n\t\tOperation: models.Query,\n\t\tInput: clusterList,\n\t\tArgs: models.ClusterListInput{\n\t\t\tProductType: models.Starter,\n\t\t},\n\t\tResponse: clusterList,\n\t}\n\treq, err := c.client.NewRequest(&graphqlRequest)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tresp, err := c.client.Do(ctx, req, &clusterList)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn &clusterList, resp, err\n}", "func getClusters(kubeconfig string) ([]string, error) {\n\tkubectlArgs := []string{\"kubectl\"}\n\tif kubeconfig != \"\" {\n\t\tkubectlArgs = append(kubectlArgs, fmt.Sprintf(\"--kubeconfig=%s\", kubeconfig))\n\t}\n\tcontextArgs := append(kubectlArgs, []string{\"config\", \"get-contexts\", \"-o=name\"}...)\n\toutput, err := runCommand(contextArgs)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error in getting contexts from kubeconfig: %s\", err)\n\t}\n\treturn strings.Split(output, \"\\n\"), nil\n}", "func (nh *NodeHost) Clusters() []*node {\n\tresult := make([]*node, 0)\n\tnh.clusterMu.RLock()\n\tnh.clusterMu.clusters.Range(func(k, v interface{}) bool {\n\t\tresult = append(result, v.(*node))\n\t\treturn true\n\t})\n\tnh.clusterMu.RUnlock()\n\n\treturn result\n}", "func (a *ClustersApiService) ListClusters(ctx _context.Context, space string) ApiListClustersRequest {\n\treturn ApiListClustersRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tspace: space,\n\t}\n}", "func GetMultipleClustersName(cmd *cobra.Command, args []string) ([]string, error) {\n\tif len(args) == 0 {\n\t\treturn nil, UserErrorf(\"requires a cluster name\")\n\t}\n\treturn args, nil\n}", "func (bc *Baiducloud) ListClusters(ctx context.Context) ([]string, error) {\n\treturn nil, fmt.Errorf(\"ListClusters unimplemented\")\n}", "func (a *ClustersApiService) ListClustersExecute(r ApiListClustersRequest) (ListClustersResponse, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue ListClustersResponse\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"ClustersApiService.ListClusters\")\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, GenericOpenAPIError{error: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/spaces/{space}/clusters\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"space\"+\"}\", _neturl.PathEscape(parameterToString(r.space, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = _ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (svc ServerlessClusterService) List(ctx context.Context) (*[]models.Cluster, *Response, error) {\n\tvar clusterList []models.Cluster\n\tgraphqlRequest := models.GraphqlRequest{\n\t\tName: \"clusters\",\n\t\tOperation: models.Query,\n\t\tInput: nil,\n\t\tArgs: models.ClusterListInput{\n\t\t\tProductType: models.Starter,\n\t\t},\n\t\tResponse: clusterList,\n\t}\n\treq, err := svc.client.NewRequest(&graphqlRequest)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tresp, err := svc.client.Do(ctx, req, &clusterList)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn &clusterList, resp, err\n}", "func (m *MockBuilder) Clusters() []string {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"Clusters\")\n\tret0, _ := ret[0].([]string)\n\treturn ret0\n}", "func (cc *CloudComb) GetClustersImages() (string, error) {\n\tresult, _, err := cc.doRESTRequest(\"GET\", \"/api/v1/apps/images\", \"\", nil, nil)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn result, nil\n}", "func (ds *DiscoveryService) ListClusters(request *restful.Request, response *restful.Response) {\n\tkey := request.Request.URL.String()\n\tout, cached := ds.cdsCache.cachedDiscoveryResponse(key)\n\tif !cached {\n\t\tif sc := request.PathParameter(ServiceCluster); sc != ds.mesh.IstioServiceCluster {\n\t\t\terrorResponse(response, http.StatusNotFound,\n\t\t\t\tfmt.Sprintf(\"Unexpected %s %q\", ServiceCluster, sc))\n\t\t\treturn\n\t\t}\n\n\t\t// service-node holds the IP address\n\t\tip := request.PathParameter(ServiceNode)\n\t\t// CDS computes clusters that are referenced by RDS routes for a particular proxy node\n\t\t// TODO: this implementation is inefficient as it is recomputing all the routes for all proxies\n\t\t// There is a lot of potential to cache and reuse cluster definitions across proxies and also\n\t\t// skip computing the actual HTTP routes\n\t\tinstances := ds.services.HostInstances(map[string]bool{ip: true})\n\t\tservices := ds.services.Services()\n\t\thttpRouteConfigs := buildOutboundHTTPRoutes(instances, services, &ProxyContext{\n\t\t\tDiscovery: ds.services,\n\t\t\tConfig: ds.config,\n\t\t\tMeshConfig: ds.mesh,\n\t\t\tIPAddress: ip,\n\t\t})\n\n\t\t// de-duplicate and canonicalize clusters\n\t\tclusters := httpRouteConfigs.clusters().normalize()\n\n\t\t// apply custom policies for HTTP clusters\n\t\tfor _, cluster := range clusters {\n\t\t\tinsertDestinationPolicy(ds.config, cluster)\n\t\t}\n\n\t\tvar err error\n\t\tif out, err = json.MarshalIndent(ClusterManager{Clusters: clusters}, \" \", \" \"); err != nil {\n\t\t\terrorResponse(response, http.StatusInternalServerError, err.Error())\n\t\t\treturn\n\t\t}\n\t\tds.cdsCache.updateCachedDiscoveryResponse(key, out)\n\t}\n\twriteResponse(response, out)\n}", "func (e *ECS) DescribeClusters(req *DescribeClustersReq) (*DescribeClustersResp, error) {\n\tif req == nil {\n\t\treturn nil, fmt.Errorf(\"The req params cannot be nil\")\n\t}\n\n\tparams := makeParams(\"DescribeClusters\")\n\tif len(req.Clusters) > 0 {\n\t\taddParamsList(params, \"clusters.member\", req.Clusters)\n\t}\n\n\tresp := new(DescribeClustersResp)\n\tif err := e.query(params, resp); err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func (c *ClustersController) List(ctx *app.ListClustersContext) error {\n\t// return a single cluster given its URL\n\tif ctx.ClusterURL != nil {\n\t\t// authorization is checked at the service level for more consistency accross the codebase.\n\t\tclustr, err := c.app.ClusterService().FindByURL(ctx, *ctx.ClusterURL)\n\t\tif err != nil {\n\t\t\tif ok, _ := errors.IsNotFoundError(err); ok {\n\t\t\t\t// no result found, return an empty array\n\t\t\t\treturn ctx.OK(&app.ClusterList{\n\t\t\t\t\tData: []*app.ClusterData{},\n\t\t\t\t})\n\t\t\t}\n\t\t\t// something wrong happened, return the error\n\t\t\treturn app.JSONErrorResponse(ctx, err)\n\t\t}\n\t\treturn ctx.OK(&app.ClusterList{\n\t\t\tData: []*app.ClusterData{convertToClusterData(*clustr)},\n\t\t})\n\t}\n\t// otherwise, list all clusters\n\tclusters, err := c.app.ClusterService().List(ctx, ctx.Type)\n\tif err != nil {\n\t\treturn app.JSONErrorResponse(ctx, err)\n\t}\n\tvar data []*app.ClusterData\n\tfor _, clustr := range clusters {\n\t\tdata = append(data, convertToClusterData(clustr))\n\t}\n\treturn ctx.OK(&app.ClusterList{\n\t\tData: data,\n\t})\n}", "func ListClusters(c *cli.Context) error {\n\tif err := printClusters(); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func GetClusterNodes(cs *framework.ClientSet) (int, error) {\n\tnodes, err := getNodesByLabel(cs, \"\")\n\tif err != nil {\n\t\treturn 0, fmt.Errorf(\"unable to get the number of cluster nodes: %v\", err)\n\t}\n\n\treturn len(nodes), nil\n}", "func (c *ClientIMPL) GetCluster(ctx context.Context) (resp Cluster, err error) {\n\tvar systemList []Cluster\n\tcluster := Cluster{}\n\tqp := c.APIClient().QueryParamsWithFields(&cluster)\n\n\tmajorMinorVersion, err := c.GetSoftwareMajorMinorVersion(ctx)\n\tif err != nil {\n\t\tlog.Errorf(\"Couldn't find the array version %s\", err.Error())\n\t} else {\n\t\tif majorMinorVersion >= 3.0 {\n\t\t\tqp.Select(\"nvm_subsystem_nqn\")\n\t\t}\n\t}\n\t_, err = c.APIClient().Query(\n\t\tctx,\n\t\tRequestConfig{\n\t\t\tMethod: \"GET\",\n\t\t\tEndpoint: clusterURL,\n\t\t\tQueryParams: qp,\n\t\t},\n\t\t&systemList)\n\terr = WrapErr(err)\n\tif err != nil {\n\t\treturn resp, err\n\t}\n\treturn systemList[0], err\n}", "func (c starterClusterServiceOp) Get(ctx context.Context, input *models.GetStarterClusterInput) (*models.Cluster, *Response, error) {\n\tvar cluster models.Cluster\n\tvar graphqlRequest = models.GraphqlRequest{\n\t\tName: \"cluster\",\n\t\tOperation: models.Query,\n\t\tInput: nil,\n\t\tArgs: *input,\n\t\tResponse: cluster,\n\t}\n\treq, err := c.client.NewRequest(&graphqlRequest)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tresp, err := c.client.Do(ctx, req, &cluster)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn &cluster, resp, err\n}", "func ExampleClustersClient_List() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armservicefabric.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := clientFactory.NewClustersClient().List(ctx, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// You could use response here. We use blank identifier for just demo purposes.\n\t_ = res\n\t// If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t// res.ClusterListResult = armservicefabric.ClusterListResult{\n\t// \tValue: []*armservicefabric.Cluster{\n\t// \t\t{\n\t// \t\t\tName: to.Ptr(\"myCluster\"),\n\t// \t\t\tType: to.Ptr(\"Microsoft.ServiceFabric/clusters\"),\n\t// \t\t\tEtag: to.Ptr(\"W/\\\"636462502169240745\\\"\"),\n\t// \t\t\tID: to.Ptr(\"/subscriptions/00000000-0000-0000-0000-000000000000/resourcegroups/resRg/providers/Microsoft.ServiceFabric/clusters/myCluster\"),\n\t// \t\t\tLocation: to.Ptr(\"eastus\"),\n\t// \t\t\tTags: map[string]*string{\n\t// \t\t\t},\n\t// \t\t\tProperties: &armservicefabric.ClusterProperties{\n\t// \t\t\t\tAddOnFeatures: []*armservicefabric.AddOnFeatures{\n\t// \t\t\t\t\tto.Ptr(armservicefabric.AddOnFeaturesRepairManager),\n\t// \t\t\t\t\tto.Ptr(armservicefabric.AddOnFeaturesDNSService),\n\t// \t\t\t\t\tto.Ptr(armservicefabric.AddOnFeaturesBackupRestoreService),\n\t// \t\t\t\t\tto.Ptr(armservicefabric.AddOnFeaturesResourceMonitorService)},\n\t// \t\t\t\t\tAvailableClusterVersions: []*armservicefabric.ClusterVersionDetails{\n\t// \t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\tCodeVersion: to.Ptr(\"6.1.480.9494\"),\n\t// \t\t\t\t\t\t\tEnvironment: to.Ptr(armservicefabric.ClusterEnvironmentWindows),\n\t// \t\t\t\t\t\t\tSupportExpiryUTC: to.Ptr(\"2018-06-15T23:59:59.9999999\"),\n\t// \t\t\t\t\t}},\n\t// \t\t\t\t\tAzureActiveDirectory: &armservicefabric.AzureActiveDirectory{\n\t// \t\t\t\t\t\tClientApplication: to.Ptr(\"d151ad89-4bce-4ae8-b3d1-1dc79679fa75\"),\n\t// \t\t\t\t\t\tClusterApplication: to.Ptr(\"5886372e-7bf4-4878-a497-8098aba608ae\"),\n\t// \t\t\t\t\t\tTenantID: to.Ptr(\"6abcc6a0-8666-43f1-87b8-172cf86a9f9c\"),\n\t// \t\t\t\t\t},\n\t// \t\t\t\t\tCertificateCommonNames: &armservicefabric.ServerCertificateCommonNames{\n\t// \t\t\t\t\t\tCommonNames: []*armservicefabric.ServerCertificateCommonName{\n\t// \t\t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\t\tCertificateCommonName: to.Ptr(\"abc.com\"),\n\t// \t\t\t\t\t\t\t\tCertificateIssuerThumbprint: to.Ptr(\"12599211F8F14C90AFA9532AD79A6F2CA1C00622\"),\n\t// \t\t\t\t\t\t}},\n\t// \t\t\t\t\t\tX509StoreName: to.Ptr(armservicefabric.StoreNameMy),\n\t// \t\t\t\t\t},\n\t// \t\t\t\t\tClientCertificateCommonNames: []*armservicefabric.ClientCertificateCommonName{\n\t// \t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\tCertificateCommonName: to.Ptr(\"abc.com\"),\n\t// \t\t\t\t\t\t\tCertificateIssuerThumbprint: to.Ptr(\"5F3660C715EBBDA31DB1FFDCF508302348DE8E7A\"),\n\t// \t\t\t\t\t\t\tIsAdmin: to.Ptr(true),\n\t// \t\t\t\t\t}},\n\t// \t\t\t\t\tClientCertificateThumbprints: []*armservicefabric.ClientCertificateThumbprint{\n\t// \t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\tCertificateThumbprint: to.Ptr(\"5F3660C715EBBDA31DB1FFDCF508302348DE8E7A\"),\n\t// \t\t\t\t\t\t\tIsAdmin: to.Ptr(false),\n\t// \t\t\t\t\t}},\n\t// \t\t\t\t\tClusterCodeVersion: to.Ptr(\"6.1.480.9494\"),\n\t// \t\t\t\t\tClusterEndpoint: to.Ptr(\"https://eastus.servicefabric.azure.com\"),\n\t// \t\t\t\t\tClusterID: to.Ptr(\"92584666-9889-4ae8-8d02-91902923d37f\"),\n\t// \t\t\t\t\tClusterState: to.Ptr(armservicefabric.ClusterStateWaitingForNodes),\n\t// \t\t\t\t\tDiagnosticsStorageAccountConfig: &armservicefabric.DiagnosticsStorageAccountConfig{\n\t// \t\t\t\t\t\tBlobEndpoint: to.Ptr(\"https://diag.blob.core.windows.net/\"),\n\t// \t\t\t\t\t\tProtectedAccountKeyName: to.Ptr(\"StorageAccountKey1\"),\n\t// \t\t\t\t\t\tQueueEndpoint: to.Ptr(\"https://diag.queue.core.windows.net/\"),\n\t// \t\t\t\t\t\tStorageAccountName: to.Ptr(\"diag\"),\n\t// \t\t\t\t\t\tTableEndpoint: to.Ptr(\"https://diag.table.core.windows.net/\"),\n\t// \t\t\t\t\t},\n\t// \t\t\t\t\tFabricSettings: []*armservicefabric.SettingsSectionDescription{\n\t// \t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\tName: to.Ptr(\"UpgradeService\"),\n\t// \t\t\t\t\t\t\tParameters: []*armservicefabric.SettingsParameterDescription{\n\t// \t\t\t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\t\t\tName: to.Ptr(\"AppPollIntervalInSeconds\"),\n\t// \t\t\t\t\t\t\t\t\tValue: to.Ptr(\"60\"),\n\t// \t\t\t\t\t\t\t}},\n\t// \t\t\t\t\t}},\n\t// \t\t\t\t\tManagementEndpoint: to.Ptr(\"https://myCluster.eastus.cloudapp.azure.com:19080\"),\n\t// \t\t\t\t\tNodeTypes: []*armservicefabric.NodeTypeDescription{\n\t// \t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\tName: to.Ptr(\"nt1vm\"),\n\t// \t\t\t\t\t\t\tApplicationPorts: &armservicefabric.EndpointRangeDescription{\n\t// \t\t\t\t\t\t\t\tEndPort: to.Ptr[int32](30000),\n\t// \t\t\t\t\t\t\t\tStartPort: to.Ptr[int32](20000),\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tClientConnectionEndpointPort: to.Ptr[int32](19000),\n\t// \t\t\t\t\t\t\tDurabilityLevel: to.Ptr(armservicefabric.DurabilityLevelBronze),\n\t// \t\t\t\t\t\t\tEphemeralPorts: &armservicefabric.EndpointRangeDescription{\n\t// \t\t\t\t\t\t\t\tEndPort: to.Ptr[int32](64000),\n\t// \t\t\t\t\t\t\t\tStartPort: to.Ptr[int32](49000),\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tHTTPGatewayEndpointPort: to.Ptr[int32](19007),\n\t// \t\t\t\t\t\t\tIsPrimary: to.Ptr(true),\n\t// \t\t\t\t\t\t\tVMInstanceCount: to.Ptr[int32](5),\n\t// \t\t\t\t\t}},\n\t// \t\t\t\t\tProvisioningState: to.Ptr(armservicefabric.ProvisioningStateSucceeded),\n\t// \t\t\t\t\tReliabilityLevel: to.Ptr(armservicefabric.ReliabilityLevelSilver),\n\t// \t\t\t\t\tReverseProxyCertificateCommonNames: &armservicefabric.ServerCertificateCommonNames{\n\t// \t\t\t\t\t\tCommonNames: []*armservicefabric.ServerCertificateCommonName{\n\t// \t\t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\t\tCertificateCommonName: to.Ptr(\"abc.com\"),\n\t// \t\t\t\t\t\t\t\tCertificateIssuerThumbprint: to.Ptr(\"12599211F8F14C90AFA9532AD79A6F2CA1C00622\"),\n\t// \t\t\t\t\t\t}},\n\t// \t\t\t\t\t\tX509StoreName: to.Ptr(armservicefabric.StoreNameMy),\n\t// \t\t\t\t\t},\n\t// \t\t\t\t\tUpgradeDescription: &armservicefabric.ClusterUpgradePolicy{\n\t// \t\t\t\t\t\tDeltaHealthPolicy: &armservicefabric.ClusterUpgradeDeltaHealthPolicy{\n\t// \t\t\t\t\t\t\tApplicationDeltaHealthPolicies: map[string]*armservicefabric.ApplicationDeltaHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\"fabric:/myApp1\": &armservicefabric.ApplicationDeltaHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\tDefaultServiceTypeDeltaHealthPolicy: &armservicefabric.ServiceTypeDeltaHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\t\tMaxPercentDeltaUnhealthyServices: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\t\t\tServiceTypeDeltaHealthPolicies: map[string]*armservicefabric.ServiceTypeDeltaHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\t\t\"myServiceType1\": &armservicefabric.ServiceTypeDeltaHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\t\t\tMaxPercentDeltaUnhealthyServices: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tMaxPercentDeltaUnhealthyApplications: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\tMaxPercentDeltaUnhealthyNodes: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\tMaxPercentUpgradeDomainDeltaUnhealthyNodes: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t},\n\t// \t\t\t\t\t\tForceRestart: to.Ptr(false),\n\t// \t\t\t\t\t\tHealthCheckRetryTimeout: to.Ptr(\"00:05:00\"),\n\t// \t\t\t\t\t\tHealthCheckStableDuration: to.Ptr(\"00:00:30\"),\n\t// \t\t\t\t\t\tHealthCheckWaitDuration: to.Ptr(\"00:00:30\"),\n\t// \t\t\t\t\t\tHealthPolicy: &armservicefabric.ClusterHealthPolicy{\n\t// \t\t\t\t\t\t\tApplicationHealthPolicies: map[string]*armservicefabric.ApplicationHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\"fabric:/myApp1\": &armservicefabric.ApplicationHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\tDefaultServiceTypeHealthPolicy: &armservicefabric.ServiceTypeHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\t\tMaxPercentUnhealthyServices: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\t\t\tServiceTypeHealthPolicies: map[string]*armservicefabric.ServiceTypeHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\t\t\"myServiceType1\": &armservicefabric.ServiceTypeHealthPolicy{\n\t// \t\t\t\t\t\t\t\t\t\t\tMaxPercentUnhealthyServices: to.Ptr[int32](100),\n\t// \t\t\t\t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tMaxPercentUnhealthyApplications: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\tMaxPercentUnhealthyNodes: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t},\n\t// \t\t\t\t\t\tUpgradeDomainTimeout: to.Ptr(\"00:15:00\"),\n\t// \t\t\t\t\t\tUpgradeReplicaSetCheckTimeout: to.Ptr(\"00:10:00\"),\n\t// \t\t\t\t\t\tUpgradeTimeout: to.Ptr(\"01:00:00\"),\n\t// \t\t\t\t\t},\n\t// \t\t\t\t\tUpgradeMode: to.Ptr(armservicefabric.UpgradeModeManual),\n\t// \t\t\t\t\tVMImage: to.Ptr(\"Windows\"),\n\t// \t\t\t\t},\n\t// \t\t\t},\n\t// \t\t\t{\n\t// \t\t\t\tName: to.Ptr(\"myCluster2\"),\n\t// \t\t\t\tType: to.Ptr(\"Microsoft.ServiceFabric/clusters\"),\n\t// \t\t\t\tEtag: to.Ptr(\"W/\\\"636462502164040075\\\"\"),\n\t// \t\t\t\tID: to.Ptr(\"/subscriptions/00000000-0000-0000-0000-000000000000/resourcegroups/resRg/providers/Microsoft.ServiceFabric/clusters/myCluster2\"),\n\t// \t\t\t\tLocation: to.Ptr(\"eastus\"),\n\t// \t\t\t\tTags: map[string]*string{\n\t// \t\t\t\t},\n\t// \t\t\t\tProperties: &armservicefabric.ClusterProperties{\n\t// \t\t\t\t\tAddOnFeatures: []*armservicefabric.AddOnFeatures{\n\t// \t\t\t\t\t\tto.Ptr(armservicefabric.AddOnFeaturesRepairManager)},\n\t// \t\t\t\t\t\tAvailableClusterVersions: []*armservicefabric.ClusterVersionDetails{\n\t// \t\t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\t\tCodeVersion: to.Ptr(\"6.1.187.1\"),\n\t// \t\t\t\t\t\t\t\tEnvironment: to.Ptr(armservicefabric.ClusterEnvironmentLinux),\n\t// \t\t\t\t\t\t\t\tSupportExpiryUTC: to.Ptr(\"2018-06-15T23:59:59.9999999\"),\n\t// \t\t\t\t\t\t}},\n\t// \t\t\t\t\t\tClientCertificateCommonNames: []*armservicefabric.ClientCertificateCommonName{\n\t// \t\t\t\t\t\t},\n\t// \t\t\t\t\t\tClientCertificateThumbprints: []*armservicefabric.ClientCertificateThumbprint{\n\t// \t\t\t\t\t\t},\n\t// \t\t\t\t\t\tClusterCodeVersion: to.Ptr(\"6.1.187.1\"),\n\t// \t\t\t\t\t\tClusterEndpoint: to.Ptr(\"https://eastus.servicefabric.azure.com\"),\n\t// \t\t\t\t\t\tClusterID: to.Ptr(\"2747e469-b24e-4039-8a0a-46151419523f\"),\n\t// \t\t\t\t\t\tClusterState: to.Ptr(armservicefabric.ClusterStateWaitingForNodes),\n\t// \t\t\t\t\t\tDiagnosticsStorageAccountConfig: &armservicefabric.DiagnosticsStorageAccountConfig{\n\t// \t\t\t\t\t\t\tBlobEndpoint: to.Ptr(\"https://diag.blob.core.windows.net/\"),\n\t// \t\t\t\t\t\t\tProtectedAccountKeyName: to.Ptr(\"StorageAccountKey1\"),\n\t// \t\t\t\t\t\t\tQueueEndpoint: to.Ptr(\"https://diag.queue.core.windows.net/\"),\n\t// \t\t\t\t\t\t\tStorageAccountName: to.Ptr(\"diag\"),\n\t// \t\t\t\t\t\t\tTableEndpoint: to.Ptr(\"https://diag.table.core.windows.net/\"),\n\t// \t\t\t\t\t\t},\n\t// \t\t\t\t\t\tFabricSettings: []*armservicefabric.SettingsSectionDescription{\n\t// \t\t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\t\tName: to.Ptr(\"UpgradeService\"),\n\t// \t\t\t\t\t\t\t\tParameters: []*armservicefabric.SettingsParameterDescription{\n\t// \t\t\t\t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\t\t\t\tName: to.Ptr(\"AppPollIntervalInSeconds\"),\n\t// \t\t\t\t\t\t\t\t\t\tValue: to.Ptr(\"60\"),\n\t// \t\t\t\t\t\t\t\t}},\n\t// \t\t\t\t\t\t}},\n\t// \t\t\t\t\t\tManagementEndpoint: to.Ptr(\"http://myCluster2.eastus.cloudapp.azure.com:19080\"),\n\t// \t\t\t\t\t\tNodeTypes: []*armservicefabric.NodeTypeDescription{\n\t// \t\t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\t\tName: to.Ptr(\"nt1vm\"),\n\t// \t\t\t\t\t\t\t\tApplicationPorts: &armservicefabric.EndpointRangeDescription{\n\t// \t\t\t\t\t\t\t\t\tEndPort: to.Ptr[int32](30000),\n\t// \t\t\t\t\t\t\t\t\tStartPort: to.Ptr[int32](20000),\n\t// \t\t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\t\tClientConnectionEndpointPort: to.Ptr[int32](19000),\n\t// \t\t\t\t\t\t\t\tDurabilityLevel: to.Ptr(armservicefabric.DurabilityLevelBronze),\n\t// \t\t\t\t\t\t\t\tEphemeralPorts: &armservicefabric.EndpointRangeDescription{\n\t// \t\t\t\t\t\t\t\t\tEndPort: to.Ptr[int32](64000),\n\t// \t\t\t\t\t\t\t\t\tStartPort: to.Ptr[int32](49000),\n\t// \t\t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\t\tHTTPGatewayEndpointPort: to.Ptr[int32](19007),\n\t// \t\t\t\t\t\t\t\tIsPrimary: to.Ptr(true),\n\t// \t\t\t\t\t\t\t\tVMInstanceCount: to.Ptr[int32](5),\n\t// \t\t\t\t\t\t}},\n\t// \t\t\t\t\t\tProvisioningState: to.Ptr(armservicefabric.ProvisioningStateSucceeded),\n\t// \t\t\t\t\t\tReliabilityLevel: to.Ptr(armservicefabric.ReliabilityLevelSilver),\n\t// \t\t\t\t\t\tUpgradeDescription: &armservicefabric.ClusterUpgradePolicy{\n\t// \t\t\t\t\t\t\tDeltaHealthPolicy: &armservicefabric.ClusterUpgradeDeltaHealthPolicy{\n\t// \t\t\t\t\t\t\t\tMaxPercentDeltaUnhealthyApplications: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\t\tMaxPercentDeltaUnhealthyNodes: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\t\tMaxPercentUpgradeDomainDeltaUnhealthyNodes: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tForceRestart: to.Ptr(false),\n\t// \t\t\t\t\t\t\tHealthCheckRetryTimeout: to.Ptr(\"00:05:00\"),\n\t// \t\t\t\t\t\t\tHealthCheckStableDuration: to.Ptr(\"00:00:30\"),\n\t// \t\t\t\t\t\t\tHealthCheckWaitDuration: to.Ptr(\"00:00:30\"),\n\t// \t\t\t\t\t\t\tHealthPolicy: &armservicefabric.ClusterHealthPolicy{\n\t// \t\t\t\t\t\t\t\tMaxPercentUnhealthyApplications: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\t\tMaxPercentUnhealthyNodes: to.Ptr[int32](0),\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tUpgradeDomainTimeout: to.Ptr(\"00:15:00\"),\n\t// \t\t\t\t\t\t\tUpgradeReplicaSetCheckTimeout: to.Ptr(\"00:10:00\"),\n\t// \t\t\t\t\t\t\tUpgradeTimeout: to.Ptr(\"01:00:00\"),\n\t// \t\t\t\t\t\t},\n\t// \t\t\t\t\t\tUpgradeMode: to.Ptr(armservicefabric.UpgradeModeManual),\n\t// \t\t\t\t\t\tVMImage: to.Ptr(\"Ubuntu\"),\n\t// \t\t\t\t\t},\n\t// \t\t\t}},\n\t// \t\t}\n}", "func List() ([]clusterapi.Cluster, error) {\n\tvar clusterList []clusterapi.Cluster\n\terr := utils.BrowseMetadataContent(clusterapi.ClusterMetadataPrefix, func(buf *bytes.Buffer) error {\n\t\tvar c clusterapi.Cluster\n\t\terr := gob.NewDecoder(buf).Decode(&c)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tclusterList = append(clusterList, c)\n\t\treturn nil\n\t})\n\treturn clusterList, err\n}", "func (store *CenterStore) GetCenters(data []core.Elemt, space core.Space, k int, clust core.Clust) (core.Clust, error) {\n\tvar centers, ok = store.centers[k]\n\n\tif !ok {\n\t\treturn store.genCenters(data, space, k, clust)\n\t}\n\n\treturn centers, nil\n}", "func (s *ocmClient) GetCluster() (*ClusterInfo, error) {\n\n\t// fetch the clusterversion, which contains the internal ID\n\tcv := &configv1.ClusterVersion{}\n\terr := s.client.Get(context.TODO(), types.NamespacedName{Name: \"version\"}, cv)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"can't get clusterversion: %v\", err)\n\t}\n\texternalID := cv.Spec.ClusterID\n\n\tcsUrl, err := url.Parse(s.ocmBaseUrl.String())\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"can't parse OCM API url: %v\", err)\n\t}\n\tcsUrl.Path = path.Join(csUrl.Path, CLUSTERS_V1_PATH)\n\n\tresponse, err := s.httpClient.R().\n\t\tSetQueryParams(map[string]string{\n\t\t\t\"page\": \"1\",\n\t\t\t\"size\": \"1\",\n\t\t\t\"search\": fmt.Sprintf(\"external_id = '%s'\", externalID),\n\t\t}).\n\t\tSetResult(&ClusterList{}).\n\t\tExpectContentType(\"application/json\").\n\t\tGet(csUrl.String())\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"can't query OCM cluster service: request to '%v' returned error '%v'\", csUrl.String(), err)\n\t}\n\n\toperationId := response.Header().Get(OPERATION_ID_HEADER)\n\tif response.IsError() {\n\t\treturn nil, fmt.Errorf(\"request to '%v' received error code %v, operation id '%v'\", csUrl.String(), response.StatusCode(), operationId)\n\t}\n\n\tlog.Info(fmt.Sprintf(\"request to '%v' received response code %v, operation id: '%v'\", csUrl.String(), response.StatusCode(), operationId))\n\n\tlistResponse := response.Result().(*ClusterList)\n\tif listResponse.Size != 1 || len(listResponse.Items) != 1 {\n\t\treturn nil, ErrClusterIdNotFound\n\t}\n\n\treturn &listResponse.Items[0], nil\n}", "func RetrieveClusters(manifests string) cluster.Map {\n\tklog.V(1).Info(\"retrieving clusters from manifests\")\n\tclusters := cluster.Map{}\n\tdocuments := yamlutils.SplitDocuments(manifests)\n\tscheme := runtime.NewScheme()\n\tif err := clusterv1alpha1.AddToScheme(scheme); err != nil {\n\t\treturn cluster.Map{}\n\t}\n\tserializer := json.NewSerializerWithOptions(json.DefaultMetaFactory, scheme, scheme, json.SerializerOptions{Yaml: true})\n\tfor _, document := range documents {\n\t\tclusterObj := clusterv1alpha1.Cluster{}\n\t\tif _, _, err := serializer.Decode([]byte(document), nil, &clusterObj); err != nil || clusterObj.TypeMeta.Kind != \"Cluster\" {\n\t\t\tcontinue\n\t\t}\n\t\tinternalCluster, err := cluster.NewClusterFromv1alpha1(&clusterObj)\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\t\tclusters[internalCluster.Name] = internalCluster\n\t}\n\treturn clusters\n}", "func (p *Provider) List() ([]string, error) {\n\treturn p.provider.ListClusters()\n}", "func Clusters(api API) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tclusters := api.Clusters()\n\t\tm := make(map[string]map[string]any, len(clusters))\n\t\tfor _, c := range clusters {\n\t\t\tm[c.ID] = c.Debug()\n\t\t}\n\n\t\tdata, err := json.Marshal(m)\n\t\tif err != nil {\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\tfmt.Fprintf(w, \"could not marshal cluster debug map: %s\\n\", err)\n\t\t\treturn\n\t\t}\n\n\t\tw.Write(data)\n\t\tw.Write([]byte(\"\\n\"))\n\t}\n}", "func (svc ServerlessClusterService) Get(ctx context.Context,\n\tinput *models.GetServerlessClusterInput) (*models.Cluster, *Response, error) {\n\tvar cluster models.Cluster\n\tvar graphqlRequest = models.GraphqlRequest{\n\t\tName: \"cluster\",\n\t\tOperation: models.Query,\n\t\tInput: nil,\n\t\tArgs: *input,\n\t\tResponse: cluster,\n\t}\n\treq, err := svc.client.NewRequest(&graphqlRequest)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tresp, err := svc.client.Do(ctx, req, &cluster)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn &cluster, resp, err\n}", "func Clusters(clusters map[string]cluster.Cluster) []cluster.Cluster {\n\tcs := make([]cluster.Cluster, 0, len(clusters))\n\tfor _, cls := range clusters {\n\t\tcs = append(cs, cls)\n\t}\n\treturn cs\n}", "func (p *v1Provider) GetCluster(w http.ResponseWriter, r *http.Request) {\n\thttpapi.IdentifyEndpoint(r, \"/v1/clusters/current\")\n\ttoken := p.CheckToken(r)\n\tif !token.Require(w, \"cluster:show_basic\") {\n\t\treturn\n\t}\n\tshowBasic := !token.Check(\"cluster:show\")\n\n\tfilter := reports.ReadFilter(r, p.Cluster.GetServiceTypesForArea)\n\tif showBasic {\n\t\tfilter.IsSubcapacityAllowed = func(serviceType, resourceName string) bool {\n\t\t\ttoken.Context.Request[\"service\"] = serviceType\n\t\t\ttoken.Context.Request[\"resource\"] = resourceName\n\t\t\treturn token.Check(\"cluster:show_subcapacity\")\n\t\t}\n\t}\n\n\tcluster, err := reports.GetClusterResources(p.Cluster, p.DB, filter)\n\tif respondwith.ErrorText(w, err) {\n\t\treturn\n\t}\n\trespondwith.JSON(w, 200, map[string]interface{}{\"cluster\": cluster})\n}", "func listClusters(w http.ResponseWriter, r *http.Request, t auth.Token) (err error) {\n\tctx := r.Context()\n\tallowed := permission.Check(t, permission.PermClusterRead)\n\tif !allowed {\n\t\treturn permission.ErrUnauthorized\n\t}\n\tclusters, err := servicemanager.Cluster.List(ctx)\n\tif err != nil {\n\t\tif err == provTypes.ErrNoCluster {\n\t\t\tw.WriteHeader(http.StatusNoContent)\n\t\t\treturn nil\n\t\t}\n\t\treturn err\n\t}\n\tadmin := permission.Check(t, permission.PermClusterAdmin)\n\tif !admin {\n\t\tfor i := range clusters {\n\t\t\tclusters[i].CleanUpSensitive()\n\t\t}\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\treturn json.NewEncoder(w).Encode(clusters)\n}", "func AzureGetClusters(subscriptionID, clientID, clientSecret, tenantID, resourceGroupName string, admin bool) (string, error) {\n\tctx := context.Background()\n\tclient := containerservice.NewManagedClustersClient(subscriptionID)\n\n\tauthorizer, err := getAzureAuthorizer(clientID, clientSecret, tenantID)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tclient.Authorizer = authorizer\n\n\tvar clusters []string\n\n\tfor list, err := client.ListComplete(ctx); list.NotDone(); err = list.Next() {\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\tvar res containerservice.CredentialResults\n\t\tname := *list.Value().Name\n\n\t\tif admin {\n\t\t\tres, err = client.ListClusterAdminCredentials(ctx, resourceGroupName, name)\n\t\t\tif err != nil {\n\t\t\t\treturn \"\", err\n\t\t\t}\n\t\t} else {\n\t\t\tres, err = client.ListClusterUserCredentials(ctx, resourceGroupName, name)\n\t\t\tif err != nil {\n\t\t\t\treturn \"\", err\n\t\t\t}\n\t\t}\n\n\t\tfor _, kubeconfig := range *res.Kubeconfigs {\n\t\t\tvar kubeconfigJSON interface{}\n\t\t\terr := yaml.Unmarshal(*kubeconfig.Value, &kubeconfigJSON)\n\t\t\tif err != nil {\n\t\t\t\treturn \"\", err\n\t\t\t}\n\n\t\t\tkubeconfigJSON = convert(kubeconfigJSON)\n\t\t\tkubeconfigJSONString, err := json.Marshal(kubeconfigJSON)\n\t\t\tif err != nil {\n\t\t\t\treturn \"\", err\n\t\t\t}\n\n\t\t\tclusters = append(clusters, fmt.Sprintf(\"{\\\"name\\\": \\\"%s_%s_%s\\\", \\\"kubeconfig\\\": %s}\", *kubeconfig.Name, resourceGroupName, name, kubeconfigJSONString))\n\t\t}\n\t}\n\n\treturn fmt.Sprintf(\"[%s]\", strings.Join(clusters, \",\")), nil\n}", "func (a *Client) VirtualizationClustersRead(params *VirtualizationClustersReadParams) (*VirtualizationClustersReadOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewVirtualizationClustersReadParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"virtualization_clusters_read\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/virtualization/clusters/{id}/\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &VirtualizationClustersReadReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*VirtualizationClustersReadOK), nil\n\n}", "func GetClusterCIDRs(lister configlistersv1.NetworkLister, recorder events.Recorder) ([]string, error) {\n\tnetwork, err := lister.Get(\"cluster\")\n\tif errors.IsNotFound(err) {\n\t\trecorder.Warningf(\"ObserveRestrictedCIDRFailed\", \"Required networks.%s/cluster not found\", configv1.GroupName)\n\t\treturn nil, nil\n\t}\n\tif err != nil {\n\t\trecorder.Warningf(\"ObserveRestrictedCIDRFailed\", \"error getting networks.%s/cluster: %v\", configv1.GroupName, err)\n\t\treturn nil, err\n\t}\n\n\tif len(network.Status.ClusterNetwork) == 0 {\n\t\trecorder.Warningf(\"ObserveClusterCIDRFailed\", \"Required status.clusterNetwork field is not set in networks.%s/cluster\", configv1.GroupName)\n\t\treturn nil, fmt.Errorf(\"networks.%s/cluster: status.clusterNetwork not found\", configv1.GroupName)\n\t}\n\n\tvar clusterCIDRs []string\n\tfor i, clusterNetwork := range network.Status.ClusterNetwork {\n\t\tif len(clusterNetwork.CIDR) == 0 {\n\t\t\trecorder.Warningf(\"ObserveRestrictedCIDRFailed\", \"Required status.clusterNetwork[%d].cidr field is not set in networks.%s/cluster\", i, configv1.GroupName)\n\t\t\treturn nil, fmt.Errorf(\"networks.%s/cluster: status.clusterNetwork[%d].cidr not found\", configv1.GroupName, i)\n\t\t}\n\t\tclusterCIDRs = append(clusterCIDRs, clusterNetwork.CIDR)\n\t}\n\t// TODO fallback to podCIDR? is that still a thing?\n\treturn clusterCIDRs, nil\n}", "func (a *ClusterControllerApiService) GetClustersUsingGET(ctx _context.Context, account string, application string, clusterName string) apiGetClustersUsingGETRequest {\n\treturn apiGetClustersUsingGETRequest{\n\t\tapiService: a,\n\t\tctx: ctx,\n\t\taccount: account,\n\t\tapplication: application,\n\t\tclusterName: clusterName,\n\t}\n}", "func NewClusters(db *gorm.DB) *Clusters {\n\treturn &Clusters{db: db}\n}", "func (o *ResourceLimits) GetK8sClustersProvisioned() *int32 {\n\tif o == nil {\n\t\treturn nil\n\t}\n\n\treturn o.K8sClustersProvisioned\n}", "func ExampleSnowball_ListClusters_shared00() {\n\tsvc := snowball.New(session.New())\n\tinput := &snowball.ListClustersInput{}\n\n\tresult, err := svc.ListClusters(input)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase snowball.ErrCodeInvalidNextTokenException:\n\t\t\t\tfmt.Println(snowball.ErrCodeInvalidNextTokenException, aerr.Error())\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t}\n\t\treturn\n\t}\n\n\tfmt.Println(result)\n}", "func (a *ClustersApiService) ClusterServiceListClusters(ctx context.Context, body Servicev1ClusterQuery) (V1Clusterlist, *http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Post\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue V1Clusterlist\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/gitops/api/v1/clusters\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tlocalVarQueryParams.Add(\"routingId\", body.AccountIdentifier)\n\t// body params\n\tlocalVarPostBody = &body\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"x-api-key\"] = key\n\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tif localVarHttpResponse.StatusCode < 300 {\n\t\t// If we succeed, return the data, otherwise pass on to decode error.\n\t\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\tif err == nil {\n\t\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t\t}\n\t}\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 200 {\n\t\t\tvar v V1Clusterlist\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 0 {\n\t\t\tvar v GatewayruntimeError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHttpResponse, nil\n}", "func AWSGetClusters(accessKeyId, secretAccessKey, region string) (string, error) {\n\tvar clusters []*eks.Cluster\n\tvar names []*string\n\tvar nextToken *string\n\n\tcred := credentials.NewStaticCredentials(accessKeyId, secretAccessKey, \"\")\n\n\tsess, err := session.NewSession(&aws.Config{Region: aws.String(region), Credentials: cred})\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\teksClient := eks.New(sess)\n\n\tfor {\n\t\tc, err := eksClient.ListClusters(&eks.ListClustersInput{NextToken: nextToken})\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\tnames = append(names, c.Clusters...)\n\n\t\tif c.NextToken == nil {\n\t\t\tbreak\n\t\t}\n\n\t\tnextToken = c.NextToken\n\t}\n\n\tfor _, name := range names {\n\t\tcluster, err := eksClient.DescribeCluster(&eks.DescribeClusterInput{Name: name})\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\tif *cluster.Cluster.Status == eks.ClusterStatusActive {\n\t\t\tclusters = append(clusters, cluster.Cluster)\n\t\t}\n\t}\n\n\tif clusters != nil {\n\t\tb, err := json.Marshal(clusters)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\treturn string(b), nil\n\t}\n\n\treturn \"\", nil\n}", "func (a *Client) VirtualizationClustersList(params *VirtualizationClustersListParams) (*VirtualizationClustersListOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewVirtualizationClustersListParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"virtualization_clusters_list\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/virtualization/clusters/\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &VirtualizationClustersListReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*VirtualizationClustersListOK), nil\n\n}", "func NewGetClustersOK() *GetClustersOK {\n\treturn &GetClustersOK{}\n}", "func getClusterNameForMultiVC(ctx context.Context, vs *multiVCvSphere,\n\tclientIndex int) ([]*object.ClusterComputeResource,\n\t*VsanClient, error) {\n\n\tvar vsanHealthClient *VsanClient\n\tvar err error\n\tc := newClientForMultiVC(ctx, vs)\n\n\tdatacenter := strings.Split(multiVCe2eVSphere.multivcConfig.Global.Datacenters, \",\")\n\n\tfor i, client := range c {\n\t\tif clientIndex == i {\n\t\t\tvsanHealthClient, err = newVsanHealthSvcClient(ctx, client.Client)\n\t\t\tgomega.Expect(err).NotTo(gomega.HaveOccurred())\n\t\t}\n\t}\n\n\tfinder := find.NewFinder(vsanHealthClient.vim25Client, false)\n\tdc, err := finder.Datacenter(ctx, datacenter[0])\n\tgomega.Expect(err).NotTo(gomega.HaveOccurred())\n\tfinder.SetDatacenter(dc)\n\n\tclusterComputeResource, err := finder.ClusterComputeResourceList(ctx, \"*\")\n\tframework.Logf(\"clusterComputeResource %v\", clusterComputeResource)\n\tgomega.Expect(err).NotTo(gomega.HaveOccurred())\n\n\treturn clusterComputeResource, vsanHealthClient, err\n}", "func (o GetClustersResultOutput) ClusterIdentifiers() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v GetClustersResult) []string { return v.ClusterIdentifiers }).(pulumi.StringArrayOutput)\n}", "func (elementConfiguration *ElementConfiguration) ListClusters() ([]string, error) {\n\t// collect names\n\tclusterConfigurations := []string{}\n\n elementConfiguration.ClustersX.RLock()\n\tfor clusterConfiguration := range elementConfiguration.Clusters {\n\t\tclusterConfigurations = append(clusterConfigurations, clusterConfiguration)\n\t}\n\telementConfiguration.ClustersX.RUnlock()\n\n\t// success\n\treturn clusterConfigurations, nil\n}", "func (a *Client) GetMsgVpnDistributedCacheClusters(params *GetMsgVpnDistributedCacheClustersParams, authInfo runtime.ClientAuthInfoWriter) (*GetMsgVpnDistributedCacheClustersOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetMsgVpnDistributedCacheClustersParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"getMsgVpnDistributedCacheClusters\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/msgVpns/{msgVpnName}/distributedCaches/{cacheName}/clusters\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &GetMsgVpnDistributedCacheClustersReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetMsgVpnDistributedCacheClustersOK), nil\n\n}", "func (cb *clientBase) GetCluster() string {\n\treturn cb.cluster\n}", "func (s *RpcClient) GetClusterNodes(ctx context.Context) ([]GetClusterNodesResponse, error) {\n\tres := struct {\n\t\tGeneralResponse\n\t\tResult []GetClusterNodesResponse `json:\"result\"`\n\t}{}\n\terr := s.request(ctx, \"getClusterNodes\", []interface{}{}, &res)\n\tif err != nil {\n\t\treturn []GetClusterNodesResponse{}, err\n\t}\n\tif res.Error != nil {\n\t\treturn []GetClusterNodesResponse{}, errors.New(res.Error.Message)\n\t}\n\treturn res.Result, nil\n}", "func (op *outputProvider) GetRemoteClusters(opts ...services.MarshalOption) ([]types.RemoteCluster, error) {\n\treturn op.impersonatedClient.GetRemoteClusters(opts...)\n}", "func (page ClusterListResultPage) Values() []Cluster {\n\tif page.clr.IsEmpty() {\n\t\treturn nil\n\t}\n\treturn *page.clr.Value\n}", "func (m *RedisProxy) GetCluster() string {\n\tif m != nil {\n\t\treturn m.Cluster\n\t}\n\treturn \"\"\n}", "func (connection *Connection) GetClusterNodes() []*URL {\n\tif connection.IsCluster() {\n\t\treturn connection.adabasToData.transactions.clusterNodes\n\t}\n\treturn make([]*URL, 0)\n}", "func GetClusterId() string {\n\treturn axClusterId\n}", "func (a *Client) V2ListClusters(ctx context.Context, params *V2ListClustersParams) (*V2ListClustersOK, error) {\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"v2ListClusters\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/v2/clusters\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &V2ListClustersReader{formats: a.formats},\n\t\tAuthInfo: a.authInfo,\n\t\tContext: ctx,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*V2ListClustersOK), nil\n\n}", "func ListAllClusters(response *JsonListClustersMap) *JsonListClustersMap {\n\tvar SIDCluster int\n\tvar SName string\n\tvar SAWSAccount int64\n\tvar SAWSRegion string\n\tvar SAWSEnvironment string\n\tvar SK8sVersion string\n\n\tvar SNodeType string\n\tvar SNodeInstance string\n\tvar STotalInstances int\n\n\tvar totalInstances int\n\n\tdescription := make(DescriptionMap)\n\n\tdb, err := sql.Open(\"mysql\", UserDB+\":\"+PassDB+\"@tcp(\"+HostDB+\":\"+PortDB+\")/\"+DatabaseDB+\"?charset=utf8\")\n\tcheckErr(err)\n\n\tdefer db.Close()\n\n\trows, err := db.Query(\"SELECT id_cluster, nome, aws_account, aws_region, aws_env, k8s_version FROM clusters ORDER BY nome\")\n\tcheckErr(err)\n\n\tfor rows.Next() {\n\t\terr = rows.Scan(&SIDCluster, &SName, &SAWSAccount, &SAWSRegion, &SAWSEnvironment, &SK8sVersion)\n\t\tcheckErr(err)\n\n\t\tdescription = DescriptionMap{}\n\t\ttotalInstances = 0\n\n\t\trows1, err := db.Query(\"SELECT node_type, node_instance, total_instances FROM nodes WHERE id_cluster=?\", SIDCluster)\n\t\tcheckErr(err)\n\n\t\tfor rows1.Next() {\n\t\t\terr = rows1.Scan(&SNodeType, &SNodeInstance, &STotalInstances)\n\t\t\tcheckErr(err)\n\n\t\t\tdescription[SNodeType] = append(\n\t\t\t\tdescription[SNodeType],\n\t\t\t\tDescriptionStruct{\n\t\t\t\t\tDescription{\n\t\t\t\t\t\tType: SNodeInstance,\n\t\t\t\t\t\tTotalTypeInstances: STotalInstances,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t)\n\n\t\t\ttotalInstances = totalInstances + STotalInstances\n\t\t}\n\n\t\t*response = append(\n\t\t\t*response,\n\t\t\tjsonListClusters{\n\t\t\t\tClusterName: SName,\n\t\t\t\tAws: AWS{\n\t\t\t\t\tAccount: SAWSAccount,\n\t\t\t\t\tRegion: SAWSRegion,\n\t\t\t\t\tEnvironment: SAWSEnvironment,\n\t\t\t\t},\n\t\t\t\tK8SVersion: SK8sVersion,\n\t\t\t\tInstances: Instances{\n\t\t\t\t\tTotalInstances: totalInstances,\n\t\t\t\t\tDescription: description,\n\t\t\t\t},\n\t\t\t},\n\t\t)\n\t}\n\n\treturn response\n}", "func (m *Manager) GetClusterList() ([]Cluster, error) {\n\tnames, err := m.specManager.List()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar clusters = []Cluster{}\n\n\tfor _, name := range names {\n\t\tmetadata, err := m.meta(name)\n\t\tif err != nil && !errors.Is(perrs.Cause(err), meta.ErrValidate) &&\n\t\t\t!errors.Is(perrs.Cause(err), spec.ErrNoTiSparkMaster) {\n\t\t\treturn nil, perrs.Trace(err)\n\t\t}\n\n\t\tbase := metadata.GetBaseMeta()\n\n\t\tclusters = append(clusters, Cluster{\n\t\t\tName: name,\n\t\t\tUser: base.User,\n\t\t\tVersion: base.Version,\n\t\t\tPath: m.specManager.Path(name),\n\t\t\tPrivateKey: m.specManager.Path(name, \"ssh\", \"id_rsa\"),\n\t\t})\n\t}\n\n\treturn clusters, nil\n}", "func (r *ProjectsInstancesClustersService) List(parent string) *ProjectsInstancesClustersListCall {\n\tc := &ProjectsInstancesClustersListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\treturn c\n}", "func GetClusterMode() string {\n\treturn masterRTCfg.clusterMode\n}", "func (p PGSQLConnection) GetAllClusters() ([]ClusterModel, error) {\n\tclusters := []ClusterModel{}\n\tif err := p.connection.Select(&clusters, \"SELECT * FROM clusters\"); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn clusters, nil\n}", "func (o GetClustersClusterOutput) ClusterId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetClustersCluster) string { return v.ClusterId }).(pulumi.StringOutput)\n}", "func clusterList() []string {\n\tif c := envy.String(\"DQLITED_CLUSTER\"); c != \"\" {\n\t\treturn strings.Split(c, \",\")\n\t}\n\treturn defaultCluster\n}", "func (d *Dao) OverlordClusters(c context.Context, zone, appid string) (ocs []*model.OverlordCluster, err error) {\n\tvar res struct {\n\t\tData []*model.OverlordApiserver `json:\"grouped_clusters\"`\n\t}\n\tif err = d.client.RESTfulGet(c, apiserverURI, \"\", nil, &res, appid); err != nil {\n\t\tlog.Error(\"overlord cluster url(%s) appid(%s) error(%v)\", apiserverURI, appid, err)\n\t\treturn\n\t}\nGETALL:\n\tfor _, oa := range res.Data {\n\t\tif zone == \"\" || oa.Group == zone {\n\t\t\tfor _, oc := range oa.Clusters {\n\t\t\t\tcluster := &model.OverlordCluster{\n\t\t\t\t\tName: oc.Name,\n\t\t\t\t\tType: oc.Type,\n\t\t\t\t\tZone: zone,\n\t\t\t\t\tHashMethod: \"fnv1a_64\",\n\t\t\t\t\tHashDistribution: \"ketama\",\n\t\t\t\t\tHashTag: \"{}\",\n\t\t\t\t\tListenProto: \"tcp\",\n\t\t\t\t\tListenAddr: net.JoinHostPort(\"0.0.0.0\", strconv.Itoa(oc.FrontEndPort)),\n\t\t\t\t\tDailTimeout: 1000,\n\t\t\t\t\tReadTimeout: 1000,\n\t\t\t\t\tWriteTimeout: 1000,\n\t\t\t\t\tNodeConn: 2,\n\t\t\t\t\tPingFailLimit: 3,\n\t\t\t\t\tPingAutoEject: true,\n\t\t\t\t}\n\t\t\t\tfor _, oci := range oc.Instances {\n\t\t\t\t\tif oc.Type == \"redis_cluster\" && oci.Role != \"master\" {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\ton := &model.OverlordNode{\n\t\t\t\t\t\tAlias: oci.Alias,\n\t\t\t\t\t\tAddr: net.JoinHostPort(oci.IP, strconv.Itoa(oci.Port)),\n\t\t\t\t\t\tWeight: oci.Weight,\n\t\t\t\t\t}\n\t\t\t\t\tcluster.Nodes = append(cluster.Nodes, on)\n\t\t\t\t}\n\t\t\t\tocs = append(ocs, cluster)\n\t\t\t}\n\t\t}\n\t}\n\tif len(ocs) == 0 && zone != \"\" {\n\t\tzone = \"\"\n\t\tgoto GETALL\n\t}\n\treturn\n}", "func fetchCluster(c *gin.Context) string {\n\tconst key = \"cluster\"\n\n\tswitch {\n\tcase len(c.Param(key)) > 0:\n\t\treturn c.Param(key)\n\tcase len(c.Query(key)) > 0:\n\t\treturn c.Query(key)\n\tcase len(c.PostForm(key)) > 0:\n\t\treturn c.PostForm(key)\n\tdefault:\n\t\treturn \"\"\n\t}\n}", "func (a *Client) GetClusterCredentials(params *GetClusterCredentialsParams, authInfo runtime.ClientAuthInfoWriter) (*GetClusterCredentialsOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetClusterCredentialsParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"GetClusterCredentials\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/v1/clusters/{name}/credentials\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &GetClusterCredentialsReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetClusterCredentialsOK), nil\n\n}", "func (o LookupResponsePolicyResultOutput) GkeClusters() ResponsePolicyGKEClusterResponseArrayOutput {\n\treturn o.ApplyT(func(v LookupResponsePolicyResult) []ResponsePolicyGKEClusterResponse { return v.GkeClusters }).(ResponsePolicyGKEClusterResponseArrayOutput)\n}", "func GetManegementCluster(version, capiImage, capdImage string) ([]runtime.Object, error) {\n\tcapiObjects, err := GetCAPI(version, capiImage)\n\tif err != nil {\n\t\treturn []runtime.Object{}, err\n\t}\n\n\tnamespaceObj := GetNamespace()\n\tstatefulSet := GetStatefulSet(capdImage)\n\tclusterRole := GetClusterRole()\n\tclusterRoleBinding := GetClusterRoleBinding()\n\n\treturn append(capiObjects,\n\t\t&namespaceObj,\n\t\t&statefulSet,\n\t\t&clusterRole,\n\t\t&clusterRoleBinding,\n\t), nil\n}", "func (c *Client) Cluster(ctx context.Context) ([]NodeInfo, error) {\n\trequest := protocol.Message{}\n\trequest.Init(16)\n\tresponse := protocol.Message{}\n\tresponse.Init(512)\n\n\tprotocol.EncodeCluster(&request, protocol.ClusterFormatV1)\n\n\tif err := c.protocol.Call(ctx, &request, &response); err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to send Cluster request\")\n\t}\n\n\tservers, err := protocol.DecodeNodes(&response)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to parse Node response\")\n\t}\n\n\treturn servers, nil\n}", "func GetClusterIPs(service *corev1.Service) []string {\n\tclusterIPs := []string{service.Spec.ClusterIP}\n\tif len(service.Spec.ClusterIPs) > 0 {\n\t\tclusterIPs = service.Spec.ClusterIPs\n\t}\n\n\t// Same IPv6 could be represented differently (as from rfc5952):\n\t// 2001:db8:0:0:aaaa::1\n\t// 2001:db8::aaaa:0:0:1\n\t// 2001:db8:0::aaaa:0:0:1\n\t// net.ParseIP(ip).String() output is used as a normalization form\n\t// for all cases above it returns 2001:db8::aaaa:0:0:1\n\t// without the normalization there could be mismatches in key lookups e.g. for PTR\n\tnormalized := make([]string, 0, len(clusterIPs))\n\tfor _, ip := range clusterIPs {\n\t\tnormalized = append(normalized, net.ParseIP(ip).String())\n\t}\n\n\treturn normalized\n}", "func (a *Client) ListAvailableClusters(ctx context.Context, params *ListAvailableClustersParams) (*ListAvailableClustersOK, error) {\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"listAvailableClusters\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/heappe/ClusterInformation/ListAvailableClusters\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &ListAvailableClustersReader{formats: a.formats},\n\t\tAuthInfo: a.authInfo,\n\t\tContext: ctx,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*ListAvailableClustersOK), nil\n\n}" ]
[ "0.7150111", "0.70848936", "0.7050372", "0.69830257", "0.69812804", "0.68225586", "0.6785897", "0.67075115", "0.67008924", "0.66717374", "0.6653548", "0.6628853", "0.65589386", "0.6548702", "0.6535888", "0.6530604", "0.6489139", "0.6475645", "0.6451163", "0.6437961", "0.64292145", "0.6420646", "0.64115655", "0.6351999", "0.6342829", "0.63368356", "0.63260955", "0.63172734", "0.6294134", "0.6274878", "0.62505454", "0.62502795", "0.6240707", "0.6152818", "0.61403", "0.61298174", "0.6111065", "0.6086631", "0.60699946", "0.605888", "0.6055227", "0.6039949", "0.60248417", "0.60127914", "0.5990848", "0.5971463", "0.5963538", "0.5935251", "0.59316564", "0.59266365", "0.59207803", "0.59110713", "0.5910659", "0.590048", "0.5897609", "0.58610535", "0.58484685", "0.5844702", "0.58280903", "0.5825456", "0.58106434", "0.5751794", "0.5746291", "0.5745103", "0.57434946", "0.5741837", "0.5721784", "0.57198244", "0.57118505", "0.5692283", "0.567878", "0.5669264", "0.56662285", "0.5651515", "0.56448567", "0.563336", "0.56233066", "0.56175965", "0.5593433", "0.55922747", "0.5577536", "0.55719817", "0.55662704", "0.5562364", "0.5552811", "0.55524254", "0.5536599", "0.55298865", "0.5524156", "0.5522188", "0.5513524", "0.55062175", "0.55043143", "0.5485468", "0.5481368", "0.5479709", "0.5474633", "0.54723483", "0.5467098", "0.5448341" ]
0.76226926
0
"GetClustersOk returns a tuple with the Clusters field value and a boolean to check if the value has(...TRUNCATED)
"GetClustersOk возвращает кортеж с значением поля Clusters и буле(...TRUNCATED)
"func (o *QueueManager) GetClustersOk() (*[]string, bool) {\n\tif o == nil {\n\t\treturn nil, false(...TRUNCATED)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
["func NewGetClustersOK() *GetClustersOK {\n\treturn &GetClustersOK{}\n}","func (o *ResourceLimits) (...TRUNCATED)
["0.6719827","0.65017885","0.6465093","0.646219","0.6398727","0.63660103","0.63552487","0.63049424",(...TRUNCATED)
0.8224033
0
GetAliasQueues returns the AliasQueues field value
GetAliasQueues возвращает значение поля AliasQueues
"func (o *QueueManager) GetAliasQueues() []AliasQueue {\n\tif o == nil {\n\t\tvar ret []AliasQueue\(...TRUNCATED)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
["func (o *QueueManager) GetAliasQueuesOk() (*[]AliasQueue, bool) {\n\tif o == nil {\n\t\treturn ni(...TRUNCATED)
["0.69151425","0.6435573","0.6247431","0.6227677","0.6098082","0.59897524","0.58581173","0.5692911",(...TRUNCATED)
0.7781025
0
"GetAliasQueuesOk returns a tuple with the AliasQueues field value and a boolean to check if the val(...TRUNCATED)
"GetAliasQueuesOk возвращает кортеж с значением поля AliasQueues и б(...TRUNCATED)
"func (o *QueueManager) GetAliasQueuesOk() (*[]AliasQueue, bool) {\n\tif o == nil {\n\t\treturn nil(...TRUNCATED)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
["func (o *QueueManager) GetAliasQueues() []AliasQueue {\n\tif o == nil {\n\t\tvar ret []AliasQueue(...TRUNCATED)
["0.62684894","0.6182187","0.6036427","0.5732854","0.55811864","0.5505796","0.5485279","0.54428786",(...TRUNCATED)
0.8438502
0
GetRemoteQueues returns the RemoteQueues field value
GetRemoteQueues возвращает значение поля RemoteQueues
"func (o *QueueManager) GetRemoteQueues() []RemoteQueue {\n\tif o == nil {\n\t\tvar ret []RemoteQue(...TRUNCATED)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
["func (o *QueueManager) GetRemoteQueuesOk() (*[]RemoteQueue, bool) {\n\tif o == nil {\n\t\treturn (...TRUNCATED)
["0.71619385","0.61695373","0.6127869","0.6031343","0.5995589","0.5916818","0.58918214","0.5793508",(...TRUNCATED)
0.76292956
0
"GetRemoteQueuesOk returns a tuple with the RemoteQueues field value and a boolean to check if the v(...TRUNCATED)
"GetRemoteQueuesOk возвращает кортеж с значением поля RemoteQueues и (...TRUNCATED)
"func (o *QueueManager) GetRemoteQueuesOk() (*[]RemoteQueue, bool) {\n\tif o == nil {\n\t\treturn n(...TRUNCATED)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
["func (o *QueueManager) GetRemoteQueues() []RemoteQueue {\n\tif o == nil {\n\t\tvar ret []RemoteQu(...TRUNCATED)
["0.63053876","0.61726433","0.5913533","0.5784514","0.57285035","0.533683","0.53098166","0.52802896"(...TRUNCATED)
0.8458737
0
GetClusterQueues returns the ClusterQueues field value
GetClusterQueues возвращает значение поля ClusterQueues
"func (o *QueueManager) GetClusterQueues() []ClusterQueue {\n\tif o == nil {\n\t\tvar ret []Cluster(...TRUNCATED)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
["func (o *QueueManager) GetClusterQueuesOk() (*[]ClusterQueue, bool) {\n\tif o == nil {\n\t\tretur(...TRUNCATED)
["0.6691833","0.6425696","0.6164038","0.6133957","0.6117693","0.6031181","0.60217136","0.5825728","0(...TRUNCATED)
0.741813
0
"GetClusterQueuesOk returns a tuple with the ClusterQueues field value and a boolean to check if the(...TRUNCATED)
"GetClusterQueuesOk возвращает кортеж с значением поля ClusterQueues (...TRUNCATED)
"func (o *QueueManager) GetClusterQueuesOk() (*[]ClusterQueue, bool) {\n\tif o == nil {\n\t\treturn(...TRUNCATED)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
["func (o *QueueManager) GetAliasQueuesOk() (*[]AliasQueue, bool) {\n\tif o == nil {\n\t\treturn ni(...TRUNCATED)
["0.64735645","0.64523965","0.6035547","0.58038056","0.53999776","0.53208125","0.5234418","0.5219659(...TRUNCATED)
0.832443
0
End of preview. Expand in Data Studio

The part of CoRNStack Dataset translated into Russian. Translation was done with Qwen3 model.

Samples that satisfy the dual consistency filtering condition (samples where the document_rank is 0 or 1 and document_score > 0.7) were translated.

Source code you can find here. For support please: fedor.yaronskiy@gmail.com

Downloads last month
46

Collection including fyaronskiy/cornstack_go_ru_en

Paper for fyaronskiy/cornstack_go_ru_en