query
stringlengths
7
3.85k
document
stringlengths
11
430k
metadata
dict
negatives
sequencelengths
0
101
negative_scores
sequencelengths
0
101
document_score
stringlengths
3
10
document_rank
stringclasses
102 values
TaskConfigSchema returns the schema for the driver configuration of the task.
func (d *Driver) TaskConfigSchema() (*hclspec.Spec, error) { return taskConfigSpec, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (d *Driver) ConfigSchema() (*hclspec.Spec, error) {\n\treturn configSpec, nil\n}", "func GetUserConfigSchema(t string) map[string]interface{} {\n\tif _, ok := getUserConfigurationOptionsSchemaFilenames()[t]; !ok {\n\t\tlog.Panicf(\"user configuration options schema type `%s` is not available\", t)\n\t}\n\n\treturn userConfigSchemas[t]\n}", "func tagMakeConfigSchema(tagAttrName schemaAttr) *schema.Schema {\n\treturn &schema.Schema{\n\t\tType: schema.TypeSet,\n\t\tOptional: true,\n\t\tElem: &schema.Schema{\n\t\t\tType: schema.TypeString,\n\t\t\tValidateFunc: validateTag,\n\t\t},\n\t}\n}", "func BackendSchema(factoryName string) (*config.ConfigurationSchema, error) {\n\tif backendRegistry[factoryName] == nil {\n\t\treturn nil, fmt.Errorf(\"The adapter %s is not registered Processor cannot be created\", factoryName)\n\t}\n\treturn backendRegistry[factoryName].configurationSchema, nil\n}", "func (o ParserConfigOutput) Schema() SchemaPackagePtrOutput {\n\treturn o.ApplyT(func(v ParserConfig) *SchemaPackage { return v.Schema }).(SchemaPackagePtrOutput)\n}", "func (o TableExternalDataConfigurationOutput) Schema() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TableExternalDataConfiguration) *string { return v.Schema }).(pulumi.StringPtrOutput)\n}", "func ImageConfigSchema() *gojsonschema.Schema {\n\treturn loadSchema(\"image-config.schema.json\")\n}", "func (j *AuroraJob) TaskConfig() *aurora.TaskConfig {\n\treturn j.jobConfig.TaskConfig\n}", "func autopilotConfigTableSchema() *memdb.TableSchema {\n\treturn &memdb.TableSchema{\n\t\tName: \"autopilot-config\",\n\t\tIndexes: map[string]*memdb.IndexSchema{\n\t\t\t\"id\": &memdb.IndexSchema{\n\t\t\t\tName: \"id\",\n\t\t\t\tAllowMissing: true,\n\t\t\t\tUnique: true,\n\t\t\t\tIndexer: &memdb.ConditionalIndex{\n\t\t\t\t\tConditional: func(obj interface{}) (bool, error) { return true, nil },\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func (o ParserConfigPtrOutput) Schema() SchemaPackagePtrOutput {\n\treturn o.ApplyT(func(v *ParserConfig) *SchemaPackage {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Schema\n\t}).(SchemaPackagePtrOutput)\n}", "func (t *TableSchema) schema() (string, error) {\n\tswitch t.Driver {\n\tcase DriverMysql:\n\t\treturn t.schemaMysql()\n\tcase DriverSQLite, DriverSQLite3:\n\t\treturn t.schemaSQLite()\n\tdefault:\n\t\treturn \"\", fmt.Errorf(\"not support driver: %s\", t.Driver)\n\t}\n}", "func (o GoogleCloudHealthcareV1beta1FhirBigQueryDestinationOutput) SchemaConfig() SchemaConfigPtrOutput {\n\treturn o.ApplyT(func(v GoogleCloudHealthcareV1beta1FhirBigQueryDestination) *SchemaConfig { return v.SchemaConfig }).(SchemaConfigPtrOutput)\n}", "func (o TableExternalDataConfigurationPtrOutput) Schema() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TableExternalDataConfiguration) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Schema\n\t}).(pulumi.StringPtrOutput)\n}", "func TaskConfiguration() TaskConfig {\n\treturn conf.TaskConfig\n}", "func (task *QueueTask) GetConfig() *TaskConfig {\n\treturn &TaskConfig{\n\t\tTaskID: task.taskID,\n\t\tTaskType: task.TaskType,\n\t\tIsRun: task.IsRun,\n\t\tHandler: task.handler,\n\t\tDueTime: task.DueTime,\n\t\tInterval: 0,\n\t\tExpress: \"\",\n\t\tTaskData: task.TaskData,\n\t}\n}", "func (v *templateTableType) Schema() string {\n\treturn v.s.SQLSchema\n}", "func (t Table) Schema() string {\n\treturn t.schema\n}", "func DeploymentSchema() string {\n\treturn deploymentSchema\n}", "func GetClusterConfigSchema(extensionSchema string) string {\n\tvar clusterConfigSchema string\n\tif clusterConfigSchema == \"\" {\n\t\tclusterConfigSchema = fmt.Sprintf(ClusterConfigSpecSchemaTemplate, \"\")\n\t} else {\n\t\tclusterConfigSchema = fmt.Sprintf(ClusterConfigSpecSchemaTemplate, \",\"+extensionSchema)\n\t}\n\treturn fmt.Sprintf(V2SchemaTemplate, MetadataSchema, clusterConfigSchema, DefaultDefinitions)\n}", "func (d *Describe) Schema() sql.Schema {\n\treturn sql.Schema{{\n\t\tName: \"name\",\n\t\tType: VarChar25000,\n\t}, {\n\t\tName: \"type\",\n\t\tType: VarChar25000,\n\t}}\n}", "func (m modelPluginTestDevice) Schema() (map[string]*yang.Entry, error) {\n\treturn td1.UnzipSchema()\n}", "func (e *commonFormatEncoder) Schema() *types.TableSchema {\n\treturn e.inSchema\n}", "func (v *permutationTableType) Schema() string {\n\treturn v.s.SQLSchema\n}", "func (m *BusinessScenarioPlanner) GetTaskConfiguration()(PlannerTaskConfigurationable) {\n val, err := m.GetBackingStore().Get(\"taskConfiguration\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(PlannerTaskConfigurationable)\n }\n return nil\n}", "func ConfigForTask(taskInfo *mesos.TaskInfo) *docker.CreateContainerOptions {\n\tconfig := &docker.CreateContainerOptions{\n\t\tName: GetContainerName(taskInfo.TaskId),\n\t\tConfig: &docker.Config{\n\t\t\tEnv: EnvForTask(taskInfo),\n\t\t\tExposedPorts: PortsForTask(taskInfo),\n\t\t\tImage: *taskInfo.Container.Docker.Image,\n\t\t\tLabels: LabelsForTask(taskInfo),\n\t\t},\n\t\tHostConfig: &docker.HostConfig{\n\t\t\tBinds: BindsForTask(taskInfo),\n\t\t\tPortBindings: PortBindingsForTask(taskInfo),\n\t\t\tNetworkMode: NetworkForTask(taskInfo),\n\t\t\tCapAdd: CapAddForTask(taskInfo),\n\t\t\tCapDrop: CapDropForTask(taskInfo),\n\t\t},\n\t}\n\n\t// Check for and calculate CPU shares\n\tcpus := getResource(\"cpus\", taskInfo)\n\tif cpus != nil {\n\t\tconfig.Config.CPUShares = int64(*cpus.Scalar.Value * float64(1024))\n\t}\n\n\t// Check for and calculate memory limit\n\tmemory := getResource(\"memoryMb\", taskInfo)\n\tif memory != nil {\n\t\tconfig.Config.Memory = int64(*memory.Scalar.Value * float64(1024*1024))\n\t}\n\n\treturn config\n}", "func Schema() *spec.Schema {\n\treturn internal.Schema\n}", "func (o ParserConfigResponseOutput) Schema() SchemaPackageResponseOutput {\n\treturn o.ApplyT(func(v ParserConfigResponse) SchemaPackageResponse { return v.Schema }).(SchemaPackageResponseOutput)\n}", "func (*MigrationConfigSet) Descriptor() ([]byte, []int) {\n\treturn file_trillian_migrillian_configpb_config_proto_rawDescGZIP(), []int{1}\n}", "func (s *ShowTableStatus) Schema() sql.Schema { return showTableStatusSchema }", "func (p *Psutil) GetConfigPolicy() (*cpolicy.ConfigPolicy, error) {\n\tc := cpolicy.New()\n\tconfig := cpolicy.NewPolicyNode()\n\tr, err := cpolicy.NewStringRule(\"mount_points\", false)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr.Description = \"Mountpoints\"\n\tconfig.Add(r)\n\tc.Add([]string{\"intel\", \"psutil\", \"disk\"}, config)\n\treturn c, nil\n}", "func (p *dfCollector) GetConfigPolicy() (*cpolicy.ConfigPolicy, error) {\n\tcp := cpolicy.New()\n\trule, _ := cpolicy.NewStringRule(\"proc_path\", false, \"/proc\")\n\tnode := cpolicy.NewPolicyNode()\n\tnode.Add(rule)\n\tcp.Add([]string{nsVendor, nsClass, PluginName}, node)\n\treturn cp, nil\n}", "func (e *BaseExecutor) GetSchema() *expression.Schema {\n\treturn e.schema\n}", "func ResourceSchema() string {\n\treturn resourceSchema\n}", "func (m *Module) SetSchemaConfig(evSchemas config.EventingSchemas) error {\n\tm.lock.Lock()\n\tdefer m.lock.Unlock()\n\n\t// Reset the existing schema\n\tm.schemas = map[string]model.Fields{}\n\n\tfor _, evSchema := range evSchemas {\n\t\tresourceID := ksuid.New().String()\n\t\tdummyDBSchema := config.DatabaseSchemas{\n\t\t\tresourceID: {\n\t\t\t\tTable: evSchema.ID,\n\t\t\t\tDbAlias: \"dummyDBName\",\n\t\t\t\tSchema: evSchema.Schema,\n\t\t\t},\n\t\t}\n\t\tschemaType, err := schemaHelpers.Parser(dummyDBSchema)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(schemaType[\"dummyDBName\"][evSchema.ID]) != 0 {\n\t\t\tm.schemas[evSchema.ID] = schemaType[\"dummyDBName\"][evSchema.ID]\n\t\t}\n\t}\n\treturn nil\n}", "func (*MySqlSyncConfig) Descriptor() ([]byte, []int) {\n\treturn file_google_cloud_sql_v1_cloud_sql_resources_proto_rawDescGZIP(), []int{8}\n}", "func (*MySqlSyncConfig) Descriptor() ([]byte, []int) {\n\treturn file_google_cloud_sql_v1beta4_cloud_sql_resources_proto_rawDescGZIP(), []int{29}\n}", "func (o CassandraTableResourceOutput) Schema() CassandraSchemaPtrOutput {\n\treturn o.ApplyT(func(v CassandraTableResource) *CassandraSchema { return v.Schema }).(CassandraSchemaPtrOutput)\n}", "func (o SchemaConfigOutput) SchemaType() SchemaConfigSchemaTypePtrOutput {\n\treturn o.ApplyT(func(v SchemaConfig) *SchemaConfigSchemaType { return v.SchemaType }).(SchemaConfigSchemaTypePtrOutput)\n}", "func (o GoogleCloudHealthcareV1beta1FhirBigQueryDestinationPtrOutput) SchemaConfig() SchemaConfigPtrOutput {\n\treturn o.ApplyT(func(v *GoogleCloudHealthcareV1beta1FhirBigQueryDestination) *SchemaConfig {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.SchemaConfig\n\t}).(SchemaConfigPtrOutput)\n}", "func (o GoogleCloudHealthcareV1beta1FhirBigQueryDestinationResponseOutput) SchemaConfig() SchemaConfigResponseOutput {\n\treturn o.ApplyT(func(v GoogleCloudHealthcareV1beta1FhirBigQueryDestinationResponse) SchemaConfigResponse {\n\t\treturn v.SchemaConfig\n\t}).(SchemaConfigResponseOutput)\n}", "func (*SqlActiveDirectoryConfig) Descriptor() ([]byte, []int) {\n\treturn file_google_cloud_sql_v1_cloud_sql_resources_proto_rawDescGZIP(), []int{30}\n}", "func DBConfig(t *testing.T, d drivers.Driver) *config.Config {\n\tc := Config(t)\n\tif d == \"\" {\n\t\td = c.DB.Driver\n\t}\n\treturn configDB(t, c, d)\n}", "func TableSchema(conn *setup.Connection, tableName string) (*Schema, *Paths) {\n\tdb := conn.Connect()\n\tschema := Schema{Fields: []Field{}}\n\tpaths := Paths{[]string{}}\n\n\tdefer db.Close()\n\n\tquery := fmt.Sprintf(\"select COLUMN_NAME as `Field`, COLUMN_TYPE as `Type`, IS_NULLABLE AS `Null` from information_schema.COLUMNS WHERE TABLE_NAME = '%s'\", tableName)\n\n\trows, err := db.Query(query)\n\tif err != nil {\n\t\tlog.WithField(\"error\", err).Fatal(\"there was an error extracting the schema for:\" + tableName)\n\t}\n\tfor rows.Next() {\n\t\tvar f Field\n\t\trows.Scan(&f.Name, &f.Type, &f.Null)\n\t\tpaths.JsonPaths = append(paths.JsonPaths, fmt.Sprintf(\"$['%s']\", f.Name))\n\t\tschema.Fields = append(schema.Fields, f)\n\t}\n\n\tschema.ColCount = len(schema.Fields)\n\n\treturn &schema, &paths\n}", "func DriverConfigs() map[string]interface{} {\n\treturn ds.config.Driver\n}", "func (v *pgStatStatementsViewType) Schema() string {\n\treturn v.s.SQLSchema\n}", "func (t *TikvHandlerTool) Schema() (infoschema.InfoSchema, error) {\n\tdom, err := session.GetDomain(t.Store)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn dom.InfoSchema(), nil\n}", "func (*SqlActiveDirectoryConfig) Descriptor() ([]byte, []int) {\n\treturn file_google_cloud_sql_v1beta4_cloud_sql_resources_proto_rawDescGZIP(), []int{64}\n}", "func (v *productTableType) Schema() string {\n\treturn v.s.SQLSchema\n}", "func (v *pgStatDatabaseViewType) Schema() string {\n\treturn v.s.SQLSchema\n}", "func (e *TableReaderExecutor) Schema() *expression.Schema {\n\treturn e.schema\n}", "func (tp *TestParams) DBConfig() *DBConf {\n\treturn &(tp.DBConf)\n}", "func (d *DescribeQuery) Schema() sql.Schema {\n\treturn DescribeSchema\n}", "func (c *DefaultConstr) GetSchema(conn *sql.DB) (string, error) {\n\treturn \"\", ErrorNotSupport\n}", "func GetSchema(eventID string) (schema *Schema, err error) {\n\tschema = &Schema{}\n\n\terr = stmtSchemaByEventID.QueryRow(eventID).Scan(&schema.EventID, &schema.Alias, &schema.Desc)\n\tif err != nil {\n\t\tlog.Println(\"coudn't get schema\")\n\t\treturn\n\t}\n\trows, err := stmtParamsByEventID.Query(eventID)\n\tif err != nil {\n\t\tlog.Println(\"coudn't get schema\")\n\t\treturn\n\t}\n\n\tfor rows.Next() {\n\t\tparams := &Parameters{}\n\t\terr = rows.Scan(&params.Name, &params.DataType, &params.Mandatory, &params.Description, &params.Format)\n\t\tif err != nil {\n\t\t\tlog.Println(\"coudn't get schema\")\n\t\t\treturn\n\t\t}\n\t\tschema.Params = append(schema.Params, params)\n\t}\n\n\treturn\n}", "func GenerateSchema(_ context.Context, opts GenerateSchemaOptions) (*GenerateSchemaResult, error) {\n\tif opts.ProviderInfo.Name == \"\" {\n\t\treturn nil, fmt.Errorf(\"opts.ProviderInfo.Name cannot be empty\")\n\t}\n\tsink := opts.DiagnosticsSink\n\tif sink == nil {\n\t\tsink = diag.DefaultSink(os.Stdout, os.Stderr, diag.FormatOptions{\n\t\t\tColor: colors.Never,\n\t\t})\n\t}\n\n\tgenerated, err := realtfgen.GenerateSchemaWithOptions(realtfgen.GenerateSchemaOptions{\n\t\tProviderInfo: opts.ProviderInfo,\n\t\tDiagnosticsSink: sink,\n\t})\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tschema, err := json.Marshal(generated.PackageSpec)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &GenerateSchemaResult{\n\t\tProviderMetadata: tfbridge.ProviderMetadata{\n\t\t\tPackageSchema: schema,\n\t\t},\n\t}, nil\n}", "func (v *recordingTable) Schema() string {\n\treturn v.s.SQLSchema\n}", "func (v *libraryTableType) Schema() string {\n\treturn v.s.SQLSchema\n}", "func (db *sqlite3) GetSchema(pBt *Btree) (p *Schema) {\n\tif pBt != nil {\n\t\tp = pBt.Schema(true, ClearSchema)\n\t} else {\n\t\tp = &Schema{}\n\t}\n\tif p == nil {\n\t\tdb.mallocFailed = true\n\t} else if p.file_format == 0 {\n\t\tp.Tables = make(map[string]*Table)\n\t\tp.Indices = make(map[string]*Index)\n\t\tp.Triggers = make(map[string]*Trigger)\n\t\tp.ForeignKeys = make(map[string]*ForeignKey)\n\t\tp.enc = SQLITE_UTF8\n\t}\n\treturn\n}", "func (s *ShowCreateDatabase) Schema() sql.Schema {\n\treturn showCreateDatabaseSchema\n}", "func (p *Person) GetSchema() string {\n\treturn `CREATE TABLE person (\n\t\t\t\t\t\t\t\t\t\tfirst_name text,\n\t\t\t\t\t\t\t\t\t\tlast_name text,\n\t\t\t\t\t\t\t\t\t\tsex bool\n\t\t\t\t\t\t\t);`\n}", "func (cft *ConfigForTask) GetDesignedConfig() string {\n\tconfig := \"\"\n\tfor _, dc := range cft.DesignedConfig {\n\t\tconfig += fmt.Sprintf(\"%s\\n\", dc.Command)\n\t}\n\treturn config\n}", "func DefaultSchemaConfig(kind string) config.SchemaConfig {\n\treturn SchemaConfig(kind, \"v9\", model.Now().Add(-time.Hour*2))\n}", "func GetEventingSchema(project, commandName string, params map[string]string) ([]*model.SpecObject, error) {\n\turl := fmt.Sprintf(\"/v1/config/projects/%s/eventing/schema\", project)\n\n\t// Get the spec from the server\n\tpayload := new(model.Response)\n\tif err := utils.Get(http.MethodGet, url, params, payload); err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar objs []*model.SpecObject\n\tfor _, item := range payload.Result {\n\t\tspec := item.(map[string]interface{})\n\t\tmeta := map[string]string{\"project\": project, \"id\": spec[\"id\"].(string)}\n\n\t\t// Delete the unwanted keys from spec\n\t\tdelete(spec, \"id\")\n\n\t\t// Generating the object\n\t\ts, err := utils.CreateSpecObject(\"/v1/config/projects/{project}/eventing/schema/{id}\", commandName, meta, spec)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tobjs = append(objs, s)\n\t}\n\treturn objs, nil\n}", "func (e *BaseExecutor) Schema() *expression.Schema {\n\tif e.schema == nil {\n\t\treturn expression.NewSchema()\n\t}\n\treturn e.schema\n}", "func (v *coachTableType) Schema() string {\n\treturn v.s.SQLSchema\n}", "func (p Postgres) Schema() string {\n\tif p.schema != \"\" {\n\t\treturn p.schema\n\t}\n\treturn getDefaultValue(p, \"schema\")\n}", "func (i *blockIter) Schema() sql.Schema {\n\treturn i.sch\n}", "func (f *nfsCollector) GetConfigPolicy() (*cpolicy.ConfigPolicy, error) {\n\tc := cpolicy.New()\n\trule, _ := cpolicy.NewStringRule(\"command\", true)\n\tp := cpolicy.NewPolicyNode()\n\tp.Add(rule)\n\tc.Add([]string{\"intel\", \"dummy\", \"exec\"}, p)\n\treturn c, nil\n}", "func (*S4ReportingPluginConfig) Descriptor() ([]byte, []int) {\n\treturn file_core_services_ocr2_plugins_functions_config_config_types_proto_rawDescGZIP(), []int{1}\n}", "func (u *Use) GetConfigPolicy() (plugin.ConfigPolicy, error) {\n\tpolicy := plugin.NewConfigPolicy()\n\tpolicy.AddNewStringRule([]string{\"intel\", \"use\"}, \"proc_path\", false, plugin.SetDefaultString(\"/proc_host\"))\n\treturn *policy, nil\n}", "func GetSchema() string {\n\treturn fmt.Sprintf(`\n\tschema {\n\t\tquery: Query\n\t\tmutation: Mutation\n\t}\n\t%s\n\t%s\n\t%s\n\t%s\n`, typeDefs, inputs, queries, mutations)\n}", "func (tr *TaskDefinition) GetTerraformSchemaVersion() int {\n\treturn 1\n}", "func hookConfigurationSchema() *schema.Schema {\n\treturn &schema.Schema{\n\t\tType: schema.TypeList,\n\t\tOptional: true,\n\t\tMaxItems: 1,\n\t\tElem: &schema.Resource{\n\t\t\tSchema: map[string]*schema.Schema{\n\t\t\t\t\"invocation_condition\": func() *schema.Schema {\n\t\t\t\t\tschema := documentAttributeConditionSchema()\n\t\t\t\t\treturn schema\n\t\t\t\t}(),\n\t\t\t\t\"lambda_arn\": {\n\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\tRequired: true,\n\t\t\t\t\tValidateFunc: verify.ValidARN,\n\t\t\t\t},\n\t\t\t\t\"s3_bucket\": {\n\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\tRequired: true,\n\t\t\t\t\tValidateFunc: validation.All(\n\t\t\t\t\t\tvalidation.StringLenBetween(3, 63),\n\t\t\t\t\t\tvalidation.StringMatch(\n\t\t\t\t\t\t\tregexp.MustCompile(`[a-z0-9][\\.\\-a-z0-9]{1,61}[a-z0-9]`),\n\t\t\t\t\t\t\t\"Must be a valid bucket name\",\n\t\t\t\t\t\t),\n\t\t\t\t\t),\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func (c *CvpClient) GetConfigForTask(workOrderForID string) (ConfigForTask, error) {\n\tcft := ConfigForTask{}\n\turl := \"/provisioning/getconfigfortask.do?workorderid=\" + workOrderForID\n\tresp, err := c.Get(url)\n\tif err != nil {\n\t\treturn cft, err\n\t}\n\terr = json.Unmarshal(resp, &cft)\n\treturn cft, err\n}", "func (db *DatabaseModel) Schema() *ovsdb.DatabaseSchema {\n\tdb.mutex.RLock()\n\tdefer db.mutex.RUnlock()\n\treturn db.schema\n}", "func (*MigrationConfig) Descriptor() ([]byte, []int) {\n\treturn file_trillian_migrillian_configpb_config_proto_rawDescGZIP(), []int{0}\n}", "func (stmt *Statement) Schema() string {\n\tif stmt.ObjectQualifier != \"\" {\n\t\treturn stmt.ObjectQualifier\n\t}\n\treturn stmt.DefaultDatabase\n}", "func (*ExternalSchedulerConfig) Descriptor() ([]byte, []int) {\n\treturn file_go_chromium_org_luci_swarming_proto_config_pools_proto_rawDescGZIP(), []int{7}\n}", "func (o ParserConfigResponsePtrOutput) Schema() SchemaPackageResponsePtrOutput {\n\treturn o.ApplyT(func(v *ParserConfigResponse) *SchemaPackageResponse {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Schema\n\t}).(SchemaPackageResponsePtrOutput)\n}", "func (*SqlServerAuditConfig) Descriptor() ([]byte, []int) {\n\treturn file_google_cloud_sql_v1_cloud_sql_resources_proto_rawDescGZIP(), []int{31}\n}", "func (*TaskTemplateDeployment) Descriptor() ([]byte, []int) {\n\treturn file_go_chromium_org_luci_swarming_proto_config_pools_proto_rawDescGZIP(), []int{5}\n}", "func (o CassandraTableResourcePtrOutput) Schema() CassandraSchemaPtrOutput {\n\treturn o.ApplyT(func(v *CassandraTableResource) *CassandraSchema {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Schema\n\t}).(CassandraSchemaPtrOutput)\n}", "func (event *Event) TableSchema() (string, error) {\n\tsb := strings.Builder{}\n\tsb.WriteString(\"column_name, column_type, description\\n\")\n\tsb.WriteString(\"event_date, date, is the event date\\n\")\n\tsb.WriteString(\"event_time, timestamp, is the event time\\n\")\n\tfor _, prop := range event.Fields {\n\t\tline, err := prop.TableSchema([]string{prop.Name})\n\t\tif err != nil {\n\t\t\treturn \"\", trace.Wrap(err)\n\t\t}\n\t\tsb.WriteString(line)\n\t}\n\treturn sb.String(), nil\n}", "func (e *UnionScanExec) Schema() expression.Schema {\n\treturn e.schema\n}", "func (ec *executionContext) ___Schema(ctx context.Context, sel ast.SelectionSet, obj *introspection.Schema) graphql.Marshaler {\n\tfields := graphql.CollectFields(ctx, sel, __SchemaImplementors)\n\n\tout := graphql.NewOrderedMap(len(fields))\n\tfor i, field := range fields {\n\t\tout.Keys[i] = field.Alias\n\n\t\tswitch field.Name {\n\t\tcase \"__typename\":\n\t\t\tout.Values[i] = graphql.MarshalString(\"__Schema\")\n\t\tcase \"types\":\n\t\t\tout.Values[i] = ec.___Schema_types(ctx, field, obj)\n\t\tcase \"queryType\":\n\t\t\tout.Values[i] = ec.___Schema_queryType(ctx, field, obj)\n\t\tcase \"mutationType\":\n\t\t\tout.Values[i] = ec.___Schema_mutationType(ctx, field, obj)\n\t\tcase \"subscriptionType\":\n\t\t\tout.Values[i] = ec.___Schema_subscriptionType(ctx, field, obj)\n\t\tcase \"directives\":\n\t\t\tout.Values[i] = ec.___Schema_directives(ctx, field, obj)\n\t\tdefault:\n\t\t\tpanic(\"unknown field \" + strconv.Quote(field.Name))\n\t\t}\n\t}\n\n\treturn out\n}", "func (b *SynchronizationTemplateRequestBuilder) Schema() *SynchronizationSchemaRequestBuilder {\n\tbb := &SynchronizationSchemaRequestBuilder{BaseRequestBuilder: b.BaseRequestBuilder}\n\tbb.baseURL += \"/schema\"\n\treturn bb\n}", "func GetTaskConfig(name, action string, conf *config.ComposeConfig) (types.TaskConfig, error) {\n\tact, err := getAction(action, name, conf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn types.NewTaskConfig(act.name, conf, act.deps, NewTask(act.Run, act.Stop)), nil\n}", "func (*TaskTemplate) Descriptor() ([]byte, []int) {\n\treturn file_go_chromium_org_luci_swarming_proto_config_pools_proto_rawDescGZIP(), []int{4}\n}", "func NewTaskBolt(conf config.Config) (*TaskBolt, error) {\n\tdb, err := bolt.Open(conf.DBPath, 0600, &bolt.Options{\n\t\tTimeout: time.Second * 5,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t//Check to make sure all the required buckets have been created\n\tdb.Update(func(tx *bolt.Tx) error {\n\t\tif tx.Bucket(TaskBucket) == nil {\n\t\t\ttx.CreateBucket(TaskBucket)\n\t\t}\n\t\tif tx.Bucket(TaskAuthBucket) == nil {\n\t\t\ttx.CreateBucket(TaskAuthBucket)\n\t\t}\n\t\tif tx.Bucket(JobsQueued) == nil {\n\t\t\ttx.CreateBucket(JobsQueued)\n\t\t}\n\t\tif tx.Bucket(JobState) == nil {\n\t\t\ttx.CreateBucket(JobState)\n\t\t}\n\t\tif tx.Bucket(JobsLog) == nil {\n\t\t\ttx.CreateBucket(JobsLog)\n\t\t}\n\t\tif tx.Bucket(Workers) == nil {\n\t\t\ttx.CreateBucket(Workers)\n\t\t}\n\t\tif tx.Bucket(JobWorker) == nil {\n\t\t\ttx.CreateBucket(JobWorker)\n\t\t}\n\t\tif tx.Bucket(WorkerJobs) == nil {\n\t\t\ttx.CreateBucket(WorkerJobs)\n\t\t}\n\t\treturn nil\n\t})\n\treturn &TaskBolt{db: db, conf: conf}, nil\n}", "func (*ProtocolTxConfiguration) Descriptor() ([]byte, []int) {\n\treturn file_router_proto_rawDescGZIP(), []int{4}\n}", "func (obj *transaction) Schema() schemas.Schema {\n\treturn obj.schema\n}", "func (s *BoltState) GetDBConfig() (*DBConfig, error) {\n\tif !s.valid {\n\t\treturn nil, define.ErrDBClosed\n\t}\n\n\tcfg := new(DBConfig)\n\n\tdb, err := s.getDBCon()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer s.deferredCloseDBCon(db)\n\n\terr = db.View(func(tx *bolt.Tx) error {\n\t\tconfigBucket, err := getRuntimeConfigBucket(tx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Some of these may be nil\n\t\t// When we convert to string, Go will coerce them to \"\"\n\t\t// That's probably fine - we could raise an error if the key is\n\t\t// missing, but just not including it is also OK.\n\t\tlibpodRoot := configBucket.Get(staticDirKey)\n\t\tlibpodTmp := configBucket.Get(tmpDirKey)\n\t\tstorageRoot := configBucket.Get(graphRootKey)\n\t\tstorageTmp := configBucket.Get(runRootKey)\n\t\tgraphDriver := configBucket.Get(graphDriverKey)\n\t\tvolumePath := configBucket.Get(volPathKey)\n\n\t\tcfg.LibpodRoot = string(libpodRoot)\n\t\tcfg.LibpodTmp = string(libpodTmp)\n\t\tcfg.StorageRoot = string(storageRoot)\n\t\tcfg.StorageTmp = string(storageTmp)\n\t\tcfg.GraphDriver = string(graphDriver)\n\t\tcfg.VolumePath = string(volumePath)\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn cfg, nil\n}", "func (v *pgUserViewType) Schema() string {\n\treturn v.s.SQLSchema\n}", "func (csvw *CSVWriter) GetSchema() schema.Schema {\n\treturn csvw.sch\n}", "func (t *JSONTable) Schema() sql.Schema {\n\treturn t.FlattenSchema(t.Cols)\n}", "func (*CAccountHardware_SteamControllerSetConfig_ControllerConfig) Descriptor() ([]byte, []int) {\n\treturn file_steammessages_accounthardware_steamclient_proto_rawDescGZIP(), []int{7}\n}", "func GetMutationSchema() graphql.ObjectConfig {\n\tmutationFields := graphql.Fields{}\n\n\trootMutation := graphql.ObjectConfig{Name: \"MutationQuery\", Fields: mutationFields}\n\n\treturn rootMutation\n}", "func (*SqlServerAuditConfig) Descriptor() ([]byte, []int) {\n\treturn file_google_cloud_sql_v1beta4_cloud_sql_resources_proto_rawDescGZIP(), []int{65}\n}", "func setupSchema(cli *cli.Context) error {\n\tparams, err := parseConnectParams(cli)\n\tif err != nil {\n\t\treturn handleErr(schema.NewConfigError(err.Error()))\n\t}\n\tconn, err := newConn(params)\n\tif err != nil {\n\t\treturn handleErr(err)\n\t}\n\tdefer conn.Close()\n\tif err := schema.Setup(cli, conn); err != nil {\n\t\treturn handleErr(err)\n\t}\n\treturn nil\n}", "func (*PostgreSQLDatabaseConfig) Descriptor() ([]byte, []int) {\n\treturn file_ml_metadata_proto_metadata_store_proto_rawDescGZIP(), []int{24}\n}" ]
[ "0.6586699", "0.58137196", "0.56230664", "0.53895205", "0.5364081", "0.53349406", "0.52917063", "0.52878046", "0.51269746", "0.50748026", "0.505897", "0.5052785", "0.5022496", "0.49787623", "0.49776208", "0.49759728", "0.4963513", "0.48683783", "0.4866191", "0.48051867", "0.47845778", "0.47662133", "0.47428307", "0.47266144", "0.47246894", "0.47070518", "0.46892133", "0.4651297", "0.4649197", "0.46455044", "0.46445644", "0.46438107", "0.46417186", "0.46304625", "0.46175632", "0.46133485", "0.4610687", "0.46040437", "0.46027017", "0.46018818", "0.45990658", "0.45911908", "0.45857975", "0.4561908", "0.45443594", "0.45435384", "0.4540303", "0.453427", "0.45318988", "0.4510839", "0.45060694", "0.45020452", "0.4492897", "0.44908565", "0.44876596", "0.448731", "0.4487204", "0.44817573", "0.44782123", "0.44750583", "0.4473016", "0.44671285", "0.4463673", "0.44620675", "0.44573206", "0.44562683", "0.4452419", "0.44512096", "0.4450556", "0.44409367", "0.44394958", "0.44286638", "0.4408446", "0.44067988", "0.43989676", "0.43833375", "0.43731314", "0.43689495", "0.436796", "0.4366222", "0.43611455", "0.43591735", "0.43559936", "0.4353121", "0.43433794", "0.4341221", "0.43279615", "0.43217105", "0.4313312", "0.43073764", "0.43053553", "0.4305284", "0.4300948", "0.43006393", "0.42991996", "0.4288893", "0.42888045", "0.42854106", "0.42845032", "0.42842883" ]
0.85169226
0
Capabilities define what features the driver implements.
func (d *Driver) Capabilities() (*drivers.Capabilities, error) { return capabilities, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewCapabilities(features ...string) Capabilities {\n\tc := Capabilities{}\n\tfor _, feature := range features {\n\t\tc.With(feature)\n\t}\n\treturn c\n}", "func (srv *Server) Capabilities(context.Context, *gnmipb.CapabilityRequest) (*gnmipb.CapabilityResponse, error) {\n\treturn nil, grpc.Errorf(codes.Unimplemented, \"Capabilities() is not implemented\")\n}", "func (s *GetSceneOutput) SetCapabilities(v []*string) *GetSceneOutput {\n\ts.Capabilities = v\n\treturn s\n}", "func (s *CreateSceneInput) SetCapabilities(v []*string) *CreateSceneInput {\n\ts.Capabilities = v\n\treturn s\n}", "func (s *UpdateSceneInput) SetCapabilities(v []*string) *UpdateSceneInput {\n\ts.Capabilities = v\n\treturn s\n}", "func (o *PluginConfigLinux) SetCapabilities(v []string) {\n\to.Capabilities = v\n}", "func (s *Server) capabilities(ctx context.Context, req *gnmipb.CapabilityRequest) (*gnmipb.CapabilityResponse, error) {\n\tver, err := GetGNMIVersion()\n\tif err != nil {\n\t\treturn nil, status.TaggedErrorf(codes.Internal, status.TagOperationFail,\n\t\t\t\"gnmi service version error: %v\", err)\n\t}\n\treturn &gnmipb.CapabilityResponse{\n\t\tSupportedModels: s.Modeldata,\n\t\tSupportedEncodings: supportedEncodings,\n\t\tGNMIVersion: *ver,\n\t}, nil\n}", "func (n *mockAgent) capabilities() types.Capabilities {\n\treturn types.Capabilities{}\n}", "func mapCapabilities(config configuration) cap.Capabilities {\n\tcaps := cap.Capabilities{\n\t\t\"features\": cap.Cap{\n\t\t\t\"vault\": cap.Cap{\n\t\t\t\t\"enabled\": config.Cluster.Vault.Enabled,\n\t\t\t\t\"managed\": config.Cluster.Vault.Managed.Enabled,\n\t\t\t},\n\t\t\t\"monitoring\": cap.Cap{\n\t\t\t\t\"enabled\": config.Cluster.Monitoring.Enabled,\n\t\t\t},\n\t\t\t\"logging\": cap.Cap{\n\t\t\t\t\"enabled\": config.Cluster.Logging.Enabled,\n\t\t\t},\n\t\t\t\"dns\": cap.Cap{\n\t\t\t\t\"enabled\": config.Cluster.DNS.Enabled,\n\t\t\t\t\"baseDomain\": config.Cluster.DNS.BaseDomain,\n\t\t\t},\n\t\t\t\"securityScan\": cap.Cap{\n\t\t\t\t\"enabled\": config.Cluster.SecurityScan.Enabled,\n\t\t\t\t\"managed\": config.Cluster.SecurityScan.Anchore.Enabled,\n\t\t\t},\n\t\t\t\"expiry\": cap.Cap{\n\t\t\t\t\"enabled\": config.Cluster.Expiry.Enabled,\n\t\t\t},\n\t\t\t\"ingress\": cap.Cap{\n\t\t\t\t\"enabled\": config.Cluster.Ingress.Enabled,\n\t\t\t\t\"controllers\": config.Cluster.Ingress.Controllers,\n\t\t\t},\n\t\t},\n\t\t\"helm\": cap.Cap{\n\t\t\t\"version\": helmVersion,\n\t\t},\n\t}\n\n\treturn caps\n}", "func allCapabilities() []string {\n\tvar capabilities []string\n\tfor _, cap := range capabilityList {\n\t\tcapabilities = append(capabilities, cap)\n\t}\n\treturn capabilities\n}", "func NewCapabilities() *Capabilities {\n\treturn &Capabilities{\n\t\tm: make(map[string]*Capability, 0),\n\t}\n}", "func (o *MicrosoftGraphVerifiedDomain) SetCapabilities(v string) {\n\to.Capabilities = &v\n}", "func (s *Server) Capabilities(ctx context.Context, req *gnmipb.CapabilityRequest) (*gnmipb.CapabilityResponse, error) {\n\tseq := s.rpcSequence()\n\tif glog.V(11) {\n\t\tglog.Infof(\"capabilities.request[%d]=%s\", seq, req)\n\t}\n\tresp, err := s.capabilities(ctx, req)\n\tif err != nil {\n\t\tif glog.V(11) {\n\t\t\tglog.Errorf(\"capabilities.response[%d]=%v\", seq, status.FromError(err))\n\t\t}\n\t} else {\n\t\tif glog.V(11) {\n\t\t\tglog.Infof(\"capabilities.response[%d]=%s\", seq, resp)\n\t\t}\n\t}\n\treturn resp, err\n}", "func (d ImagefsDriver) Capabilities() *volume.CapabilitiesResponse {\n\tfmt.Printf(\"-> Capabilities\\n\")\n\tresponse := volume.CapabilitiesResponse{Capabilities: volume.Capability{Scope: \"local\"}}\n\tfmt.Printf(\"<- %+v\\n\", response)\n\treturn &response\n}", "func (b *BaseConn) writeCapabilities() error {\n\n\tcapMsg := new(baseproto.Capabilities)\n\n\tmsg := new(baseproto.Message)\n\tmsg.MsgType = new(uint32)\n\t*msg.MsgType = 1\n\n\tvar err error\n\tif msg.Content, err = proto.Marshal(capMsg); err != nil {\n\t\treturn err\n\t}\n\n\treturn b.writeMsg(msg)\n}", "func (s *LinuxParameters) SetCapabilities(v *KernelCapabilities) *LinuxParameters {\n\ts.Capabilities = v\n\treturn s\n}", "func Capabilities() *volume.CapabilitiesResponse {\n\tlog.Debugf(\"Entering Capabilities\")\n\treturn &volume.CapabilitiesResponse{\n\t\tCapabilities: volume.Capability{\n\t\t\tScope: \"global\",\n\t\t},\n\t}\n}", "func SetCapabilities(config *configv1.ClusterVersion,\n\texistingEnabled map[configv1.ClusterVersionCapability]struct{}) ClusterCapabilities {\n\n\tvar capabilities ClusterCapabilities\n\tcapabilities.KnownCapabilities = setKnownCapabilities()\n\n\tcapabilities.EnabledCapabilities, capabilities.ImplicitlyEnabledCapabilities = setEnabledCapabilities(config.Spec.Capabilities,\n\t\texistingEnabled)\n\n\treturn capabilities\n}", "func (fs *Mysqlfs) Capabilities() billy.Capability {\n\treturn billy.WriteCapability |\n\t\tbilly.ReadCapability |\n\t\tbilly.ReadAndWriteCapability |\n\t\tbilly.SeekCapability |\n\t\tbilly.TruncateCapability\n}", "func (l *Libvirt) Capabilities() ([]byte, error) {\n\tcaps, err := l.ConnectGetCapabilities()\n\treturn []byte(caps), err\n}", "func (o *ApplianceSetupInfoAllOf) SetCapabilities(v []ApplianceKeyValuePair) {\n\to.Capabilities = v\n}", "func (d *MinioDriver) Capabilities(r volume.Request) volume.Response {\n\tlocalCapability := volume.Capability{\n\t\tScope: \"local\",\n\t}\n\tglog.V(1).Infof(\"Capabilities request: %#v\", r)\n\treturn volumeResp(\"\", \"\", nil, localCapability, \"\")\n}", "func (d *EmulatedBTPeerDevice) evaluateCapabilities(ctx context.Context) error {\n\t// Refresh stored capabilities.\n\tcapabilities, err := d.rpc.GetCapabilities(ctx)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to get device capabilities\")\n\t}\n\td.cache.capabilities = capabilities\n\n\t// Evaluate hasPinCode.\n\thasPinCode, ok := d.cache.capabilities[cbt.DeviceCapabilityHasPin.String()]\n\tif !ok {\n\t\treturn errors.Errorf(\"device capabilities missing %q\", cbt.DeviceCapabilityHasPin)\n\t}\n\thasPinCodeBool, ok := hasPinCode.(bool)\n\tif !ok {\n\t\treturn errors.Errorf(\"expected device capability %q to be a bool, got %v\", cbt.DeviceCapabilityHasPin, hasPinCode)\n\t}\n\td.cache.hasPinCode = hasPinCodeBool\n\n\t// Evaluate supportsInitConnect.\n\tsupportsInitConnect, ok := d.cache.capabilities[cbt.DeviceCapabilityInitConnect.String()]\n\tif !ok {\n\t\treturn errors.Errorf(\"device capabilities missing %q\", cbt.DeviceCapabilityInitConnect)\n\t}\n\tsupportsInitConnectBool, ok := supportsInitConnect.(bool)\n\tif !ok {\n\t\treturn errors.Errorf(\"expected device capability %q to be a bool, got %v\", cbt.DeviceCapabilityInitConnect, supportsInitConnect)\n\t}\n\td.cache.supportsInitConnect = supportsInitConnectBool\n\n\t// Evaluate supportedTransportMethods.\n\tsupportedTransportMethods, ok := d.cache.capabilities[cbt.DeviceCapabilityTransports.String()]\n\tif !ok {\n\t\treturn errors.Errorf(\"device capabilities missing %q\", cbt.DeviceCapabilityTransports)\n\t}\n\tsupportedTransportMethodsSlice, ok := supportedTransportMethods.([]interface{})\n\tif !ok {\n\t\treturn errors.Errorf(\"expected device capability %q to be an []interface{}, got %v\", cbt.DeviceCapabilityTransports, supportedTransportMethods)\n\t}\n\tfor _, method := range supportedTransportMethodsSlice {\n\t\tmethodStr, ok := method.(string)\n\t\tif !ok {\n\t\t\treturn errors.Errorf(\"expected device capability %q to be an []interface{} of strings, got %v\", cbt.DeviceCapabilityTransports, supportedTransportMethods)\n\t\t}\n\t\td.cache.supportedTransportMethods = append(d.cache.supportedTransportMethods, cbt.TransportMethod(methodStr))\n\t}\n\n\treturn nil\n}", "func GetCapabilities() Capabilities {\n\treturn Client().GetCapabilities()\n}", "func WithCapabilities(capabilities consumer.Capabilities) Option {\n\treturn func(o *baseConsumer) {\n\t\to.capabilities = capabilities\n\t}\n}", "func (d *VolumeDriver) Capabilities(r volume.Request) volume.Response {\n\treturn volume.Response{Capabilities: volume.Capability{Scope: \"global\"}}\n}", "func (conn *Conn) SupportsCapability(cap string) bool {\n\treturn conn.supportedCaps.Has(cap)\n}", "func (s *SCEP) GetCapabilities() []string {\n\treturn s.Capabilities\n}", "func (s *SCEP) GetCapabilities() []string {\n\treturn s.Capabilities\n}", "func (o *PluginConfigLinux) GetCapabilities() []string {\n\tif o == nil {\n\t\tvar ret []string\n\t\treturn ret\n\t}\n\n\treturn o.Capabilities\n}", "func (mysqlFlavor80) supportsCapability(serverVersion string, capability FlavorCapability) (bool, error) {\n\tswitch capability {\n\tcase InstantDDLFlavorCapability,\n\t\tInstantExpandEnumCapability,\n\t\tInstantAddLastColumnFlavorCapability,\n\t\tInstantAddDropVirtualColumnFlavorCapability,\n\t\tInstantChangeColumnDefaultFlavorCapability:\n\t\treturn true, nil\n\tcase InstantAddDropColumnFlavorCapability:\n\t\treturn ServerVersionAtLeast(serverVersion, 8, 0, 29)\n\tcase TransactionalGtidExecutedFlavorCapability:\n\t\treturn ServerVersionAtLeast(serverVersion, 8, 0, 17)\n\tcase FastDropTableFlavorCapability:\n\t\treturn ServerVersionAtLeast(serverVersion, 8, 0, 23)\n\tcase MySQLJSONFlavorCapability:\n\t\treturn true, nil\n\tcase MySQLUpgradeInServerFlavorCapability:\n\t\treturn ServerVersionAtLeast(serverVersion, 8, 0, 16)\n\tcase DynamicRedoLogCapacityFlavorCapability:\n\t\treturn ServerVersionAtLeast(serverVersion, 8, 0, 30)\n\tcase DisableRedoLogFlavorCapability:\n\t\treturn ServerVersionAtLeast(serverVersion, 8, 0, 21)\n\tdefault:\n\t\treturn false, nil\n\t}\n}", "func (d *GitDir) Capabilities() (*common.Capabilities, error) {\n\tc := common.NewCapabilities()\n\n\terr := d.addSymRefCapability(c)\n\n\treturn c, err\n}", "func (s *Source) GetCapabilities() map[string]*utils.TaskDescription {\n\tavailableAction := make(map[string]*utils.TaskDescription)\n\treturn availableAction\n}", "func NewCapabilitiesCaveat(serviceName string, capabilities string) Caveat {\n\treturn Caveat{\n\t\tCondition: serviceName + CondCapabilitiesSuffix,\n\t\tValue: capabilities,\n\t}\n}", "func (provider *NetconfServiceProvider) GetCapabilities() []string {\n\tsession := path.ServiceProviderGetSession(provider.Private)\n\treturn path.GetSessionCapabilities(session)\n}", "func (cc *ChannelConfig) Capabilities() ChannelCapabilities {\n\t_ = cc.protos\n\t_ = cc.protos.Capabilities\n\t_ = cc.protos.Capabilities.Capabilities\n\treturn capabilities.NewChannelProvider(cc.protos.Capabilities.Capabilities)\n}", "func (mysqlFlavor56) supportsCapability(serverVersion string, capability FlavorCapability) (bool, error) {\n\tswitch capability {\n\tdefault:\n\t\treturn false, nil\n\t}\n}", "func (a *Client) ListCapabilities(params *ListCapabilitiesParams, authInfo runtime.ClientAuthInfoWriter) (*ListCapabilitiesOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListCapabilitiesParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"listCapabilities\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/capabilities\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &ListCapabilitiesReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*ListCapabilitiesOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*ListCapabilitiesDefault)\n\treturn nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func (_class PIFClass) GetCapabilities(sessionID SessionRef, self PIFRef) (_retval []string, _err error) {\n\t_method := \"PIF.get_capabilities\"\n\t_sessionIDArg, _err := convertSessionRefToXen(fmt.Sprintf(\"%s(%s)\", _method, \"session_id\"), sessionID)\n\tif _err != nil {\n\t\treturn\n\t}\n\t_selfArg, _err := convertPIFRefToXen(fmt.Sprintf(\"%s(%s)\", _method, \"self\"), self)\n\tif _err != nil {\n\t\treturn\n\t}\n\t_result, _err := _class.client.APICall(_method, _sessionIDArg, _selfArg)\n\tif _err != nil {\n\t\treturn\n\t}\n\t_retval, _err = convertStringSetToGo(_method + \" -> \", _result.Value)\n\treturn\n}", "func GetCapabilities(cap string) []*string {\n\tx := []*string{}\n\tif strings.Compare(cap, \"\") == 0 {\n\t\treturn nil\n\t}\n\tfor _, c := range strings.Split(cap, \",\") {\n\t\tx = append(x, &c)\n\t}\n\treturn x\n}", "func (d *DirDriver) Capabilities() *volume.CapabilitiesResponse {\n\tlogrus.Infof(\"Hit Capabilities() endpoint\")\n\n\treturn &volume.CapabilitiesResponse{\n\t\tCapabilities: volume.Capability{\n\t\t\tScope: \"local\",\n\t\t},\n\t}\n}", "func detectCapabilities(d diag.Sink, client *client.Client) func(ctx context.Context) capabilities {\n\tvar once sync.Once\n\tvar caps capabilities\n\tdone := make(chan struct{})\n\tget := func(ctx context.Context) capabilities {\n\t\tonce.Do(func() {\n\t\t\tcaps = doDetectCapabilities(ctx, d, client)\n\t\t\tclose(done)\n\t\t})\n\t\t<-done\n\t\treturn caps\n\t}\n\treturn get\n}", "func (o *ApplianceSetupInfoAllOf) GetCapabilities() []ApplianceKeyValuePair {\n\tif o == nil {\n\t\tvar ret []ApplianceKeyValuePair\n\t\treturn ret\n\t}\n\treturn o.Capabilities\n}", "func (c *Config) SetCapabilities(ctype string, caps []string) {\n\tswitch ctype {\n\tcase capabilities.Permitted:\n\t\tc.config.capPermitted = 0\n\t\tfor _, v := range caps {\n\t\t\tc.config.capPermitted |= C.ulonglong(1 << capabilities.Map[v].Value)\n\t\t}\n\tcase capabilities.Effective:\n\t\tc.config.capEffective = 0\n\t\tfor _, v := range caps {\n\t\t\tc.config.capEffective |= C.ulonglong(1 << capabilities.Map[v].Value)\n\t\t}\n\tcase capabilities.Inheritable:\n\t\tc.config.capInheritable = 0\n\t\tfor _, v := range caps {\n\t\t\tc.config.capInheritable |= C.ulonglong(1 << capabilities.Map[v].Value)\n\t\t}\n\tcase capabilities.Bounding:\n\t\tc.config.capBounding = 0\n\t\tfor _, v := range caps {\n\t\t\tc.config.capBounding |= C.ulonglong(1 << capabilities.Map[v].Value)\n\t\t}\n\tcase capabilities.Ambient:\n\t\tc.config.capAmbient = 0\n\t\tfor _, v := range caps {\n\t\t\tc.config.capAmbient |= C.ulonglong(1 << capabilities.Map[v].Value)\n\t\t}\n\t}\n}", "func (mysqlFlavor57) supportsCapability(serverVersion string, capability FlavorCapability) (bool, error) {\n\tswitch capability {\n\tcase MySQLJSONFlavorCapability:\n\t\treturn true, nil\n\tdefault:\n\t\treturn false, nil\n\t}\n}", "func (bs baseConsumer) Capabilities() consumer.Capabilities {\n\treturn bs.capabilities\n}", "func (s *EtcdStorage) Capabilities() storage.Capability {\n\treturn storage.CapabilityDistributedLocking | storage.CapabilityWatching\n}", "func (d *EmulatedBTPeerDevice) Capabilities() map[string]interface{} {\n\treturn d.cache.capabilities\n}", "func (proxy *remoteDriverProxy) Capabilities() (*remoteVolumeCapability, error) {\n\tvar req remoteVolumeCapabilitiesReq\n\tvar resp remoteVolumeCapabilitiesResp\n\n\tif err := proxy.client.CallService(remoteVolumeCapabilitiesService, &req, &resp, true); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif resp.Err != \"\" {\n\t\treturn nil, errors.New(resp.Err)\n\t}\n\n\treturn resp.Capabilities, nil\n}", "func (mariadbFlavor) supportsCapability(serverVersion string, capability FlavorCapability) (bool, error) {\n\tswitch capability {\n\tdefault:\n\t\treturn false, nil\n\t}\n}", "func (d *Driver) GetCapabilities() (*sdk.CapabilitiesResponse, error) {\n\tscope := &sdk.CapabilitiesResponse{Scope: sdk.LocalScope}\n\treturn scope, nil\n}", "func (c *Config) SetCapabilities(ctype string, caps []string) {\n\tswitch ctype {\n\tcase capabilities.Permitted:\n\t\tc.config.capabilities.permitted = 0\n\t\tfor _, v := range caps {\n\t\t\tc.config.capabilities.permitted |= C.ulonglong(1 << capabilities.Map[v].Value)\n\t\t}\n\tcase capabilities.Effective:\n\t\tc.config.capabilities.effective = 0\n\t\tfor _, v := range caps {\n\t\t\tc.config.capabilities.effective |= C.ulonglong(1 << capabilities.Map[v].Value)\n\t\t}\n\tcase capabilities.Inheritable:\n\t\tc.config.capabilities.inheritable = 0\n\t\tfor _, v := range caps {\n\t\t\tc.config.capabilities.inheritable |= C.ulonglong(1 << capabilities.Map[v].Value)\n\t\t}\n\tcase capabilities.Bounding:\n\t\tc.config.capabilities.bounding = 0\n\t\tfor _, v := range caps {\n\t\t\tc.config.capabilities.bounding |= C.ulonglong(1 << capabilities.Map[v].Value)\n\t\t}\n\tcase capabilities.Ambient:\n\t\tc.config.capabilities.ambient = 0\n\t\tfor _, v := range caps {\n\t\t\tc.config.capabilities.ambient |= C.ulonglong(1 << capabilities.Map[v].Value)\n\t\t}\n\t}\n}", "func (o *ARVRInterface) GetCapabilities() gdnative.Int {\n\t//log.Println(\"Calling ARVRInterface.GetCapabilities()\")\n\n\t// Build out the method's arguments\n\tptrArguments := make([]gdnative.Pointer, 0, 0)\n\n\t// Get the method bind\n\tmethodBind := gdnative.NewMethodBind(\"ARVRInterface\", \"get_capabilities\")\n\n\t// Call the parent method.\n\t// int\n\tretPtr := gdnative.NewEmptyInt()\n\tgdnative.MethodBindPtrCall(methodBind, o.GetBaseObject(), ptrArguments, retPtr)\n\n\t// If we have a return type, convert it from a pointer into its actual object.\n\tret := gdnative.NewIntFromPointer(retPtr)\n\treturn ret\n}", "func setEnabledCapabilities(capabilitiesSpec *configv1.ClusterVersionCapabilitiesSpec,\n\tpriorEnabled map[configv1.ClusterVersionCapability]struct{}) (map[configv1.ClusterVersionCapability]struct{},\n\t[]configv1.ClusterVersionCapability) {\n\n\tcapSet := DefaultCapabilitySet\n\n\tif capabilitiesSpec != nil && len(capabilitiesSpec.BaselineCapabilitySet) > 0 {\n\t\tcapSet = capabilitiesSpec.BaselineCapabilitySet\n\t}\n\tenabled := GetCapabilitiesAsMap(configv1.ClusterVersionCapabilitySets[capSet])\n\n\tif capabilitiesSpec != nil {\n\t\tfor _, v := range capabilitiesSpec.AdditionalEnabledCapabilities {\n\t\t\tif _, ok := enabled[v]; ok {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tenabled[v] = struct{}{}\n\t\t}\n\t}\n\tvar implicitlyEnabled []configv1.ClusterVersionCapability\n\tfor k := range priorEnabled {\n\t\tif _, ok := enabled[k]; !ok {\n\t\t\timplicitlyEnabled = append(implicitlyEnabled, k)\n\t\t\tenabled[k] = struct{}{}\n\t\t}\n\t}\n\tsort.Sort(capabilitiesSort(implicitlyEnabled))\n\treturn enabled, implicitlyEnabled\n}", "func GetCapabilities(\n\treq *http.Request, rsAPI roomserverAPI.ClientRoomserverAPI,\n) util.JSONResponse {\n\troomVersionsQueryReq := roomserverAPI.QueryRoomVersionCapabilitiesRequest{}\n\troomVersionsQueryRes := roomserverAPI.QueryRoomVersionCapabilitiesResponse{}\n\tif err := rsAPI.QueryRoomVersionCapabilities(\n\t\treq.Context(),\n\t\t&roomVersionsQueryReq,\n\t\t&roomVersionsQueryRes,\n\t); err != nil {\n\t\tutil.GetLogger(req.Context()).WithError(err).Error(\"queryAPI.QueryRoomVersionCapabilities failed\")\n\t\treturn jsonerror.InternalServerError()\n\t}\n\n\tresponse := map[string]interface{}{\n\t\t\"capabilities\": map[string]interface{}{\n\t\t\t\"m.change_password\": map[string]bool{\n\t\t\t\t\"enabled\": true,\n\t\t\t},\n\t\t\t\"m.room_versions\": roomVersionsQueryRes,\n\t\t},\n\t}\n\n\treturn util.JSONResponse{\n\t\tCode: http.StatusOK,\n\t\tJSON: response,\n\t}\n}", "func getCapabilities(attributes bascule.Attributes) ([]string, string, error) {\n\tif attributes == nil {\n\t\treturn []string{}, UndeterminedCapabilities, ErrNilAttributes\n\t}\n\n\tval, ok := attributes.Get(CapabilityKey)\n\tif !ok {\n\t\treturn []string{}, UndeterminedCapabilities, fmt.Errorf(\"couldn't get capabilities using key %v\", CapabilityKey)\n\t}\n\n\tvals, err := cast.ToStringSliceE(val)\n\tif err != nil {\n\t\treturn []string{}, UndeterminedCapabilities, fmt.Errorf(\"capabilities \\\"%v\\\" not the expected string slice: %v\", val, err)\n\t}\n\n\tif len(vals) == 0 {\n\t\treturn []string{}, EmptyCapabilitiesList, ErrNoVals\n\t}\n\n\treturn vals, \"\", nil\n\n}", "func getCapabilities(client kubernetes.Interface) (Capabilities, error) {\n\n\tgroupList, err := client.Discovery().ServerGroups()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcaps := Capabilities{}\n\tfor _, g := range groupList.Groups {\n\t\tfor _, gv := range g.Versions {\n\t\t\tcaps[gv.GroupVersion] = true\n\t\t}\n\t}\n\n\treturn caps, nil\n}", "func (self *MyDockerNetworkPlugin) GetCapabilities() (*n.CapabilitiesResponse,\n\terror) {\n\tlog.Printf(\"Received GetCapabilities req\")\n\tcapabilities := &n.CapabilitiesResponse{\n\t\tScope: self.scope,\n\t}\n\treturn capabilities, nil\n}", "func (o *ApplianceSetupInfoAllOf) HasCapabilities() bool {\n\tif o != nil && o.Capabilities != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (c *Capabilities) Supports(capability string) bool {\n\t_, ok := c.m[capability]\n\treturn ok\n}", "func (s RainsMessage) NewCapabilities(n int32) (capnp.TextList, error) {\n\tl, err := capnp.NewTextList(s.Struct.Segment(), n)\n\tif err != nil {\n\t\treturn capnp.TextList{}, err\n\t}\n\terr = s.Struct.SetPtr(1, l.List.ToPtr())\n\treturn l, err\n}", "func (fmp *filterMetricProcessor) GetCapabilities() component.ProcessorCapabilities {\n\treturn component.ProcessorCapabilities{MutatesConsumedData: false}\n}", "func (c *Client) GetCapabilities(ctx context.Context) (res *repb.ServerCapabilities, err error) {\n\treturn c.GetCapabilitiesForInstance(ctx, c.InstanceName)\n}", "func validateCapabilities(caps []*csi.VolumeCapability) bool {\n\n\tvar supported bool\n\tfor _, cap := range caps {\n\t\tif IsSupportedVolumeCapabilityAccessMode(cap.AccessMode.Mode) {\n\t\t\tsupported = true\n\t\t} else {\n\t\t\tsupported = false\n\t\t}\n\t}\n\n\treturn supported\n}", "func (c *Client) Capability() (caps []string, err error) {\n\ttag, err := c.prepareCmd(\"CAPABILITY\")\n\tif err != nil {\n\t\treturn\n\t}\n\tdefer c.cleanCmd()\n\n\terr = c.writeString(tag + \" CAPABILITY\\r\\n\")\n\tif err != nil {\n\t\treturn\n\t}\n\n\trep := <-c.rep\n\tif rep == nil {\n\t\terr = ErrNilRep\n\t\treturn\n\t}\n\tif rep.err != nil {\n\t\terr = rep.err\n\t\treturn\n\t}\n\tfor _, line := range rep.data {\n\t\tlineCaps := strings.Split(string(line[:len(line)-2]), \" \")\n\t\tif lineCaps[0] != \"CAPABILITY\" {\n\t\t\tcontinue\n\t\t}\n\t\tcaps = append(caps, lineCaps[1:]...)\n\t}\n\treturn\n}", "func WithCapabilities(capabilities consumer.Capabilities) Option {\n\treturn func(o *baseSettings) {\n\t\to.consumerOptions = append(o.consumerOptions, consumer.WithCapabilities(capabilities))\n\t}\n}", "func (registry *ModelRegistry) Capabilities() []*gnmi.ModelData {\n\t// Make a map - if we get duplicates overwrite them\n\tmodelMap := make(map[string]*gnmi.ModelData)\n\tfor _, model := range registry.ModelPlugins {\n\t\t_, _, modelItem, _ := model.ModelData()\n\t\tfor _, mi := range modelItem {\n\t\t\tmodelName := utils.ToModelName(devicetype.Type(mi.Name), devicetype.Version(mi.Version))\n\t\t\tmodelMap[modelName] = mi\n\t\t}\n\t}\n\n\toutputList := make([]*gnmi.ModelData, len(modelMap))\n\ti := 0\n\tfor _, modelItem := range modelMap {\n\t\toutputList[i] = modelItem\n\t\ti++\n\t}\n\treturn outputList\n}", "func (p *cliModules) Features() modules.Features {\n\treturn modules.Features{\n\t\tKubernetes: true,\n\t\tDB: true,\n\t\tApp: true,\n\t\tAdvancedAccessWorkflows: true,\n\t\tAccessControls: true,\n\t}\n}", "func PrintCapability(cap int32) string {\n\tvar capabilities = map[int32]string{\n\t\t0: \"CAP_CHOWN\",\n\t\t1: \"CAP_DAC_OVERRIDE\",\n\t\t2: \"CAP_DAC_READ_SEARCH\",\n\t\t3: \"CAP_FOWNER\",\n\t\t4: \"CAP_FSETID\",\n\t\t5: \"CAP_KILL\",\n\t\t6: \"CAP_SETGID\",\n\t\t7: \"CAP_SETUID\",\n\t\t8: \"CAP_SETPCAP\",\n\t\t9: \"CAP_LINUX_IMMUTABLE\",\n\t\t10: \"CAP_NET_BIND_SERVICE\",\n\t\t11: \"CAP_NET_BROADCAST\",\n\t\t12: \"CAP_NET_ADMIN\",\n\t\t13: \"CAP_NET_RAW\",\n\t\t14: \"CAP_IPC_LOCK\",\n\t\t15: \"CAP_IPC_OWNER\",\n\t\t16: \"CAP_SYS_MODULE\",\n\t\t17: \"CAP_SYS_RAWIO\",\n\t\t18: \"CAP_SYS_CHROOT\",\n\t\t19: \"CAP_SYS_PTRACE\",\n\t\t20: \"CAP_SYS_PACCT\",\n\t\t21: \"CAP_SYS_ADMIN\",\n\t\t22: \"CAP_SYS_BOOT\",\n\t\t23: \"CAP_SYS_NICE\",\n\t\t24: \"CAP_SYS_RESOURCE\",\n\t\t25: \"CAP_SYS_TIME\",\n\t\t26: \"CAP_SYS_TTY_CONFIG\",\n\t\t27: \"CAP_MKNOD\",\n\t\t28: \"CAP_LEASE\",\n\t\t29: \"CAP_AUDIT_WRITE\",\n\t\t30: \"CAP_AUDIT_CONTROL\",\n\t\t31: \"CAP_SETFCAP\",\n\t\t32: \"CAP_MAC_OVERRIDE\",\n\t\t33: \"CAP_MAC_ADMIN\",\n\t\t34: \"CAP_SYSLOG\",\n\t\t35: \"CAP_WAKE_ALARM\",\n\t\t36: \"CAP_BLOCK_SUSPEND\",\n\t\t37: \"CAP_AUDIT_READ\",\n\t}\n\tvar res string\n\tif capName, ok := capabilities[cap]; ok {\n\t\tres = capName\n\t} else {\n\t\tres = strconv.Itoa(int(cap))\n\t}\n\treturn res\n}", "func (rmp *resourceMetricProcessor) GetCapabilities() component.ProcessorCapabilities {\n\treturn component.ProcessorCapabilities{MutatesConsumedData: true}\n}", "func (l *RemoteProvider) loadCapabilities(token string) {\n\tvar resp *http.Response\n\tvar err error\n\n\tversion := viper.GetString(\"BUILD\")\n\tos := viper.GetString(\"OS\")\n\tfinalURL := fmt.Sprintf(\"%s/%s/capabilities?os=%s\", l.RemoteProviderURL, version, os)\n\tfinalURL = strings.TrimSuffix(finalURL, \"\\n\")\n\tremoteProviderURL, err := url.Parse(finalURL)\n\tif err != nil {\n\t\tlogrus.Errorf(\"Error while constructing url: %s\", err)\n\t\treturn\n\t}\n\n\treq, _ := http.NewRequest(http.MethodGet, remoteProviderURL.String(), nil)\n\n\t// If not token is provided then make a simple GET request\n\tif token == \"\" {\n\t\tc := &http.Client{}\n\t\tresp, err = c.Do(req)\n\t} else {\n\t\t// Proceed to make a request with the token\n\t\tresp, err = l.DoRequest(req, token)\n\t}\n\n\tif err != nil || resp.StatusCode != http.StatusOK {\n\t\tlogrus.Errorf(\"[Initialize Provider]: Failed to get capabilities %s\", err)\n\t\treturn\n\t}\n\tdefer func() {\n\t\terr := resp.Body.Close()\n\t\tif err != nil {\n\t\t\tlogrus.Errorf(\"[Initialize]: Failed to close response body %s\", err)\n\t\t}\n\t}()\n\n\t// Clear the previous capabilities before writing new one\n\tl.ProviderProperties = ProviderProperties{\n\t\tProviderURL: l.RemoteProviderURL,\n\t}\n\tdecoder := json.NewDecoder(resp.Body)\n\tif err := decoder.Decode(&l.ProviderProperties); err != nil {\n\t\tlogrus.Errorf(\"[Initialize]: Failed to decode provider properties %s\", err)\n\t}\n}", "func (bq *InMemoryBuildQueue) GetCapabilities(ctx context.Context, in *remoteexecution.GetCapabilitiesRequest) (*remoteexecution.ServerCapabilities, error) {\n\tinstanceName, err := digest.NewInstanceName(in.InstanceName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\texecEnabled := true\n\tauthErr := auth.AuthorizeSingleInstanceName(ctx, bq.executeAuthorizer, instanceName)\n\tswitch status.Code(authErr) {\n\tcase codes.OK:\n\t\t// Nothing to do.\n\tcase codes.PermissionDenied:\n\t\texecEnabled = false\n\tdefault:\n\t\treturn nil, util.StatusWrap(authErr, \"Authorization\")\n\t}\n\n\treturn &remoteexecution.ServerCapabilities{\n\t\tCacheCapabilities: &remoteexecution.CacheCapabilities{\n\t\t\tDigestFunctions: digest.SupportedDigestFunctions,\n\t\t\tActionCacheUpdateCapabilities: &remoteexecution.ActionCacheUpdateCapabilities{\n\t\t\t\tUpdateEnabled: false,\n\t\t\t},\n\t\t\t// CachePriorityCapabilities: Priorities not supported.\n\t\t\t// MaxBatchTotalSize: Not used by Bazel yet.\n\t\t\tSymlinkAbsolutePathStrategy: remoteexecution.SymlinkAbsolutePathStrategy_ALLOWED,\n\t\t},\n\t\tExecutionCapabilities: &remoteexecution.ExecutionCapabilities{\n\t\t\tDigestFunction: remoteexecution.DigestFunction_SHA256,\n\t\t\tExecEnabled: execEnabled,\n\t\t\tExecutionPriorityCapabilities: &remoteexecution.PriorityCapabilities{\n\t\t\t\tPriorities: []*remoteexecution.PriorityCapabilities_PriorityRange{\n\t\t\t\t\t{MinPriority: math.MinInt32, MaxPriority: math.MaxInt32},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t// TODO: DeprecatedApiVersion.\n\t\tLowApiVersion: &semver.SemVer{Major: 2},\n\t\tHighApiVersion: &semver.SemVer{Major: 2},\n\t}, nil\n}", "func (c CapabilitiesValidator) checkCapabilities(capabilities []string, reqURL string, method string) error {\n\tfor _, val := range capabilities {\n\t\tif c.Checker.Authorized(val, reqURL, method) {\n\t\t\treturn nil\n\t\t}\n\t}\n\treturn emperror.With(ErrNoValidCapabilityFound, \"capabilitiesFound\", capabilities, \"urlToMatch\", reqURL, \"methodToMatch\", method)\n\n}", "func AdditionalCapabilitiesGenerator() gopter.Gen {\n\tif additionalCapabilitiesGenerator != nil {\n\t\treturn additionalCapabilitiesGenerator\n\t}\n\n\tgenerators := make(map[string]gopter.Gen)\n\tAddIndependentPropertyGeneratorsForAdditionalCapabilities(generators)\n\tadditionalCapabilitiesGenerator = gen.Struct(reflect.TypeOf(AdditionalCapabilities{}), generators)\n\n\treturn additionalCapabilitiesGenerator\n}", "func (o *PluginConfigLinux) GetCapabilitiesOk() ([]string, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Capabilities, true\n}", "func (o *MicrosoftGraphVerifiedDomain) GetCapabilities() string {\n\tif o == nil || o.Capabilities == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Capabilities\n}", "func (ns *NodeServer) NodeGetCapabilities(ctx context.Context, req *csi.NodeGetCapabilitiesRequest) (*csi.NodeGetCapabilitiesResponse, error) {\n\tutil.DebugLog(ctx, \"get node Capabilities\")\n\treturn &csi.NodeGetCapabilitiesResponse{\n\t\tCapabilities: []*csi.NodeServiceCapability{\n\t\t\t{\n\t\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\t\tType: csi.NodeServiceCapability_RPC_STAGE_UNSTAGE_VOLUME,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\t\tType: csi.NodeServiceCapability_RPC_GET_VOLUME_STATS,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\t\tType: csi.NodeServiceCapability_RPC_EXPAND_VOLUME,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}, nil\n}", "func (o *MicrosoftGraphVerifiedDomain) HasCapabilities() bool {\n\tif o != nil && o.Capabilities != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func GetCapabilitiesAsMap(capabilities []configv1.ClusterVersionCapability) map[configv1.ClusterVersionCapability]struct{} {\n\tcaps := make(map[configv1.ClusterVersionCapability]struct{}, len(capabilities))\n\tfor _, c := range capabilities {\n\t\tcaps[c] = struct{}{}\n\t}\n\treturn caps\n}", "func (o GetRegionResultOutput) Capabilities() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v GetRegionResult) []string { return v.Capabilities }).(pulumi.StringArrayOutput)\n}", "func (d *Driver) NodeGetCapabilities(ctx context.Context, req *csi.NodeGetCapabilitiesRequest) (*csi.NodeGetCapabilitiesResponse, error) {\n\tnscaps := []*csi.NodeServiceCapability{\n\t\t&csi.NodeServiceCapability{\n\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\tType: csi.NodeServiceCapability_RPC_STAGE_UNSTAGE_VOLUME,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t&csi.NodeServiceCapability{\n\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\tType: csi.NodeServiceCapability_RPC_EXPAND_VOLUME,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t&csi.NodeServiceCapability{\n\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\tType: csi.NodeServiceCapability_RPC_GET_VOLUME_STATS,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\td.log.WithFields(logrus.Fields{\n\t\t\"node_capabilities\": nscaps,\n\t\t\"method\": \"node_get_capabilities\",\n\t}).Info(\"node get capabilities called\")\n\treturn &csi.NodeGetCapabilitiesResponse{\n\t\tCapabilities: nscaps,\n\t}, nil\n}", "func (c *logsConnector) Capabilities() consumer.Capabilities {\n\treturn consumer.Capabilities{MutatesData: false}\n}", "func (c *Client) CheckCapabilities(ctx context.Context) (err error) {\n\t// Only query the server once. There is no need for a lock, because we will\n\t// usually make the call on startup.\n\tif c.serverCaps == nil {\n\t\tif c.serverCaps, err = c.GetCapabilities(ctx); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif err := digest.CheckCapabilities(c.serverCaps); err != nil {\n\t\treturn errors.Wrapf(err, \"digest function mismatch\")\n\t}\n\n\tif c.serverCaps.CacheCapabilities != nil {\n\t\tc.MaxBatchSize = MaxBatchSize(c.serverCaps.CacheCapabilities.MaxBatchTotalSizeBytes)\n\t}\n\n\tif useCompression := c.CompressedBytestreamThreshold >= 0; useCompression {\n\t\tif c.serverCaps.CacheCapabilities.SupportedCompressors == nil {\n\t\t\treturn errors.New(\"the server does not support compression\")\n\t\t}\n\n\t\tfoundZstd := false\n\t\tfor _, sComp := range c.serverCaps.CacheCapabilities.SupportedCompressors {\n\t\t\tif sComp == repb.Compressor_ZSTD {\n\t\t\t\tfoundZstd = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !foundZstd {\n\t\t\treturn errors.New(\"zstd is not supported by server, while the SDK only supports ZSTD compression\")\n\t\t}\n\t\tfor _, compressor := range c.serverCaps.CacheCapabilities.SupportedBatchUpdateCompressors {\n\t\t\tif compressor == repb.Compressor_ZSTD {\n\t\t\t\tc.batchCompression = true\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (ns *nodeServer) NodeGetCapabilities(ctx context.Context, req *csi.NodeGetCapabilitiesRequest) (*csi.NodeGetCapabilitiesResponse, error) {\n\treturn &csi.NodeGetCapabilitiesResponse{\n\t\tCapabilities: []*csi.NodeServiceCapability{\n\t\t\t{\n\t\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\t\tType: csi.NodeServiceCapability_RPC_STAGE_UNSTAGE_VOLUME,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\t\tType: csi.NodeServiceCapability_RPC_GET_VOLUME_STATS,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t}, {\n\t\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\t\tType: csi.NodeServiceCapability_RPC_EXPAND_VOLUME,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}, nil\n}", "func (rtp *resourceTraceProcessor) GetCapabilities() component.ProcessorCapabilities {\n\treturn component.ProcessorCapabilities{MutatesConsumedData: true}\n}", "func (c *ServerConn) Features() map[string]string {\n\treturn c.features\n}", "func (l *RemoteProvider) GetProviderCapabilities(w http.ResponseWriter, r *http.Request) {\n\tencoder := json.NewEncoder(w)\n\tif err := encoder.Encode(l.ProviderProperties); err != nil {\n\t\thttp.Error(w, ErrEncoding(err, \"Provider Capablity\").Error(), http.StatusInternalServerError)\n\t}\n}", "func GetCapabilities(dc discovery.DiscoveryInterface) (*Capabilities, error) {\n\tkubeVersion, err := dc.ServerVersion()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not get server version from Kubernetes\")\n\t}\n\t// Issue #6361:\n\t// Client-Go emits an error when an API service is registered but unimplemented.\n\t// We trap that error here and print a warning. But since the discovery client continues\n\t// building the API object, it is correctly populated with all valid APIs.\n\t// See https://github.com/kubernetes/kubernetes/issues/72051#issuecomment-521157642\n\tapiVersions, err := GetVersionSet(dc)\n\tif err != nil {\n\t\tif discovery.IsGroupDiscoveryFailedError(err) {\n\t\t\tblog.Infof(\"WARNING: The Kubernetes server has an orphaned API service. Server reports: %s\", err)\n\t\t\tblog.Infof(\"WARNING: To fix this, kubectl delete apiservice <service-name>\")\n\t\t} else {\n\t\t\treturn nil, errors.Wrap(err, \"could not get apiVersions from Kubernetes\")\n\t\t}\n\t}\n\treturn &Capabilities{\n\t\tKubeVersion: KubeVersion{\n\t\t\tMajor: kubeVersion.Major,\n\t\t\tMinor: kubeVersion.Minor,\n\t\t\tVersion: kubeVersion.GitVersion,\n\t\t},\n\t\tAPIVersions: apiVersions,\n\t}, nil\n}", "func newControllerCapabilities() []*csi.ControllerServiceCapability {\n\tfromType := func(cap csi.ControllerServiceCapability_RPC_Type) *csi.ControllerServiceCapability {\n\t\treturn &csi.ControllerServiceCapability{\n\t\t\tType: &csi.ControllerServiceCapability_Rpc{\n\t\t\t\tRpc: &csi.ControllerServiceCapability_RPC{\n\t\t\t\t\tType: cap,\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t}\n\n\tvar capabilities []*csi.ControllerServiceCapability\n\tfor _, cap := range []csi.ControllerServiceCapability_RPC_Type{\n\t\tcsi.ControllerServiceCapability_RPC_CREATE_DELETE_VOLUME,\n\t\tcsi.ControllerServiceCapability_RPC_CREATE_DELETE_SNAPSHOT,\n\t\tcsi.ControllerServiceCapability_RPC_LIST_SNAPSHOTS,\n\t\tcsi.ControllerServiceCapability_RPC_LIST_VOLUMES,\n\t} {\n\t\tcapabilities = append(capabilities, fromType(cap))\n\t}\n\treturn capabilities\n}", "func (ns *nodeServer) NodeGetCapabilities(ctx context.Context, req *csi.NodeGetCapabilitiesRequest) (*csi.NodeGetCapabilitiesResponse, error) {\n\treturn &csi.NodeGetCapabilitiesResponse{\n\t\tCapabilities: []*csi.NodeServiceCapability{\n\t\t\t{\n\t\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\t\tType: csi.NodeServiceCapability_RPC_STAGE_UNSTAGE_VOLUME,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\t\tType: csi.NodeServiceCapability_RPC_EXPAND_VOLUME,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}, nil\n}", "func New(capabilities ...string) *Capability {\n\treturn &Capability{\n\t\tCapabilities: capabilities,\n\t}\n}", "func (c *Channel) Capabilities() channelconfig.ApplicationCapabilities {\n\tac, ok := c.Resources().ApplicationConfig()\n\tif !ok {\n\t\treturn nil\n\t}\n\treturn ac.Capabilities()\n}", "func (mt *Mytoken) VerifyCapabilities(required ...api.Capability) bool {\n\tif mt.Capabilities == nil || len(mt.Capabilities) == 0 {\n\t\treturn false\n\t}\n\tfor _, c := range required {\n\t\tif !mt.Capabilities.Has(c) {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (mt *Mytoken) VerifyCapabilities(required ...api.Capability) bool {\n\tif mt.Capabilities == nil || len(mt.Capabilities) == 0 {\n\t\treturn false\n\t}\n\tfor _, c := range required {\n\t\tif !mt.Capabilities.Has(c) {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (d *Driver) NodeGetCapabilities(ctx context.Context, request *csi.NodeGetCapabilitiesRequest) (*csi.NodeGetCapabilitiesResponse, error) {\n\tcaps := []*csi.NodeServiceCapability{\n\t\t{\n\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\tType: csi.NodeServiceCapability_RPC_STAGE_UNSTAGE_VOLUME,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\tType: csi.NodeServiceCapability_RPC_GET_VOLUME_STATS,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tType: &csi.NodeServiceCapability_Rpc{\n\t\t\t\tRpc: &csi.NodeServiceCapability_RPC{\n\t\t\t\t\tType: csi.NodeServiceCapability_RPC_UNKNOWN,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\td.log.WithFields(logrus.Fields{\n\t\t\"node_capabilities\": caps,\n\t\t\"method\": \"node_get_capabilities\",\n\t}).Info(\"node get capabilities called\")\n\treturn &csi.NodeGetCapabilitiesResponse{\n\t\tCapabilities: caps,\n\t}, nil\n}", "func GetVolumeCapabilities(pvSpec v1.PersistentVolumeSpec, singleNodeMultiWriterCapable bool) (*csilib.VolumeCapability, error) {\n\tif pvSpec.CSI == nil {\n\t\treturn nil, errors.New(\"CSI volume source was nil\")\n\t}\n\n\tvar cap *csilib.VolumeCapability\n\tif pvSpec.VolumeMode != nil && *pvSpec.VolumeMode == v1.PersistentVolumeBlock {\n\t\tcap = &csilib.VolumeCapability{\n\t\t\tAccessType: &csilib.VolumeCapability_Block{\n\t\t\t\tBlock: &csilib.VolumeCapability_BlockVolume{},\n\t\t\t},\n\t\t\tAccessMode: &csilib.VolumeCapability_AccessMode{},\n\t\t}\n\n\t} else {\n\t\tfsType := pvSpec.CSI.FSType\n\n\t\tcap = &csilib.VolumeCapability{\n\t\t\tAccessType: &csilib.VolumeCapability_Mount{\n\t\t\t\tMount: &csilib.VolumeCapability_MountVolume{\n\t\t\t\t\tFsType: fsType,\n\t\t\t\t\tMountFlags: pvSpec.MountOptions,\n\t\t\t\t},\n\t\t\t},\n\t\t\tAccessMode: &csilib.VolumeCapability_AccessMode{},\n\t\t}\n\t}\n\n\tam, err := accessmodes.ToCSIAccessMode(pvSpec.AccessModes, singleNodeMultiWriterCapable)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcap.AccessMode.Mode = am\n\treturn cap, nil\n}", "func (sc *ServerConn) Features(ctx context.Context) (*ServerFeatures, error) {\n\tvar feats ServerFeatures\n\terr := sc.Request(ctx, \"server.features\", nil, &feats)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &feats, nil\n}", "func Caps() Capabilities {\n\tcaps := C.bgfx_get_caps()\n\treturn Capabilities{\n\t\tRendererType: RendererType(caps.rendererType),\n\t\tSupported: CapFlags(caps.supported),\n\t\tMaxTextureSize: uint16(caps.maxTextureSize),\n\t\tMaxDrawCalls: uint16(caps.maxDrawCalls),\n\t\tMaxFBAttachments: uint8(caps.maxFBAttachments),\n\t\tFormats: *(*[C.BGFX_TEXTURE_FORMAT_COUNT]uint8)(unsafe.Pointer(&caps.formats)),\n\t}\n}", "func (c CapabilitiesRegistry) Add(capability interface{}) {\n\tswitch capability.(type) {\n\tcase types.AttachedCapability:\n\t\tlogging.Debug(\"[Registry] Adding ATTACHED capability to registry: %s\\n\", capability.(types.AttachedCapability).Title)\n\t\tRegistry.attached = append(Registry.attached, capability.(types.AttachedCapability))\n\tcase dockeradapter.DockerAPICapability:\n\t\tlogging.Debug(\"[Registry] Adding DOCKER API capability to registry: %s\\n\", capability.(dockeradapter.DockerAPICapability).Title)\n\t\tRegistry.dockerAPI = append(Registry.dockerAPI, capability.(dockeradapter.DockerAPICapability))\n\tdefault:\n\t\tlogging.Info(\"[Registry] Invalid capability type detected. Exiting..\")\n\t\tos.Exit(1)\n\t}\n}", "func (l *Libvirt) ConnectGetCapabilities() (rCapabilities string, err error) {\n\tvar buf []byte\n\n\tvar r response\n\tr, err = l.requestStream(7, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Return value unmarshaling\n\ttpd := typedParamDecoder{}\n\tct := map[string]xdr.TypeDecoder{\"libvirt.TypedParam\": tpd}\n\trdr := bytes.NewReader(r.Payload)\n\tdec := xdr.NewDecoderCustomTypes(rdr, 0, ct)\n\t// Capabilities: string\n\t_, err = dec.Decode(&rCapabilities)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}" ]
[ "0.7242673", "0.6908501", "0.68707454", "0.67757386", "0.6752177", "0.6611623", "0.6568753", "0.6544832", "0.6525126", "0.6461981", "0.64540225", "0.64472747", "0.64395314", "0.63170815", "0.62939256", "0.62640125", "0.6262464", "0.6258976", "0.62495536", "0.6239869", "0.62083316", "0.6203329", "0.6202581", "0.6164683", "0.615461", "0.61346865", "0.60399497", "0.6038149", "0.6038149", "0.6021698", "0.60024375", "0.59593487", "0.59503865", "0.5946005", "0.59404945", "0.58947384", "0.58873093", "0.58657503", "0.5860465", "0.58443344", "0.5839894", "0.5803541", "0.57917124", "0.5791345", "0.5785931", "0.5772292", "0.5764793", "0.5758362", "0.5756547", "0.5721204", "0.5712479", "0.57081217", "0.56902355", "0.56885064", "0.56714153", "0.56294227", "0.5597979", "0.5588034", "0.5574235", "0.55634516", "0.555615", "0.5500445", "0.54932004", "0.5486496", "0.5485731", "0.54782325", "0.54765207", "0.5475724", "0.54513466", "0.5444694", "0.54218036", "0.53967935", "0.5393943", "0.53890103", "0.53859127", "0.53828955", "0.5377395", "0.5376304", "0.5372389", "0.5331993", "0.5328301", "0.5301329", "0.5295551", "0.5290416", "0.5285516", "0.52744555", "0.52710944", "0.5262468", "0.5257554", "0.52280277", "0.52139616", "0.52126914", "0.5208376", "0.5208376", "0.51997596", "0.5197949", "0.51754445", "0.517274", "0.51538444", "0.5149843" ]
0.6607212
6
Fingerprint is called by the client when the plugin is started. It allows the driver to indicate its health to the client. The channel returned should immediately send an initial Fingerprint, then send periodic updates at an interval that is appropriate for the driver until the context is canceled.
func (d *Driver) Fingerprint(ctx context.Context) (<-chan *drivers.Fingerprint, error) { err := shelpers.Init() if err != nil { d.logger.Error("Could not init stats helper", "err", err) return nil, err } ch := make(chan *drivers.Fingerprint) go d.handleFingerprint(ctx, ch) return ch, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (i *instanceManager) fingerprint() {\n\tfpChan, cancel, err := i.dispenseFingerprintCh()\n\tif err != nil {\n\t\ti.logger.Error(\"failed to dispense driver plugin\", \"error\", err)\n\t}\n\n\t// backoff and retry used if the RPC is closed by the other end\n\tvar backoff time.Duration\n\tvar retry int\n\tfor {\n\t\tif backoff > 0 {\n\t\t\tselect {\n\t\t\tcase <-time.After(backoff):\n\t\t\tcase <-i.ctx.Done():\n\t\t\t\tcancel()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\tselect {\n\t\tcase <-i.ctx.Done():\n\t\t\tcancel()\n\t\t\treturn\n\t\tcase fp, ok := <-fpChan:\n\t\t\tif ok {\n\t\t\t\tif fp.Err == nil {\n\t\t\t\t\ti.handleFingerprint(fp)\n\t\t\t\t} else {\n\t\t\t\t\ti.logger.Warn(\"received fingerprint error from driver\", \"error\", fp.Err)\n\t\t\t\t\ti.handleFingerprintError()\n\t\t\t\t}\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// avoid fingerprinting again if ctx and fpChan both close\n\t\t\tif i.ctx.Err() != nil {\n\t\t\t\tcancel()\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// if the channel is closed attempt to open a new one\n\t\t\tnewFpChan, newCancel, err := i.dispenseFingerprintCh()\n\t\t\tif err != nil {\n\t\t\t\ti.logger.Warn(\"error fingerprinting driver\", \"error\", err, \"retry\", retry)\n\t\t\t\ti.handleFingerprintError()\n\n\t\t\t\t// Calculate the new backoff\n\t\t\t\tbackoff = (1 << (2 * uint64(retry))) * driverFPBackoffBaseline\n\t\t\t\tif backoff > driverFPBackoffLimit {\n\t\t\t\t\tbackoff = driverFPBackoffLimit\n\t\t\t\t}\n\t\t\t\t// Increment retry counter\n\t\t\t\tretry++\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tcancel()\n\t\t\tfpChan = newFpChan\n\t\t\tcancel = newCancel\n\n\t\t\t// Reset backoff\n\t\t\tbackoff = 0\n\t\t\tretry = 0\n\t\t}\n\t}\n}", "func (fm *FingerprintManager) watchDriverFingerprint(fpChan <-chan *drivers.Fingerprint, name string, cancel context.CancelFunc) {\n\tvar backoff time.Duration\n\tvar retry int\n\tfor {\n\t\tif backoff > 0 {\n\t\t\ttime.Sleep(backoff)\n\t\t}\n\t\tselect {\n\t\tcase <-fm.shutdownCh:\n\t\t\tcancel()\n\t\t\treturn\n\t\tcase fp, ok := <-fpChan:\n\t\t\tif ok && fp.Err == nil {\n\t\t\t\tfm.processDriverFingerprint(fp, name)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// if the channel is closed attempt to open a new one\n\t\t\tnewFpChan, newCancel, err := fm.dispenseDriverFingerprint(name)\n\t\t\tif err != nil {\n\t\t\t\tfm.logger.Warn(\"failed to fingerprint driver, retrying in 30s\", \"error\", err, \"retry\", retry)\n\t\t\t\tdi := &structs.DriverInfo{\n\t\t\t\t\tHealthy: false,\n\t\t\t\t\tHealthDescription: \"failed to fingerprint driver\",\n\t\t\t\t\tUpdateTime: time.Now(),\n\t\t\t\t}\n\t\t\t\tif n := fm.updateNodeFromDriver(name, di); n != nil {\n\t\t\t\t\tfm.setNode(n)\n\t\t\t\t}\n\n\t\t\t\t// Calculate the new backoff\n\t\t\t\tbackoff = (1 << (2 * uint64(retry))) * driverFPBackoffBaseline\n\t\t\t\tif backoff > driverFPBackoffLimit {\n\t\t\t\t\tbackoff = driverFPBackoffLimit\n\t\t\t\t}\n\t\t\t\tretry++\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tcancel()\n\t\t\tfpChan = newFpChan\n\t\t\tcancel = newCancel\n\n\t\t\t// Reset backoff\n\t\t\tbackoff = 0\n\t\t\tretry = 0\n\t\t}\n\t}\n}", "func (i *instanceManager) handleFingerprint(fp *drivers.Fingerprint) {\n\tattrs := make(map[string]string, len(fp.Attributes))\n\tfor key, attr := range fp.Attributes {\n\t\tattrs[key] = attr.GoString()\n\t}\n\tdi := &structs.DriverInfo{\n\t\tAttributes: attrs,\n\t\tDetected: fp.Health != drivers.HealthStateUndetected,\n\t\tHealthy: fp.Health == drivers.HealthStateHealthy,\n\t\tHealthDescription: fp.HealthDescription,\n\t\tUpdateTime: time.Now(),\n\t}\n\ti.updateNodeFromDriver(i.id.Name, di)\n\n\t// log detected/undetected state changes after the initial fingerprint\n\ti.lastHealthStateMu.Lock()\n\tif i.hasFingerprinted {\n\t\tif i.lastHealthState != fp.Health {\n\t\t\ti.logger.Info(\"driver health state has changed\", \"previous\", i.lastHealthState, \"current\", fp.Health, \"description\", fp.HealthDescription)\n\t\t}\n\t}\n\ti.lastHealthState = fp.Health\n\ti.lastHealthStateMu.Unlock()\n\n\t// if this is the first fingerprint, mark that we have received it\n\tif !i.hasFingerprinted {\n\t\ti.logger.Debug(\"initial driver fingerprint\", \"health\", fp.Health, \"description\", fp.HealthDescription)\n\t\tclose(i.firstFingerprintCh)\n\t\ti.hasFingerprinted = true\n\t}\n}", "func (fm *FingerprintManager) dispenseDriverFingerprint(driverName string) (<-chan *drivers.Fingerprint, context.CancelFunc, error) {\n\tplug, err := fm.singletonLoader.Dispense(driverName, base.PluginTypeDriver, fm.getConfig().NomadPluginConfig(), fm.logger)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tdriver, ok := plug.Plugin().(drivers.DriverPlugin)\n\tif !ok {\n\t\treturn nil, nil, fmt.Errorf(\"registered driver plugin %q does not implement DriverPlugin interface\", driverName)\n\t}\n\n\tctx, cancel := context.WithCancel(context.Background())\n\tfingerCh, err := driver.Fingerprint(ctx)\n\tif err != nil {\n\t\tcancel()\n\t\treturn nil, nil, err\n\t}\n\n\treturn fingerCh, cancel, nil\n}", "func (i *instanceManager) dispenseFingerprintCh() (<-chan *drivers.Fingerprint, context.CancelFunc, error) {\n\tdriver, err := i.dispense()\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tctx, cancel := context.WithCancel(i.ctx)\n\tfingerCh, err := driver.Fingerprint(ctx)\n\tif err != nil {\n\t\tcancel()\n\t\treturn nil, nil, err\n\t}\n\n\treturn fingerCh, cancel, nil\n}", "func (i *instanceManager) WaitForFirstFingerprint(ctx context.Context) {\n\tselect {\n\tcase <-i.ctx.Done():\n\tcase <-ctx.Done():\n\tcase <-i.firstFingerprintCh:\n\t}\n}", "func TestFingerprintManager_HealthCheck_Driver(t *testing.T) {\n\tt.Parallel()\n\trequire := require.New(t)\n\ttestClient := TestClient(t, func(c *config.Config) {\n\t\tc.Options = map[string]string{\n\t\t\t\"driver.raw_exec.enable\": \"1\",\n\t\t\t\"test.shutdown_periodic_after\": \"true\",\n\t\t\t\"test.shutdown_periodic_duration\": \"2\",\n\t\t}\n\t})\n\n\ttestClient.logger = testLogger()\n\tdefer testClient.Shutdown()\n\n\tfm := NewFingerprintManager(\n\t\ttestClient.GetConfig,\n\t\ttestClient.config.Node,\n\t\ttestClient.shutdownCh,\n\t\ttestClient.updateNodeFromFingerprint,\n\t\ttestClient.updateNodeFromDriver,\n\t\ttestClient.logger,\n\t)\n\n\terr := fm.Run()\n\trequire.Nil(err)\n\n\t// Ensure the mock driver is registered and healthy on the client\n\ttestutil.WaitForResult(func() (bool, error) {\n\t\tfm.nodeLock.Lock()\n\t\tnode := fm.node\n\t\tdefer fm.nodeLock.Unlock()\n\n\t\tmockDriverAttribute := node.Attributes[\"driver.mock_driver\"]\n\t\tif mockDriverAttribute == \"\" {\n\t\t\treturn false, fmt.Errorf(\"mock driver info should be set on the client attributes\")\n\t\t}\n\t\tmockDriverInfo := node.Drivers[\"mock_driver\"]\n\t\tif mockDriverInfo == nil {\n\t\t\treturn false, fmt.Errorf(\"mock driver info should be set on the client\")\n\t\t}\n\t\tif !mockDriverInfo.Healthy {\n\t\t\treturn false, fmt.Errorf(\"mock driver info should be healthy\")\n\t\t}\n\t\treturn true, nil\n\t}, func(err error) {\n\t\tt.Fatalf(\"err: %v\", err)\n\t})\n\n\t// Ensure that a default driver without health checks enabled is registered and healthy on the client\n\ttestutil.WaitForResult(func() (bool, error) {\n\t\tfm.nodeLock.Lock()\n\t\tnode := fm.node\n\t\tdefer fm.nodeLock.Unlock()\n\n\t\trawExecAttribute := node.Attributes[\"driver.raw_exec\"]\n\t\tif rawExecAttribute == \"\" {\n\t\t\treturn false, fmt.Errorf(\"raw exec info should be set on the client attributes\")\n\t\t}\n\t\trawExecInfo := node.Drivers[\"raw_exec\"]\n\t\tif rawExecInfo == nil {\n\t\t\treturn false, fmt.Errorf(\"raw exec driver info should be set on the client\")\n\t\t}\n\t\tif !rawExecInfo.Detected {\n\t\t\treturn false, fmt.Errorf(\"raw exec driver should be detected\")\n\t\t}\n\t\treturn true, nil\n\t}, func(err error) {\n\t\tt.Fatalf(\"err: %v\", err)\n\t})\n\n\t// Ensure the mock driver is registered\n\ttestutil.WaitForResult(func() (bool, error) {\n\t\tfm.nodeLock.Lock()\n\t\tnode := fm.node\n\t\tdefer fm.nodeLock.Unlock()\n\n\t\tmockDriverAttribute := node.Attributes[\"driver.mock_driver\"]\n\t\tif mockDriverAttribute == \"\" {\n\t\t\treturn false, fmt.Errorf(\"mock driver info should set on the client attributes\")\n\t\t}\n\t\tmockDriverInfo := node.Drivers[\"mock_driver\"]\n\t\tif mockDriverInfo == nil {\n\t\t\treturn false, fmt.Errorf(\"mock driver info should be set on the client\")\n\t\t}\n\t\tif !mockDriverInfo.Healthy {\n\t\t\treturn false, fmt.Errorf(\"mock driver info should not be healthy\")\n\t\t}\n\t\treturn true, nil\n\t}, func(err error) {\n\t\tt.Fatalf(\"err: %v\", err)\n\t})\n\n\t// Ensure that we don't duplicate health check information on the driver\n\t// health information\n\tfm.nodeLock.Lock()\n\tnode := fm.node\n\tfm.nodeLock.Unlock()\n\tmockDriverAttributes := node.Drivers[\"mock_driver\"].Attributes\n\trequire.NotContains(mockDriverAttributes, \"driver.mock_driver\")\n}", "func (m *manager) fingerprint() {\n\tfor {\n\t\tselect {\n\t\tcase <-m.ctx.Done():\n\t\t\treturn\n\t\tcase <-m.fingerprintResCh:\n\t\t}\n\n\t\t// Collect the data\n\t\tvar fingerprinted []*device.DeviceGroup\n\t\tfor _, i := range m.instances {\n\t\t\tfingerprinted = append(fingerprinted, i.Devices()...)\n\t\t}\n\n\t\t// Convert and update\n\t\tout := make([]*structs.NodeDeviceResource, len(fingerprinted))\n\t\tfor i, f := range fingerprinted {\n\t\t\tout[i] = convertDeviceGroup(f)\n\t\t}\n\n\t\t// Call the updater\n\t\tm.updater(out)\n\t}\n}", "func (fm *FingerprintManager) processDriverFingerprint(fp *drivers.Fingerprint, driverName string) {\n\tdi := &structs.DriverInfo{\n\t\tAttributes: fp.Attributes,\n\t\tDetected: fp.Health != drivers.HealthStateUndetected,\n\t\tHealthy: fp.Health == drivers.HealthStateHealthy,\n\t\tHealthDescription: fp.HealthDescription,\n\t\tUpdateTime: time.Now(),\n\t}\n\tif n := fm.updateNodeFromDriver(driverName, di); n != nil {\n\t\tfm.setNode(n)\n\t}\n}", "func (fm *FingerprintManager) runFingerprint(f fingerprint.Fingerprint, period time.Duration, name string) {\n\tfm.logger.Debug(\"fingerprinting periodically\", \"fingerprinter\", name, \"period\", period)\n\n\ttimer := time.NewTimer(period)\n\tdefer timer.Stop()\n\n\tfor {\n\t\tselect {\n\t\tcase <-timer.C:\n\t\t\ttimer.Reset(period)\n\n\t\t\t_, err := fm.fingerprint(name, f)\n\t\t\tif err != nil {\n\t\t\t\tfm.logger.Debug(\"error periodic fingerprinting\", \"error\", err, \"fingerprinter\", name)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\tcase <-fm.shutdownCh:\n\t\t\treturn\n\t\t}\n\t}\n}", "func (c *AccountsUpdateCall) Fingerprint(fingerprint string) *AccountsUpdateCall {\n\tc.params_.Set(\"fingerprint\", fmt.Sprintf(\"%v\", fingerprint))\n\treturn c\n}", "func (f *CGroupFingerprint) Fingerprint(req *FingerprintRequest, resp *FingerprintResponse) error {\n\tmount, err := f.mountPointDetector.MountPoint()\n\tif err != nil {\n\t\tf.clearCGroupAttributes(resp)\n\t\treturn fmt.Errorf(\"failed to discover cgroup mount point: %s\", err)\n\t}\n\n\t// Check if a cgroup mount point was found.\n\tif mount == \"\" {\n\t\tf.clearCGroupAttributes(resp)\n\t\tif f.lastState == cgroupAvailable {\n\t\t\tf.logger.Warn(\"cgroups are now unavailable\")\n\t\t}\n\t\tf.lastState = cgroupUnavailable\n\t\treturn nil\n\t}\n\n\t// Check the version in use.\n\tversion := f.versionDetector.CgroupVersion()\n\n\tresp.AddAttribute(cgroupMountPointAttribute, mount)\n\tresp.AddAttribute(cgroupVersionAttribute, version)\n\tresp.Detected = true\n\n\tif f.lastState == cgroupUnavailable {\n\t\tf.logger.Info(\"cgroups are available\")\n\t}\n\tf.lastState = cgroupAvailable\n\treturn nil\n}", "func Fingerprint(scope *Scope, data tf.Output, method tf.Output) (fingerprint tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"Fingerprint\",\n\t\tInput: []tf.Input{\n\t\t\tdata, method,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func (c *AccountsContainersMacrosUpdateCall) Fingerprint(fingerprint string) *AccountsContainersMacrosUpdateCall {\n\tc.params_.Set(\"fingerprint\", fmt.Sprintf(\"%v\", fingerprint))\n\treturn c\n}", "func (c *AccountsContainersTriggersUpdateCall) Fingerprint(fingerprint string) *AccountsContainersTriggersUpdateCall {\n\tc.params_.Set(\"fingerprint\", fmt.Sprintf(\"%v\", fingerprint))\n\treturn c\n}", "func (u *User) RegisterFingerprint(fp string) (map[string]interface{}, error) {\n\tlog.info(\"========== REGISTER FINGERPRINT ==========\")\n\turl := buildURL(path[\"auth\"], u.UserID)\n\n\tu.request.fingerprint = fp\n\n\tdata := `{ \"refresh_token\": \"` + u.RefreshToken + `\" }`\n\n\tres, err := u.do(\"POST\", url, data, nil)\n\n\treturn res, err\n}", "func (fp *FingerprintManager) Run() error {\n\t// First, set up all fingerprints\n\tcfg := fp.getConfig()\n\twhitelistFingerprints := cfg.ReadStringListToMap(\"fingerprint.whitelist\")\n\twhitelistFingerprintsEnabled := len(whitelistFingerprints) > 0\n\tblacklistFingerprints := cfg.ReadStringListToMap(\"fingerprint.blacklist\")\n\n\tfp.logger.Debug(\"built-in fingerprints\", \"fingerprinters\", fingerprint.BuiltinFingerprints())\n\n\tvar availableFingerprints []string\n\tvar skippedFingerprints []string\n\tfor _, name := range fingerprint.BuiltinFingerprints() {\n\t\t// Skip modules that are not in the whitelist if it is enabled.\n\t\tif _, ok := whitelistFingerprints[name]; whitelistFingerprintsEnabled && !ok {\n\t\t\tskippedFingerprints = append(skippedFingerprints, name)\n\t\t\tcontinue\n\t\t}\n\t\t// Skip modules that are in the blacklist\n\t\tif _, ok := blacklistFingerprints[name]; ok {\n\t\t\tskippedFingerprints = append(skippedFingerprints, name)\n\t\t\tcontinue\n\t\t}\n\n\t\tavailableFingerprints = append(availableFingerprints, name)\n\t}\n\n\tif err := fp.setupFingerprinters(availableFingerprints); err != nil {\n\t\treturn err\n\t}\n\n\tif len(skippedFingerprints) != 0 {\n\t\tfp.logger.Debug(\"fingerprint modules skipped due to white/blacklist\",\n\t\t\t\"skipped_fingerprinters\", skippedFingerprints)\n\t}\n\n\t// Next, set up drivers\n\t// Build the white/blacklists of drivers.\n\twhitelistDrivers := cfg.ReadStringListToMap(\"driver.whitelist\")\n\twhitelistDriversEnabled := len(whitelistDrivers) > 0\n\tblacklistDrivers := cfg.ReadStringListToMap(\"driver.blacklist\")\n\n\tvar availDrivers []string\n\tvar skippedDrivers []string\n\n\tfor _, pl := range fp.singletonLoader.Catalog()[base.PluginTypeDriver] {\n\t\tname := pl.Name\n\t\t// Skip fingerprinting drivers that are not in the whitelist if it is\n\t\t// enabled.\n\t\tif _, ok := whitelistDrivers[name]; whitelistDriversEnabled && !ok {\n\t\t\tskippedDrivers = append(skippedDrivers, name)\n\t\t\tcontinue\n\t\t}\n\t\t// Skip fingerprinting drivers that are in the blacklist\n\t\tif _, ok := blacklistDrivers[name]; ok {\n\t\t\tskippedDrivers = append(skippedDrivers, name)\n\t\t\tcontinue\n\t\t}\n\n\t\tavailDrivers = append(availDrivers, name)\n\t}\n\n\tif err := fp.setupDrivers(availDrivers); err != nil {\n\t\treturn err\n\t}\n\n\tif len(skippedDrivers) > 0 {\n\t\tfp.logger.Debug(\"drivers skipped due to white/blacklist\", \"skipped_drivers\", skippedDrivers)\n\t}\n\treturn nil\n}", "func (c *AccountsContainersUpdateCall) Fingerprint(fingerprint string) *AccountsContainersUpdateCall {\n\tc.params_.Set(\"fingerprint\", fmt.Sprintf(\"%v\", fingerprint))\n\treturn c\n}", "func (c *AccountsContainersVersionsUpdateCall) Fingerprint(fingerprint string) *AccountsContainersVersionsUpdateCall {\n\tc.params_.Set(\"fingerprint\", fmt.Sprintf(\"%v\", fingerprint))\n\treturn c\n}", "func NewFingerprintManager(\n\tsingletonLoader loader.PluginCatalog,\n\tgetConfig func() *config.Config,\n\tnode *structs.Node,\n\tshutdownCh chan struct{},\n\tupdateNodeAttributes func(*cstructs.FingerprintResponse) *structs.Node,\n\tupdateNodeFromDriver func(string, *structs.DriverInfo) *structs.Node,\n\tlogger log.Logger) *FingerprintManager {\n\n\treturn &FingerprintManager{\n\t\tsingletonLoader: singletonLoader,\n\t\tgetConfig: getConfig,\n\t\tupdateNodeAttributes: updateNodeAttributes,\n\t\tupdateNodeFromDriver: updateNodeFromDriver,\n\t\tnode: node,\n\t\tshutdownCh: shutdownCh,\n\t\tlogger: logger.Named(\"fingerprint_mgr\"),\n\t}\n}", "func (fm *FingerprintManager) fingerprint(name string, f fingerprint.Fingerprint) (bool, error) {\n\tvar response cstructs.FingerprintResponse\n\n\tfm.nodeLock.Lock()\n\trequest := &cstructs.FingerprintRequest{Config: fm.getConfig(), Node: fm.node}\n\terr := f.Fingerprint(request, &response)\n\tfm.nodeLock.Unlock()\n\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tif node := fm.updateNodeAttributes(&response); node != nil {\n\t\tfm.setNode(node)\n\t}\n\n\treturn response.Detected, nil\n}", "func (m *Monitor) Start() (chan<- StartPlugin, error) {\n\tm.lock.Lock()\n\tdefer m.lock.Unlock()\n\tif m.startChan != nil {\n\t\treturn m.inputChan, nil\n\t}\n\n\tch := make(chan StartPlugin)\n\tm.startChan = ch\n\tm.inputChan = ch\n\n\tgo func() {\n\tloop:\n\t\tfor {\n\t\t\treq, open := <-m.startChan\n\t\t\tif !open {\n\t\t\t\tm.inputChan = nil\n\t\t\t\tlog.Debug(\"Plugin activation input closed. Stopping.\")\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tconfigCopy := types.AnyBytes(nil)\n\n\t\t\tif req.Options == nil {\n\t\t\t\t// match first by full name of the form lookup/type -- 'specialization'\n\t\t\t\tproperties, has := m.rules[req.Key][req.Exec]\n\t\t\t\tif !has {\n\t\t\t\t\tlog.Warn(\"no plugin kind defined\", \"key\", req.Key)\n\t\t\t\t\treq.reportError(req.Key, plugin.Name(\"\"), nil, errNoConfig)\n\t\t\t\t\tcontinue loop\n\t\t\t\t}\n\t\t\t\tif properties != nil {\n\t\t\t\t\t*configCopy = *properties\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t*configCopy = *req.Options\n\t\t\t}\n\n\t\t\texec, has := m.execs[req.Exec]\n\t\t\tif !has {\n\t\t\t\treq.reportError(req.Key, plugin.Name(\"\"), configCopy, fmt.Errorf(\"no exec:%v\", req.Exec))\n\t\t\t\tcontinue loop\n\t\t\t}\n\n\t\t\tlog.Info(\"Starting plugin\", \"executor\", exec.Name(), \"key\", req.Key, \"name\", req.Name, \"exec\", req.Exec)\n\n\t\t\tname, block, err := exec.Exec(req.Key, req.Name, configCopy)\n\t\t\tif err != nil {\n\t\t\t\tlog.Warn(\"error starting plugin\", \"err\", err, \"config\", configCopy,\n\t\t\t\t\t\"key\", req.Key, \"name\", req.Name, \"as\", name)\n\t\t\t\treq.reportError(req.Key, req.Name, configCopy, err)\n\t\t\t\tcontinue loop\n\t\t\t}\n\n\t\t\tlog.Info(\"Waiting for startup\", \"key\", req.Key, \"name\", req.Name,\n\t\t\t\t\"config\", configCopy.String(), \"as\", name)\n\t\t\terr = <-block\n\t\t\tif err != nil {\n\t\t\t\tlog.Warn(\"error startup\", \"err\", err, \"config\", configCopy, \"key\", req.Key, \"name\", req.Name)\n\t\t\t\treq.reportError(req.Key, name, configCopy, err)\n\t\t\t\tcontinue loop\n\t\t\t}\n\n\t\t\treq.reportSuccess(req.Key, name, configCopy)\n\t\t}\n\t}()\n\n\treturn m.inputChan, nil\n}", "func (s *EBPFSocketInfoProbe) Start() {\n\ts.tracer.Start()\n\n\ts.scanProc()\n\ts.updateMetadata()\n\n\tgo func() {\n\t\tseconds := config.GetInt(\"agent.topology.socketinfo.host_update\")\n\t\tticker := time.NewTicker(time.Duration(seconds) * time.Second)\n\t\tdefer ticker.Stop()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-s.quit:\n\t\t\t\treturn\n\t\t\tcase <-ticker.C:\n\t\t\t\ts.updateMetadata()\n\t\t\t}\n\t\t}\n\t}()\n}", "func (m *CambriconDevicePlugin) Start() error {\n\terr := m.cleanup()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsock, err := net.Listen(\"unix\", m.socket)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tm.server = grpc.NewServer([]grpc.ServerOption{}...)\n\tpluginapi.RegisterDevicePluginServer(m.server, m)\n\n\tgo m.server.Serve(sock)\n\n\t// Wait for server to start by launching a blocking connection\n\tconn, err := dial(m.socket, 5*time.Second)\n\tif err != nil {\n\t\treturn err\n\t}\n\tconn.Close()\n\n\tif !m.options.DisableHealthCheck {\n\t\tgo m.healthcheck()\n\t}\n\n\treturn nil\n}", "func probe(c *dns.Client, addr string, f *fingerprint) *fingerprint {\n\tm := f.msg()\n\tr, err := c.Exchange(m, addr)\n\tif err != nil {\n\t\treturn errorToFingerprint(err)\n\t}\n\treturn toFingerprint(r)\n}", "func (s *Driver) Initialize(lc logger.LoggingClient, asyncCh chan<- *dsModels.AsyncValues, deviceCh chan<- []dsModels.DiscoveredDevice) error {\n\ts.lc = lc\n\ts.asyncCh = asyncCh\n\ts.deviceCh = deviceCh\n\ts.temperatureSensor = NewTemperatureSensor(lc)\n\treturn nil\n}", "func (o *ApplianceImageBundleAllOf) SetFingerprint(v string) {\n\to.Fingerprint = &v\n}", "func (p *rpmPlugin) Run() {\n\tif p.frequency <= config.FREQ_DISABLE_SAMPLING {\n\t\trpmlog.Debug(\"Disabled.\")\n\t\treturn\n\t}\n\n\t// Subscribe to filesystem events are care about\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\trpmlog.WithError(err).Error(\"can't instantiate rpm watcher\")\n\t\tp.Unregister()\n\t\treturn\n\t}\n\n\terr = watcher.Add(\"/var/lib/rpm/.rpm.lock\")\n\tif err != nil {\n\t\t// Some old distros, like SLES 11, do not provide .rpm.lock file, but the same\n\t\t// effect can be achieved by listening some standard files from the RPM database\n\t\terr = watcher.Add(\"/var/lib/rpm/Installtid\")\n\t\tif err != nil {\n\t\t\trpmlog.WithError(err).Error(\"can't setup trigger file watcher for rpm\")\n\t\t\tp.Unregister()\n\t\t\treturn\n\t\t}\n\t}\n\n\tcounter := 1\n\tticker := time.NewTicker(1)\n\tfor {\n\t\tselect {\n\t\tcase event, ok := <-watcher.Events:\n\t\t\tif ok {\n\t\t\t\tif event.Op&fsnotify.Write == fsnotify.Write {\n\t\t\t\t\tcounter = counter + 1\n\t\t\t\t\tif counter > 1 {\n\t\t\t\t\t\trpmlog.WithFields(logrus.Fields{\n\t\t\t\t\t\t\t\"frequency\": p.frequency,\n\t\t\t\t\t\t\t\"counter\": counter,\n\t\t\t\t\t\t}).Debug(\"rpm plugin oversampling.\")\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\trpmlog.Debug(\"rpm lock watcher closed.\")\n\t\t\t\treturn\n\t\t\t}\n\t\tcase <-ticker.C:\n\t\t\tticker.Stop()\n\t\t\tticker = time.NewTicker(p.frequency)\n\t\t\tif counter > 0 {\n\t\t\t\tdata, err := p.fetchPackageInfo()\n\t\t\t\tif err != nil {\n\t\t\t\t\trpmlog.WithError(err).Error(\"fetching rpm data\")\n\t\t\t\t} else {\n\t\t\t\t\tp.EmitInventory(data, entity.NewFromNameWithoutID(p.Context.EntityKey()))\n\t\t\t\t}\n\t\t\t\tcounter = 0\n\t\t\t}\n\t\t}\n\t}\n}", "func (d *Driver) setFingerprintSuccess() {\n\td.fingerprintLock.Lock()\n\td.fingerprintSuccess = helper.BoolToPtr(true)\n\td.fingerprintLock.Unlock()\n}", "func (*IntSchema) Fingerprint() (*Fingerprint, error) {\n\treturn &Fingerprint{\n\t\t63, 43, 135, 169, 254, 124, 201, 177, 56, 53, 89, 140, 57, 129, 205, 69,\n\t\t227, 227, 85, 48, 158, 80, 144, 170, 9, 51, 215, 190, 203, 111, 186, 69,\n\t}, nil\n}", "func (fn genericResourceInformerFunc) Start(stopCh <-chan struct{}) {}", "func (s *EnumSchema) Fingerprint() (*Fingerprint, error) {\n\tif s.fingerprint == nil {\n\t\tif f, err := calculateSchemaFingerprint(s); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\ts.fingerprint = f\n\t\t}\n\t}\n\treturn s.fingerprint, nil\n}", "func (*BytesSchema) Fingerprint() (*Fingerprint, error) {\n\treturn &Fingerprint{\n\t\t154, 229, 7, 169, 221, 57, 238, 91, 124, 126, 40, 93, 162, 192, 132, 101,\n\t\t33, 200, 174, 141, 128, 254, 234, 229, 80, 78, 12, 152, 29, 83, 245, 250,\n\t}, nil\n}", "func (n *Notifier) Start() error {\n\tgocron.Every(1).Day().At(n.Config.ReportTime).Do(n.RevealRooks)\n\tgocron.Every(60).Seconds().Do(n.NotifyChannels)\n\tchannel := gocron.Start()\n\tfor {\n\t\treport := <-channel\n\t\tlogrus.Println(report)\n\t}\n}", "func sendStartupEvent() {\n\t// grace wait time for plugin to be functional\n\ttime.Sleep(3 * time.Second)\n\n\tvar pluginIP string\n\tif pluginIP = os.Getenv(\"ASSIGNED_POD_IP\"); pluginIP == \"\" {\n\t\tpluginIP = config.Data.PluginConf.Host\n\t}\n\n\tstartupEvt := common.PluginStatusEvent{\n\t\tName: \"Plugin startup event\",\n\t\tType: \"PluginStarted\",\n\t\tTimestamp: time.Now().String(),\n\t\tOriginatorID: pluginIP,\n\t}\n\n\trequest, _ := json.Marshal(startupEvt)\n\tevent := common.Events{\n\t\tIP: net.JoinHostPort(config.Data.PluginConf.Host, config.Data.PluginConf.Port),\n\t\tRequest: request,\n\t\tEventType: \"PluginStartUp\",\n\t}\n\n\tdone := make(chan bool)\n\tevents := []interface{}{event}\n\tgo common.RunWriteWorkers(caphandler.In, events, 1, done)\n\tlog.Info(\"successfully sent startup event\")\n}", "func (f *Fingerprinter) FingerprintClientHello(data []byte) (clientHelloSpec *ClientHelloSpec, err error) {\n\treturn f.RawClientHello(data)\n}", "func (c *AccountsContainersVersionsPublishCall) Fingerprint(fingerprint string) *AccountsContainersVersionsPublishCall {\n\tc.params_.Set(\"fingerprint\", fmt.Sprintf(\"%v\", fingerprint))\n\treturn c\n}", "func (*DoubleSchema) Fingerprint() (*Fingerprint, error) {\n\treturn &Fingerprint{\n\t\t115, 10, 154, 140, 97, 22, 129, 215, 238, 244, 66, 224, 60, 22, 199, 13,\n\t\t19, 188, 163, 235, 139, 151, 123, 180, 3, 234, 255, 82, 23, 106, 242, 84,\n\t}, nil\n}", "func (s *SimpleDriver) Initialize(svc *device.Service, lc logger.LoggingClient, asyncCh <-chan *device.CommandResult) error {\n\ts.lc = lc\n\ts.lc.Debug(fmt.Sprintf(\"SimpleHandler.Initialize called!\"))\n\treturn nil\n}", "func main() {\n\t// Local (127.0.0.1) hardcoded IPs to simplify testing.\n\tlocalIpPort := \"127.0.0.1:8080\"\n\tvar lostMsgThresh uint8 = 3\n\n\t// TODO: generate a new random epoch nonce on each run\n\tvar epochNonce uint64 = 12345\n\tvar chCapacity uint8 = 5\n\n\n\tfd, notifyCh, err := fdlib.Initialize(epochNonce, chCapacity)\n\tif common.CheckError(err) != nil {\n\t\treturn\n\t}\n\n\t// Stop monitoring and stop responding on exit.\n\t// Defers are really cool, check out: https://blog.golang.org/defer-panic-and-recover\n\tdefer fd.StopMonitoring()\n\t//defer fd.StopResponding()\n\n\terr = fd.StartResponding(localIpPort)\n\tif common.CheckError(err) != nil {\n\t\treturn\n\t}\n\n\tfmt.Println(\"Started responding to heartbeats.\")\n\n\t// Add a monitor for a remote node.\n\tlocalIpPortMon := \"127.0.0.1:9001\"\n\terr = fd.AddMonitor(localIpPortMon, localIpPort, lostMsgThresh)\n\tif common.CheckError(err) != nil {\n\t\treturn\n\t}\n\n\tfmt.Println(\"Started to monitor node: \", localIpPort)\n\n\t// Wait indefinitely, blocking on the notify channel, to detect a\n\t// failure.\n\tstopRespondingTicker := time.NewTicker(5*time.Second)\n\tfor {\n\t\tselect {\n\t\tcase notify := <-notifyCh:\n\t\t\tfmt.Println(\"Success - Detected a failure of\", notify)\n\t\t\treturn\n\t\tcase <-stopRespondingTicker.C:\n\t\t\tstopRespondingTicker.Stop()\n\t\t\tfd.StopResponding()\n\t\t}\n\t}\n\n}", "func (c *AccountsContainersRulesUpdateCall) Fingerprint(fingerprint string) *AccountsContainersRulesUpdateCall {\n\tc.params_.Set(\"fingerprint\", fmt.Sprintf(\"%v\", fingerprint))\n\treturn c\n}", "func HandleCertFingerprintRequest(w http.ResponseWriter, req *http.Request) {\n\t//logger.Info.Printf(\"Cert Hash\")\n\tfingerprint, err := GetCertificateFingerprint()\n\tif err != nil {\n\t\tlogger.Error.Printf(\"%s\\n\", err.Error())\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\tw.WriteHeader(http.StatusOK)\n\tw.Write(fingerprint)\n}", "func (s *RecordSchema) Fingerprint() (*Fingerprint, error) {\n\tif s.fingerprint == nil {\n\t\tif f, err := calculateSchemaFingerprint(s); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\ts.fingerprint = f\n\t\t}\n\t}\n\treturn s.fingerprint, nil\n}", "func (u *processorImpl) Start(ctx context.Context) error {\n\tu.resourceChan = make(chan *ObjectResourceMessage)\n\tu.meterDefStore.RegisterListener(u.name, u.resourceChan)\n\n\tvar wg sync.WaitGroup\n\n\twg.Add(u.digestersSize)\n\n\tfor i := 0; i < u.digestersSize; i++ {\n\t\tgo func() {\n\t\t\tfor data := range u.resourceChan {\n\t\t\t\terr := utils.Retry(\n\t\t\t\t\tfunc() error {\n\t\t\t\t\t\treturn u.processor.Process(ctx, data)\n\t\t\t\t\t}, u.retryCount)\n\t\t\t\tif err != nil {\n\t\t\t\t\tu.log.Error(err, \"error processing message\")\n\t\t\t\t}\n\t\t\t}\n\t\t\twg.Done()\n\t\t}()\n\t}\n\n\t<-ctx.Done()\n\tclose(u.resourceChan)\n\n\twg.Wait()\n\treturn nil\n}", "func (*BooleanSchema) Fingerprint() (*Fingerprint, error) {\n\treturn &Fingerprint{\n\t\t165, 176, 49, 171, 98, 188, 65, 109, 114, 12, 4, 16, 216, 2, 234, 70,\n\t\t185, 16, 196, 251, 232, 92, 80, 169, 70, 204, 198, 88, 183, 78, 103, 126,\n\t}, nil\n}", "func (p *fleetPlugin) Start(config service.ServiceConfig, trigger chan string) error {\n\tif err := util.SetLogLevel(p.FleetLogLevel, config.LogLevel, logName); err != nil {\n\t\treturn maskAny(err)\n\t}\n\tp.nodeExporterPort = config.NodeExporterPort\n\t// No custom triggers here, just update once in a while.\n\treturn nil\n}", "func sendStartupEvent() {\n\t// grace wait time for plugin to be functional\n\ttime.Sleep(3 * time.Second)\n\n\tvar pluginIP string\n\tif pluginIP = os.Getenv(\"ASSIGNED_POD_IP\"); pluginIP == \"\" {\n\t\tpluginIP = config.Data.PluginConf.Host\n\t}\n\n\tstartupEvt := common.PluginStatusEvent{\n\t\tName: \"Plugin startup event\",\n\t\tType: \"PluginStarted\",\n\t\tTimestamp: time.Now().String(),\n\t\tOriginatorID: pluginIP,\n\t}\n\n\trequest, _ := json.Marshal(startupEvt)\n\tevent := common.Events{\n\t\tIP: net.JoinHostPort(config.Data.PluginConf.Host, config.Data.PluginConf.Port),\n\t\tRequest: request,\n\t\tEventType: \"PluginStartUp\",\n\t}\n\n\tdone := make(chan bool)\n\tevents := []interface{}{event}\n\tgo common.RunWriteWorkers(dphandler.In, events, 1, done)\n\tlogs.Log.Info(\"successfully sent startup event\")\n\n}", "func (i *instanceManager) run() {\n\t// Dispense once to ensure we are given a valid plugin\n\tif _, err := i.dispense(); err != nil {\n\t\ti.logger.Error(\"dispensing initial plugin failed\", \"error\", err)\n\t\treturn\n\t}\n\n\t// Create a waitgroup to block on shutdown for all created goroutines to\n\t// exit\n\tvar wg sync.WaitGroup\n\n\t// Start the fingerprinter\n\twg.Add(1)\n\tgo func() {\n\t\ti.fingerprint()\n\t\twg.Done()\n\t}()\n\n\t// Start event handler\n\twg.Add(1)\n\tgo func() {\n\t\ti.handleEvents()\n\t\twg.Done()\n\t}()\n\n\t// Do a final cleanup\n\twg.Wait()\n\ti.cleanup()\n}", "func (f *Ferry) Run() {\n\tf.logger.Info(\"starting ferry run\")\n\tf.OverallState.Store(StateCopying)\n\n\tif f.Config.EnablePProf {\n\t\tgo func() {\n\t\t\terr := http.ListenAndServe(\"localhost:6060\", nil)\n\t\t\tif err != nil {\n\t\t\t\tf.logger.WithError(err).Warn(\"pprof server finished\")\n\t\t\t}\n\t\t}()\n\t}\n\n\tctx, shutdown := context.WithCancel(context.Background())\n\n\thandleError := func(name string, err error) {\n\t\tif err != nil && err != context.Canceled {\n\t\t\tf.ErrorHandler.Fatal(name, err)\n\t\t}\n\t}\n\n\tsupportingServicesWg := &sync.WaitGroup{}\n\tsupportingServicesWg.Add(1)\n\n\tgo func() {\n\t\tdefer supportingServicesWg.Done()\n\t\thandleError(\"throttler\", f.Throttler.Run(ctx))\n\t}()\n\n\tif f.Config.ControlServerConfig.Enabled {\n\t\tgo f.ControlServer.Run()\n\t}\n\n\tif f.Config.ProgressCallback.URI != \"\" {\n\t\tsupportingServicesWg.Add(1)\n\t\tgo func() {\n\t\t\tdefer supportingServicesWg.Done()\n\n\t\t\tfrequency := time.Duration(f.Config.ProgressReportFrequency) * time.Millisecond\n\n\t\t\tfor {\n\t\t\t\tselect {\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\treturn\n\t\t\t\tcase <-time.After(frequency):\n\t\t\t\t\tf.ReportProgress()\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\t}\n\n\tif f.Config.StateCallback.URI != \"\" {\n\t\tsupportingServicesWg.Add(1)\n\t\tgo func() {\n\t\t\tdefer supportingServicesWg.Done()\n\n\t\t\tfrequency := time.Duration(f.Config.StateReportFrequency) * time.Millisecond\n\n\t\t\tfor {\n\t\t\t\tselect {\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\treturn\n\t\t\t\tcase <-time.After(frequency):\n\t\t\t\t\tf.ReportState()\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\t}\n\n\tif f.DumpStateOnSignal {\n\t\tgo func() {\n\t\t\tc := make(chan os.Signal, 1)\n\t\t\tsignal.Notify(c, syscall.SIGINT, syscall.SIGTERM)\n\n\t\t\ts := <-c\n\t\t\tif ctx.Err() == nil {\n\t\t\t\t// Ghostferry is still running\n\t\t\t\tif f.OverallState.Load() != StateCutover {\n\t\t\t\t\t// Save state dump and exit if not during the cutover stage\n\t\t\t\t\tf.ErrorHandler.Fatal(\"user_interrupt\", fmt.Errorf(\"signal received: %v\", s.String()))\n\t\t\t\t} else {\n\t\t\t\t\t// Log and ignore the signal during cutover\n\t\t\t\t\tf.logger.Warnf(\"Received signal: %s during cutover. \"+\n\t\t\t\t\t\t\"Refusing to interrupt and will attempt to complete the run.\", s.String())\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t// shutdown() has been called and Ghostferry is done.\n\t\t\t\tos.Exit(0)\n\t\t\t}\n\t\t}()\n\t}\n\n\tinlineVerifierWg := &sync.WaitGroup{}\n\tinlineVerifierContext, stopInlineVerifier := context.WithCancel(ctx)\n\tif f.inlineVerifier != nil {\n\t\tinlineVerifierWg.Add(1)\n\t\tgo func() {\n\t\t\tdefer inlineVerifierWg.Done()\n\t\t\tf.inlineVerifier.PeriodicallyVerifyBinlogEvents(inlineVerifierContext)\n\t\t}()\n\t}\n\n\tbinlogWg := &sync.WaitGroup{}\n\tbinlogWg.Add(1)\n\tgo func() {\n\t\tdefer binlogWg.Done()\n\t\tf.BinlogWriter.Run()\n\t}()\n\n\tbinlogWg.Add(1)\n\tgo func() {\n\t\tdefer binlogWg.Done()\n\n\t\tf.BinlogStreamer.Run()\n\t\tf.BinlogWriter.Stop()\n\t}()\n\n\tif !f.Config.SkipTargetVerification {\n\t\tf.targetVerifierWg = &sync.WaitGroup{}\n\t\tf.targetVerifierWg.Add(1)\n\t\tgo func() {\n\t\t\tdefer f.targetVerifierWg.Done()\n\t\t\tf.TargetVerifier.BinlogStreamer.Run()\n\t\t}()\n\t}\n\n\tdataIteratorWg := &sync.WaitGroup{}\n\tdataIteratorWg.Add(1)\n\tgo func() {\n\t\tdefer dataIteratorWg.Done()\n\t\tf.DataIterator.Run(f.Tables.AsSlice())\n\t}()\n\n\tdataIteratorWg.Wait()\n\n\tf.logger.Info(\"data copy is complete, waiting for cutover\")\n\tf.OverallState.Store(StateWaitingForCutover)\n\tf.waitUntilAutomaticCutoverIsTrue()\n\n\tif f.inlineVerifier != nil {\n\t\t// Stops the periodic verification of binlogs in the inline verifier\n\t\t// This should be okay as we enqueue the binlog events into the verifier,\n\t\t// which will be verified both in VerifyBeforeCutover and\n\t\t// VerifyDuringCutover.\n\t\tstopInlineVerifier()\n\t\tinlineVerifierWg.Wait()\n\t}\n\n\tif f.Verifier != nil {\n\t\tf.logger.Info(\"calling VerifyBeforeCutover\")\n\t\tf.OverallState.Store(StateVerifyBeforeCutover)\n\n\t\tmetrics.Measure(\"VerifyBeforeCutover\", nil, 1.0, func() {\n\t\t\terr := f.Verifier.VerifyBeforeCutover()\n\t\t\tif err != nil {\n\t\t\t\tf.logger.WithError(err).Error(\"VerifyBeforeCutover failed\")\n\t\t\t\tf.ErrorHandler.Fatal(\"verifier\", err)\n\t\t\t}\n\t\t})\n\t}\n\n\t// Cutover is a cooperative activity between the Ghostferry library and\n\t// applications built on Ghostferry:\n\t//\n\t// 1. At this point (before stopping the binlog streaming), the application\n\t// should prepare to cutover, such as setting the source database to\n\t// READONLY.\n\t// 2. Once that is complete, trigger the cutover by requesting the\n\t// BinlogStreamer to stop (WaitUntilBinlogStreamerCatchesUp and\n\t// FlushBinlogAndStopStreaming).\n\t// 3. Once the binlog stops, this Run function will return and the cutover\n\t// will be completed. Application and human operators can perform\n\t// additional operations, such as additional verification, and repointing\n\t// any consumers of the source database to use the target database.\n\t//\n\t// During cutover, if verifiers are enabled, VerifyDuringCutover should be\n\t// called. This can be performed as a part of the ControlServer, if that\n\t// component is used.\n\n\tf.logger.Info(\"entering cutover phase, notifying caller that row copy is complete\")\n\tf.OverallState.Store(StateCutover)\n\tif f.Config.ProgressCallback.URI != \"\" {\n\t\tf.ReportProgress()\n\t}\n\tf.notifyRowCopyComplete()\n\n\tbinlogWg.Wait()\n\n\tf.logger.Info(\"ghostferry run is complete, shutting down auxiliary services\")\n\n\tf.OverallState.Store(StateDone)\n\tf.DoneTime = time.Now()\n\n\tshutdown()\n\tsupportingServicesWg.Wait()\n\n\tif f.Config.ProgressCallback.URI != \"\" {\n\t\tf.ReportProgress()\n\t}\n}", "func (s *refSchema) Fingerprint() (*Fingerprint, error) {\n\treturn s.Fingerprint()\n}", "func NewFingerprint(devices ...*DeviceGroup) *FingerprintResponse {\n\treturn &FingerprintResponse{\n\t\tDevices: devices,\n\t}\n}", "func (p *APIPingProbe) Start() error {\n\tif p.Active {\n\t\treturn nil\n\t}\n\n\tt := time.Duration(p.Config.Frequency * float64(time.Second))\n\twriteTicker := time.NewTicker(t)\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-p.StopChannel:\n\t\t\t\treturn\n\t\t\tcase <-writeTicker.C:\n\t\t\t\tgo p.PingAPI()\n\t\t\t}\n\t\t}\n\t}()\n\n\tp.Active = true\n\treturn nil\n}", "func (s *EBPFSocketInfoEnhancer) Start() error {\n\ts.tracer.Start()\n\treturn s.scanProc()\n}", "func (c *AccountsContainersVariablesUpdateCall) Fingerprint(fingerprint string) *AccountsContainersVariablesUpdateCall {\n\tc.params_.Set(\"fingerprint\", fmt.Sprintf(\"%v\", fingerprint))\n\treturn c\n}", "func (p *SearchProbe) Start() error {\n\tif p.Active {\n\t\treturn nil\n\t}\n\n\tt := time.Duration(p.Config.Frequency * float64(time.Second))\n\twriteTicker := time.NewTicker(t)\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-p.StopChannel:\n\t\t\t\treturn\n\t\t\tcase <-writeTicker.C:\n\t\t\t\tif p.Config.ChannelEnabled {\n\t\t\t\t\tgo p.SearchChannels()\n\t\t\t\t}\n\t\t\t\tif p.Config.UserEnabled {\n\t\t\t\t\tgo p.SearchUsers()\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\tp.Active = true\n\treturn nil\n}", "func (p *ProbeHandler) Start() error {\n\t// start a goroutine in order to update the graph\n\tgo func() {\n\t\t// update the graph each five seconds\n\t\tticker := time.NewTicker(5 * time.Second)\n\t\tdefer ticker.Stop()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-p.quit:\n\t\t\t\t// got a message on the quit chan\n\t\t\t\treturn\n\t\t\tcase <-ticker.C:\n\t\t\t\tv, err := mem.VirtualMemory()\n\t\t\t\tif err != nil {\n\t\t\t\t\tp.Ctx.Logger.Errorf(\"unable to retrieve memory information: %s\", err)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tusage := &Usage{\n\t\t\t\t\tTotal: int64(v.Total),\n\t\t\t\t\tFree: int64(v.Free),\n\t\t\t\t\tUsedPercent: int64(v.UsedPercent),\n\t\t\t\t}\n\n\t\t\t\t// lock the graph for modification\n\t\t\t\tp.Ctx.Graph.Lock()\n\n\t\t\t\t// add metadata entry to the root node\n\t\t\t\tp.Ctx.Graph.AddMetadata(p.Ctx.RootNode, \"Memory\", usage)\n\n\t\t\t\t// release the graph lock\n\t\t\t\tp.Ctx.Graph.Unlock()\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn nil\n}", "func fixFingers() {\n\tfixfingersTicker := time.NewTicker(time.Duration(fixfingersInterval) *\n\t\ttime.Second)\n\tfixfingersChan = make(chan struct{})\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-fixfingersTicker.C:\n\t\t\t\tnext = next + 1\n\t\t\t\tif next > 5 {\n\t\t\t\t\tnext = 1\n\t\t\t\t}\n\t\t\t\tfingertable[next] = findSuccessorFT((chordid +\n\t\t\t\t\tint(math.Pow(2, float64(next-1)))) % 32)\n\t\t\t\t//fmt.Println(\"finger \", next, fingertable[next])\n\t\t\tcase <-fixfingersChan:\n\t\t\t\tfixfingersTicker.Stop()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n}", "func (ks *Kind) Start(ctx context.Context, handler handler.EventHandler, queue workqueue.RateLimitingInterface,\n\tprct ...predicate.Predicate) error {\n\t// Type should have been specified by the user.\n\tif ks.Type == nil {\n\t\treturn fmt.Errorf(\"must specify Kind.Type\")\n\t}\n\n\t// cache should have been injected before Start was called\n\tif ks.cache == nil {\n\t\treturn fmt.Errorf(\"must call CacheInto on Kind before calling Start\")\n\t}\n\n\t// cache.GetInformer will block until its context is cancelled if the cache was already started and it can not\n\t// sync that informer (most commonly due to RBAC issues).\n\tctx, ks.startCancel = context.WithCancel(ctx)\n\tks.started = make(chan error)\n\tgo func() {\n\t\tvar (\n\t\t\ti cache.Informer\n\t\t\tlastErr error\n\t\t)\n\n\t\t// Tries to get an informer until it returns true,\n\t\t// an error or the specified context is cancelled or expired.\n\t\tif err := wait.PollImmediateUntilWithContext(ctx, 10*time.Second, func(ctx context.Context) (bool, error) {\n\t\t\t// Lookup the Informer from the Cache and add an EventHandler which populates the Queue\n\t\t\ti, lastErr = ks.cache.GetInformer(ctx, ks.Type)\n\t\t\tif lastErr != nil {\n\t\t\t\tkindMatchErr := &meta.NoKindMatchError{}\n\t\t\t\tswitch {\n\t\t\t\tcase errors.As(lastErr, &kindMatchErr):\n\t\t\t\t\tlog.Error(lastErr, \"if kind is a CRD, it should be installed before calling Start\",\n\t\t\t\t\t\t\"kind\", kindMatchErr.GroupKind)\n\t\t\t\tcase runtime.IsNotRegisteredError(lastErr):\n\t\t\t\t\tlog.Error(lastErr, \"kind must be registered to the Scheme\")\n\t\t\t\tdefault:\n\t\t\t\t\tlog.Error(lastErr, \"failed to get informer from cache\")\n\t\t\t\t}\n\t\t\t\treturn false, nil // Retry.\n\t\t\t}\n\t\t\treturn true, nil\n\t\t}); err != nil {\n\t\t\tif lastErr != nil {\n\t\t\t\tks.started <- fmt.Errorf(\"failed to get informer from cache: %w\", lastErr)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tks.started <- err\n\t\t\treturn\n\t\t}\n\n\t\t_, err := i.AddEventHandler(internal.EventHandler{Queue: queue, EventHandler: handler, Predicates: prct})\n\t\tif err != nil {\n\t\t\tks.started <- err\n\t\t\treturn\n\t\t}\n\t\tif !ks.cache.WaitForCacheSync(ctx) {\n\t\t\t// Would be great to return something more informative here\n\t\t\tks.started <- errors.New(\"cache did not sync\")\n\t\t}\n\t\tclose(ks.started)\n\t}()\n\n\treturn nil\n}", "func (f *Function) Serve(ctx context.Context, fID, imageName, reqPayload string) (*hpb.FwdHelloResp, *metrics.Metric, error) {\n\tvar (\n\t\tserveMetric *metrics.Metric = metrics.NewMetric()\n\t\ttStart time.Time\n\t\tsyncID int64 = -1 // default is no synchronization\n\t\tisColdStart bool = false\n\t)\n\n\tlogger := log.WithFields(log.Fields{\"fID\": f.fID})\n\n\tif !f.isPinnedInMem {\n\t\tif err := f.sem.Acquire(context.Background(), 1); err != nil {\n\t\t\tlogger.Panic(\"Failed to acquire semaphore for serving\")\n\t\t}\n\n\t\tsyncID = atomic.AddInt64(&f.servedSyncCounter, -1) // unique number for goroutines acquiring the semaphore\n\t}\n\n\tf.stats.IncServed(f.fID)\n\n\tf.OnceAddInstance.Do(\n\t\tfunc() {\n\t\t\tvar metr *metrics.Metric\n\t\t\tisColdStart = true\n\t\t\tlogger.Debug(\"Function is inactive, starting the instance...\")\n\t\t\ttStart = time.Now()\n\t\t\tmetr = f.AddInstance()\n\t\t\tserveMetric.MetricMap[metrics.AddInstance] = metrics.ToUS(time.Since(tStart))\n\n\t\t\tif metr != nil {\n\t\t\t\tfor k, v := range metr.MetricMap {\n\t\t\t\t\tserveMetric.MetricMap[k] = v\n\t\t\t\t}\n\t\t\t}\n\t\t})\n\n\tf.RLock()\n\n\t// FIXME: keep a strict deadline for forwarding RPCs to a warm function\n\t// Eventually, it needs to be RPC-dependent and probably client-defined\n\tctxFwd, cancel := context.WithDeadline(context.Background(), time.Now().Add(20*time.Second))\n\tdefer cancel()\n\n\ttStart = time.Now()\n\tresp, err := f.fwdRPC(ctxFwd, reqPayload)\n\tserveMetric.MetricMap[metrics.FuncInvocation] = metrics.ToUS(time.Since(tStart))\n\n\tif err != nil && ctxFwd.Err() == context.Canceled {\n\t\t// context deadline exceeded\n\t\tf.RUnlock()\n\t\treturn &hpb.FwdHelloResp{IsColdStart: isColdStart, Payload: \"\"}, serveMetric, err\n\t} else if err != nil {\n\t\tif e, ok := status.FromError(err); ok {\n\t\t\tswitch e.Code() {\n\t\t\tcase codes.DeadlineExceeded:\n\t\t\t\t// deadline exceeded\n\t\t\t\tf.RUnlock()\n\t\t\t\treturn &hpb.FwdHelloResp{IsColdStart: isColdStart, Payload: \"\"}, serveMetric, err\n\t\t\tdefault:\n\t\t\t\tlogger.Warn(\"Function returned error: \", err)\n\t\t\t\tf.RUnlock()\n\t\t\t\treturn &hpb.FwdHelloResp{IsColdStart: isColdStart, Payload: \"\"}, serveMetric, err\n\t\t\t}\n\t\t} else {\n\t\t\tlogger.Panic(\"Not able to parse error returned \", err)\n\t\t}\n\t}\n\n\tif orch.GetSnapshotsEnabled() {\n\t\tf.OnceCreateSnapInstance.Do(\n\t\t\tfunc() {\n\t\t\t\tlogger.Debug(\"First time offloading, need to create a snapshot first\")\n\t\t\t\tf.CreateInstanceSnapshot()\n\t\t\t\tf.isSnapshotReady = true\n\t\t\t})\n\t}\n\n\tf.RUnlock()\n\n\tif !f.isPinnedInMem && syncID == 0 {\n\t\tlogger.Debugf(\"Function has to shut down its instance, served %d requests\", f.GetStatServed())\n\t\ttStart = time.Now()\n\t\tif _, err := f.RemoveInstance(false); err != nil {\n\t\t\tlogger.Panic(\"Failed to remove instance after servedTh expired\", err)\n\t\t}\n\t\tserveMetric.MetricMap[metrics.RetireOld] = metrics.ToUS(time.Since(tStart))\n\t\tf.ZeroServedStat()\n\t\tf.servedSyncCounter = int64(f.servedTh) // reset counter\n\t\tf.sem.Release(int64(f.servedTh))\n\t}\n\n\treturn &hpb.FwdHelloResp{IsColdStart: isColdStart, Payload: resp.Message}, serveMetric, err\n}", "func (c *client) Run() {\n\tstream, err := c.client.ListAndWatch(context.Background(), &api.Empty{})\n\tif err != nil {\n\t\tklog.ErrorS(err, \"ListAndWatch ended unexpectedly for device plugin\", \"resource\", c.resource)\n\t\treturn\n\t}\n\n\tfor {\n\t\tresponse, err := stream.Recv()\n\t\tif err != nil {\n\t\t\tklog.ErrorS(err, \"ListAndWatch ended unexpectedly for device plugin\", \"resource\", c.resource)\n\t\t\treturn\n\t\t}\n\t\tklog.V(2).InfoS(\"State pushed for device plugin\", \"resource\", c.resource, \"resourceCapacity\", len(response.Devices))\n\t\tc.handler.PluginListAndWatchReceiver(c.resource, response)\n\t}\n}", "func (ht *HeadTracker) Start() error {\n\tht.headMutex.Lock()\n\tdefer ht.headMutex.Unlock()\n\n\tif ht.started {\n\t\treturn nil\n\t}\n\n\tif err := ht.updateHeadFromDb(); err != nil {\n\t\treturn err\n\t}\n\tnumber := ht.head\n\tif number != nil {\n\t\tlogger.Debug(\"Tracking logs from last block \", presenters.FriendlyBigInt(number.ToInt()), \" with hash \", number.Hash.Hex())\n\t}\n\n\tht.done = make(chan struct{})\n\tht.subscriptionSucceeded = make(chan struct{})\n\n\tht.listenForNewHeadsWg.Add(1)\n\tgo ht.listenForNewHeads()\n\n\tht.started = true\n\treturn nil\n}", "func (c *AccountsContainersTagsUpdateCall) Fingerprint(fingerprint string) *AccountsContainersTagsUpdateCall {\n\tc.params_.Set(\"fingerprint\", fmt.Sprintf(\"%v\", fingerprint))\n\treturn c\n}", "func (d *KVS) Initialize(localTracer *tracing.Tracer, clientId string, frontEndAddr string, chCapacity uint) (NotifyChannel, error) {\n\tvar trce *tracing.Trace\n\tif localTracer != nil {\n\t\ttrce = localTracer.CreateTrace()\n\t}\n\n\td.ltrace = trce\n\td.ltrace.RecordAction(KvslibBegin{ClientId: clientId})\n\t//connect to front-end\n\n\tvar wg sync.WaitGroup\n\td.wg = &wg\n\n\td.notifyCh = make(NotifyChannel, chCapacity)\n\td.clientId = clientId\n\td.mapKeys = make(map[string]*sync.Mutex)\n\td.mapProgress = make(map[string]chan uint32)\n\t//d.mapTasks = make(map[string]chan bool)\n\td.opId = 0\n\t//make the new channel\n\tfrontend, err := rpc.Dial(\"tcp\", frontEndAddr)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error dialing front-end: %s\", err)\n\t}\n\td.frontEnd = frontend\n\treturn d.notifyCh, nil\n}", "func (o PartnerAccountOutput) Fingerprint() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *PartnerAccount) pulumi.StringOutput { return v.Fingerprint }).(pulumi.StringOutput)\n}", "func (s *RecursiveSchema) Fingerprint() (*Fingerprint, error) {\n\treturn s.Actual.Fingerprint()\n}", "func (m *manager) Run() {\n\t// Check if there are any plugins that didn't get cleanly shutdown before\n\t// and if there are shut them down.\n\tm.cleanupStalePlugins()\n\n\t// Get device plugins\n\tdevices := m.loader.Catalog()[base.PluginTypeDevice]\n\tif len(devices) == 0 {\n\t\tm.logger.Debug(\"exiting since there are no device plugins\")\n\t\tm.cancel()\n\t\treturn\n\t}\n\n\tfor _, d := range devices {\n\t\tid := loader.PluginInfoID(d)\n\t\tstoreFn := func(c *plugin.ReattachConfig) error {\n\t\t\tid := id\n\t\t\treturn m.storePluginReattachConfig(id, c)\n\t\t}\n\t\tm.instances[id] = newInstanceManager(&instanceManagerConfig{\n\t\t\tLogger: m.logger,\n\t\t\tCtx: m.ctx,\n\t\t\tLoader: m.loader,\n\t\t\tStoreReattach: storeFn,\n\t\t\tPluginConfig: m.pluginConfig,\n\t\t\tId: &id,\n\t\t\tFingerprintOutCh: m.fingerprintResCh,\n\t\t\tStatsInterval: m.statsInterval,\n\t\t})\n\t}\n\n\t// Now start the fingerprint handler\n\tgo m.fingerprint()\n}", "func (r *Registration) Start(ctx context.Context) {\n\tif r.apiKey != \"\" {\n\t\tctx = metadata.AppendToOutgoingContext(ctx, auth.APIKeyHeader, r.apiKey)\n\t}\n\n\tgo func() {\n\t\tr.maintainRegistrationAndStreamWork(ctx)\n\t}()\n}", "func (o NetworkInterfaceResponseOutput) Fingerprint() pulumi.StringOutput {\n\treturn o.ApplyT(func(v NetworkInterfaceResponse) string { return v.Fingerprint }).(pulumi.StringOutput)\n}", "func (s *FixedSchema) Fingerprint() (*Fingerprint, error) {\n\tif s.fingerprint == nil {\n\t\tif f, err := calculateSchemaFingerprint(s); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\ts.fingerprint = f\n\t\t}\n\t}\n\treturn s.fingerprint, nil\n}", "func (rf *Raft) HeartbeatThread() {\n\tfor {\n\t\trf.lock()\n\t\tif rf.CurrentElectionState != Leader {\n\t\t\tad.DebugObj(rf, ad.TRACE, \"HeartbeatThread waiting until is leader\")\n\t\t\trf.unlock()\n\n\t\t\t// blocking read\n\t\twaitForBecomeLeader:\n\t\t\tterm := <-rf.becomeLeader\n\n\t\t\trf.lock()\n\t\t\tif term < rf.CurrentTerm {\n\t\t\t\t// I became leader in a previous term but then advanced to my current term before I\n\t\t\t\t// noticed I became a leader, so instead I should become a follower.\n\t\t\t\tad.DebugObj(rf, ad.WARN, \"Just noticed that I won election in term %d, but it's now term %d, so I'll stay a follower\",\n\t\t\t\t\tterm, rf.CurrentTerm)\n\t\t\t\tassert(rf.CurrentElectionState != Leader)\n\t\t\t\trf.unlock()\n\t\t\t\tgoto waitForBecomeLeader\n\t\t\t}\n\n\t\t\t// term > rf.CurrentTerm wouldn't make any sense\n\t\t\tassert(term == rf.CurrentTerm)\n\t\t\tad.DebugObj(rf, ad.RPC, \"Becoming leader\")\n\t\t\trf.CurrentElectionState = Leader\n\t\t\trf.writePersist()\n\t\t\tfor peerNum, _ := range rf.peers {\n\t\t\t\trf.nextIndex[peerNum] = rf.lastLogIndex() + 1\n\t\t\t\trf.matchIndex[peerNum] = 0\n\t\t\t}\n\t\t\trf.matchIndex[rf.me] = rf.Log.length()\n\t\t}\n\n\t\tif !rf.isAlive {\n\t\t\trf.unlock()\n\t\t\treturn\n\t\t}\n\n\t\tad.DebugObj(rf, ad.RPC, \"Sending heartbeats. commitIndex=%+v, nextIndex=%+v, matchIndex=%+v\",\n\t\t\trf.commitIndex, rf.nextIndex, rf.matchIndex)\n\t\tfor peerNum, _ := range rf.peers {\n\t\t\tgo rf.sendAppendEntries(peerNum, true)\n\t\t}\n\t\trf.unlock()\n\t\ttime.Sleep(getHeartbeatTimeout())\n\n\t}\n}", "func (o LookupPartnerAccountResultOutput) Fingerprint() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v LookupPartnerAccountResult) *string { return v.Fingerprint }).(pulumi.StringPtrOutput)\n}", "func (*FloatSchema) Fingerprint() (*Fingerprint, error) {\n\treturn &Fingerprint{\n\t\t30, 113, 249, 236, 5, 29, 102, 63, 86, 176, 216, 225, 252, 132, 215, 26,\n\t\t165, 108, 207, 233, 250, 147, 170, 32, 209, 5, 71, 167, 171, 235, 92, 192,\n\t}, nil\n}", "func Fingerprint(certificate []byte) FingerprintBytes {\n\treturn sha256.Sum256(certificate)\n}", "func (rf *Raft) sendInstallSnapshot(peerIdx int) {\n\trf.mu.Lock()\n\targs := InstallSnapshotArgs{\n\t\tTerm: rf.currentTerm,\n\t\tLeaderId: rf.me,\n\t\tLastIncludedIndex: rf.lastSnapshotIndex,\n\t\tLastIncludedTerm: rf.lastSnapshotTerm,\n\t\tData: rf.persister.ReadSnapshot(),\n\t}\n\trf.mu.Unlock()\n\ttimer := time.NewTimer(RPCTimeout)\n\tdefer timer.Stop()\n\n\tfor !rf.killed() {\n\t\ttimer.Stop()\n\t\ttimer.Reset(RPCTimeout)\n\t\tokCh := make(chan bool, 1)\n\t\treply := InstallSnapshotReply{}\n\t\tgo func() {\n\t\t\tr := rf.peers[peerIdx].Call(\"Raft.InstallSnapshot\", &args, &reply)\n\t\t\tif !r {\n\t\t\t\ttime.Sleep(time.Millisecond * 10)\n\t\t\t}\n\t\t\tokCh <- r\n\t\t}()\n\n\t\tok := false\n\t\tselect {\n\t\tcase <-timer.C: // RPC 超时\n\t\t\tcontinue\n\t\tcase ok = <-okCh:\n\t\t\tif !ok { // RPC 失败\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\t// ok == true\n\t\trf.mu.Lock()\n\t\tdefer rf.mu.Unlock()\n\t\tif rf.currentTerm != args.Term || rf.role != Leader {\n\t\t\treturn\n\t\t} else if reply.Term > rf.currentTerm {\n\t\t\trf.changeRole(Follower)\n\t\t\trf.resetElectionTimer()\n\t\t\trf.currentTerm = reply.Term\n\t\t\trf.persist()\n\t\t\treturn\n\t\t}\n\t\t// success\n\t\tif args.LastIncludedIndex > rf.matchIndex[peerIdx] {\n\t\t\trf.matchIndex[peerIdx] = args.LastIncludedIndex\n\t\t}\n\t\tif args.LastIncludedIndex+1 > rf.nextIndex[peerIdx] {\n\t\t\trf.nextIndex[peerIdx] = args.LastIncludedIndex + 1\n\t\t}\n\t\treturn\n\t}\n}", "func (f *fakeDiskUpdateWatchServer) Send(event *vosinternalprotos.DiskUpdate) error {\n\ttestChannel <- event\n\treturn nil\n}", "func (rf *Raft) ticker() {\n\n\ttime.Sleep(time.Duration(1000000 * rand.Intn(150)))\n\n\tfor !rf.killed() {\n\n\t\t// Your code here to check if a leader election should\n\t\t// be started and to randomize sleeping time using\n\t\t// time.Sleep().\n\n\t\trf.mu.Lock()\n\t\tvar sleepTime time.Duration\n\t\tisLeader := rf.IsLeader()\n\t\tif isLeader {\n\t\t\t// rf.checkCommit()\n\t\t\t// rf.debugCommit()\n\t\t\trf.canSend = false\n\t\t\trf.broadcast() // heartbeat all followers\n\t\t\tsleepTime = time.Millisecond * 100 // fixed time, 10 times per second max\n\t\t} else {\n\t\t\tif !rf.gotContacted {\n\t\t\t\t// start election\n\t\t\t\trf.currentTerm += 1\n\t\t\t\trf.votedFor = rf.me\n\t\t\t\tatomic.StoreInt32(&rf.state, CANDIDATE)\n\t\t\t\t// rf.state = CANDIDATE\n\n\t\t\t\t// reset vote count\n\t\t\t\tfor i := range rf.voteCount {\n\t\t\t\t\trf.voteCount[i] = false\n\t\t\t\t}\n\t\t\t\trf.voteCount[rf.me] = true // vote for itself\n\n\t\t\t\trf.debug(\"Failed to be contacted, initiating election in term %v\\n\", rf.currentTerm)\n\n\t\t\t\t// request vote from all others\n\t\t\t\tfor i := range rf.peers {\n\t\t\t\t\tif i != rf.me {\n\t\t\t\t\t\tgo rf.candidateNotify(i, rf.currentTerm, rf.me, rf.lastEntryIndex(), rf.index(rf.lastEntryIndex()).Term)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tsleepTime = time.Duration(1000000 * (100 + rand.Intn(200))) // 100-300ms\n\t\t\t} else {\n\t\t\t\trf.debug(\"Got contacted and will sleep again...\\n\")\n\t\t\t\tsleepTime = time.Duration(1000000 * (600 + rand.Intn(350))) // 600-950ms\n\t\t\t}\n\t\t\trf.gotContacted = false\n\t\t}\n\n\t\trf.mu.Unlock()\n\n\t\ttime.Sleep(sleepTime)\n\t}\n}", "func (o *RevokeDeviceCertificateUsingDELETEParams) SetFingerprint(fingerprint string) {\n\to.Fingerprint = fingerprint\n}", "func (w *XPubWallet) Fingerprint() string {\n\t// Note: the xpub key is not used as the fingerprint, because it is\n\t// partially sensitive data\n\taddr := \"\"\n\tif len(w.Entries) == 0 {\n\t\tif !w.IsEncrypted() {\n\t\t\tentries, err := w.generateEntries(1, 0)\n\t\t\tif err != nil {\n\t\t\t\tlogger.WithError(err).Panic(\"Fingerprint failed to generate initial entry for empty wallet\")\n\t\t\t}\n\t\t\taddr = entries[0].Address.String()\n\t\t}\n\t} else {\n\t\taddr = w.Entries[0].Address.String()\n\t}\n\n\treturn fmt.Sprintf(\"%s-%s\", w.Type(), addr)\n}", "func StartPlugin(impl ComputationImplementation) {\n\tprovider := pie.NewProvider()\n\tcomputationPlugin := &ComputationPlugin{impl, sync.RWMutex{}}\n\tif err := provider.RegisterName(\"Computation\", computationPlugin); err != nil {\n\t\tlog.Fatalf(\"failed to register computation Plugin: %s\", err)\n\t}\n\tprovider.ServeCodec(jsonrpc.NewServerCodec)\n}", "func (p *Probe) Start(ctx context.Context, dataChan chan *metrics.EventMetrics) {\n\tdefer p.wait()\n\n\tp.updateTargetsAndStartProbes(ctx, dataChan)\n\n\t// Do more frequent listing of targets until we get a non-zero list of\n\t// targets.\n\tinitialRefreshInterval := p.opts.Interval\n\t// Don't wait too long if p.opts.Interval is large.\n\tif initialRefreshInterval > time.Second {\n\t\tinitialRefreshInterval = time.Second\n\t}\n\n\tfor {\n\t\tif ctxDone(ctx) {\n\t\t\treturn\n\t\t}\n\t\tif len(p.targets) != 0 {\n\t\t\tbreak\n\t\t}\n\t\tp.updateTargetsAndStartProbes(ctx, dataChan)\n\t\ttime.Sleep(initialRefreshInterval)\n\t}\n\n\ttargetsUpdateTicker := time.NewTicker(p.targetsUpdateInterval)\n\tdefer targetsUpdateTicker.Stop()\n\n\tfor {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn\n\t\tcase <-targetsUpdateTicker.C:\n\t\t\tp.updateOauthToken()\n\t\t\tp.updateTargetsAndStartProbes(ctx, dataChan)\n\t\t}\n\t}\n}", "func (*NullSchema) Fingerprint() (*Fingerprint, error) {\n\treturn &Fingerprint{\n\t\t240, 114, 203, 236, 59, 248, 132, 24, 113, 212, 40, 66, 48, 197, 233, 131,\n\t\t220, 33, 26, 86, 131, 122, 237, 134, 36, 135, 20, 143, 148, 125, 26, 31,\n\t}, nil\n}", "func (p *Probe) Start(ctx context.Context, dataChan chan *metrics.EventMetrics) {\n\tif p.conn == nil {\n\t\tp.l.Critical(\"Probe has not been properly initialized yet.\")\n\t}\n\tdefer p.conn.close()\n\tfor ts := range time.Tick(p.opts.Interval) {\n\t\t// Don't run another probe if context is canceled already.\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn\n\t\tdefault:\n\t\t}\n\n\t\tp.runProbe()\n\t\tp.l.Debugf(\"%s: Probe finished.\", p.name)\n\t\tif (p.runCnt % uint64(p.c.GetStatsExportInterval())) != 0 {\n\t\t\tcontinue\n\t\t}\n\t\tfor _, t := range p.targets {\n\t\t\tem := metrics.NewEventMetrics(ts).\n\t\t\t\tAddMetric(\"total\", metrics.NewInt(p.sent[t])).\n\t\t\t\tAddMetric(\"success\", metrics.NewInt(p.received[t])).\n\t\t\t\tAddMetric(\"latency\", metrics.NewFloat(p.latency[t].Seconds()/p.opts.LatencyUnit.Seconds())).\n\t\t\t\tAddLabel(\"ptype\", \"ping\").\n\t\t\t\tAddLabel(\"probe\", p.name).\n\t\t\t\tAddLabel(\"dst\", t)\n\n\t\t\tdataChan <- em.Clone()\n\t\t\tp.l.Info(em.String())\n\t\t}\n\t}\n}", "func (m *RdmaDevPlugin) Start() error {\n\terr := m.cleanup()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsock, err := net.Listen(\"unix\", m.socketPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tm.server = grpc.NewServer([]grpc.ServerOption{}...)\n\n\tif m.watchMode {\n\t\tregisterapi.RegisterRegistrationServer(m.server, m)\n\t}\n\tpluginapi.RegisterDevicePluginServer(m.server, m)\n\n\tgo m.server.Serve(sock)\n\n\t// Wait for server to start by launching a blocking connexion\n\tconn, err := dial(m.socketPath, 5*time.Second)\n\tif err != nil {\n\t\treturn err\n\t}\n\tconn.Close()\n\n\tif !m.watchMode {\n\t\tif err = m.register(); err != nil {\n\t\t\tm.server.Stop()\n\t\t\tlog.Fatal(err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// go m.healthcheck()\n\n\treturn nil\n}", "func (l *ProgressBarNoTTYLogger) Start(ctx context.Context, progressChan <-chan regv1.Update) {\n\tctx, cancelFunc := context.WithCancel(ctx)\n\tl.cancelFunc = cancelFunc\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn\n\t\t\tcase <-progressChan:\n\t\t\t}\n\t\t}\n\t}()\n}", "func StartHC() {\n\ttfhc.StartHC()\n}", "func Fingerprint(input string) (result string, err error) {\n\treturn parser.FingerprintToHexStr(input)\n}", "func (zd *ZFSSADriver) Probe(ctx context.Context, req *csi.ProbeRequest) (\n\t*csi.ProbeResponse, error) {\n\n\tutils.GetLogIDTY(ctx, 5).Println(\"Probe\")\n\n\t// Check that the appliance is responsive, if it is not, we are on hold\n\tuser, password, err := zd.getUserLogin(ctx, nil)\n\tif err != nil {\n\t\treturn nil, grpcStatus.Error(codes.Unauthenticated, \"Invalid credentials\")\n\t}\n\ttoken := zfssarest.LookUpToken(user, password)\n\t_, err = zfssarest.GetServices(ctx, token)\n\tif err != nil {\n\t\treturn &csi.ProbeResponse{\n\t\t\tReady: &wrappers.BoolValue{Value: false},\n\t\t}, grpcStatus.Error(codes.FailedPrecondition, \"Failure creating token\")\n\t}\n\n\treturn &csi.ProbeResponse{\n\t\tReady: &wrappers.BoolValue{Value: true},\n\t}, nil\n}", "func (s *SimpleDriver) Initialize(lc logger.LoggingClient, asyncCh chan<- *dsModels.AsyncValues) error {\n\tcommand_list = make(map[string][2]string)\n\tcurrent_running_task = make(map[string][]string)\n\tvar opts Options\n\tflags.Parse(&opts)\n\tif opts.ConfProfile == \"docker\" {\n\t\tIPE_addr = \"dockerhost:8700\"\n\t}\n\tfmt.Println(IPE_addr)\n\tconn, _ = net.Dial(\"tcp\",IPE_addr)\n\n\tgo IPEMessageHandler()\n\n\ts.lc = lc\n\ts.asyncCh = asyncCh\n\treturn nil\n}", "func (o TagsResponsePtrOutput) Fingerprint() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TagsResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Fingerprint\n\t}).(pulumi.StringPtrOutput)\n}", "func Register(ctx context.Context, service string, host Host) *Registration {\n\tif !host.Public && len(host.PrivateHostname) == 0 {\n\t\thost.PrivateHostname = host.Hostname\n\t}\n\n\tif len(host.PrivateHostname) == 0 {\n\t\thost.PrivateHostname = hostname\n\t}\n\thost.Name = service\n\n\tif len(host.PrivateHostname) != 0 && len(host.PrivatePorts) == 0 {\n\t\thost.PrivatePorts = host.Ports\n\t}\n\n\tuuidV4, _ := uuid.NewV4()\n\thostUuid := fmt.Sprintf(\"%s-%s\", uuidV4.String(), host.PrivateHostname)\n\thost.UUID = hostUuid\n\n\tserviceInfos := &Service{\n\t\tName: service,\n\t\tCritical: host.Critical,\n\t\tPublic: host.Public,\n\t}\n\n\tif host.Public {\n\t\tserviceInfos.Hostname = host.Hostname\n\t\tserviceInfos.Ports = host.Ports\n\t\tserviceInfos.Password = host.Password\n\t\tserviceInfos.User = host.User\n\t}\n\n\tpublicCredentialsChan := make(chan Credentials, 1) // Communication between register and the client\n\tprivateCredentialsChan := make(chan Credentials, 1) // Communication between watcher and register\n\n\thostKey := fmt.Sprintf(\"/services/%s/%s\", service, hostUuid)\n\thostJson, _ := json.Marshal(&host)\n\thostValue := string(hostJson)\n\n\tserviceKey := fmt.Sprintf(\"/services_infos/%s\", service)\n\tserviceJson, _ := json.Marshal(serviceInfos)\n\tserviceValue := string(serviceJson)\n\n\tgo func() {\n\t\tticker := time.NewTicker((HEARTBEAT_DURATION - 1) * time.Second)\n\n\t\t// id is the current modification index of the service key.\n\t\t// this is used for the watcher.\n\t\tid, err := serviceRegistration(serviceKey, serviceValue)\n\t\tfor err != nil {\n\t\t\tid, err = serviceRegistration(serviceKey, serviceValue)\n\t\t}\n\n\t\terr = hostRegistration(hostKey, hostValue)\n\t\tfor err != nil {\n\t\t\terr = hostRegistration(hostKey, hostValue)\n\t\t}\n\n\t\tpublicCredentialsChan <- Credentials{\n\t\t\tUser: serviceInfos.User,\n\t\t\tPassword: serviceInfos.Password,\n\t\t}\n\n\t\tif host.Public {\n\t\t\tgo watch(ctx, serviceKey, id, privateCredentialsChan)\n\t\t}\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\t_, err := KAPI().Delete(context.Background(), hostKey, &etcd.DeleteOptions{Recursive: false})\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogger.Println(\"fail to remove key\", hostKey)\n\t\t\t\t}\n\t\t\t\tticker.Stop()\n\t\t\t\treturn\n\t\t\tcase credentials := <-privateCredentialsChan: // If the credentials has benn changed\n\t\t\t\t// We update our cache\n\t\t\t\thost.User = credentials.User\n\t\t\t\thost.Password = credentials.Password\n\t\t\t\tserviceInfos.User = credentials.User\n\t\t\t\tserviceInfos.Password = credentials.Password\n\n\t\t\t\t// Re-marshal the host\n\t\t\t\thostJson, _ = json.Marshal(&host)\n\t\t\t\thostValue = string(hostJson)\n\n\t\t\t\t// synchro the host informations\n\t\t\t\thostRegistration(hostKey, hostValue)\n\t\t\t\t// and transmit them to the client\n\t\t\t\tpublicCredentialsChan <- credentials\n\t\t\tcase <-ticker.C:\n\t\t\t\terr := hostRegistration(hostKey, hostValue)\n\t\t\t\t// If for any random reason, there is an error,\n\t\t\t\t// we retry every second until it's ok.\n\t\t\t\tfor err != nil {\n\t\t\t\t\tlogger.Printf(\"lost registration of '%v': %v (%v)\", service, err, Client().Endpoints())\n\t\t\t\t\ttime.Sleep(1 * time.Second)\n\n\t\t\t\t\terr = hostRegistration(hostKey, hostValue)\n\t\t\t\t\tif err == nil {\n\t\t\t\t\t\tlogger.Printf(\"recover registration of '%v'\", service)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn NewRegistration(ctx, hostUuid, publicCredentialsChan)\n}", "func (rf *Raft) InstallSnapshot(args *InstallSnapshotArgs, reply *InstallSnapshotReply) {\n rf.mu.Lock()\n defer rf.mu.Unlock()\n\n DISPrintf(\"Server(%d) receive InstallSnapshot RPC from Leader(%d)[LeaderTerm=%d, LastIncludedIndex=%d, LastIncludedTerm=%d]\", rf.me, args.LeaderId,\n args.LeaderTerm, args.LastIncludedIndex, args.LastIncludedTerm)\n raftViewIndex := rf.convertToRaftLogViewIndex(args.LastIncludedIndex)\n\n if args.LeaderTerm < rf.currentTerm || raftViewIndex < 0 {\n reply.FollowerTerm = rf.currentTerm\n return\n }\n\n if args.LeaderTerm > rf.currentTerm {\n rf.currentTerm = args.LeaderTerm\n rf.mu.Unlock()\n rf.convertToFollower()\n rf.mu.Lock()\n }\n\n reply.FollowerTerm = rf.currentTerm\n\n if raftViewIndex < len(rf.log) && args.LastIncludedTerm == rf.log[raftViewIndex].Term {\n rf.log = rf.log[raftViewIndex:]\n }else{\n rf.log = make([]Entry, 1, 100)\n rf.log[0].Term = args.LastIncludedTerm\n }\n\n rf.snapshottedIndex = args.LastIncludedIndex\n if rf.commitIndex < args.LastIncludedIndex {\n rf.commitIndex = args.LastIncludedIndex\n \n }\n if rf.lastApplied < args.LastIncludedIndex {\n rf.lastApplied = rf.commitIndex\n }\n\n rf.persistWithSnapshotInBytes(args.Data)\n\n if rf.lastApplied > args.LastIncludedIndex {\n return\n }\n\n applyMsg := ApplyMsg{}\n applyMsg.CommandValid = false\n applyMsg.CommandSnapshot = args.Data\n\n go func() {\n DISPrintf(\"Follower(%d) send a snapshot to its applyCh\", rf.me)\n rf.applyCh <- applyMsg\n }()\n}", "func (s *ProcProbe) Start() error {\n\tif err := s.scanProc(); err != nil {\n\t\treturn err\n\t}\n\n\ts.updateMetadata()\n\n\tgo func() {\n\t\tseconds := s.Ctx.Config.GetInt(\"agent.topology.socketinfo.host_update\")\n\t\tticker := time.NewTicker(time.Duration(seconds) * time.Second)\n\t\tdefer ticker.Stop()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-s.quit:\n\t\t\t\treturn\n\t\t\tcase <-ticker.C:\n\t\t\t\ts.scanProc()\n\t\t\t\ts.updateMetadata()\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn nil\n}", "func (o LookupInstanceResultOutput) Fingerprint() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupInstanceResult) string { return v.Fingerprint }).(pulumi.StringOutput)\n}", "func (o MetadataResponsePtrOutput) Fingerprint() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *MetadataResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Fingerprint\n\t}).(pulumi.StringPtrOutput)\n}", "func (m *DevicePluginStub) Start() error {\n\tos.MkdirAll(filepath.Dir(m.socket), 0755)\n\n\tif err := m.cleanup(); err != nil {\n\t\treturn err\n\t}\n\n\tsock, err := net.Listen(\"unix\", m.socket)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tm.server = grpc.NewServer([]grpc.ServerOption{}...)\n\tpluginapi.RegisterDevicePluginServer(m.server, m)\n\tpluginregistration.RegisterIdentityServer(m.server, m)\n\n\tm.wg.Add(1)\n\tgo func() {\n\t\tdefer m.wg.Done()\n\t\tm.server.Serve(sock)\n\t}()\n\n\t// Wait for server to start by launching a blocking connexion\n\tc, err := dial(m.socket, 5*time.Second)\n\tif err != nil {\n\t\treturn err\n\t}\n\tc.Close()\n\n\tglog.V(2).Infof(\"Starting to serve on %s\", m.socket)\n\n\treturn nil\n}", "func (self *Notifier) InitializeNotifier() {\n\tself.Facade = GetInstance(func() interfaces.IFacade { return &Facade{} })\n}", "func (a *Age) Fingerprint(ctx context.Context, id string) string {\n\treturn id\n}", "func (a *Age) Fingerprint(ctx context.Context, id string) string {\n\treturn id\n}", "func (r *ring) Start() {\n\tif !atomic.CompareAndSwapInt32(\n\t\t&r.status,\n\t\tcommon.DaemonStatusInitialized,\n\t\tcommon.DaemonStatusStarted,\n\t) {\n\t\treturn\n\t}\n\tif err := r.peerProvider.Subscribe(r.service, r.refreshChan); err != nil {\n\t\tr.logger.Fatal(\"subscribing to peer provider\", tag.Error(err))\n\t}\n\n\tif err := r.refresh(); err != nil {\n\t\tr.logger.Fatal(\"failed to start service resolver\", tag.Error(err))\n\t}\n\n\tr.shutdownWG.Add(1)\n\tgo r.refreshRingWorker()\n}", "func (o NetworkAttachmentOutput) Fingerprint() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *NetworkAttachment) pulumi.StringOutput { return v.Fingerprint }).(pulumi.StringOutput)\n}" ]
[ "0.67849284", "0.5897063", "0.5688403", "0.5526209", "0.54947096", "0.54665834", "0.53494525", "0.51623464", "0.5127396", "0.5124749", "0.5117888", "0.49343532", "0.48946166", "0.4862256", "0.47582838", "0.4736561", "0.4724519", "0.47087526", "0.466834", "0.46634454", "0.4586778", "0.4550135", "0.4525737", "0.4466943", "0.4465204", "0.44611245", "0.4460576", "0.44463235", "0.44444984", "0.4441748", "0.44261402", "0.44142184", "0.4412242", "0.43842974", "0.43705264", "0.4365422", "0.43572924", "0.4351877", "0.4351478", "0.43484622", "0.43483663", "0.4332941", "0.43218595", "0.43069226", "0.4298474", "0.42984462", "0.42927745", "0.42923", "0.42903873", "0.42876232", "0.42808053", "0.42794287", "0.42639336", "0.42421257", "0.4237432", "0.423307", "0.42260873", "0.42255804", "0.41922984", "0.41891232", "0.41876793", "0.41852403", "0.41847932", "0.41651833", "0.41571334", "0.41513914", "0.41491354", "0.41440117", "0.4141246", "0.41338995", "0.41309735", "0.4130134", "0.4110368", "0.40966398", "0.4089771", "0.40876365", "0.4084617", "0.40834782", "0.40804636", "0.4062094", "0.40610018", "0.40501463", "0.40374094", "0.40351835", "0.403238", "0.40315428", "0.40313062", "0.40226287", "0.4019886", "0.40195358", "0.4004443", "0.4003595", "0.39989913", "0.39934266", "0.39921886", "0.3991892", "0.39860448", "0.39860448", "0.39742136", "0.39740947" ]
0.6965612
0
RecoverTask detects running tasks when nomad client or task driver is restarted. When a driver is restarted it is not expected to persist any internal state to disk. To support this, Nomad will attempt to recover a task that was previously started if the driver does not recognize the task ID. During task recovery, Nomad calls RecoverTask passing the TaskHandle that was returned by the StartTask function.
func (d *Driver) RecoverTask(handle *drivers.TaskHandle) error { if handle == nil { return fmt.Errorf("error: handle cannot be nil") } if _, ok := d.tasks.Get(handle.Config.ID); ok { return nil } var taskState TaskState if err := handle.GetDriverState(&taskState); err != nil { return fmt.Errorf("failed to decode task state from handle: %v", err) } d.logger.Debug("Checking for recoverable task", "task", handle.Config.Name, "taskid", handle.Config.ID, "container", taskState.ContainerID) inspectData, err := d.podman.ContainerInspect(d.ctx, taskState.ContainerID) if err != nil { d.logger.Warn("Recovery lookup failed", "task", handle.Config.ID, "container", taskState.ContainerID, "err", err) return nil } h := &TaskHandle{ containerID: taskState.ContainerID, driver: d, taskConfig: taskState.TaskConfig, procState: drivers.TaskStateUnknown, startedAt: taskState.StartedAt, exitResult: &drivers.ExitResult{}, logger: d.logger.Named("podmanHandle"), totalCPUStats: stats.NewCpuStats(), userCPUStats: stats.NewCpuStats(), systemCPUStats: stats.NewCpuStats(), removeContainerOnExit: d.config.GC.Container, } if inspectData.State.Running { d.logger.Info("Recovered a still running container", "container", inspectData.State.Pid) h.procState = drivers.TaskStateRunning } else if inspectData.State.Status == "exited" { // are we allowed to restart a stopped container? if d.config.RecoverStopped { d.logger.Debug("Found a stopped container, try to start it", "container", inspectData.State.Pid) if err = d.podman.ContainerStart(d.ctx, inspectData.ID); err != nil { d.logger.Warn("Recovery restart failed", "task", handle.Config.ID, "container", taskState.ContainerID, "err", err) } else { d.logger.Info("Restarted a container during recovery", "container", inspectData.ID) h.procState = drivers.TaskStateRunning } } else { // no, let's cleanup here to prepare for a StartTask() d.logger.Debug("Found a stopped container, removing it", "container", inspectData.ID) if err = d.podman.ContainerStart(d.ctx, inspectData.ID); err != nil { d.logger.Warn("Recovery cleanup failed", "task", handle.Config.ID, "container", inspectData.ID) } h.procState = drivers.TaskStateExited } } else { d.logger.Warn("Recovery restart failed, unknown container state", "state", inspectData.State.Status, "container", taskState.ContainerID) h.procState = drivers.TaskStateUnknown } d.tasks.Set(taskState.TaskConfig.ID, h) go h.runContainerMonitor() d.logger.Debug("Recovered container handle", "container", taskState.ContainerID) return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func TestLibvirtDriver_Start_Stop_Recover_Task(t *testing.T) {\n\tutil.RequireLibvirt(t)\n\n\trequire := require.New(t)\n\n\td := NewLibvirtDriver(testlog.HCLogger(t)).(*Driver)\n\tharness := dtestutil.NewDriverHarness(t, d)\n\n\trequire.NoError(task.EncodeConcreteDriverConfig(&taskCfg))\n\n\tcleanup := harness.MkAllocDir(task, false)\n\tdefer cleanup()\n\n\t// Test start task\n\thandle, _, err := harness.StartTask(task)\n\trequire.NoError(err)\n\trequire.NotNil(handle)\n\n\tlibvirtHandle, ok := d.tasks.Get(task.ID)\n\trequire.NotNil(libvirtHandle)\n\trequire.True(ok)\n\n\ttestutil.WaitForResult(func() (bool, error) {\n\t\tstatus, err := d.InspectTask(task.ID)\n\t\trequire.NoError(err)\n\t\tif status.State == drivers.TaskStateRunning {\n\t\t\treturn true, nil\n\t\t}\n\t\treturn false, fmt.Errorf(\"task in state: %v\", status.State)\n\t}, func(err error) {\n\t\tt.Fatalf(\"task failed to start: %v\", err)\n\t})\n\n\t// Missing the task handle\n\td.tasks.Delete(task.ID)\n\n\t// Test recover the missed task\n\trecoverHandle := handle.Copy()\n\trequire.NoError(d.RecoverTask(recoverHandle))\n\n\td.StopTask(task.ID, 5*time.Second, \"kill\")\n\n\t// Destroy the task/vm after test\n\tdefer d.DestroyTask(task.ID, false)\n\n\t// Test after recovery and stop task\n\ttestutil.WaitForResult(func() (bool, error) {\n\t\tstatus, err := d.InspectTask(task.ID)\n\t\trequire.NoError(err)\n\t\tif status.State == drivers.TaskStateExited {\n\t\t\treturn true, nil\n\t\t}\n\t\treturn false, fmt.Errorf(\"task in state: %v\", status.State)\n\t}, func(err error) {\n\t\tt.Fatalf(\"task failed to stop: %v\", err)\n\t})\n}", "func (mgr *DataCheckMgr) doRecover() {\n\n\tnow := time.Now().Unix()\n\tblog.Info(\"data checker: doRecover begin.... now(%d)\", now)\n\tdefer func() {\n\t\tnow = time.Now().Unix()\n\t\tblog.Info(\"data checker: doRecover end.... now(%d)\", now)\n\t}()\n\n\trunAses, err := mgr.store.ListRunAs()\n\tif err != nil {\n\t\tblog.Error(\"data checker: fail to list runAses, err:%s\", err.Error())\n\t\treturn\n\t}\n\n\tfor _, runAs := range runAses {\n\t\tappIDs, err := mgr.store.ListApplicationNodes(runAs)\n\t\tif err != nil {\n\t\t\tblog.Error(\"data checker: fail to list %s, err:%s\", runAs, err.Error())\n\t\t\tcontinue\n\t\t}\n\t\tif nil == appIDs {\n\t\t\tblog.Warn(\"data checker: no application nodes under runAs:%s\", runAs)\n\t\t\tcontinue\n\t\t}\n\t\tfor _, appID := range appIDs {\n\t\t\tblog.Info(\"data checker: to recover application:%s.%s \", runAs, appID)\n\t\t\tmgr.recoverTaskgroup(runAs, appID)\n\t\t}\n\t}\n\n\treturn\n}", "func (broadcast *Broadcast) Recover(ctx context.Context, username, newResetPubKeyHex,\n\tnewTransactionPubKeyHex, newAppPubKeyHex, privKeyHex string, seq int64) (*model.BroadcastResponse, error) {\n\tresetPubKey, err := transport.GetPubKeyFromHex(newResetPubKeyHex)\n\tif err != nil {\n\t\treturn nil, errors.FailedToGetPubKeyFromHexf(\"Recover: failed to get Reset pub key\").AddCause(err)\n\t}\n\ttxPubKey, err := transport.GetPubKeyFromHex(newTransactionPubKeyHex)\n\tif err != nil {\n\t\treturn nil, errors.FailedToGetPubKeyFromHexf(\"Recover: failed to get Tx pub key\").AddCause(err)\n\t}\n\tappPubKey, err := transport.GetPubKeyFromHex(newAppPubKeyHex)\n\tif err != nil {\n\t\treturn nil, errors.FailedToGetPubKeyFromHexf(\"Recover: failed to get App pub key\").AddCause(err)\n\t}\n\n\tmsg := model.RecoverMsg{\n\t\tUsername: username,\n\t\tNewResetPubKey: resetPubKey,\n\t\tNewTransactionPubKey: txPubKey,\n\t\tNewAppPubKey: appPubKey,\n\t}\n\treturn broadcast.broadcastTransaction(ctx, msg, privKeyHex, seq, \"\", false)\n}", "func (oi *offsetInjector) recover(ctx context.Context, nodeID int) {\n\tif !oi.deployed {\n\t\toi.c.t.Fatal(\"Offset injector must be deployed before recovering from clock offsets\")\n\t}\n\n\tsyncCmds := [][]string{\n\t\t{\"sudo\", \"service\", \"ntp\", \"stop\"},\n\t\t{\"sudo\", \"ntpdate\", \"-u\", \"time.google.com\"},\n\t\t{\"sudo\", \"service\", \"ntp\", \"start\"},\n\t}\n\tfor _, cmd := range syncCmds {\n\t\toi.c.Run(\n\t\t\tctx,\n\t\t\toi.c.Node(nodeID),\n\t\t\tcmd...,\n\t\t)\n\t}\n}", "func (s *Stopper) Recover(ctx context.Context) {\n\tif r := recover(); r != nil {\n\t\tif s.onPanic != nil {\n\t\t\ts.onPanic(r)\n\t\t\treturn\n\t\t}\n\t\tif sv := settings.TODO(); sv != nil {\n\t\t\tlog.ReportPanic(ctx, sv, r, 1)\n\t\t}\n\t\tpanic(r)\n\t}\n}", "func Recover() {\n\tdoRecover(recover(), false)\n}", "func TestRktDriver_StartWaitRecoverWaitStop(t *testing.T) {\n\tctestutil.RktCompatible(t)\n\tif !testutil.IsCI() {\n\t\tt.Parallel()\n\t}\n\n\trequire := require.New(t)\n\td := NewRktDriver(testlog.HCLogger(t))\n\tharness := dtestutil.NewDriverHarness(t, d)\n\n\ttask := &drivers.TaskConfig{\n\t\tID: uuid.Generate(),\n\t\tAllocID: uuid.Generate(),\n\t\tName: \"etcd\",\n\t\tResources: &drivers.Resources{\n\t\t\tNomadResources: &structs.AllocatedTaskResources{\n\t\t\t\tMemory: structs.AllocatedMemoryResources{\n\t\t\t\t\tMemoryMB: 128,\n\t\t\t\t},\n\t\t\t\tCpu: structs.AllocatedCpuResources{\n\t\t\t\t\tCpuShares: 100,\n\t\t\t\t},\n\t\t\t},\n\t\t\tLinuxResources: &drivers.LinuxResources{\n\t\t\t\tMemoryLimitBytes: 134217728,\n\t\t\t\tCPUShares: 100,\n\t\t\t},\n\t\t},\n\t}\n\n\ttc := &TaskConfig{\n\t\tImageName: \"coreos.com/etcd:v2.0.4\",\n\t\tCommand: \"/etcd\",\n\t}\n\trequire.NoError(task.EncodeConcreteDriverConfig(&tc))\n\n\tcleanup := harness.MkAllocDir(task, true)\n\tdefer cleanup()\n\n\thandle, _, err := harness.StartTask(task)\n\trequire.NoError(err)\n\n\tch, err := harness.WaitTask(context.Background(), task.ID)\n\trequire.NoError(err)\n\n\tvar waitDone bool\n\tvar wg sync.WaitGroup\n\twg.Add(1)\n\tgo func() {\n\t\tdefer wg.Done()\n\t\tresult := <-ch\n\t\trequire.Error(result.Err)\n\t\twaitDone = true\n\t}()\n\n\toriginalStatus, err := d.InspectTask(task.ID)\n\trequire.NoError(err)\n\n\td.(*Driver).tasks.Delete(task.ID)\n\n\twg.Wait()\n\trequire.True(waitDone)\n\t_, err = d.InspectTask(task.ID)\n\trequire.Equal(drivers.ErrTaskNotFound, err)\n\n\terr = d.RecoverTask(handle)\n\trequire.NoError(err)\n\n\tstatus, err := d.InspectTask(task.ID)\n\trequire.NoError(err)\n\trequire.Exactly(originalStatus, status)\n\n\tch, err = harness.WaitTask(context.Background(), task.ID)\n\trequire.NoError(err)\n\n\trequire.NoError(d.StopTask(task.ID, 0, \"SIGKILL\"))\n\n\tselect {\n\tcase result := <-ch:\n\t\trequire.NoError(result.Err)\n\t\trequire.NotZero(result.ExitCode)\n\n\t\t// when killing a task, signal might not propagate\n\t\t// when executor proc.Wait() call gets \"wait: no child processes\" error\n\t\t//require.Equal(9, result.Signal)\n\tcase <-time.After(time.Duration(testutil.TestMultiplier()*5) * time.Second):\n\t\trequire.Fail(\"WaitTask timeout\")\n\t}\n\n\trequire.NoError(d.DestroyTask(task.ID, false))\n}", "func (ctb *Crontab) Recover(recoverable Recoverable, err interface{}) {\n\tlog.Printf(\"Recovering crontab backend after error '%v'!\", err)\n\n\tgo ctb.backend()\n}", "func (s *Stopper) Recover(ctx context.Context) {\n\tif r := recover(); r != nil {\n\t\tif s.onPanic != nil {\n\t\t\ts.onPanic(r)\n\t\t\treturn\n\t\t}\n\t\tif sv := settings.TODO(); sv != nil {\n\t\t\tlogcrash.ReportPanic(ctx, sv, r, 1)\n\t\t}\n\t\tpanic(r)\n\t}\n}", "func (s *Worker) Start() error {\n\tclient, err := worker.InitRPCChannel(*s.cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\ts.rc = client\n\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tif client != nil {\n\t\t\t\t// we dont really care about the error here...\n\t\t\t\thostname, _ := os.Hostname()\n\t\t\t\terrStr := fmt.Sprintf(\"A panic occurred. Check logs on %s for more details\", hostname)\n\t\t\t\tclient.ChangeTaskStatus(rpc.ChangeTaskStatusRequest{\n\t\t\t\t\tTaskID: s.taskid,\n\t\t\t\t\tNewStatus: storage.TaskStatusError,\n\t\t\t\t\tError: &errStr,\n\t\t\t\t})\n\t\t\t}\n\t\t\tlog.Error().Str(\"task_id\", s.taskid).Msg(\"A critical error occurred while running task (panic)\")\n\t\t}\n\t}()\n\n\ts.t = NewTask(s.taskid, s.devices, s.cfg, s.rc) //Get the task in order to collect the task duration\n\tresp, err := s.t.c.GetTask(rpc.RequestTaskPayload{\n\t\tTaskID: s.t.taskid,\n\t})\n\n\tif resp.TaskDuration != 0 { //If the task duration is 0 (not set), we don't run the timer\n\t\ttimer := time.NewTimer(time.Second * time.Duration(resp.TaskDuration))\n\t\tgo func() {\n\t\t\t<-timer.C\n\t\t\tlog.Warn().Msg(\"Timer expired, stopping task\")\n\t\t\ts.t.Stop()\n\t\t\ttimer.Stop()\n\t\t}()\n\t}\n\n\tif err := s.t.Start(); err != nil {\n\t\tlog.Error().Err(err).Str(\"task_id\", s.taskid).Msg(\"An error occurred while processing a task\")\n\t\terrptr := err.Error()\n\t\tif rpcerr := client.ChangeTaskStatus(rpc.ChangeTaskStatusRequest{\n\t\t\tTaskID: s.taskid,\n\t\t\tNewStatus: storage.TaskStatusError,\n\t\t\tError: &errptr,\n\t\t}); rpcerr != nil {\n\t\t\tlog.Error().Err(rpcerr).Msg(\"Failed to change tasks status to error\")\n\t\t}\n\t}\n\treturn nil\n}", "func Recover() {\n\tdoRecover(recover())\n}", "func (t *trial) recover() error {\n\trunID, restarts, err := t.db.TrialRunIDAndRestarts(t.id)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"restoring old trial state\")\n\t}\n\tt.runID = runID\n\tt.restarts = restarts\n\treturn nil\n}", "func resetTask(ctx context.Context, settings *evergreen.Settings, taskId, username string, failedOnly bool) error {\n\tt, err := task.FindOneId(taskId)\n\tif err != nil {\n\t\treturn gimlet.ErrorResponse{\n\t\t\tStatusCode: http.StatusInternalServerError,\n\t\t\tMessage: errors.Wrapf(err, \"finding task '%s'\", t).Error(),\n\t\t}\n\t}\n\tif t == nil {\n\t\treturn gimlet.ErrorResponse{\n\t\t\tStatusCode: http.StatusNotFound,\n\t\t\tMessage: fmt.Sprintf(\"task '%s' not found\", taskId),\n\t\t}\n\t}\n\treturn errors.Wrapf(serviceModel.ResetTaskOrDisplayTask(ctx, settings, t, username, evergreen.RESTV2Package, failedOnly, nil), \"resetting task '%s'\", taskId)\n}", "func (c *Cleanup) Recover() {\n\tif err := recover(); err != nil {\n\t\tc.run()\n\t\tpanic(err)\n\t}\n}", "func Recover() arpc.HandlerFunc {\n\treturn func(ctx *arpc.Context) {\n\t\tdefer util.Recover()\n\t\tctx.Next()\n\t}\n}", "func (s *Session) RetryTask(t util.Task) error {\n\ttask := util.Task{\n\t\tName: t.Name,\n\t\tOriginalTaskID: t.OriginalTaskID,\n\t\tPayload: t.Payload,\n\t\tPriority: t.Priority,\n\t\tStatus: util.StatusRetry,\n\t}\n\n\t// updating original task id counter\n\ts.taskRepo.UpdateRetryCount(t.OriginalTaskID, -1)\n\tif err := s.SendTask(task); err != nil {\n\t\ts.lgr.Error(\"failed to retry\", err, util.Object{Key: \"TaskID\", Val: task.TaskID}, util.Object{Key: \"OriginalTaskID\", Val: task.OriginalTaskID})\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (sr SecureRecoverer) Recover(message string) error {\n\tif sr.Sync {\n\t\tsyscall.Sync()\n\t}\n\n\tif sr.Debug {\n\t\tif message != \"\" {\n\t\t\tlog.Print(message)\n\t\t}\n\t\ttime.Sleep(DebugTimeout * time.Second)\n\t}\n\n\tif sr.Reboot {\n\t\tif err := syscall.Reboot(syscall.LINUX_REBOOT_CMD_RESTART); err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tif err := syscall.Reboot(syscall.LINUX_REBOOT_CMD_POWER_OFF); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (mpt *MapPinTracker) Recover(c *cid.Cid) (api.PinInfo, error) {\n\tp := mpt.get(c)\n\tif p.Status != api.TrackerStatusPinError &&\n\t\tp.Status != api.TrackerStatusUnpinError {\n\t\treturn p, nil\n\t}\n\tlogger.Infof(\"Recovering %s\", c)\n\tvar err error\n\tswitch p.Status {\n\tcase api.TrackerStatusPinError:\n\t\terr = mpt.pin(api.Pin{Cid: c})\n\tcase api.TrackerStatusUnpinError:\n\t\terr = mpt.unpin(api.Pin{Cid: c})\n\t}\n\tif err != nil {\n\t\tlogger.Errorf(\"error recovering %s: %s\", c, err)\n\t}\n\treturn mpt.get(c), err\n}", "func (e *dockerEngine) Recover() {\n\t// clean old services in docker mode\n\te.clean()\n}", "func (suite *TaskFailRetryTestSuite) TestLostTaskRetry() {\n\ttaskConfig := pbtask.TaskConfig{\n\t\tRestartPolicy: &pbtask.RestartPolicy{\n\t\t\tMaxFailures: 3,\n\t\t},\n\t}\n\n\tsuite.cachedTask.EXPECT().\n\t\tID().\n\t\tReturn(uint32(0)).\n\t\tAnyTimes()\n\n\tsuite.jobFactory.EXPECT().\n\t\tGetJob(suite.jobID).Return(suite.cachedJob)\n\n\tsuite.cachedJob.EXPECT().\n\t\tGetTask(suite.instanceID).Return(suite.cachedTask)\n\n\tsuite.cachedJob.EXPECT().\n\t\tID().Return(suite.jobID)\n\n\tsuite.cachedTask.EXPECT().\n\t\tGetRuntime(gomock.Any()).Return(suite.lostTaskRuntime, nil)\n\n\tsuite.taskConfigV2Ops.EXPECT().\n\t\tGetTaskConfig(gomock.Any(), suite.jobID, suite.instanceID, gomock.Any()).\n\t\tReturn(&taskConfig, &models.ConfigAddOn{}, nil)\n\n\tsuite.cachedJob.EXPECT().\n\t\tPatchTasks(gomock.Any(), gomock.Any(), false).\n\t\tDo(func(ctx context.Context,\n\t\t\truntimeDiffs map[uint32]jobmgrcommon.RuntimeDiff,\n\t\t\t_ bool) {\n\t\t\truntimeDiff := runtimeDiffs[suite.instanceID]\n\t\t\tsuite.True(\n\t\t\t\truntimeDiff[jobmgrcommon.MesosTaskIDField].(*mesosv1.TaskID).GetValue() != suite.mesosTaskID)\n\t\t\tsuite.True(\n\t\t\t\truntimeDiff[jobmgrcommon.PrevMesosTaskIDField].(*mesosv1.TaskID).GetValue() == suite.mesosTaskID)\n\t\t\tsuite.True(\n\t\t\t\truntimeDiff[jobmgrcommon.StateField].(pbtask.TaskState) == pbtask.TaskState_INITIALIZED)\n\t\t}).Return(nil, nil, nil)\n\n\tsuite.cachedJob.EXPECT().\n\t\tGetJobType().Return(pbjob.JobType_BATCH)\n\n\tsuite.taskGoalStateEngine.EXPECT().\n\t\tEnqueue(gomock.Any(), gomock.Any()).\n\t\tReturn()\n\n\tsuite.jobGoalStateEngine.EXPECT().\n\t\tEnqueue(gomock.Any(), gomock.Any()).\n\t\tReturn()\n\n\terr := TaskFailRetry(context.Background(), suite.taskEnt)\n\tsuite.NoError(err)\n}", "func (w *Worker) handleTask() {\n\tvar handleTaskInterval = time.Second\n\tfailpoint.Inject(\"handleTaskInterval\", func(val failpoint.Value) {\n\t\tif milliseconds, ok := val.(int); ok {\n\t\t\thandleTaskInterval = time.Duration(milliseconds) * time.Millisecond\n\t\t\tw.l.Info(\"set handleTaskInterval\", zap.String(\"failpoint\", \"handleTaskInterval\"), zap.Int(\"value\", milliseconds))\n\t\t}\n\t})\n\tticker := time.NewTicker(handleTaskInterval)\n\tdefer ticker.Stop()\n\n\tretryCnt := 0\n\nLoop:\n\tfor {\n\t\tselect {\n\t\tcase <-w.ctx.Done():\n\t\t\tw.l.Info(\"handle task process exits!\")\n\t\t\treturn\n\t\tcase <-ticker.C:\n\t\t\tif w.closed.Get() == closedTrue {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\topLog := w.meta.PeekLog()\n\t\t\tif opLog == nil {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tw.l.Info(\"start to execute operation\", zap.Reflect(\"oplog\", opLog))\n\n\t\t\tst := w.subTaskHolder.findSubTask(opLog.Task.Name)\n\t\t\tvar err error\n\t\t\tswitch opLog.Task.Op {\n\t\t\tcase pb.TaskOp_Start:\n\t\t\t\tif st != nil {\n\t\t\t\t\terr = terror.ErrWorkerSubTaskExists.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tif w.relayPurger.Purging() {\n\t\t\t\t\tif retryCnt < maxRetryCount {\n\t\t\t\t\t\tretryCnt++\n\t\t\t\t\t\tw.l.Warn(\"relay log purger is purging, cannot start subtask, would try again later\", zap.String(\"task\", opLog.Task.Name))\n\t\t\t\t\t\tcontinue Loop\n\t\t\t\t\t}\n\n\t\t\t\t\tretryCnt = 0\n\t\t\t\t\terr = terror.ErrWorkerRelayIsPurging.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tretryCnt = 0\n\t\t\t\ttaskCfg := new(config.SubTaskConfig)\n\t\t\t\tif err1 := taskCfg.Decode(string(opLog.Task.Task)); err1 != nil {\n\t\t\t\t\terr = terror.Annotate(err1, \"decode subtask config error in handleTask\")\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tvar cfgDecrypted *config.SubTaskConfig\n\t\t\t\tcfgDecrypted, err = taskCfg.DecryptPassword()\n\t\t\t\tif err != nil {\n\t\t\t\t\terr = terror.WithClass(err, terror.ClassDMWorker)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tw.l.Info(\"started sub task\", zap.Stringer(\"config\", cfgDecrypted))\n\t\t\t\tst = NewSubTask(cfgDecrypted)\n\t\t\t\tw.subTaskHolder.recordSubTask(st)\n\t\t\t\tst.Run()\n\n\t\t\tcase pb.TaskOp_Update:\n\t\t\t\tif st == nil {\n\t\t\t\t\terr = terror.ErrWorkerSubTaskNotFound.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\ttaskCfg := new(config.SubTaskConfig)\n\t\t\t\tif err1 := taskCfg.Decode(string(opLog.Task.Task)); err1 != nil {\n\t\t\t\t\terr = terror.Annotate(err1, \"decode subtask config error in handleTask\")\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tw.l.Info(\"updated sub task\", zap.String(\"task\", opLog.Task.Name), zap.Stringer(\"new config\", taskCfg))\n\t\t\t\terr = st.Update(taskCfg)\n\t\t\tcase pb.TaskOp_Stop:\n\t\t\t\tif st == nil {\n\t\t\t\t\terr = terror.ErrWorkerSubTaskNotFound.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tw.l.Info(\"stop sub task\", zap.String(\"task\", opLog.Task.Name))\n\t\t\t\tst.Close()\n\t\t\t\tw.subTaskHolder.removeSubTask(opLog.Task.Name)\n\t\t\tcase pb.TaskOp_Pause:\n\t\t\t\tif st == nil {\n\t\t\t\t\terr = terror.ErrWorkerSubTaskNotFound.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tw.l.Info(\"pause sub task\", zap.String(\"task\", opLog.Task.Name))\n\t\t\t\terr = st.Pause()\n\t\t\tcase pb.TaskOp_Resume:\n\t\t\t\tif st == nil {\n\t\t\t\t\terr = terror.ErrWorkerSubTaskNotFound.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tw.l.Info(\"resume sub task\", zap.String(\"task\", opLog.Task.Name))\n\t\t\t\terr = st.Resume()\n\t\t\tcase pb.TaskOp_AutoResume:\n\t\t\t\tif st == nil {\n\t\t\t\t\terr = terror.ErrWorkerSubTaskNotFound.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tw.l.Info(\"auto_resume sub task\", zap.String(\"task\", opLog.Task.Name))\n\t\t\t\terr = st.Resume()\n\t\t\t}\n\n\t\t\tw.l.Info(\"end to execute operation\", zap.Int64(\"oplog ID\", opLog.Id), log.ShortError(err))\n\n\t\t\tif err != nil {\n\t\t\t\topLog.Message = err.Error()\n\t\t\t} else {\n\t\t\t\topLog.Task.Stage = st.Stage()\n\t\t\t\topLog.Success = true\n\t\t\t}\n\n\t\t\t// fill current task config\n\t\t\tif len(opLog.Task.Task) == 0 {\n\t\t\t\ttm := w.meta.GetTask(opLog.Task.Name)\n\t\t\t\tif tm == nil {\n\t\t\t\t\tw.l.Warn(\"task meta not found\", zap.String(\"task\", opLog.Task.Name))\n\t\t\t\t} else {\n\t\t\t\t\topLog.Task.Task = append([]byte{}, tm.Task...)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\terr = w.meta.MarkOperation(opLog)\n\t\t\tif err != nil {\n\t\t\t\tw.l.Error(\"fail to mark subtask operation\", zap.Reflect(\"oplog\", opLog))\n\t\t\t}\n\t\t}\n\t}\n}", "func (x Go) Recover(recoverFunc func(interface{})) Go {\n\tx.recoverFunc = recoverFunc\n\treturn x\n}", "func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error {\n\ttr.logger.Trace(\"Restart requested\", \"failure\", failure, \"event\", event.GoString())\n\n\ttaskState := tr.TaskState()\n\tif taskState == nil {\n\t\treturn ErrTaskNotRunning\n\t}\n\n\tswitch taskState.State {\n\tcase structs.TaskStatePending, structs.TaskStateDead:\n\t\treturn ErrTaskNotRunning\n\t}\n\n\treturn tr.restartImpl(ctx, event, failure)\n}", "func (tc *DBTaskConnector) ResetTask(taskId, username string, proj *serviceModel.Project) error {\n\treturn errors.Wrap(serviceModel.TryResetTask(taskId, username, evergreen.RESTV2Package, proj, nil),\n\t\t\"Reset task error\")\n}", "func (k *KubernetesExecutor) reportLostTask(driver bindings.ExecutorDriver, tid, reason string) {\n\tk.removePodTask(driver, tid, reason, mesos.TaskState_TASK_LOST)\n}", "func (k *KubernetesExecutor) reportLostTask(driver bindings.ExecutorDriver, tid, reason string) {\n\tk.removePodTask(driver, tid, reason, mesos.TaskState_TASK_LOST)\n}", "func (suite *TaskFailRetryTestSuite) TestLostTaskNoRetry() {\n\ttaskConfig := pbtask.TaskConfig{\n\t\tRestartPolicy: &pbtask.RestartPolicy{\n\t\t\tMaxFailures: 0,\n\t\t},\n\t}\n\tsuite.jobFactory.EXPECT().\n\t\tGetJob(suite.jobID).Return(suite.cachedJob)\n\n\tsuite.cachedJob.EXPECT().\n\t\tGetTask(suite.instanceID).Return(suite.cachedTask)\n\n\tsuite.cachedTask.EXPECT().\n\t\tGetRuntime(gomock.Any()).Return(suite.lostTaskRuntime, nil)\n\n\tsuite.taskConfigV2Ops.EXPECT().\n\t\tGetTaskConfig(gomock.Any(), suite.jobID, suite.instanceID, gomock.Any()).\n\t\tReturn(&taskConfig, &models.ConfigAddOn{}, nil)\n\n\terr := TaskFailRetry(context.Background(), suite.taskEnt)\n\tsuite.NoError(err)\n}", "func (trh *taskRestartHandler) Run(ctx context.Context) gimlet.Responder {\n\terr := resetTask(ctx, evergreen.GetEnvironment().Settings(), trh.taskId, trh.username, trh.FailedOnly)\n\tif err != nil {\n\t\treturn gimlet.MakeJSONErrorResponder(err)\n\t}\n\n\trefreshedTask, err := task.FindOneId(trh.taskId)\n\tif err != nil {\n\t\treturn gimlet.MakeJSONInternalErrorResponder(errors.Wrapf(err, \"finding updated task '%s'\", trh.taskId))\n\t}\n\tif refreshedTask == nil {\n\t\treturn gimlet.MakeJSONErrorResponder(gimlet.ErrorResponse{\n\t\t\tStatusCode: http.StatusNotFound,\n\t\t\tMessage: fmt.Sprintf(\"task '%s' not found\", trh.taskId),\n\t\t})\n\t}\n\n\ttaskModel := &model.APITask{}\n\terr = taskModel.BuildFromService(ctx, refreshedTask, &model.APITaskArgs{IncludeProjectIdentifier: true, IncludeAMI: true})\n\tif err != nil {\n\t\treturn gimlet.MakeJSONInternalErrorResponder(errors.Wrapf(err, \"converting task '%s' to API model\", trh.taskId))\n\t}\n\treturn gimlet.NewJSONResponse(taskModel)\n}", "func (gWal *GenericWAL) Recover() error {\n entryCh := make(chan *Entry)\n quitCh := make(chan struct{})\n defer close(quitCh)\n go func() {\n if err := gWal.wal.StartRecovery(entryCh, quitCh); err != nil {\n glog.Errorf(\"could not recover state from local wal :: %v\", err)\n entryCh <- nil\n }\n }()\n\n count := 0\n for entry := range entryCh {\n if entry == nil {\n glog.Errorf(\"wal recovery channel is closed unexpectedly\")\n return fmt.Errorf(\"wal error\")\n }\n count++\n\n switch entry.Type {\n case CTypeDone:\n glog.Infof(\"wal recovery is complete because last record is read\")\n close(entryCh)\n\n case CTypeData:\n gWal.updateRecoveryMap(false /* checkpoint */, int64(entry.LSN),\n entry.Data)\n glog.V(1).Infof(\"recovered a delta record with lsn %v\", entry.LSN)\n\n case CTypeCheckpoint:\n gWal.updateRecoveryMap(true /* checkpoint */, -1, entry.Data)\n glog.V(1).Infof(\"recovered a checkpoint record with lsn %v\", entry.LSN)\n\n case CTypeError:\n glog.Errorf(\"wal recovery encountered an unrecoverable error\")\n return fmt.Errorf(\"wal error\")\n\n default:\n glog.Errorf(\"wal recovery received an unknown or invalid record\")\n return fmt.Errorf(\"wal error\")\n }\n }\n\n return nil\n}", "func (s *Stargate) Recover() {\n\tif err := recover(); err != nil {\n\t\tfmt.Printf(\"Recovered from panic: %v\\n\", err)\n\t}\n}", "func (srv *PBServer) RecoverFromPeer(peer int, args *RecoveryArgs) {\n\treply := new(RecoveryReply)\n\tok := srv.sendRecovery(peer, args, reply)\n\n\tsrv.mu.Lock()\n\tdefer srv.mu.Unlock()\n\n\tsuccess := ok && reply.Success && srv.status == RECOVERING\n\n\tif success && reply.View >= srv.currentView {\n\t\tlog.Printf(\"Node %v - will recover with commit index %d and op index %d and log %v.\\n\", srv.me, srv.commitIndex, srv.opIndex, srv.log)\n\n\t\tif reply.View == srv.currentView {\n\t\t\tfor i := len(srv.log); i < len(reply.Entries); i++ {\n\t\t\t\tsrv.opIndex++\n\t\t\t\tsrv.log = append(srv.log, reply.Entries[i])\n\t\t\t}\n\t\t} else {\n\t\t\tsrv.log = reply.Entries\n\t\t}\n\n\t\tsrv.status = NORMAL\n\t\tsrv.opIndex = len(reply.Entries) - 1\n\t\tsrv.commitIndex = reply.PrimaryCommit\n\t\tsrv.currentView = reply.View\n\t\tsrv.lastNormalView = reply.View\n\n\t\tlog.Printf(\"Node %v - recovered with commit index %d and op index %d and log %v.\\n\", srv.me, srv.commitIndex, srv.opIndex, srv.log)\n\n\t\tgo srv.prepareUncommittedOperations()\n\t}\n}", "func (r *rpcClient) Recover(ctx context.Context) error {\n\tctx, span := r.startSpanFromContext(ctx, \"sb.rpcClient.Recover\")\n\tdefer span.End()\n\n\t_ = r.Close()\n\treturn r.ensureConn(ctx)\n}", "func (srv *PBServer) StartRecovery() {\n\tsrv.mu.Lock()\n\tdefer srv.mu.Unlock()\n\n\tif srv.status != NORMAL {\n\t\tlog.Printf(\"Node %v - not in normal status (view: %v op: %v commit: %v, status: %d)\", srv.me, srv.currentView, srv.opIndex, srv.commitIndex, srv.status)\n\t\treturn\n\t}\n\n\tlog.Printf(\"Node %v - Recovering (view: %v op: %v commit: %v)\", srv.me, srv.currentView, srv.opIndex, srv.commitIndex)\n\n\tsrv.status = RECOVERING\n\n\targs := &RecoveryArgs{View: srv.currentView, Server: srv.me}\n\n\t// Send recovery requests to all peers.\n\tfor peer := range srv.peers {\n\t\tif peer != srv.me {\n\t\t\tgo srv.RecoverFromPeer(peer, args)\n\t\t}\n\t}\n}", "func Recover(cb func()) (err error) {\n\tdefer func() {\n\t\tif e := recover(); e != nil {\n\t\t\terr = errors.E(\"panic %v: %v\", e, string(debug.Stack()))\n\t\t}\n\t}()\n\tcb()\n\treturn nil\n}", "func PanicRecover(h juggler.Handler, vars *expvar.Map) juggler.Handler {\n\treturn juggler.HandlerFunc(func(ctx context.Context, c *juggler.Conn, m message.Msg) {\n\t\tdefer func() {\n\t\t\tif e := recover(); e != nil {\n\t\t\t\tif vars != nil {\n\t\t\t\t\tvars.Add(\"RecoveredPanics\", 1)\n\t\t\t\t}\n\n\t\t\t\tvar err error\n\t\t\t\tswitch e := e.(type) {\n\t\t\t\tcase error:\n\t\t\t\t\terr = e\n\t\t\t\tdefault:\n\t\t\t\t\terr = fmt.Errorf(\"%v\", e)\n\t\t\t\t}\n\t\t\t\tc.Close(err)\n\t\t\t}\n\t\t}()\n\t\th.Handle(ctx, c, m)\n\t})\n}", "func (r NopReporter) Recover(ctx context.Context) { _ = recover() }", "func (p *AuroraAdminClient) DeleteRecoveryTasks(ctx context.Context, query *TaskQuery) (r *Response, err error) {\n var _args329 AuroraAdminDeleteRecoveryTasksArgs\n _args329.Query = query\n var _result330 AuroraAdminDeleteRecoveryTasksResult\n if err = p.Client_().Call(ctx, \"deleteRecoveryTasks\", &_args329, &_result330); err != nil {\n return\n }\n return _result330.GetSuccess(), nil\n}", "func (p *AuroraAdminClient) DeleteRecoveryTasks(ctx context.Context, query *TaskQuery) (r *Response, err error) {\n var _args379 AuroraAdminDeleteRecoveryTasksArgs\n _args379.Query = query\n var _result380 AuroraAdminDeleteRecoveryTasksResult\n var meta thrift.ResponseMeta\n meta, err = p.Client_().Call(ctx, \"deleteRecoveryTasks\", &_args379, &_result380)\n p.SetLastResponseMeta_(meta)\n if err != nil {\n return\n }\n return _result380.GetSuccess(), nil\n}", "func (e *bcsExecutor) RestartTasks() error {\n\tfor _, task := range e.tasks {\n\t\tblog.Infof(\"reload task %s start...\", task.TaskId)\n\t\terr := e.procDaemon.RestartProcess(task.TaskId)\n\t\tif err != nil {\n\t\t\tblog.Errorf(\"reload process %s error %s\", task.TaskId, err.Error())\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (m *TaskManager) ResetOverdueTask() {\n\tm.WIPTable.Range(func(key, value interface{}) bool {\n\t\tif t := value.(*Task); t.IsTimeout() {\n\t\t\tif t.LifeCycle != WIP {\n\t\t\t\tlog.Logger.Fatalf(\"the LifeCycle of the task under check is %d, but `WIP` is expected\", t.LifeCycle)\n\t\t\t}\n\t\t\tt.LifeCycle = READY\n\t\t\tm.ReadyQueue <- t\n\t\t\tm.WIPTable.Delete(key)\n\t\t\tlog.Logger.WithFields(logrus.Fields{\n\t\t\t\t\"ID\": t.ID,\n\t\t\t\t\"TaskType\": t.TaskType,\n\t\t\t}).Warn(\"reset an overdue task\")\n\t\t\treturn false\n\t\t}\n\t\treturn true\n\t})\n}", "func (tr *TaskRunner) ForceRestart(ctx context.Context, event *structs.TaskEvent, failure bool) error {\n\ttr.logger.Trace(\"Force restart requested\", \"failure\", failure, \"event\", event.GoString())\n\n\ttaskState := tr.TaskState()\n\tif taskState == nil {\n\t\treturn ErrTaskNotRunning\n\t}\n\n\ttr.stateLock.Lock()\n\tlocalState := tr.localState.Copy()\n\ttr.stateLock.Unlock()\n\n\tif localState == nil {\n\t\treturn ErrTaskNotRunning\n\t}\n\n\tswitch taskState.State {\n\tcase structs.TaskStatePending:\n\t\treturn ErrTaskNotRunning\n\n\tcase structs.TaskStateDead:\n\t\t// Tasks that are in the \"dead\" state are only allowed to restart if\n\t\t// their Run() method is still active.\n\t\tif localState.RunComplete {\n\t\t\treturn ErrTaskNotRunning\n\t\t}\n\t}\n\n\treturn tr.restartImpl(ctx, event, failure)\n}", "func (pr PermissiveRecoverer) Recover(message string) error {\n\tlog.Print(message)\n\n\tif pr.RecoveryCommand != \"\" {\n\t\tcmd := exec.Command(pr.RecoveryCommand)\n\t\tif err := cmd.Run(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (_DelegatableDai *DelegatableDaiCaller) Recover(opts *bind.CallOpts, hash [32]byte, sig []byte) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _DelegatableDai.contract.Call(opts, out, \"recover\", hash, sig)\n\treturn *ret0, err\n}", "func (e *Eval) retryTask(ctx context.Context, f *Flow, resources reflow.Resources, retryType, msg string) (*sched.Task, error) {\n\t// Apply ExecReset so that the exec can be resubmitted to the scheduler with the flow's\n\t// exec runtime parameters reset.\n\tf.ExecReset()\n\tcappedR, capped, err := e.capMemory(resources)\n\tswitch {\n\tcase err != nil:\n\t\treturn nil, err\n\tcase capped:\n\t\te.Log.Printf(\"flow %s: retryTask (reason: %s): capping resources from %s to %s (max available %s)\", f.Digest().Short(), retryType, resources, cappedR, e.MaxResources)\n\t\tresources.Set(cappedR)\n\t}\n\te.Mutate(f, SetReserved(resources), Execing)\n\ttask := e.newTask(f)\n\te.Log.Printf(\"flow %s: retryTask (reason: %s): re-submitting task with %s\", f.Digest().Short(), retryType, msg)\n\te.Scheduler.Submit(task)\n\treturn task, e.taskWait(ctx, f, task)\n}", "func (t *Task) Exec(agent *Agent) {\n\tdefer func() {\n\t\tif e := recover(); e != nil {\n\n\t\t\t//todo send task status to DCMS-agent\n\t\t\t// log.Warningf(\"run task: %s jobname: failed : %s\", t.TaskId, t.Job.Name, e)\n\t\t\tts := &TaskStatus{\n\t\t\t\tTaskPtr: t,\n\t\t\t\tCommand: nil,\n\t\t\t\tStatus: StatusFailed,\n\t\t\t\tCreateAt: time.Now().Unix(),\n\t\t\t\tErr: fmt.Errorf(\"run task: %s jobname: failed : %s\", t.TaskId, t.Job.Name, e),\n\t\t\t}\n\n\t\t\terrstr := fmt.Sprintf(\"%s\", e)\n\t\t\tif errstr == \"signal: killed\" {\n\t\t\t\tts.Status = StatusKilled\n\t\t\t}\n\t\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\t}\n\t}()\n\n\tvar ts *TaskStatus\n\tvar err error\n\t// log.Info(\"task run Exec function in goroutine\")\n\n\tt.genLogFile()\n\t// check file signature\n\ttmp_md5 := util.Md5File(t.Job.Executor)\n\tif t.Job.Signature != tmp_md5 {\n\t\tts = &TaskStatus{\n\t\t\tTaskPtr: t,\n\t\t\tCommand: nil,\n\t\t\tStatus: StatusFailed,\n\t\t\tCreateAt: time.Now().Unix(),\n\t\t\tErr: fmt.Errorf(\"cronjob: %s executor: %s signature:%s does't match db's sig:%s\", t.Job.Name, t.Job.Executor, tmp_md5, t.Job.Signature),\n\t\t}\n\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\treturn\n\t} else {\n\t\tlog.Info(\"cronjob signature match for \", t.Job.Name, t.Job.ExecutorFlags)\n\t}\n\n\tvar u *user.User\n\tu, err = user.Lookup(t.Job.Runner)\n\tif err != nil {\n\t\t// log.Warningf(\"user %s not exists, task %s quit \", err, t.TaskId)\n\t\tts = &TaskStatus{\n\t\t\tTaskPtr: t,\n\t\t\tCommand: nil,\n\t\t\tStatus: StatusFailed,\n\t\t\tCreateAt: time.Now().Unix(),\n\t\t\tErr: fmt.Errorf(\"user %s not exists, task %s quit \", err, t.TaskId),\n\t\t}\n\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\treturn\n\t}\n\n\tvar uid int\n\tuid, err = strconv.Atoi(u.Uid)\n\tif err != nil {\n\t\t// log.Warningf(\"uid %s conver to int failed \", uid)\n\t\tts = &TaskStatus{\n\t\t\tTaskPtr: t,\n\t\t\tCommand: nil,\n\t\t\tStatus: StatusFailed,\n\t\t\tCreateAt: time.Now().Unix(),\n\t\t\tErr: fmt.Errorf(\"uid %s conver to int failed \", uid),\n\t\t}\n\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\treturn\n\t}\n\n\t// chown log file to specific t.Job.Runner user\n\tif err = t.logfile.Chown(uid, uid); err != nil {\n\t\t// log.Warningf(\"chown logfile: %s to uid: %s failed, %s\", t.logfile.Name(), u.Uid, err)\n\t\tt.logfile = nil\n\t}\n\tvar cmd *exec.Cmd\n\tif t.Job.Executor != \"\" && t.Job.ExecutorFlags != \"\" {\n\t\tcmd = exec.Command(t.Job.Executor, t.Job.ExecutorFlags)\n\t} else if t.Job.Executor != \"\" && t.Job.ExecutorFlags == \"\" {\n\t\tcmd = exec.Command(t.Job.Executor)\n\t} else {\n\t\tts = &TaskStatus{\n\t\t\tTaskPtr: t,\n\t\t\tCommand: cmd,\n\t\t\tStatus: StatusFailed,\n\t\t\tCreateAt: time.Now().Unix(),\n\t\t\tErr: fmt.Errorf(\"job %s must have Executor \", t.Job.Name),\n\t\t}\n\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\treturn\n\t}\n\n\tcmd.SysProcAttr = &syscall.SysProcAttr{}\n\tcmd.SysProcAttr.Credential = &syscall.Credential{Uid: uint32(uid)}\n\tcmd.SysProcAttr.Setsid = true\n\t// Pdeathsig only valid on linux system\n\t//\n\tcmd.SysProcAttr.Pdeathsig = syscall.SIGUSR1\n\n\tcmd.Stderr = t.logfile\n\tcmd.Stdout = t.logfile\n\n\tif err = cmd.Start(); err != nil {\n\t\t// log.Warningf(\"taskid:%s cmd Start failed: %s\", t.TaskId, err)\n\t\tts = &TaskStatus{\n\t\t\tTaskPtr: t,\n\t\t\tCommand: cmd,\n\t\t\tStatus: StatusFailed,\n\t\t\tCreateAt: time.Now().Unix(),\n\t\t\tErr: fmt.Errorf(\"taskid:%s cmd Start failed: %s\", t.TaskId, err),\n\t\t}\n\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\treturn\n\t}\n\n\tts = &TaskStatus{\n\t\tTaskPtr: t,\n\t\tCommand: cmd,\n\t\tStatus: StatusRunning,\n\t\tCreateAt: time.Now().Unix(),\n\t\tErr: nil,\n\t}\n\tt.Job.Dcms.JobStatusChan <- ts\n\t// send cmd.process to dcms-agent\n\n\tif err = cmd.Wait(); err != nil {\n\t\t// log.Warningf(\"taskid:%s cmd Wait failed: %s\", t.TaskId, err)\n\t\tts = &TaskStatus{\n\t\t\tTaskPtr: t,\n\t\t\tCommand: cmd,\n\t\t\tStatus: StatusFailed,\n\t\t\tCreateAt: time.Now().Unix(),\n\t\t\tErr: fmt.Errorf(\"taskid:%s cmd Wait failed: %s\", t.TaskId, err),\n\t\t}\n\t\terrstr := fmt.Sprintf(\"%s\", err.Error())\n\t\tif errstr == \"signal: killed\" {\n\t\t\tts.Status = StatusKilled\n\t\t}\n\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\treturn\n\t}\n\t// log.Warning(\"task run DONE\")\n\tts = &TaskStatus{\n\t\tTaskPtr: t,\n\t\tCommand: cmd,\n\t\tStatus: StatusSuccess,\n\t\tCreateAt: time.Now().Unix(),\n\t\tErr: nil,\n\t}\n\tt.Job.Dcms.JobStatusChan <- ts\n\treturn\n}", "func RestartVersion(versionId string, taskIds []string, abortInProgress bool, caller string) error {\n\tif abortInProgress {\n\t\tif err := task.AbortTasksForVersion(versionId, taskIds, caller); err != nil {\n\t\t\treturn errors.WithStack(err)\n\t\t}\n\t}\n\tfinishedTasks, err := task.FindAll(task.ByIdsAndStatus(taskIds, evergreen.CompletedStatuses))\n\tif err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\tfinishedTasks, err = task.AddParentDisplayTasks(finishedTasks)\n\tif err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\t// remove execution tasks in case the caller passed both display and execution tasks\n\t// the functions below are expected to work if just the display task is passed\n\tfor i := len(finishedTasks) - 1; i >= 0; i-- {\n\t\tt := finishedTasks[i]\n\t\tif t.DisplayTask != nil {\n\t\t\tfinishedTasks = append(finishedTasks[:i], finishedTasks[i+1:]...)\n\t\t}\n\t}\n\n\t// archive all the finished tasks\n\ttoArchive := []task.Task{}\n\tfor _, t := range finishedTasks {\n\t\tif !t.IsPartOfSingleHostTaskGroup() { // for single host task groups we don't archive until fully restarting\n\t\t\ttoArchive = append(toArchive, t)\n\t\t}\n\t}\n\tif err = task.ArchiveMany(toArchive); err != nil {\n\t\treturn errors.Wrap(err, \"unable to archive tasks\")\n\t}\n\n\ttype taskGroupAndBuild struct {\n\t\tBuild string\n\t\tTaskGroup string\n\t}\n\t// only need to check one task per task group / build combination\n\ttaskGroupsToCheck := map[taskGroupAndBuild]task.Task{}\n\ttasksToRestart := finishedTasks\n\tif abortInProgress {\n\t\taborted, err := task.Find(task.BySubsetAborted(taskIds))\n\t\tif err != nil {\n\t\t\treturn errors.WithStack(err)\n\t\t}\n\t\tcatcher := grip.NewBasicCatcher()\n\t\tfor _, t := range aborted {\n\t\t\tcatcher.Add(t.SetResetWhenFinished())\n\t\t}\n\t\tif catcher.HasErrors() {\n\t\t\treturn catcher.Resolve()\n\t\t}\n\t}\n\n\trestartIds := []string{}\n\tfor _, t := range tasksToRestart {\n\t\tif t.IsPartOfSingleHostTaskGroup() {\n\t\t\tif err = t.SetResetWhenFinished(); err != nil {\n\t\t\t\treturn errors.Wrapf(err, \"unable to mark '%s' for restart when finished\", t.Id)\n\t\t\t}\n\t\t\ttaskGroupsToCheck[taskGroupAndBuild{\n\t\t\t\tBuild: t.BuildId,\n\t\t\t\tTaskGroup: t.TaskGroup,\n\t\t\t}] = t\n\t\t} else {\n\t\t\t// only hard restart non-single host task group tasks\n\t\t\trestartIds = append(restartIds, t.Id)\n\t\t\tif t.DisplayOnly {\n\t\t\t\trestartIds = append(restartIds, t.ExecutionTasks...)\n\t\t\t}\n\t\t}\n\t}\n\n\tfor tg, t := range taskGroupsToCheck {\n\t\tif err = checkResetSingleHostTaskGroup(&t, caller); err != nil {\n\t\t\treturn errors.Wrapf(err, \"error resetting task group '%s' for build '%s'\", tg.TaskGroup, tg.Build)\n\t\t}\n\t}\n\n\t// Set all the task fields to indicate restarted\n\tif err = MarkTasksReset(restartIds); err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\tfor _, t := range tasksToRestart {\n\t\tif !t.IsPartOfSingleHostTaskGroup() { // this will be logged separately if task group is restarted\n\t\t\tevent.LogTaskRestarted(t.Id, t.Execution, caller)\n\t\t}\n\t}\n\tif err = build.SetBuildStartedForTasks(tasksToRestart, caller); err != nil {\n\t\treturn errors.Wrapf(err, \"error setting builds started\")\n\t}\n\tversion, err := VersionFindOneId(versionId)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"unable to find version\")\n\t}\n\treturn errors.Wrap(version.UpdateStatus(evergreen.VersionStarted), \"unable to change version status\")\n\n}", "func (m TaskManager) LaunchTask(c context.Context, ctl task.Controller) error {\n\tcfg := ctl.Task().(*messages.GitilesTask)\n\n\tctl.DebugLog(\"Repo: %s, Refs: %s\", cfg.Repo, cfg.Refs)\n\tu, err := url.Parse(cfg.Repo)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\twatchedRefs := watchedRefs{}\n\twatchedRefs.init(cfg.GetRefs())\n\n\tvar wg sync.WaitGroup\n\n\tvar heads map[string]string\n\tvar headsErr error\n\twg.Add(1)\n\tgo func() {\n\t\tdefer wg.Done()\n\t\theads, headsErr = loadState(c, ctl.JobID(), u)\n\t}()\n\n\tvar refs map[string]string\n\tvar refsErr error\n\twg.Add(1)\n\tgo func() {\n\t\tdefer wg.Done()\n\t\trefs, refsErr = m.getRefsTips(c, ctl, cfg.Repo, watchedRefs)\n\t}()\n\n\twg.Wait()\n\n\tif headsErr != nil {\n\t\tctl.DebugLog(\"Failed to fetch heads - %s\", headsErr)\n\t\treturn fmt.Errorf(\"failed to fetch heads: %v\", headsErr)\n\t}\n\tif refsErr != nil {\n\t\tctl.DebugLog(\"Failed to fetch refs - %s\", refsErr)\n\t\treturn fmt.Errorf(\"failed to fetch refs: %v\", refsErr)\n\t}\n\n\trefsChanged := 0\n\n\t// Delete all previously known refs which are either no longer watched or no\n\t// longer exist in repo.\n\tfor ref := range heads {\n\t\tswitch {\n\t\tcase !watchedRefs.hasRef(ref):\n\t\t\tctl.DebugLog(\"Ref %s is no longer watched\", ref)\n\t\t\tdelete(heads, ref)\n\t\t\trefsChanged++\n\t\tcase refs[ref] == \"\":\n\t\t\tctl.DebugLog(\"Ref %s deleted\", ref)\n\t\t\tdelete(heads, ref)\n\t\t\trefsChanged++\n\t\t}\n\t}\n\t// For determinism, sort keys of current refs.\n\tsortedRefs := make([]string, 0, len(refs))\n\tfor ref := range refs {\n\t\tsortedRefs = append(sortedRefs, ref)\n\t}\n\tsort.Strings(sortedRefs)\n\n\temittedTriggers := 0\n\tmaxTriggersPerInvocation := m.maxTriggersPerInvocation\n\tif maxTriggersPerInvocation == 0 {\n\t\tmaxTriggersPerInvocation = defaultMaxTriggersPerInvocation\n\t}\n\t// Note, that current `refs` contain only watched refs (see getRefsTips).\n\tfor _, ref := range sortedRefs {\n\t\tnewHead := refs[ref]\n\t\toldHead, existed := heads[ref]\n\t\tswitch {\n\t\tcase !existed:\n\t\t\tctl.DebugLog(\"Ref %s is new: %s\", ref, newHead)\n\t\tcase oldHead != newHead:\n\t\t\tctl.DebugLog(\"Ref %s updated: %s => %s\", ref, oldHead, newHead)\n\t\tdefault:\n\t\t\t// No change.\n\t\t\tcontinue\n\t\t}\n\t\theads[ref] = newHead\n\t\trefsChanged++\n\t\temittedTriggers++\n\t\t// TODO(tandrii): actually look at commits between current and previously\n\t\t// known tips of each ref.\n\t\t// In current (v1) engine, all triggers emitted around the same time will\n\t\t// result in just 1 invocation of each triggered job. Therefore,\n\t\t// passing just HEAD's revision is good enough.\n\t\t// For the same reason, only 1 of the refs will actually be processed if\n\t\t// several refs changed at the same time.\n\t\tctl.EmitTrigger(c, &internal.Trigger{\n\t\t\tId: fmt.Sprintf(\"%s/+/%s@%s\", cfg.Repo, ref, newHead),\n\t\t\tTitle: newHead,\n\t\t\tUrl: fmt.Sprintf(\"%s/+/%s\", cfg.Repo, newHead),\n\t\t\tPayload: &internal.Trigger_Gitiles{\n\t\t\t\tGitiles: &api.GitilesTrigger{Repo: cfg.Repo, Ref: ref, Revision: newHead},\n\t\t\t},\n\t\t})\n\n\t\t// Safeguard against too many changes such as the first run after\n\t\t// config change to watch many more refs than before.\n\t\tif emittedTriggers >= maxTriggersPerInvocation {\n\t\t\tctl.DebugLog(\"Emitted %d triggers, postponing the rest\", emittedTriggers)\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif refsChanged == 0 {\n\t\tctl.DebugLog(\"No changes detected\")\n\t} else {\n\t\tctl.DebugLog(\"%d refs changed\", refsChanged)\n\t\t// Force save to ensure triggers are actually emitted.\n\t\tif err := ctl.Save(c); err != nil {\n\t\t\t// At this point, triggers have not been sent, so bail now and don't save\n\t\t\t// the refs' heads newest values.\n\t\t\treturn err\n\t\t}\n\t\tif err := saveState(c, ctl.JobID(), u, heads); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tctl.DebugLog(\"Saved %d known refs\", len(heads))\n\t}\n\n\tctl.State().Status = task.StatusSucceeded\n\treturn nil\n}", "func (recovery *Recovery) RecoverRegions(ctx context.Context) (err error) {\n\teg, ectx := errgroup.WithContext(ctx)\n\ttotalRecoveredStores := len(recovery.RecoveryPlan)\n\tworkers := utils.NewWorkerPool(uint(mathutil.Min(totalRecoveredStores, common.MaxStoreConcurrency)), \"Recover Regions\")\n\n\tfor storeId, plan := range recovery.RecoveryPlan {\n\t\tif err := ectx.Err(); err != nil {\n\t\t\tbreak\n\t\t}\n\n\t\tstoreAddr := getStoreAddress(recovery.allStores, storeId)\n\t\trecoveryPlan := plan\n\t\trecoveryStoreId := storeId\n\t\tworkers.ApplyOnErrorGroup(eg, func() error {\n\t\t\trecoveryClient, conn, err := recovery.newRecoveryClient(ectx, storeAddr)\n\t\t\tif err != nil {\n\t\t\t\tlog.Error(\"create tikv client failed\", zap.Uint64(\"store id\", recoveryStoreId))\n\t\t\t\treturn errors.Trace(err)\n\t\t\t}\n\t\t\tdefer conn.Close()\n\t\t\tlog.Info(\"send recover region to tikv\", zap.String(\"tikv address\", storeAddr), zap.Uint64(\"store id\", recoveryStoreId))\n\t\t\tstream, err := recoveryClient.RecoverRegion(ectx)\n\t\t\tif err != nil {\n\t\t\t\tlog.Error(\"create recover region failed\", zap.Uint64(\"store id\", recoveryStoreId))\n\t\t\t\treturn errors.Trace(err)\n\t\t\t}\n\n\t\t\t// for a TiKV, send the stream\n\t\t\tfor _, s := range recoveryPlan {\n\t\t\t\tif err = stream.Send(s); err != nil {\n\t\t\t\t\tlog.Error(\"send recover region failed\", zap.Error(err))\n\t\t\t\t\treturn errors.Trace(err)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\treply, err := stream.CloseAndRecv()\n\t\t\tif err != nil {\n\t\t\t\tlog.Error(\"close the stream failed\")\n\t\t\t\treturn errors.Trace(err)\n\t\t\t}\n\t\t\trecovery.progress.Inc()\n\t\t\tlog.Info(\"recover region execution success\", zap.Uint64(\"store id\", reply.GetStoreId()))\n\t\t\treturn nil\n\t\t})\n\t}\n\t// Wait for all TiKV instances force leader and wait apply to last log.\n\treturn eg.Wait()\n}", "func (m *Master) GetTask(_ *ExampleArgs, reply *GetTaskReply) error {\n\tswitch m.masterState {\n\tcase newMaster:\n\t\tfor i, task := range m.mapTask {\n\t\t\tif task.State == initialState {\n\t\t\t\treply.Task.Type_ = task.Type_\n\t\t\t\treply.Task.Filename = task.Filename\n\t\t\t\treply.Task.Id = task.Id\n\t\t\t\treply.Task.NReduce = task.NReduce\n\t\t\t\treply.Flag = 0\n\t\t\t\treply.Task.State = task.State\n\n\t\t\t\tm.mapTask[i].State = inProgress\n\t\t\t\tm.mapTask[i].Time = time.Now()\n\t\t\t\t//reply.Task.State=m.mapTask[i].State\n\n\t\t\t\treturn nil\n\t\t\t} else if task.State == inProgress && time.Now().Sub(m.mapTask[i].Time) > time.Duration(5)*time.Second {\n\t\t\t\treply.Task.Type_ = task.Type_\n\t\t\t\treply.Task.Filename = task.Filename\n\t\t\t\treply.Task.Id = task.Id\n\t\t\t\treply.Task.NReduce = task.NReduce\n\t\t\t\treply.Task.State = task.State\n\t\t\t\treply.Flag = 0\n\n\t\t\t\tm.mapTask[i].State = inProgress\n\t\t\t\tm.mapTask[i].Time = time.Now()\n\t\t\t\t//reply.Task.State=m.mapTask[i].State\n\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t\treply.Flag = 1 // map not finished but in progress\n\tcase completeMap:\n\t\tfor i, task := range m.reduceTask {\n\t\t\tif task.State == initialState {\n\t\t\t\treply.Task.Type_ = task.Type_\n\t\t\t\treply.Task.Filename = task.Filename\n\t\t\t\treply.Task.Id = task.Id\n\t\t\t\treply.Task.NReduce = task.NReduce\n\t\t\t\treply.Flag = 0\n\t\t\t\treply.Task.Files = task.Files\n\t\t\t\treply.Task.State = task.State\n\n\t\t\t\tm.reduceTask[i].State = inProgress\n\t\t\t\tm.reduceTask[i].Time = time.Now()\n\t\t\t\t//reply.Task.State=m.mapTask[i].State\n\n\t\t\t\treturn nil\n\t\t\t} else if task.State == inProgress && time.Now().Sub(m.reduceTask[i].Time) > time.Duration(5)*time.Second {\n\t\t\t\treply.Task.Type_ = task.Type_\n\t\t\t\treply.Task.Filename = task.Filename\n\t\t\t\treply.Task.Id = task.Id\n\t\t\t\treply.Task.NReduce = task.NReduce\n\t\t\t\treply.Flag = 0\n\t\t\t\treply.Task.Files = task.Files\n\t\t\t\treply.Task.State = task.State\n\n\t\t\t\tm.reduceTask[i].State = inProgress\n\t\t\t\tm.reduceTask[i].Time = time.Now()\n\t\t\t\t//reply.Task.State=m.mapTask[i].State\n\n\t\t\t\treturn nil\n\n\t\t\t}\n\t\t}\n\t\treply.Flag = 1 // reduce not finished but in progress\n\tcase completeReduce:\n\t\treply.Flag = 2 // all task have been finished\n\n\t}\n\n\treturn nil\n}", "func (p *AuroraAdminClient) QueryRecovery(ctx context.Context, query *TaskQuery) (r *Response, err error) {\n var _args327 AuroraAdminQueryRecoveryArgs\n _args327.Query = query\n var _result328 AuroraAdminQueryRecoveryResult\n if err = p.Client_().Call(ctx, \"queryRecovery\", &_args327, &_result328); err != nil {\n return\n }\n return _result328.GetSuccess(), nil\n}", "func RecoverCluster(conf *Config, fsm FSM, logs LogStore, stable StableStore,\n\tsnaps SnapshotStore, trans Transport, configuration Configuration) error {\n\t// Validate the Raft server config.\n\tif err := ValidateConfig(conf); err != nil {\n\t\treturn err\n\t}\n\n\t// Sanity check the Raft peer configuration.\n\tif err := checkConfiguration(configuration); err != nil {\n\t\treturn err\n\t}\n\n\t// Make sure the cluster is in a clean state.\n\thasState, err := HasExistingState(logs, stable, snaps)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to check for existing state: %v\", err)\n\t}\n\tif !hasState {\n\t\treturn fmt.Errorf(\"refused to recover cluster with no initial state, this is probably an operator error\")\n\t}\n\n\t// Attempt to restore any snapshotStore we find, newest to oldest.\n\tvar (\n\t\tsnapshotIndex uint64\n\t\tsnapshotTerm uint64\n\t)\n\tsnapshots, err := snaps.List()\n\tif err != nil {\n\t\tklog.Errorf(fmt.Sprintf(\"failed to list snapshotStore err:%v\", err))\n\t\treturn err\n\t}\n\t// Try to load in order of newest to oldest\n\tfor _, snapshot := range snapshots {\n\t\tif !conf.NoSnapshotRestoreOnStart {\n\t\t\t_, source, err := snaps.Open(snapshot.ID)\n\t\t\tif err != nil {\n\t\t\t\tklog.Errorf(fmt.Sprintf(\"failed to open snapshot id:%s err:%v\", snapshot.ID, err))\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif err := fsm.Restore(source); err != nil {\n\t\t\t\tsource.Close()\n\t\t\t\tklog.Errorf(fmt.Sprintf(\"failed to restore snapshot id:%s err:%v\", snapshot.ID, err))\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tsource.Close()\n\t\t\tklog.Infof(fmt.Sprintf(\"restored from snapshot id:%s\", snapshot.ID))\n\t\t}\n\n\t\tsnapshotIndex = snapshot.Index\n\t\tsnapshotTerm = snapshot.Term\n\t\tbreak\n\t}\n\tif len(snapshots) > 0 && (snapshotIndex == 0 || snapshotTerm == 0) {\n\t\treturn fmt.Errorf(\"failed to restore any of the available snapshotStore\")\n\t}\n\n\t// The snapshot information is the best known end point for the data\n\t// until we play back the Raft log entries.\n\tlastIndex := snapshotIndex\n\tlastTerm := snapshotTerm\n\t// Apply any Raft log entries from the snapshot index to last log index.\n\tlastLogIndex, err := logs.LastIndex()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to find last log: %v\", err)\n\t}\n\tfor index := snapshotIndex + 1; index <= lastLogIndex; index++ {\n\t\tvar entry pb.Log\n\t\tif err = logs.GetLog(index, &entry); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to get log at index %d: %v\", index, err)\n\t\t}\n\t\tif entry.Type == pb.LogType_COMMAND {\n\t\t\t_ = fsm.Apply(&entry)\n\t\t}\n\n\t\tlastIndex = entry.Index\n\t\tlastTerm = entry.Term\n\t}\n\n\tif lastIndex != lastLogIndex {\n\t\tklog.Fatalf(fmt.Sprintf(\"lastIndex:%d should be equal to lastLogIndex:%d\", lastIndex, lastLogIndex))\n\t}\n\n\t// snapshot fsm\n\tsnapshot, err := fsm.Snapshot()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to snapshot FSM: %v\", err)\n\t}\n\tsink, err := snaps.Create(lastIndex, lastTerm, configuration, 1)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to create snapshot: %v\", err)\n\t}\n\tif err = snapshot.Persist(sink); err != nil {\n\t\treturn fmt.Errorf(\"failed to persist snapshot: %v\", err)\n\t}\n\tif err = sink.Close(); err != nil {\n\t\treturn fmt.Errorf(\"failed to finalize snapshot: %v\", err)\n\t}\n\t// compact logs\n\tfirstLogIndex, err := logs.FirstIndex()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get first log index: %v\", err)\n\t}\n\tif err := logs.DeleteRange(firstLogIndex, lastLogIndex); err != nil {\n\t\treturn fmt.Errorf(\"log compaction failed: %v\", err)\n\t}\n\n\treturn nil\n}", "func (mc *mgmtClient) recover(ctx context.Context) error {\n\tmc.clientMu.Lock()\n\tdefer mc.clientMu.Unlock()\n\n\tctx, span := mc.startSpanFromContext(ctx, string(tracing.SpanNameRecover))\n\tdefer span.End()\n\n\tif mc.rpcLink != nil {\n\t\tif err := mc.rpcLink.Close(ctx); err != nil {\n\t\t\ttab.For(ctx).Debug(fmt.Sprintf(\"Error while closing old link in recovery: %s\", err.Error()))\n\t\t}\n\t\tmc.rpcLink = nil\n\t}\n\n\tif _, err := mc.getLinkWithoutLock(ctx); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func recoverWorkflow(ctx workflow.Context, params Params) error {\n\tlogger := workflow.GetLogger(ctx)\n\tlogger.Info(\"Recover workflow started.\")\n\n\tao := workflow.ActivityOptions{\n\t\tScheduleToStartTimeout: 10 * time.Minute,\n\t\tStartToCloseTimeout: 10 * time.Minute,\n\t\tHeartbeatTimeout: time.Second * 30,\n\t}\n\tctx = workflow.WithActivityOptions(ctx, ao)\n\n\tvar result ListOpenExecutionsResult\n\terr := workflow.ExecuteActivity(ctx, listOpenExecutions, params.Type).Get(ctx, &result)\n\tif err != nil {\n\t\tlogger.Error(\"Failed to list open workflow executions.\", zap.Error(err))\n\t\treturn err\n\t}\n\n\tconcurrency := 1\n\tif params.Concurrency > 0 {\n\t\tconcurrency = params.Concurrency\n\t}\n\n\tif result.Count < concurrency {\n\t\tconcurrency = result.Count\n\t}\n\n\tbatchSize := result.Count / concurrency\n\tif result.Count%concurrency != 0 {\n\t\tbatchSize++\n\t}\n\n\t// Setup retry policy for recovery activity\n\tinfo := workflow.GetInfo(ctx)\n\texpiration := time.Duration(info.ExecutionStartToCloseTimeoutSeconds) * time.Second\n\tretryPolicy := &cadence.RetryPolicy{\n\t\tInitialInterval: time.Second,\n\t\tBackoffCoefficient: 2,\n\t\tMaximumInterval: 10 * time.Second,\n\t\tExpirationInterval: expiration,\n\t\tMaximumAttempts: 100,\n\t}\n\tao = workflow.ActivityOptions{\n\t\tScheduleToStartTimeout: expiration,\n\t\tStartToCloseTimeout: expiration,\n\t\tHeartbeatTimeout: time.Second * 30,\n\t\tRetryPolicy: retryPolicy,\n\t}\n\tctx = workflow.WithActivityOptions(ctx, ao)\n\n\tdoneCh := workflow.NewChannel(ctx)\n\tfor i := 0; i < concurrency; i++ {\n\t\tstartIndex := i * batchSize\n\n\t\tworkflow.Go(ctx, func(ctx workflow.Context) {\n\t\t\terr = workflow.ExecuteActivity(ctx, recoverExecutions, result.ID, startIndex, batchSize).Get(ctx, nil)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Error(\"Recover executions failed.\", zap.Int(\"StartIndex\", startIndex), zap.Error(err))\n\t\t\t} else {\n\t\t\t\tlogger.Info(\"Recover executions completed.\", zap.Int(\"StartIndex\", startIndex))\n\t\t\t}\n\n\t\t\tdoneCh.Send(ctx, \"done\")\n\t\t})\n\t}\n\n\tfor i := 0; i < concurrency; i++ {\n\t\tdoneCh.Receive(ctx, nil)\n\t}\n\n\tlogger.Info(\"Workflow completed.\", zap.Int(\"Result\", result.Count))\n\n\treturn nil\n}", "func (ra *RecoverableAction) Recover(r Recoverable, err interface{}) {\n\tif ra == r {\n\t\tlog.Printf(\"Recovering error '%v'!\", err)\n\n\t\tra.replyChan <- \"Recovered\"\n\n\t\tgo ra.backend()\n\t}\n}", "func (conf *Confirmer) StartConfirmerTask(bundleTrytes []Trytes) (chan *ConfirmerUpdate, func(), error) {\n\tconf.runningMutex.Lock()\n\tdefer conf.runningMutex.Unlock()\n\n\tif conf.running {\n\t\treturn nil, nil, fmt.Errorf(\"Confirmer task is already running\")\n\t}\n\n\ttail, err := utils.TailFromBundleTrytes(bundleTrytes)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tbundleHash := tail.Bundle\n\tnowis := time.Now()\n\n\t// no need to lock state because no routine is running\n\tconf.lastBundleTrytes = bundleTrytes\n\tconf.bundleHash = bundleHash\n\tconf.nextForceReattachTime = nowis.Add(time.Duration(conf.ForceReattachAfterMin) * time.Minute)\n\tconf.nextPromoTime = nowis\n\tconf.nextTailHashToPromote = tail.Hash\n\tconf.isNotPromotable = false\n\tconf.chanUpdate = make(chan *ConfirmerUpdate, 1) // not to block each time\n\tconf.numAttach = 0\n\tconf.numPromote = 0\n\tconf.totalDurationGTTAMsec = 0\n\tconf.totalDurationATTMsec = 0\n\tif conf.AEC == nil {\n\t\tconf.AEC = &utils.DummyAEC{}\n\t}\n\tif conf.SlowDownThreshold == 0 {\n\t\tconf.SlowDownThreshold = defaultSlowDownThresholdNumGoroutine\n\t}\n\n\t// starting 3 routines\n\tcancelPromoCheck := conf.goPromotabilityCheck()\n\tcancelPromo := conf.goPromote()\n\tcancelReattach := conf.goReattach()\n\n\t// confirmation monitor starts yet another routine\n\tconf.confMon.OnConfirmation(bundleHash, func(nowis time.Time) {\n\t\tconf.postConfirmerUpdate(UPD_CONFIRM, \"\", nil)\n\t})\n\n\tconf.running = true\n\n\treturn conf.chanUpdate, func() {\n\t\tconf.stopConfirmerTask(cancelPromoCheck, cancelPromo, cancelReattach)\n\t\tconf.confMon.CancelConfirmationPolling(bundleHash)\n\t}, nil\n}", "func CleanTask() {\n\tfor taskID, t := range kv.DefaultClient.GetStorage().Tasks {\n\t\tflag := true\n\t\tfor nid := range kv.DefaultClient.GetStorage().Nodes {\n\t\t\tif t.NodeID == nid {\n\t\t\t\tflag = false\n\t\t\t}\n\t\t}\n\t\tif flag {\n\t\t\tif t.Timer {\n\t\t\t\tlog.Info(\"clean timer:\", taskID)\n\t\t\t\tormTimer := new(orm.Timer)\n\t\t\t\tormTimer.ID = taskID\n\t\t\t\tormTimer.Status = false\n\t\t\t\terr := orm.UpdateTimerStatus(ormTimer)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error(err)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tlog.Info(\"clean task:\", taskID)\n\t\t\t\tormTask := new(orm.Task)\n\t\t\t\tormTask.ID = taskID\n\t\t\t\tormTask.Status = \"error\"\n\t\t\t\terr := orm.UpdateTask(ormTask)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error(err)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tkv.DefaultClient.DeleteTask(taskID)\n\t\t}\n\t}\n}", "func (rm *ResourceManager) AutoRecoverResourceGroup(rgName string) ([]int64, error) {\n\trm.rwmutex.Lock()\n\tdefer rm.rwmutex.Unlock()\n\n\tif rm.groups[rgName] == nil {\n\t\treturn nil, merr.WrapErrResourceGroupNotFound(rgName)\n\t}\n\n\tret := make([]int64, 0)\n\n\trm.checkRGNodeStatus(DefaultResourceGroupName)\n\trm.checkRGNodeStatus(rgName)\n\tlackNodesNum := rm.groups[rgName].LackOfNodes()\n\tnodesInDefault := rm.groups[DefaultResourceGroupName].GetNodes()\n\tfor i := 0; i < len(nodesInDefault) && i < lackNodesNum; i++ {\n\t\t//todo: a better way to choose a node with least balance cost\n\t\tnode := nodesInDefault[i]\n\t\terr := rm.unassignNode(DefaultResourceGroupName, node)\n\t\tif err != nil {\n\t\t\t// interrupt transfer, unreachable logic path\n\t\t\treturn ret, err\n\t\t}\n\n\t\terr = rm.groups[rgName].assignNode(node, 0)\n\t\tif err != nil {\n\t\t\t// roll back, unreachable logic path\n\t\t\trm.assignNode(DefaultResourceGroupName, node)\n\t\t\treturn ret, err\n\t\t}\n\n\t\tlog.Info(\"move node from default rg to recover\",\n\t\t\tzap.String(\"targetRG\", rgName),\n\t\t\tzap.Int64(\"nodeID\", node),\n\t\t)\n\n\t\tret = append(ret, node)\n\t}\n\n\treturn ret, nil\n}", "func NewTaskRetry() *TaskRetry {\n\treturn &TaskRetry{}\n}", "func (k *KubernetesExecutor) KillTask(driver bindings.ExecutorDriver, taskId *mesos.TaskID) {\n\tif k.isDone() {\n\t\treturn\n\t}\n\tlog.Infof(\"Kill task %v\\n\", taskId)\n\n\tif !k.isConnected() {\n\t\t//TODO(jdefelice) sent TASK_LOST here?\n\t\tlog.Warningf(\"Ignore kill task because the executor is disconnected\\n\")\n\t\treturn\n\t}\n\n\tk.lock.Lock()\n\tdefer k.lock.Unlock()\n\tk.killPodForTask(driver, taskId.GetValue(), messages.TaskKilled)\n}", "func (krct *keyRegistrationConfirmationTask) RunTask() (interface{}, error) {\n\tlog.Infof(\"Waiting for confirmation for the Key [%x]\", krct.key)\n\tif krct.ctx == nil {\n\t\tkrct.ctx, _ = krct.contextInitializer(krct.timeout)\n\t}\n\n\tid := newEthereumIdentity(krct.centID, krct.contract, krct.config, krct.queue, krct.gethClientFinder, krct.contractProvider)\n\tcontract, err := id.getContract()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tkrct.filterer = contract\n\tfOpts := &bind.FilterOpts{\n\t\tContext: krct.ctx,\n\t\tStart: krct.blockHeight,\n\t}\n\n\tfor {\n\t\titer, err := krct.filterer.FilterKeyAdded(fOpts, [][32]byte{krct.key}, []*big.Int{big.NewInt(int64(krct.keyPurpose))})\n\t\tif err != nil {\n\t\t\treturn nil, centerrors.Wrap(err, \"failed to start filtering key event logs\")\n\t\t}\n\n\t\terr = utils.LookForEvent(iter)\n\t\tif err == nil {\n\t\t\tlog.Infof(\"Received filtered event Key Registration Confirmation for CentrifugeID [%s] and key [%x] with purpose [%d]\\n\", krct.centID.String(), krct.key, krct.keyPurpose)\n\t\t\treturn iter.Event, nil\n\t\t}\n\n\t\tif err != utils.ErrEventNotFound {\n\t\t\treturn nil, err\n\t\t}\n\t\ttime.Sleep(100 * time.Millisecond)\n\t}\n}", "func (tr *TaskRunner) restartImpl(ctx context.Context, event *structs.TaskEvent, failure bool) error {\n\n\t// Check if the task is able to restart based on its state and the type of\n\t// restart event that was triggered.\n\ttaskState := tr.TaskState()\n\tif taskState == nil {\n\t\treturn ErrTaskNotRunning\n\t}\n\n\t// Emit the event since it may take a long time to kill\n\ttr.EmitEvent(event)\n\n\t// Tell the restart tracker that a restart triggered the exit\n\ttr.restartTracker.SetRestartTriggered(failure)\n\n\t// Signal a restart to unblock tasks that are in the \"dead\" state, but\n\t// don't block since the channel is buffered. Only one signal is enough to\n\t// notify the tr.Run() loop.\n\t// The channel must be signaled after SetRestartTriggered is called so the\n\t// tr.Run() loop runs again.\n\tif taskState.State == structs.TaskStateDead {\n\t\tselect {\n\t\tcase tr.restartCh <- struct{}{}:\n\t\tdefault:\n\t\t}\n\t}\n\n\t// Grab the handle to see if the task is still running and needs to be\n\t// killed.\n\thandle := tr.getDriverHandle()\n\tif handle == nil {\n\t\treturn nil\n\t}\n\n\t// Run the pre-kill hooks prior to restarting the task\n\ttr.preKill()\n\n\t// Grab a handle to the wait channel that will timeout with context cancelation\n\t// _before_ killing the task.\n\twaitCh, err := handle.WaitCh(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Kill the task using an exponential backoff in-case of failures.\n\tif _, err := tr.killTask(handle, waitCh); err != nil {\n\t\t// We couldn't successfully destroy the resource created.\n\t\ttr.logger.Error(\"failed to kill task. Resources may have been leaked\", \"error\", err)\n\t}\n\n\tselect {\n\tcase <-waitCh:\n\tcase <-ctx.Done():\n\t}\n\treturn nil\n}", "func (j *TxGuardedJob[Tx, P]) discoverTask(ctx context.Context, labels prometheus.Labels) (task *txGuardedTask[Tx, P], returnedError error) {\n\ttx, err := j.BeginTx()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer func() {\n\t\tif returnedError != nil {\n\t\t\tsqlext.RollbackUnlessCommitted(tx)\n\t\t}\n\t}()\n\n\tpayload, err := j.DiscoverRow(ctx, tx, labels)\n\tif err != nil {\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\t//nolint:errcheck\n\t\t\ttx.Rollback() //avoid the log line generated by sqlext.RollbackUnlessCommitted()\n\t\t}\n\t\treturn nil, err\n\t}\n\n\treturn &txGuardedTask[Tx, P]{\n\t\tTransaction: tx,\n\t\tPayload: payload,\n\t}, nil\n}", "func main() {\n\ttryRecover()\n}", "func (sd *ScanDiapasons) NotifyDpnsTask(ctx context.Context) ([]byte, error) {\n\trequest, err := http.NewRequest(\"POST\", sd.client.Server+\"/api/v1.0/ScanDiapasons.NotifyDpnsTask\", nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\traw, err := sd.client.Do(ctx, request, nil)\n\treturn raw, err\n}", "func (task *QueueTask) Reset(conf *TaskConfig) error {\n\tif conf.Interval <= 0 {\n\t\terrmsg := \"interval is wrong format => must bigger then zero\"\n\t\ttask.taskService.Logger().Debug(fmt.Sprint(\"TaskInfo:Reset \", task, conf, \"error\", errmsg))\n\t\treturn errors.New(errmsg)\n\t}\n\n\t//restart task\n\ttask.Stop()\n\ttask.IsRun = conf.IsRun\n\tif conf.TaskData != nil {\n\t\ttask.TaskData = conf.TaskData\n\t}\n\tif conf.Handler != nil {\n\t\ttask.handler = conf.Handler\n\t}\n\ttask.Interval = conf.Interval\n\ttask.Start()\n\ttask.taskService.Logger().Debug(fmt.Sprint(\"TaskInfo:Reset \", task, conf, \"success\"))\n\treturn nil\n}", "func (syncer *MerkleSyncer) recover(begin, end uint64) {\n\tlogger.WithFields(logrus.Fields{\n\t\t\"begin\": begin,\n\t\t\"end\": end,\n\t}).Info(\"Syncer recover\")\n\n\tch, err := syncer.agent.GetMerkleWrapper(begin, end)\n\tif err != nil {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"begin\": begin,\n\t\t\t\"end\": end,\n\t\t\t\"error\": err,\n\t\t}).Warn(\"get merkle wrapper\")\n\t}\n\n\tfor w := range ch {\n\t\tsyncer.handleMerkleWrapper(w)\n\t}\n}", "func ResetTasks() {\n\tconn := getConnection(\"mflow\")\n\tdb, err := sql.Open(\"godror\", conn.User+\"/\"+conn.Password+\"@\"+conn.ConnectionString)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\tdefer db.Close()\n\ttx, err := db.Begin()\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\tconst command string = `delete mflow.tasks where id_master = :id_master and status = :status`\n\t_, err = tx.Exec(command, sql.Named(\"id_master\", global.IDMaster), sql.Named(\"status\", runningStatus))\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\terr = tx.Commit()\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n}", "func (tsi *TaskerServerImpl) TriggerRepairOnRepairFailed(ctx context.Context, req *fleet.TriggerRepairOnRepairFailedRequest) (resp *fleet.TaskerTasksResponse, err error) {\n\tdefer func() {\n\t\terr = grpcutil.GRPCifyAndLogErr(ctx, err)\n\t}()\n\n\tif err = req.Validate(); err != nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, err.Error())\n\t}\n\tsc, err := tsi.newSwarmingClient(ctx, config.Get(ctx).Swarming.Host)\n\tif err != nil {\n\t\treturn nil, errors.Annotate(err, \"failed to obtain Swarming client\").Err()\n\t}\n\n\tbses, err := botsummary.Get(ctx, req.Selectors)\n\tif err != nil {\n\t\treturn nil, errors.Annotate(err, \"failed to obtain requested bots from datastore\").Err()\n\t}\n\treturn createTasksPerBot(bses, func(bse *botsummary.Entity) (*fleet.TaskerBotTasks, error) {\n\t\treturn triggerRepairOnRepairFailedForBot(ctx, sc, req, bse)\n\t})\n}", "func (db *TaskDB) ReadTask(id string) (eremetic.Task, error) {\n\ttask, err := db.ReadUnmaskedTask(id)\n\n\teremetic.ApplyMask(&task)\n\n\treturn task, err\n}", "func (_m *StateOps) RefreshTaskState() {\n\t_m.Called()\n}", "func (d Dependency) DoRecover() (interface{}, error) {\n\tdLog.V(utils.Debug).Info(\"can not recover dependency now\", \"name\", d.Name, \"version\", d.Version)\n\treturn \"\", nil\n}", "func (n *NoOP) RecoverFromSnapshot(r io.Reader,\n\tfiles []statemachine.SnapshotFile,\n\tdone <-chan struct{}) error {\n\tvar sn NoOP\n\tdata, err := ioutil.ReadAll(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = json.Unmarshal(data, &sn)\n\tif err != nil {\n\t\tpanic(\"failed to unmarshal snapshot\")\n\t}\n\n\treturn nil\n}", "func (p *AuroraAdminClient) QueryRecovery(ctx context.Context, query *TaskQuery) (r *Response, err error) {\n var _args377 AuroraAdminQueryRecoveryArgs\n _args377.Query = query\n var _result378 AuroraAdminQueryRecoveryResult\n var meta thrift.ResponseMeta\n meta, err = p.Client_().Call(ctx, \"queryRecovery\", &_args377, &_result378)\n p.SetLastResponseMeta_(meta)\n if err != nil {\n return\n }\n return _result378.GetSuccess(), nil\n}", "func (orc *Oracle) RunWithRecovery() {\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\torc.logger.Error(\"recovered from panic in Gateway Oracle\", \"r\", r)\n\t\t\t// Unless it's a runtime error restart the goroutine\n\t\t\tif _, ok := r.(runtime.Error); !ok {\n\t\t\t\ttime.Sleep(30 * time.Second)\n\t\t\t\torc.logger.Info(\"Restarting Gateway Oracle...\")\n\t\t\t\tgo orc.RunWithRecovery()\n\t\t\t}\n\t\t}\n\t}()\n\n\t// When running in-process give the node a bit of time to spin up.\n\tif orc.startupDelay > 0 {\n\t\ttime.Sleep(orc.startupDelay)\n\t}\n\n\torc.Run()\n}", "func (agent *TestAgent) ResolveTaskDockerID(task *TestTask, containerName string) (string, error) {\n\tvar err error\n\tvar dockerId string\n\tfor i := 0; i < 5; i++ {\n\t\tdockerId, err = agent.resolveTaskDockerID(task, containerName)\n\t\tif err == nil {\n\t\t\tbreak\n\t\t}\n\t\ttime.Sleep(100 * time.Millisecond)\n\t}\n\treturn dockerId, err\n}", "func (k *KubernetesExecutor) KillTask(driver bindings.ExecutorDriver, taskId *mesos.TaskID) {\n\tif k.isDone() {\n\t\treturn\n\t}\n\tlog.Infof(\"Kill task %v\\n\", taskId)\n\n\tif !k.isConnected() {\n\t\t//TODO(jdefelice) sent TASK_LOST here?\n\t\tlog.Warningf(\"Ignore kill task because the executor is disconnected\\n\")\n\t\treturn\n\t}\n\n\tk.lock.Lock()\n\tdefer k.lock.Unlock()\n\tk.removePodTask(driver, taskId.GetValue(), messages.TaskKilled, mesos.TaskState_TASK_KILLED)\n}", "func (mgr *ClientMgr) onTaskEnd(ctx context.Context, client *Client, task *Task,\n\terr error, reply *jarviscrawlercore.ReplyCrawler, endChan chan int) {\n\n\tif err != nil {\n\t\tif task.Logger != nil {\n\t\t\ttask.Logger.Warn(\"onTaskEnd: error\",\n\t\t\t\tzap.Error(err),\n\t\t\t\tzap.String(\"servaddr\", client.servAddr),\n\t\t\t\tJSON(\"task\", task))\n\t\t}\n\n\t\t// if !(strings.Index(err.Error(), \"Error: noretry:\") == 0 ||\n\t\t// \tstrings.Index(err.Error(), \"noretry:\") == 0) {\n\t\tif !IsNoRetryError(err) {\n\n\t\t\tif task.RetryNums > 0 {\n\t\t\t\ttask.RetryNums--\n\n\t\t\t\t// time.Sleep(time.Second * time.Duration(mgr.cfg.SleepTime))\n\n\t\t\t\ttask.Running = false\n\t\t\t\tclient.Running = false\n\t\t\t\tendChan <- 0\n\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// task.Fail = true\n\t\t\t// task.running = false\n\t\t}\n\n\t\ttask.Fail = true\n\t}\n\n\tgo task.Callback(ctx, task, err, reply)\n\n\t// time.Sleep(time.Second * time.Duration(mgr.cfg.SleepTime))\n\n\ttask.Running = false\n\tclient.Running = false\n\tendChan <- task.TaskID\n}", "func Rebalance() error {\n\ttargetQuota, err := getQuotaMap(safeZkConn)\n\tif err != nil {\n\t\treturn errors.Trace(err)\n\t}\n\tlivingNodes, err := getLivingNodeInfos(safeZkConn)\n\tif err != nil {\n\t\treturn errors.Trace(err)\n\t}\n\tlog.Infof(\"start rebalance\")\n\tfor _, node := range livingNodes {\n\t\tfor len(node.CurSlots) > targetQuota[node.GroupId] {\n\t\t\tfor _, dest := range livingNodes {\n\t\t\t\tif dest.GroupId != node.GroupId && len(dest.CurSlots) < targetQuota[dest.GroupId] {\n\t\t\t\t\tslot := node.CurSlots[len(node.CurSlots)-1]\n\t\t\t\t\t// create a migration task\n\t\t\t\t\tinfo := &MigrateTaskInfo{\n\t\t\t\t\t\tDelay: 0,\n\t\t\t\t\t\tSlotId: slot,\n\t\t\t\t\t\tNewGroupId: dest.GroupId,\n\t\t\t\t\t\tStatus: MIGRATE_TASK_PENDING,\n\t\t\t\t\t\tCreateAt: strconv.FormatInt(time.Now().Unix(), 10),\n\t\t\t\t\t}\n\t\t\t\t\tglobalMigrateManager.PostTask(info)\n\n\t\t\t\t\tnode.CurSlots = node.CurSlots[0 : len(node.CurSlots)-1]\n\t\t\t\t\tdest.CurSlots = append(dest.CurSlots, slot)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tlog.Infof(\"rebalance tasks submit finish\")\n\treturn nil\n}", "func (b *configuratorBehavior) Recover(err interface{}) error {\n\treturn nil\n}", "func RestoreTaskFunc(w http.ResponseWriter, r *http.Request) {\n\tif r.Method != \"GET\" {\n\t\thttp.Redirect(w, r, \"/\", http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tid, err := strconv.Atoi(r.URL.Path[len(\"/restore/\"):])\n\tif err != nil {\n\t\tlog.Println(err)\n\t\thttp.Redirect(w, r, \"/deleted\", http.StatusBadRequest)\n\t} else {\n\t\tusername := sessions.GetCurrentUserName(r)\n\t\terr = db.RestoreTask(username, id)\n\t\tif err != nil {\n\t\t\tmessage = \"Restore failed\"\n\t\t} else {\n\t\t\tmessage = \"Task restored\"\n\t\t}\n\t\thttp.Redirect(w, r, \"/deleted/\", http.StatusFound)\n\t}\n\n}", "func Recover(yes bool) GroupOption {\n\treturn func(g *Group) {\n\t\tg.recover = yes\n\t}\n}", "func (ctl Controller) Restore(name string) *pitr.Error {\n\terr := ctl.cluster.Stop()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = ctl.cluster.Clear()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tstdout, stderr, runErr := ctl.runner.Run(\"sudo --login --user postgres wal-g backup-fetch %s %s\", ctl.cluster.DataDirectory(), name)\n\n\tif runErr != nil {\n\t\treturn &pitr.Error{\n\t\t\tMessage: runErr.Error(),\n\t\t\tStdout: stdout,\n\t\t\tStderr: stderr,\n\t\t}\n\t}\n\n\tctl.createRecoveryConf(`restore_command = 'bash --login -c \\\"wal-g wal-fetch %f %p\\\"'`)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = ctl.cluster.Start()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *Consumer) RebalanceTaskProgress() *cm.RebalanceProgress {\n\tlogPrefix := \"Consumer::RebalanceTaskProgress\"\n\n\tprogress := &cm.RebalanceProgress{}\n\n\tvbsRemainingToCloseStream := c.getVbRemainingToCloseStream()\n\tvbsRemainingToStreamReq := c.getVbRemainingToStreamReq()\n\n\tlogging.Infof(\"%s [%s:%s:%d] vbsRemainingToCloseStream len: %d dump: %v vbsRemainingToStreamReq len: %d dump: %v\",\n\t\tlogPrefix, c.workerName, c.tcpPort, c.Pid(), len(vbsRemainingToCloseStream),\n\t\tutil.Condense(vbsRemainingToCloseStream), len(vbsRemainingToStreamReq),\n\t\tutil.Condense(vbsRemainingToStreamReq))\n\n\tif len(vbsRemainingToCloseStream) > 0 || len(vbsRemainingToStreamReq) > 0 {\n\t\tvbsOwnedPerPlan := c.getVbsOwned()\n\n\t\tprogress.CloseStreamVbsLen = len(vbsRemainingToCloseStream)\n\t\tprogress.StreamReqVbsLen = len(vbsRemainingToStreamReq)\n\n\t\tprogress.VbsOwnedPerPlan = len(vbsOwnedPerPlan)\n\t\tprogress.VbsRemainingToShuffle = len(vbsRemainingToCloseStream) + len(vbsRemainingToStreamReq)\n\t}\n\n\tif len(vbsRemainingToCloseStream) == 0 && len(vbsRemainingToStreamReq) == 0 && c.vbsStateUpdateRunning {\n\t\tc.isRebalanceOngoing = false\n\t\tlogging.Infof(\"%s [%s:%s:%d] Updated isRebalanceOngoing to %t\",\n\t\t\tlogPrefix, c.workerName, c.tcpPort, c.Pid(), c.isRebalanceOngoing)\n\t}\n\n\treturn progress\n}", "func (suite *TaskFailRetryTestSuite) TestTaskFailRetry() {\n\ttaskConfig := pbtask.TaskConfig{\n\t\tRestartPolicy: &pbtask.RestartPolicy{\n\t\t\tMaxFailures: 3,\n\t\t},\n\t}\n\n\tsuite.cachedTask.EXPECT().\n\t\tID().\n\t\tReturn(uint32(0)).\n\t\tAnyTimes()\n\n\tsuite.jobFactory.EXPECT().\n\t\tGetJob(suite.jobID).Return(suite.cachedJob)\n\n\tsuite.cachedJob.EXPECT().\n\t\tGetTask(suite.instanceID).Return(suite.cachedTask)\n\n\tsuite.cachedJob.EXPECT().\n\t\tID().Return(suite.jobID)\n\n\tsuite.cachedTask.EXPECT().\n\t\tGetRuntime(gomock.Any()).Return(suite.taskRuntime, nil)\n\n\tsuite.taskConfigV2Ops.EXPECT().\n\t\tGetTaskConfig(gomock.Any(), suite.jobID, suite.instanceID, gomock.Any()).\n\t\tReturn(&taskConfig, &models.ConfigAddOn{}, nil)\n\n\tsuite.cachedJob.EXPECT().\n\t\tPatchTasks(gomock.Any(), gomock.Any(), false).\n\t\tDo(func(ctx context.Context,\n\t\t\truntimeDiffs map[uint32]jobmgrcommon.RuntimeDiff,\n\t\t\t_ bool) {\n\t\t\truntimeDiff := runtimeDiffs[suite.instanceID]\n\t\t\tsuite.True(\n\t\t\t\truntimeDiff[jobmgrcommon.MesosTaskIDField].(*mesosv1.TaskID).GetValue() != suite.mesosTaskID)\n\t\t\tsuite.True(\n\t\t\t\truntimeDiff[jobmgrcommon.PrevMesosTaskIDField].(*mesosv1.TaskID).GetValue() == suite.mesosTaskID)\n\t\t\tsuite.True(\n\t\t\t\truntimeDiff[jobmgrcommon.StateField].(pbtask.TaskState) == pbtask.TaskState_INITIALIZED)\n\t\t}).Return(nil, nil, nil)\n\n\tsuite.cachedJob.EXPECT().\n\t\tGetJobType().Return(pbjob.JobType_BATCH)\n\n\tsuite.taskGoalStateEngine.EXPECT().\n\t\tEnqueue(gomock.Any(), gomock.Any()).\n\t\tReturn()\n\n\tsuite.jobGoalStateEngine.EXPECT().\n\t\tEnqueue(gomock.Any(), gomock.Any()).\n\t\tReturn()\n\n\terr := TaskFailRetry(context.Background(), suite.taskEnt)\n\tsuite.NoError(err)\n}", "func (s *OnDiskStateMachine) RecoverFromSnapshot(r io.Reader,\n\tfs []sm.SnapshotFile, stopc <-chan struct{}) error {\n\tif !s.opened {\n\t\tpanic(\"RecoverFromSnapshot called when not opened\")\n\t}\n\t/*\n\t\trollback := ss.Imported && init\n\t\tif !rollback {\n\t\t\tif ss.StateMachineIndex <= s.index {\n\t\t\t\tplog.Panicf(\"recover snapshot moving applied index backwards, %d, %d\",\n\t\t\t\t\tss.StateMachineIndex, s.index)\n\t\t\t}\n\t\t} else {\n\t\t\ts.initialIndex = ss.StateMachineIndex\n\t\t}\n\t\ts.index = ss.StateMachineIndex\n\t*/\n\treturn s.sm.RecoverFromSnapshot(r, stopc)\n}", "func ReturnIDCNodeToResourcePoolTask(taskID, stepName string) error {\n\tstart := time.Now()\n\n\t// get task and task current step\n\tstate, step, err := cloudprovider.GetTaskStateAndCurrentStep(taskID, stepName)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// previous step successful when retry task\n\tif step == nil {\n\t\tblog.Infof(\"ReturnIDCNodeToResourcePoolTask[%s]: current step[%s] successful and skip\", taskID, stepName)\n\t\treturn nil\n\t}\n\tblog.Infof(\"ReturnIDCNodeToResourcePoolTask[%s] task %s run step %s, system: %s, old state: %s, params %v\",\n\t\ttaskID, taskID, stepName, step.System, step.Status, step.Params)\n\n\t// extract valid parameter\n\tclusterID := step.Params[cloudprovider.ClusterIDKey.String()]\n\tnodeGroupID := step.Params[cloudprovider.NodeGroupIDKey.String()]\n\tcloudID := step.Params[cloudprovider.CloudIDKey.String()]\n\n\toperator := step.Params[cloudprovider.OperatorKey.String()]\n\tnodeIPList := cloudprovider.ParseNodeIpOrIdFromCommonMap(step.Params, cloudprovider.NodeIPsKey.String(), \",\")\n\tdeviceList := cloudprovider.ParseNodeIpOrIdFromCommonMap(step.Params, cloudprovider.DeviceIDsKey.String(), \",\")\n\n\tdependInfo, err := cloudprovider.GetClusterDependBasicInfo(cloudprovider.GetBasicInfoReq{\n\t\tClusterID: clusterID,\n\t\tCloudID: cloudID,\n\t\tNodeGroupID: nodeGroupID,\n\t})\n\tif err != nil {\n\t\tblog.Errorf(\"ReturnIDCNodeToResourcePoolTask[%s] GetClusterDependBasicInfo for NodeGroup %s to clean Node in task %s \"+\n\t\t\t\"step %s failed, %s\", taskID, nodeGroupID, taskID, stepName, err.Error())\n\t\tretErr := fmt.Errorf(\"get cloud/project information failed, %s\", err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\treturn retErr\n\t}\n\n\t// inject taskID\n\tctx := cloudprovider.WithTaskIDForContext(context.Background(), taskID)\n\n\t// return IDC device to resource-manager\n\torderID, err := destroyIDCDeviceList(ctx, dependInfo, deviceList, operator)\n\tif err != nil {\n\t\tblog.Errorf(\"ReturnIDCNodeToResourcePoolTask[%s] destroyIDCDeviceList[%v] from NodeGroup %s failed: %v\",\n\t\t\ttaskID, nodeIPList, nodeGroupID, err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, err)\n\t\treturn fmt.Errorf(\"ReturnIDCNodeToResourcePoolTask destroyIDCDeviceList failed %s\", err.Error())\n\t}\n\n\t// update response information to task common params\n\tif state.Task.CommonParams == nil {\n\t\tstate.Task.CommonParams = make(map[string]string)\n\t}\n\tstate.Task.CommonParams[cloudprovider.OrderIDKey.String()] = orderID\n\n\t// update step\n\tif err := state.UpdateStepSucc(start, stepName); err != nil {\n\t\tblog.Errorf(\"task %s %s update to storage fatal\", taskID, stepName)\n\t\treturn err\n\t}\n\treturn nil\n}", "func panicRecover(input *models.RunningInput) {\n\tif err := recover(); err != nil {\n\t\ttrace := make([]byte, 2048)\n\t\truntime.Stack(trace, true)\n\t\tlog.Printf(\"E! FATAL: [%s] panicked: %s, Stack:\\n%s\",\n\t\t\tinput.LogName(), err, trace)\n\t\tlog.Println(\"E! PLEASE REPORT THIS PANIC ON GITHUB with \" +\n\t\t\t\"stack trace, configuration, and OS information: \" +\n\t\t\t\"https://github.com/influxdata/telegraf/issues/new/choose\")\n\t}\n}", "func Recoverer() {\n\tdebug := DebugMode()\n\tif r := recover(); r != nil {\n\t\tlog.Criticalf(\"PANIC: %s\", r)\n\t\tif debug {\n\t\t\tpanic(r)\n\t\t} else {\n\t\t\tlog.Info(\"Hint: For debug info, please set \\\"NMZ_DEBUG\\\" to 1.\")\n\t\t\tos.Exit(1)\n\t\t}\n\t}\n}", "func CheckRecurredTaskBegin(name string) error {\n\tlastRc := NewResultContainer()\n\terr := lastRc.LoadLatest()\n\tif err != nil {\n\t\tif lastRc == nil {\n\t\t\tnegativeCheck(\"internal error\")\n\t\t\treturn err\n\t\t}\n\t}\n\n\tlastBegin, err := persistence.GetLastBegin(name)\n\tif err != nil {\n\t\tnegativeCheck(\"internal error\")\n\t\treturn err\n\t}\n\n\tres := lastRc.Get(name)\n\tif res.IntervalAvg < 1 {\n\t\treturn nil\n\t\t//no interval data\n\t}\n\n\tif time.Now().Sub(lastBegin.Time) > (res.IntervalAvg+res.IntervalStdDev) && alertBlocker.IsUnknown(name, \"not recurred\") {\n\t\tmsg := `Task \"` + name + `\" has not recurred in time`\n\t\tidl.Err(msg, lastBegin)\n\t\tnegativeCheck(msg)\n\t\talertBlocker.Occurred(name, \"not recurred\")\n\t}\n\n\treturn nil\n}", "func (d *dispatcher) monitorTask(taskID int64) (finished bool, subTaskErrs []error) {\n\t// TODO: Consider putting the following operations into a transaction.\n\tvar err error\n\td.task, err = d.taskMgr.GetGlobalTaskByID(taskID)\n\tif err != nil {\n\t\tlogutil.BgLogger().Error(\"check task failed\", zap.Int64(\"task ID\", d.task.ID), zap.Error(err))\n\t\treturn false, nil\n\t}\n\tswitch d.task.State {\n\tcase proto.TaskStateCancelling:\n\t\treturn false, []error{errors.New(\"cancel\")}\n\tcase proto.TaskStateReverting:\n\t\tcnt, err := d.taskMgr.GetSubtaskInStatesCnt(d.task.ID, proto.TaskStateRevertPending, proto.TaskStateReverting)\n\t\tif err != nil {\n\t\t\tlogutil.BgLogger().Warn(\"check task failed\", zap.Int64(\"task ID\", d.task.ID), zap.Error(err))\n\t\t\treturn false, nil\n\t\t}\n\t\treturn cnt == 0, nil\n\tdefault:\n\t\tsubTaskErrs, err = d.taskMgr.CollectSubTaskError(d.task.ID)\n\t\tif err != nil {\n\t\t\tlogutil.BgLogger().Warn(\"collect subtask error failed\", zap.Int64(\"task ID\", d.task.ID), zap.Error(err))\n\t\t\treturn false, nil\n\t\t}\n\t\tif len(subTaskErrs) > 0 {\n\t\t\treturn false, subTaskErrs\n\t\t}\n\t\t// check subtasks pending or running.\n\t\tcnt, err := d.taskMgr.GetSubtaskInStatesCnt(d.task.ID, proto.TaskStatePending, proto.TaskStateRunning)\n\t\tif err != nil {\n\t\t\tlogutil.BgLogger().Warn(\"check task failed\", zap.Int64(\"task ID\", d.task.ID), zap.Error(err))\n\t\t\treturn false, nil\n\t\t}\n\t\treturn cnt == 0, nil\n\t}\n}", "func Recover(files []string) {\n\t//gurantee the time sequence is in right ordered\n\tsort.Slice(files, func(i, j int) bool {\n\t\treturn files[i] < files[j]\n\t})\n\n\t//read log items and recover them into DataItem\n\tdataJson := make([]logger.DataItem, 0)\n\tfor _, filename := range files {\n\t\tfile, err := os.Open(filename)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"file can't open\")\n\t\t}\n\n\t\tcontent, err := ioutil.ReadAll(file)\n\t\tif err != nil {\n\t\t\tlog.Println(\"file can't read\")\n\t\t}\n\n\t\titems := strings.Split(string(content), \"\\r\\n\")\n\t\tfor _, jsonitem := range items {\n\t\t\tdata := logger.DataItem{}\n\n\t\t\terr = json.Unmarshal([]byte(jsonitem), &data)\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(\"unmarshal some data fail\")\n\t\t\t}\n\n\t\t\tdataJson = append(dataJson, data)\n\t\t}\n\t}\n\n\t//recover the dataset status\n\tfor _, data := range dataJson {\n\t\tif data.Commandtype == logger.SET {\n\t\t\tnow := time.Now().Unix()\n\t\t\tif now < data.Expire {\n\t\t\t\tstore.SetValue(data.Key, data.Value, time.Duration(data.Expire-now))\n\t\t\t}\n\t\t}\n\t}\n}", "func asyncRestart(ctx context.Context, logger hclog.Logger, task WorkloadRestarter, event *structs.TaskEvent) {\n\t// Check watcher restarts are always failures\n\tconst failure = true\n\n\t// Restarting is asynchronous so there's no reason to allow this\n\t// goroutine to block indefinitely.\n\tctx, cancel := context.WithTimeout(ctx, 10*time.Second)\n\tdefer cancel()\n\n\tif err := task.Restart(ctx, event, failure); err != nil {\n\t\t// Restart errors are not actionable and only relevant when\n\t\t// debugging allocation lifecycle management.\n\t\tlogger.Debug(\"failed to restart task\", \"error\", err, \"event_time\", event.Time, \"event_type\", event.Type)\n\t}\n}", "func (r *ReconcileAerospikeCluster) recoverFailedCreate(aeroCluster *aerospikev1alpha1.AerospikeCluster) error {\n\tlogger := pkglog.New(log.Ctx{\"AerospikeCluster\": utils.ClusterNamespacedName(aeroCluster)})\n\tlogger.Info(\"Forcing a cluster recreate as status is nil. The cluster could be unreachable due to bad configuration.\")\n\n\t// Delete all statefulsets and everything related so that it can be properly created and updated in next run.\n\tstatefulSetList, err := r.getClusterStatefulSets(aeroCluster)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error getting statefulsets while forcing recreate of the cluster as status is nil: %v\", err)\n\t}\n\n\tlogger.Debug(\"Found statefulset for cluster. Need to delete them\", log.Ctx{\"nSTS\": len(statefulSetList.Items)})\n\tfor _, statefulset := range statefulSetList.Items {\n\t\tif err := r.deleteStatefulSet(aeroCluster, &statefulset); err != nil {\n\t\t\treturn fmt.Errorf(\"Error deleting statefulset while forcing recreate of the cluster as status is nil: %v\", err)\n\t\t}\n\t}\n\n\t// Clear pod status as well in status since we want to be re-initializing or cascade deleting devices if any.\n\t// This is not necessary since scale-up would cleanup danglin pod status. However done here for general\n\t// cleanliness.\n\trackStateList := getNewRackStateList(aeroCluster)\n\tfor _, state := range rackStateList {\n\t\tpods, err := r.getRackPodList(aeroCluster, state.Rack.ID)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Failed recover failed cluster: %v\", err)\n\t\t}\n\n\t\tnewPodNames := []string{}\n\t\tfor i := 0; i < len(pods.Items); i++ {\n\t\t\tnewPodNames = append(newPodNames, pods.Items[i].Name)\n\t\t}\n\n\t\terr = r.cleanupPods(aeroCluster, newPodNames, state)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Failed recover failed cluster: %v\", err)\n\t\t}\n\t}\n\n\treturn fmt.Errorf(\"Forcing recreate of the cluster as status is nil\")\n}", "func (_DelegatableDai *DelegatableDaiCallerSession) Recover(hash [32]byte, sig []byte) (common.Address, error) {\n\treturn _DelegatableDai.Contract.Recover(&_DelegatableDai.CallOpts, hash, sig)\n}", "func Recover() *PanicError {\n\terr := recover()\n\tif err != nil {\n\t\treturn &PanicError{err}\n\t}\n\treturn nil\n}", "func (*FakeReconcilerClient) GetTask(string) (swarm.Task, error) {\n\treturn swarm.Task{}, FakeUnimplemented\n}", "func (dyn *Dynamo) checkAndExitRecoveryMode() {\n\tdyn.logger.Info(\"Starting background task for exiting recovery mode\")\n\n\t// block indefinitely while the node is initializing\n\tfor dyn.isInitializing {\n\t\tdyn.logger.Debug(\"Delaying exiting recovery mode while initializing\")\n\t\ttime.Sleep(time.Second * 3)\n\t}\n\n\tfor {\n\t\tdyn.recoveryLock.RLock()\n\t\trecovering := dyn.inRecovery\n\t\tlastWrite := dyn.lastRecoveryTimestamp\n\t\tdyn.recoveryLock.RUnlock()\n\n\t\tif recovering {\n\t\t\tif time.Since(lastWrite) >= (time.Second * time.Duration(dyn.cfg.RecoveryGracePeriod)) {\n\t\t\t\tdyn.logger.Debug(\"Exiting recovery mode\")\n\t\t\t\tdyn.recoveryLock.Lock()\n\t\t\t\tdyn.inRecovery = false\n\t\t\t\tdyn.recoveryLock.Unlock()\n\t\t\t}\n\t\t}\n\n\t\tdyn.logger.Debug(\"Sleeping for checking recovery mode\", zap.Int(\"time\", dyn.cfg.RecoveryGracePeriod))\n\t\ttime.Sleep(time.Second * time.Duration(dyn.cfg.RecoveryGracePeriod))\n\t}\n}", "func (suite *TaskFailRetryTestSuite) TestTaskFailNoRetry() {\n\n\ttaskConfig := pbtask.TaskConfig{\n\t\tRestartPolicy: &pbtask.RestartPolicy{\n\t\t\tMaxFailures: 0,\n\t\t},\n\t}\n\n\tsuite.jobFactory.EXPECT().\n\t\tGetJob(suite.jobID).Return(suite.cachedJob)\n\n\tsuite.cachedJob.EXPECT().\n\t\tGetTask(suite.instanceID).Return(suite.cachedTask)\n\n\tsuite.cachedTask.EXPECT().\n\t\tGetRuntime(gomock.Any()).Return(suite.taskRuntime, nil)\n\n\tsuite.taskConfigV2Ops.EXPECT().\n\t\tGetTaskConfig(gomock.Any(), suite.jobID, suite.instanceID, gomock.Any()).\n\t\tReturn(&taskConfig, &models.ConfigAddOn{}, nil)\n\n\terr := TaskFailRetry(context.Background(), suite.taskEnt)\n\tsuite.NoError(err)\n}", "func deleteTasks() {\n\tcounter := 0\n\n\t// Create AWS session\n\ts, err := session.NewSession()\n\tif err != nil {\n\t\tlog.Fatal(\"Couldn't create AWS Session.\")\n\t}\n\n\t// Create the AWS Service\n\tsvc := databasemigrationservice.New(s, &aws.Config{Region: &region})\n\n\t// Read the defaults file\n\treadTasks, err := ioutil.ReadFile(tasksFile)\n\tif err != nil {\n\t\tlog.Fatal(\"Couldn't read file \"+tasksFile, err)\n\t}\n\n\t// Create tasks and unmarshal the JSON\n\ttasks := new([]ReplicationTask)\n\tremainingTasks := new([]ReplicationTask) // Tasks that will be saved (if they couldn't be removed for example)\n\terr = json.Unmarshal(readTasks, tasks)\n\tif err != nil {\n\t\tlog.Fatal(\"Couldn't JSON unmarshal file \"+tasksFile, err)\n\t}\n\n\t// Start all the tasks stored in tasks\n\tfor _, task := range *tasks {\n\t\tparams := &databasemigrationservice.DeleteReplicationTaskInput{\n\t\t\tReplicationTaskArn: aws.String(task.ReplicationTaskArn),\n\t\t}\n\n\t\t_, err := svc.DeleteReplicationTask(params)\n\t\tif err != nil {\n\t\t\t// If the task doesn't exists we shouldn't keep it in the tasks.json file - just continue\n\t\t\tif strings.Contains(err.Error(), notFound) {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// If the task errored and not because it doesn't exists, keep it in the tasks.json file\n\t\t\t*remainingTasks = append(*remainingTasks, task)\n\n\t\t\t// Go through the different statuses that might have made the action failed\n\t\t\tswitch {\n\t\t\tcase strings.Contains(err.Error(), isRunning):\n\t\t\t\tfmt.Println(\"Please stop task\", task.ReplicationTaskIdentifier, stopBeforeDeleting)\n\t\t\t\tcontinue\n\t\t\tcase strings.Contains(err.Error(), \"is currently being stopped\"):\n\t\t\t\tfmt.Println(\"Please wait until task\", task.ReplicationTaskIdentifier, waitForStop)\n\t\t\t\tcontinue\n\t\t\tcase strings.Contains(err.Error(), \"is already being deleted\"):\n\t\t\t\tfmt.Println(\"Task\", task.ReplicationTaskIdentifier, beingDeleted)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tfmt.Println(\"Couldn't delete Replication Task\", err)\n\t\t\tcontinue\n\t\t}\n\n\t\tcounter++\n\t\tfmt.Println(\"Task deleted: \" + task.ReplicationTaskIdentifier)\n\t}\n\n\t// If we have no tasks left, delete the whole file\n\tswitch {\n\tcase len(*remainingTasks) == 0:\n\t\terr := os.Remove(tasksFile)\n\t\tif err != nil {\n\t\t\tfmt.Println(\"Couldn't remove tasks files\", err)\n\t\t}\n\n\tdefault:\n\t\t// Write remaining tasks to tasks-file\n\t\twriteTaskFile(remainingTasks)\n\t}\n\n\tfmt.Println(\"\\nDONE! Deleted\", counter, \"tasks.\")\n}", "func StartDMTask(fw portforward.PortForward, ns, masterSvcName, taskConf, errSubStr string) error {\n\tapiPath := \"/apis/v1alpha1/tasks\"\n\n\ttype Req struct {\n\t\tTask string `json:\"task\"`\n\t}\n\ttype Resp struct {\n\t\tResult bool `json:\"result\"`\n\t\tMsg string `json:\"msg\"`\n\t\tCheckResult string `json:\"checkResult\"`\n\t}\n\n\tvar req = Req{\n\t\tTask: fmt.Sprintf(taskConf, DMTaskName(ns), v1alpha1.DefaultTiDBServerPort, DMTaskName(ns)),\n\t}\n\tdata, err := json.Marshal(req)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to marshal task start request, %v, %v\", req, err)\n\t}\n\n\treturn wait.Poll(5*time.Second, time.Minute, func() (bool, error) {\n\t\tlocalHost, localPort, cancel, err := portforward.ForwardOnePort(\n\t\t\tfw, ns, fmt.Sprintf(\"svc/%s\", masterSvcName), dmMasterSvcPort)\n\t\tif err != nil {\n\t\t\tlog.Logf(\"failed to forward dm-master svc: %v\", err)\n\t\t\treturn false, nil\n\t\t}\n\t\tdefer cancel()\n\n\t\tbody, err := httputil.DoBodyOK(\n\t\t\t&http.Client{Transport: &http.Transport{}},\n\t\t\tfmt.Sprintf(\"http://%s:%d%s\", localHost, localPort, apiPath),\n\t\t\t\"POST\",\n\t\t\tbytes.NewReader(data))\n\t\tif err != nil {\n\t\t\tlog.Logf(\"failed to start DM task: %v\", err)\n\t\t\treturn false, nil\n\t\t}\n\t\tvar resp Resp\n\t\tif err = json.Unmarshal(body, &resp); err != nil {\n\t\t\tlog.Logf(\"failed to unmarshal DM task start response, %s: %v\", string(body), err)\n\t\t\treturn false, nil\n\t\t} else if !resp.Result && !strings.Contains(resp.Msg, \"already exists\") {\n\t\t\tif errSubStr != \"\" && strings.Contains(resp.Msg, errSubStr) {\n\t\t\t\tlog.Logf(\"start DM task match the error sub string %q: %s\", errSubStr, resp.Msg)\n\t\t\t\treturn true, nil\n\t\t\t}\n\t\t\tlog.Logf(\"failed to start DM task, msg: %s, err: %v, checkResult: %s\", resp.Msg, err, resp.CheckResult)\n\t\t\treturn false, nil\n\t\t}\n\t\treturn true, nil\n\t})\n}" ]
[ "0.58838654", "0.5342246", "0.4967397", "0.49165246", "0.48566574", "0.48527366", "0.48488343", "0.4843941", "0.48418212", "0.4827211", "0.481903", "0.47926974", "0.47663003", "0.47648853", "0.47385824", "0.47373864", "0.47090328", "0.47073582", "0.4702934", "0.46741527", "0.46718264", "0.46462384", "0.46432397", "0.4636801", "0.4632619", "0.4632619", "0.46323818", "0.46107763", "0.46076995", "0.46048504", "0.45881578", "0.455844", "0.45480078", "0.45334917", "0.45299217", "0.45278898", "0.4523431", "0.44760266", "0.44696766", "0.44650427", "0.44616526", "0.44594097", "0.4457134", "0.44409773", "0.44336438", "0.44282997", "0.44252503", "0.4416343", "0.44075814", "0.44045067", "0.43794793", "0.43758187", "0.4372764", "0.4367623", "0.43644828", "0.43626356", "0.43547988", "0.43495053", "0.43439603", "0.4343559", "0.4343055", "0.4335428", "0.43221122", "0.43211946", "0.43204084", "0.4314417", "0.4309435", "0.43087482", "0.43062946", "0.429755", "0.42945042", "0.42938656", "0.42904478", "0.42862546", "0.4281773", "0.42770818", "0.42734993", "0.42720345", "0.42702156", "0.4270006", "0.4262931", "0.42459667", "0.42420658", "0.4241359", "0.42351967", "0.42289576", "0.4224809", "0.42239574", "0.4219015", "0.42189053", "0.42152694", "0.4213842", "0.42100412", "0.42050502", "0.41974622", "0.41872245", "0.41734666", "0.41713476", "0.4167909", "0.41650474" ]
0.7950789
0
BuildContainerName returns the podman container name for a given TaskConfig
func BuildContainerName(cfg *drivers.TaskConfig) string { return fmt.Sprintf("%s-%s", cfg.Name, cfg.AllocID) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (e *dockerExec) containerName() string {\n\tpathHex := reflow.Digester.FromString(e.path()).Short()\n\treturn fmt.Sprintf(\"reflow-%s-%s-%s\", e.Executor.ID, e.id.Hex(), pathHex)\n}", "func (o ClusterBuildStrategySpecBuildStepsEnvValueFromResourceFieldRefOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ClusterBuildStrategySpecBuildStepsEnvValueFromResourceFieldRef) *string { return v.ContainerName }).(pulumi.StringPtrOutput)\n}", "func (o ClusterBuildStrategySpecBuildStepsEnvValueFromResourceFieldRefPtrOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ClusterBuildStrategySpecBuildStepsEnvValueFromResourceFieldRef) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ContainerName\n\t}).(pulumi.StringPtrOutput)\n}", "func makeContainerName(c *runtime.ContainerMetadata, s *runtime.PodSandboxMetadata) string {\n\treturn strings.Join([]string{\n\t\tc.Name, // 0\n\t\ts.Name, // 1: sandbox name\n\t\ts.Namespace, // 2: sandbox namespace\n\t\ts.Uid, // 3: sandbox uid\n\t\tfmt.Sprintf(\"%d\", c.Attempt), // 4\n\t}, nameDelimiter)\n}", "func (o BuildStrategySpecBuildStepsEnvValueFromResourceFieldRefOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BuildStrategySpecBuildStepsEnvValueFromResourceFieldRef) *string { return v.ContainerName }).(pulumi.StringPtrOutput)\n}", "func (s Service) ContainerName() string {\n\treturn fmt.Sprintf(\"coz_neo-local_%s\", s.Image)\n}", "func (n *natsDependency) GetContainerName() string {\n\treturn n.config.Name + n.config.Version\n}", "func (o BuildStrategySpecBuildStepsEnvValueFromResourceFieldRefPtrOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BuildStrategySpecBuildStepsEnvValueFromResourceFieldRef) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ContainerName\n\t}).(pulumi.StringPtrOutput)\n}", "func (o VirtualDatabaseSpecBuildEnvValueFromResourceFieldRefOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v VirtualDatabaseSpecBuildEnvValueFromResourceFieldRef) *string { return v.ContainerName }).(pulumi.StringPtrOutput)\n}", "func (o VirtualDatabaseSpecBuildEnvValueFromResourceFieldRefPtrOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *VirtualDatabaseSpecBuildEnvValueFromResourceFieldRef) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ContainerName\n\t}).(pulumi.StringPtrOutput)\n}", "func (e *localExocomDependency) GetContainerName() string {\n\treturn e.config.Name + e.config.Version\n}", "func (cni *ContrailCni) buildContainerIntfName(\n index int, isMetaPlugin bool) string {\n var intfName string\n if isMetaPlugin == true {\n intfName = cni.cniArgs.IfName\n } else {\n intfName = VIF_TYPE_ETH + strconv.Itoa(index)\n }\n log.Infof(\"Built container interface name - %s\", intfName)\n return intfName\n}", "func (o VirtualDatabaseSpecEnvValueFromResourceFieldRefOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v VirtualDatabaseSpecEnvValueFromResourceFieldRef) *string { return v.ContainerName }).(pulumi.StringPtrOutput)\n}", "func MakePodName(jobName string, taskName string, index int) string {\n\treturn fmt.Sprintf(jobhelpers.PodNameFmt, jobName, taskName, index)\n}", "func containerNameFromSpec(debugger string) string {\n\tcontainerVariant := \"gdb\"\n\tif debugger == \"dlv\" {\n\t\tcontainerVariant = \"dlv\"\n\t}\n\treturn fmt.Sprintf(\"%v-%v\", sqOpts.ParticularContainerRootName, containerVariant)\n}", "func (o IopingSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsResourceFieldRefOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsResourceFieldRef) *string {\n\t\treturn v.ContainerName\n\t}).(pulumi.StringPtrOutput)\n}", "func (c *ContainerConfigMapSpec) GetName() string {\n\tsuffix := fmt.Sprintf(\"%s-%s\", c.ContainerName, c.Subdir)\n\treturn util.PrefixConfigmap(c.qserv, suffix)\n}", "func containerIfName(container garden.Container) string {\n\treturn ifNamePrefix(container) + \"-1\"\n}", "func (o IopingSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsResourceFieldRefPtrOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *IopingSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsResourceFieldRef) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ContainerName\n\t}).(pulumi.StringPtrOutput)\n}", "func (o FioSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsResourceFieldRefOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsResourceFieldRef) *string {\n\t\treturn v.ContainerName\n\t}).(pulumi.StringPtrOutput)\n}", "func (m *TestcontainersModule) ContainerName() string {\n\tname := m.Lower()\n\n\tif m.IsModule {\n\t\tname = m.Title()\n\t} else {\n\t\tif m.TitleName != \"\" {\n\t\t\tr, n := utf8.DecodeRuneInString(m.TitleName)\n\t\t\tname = string(unicode.ToLower(r)) + m.TitleName[n:]\n\t\t}\n\t}\n\n\treturn name + \"Container\"\n}", "func machineContainerName(cluster, machine string) string {\n\tif strings.HasPrefix(machine, cluster) {\n\t\treturn machine\n\t}\n\treturn fmt.Sprintf(\"%s-%s\", cluster, machine)\n}", "func (o IopingSpecVolumeVolumeSourceDownwardAPIItemsResourceFieldRefOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceDownwardAPIItemsResourceFieldRef) *string { return v.ContainerName }).(pulumi.StringPtrOutput)\n}", "func (o GetAppTemplateContainerOutput) Name() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetAppTemplateContainer) string { return v.Name }).(pulumi.StringOutput)\n}", "func (j SubmittedJob) ContainerName() string {\n\tvar nameFragment string\n\tif j.Name != nil {\n\t\tnameFragment = *j.Name\n\t} else {\n\t\tnameFragment = \"unnamed\"\n\t}\n\n\treturn fmt.Sprintf(\"job_%d_%s\", j.JID, nameFragment)\n}", "func createVolumeUnitContainerName(t *jobs.Task, volIndex int, scalingGroup uint) string {\n\treturn fmt.Sprintf(\"%s-vl%d\", t.ContainerName(scalingGroup), volIndex)\n}", "func (o VirtualDatabaseSpecEnvValueFromResourceFieldRefPtrOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *VirtualDatabaseSpecEnvValueFromResourceFieldRef) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ContainerName\n\t}).(pulumi.StringPtrOutput)\n}", "func (o VirtualDatabaseSpecDatasourcesPropertiesValueFromResourceFieldRefOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v VirtualDatabaseSpecDatasourcesPropertiesValueFromResourceFieldRef) *string {\n\t\treturn v.ContainerName\n\t}).(pulumi.StringPtrOutput)\n}", "func (o IopingSpecVolumeVolumeSourceDownwardAPIItemsResourceFieldRefPtrOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *IopingSpecVolumeVolumeSourceDownwardAPIItemsResourceFieldRef) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ContainerName\n\t}).(pulumi.StringPtrOutput)\n}", "func (p *DockerPod) GetContainerByName(name string) *container.BcsContainerInfo {\n\tif info, ok := p.conTasks[name]; ok {\n\t\treturn info.RuntimeConf\n\t}\n\treturn nil\n}", "func (o FioSpecVolumeVolumeSourceDownwardAPIItemsResourceFieldRefOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceDownwardAPIItemsResourceFieldRef) *string { return v.ContainerName }).(pulumi.StringPtrOutput)\n}", "func BuildAtlantisProjectNameFromComponentConfig(\n\tcliConfig schema.CliConfiguration,\n\tcomponentName string,\n\tcomponentSettingsSection map[any]any,\n\tcomponentVarsSection map[any]any,\n) (string, error) {\n\n\tvar atlantisProjectTemplate schema.AtlantisProjectConfig\n\tvar atlantisProjectName string\n\n\tif atlantisSettingsSection, ok := componentSettingsSection[\"atlantis\"].(map[any]any); ok {\n\t\t// 'settings.atlantis.project_template' has higher priority than 'settings.atlantis.project_template_name'\n\t\tif atlantisSettingsProjectTemplate, ok := atlantisSettingsSection[\"project_template\"].(map[any]any); ok {\n\t\t\terr := mapstructure.Decode(atlantisSettingsProjectTemplate, &atlantisProjectTemplate)\n\t\t\tif err != nil {\n\t\t\t\treturn \"\", err\n\t\t\t}\n\t\t} else if atlantisSettingsProjectTemplateName, ok := atlantisSettingsSection[\"project_template_name\"].(string); ok && atlantisSettingsProjectTemplateName != \"\" {\n\t\t\tif pt, ok := cliConfig.Integrations.Atlantis.ProjectTemplates[atlantisSettingsProjectTemplateName]; ok {\n\t\t\t\tatlantisProjectTemplate = pt\n\t\t\t}\n\t\t}\n\n\t\tcontext := cfg.GetContextFromVars(componentVarsSection)\n\t\tcontext.Component = strings.Replace(componentName, \"/\", \"-\", -1)\n\n\t\t// If Atlantis project template is defined and has a name, replace tokens in the name and add the Atlantis project to the output\n\t\tif !reflect.ValueOf(atlantisProjectTemplate).IsZero() && atlantisProjectTemplate.Name != \"\" {\n\t\t\tatlantisProjectName = BuildAtlantisProjectName(context, atlantisProjectTemplate.Name)\n\t\t}\n\t}\n\n\treturn atlantisProjectName, nil\n}", "func (c *Container) Name() string {\n\treturn c.config.Name\n}", "func (e *dockerExec) Name() string {\n\treturn e.containerName()\n}", "func (cfg *Config) PodName(nodeName string, id string) string {\n\tnameParts := strings.Split(nodeName, \".\")\n\tpodName := fmt.Sprintf(\"%s-job-%s-%s\", cfg.Name, nameParts[0], id)\n\treturn podName\n}", "func (o FioSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsResourceFieldRefPtrOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FioSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsResourceFieldRef) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ContainerName\n\t}).(pulumi.StringPtrOutput)\n}", "func (o GetAppTemplateContainerEnvOutput) Name() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetAppTemplateContainerEnv) string { return v.Name }).(pulumi.StringOutput)\n}", "func (o VirtualDatabaseSpecDatasourcesPropertiesValueFromResourceFieldRefPtrOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *VirtualDatabaseSpecDatasourcesPropertiesValueFromResourceFieldRef) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ContainerName\n\t}).(pulumi.StringPtrOutput)\n}", "func (o AppTemplateContainerOutput) Name() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AppTemplateContainer) string { return v.Name }).(pulumi.StringOutput)\n}", "func (c *CvpClient) GetContainerNameById(query string) (string, error) {\n\turl := \"/provisioning/getContainerInfoById.do?containerId=\" + query\n\trespbody, err := c.Get(url)\n\trespContainer := struct {\n\t\tName string `json:\"name\"`\n\t}{}\n\terr = json.Unmarshal(respbody, &respContainer)\n\tif err != nil {\n\t\tlog.Printf(\"Error decoding getcontainerbyid :%s\\n\", err)\n\t\treturn \"\", err\n\t}\n\treturn respContainer.Name, nil\n}", "func (p *DockerPod) GetContainerID() string {\n\tif p.netTask == nil {\n\t\treturn \"\"\n\t}\n\treturn p.netTask.RuntimeConf.ID\n}", "func ConfigNameForBuild(build *buildapi.Build) string {\n\tif build == nil {\n\t\treturn \"\"\n\t}\n\tif build.Annotations != nil {\n\t\tif _, exists := build.Annotations[buildapi.BuildConfigAnnotation]; exists {\n\t\t\treturn build.Annotations[buildapi.BuildConfigAnnotation]\n\t\t}\n\t}\n\tif _, exists := build.Labels[buildapi.BuildConfigLabel]; exists {\n\t\treturn build.Labels[buildapi.BuildConfigLabel]\n\t}\n\treturn build.Labels[buildapi.BuildConfigLabelDeprecated]\n}", "func buildPodMapTask(task *idlCore.TaskTemplate, metadata core.TaskExecutionMetadata) (v1.Pod, error) {\n\tif task.GetK8SPod() == nil || task.GetK8SPod().PodSpec == nil {\n\t\treturn v1.Pod{}, errors.Errorf(errors.BadTaskSpecification, \"Missing pod spec for task\")\n\t}\n\tvar podSpec = &v1.PodSpec{}\n\terr := utils.UnmarshalStructToObj(task.GetK8SPod().PodSpec, &podSpec)\n\tif err != nil {\n\t\treturn v1.Pod{}, errors.Errorf(errors.BadTaskSpecification,\n\t\t\t\"Unable to unmarshal task custom [%v], Err: [%v]\", task.GetCustom(), err.Error())\n\t}\n\tprimaryContainerName, ok := task.GetConfig()[primaryContainerKey]\n\tif !ok {\n\t\treturn v1.Pod{}, errors.Errorf(errors.BadTaskSpecification,\n\t\t\t\"invalid TaskSpecification, config missing [%s] key in [%v]\", primaryContainerKey, task.GetConfig())\n\t}\n\n\tvar pod = v1.Pod{\n\t\tSpec: *podSpec,\n\t}\n\tif task.GetK8SPod().Metadata != nil {\n\t\tif task.GetK8SPod().Metadata.Annotations != nil {\n\t\t\tpod.Annotations = task.GetK8SPod().Metadata.Annotations\n\t\t}\n\t\tif task.GetK8SPod().Metadata.Labels != nil {\n\t\t\tpod.Labels = task.GetK8SPod().Metadata.Labels\n\t\t}\n\t}\n\tif len(pod.Annotations) == 0 {\n\t\tpod.Annotations = make(map[string]string)\n\t}\n\tpod.Annotations[primaryContainerKey] = primaryContainerName\n\n\t// Set the restart policy to *not* inherit from the default so that a completed pod doesn't get caught in a\n\t// CrashLoopBackoff after the initial job completion.\n\tpod.Spec.RestartPolicy = v1.RestartPolicyNever\n\tflytek8s.GetServiceAccountNameFromTaskExecutionMetadata(metadata)\n\treturn pod, nil\n}", "func (o FioSpecVolumeVolumeSourceDownwardAPIItemsResourceFieldRefPtrOutput) ContainerName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FioSpecVolumeVolumeSourceDownwardAPIItemsResourceFieldRef) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ContainerName\n\t}).(pulumi.StringPtrOutput)\n}", "func (f *Function) GetDockerFileName() string {\n\treturn \"Dockerfile\"\n}", "func (d *Docker) GetWorkloadName(ctx context.Context, ID string) (string, error) {\n\tvar containerJSON enginetypes.ContainerJSON\n\tvar err error\n\tutils.WithTimeout(ctx, d.config.GlobalConnectionTimeout, func(ctx context.Context) {\n\t\tcontainerJSON, err = d.client.ContainerInspect(ctx, ID)\n\t})\n\tif err != nil {\n\t\tlog.WithFunc(\"GetWorkloadName\").WithField(\"ID\", ID).Error(ctx, err, \"failed to get container by id\")\n\t\treturn \"\", err\n\t}\n\n\treturn containerJSON.Name, nil\n}", "func GetContainerIDByName(pod *corev1.Pod, containerName string) (string, error) {\n\tupdatedPod := &corev1.Pod{}\n\tkey := types.NamespacedName{\n\t\tName: pod.Name,\n\t\tNamespace: pod.Namespace,\n\t}\n\tif err := testclient.Client.Get(context.TODO(), key, updatedPod); err != nil {\n\t\treturn \"\", err\n\t}\n\tfor _, containerStatus := range updatedPod.Status.ContainerStatuses {\n\t\tif containerStatus.Name == containerName {\n\t\t\treturn strings.Trim(containerStatus.ContainerID, \"cri-o://\"), nil\n\t\t}\n\t}\n\treturn \"\", fmt.Errorf(\"failed to find the container ID for the container %q under the pod %q\", containerName, pod.Name)\n}", "func (o AppTemplateContainerEnvOutput) Name() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AppTemplateContainerEnv) string { return v.Name }).(pulumi.StringOutput)\n}", "func (f *Function) GetDockerTemplateFileName() string {\n\tif f.Runtime == NodeRuntime {\n\t\treturn \"node.Dockerfile\"\n\t}\n\treturn \"golang.Dockerfile\"\n}", "func BuildNameForConfigVersion(name string, version int) string {\n\treturn fmt.Sprintf(\"%s-%d\", name, version)\n}", "func (fh *FilesystemHandler) generateAzureContainerName(blob *models.SimpleBlob) string {\n\tcurrentContainer := blob.ParentContainer\n\treturn currentContainer.Name\n}", "func (o ArgoCDSpecTlsCaOutput) ConfigMapName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ArgoCDSpecTlsCa) *string { return v.ConfigMapName }).(pulumi.StringPtrOutput)\n}", "func (t taskInst) getContainer(named string) *ecs.Container {\n\tfor _, c := range t.Containers {\n\t\tif c != nil && ptr.StringValue(c.Name) == named {\n\t\t\treturn c\n\t\t}\n\t}\n\n\treturn nil\n}", "func namespacedPodName(namespace, name string) string {\n\treturn fmt.Sprintf(\"%s:%s\", namespace, name)\n}", "func GenerateContainerInterfaceName(podName, podNamespace, containerID string) string {\n\t// Use the podName as the prefix and the containerID as the hashing key.\n\t// podNamespace is not used currently.\n\treturn generateInterfaceName(containerID, podName, true)\n}", "func GetContainerId() string {\n\tcmd := \"cat /proc/self/cgroup | grep \\\"docker\\\" | sed s/\\\\\\\\//\\\\\\\\n/g | tail -1\"\n\tout, err := exec.Command(\"bash\", \"-c\", cmd).Output()\n\tif err != nil {\n\t\tlog.Printf(\"Container Id err is %s\\n\", err)\n\t\treturn \"\"\n\t}\n\tlog.Printf(\"The container id is %s\\n\", out)\n\treturn strings.TrimSpace(string(out))\n}", "func GetContainerId() string {\n\tcmd := \"cat /proc/self/cgroup | grep \\\"docker\\\" | sed s/\\\\\\\\//\\\\\\\\n/g | tail -1\"\n\tout, err := exec.Command(\"bash\", \"-c\", cmd).Output()\n\tif err != nil {\n\t\tlog.Printf(\"Container Id err is %s\\n\", err)\n\t\treturn \"\"\n\t}\n\tlog.Printf(\"The container id is %s\\n\", out)\n\treturn strings.TrimSpace(string(out))\n}", "func (c *ConfigurationData) GetPodName() string {\n\treturn c.v.GetString(varPodName)\n}", "func (c *Config) GetPodName() string {\n\treturn c.v.GetString(varPodName)\n}", "func (b *MySQLBox) ContainerName() string {\n\treturn b.containerName\n}", "func ConfigMapName() string {\n\tcm := os.Getenv(configMapNameEnv)\n\tif cm == \"\" {\n\t\treturn \"config-redis\"\n\t}\n\treturn cm\n}", "func (o FluxConfigurationBlobStorageOutput) ContainerId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FluxConfigurationBlobStorage) string { return v.ContainerId }).(pulumi.StringOutput)\n}", "func GetProfileNameFromPodAnnotations(annotations map[string]string, containerName string) string {\n\treturn annotations[ContainerAnnotationKeyPrefix+containerName]\n}", "func (c ArtifactConfig) Name() string {\n\treturn c.BuilderName\n}", "func getUnitNameFromDockerContainer(ctr string) (a string, b string) {\n\tre, _ := regexp.Compile(\"^.*(-[0-9.]+)$\")\n\n\tarr := strings.Split(ctr, \":\")\n\timage := arr[0]\n\ttag := arr[len(arr)-1]\n\tname := strings.Split(image, \"/\")\n\tlast := name[len(name)-1]\n\tfev := re.FindStringSubmatch(last)[1]\n\tsanitzed := strings.Replace(last, fev, \"\", 1)\n\n\treturn sanitzed, tag\n}", "func (o ArgoCDSpecTlsCaPtrOutput) ConfigMapName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ArgoCDSpecTlsCa) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ConfigMapName\n\t}).(pulumi.StringPtrOutput)\n}", "func (cmd ConfigCmd) Name() string {\n\treturn \"config\"\n}", "func (o ClusterBuildStrategySpecBuildStepsEnvFromConfigMapRefOutput) Name() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ClusterBuildStrategySpecBuildStepsEnvFromConfigMapRef) *string { return v.Name }).(pulumi.StringPtrOutput)\n}", "func GetMachineConfigName(profile *performancev2.PerformanceProfile) string {\n\tname := components.GetComponentName(profile.Name, components.ComponentNamePrefix)\n\treturn fmt.Sprintf(\"50-%s\", name)\n}", "func (c *container) Name() string {\n\treturn c.name\n}", "func (c *container) Name() string {\n\treturn c.name\n}", "func (c *container) Name() string {\n\treturn c.name\n}", "func (c wrappedContainer) userContainerName() string {\n\t// The name we want is the one in c.pod.Spec.Containers[0].Name, but this doesn't work\n\t// if the pod has multiple containers, so we parse the container name instead.\n\t// The name has the format \"k8s_<container_name>_[...]\".\n\tsplitName := strings.SplitN(c.ContainerName(), \"_\", 3)\n\tif len(splitName) < 3 {\n\t\treturn \"\"\n\t}\n\n\treturn splitName[1]\n}", "func launchConfigurationName(cluster v1alpha1.AWSConfig, prefix, securityGroupID string) (string, error) {\n\tif key.ClusterID(cluster) == \"\" {\n\t\treturn \"\", microerror.Maskf(missingCloudConfigKeyError, \"spec.cluster.cluster.id\")\n\t}\n\n\tif prefix == \"\" {\n\t\treturn \"\", microerror.Maskf(missingCloudConfigKeyError, \"launchConfiguration prefix\")\n\t}\n\n\tif securityGroupID == \"\" {\n\t\treturn \"\", microerror.Maskf(missingCloudConfigKeyError, \"launchConfiguration securityGroupID\")\n\t}\n\n\treturn fmt.Sprintf(\"%s-%s-%s\", key.ClusterID(cluster), prefix, securityGroupID), nil\n}", "func BuildAtlantisProjectName(context schema.Context, projectNameTemplate string) string {\n\treturn cfg.ReplaceContextTokens(context, projectNameTemplate)\n}", "func (s *BoltState) GetContainerName(id string) (string, error) {\n\tif id == \"\" {\n\t\treturn \"\", define.ErrEmptyID\n\t}\n\n\tif !s.valid {\n\t\treturn \"\", define.ErrDBClosed\n\t}\n\n\tidBytes := []byte(id)\n\n\tdb, err := s.getDBCon()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer s.deferredCloseDBCon(db)\n\n\tname := \"\"\n\n\terr = db.View(func(tx *bolt.Tx) error {\n\t\tidBkt, err := getIDBucket(tx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tctrsBkt, err := getCtrBucket(tx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tnameBytes := idBkt.Get(idBytes)\n\t\tif nameBytes == nil {\n\t\t\treturn define.ErrNoSuchCtr\n\t\t}\n\n\t\tctrExists := ctrsBkt.Bucket(idBytes)\n\t\tif ctrExists == nil {\n\t\t\treturn define.ErrNoSuchCtr\n\t\t}\n\n\t\tname = string(nameBytes)\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn name, nil\n}", "func (o ClusterBuildStrategySpecBuildStepsEnvValueFromConfigMapKeyRefOutput) Name() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ClusterBuildStrategySpecBuildStepsEnvValueFromConfigMapKeyRef) *string { return v.Name }).(pulumi.StringPtrOutput)\n}", "func (o *CreateDeploymentOptions) buildPodSpec() corev1.PodSpec {\n\tpodSpec := corev1.PodSpec{Containers: []corev1.Container{}}\n\tfor _, imageString := range o.Images {\n\t\t// Retain just the image name\n\t\timageSplit := strings.Split(imageString, \"/\")\n\t\tname := imageSplit[len(imageSplit)-1]\n\t\t// Remove any tag or hash\n\t\tif strings.Contains(name, \":\") {\n\t\t\tname = strings.Split(name, \":\")[0]\n\t\t}\n\t\tif strings.Contains(name, \"@\") {\n\t\t\tname = strings.Split(name, \"@\")[0]\n\t\t}\n\t\tname = sanitizeAndUniquify(name)\n\t\tpodSpec.Containers = append(podSpec.Containers, corev1.Container{\n\t\t\tName: name,\n\t\t\tImage: imageString,\n\t\t\tCommand: o.Command,\n\t\t})\n\t}\n\treturn podSpec\n}", "func (c Container) ImageName() string {\n\t// Compatibility w/ Zodiac deployments\n\timageName, ok := c.getLabelValue(zodiacLabel)\n\tif !ok {\n\t\timageName = c.containerInfo.Config.Image\n\t}\n\n\tif !strings.Contains(imageName, \":\") {\n\t\timageName = fmt.Sprintf(\"%s:latest\", imageName)\n\t}\n\n\treturn imageName\n}", "func getContainerNameVersion(c *v1.Container) (version string) {\n\tparts := strings.Split(c.Image, \":\")\n\tif len(parts) != 2 {\n\t\treturn \"\"\n\t}\n\tversion = parts[1]\n\treturn\n}", "func (m *Metadata) getContainerID(podUID string, containerName string) (string, error) {\n\tuid := types.UID(podUID)\n\tfor _, pod := range m.PodsMetadata.Items {\n\t\tif pod.UID == uid {\n\t\t\tfor _, containerStatus := range pod.Status.ContainerStatuses {\n\t\t\t\tif containerName == containerStatus.Name {\n\t\t\t\t\treturn stripContainerID(containerStatus.ContainerID), nil\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn \"\", fmt.Errorf(\"pod %q with container %q not found in the fetched metadata\", podUID, containerName)\n}", "func buildRuntimeName(spec SupervisorSpec, parentName string) string {\n\tvar runtimeName string\n\tif parentName == rootSupervisorName {\n\t\t// We are the root supervisor, no need to add prefix\n\t\truntimeName = spec.GetName()\n\t} else {\n\t\truntimeName = strings.Join([]string{parentName, spec.GetName()}, \"/\")\n\t}\n\treturn runtimeName\n}", "func (o BuildRunSpecBuildRefOutput) Name() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BuildRunSpecBuildRef) string { return v.Name }).(pulumi.StringOutput)\n}", "func makeContainerConfig(info enginetypes.ContainerJSON, image string) (\n\t*enginecontainer.Config,\n\t*enginecontainer.HostConfig,\n\t*enginenetwork.NetworkingConfig,\n\tstring,\n\terror) {\n\n\t// we use `_` to join container name\n\t// since we don't support `_` in entrypoint, and no `_` is in suffix,\n\t// the last part will be suffix and second last part will be entrypoint,\n\t// the rest will be the appname\n\tparts := strings.Split(trimLeftSlash(info.Name), \"_\")\n\tlength := len(parts)\n\tif length < 3 {\n\t\treturn nil, nil, nil, \"\", fmt.Errorf(\"Bad container name format: %q\", info.Name)\n\t}\n\n\tentrypoint := parts[length-2]\n\tappname := strings.Join(parts[:length-2], \"_\")\n\n\tsuffix := utils.RandomString(6)\n\tcontainerName := strings.Join([]string{appname, entrypoint, suffix}, \"_\")\n\n\tconfig := info.Config\n\tconfig.Image = image\n\n\thostConfig := info.HostConfig\n\tnetworkConfig := &enginenetwork.NetworkingConfig{\n\t\tEndpointsConfig: info.NetworkSettings.Networks,\n\t}\n\treturn config, hostConfig, networkConfig, containerName, nil\n}", "func (p *DockerPod) GetNetworkName() string {\n\treturn p.netTask.NetworkName\n}", "func containerID(path, co string) string {\n\tns, n := client.Namespaced(path)\n\tpo := strings.Split(n, \"-\")[0]\n\n\treturn ns + \"/\" + po + \":\" + co\n}", "func (o VirtualDatabaseSpecBuildEnvValueFromConfigMapKeyRefOutput) Name() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v VirtualDatabaseSpecBuildEnvValueFromConfigMapKeyRef) *string { return v.Name }).(pulumi.StringPtrOutput)\n}", "func (o *DeploymentOpts) buildPodSpec() v1.PodSpec {\n\tpodSpec := v1.PodSpec{Containers: []v1.Container{}}\n\tfor _, imageString := range o.Images {\n\t\t// Retain just the image name\n\t\timageSplit := strings.Split(imageString, \"/\")\n\t\tname := imageSplit[len(imageSplit)-1]\n\t\t// Remove any tag or hash\n\t\tif strings.Contains(name, \":\") {\n\t\t\tname = strings.Split(name, \":\")[0]\n\t\t}\n\t\tif strings.Contains(name, \"@\") {\n\t\t\tname = strings.Split(name, \"@\")[0]\n\t\t}\n\t\tname = sanitizeAndUniquify(name)\n\t\tpodSpec.Containers = append(podSpec.Containers, v1.Container{Name: name, Image: imageString})\n\t}\n\treturn podSpec\n}", "func GetPodName() string {\n\treturn os.Getenv(PodName)\n}", "func GetPodName() string {\n\treturn os.Getenv(PodName)\n}", "func (o ClusterBuildStrategySpecBuildStepsEnvFromConfigMapRefPtrOutput) Name() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ClusterBuildStrategySpecBuildStepsEnvFromConfigMapRef) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Name\n\t}).(pulumi.StringPtrOutput)\n}", "func ConfigForTask(taskInfo *mesos.TaskInfo) *docker.CreateContainerOptions {\n\tconfig := &docker.CreateContainerOptions{\n\t\tName: GetContainerName(taskInfo.TaskId),\n\t\tConfig: &docker.Config{\n\t\t\tEnv: EnvForTask(taskInfo),\n\t\t\tExposedPorts: PortsForTask(taskInfo),\n\t\t\tImage: *taskInfo.Container.Docker.Image,\n\t\t\tLabels: LabelsForTask(taskInfo),\n\t\t},\n\t\tHostConfig: &docker.HostConfig{\n\t\t\tBinds: BindsForTask(taskInfo),\n\t\t\tPortBindings: PortBindingsForTask(taskInfo),\n\t\t\tNetworkMode: NetworkForTask(taskInfo),\n\t\t\tCapAdd: CapAddForTask(taskInfo),\n\t\t\tCapDrop: CapDropForTask(taskInfo),\n\t\t},\n\t}\n\n\t// Check for and calculate CPU shares\n\tcpus := getResource(\"cpus\", taskInfo)\n\tif cpus != nil {\n\t\tconfig.Config.CPUShares = int64(*cpus.Scalar.Value * float64(1024))\n\t}\n\n\t// Check for and calculate memory limit\n\tmemory := getResource(\"memoryMb\", taskInfo)\n\tif memory != nil {\n\t\tconfig.Config.Memory = int64(*memory.Scalar.Value * float64(1024*1024))\n\t}\n\n\treturn config\n}", "func (o ContainerServiceOutput) Name() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *ContainerService) pulumi.StringOutput { return v.Name }).(pulumi.StringOutput)\n}", "func (p *pluginConfig) GetConfigName() string {\n\tplugCfg := p.pluginName + \"-config\"\n\tflg := flag.CommandLine.Lookup(plugCfg)\n\tif flg != nil {\n\t\tval := flg.Value.String()\n\n\t\tif val != \"\" {\n\t\t\tplugCfg = val\n\t\t}\n\t}\n\n\treturn plugCfg\n}", "func poolName(endpointsNamespace, endpointsName string) string {\n\treturn fmt.Sprintf(\"openshift_%s_%s\", endpointsNamespace, endpointsName)\n}", "func (o DockerExecutorOutput) ImageName() pulumi.StringOutput {\n\treturn o.ApplyT(func(v DockerExecutor) string { return v.ImageName }).(pulumi.StringOutput)\n}", "func (e *EndToEndTest) GetContainer(t string) string {\n\treturn fmt.Sprintf(\"%s-%s\", e.Identity, t)\n}", "func GetContainerID() (string, error) {\n\tf, err := os.Open(cgroupsFile)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"redpill: error detecting container id: %v\", err)\n\t}\n\n\tdefer f.Close()\n\tscanner := bufio.NewScanner(f)\n\n\t// going to get lines like \"11:pids:/docker/ed807a7d59accf3b9e70c05d52cc494fc125377b82e4\"\n\t// or \"11:pids:/\" if not in container\n\n\tfor scanner.Scan() {\n\t\tline := strings.Split(scanner.Text(), \":\")\n\n\t\tif len(line) != 3 {\n\t\t\t// don't really know what's going on here so ignore\n\t\t\tcontinue\n\t\t}\n\n\t\tns := strings.TrimSpace(line[2])\n\n\t\tif strings.HasPrefix(ns, dockerNsPrefix) {\n\t\t\treturn strings.TrimPrefix(ns, dockerNsPrefix), nil\n\t\t}\n\t}\n\n\treturn \"\", nil\n}", "func (mgr *deploymentManager) imageName() string {\n\treturn fmt.Sprintf(\"%s:%s\", mgr.serverImageName, mgr.serverVersion)\n}", "func (o BuildRunStatusBuildSpecStrategyOutput) Name() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BuildRunStatusBuildSpecStrategy) string { return v.Name }).(pulumi.StringOutput)\n}" ]
[ "0.6439012", "0.62852246", "0.622028", "0.6200449", "0.6197181", "0.61668086", "0.61047316", "0.6056285", "0.60304743", "0.5793392", "0.5791889", "0.57634854", "0.57511026", "0.5727425", "0.5724323", "0.56959724", "0.56770957", "0.56690484", "0.56192434", "0.5603206", "0.5591932", "0.5569635", "0.55652946", "0.5559308", "0.552347", "0.55095756", "0.55065435", "0.55046165", "0.54978764", "0.5472967", "0.5468492", "0.5462318", "0.5451199", "0.54448706", "0.54446214", "0.54418856", "0.5433708", "0.53799933", "0.5370146", "0.5353772", "0.53490806", "0.5336037", "0.53229433", "0.5318941", "0.5316565", "0.52678823", "0.5252197", "0.5251731", "0.5237923", "0.52195203", "0.5200237", "0.5185175", "0.5160784", "0.51580554", "0.515319", "0.5152357", "0.5152357", "0.50906223", "0.5090477", "0.50886506", "0.508398", "0.50834686", "0.50830173", "0.5080693", "0.50709516", "0.5070011", "0.50655687", "0.5063393", "0.5061059", "0.5037244", "0.5037244", "0.5037244", "0.5033903", "0.5029119", "0.5015825", "0.50131", "0.50025433", "0.50023085", "0.50022537", "0.5001889", "0.49984562", "0.4988718", "0.49770015", "0.49696356", "0.49514276", "0.4947053", "0.4945861", "0.49425483", "0.49388728", "0.49388728", "0.49376607", "0.49324864", "0.49272668", "0.49272397", "0.4917837", "0.4915611", "0.4912177", "0.49106383", "0.49076572", "0.49059963" ]
0.7562802
0
StartTask creates and starts a new Container based on the given TaskConfig.
func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *drivers.DriverNetwork, error) { if _, ok := d.tasks.Get(cfg.ID); ok { return nil, nil, fmt.Errorf("task with ID %q already started", cfg.ID) } var driverConfig TaskConfig if err := cfg.DecodeDriverConfig(&driverConfig); err != nil { return nil, nil, fmt.Errorf("failed to decode driver config: %v", err) } handle := drivers.NewTaskHandle(taskHandleVersion) handle.Config = cfg if driverConfig.Image == "" { return nil, nil, fmt.Errorf("image name required") } createOpts := api.SpecGenerator{} createOpts.ContainerBasicConfig.LogConfiguration = &api.LogConfig{} allArgs := []string{} if driverConfig.Command != "" { allArgs = append(allArgs, driverConfig.Command) } allArgs = append(allArgs, driverConfig.Args...) if driverConfig.Entrypoint != "" { createOpts.ContainerBasicConfig.Entrypoint = append(createOpts.ContainerBasicConfig.Entrypoint, driverConfig.Entrypoint) } containerName := BuildContainerName(cfg) // ensure to include port_map into tasks environment map cfg.Env = taskenv.SetPortMapEnvs(cfg.Env, driverConfig.PortMap) // Basic config options createOpts.ContainerBasicConfig.Name = containerName createOpts.ContainerBasicConfig.Command = allArgs createOpts.ContainerBasicConfig.Env = cfg.Env createOpts.ContainerBasicConfig.Hostname = driverConfig.Hostname createOpts.ContainerBasicConfig.Sysctl = driverConfig.Sysctl createOpts.ContainerBasicConfig.LogConfiguration.Path = cfg.StdoutPath // Storage config options createOpts.ContainerStorageConfig.Init = driverConfig.Init createOpts.ContainerStorageConfig.Image = driverConfig.Image createOpts.ContainerStorageConfig.InitPath = driverConfig.InitPath createOpts.ContainerStorageConfig.WorkDir = driverConfig.WorkingDir allMounts, err := d.containerMounts(cfg, &driverConfig) if err != nil { return nil, nil, err } createOpts.ContainerStorageConfig.Mounts = allMounts // Resources config options createOpts.ContainerResourceConfig.ResourceLimits = &spec.LinuxResources{ Memory: &spec.LinuxMemory{}, CPU: &spec.LinuxCPU{}, } if driverConfig.MemoryReservation != "" { reservation, err := memoryInBytes(driverConfig.MemoryReservation) if err != nil { return nil, nil, err } createOpts.ContainerResourceConfig.ResourceLimits.Memory.Reservation = &reservation } if cfg.Resources.NomadResources.Memory.MemoryMB > 0 { limit := cfg.Resources.NomadResources.Memory.MemoryMB * 1024 * 1024 createOpts.ContainerResourceConfig.ResourceLimits.Memory.Limit = &limit } if driverConfig.MemorySwap != "" { swap, err := memoryInBytes(driverConfig.MemorySwap) if err != nil { return nil, nil, err } createOpts.ContainerResourceConfig.ResourceLimits.Memory.Swap = &swap } if !d.cgroupV2 { swappiness := uint64(driverConfig.MemorySwappiness) createOpts.ContainerResourceConfig.ResourceLimits.Memory.Swappiness = &swappiness } // FIXME: can fail for nonRoot due to missing cpu limit delegation permissions, // see https://github.com/containers/podman/blob/master/troubleshooting.md if !d.systemInfo.Host.Rootless { cpuShares := uint64(cfg.Resources.LinuxResources.CPUShares) createOpts.ContainerResourceConfig.ResourceLimits.CPU.Shares = &cpuShares } // Security config options createOpts.ContainerSecurityConfig.CapAdd = driverConfig.CapAdd createOpts.ContainerSecurityConfig.CapDrop = driverConfig.CapDrop createOpts.ContainerSecurityConfig.User = cfg.User // Network config options for _, strdns := range driverConfig.Dns { ipdns := net.ParseIP(strdns) if ipdns == nil { return nil, nil, fmt.Errorf("Invald dns server address") } createOpts.ContainerNetworkConfig.DNSServers = append(createOpts.ContainerNetworkConfig.DNSServers, ipdns) } // Configure network if cfg.NetworkIsolation != nil && cfg.NetworkIsolation.Path != "" { createOpts.ContainerNetworkConfig.NetNS.NSMode = api.Path createOpts.ContainerNetworkConfig.NetNS.Value = cfg.NetworkIsolation.Path } else { if driverConfig.NetworkMode == "" { createOpts.ContainerNetworkConfig.NetNS.NSMode = api.Bridge } else if driverConfig.NetworkMode == "bridge" { createOpts.ContainerNetworkConfig.NetNS.NSMode = api.Bridge } else if driverConfig.NetworkMode == "host" { createOpts.ContainerNetworkConfig.NetNS.NSMode = api.Host } else if driverConfig.NetworkMode == "none" { createOpts.ContainerNetworkConfig.NetNS.NSMode = api.NoNetwork } else if driverConfig.NetworkMode == "slirp4netns" { createOpts.ContainerNetworkConfig.NetNS.NSMode = api.Slirp } else { return nil, nil, fmt.Errorf("Unknown/Unsupported network mode: %s", driverConfig.NetworkMode) } } portMappings, err := d.portMappings(cfg, driverConfig) if err != nil { return nil, nil, err } createOpts.ContainerNetworkConfig.PortMappings = portMappings containerID := "" recoverRunningContainer := false // check if there is a container with same name otherContainerInspect, err := d.podman.ContainerInspect(d.ctx, containerName) if err == nil { // ok, seems we found a container with similar name if otherContainerInspect.State.Running { // it's still running. So let's use it instead of creating a new one d.logger.Info("Detect running container with same name, we reuse it", "task", cfg.ID, "container", otherContainerInspect.ID) containerID = otherContainerInspect.ID recoverRunningContainer = true } else { // let's remove the old, dead container d.logger.Info("Detect stopped container with same name, removing it", "task", cfg.ID, "container", otherContainerInspect.ID) if err = d.podman.ContainerDelete(d.ctx, otherContainerInspect.ID, true, true); err != nil { return nil, nil, nstructs.WrapRecoverable(fmt.Sprintf("failed to remove dead container: %v", err), err) } } } if !recoverRunningContainer { // FIXME: there are more variations of image sources, we should handle it // e.g. oci-archive:/... etc // see also https://github.com/hashicorp/nomad-driver-podman/issues/69 // do we already have this image in local storage? haveImage, err := d.podman.ImageExists(d.ctx, createOpts.Image) if err != nil { return nil, nil, fmt.Errorf("failed to start task, unable to check for local image: %v", err) } if !haveImage { // image is not in local storage, so we need to pull it if err = d.podman.ImagePull(d.ctx, createOpts.Image); err != nil { return nil, nil, fmt.Errorf("failed to start task, unable to pull image %s: %v", createOpts.Image, err) } } createResponse, err := d.podman.ContainerCreate(d.ctx, createOpts) for _, w := range createResponse.Warnings { d.logger.Warn("Create Warning", "warning", w) } if err != nil { return nil, nil, fmt.Errorf("failed to start task, could not create container: %v", err) } containerID = createResponse.Id } cleanup := func() { d.logger.Debug("Cleaning up", "container", containerID) if err := d.podman.ContainerDelete(d.ctx, containerID, true, true); err != nil { d.logger.Error("failed to clean up from an error in Start", "error", err) } } if !recoverRunningContainer { if err = d.podman.ContainerStart(d.ctx, containerID); err != nil { cleanup() return nil, nil, fmt.Errorf("failed to start task, could not start container: %v", err) } } inspectData, err := d.podman.ContainerInspect(d.ctx, containerID) if err != nil { d.logger.Error("failed to inspect container", "err", err) cleanup() return nil, nil, fmt.Errorf("failed to start task, could not inspect container : %v", err) } net := &drivers.DriverNetwork{ PortMap: driverConfig.PortMap, IP: inspectData.NetworkSettings.IPAddress, AutoAdvertise: true, } h := &TaskHandle{ containerID: containerID, driver: d, taskConfig: cfg, procState: drivers.TaskStateRunning, exitResult: &drivers.ExitResult{}, startedAt: time.Now().Round(time.Millisecond), logger: d.logger.Named("podmanHandle"), totalCPUStats: stats.NewCpuStats(), userCPUStats: stats.NewCpuStats(), systemCPUStats: stats.NewCpuStats(), removeContainerOnExit: d.config.GC.Container, } driverState := TaskState{ ContainerID: containerID, TaskConfig: cfg, StartedAt: h.startedAt, Net: net, } if err := handle.SetDriverState(&driverState); err != nil { d.logger.Error("failed to start task, error setting driver state", "error", err) cleanup() return nil, nil, fmt.Errorf("failed to set driver state: %v", err) } d.tasks.Set(cfg.ID, h) go h.runContainerMonitor() d.logger.Info("Completely started container", "taskID", cfg.ID, "container", containerID, "ip", inspectData.NetworkSettings.IPAddress) return handle, net, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (e *ECS) StartTask(req *StartTaskReq) (*StartTaskResp, error) {\n\tif req == nil {\n\t\treturn nil, fmt.Errorf(\"The req params cannot be nil\")\n\t}\n\n\tparams := makeParams(\"StartTask\")\n\tif req.Cluster != \"\" {\n\t\tparams[\"cluster\"] = req.Cluster\n\t}\n\tif req.TaskDefinition != \"\" {\n\t\tparams[\"taskDefinition\"] = req.TaskDefinition\n\t}\n\tfor i, ci := range req.ContainerInstances {\n\t\tparams[fmt.Sprintf(\"containerInstances.member.%d\", i+1)] = ci\n\t}\n\tfor i, co := range req.Overrides.ContainerOverrides {\n\t\tkey := fmt.Sprintf(\"overrides.containerOverrides.member.%d\", i+1)\n\t\tparams[fmt.Sprintf(\"%s.name\", key)] = co.Name\n\t\tfor k, cmd := range co.Command {\n\t\t\tparams[fmt.Sprintf(\"%s.command.member.%d\", key, k+1)] = cmd\n\t\t}\n\t}\n\n\tresp := new(StartTaskResp)\n\tif err := e.query(params, resp); err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func (d *DockerDriver) Start(TaskUUID string, image string, cmd []string, target string) string {\n\tlog.Println(\"Starting container \" + TaskUUID)\n\tcli := d.getClient()\n\n\tportmaps := make(nat.PortMap, 1)\n\tportbindings := make([]nat.PortBinding, 1)\n\tportbindings[0] = nat.PortBinding{\n\t\tHostPort: \"\",\n\t}\n\tportmaps[\"8000/tcp\"] = portbindings\n\n\tcreatedContainer, err := cli.ContainerCreate(\n\t\tcontext.TODO(),\n\t\t&container.Config{\n\t\t\tHostname: TaskUUID,\n\t\t\tEnv: []string{},\n\t\t\tImage: image,\n\t\t\tCmd: cmd,\n\t\t},\n\t\t&container.HostConfig{\n\t\t\tPortBindings: portmaps,\n\t\t\tAutoRemove: true,\n\t\t\tBinds: []string{\n\t\t\t\ttarget + \":/srv\",\n\t\t\t},\n\t\t},\n\t\t&network.NetworkingConfig{},\n\t\tTaskUUID+\".\"+strconv.Itoa(rand.Int()), // for multiple containers per task per server\n\t)\n\tif err != nil {\n\t\tpanic(shadowerrors.ShadowError{\n\t\t\tOrigin: err,\n\t\t\tVisibleMessage: \"start containers container error\",\n\t\t})\n\t}\n\n\tcontainerId := createdContainer.ID\n\n\terr = cli.ContainerStart(context.TODO(), createdContainer.ID, types.ContainerStartOptions{})\n\n\treturn containerId\n}", "func (lenc *Lencak) StartTask(workSpaceName, taskName string, asService bool) bool {\n\treturn lenc.WithWorkspaceTask(workSpaceName, taskName, func(task *Task) {\n\t\tif asService {\n\t\t\ttask.serviceMu.Lock()\n\t\t\ttask.Service = true\n\t\t\ttask.serviceMu.Unlock()\n\t\t\tif task.ActiveTask == nil {\n\t\t\t\ttask.Start(lenc.sync)\n\t\t\t}\n\t\t} else {\n\t\t\ttask.Start(lenc.sync)\n\t\t}\n\t})\n}", "func (c *ECS) StartTask(input *StartTaskInput) (output *StartTaskOutput, err error) {\n\treq, out := c.StartTaskRequest(input)\n\toutput = out\n\terr = req.Send()\n\treturn\n}", "func (c *Client) StartTask(ctx context.Context, params *StartTaskInput, optFns ...func(*Options)) (*StartTaskOutput, error) {\n\tif params == nil {\n\t\tparams = &StartTaskInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"StartTask\", params, optFns, addOperationStartTaskMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*StartTaskOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (c Control) ServeStartTask(w http.ResponseWriter, r *http.Request) {\n\tc.ServeTaskAction(w, r, true)\n}", "func (s *Session) Start(tags map[string]interface{}) error {\n\tif s.task != nil {\n\t\treturn errors.New(\"task already running\")\n\t}\n\n\tt := &task{\n\t\tName: s.TaskName,\n\t\tVersion: 1,\n\t\tSchedule: s.Schedule,\n\t}\n\n\twf := wmap.NewWorkflowMap()\n\n\tsnapTags := make(map[string]string)\n\tfor key, value := range tags {\n\t\tsnapTags[key] = fmt.Sprintf(\"%v\", value)\n\t}\n\twf.CollectNode.Tags = map[string]map[string]string{\"\": snapTags}\n\n\tfor _, metric := range s.Metrics {\n\t\twf.CollectNode.AddMetric(metric, -1)\n\t}\n\n\tfor _, configItem := range s.CollectNodeConfigItems {\n\t\twf.CollectNode.AddConfigItem(configItem.Ns, configItem.Key, configItem.Value)\n\t}\n\n\tloaderConfig := DefaultPluginLoaderConfig()\n\tloaderConfig.SnapteldAddress = s.pClient.URL\n\n\t// Add specified publisher to workflow as well.\n\twf.CollectNode.Add(s.Publisher)\n\n\tt.Workflow = wf\n\n\tr := s.pClient.CreateTask(t.Schedule, t.Workflow, t.Name, t.Deadline, true, 10)\n\tif r.Err != nil {\n\t\treturn errors.Wrapf(r.Err, \"could not create task %q\", t.Name)\n\t}\n\n\t// Save a copy of the task so we can stop it again.\n\tt.ID = r.ID\n\tt.State = r.State\n\ts.task = t\n\n\treturn nil\n}", "func (i *TaskRegisterUpdater) StartTask(ctx context.Context, action string, age time.Duration) (models.Task, error) {\n\n\treturn i.repository.GetTask(ctx, action, age)\n}", "func (ts *TaskService) Create(ctx context.Context, req *taskAPI.CreateTaskRequest) (*taskAPI.CreateTaskResponse, error) {\n\tdefer logPanicAndDie(log.G(ctx))\n\tlogger := log.G(ctx).WithFields(logrus.Fields{\"id\": req.ID, \"bundle\": req.Bundle})\n\tlogger.Info(\"create\")\n\n\textraData, err := unmarshalExtraData(req.Options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbundleDir := bundle.Dir(filepath.Join(containerRootDir, req.ID))\n\terr = bundleDir.Create()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = bundleDir.OCIConfig().Write(extraData.JsonSpec)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// TODO replace with proper drive mounting once that PR is merged. Right now, all containers in\n\t// this VM start up with the same rootfs image no matter their configuration\n\terr = bundleDir.MountRootfs(\"/dev/vdb\", \"ext4\", nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Create a runc shim to manage this task\n\t// TODO if we update to the v2 runc implementation in containerd, we can use a single\n\t// runc service instance to manage all tasks instead of creating a new one for each\n\truncService, err := runc.New(ctx, req.ID, ts.eventExchange)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Override the incoming stdio FIFOs, which have paths from the host that we can't use\n\tfifoSet, err := cio.NewFIFOSetInDir(bundleDir.RootPath(), req.ID, req.Terminal)\n\tif err != nil {\n\t\tlog.G(ctx).WithError(err).Error(\"failed opening stdio FIFOs\")\n\t\treturn nil, err\n\t}\n\n\t// Don't try to connect any io streams that weren't requested by the client\n\tif req.Stdin == \"\" {\n\t\tfifoSet.Stdin = \"\"\n\t}\n\n\tif req.Stdout == \"\" {\n\t\tfifoSet.Stdout = \"\"\n\t}\n\n\tif req.Stderr == \"\" {\n\t\tfifoSet.Stderr = \"\"\n\t}\n\n\ttask, err := ts.taskManager.AddTask(ctx, req.ID, runcService, bundleDir, extraData, fifoSet)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlogger.Debug(\"calling runc create\")\n\n\t// Override some of the incoming paths, which were set on the Host and thus not valid here in the Guest\n\treq.Bundle = bundleDir.RootPath()\n\treq.Rootfs = nil\n\treq.Stdin = fifoSet.Stdin\n\treq.Stdout = fifoSet.Stdout\n\treq.Stderr = fifoSet.Stderr\n\n\t// Just provide runc the options it knows about, not our wrapper\n\treq.Options = task.ExtraData().GetRuncOptions()\n\n\t// Start the io proxy and wait for initialization to complete before starting\n\t// the task to ensure we capture all task output\n\terr = <-task.StartStdioProxy(ctx, vm.VSockToFIFO, acceptVSock)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp, err := task.Create(ctx, req)\n\tif err != nil {\n\t\tlogger.WithError(err).Error(\"error creating container\")\n\t\treturn nil, err\n\t}\n\n\tlogger.WithField(\"pid\", resp.Pid).Debugf(\"create succeeded\")\n\treturn resp, nil\n}", "func (h *Hub) StartTask(ctx context.Context, request *pb.HubStartTaskRequest) (*pb.HubStartTaskReply, error) {\n\tlog.G(h.ctx).Info(\"handling StartTask request\", zap.Any(\"req\", request))\n\n\ttaskID := uuid.New()\n\tminer, err := h.selectMiner(request)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar startRequest = &pb.MinerStartRequest{\n\t\tId: taskID,\n\t\tRegistry: request.Registry,\n\t\tImage: request.Image,\n\t\tAuth: request.Auth,\n\t\tPublicKeyData: request.PublicKeyData,\n\t\tCommitOnStop: request.CommitOnStop,\n\t\tEnv: request.Env,\n\t\tUsage: request.Requirements.GetResources(),\n\t\tRestartPolicy: &pb.ContainerRestartPolicy{\n\t\t\tName: \"\",\n\t\t\tMaximumRetryCount: 0,\n\t\t},\n\t}\n\n\tresp, err := miner.Client.Start(ctx, startRequest)\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.Internal, \"failed to start %v\", err)\n\t}\n\n\troutes := []extRoute{}\n\tfor k, v := range resp.Ports {\n\t\t_, protocol, err := decodePortBinding(k)\n\t\tif err != nil {\n\t\t\tlog.G(h.ctx).Warn(\"failed to decode miner's port mapping\",\n\t\t\t\tzap.String(\"mapping\", k),\n\t\t\t\tzap.Error(err),\n\t\t\t)\n\t\t\tcontinue\n\t\t}\n\n\t\trealPort, err := strconv.ParseUint(v.Port, 10, 16)\n\t\tif err != nil {\n\t\t\tlog.G(h.ctx).Warn(\"failed to convert real port to uint16\",\n\t\t\t\tzap.Error(err),\n\t\t\t\tzap.String(\"port\", v.Port),\n\t\t\t)\n\t\t\tcontinue\n\t\t}\n\n\t\troute, err := miner.router.RegisterRoute(taskID, protocol, v.IP, uint16(realPort))\n\t\tif err != nil {\n\t\t\tlog.G(h.ctx).Warn(\"failed to register route\", zap.Error(err))\n\t\t\tcontinue\n\t\t}\n\t\troutes = append(routes, extRoute{\n\t\t\tcontainerPort: k,\n\t\t\troute: route,\n\t\t})\n\t}\n\n\th.setMinerTaskID(miner.ID(), taskID)\n\n\tresources := request.GetRequirements().GetResources()\n\tcpuCount := resources.GetCPUCores()\n\tmemoryCount := resources.GetMaxMemory()\n\n\tvar usage = resource.NewResources(int(cpuCount), int64(memoryCount))\n\tif err := miner.Consume(taskID, &usage); err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar reply = pb.HubStartTaskReply{\n\t\tId: taskID,\n\t}\n\n\tfor _, route := range routes {\n\t\treply.Endpoint = append(\n\t\t\treply.Endpoint,\n\t\t\tfmt.Sprintf(\"%s->%s:%d\", route.containerPort, route.route.Host, route.route.Port),\n\t\t)\n\t}\n\n\treturn &reply, nil\n}", "func (ts *TaskService) Start(ctx context.Context, req *taskAPI.StartRequest) (*taskAPI.StartResponse, error) {\n\tdefer logPanicAndDie(log.G(ctx))\n\n\tlog.G(ctx).WithFields(logrus.Fields{\"id\": req.ID, \"exec_id\": req.ExecID}).Debug(\"start\")\n\ttask, err := ts.taskManager.Task(req.ID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx = namespaces.WithNamespace(ctx, defaultNamespace)\n\tresp, err := task.Start(ctx, req)\n\tif err != nil {\n\t\tlog.G(ctx).WithError(err).Error(\"start failed\")\n\t\treturn nil, err\n\t}\n\n\tlog.G(ctx).WithField(\"pid\", resp.Pid).Debug(\"start succeeded\")\n\treturn resp, nil\n}", "func Start(ctx context.Context, name string, args ...interface{}) context.Context {\n\tparent := GetTask(ctx)\n\tif parent == nil {\n\t\tparent = &app\n\t}\n\tt := &Task{\n\t\tid: atomic.AddUint64(&nextID, 1),\n\t\tname: fmt.Sprintf(name, args...),\n\t\tbegun: time.Now(),\n\t\tparent: parent,\n\t\tchildren: map[*Task]struct{}{},\n\t\tbackground: false,\n\t}\n\tlog.D(ctx, \"Starting task: %s\", t.name)\n\n\tparent.add(t)\n\tonTaskStart(ctx, t)\n\treturn PutTask(ctx, t)\n}", "func (ts *TaskService) Create(requestCtx context.Context, req *taskAPI.CreateTaskRequest) (*taskAPI.CreateTaskResponse, error) {\n\tdefer logPanicAndDie(log.G(requestCtx))\n\ttaskID := req.ID\n\texecID := \"\" // the exec ID of the initial process in a task is an empty string by containerd convention\n\n\tlogger := log.G(requestCtx).WithField(\"TaskID\", taskID).WithField(\"ExecID\", execID)\n\tlogger.Info(\"create\")\n\n\textraData, err := unmarshalExtraData(req.Options)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to unmarshal extra data\")\n\t}\n\n\t// Just provide runc the options it knows about, not our wrapper\n\treq.Options = extraData.RuncOptions\n\n\t// Override the bundle dir and rootfs paths, which were set on the Host and thus not valid here in the Guest\n\tbundleDir := bundle.Dir(filepath.Join(containerRootDir, taskID))\n\terr = bundleDir.Create()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to create bundle dir\")\n\t}\n\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tremoveErr := os.RemoveAll(bundleDir.RootPath())\n\t\t\tif removeErr != nil {\n\t\t\t\tlogger.WithError(removeErr).Error(\"failed to cleanup bundle dir\")\n\t\t\t}\n\t\t}\n\t}()\n\n\terr = bundleDir.OCIConfig().Write(extraData.JsonSpec)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to write oci config file\")\n\t}\n\n\tdriveID := strings.TrimSpace(extraData.DriveID)\n\tdrive, ok := ts.driveHandler.GetDrive(driveID)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"Drive %q could not be found\", driveID)\n\t}\n\n\tconst (\n\t\tmaxRetries = 100\n\t\tretryDelay = 10 * time.Millisecond\n\t)\n\n\t// We retry here due to guest kernel needing some time to populate the guest\n\t// drive.\n\t// https://github.com/firecracker-microvm/firecracker/issues/1159\n\tfor i := 0; i < maxRetries; i++ {\n\t\tif err = bundleDir.MountRootfs(drive.Path(), \"ext4\", nil); isRetryableMountError(err) {\n\t\t\tlogger.WithError(err).Warnf(\"retrying to mount rootfs %q\", drive.Path())\n\n\t\t\ttime.Sleep(retryDelay)\n\t\t\tcontinue\n\t\t}\n\n\t\tbreak\n\t}\n\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to mount rootfs %q\", drive.Path())\n\t}\n\n\treq.Bundle = bundleDir.RootPath()\n\treq.Rootfs = nil\n\n\tvar ioConnectorSet vm.IOProxy\n\n\tif vm.IsAgentOnlyIO(req.Stdout, logger) {\n\t\tioConnectorSet = vm.NewNullIOProxy()\n\t} else {\n\t\t// Override the incoming stdio FIFOs, which have paths from the host that we can't use\n\t\tfifoSet, err := cio.NewFIFOSetInDir(bundleDir.RootPath(), fifoName(taskID, execID), req.Terminal)\n\t\tif err != nil {\n\t\t\tlogger.WithError(err).Error(\"failed opening stdio FIFOs\")\n\t\t\treturn nil, errors.Wrap(err, \"failed to open stdio FIFOs\")\n\t\t}\n\n\t\tvar stdinConnectorPair *vm.IOConnectorPair\n\t\tif req.Stdin != \"\" {\n\t\t\treq.Stdin = fifoSet.Stdin\n\t\t\tstdinConnectorPair = &vm.IOConnectorPair{\n\t\t\t\tReadConnector: vm.VSockAcceptConnector(extraData.StdinPort),\n\t\t\t\tWriteConnector: vm.FIFOConnector(fifoSet.Stdin),\n\t\t\t}\n\t\t}\n\n\t\tvar stdoutConnectorPair *vm.IOConnectorPair\n\t\tif req.Stdout != \"\" {\n\t\t\treq.Stdout = fifoSet.Stdout\n\t\t\tstdoutConnectorPair = &vm.IOConnectorPair{\n\t\t\t\tReadConnector: vm.FIFOConnector(fifoSet.Stdout),\n\t\t\t\tWriteConnector: vm.VSockAcceptConnector(extraData.StdoutPort),\n\t\t\t}\n\t\t}\n\n\t\tvar stderrConnectorPair *vm.IOConnectorPair\n\t\tif req.Stderr != \"\" {\n\t\t\treq.Stderr = fifoSet.Stderr\n\t\t\tstderrConnectorPair = &vm.IOConnectorPair{\n\t\t\t\tReadConnector: vm.FIFOConnector(fifoSet.Stderr),\n\t\t\t\tWriteConnector: vm.VSockAcceptConnector(extraData.StderrPort),\n\t\t\t}\n\t\t}\n\n\t\tioConnectorSet = vm.NewIOConnectorProxy(stdinConnectorPair, stdoutConnectorPair, stderrConnectorPair)\n\t}\n\n\tresp, err := ts.taskManager.CreateTask(requestCtx, req, ts.runcService, ioConnectorSet)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to create runc shim for task\")\n\t}\n\n\tlogger.WithField(\"pid\", resp.Pid).Debugf(\"create succeeded\")\n\treturn resp, nil\n}", "func StartTaskService(brain *brain.Manager, errChan chan error) {\n\tlis, err := net.Listen(\"tcp\", taskServicePort)\n\tif err != nil {\n\t\terrChan <- err\n\t\treturn\n\t}\n\n\tgrpcServer := grpc.NewServer()\n\n\tRegisterTaskServiceServer(grpcServer, TaskService{Manager: brain})\n\n\tlog.LogInfo(\"starting taask-server task service on :3688\")\n\tif err := grpcServer.Serve(lis); err != nil {\n\t\terrChan <- err\n\t}\n}", "func StartDMTask(fw portforward.PortForward, ns, masterSvcName, taskConf, errSubStr string) error {\n\tapiPath := \"/apis/v1alpha1/tasks\"\n\n\ttype Req struct {\n\t\tTask string `json:\"task\"`\n\t}\n\ttype Resp struct {\n\t\tResult bool `json:\"result\"`\n\t\tMsg string `json:\"msg\"`\n\t\tCheckResult string `json:\"checkResult\"`\n\t}\n\n\tvar req = Req{\n\t\tTask: fmt.Sprintf(taskConf, DMTaskName(ns), v1alpha1.DefaultTiDBServerPort, DMTaskName(ns)),\n\t}\n\tdata, err := json.Marshal(req)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to marshal task start request, %v, %v\", req, err)\n\t}\n\n\treturn wait.Poll(5*time.Second, time.Minute, func() (bool, error) {\n\t\tlocalHost, localPort, cancel, err := portforward.ForwardOnePort(\n\t\t\tfw, ns, fmt.Sprintf(\"svc/%s\", masterSvcName), dmMasterSvcPort)\n\t\tif err != nil {\n\t\t\tlog.Logf(\"failed to forward dm-master svc: %v\", err)\n\t\t\treturn false, nil\n\t\t}\n\t\tdefer cancel()\n\n\t\tbody, err := httputil.DoBodyOK(\n\t\t\t&http.Client{Transport: &http.Transport{}},\n\t\t\tfmt.Sprintf(\"http://%s:%d%s\", localHost, localPort, apiPath),\n\t\t\t\"POST\",\n\t\t\tbytes.NewReader(data))\n\t\tif err != nil {\n\t\t\tlog.Logf(\"failed to start DM task: %v\", err)\n\t\t\treturn false, nil\n\t\t}\n\t\tvar resp Resp\n\t\tif err = json.Unmarshal(body, &resp); err != nil {\n\t\t\tlog.Logf(\"failed to unmarshal DM task start response, %s: %v\", string(body), err)\n\t\t\treturn false, nil\n\t\t} else if !resp.Result && !strings.Contains(resp.Msg, \"already exists\") {\n\t\t\tif errSubStr != \"\" && strings.Contains(resp.Msg, errSubStr) {\n\t\t\t\tlog.Logf(\"start DM task match the error sub string %q: %s\", errSubStr, resp.Msg)\n\t\t\t\treturn true, nil\n\t\t\t}\n\t\t\tlog.Logf(\"failed to start DM task, msg: %s, err: %v, checkResult: %s\", resp.Msg, err, resp.CheckResult)\n\t\t\treturn false, nil\n\t\t}\n\t\treturn true, nil\n\t})\n}", "func (rm *ResponseManager) StartTask(task *peertask.Task, responseTaskDataChan chan<- ResponseTaskData) {\n\trm.send(&startTaskRequest{task, responseTaskDataChan}, nil)\n}", "func (c *Client) StartImportTask(ctx context.Context, params *StartImportTaskInput, optFns ...func(*Options)) (*StartImportTaskOutput, error) {\n\tif params == nil {\n\t\tparams = &StartImportTaskInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"StartImportTask\", params, optFns, addOperationStartImportTaskMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*StartImportTaskOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (w *Worker) StartSubTask(cfg *config.SubTaskConfig) (int64, error) {\n\tw.Lock()\n\tdefer w.Unlock()\n\n\t// copy some config item from dm-worker's config\n\tw.copyConfigFromWorker(cfg)\n\tcfgStr, err := cfg.Toml()\n\tif err != nil {\n\t\treturn 0, terror.Annotatef(err, \"encode subtask %+v into toml format\", cfg)\n\t}\n\n\topLogID, err := w.operateSubTask(&pb.TaskMeta{\n\t\tOp: pb.TaskOp_Start,\n\t\tName: cfg.Name,\n\t\tTask: append([]byte{}, cfgStr...),\n\t})\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn opLogID, nil\n}", "func CtrStartContainer(domainName string) (int, error) {\n\tif err := verifyCtr(); err != nil {\n\t\treturn 0, fmt.Errorf(\"CtrStartContainer: exception while verifying ctrd client: %s\", err.Error())\n\t}\n\tctr, err := CtrLoadContainer(domainName)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tlogger := GetLog()\n\n\tio := func(id string) (cio.IO, error) {\n\t\tstdoutFile := logger.Path(domainName + \".out\")\n\t\tstderrFile := logger.Path(domainName)\n\t\treturn &logio{\n\t\t\tcio.Config{\n\t\t\t\tStdin: \"/dev/null\",\n\t\t\t\tStdout: stdoutFile,\n\t\t\t\tStderr: stderrFile,\n\t\t\t\tTerminal: false,\n\t\t\t},\n\t\t}, nil\n\t}\n\ttask, err := ctr.NewTask(ctrdCtx, io)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tif err := prepareProcess(int(task.Pid()), nil); err != nil {\n\t\treturn 0, err\n\t}\n\n\tif err := task.Start(ctrdCtx); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn int(task.Pid()), nil\n}", "func (c *Client) StartBkOpsTask(url string, paras *TaskPathParas,\n\trequest *StartTaskRequest) (*StartTaskResponse, error) {\n\tif c == nil {\n\t\treturn nil, ErrServerNotInit\n\t}\n\n\tvar (\n\t\treqURL = fmt.Sprintf(\"/start_task/%s/%s/\", paras.TaskID, paras.BkBizID)\n\t\trespData = &StartTaskResponse{}\n\t)\n\n\trequest.Scope = string(CmdbBizScope)\n\tuserAuth, err := c.generateGateWayAuth(paras.Operator)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"bksops StartBkOpsTask generateGateWayAuth failed: %v\", err)\n\t}\n\n\t_, _, errs := gorequest.New().\n\t\tTimeout(defaultTimeOut).\n\t\tPost(c.server+reqURL).\n\t\tSet(\"Content-Type\", \"application/json\").\n\t\tSet(\"Accept\", \"application/json\").\n\t\tSet(\"X-Bkapi-Authorization\", userAuth).\n\t\tSetDebug(c.serverDebug).\n\t\tSend(request).\n\t\tEndStruct(&respData)\n\tif len(errs) > 0 {\n\t\tblog.Errorf(\"call api StartBkOpsTask failed: %v\", errs[0])\n\t\treturn nil, errs[0]\n\t}\n\n\tif !respData.Result {\n\t\tblog.Errorf(\"call api StartBkOpsTask failed: %v\", respData.Message)\n\t\treturn nil, fmt.Errorf(respData.Message)\n\t}\n\n\t//successfully request\n\tblog.Infof(\"call api StartBkOpsTask with url(%s) successfully\", reqURL)\n\treturn respData, nil\n}", "func Start(ctx context.Context, fun TaskFunc) Task {\n\treturn (&taskGroup{ctx: ctx}).Task(fun)\n}", "func (c *ECS) StartTaskRequest(input *StartTaskInput) (req *aws.Request, output *StartTaskOutput) {\n\toprw.Lock()\n\tdefer oprw.Unlock()\n\n\tif opStartTask == nil {\n\t\topStartTask = &aws.Operation{\n\t\t\tName: \"StartTask\",\n\t\t\tHTTPMethod: \"POST\",\n\t\t\tHTTPPath: \"/\",\n\t\t}\n\t}\n\n\treq = c.newRequest(opStartTask, input, output)\n\toutput = &StartTaskOutput{}\n\treq.Data = output\n\treturn\n}", "func (ts *TaskService) Start(requestCtx context.Context, req *taskAPI.StartRequest) (*taskAPI.StartResponse, error) {\n\tdefer logPanicAndDie(log.G(requestCtx))\n\tlog.G(requestCtx).WithFields(logrus.Fields{\"id\": req.ID, \"exec_id\": req.ExecID}).Debug(\"start\")\n\n\tresp, err := ts.runcService.Start(requestCtx, req)\n\tif err != nil {\n\t\tlog.G(requestCtx).WithError(err).Error(\"start failed\")\n\t\treturn nil, err\n\t}\n\n\tlog.G(requestCtx).WithField(\"pid\", resp.Pid).Debug(\"start succeeded\")\n\treturn resp, nil\n}", "func InitiateRakeTask(taskName string, settings *models.Settings) {\n\trakeTask := map[string]string{}\n\tb, err := json.Marshal(rakeTask)\n\tif err != nil {\n\t\tfmt.Println(err.Error())\n\t\tos.Exit(1)\n\t}\n\tencodedTaskName, err := url.Parse(taskName)\n\tif err != nil {\n\t\tfmt.Println(err.Error())\n\t\tos.Exit(1)\n\t}\n\thttpclient.Post(b, fmt.Sprintf(\"%s/v1/environments/%s/services/%s/rake/%s\", settings.PaasHost, settings.EnvironmentID, settings.ServiceID, encodedTaskName), true, settings)\n}", "func NewTask(taskConfig TaskConfig, displayStartIdx int, replicaValue string) *Task {\n\ttask := Task{Config: taskConfig}\n\ttask.inflate(displayStartIdx, replicaValue)\n\n\tfor subIndex := range taskConfig.ParallelTasks {\n\t\tsubTaskConfig := &taskConfig.ParallelTasks[subIndex]\n\n\t\tsubTask := NewTask(*subTaskConfig, nextDisplayIdx, replicaValue)\n\t\tsubTask.Display.Template = lineParallelTemplate\n\t\ttask.Children = append(task.Children, subTask)\n\t\tnextDisplayIdx++\n\t}\n\n\tif len(task.Children) > 0 {\n\t\ttask.Children[len(task.Children)-1].Display.Template = lineLastParallelTemplate\n\t}\n\treturn &task\n}", "func runTask(ctx context.Context, cfg *taskConfig, f TaskFunc, args ...interface{}) *Task {\n\ttask := newTask(ctx, cfg, f, args...)\n\ttask.Start()\n\treturn task\n}", "func NewTask(service *Service) *Task {\n\treturn &Task{\n\t\tTask: task.New(StatusTaskStarted),\n\t\tService: service,\n\t}\n}", "func (ecs *ECS) CreateTaskDefinition(input *CreateTaskDefinitionInput) string {\n\tconsole.Debug(\"Creating ECS task definition\")\n\n\tlogConfiguration := &awsecs.LogConfiguration{\n\t\tLogDriver: aws.String(awsecs.LogDriverAwslogs),\n\t\tOptions: map[string]*string{\n\t\t\t\"awslogs-region\": aws.String(input.LogRegion),\n\t\t\t\"awslogs-group\": aws.String(input.LogGroupName),\n\t\t\t\"awslogs-stream-prefix\": aws.String(logStreamPrefix),\n\t\t},\n\t}\n\n\tcontainerDefinition := &awsecs.ContainerDefinition{\n\t\tEnvironment: input.Environment(),\n\t\tEssential: aws.Bool(true),\n\t\tImage: aws.String(input.Image),\n\t\tLogConfiguration: logConfiguration,\n\t\tName: aws.String(input.Name),\n\t\tSecrets: input.Secrets(),\n\t}\n\n\tif input.Port != 0 {\n\t\tcontainerDefinition.SetPortMappings(\n\t\t\t[]*awsecs.PortMapping{\n\t\t\t\t&awsecs.PortMapping{\n\t\t\t\t\tContainerPort: aws.Int64(int64(input.Port)),\n\t\t\t\t},\n\t\t\t},\n\t\t)\n\t}\n\n\tresp, err := ecs.svc.RegisterTaskDefinition(\n\t\t&awsecs.RegisterTaskDefinitionInput{\n\t\t\tContainerDefinitions: []*awsecs.ContainerDefinition{containerDefinition},\n\t\t\tCpu: aws.String(input.Cpu),\n\t\t\tExecutionRoleArn: aws.String(input.ExecutionRoleArn),\n\t\t\tFamily: aws.String(fmt.Sprintf(\"%s_%s\", input.Type, input.Name)),\n\t\t\tMemory: aws.String(input.Memory),\n\t\t\tNetworkMode: aws.String(awsecs.NetworkModeAwsvpc),\n\t\t\tRequiresCompatibilities: aws.StringSlice([]string{awsecs.CompatibilityFargate}),\n\t\t\tTaskRoleArn: aws.String(input.TaskRole),\n\t\t\tTags: input.Tags,\n\t\t},\n\t)\n\n\tif err != nil {\n\t\tconsole.ErrorExit(err, \"Couldn't register ECS task definition\")\n\t}\n\n\ttd := resp.TaskDefinition\n\n\tconsole.Debug(\"Created ECS task definition [%s:%d]\", aws.StringValue(td.Family), aws.Int64Value(td.Revision))\n\n\treturn aws.StringValue(td.TaskDefinitionArn)\n}", "func (t *TaskService) CreateTask(path string, newTaskDef Definition, overwrite bool) (*RegisteredTask, bool, error) {\n\treturn t.CreateTaskEx(path, newTaskDef, \"\", \"\", newTaskDef.Principal.LogonType, overwrite)\n}", "func (db *PgDB) StartTaskSession(taskID string) (string, error) {\n\ttaskSession := &model.TaskSession{\n\t\tTaskID: taskID,\n\t}\n\n\tquery := \"INSERT INTO task_sessions (task_id) VALUES (:task_id) RETURNING id\"\n\tif err := db.namedGet(&taskSession.ID, query, *taskSession); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tv2 := paseto.NewV2()\n\ttoken, err := v2.Sign(db.tokenKeys.PrivateKey, taskSession, nil)\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"failed to generate task authentication token\")\n\t}\n\treturn token, nil\n}", "func ConfigForTask(taskInfo *mesos.TaskInfo) *docker.CreateContainerOptions {\n\tconfig := &docker.CreateContainerOptions{\n\t\tName: GetContainerName(taskInfo.TaskId),\n\t\tConfig: &docker.Config{\n\t\t\tEnv: EnvForTask(taskInfo),\n\t\t\tExposedPorts: PortsForTask(taskInfo),\n\t\t\tImage: *taskInfo.Container.Docker.Image,\n\t\t\tLabels: LabelsForTask(taskInfo),\n\t\t},\n\t\tHostConfig: &docker.HostConfig{\n\t\t\tBinds: BindsForTask(taskInfo),\n\t\t\tPortBindings: PortBindingsForTask(taskInfo),\n\t\t\tNetworkMode: NetworkForTask(taskInfo),\n\t\t\tCapAdd: CapAddForTask(taskInfo),\n\t\t\tCapDrop: CapDropForTask(taskInfo),\n\t\t},\n\t}\n\n\t// Check for and calculate CPU shares\n\tcpus := getResource(\"cpus\", taskInfo)\n\tif cpus != nil {\n\t\tconfig.Config.CPUShares = int64(*cpus.Scalar.Value * float64(1024))\n\t}\n\n\t// Check for and calculate memory limit\n\tmemory := getResource(\"memoryMb\", taskInfo)\n\tif memory != nil {\n\t\tconfig.Config.Memory = int64(*memory.Scalar.Value * float64(1024*1024))\n\t}\n\n\treturn config\n}", "func (cl *tektonClient) StartTaskRun(taskrun *v1alpha1.TaskRun) (string, error) {\n\tnewTaskRun, err := cl.pipelineClient.TaskRuns(cl.namespace).Create(taskrun)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\ttime.Sleep(5 * time.Second)\n\ti := 0\n\tfor i < BuildTimeout {\n\t\ttaskrun, err = cl.pipelineClient.TaskRuns(cl.namespace).Get(newTaskRun.Name, metav1.GetOptions{})\n\t\tif taskrun.Status.Conditions[0].Type == apis.ConditionSucceeded && taskrun.Status.Conditions[0].Status == \"True\" {\n\t\t\tfmt.Println(\"[INFO] Build task run\", taskrun.Name, \"is ready from\", taskrun.Status.StartTime, \"to\", taskrun.Status.CompletionTime)\n\t\t\treturn taskrun.Name, nil\n\t\t} else {\n\t\t\tfmt.Println(\"[INFO] Build task run\", taskrun.Name, \"is still\", taskrun.Status.Conditions[0].Reason, \", waiting\")\n\t\t\ttime.Sleep(5 * time.Second)\n\t\t}\n\t\ti += 5\n\t\ttime.Sleep(5 * time.Second)\n\t}\n\n\treturn taskrun.Name, fmt.Errorf(\"the build taskrun is not ready after timeout\")\n}", "func HandleStartTask(w http.ResponseWriter, r *http.Request) {\n\tlog.Root.Info(\"HandleStartTask BEGIN\")\n\n\tif r.Method != http.MethodPost {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tHttpResponseError(w, ErrNotFound)\n\t\treturn\n\t}\n\n\tdefer r.Body.Close()\n\tbody, _ := ioutil.ReadAll(r.Body)\n\n\tdata := make(map[string]interface{})\n\terr := json.Unmarshal(body, &data)\n\tif err != nil {\n\t\tlog.Root.Error(\"HandleStartTask Parse HTTP request body error\")\n\t\tHttpResponseError(w, ErrForm)\n\t\treturn\n\t}\n\n\telem, ok := data[\"taskID\"]\n\tif !ok {\n\t\tlog.Root.Error(\"HandleStartTask HTTP form data error\")\n\t\tHttpResponseError(w, ErrForm)\n\t\treturn\n\t}\n\n\ttaskID := elem.(string)\n\ttaskCapacity, err := node.StartTask(taskID)\n\tif err != nil {\n\t\tlog.Root.Error(\"HandleStartTask Start task error. TaskID: %v\", taskID)\n\t\tHttpResponseError(w, ErrServer)\n\t\treturn\n\t}\n\n\tif taskCapacity < 0 {\n\t\tlog.Root.Error(\"HandleStartTask Lack of computing resources\")\n\t\tHttpResponseError(w, ErrLackResources)\n\t\treturn\n\t}\n\n\tlog.Root.Info(\"HandleStartTask END\")\n\tHttpResponseData(w, H{\n\t\t\"taskCapacity\": taskCapacity,\n\t})\n\treturn\n}", "func (c *DockerContainerRuntime) StartContainer(cfg RunConfig, id string) error {\n\tctx, cancel := cfg.getContextWithTimeout()\n\tdefer cancel()\n\terr := c.Client.ContainerStart(ctx, id, types.ContainerStartOptions{})\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (t TaskService) CreateTask(ctx context.Context, tc taskmodel.TaskCreate) (*taskmodel.Task, error) {\n\tspan, _ := tracing.StartSpanFromContext(ctx)\n\tdefer span.Finish()\n\tvar tr taskResponse\n\n\terr := t.Client.\n\t\tPostJSON(tc, prefixTasks).\n\t\tDecodeJSON(&tr).\n\t\tDo(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn convertTask(tr.Task), nil\n}", "func (p SourceProvider) TaskStarted(t *provider.Task) error {\n\tt.Running = true\n\tif !p.Config.Enabled {\n\t\treturn nil\n\t}\n\tif p.Connection.KAPI == nil {\n\t\tif err := p.Connection.Connect(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tif err := p.Connection.WriteTask(t); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *Client) CreateTask(ctx context.Context, req *swarming_api.SwarmingRpcsNewTaskRequest) (*swarming_api.SwarmingRpcsTaskRequestMetadata, error) {\n\tvar resp *swarming_api.SwarmingRpcsTaskRequestMetadata\n\tcreateTask := func() error {\n\t\tvar err error\n\t\tresp, err = c.SwarmingService.Tasks.New(req).Context(ctx).Do()\n\t\treturn err\n\t}\n\n\tif err := callWithRetries(ctx, createTask); err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func (c *Client) StartImportFileTask(ctx context.Context, params *StartImportFileTaskInput, optFns ...func(*Options)) (*StartImportFileTaskOutput, error) {\n\tif params == nil {\n\t\tparams = &StartImportFileTaskInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"StartImportFileTask\", params, optFns, c.addOperationStartImportFileTaskMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*StartImportFileTaskOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func NewTask(service *Service, request *userdata.ProcessRequest, userpastelid string) *Task {\n\treturn &Task{\n\t\tTask: task.New(StatusTaskStarted),\n\t\tService: service,\n\t\trequest: request,\n\t\tuserpastelid: userpastelid,\n\t\tresultChan: make(chan *userdata.ProcessResult),\n\t\tresultChanGet: make(chan *userdata.ProcessRequest),\n\t}\n}", "func (executable *AWSECS) startECSContainer(messageBody *string, messageID *string) (string, error) {\n\te := &ECSMetadata{}\n\tm := &InstanceMetadata{}\n\tm.init()\n\te.init()\n\tvar ecsCluster *string\n\tvar containerInstanceID *string\n\n\tecsCluster = aws.String(e.Cluster)\n\tcontainerInstanceID = aws.String(e.ContainerInstanceArn)\n\n\t// Start ECS task on self\n\tsess, err := session.NewSession(&aws.Config{Region: aws.String(\"us-west-2\")})\n\tif err != nil {\n\t\tfmt.Println(\"failed to create session,\", err)\n\t\treturn \"\", err\n\t}\n\n\tsvc := ecs.New(sess)\n\n\tparams := &ecs.StartTaskInput{\n\t\tContainerInstances: []*string{\n\t\t\tcontainerInstanceID,\n\t\t},\n\t\tTaskDefinition: executable.ecsTaskDefinition,\n\t\tCluster: ecsCluster,\n\t\tOverrides: &ecs.TaskOverride{\n\t\t\tContainerOverrides: []*ecs.ContainerOverride{\n\t\t\t\t{\n\t\t\t\t\tEnvironment: []*ecs.KeyValuePair{\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tName: executable.overridePayloadKey,\n\t\t\t\t\t\t\tValue: aws.String(*messageBody),\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\tName: executable.overrideContainerName,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tStartedBy: aws.String(\"tasque\"),\n\t}\n\tresp, err := svc.StartTask(params)\n\n\tif err != nil {\n\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t// Message from an error.\n\t\tfmt.Println(\"Error:\", err.Error())\n\t\treturn \"\", err\n\t}\n\n\t// Pretty-print the response data.\n\tfmt.Println(resp)\n\tif len(resp.Failures) > 0 {\n\t\tvar err error\n\t\t// There were errors starting the container\n\t\treason := resp.Failures[0].Reason\n\t\tif strings.Contains(*reason, \"CPU\") {\n\t\t\texecutable.result.SetExit(\"CPU\")\n\t\t\terr = fmt.Errorf(\"%s %s The cpu requested by the task is unavailable on the given container instance. You may need to add container instances to your cluster\", *reason, *resp.Failures[0].Arn)\n\t\t} else if strings.Contains(*reason, \"MEMORY\") {\n\t\t\texecutable.result.SetExit(\"MEMORY\")\n\t\t\terr = fmt.Errorf(\"%s %s The memory requested by the task is unavailable on the given container instance. You may need to add container instances to your cluster\", *reason, *resp.Failures[0].Arn)\n\t\t} else if strings.Contains(*reason, \"RESOURCE\") {\n\t\t\texecutable.result.SetExit(\"RESOURCE\")\n\t\t\terr = fmt.Errorf(\"%s %s The resource or resources requested by the task are unavailable on the given container instance. If the resource is CPU or memory, you may need to add container instances to your cluster\", *reason, *resp.Failures[0].Arn)\n\t\t} else if strings.Contains(*reason, \"AGENT\") {\n\t\t\texecutable.result.SetExit(\"AGENT\")\n\t\t\terr = fmt.Errorf(\"%s %s The container instance that you attempted to launch a task onto has an agent which is currently disconnected. In order to prevent extended wait times for task placement, the request was rejected\", *reason, *resp.Failures[0].Arn)\n\t\t} else if strings.Contains(*reason, \"ATTRIBUTE\") {\n\t\t\texecutable.result.SetExit(\"ATTRIBUTE\")\n\t\t\terr = fmt.Errorf(\"%s %s Your task definition contains a parameter that requires a specific container instance attribute that is not available on your container instances. For more information on which attributes are required for specific task definition parameters and agent configuration variables, see Task Definition Parameters and Amazon ECS Container Agent Configuration\", *reason, *resp.Failures[0].Arn)\n\t\t} else {\n\t\t\t// Unrecognized error\n\t\t\texecutable.result.SetExit(\"UNKNOWN\")\n\t\t\terr = fmt.Errorf(\"Unrecognized error: '%s' %+v\", *reason, resp)\n\t\t}\n\t\treturn \"\", err\n\t}\n\ttaskArn := resp.Tasks[0].Containers[0].TaskArn\n\treturn *taskArn, nil\n}", "func (client *Client) StartRecordTask(request *StartRecordTaskRequest) (response *StartRecordTaskResponse, err error) {\n\tresponse = CreateStartRecordTaskResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func (c *Client) CreateTask(ctx context.Context, in *todopb.TaskCreateRequest, opts ...grpc.CallOption) (*todopb.TaskResponse, error) {\n\treturn c.client.CreateTask(ctx, in, opts...)\n}", "func (s *Storage) CreateTask(t creation.Task) error {\r\n\tnewTask := Task{\r\n\t\tID: uuid.New().String(),\r\n\t\tDescription: t.Description,\r\n\t\tCompleted: false,\r\n\t}\r\n\r\n\ts.tasks = append(s.tasks, &newTask)\r\n\r\n\treturn nil\r\n}", "func CreateTask(c *gin.Context) {\n\tsrv := server.GetServer()\n\n\tvar param TaskParams\n\tc.BindJSON(&param)\n\n\targs := make([]tasks.Arg, len(param.Args))\n\tfor idx, arg := range param.Args {\n\t\targs[idx] = tasks.Arg{\n\t\t\tType: \"int64\",\n\t\t\tValue: arg,\n\t\t}\n\t}\n\tsignature := &tasks.Signature{\n\t\tName: param.TaskName,\n\t\tArgs: args,\n\t\tRetryCount: 3,\n\t}\n\n\tasyncResult, err := srv.SendTask(signature)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tlog.Println(\"asyncResult:\", asyncResult)\n\n\tc.JSON(http.StatusOK, gin.H{\"Status\": \"In progress\", \"Job\": asyncResult})\n}", "func (nsr *NatsServiceRepository) Start(ctx context.Context, natsSvc NatsServiceIntf, params ...interface{}) *task.Task {\n\tnatsSvc.SetLogger(nsr.logger)\n\tnatsSvc.SetTransport(nsr.transport)\n\treturn nsr.RunnerRepository.Start(ctx, natsSvc, params...)\n}", "func NewTask(task string, context *Context, loadedTasks map[string]Task) {\n\tdir, t, tasks := FetchTask(task, context, loadedTasks)\n\n\t// register plugins ...\n\tplugin(t, context, tasks)\n\tevent.Trigger(\"dir\", &dir)\n\tevent.Trigger(\"t\", &t)\n\tevent.Trigger(\"tasks\", &tasks)\n\n\t// Skip the task, if we need to skip\n\tif t.Skip {\n\t\treturn\n\t}\n\n\t// innocent until proven guilty\n\tcontext.Ok = true\n\n\t// set our taskname\n\t_, context.TaskName = TaskParser(task, \"alfred:list\")\n\n\t// interactive mode?\n\tcontext.Interactive = t.Interactive\n\n\tif !context.hasBeenInited {\n\t\tcontext.hasBeenInited = true\n\t\tNewTask(MagicTaskURL(task)+\"__init\", context, tasks)\n\t}\n\n\tcomponents := []Component{\n\t\tComponent{\"log\", log},\n\t\tComponent{\"summary\", summary},\n\t\tComponent{\"prompt\", prompt},\n\t\tComponent{\"register\", register},\n\t\tComponent{\"defaults\", defaults},\n\t\tComponent{\"stdin\", stdin},\n\t\tComponent{\"config\", configC},\n\t\tComponent{\"env\", env},\n\t\tComponent{\"check\", check},\n\t\tComponent{\"watch\", watch},\n\t\tComponent{\"serve\", serve},\n\t\tComponent{\"setup\", setup},\n\t\tComponent{\"include\", include},\n\t\tComponent{\"multitask\", multitask},\n\t\tComponent{\"tasks\", tasksC},\n\t\tComponent{\"for\", forC},\n\t\tComponent{\"command\", commandC},\n\t\tComponent{\"commands\", commands},\n\t\tComponent{\"httptasks\", httptasks},\n\t\tComponent{\"result\", result},\n\t\tComponent{\"ok\", ok},\n\t\tComponent{\"fail\", fail},\n\t\tComponent{\"wait\", wait},\n\t\tComponent{\"every\", every},\n\t}\n\n\t// cycle through our components ...\n\tevent.Trigger(\"task.started\", t, context, tasks)\n\tfor _, component := range components {\n\t\tcontext.Component = component.Name\n\t\tevent.Trigger(\"before.\"+component.Name, context)\n\t\tcomponent.F(t, context, tasks)\n\t\tevent.Trigger(\"after.\"+component.Name, context)\n\n\t\tif context.Skip != \"\" {\n\t\t\toutOK(context.Skip, \"skipped\", context)\n\t\t\tevent.Trigger(\"task.skipped\", context)\n\t\t\treturn\n\t\t}\n\t}\n\tevent.Trigger(\"task.completed\", context)\n}", "func (*container) AttachTask(context.Context, libcontainerdtypes.StdioCallback) (libcontainerdtypes.Task, error) {\n\treturn nil, errdefs.NotFound(cerrdefs.ErrNotImplemented)\n}", "func NewTask(run actionFunc, stop actionFunc) func(task.Name, config.Resource) types.Task {\n\treturn func(name task.Name, res config.Resource) types.Task {\n\t\treturn &Task{\n\t\t\tname: name,\n\t\t\tconfig: res.(*config.ComposeConfig),\n\t\t\trun: run,\n\t\t\tstop: stop,\n\t\t}\n\t}\n}", "func NewInitTask(args []string) (Task, error) {\n\n\tflags := flag.NewFlagSet(\"init\", flag.ExitOnError)\n\n\tgoPath := flags.String(\"g\", \"\", \"the GOPATH to create\")\n\tprojectName := flags.String(\"n\", \"\", \"the project name\")\n\tscriptPath := flags.String(\"s\", \"./goenv/activate\", \"the full path to the initialization script\")\n\tprojectPath := flags.String(\"p\", \"\", \"the project path\")\n\n\tflags.Parse(args)\n\targs = flags.Args()\n\n\tif len(args) < 1 {\n\t\treturn nil, errors.New(\"no import path specified\")\n\t}\n\n\tworkingDir, err := os.Getwd()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttask := InitTask{\n\t\tImportPath: args[0],\n\t\tGoPath: *goPath,\n\t\tProjectName: *projectName,\n\t\tProjectPath: *projectPath,\n\t\tScriptPath: *scriptPath,\n\t}\n\n\tif task.ProjectName == \"\" {\n\t\ttask.ProjectName = filepath.Base(workingDir)\n\t}\n\n\tif task.ProjectPath == \"\" {\n\t\ttask.ProjectPath = workingDir\n\t}\n\n\tif task.GoPath == \"\" {\n\t\tusr, err := user.Current()\n\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\ttask.GoPath = filepath.Join(usr.HomeDir, \".goenv/\", task.ProjectName)\n\t}\n\n\treturn &task, nil\n}", "func (s *Worker) Start() error {\n\tclient, err := worker.InitRPCChannel(*s.cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\ts.rc = client\n\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tif client != nil {\n\t\t\t\t// we dont really care about the error here...\n\t\t\t\thostname, _ := os.Hostname()\n\t\t\t\terrStr := fmt.Sprintf(\"A panic occurred. Check logs on %s for more details\", hostname)\n\t\t\t\tclient.ChangeTaskStatus(rpc.ChangeTaskStatusRequest{\n\t\t\t\t\tTaskID: s.taskid,\n\t\t\t\t\tNewStatus: storage.TaskStatusError,\n\t\t\t\t\tError: &errStr,\n\t\t\t\t})\n\t\t\t}\n\t\t\tlog.Error().Str(\"task_id\", s.taskid).Msg(\"A critical error occurred while running task (panic)\")\n\t\t}\n\t}()\n\n\ts.t = NewTask(s.taskid, s.devices, s.cfg, s.rc) //Get the task in order to collect the task duration\n\tresp, err := s.t.c.GetTask(rpc.RequestTaskPayload{\n\t\tTaskID: s.t.taskid,\n\t})\n\n\tif resp.TaskDuration != 0 { //If the task duration is 0 (not set), we don't run the timer\n\t\ttimer := time.NewTimer(time.Second * time.Duration(resp.TaskDuration))\n\t\tgo func() {\n\t\t\t<-timer.C\n\t\t\tlog.Warn().Msg(\"Timer expired, stopping task\")\n\t\t\ts.t.Stop()\n\t\t\ttimer.Stop()\n\t\t}()\n\t}\n\n\tif err := s.t.Start(); err != nil {\n\t\tlog.Error().Err(err).Str(\"task_id\", s.taskid).Msg(\"An error occurred while processing a task\")\n\t\terrptr := err.Error()\n\t\tif rpcerr := client.ChangeTaskStatus(rpc.ChangeTaskStatusRequest{\n\t\t\tTaskID: s.taskid,\n\t\t\tNewStatus: storage.TaskStatusError,\n\t\t\tError: &errptr,\n\t\t}); rpcerr != nil {\n\t\t\tlog.Error().Err(rpcerr).Msg(\"Failed to change tasks status to error\")\n\t\t}\n\t}\n\treturn nil\n}", "func TrackTaskStart(payload string, duration int64) {\n\tpersist.SetValue(taskPayloadKey, payload)\n\tpersist.SetValue(taskEndTimeKey, time.Now().Unix()+duration)\n}", "func (s *K8sSvc) RunTask(ctx context.Context, opts *containersvc.RunTaskOptions) (taskID string, err error) {\n\trequuid := utils.GetReqIDFromContext(ctx)\n\n\ttaskID = opts.Common.ServiceName + common.NameSeparator + opts.TaskType\n\n\tlabels := make(map[string]string)\n\tlabels[serviceNameLabel] = opts.Common.ServiceName\n\tlabels[serviceUUIDLabel] = opts.Common.ServiceUUID\n\n\tenvs := make([]corev1.EnvVar, len(opts.Envkvs))\n\tfor i, e := range opts.Envkvs {\n\t\tenvs[i] = corev1.EnvVar{\n\t\t\tName: e.Name,\n\t\t\tValue: e.Value,\n\t\t}\n\t}\n\n\tjob := &batchv1.Job{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: taskID,\n\t\t\tNamespace: s.namespace,\n\t\t\tLabels: labels,\n\t\t},\n\t\tSpec: batchv1.JobSpec{\n\t\t\tParallelism: utils.Int32Ptr(1),\n\t\t\tCompletions: utils.Int32Ptr(1),\n\t\t\t// allow restarting the job twice before mark the job failed.\n\t\t\tBackoffLimit: utils.Int32Ptr(2),\n\t\t\tTemplate: corev1.PodTemplateSpec{\n\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\tName: taskID,\n\t\t\t\t\tNamespace: s.namespace,\n\t\t\t\t\tLabels: labels,\n\t\t\t\t},\n\t\t\t\tSpec: corev1.PodSpec{\n\t\t\t\t\tContainers: []corev1.Container{\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tName: taskID,\n\t\t\t\t\t\t\tImage: opts.Common.ContainerImage,\n\t\t\t\t\t\t\tEnv: envs,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\tRestartPolicy: corev1.RestartPolicyNever,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\t_, err = s.cliset.BatchV1().Jobs(s.namespace).Create(job)\n\tif err != nil {\n\t\tif k8errors.IsAlreadyExists(err) {\n\t\t\tglog.Infoln(\"service task exist\", taskID, \"requuid\", requuid)\n\t\t\treturn taskID, nil\n\t\t}\n\t\tglog.Errorln(\"create service task error\", taskID, \"requuid\", requuid)\n\t\treturn \"\", err\n\t}\n\n\tglog.Infoln(\"created service task\", taskID, \"requuid\", requuid)\n\treturn taskID, nil\n}", "func (c *Client) CreateTask(tr TaskRequest) (task Task, err error) {\n\tbodyReader, err := createReader(tr)\n\tif err != nil {\n\t\treturn task, err\n\t}\n\n\trequest := fmt.Sprintf(\"/v3/apps/%s/tasks\", tr.DropletGUID)\n\treq := c.NewRequestWithBody(\"POST\", request, bodyReader)\n\n\tresp, err := c.DoRequest(req)\n\tif err != nil {\n\t\treturn task, errors.Wrap(err, \"Error creating task\")\n\t}\n\tdefer resp.Body.Close()\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn task, errors.Wrap(err, \"Error reading task after creation\")\n\t}\n\n\terr = json.Unmarshal(body, &task)\n\tif err != nil {\n\t\treturn task, errors.Wrap(err, \"Error unmarshaling task\")\n\t}\n\treturn task, err\n}", "func NewTask(TskName string, Capacity WLSpec, Type string, SlaveOf string) *Task {\n\n\tvar tmp Task\n\tiname, id := TaskSplitNames(TskName)\n\ttmp.IName = iname\n\ttmp.ID = id\n\ttmp.Name = TskName\n\ttmp.Stats.Capacity.Copy(Capacity)\n\n\ttmp.Nodename = etcd.ETCD_INSTDIR + \"/\" + tmp.IName + \"/Procs/\" + tmp.ID\n\treturn &tmp\n}", "func (atr *ActiveTaskRun) Start() (int, time.Duration, error) {\n var env []string\n var err error\n\n for _, param := range atr.Task.Params {\n env = append(env, fmt.Sprintf(\"%s=%s\", param.Name, param.Value))\n }\n\n env = append(env, fmt.Sprintf(\"WAYFINDER_TOTAL_CORES=%d\", len(atr.CoreIds)))\n env = append(env, fmt.Sprintf(\"WAYFINDER_CORES=%s\", strings.Trim(\n strings.Join(strings.Fields(fmt.Sprint(atr.CoreIds)), \" \"), \"[]\",\n )))\n for i, coreId := range atr.CoreIds {\n env = append(env, fmt.Sprintf(\"WAYFINDER_CORE_ID%d=%d\", i, coreId))\n }\n\n config := &run.RunnerConfig{\n Log: atr.log,\n CacheDir: atr.Task.cacheDir,\n ResultsDir: atr.Task.resultsDir,\n AllowOverride: atr.Task.AllowOverride,\n Name: atr.run.Name,\n Image: atr.run.Image,\n CoreIds: atr.CoreIds,\n Devices: atr.run.Devices,\n Inputs: atr.Task.Inputs,\n Outputs: atr.Task.Outputs,\n Env: env,\n Capabilities: atr.run.Capabilities,\n }\n if atr.run.Path != \"\" {\n config.Path = atr.run.Path\n } else if atr.run.Cmd != \"\" {\n config.Cmd = atr.run.Cmd\n } else {\n return 1, -1, fmt.Errorf(\"Run did not specify path or cmd: %s\", atr.run.Name)\n }\n\n atr.Runner, err = run.NewRunner(config, atr.bridge, atr.dryRun)\n if err != nil {\n return 1, -1, err\n }\n\n atr.log.Infof(\"Starting run...\")\n exitCode, timeElapsed, err := atr.Runner.Run()\n atr.Runner.Destroy()\n if err != nil {\n return 1, -1, fmt.Errorf(\"Could not start runner: %s\", err)\n }\n\n return exitCode, timeElapsed, nil\n}", "func (k *KubernetesExecutor) LaunchTask(driver bindings.ExecutorDriver, taskInfo *mesos.TaskInfo) {\n\tif k.isDone() {\n\t\treturn\n\t}\n\tlog.Infof(\"Launch task %v\\n\", taskInfo)\n\n\tif !k.isConnected() {\n\t\tlog.Errorf(\"Ignore launch task because the executor is disconnected\\n\")\n\t\tk.sendStatus(driver, newStatus(taskInfo.GetTaskId(), mesos.TaskState_TASK_FAILED,\n\t\t\tmessages.ExecutorUnregistered))\n\t\treturn\n\t}\n\n\tobj, err := api.Codec.Decode(taskInfo.GetData())\n\tif err != nil {\n\t\tlog.Errorf(\"failed to extract yaml data from the taskInfo.data %v\", err)\n\t\tk.sendStatus(driver, newStatus(taskInfo.GetTaskId(), mesos.TaskState_TASK_FAILED,\n\t\t\tmessages.UnmarshalTaskDataFailure))\n\t\treturn\n\t}\n\tpod, ok := obj.(*api.Pod)\n\tif !ok {\n\t\tlog.Errorf(\"expected *api.Pod instead of %T: %+v\", pod, pod)\n\t\tk.sendStatus(driver, newStatus(taskInfo.GetTaskId(), mesos.TaskState_TASK_FAILED,\n\t\t\tmessages.UnmarshalTaskDataFailure))\n\t\treturn\n\t}\n\n\tk.lock.Lock()\n\tdefer k.lock.Unlock()\n\n\ttaskId := taskInfo.GetTaskId().GetValue()\n\tif _, found := k.tasks[taskId]; found {\n\t\tlog.Errorf(\"task already launched\\n\")\n\t\t// Not to send back TASK_RUNNING here, because\n\t\t// may be duplicated messages or duplicated task id.\n\t\treturn\n\t}\n\t// remember this task so that:\n\t// (a) we ignore future launches for it\n\t// (b) we have a record of it so that we can kill it if needed\n\t// (c) we're leaving podName == \"\" for now, indicates we don't need to delete containers\n\tk.tasks[taskId] = &kuberTask{\n\t\tmesosTaskInfo: taskInfo,\n\t}\n\tk.resetSuicideWatch(driver)\n\n\tgo k.launchTask(driver, taskId, pod)\n}", "func (e *ECS) RunTask(req *RunTaskReq) (*RunTaskResp, error) {\n\tif req == nil {\n\t\treturn nil, fmt.Errorf(\"The req params cannot be nil\")\n\t}\n\n\tparams := makeParams(\"RunTask\")\n\tif req.Count > 0 {\n\t\tparams[\"count\"] = strconv.Itoa(int(req.Count))\n\t}\n\tif req.Cluster != \"\" {\n\t\tparams[\"cluster\"] = req.Cluster\n\t}\n\tif req.TaskDefinition != \"\" {\n\t\tparams[\"taskDefinition\"] = req.TaskDefinition\n\t}\n\n\tfor i, co := range req.Overrides.ContainerOverrides {\n\t\tkey := fmt.Sprintf(\"overrides.containerOverrides.member.%d\", i+1)\n\t\tparams[fmt.Sprintf(\"%s.name\", key)] = co.Name\n\t\tfor k, cmd := range co.Command {\n\t\t\tparams[fmt.Sprintf(\"%s.command.member.%d\", key, k+1)] = cmd\n\t\t}\n\t\tfor k, env := range co.Environment {\n\t\t\tparams[fmt.Sprintf(\"%s.environment.member.%d.name\", key, k+1)] = env.Name\n\t\t\tparams[fmt.Sprintf(\"%s.environment.member.%d.value\", key, k+1)] = env.Value\n\t\t}\n\t}\n\n\tresp := new(RunTaskResp)\n\tif err := e.query(params, resp); err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func initTask() {\n\tcwd, err := os.Getwd()\n\tif err != nil {\n\t\tcwd = os.Getenv(\"PWD\")\n\t}\n\n\tcfgPath := filepath.Join(filepath.Clean(cwd), \"bonclay.conf.yaml\")\n\n\t_, err = os.Lstat(cfgPath)\n\tif err == nil {\n\t\tfmt.Fprintf(os.Stderr, \"bonclay: error: config file already exists: %s\\n\", cfgPath)\n\t\tos.Exit(1)\n\t}\n\n\tf, err := os.Create(cfgPath)\n\tif err != nil {\n\t\tfmt.Fprintf(os.Stderr, \"bonclay: error: could not create config file: %s\\n\", cfgPath)\n\t\tos.Exit(1)\n\t}\n\tdefer f.Close()\n\n\t_, err = f.WriteString(strings.TrimSpace(exampleConfig))\n\tif err != nil {\n\t\tfmt.Fprintf(os.Stderr, \"bonclay: error: could not create config file: %s\\n\", cfgPath)\n\t\tos.Exit(1)\n\t}\n\tf.Sync()\n\n\tfmt.Printf(\"bonclay: config file created: %s\\n\", cfgPath)\n}", "func startContainer(config *startConfig) (int, error) {\n\tconn, err := runvRequest(config.Root, config.Name, RUNV_STARTCONTAINER, config)\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\n\treturn containerTtySplice(config.Root, config.Name, conn, true)\n}", "func (k *KubernetesExecutor) LaunchTask(driver bindings.ExecutorDriver, taskInfo *mesos.TaskInfo) {\n\tif k.isDone() {\n\t\treturn\n\t}\n\tlog.Infof(\"Launch task %v\\n\", taskInfo)\n\n\tif !k.isConnected() {\n\t\tlog.Warningf(\"Ignore launch task because the executor is disconnected\\n\")\n\t\tk.sendStatus(driver, newStatus(taskInfo.GetTaskId(), mesos.TaskState_TASK_FAILED,\n\t\t\tmessages.ExecutorUnregistered))\n\t\treturn\n\t}\n\n\tvar pod api.BoundPod\n\tif err := yaml.Unmarshal(taskInfo.GetData(), &pod); err != nil {\n\t\tlog.Warningf(\"Failed to extract yaml data from the taskInfo.data %v\\n\", err)\n\t\tk.sendStatus(driver, newStatus(taskInfo.GetTaskId(), mesos.TaskState_TASK_FAILED,\n\t\t\tmessages.UnmarshalTaskDataFailure))\n\t\treturn\n\t}\n\n\tk.lock.Lock()\n\tdefer k.lock.Unlock()\n\n\ttaskId := taskInfo.GetTaskId().GetValue()\n\tif _, found := k.tasks[taskId]; found {\n\t\tlog.Warningf(\"task already launched\\n\")\n\t\t// Not to send back TASK_RUNNING here, because\n\t\t// may be duplicated messages or duplicated task id.\n\t\treturn\n\t}\n\t// remember this task so that:\n\t// (a) we ignore future launches for it\n\t// (b) we have a record of it so that we can kill it if needed\n\t// (c) we're leaving podName == \"\" for now, indicates we don't need to delete containers\n\tk.tasks[taskId] = &kuberTask{\n\t\tmesosTaskInfo: taskInfo,\n\t}\n\tgo k.launchTask(driver, taskId, &pod)\n}", "func NewTask(cluster *api.Cluster, service *api.Service, slot uint64, nodeID string) *api.Task {\n\tvar logDriver *api.Driver\n\tif service.Spec.Task.LogDriver != nil {\n\t\t// use the log driver specific to the task, if we have it.\n\t\tlogDriver = service.Spec.Task.LogDriver\n\t} else if cluster != nil {\n\t\t// pick up the cluster default, if available.\n\t\tlogDriver = cluster.Spec.TaskDefaults.LogDriver // nil is okay here.\n\t}\n\n\ttaskID := identity.NewID()\n\ttask := api.Task{\n\t\tID: taskID,\n\t\tServiceAnnotations: service.Spec.Annotations,\n\t\tSpec: service.Spec.Task,\n\t\tServiceID: service.ID,\n\t\tSlot: slot,\n\t\tStatus: api.TaskStatus{\n\t\t\tState: api.TaskStateNew,\n\t\t\tTimestamp: ptypes.MustTimestampProto(time.Now()),\n\t\t\tMessage: \"created\",\n\t\t},\n\t\tEndpoint: &api.Endpoint{\n\t\t\tSpec: service.Spec.Endpoint.Copy(),\n\t\t},\n\t\tDesiredState: api.TaskStateRunning,\n\t\tLogDriver: logDriver,\n\t}\n\n\t// In global mode we also set the NodeID\n\tif nodeID != \"\" {\n\t\ttask.NodeID = nodeID\n\t}\n\n\treturn &task\n}", "func (p *DockerPod) Start() error {\n\tp.status = container.PodStatus_STARTING\n\tp.message = \"Pod is starting\"\n\t//add pod ipaddr to ENV\n\tenvHost := container.BcsKV{\n\t\tKey: \"BCS_CONTAINER_IP\",\n\t\tValue: util.GetIPAddress(),\n\t}\n\n\tlogs.Infof(\"docker pod start container...\")\n\n\tfor name, task := range p.conTasks {\n\t\t//create container attach to network infrastructure\n\t\ttask.NetworkName = \"container:\" + p.GetContainerID()\n\t\ttask.RuntimeConf = &container.BcsContainerInfo{\n\t\t\tName: name,\n\t\t}\n\t\ttask.Env = append(task.Env, envHost)\n\t\t//assignment for environments\n\t\tcontainer.EnvOperCopy(task)\n\t\tvar extendedErr error\n\t\t//if task contains extended resources, need connect device plugin to allocate resources\n\t\tfor _, ex := range task.ExtendedResources {\n\t\t\tlogs.Infof(\"task %s contains extended resource %s, then allocate it\", task.TaskId, ex.Name)\n\t\t\tenvs, err := p.resourceManager.ApplyExtendedResources(ex, p.netTask.TaskId)\n\t\t\tif err != nil {\n\t\t\t\tlogs.Errorf(\"apply extended resource failed, err %s\", err.Error())\n\t\t\t\textendedErr = err\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tlogs.Infof(\"add env %v for task %s\", envs, task.TaskId)\n\n\t\t\t//append response docker envs to task.envs\n\t\t\tfor k, v := range envs {\n\t\t\t\tkv := container.BcsKV{\n\t\t\t\t\tKey: k,\n\t\t\t\t\tValue: v,\n\t\t\t\t}\n\t\t\t\ttask.Env = append(task.Env, kv)\n\t\t\t}\n\t\t}\n\n\t\t//if allocate extended resource failed, then return and exit\n\t\tif extendedErr != nil {\n\t\t\tlogs.Errorf(extendedErr.Error())\n\t\t\ttask.RuntimeConf.Status = container.ContainerStatus_EXITED\n\t\t\ttask.RuntimeConf.Message = extendedErr.Error()\n\t\t\tp.startFailedStop(extendedErr)\n\t\t\treturn extendedErr\n\t\t}\n\t\tcreatedInst, createErr := p.conClient.CreateContainer(name, task)\n\t\tif createErr != nil {\n\t\t\tlogs.Errorf(\"DockerPod create %s with name %s failed, err: %s\\n\", task.Image, name, createErr.Error())\n\t\t\ttask.RuntimeConf.Status = container.ContainerStatus_EXITED\n\t\t\ttask.RuntimeConf.Message = createErr.Error()\n\t\t\tp.startFailedStop(createErr)\n\t\t\treturn createErr\n\t\t}\n\t\ttask.RuntimeConf.ID = createdInst.ID\n\t\ttask.RuntimeConf.NodeAddress = util.GetIPAddress()\n\t\ttask.RuntimeConf.IPAddress = p.cnmIPAddr\n\t\ttask.RuntimeConf.Status = container.ContainerStatus_CREATED\n\t\ttask.RuntimeConf.Message = \"container created\"\n\t\ttask.RuntimeConf.Resource = task.Resource\n\n\t\tlogs.Infof(\"task %s cpu %f mem %f\", task.TaskId, task.RuntimeConf.Resource.Cpus, task.RuntimeConf.Resource.Mem)\n\n\t\t//crate success, event callback before start\n\t\tif p.events != nil && p.events.PreStart != nil {\n\t\t\tpreErr := p.events.PreStart(task)\n\t\t\tif preErr != nil {\n\t\t\t\tlogs.Errorf(\"DockerPod PreStart setting container %s err: %s\\n\", task.RuntimeConf.ID, preErr.Error())\n\t\t\t\tp.conClient.RemoveContainer(task.RuntimeConf.ID, true)\n\t\t\t\ttask.RuntimeConf.Status = container.ContainerStatus_EXITED\n\t\t\t\ttask.RuntimeConf.Message = preErr.Error()\n\t\t\t\tp.startFailedStop(createErr)\n\t\t\t\treturn preErr\n\t\t\t}\n\t\t}\n\t\t//ready to starting\n\t\tif err := p.conClient.StartContainer(task.RuntimeConf.Name); err != nil {\n\t\t\tlogs.Errorf(\"DockerPod Start %s with name %s failed, err: %s\\n\", task.Image, name, err.Error())\n\t\t\ttask.RuntimeConf.Status = container.ContainerStatus_EXITED\n\t\t\ttask.RuntimeConf.Message = err.Error()\n\t\t\tp.conClient.RemoveContainer(task.RuntimeConf.Name, true)\n\t\t\tp.startFailedStop(err)\n\t\t\treturn err\n\t\t}\n\t\ttask.RuntimeConf.Message = \"container is starting\"\n\t\tif p.events != nil && p.events.PostStart != nil {\n\t\t\tp.events.PostStart(task)\n\t\t}\n\t\tp.runningContainer[task.RuntimeConf.Name] = task.RuntimeConf\n\t\tlogs.Infof(\"Pod add container %s in running container.\\n\", task.RuntimeConf.Name)\n\t}\n\tp.conTasks[p.netTask.Name] = p.netTask\n\t//all container starting, start containerWatch\n\twatchCxt, _ := context.WithCancel(p.podCxt)\n\tgo p.containersWatch(watchCxt)\n\treturn nil\n}", "func (e *ContainerRunner) Start(ctx context.Context) error {\n\tvar err error\n\te.client, err = client.NewEnvClient()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"creating env client: %w\", err)\n\t}\n\n\tlog.Infoln(\"pulling image\")\n\t_, err = e.client.ImagePull(ctx, e.image, types.ImagePullOptions{})\n\tif err != nil {\n\t\treturn fmt.Errorf(\"pulling image: %w\", err)\n\t}\n\n\tlog.Infoln(\"creating container\")\n\tresp, err := e.client.ContainerCreate(ctx, &container.Config{\n\t\tImage: e.image,\n\t\tExposedPorts: e.exposedPorts,\n\t\tEnv: e.env,\n\t}, &container.HostConfig{\n\t\tPortBindings: e.portBindings,\n\t}, nil, e.name)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"creating container: %w\", err)\n\t}\n\n\t// Save the container id\n\te.id = resp.ID\n\n\tlog.Infoln(\"starting container\")\n\tif err := e.client.ContainerStart(ctx, e.id, types.ContainerStartOptions{}); err != nil {\n\t\treturn fmt.Errorf(\"starting container: %w\", err)\n\t}\n\tlog.Infoln(\"container started\")\n\treturn nil\n}", "func Start(config *config.Config) error {\n\tCamsPoolInstance = &CamsPool{config: config}\n\tCamsPoolInstance.boot()\n\n\tcc := &CommandCenter{Port: config.Port}\n\tstartError := cc.Start()\n\n\tif startError != nil {\n\t\treturn startError\n\t}\n\n\treturn nil\n}", "func (c *jxTasks) Create(jxTask *v1alpha1.JxTask) (result *v1alpha1.JxTask, err error) {\n\tresult = &v1alpha1.JxTask{}\n\terr = c.client.Post().\n\t\tNamespace(c.ns).\n\t\tResource(\"jxtasks\").\n\t\tBody(jxTask).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func CreateTask(w http.ResponseWriter, r *http.Request, _ httprouter.Params) {\n\tt := Task{}\n\tdecoder := json.NewDecoder(r.Body)\n\terr := decoder.Decode(&t)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\terr = store.CreateTask(&t)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tw.WriteHeader(http.StatusCreated)\n}", "func newTask(ctx context.Context, cfg *taskConfig, f TaskFunc, args ...interface{}) *Task {\n\tctx, cancelCtx := context.WithCancel(ctx)\n\n\ttask := &Task{\n\t\tcfg: cfg,\n\n\t\tctx: ctx,\n\t\tcancelCtx: cancelCtx,\n\n\t\tf: f,\n\t\targs: args,\n\n\t\tstartedChan: make(chan struct{}),\n\t\trunningChan: make(chan struct{}),\n\t\tfinishedChan: make(chan struct{}),\n\n\t\tresultChan: make(chan TaskResult),\n\t}\n\n\treturn task\n}", "func NewTask(t *TestCase) *Task {\n\treturn &Task{\n\t\tName: t.Name,\n\t\tMethod: t.Method,\n\t\tBaseURL: t.BaseURL,\n\t\tConfig: &Config{\n\t\t\tInsecure: false,\n\t\t\tDetail: false,\n\t\t},\n\t\tPath: t.Path,\n\t\tContentType: t.ContentType,\n\t\tRequestBody: t.RequestBody,\n\t\tHeaders: t.Headers,\n\t\tCaptures: t.Captures,\n\t\tVariables: map[string]string{},\n\t\tCaptured: map[string]string{},\n\t\tCookies: []*http.Cookie{},\n\t\tUploadList: t.UploadList,\n\t}\n}", "func (p k8sPodResources) Start(\n\tctx *actor.System, logCtx logger.Context, spec tasks.TaskSpec, rri sproto.ResourcesRuntimeInfo,\n) error {\n\tp.setPosition(&spec)\n\tspec.ContainerID = string(p.containerID)\n\tspec.ResourcesID = string(p.containerID)\n\tspec.AllocationID = string(p.req.AllocationID)\n\tspec.AllocationSessionToken = rri.Token\n\tspec.TaskID = string(p.req.TaskID)\n\tspec.UseHostMode = rri.IsMultiAgent\n\tspec.ResourcesConfig.SetPriority(p.group.Priority)\n\tif spec.LoggingFields == nil {\n\t\tspec.LoggingFields = map[string]string{}\n\t}\n\tspec.LoggingFields[\"allocation_id\"] = spec.AllocationID\n\tspec.LoggingFields[\"task_id\"] = spec.TaskID\n\tspec.ExtraEnvVars[sproto.ResourcesTypeEnvVar] = string(sproto.ResourcesTypeK8sPod)\n\tspec.ExtraEnvVars[resourcePoolEnvVar] = p.req.ResourcePool\n\treturn ctx.Ask(p.podsActor, StartTaskPod{\n\t\tAllocationID: p.req.AllocationID,\n\t\tSpec: spec,\n\t\tSlots: p.slots,\n\t\tRank: rri.AgentRank,\n\t\tNamespace: p.namespace,\n\t\tLogContext: logCtx,\n\t}).Error()\n}", "func (f *dockerBuildFixture) startContainer(ctx context.Context, config *container.Config) wmcontainer.ID {\n\tresp, err := f.dCli.ContainerCreate(ctx, config, nil, nil, nil, \"\")\n\tif err != nil {\n\t\tf.t.Fatalf(\"startContainer: %v\", err)\n\t}\n\tcID := resp.ID\n\n\terr = f.dCli.ContainerStart(ctx, cID, types.ContainerStartOptions{})\n\tif err != nil {\n\t\tf.t.Fatalf(\"startContainer: %v\", err)\n\t}\n\n\tresult := wmcontainer.ID(cID)\n\tf.containerIDs = append(f.containerIDs, result)\n\treturn result\n}", "func (client *Client) CreateTagTask(request *CreateTagTaskRequest) (response *CreateTagTaskResponse, err error) {\n\tresponse = CreateCreateTagTaskResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func (c *Client) CreateBkOpsTask(url string, paras *CreateTaskPathParas,\n\trequest *CreateTaskRequest) (*CreateTaskResponse, error) {\n\tif c == nil {\n\t\treturn nil, ErrServerNotInit\n\t}\n\n\tvar (\n\t\treqURL = fmt.Sprintf(\"/create_task/%s/%s/\", paras.TemplateID, paras.BkBizID)\n\t\trespData = &CreateTaskResponse{}\n\t)\n\n\tuserAuth, err := c.generateGateWayAuth(paras.Operator)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"bksops CreateBkOpsTask generateGateWayAuth failed: %v\", err)\n\t}\n\trequest.FlowType = string(CommonFlow)\n\t// TemplateSource 模版来源, 默认是业务流程; 可由用户自定义\n\tif request.TemplateSource == \"\" {\n\t\trequest.TemplateSource = string(BusinessTpl)\n\t}\n\n\t_, _, errs := gorequest.New().\n\t\tTimeout(defaultTimeOut).\n\t\tPost(c.server+reqURL).\n\t\tSet(\"Content-Type\", \"application/json\").\n\t\tSet(\"Accept\", \"application/json\").\n\t\tSet(\"X-Bkapi-Authorization\", userAuth).\n\t\tSetDebug(c.serverDebug).\n\t\tSend(request).\n\t\tEndStruct(&respData)\n\tif len(errs) > 0 {\n\t\tblog.Errorf(\"call api CreateBkOpsTask failed: %v\", errs[0])\n\t\treturn nil, errs[0]\n\t}\n\n\tif !respData.Result {\n\t\tblog.Errorf(\"call api CreateBkOpsTask failed: %v\", respData.Message)\n\t\treturn nil, fmt.Errorf(respData.Message)\n\t}\n\t//successfully request\n\tblog.Infof(\"call api CreateBkOpsTask with url(%s) successfully\", reqURL)\n\treturn respData, nil\n}", "func (n *BaseNode) Start(ctx context.Context, tc CreateTaskFunc, execFunc BaseNodeExecFunc) {\n\t// Make sure the node can only be started once\n\tn.oStart.Do(func() {\n\t\t// Check context\n\t\tif ctx.Err() != nil {\n\t\t\treturn\n\t\t}\n\n\t\t// Create task\n\t\tt := tc()\n\n\t\t// Reset context\n\t\tn.ctx, n.cancel = context.WithCancel(ctx)\n\n\t\t// Reset once\n\t\tn.oStop = &sync.Once{}\n\n\t\t// Loop through children\n\t\tfor _, c := range n.Children() {\n\t\t\tc.ParentIsStarted(n.o.Metadata)\n\t\t}\n\n\t\t// Loop through parents\n\t\tfor _, p := range n.Parents() {\n\t\t\tp.ChildIsStarted(n.o.Metadata)\n\t\t}\n\n\t\t// Update status\n\t\tn.m.Lock()\n\t\tn.status = StatusRunning\n\t\tn.m.Unlock()\n\n\t\t// Send started event\n\t\tn.eh.Emit(Event{\n\t\t\tName: n.et(EventTypeStarted),\n\t\t\tTarget: n.target,\n\t\t})\n\n\t\t// Execute the rest in a goroutine\n\t\tgo func() {\n\t\t\t// Task is done\n\t\t\tdefer t.Done()\n\n\t\t\t// Send stopped event\n\t\t\tdefer n.eh.Emit(Event{\n\t\t\t\tName: n.et(EventTypeStopped),\n\t\t\t\tTarget: n.target,\n\t\t\t})\n\n\t\t\t// Make sure the status is updated once everything is done\n\t\t\tdefer func() {\n\t\t\t\tn.m.Lock()\n\t\t\t\tdefer n.m.Unlock()\n\t\t\t\tn.status = StatusStopped\n\t\t\t}()\n\n\t\t\t// Let children and parents know the node is stopped\n\t\t\tdefer func() {\n\t\t\t\t// Loop through children\n\t\t\t\tfor _, c := range n.Children() {\n\t\t\t\t\tc.ParentIsStopped(n.o.Metadata)\n\t\t\t\t}\n\n\t\t\t\t// Loop through parents\n\t\t\t\tfor _, p := range n.Parents() {\n\t\t\t\t\tp.ChildIsStopped(n.o.Metadata)\n\t\t\t\t}\n\t\t\t}()\n\n\t\t\t// Make sure the node is stopped properly\n\t\t\tdefer n.Stop()\n\n\t\t\t// Handle stats\n\t\t\tif n.s != nil {\n\t\t\t\t// Make sure to delete stats\n\t\t\t\tdefer func() {\n\t\t\t\t\t// Lock\n\t\t\t\t\tn.m.Lock()\n\t\t\t\t\tdefer n.m.Unlock()\n\n\t\t\t\t\t// Delete stats\n\t\t\t\t\tn.s.DelStats(n.target, n.ss...)\n\t\t\t\t}()\n\n\t\t\t\t// Add stats\n\t\t\t\tn.m.Lock()\n\t\t\t\tn.s.AddStats(n.target, n.ss...)\n\t\t\t\tn.m.Unlock()\n\t\t\t}\n\n\t\t\t// Exec func\n\t\t\texecFunc(t)\n\t\t}()\n\t})\n}", "func (cl *tektonClient) ConstructTaskRun(name, builder, gitResourceName, imageResourceName, serviceAccount, namespace string) *pipelinev1alpha1.TaskRun {\n\n\tbuilderImage := \"\"\n\tif builder == BuildpacksBuilderName {\n\t\tbuilderImage = BuildpacksBuilderImage\n\t} else if builder == KanikoBuilderName {\n\t\tbuilderImage = KanikoBuilderImage\n\t}\n\n\treturn &pipelinev1alpha1.TaskRun{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tNamespace: namespace,\n\t\t\tGenerateName: name + \"-build-\",\n\t\t},\n\t\tSpec: pipelinev1alpha1.TaskRunSpec{\n\t\t\tTaskRef: &pipelinev1alpha1.TaskRef{\n\t\t\t\tName: builder,\n\t\t\t},\n\t\t\tInputs: pipelinev1alpha1.TaskRunInputs{\n\t\t\t\tResources: []pipelinev1alpha1.TaskResourceBinding{\n\t\t\t\t\t{\n\t\t\t\t\t\tPipelineResourceBinding: pipelinev1alpha1.PipelineResourceBinding{\n\t\t\t\t\t\t\tResourceRef: &pipelinev1alpha1.PipelineResourceRef{\n\t\t\t\t\t\t\t\tName: gitResourceName,\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tName: \"source\",\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tParams: []pipelinev1alpha1.Param{\n\t\t\t\t\t{\n\t\t\t\t\t\tName: \"BUILDER_IMAGE\",\n\t\t\t\t\t\tValue: *ArrayOrString(builderImage),\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\tOutputs: pipelinev1alpha1.TaskRunOutputs{\n\t\t\t\tResources: []pipelinev1alpha1.TaskResourceBinding{\n\t\t\t\t\t{\n\t\t\t\t\t\tPipelineResourceBinding: pipelinev1alpha1.PipelineResourceBinding{\n\t\t\t\t\t\t\tResourceRef: &pipelinev1alpha1.PipelineResourceRef{\n\t\t\t\t\t\t\t\tName: imageResourceName,\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tName: \"image\",\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\tServiceAccountName: serviceAccount,\n\t\t},\n\t}\n\treturn nil\n}", "func NewDeployMasterTask(taskName string, taskConfig *DeployMasterTaskConfig) (Task, error) {\n\tvar err error\n\tif taskConfig == nil {\n\t\terr = fmt.Errorf(\"invalid task config: nil\")\n\t} else if len(taskConfig.Nodes) == 0 {\n\t\terr = fmt.Errorf(\"invalid task config: nodes is empty\")\n\t} else if len(taskConfig.EtcdNodes) == 0 {\n\t\terr = fmt.Errorf(\"invalid task config: etcd nodes is empty\")\n\t} else if taskConfig.ClusterConfig == nil {\n\t\terr = fmt.Errorf(\"nil cluster config\")\n\t}\n\n\tif err != nil {\n\t\tlogrus.Error(err)\n\t\treturn nil, err\n\t}\n\n\ttask := &deployMasterTask{\n\t\tBase: Base{\n\t\t\tName: taskName,\n\t\t\tTaskType: TaskTypeDeployMaster,\n\t\t\tStatus: TaskPending,\n\t\t\tLogFileDir: GenTaskLogFileDir(taskConfig.LogFileBasePath, taskName),\n\t\t\tCreationTimestamp: time.Now(),\n\t\t\tPriority: taskConfig.Priority,\n\t\t\tParent: taskConfig.Parent,\n\t\t},\n\t\tCertKey: taskConfig.CertKey,\n\t\tNodes: taskConfig.Nodes,\n\t\tEtcdNodes: taskConfig.EtcdNodes,\n\t\tClusterConfig: taskConfig.ClusterConfig,\n\t}\n\n\treturn task, nil\n}", "func StartContainer(url string, imageName string, done chan bool, ef util.ExchangeFolder, a action.ActionID) (int, error) {\n\tenvVar := []string{}\n\tenvVar = append(envVar, util.StarterEnvVariableKey+\"=\"+url)\n\tenvVar = append(envVar, util.StarterEnvNameVariableKey+\"=\"+common.Flags.Descriptor.File)\n\tenvVar = append(envVar, util.StarterEnvLoginVariableKey+\"=\"+common.Flags.Descriptor.Login)\n\tenvVar = append(envVar, util.StarterEnvPasswordVariableKey+\"=\"+common.Flags.Descriptor.Password)\n\tenvVar = append(envVar, util.StarterVerbosityVariableKey+\"=\"+strconv.Itoa(common.Flags.Logging.VerbosityLevel()))\n\tenvVar = append(envVar, util.ActionEnvVariableSkip+\"=\"+strconv.Itoa(common.Flags.Skipping.SkippingLevel()))\n\tenvVar = append(envVar, util.ActionEnvVariableKey+\"=\"+a.String())\n\tenvVar = append(envVar, \"http_proxy=\"+common.Flags.Proxy.HTTP)\n\tenvVar = append(envVar, \"https_proxy=\"+common.Flags.Proxy.HTTPS)\n\tenvVar = append(envVar, \"no_proxy=\"+common.Flags.Proxy.Exclusions)\n\n\tcommon.Logger.Printf(common.LOG_PASSING_CONTAINER_ENVARS, envVar)\n\n\t// Check if we need to load parameters from the comand line\n\tif common.Flags.Descriptor.ParamFile != \"\" {\n\t\tcopyExtraParameters(common.Flags.Descriptor.ParamFile, ef)\n\t}\n\n\tstartedAt := time.Now().UTC()\n\tstartedAt = startedAt.Add(time.Second * -2)\n\tresp, err := client.ContainerCreate(context.Background(), &container.Config{\n\t\tImage: imageName,\n\t\tWorkingDir: util.InstallerVolume,\n\t\tEnv: envVar,\n\t}, &container.HostConfig{\n\t\tMounts: []mount.Mount{\n\t\t\t{\n\t\t\t\tType: mount.TypeBind,\n\t\t\t\tSource: ef.Location.AdaptedPath(),\n\t\t\t\tTarget: util.InstallerVolume,\n\t\t\t},\n\t\t\t{\n\t\t\t\tType: mount.TypeBind,\n\t\t\t\tSource: \"/var/run/docker.sock\",\n\t\t\t\tTarget: \"/var/run/docker.sock\",\n\t\t\t},\n\t\t},\n\t}, nil, \"\")\n\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\t// Chan used to turn off the rolling log\n\tstopLogReading := make(chan bool)\n\n\t// Rolling output of the container logs\n\tgo func(start time.Time, exit chan bool) {\n\t\tlogMap := make(map[string]string)\n\t\t// Trick to avoid tracing twice the same log line\n\t\tnotExist := func(s string) (bool, string) {\n\t\t\ttab := strings.Split(s, util.InstallerLogPrefix)\n\t\t\tif len(tab) > 1 {\n\t\t\t\tsTrim := strings.Trim(tab[1], \" \")\n\t\t\t\tif _, ok := logMap[sTrim]; ok {\n\t\t\t\t\treturn false, \"\"\n\t\t\t\t}\n\t\t\t\tlogMap[sTrim] = \"\"\n\t\t\t\treturn true, util.InstallerLogPrefix + sTrim\n\t\t\t} else {\n\t\t\t\treturn true, s\n\t\t\t}\n\t\t}\n\n\t\t// Request to get the logs content from the container\n\t\treq := func(sr string) {\n\t\t\tout, err := client.ContainerLogs(context.Background(), resp.ID, types.ContainerLogsOptions{Since: sr, ShowStdout: true, ShowStderr: true})\n\t\t\tif err != nil {\n\t\t\t\tstopLogReading <- true\n\t\t\t}\n\t\t\ts := bufio.NewScanner(out)\n\t\t\tfor s.Scan() {\n\t\t\t\tstr := s.Text()\n\t\t\t\tif b, sTrim := notExist(str); b {\n\t\t\t\t\tidx := strings.Index(sTrim, util.FeedbackPrefix)\n\t\t\t\t\tif idx != -1 {\n\t\t\t\t\t\tfU := util.FeedbackUpdate{}\n\t\t\t\t\t\terr = json.Unmarshal([]byte(sTrim[idx+len(util.FeedbackPrefix):]), &fU)\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\tcommon.Logger.Println(\"Unable to parse progress update: \" + err.Error())\n\t\t\t\t\t\t} else if !common.Flags.Logging.ShouldOutputLogs() {\n\t\t\t\t\t\t\tswitch fU.Type {\n\t\t\t\t\t\t\tcase \"I\":\n\t\t\t\t\t\t\t\tcommon.CliFeedbackNotifier.Info(fU.Message)\n\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\tcase \"E\":\n\t\t\t\t\t\t\t\tcommon.CliFeedbackNotifier.Error(fU.Message)\n\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\tcase \"P\":\n\t\t\t\t\t\t\t\tcommon.CliFeedbackNotifier.ProgressG(fU.Key, fU.Goal, fU.Message)\n\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\tcase \"D\":\n\t\t\t\t\t\t\t\tcommon.CliFeedbackNotifier.Detail(fU.Message)\n\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t} else if common.Flags.Logging.ShouldOutputLogs() {\n\t\t\t\t\t\tfmt.Println(sTrim)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\terr = out.Close()\n\t\t\tif err != nil {\n\t\t\t\tcommon.Logger.Println(\"Unable to close container log reader: \" + err.Error())\n\t\t\t}\n\t\t}\n\tLoop:\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-exit:\n\t\t\t\t// Last call to be sure to get the end of the logs content\n\t\t\t\tnow := time.Now()\n\t\t\t\tnow = now.Add(time.Second * -1)\n\t\t\t\tsinceReq := strconv.FormatInt(now.Unix(), 10)\n\t\t\t\treq(sinceReq)\n\t\t\t\tbreak Loop\n\t\t\tdefault:\n\t\t\t\t// Running call to trace the container logs every 500ms\n\t\t\t\tsinceReq := strconv.FormatInt(start.Unix(), 10)\n\t\t\t\tstart = start.Add(time.Millisecond * 500)\n\t\t\t\treq(sinceReq)\n\t\t\t\ttime.Sleep(time.Millisecond * 500)\n\t\t\t}\n\t\t}\n\t}(startedAt, stopLogReading)\n\n\tdefer func() {\n\t\tif err := LogAllFromContainer(resp.ID, ef, done); err != nil {\n\t\t\tcommon.Logger.Println(\"Unable to fetch logs from container\")\n\t\t}\n\t}()\n\n\tif err := client.ContainerStart(context.Background(), resp.ID, types.ContainerStartOptions{}); err != nil {\n\t\tcommon.CliFeedbackNotifier.Error(\"Unable to start container: %s\", err.Error())\n\t\treturn 0, err\n\t}\n\n\tstatusCh, errCh := client.ContainerWait(context.Background(), resp.ID, container.WaitConditionNotRunning)\n\tselect {\n\tcase err := <-errCh:\n\t\tstopLogReading <- true\n\t\treturn 0, err\n\tcase status := <-statusCh:\n\t\tstopLogReading <- true\n\t\treturn int(status.StatusCode), nil\n\t}\n}", "func (c *Client) StartContainer(id string, hostConfig *HostConfig) error {\n\treturn c.startContainer(id, hostConfig, doOptions{})\n}", "func (p *Pool) Start(task interface{}) error {\n\tswitch t := task.(type) {\n\tcase func():\n\t\tvar w funcWrapper = t\n\t\tp.taskQueue <- &w\n\n\tcase Runnable:\n\t\tp.taskQueue <- t\n\n\tdefault:\n\t\treturn fmt.Errorf(\"invalid task type for Pool.Start, Runnable or func() expected, %T given\", t)\n\t}\n\n\treturn nil\n}", "func (t *TaskInfo) create() (containerInfo, error) {\n\targs := []string{\"create\"}\n\n\t// enable interactive\n\targs = append(args, \"-i\")\n\n\targs = append(args, \"--init\")\n\n\t// cpuset\n\tif len(t.Cpuset) != 0 {\n\t\tcpus := []string{}\n\t\tfor c := range t.Cpuset {\n\t\t\tcpus = append(cpus, strconv.Itoa(c))\n\t\t}\n\t\targs = append(args, \"--cpuset-cpus=\"+strings.Join(cpus, \",\"))\n\t}\n\n\t// network\n\tif !t.EnableNetwork {\n\t\targs = append(args, \"--net=none\")\n\t}\n\n\t// logging driver\n\tif !t.EnableLoggingDriver {\n\t\targs = append(args, \"--log-driver=none\")\n\t}\n\n\t// memory limit\n\tif t.MemoryLimitMB != 0 {\n\t\targs = append(args, fmt.Sprintf(\"--memory=%dm\", t.MemoryLimitMB))\n\t\targs = append(args, fmt.Sprintf(\"--memory-swap=%dm\", t.MemoryLimitMB))\n\t}\n\n\t// pids limit\n\tif t.PidsLimit != 0 {\n\t\targs = append(args, \"--pids-limit\")\n\t\targs = append(args, strconv.Itoa(t.PidsLimit))\n\t}\n\n\t// stack size\n\tif t.StackLimitKB != 0 {\n\t\targs = append(args, \"--ulimit\")\n\t\targs = append(args, fmt.Sprintf(\"stack=%d:%d\", t.StackLimitKB, t.StackLimitKB))\n\t}\n\n\t// workdir\n\tif t.WorkDir != \"\" {\n\t\targs = append(args, \"-w\")\n\t\targs = append(args, t.WorkDir)\n\t}\n\n\t// mount volume\n\tfor _, volumeMount := range t.VolumeMountInfo {\n\t\targs = append(args, \"-v\")\n\t\targs = append(args, fmt.Sprintf(\"%s:%s\", volumeMount.Volume.Name, volumeMount.Path))\n\t}\n\n\t// cgroup parent\n\tif t.cgroupParent != \"\" {\n\t\targs = append(args, fmt.Sprintf(\"--cgroup-parent=%s\", t.cgroupParent))\n\t}\n\n\t// container name\n\targs = append(args, t.Name)\n\n\t// extra arguments\n\targs = append(args, t.Argments...)\n\n\tcmd := exec.Command(\"docker\", args...)\n\n\tcmd.Stderr = os.Stderr\n\n\toutput, err := cmd.Output()\n\n\tif err != nil {\n\t\tlog.Println(\"create failed:\", err.Error())\n\t\treturn containerInfo{}, err\n\t}\n\n\tcontainerId := strings.TrimSpace(string(output))\n\n\treturn containerInfo{\n\t\tcontainerID: containerId,\n\t\tcgroupParent: t.cgroupParent,\n\t}, nil\n}", "func startService(sess *session.Session, subscription APISubscription, tasks []*string) error {\n\n\tsvc := ecs.New(sess)\n\n\tvar varClusterName = os.Getenv(\"CLUSTER_NAME\")\n\n\tfor _, task := range tasks {\n\t\tvTask := &ecs.StopTaskInput{\n\t\t\tCluster: aws.String(varClusterName),\n\t\t\tTask: task,\n\t\t}\n\t\twaiting := &ecs.DescribeTasksInput{\n\t\t\tCluster: varClusterName,\n\t\t\tTasks: tasks,\n\t\t}\n\n\t}\n\n\tupdateECS := &ecs.UpdateServiceInput{\n\t\tCluster: aws.String(varClusterName),\n\t\tDesiredCount: aws.Int64(1),\n\t\tForceNewDeployment: aws.Bool(true),\n\t\tService: aws.String(subscription.ID),\n\t}\n\t_, err := svc.UpdateService(updateECS)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase ecs.ErrCodeServerException:\n\t\t\t\tfmt.Println(ecs.ErrCodeServerException, aerr.Error())\n\t\t\tcase ecs.ErrCodeClientException:\n\t\t\t\tfmt.Println(ecs.ErrCodeClientException, aerr.Error())\n\t\t\tcase ecs.ErrCodeInvalidParameterException:\n\t\t\t\tfmt.Println(ecs.ErrCodeInvalidParameterException, aerr.Error())\n\t\t\tcase ecs.ErrCodeClusterNotFoundException:\n\t\t\t\tfmt.Println(ecs.ErrCodeClusterNotFoundException, aerr.Error())\n\t\t\tcase ecs.ErrCodeServiceNotFoundException:\n\t\t\t\tfmt.Println(ecs.ErrCodeServiceNotFoundException, aerr.Error())\n\t\t\tcase ecs.ErrCodeServiceNotActiveException:\n\t\t\t\tfmt.Println(ecs.ErrCodeServiceNotActiveException, aerr.Error())\n\t\t\tcase ecs.ErrCodePlatformUnknownException:\n\t\t\t\tfmt.Println(ecs.ErrCodePlatformUnknownException, aerr.Error())\n\t\t\tcase ecs.ErrCodePlatformTaskDefinitionIncompatibilityException:\n\t\t\t\tfmt.Println(ecs.ErrCodePlatformTaskDefinitionIncompatibilityException, aerr.Error())\n\t\t\tcase ecs.ErrCodeAccessDeniedException:\n\t\t\t\tfmt.Println(ecs.ErrCodeAccessDeniedException, aerr.Error())\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t}\n\t\treturn err\n\t}\n\n\treturn err\n}", "func (c *iperfTasks) Create(iperfTask *alpha1.IperfTask) (result *alpha1.IperfTask, err error) {\n\tresult = &alpha1.IperfTask{}\n\terr = c.client.Post().\n\t\tNamespace(c.ns).\n\t\tResource(\"iperftasks\").\n\t\tBody(iperfTask).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func NewTaskContainerWithSingleTask(taskName string, parameters map[string]string) *pb.TaskContainer {\n\treturn &pb.TaskContainer{\n\t\tParallelTasks: []*pb.Task{\n\t\t\tNewTask(taskName, parameters),\n\t\t},\n\t}\n}", "func (t TaskService) NewTaskDefinition() Definition {\n\tvar newDef Definition\n\n\tnewDef.Principal.LogonType = TASK_LOGON_INTERACTIVE_TOKEN\n\tnewDef.Principal.RunLevel = TASK_RUNLEVEL_LUA\n\n\tnewDef.RegistrationInfo.Author = t.connectedDomain + \"\\\\\" + t.connectedUser\n\tnewDef.RegistrationInfo.Date = time.Now()\n\n\tnewDef.Settings.AllowDemandStart = true\n\tnewDef.Settings.AllowHardTerminate = true\n\tnewDef.Settings.Compatibility = TASK_COMPATIBILITY_V2\n\tnewDef.Settings.DontStartOnBatteries = true\n\tnewDef.Settings.Enabled = true\n\tnewDef.Settings.Hidden = false\n\tnewDef.Settings.IdleSettings.IdleDuration = period.NewHMS(0, 10, 0) // PT10M\n\tnewDef.Settings.IdleSettings.WaitTimeout = period.NewHMS(1, 0, 0) // PT1H\n\tnewDef.Settings.MultipleInstances = TASK_INSTANCES_IGNORE_NEW\n\tnewDef.Settings.Priority = 7\n\tnewDef.Settings.RestartCount = 0\n\tnewDef.Settings.RestartOnIdle = false\n\tnewDef.Settings.RunOnlyIfIdle = false\n\tnewDef.Settings.RunOnlyIfNetworkAvailable = false\n\tnewDef.Settings.StartWhenAvailable = false\n\tnewDef.Settings.StopIfGoingOnBatteries = true\n\tnewDef.Settings.StopOnIdleEnd = true\n\tnewDef.Settings.TimeLimit = period.NewHMS(72, 0, 0) // PT72H\n\tnewDef.Settings.WakeToRun = false\n\n\treturn newDef\n}", "func (c *BasicECSClient) RunTask(ctx context.Context, in *ecs.RunTaskInput) (*ecs.RunTaskOutput, error) {\n\tif err := c.setup(); err != nil {\n\t\treturn nil, errors.Wrap(err, \"setting up client\")\n\t}\n\n\tvar out *ecs.RunTaskOutput\n\tvar err error\n\tmsg := awsutil.MakeAPILogMessage(\"RunTask\", in)\n\tif err := utility.Retry(ctx,\n\t\tfunc() (bool, error) {\n\t\t\tout, err = c.ecs.RunTaskWithContext(ctx, in)\n\t\t\tif awsErr, ok := err.(awserr.Error); ok {\n\t\t\t\tgrip.Debug(message.WrapError(awsErr, msg))\n\t\t\t\tif c.isNonRetryableErrorCode(awsErr.Code()) {\n\t\t\t\t\treturn false, err\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn true, err\n\t\t}, *c.opts.RetryOpts); err != nil {\n\t\treturn nil, err\n\t}\n\treturn out, nil\n}", "func Make(t *v2alpha2.TaskSpec) (core.Task, error) {\n\tif *t.Task != TaskName {\n\t\treturn nil, fmt.Errorf(\"task need to be '%s'\", TaskName)\n\t}\n\tvar jsonBytes []byte\n\tvar task Task\n\t// convert t to jsonBytes\n\tjsonBytes, err := json.Marshal(t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// convert jsonString to ExecTask\n\ttask = Task{}\n\terr = json.Unmarshal(jsonBytes, &task)\n\treturn &task, err\n}", "func (operation *Operation) CreateTask(content string) *Task {\n\ttask := &Task{\n\t\tOperation: operation,\n\t\tAction: NewAction(true, content),\n\t}\n\tcolumns := \"(`process`, `start`, `operation`, `result`, `content`)\"\n\tvalues := \"(?, ?, ?, ?, ?)\"\n\tsql := \"INSERT INTO `task` \" + columns + \" VALUES \" + values\n\tquery, err := database.Connection.Prepare(sql)\n\tif err != nil {\n\t\tfmt.Println(\"Error #1 when creating the task:\")\n\t\tfmt.Println(err)\n\t}\n\n\t_, err = query.Exec(operation.Process.ID, task.Start, operation.ID, task.GetResult(), task.Content)\n\tif err != nil {\n\t\tfmt.Println(\"Error #2 when creating the task:\")\n\t\tfmt.Println(err)\n\t}\n\n\t// find its id\n\twhereClause := \"start=? AND operation=? AND result=? AND is_running=? AND content=?\"\n\tsql = \"SELECT `id` FROM `task` WHERE \" + whereClause\n\tquery, err = database.Connection.Prepare(sql)\n\tquery.QueryRow(task.Start, task.Operation.ID, task.result, strconv.FormatBool(task.IsRunning), task.Content).Scan(&task.ID)\n\n\tif err != nil {\n\t\tfmt.Println(\"Error when selecting the task id:\")\n\t\tfmt.Println(err)\n\t}\n\n\treturn task\n}", "func NewTaskService(shimCtx context.Context, shimCancel context.CancelFunc, publisher shim.Publisher) (taskAPI.TaskService, error) {\n\t// We provide an empty string for \"id\" as the service manages multiple tasks; there is no single\n\t// \"id\" being managed. As noted in the comments of the called code, the \"id\" arg is only used by\n\t// the Cleanup function, so it will never be invoked as part of the task service API, which is all\n\t// we need.\n\truncService, err := runc.New(shimCtx, \"\", publisher, shimCancel)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdh, err := newDriveHandler(blockPath, drivePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &TaskService{\n\t\ttaskManager: vm.NewTaskManager(shimCtx, log.G(shimCtx)),\n\t\truncService: runcService,\n\n\t\tpublisher: publisher,\n\t\tshimCtx: shimCtx,\n\t\tshimCancel: shimCancel,\n\t\tdriveHandler: dh,\n\t}, nil\n}", "func StartContainer(connText *context.Context, containerID string) {\n\texists, err := containers.Exists(*connText, containerID, false)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\tif exists {\n\t\terr := containers.Start(*connText, containerID, nil)\n\t\tif err != nil {\n\t\t\tlog.Fatalln(err)\n\t\t}\n\n\t\trunning := define.ContainerStateRunning\n\t\t_, err = containers.Wait(*connText, containerID, &running)\n\t\tif err != nil {\n\t\t\tlog.Fatalln(err)\n\t\t}\n\t}\n}", "func (Tests) Start(ctx context.Context) error {\n\targ := BuildDockerComposeArgs(ProjectName, ProjectType, \"test\", DockerComposeTestFile)\n\targ = append(arg, \"up\", \"-d\")\n\targ = append(arg, DockerComposeTestDependencies...)\n\treturn Exec(ComposeBin, arg...)\n}", "func CommandStart(conf Config, ctx, query Query) error {\n\tts, err := LoadTaskSet(conf.Repo, conf.IDsFile, false)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif query.Template > 0 {\n\t\treturn errors.New(\"templates not yet supported for start command\")\n\t}\n\n\tif len(query.IDs) > 0 {\n\t\t// start given tasks by IDs\n\t\tfor _, id := range query.IDs {\n\t\t\ttask := ts.MustGetByID(id)\n\t\t\ttask.Status = STATUS_ACTIVE\n\t\t\tif query.Text != \"\" {\n\t\t\t\ttask.Notes += \"\\n\" + query.Text\n\t\t\t}\n\t\t\tts.MustUpdateTask(task)\n\n\t\t\tts.SavePendingChanges()\n\t\t\tMustGitCommit(conf.Repo, \"Started %s\", task)\n\n\t\t\tif task.Notes != \"\" {\n\t\t\t\tfmt.Printf(\"\\nNotes on task %d:\\n\\033[38;5;245m%s\\033[0m\\n\\n\", task.ID, task.Notes)\n\t\t\t}\n\t\t}\n\t} else if query.Text != \"\" {\n\t\t// create a new task that is already active (started)\n\t\tquery = query.Merge(ctx)\n\t\ttask := Task{\n\t\t\tWritePending: true,\n\t\t\tStatus: STATUS_ACTIVE,\n\t\t\tSummary: query.Text,\n\t\t\tTags: query.Tags,\n\t\t\tProject: query.Project,\n\t\t\tPriority: query.Priority,\n\t\t\tNotes: query.Note,\n\t\t}\n\t\ttask = ts.MustLoadTask(task)\n\t\tts.SavePendingChanges()\n\t\tMustGitCommit(conf.Repo, \"Added and started %s\", task)\n\t} else {\n\t\treturn errors.New(\"nothing to do -- specify an ID or describe a task\")\n\t}\n\treturn nil\n\n}", "func (s *Supervisor) AddTask(name string, startStopper StartStopper, policyOptions ...PolicyOption) {\n\tkey := fmt.Sprintf(\"%s-%s\", \"task\", name)\n\ts.mux.Lock()\n\tdefer s.mux.Unlock()\n\n\tif _, exists := s.processes[key]; exists {\n\t\ts.logger(Error, loggerData{\"name\": name}, \"task already exists\")\n\t\treturn\n\t}\n\n\tt := &task{\n\t\tStartStopper: startStopper,\n\t\tname: name,\n\t\tlogger: s.logger,\n\t}\n\n\tp := Policy{\n\t\tRestart: s.policy.Restart,\n\t}\n\tp.Reconfigure(policyOptions...)\n\n\tt.restartPolicy = p.Restart\n\n\ts.processes[key] = t\n}", "func (wc *workflowClient) StartWorkflow(\n\tctx context.Context,\n\toptions StartWorkflowOptions,\n\tworkflowFunc interface{},\n\targs ...interface{},\n) (*WorkflowExecution, error) {\n\tworkflowID := options.ID\n\tif len(workflowID) == 0 {\n\t\tworkflowID = uuid.NewRandom().String()\n\t}\n\n\tif options.TaskList == \"\" {\n\t\treturn nil, errors.New(\"missing TaskList\")\n\t}\n\n\texecutionTimeout := common.Int32Ceil(options.ExecutionStartToCloseTimeout.Seconds())\n\tif executionTimeout <= 0 {\n\t\treturn nil, errors.New(\"missing or invalid ExecutionStartToCloseTimeout\")\n\t}\n\n\tdecisionTaskTimeout := common.Int32Ceil(options.DecisionTaskStartToCloseTimeout.Seconds())\n\tif decisionTaskTimeout < 0 {\n\t\treturn nil, errors.New(\"negative DecisionTaskStartToCloseTimeout provided\")\n\t}\n\tif decisionTaskTimeout == 0 {\n\t\tdecisionTaskTimeout = defaultDecisionTaskTimeoutInSecs\n\t}\n\n\t// Validate type and its arguments.\n\tworkflowType, input, err := getValidatedWorkflowFunction(workflowFunc, args, wc.dataConverter, wc.registry)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmemo, err := getWorkflowMemo(options.Memo, wc.dataConverter)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsearchAttr, err := serializeSearchAttributes(options.SearchAttributes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdelayStartSeconds := common.Int32Ceil(options.DelayStart.Seconds())\n\tif delayStartSeconds < 0 {\n\t\treturn nil, errors.New(\"Invalid DelayStart option\")\n\t}\n\n\tjitterStartSeconds := common.Int32Ceil(options.JitterStart.Seconds())\n\tif jitterStartSeconds < 0 {\n\t\treturn nil, errors.New(\"Invalid JitterStart option\")\n\t}\n\n\t// create a workflow start span and attach it to the context object.\n\t// N.B. we need to finish this immediately as jaeger does not give us a way\n\t// to recreate a span given a span context - which means we will run into\n\t// issues during replay. we work around this by creating and ending the\n\t// workflow start span and passing in that context to the workflow. So\n\t// everything beginning with the StartWorkflowExecutionRequest will be\n\t// parented by the created start workflow span.\n\tctx, span := createOpenTracingWorkflowSpan(ctx, wc.tracer, time.Now(), fmt.Sprintf(\"StartWorkflow-%s\", workflowType.Name), workflowID)\n\tspan.Finish()\n\n\t// get workflow headers from the context\n\theader := wc.getWorkflowHeader(ctx)\n\n\t// run propagators to extract information about tracing and other stuff, store in headers field\n\tstartRequest := &s.StartWorkflowExecutionRequest{\n\t\tDomain: common.StringPtr(wc.domain),\n\t\tRequestId: common.StringPtr(uuid.New()),\n\t\tWorkflowId: common.StringPtr(workflowID),\n\t\tWorkflowType: workflowTypePtr(*workflowType),\n\t\tTaskList: common.TaskListPtr(s.TaskList{Name: common.StringPtr(options.TaskList)}),\n\t\tInput: input,\n\t\tExecutionStartToCloseTimeoutSeconds: common.Int32Ptr(executionTimeout),\n\t\tTaskStartToCloseTimeoutSeconds: common.Int32Ptr(decisionTaskTimeout),\n\t\tIdentity: common.StringPtr(wc.identity),\n\t\tWorkflowIdReusePolicy: options.WorkflowIDReusePolicy.toThriftPtr(),\n\t\tRetryPolicy: convertRetryPolicy(options.RetryPolicy),\n\t\tCronSchedule: common.StringPtr(options.CronSchedule),\n\t\tMemo: memo,\n\t\tSearchAttributes: searchAttr,\n\t\tHeader: header,\n\t\tDelayStartSeconds: common.Int32Ptr(delayStartSeconds),\n\t\tJitterStartSeconds: common.Int32Ptr(jitterStartSeconds),\n\t}\n\n\tvar response *s.StartWorkflowExecutionResponse\n\n\t// Start creating workflow request.\n\terr = backoff.Retry(ctx,\n\t\tfunc() error {\n\t\t\ttchCtx, cancel, opt := newChannelContext(ctx, wc.featureFlags)\n\t\t\tdefer cancel()\n\n\t\t\tvar err1 error\n\t\t\tresponse, err1 = wc.workflowService.StartWorkflowExecution(tchCtx, startRequest, opt...)\n\t\t\treturn err1\n\t\t}, createDynamicServiceRetryPolicy(ctx), isServiceTransientError)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif wc.metricsScope != nil {\n\t\tscope := wc.metricsScope.GetTaggedScope(tagTaskList, options.TaskList, tagWorkflowType, workflowType.Name)\n\t\tscope.Counter(metrics.WorkflowStartCounter).Inc(1)\n\t}\n\n\texecutionInfo := &WorkflowExecution{\n\t\tID: workflowID,\n\t\tRunID: response.GetRunId()}\n\treturn executionInfo, nil\n}", "func start(tasks []string) {\n\t// Iterate over all available tasks.\n\tfor i := range tasks {\n\t\tn, as := parseTask(tasks[i])\n\t\tcmd := exec.Command(n, as...)\n\t\tcmd.Stderr = os.Stderr\n\t\tcmd.Stdout = os.Stdout\n\t\tt := replaceVars(tasks[i])\n\t\tlog.Trace.Printf(\"Starting `%s`...\", t)\n\t\terr := cmd.Start()\n\t\tif err != nil {\n\t\t\tlog.Error.Printf(`Failed to start a command \"%s\", error: %v.`, t, err)\n\t\t}\n\t}\n}", "func TaskFromThrift(config *aurora.TaskConfig) *AuroraTask {\n\n\tnewTask := NewTask()\n\n\t// Pass values using receivers as much as possible\n\tnewTask.\n\t\tEnvironment(config.Job.Environment).\n\t\tRole(config.Job.Role).\n\t\tName(config.Job.Name).\n\t\tMaxFailure(config.MaxTaskFailures).\n\t\tIsService(config.IsService).\n\t\tPriority(config.Priority)\n\n\tif config.Tier != nil {\n\t\tnewTask.Tier(*config.Tier)\n\t}\n\n\tif config.Production != nil {\n\t\tnewTask.Production(*config.Production)\n\t}\n\n\tif config.ExecutorConfig != nil {\n\t\tnewTask.\n\t\t\tExecutorName(config.ExecutorConfig.Name).\n\t\t\tExecutorData(config.ExecutorConfig.Data)\n\t}\n\n\tif config.PartitionPolicy != nil {\n\t\tnewTask.PartitionPolicy(\n\t\t\taurora.PartitionPolicy{\n\t\t\t\tReschedule: config.PartitionPolicy.Reschedule,\n\t\t\t\tDelaySecs: thrift.Int64Ptr(*config.PartitionPolicy.DelaySecs),\n\t\t\t})\n\t}\n\n\t// Make a deep copy of the task's container\n\tif config.Container != nil {\n\t\tif config.Container.Mesos != nil {\n\t\t\tmesosContainer := NewMesosContainer()\n\n\t\t\tif config.Container.Mesos.Image != nil {\n\t\t\t\tif config.Container.Mesos.Image.Appc != nil {\n\t\t\t\t\tmesosContainer.AppcImage(config.Container.Mesos.Image.Appc.Name, config.Container.Mesos.Image.Appc.ImageId)\n\t\t\t\t} else if config.Container.Mesos.Image.Docker != nil {\n\t\t\t\t\tmesosContainer.DockerImage(config.Container.Mesos.Image.Docker.Name, config.Container.Mesos.Image.Docker.Tag)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tfor _, vol := range config.Container.Mesos.Volumes {\n\t\t\t\tmesosContainer.AddVolume(vol.ContainerPath, vol.HostPath, vol.Mode)\n\t\t\t}\n\n\t\t\tnewTask.Container(mesosContainer)\n\t\t} else if config.Container.Docker != nil {\n\t\t\tdockerContainer := NewDockerContainer()\n\t\t\tdockerContainer.Image(config.Container.Docker.Image)\n\n\t\t\tfor _, param := range config.Container.Docker.Parameters {\n\t\t\t\tdockerContainer.AddParameter(param.Name, param.Value)\n\t\t\t}\n\n\t\t\tnewTask.Container(dockerContainer)\n\t\t}\n\t}\n\n\t// Copy all ports\n\tfor _, resource := range config.Resources {\n\t\t// Copy only ports. Set CPU, RAM, DISK, and GPU\n\t\tif resource != nil {\n\t\t\tif resource.NamedPort != nil {\n\t\t\t\tnewTask.task.Resources = append(\n\t\t\t\t\tnewTask.task.Resources,\n\t\t\t\t\t&aurora.Resource{NamedPort: thrift.StringPtr(*resource.NamedPort)},\n\t\t\t\t)\n\t\t\t\tnewTask.portCount++\n\t\t\t}\n\n\t\t\tif resource.RamMb != nil {\n\t\t\t\tnewTask.RAM(*resource.RamMb)\n\t\t\t}\n\n\t\t\tif resource.NumCpus != nil {\n\t\t\t\tnewTask.CPU(*resource.NumCpus)\n\t\t\t}\n\n\t\t\tif resource.DiskMb != nil {\n\t\t\t\tnewTask.Disk(*resource.DiskMb)\n\t\t\t}\n\n\t\t\tif resource.NumGpus != nil {\n\t\t\t\tnewTask.GPU(*resource.NumGpus)\n\t\t\t}\n\t\t}\n\t}\n\n\t// Copy constraints\n\tfor _, constraint := range config.Constraints {\n\t\tif constraint != nil && constraint.Constraint != nil {\n\n\t\t\tnewConstraint := aurora.Constraint{Name: constraint.Name}\n\n\t\t\ttaskConstraint := constraint.Constraint\n\t\t\tif taskConstraint.Limit != nil {\n\t\t\t\tnewConstraint.Constraint = &aurora.TaskConstraint{\n\t\t\t\t\tLimit: &aurora.LimitConstraint{Limit: taskConstraint.Limit.Limit},\n\t\t\t\t}\n\t\t\t\tnewTask.task.Constraints = append(newTask.task.Constraints, &newConstraint)\n\n\t\t\t} else if taskConstraint.Value != nil {\n\n\t\t\t\tvalues := make([]string, 0)\n\t\t\t\tfor _, val := range taskConstraint.Value.Values {\n\t\t\t\t\tvalues = append(values, val)\n\t\t\t\t}\n\n\t\t\t\tnewConstraint.Constraint = &aurora.TaskConstraint{\n\t\t\t\t\tValue: &aurora.ValueConstraint{Negated: taskConstraint.Value.Negated, Values: values}}\n\n\t\t\t\tnewTask.task.Constraints = append(newTask.task.Constraints, &newConstraint)\n\t\t\t}\n\t\t}\n\t}\n\n\t// Copy labels\n\tfor _, label := range config.Metadata {\n\t\tnewTask.task.Metadata = append(newTask.task.Metadata, &aurora.Metadata{Key: label.Key, Value: label.Value})\n\t}\n\n\t// Copy Mesos fetcher URIs\n\tfor _, uri := range config.MesosFetcherUris {\n\t\tnewTask.task.MesosFetcherUris = append(\n\t\t\tnewTask.task.MesosFetcherUris,\n\t\t\t&aurora.MesosFetcherURI{\n\t\t\t\tValue: uri.Value,\n\t\t\t\tExtract: thrift.BoolPtr(*uri.Extract),\n\t\t\t\tCache: thrift.BoolPtr(*uri.Cache),\n\t\t\t})\n\t}\n\n\treturn newTask\n}", "func CreateTask(task Task) (Task, error) {\n\tres, err := makeRequest(http.MethodPost, \"tasks\", task.taskSave())\n\tif err != nil {\n\t\treturn Task{}, err\n\t}\n\n\treturn decodeTask(res.Body)\n}", "func ConvertToTaskDefinition(params ConvertTaskDefParams) (*ecs.TaskDefinition, error) {\n\tif len(params.ContainerConfigs) == 0 {\n\t\treturn nil, errors.New(\"cannot create a task definition with no containers; invalid service config\")\n\t}\n\n\t// Instantiates zero values for fields on task def specified by ecs-params\n\ttaskDefParams, err := convertTaskDefParams(params.ECSParams)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// The task-role-arn flag takes precedence over a taskRoleArn value specified in ecs-params file.\n\tif params.TaskRoleArn == \"\" {\n\t\tparams.TaskRoleArn = taskDefParams.taskRoleArn\n\t}\n\n\t// Create containerDefinitions\n\tcontainerDefinitions := []*ecs.ContainerDefinition{}\n\n\tfor _, containerConfig := range params.ContainerConfigs {\n\t\tname := containerConfig.Name\n\t\t// Check if there are ecs-params specified for the container\n\t\tecsContainerDef := &ContainerDef{Essential: true}\n\t\tif cd, ok := taskDefParams.containerDefs[name]; ok {\n\t\t\tecsContainerDef = &cd\n\t\t}\n\n\t\t// Validate essential containers\n\t\tcount := len(params.ContainerConfigs)\n\t\tif !hasEssential(taskDefParams.containerDefs, count) {\n\t\t\treturn nil, errors.New(\"Task definition does not have any essential containers\")\n\t\t}\n\n\t\ttaskVals := taskLevelValues{\n\t\t\tMemLimit: taskDefParams.memory,\n\t\t}\n\n\t\tcontainerDef, err := reconcileContainerDef(&containerConfig, ecsContainerDef, taskVals)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tcontainerDefinitions = append(containerDefinitions, containerDef)\n\t}\n\n\tecsVolumes, err := convertToECSVolumes(params.Volumes, params.ECSParams)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\texecutionRoleArn := taskDefParams.executionRoleArn\n\n\tplacementConstraints := convertToTaskDefinitionConstraints(params.ECSParams)\n\n\t// Check for and apply provided ecs-registry-creds values\n\tif params.ECSRegistryCreds != nil {\n\t\terr := addRegistryCredsToContainerDefs(containerDefinitions, params.ECSRegistryCreds.CredentialResources.ContainerCredentials)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// if provided, add or replace existing executionRoleArn with value from cred file\n\t\tif params.ECSRegistryCreds.CredentialResources.TaskExecutionRole != \"\" {\n\t\t\tnewExecutionRole := params.ECSRegistryCreds.CredentialResources.TaskExecutionRole\n\n\t\t\tif executionRoleArn != \"\" {\n\t\t\t\t// TODO: refactor 'showResourceOverrideMsg()' to take in override src and use here\n\t\t\t\tlog.WithFields(log.Fields{\n\t\t\t\t\t\"option name\": \"task_execution_role\",\n\t\t\t\t}).Infof(\"Using \"+regcredio.ECSCredFileBaseName+\" value as override (was %s but is now %s)\", executionRoleArn, newExecutionRole)\n\t\t\t} else {\n\t\t\t\tlog.WithFields(log.Fields{\n\t\t\t\t\t\"option name\": \"task_execution_role\",\n\t\t\t\t}).Infof(\"Using \"+regcredio.ECSCredFileBaseName+\" value %s\", newExecutionRole)\n\t\t\t}\n\t\t\texecutionRoleArn = newExecutionRole\n\t\t}\n\t}\n\n\t// Note: this is later converted into an ecs.RegisterTaskDefinitionInput in entity_helper.go\n\ttaskDefinition := &ecs.TaskDefinition{\n\t\tFamily: aws.String(params.TaskDefName),\n\t\tContainerDefinitions: containerDefinitions,\n\t\tVolumes: ecsVolumes,\n\t\tTaskRoleArn: aws.String(params.TaskRoleArn),\n\t\tNetworkMode: aws.String(taskDefParams.networkMode),\n\t\tCpu: aws.String(taskDefParams.cpu),\n\t\tMemory: aws.String(taskDefParams.memory),\n\t\tExecutionRoleArn: aws.String(executionRoleArn),\n\t\tPlacementConstraints: placementConstraints,\n\t}\n\n\t// Set launch type\n\tif params.RequiredCompatibilites != \"\" {\n\t\ttaskDefinition.RequiresCompatibilities = []*string{aws.String(params.RequiredCompatibilites)}\n\t}\n\tif taskDefParams.pidMode != \"\" {\n\t\ttaskDefinition.SetPidMode(taskDefParams.pidMode)\n\t}\n\tif taskDefParams.ipcMode != \"\" {\n\t\ttaskDefinition.SetIpcMode(taskDefParams.ipcMode)\n\t}\n\treturn taskDefinition, nil\n}", "func (c *gcsCore) CreateContainer(id string, settings prot.VMHostedContainerSettings) error {\n\tc.containerCacheMutex.Lock()\n\tdefer c.containerCacheMutex.Unlock()\n\n\tif c.getContainer(id) != nil {\n\t\treturn gcserr.NewHresultError(gcserr.HrVmcomputeSystemAlreadyExists)\n\t}\n\n\tcontainerEntry := newContainerCacheEntry(id)\n\t// We need to only allow exited notifications when at least one WaitProcess\n\t// call has been written. We increment the writers here which is safe even\n\t// on failure because this entry will not be in the map on failure.\n\tlogrus.Debugf(\"+1 initprocess.writersWg [gcsCore::CreateContainer]\")\n\tcontainerEntry.initProcess.writersWg.Add(1)\n\n\t// Set up mapped virtual disks.\n\tif err := c.setupMappedVirtualDisks(id, settings.MappedVirtualDisks); err != nil {\n\t\treturn errors.Wrapf(err, \"failed to set up mapped virtual disks during create for container %s\", id)\n\t}\n\tfor _, disk := range settings.MappedVirtualDisks {\n\t\tcontainerEntry.AddMappedVirtualDisk(disk)\n\t}\n\t// Set up mapped directories.\n\tif err := c.setupMappedDirectories(id, settings.MappedDirectories); err != nil {\n\t\treturn errors.Wrapf(err, \"failed to set up mapped directories during create for container %s\", id)\n\t}\n\tfor _, dir := range settings.MappedDirectories {\n\t\tcontainerEntry.AddMappedDirectory(dir)\n\t}\n\n\t// Set up layers.\n\tscratch, layers, err := c.getLayerMounts(settings.SandboxDataPath, settings.Layers)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to get layer devices for container %s\", id)\n\t}\n\tcontainerEntry.Index, err = c.getOrAddContainerIndex(id)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to get a valid container index\")\n\t}\n\n\tif err := c.mountLayers(containerEntry.Index, scratch, layers); err != nil {\n\t\treturn errors.Wrapf(err, \"failed to mount layers for container %s\", id)\n\t}\n\n\t// Stash network adapters away\n\tfor _, adapter := range settings.NetworkAdapters {\n\t\tcontainerEntry.AddNetworkAdapter(adapter)\n\t}\n\t// Create the directory that will contain the resolv.conf file.\n\t//\n\t// TODO(rn): This isn't quite right but works. Basically, when\n\t// we do the network config in ExecProcess() the overlay for\n\t// the rootfs has already been created. When we then write\n\t// /etc/resolv.conf to the base layer it won't show up unless\n\t// /etc exists when the overlay is created. This is a bit\n\t// problematic as we basically later write to a what is\n\t// supposed to be read-only layer in the overlay... Ideally,\n\t// dockerd would pass a runc config with a bind mount for\n\t// /etc/resolv.conf like it does on unix.\n\tif err := os.MkdirAll(filepath.Join(baseFilesPath, \"etc\"), 0755); err != nil {\n\t\treturn errors.Wrapf(err, \"failed to create resolv.conf directory\")\n\t}\n\n\tc.containerCache[id] = containerEntry\n\n\treturn nil\n}", "func (m *MockManager) CreateInstanceTask(arg0 *config.Config) (*task.Task, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"CreateInstanceTask\", arg0)\n\tret0, _ := ret[0].(*task.Task)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (r *remoteRuntimeService) StartContainer(ctx context.Context, containerID string) (err error) {\n\tklog.V(10).InfoS(\"[RemoteRuntimeService] StartContainer\", \"containerID\", containerID, \"timeout\", r.timeout)\n\tctx, cancel := context.WithTimeout(ctx, r.timeout)\n\tdefer cancel()\n\n\tif _, err := r.runtimeClient.StartContainer(ctx, &runtimeapi.StartContainerRequest{\n\t\tContainerId: containerID,\n\t}); err != nil {\n\t\tklog.ErrorS(err, \"StartContainer from runtime service failed\", \"containerID\", containerID)\n\t\treturn err\n\t}\n\tklog.V(10).InfoS(\"[RemoteRuntimeService] StartContainer Response\", \"containerID\", containerID)\n\n\treturn nil\n}", "func (s *Service) CreateTask(ctx context.Context, req *taskspb.CreateTaskReq) (*taskspb.CreateTaskRes, error) {\n\ttaskID := uuid.NewV4().String()\n\tdata := req.GetData()\n\n\tnewTask := types.Task{\n\t\tID: taskID,\n\t\tData: data,\n\t}\n\n\t// insert newTask to DB\n\t_, err := s.db.tasks.InsertOne(ctx, newTask)\n\tif err != nil {\n\t\treturn nil, status.Errorf(\n\t\t\tcodes.Internal,\n\t\t\tfmt.Sprintf(\"unexpected error: %v\", err),\n\t\t)\n\t}\n\n\tres := &taskspb.CreateTaskRes{\n\t\tId: taskID,\n\t}\n\n\treturn res, nil\n}" ]
[ "0.65428984", "0.63425386", "0.618552", "0.6107186", "0.6064747", "0.6054203", "0.604236", "0.6014893", "0.59061396", "0.5791596", "0.5766864", "0.5757924", "0.5691278", "0.56747615", "0.56268", "0.55775166", "0.5566711", "0.55478686", "0.55056304", "0.54779893", "0.54329526", "0.542788", "0.5422692", "0.5339538", "0.53301895", "0.52663946", "0.5250949", "0.5221968", "0.5206786", "0.5180475", "0.5170479", "0.5168469", "0.51642007", "0.51584524", "0.5154574", "0.5153702", "0.51080674", "0.5081849", "0.5074305", "0.5069516", "0.5047581", "0.50454706", "0.5042991", "0.5038277", "0.5035246", "0.502265", "0.50139475", "0.5004779", "0.50023216", "0.49972278", "0.4994446", "0.49741602", "0.4958928", "0.49551874", "0.49533126", "0.49375334", "0.49316084", "0.49168235", "0.49140316", "0.4914031", "0.49139467", "0.4888416", "0.48875654", "0.48779222", "0.48682404", "0.48563665", "0.4843448", "0.48369265", "0.483382", "0.4827392", "0.4822119", "0.48217848", "0.48114258", "0.4787288", "0.47817934", "0.47727925", "0.4770126", "0.47694576", "0.47592944", "0.47527313", "0.47491503", "0.47417244", "0.47377193", "0.47359756", "0.47276184", "0.4718575", "0.47129774", "0.46968645", "0.46947733", "0.4694232", "0.46919078", "0.46844855", "0.46832594", "0.46736374", "0.467317", "0.4670889", "0.46599284", "0.46571788", "0.4654566", "0.46423367" ]
0.6873531
0
WaitTask function is expected to return a channel that will send an ExitResult when the task exits or close the channel when the context is canceled. It is also expected that calling WaitTask on an exited task will immediately send an ExitResult on the returned channel. A call to WaitTask after StopTask is valid and should be handled. If WaitTask is called after DestroyTask, it should return drivers.ErrTaskNotFound as no task state should exist after DestroyTask is called.
func (d *Driver) WaitTask(ctx context.Context, taskID string) (<-chan *drivers.ExitResult, error) { d.logger.Debug("WaitTask called", "task", taskID) handle, ok := d.tasks.Get(taskID) if !ok { return nil, drivers.ErrTaskNotFound } ch := make(chan *drivers.ExitResult) go handle.runExitWatcher(ctx, ch) return ch, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (t *Task) Wait(timeout time.Duration) (TaskResult, error) {\n\tselect {\n\tcase <-t.startedChan:\n\tdefault:\n\t\treturn nil, ErrNotExecuting\n\t}\n\n\tt.resultLock.RLock()\n\tif t.waitResult != nil {\n\t\tt.resultLock.RUnlock()\n\t\treturn t.waitResult, nil\n\t}\n\tt.resultLock.RUnlock()\n\n\tvar timeoutChan <-chan time.Time = make(chan time.Time)\n\tif timeout > 0 {\n\t\ttimeoutChan = t.cfg.clock.After(timeout)\n\t}\n\n\tselect {\n\tcase res := <-t.resultChan:\n\t\tt.completed(res)\n\t\tt.SetRunning(false)\n\t\treturn res, nil\n\tcase <-timeoutChan:\n\t\treturn nil, ErrTimeout\n\t}\n}", "func (t *Task) Wait() error {\n\t<-*t.Done\n\tif t.err != nil {\n\t\treturn goerr.Wrap(t.err)\n\t}\n\treturn nil\n}", "func (ts *TaskService) Wait(ctx context.Context, req *taskAPI.WaitRequest) (*taskAPI.WaitResponse, error) {\n\tdefer logPanicAndDie(log.G(ctx))\n\n\tlog.G(ctx).WithFields(logrus.Fields{\"id\": req.ID, \"exec_id\": req.ExecID}).Debug(\"wait\")\n\ttask, err := ts.taskManager.Task(req.ID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx = namespaces.WithNamespace(ctx, defaultNamespace)\n\tresp, err := task.Wait(ctx, req)\n\tif err != nil {\n\t\tlog.G(ctx).WithError(err).Error(\"wait failed\")\n\t\treturn nil, err\n\t}\n\n\tlog.G(ctx).WithField(\"exit_status\", resp.ExitStatus).Debug(\"wait succeeded\")\n\treturn resp, nil\n}", "func (ft *FutureTask) Wait(timeout time.Duration) (res *[]byte, err error) {\n\tselect {\n\tcase res = <-ft.out:\n\tcase <-time.After(timeout):\n\t\terr = fmt.Errorf(\"task(%+v) timeout\", ft)\n\t}\n\treturn\n}", "func (s *Session) Wait() error {\n\tfor {\n\t\tt := s.pClient.GetTask(s.task.ID)\n\t\tif t.Err != nil {\n\t\t\treturn errors.Wrapf(t.Err, \"getting task %q failed\", s.task.ID)\n\t\t}\n\n\t\tif t.State == \"Stopped\" || t.State == \"Disabled\" {\n\t\t\treturn errors.Errorf(\"failed to wait for task: task %q is in state %q (last failure: %q)\",\n\t\t\t\ts.task.ID,\n\t\t\t\tt.State,\n\t\t\t\tt.LastFailureMessage)\n\n\t\t}\n\n\t\tif (t.HitCount - (t.FailedCount + t.MissCount)) > 0 {\n\t\t\treturn nil\n\t\t}\n\n\t\t// Make sure that data is published.\n\t\ttime.Sleep(100 * time.Millisecond)\n\t}\n}", "func (t *TaskBox[T, U, C, CT, TF]) GetTaskCh() chan Task[T] {\n\treturn t.task\n}", "func (s *InvokeSync) Wait(ctx context.Context) error {\n\tif !s.wait.Wait(ctx) {\n\t\treturn task.StopReason(ctx)\n\t}\n\treturn s.err\n}", "func (t *Task) Wait() {\n\t<-t.done\n}", "func (t *TaskController[T, U, C, CT, TF]) Wait() {\n\tt.wg.Wait()\n\tclose(t.resultCh)\n\tt.pool.DeleteTask(t.taskID)\n}", "func WaitForAsyncTaskDone(ctx context.Context, taskID string, client *goqedit.APIClient) (goqedit.GetTaskStatusResponse, error) {\n\tvar taskStatus goqedit.GetTaskStatusResponse\n\ttaskStatusRequest := goqedit.GetTaskStatusRequest{Id: taskID}\n\tfor i := 0; i < asyncTaskRetries; i++ {\n\t\ttaskStatus, _, err := client.NodeApi.NodeGetTaskStatusPost(ctx, taskStatusRequest)\n\t\tif err != nil {\n\t\t\treturn taskStatus, fmt.Errorf(\"couldn't get task status: %v\", ErrorResponseString(err))\n\t\t}\n\t\tif taskStatus.Result != \"pending\" && taskStatus.Result != \"in_progress\" {\n\t\t\treturn taskStatus, nil\n\t\t}\n\t\tfmt.Println(\"Waiting for task to be done\")\n\t\ttime.Sleep(asyncTaskWaitTime)\n\t}\n\treturn taskStatus, fmt.Errorf(\"waiting for task timed out after %d iterations\", asyncTaskRetries)\n}", "func NewWaitTask(name string, ids object.ObjMetadataSet, cond Condition, timeout time.Duration, mapper meta.RESTMapper) *WaitTask {\n\treturn &WaitTask{\n\t\tTaskName: name,\n\t\tIds: ids,\n\t\tCondition: cond,\n\t\tTimeout: timeout,\n\t\tMapper: mapper,\n\t}\n}", "func (c *TaskChain) Wait() error {\n\treturn errors.EnsureStack(c.eg.Wait())\n}", "func (t *Task) Wait() {\n\tt.ow.Do(func() {\n\t\tt.wg.Wait()\n\t})\n}", "func (d *Driver) DestroyTask(taskID string, force bool) error {\n\thandle, ok := d.tasks.Get(taskID)\n\tif !ok {\n\t\treturn drivers.ErrTaskNotFound\n\t}\n\n\tif handle.isRunning() && !force {\n\t\treturn fmt.Errorf(\"cannot destroy running task\")\n\t}\n\n\tif handle.isRunning() {\n\t\td.logger.Debug(\"Have to destroyTask but container is still running\", \"containerID\", handle.containerID)\n\t\t// we can not do anything, so catching the error is useless\n\t\terr := d.podman.ContainerStop(d.ctx, handle.containerID, 60)\n\t\tif err != nil {\n\t\t\td.logger.Warn(\"failed to stop/kill container during destroy\", \"error\", err)\n\t\t}\n\t\t// wait a while for stats emitter to collect exit code etc.\n\t\tfor i := 0; i < 20; i++ {\n\t\t\tif !handle.isRunning() {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\ttime.Sleep(time.Millisecond * 250)\n\t\t}\n\t\tif handle.isRunning() {\n\t\t\td.logger.Warn(\"stats emitter did not exit while stop/kill container during destroy\", \"error\", err)\n\t\t}\n\t}\n\n\tif handle.removeContainerOnExit {\n\t\terr := d.podman.ContainerDelete(d.ctx, handle.containerID, true, true)\n\t\tif err != nil {\n\t\t\td.logger.Warn(\"Could not remove container\", \"container\", handle.containerID, \"error\", err)\n\t\t}\n\t}\n\n\td.tasks.Delete(taskID)\n\treturn nil\n}", "func (ts *TaskService) Wait(requestCtx context.Context, req *taskAPI.WaitRequest) (*taskAPI.WaitResponse, error) {\n\tdefer logPanicAndDie(log.G(requestCtx))\n\tlog.G(requestCtx).WithFields(logrus.Fields{\"id\": req.ID, \"exec_id\": req.ExecID}).Debug(\"wait\")\n\n\tresp, err := ts.runcService.Wait(requestCtx, req)\n\tif err != nil {\n\t\tlog.G(requestCtx).WithError(err).Error(\"wait failed\")\n\t\treturn nil, err\n\t}\n\n\tlog.G(requestCtx).WithField(\"exit_status\", resp.ExitStatus).Debug(\"wait succeeded\")\n\treturn resp, nil\n}", "func waitForTask() {\n\t// create UNBUFFERED channel for messages of type string\n\tch := make(chan string)\n\n\tgo func() {\n\t\t// wait here until manager gives us a task\n\t\t// (because channel is unbuffered/blocking)\n\t\tp := <-ch // channel receive unary operator\n\t\tfmt.Println(\"worker: received signal:\", p)\n\t}()\n\n\t// wait here for a bit\n\ttime.Sleep(time.Duration(rand.Intn(500)) * time.Millisecond)\n\t// send message\n\tch <- \"paper\"\n\tfmt.Println(\"manager: sent signal\")\n\n\ttime.Sleep(time.Second)\n\tfmt.Println(\"------------ done ---------\")\n}", "func monitorTask(ctx context.Context, task *model.Task, channel chan model.Event) {\n\t// derive new timeout context\n\tmonitorCtx, cancel := context.WithTimeout(ctx, 10 * time.Second)\n\tdefer cancel()\n\n\tselect {\n\tcase <- monitorCtx.Done():\n\t\t// check status of task\n\t\tstatus := task.GetStatus()\n\n\t\tif status != model.TaskStatusInitial && status != model.TaskStatusExecuting {\n\t\t\treturn\n\t\t}\n\n\t\t// task may still be active\n\t\tswitch monitorCtx.Err().Error() {\n\t\tcase \"context canceled\": // termination of processes\n\t\t\tutil.LogInfo(task.UUID, \"ENG\", \"termination\")\n\t\t\tchannel <- model.NewEvent(task.Domain, task.UUID, model.EventTypeTaskTermination, task.UUID, \"termination\")\n\t\tdefault: // timeout\n\t\t\tutil.LogInfo(task.UUID, \"ENG\", \"timeout\")\n\t\t\tchannel <- model.NewEvent(task.Domain, task.UUID, model.EventTypeTaskTimeout, task.UUID, \"timeout\")\n\t\t}\n\t}\n}", "func completeIfWaitTask(currentTask Task, taskContext *TaskContext) {\n\tif wt, ok := currentTask.(*WaitTask); ok {\n\t\twt.complete(taskContext)\n\t}\n}", "func (e *Eval) taskWait(ctx context.Context, f *Flow, task *sched.Task) error {\n\tif err := task.Wait(ctx, sched.TaskRunning); err != nil {\n\t\treturn err\n\t}\n\t// Grab the task's exec so that it can be logged properly.\n\tf.Exec = task.Exec\n\tif f.Op == Exec && f.Argmap != nil {\n\t\t// If this is an Exec and f.Argmap is defined, then\n\t\t// update the flow's resolved filesets.\n\t\tn := f.NExecArg()\n\t\tf.resolvedFs = make([]*reflow.Fileset, n)\n\t\tfor i := 0; i < n; i++ {\n\t\t\tearg, arg := f.ExecArg(i), task.Config.Args[i]\n\t\t\tif earg.Out {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tf.resolvedFs[earg.Index] = arg.Fileset\n\t\t}\n\t}\n\te.LogFlow(ctx, f)\n\tif err := task.Wait(ctx, sched.TaskDone); err != nil {\n\t\treturn err\n\t}\n\tf.RunInfo = task.RunInfo\n\tif task.Err != nil {\n\t\te.Mutate(f, task.Err, Done)\n\t} else {\n\t\te.Mutate(f, task.Result.Err, task.Result.Fileset, Propagate, Done)\n\t}\n\treturn nil\n}", "func (d *Driver) StopTask(taskID string, timeout time.Duration, signal string) error {\n\td.logger.Info(\"Stopping task\", \"taskID\", taskID, \"signal\", signal)\n\thandle, ok := d.tasks.Get(taskID)\n\tif !ok {\n\t\treturn drivers.ErrTaskNotFound\n\t}\n\t// fixme send proper signal to container\n\terr := d.podman.ContainerStop(d.ctx, handle.containerID, int(timeout.Seconds()))\n\tif err != nil {\n\t\td.logger.Error(\"Could not stop/kill container\", \"containerID\", handle.containerID, \"err\", err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (w *WaitTask) Start(taskContext *TaskContext) {\n\tklog.V(2).Infof(\"wait task starting (name: %q, objects: %d)\",\n\t\tw.Name(), len(w.Ids))\n\n\t// TODO: inherit context from task runner, passed through the TaskContext\n\tctx := context.Background()\n\n\t// use a context wrapper to handle complete/cancel/timeout\n\tif w.Timeout > 0 {\n\t\tctx, w.cancelFunc = context.WithTimeout(ctx, w.Timeout)\n\t} else {\n\t\tctx, w.cancelFunc = context.WithCancel(ctx)\n\t}\n\n\tw.startInner(taskContext)\n\n\t// A goroutine to handle ending the WaitTask.\n\tgo func() {\n\t\t// Block until complete/cancel/timeout\n\t\t<-ctx.Done()\n\t\t// Err is always non-nil when Done channel is closed.\n\t\terr := ctx.Err()\n\n\t\tklog.V(2).Infof(\"wait task completing (name: %q,): %v\", w.TaskName, err)\n\n\t\tswitch err {\n\t\tcase context.Canceled:\n\t\t\t// happy path - cancelled or completed (not considered an error)\n\t\tcase context.DeadlineExceeded:\n\t\t\t// timed out\n\t\t\tw.sendTimeoutEvents(taskContext)\n\t\t}\n\n\t\t// Update RESTMapper to pick up new custom resource types\n\t\tw.updateRESTMapper(taskContext)\n\n\t\t// Done here. signal completion to the task runner\n\t\ttaskContext.TaskChannel() <- TaskResult{}\n\t}()\n}", "func (c *EC2) WaitUntilExportTaskCancelled(input *DescribeExportTasksInput) error {\n\treturn c.WaitUntilExportTaskCancelledWithContext(aws.BackgroundContext(), input)\n}", "func (p *WorkPool) DoWait(task TaskHandler) { // 添加到工作池,并等待执行完成之后再返回\n\tif p.IsClosed() { // closed\n\t\treturn\n\t}\n\n\tdoneChan := make(chan struct{})\n\tp.waitingQueue.Push(TaskHandler(func() error {\n\t\tdefer close(doneChan)\n\t\treturn task()\n\t}))\n\t<-doneChan\n}", "func (p *WorkPool) Wait() error { // 等待工作线程执行结束\n\tp.waitingQueue.Wait() // 等待队列结束\n\tp.waitingQueue.Close() //\n\tp.waitTask() // wait que down\n\tclose(p.task)\n\tp.wg.Wait() // 等待结束\n\tselect {\n\tcase err := <-p.errChan:\n\t\treturn err\n\tdefault:\n\t\treturn nil\n\t}\n}", "func TestRktDriver_Start_Wait_Stop(t *testing.T) {\n\tctestutil.RktCompatible(t)\n\tif !testutil.IsCI() {\n\t\tt.Parallel()\n\t}\n\n\trequire := require.New(t)\n\td := NewRktDriver(testlog.HCLogger(t))\n\tharness := dtestutil.NewDriverHarness(t, d)\n\n\ttask := &drivers.TaskConfig{\n\t\tID: uuid.Generate(),\n\t\tAllocID: uuid.Generate(),\n\t\tName: \"etcd\",\n\t\tResources: &drivers.Resources{\n\t\t\tNomadResources: &structs.AllocatedTaskResources{\n\t\t\t\tMemory: structs.AllocatedMemoryResources{\n\t\t\t\t\tMemoryMB: 128,\n\t\t\t\t},\n\t\t\t\tCpu: structs.AllocatedCpuResources{\n\t\t\t\t\tCpuShares: 100,\n\t\t\t\t},\n\t\t\t},\n\t\t\tLinuxResources: &drivers.LinuxResources{\n\t\t\t\tMemoryLimitBytes: 134217728,\n\t\t\t\tCPUShares: 100,\n\t\t\t},\n\t\t},\n\t}\n\n\ttc := &TaskConfig{\n\t\tTrustPrefix: \"coreos.com/etcd\",\n\t\tImageName: \"coreos.com/etcd:v2.0.4\",\n\t\tCommand: \"/etcd\",\n\t\tArgs: []string{\"--version\"},\n\t\tNet: []string{\"none\"},\n\t\tDebug: true,\n\t}\n\trequire.NoError(task.EncodeConcreteDriverConfig(&tc))\n\tcleanup := harness.MkAllocDir(task, true)\n\tdefer cleanup()\n\n\thandle, _, err := harness.StartTask(task)\n\trequire.NoError(err)\n\n\t// Wait on the task, it should exit since we are only asking for etcd version here\n\tch, err := harness.WaitTask(context.Background(), handle.Config.ID)\n\trequire.NoError(err)\n\tresult := <-ch\n\trequire.Nil(result.Err)\n\n\trequire.Zero(result.ExitCode)\n\n\trequire.NoError(harness.DestroyTask(task.ID, true))\n\n}", "func (c *Cond) Wait() {\n\tcur := task.Current()\n\tfor {\n\t\tt := (*task.Task)(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(&c.t))))\n\t\tswitch t {\n\t\tcase nil:\n\t\t\t// Condition variable has not been notified.\n\t\t\t// Block the current task on the condition variable.\n\t\t\tif atomic.CompareAndSwapPointer((*unsafe.Pointer)(unsafe.Pointer(&c.t)), nil, unsafe.Pointer(cur)) {\n\t\t\t\ttask.Pause()\n\t\t\t\treturn\n\t\t\t}\n\t\tcase &notifiedPlaceholder:\n\t\t\t// A notification arrived and there is no waiting goroutine.\n\t\t\t// Clear the notification and return.\n\t\t\tif atomic.CompareAndSwapPointer((*unsafe.Pointer)(unsafe.Pointer(&c.t)), unsafe.Pointer(t), nil) {\n\t\t\t\treturn\n\t\t\t}\n\t\tdefault:\n\t\t\tpanic(\"interrupt.Cond: condition variable in use by another goroutine\")\n\t\t}\n\t}\n}", "func (b *baseRunner) nextTask(taskQueue chan Task,\n\ttaskContext *TaskContext) (Task, bool) {\n\tvar tsk Task\n\tselect {\n\t// If there is any tasks left in the queue, this\n\t// case statement will be executed.\n\tcase t := <-taskQueue:\n\t\ttsk = t\n\tdefault:\n\t\t// Only happens when the channel is empty.\n\t\treturn nil, true\n\t}\n\n\tswitch st := tsk.(type) {\n\tcase *WaitTask:\n\t\t// The wait tasks need to be handled specifically here. Before\n\t\t// starting a new wait task, we check if the condition is already\n\t\t// met. Without this check, a task might end up waiting for\n\t\t// status events when the condition is in fact already met.\n\t\tif st.checkCondition(taskContext, b.collector) {\n\t\t\tst.startAndComplete(taskContext)\n\t\t} else {\n\t\t\tst.Start(taskContext)\n\t\t}\n\tdefault:\n\t\ttsk.Start(taskContext)\n\t}\n\treturn tsk, false\n}", "func (task *Task) IsWaiting() bool {\n\treturn task.status == TaskStatusWaiting\n}", "func TestRktDriver_StartWaitRecoverWaitStop(t *testing.T) {\n\tctestutil.RktCompatible(t)\n\tif !testutil.IsCI() {\n\t\tt.Parallel()\n\t}\n\n\trequire := require.New(t)\n\td := NewRktDriver(testlog.HCLogger(t))\n\tharness := dtestutil.NewDriverHarness(t, d)\n\n\ttask := &drivers.TaskConfig{\n\t\tID: uuid.Generate(),\n\t\tAllocID: uuid.Generate(),\n\t\tName: \"etcd\",\n\t\tResources: &drivers.Resources{\n\t\t\tNomadResources: &structs.AllocatedTaskResources{\n\t\t\t\tMemory: structs.AllocatedMemoryResources{\n\t\t\t\t\tMemoryMB: 128,\n\t\t\t\t},\n\t\t\t\tCpu: structs.AllocatedCpuResources{\n\t\t\t\t\tCpuShares: 100,\n\t\t\t\t},\n\t\t\t},\n\t\t\tLinuxResources: &drivers.LinuxResources{\n\t\t\t\tMemoryLimitBytes: 134217728,\n\t\t\t\tCPUShares: 100,\n\t\t\t},\n\t\t},\n\t}\n\n\ttc := &TaskConfig{\n\t\tImageName: \"coreos.com/etcd:v2.0.4\",\n\t\tCommand: \"/etcd\",\n\t}\n\trequire.NoError(task.EncodeConcreteDriverConfig(&tc))\n\n\tcleanup := harness.MkAllocDir(task, true)\n\tdefer cleanup()\n\n\thandle, _, err := harness.StartTask(task)\n\trequire.NoError(err)\n\n\tch, err := harness.WaitTask(context.Background(), task.ID)\n\trequire.NoError(err)\n\n\tvar waitDone bool\n\tvar wg sync.WaitGroup\n\twg.Add(1)\n\tgo func() {\n\t\tdefer wg.Done()\n\t\tresult := <-ch\n\t\trequire.Error(result.Err)\n\t\twaitDone = true\n\t}()\n\n\toriginalStatus, err := d.InspectTask(task.ID)\n\trequire.NoError(err)\n\n\td.(*Driver).tasks.Delete(task.ID)\n\n\twg.Wait()\n\trequire.True(waitDone)\n\t_, err = d.InspectTask(task.ID)\n\trequire.Equal(drivers.ErrTaskNotFound, err)\n\n\terr = d.RecoverTask(handle)\n\trequire.NoError(err)\n\n\tstatus, err := d.InspectTask(task.ID)\n\trequire.NoError(err)\n\trequire.Exactly(originalStatus, status)\n\n\tch, err = harness.WaitTask(context.Background(), task.ID)\n\trequire.NoError(err)\n\n\trequire.NoError(d.StopTask(task.ID, 0, \"SIGKILL\"))\n\n\tselect {\n\tcase result := <-ch:\n\t\trequire.NoError(result.Err)\n\t\trequire.NotZero(result.ExitCode)\n\n\t\t// when killing a task, signal might not propagate\n\t\t// when executor proc.Wait() call gets \"wait: no child processes\" error\n\t\t//require.Equal(9, result.Signal)\n\tcase <-time.After(time.Duration(testutil.TestMultiplier()*5) * time.Second):\n\t\trequire.Fail(\"WaitTask timeout\")\n\t}\n\n\trequire.NoError(d.DestroyTask(task.ID, false))\n}", "func (c *EC2) WaitUntilConversionTaskCancelled(input *DescribeConversionTasksInput) error {\n\treturn c.WaitUntilConversionTaskCancelledWithContext(aws.BackgroundContext(), input)\n}", "func getWaitChan(ctx context.Context) (wait waitChan, ok bool) {\n\twait, ok = ctx.Value(waitKey{}).(waitChan)\n\n\treturn\n}", "func (a *agent) GetTask(ctx context.Context, msg *api.AgentID) (*api.Task, error) {\n\tvar task *api.Task = new(api.Task)\n\tselect {\n\tcase task, ok := <-a.work[msg.GetAgentID()]:\n\t\tif ok {\n\t\t\treturn task, nil\n\t\t}\n\t\treturn task, errors.New(\"channel closed\")\n\tdefault:\n\t\treturn task, nil\n\t}\n}", "func (m *Module) Wait(ctx context.Context) error {\n\tselect {\n\tcase <-m.done:\n\t\tif m.err != nil {\n\t\t\treturn m.err\n\t\t}\n\t\treturn nil\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\t}\n}", "func WaitContext(ctx context.Context, cmd *exec.Cmd) error {\n\t// We use cmd.Process.Wait instead of cmd.Wait because cmd.Wait is not reenterable\n\tc := make(chan error, 1)\n\tgo func() {\n\t\tif cmd == nil || cmd.Process == nil {\n\t\t\tc <- nil\n\t\t} else {\n\t\t\t_, err := cmd.Process.Wait()\n\t\t\tc <- err\n\t\t}\n\t}()\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn ErrorWaitTimeout\n\tcase err := <-c:\n\t\treturn err\n\t}\n}", "func Wait(group TaskGroup, fun TaskFunc) error {\n\treturn waitFor(group.Task(fun))\n}", "func (m *MockTaskDao) GetExecWaitTask() ([]task.OwlTask, int, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"GetExecWaitTask\")\n\tret0, _ := ret[0].([]task.OwlTask)\n\tret1, _ := ret[1].(int)\n\tret2, _ := ret[2].(error)\n\treturn ret0, ret1, ret2\n}", "func waitForTask() {\n\tch := make(chan string)\n\n\tgo func() {\n\t\td := <-ch\n\t\tfmt.Println(\"child : recv'd signal :\", d)\n\t}()\n\n\ttime.Sleep(time.Duration(rand.Intn(500)) * time.Millisecond)\n\tch <- \"data\"\n\tfmt.Println(\"parent : sent signal\")\n\n\ttime.Sleep(time.Second)\n\tfmt.Println(\"-------------------------------------------------\")\n}", "func (c *EC2) WaitUntilConversionTaskDeleted(input *DescribeConversionTasksInput) error {\n\treturn c.WaitUntilConversionTaskDeletedWithContext(aws.BackgroundContext(), input)\n}", "func pollTaskAndWait(task enamlbosh.BoshTask, client *enamlbosh.Client, tries int) error {\n\tUIPrint(\"polling task...\")\n\tdefer UIPrint(fmt.Sprintf(\"Finished with Task %s\", task.Description))\n\tticker := time.Tick(time.Second)\n\tcount := 0\n\tfor {\n\t\t<-ticker\n\t\tvar err error\n\t\ttask, err = client.GetTask(task.ID)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tswitch task.State {\n\t\tcase enamlbosh.StatusDone:\n\t\t\tUIPrintStatus(fmt.Sprintf(\"task state %s\", task.State))\n\t\t\treturn nil\n\t\tcase enamlbosh.StatusCancelled, enamlbosh.StatusError:\n\t\t\terr := fmt.Errorf(\"%s - %s\", task.State, task.Description)\n\t\t\tlo.G.Error(\"task error: \" + err.Error())\n\t\t\treturn err\n\t\tdefault:\n\t\t\tUIPrintStatus(fmt.Sprintf(\"task '%s' is %s\", task.Description, task.State))\n\t\t}\n\t\tcount++\n\n\t\tif tries != -1 && count >= tries {\n\t\t\tUIPrintStatus(\"hit poll limit, exiting task poller without error\")\n\t\t\treturn nil\n\t\t}\n\t}\n}", "func TaskWrapper(ctx context.Context, logger Logger, processors Processors, wg *sync.WaitGroup, task Task, host *Host, results chan *JobResult) error {\n\tif err := processors.TaskInstanceStarted(ctx, logger, host, task); err != nil {\n\t\terr = errors.Wrap(err, \"problem running HostStart\")\n\t\tlogger.Error(err.Error())\n\t\treturn err\n\t}\n\n\tdefer wg.Done()\n\tres, err := task.Run(ctx, logger, host)\n\thost.SetErr(err)\n\n\tjobResult := NewJobResult(ctx, host, res, err)\n\n\tif err := processors.TaskInstanceCompleted(ctx, logger, jobResult, host, task); err != nil {\n\t\terr = errors.Wrap(err, \"problem running HostCompleted\")\n\t\tlogger.Error(err.Error())\n\t\treturn err\n\t}\n\n\tresults <- jobResult\n\treturn nil\n}", "func (s *FutureTree) Wait(ctx context.Context) {\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn\n\tcase res, ok := <-s.ch:\n\t\tif ok {\n\t\t\ts.res = res\n\t\t}\n\t}\n}", "func (cs *CheckinSession) Wait(ctx context.Context) error {\n\t// Once wait terminates, the checkin session should be removed from the event\n\t// server. It might have already been partially removed by a successful round\n\t// or the event being ended.\n\tdefer func() {\n\t\tcs.server.lock.Lock()\n\t\tdefer cs.server.lock.Unlock()\n\n\t\tcs.close()\n\t}()\n\t// Wait for the session to succeed, fail or time out\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn errors.New(\"context cancelled\")\n\tcase err := <-cs.result:\n\t\treturn err\n\t}\n}", "func (t *Task) StopAndWait(timeout time.Duration) (TaskResult, error) {\n\terr := t.Stop()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tres, err := t.Wait(timeout)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn res, nil\n}", "func (f *FakeCmdRunner) Wait() error {\n\treturn f.Err\n}", "func (krct *keyRegistrationConfirmationTask) RunTask() (interface{}, error) {\n\tlog.Infof(\"Waiting for confirmation for the Key [%x]\", krct.key)\n\tif krct.ctx == nil {\n\t\tkrct.ctx, _ = krct.contextInitializer(krct.timeout)\n\t}\n\n\tid := newEthereumIdentity(krct.centID, krct.contract, krct.config, krct.queue, krct.gethClientFinder, krct.contractProvider)\n\tcontract, err := id.getContract()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tkrct.filterer = contract\n\tfOpts := &bind.FilterOpts{\n\t\tContext: krct.ctx,\n\t\tStart: krct.blockHeight,\n\t}\n\n\tfor {\n\t\titer, err := krct.filterer.FilterKeyAdded(fOpts, [][32]byte{krct.key}, []*big.Int{big.NewInt(int64(krct.keyPurpose))})\n\t\tif err != nil {\n\t\t\treturn nil, centerrors.Wrap(err, \"failed to start filtering key event logs\")\n\t\t}\n\n\t\terr = utils.LookForEvent(iter)\n\t\tif err == nil {\n\t\t\tlog.Infof(\"Received filtered event Key Registration Confirmation for CentrifugeID [%s] and key [%x] with purpose [%d]\\n\", krct.centID.String(), krct.key, krct.keyPurpose)\n\t\t\treturn iter.Event, nil\n\t\t}\n\n\t\tif err != utils.ErrEventNotFound {\n\t\t\treturn nil, err\n\t\t}\n\t\ttime.Sleep(100 * time.Millisecond)\n\t}\n}", "func (w *WaitTask) Cancel(_ *TaskContext) {\n\tw.cancelFunc()\n}", "func TestRktDriver_Start_Wait_Skip_Trust(t *testing.T) {\n\tctestutil.RktCompatible(t)\n\tif !testutil.IsCI() {\n\t\tt.Parallel()\n\t}\n\n\trequire := require.New(t)\n\td := NewRktDriver(testlog.HCLogger(t))\n\tharness := dtestutil.NewDriverHarness(t, d)\n\n\ttask := &drivers.TaskConfig{\n\t\tID: uuid.Generate(),\n\t\tAllocID: uuid.Generate(),\n\t\tName: \"etcd\",\n\t\tResources: &drivers.Resources{\n\t\t\tNomadResources: &structs.AllocatedTaskResources{\n\t\t\t\tMemory: structs.AllocatedMemoryResources{\n\t\t\t\t\tMemoryMB: 128,\n\t\t\t\t},\n\t\t\t\tCpu: structs.AllocatedCpuResources{\n\t\t\t\t\tCpuShares: 100,\n\t\t\t\t},\n\t\t\t},\n\t\t\tLinuxResources: &drivers.LinuxResources{\n\t\t\t\tMemoryLimitBytes: 134217728,\n\t\t\t\tCPUShares: 100,\n\t\t\t},\n\t\t},\n\t}\n\n\ttc := &TaskConfig{\n\t\tImageName: \"coreos.com/etcd:v2.0.4\",\n\t\tCommand: \"/etcd\",\n\t\tArgs: []string{\"--version\"},\n\t\tNet: []string{\"none\"},\n\t\tDebug: true,\n\t}\n\trequire.NoError(task.EncodeConcreteDriverConfig(&tc))\n\ttesttask.SetTaskConfigEnv(task)\n\n\tcleanup := harness.MkAllocDir(task, true)\n\tdefer cleanup()\n\n\thandle, _, err := harness.StartTask(task)\n\trequire.NoError(err)\n\n\t// Wait on the task, it should exit since we are only asking for etcd version here\n\tch, err := harness.WaitTask(context.Background(), handle.Config.ID)\n\trequire.NoError(err)\n\tresult := <-ch\n\trequire.Nil(result.Err)\n\trequire.Zero(result.ExitCode)\n\n\trequire.NoError(harness.DestroyTask(task.ID, true))\n\n}", "func (p *process) Wait(ctx context.Context) error {\n\treturn WaitContext(ctx, p.cmd)\n}", "func (t *Task) MustWait() {\n\tgoerr.Check(t.Wait())\n}", "func (t *Task) WaitForRunning(timeout time.Duration) error {\n\tvar timeoutChan <-chan time.Time = make(chan time.Time)\n\tif timeout > 0 {\n\t\ttimeoutChan = t.cfg.clock.After(timeout)\n\t}\n\tselect {\n\tcase res := <-t.resultChan:\n\t\tt.completed(res)\n\t\treturn nil\n\tcase <-t.runningChan:\n\t\treturn nil\n\tcase <-timeoutChan:\n\t\treturn ErrTimeout\n\t}\n}", "func (client *Client) DescribeSyncEcsHostTaskWithCallback(request *DescribeSyncEcsHostTaskRequest, callback func(response *DescribeSyncEcsHostTaskResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeSyncEcsHostTaskResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeSyncEcsHostTask(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func TestTaskWaitForHostResources(t *testing.T) {\n\tctx, cancel := context.WithCancel(context.TODO())\n\tdefer cancel()\n\n\t// 1 vCPU available on host\n\thostResourceManager := NewHostResourceManager(getTestHostResources())\n\ttaskEngine := &DockerTaskEngine{\n\t\tmanagedTasks: make(map[string]*managedTask),\n\t\tmonitorQueuedTaskEvent: make(chan struct{}, 1),\n\t\thostResourceManager: &hostResourceManager,\n\t}\n\tgo taskEngine.monitorQueuedTasks(ctx)\n\t// 3 tasks requesting 0.5 vCPUs each\n\ttasks := []*apitask.Task{}\n\tfor i := 0; i < 3; i++ {\n\t\ttask := testdata.LoadTask(\"sleep5\")\n\t\ttask.Arn = fmt.Sprintf(\"arn%d\", i)\n\t\ttask.CPU = float64(0.5)\n\t\tmtask := &managedTask{\n\t\t\tTask: task,\n\t\t\tengine: taskEngine,\n\t\t\tconsumedHostResourceEvent: make(chan struct{}, 1),\n\t\t}\n\t\ttasks = append(tasks, task)\n\t\ttaskEngine.managedTasks[task.Arn] = mtask\n\t}\n\n\t// acquire for host resources order arn0, arn1, arn2\n\tgo func() {\n\t\ttaskEngine.managedTasks[\"arn0\"].waitForHostResources()\n\t\ttaskEngine.managedTasks[\"arn1\"].waitForHostResources()\n\t\ttaskEngine.managedTasks[\"arn2\"].waitForHostResources()\n\t}()\n\ttime.Sleep(500 * time.Millisecond)\n\n\t// Verify waiting queue is waiting at arn2\n\ttopTask, err := taskEngine.topTask()\n\tassert.NoError(t, err)\n\tassert.Equal(t, topTask.Arn, \"arn2\")\n\n\t// Remove 1 task\n\ttaskResources := taskEngine.managedTasks[\"arn0\"].ToHostResources()\n\ttaskEngine.hostResourceManager.release(\"arn0\", taskResources)\n\ttaskEngine.wakeUpTaskQueueMonitor()\n\n\ttime.Sleep(500 * time.Millisecond)\n\n\t// Verify arn2 got dequeued\n\ttopTask, err = taskEngine.topTask()\n\tassert.Error(t, err)\n}", "func (m *etcdMinion) TaskListener(c chan<- *task.Task) error {\n\tlog.Printf(\"Task listener is watching %s\\n\", m.queueDir)\n\n\trand.Seed(time.Now().UTC().UnixNano())\n\tb := backoff.Backoff{\n\t\tMin: 1 * time.Second,\n\t\tMax: 10 * time.Minute,\n\t\tFactor: 2.0,\n\t\tJitter: true,\n\t}\n\n\twatcherOpts := &etcdclient.WatcherOptions{\n\t\tRecursive: true,\n\t}\n\twatcher := m.kapi.Watcher(m.queueDir, watcherOpts)\n\n\tfor {\n\t\tresp, err := watcher.Next(context.Background())\n\t\tif err != nil {\n\t\t\t// Use a backoff and retry later again\n\t\t\tduration := b.Duration()\n\t\t\tlog.Printf(\"%s, retrying in %s\\n\", err, duration)\n\t\t\ttime.Sleep(duration)\n\t\t\tcontinue\n\t\t}\n\n\t\t// Reset the backoff counter on successful receive\n\t\tb.Reset()\n\n\t\t// Ignore \"delete\" events when removing a task from the queue\n\t\taction := strings.ToLower(resp.Action)\n\t\tif strings.EqualFold(action, \"delete\") {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Unmarshal and remove task from the queue\n\t\tt, err := EtcdUnmarshalTask(resp.Node)\n\t\tm.kapi.Delete(context.Background(), resp.Node.Key, nil)\n\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Received invalid task %s: %s\\n\", resp.Node.Key, err)\n\t\t\tcontinue\n\t\t}\n\n\t\t// Send the task for processing\n\t\tlog.Printf(\"Received task %s\\n\", t.ID)\n\t\tt.State = task.TaskStateQueued\n\t\tt.TimeReceived = time.Now().Unix()\n\t\tif err := m.SaveTaskResult(t); err != nil {\n\t\t\tlog.Printf(\"Unable to save task state: %s\\n\", err)\n\t\t\tcontinue\n\t\t}\n\n\t\tc <- t\n\t}\n\n\treturn nil\n}", "func (t *TaskBox[T, U, C, CT, TF]) GetResultCh() chan U {\n\treturn t.resultCh\n}", "func (j *Job) Wait(ctx context.Context) error {\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\tcase <-j.doneCh:\n\t\treturn nil\n\t}\n}", "func (d *dockerWaiter) wait(ctx context.Context, containerID string, stopFn func()) error {\n\tstatusCh, errCh := d.client.ContainerWait(ctx, containerID, container.WaitConditionNotRunning)\n\n\tif stopFn != nil {\n\t\tstopFn()\n\t}\n\n\tfor {\n\t\tselect {\n\t\tcase <-time.After(time.Second):\n\t\t\tif stopFn != nil {\n\t\t\t\tstopFn()\n\t\t\t}\n\n\t\tcase err := <-errCh:\n\t\t\treturn err\n\n\t\tcase status := <-statusCh:\n\t\t\tif status.StatusCode != 0 {\n\t\t\t\treturn &common.BuildError{\n\t\t\t\t\tInner: fmt.Errorf(\"exit code %d\", status.StatusCode),\n\t\t\t\t\tExitCode: int(status.StatusCode),\n\t\t\t\t}\n\t\t\t}\n\n\t\t\treturn nil\n\t\t}\n\t}\n}", "func (f WaiterFunc) Wait(ctx context.Context) error {\n\treturn f(ctx)\n}", "func (c *BasicECSClient) StopTask(ctx context.Context, in *ecs.StopTaskInput) (*ecs.StopTaskOutput, error) {\n\tif err := c.setup(); err != nil {\n\t\treturn nil, errors.Wrap(err, \"setting up client\")\n\t}\n\n\tvar out *ecs.StopTaskOutput\n\tvar err error\n\tmsg := awsutil.MakeAPILogMessage(\"StopTask\", in)\n\tif err := utility.Retry(ctx,\n\t\tfunc() (bool, error) {\n\t\t\tout, err = c.ecs.StopTaskWithContext(ctx, in)\n\t\t\tif awsErr, ok := err.(awserr.Error); ok {\n\t\t\t\tgrip.Debug(message.WrapError(awsErr, msg))\n\t\t\t\tif c.isNonRetryableErrorCode(awsErr.Code()) {\n\t\t\t\t\treturn false, err\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn true, err\n\t\t}, *c.opts.RetryOpts); err != nil {\n\t\treturn nil, err\n\t}\n\treturn out, nil\n}", "func (i *instanceManager) dispenseTaskEventsCh() (<-chan *drivers.TaskEvent, context.CancelFunc, error) {\n\tdriver, err := i.dispense()\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tctx, cancel := context.WithCancel(i.ctx)\n\teventsCh, err := driver.TaskEvents(ctx)\n\tif err != nil {\n\t\tcancel()\n\t\treturn nil, nil, err\n\t}\n\n\treturn eventsCh, cancel, nil\n}", "func (t *Task) Reject() (interface{}, error) {\n\tpar := map[string]interface{}{\n\t\t\"taskid\": t.taskId,\n\t}\n\treturn t.nc.Exec(\"task.reject\", par)\n}", "func (ret *OpRet) Wait() error {\n\tif ret.delayed == nil {\n\t\treturn nil\n\t}\n\n\t<-ret.delayed\n\treturn ret.error\n}", "func (s *sshSessionExternal) Wait() error {\n\tif s.exited() {\n\t\treturn nil\n\t}\n\terr := s.cmd.Wait()\n\tif err == nil {\n\t\tfs.Debugf(s.f, \"ssh external: command exited OK\")\n\t} else {\n\t\tfs.Debugf(s.f, \"ssh external: command exited with error: %v\", err)\n\t}\n\treturn err\n}", "func wait(ctx context.Context, c TimedActuator,\n\tresChan chan error, cancel context.CancelFunc) error {\n\tif timeout := c.GetTimeout(); timeout != nil {\n\t\treturn waitWithTimeout(ctx, resChan, *timeout, cancel)\n\t}\n\n\tfor {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn nil\n\t\tcase err := <-resChan:\n\t\t\tif err != nil {\n\t\t\t\tcancel()\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n}", "func (_Contract *ContractCallerSession) GetTask(i *big.Int) (struct {\n\tActive bool\n\tAssignment *big.Int\n\tProposalID *big.Int\n}, error) {\n\treturn _Contract.Contract.GetTask(&_Contract.CallOpts, i)\n}", "func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *drivers.DriverNetwork, error) {\n\tif _, ok := d.tasks.Get(cfg.ID); ok {\n\t\treturn nil, nil, fmt.Errorf(\"task with ID %q already started\", cfg.ID)\n\t}\n\n\tvar driverConfig TaskConfig\n\tif err := cfg.DecodeDriverConfig(&driverConfig); err != nil {\n\t\treturn nil, nil, fmt.Errorf(\"failed to decode driver config: %v\", err)\n\t}\n\n\thandle := drivers.NewTaskHandle(taskHandleVersion)\n\thandle.Config = cfg\n\n\tif driverConfig.Image == \"\" {\n\t\treturn nil, nil, fmt.Errorf(\"image name required\")\n\t}\n\n\tcreateOpts := api.SpecGenerator{}\n\tcreateOpts.ContainerBasicConfig.LogConfiguration = &api.LogConfig{}\n\tallArgs := []string{}\n\tif driverConfig.Command != \"\" {\n\t\tallArgs = append(allArgs, driverConfig.Command)\n\t}\n\tallArgs = append(allArgs, driverConfig.Args...)\n\n\tif driverConfig.Entrypoint != \"\" {\n\t\tcreateOpts.ContainerBasicConfig.Entrypoint = append(createOpts.ContainerBasicConfig.Entrypoint, driverConfig.Entrypoint)\n\t}\n\n\tcontainerName := BuildContainerName(cfg)\n\n\t// ensure to include port_map into tasks environment map\n\tcfg.Env = taskenv.SetPortMapEnvs(cfg.Env, driverConfig.PortMap)\n\n\t// Basic config options\n\tcreateOpts.ContainerBasicConfig.Name = containerName\n\tcreateOpts.ContainerBasicConfig.Command = allArgs\n\tcreateOpts.ContainerBasicConfig.Env = cfg.Env\n\tcreateOpts.ContainerBasicConfig.Hostname = driverConfig.Hostname\n\tcreateOpts.ContainerBasicConfig.Sysctl = driverConfig.Sysctl\n\n\tcreateOpts.ContainerBasicConfig.LogConfiguration.Path = cfg.StdoutPath\n\n\t// Storage config options\n\tcreateOpts.ContainerStorageConfig.Init = driverConfig.Init\n\tcreateOpts.ContainerStorageConfig.Image = driverConfig.Image\n\tcreateOpts.ContainerStorageConfig.InitPath = driverConfig.InitPath\n\tcreateOpts.ContainerStorageConfig.WorkDir = driverConfig.WorkingDir\n\tallMounts, err := d.containerMounts(cfg, &driverConfig)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tcreateOpts.ContainerStorageConfig.Mounts = allMounts\n\n\t// Resources config options\n\tcreateOpts.ContainerResourceConfig.ResourceLimits = &spec.LinuxResources{\n\t\tMemory: &spec.LinuxMemory{},\n\t\tCPU: &spec.LinuxCPU{},\n\t}\n\tif driverConfig.MemoryReservation != \"\" {\n\t\treservation, err := memoryInBytes(driverConfig.MemoryReservation)\n\t\tif err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t\tcreateOpts.ContainerResourceConfig.ResourceLimits.Memory.Reservation = &reservation\n\t}\n\n\tif cfg.Resources.NomadResources.Memory.MemoryMB > 0 {\n\t\tlimit := cfg.Resources.NomadResources.Memory.MemoryMB * 1024 * 1024\n\t\tcreateOpts.ContainerResourceConfig.ResourceLimits.Memory.Limit = &limit\n\t}\n\tif driverConfig.MemorySwap != \"\" {\n\t\tswap, err := memoryInBytes(driverConfig.MemorySwap)\n\t\tif err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t\tcreateOpts.ContainerResourceConfig.ResourceLimits.Memory.Swap = &swap\n\t}\n\tif !d.cgroupV2 {\n\t\tswappiness := uint64(driverConfig.MemorySwappiness)\n\t\tcreateOpts.ContainerResourceConfig.ResourceLimits.Memory.Swappiness = &swappiness\n\t}\n\t// FIXME: can fail for nonRoot due to missing cpu limit delegation permissions,\n\t// see https://github.com/containers/podman/blob/master/troubleshooting.md\n\tif !d.systemInfo.Host.Rootless {\n\t\tcpuShares := uint64(cfg.Resources.LinuxResources.CPUShares)\n\t\tcreateOpts.ContainerResourceConfig.ResourceLimits.CPU.Shares = &cpuShares\n\t}\n\n\t// Security config options\n\tcreateOpts.ContainerSecurityConfig.CapAdd = driverConfig.CapAdd\n\tcreateOpts.ContainerSecurityConfig.CapDrop = driverConfig.CapDrop\n\tcreateOpts.ContainerSecurityConfig.User = cfg.User\n\n\t// Network config options\n\tfor _, strdns := range driverConfig.Dns {\n\t\tipdns := net.ParseIP(strdns)\n\t\tif ipdns == nil {\n\t\t\treturn nil, nil, fmt.Errorf(\"Invald dns server address\")\n\t\t}\n\t\tcreateOpts.ContainerNetworkConfig.DNSServers = append(createOpts.ContainerNetworkConfig.DNSServers, ipdns)\n\t}\n\t// Configure network\n\tif cfg.NetworkIsolation != nil && cfg.NetworkIsolation.Path != \"\" {\n\t\tcreateOpts.ContainerNetworkConfig.NetNS.NSMode = api.Path\n\t\tcreateOpts.ContainerNetworkConfig.NetNS.Value = cfg.NetworkIsolation.Path\n\t} else {\n\t\tif driverConfig.NetworkMode == \"\" {\n\t\t\tcreateOpts.ContainerNetworkConfig.NetNS.NSMode = api.Bridge\n\t\t} else if driverConfig.NetworkMode == \"bridge\" {\n\t\t\tcreateOpts.ContainerNetworkConfig.NetNS.NSMode = api.Bridge\n\t\t} else if driverConfig.NetworkMode == \"host\" {\n\t\t\tcreateOpts.ContainerNetworkConfig.NetNS.NSMode = api.Host\n\t\t} else if driverConfig.NetworkMode == \"none\" {\n\t\t\tcreateOpts.ContainerNetworkConfig.NetNS.NSMode = api.NoNetwork\n\t\t} else if driverConfig.NetworkMode == \"slirp4netns\" {\n\t\t\tcreateOpts.ContainerNetworkConfig.NetNS.NSMode = api.Slirp\n\t\t} else {\n\t\t\treturn nil, nil, fmt.Errorf(\"Unknown/Unsupported network mode: %s\", driverConfig.NetworkMode)\n\t\t}\n\t}\n\n\tportMappings, err := d.portMappings(cfg, driverConfig)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tcreateOpts.ContainerNetworkConfig.PortMappings = portMappings\n\n\tcontainerID := \"\"\n\trecoverRunningContainer := false\n\t// check if there is a container with same name\n\totherContainerInspect, err := d.podman.ContainerInspect(d.ctx, containerName)\n\tif err == nil {\n\t\t// ok, seems we found a container with similar name\n\t\tif otherContainerInspect.State.Running {\n\t\t\t// it's still running. So let's use it instead of creating a new one\n\t\t\td.logger.Info(\"Detect running container with same name, we reuse it\", \"task\", cfg.ID, \"container\", otherContainerInspect.ID)\n\t\t\tcontainerID = otherContainerInspect.ID\n\t\t\trecoverRunningContainer = true\n\t\t} else {\n\t\t\t// let's remove the old, dead container\n\t\t\td.logger.Info(\"Detect stopped container with same name, removing it\", \"task\", cfg.ID, \"container\", otherContainerInspect.ID)\n\t\t\tif err = d.podman.ContainerDelete(d.ctx, otherContainerInspect.ID, true, true); err != nil {\n\t\t\t\treturn nil, nil, nstructs.WrapRecoverable(fmt.Sprintf(\"failed to remove dead container: %v\", err), err)\n\t\t\t}\n\t\t}\n\t}\n\n\tif !recoverRunningContainer {\n\t\t// FIXME: there are more variations of image sources, we should handle it\n\t\t// e.g. oci-archive:/... etc\n\t\t// see also https://github.com/hashicorp/nomad-driver-podman/issues/69\n\t\t// do we already have this image in local storage?\n\t\thaveImage, err := d.podman.ImageExists(d.ctx, createOpts.Image)\n\t\tif err != nil {\n\t\t\treturn nil, nil, fmt.Errorf(\"failed to start task, unable to check for local image: %v\", err)\n\t\t}\n\t\tif !haveImage {\n\t\t\t// image is not in local storage, so we need to pull it\n\t\t\tif err = d.podman.ImagePull(d.ctx, createOpts.Image); err != nil {\n\t\t\t\treturn nil, nil, fmt.Errorf(\"failed to start task, unable to pull image %s: %v\", createOpts.Image, err)\n\t\t\t}\n\t\t}\n\n\t\tcreateResponse, err := d.podman.ContainerCreate(d.ctx, createOpts)\n\t\tfor _, w := range createResponse.Warnings {\n\t\t\td.logger.Warn(\"Create Warning\", \"warning\", w)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn nil, nil, fmt.Errorf(\"failed to start task, could not create container: %v\", err)\n\t\t}\n\t\tcontainerID = createResponse.Id\n\t}\n\n\tcleanup := func() {\n\t\td.logger.Debug(\"Cleaning up\", \"container\", containerID)\n\t\tif err := d.podman.ContainerDelete(d.ctx, containerID, true, true); err != nil {\n\t\t\td.logger.Error(\"failed to clean up from an error in Start\", \"error\", err)\n\t\t}\n\t}\n\n\tif !recoverRunningContainer {\n\t\tif err = d.podman.ContainerStart(d.ctx, containerID); err != nil {\n\t\t\tcleanup()\n\t\t\treturn nil, nil, fmt.Errorf(\"failed to start task, could not start container: %v\", err)\n\t\t}\n\t}\n\n\tinspectData, err := d.podman.ContainerInspect(d.ctx, containerID)\n\tif err != nil {\n\t\td.logger.Error(\"failed to inspect container\", \"err\", err)\n\t\tcleanup()\n\t\treturn nil, nil, fmt.Errorf(\"failed to start task, could not inspect container : %v\", err)\n\t}\n\n\tnet := &drivers.DriverNetwork{\n\t\tPortMap: driverConfig.PortMap,\n\t\tIP: inspectData.NetworkSettings.IPAddress,\n\t\tAutoAdvertise: true,\n\t}\n\n\th := &TaskHandle{\n\t\tcontainerID: containerID,\n\t\tdriver: d,\n\t\ttaskConfig: cfg,\n\t\tprocState: drivers.TaskStateRunning,\n\t\texitResult: &drivers.ExitResult{},\n\t\tstartedAt: time.Now().Round(time.Millisecond),\n\t\tlogger: d.logger.Named(\"podmanHandle\"),\n\n\t\ttotalCPUStats: stats.NewCpuStats(),\n\t\tuserCPUStats: stats.NewCpuStats(),\n\t\tsystemCPUStats: stats.NewCpuStats(),\n\n\t\tremoveContainerOnExit: d.config.GC.Container,\n\t}\n\n\tdriverState := TaskState{\n\t\tContainerID: containerID,\n\t\tTaskConfig: cfg,\n\t\tStartedAt: h.startedAt,\n\t\tNet: net,\n\t}\n\n\tif err := handle.SetDriverState(&driverState); err != nil {\n\t\td.logger.Error(\"failed to start task, error setting driver state\", \"error\", err)\n\t\tcleanup()\n\t\treturn nil, nil, fmt.Errorf(\"failed to set driver state: %v\", err)\n\t}\n\n\td.tasks.Set(cfg.ID, h)\n\n\tgo h.runContainerMonitor()\n\n\td.logger.Info(\"Completely started container\", \"taskID\", cfg.ID, \"container\", containerID, \"ip\", inspectData.NetworkSettings.IPAddress)\n\n\treturn handle, net, nil\n}", "func (operation *Operation) GetTaskConduit() chan Messager {\n\treturn operation.conduit\n}", "func (o AppV2Output) Wait() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v *AppV2) pulumi.BoolPtrOutput { return v.Wait }).(pulumi.BoolPtrOutput)\n}", "func (*FakeReconcilerClient) GetTask(string) (swarm.Task, error) {\n\treturn swarm.Task{}, FakeUnimplemented\n}", "func (p SourceProvider) TaskDone(t *provider.Task) error {\n\tt.LastRunAt = time.Now()\n\tt.Running = false\n\tt.LastError = \"\"\n\tt.CurrentRetryCount = 0\n\tif !p.Config.Enabled {\n\t\treturn nil\n\t}\n\tif p.Connection.KAPI == nil {\n\t\tif err := p.Connection.Connect(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif err := p.Connection.WriteTask(t); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *TowerClient) taskRejected(task *backupTask, curStatus reserveStatus) {\n\tswitch curStatus {\n\n\t// The sessionQueue has available capacity but the task was rejected,\n\t// this indicates that the task was ineligible for backup.\n\tcase reserveAvailable:\n\t\tc.stats.taskIneligible()\n\n\t\tlog.Infof(\"Backup chanid=%s commit-height=%d is ineligible\",\n\t\t\ttask.id.ChanID, task.id.CommitHeight)\n\n\t\terr := c.cfg.DB.MarkBackupIneligible(\n\t\t\ttask.id.ChanID, task.id.CommitHeight,\n\t\t)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Unable to mark task chanid=%s \"+\n\t\t\t\t\"commit-height=%d ineligible: %v\",\n\t\t\t\ttask.id.ChanID, task.id.CommitHeight, err)\n\n\t\t\t// It is safe to not handle this error, even if we could\n\t\t\t// not persist the result. At worst, this task may be\n\t\t\t// reprocessed on a subsequent start up, and will either\n\t\t\t// succeed do a change in session parameters or fail in\n\t\t\t// the same manner.\n\t\t}\n\n\t\t// If this task was rejected *and* the session had available\n\t\t// capacity, we discard anything held in the prevTask. Either it\n\t\t// was nil before, or is the task which was just rejected.\n\t\tc.prevTask = nil\n\n\t// The sessionQueue rejected the task because it is full, we will stash\n\t// this task and try to add it to the next available sessionQueue.\n\tcase reserveExhausted:\n\t\tc.stats.sessionExhausted()\n\n\t\tlog.Debugf(\"Session %s exhausted, backup chanid=%s \"+\n\t\t\t\"commit-height=%d queued for next session\",\n\t\t\tc.sessionQueue.ID(), task.id.ChanID,\n\t\t\ttask.id.CommitHeight)\n\n\t\t// Cache the task that we pulled off, so that we can process it\n\t\t// once a new session queue is available.\n\t\tc.sessionQueue = nil\n\t\tc.prevTask = task\n\t}\n}", "func WaitTasks(taskBroker *schedule.TaskBroker, run *models.Run) error {\n\tlogrus.Info(\"Begin monitoring task execution ...\")\n\n\tch, err := taskBroker.GetChannel()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tjobName := run.Details[common.KeyJobName]\n\tpodListOpt := metav1.ListOptions{LabelSelector: fmt.Sprintf(\"job-name=%s\", jobName)}\n\tapi := clientset.CoreV1()\n\n\tfor {\n\t\ttime.Sleep(interval)\n\n\t\tqueue, err := ch.QueueInspect(jobName)\n\t\tif err != nil {\n\t\t\tlogrus.Info(\"The queue doesn't exist. All tasks have been executed.\")\n\t\t\tbreak\n\t\t}\n\t\tlogrus.Infof(\"Queue: messages %d.\", queue.Messages)\n\n\t\tif queue.Messages != 0 {\n\t\t\t// there are tasks to be run\n\t\t\tcontinue\n\t\t}\n\n\t\t// the number of the message in the queue is zero. make sure all the\n\t\t// pods in this job have finished\n\t\tpodList, err := api.Pods(namespace).List(podListOpt)\n\t\tif err != nil {\n\t\t\tlogrus.Warnf(\"Fail to list pod of %s: %s\", jobName, err)\n\t\t\tcontinue\n\t\t}\n\n\t\trunningPods := 0\n\t\tfor _, pod := range podList.Items {\n\t\t\tif pod.Status.Phase == corev1.PodRunning {\n\t\t\t\trunningPods++\n\t\t\t}\n\t\t}\n\n\t\tif runningPods != 0 {\n\t\t\tlogrus.Infof(\"%d pod are still running.\", runningPods)\n\t\t\tcontinue\n\t\t}\n\n\t\t// zero task in the queue and all pod stop.\n\t\tbreak\n\t}\n\n\treturn nil\n}", "func (client *Client) DescribeSyncEcsHostTaskWithChan(request *DescribeSyncEcsHostTaskRequest) (<-chan *DescribeSyncEcsHostTaskResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeSyncEcsHostTaskResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeSyncEcsHostTask(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (d *dispatcher) monitorTask(taskID int64) (finished bool, subTaskErrs []error) {\n\t// TODO: Consider putting the following operations into a transaction.\n\tvar err error\n\td.task, err = d.taskMgr.GetGlobalTaskByID(taskID)\n\tif err != nil {\n\t\tlogutil.BgLogger().Error(\"check task failed\", zap.Int64(\"task ID\", d.task.ID), zap.Error(err))\n\t\treturn false, nil\n\t}\n\tswitch d.task.State {\n\tcase proto.TaskStateCancelling:\n\t\treturn false, []error{errors.New(\"cancel\")}\n\tcase proto.TaskStateReverting:\n\t\tcnt, err := d.taskMgr.GetSubtaskInStatesCnt(d.task.ID, proto.TaskStateRevertPending, proto.TaskStateReverting)\n\t\tif err != nil {\n\t\t\tlogutil.BgLogger().Warn(\"check task failed\", zap.Int64(\"task ID\", d.task.ID), zap.Error(err))\n\t\t\treturn false, nil\n\t\t}\n\t\treturn cnt == 0, nil\n\tdefault:\n\t\tsubTaskErrs, err = d.taskMgr.CollectSubTaskError(d.task.ID)\n\t\tif err != nil {\n\t\t\tlogutil.BgLogger().Warn(\"collect subtask error failed\", zap.Int64(\"task ID\", d.task.ID), zap.Error(err))\n\t\t\treturn false, nil\n\t\t}\n\t\tif len(subTaskErrs) > 0 {\n\t\t\treturn false, subTaskErrs\n\t\t}\n\t\t// check subtasks pending or running.\n\t\tcnt, err := d.taskMgr.GetSubtaskInStatesCnt(d.task.ID, proto.TaskStatePending, proto.TaskStateRunning)\n\t\tif err != nil {\n\t\t\tlogutil.BgLogger().Warn(\"check task failed\", zap.Int64(\"task ID\", d.task.ID), zap.Error(err))\n\t\t\treturn false, nil\n\t\t}\n\t\treturn cnt == 0, nil\n\t}\n}", "func (p *process) Wait(ctx context.Context) error {\n\tselect {\n\tcase err, ok := <-p.waitC:\n\t\t// Process exited\n\t\tif ok {\n\t\t\treturn err\n\t\t}\n\t\treturn errWaitAlreadyCalled\n\tcase <-ctx.Done():\n\t\t// Timed out. Send a kill signal and release our handle to it.\n\t\treturn multierr.Combine(ctx.Err(), p.cmd.Process.Kill())\n\t}\n}", "func (n *Node) Wait() (int, error) {\n\tctx := context.TODO()\n\n\tclient, err := client.NewEnvClient()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn client.ContainerWait(ctx, n.id)\n}", "func (c *lockBased) Wait(ctx context.Context) error {\n\tc.mux.RLock()\n\tdefer c.mux.RUnlock()\n\treturn ctx.Err()\n}", "func (s *testSignaler) wait() bool {\n\tselect {\n\tcase s := <-s.nonBlockingStatus:\n\t\treturn s\n\tcase s := <-s.status:\n\t\treturn s\n\t}\n}", "func (c *EC2) WaitUntilExportTaskCancelledWithContext(ctx aws.Context, input *DescribeExportTasksInput, opts ...request.WaiterOption) error {\n\tw := request.Waiter{\n\t\tName: \"WaitUntilExportTaskCancelled\",\n\t\tMaxAttempts: 40,\n\t\tDelay: request.ConstantWaiterDelay(15 * time.Second),\n\t\tAcceptors: []request.WaiterAcceptor{\n\t\t\t{\n\t\t\t\tState: request.SuccessWaiterState,\n\t\t\t\tMatcher: request.PathAllWaiterMatch, Argument: \"ExportTasks[].State\",\n\t\t\t\tExpected: \"cancelled\",\n\t\t\t},\n\t\t},\n\t\tLogger: c.Config.Logger,\n\t\tNewRequest: func(opts []request.Option) (*request.Request, error) {\n\t\t\tvar inCpy *DescribeExportTasksInput\n\t\t\tif input != nil {\n\t\t\t\ttmp := *input\n\t\t\t\tinCpy = &tmp\n\t\t\t}\n\t\t\treq, _ := c.DescribeExportTasksRequest(inCpy)\n\t\t\treq.SetContext(ctx)\n\t\t\treq.ApplyOptions(opts...)\n\t\t\treturn req, nil\n\t\t},\n\t}\n\tw.ApplyOptions(opts...)\n\n\treturn w.WaitWithContext(ctx)\n}", "func (m *Manager) Wait(ctx context.Context, uuid string) error {\n\t// Find the workflow.\n\trw, err := m.runningWorkflow(uuid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Just wait for it.\n\tselect {\n\tcase <-rw.done:\n\t\tbreak\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\t}\n\treturn nil\n}", "func (db *DynamoDB) CancelTask(ctx context.Context, req *tes.CancelTaskRequest) (*tes.CancelTaskResponse, error) {\n\n\t// call GetTask prior to cancel to ensure that the task exists\n\tt, err := db.GetTask(ctx, &tes.GetTaskRequest{Id: req.Id, View: tes.TaskView_MINIMAL})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tswitch t.GetState() {\n\tcase tes.State_COMPLETE, tes.State_EXECUTOR_ERROR, tes.State_SYSTEM_ERROR:\n\t\terr = fmt.Errorf(\"illegal state transition from %s to %s\", t.GetState().String(), tes.State_CANCELED.String())\n\t\treturn nil, fmt.Errorf(\"cannot cancel task: %s\", err)\n\tcase tes.State_CANCELED:\n\t\treturn &tes.CancelTaskResponse{}, nil\n\t}\n\n\titem := &dynamodb.UpdateItemInput{\n\t\tTableName: aws.String(db.taskTable),\n\t\tKey: map[string]*dynamodb.AttributeValue{\n\t\t\tdb.partitionKey: {\n\t\t\t\tS: aws.String(db.partitionValue),\n\t\t\t},\n\t\t\t\"id\": {\n\t\t\t\tS: aws.String(req.Id),\n\t\t\t},\n\t\t},\n\t\tExpressionAttributeNames: map[string]*string{\n\t\t\t\"#state\": aws.String(\"state\"),\n\t\t},\n\t\tUpdateExpression: aws.String(\"SET #state = :to\"),\n\t\tExpressionAttributeValues: map[string]*dynamodb.AttributeValue{\n\t\t\t\":to\": {\n\t\t\t\tN: aws.String(strconv.Itoa(int(tes.State_CANCELED))),\n\t\t\t},\n\t\t},\n\t\tReturnValues: aws.String(\"UPDATED_NEW\"),\n\t}\n\n\t_, err = db.client.UpdateItemWithContext(ctx, item)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &tes.CancelTaskResponse{}, nil\n}", "func (s *Service) waitChan(cmd *exec.Cmd) <-chan waitResult {\n\tvar wr waitResult\n\tc := make(chan waitResult, 1)\n\tgo func() {\n\t\twr.ps, wr.err = cmd.Process.Wait()\n\t\tc <- wr\n\t}()\n\treturn c\n}", "func (t *Transport) Wait() <-chan error {\n\tc := make(chan error, 1)\n\tc <- errors.New(\"TODO\")\n\treturn c\n}", "func (_Contract *ContractSession) GetTask(i *big.Int) (struct {\n\tActive bool\n\tAssignment *big.Int\n\tProposalID *big.Int\n}, error) {\n\treturn _Contract.Contract.GetTask(&_Contract.CallOpts, i)\n}", "func (p *Pairing) Wait(ctx context.Context) (tornet.RemoteKeyRing, error) {\n\tdefer p.peerset.Close()\n\tif p.server != nil {\n\t\tdefer p.server.Close()\n\t}\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn tornet.RemoteKeyRing{}, errors.New(\"context cancelled\")\n\tcase <-p.finished:\n\t\tif p.failure != nil {\n\t\t\treturn tornet.RemoteKeyRing{}, p.failure\n\t\t}\n\t\treturn p.peer, nil\n\t}\n}", "func Wait(exitChannel chan error, cancel context.CancelFunc) error {\n\terr := <-exitChannel\n\t// cancel the context\n\tcancel()\n\treturn err\n}", "func (g *Group) Wait() error {\n\tif v := atomic.SwapUint32(&g.waiting, 1); v != 0 {\n\t\tpanic(\"schedgroup: multiple calls to Group.Wait\")\n\t}\n\n\t// Context cancelation takes priority.\n\tif err := g.ctx.Err(); err != nil {\n\t\treturn err\n\t}\n\n\t// See if the task heap is already empty. If so, we can exit early.\n\tg.mu.Lock()\n\tif g.tasks.Len() == 0 {\n\t\t// Release the mutex immediately so that any running jobs are able to\n\t\t// complete and send on g.lenC.\n\t\tg.mu.Unlock()\n\t\tg.cancel()\n\t\tg.wg.Wait()\n\t\treturn nil\n\t}\n\tg.mu.Unlock()\n\n\t// Wait on context cancelation or for the number of items in the heap\n\t// to reach 0.\n\tvar n int\n\tfor {\n\t\tselect {\n\t\tcase <-g.ctx.Done():\n\t\t\treturn g.ctx.Err()\n\t\tcase n = <-g.lenC:\n\t\t\t// Context cancelation takes priority.\n\t\t\tif err := g.ctx.Err(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t\tif n == 0 {\n\t\t\t// No more tasks left, cancel the monitor goroutine and wait for\n\t\t\t// all tasks to complete.\n\t\t\tg.cancel()\n\t\t\tg.wg.Wait()\n\t\t\treturn nil\n\t\t}\n\t}\n}", "func (cmd *Command) Wait() error {\n\t// According to https://github.com/golang/go/issues/28461,\n\t// exec.Cmd#Wait is not thread-safe, so we need to implement\n\t// our own version.\n\tcmd.waitOnce.Do(func() {\n\t\tcmd.waitResult = cmd.c.Wait()\n\t\tclose(cmd.waitDoneCh)\n\t})\n\treturn cmd.waitResult\n}", "func (r *Robot) FailTask(name string, args ...string) RetVal {\n\treturn r.pipeTask(flavorFail, typeTask, name, args...)\n}", "func (e *engine) Wait(context.Context, *Spec, *Step) (*State, error) {\n\treturn nil, nil // no-op for bash implementation\n}", "func (m *Manager) PauseWait() <-chan struct{} {\n\ts := <-m.status\n\tch := m.pauseStart\n\tm.status <- s\n\treturn ch\n}", "func (t *Transaction) WaitContext(ctx context.Context) (tx *interfaces.TxWithBlock, err error) {\n\tselect {\n\tcase tx = <-t.Transaction(): // wait for the tx\n\tcase err2, ok := <-t.Err(): // wait for an error\n\t\tif !ok { // in case the channel is closed, then return status.ErrShutdown\n\t\t\terr = status.ErrShutdown\n\t\t\treturn\n\t\t}\n\t\terr = err2\n\tcase <-ctx.Done(): // wait for context cancellation\n\t\terr = ctx.Err() // TODO should we close the instance?\n\t}\n\treturn\n}", "func (tm *Manager) WaitForPendingTasks() {\n\ttm.wg.Wait()\n}", "func (sb *shardBuffer) wait(ctx context.Context, e *entry) (RetryDoneFunc, error) {\n\tselect {\n\tcase <-ctx.Done():\n\t\tsb.remove(e)\n\t\treturn nil, vterrors.Errorf(vterrors.Code(contextCanceledError), \"%v: %v\", contextCanceledError, ctx.Err())\n\tcase <-e.done:\n\t\treturn e.bufferCancel, e.err\n\t}\n}", "func (t *Task) Stop() error {\n\tselect {\n\tcase <-t.startedChan:\n\tdefault:\n\t\treturn ErrNotExecuting\n\t}\n\n\tt.cancelCtx()\n\n\treturn nil\n}", "func (r *Runsc) Wait(context context.Context, id string) (int, error) {\n\tdata, stderr, err := cmdOutput(r.command(context, \"wait\", id), false)\n\tif err != nil {\n\t\treturn 0, fmt.Errorf(\"%w: %s\", err, stderr)\n\t}\n\tvar res waitResult\n\tif err := json.Unmarshal(data, &res); err != nil {\n\t\treturn 0, err\n\t}\n\treturn res.ExitStatus, nil\n}", "func (op *DeleteTensorboardRunOperation) Wait(ctx context.Context, opts ...gax.CallOption) error {\n\topts = append([]gax.CallOption{gax.WithPath(op.pollPath)}, opts...)\n\treturn op.lro.WaitWithInterval(ctx, nil, time.Minute, opts...)\n}", "func newTask(opts taskOptions) *task {\n\tctx, cancel := context.WithCancel(opts.Context)\n\n\tt := &task{\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t\texited: make(chan struct{}),\n\t}\n\n\tgo func() {\n\t\tdefer opts.OnDone()\n\t\tdefer close(t.exited)\n\t\t_ = opts.Runnable.Run(t.ctx)\n\t}()\n\treturn t\n}", "func (d *dispatcher) scheduleTask(taskID int64) {\n\tticker := time.NewTicker(checkTaskFinishedInterval)\n\tdefer ticker.Stop()\n\tfor {\n\t\tselect {\n\t\tcase <-d.ctx.Done():\n\t\t\tlogutil.BgLogger().Info(\"schedule task exits\", zap.Int64(\"task ID\", taskID), zap.Error(d.ctx.Err()))\n\t\t\treturn\n\t\tcase <-ticker.C:\n\t\t\tstepIsFinished, errs := d.monitorTask(taskID)\n\t\t\tfailpoint.Inject(\"cancelTaskAfterMonitorTask\", func(val failpoint.Value) {\n\t\t\t\tif val.(bool) && d.task.State == proto.TaskStateRunning {\n\t\t\t\t\terr := d.taskMgr.CancelGlobalTask(taskID)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlogutil.BgLogger().Error(\"cancel task failed\", zap.Error(err))\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t})\n\t\t\t// The global task isn't finished and not failed.\n\t\t\tif !stepIsFinished && len(errs) == 0 {\n\t\t\t\tGetTaskFlowHandle(d.task.Type).OnTicker(d.ctx, d.task)\n\t\t\t\tlogutil.BgLogger().Debug(\"schedule task, this task keeps current state\",\n\t\t\t\t\tzap.Int64(\"task-id\", d.task.ID), zap.String(\"state\", d.task.State))\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\terr := d.processFlow(d.task, errs)\n\t\t\tif err == nil && d.task.IsFinished() {\n\t\t\t\tlogutil.BgLogger().Info(\"schedule task, task is finished\",\n\t\t\t\t\tzap.Int64(\"task-id\", d.task.ID), zap.String(\"state\", d.task.State))\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\tfailpoint.Inject(\"mockOwnerChange\", func(val failpoint.Value) {\n\t\t\tif val.(bool) {\n\t\t\t\tlogutil.BgLogger().Info(\"mockOwnerChange called\")\n\t\t\t\tMockOwnerChange()\n\t\t\t\ttime.Sleep(time.Second)\n\t\t\t}\n\t\t})\n\t}\n}", "func (mgr *ClientMgr) onTaskEnd(ctx context.Context, client *Client, task *Task,\n\terr error, reply *jarviscrawlercore.ReplyCrawler, endChan chan int) {\n\n\tif err != nil {\n\t\tif task.Logger != nil {\n\t\t\ttask.Logger.Warn(\"onTaskEnd: error\",\n\t\t\t\tzap.Error(err),\n\t\t\t\tzap.String(\"servaddr\", client.servAddr),\n\t\t\t\tJSON(\"task\", task))\n\t\t}\n\n\t\t// if !(strings.Index(err.Error(), \"Error: noretry:\") == 0 ||\n\t\t// \tstrings.Index(err.Error(), \"noretry:\") == 0) {\n\t\tif !IsNoRetryError(err) {\n\n\t\t\tif task.RetryNums > 0 {\n\t\t\t\ttask.RetryNums--\n\n\t\t\t\t// time.Sleep(time.Second * time.Duration(mgr.cfg.SleepTime))\n\n\t\t\t\ttask.Running = false\n\t\t\t\tclient.Running = false\n\t\t\t\tendChan <- 0\n\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// task.Fail = true\n\t\t\t// task.running = false\n\t\t}\n\n\t\ttask.Fail = true\n\t}\n\n\tgo task.Callback(ctx, task, err, reply)\n\n\t// time.Sleep(time.Second * time.Duration(mgr.cfg.SleepTime))\n\n\ttask.Running = false\n\tclient.Running = false\n\tendChan <- task.TaskID\n}", "func TestTaskContext(t *testing.T) {\n\tt.Run(\"reflect external context\", func(t *testing.T) {\n\t\tctx, cancel := context.WithCancel(context.Background())\n\t\ttask := Start(ctx, TaskFuncDoneErr)\n\n\t\tnotDone(task)\n\t\teq(nil, task.Err())\n\n\t\tcancel()\n\t\twaitDone(task)\n\t\teq(true, errors.Is(task.Err(), context.Canceled))\n\t})\n\n\tt.Run(\"wait on external context, override error\", func(t *testing.T) {\n\t\tsentinel := fmt.Errorf(`sentinel`)\n\n\t\tfun := func(ctx Task) error {\n\t\t\twaitDone(ctx)\n\t\t\treturn sentinel\n\t\t}\n\n\t\tctx, cancel := context.WithCancel(context.Background())\n\t\ttask := Start(ctx, fun)\n\n\t\tnotDone(task)\n\t\teq(nil, task.Err())\n\n\t\tcancel()\n\t\twaitDone(task)\n\t\teq(true, errors.Is(task.Err(), sentinel))\n\t})\n}" ]
[ "0.68873113", "0.6370824", "0.6234808", "0.6090943", "0.60882926", "0.60077477", "0.59856975", "0.5946728", "0.57896346", "0.5758922", "0.5743035", "0.57364565", "0.5708013", "0.56723464", "0.5585439", "0.5556589", "0.54844946", "0.54625577", "0.5353345", "0.5346748", "0.5334011", "0.52867043", "0.5280479", "0.52784675", "0.523096", "0.5144518", "0.51311857", "0.5104822", "0.5084476", "0.50472355", "0.5033292", "0.5029276", "0.4999667", "0.49661478", "0.49404478", "0.49349874", "0.49334714", "0.49268", "0.49101368", "0.49028888", "0.49000618", "0.4899081", "0.4892145", "0.4877008", "0.48608527", "0.48507208", "0.48488528", "0.48448238", "0.4840617", "0.48394042", "0.48359224", "0.48334652", "0.4790766", "0.47819018", "0.47799274", "0.47621638", "0.4749033", "0.47453642", "0.4743298", "0.4742895", "0.47344464", "0.47295034", "0.47117302", "0.47039446", "0.47012135", "0.47007248", "0.46943593", "0.46943185", "0.46771812", "0.4665864", "0.46579868", "0.4657349", "0.46568346", "0.4641224", "0.46397376", "0.4639023", "0.46325785", "0.4628944", "0.4627311", "0.4598577", "0.45903948", "0.45880347", "0.4587894", "0.45809475", "0.45763236", "0.45717672", "0.4570821", "0.4570056", "0.45693114", "0.45683238", "0.45489994", "0.45472232", "0.45470476", "0.45453542", "0.4542372", "0.4537725", "0.45297983", "0.45289338", "0.45275784", "0.45212483" ]
0.76350826
0
StopTask function is expected to stop a running task by sending the given signal to it. If the task does not stop during the given timeout, the driver must forcefully kill the task. StopTask does not clean up resources of the task or remove it from the driver's internal state.
func (d *Driver) StopTask(taskID string, timeout time.Duration, signal string) error { d.logger.Info("Stopping task", "taskID", taskID, "signal", signal) handle, ok := d.tasks.Get(taskID) if !ok { return drivers.ErrTaskNotFound } // fixme send proper signal to container err := d.podman.ContainerStop(d.ctx, handle.containerID, int(timeout.Seconds())) if err != nil { d.logger.Error("Could not stop/kill container", "containerID", handle.containerID, "err", err) return err } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (t *Task) StopWithTimeout(timeout time.Duration) error {\n\tdefer func() { recover() }()\n\tclose(*t.Stopper)\n\n\tselect {\n\tcase <-*t.Done:\n\t\treturn nil\n\tcase <-time.After(timeout):\n\t\treturn goerr.Wrap(&ErrStoppingTaskTimeout{})\n\t}\n}", "func (h *Hub) StopTask(ctx context.Context, request *pb.StopTaskRequest) (*pb.StopTaskReply, error) {\n\tlog.G(h.ctx).Info(\"handling StopTask request\", zap.Any(\"req\", request))\n\ttaskID := request.Id\n\tminerID, ok := h.getMinerByTaskID(taskID)\n\tif !ok {\n\t\treturn nil, status.Errorf(codes.NotFound, \"no such task %s\", taskID)\n\t}\n\n\tminer, ok := h.getMinerByID(minerID)\n\tif !ok {\n\t\treturn nil, status.Errorf(codes.NotFound, \"no miner with task %s\", minerID)\n\t}\n\n\t_, err := miner.Client.Stop(ctx, &pb.StopTaskRequest{Id: taskID})\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.NotFound, \"failed to stop the task %s\", taskID)\n\t}\n\n\tminer.deregisterRoute(taskID)\n\tminer.Retain(taskID)\n\n\th.deleteTaskByID(taskID)\n\n\treturn &pb.StopTaskReply{}, nil\n}", "func (s *Session) Stop() error {\n\tif s.task == nil {\n\t\treturn errors.New(\"snap task not running or not found\")\n\t}\n\n\trs := s.pClient.StopTask(s.task.ID)\n\tif rs.Err != nil {\n\t\treturn errors.Wrapf(rs.Err, \"could not send stop signal to task %q\", s.task.ID)\n\t}\n\n\terr := s.waitForStop()\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"could not stop task %q\", s.task.ID)\n\t}\n\n\trr := s.pClient.RemoveTask(s.task.ID)\n\tif rr.Err != nil {\n\t\treturn errors.Wrapf(rr.Err, \"could not remove task %q\", s.task.ID)\n\t}\n\n\ts.task = nil\n\n\treturn nil\n}", "func (c *BasicECSClient) StopTask(ctx context.Context, in *ecs.StopTaskInput) (*ecs.StopTaskOutput, error) {\n\tif err := c.setup(); err != nil {\n\t\treturn nil, errors.Wrap(err, \"setting up client\")\n\t}\n\n\tvar out *ecs.StopTaskOutput\n\tvar err error\n\tmsg := awsutil.MakeAPILogMessage(\"StopTask\", in)\n\tif err := utility.Retry(ctx,\n\t\tfunc() (bool, error) {\n\t\t\tout, err = c.ecs.StopTaskWithContext(ctx, in)\n\t\t\tif awsErr, ok := err.(awserr.Error); ok {\n\t\t\t\tgrip.Debug(message.WrapError(awsErr, msg))\n\t\t\t\tif c.isNonRetryableErrorCode(awsErr.Code()) {\n\t\t\t\t\treturn false, err\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn true, err\n\t\t}, *c.opts.RetryOpts); err != nil {\n\t\treturn nil, err\n\t}\n\treturn out, nil\n}", "func (d *Driver) SignalTask(taskID string, signal string) error {\n\thandle, ok := d.tasks.Get(taskID)\n\tif !ok {\n\t\treturn drivers.ErrTaskNotFound\n\t}\n\n\treturn d.podman.ContainerKill(d.ctx, handle.containerID, signal)\n}", "func (c Control) ServeStopTask(w http.ResponseWriter, r *http.Request) {\n\tc.ServeTaskAction(w, r, false)\n}", "func (t *Task) Stop() error {\n\tselect {\n\tcase <-t.startedChan:\n\tdefault:\n\t\treturn ErrNotExecuting\n\t}\n\n\tt.cancelCtx()\n\n\treturn nil\n}", "func (e *ECS) StopTask(req *StopTaskReq) (*StopTaskResp, error) {\n\tif req == nil {\n\t\treturn nil, fmt.Errorf(\"The req params cannot be nil\")\n\t}\n\n\tparams := makeParams(\"StopTask\")\n\tif req.Cluster != \"\" {\n\t\tparams[\"cluster\"] = req.Cluster\n\t}\n\tif req.Task != \"\" {\n\t\tparams[\"task\"] = req.Task\n\t}\n\n\tresp := new(StopTaskResp)\n\tif err := e.query(params, resp); err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func (c *Client) TerminateTask(guid string) error {\n\treq := c.NewRequest(\"PUT\", fmt.Sprintf(\"/v3/tasks/%s/cancel\", guid))\n\tresp, err := c.DoRequest(req)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Error terminating task\")\n\t}\n\tdefer resp.Body.Close()\n\n\tif resp.StatusCode != 202 {\n\t\treturn errors.Wrapf(err, \"Failed terminating task, response status code %d\", resp.StatusCode)\n\t}\n\treturn nil\n}", "func (c *ECS) StopTask(input *StopTaskInput) (output *StopTaskOutput, err error) {\n\treq, out := c.StopTaskRequest(input)\n\toutput = out\n\terr = req.Send()\n\treturn\n}", "func (t *Task) Stop() {\n\tdefer func() { recover() }()\n\tclose(*t.Stopper)\n\t<-*t.Done\n}", "func HandleStopTask(w http.ResponseWriter, r *http.Request) {\n\tlog.Root.Info(\"HandleStopTask BEGIN\")\n\n\tif r.Method != http.MethodPost {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tHttpResponseError(w, ErrNotFound)\n\t\treturn\n\t}\n\n\tdefer r.Body.Close()\n\tbody, _ := ioutil.ReadAll(r.Body)\n\n\tdata := make(map[string]interface{})\n\terr := json.Unmarshal(body, &data)\n\tif err != nil {\n\t\tlog.Root.Error(\"HandleStopTask Parse HTTP request body error\")\n\t\tHttpResponseError(w, ErrForm)\n\t\treturn\n\t}\n\n\telem, ok := data[\"taskID\"]\n\tif !ok {\n\t\tlog.Root.Error(\"HandleStopTask HTTP form data error\")\n\t\tHttpResponseError(w, ErrForm)\n\t\treturn\n\t}\n\n\ttaskID := elem.(string)\n\ttaskCapacity, err := node.StopTask(taskID)\n\tif err != nil {\n\t\tlog.Root.Error(\"HandleStopTask Stop task error. TaskID: %v\", taskID)\n\t\tHttpResponseError(w, ErrServer)\n\t\treturn\n\t}\n\n\tlog.Root.Info(\"HandleStopTask END\")\n\tHttpResponseData(w, H{\n\t\t\"taskCapacity\": taskCapacity,\n\t})\n\treturn\n}", "func (k *KubernetesExecutor) KillTask(driver bindings.ExecutorDriver, taskId *mesos.TaskID) {\n\tif k.isDone() {\n\t\treturn\n\t}\n\tlog.Infof(\"Kill task %v\\n\", taskId)\n\n\tif !k.isConnected() {\n\t\t//TODO(jdefelice) sent TASK_LOST here?\n\t\tlog.Warningf(\"Ignore kill task because the executor is disconnected\\n\")\n\t\treturn\n\t}\n\n\tk.lock.Lock()\n\tdefer k.lock.Unlock()\n\tk.killPodForTask(driver, taskId.GetValue(), messages.TaskKilled)\n}", "func (k *KubernetesExecutor) KillTask(driver bindings.ExecutorDriver, taskId *mesos.TaskID) {\n\tif k.isDone() {\n\t\treturn\n\t}\n\tlog.Infof(\"Kill task %v\\n\", taskId)\n\n\tif !k.isConnected() {\n\t\t//TODO(jdefelice) sent TASK_LOST here?\n\t\tlog.Warningf(\"Ignore kill task because the executor is disconnected\\n\")\n\t\treturn\n\t}\n\n\tk.lock.Lock()\n\tdefer k.lock.Unlock()\n\tk.removePodTask(driver, taskId.GetValue(), messages.TaskKilled, mesos.TaskState_TASK_KILLED)\n}", "func (t *TaskController[T, U, C, CT, TF]) Stop() {\n\tclose(t.productExitCh)\n\t// Clear all the task in the task queue and mark all task complete.\n\t// so that ```t.Wait``` is able to close resultCh\n\tfor range t.inputCh {\n\t\tt.wg.Done()\n\t}\n\tt.pool.StopTask(t.TaskID())\n\t// Clear the resultCh to avoid blocking the consumer put result into the channel and cannot exit.\n\tchannel.Clear(t.resultCh)\n}", "func (c *Cmd) StopWithTimeout(to time.Duration) error {\n\texited := make(chan bool)\n\tgo func() {\n\t\tc.Wait()\n\t\texited <- true\n\t}()\n\n\tps, err := c.expandToChildProcesses(c.Process)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t//signal all (child)processes to better simulate\n\t//shell behaviour\n\tfor _, p := range ps {\n\t\terr := p.Signal(os.Interrupt)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tselect {\n\tcase <-exited:\n\t\treturn nil //process exited by itself\n\tcase <-time.After(to):\n\t\tfor _, p := range ps {\n\t\t\terr := p.Kill()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t}\n}", "func (f *Failer) KillTask(host, task string) error {\n\tscript := \"sudo pkill -x %s\"\n\tlog.V(1).Infof(\"Killing task %s on host %s\", task, host)\n\treturn f.runWithEvilTag(host, fmt.Sprintf(script, task))\n}", "func (t *Task) StopAndWait(timeout time.Duration) (TaskResult, error) {\n\terr := t.Stop()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tres, err := t.Wait(timeout)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn res, nil\n}", "func (r *subdomainTask) Stop() error {\n\tclose(r.done)\n\tr.queue = queue.NewQueue()\n\treturn nil\n}", "func (o *DeleteLTENetworkIDNetworkProbeTasksTaskIDParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func StopTimeout(t time.Duration) Option {\n\treturn func(o *options) { o.stopTimeout = t }\n}", "func (m TaskManager) AbortTask(c context.Context, ctl task.Controller) error {\n\treturn nil\n}", "func (m TaskManager) AbortTask(c context.Context, ctl task.Controller) error {\n\treturn nil\n}", "func (t *task) Kill(_ context.Context, signal syscall.Signal) error {\n\thcsContainer, err := t.getHCSContainer()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlogger := t.ctr.client.logger.WithFields(log.Fields{\n\t\t\"container\": t.ctr.id,\n\t\t\"process\": t.id,\n\t\t\"pid\": t.Pid(),\n\t\t\"signal\": signal,\n\t})\n\tlogger.Debug(\"Signal()\")\n\n\tvar op string\n\tif signal == syscall.SIGKILL {\n\t\t// Terminate the compute system\n\t\tt.ctr.mu.Lock()\n\t\tt.ctr.terminateInvoked = true\n\t\tt.ctr.mu.Unlock()\n\t\top, err = \"terminate\", hcsContainer.Terminate()\n\t} else {\n\t\t// Shut down the container\n\t\top, err = \"shutdown\", hcsContainer.Shutdown()\n\t}\n\tif err != nil {\n\t\tif !hcsshim.IsPending(err) && !hcsshim.IsAlreadyStopped(err) {\n\t\t\t// ignore errors\n\t\t\tlogger.WithError(err).Errorf(\"failed to %s hccshim container\", op)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (c *ECS) StopTaskRequest(input *StopTaskInput) (req *aws.Request, output *StopTaskOutput) {\n\toprw.Lock()\n\tdefer oprw.Unlock()\n\n\tif opStopTask == nil {\n\t\topStopTask = &aws.Operation{\n\t\t\tName: \"StopTask\",\n\t\t\tHTTPMethod: \"POST\",\n\t\t\tHTTPPath: \"/\",\n\t\t}\n\t}\n\n\treq = c.newRequest(opStopTask, input, output)\n\toutput = &StopTaskOutput{}\n\treq.Data = output\n\treturn\n}", "func KillTask(tid int) Errno {\n\t_, e := internal.Syscall1(KILLTASK, uintptr(tid))\n\treturn Errno(e)\n}", "func (md *ManagementNode) StopTaskDeadTimeout(ctx context.Context, id string) error {\n\n\tmd.scheduledTasksMtx.RLock()\n\tdefer md.scheduledTasksMtx.RUnlock()\n\n\ttask, ok := md.scheduledTasks[id]\n\tif !ok {\n\t\treturn trace.Errorf(\"There is no task %s in the scheduled task storage\", id)\n\t}\n\n\ttask.StopDeadTimeout(ctx)\n\n\treturn nil\n\n}", "func (timeout *Timeout) Stop() {\n\ttimeout.state = Inactive\n\ttimeout.start = time.Now()\n}", "func cleanupTask(ctx context.Context, t *testing.T, c cocoa.ECSClient, runOut *ecs.RunTaskOutput) {\n\tif runOut != nil && len(runOut.Tasks) > 0 && runOut.Tasks[0].TaskArn != nil {\n\t\tout, err := c.StopTask(ctx, &ecs.StopTaskInput{\n\t\t\tCluster: aws.String(testutil.ECSClusterName()),\n\t\t\tTask: aws.String(*runOut.Tasks[0].TaskArn),\n\t\t})\n\t\trequire.NoError(t, err)\n\t\trequire.NotZero(t, out)\n\t}\n}", "func (t *DeferredRecordingTaskImpl) Stop(blocking bool) error {\n\treturn t.task.Stop(blocking)\n}", "func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error {\n\ttr.logger.Trace(\"Kill requested\")\n\n\t// Cancel the task runner to break out of restart delay or the main run\n\t// loop.\n\ttr.killCtxCancel()\n\n\t// Emit kill event\n\tif event != nil {\n\t\ttr.logger.Trace(\"Kill event\", \"event_type\", event.Type, \"event_reason\", event.KillReason)\n\t\ttr.EmitEvent(event)\n\t}\n\n\tselect {\n\tcase <-tr.WaitCh():\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\t}\n\n\treturn tr.getKillErr()\n}", "func (b *B) StopTimer()", "func (m *Mock) StopPlugin(name string, cancelFlag task.CancelFlag) (err error) {\n\treturn nil\n}", "func (w *WaitTask) Cancel(_ *TaskContext) {\n\tw.cancelFunc()\n}", "func (t *TimeTask) DeleteTask(task *RawTask) {\n\tt.deleteChan <- task\n}", "func (d *Driver) DestroyTask(taskID string, force bool) error {\n\thandle, ok := d.tasks.Get(taskID)\n\tif !ok {\n\t\treturn drivers.ErrTaskNotFound\n\t}\n\n\tif handle.isRunning() && !force {\n\t\treturn fmt.Errorf(\"cannot destroy running task\")\n\t}\n\n\tif handle.isRunning() {\n\t\td.logger.Debug(\"Have to destroyTask but container is still running\", \"containerID\", handle.containerID)\n\t\t// we can not do anything, so catching the error is useless\n\t\terr := d.podman.ContainerStop(d.ctx, handle.containerID, 60)\n\t\tif err != nil {\n\t\t\td.logger.Warn(\"failed to stop/kill container during destroy\", \"error\", err)\n\t\t}\n\t\t// wait a while for stats emitter to collect exit code etc.\n\t\tfor i := 0; i < 20; i++ {\n\t\t\tif !handle.isRunning() {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\ttime.Sleep(time.Millisecond * 250)\n\t\t}\n\t\tif handle.isRunning() {\n\t\t\td.logger.Warn(\"stats emitter did not exit while stop/kill container during destroy\", \"error\", err)\n\t\t}\n\t}\n\n\tif handle.removeContainerOnExit {\n\t\terr := d.podman.ContainerDelete(d.ctx, handle.containerID, true, true)\n\t\tif err != nil {\n\t\t\td.logger.Warn(\"Could not remove container\", \"container\", handle.containerID, \"error\", err)\n\t\t}\n\t}\n\n\td.tasks.Delete(taskID)\n\treturn nil\n}", "func (s *Syncthing) Stop() error {\n\tdefer s.cmd.Process.Wait() // nolint: errcheck\n\n\treturn s.cmd.Process.Signal(os.Interrupt)\n}", "func (d *Driver) Stop() error {\n\tif err := d.verifyRootPermissions(); err != nil {\n\t\treturn err\n\t}\n\n\ts, err := d.GetState()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif s != state.Stopped {\n\t\terr := d.sendSignal(syscall.SIGTERM)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"hyperkit sigterm failed\")\n\t\t}\n\t\t// wait 120s for graceful shutdown\n\t\tfor i := 0; i < 60; i++ {\n\t\t\ttime.Sleep(2 * time.Second)\n\t\t\ts, _ := d.GetState()\n\t\t\tlog.Debugf(\"VM state: %s\", s)\n\t\t\tif s == state.Stopped {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t\treturn errors.New(\"VM Failed to gracefully shutdown, try the kill command\")\n\t}\n\treturn nil\n}", "func (c *Command) Stop() error {\n\tif c.cmd.Process == nil {\n\t\t// not yet started\n\t\treturn nil\n\t}\n\n\tpid := c.cmd.Process.Pid\n\tif c.timeout > 0 {\n\t\t// stop listening on os.Interrupt signal\n\t\tsignal.Stop(c.interruptChan)\n\t\t// stop the timeout timer\n\t\tc.timeoutTimer.Stop()\n\n\t\t// use the negative process group id, to kill the whole process group\n\t\tpgid, err := syscall.Getpgid(c.cmd.Process.Pid)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tpid = -1 * pgid\n\t}\n\n\t// kill the process\n\treturn syscall.Kill(pid, syscall.SIGKILL)\n}", "func (c *Client) CancelTask(ctx context.Context, taskID string) error {\n\tctx, cf := context.WithTimeout(ctx, 60*time.Second)\n\tdefer cf()\n\tvar tc *swarming_api.SwarmingRpcsCancelResponse\n\tgetResult := func() error {\n\t\tvar err error\n\t\treq := &swarming_api.SwarmingRpcsTaskCancelRequest{\n\t\t\tKillRunning: true,\n\t\t}\n\t\ttc, err = c.SwarmingService.Task.Cancel(taskID, req).Context(ctx).Do()\n\t\treturn err\n\t}\n\tif err := callWithRetries(ctx, getResult); err != nil {\n\t\treturn errors.Annotate(err, fmt.Sprintf(\"cancel task %s\", taskID)).Err()\n\t}\n\tif !tc.Ok {\n\t\treturn errors.New(fmt.Sprintf(\"task %s is not successfully canceled\", taskID))\n\t}\n\treturn nil\n}", "func (t *Task) StopDeadTimeout(ctx context.Context) {\n\tc, cancel := context.WithCancel(ctx)\n\tdefer cancel()\n\n\tselect {\n\tcase t.rxed <- struct{}{}:\n\t\t{\n\t\t\tlog.Debugf(\"Dead timeout is stopped for task %s\",\n\t\t\t\tt.task.Id)\n\t\t}\n\tcase <-c.Done():\n\t\t{\n\t\t\tlog.Warningf(\"context declined for task %s\", t.task.Id)\n\t\t}\n\t}\n}", "func execStop(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tsignal.Stop(args[0].(chan<- os.Signal))\n}", "func (b Bot) CancelTask(taskID string) error {\n\ttask, err := b.repository.GetTask(taskID)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif task.executed {\n\t\treturn fmt.Errorf(\"task %s has already been executed\", taskID)\n\t}\n\n\ttask.cancelled = true\n\tb.repository.UpdateTask(task)\n\treturn nil\n}", "func (k *KubernetesExecutor) reportLostTask(driver bindings.ExecutorDriver, tid, reason string) {\n\tk.removePodTask(driver, tid, reason, mesos.TaskState_TASK_LOST)\n}", "func (k *KubernetesExecutor) reportLostTask(driver bindings.ExecutorDriver, tid, reason string) {\n\tk.removePodTask(driver, tid, reason, mesos.TaskState_TASK_LOST)\n}", "func SignalStopEmission(instance *Object, signalId uint32, detail glib.Quark) {\n\tc_instance := (C.gpointer)(C.NULL)\n\tif instance != nil {\n\t\tc_instance = (C.gpointer)(instance.ToC())\n\t}\n\n\tc_signal_id := (C.guint)(signalId)\n\n\tc_detail := (C.GQuark)(detail)\n\n\tC.g_signal_stop_emission(c_instance, c_signal_id, c_detail)\n\n\treturn\n}", "func runTask(bFunc taskFunc, tFunc taskFunc, name string, interval time.Duration) {\n\tif err := bFunc(name); err != nil {\n\t\tglog.Errorf(\"%s: %s\", name, err)\n\t}\n\tfor {\n\t\tselect {\n\t\tcase <-killTask:\n\t\t\tglog.V(vvLevel).Infof(\"Exiting %s\", name)\n\t\t\theartbeatWG.Done()\n\t\t\treturn\n\t\tdefault:\n\t\t\tif err := tFunc(name); err != nil {\n\t\t\t\tglog.Errorf(\"%s: %s\", name, err)\n\t\t\t}\n\t\t\ttime.Sleep(interval)\n\t\t}\n\t}\n}", "func (NilTimer) Stop() {}", "func (sys Systemd) Stop(unit string) error {\n\tns := fmt.Sprintf(\"project_%s_%s\", sys.p.ID, sys.kind)\n\ttarget := ns + \"_\" + unit\n\n\tif err := exec.Command(\"systemctl\", \"stop\", target).Run(); err != nil {\n\t\treturn fmt.Errorf(\"failed to stop systemd unit %s: %s\", target, err)\n\t}\n\treturn nil\n}", "func (o *AutoscaleStopInstancesByCrnParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (vm *VirtualMachine) Stop(args *DomainXML, reply *bool) error {\n\t// Passing the true parameter to ensure the stop vm task is added to waitgroup as this action needs to be completed\n\t// even if there is pending signal termination on rpc\n\t_, err := proc.AddTask(true)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rpc/server:Stop() Could not add task for vm stop\")\n\t}\n\tdefer proc.TaskDone()\n\n\tlog.Trace(\"rpc/server:Stop() Entering\")\n\tdefer log.Trace(\"rpc/server:Stop() Leaving\")\n\n\twlaMtx.Lock()\n\tdefer wlaMtx.Unlock()\n\n\tif err = validation.ValidateXMLString(args.XML); err != nil {\n\t\tsecLog.Errorf(\"rpc:server() Stop: %s, Invalid domain XML format\", message.InvalidInputBadParam)\n\t\treturn nil\n\t}\n\n\t// pass in vm.Watcher to get the instance to the File System Watcher\n\t*reply = wlavm.Stop(args.XML, vm.Watcher)\n\treturn nil\n}", "func (p *GoshPool) StopWait() {\n\tp.stop(true)\n}", "func (_m *MockApiClient) CancelTask(id string) error {\n\tret := _m.Called(id)\n\n\tvar r0 error\n\tif rf, ok := ret.Get(0).(func(string) error); ok {\n\t\tr0 = rf(id)\n\t} else {\n\t\tr0 = ret.Error(0)\n\t}\n\n\treturn r0\n}", "func (t Task) Close() error {\n\tpath := fmt.Sprintf(\"tasks/%d/close\", t.ID)\n\t_, err := makeRequest(http.MethodPost, path, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *DeleteLTENetworkIDNetworkProbeTasksTaskIDParams) WithTimeout(timeout time.Duration) *DeleteLTENetworkIDNetworkProbeTasksTaskIDParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (i *DeleteOrUpdateInvTask) Cancel(_ *taskrunner.TaskContext) {}", "func (o *StopGatewayBundleUsingPOSTParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (tw *TimeWheel) removeTask(key interface{}) {\n\t// get the timer slot\n\tposition, ok := tw.timer[key]\n\tif !ok {\n\t\treturn\n\t}\n\t// get slot list\n\tl := tw.slots[position]\n\tfor e := l.Front(); e != nil; {\n\t\ttask := e.Value.(*Task)\n\t\tif task.key == key {\n\t\t\tdelete(tw.timer, task.key)\n\t\t\tl.Remove(e)\n\t\t}\n\n\t\te = e.Next()\n\t}\n}", "func (c *hostCommunicator) EndTask(ctx context.Context, detail *apimodels.TaskEndDetail, taskData TaskData) (*apimodels.EndTaskResponse, error) {\n\tgrip.Info(message.Fields{\n\t\t\"message\": \"started EndTask\",\n\t\t\"task_id\": taskData.ID,\n\t})\n\ttaskEndResp := &apimodels.EndTaskResponse{}\n\tinfo := requestInfo{\n\t\tmethod: http.MethodPost,\n\t\ttaskData: &taskData,\n\t\tpath: fmt.Sprintf(\"hosts/%s/task/%s/end\", c.hostID, taskData.ID),\n\t}\n\tresp, err := c.retryRequest(ctx, info, detail)\n\tif err != nil {\n\t\treturn nil, util.RespErrorf(resp, errors.Wrap(err, \"ending task\").Error())\n\t}\n\tdefer resp.Body.Close()\n\tif err = utility.ReadJSON(resp.Body, taskEndResp); err != nil {\n\t\treturn nil, errors.Wrap(err, \"reading end task reply from response\")\n\t}\n\tgrip.Info(message.Fields{\n\t\t\"message\": \"finished EndTask\",\n\t\t\"task_id\": taskData.ID,\n\t})\n\treturn taskEndResp, nil\n}", "func (o *CleanUpTasksDeprecatedParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (p *PruneTask) ClearTimeout() {}", "func (a API) StopWait(cmd *None) (out *None, e error) {\n\tRPCHandlers[\"stop\"].Call <-API{a.Ch, cmd, nil}\n\tselect {\n\tcase <-time.After(time.Second*5):\n\t\tbreak\n\tcase o := <-a.Ch.(chan StopRes):\n\t\tout, e = o.Res, o.Err\n\t}\n\treturn\n}", "func (t *Task) Cancel() {\n log.Warnf(\"Cancelling task and all subsequent runs\")\n\n // Clear queue of subsequent runs\n t.runs.Clear()\n}", "func TrackTaskEnd() {\n\tpersist.DeleteValue(taskPayloadKey)\n\tpersist.DeleteValue(taskEndTimeKey)\n}", "func TestLibvirtDriver_Start_Stop_Recover_Task(t *testing.T) {\n\tutil.RequireLibvirt(t)\n\n\trequire := require.New(t)\n\n\td := NewLibvirtDriver(testlog.HCLogger(t)).(*Driver)\n\tharness := dtestutil.NewDriverHarness(t, d)\n\n\trequire.NoError(task.EncodeConcreteDriverConfig(&taskCfg))\n\n\tcleanup := harness.MkAllocDir(task, false)\n\tdefer cleanup()\n\n\t// Test start task\n\thandle, _, err := harness.StartTask(task)\n\trequire.NoError(err)\n\trequire.NotNil(handle)\n\n\tlibvirtHandle, ok := d.tasks.Get(task.ID)\n\trequire.NotNil(libvirtHandle)\n\trequire.True(ok)\n\n\ttestutil.WaitForResult(func() (bool, error) {\n\t\tstatus, err := d.InspectTask(task.ID)\n\t\trequire.NoError(err)\n\t\tif status.State == drivers.TaskStateRunning {\n\t\t\treturn true, nil\n\t\t}\n\t\treturn false, fmt.Errorf(\"task in state: %v\", status.State)\n\t}, func(err error) {\n\t\tt.Fatalf(\"task failed to start: %v\", err)\n\t})\n\n\t// Missing the task handle\n\td.tasks.Delete(task.ID)\n\n\t// Test recover the missed task\n\trecoverHandle := handle.Copy()\n\trequire.NoError(d.RecoverTask(recoverHandle))\n\n\td.StopTask(task.ID, 5*time.Second, \"kill\")\n\n\t// Destroy the task/vm after test\n\tdefer d.DestroyTask(task.ID, false)\n\n\t// Test after recovery and stop task\n\ttestutil.WaitForResult(func() (bool, error) {\n\t\tstatus, err := d.InspectTask(task.ID)\n\t\trequire.NoError(err)\n\t\tif status.State == drivers.TaskStateExited {\n\t\t\treturn true, nil\n\t\t}\n\t\treturn false, fmt.Errorf(\"task in state: %v\", status.State)\n\t}, func(err error) {\n\t\tt.Fatalf(\"task failed to stop: %v\", err)\n\t})\n}", "func (s *T) Stop() {\n\tclose(s.stopCh)\n\ts.wg.Wait()\n}", "func (s *stateMachine) Stop() {\n\tselect {\n\tcase s.cmds <- ReleaseMessage():\n\t\t// Also inform the state machine it should exit since the internal handler\n\t\t// may override the release message causing the task to be unreleaseable.\n\t\ts.stop()\n\tcase <-s.stopped:\n\t\t// Already stopped!\n\t}\n}", "func (b *B) StopTimer() {}", "func (c *Controller) Stop(ctx hive.HookContext) error {\n\tdoneChan := make(chan struct{})\n\tgo func() {\n\t\tc.workerpool.Close()\n\t\tclose(doneChan)\n\t}()\n\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\tcase <-doneChan:\n\t}\n\n\treturn nil\n}", "func (c *Controller) Stop(ctx hive.HookContext) error {\n\tdoneChan := make(chan struct{})\n\tgo func() {\n\t\tc.workerpool.Close()\n\t\tclose(doneChan)\n\t}()\n\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\tcase <-doneChan:\n\t}\n\n\treturn nil\n}", "func (sh *SignalHandler) Stop() {\n\tsh.lock.Lock()\n\tdefer sh.lock.Unlock()\n\n\tif sh.isRunning == true {\n\t\tsh.isRunning = false\n\t}\n}", "func (t *Task) Reject() (interface{}, error) {\n\tpar := map[string]interface{}{\n\t\t\"taskid\": t.taskId,\n\t}\n\treturn t.nc.Exec(\"task.reject\", par)\n}", "func (h *TCPProbe) Stop() {\n\th.Cancel()\n}", "func (s *Slot) Stop() {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\tif s.cancel != nil {\n\t\ts.cancel()\n\t\ts.cancel = nil\n\t}\n}", "func (db *DynamoDB) CancelTask(ctx context.Context, req *tes.CancelTaskRequest) (*tes.CancelTaskResponse, error) {\n\n\t// call GetTask prior to cancel to ensure that the task exists\n\tt, err := db.GetTask(ctx, &tes.GetTaskRequest{Id: req.Id, View: tes.TaskView_MINIMAL})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tswitch t.GetState() {\n\tcase tes.State_COMPLETE, tes.State_EXECUTOR_ERROR, tes.State_SYSTEM_ERROR:\n\t\terr = fmt.Errorf(\"illegal state transition from %s to %s\", t.GetState().String(), tes.State_CANCELED.String())\n\t\treturn nil, fmt.Errorf(\"cannot cancel task: %s\", err)\n\tcase tes.State_CANCELED:\n\t\treturn &tes.CancelTaskResponse{}, nil\n\t}\n\n\titem := &dynamodb.UpdateItemInput{\n\t\tTableName: aws.String(db.taskTable),\n\t\tKey: map[string]*dynamodb.AttributeValue{\n\t\t\tdb.partitionKey: {\n\t\t\t\tS: aws.String(db.partitionValue),\n\t\t\t},\n\t\t\t\"id\": {\n\t\t\t\tS: aws.String(req.Id),\n\t\t\t},\n\t\t},\n\t\tExpressionAttributeNames: map[string]*string{\n\t\t\t\"#state\": aws.String(\"state\"),\n\t\t},\n\t\tUpdateExpression: aws.String(\"SET #state = :to\"),\n\t\tExpressionAttributeValues: map[string]*dynamodb.AttributeValue{\n\t\t\t\":to\": {\n\t\t\t\tN: aws.String(strconv.Itoa(int(tes.State_CANCELED))),\n\t\t\t},\n\t\t},\n\t\tReturnValues: aws.String(\"UPDATED_NEW\"),\n\t}\n\n\t_, err = db.client.UpdateItemWithContext(ctx, item)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &tes.CancelTaskResponse{}, nil\n}", "func (n *SQSNotify) Stop() {\n\tn.running = false\n\t_ = n.flushDeleteQueue0()\n}", "func stopTimeoutHandler(args []string) {\n\tvar exitCode = 0\n\ttimeout, _ := strconv.Atoi(args[0])\n\tif len(args) > 1 {\n\t\texitCode, _ = strconv.Atoi(args[1])\n\t}\n\n\tfmt.Printf(\"will finish in %dms\\n\", timeout)\n\t<-time.After(time.Millisecond * time.Duration(timeout))\n\n\tos.Exit(exitCode)\n}", "func (pomo *Pomo) Stop() {\n\n\tif pomo.Status == ON {\n\t\tpomo.Cancel <- true\n\t\tpomo.Time = DEFAULT_DURATION\n\t}\n\n}", "func (task *Task) Kill() {\n\tif task.Config.CmdString != \"\" && task.Command.Started && !task.Command.Complete {\n\t\tsyscall.Kill(-task.Command.Cmd.Process.Pid, syscall.SIGKILL)\n\t}\n\n\tfor _, subTask := range task.Children {\n\t\tif subTask.Config.CmdString != \"\" && subTask.Command.Started && !subTask.Command.Complete {\n\t\t\tsyscall.Kill(-subTask.Command.Cmd.Process.Pid, syscall.SIGKILL)\n\t\t}\n\t}\n\n}", "func (app *App) Stop(ctx context.Context) error {\n\treturn withTimeout(ctx, app.lifecycle.Stop)\n}", "func (device *SilentStepperBrick) Stop() (err error) {\n\tvar buf bytes.Buffer\n\n\tresultBytes, err := device.device.Set(uint8(FunctionStop), buf.Bytes())\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(resultBytes) > 0 {\n\t\tvar header PacketHeader\n\n\t\theader.FillFromBytes(resultBytes)\n\n\t\tif header.Length != 8 {\n\t\t\treturn fmt.Errorf(\"Received packet of unexpected size %d, instead of %d\", header.Length, 8)\n\t\t}\n\n\t\tif header.ErrorCode != 0 {\n\t\t\treturn DeviceError(header.ErrorCode)\n\t\t}\n\n\t\tbytes.NewBuffer(resultBytes[8:])\n\n\t}\n\n\treturn nil\n}", "func (pool *SubPool) Stop() {\n\tpool.StopChan <- struct{}{}\n}", "func (n *Node) Stop() error {\n\tn.opMx.RLock()\n\tdefer n.opMx.RUnlock()\n\tif n.running {\n\t\tsafeCloseChan(n.close)\n\t\treturn nil\n\t}\n\treturn ErrStopNotStarted\n}", "func (d *Driver) Stop() error {\n\tcs := d.client()\n\t_, err := cs.AsyncRequest(&egoscale.StopVirtualMachine{\n\t\tID: d.ID,\n\t}, d.async)\n\n\treturn err\n}", "func (r *Robot) FailTask(name string, args ...string) RetVal {\n\treturn r.pipeTask(flavorFail, typeTask, name, args...)\n}", "func (sn *SimNode) Stop() error {\n\tsn.lock.Lock()\n\tif sn.client != nil {\n\t\tsn.client.Close()\n\t\tsn.client = nil\n\t}\n\tsn.lock.Unlock()\n\treturn sn.node.Stop()\n}", "func (p *Plugin) Stop() {\n\twr := newWaiter()\n\tp.killCh <- wr\n\twr.wait()\n\tp.exitCh <- struct{}{}\n}", "func (pool *TxPool) Stop() {\n\tclose(pool.close)\n\tpool.wg.Wait()\n\n\tlog.Infof(\"TxPool Stopped\")\n}", "func (t *Tuner) Stop() error {\n\tt.mu.Lock()\n\tdefer t.mu.Unlock()\n\n\terr := t.destroyAnyRunningPipeline()\n\tt.status.Set(Status{Error: err})\n\tt.tracks.Set(Tracks{})\n\treturn err\n}", "func (ep *ExpectProcess) Stop() error {\n\terr := ep.Signal(syscall.SIGTERM)\n\tif err != nil && errors.Is(err, os.ErrProcessDone) {\n\t\treturn nil\n\t}\n\treturn err\n}", "func (c *Client) StopNode(nodeID string) error {\n\treturn c.Post(fmt.Sprintf(\"/nodes/%s/stop\", nodeID), nil, nil)\n}", "func (self *JobPipeline) Stop() {\n\tlog.Printf(\"Stopping %s\", self.TaskName)\n\tfor i := range(self.workers) {\n\t\tself.workers[i].Stop()\n\t}\n\tfor i := range(self.workers) {\n\t\tself.workers[i].WaitForStop()\n\t}\n\tlog.Printf(\"%s Stopped\", self.TaskName)\n}", "func (tw *TopologyWatcher) Stop() {\n\ttw.cancelFunc()\n\t// wait for watch goroutine to finish.\n\ttw.wg.Wait()\n}", "func handleTimeoutNotification(task *task.MessageTask, env *task.Env) {\n\tkey := fmt.Sprintf(\"%x\", task.GetMessage().Token)\n\tdelete(env.Requests(), key)\n\tlog.Info(\"<<< handleTimeout Notification>>>\")\n}", "func (p *Pool) Stop() {\n\tp.lock.Lock()\n\tdefer p.lock.Unlock()\n\tp.cancel()\n\tfor _, routine := range p.routines {\n\t\troutine.stop <- true\n\t}\n\tp.waitGroup.Wait()\n\tfor _, routine := range p.routines {\n\t\tclose(routine.stop)\n\t}\n}", "func (t *TCPTest) Stop() {\n\tt.exit<-struct{}{}\n}", "func (t *Timer) Stop() {\n\tt.tickObj.Stop()\n}", "func (s *Server) Stop(timeout int) error {\n\tctx, cancel := context.WithTimeout(context.Background(), time.Second*time.Duration(timeout))\n\tdefer cancel()\n\treturn s.HTTP.Shutdown(ctx)\n}", "func (s *SignalMonitor) Stop() {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\n\tif s.isOn {\n\t\ts.isOn = false\n\t\ts.offc <- struct{}{}\n\t}\n}", "func (o *StopGatewayBundleUsingPOSTParams) WithTimeout(timeout time.Duration) *StopGatewayBundleUsingPOSTParams {\n\to.SetTimeout(timeout)\n\treturn o\n}" ]
[ "0.6536051", "0.61022145", "0.5902817", "0.57699805", "0.5755604", "0.5728442", "0.56008637", "0.549826", "0.5445834", "0.5424107", "0.5390439", "0.53785425", "0.53028846", "0.5296322", "0.52536076", "0.5202666", "0.5167302", "0.50993377", "0.507832", "0.50076854", "0.49846232", "0.49749953", "0.49749953", "0.49307266", "0.4928855", "0.49279425", "0.48768735", "0.4853322", "0.48379624", "0.4831443", "0.48140404", "0.47923818", "0.47759488", "0.47463197", "0.4722326", "0.47043425", "0.4686091", "0.46621978", "0.4643309", "0.46411923", "0.4629522", "0.46124867", "0.45941493", "0.45401618", "0.45401618", "0.45292267", "0.4518759", "0.4516687", "0.45070264", "0.4507", "0.4490949", "0.44791627", "0.44737315", "0.44702205", "0.44687346", "0.44519174", "0.44498596", "0.4445983", "0.44444862", "0.44358045", "0.4432474", "0.44218045", "0.44203535", "0.4406673", "0.4394331", "0.43890363", "0.4387445", "0.43699595", "0.43683076", "0.43683076", "0.4352323", "0.4350847", "0.43239987", "0.432381", "0.43208435", "0.4317084", "0.43059212", "0.42878434", "0.4281372", "0.42649144", "0.42580542", "0.42521518", "0.4247513", "0.42473304", "0.42426768", "0.42388314", "0.42335173", "0.42331445", "0.42302957", "0.42292923", "0.42260396", "0.42246574", "0.42245987", "0.421451", "0.4212215", "0.42096788", "0.42086366", "0.4208215", "0.42068964", "0.4206015" ]
0.7427215
0
DestroyTask function cleans up and removes a task that has terminated. If force is set to true, the driver must destroy the task even if it is still running.
func (d *Driver) DestroyTask(taskID string, force bool) error { handle, ok := d.tasks.Get(taskID) if !ok { return drivers.ErrTaskNotFound } if handle.isRunning() && !force { return fmt.Errorf("cannot destroy running task") } if handle.isRunning() { d.logger.Debug("Have to destroyTask but container is still running", "containerID", handle.containerID) // we can not do anything, so catching the error is useless err := d.podman.ContainerStop(d.ctx, handle.containerID, 60) if err != nil { d.logger.Warn("failed to stop/kill container during destroy", "error", err) } // wait a while for stats emitter to collect exit code etc. for i := 0; i < 20; i++ { if !handle.isRunning() { break } time.Sleep(time.Millisecond * 250) } if handle.isRunning() { d.logger.Warn("stats emitter did not exit while stop/kill container during destroy", "error", err) } } if handle.removeContainerOnExit { err := d.podman.ContainerDelete(d.ctx, handle.containerID, true, true) if err != nil { d.logger.Warn("Could not remove container", "container", handle.containerID, "error", err) } } d.tasks.Delete(taskID) return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func CleanTask() {\n\tfor taskID, t := range kv.DefaultClient.GetStorage().Tasks {\n\t\tflag := true\n\t\tfor nid := range kv.DefaultClient.GetStorage().Nodes {\n\t\t\tif t.NodeID == nid {\n\t\t\t\tflag = false\n\t\t\t}\n\t\t}\n\t\tif flag {\n\t\t\tif t.Timer {\n\t\t\t\tlog.Info(\"clean timer:\", taskID)\n\t\t\t\tormTimer := new(orm.Timer)\n\t\t\t\tormTimer.ID = taskID\n\t\t\t\tormTimer.Status = false\n\t\t\t\terr := orm.UpdateTimerStatus(ormTimer)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error(err)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tlog.Info(\"clean task:\", taskID)\n\t\t\t\tormTask := new(orm.Task)\n\t\t\t\tormTask.ID = taskID\n\t\t\t\tormTask.Status = \"error\"\n\t\t\t\terr := orm.UpdateTask(ormTask)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error(err)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tkv.DefaultClient.DeleteTask(taskID)\n\t\t}\n\t}\n}", "func (t *task) deleteTask() {\n\t// There is no state to clean up as of now.\n\t// If the goal state was set to DELETED, then let the\n\t// listeners know that the task has been deleted.\n\n\tvar runtimeCopy *pbtask.RuntimeInfo\n\tvar labelsCopy []*peloton.Label\n\n\t// notify listeners after dropping the lock\n\tdefer func() {\n\t\tif runtimeCopy != nil {\n\t\t\tt.jobFactory.notifyTaskRuntimeChanged(\n\t\t\t\tt.jobID,\n\t\t\t\tt.id,\n\t\t\t\tt.jobType,\n\t\t\t\truntimeCopy,\n\t\t\t\tlabelsCopy,\n\t\t\t)\n\t\t}\n\t}()\n\n\tt.RLock()\n\tdefer t.RUnlock()\n\n\tif t.runtime == nil {\n\t\treturn\n\t}\n\n\tif t.runtime.GetGoalState() != pbtask.TaskState_DELETED {\n\t\treturn\n\t}\n\n\truntimeCopy = proto.Clone(t.runtime).(*pbtask.RuntimeInfo)\n\truntimeCopy.State = pbtask.TaskState_DELETED\n\tlabelsCopy = t.copyLabelsInCache()\n}", "func cleanupTask(ctx context.Context, t *testing.T, c cocoa.ECSClient, runOut *ecs.RunTaskOutput) {\n\tif runOut != nil && len(runOut.Tasks) > 0 && runOut.Tasks[0].TaskArn != nil {\n\t\tout, err := c.StopTask(ctx, &ecs.StopTaskInput{\n\t\t\tCluster: aws.String(testutil.ECSClusterName()),\n\t\t\tTask: aws.String(*runOut.Tasks[0].TaskArn),\n\t\t})\n\t\trequire.NoError(t, err)\n\t\trequire.NotZero(t, out)\n\t}\n}", "func (t *TimeTask) DeleteTask(task *RawTask) {\n\tt.deleteChan <- task\n}", "func (task SchTask) Delete(taskname string, own, force bool) string {\n\tcmd := &exec.Cmd{}\n\n\tif Debug {\n\t\treturn dbgMessage\n\t}\n\n\tif own {\n\t\ttaskname = task.prefix + taskname\n\t}\n\n\tif !force {\n\t\tcmd = exec.Command(task.bin, _Delete.Command, _Delete.taskname, taskname)\n\t} else {\n\t\tcmd = exec.Command(task.bin, _Delete.Command, _Delete.taskname, taskname, _Delete.force)\n\t}\n\n\toutput, err := cmd.CombinedOutput()\n\tcatch(output, err)\n\n\treturn string(output)\n}", "func (task *Task) Kill() {\n\tif task.Config.CmdString != \"\" && task.Command.Started && !task.Command.Complete {\n\t\tsyscall.Kill(-task.Command.Cmd.Process.Pid, syscall.SIGKILL)\n\t}\n\n\tfor _, subTask := range task.Children {\n\t\tif subTask.Config.CmdString != \"\" && subTask.Command.Started && !subTask.Command.Complete {\n\t\t\tsyscall.Kill(-subTask.Command.Cmd.Process.Pid, syscall.SIGKILL)\n\t\t}\n\t}\n\n}", "func DeleteTask(w http.ResponseWriter, r *http.Request, _ httprouter.Params) {\n\tfmt.Fprint(w, \"DeleteTask\\n\")\n}", "func RemoveForced(in io.Reader, out io.Writer, args []string, server DeleteServer) error {\n\tpArg, err := strconv.Atoi(args[0])\n\tif err != nil {\n\t\treturn err\n\t}\n\tt, err := GetTaskAtPosition(server, pArg-1)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = server.DeleteTask(t)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfmt.Fprintln(out, \"Removed tasks:\")\n\tfmt.Fprintln(out, fmt.Sprintf(\"%d%s %s \", t.Order, \"X\", t.Title))\n\n\treturn nil\n}", "func cleanupTaskDefinition(ctx context.Context, t *testing.T, c cocoa.ECSClient, out *ecs.RegisterTaskDefinitionOutput) {\n\tif out != nil && out.TaskDefinition != nil && out.TaskDefinition.TaskDefinitionArn != nil {\n\t\tout, err := c.DeregisterTaskDefinition(ctx, &ecs.DeregisterTaskDefinitionInput{\n\t\t\tTaskDefinition: out.TaskDefinition.TaskDefinitionArn,\n\t\t})\n\t\trequire.NoError(t, err)\n\t\trequire.NotZero(t, out)\n\t}\n}", "func (na *cnmNetworkAllocator) DeallocateTask(t *api.Task) error {\n\tdelete(na.tasks, t.ID)\n\treturn na.releaseEndpoints(t.Networks)\n}", "func (md *ManagementNode) DelTask(id string) {\n\tmd.scheduledTasksMtx.Lock()\n\tdefer md.scheduledTasksMtx.Unlock()\n\n\tdelete(md.scheduledTasks, id)\n\n}", "func (r *TaskRepository) DeleteTask(id int64) error{\n\tvar task, _ = r.GetTask(id)\n\n\t// soft delete, according to doc UpdatedAt will be set automatically\n\tr.DB.Model(&task).Update(\"IsDeleted\", true)\n\n return nil\n}", "func CleanTask() {\n\tvar wg sync.WaitGroup\n\tremove := []string{\n\t\t\"pkg/res/data.go\",\n\t\t\"res/generated/bundle.js\",\n\t\t\"res/generated/style.css\",\n\t\t\"browser/chrome-ext/src/src.zip\",\n\t}\n\tremoveAll := []string{\n\t\t\"dist/\",\n\t\t\"dist-archives/\",\n\t\t\"site/\",\n\t\t\"build/\",\n\t\t\"res/generated/\",\n\t\t\"res/messages/_ref\",\n\t\t\"browser/chrome-ext/src/javascripts\",\n\t\t\"AlkasirChromeExtension/\",\n\t}\n\twg.Add(len(remove))\n\twg.Add(len(removeAll))\n\tfor _, v := range remove {\n\t\tgo func(f string) {\n\t\t\tdefer wg.Done()\n\t\t\tos.Remove(f)\n\t\t}(v)\n\t}\n\tfor _, v := range removeAll {\n\t\tgo func(f string) {\n\t\t\tdefer wg.Done()\n\t\t\tos.RemoveAll(f)\n\t\t}(v)\n\t}\n\twg.Wait()\n}", "func (t TaskService) DeleteTask(ctx context.Context, id platform.ID) error {\n\tspan, _ := tracing.StartSpanFromContext(ctx)\n\tdefer span.Finish()\n\n\treturn t.Client.\n\t\tDelete(taskIDPath(id)).\n\t\tDo(ctx)\n}", "func (f *Failer) KillTask(host, task string) error {\n\tscript := \"sudo pkill -x %s\"\n\tlog.V(1).Infof(\"Killing task %s on host %s\", task, host)\n\treturn f.runWithEvilTag(host, fmt.Sprintf(script, task))\n}", "func (z *zpoolctl) Destroy(ctx context.Context, name string, force bool) *execute {\n\targs := []string{\"destroy\"}\n\tif force {\n\t\targs = append(args, \"-f\")\n\t}\n\targs = append(args, name)\n\treturn &execute{ctx: ctx, name: z.cmd, args: args}\n}", "func (o *DeleteNodeParams) SetForce(force *bool) {\n\to.Force = force\n}", "func (c *Client) DeleteTask(ctx context.Context, in *todopb.TaskQuery, opts ...grpc.CallOption) (*empty.Empty, error) {\n\treturn c.client.DeleteTask(ctx, in, opts...)\n}", "func (t *TaskService) DeleteTask(path string) error {\n\tvar err error\n\n\tif path[0] != '\\\\' {\n\t\treturn errors.New(\"path must start with root folder '\\\\'\")\n\t}\n\n\tif !t.registeredTaskExist(path) {\n\t\treturn errors.New(\"registered task doesn't exist\")\n\t}\n\n\t_, err = oleutil.CallMethod(t.rootFolderObj, \"DeleteTask\", path, 0)\n\tif err != nil {\n\t\tif GetOLEErrorCode(err) == 0x80070005 {\n\t\t\treturn fmt.Errorf(\"error deleting %s task: access is denied\", path)\n\t\t}\n\t\treturn fmt.Errorf(\"error deleting %s task: %s\", path, err)\n\t}\n\n\treturn nil\n}", "func (c *Client) TerminateTask(guid string) error {\n\treq := c.NewRequest(\"PUT\", fmt.Sprintf(\"/v3/tasks/%s/cancel\", guid))\n\tresp, err := c.DoRequest(req)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Error terminating task\")\n\t}\n\tdefer resp.Body.Close()\n\n\tif resp.StatusCode != 202 {\n\t\treturn errors.Wrapf(err, \"Failed terminating task, response status code %d\", resp.StatusCode)\n\t}\n\treturn nil\n}", "func (domain *Domain) DeleteTask(uuid string) error {\n\t// determine task\n\tdomain.TasksX.RLock()\n\t_, ok := domain.Tasks[uuid]\n\tdomain.TasksX.RUnlock()\n\n\tif !ok {\n\t\treturn errors.New(\"task not found\")\n\t}\n\n\t// remove task\n\tdomain.TasksX.Lock()\n\tdelete(domain.Tasks, uuid)\n\tdomain.TasksX.Unlock()\n\n\t// success\n\treturn nil\n}", "func DeleteTask(w http.ResponseWriter, r *http.Request) {\n\tparams := mux.Vars(r)\n\n\tif params[\"id\"] == \"\" {\n\t\thttp.Error(w, http.StatusText(400), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\ttaskID := bson.ObjectIdHex(params[\"id\"])\n\n\tdeleted, err := repository.DeleteTask(taskID)\n\n\tif err != nil {\n\t\thttp.Error(w, http.StatusText(500), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tif !deleted {\n\t\thttp.Error(w, http.StatusText(500), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tw.WriteHeader(http.StatusOK)\n}", "func DeleteTask(id int) error {\n\treturn db.Update(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket(taskBucket)\n\t\treturn b.Delete(itob(id))\n\t})\n}", "func (c Control) ServeDeleteTask(w http.ResponseWriter, r *http.Request) {\n\tid, err := strconv.ParseInt(r.URL.Query().Get(\"id\"), 10, 64)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\tc.Config.Lock()\n\tdefer c.Config.Unlock()\n\tindex, task := c.findTaskById(id)\n\tif task == nil {\n\t\thttp.Error(w, \"Invalid task ID\", http.StatusBadRequest)\n\t\treturn\n\t}\n\ttask.StopLoop()\n\tfor i := index; i < len(c.Config.Tasks)-1; i++ {\n\t\tc.Config.Tasks[i] = c.Config.Tasks[i+1]\n\t}\n\tc.Config.Tasks = c.Config.Tasks[0 : len(c.Config.Tasks)-1]\n\tc.Config.Save()\n\n\thttp.Redirect(w, r, \"/\", http.StatusTemporaryRedirect)\n}", "func (o InstanceOutput) ForceDelete() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v *Instance) pulumi.BoolPtrOutput { return v.ForceDelete }).(pulumi.BoolPtrOutput)\n}", "func (s *Storage) DeleteTask(id uint) error {\n\tif _, err := s.db.Exec(\"DELETE FROM tasks WHERE id=$1\", id); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (e *Endpoints) TaskClean(interval time.Duration) {\n\tl := loop.New(loop.WithInterval(interval))\n\tl.Do(func() (bool, error) {\n\t\ttimeUnix := time.Now().Unix()\n\t\tfmt.Println(timeUnix)\n\n\t\tstartTimestamp := timeUnix - TaskCleanDurationTimestamp\n\n\t\tstartTime := time.Unix(startTimestamp, 0).Format(\"2006-01-02 15:04:05\")\n\n\t\t// clean job resource\n\t\tjobs := e.dbclient.ListExpiredJobs(startTime)\n\n\t\tfor _, job := range jobs {\n\t\t\terr := e.dbclient.DeleteJob(strconv.FormatUint(job.OrgID, 10), job.TaskID)\n\t\t\tif err != nil {\n\t\t\t\terr = e.dbclient.DeleteJob(strconv.FormatUint(job.OrgID, 10), job.TaskID)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogrus.Errorf(\"failed to delete job, job: %+v, (%+v)\", job, err)\n\t\t\t\t}\n\t\t\t}\n\t\t\tlogrus.Debugf(\"[clean] expired job: %+v\", job)\n\t\t}\n\n\t\t// clean deployment resource\n\t\tdeployments := e.dbclient.ListExpiredDeployments(startTime)\n\n\t\tfor _, deployment := range deployments {\n\t\t\terr := e.dbclient.DeleteDeployment(strconv.FormatUint(deployment.OrgID, 10), deployment.TaskID)\n\t\t\tif err != nil {\n\t\t\t\terr = e.dbclient.DeleteDeployment(strconv.FormatUint(deployment.OrgID, 10), deployment.TaskID)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogrus.Errorf(\"failed to delete deployment, deployment: %+v, (%+v)\", deployment, err)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tlogrus.Debugf(\"[clean] expired deployment: %+v\", deployment)\n\t\t}\n\n\t\treturn false, nil\n\t})\n}", "func (k *KubernetesExecutor) KillTask(driver bindings.ExecutorDriver, taskId *mesos.TaskID) {\n\tif k.isDone() {\n\t\treturn\n\t}\n\tlog.Infof(\"Kill task %v\\n\", taskId)\n\n\tif !k.isConnected() {\n\t\t//TODO(jdefelice) sent TASK_LOST here?\n\t\tlog.Warningf(\"Ignore kill task because the executor is disconnected\\n\")\n\t\treturn\n\t}\n\n\tk.lock.Lock()\n\tdefer k.lock.Unlock()\n\tk.killPodForTask(driver, taskId.GetValue(), messages.TaskKilled)\n}", "func (s *K8sSvc) DeleteTask(ctx context.Context, cluster string, service string, taskType string) error {\n\trequuid := utils.GetReqIDFromContext(ctx)\n\n\ttaskID := service + common.NameSeparator + taskType\n\n\terr := s.cliset.BatchV1().Jobs(s.namespace).Delete(taskID, &metav1.DeleteOptions{})\n\tif err != nil {\n\t\tif k8errors.IsNotFound(err) {\n\t\t\tglog.Infoln(\"task not found\", taskID, \"requuid\", requuid)\n\t\t\treturn nil\n\t\t}\n\t\tglog.Errorln(\"delete task error\", err, \"taskID\", taskID, \"requuid\", requuid)\n\t\treturn err\n\t}\n\n\tglog.Infoln(\"deleted task\", taskID, \"requuid\", requuid)\n\treturn nil\n}", "func (d *Driver) StopTask(taskID string, timeout time.Duration, signal string) error {\n\td.logger.Info(\"Stopping task\", \"taskID\", taskID, \"signal\", signal)\n\thandle, ok := d.tasks.Get(taskID)\n\tif !ok {\n\t\treturn drivers.ErrTaskNotFound\n\t}\n\t// fixme send proper signal to container\n\terr := d.podman.ContainerStop(d.ctx, handle.containerID, int(timeout.Seconds()))\n\tif err != nil {\n\t\td.logger.Error(\"Could not stop/kill container\", \"containerID\", handle.containerID, \"err\", err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (k *KubernetesExecutor) KillTask(driver bindings.ExecutorDriver, taskId *mesos.TaskID) {\n\tif k.isDone() {\n\t\treturn\n\t}\n\tlog.Infof(\"Kill task %v\\n\", taskId)\n\n\tif !k.isConnected() {\n\t\t//TODO(jdefelice) sent TASK_LOST here?\n\t\tlog.Warningf(\"Ignore kill task because the executor is disconnected\\n\")\n\t\treturn\n\t}\n\n\tk.lock.Lock()\n\tdefer k.lock.Unlock()\n\tk.removePodTask(driver, taskId.GetValue(), messages.TaskKilled, mesos.TaskState_TASK_KILLED)\n}", "func (pv *Provisioner) removeTaskRef(machine *clusterv1.Machine) error {\n\tnmachine := machine.DeepCopy()\n\tif nmachine.ObjectMeta.Annotations == nil {\n\t\treturn nil\n\t}\n\tdelete(nmachine.ObjectMeta.Annotations, constants.VirtualMachineTaskRef)\n\t_, err := pv.clusterV1alpha1.Machines(nmachine.Namespace).Update(nmachine)\n\treturn err\n}", "func KillTask(tid int) Errno {\n\t_, e := internal.Syscall1(KILLTASK, uintptr(tid))\n\treturn Errno(e)\n}", "func (t *TaskService) Delete(w http.ResponseWriter, r *http.Request) {\n\tvar (\n\t\temptyUUID gocql.UUID\n\t\ttaskIDStr = mux.Vars(r)[\"taskID\"]\n\t\tpartnerID = mux.Vars(r)[\"partnerID\"]\n\t\tctx = r.Context()\n\t\tcurrentUser = t.userService.GetUser(r, t.httpClient)\n\t)\n\n\ttaskID, err := gocql.ParseUUID(taskIDStr)\n\tif err != nil || taskID == emptyUUID {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorTaskIDHasBadFormat, \"TaskService.Delete: task ID(UUID=%s) has bad format or empty. err=%v\", taskIDStr, err)\n\t\tcommon.SendBadRequest(w, r, errorcode.ErrorTaskIDHasBadFormat)\n\t\treturn\n\t}\n\n\tinternalTasks, err := t.taskPersistence.GetByIDs(ctx, nil, partnerID, false, taskID)\n\tif err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantGetTaskByTaskID, \"TaskService.Delete: can't get internal tasks by task ID %v. err=%v\", taskID, err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantGetTaskByTaskID)\n\t\treturn\n\t}\n\n\tif len(internalTasks) == 0 {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorTaskIsNotFoundByTaskID, \"TaskService.Delete: task with ID %v not found.\", taskID)\n\t\tcommon.SendBadRequest(w, r, errorcode.ErrorTaskIsNotFoundByTaskID)\n\t\treturn\n\t}\n\n\tcommonTaskData := internalTasks[0]\n\tif currentUser.HasNOCAccess() != commonTaskData.IsRequireNOCAccess {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorAccessDenied, \"TaskService.Delete: current user %s is not authorized to delete task with ID %v for partnerID %v\", currentUser.UID(), commonTaskData.ID, commonTaskData.PartnerID)\n\t\tcommon.SendForbidden(w, r, errorcode.ErrorAccessDenied)\n\t\treturn\n\t}\n\n\tdto, err := t.getDataToDelete(ctx, taskID, r, w, partnerID)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tdto.tasks = internalTasks\n\tif err = t.executeDeleting(dto); err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantDeleteTask, \"TaskService.Delete: can't process deleting of the task. err=%v\", err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantDeleteTask)\n\t\treturn\n\t}\n\n\tif !currentUser.HasNOCAccess() {\n\t\t// update counters for tasks in separate goroutine\n\t\tgo func(ctx context.Context, iTasks []models.Task) {\n\t\t\tcounters := getCountersForInternalTasks(iTasks)\n\t\t\tif len(counters) == 0 {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\terr := t.taskCounterRepo.DecreaseCounter(commonTaskData.PartnerID, counters, false)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Log.ErrfCtx(ctx, errorcode.ErrorCantProcessData, \"Delete: error while trying to increase counter: \", err)\n\t\t\t}\n\t\t}(ctx, internalTasks)\n\t}\n\n\tlogger.Log.InfofCtx(r.Context(), \"TaskService.Delete: successfully deleted task with ID = %v\", taskID)\n\tcommon.SendNoContent(w)\n}", "func deleteClusterTask(ctx context.Context, t *testing.T, c *clients, name string) {\n\tt.Logf(\"Deleting clustertask %s\", name)\n\tif err := c.ClusterTaskClient.Delete(ctx, name, metav1.DeleteOptions{}); err != nil {\n\t\tt.Fatalf(\"Failed to delete clustertask: %v\", err)\n\t}\n}", "func (t Task) Delete() error {\n\tpath := fmt.Sprintf(\"tasks/%d\", t.ID)\n\t_, err := makeRequest(http.MethodDelete, path, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *ProjectsLocationsMigrationJobsDeleteCall) Force(force bool) *ProjectsLocationsMigrationJobsDeleteCall {\n\tc.urlParams_.Set(\"force\", fmt.Sprint(force))\n\treturn c\n}", "func (access *TaskAccess) DeleteTask(id string) error {\n\treturn access.DB.Where(\"id = ?\", id).First(&models.Task{}).Delete(&models.Task{}).Error\n}", "func DeleteTask(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tw.Header().Set(\"Access-Control-Allow-Methods\", \"DELETE\")\n\tw.Header().Set(\"Access-Control-Allow-Headers\", \"Content-Type\")\n\tparams := mux.Vars(r)\n\tdeleteOneTask(params[\"id\"])\n\tjson.NewEncoder(w).Encode(params[\"id\"])\n}", "func (m TaskManager) AbortTask(c context.Context, ctl task.Controller) error {\n\treturn nil\n}", "func (m TaskManager) AbortTask(c context.Context, ctl task.Controller) error {\n\treturn nil\n}", "func (c Control) ServeStopTask(w http.ResponseWriter, r *http.Request) {\n\tc.ServeTaskAction(w, r, false)\n}", "func (o BucketOutput) ForceDestroy() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v *Bucket) pulumi.BoolPtrOutput { return v.ForceDestroy }).(pulumi.BoolPtrOutput)\n}", "func (o *DeleteNodeParams) WithForce(force *bool) *DeleteNodeParams {\n\to.SetForce(force)\n\treturn o\n}", "func (o OceanLaunchSpecDeleteOptionsOutput) ForceDelete() pulumi.BoolOutput {\n\treturn o.ApplyT(func(v OceanLaunchSpecDeleteOptions) bool { return v.ForceDelete }).(pulumi.BoolOutput)\n}", "func DeleteCloudNodeGroupTask(taskID string, stepName string) error {\n\tstart := time.Now()\n\t//get task information and validate\n\tstate, step, err := cloudprovider.GetTaskStateAndCurrentStep(taskID, stepName)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif step == nil {\n\t\treturn nil\n\t}\n\n\t// step login started here\n\tcloudID := step.Params[\"CloudID\"]\n\tnodeGroupID := step.Params[\"NodeGroupID\"]\n\tkeepInstance := false\n\tif step.Params[\"KeepInstance\"] == \"true\" {\n\t\tkeepInstance = true\n\t}\n\tgroup, err := cloudprovider.GetStorageModel().GetNodeGroup(context.Background(), nodeGroupID)\n\tif err != nil {\n\t\tblog.Errorf(\"DeleteCloudNodeGroupTask[%s]: get nodegroup for %s failed\", taskID, nodeGroupID)\n\t\tretErr := fmt.Errorf(\"get nodegroup information failed, %s\", err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\treturn retErr\n\t}\n\n\t// get cloud and cluster info\n\tcloud, cluster, err := actions.GetCloudAndCluster(cloudprovider.GetStorageModel(), cloudID, group.ClusterID)\n\tif err != nil {\n\t\tblog.Errorf(\"DeleteCloudNodeGroupTask[%s]: get cloud/cluster for nodegroup %s in task %s step %s failed, %s\",\n\t\t\ttaskID, nodeGroupID, taskID, stepName, err.Error())\n\t\tretErr := fmt.Errorf(\"get cloud/cluster information failed, %s\", err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\treturn retErr\n\t}\n\n\t// get dependency resource for cloudprovider operation\n\tcmOption, err := cloudprovider.GetCredential(&cloudprovider.CredentialData{\n\t\tCloud: cloud,\n\t\tAccountID: cluster.CloudAccountID,\n\t})\n\tif err != nil {\n\t\tblog.Errorf(\"DeleteCloudNodeGroupTask[%s]: get credential for nodegroup %s in task %s step %s failed, %s\",\n\t\t\ttaskID, nodeGroupID, taskID, stepName, err.Error())\n\t\tretErr := fmt.Errorf(\"get cloud credential err, %s\", err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\treturn retErr\n\t}\n\tcmOption.Region = group.Region\n\n\t// create node group\n\ttkeCli, err := api.NewTkeClient(cmOption)\n\tif err != nil {\n\t\tblog.Errorf(\"DeleteCloudNodeGroupTask[%s]: get tke client for nodegroup[%s] in task %s step %s failed, %s\",\n\t\t\ttaskID, nodeGroupID, taskID, stepName, err.Error())\n\t\tretErr := fmt.Errorf(\"get cloud tke client err, %s\", err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\treturn err\n\t}\n\terr = tkeCli.DeleteClusterNodePool(cluster.SystemID, []string{group.CloudNodeGroupID}, keepInstance)\n\tif err != nil {\n\t\tblog.Errorf(\"DeleteCloudNodeGroupTask[%s]: call DeleteClusterNodePool[%s] api in task %s step %s failed, %s\",\n\t\t\ttaskID, nodeGroupID, taskID, stepName, err.Error())\n\t\tretErr := fmt.Errorf(\"call DeleteClusterNodePool[%s] api err, %s\", nodeGroupID, err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\treturn retErr\n\t}\n\tblog.Infof(\"DeleteCloudNodeGroupTask[%s]: call DeleteClusterNodePool successful\", taskID)\n\n\t// update response information to task common params\n\tif state.Task.CommonParams == nil {\n\t\tstate.Task.CommonParams = make(map[string]string)\n\t}\n\n\t// update step\n\tif err := state.UpdateStepSucc(start, stepName); err != nil {\n\t\tblog.Errorf(\"DeleteCloudNodeGroupTask[%s] task %s %s update to storage fatal\", taskID, taskID, stepName)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (o *VolumeDeleteParams) SetForce(force *bool) {\n\to.Force = force\n}", "func (o EciScalingConfigurationOutput) ForceDelete() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v *EciScalingConfiguration) pulumi.BoolPtrOutput { return v.ForceDelete }).(pulumi.BoolPtrOutput)\n}", "func (ctrl *TaskController) DeleteTask(w http.ResponseWriter, r *http.Request) {\n\ttaskId := ParamAsString(\"id\", r)\n\tlogrus.Println(\"delete task : \", taskId)\n\n\terr := ctrl.taskDao.Delete(taskId)\n\tif err != nil {\n\t\tlogrus.Error(err)\n\t\tSendJSONError(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tlogrus.Println(\"deleted task : \", taskId)\n\tSendJSONWithHTTPCode(w, nil, http.StatusNoContent)\n}", "func DeleteTask(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Context-Type\", \"application/x-www-form-urlencoded\")\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tw.Header().Set(\"Access-Control-Allow-Methods\", \"DELETE\")\n\tw.Header().Set(\"Access-Control-Allow-Headers\", \"Content-Type\")\n\tparams := mux.Vars(r)\n\tdeleteOneTask(params[\"id\"])\n\tjson.NewEncoder(w).Encode(params[\"id\"])\n\t// json.NewEncoder(w).Encode(\"Task not found\")\n\n}", "func (r *DeviceAppManagementTaskRequest) Delete(ctx context.Context) error {\n\treturn r.JSONRequest(ctx, \"DELETE\", \"\", nil, nil)\n}", "func (c *Cluster) ForceDestoryVM(opts compute.VirtualMachine) error {\n\n\tnode, err := c.getNodeRegion(opts.Region)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\topts.T = node.Client\n\tdefer node.Client.Client.Close()\n\n\t_, err = opts.RecoverDelete()\n\tif err != nil {\n\t\treturn wrapErrorWithCmd(node, err, \"DestroyVM\")\n\t}\n\n\treturn nil\n}", "func deleteTask(id interface{}) {\n\ti, ok := getTaskPosition(id)\n\tif ok == true {\n\t\ttasks_mutex.Lock()\n\t\ttasks = append(tasks[:i], tasks[i+1:]...)\n\t\ttasks_mutex.Unlock()\n\t}\n}", "func DeleteTask(c *gin.Context) {\n\tfmt.Println(\"deleteTask\")\n\ttask := c.Param(\"id\")\n\tfmt.Println(\"task_id: \", task)\n\tdeleteOneTask(task)\n\tc.JSON(http.StatusOK, task)\n\t// json.NewEncoder(w).Encode(\"Task not found\")\n\n}", "func (s *StubTodoStore) DeleteTask(task model.Task) error {\n\tfor i, storeTask := range s.Tasks {\n\t\tif storeTask.Name == task.Name {\n\t\t\ts.Tasks = append(s.Tasks[:i], s.Tasks[(i+1):]...)\n\t\t\treturn nil\n\t\t}\n\t}\n\treturn nil\n}", "func DeleteTask(taskKey int) error {\n\t// connect to database\n\tdb, err := connect(dbString)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Could Not Connect to db\")\n\t}\n\tdefer db.Close()\n\n\t// read/write transaction\n\terr = db.Update(func(tx *bolt.Tx) error {\n\t\tbucket, err := tx.CreateBucketIfNotExists(bucketName)\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"Failed to create non existant bucket: %s\", string(bucketName))\n\t\t}\n\n\t\t// delete task from database\n\t\terr = bucket.Delete(itob(taskKey))\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"Could not Delete item %d\", taskKey)\n\t\t}\n\n\t\treturn nil\n\t})\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Could not complete transaction\")\n\t}\n\n\treturn nil\n}", "func (t *TaskController[T, U, C, CT, TF]) Wait() {\n\tt.wg.Wait()\n\tclose(t.resultCh)\n\tt.pool.DeleteTask(t.taskID)\n}", "func (o OceanLaunchSpecDeleteOptionsPtrOutput) ForceDelete() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v *OceanLaunchSpecDeleteOptions) *bool {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.ForceDelete\n\t}).(pulumi.BoolPtrOutput)\n}", "func DeleteTKEClusterTask(taskID string, stepName string) error {\n\tstart := time.Now()\n\t// get task and task current step\n\tstate, step, err := cloudprovider.GetTaskStateAndCurrentStep(taskID, stepName)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// previous step successful when retry task\n\tif step == nil {\n\t\tblog.Infof(\"DeleteTKEClusterTask[%s]: current step[%s] successful and skip\", taskID, stepName)\n\t\treturn nil\n\t}\n\tblog.Infof(\"DeleteTKEClusterTask[%s]: task %s run step %s, system: %s, old state: %s, params %v\",\n\t\ttaskID, taskID, stepName, step.System, step.Status, step.Params)\n\n\t// step login started here\n\tclusterID := step.Params[cloudprovider.ClusterIDKey.String()]\n\tcloudID := step.Params[cloudprovider.CloudIDKey.String()]\n\tdeleteMode := step.Params[cloudprovider.DeleteModeKey.String()]\n\n\t// only support retain mode\n\tif deleteMode != cloudprovider.Retain.String() {\n\t\tdeleteMode = cloudprovider.Retain.String()\n\t}\n\n\tdependInfo, err := cloudprovider.GetClusterDependBasicInfo(cloudprovider.GetBasicInfoReq{\n\t\tClusterID: clusterID,\n\t\tCloudID: cloudID,\n\t})\n\tif err != nil {\n\t\tblog.Errorf(\"DeleteTKEClusterTask[%s]: GetClusterDependBasicInfo for cluster %s \"+\n\t\t\t\"in task %s step %s failed, %s\", taskID, clusterID, taskID, stepName, err.Error())\n\t\tretErr := fmt.Errorf(\"get cloud/project information failed, %s\", err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\treturn retErr\n\t}\n\n\t// get qcloud client\n\tcli, err := api.NewTkeClient(dependInfo.CmOption)\n\tif err != nil {\n\t\tblog.Errorf(\"DeleteTKEClusterTask[%s]: get tke client for cluster[%s] in task %s step %s failed, %s\",\n\t\t\ttaskID, clusterID, taskID, stepName, err.Error())\n\t\tretErr := fmt.Errorf(\"get cloud tke client err, %s\", err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\treturn retErr\n\t}\n\n\tif dependInfo.Cluster.SystemID != \"\" {\n\t\terr = cli.DeleteTKECluster(dependInfo.Cluster.SystemID, api.DeleteMode(deleteMode))\n\t\tif err != nil {\n\t\t\tblog.Errorf(\"DeleteTKEClusterTask[%s]: task[%s] step[%s] call qcloud DeleteTKECluster failed: %v\",\n\t\t\t\ttaskID, taskID, stepName, err)\n\t\t\tretErr := fmt.Errorf(\"call qcloud DeleteTKECluster failed: %s\", err.Error())\n\t\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\t\treturn retErr\n\t\t}\n\t\t_ = updateClusterSystemID(clusterID, \"\")\n\t\tblog.Infof(\"DeleteTKEClusterTask[%s]: task %s DeleteTKECluster[%s] successful\",\n\t\t\ttaskID, taskID, dependInfo.Cluster.SystemID)\n\t} else {\n\t\tblog.Infof(\"DeleteTKEClusterTask[%s]: task %s DeleteTKECluster skip current step \"+\n\t\t\t\"because SystemID empty\", taskID, taskID)\n\t}\n\n\tif err := state.UpdateStepSucc(start, stepName); err != nil {\n\t\tblog.Errorf(\"DeleteTKEClusterTask[%s]: task %s %s update to storage fatal\", taskID, taskID, stepName)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *restClient) DeleteTask(ctx context.Context, req *cloudtaskspb.DeleteTaskRequest, opts ...gax.CallOption) error {\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v2beta3/%v\", req.GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"name\", url.QueryEscape(req.GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\treturn gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"DELETE\", baseUrl.String(), nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\t// Returns nil if there is no error, otherwise wraps\n\t\t// the response code and body into a non-nil error\n\t\treturn googleapi.CheckResponse(httpRsp)\n\t}, opts...)\n}", "func (r ApiApiDeploymentsIdDeleteRequest) Force(force string) ApiApiDeploymentsIdDeleteRequest {\n\tr.force = &force\n\treturn r\n}", "func (a *Client) DeleteTask(params *DeleteTaskParams, authInfo runtime.ClientAuthInfoWriter, opts ...ClientOption) (*DeleteTaskNoContent, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewDeleteTaskParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"deleteTask\",\n\t\tMethod: \"DELETE\",\n\t\tPathPattern: \"/tasks/{id}\",\n\t\tProducesMediaTypes: []string{\"application/vnd.goswagger.examples.task-tracker.v1+json\"},\n\t\tConsumesMediaTypes: []string{\"application/vnd.goswagger.examples.task-tracker.v1+json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &DeleteTaskReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*DeleteTaskNoContent)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*DeleteTaskDefault)\n\treturn nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func DeleteTask(w http.ResponseWriter, r *http.Request, repo *tasks.TaskRepository) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tparams := mux.Vars(r)\n\ttaskID, err := strconv.Atoi(params[\"id\"])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\ttask, err := repo.DeleteTask(taskID)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tjson.NewEncoder(w).Encode(apiIndexTask(task))\n}", "func (c *ProjectsLocationsConnectionProfilesDeleteCall) Force(force bool) *ProjectsLocationsConnectionProfilesDeleteCall {\n\tc.urlParams_.Set(\"force\", fmt.Sprint(force))\n\treturn c\n}", "func DeleteJobTask(w http.ResponseWriter, r *http.Request) {\n\tresponse := services.DeleteJobTask(r)\n\n\trender.Status(r, response.Code)\n\trender.JSON(w, r, response)\n}", "func (ts *TaskService) Delete(ctx context.Context, req *taskAPI.DeleteRequest) (*taskAPI.DeleteResponse, error) {\n\tdefer logPanicAndDie(log.G(ctx))\n\n\tlog.G(ctx).WithFields(logrus.Fields{\"id\": req.ID, \"exec_id\": req.ExecID}).Debug(\"delete\")\n\ttask, err := ts.taskManager.Task(req.ID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx = namespaces.WithNamespace(ctx, defaultNamespace)\n\tresp, err := task.Delete(ctx, req)\n\tif err != nil {\n\t\tlog.G(ctx).WithError(err).Error(\"delete failed\")\n\t\treturn nil, err\n\t}\n\n\tlog.G(ctx).WithFields(logrus.Fields{\n\t\t\"pid\": resp.Pid,\n\t\t\"exit_status\": resp.ExitStatus,\n\t}).Debug(\"delete succeeded\")\n\treturn resp, nil\n}", "func (c *Client) DeleteWorkforce(ctx context.Context, params *DeleteWorkforceInput, optFns ...func(*Options)) (*DeleteWorkforceOutput, error) {\n\tif params == nil {\n\t\tparams = &DeleteWorkforceInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"DeleteWorkforce\", params, optFns, addOperationDeleteWorkforceMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*DeleteWorkforceOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func deleteTaskMeta(db *leveldb.DB, name string) error {\n\tif db == nil {\n\t\treturn terror.ErrWorkerLogInvalidHandler.Generate()\n\t}\n\n\terr := db.Delete(encodeTaskMetaKey(name), nil)\n\tif err != nil {\n\t\treturn terror.ErrWorkerLogDeleteTaskMeta.Delegate(err, name)\n\t}\n\n\treturn nil\n}", "func (t *PodTask) ClearTaskInfo() {\n\tt.OfferIds = nil\n\tt.TaskInfo.TaskId = nil\n\tt.TaskInfo.SlaveId = nil\n\tt.TaskInfo.Resources = nil\n\tt.TaskInfo.Data = nil\n}", "func (t *task) cleanTaskCache() {\n\tt.runtime = nil\n\tt.config = nil\n}", "func (sysload *SystemLoad) RemoveTask(id uint64) {\n\ttaskLoad := sysload.findTask(id)\n\tif taskLoad != nil {\n\t\tsysload.tasksLoad.Remove(taskLoad)\n\t}\n}", "func (t *task) replaceTask(\n\truntime *pbtask.RuntimeInfo,\n\ttaskConfig *pbtask.TaskConfig,\n\tforceReplace bool) error {\n\tif runtime == nil || runtime.GetRevision() == nil {\n\t\treturn yarpcerrors.InvalidArgumentErrorf(\n\t\t\t\"replaceTask expects a non-nil runtime with non-nil Revision\")\n\t}\n\n\tvar runtimeCopy *pbtask.RuntimeInfo\n\tvar labelsCopy []*peloton.Label\n\n\t// notify listeners after dropping the lock\n\tdefer func() {\n\t\tt.jobFactory.notifyTaskRuntimeChanged(\n\t\t\tt.JobID(),\n\t\t\tt.ID(),\n\t\t\tt.jobType,\n\t\t\truntimeCopy,\n\t\t\tlabelsCopy,\n\t\t)\n\t}()\n\n\tt.Lock()\n\tdefer t.Unlock()\n\n\t// update the cache if,\n\t// 1. it is a force replace, or\n\t// 2. there is no existing runtime cache,\n\t// 3. new runtime has a higher version number than the existing\n\tif forceReplace ||\n\t\tt.runtime == nil ||\n\t\truntime.GetRevision().GetVersion() > t.runtime.GetRevision().GetVersion() {\n\t\t// Update task config and runtime\n\t\tt.config = &taskConfigCache{\n\t\t\tconfigVersion: runtime.GetConfigVersion(),\n\t\t\tlabels: taskConfig.GetLabels(),\n\t\t}\n\t\tt.runtime = runtime\n\t\truntimeCopy = proto.Clone(t.runtime).(*pbtask.RuntimeInfo)\n\t\tlabelsCopy = t.copyLabelsInCache()\n\t}\n\n\treturn nil\n}", "func (o *Orchestrator) RemoveTask(taskDefinition interface{}) {\n\to.mu.Lock()\n\tdefer o.mu.Unlock()\n\n\tidx := containsTask(taskDefinition, o.expectedTasks)\n\tif idx < 0 {\n\t\treturn\n\t}\n\n\to.expectedTasks = append(o.expectedTasks[:idx], o.expectedTasks[idx+1:]...)\n}", "func HandleDeleteTask(w http.ResponseWriter, r *http.Request) {\n\tlog.Root.Info(\"HandleDeleteTask BEGIN\")\n\n\tif r.Method != http.MethodPost {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tHttpResponseError(w, ErrNotFound)\n\t\treturn\n\t}\n\n\tdefer r.Body.Close()\n\tbody, _ := ioutil.ReadAll(r.Body)\n\n\tdata := make(map[string]interface{})\n\terr := json.Unmarshal(body, &data)\n\tif err != nil {\n\t\tlog.Root.Error(\"HandleDeleteTask Parse HTTP request body error\")\n\t\tHttpResponseError(w, ErrForm)\n\t\treturn\n\t}\n\n\telem, ok := data[\"taskID\"]\n\tif !ok {\n\t\tlog.Root.Error(\"HandleDeleteTask HTTP form data error\")\n\t\tHttpResponseError(w, ErrForm)\n\t\treturn\n\t}\n\n\ttaskID := elem.(string)\n\terr = node.DeleteTask(taskID)\n\tif err != nil {\n\t\tlog.Root.Error(\"HandleDeleteTask Delete task error. TaskID: %v\", taskID)\n\t\tHttpResponseError(w, ErrServer)\n\t\treturn\n\t}\n\n\tlog.Root.Info(\"HandleDeleteTask END\")\n\tHttpResponseOk(w)\n\treturn\n}", "func (o *DeleteBlueprintRequestUsingDELETE1Params) SetForce(force *bool) {\n\to.Force = force\n}", "func CtrStopContainer(containerID string, force bool) error {\n\tif err := verifyCtr(); err != nil {\n\t\treturn fmt.Errorf(\"CtrStopContainer: exception while verifying ctrd client: %s\", err.Error())\n\t}\n\tctr, err := CtrLoadContainer(containerID)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"can't find cotainer %s (%v)\", containerID, err)\n\t}\n\n\tsignal, err := containerd.ParseSignal(defaultSignal)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif signal, err = containerd.GetStopSignal(ctrdCtx, ctr, signal); err != nil {\n\t\treturn err\n\t}\n\n\ttask, err := ctr.Task(ctrdCtx, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// it is unclear whether we have to wait after this or proceed\n\t// straight away. It is also unclear whether paying any attention\n\t// to the err returned is worth anything at this point\n\t_ = task.Kill(ctrdCtx, signal, containerd.WithKillAll)\n\n\tif force {\n\t\t_, err = task.Delete(ctrdCtx, containerd.WithProcessKill)\n\t} else {\n\t\t_, err = task.Delete(ctrdCtx)\n\t}\n\n\treturn err\n}", "func deleteTask(writer http.ResponseWriter, req *http.Request) {\n\tvars := mux.Vars(req)\n\tcreatedAt, err := time.Parse(time.RFC3339, vars[\"createdAt\"])\n\tif err != nil {\n\t\tlog.Print(\"error:\", err)\n\t\thttp.Error(writer, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tdatabase, err := loadJsonFile()\n\tif err != nil {\n\t\tlog.Print(\"error:\", err)\n\t\thttp.Error(writer, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tfor i, _ := range database.Tasks {\n\t\tif database.Tasks[i].CreatedAt.Equal(createdAt) {\n\t\t\tdatabase.Tasks = append(database.Tasks[:i], database.Tasks[i+1:]...)\n\t\t\treturnJson(database, writer)\n\t\t\treturn\n\t\t}\n\t}\n\t//this code runs only if no taks was found with the correct createdAt timestamp\n\thttp.Error(writer, err.Error(), http.StatusBadRequest)\n}", "func deleteTask(w http.ResponseWriter, r *http.Request){\n\t//definimos variable de vars que devuelve las variables de ruta\n\tvars := mux.Vars(r)\n\n\ttaskID, err := strconv.Atoi(vars[\"id\"])\n\tif err != nil{\n\t\tfmt.Fprintf(w, \"Invalid ID\")\n\t\treturn\n\t}\n\n\t//Se elimina la task a la lista, guardando todas las que estan hasta su indice, y la que le sigue en adelante.\n\tfor i, task := range tasks {\n\t\tif task.ID == taskID {\n\t\t\ttasks = append(tasks[:i], tasks[i + 1:] ...)\n\t\t\tfmt.Fprintf(w, \"The task with ID %v has been removed succesfully\", taskID)\n\t\t}\n\t}\n}", "func (c *clustermgrClient) DeleteMigrateTask(ctx context.Context, key string) (err error) {\n\treturn c.client.DeleteKV(ctx, key)\n}", "func (options *DeleteJobOptions) SetForce(force bool) *DeleteJobOptions {\n\toptions.Force = core.BoolPtr(force)\n\treturn options\n}", "func (p *PruneTask) ClearTimeout() {}", "func (ts *TaskService) Delete(requestCtx context.Context, req *taskAPI.DeleteRequest) (*taskAPI.DeleteResponse, error) {\n\tdefer logPanicAndDie(log.G(requestCtx))\n\tlog.G(requestCtx).WithFields(logrus.Fields{\"id\": req.ID, \"exec_id\": req.ExecID}).Debug(\"delete\")\n\n\tresp, err := ts.taskManager.DeleteProcess(requestCtx, req, ts.runcService)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlog.G(requestCtx).WithFields(logrus.Fields{\n\t\t\"pid\": resp.Pid,\n\t\t\"exit_status\": resp.ExitStatus,\n\t}).Debug(\"delete succeeded\")\n\treturn resp, nil\n}", "func (o TaskDefinitionOutput) SkipDestroy() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v *TaskDefinition) pulumi.BoolPtrOutput { return v.SkipDestroy }).(pulumi.BoolPtrOutput)\n}", "func ReleaseTask() {\n\tRun(\"clean\")\n\toptimizeFlag = true\n\txcnFlag = true\n\tbs := getGoBuilders()\n\tif bs.needsBrowser() {\n\t\tRun(\"deps\")\n\t\tRun(\"browser\")\n\t\tRun(\"chrome\")\n\t\tRun(\"bindata\")\n\t}\n\tbs.RunAll()\n\tbs.CreateReleaseArchives()\n}", "func CleanClusterDBInfoTask(taskID string, stepName string) error {\n\tstart := time.Now()\n\t// get task and task current step\n\tstate, step, err := cloudprovider.GetTaskStateAndCurrentStep(taskID, stepName)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// previous step successful when retry task\n\tif step == nil {\n\t\tblog.Infof(\"CleanClusterDBInfoTask[%s]: current step[%s] successful and skip\", taskID, stepName)\n\t\treturn nil\n\t}\n\tblog.Infof(\"CleanClusterDBInfoTask[%s]: task %s run step %s, system: %s, old state: %s, params %v\",\n\t\ttaskID, taskID, stepName, step.System, step.Status, step.Params)\n\n\t// step login started here\n\tclusterID := step.Params[cloudprovider.ClusterIDKey.String()]\n\tcluster, err := cloudprovider.GetStorageModel().GetCluster(context.Background(), clusterID)\n\tif err != nil {\n\t\tblog.Errorf(\"CleanClusterDBInfoTask[%s]: get cluster for %s failed\", taskID, clusterID)\n\t\tretErr := fmt.Errorf(\"get cluster information failed, %s\", err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\treturn retErr\n\t}\n\n\t// delete cluster autoscalingOption\n\terr = cloudprovider.GetStorageModel().DeleteAutoScalingOption(context.Background(), cluster.ClusterID)\n\tif err != nil {\n\t\tblog.Errorf(\"CleanClusterDBInfoTask[%s]: clean cluster[%s] \"+\n\t\t\t\"autoscalingOption failed: %v\", taskID, cluster.ClusterID, err)\n\t}\n\n\t// delete nodes\n\terr = cloudprovider.GetStorageModel().DeleteNodesByClusterID(context.Background(), cluster.ClusterID)\n\tif err != nil {\n\t\tblog.Errorf(\"CleanClusterDBInfoTask[%s]: delete nodes for %s failed\", taskID, clusterID)\n\t\tretErr := fmt.Errorf(\"delete node for %s failed, %s\", clusterID, err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\treturn retErr\n\t}\n\tblog.Infof(\"CleanClusterDBInfoTask[%s]: delete nodes for cluster[%s] in DB successful\", taskID, clusterID)\n\n\t// delete nodeGroup\n\terr = cloudprovider.GetStorageModel().DeleteNodeGroupByClusterID(context.Background(), cluster.ClusterID)\n\tif err != nil {\n\t\tblog.Errorf(\"CleanClusterDBInfoTask[%s]: delete nodeGroups for %s failed\", taskID, clusterID)\n\t\tretErr := fmt.Errorf(\"delete nodeGroups for %s failed, %s\", clusterID, err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\treturn retErr\n\t}\n\tblog.Infof(\"CleanClusterDBInfoTask[%s]: delete nodeGroups for cluster[%s] in DB successful\",\n\t\ttaskID, clusterID)\n\n\t// delete CIDR and only print logInfo\n\terr = releaseClusterCIDR(cluster)\n\tif err != nil {\n\t\tblog.Errorf(\"CleanClusterDBInfoTask[%s]: releaseClusterCIDR[%s] cidr failed\", taskID, clusterID)\n\t} else {\n\t\tblog.Infof(\"CleanClusterDBInfoTask[%s]: releaseClusterCIDR[%s] cidr successful\", taskID, clusterID)\n\t}\n\n\t// delete cluster\n\tcluster.Status = icommon.StatusDeleting\n\terr = cloudprovider.GetStorageModel().UpdateCluster(context.Background(), cluster)\n\tif err != nil {\n\t\tblog.Errorf(\"CleanClusterDBInfoTask[%s]: delete cluster for %s failed\", taskID, clusterID)\n\t\tretErr := fmt.Errorf(\"delete cluster for %s failed, %s\", clusterID, err.Error())\n\t\t_ = state.UpdateStepFailure(start, stepName, retErr)\n\t\treturn retErr\n\t}\n\tblog.Infof(\"CleanClusterDBInfoTask[%s]: delete cluster[%s] in DB successful\", taskID, clusterID)\n\n\tutils.SyncDeletePassCCCluster(taskID, cluster)\n\t_ = utils.DeleteClusterCredentialInfo(cluster.ClusterID)\n\n\t// virtual cluster need to clean cluster token\n\tif cluster.ClusterType == icommon.ClusterTypeVirtual {\n\t\t_ = utils.DeleteBcsAgentToken(clusterID)\n\t}\n\n\tif err := state.UpdateStepSucc(start, stepName); err != nil {\n\t\tblog.Errorf(\"CleanClusterDBInfoTask[%s]: task %s %s update to storage fatal\", taskID, taskID, stepName)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *Job) AfterExecuteTask() *Job {\n\tif c.delayUnit == delayNone {\n\t\tc.timingMode = afterExecuteTask\n\t}\n\treturn c\n}", "func DelContainerForce(c *check.C, cname string) (*http.Response, error) {\n\tq := url.Values{}\n\tq.Add(\"force\", \"true\")\n\tq.Add(\"v\", \"true\")\n\treturn request.Delete(\"/containers/\"+cname, request.WithQuery(q))\n}", "func (s *DeleteTaskSetInput) SetForce(v bool) *DeleteTaskSetInput {\n\ts.Force = &v\n\treturn s\n}", "func (options *DeleteActionOptions) SetForce(force bool) *DeleteActionOptions {\n\toptions.Force = core.BoolPtr(force)\n\treturn options\n}", "func (t Task) Delete(id string) error {\n\terr := os.RemoveAll(path.Dir(t.Path))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, user := range DB.Users {\n\t\tdelete(user.Tasks, id)\n\t}\n\n\tdelete(DB.Tasks, id)\n\tWriteJSON(\"db.json\", DB)\n\n\treturn nil\n}", "func deleteTask(taskID int) {\n\n\ttrepo := sqlite.NewTaskRepo()\n\n\ttrepo.DeleteTask(taskID)\n\tfmt.Println(\"Tarea borrada correctamente\")\n}", "func (dtm *DfgetTaskManager) Delete(ctx context.Context, clientID, taskID string) error {\n\tkey, err := generateKey(clientID, taskID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn dtm.dfgetTaskStore.Delete(key)\n}", "func deleteTask(w http.ResponseWriter, r *http.Request) {\r\n\tvars := mux.Vars(r) //copio del anterior porque el metodo de busqueda es el mismo\r\n\ttaskID, err := strconv.Atoi(vars[\"id\"])\r\n\tif err != nil {\r\n\t\tfmt.Fprintf(w, \"Invalid ID\")\r\n\t\treturn\r\n\t}\r\n\tfor i, task := range tasks { //misma busqueda que en el caso anterior\r\n\t\tif task.ID == taskID {\r\n\t\t\ttasks = append(tasks[:i], tasks[i+1:]...) //en realidad no borra sino que arma un slice con los elementos previos y posteriores al indice dado\r\n\t\t\tfmt.Fprintf(w, \"The task with ID: %v has been successfully removed.\", taskID)\r\n\t\t}\r\n\t}\r\n}", "func (k *KubernetesExecutor) killPodForTask(driver bindings.ExecutorDriver, tid, reason string) {\n\tk.removePodTask(driver, tid, reason, mesos.TaskState_TASK_KILLED)\n}", "func (h *Heartbeat) RemoveTask(name string) error {\n\th.lock <- struct{}{}\n\tdefer func() {\n\t\t<-h.lock\n\t}()\n\tif _, ok := h.slavesTs[name]; !ok {\n\t\treturn terror.ErrSyncerUnitHeartbeatRecordNotFound.Generate(name)\n\t}\n\tdelete(h.slavesTs, name)\n\n\tif len(h.slavesTs) == 0 {\n\t\t// cancel work\n\t\th.cancel()\n\t\th.cancel = nil\n\t\th.wg.Wait()\n\n\t\t// close DB\n\t\th.master.Close()\n\t\th.master = nil\n\t}\n\n\treturn nil\n}", "func FinishTask(taskID string) error {\n\ttasks.Lock()\n\tdefer tasks.Unlock()\n\n\tif _, ok := tasks.m[taskID]; ok { // key exist\n\t\tdelete(tasks.m, taskID)\n\t\treturn nil\n\t}\n\n\treturn errors.New(\"Task not exist\")\n}", "func (r *Bucket) ForceDestroy() pulumi.BoolOutput {\n\treturn (pulumi.BoolOutput)(r.s.State[\"forceDestroy\"])\n}", "func deleteOneTask(task string) {\n\tfmt.Println(task)\n\tid, _ := primitive.ObjectIDFromHex(task)\n\tfilter := bson.M{\"_id\": id}\n\t_, err := collection.DeleteOne(context.Background(), filter)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tfmt.Println(\"Deleted task\", id)\n}", "func Run(task structs.Task, threadChannel chan<- structs.ThreadMsg) {\n\ttMsg := structs.ThreadMsg{}\n\ttMsg.Error = false\n\ttMsg.TaskItem = task\n\n\tparams := strings.TrimSpace(task.Params)\n\terr := os.Unsetenv(params)\n\n\tif err != nil {\n\t\ttMsg.Error = true\n\t\ttMsg.TaskResult = []byte(err.Error())\n\t\tthreadChannel <- tMsg\n\t\treturn\n\t}\n\n\ttMsg.TaskResult = []byte(fmt.Sprintf(\"Successfully cleared %s\", params))\n\tthreadChannel <- tMsg\n}", "func (h *Hub) StopTask(ctx context.Context, request *pb.StopTaskRequest) (*pb.StopTaskReply, error) {\n\tlog.G(h.ctx).Info(\"handling StopTask request\", zap.Any(\"req\", request))\n\ttaskID := request.Id\n\tminerID, ok := h.getMinerByTaskID(taskID)\n\tif !ok {\n\t\treturn nil, status.Errorf(codes.NotFound, \"no such task %s\", taskID)\n\t}\n\n\tminer, ok := h.getMinerByID(minerID)\n\tif !ok {\n\t\treturn nil, status.Errorf(codes.NotFound, \"no miner with task %s\", minerID)\n\t}\n\n\t_, err := miner.Client.Stop(ctx, &pb.StopTaskRequest{Id: taskID})\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.NotFound, \"failed to stop the task %s\", taskID)\n\t}\n\n\tminer.deregisterRoute(taskID)\n\tminer.Retain(taskID)\n\n\th.deleteTaskByID(taskID)\n\n\treturn &pb.StopTaskReply{}, nil\n}" ]
[ "0.60128415", "0.59829104", "0.59223276", "0.5853163", "0.5822631", "0.5637229", "0.55053747", "0.5499162", "0.54739857", "0.54706156", "0.5426137", "0.54249454", "0.5411983", "0.5394763", "0.53803176", "0.5303343", "0.529844", "0.52842695", "0.52794904", "0.526981", "0.52620506", "0.52614605", "0.52438265", "0.5229123", "0.52031994", "0.52016747", "0.5198306", "0.5193687", "0.5182462", "0.516419", "0.5158677", "0.5154214", "0.51256126", "0.51209474", "0.5080228", "0.507524", "0.50470334", "0.5025553", "0.5022043", "0.5003506", "0.5003506", "0.49824268", "0.49779436", "0.49666533", "0.49641287", "0.49600813", "0.49511868", "0.49462786", "0.49231607", "0.4920158", "0.4897834", "0.48959228", "0.48950416", "0.4893433", "0.48785496", "0.4878278", "0.48781827", "0.48748326", "0.48673394", "0.48550013", "0.48445103", "0.48237336", "0.48209566", "0.48189393", "0.4788337", "0.47846553", "0.47768438", "0.4771809", "0.4770359", "0.47581577", "0.47552618", "0.4751539", "0.47435454", "0.47347307", "0.4725758", "0.4714502", "0.47056973", "0.47044498", "0.46800032", "0.46674934", "0.4664395", "0.46612114", "0.46532768", "0.46527845", "0.464209", "0.46356976", "0.46270242", "0.46268144", "0.46195802", "0.46157256", "0.46102428", "0.4605348", "0.46036488", "0.45955518", "0.45602873", "0.4557687", "0.4554907", "0.45543045", "0.45530924", "0.45441574" ]
0.7716327
0
InspectTask function returns detailed status information for the referenced taskID.
func (d *Driver) InspectTask(taskID string) (*drivers.TaskStatus, error) { d.logger.Debug("InspectTask called") handle, ok := d.tasks.Get(taskID) if !ok { return nil, drivers.ErrTaskNotFound } return handle.taskStatus(), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *Client) InspectTask(id string) (*swarm.Task, error) {\n\tresp, err := c.do(http.MethodGet, \"/tasks/\"+id, doOptions{})\n\tif err != nil {\n\t\tvar e *Error\n\t\tif errors.As(err, &e) && e.Status == http.StatusNotFound {\n\t\t\treturn nil, &NoSuchTask{ID: id}\n\t\t}\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\tvar task swarm.Task\n\tif err := json.NewDecoder(resp.Body).Decode(&task); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &task, nil\n}", "func (ts *TasksRPC) GetTaskMonitor(ctx context.Context, req *taskproto.GetTaskRequest) (*taskproto.TaskResponse, error) {\n\tvar rsp taskproto.TaskResponse\n\tctx = common.GetContextData(ctx)\n\tctx = common.ModifyContext(ctx, common.TaskService, podName)\n\n\tl.LogWithFields(ctx).Debugf(\"Incoming request to get the task details and response body for the task %v\", req.TaskID)\n\trsp.Header = map[string]string{\n\t\t\"Date\": time.Now().Format(http.TimeFormat),\n\t}\n\tprivileges := []string{common.PrivilegeLogin}\n\tauthResp, err := ts.AuthenticationRPC(ctx, req.SessionToken, privileges)\n\tif authResp.StatusCode != http.StatusOK {\n\t\tif err != nil {\n\t\t\tl.LogWithFields(ctx).Errorf(\"Error while authorizing the session token : %s\", err.Error())\n\t\t}\n\t\tfillProtoResponse(ctx, &rsp, authResp)\n\t\treturn &rsp, nil\n\t}\n\t_, err = ts.GetSessionUserNameRPC(ctx, req.SessionToken)\n\tif err != nil {\n\t\tl.LogWithFields(ctx).Printf(authErrorMessage)\n\t\tfillProtoResponse(ctx, &rsp, common.GeneralError(http.StatusUnauthorized, response.NoValidSession, authErrorMessage, nil, nil))\n\t\treturn &rsp, nil\n\t}\n\t// get task status from database using task id\n\ttask, err := ts.GetTaskStatusModel(ctx, req.TaskID, common.InMemory)\n\tif err != nil {\n\t\tl.LogWithFields(ctx).Printf(\"error getting task status : %v\", err)\n\t\tfillProtoResponse(ctx, &rsp, common.GeneralError(http.StatusNotFound, response.ResourceNotFound, err.Error(), []interface{}{\"Task\", req.TaskID}, nil))\n\t\treturn &rsp, nil\n\t}\n\n\t// Check the state of the task\n\tif task.TaskState == \"Completed\" || task.TaskState == \"Cancelled\" || task.TaskState == \"Killed\" || task.TaskState == \"Exception\" {\n\t\t// return with the actual status code, along with response header and response body\n\t\t//Build the response Body\n\t\trsp.Header = task.Payload.HTTPHeaders\n\t\trsp.Body = task.TaskResponse\n\t\trsp.StatusCode = task.StatusCode\n\t\t// Delete the task from db as it is completed and user requested for the details.\n\t\t// return the user with task details by deleting the task from db\n\t\t// User should be careful as this is the last call to Task monitor API.\n\t\t/*\n\t\t\terr := task.Delete()\n\t\t\tif err != nil {\n\t\t\t\tl.Log.Printf(\"error while deleting the task from db: %v\", err)\n\t\t\t}\n\t\t*/\n\t\treturn &rsp, nil\n\t}\n\t// Construct the Task object to return as long as 202 code is being returned.\n\n\tmessageList := []tresponse.Messages{}\n\tfor _, element := range task.Messages {\n\t\tmessage := tresponse.Messages{\n\t\t\tMessageID: element.MessageID,\n\t\t\tRelatedProperties: element.RelatedProperties,\n\t\t\tMessage: element.Message,\n\t\t\tMessageArgs: element.MessageArgs,\n\t\t\tSeverity: element.Severity,\n\t\t}\n\t\tmessageList = append(messageList, message)\n\t}\n\n\tcommonResponse := response.Response{\n\t\tOdataType: common.TaskType,\n\t\tID: task.ID,\n\t\tName: task.Name,\n\t\tOdataContext: \"/redfish/v1/$metadata#Task.Task\",\n\t\tOdataID: \"/redfish/v1/TaskService/Tasks/\" + task.ID,\n\t}\n\trsp.StatusCode = http.StatusAccepted\n\trsp.StatusMessage = response.TaskStarted\n\tcommonResponse.MessageArgs = []string{task.ID}\n\tcommonResponse.CreateGenericResponse(rsp.StatusMessage)\n\n\thttpHeaders := []string{}\n\tfor key, value := range task.Payload.HTTPHeaders {\n\t\thttpHeaders = append(httpHeaders, fmt.Sprintf(\"%v: %v\", key, value))\n\t}\n\n\ttaskResponse := tresponse.Task{\n\t\tResponse: commonResponse,\n\t\tTaskState: task.TaskState,\n\t\tStartTime: task.StartTime.UTC(),\n\t\tEndTime: task.EndTime.UTC(),\n\t\tTaskStatus: task.TaskStatus,\n\t\tMessages: messageList,\n\t\tTaskMonitor: task.TaskMonitor,\n\t\tPayload: tresponse.Payload{\n\t\t\tHTTPHeaders: httpHeaders,\n\t\t\tHTTPOperation: task.Payload.HTTPOperation,\n\t\t\tJSONBody: string(task.Payload.JSONBody),\n\t\t\tTargetURI: task.Payload.TargetURI,\n\t\t},\n\t\tPercentComplete: task.PercentComplete,\n\t}\n\tif task.ParentID == \"\" {\n\t\tvar subTask = tresponse.ListMember{\n\t\t\tOdataID: \"/redfish/v1/TaskService/Tasks/\" + task.ID + \"/SubTasks\",\n\t\t}\n\t\ttaskResponse.SubTasks = &subTask\n\t}\n\trespBody := generateResponse(ctx, taskResponse)\n\trsp.Body = respBody\n\tl.LogWithFields(ctx).Debugf(\"Outgoing response for getting subtasks: %v\", string(respBody))\n\n\trsp.Header[\"location\"] = task.TaskMonitor\n\treturn &rsp, nil\n}", "func (c *Client) DescribeTaskDetail(request *DescribeTaskDetailRequest) (response *DescribeTaskDetailResponse, err error) {\n if request == nil {\n request = NewDescribeTaskDetailRequest()\n }\n response = NewDescribeTaskDetailResponse()\n err = c.Send(request, response)\n return\n}", "func ShowTaskStat(taskJSON []byte, scale float64) error {\n\ttask := Task{}\n\n\terror := json.Unmarshal([]byte(taskJSON), &task)\n\tif error != nil {\n\t\treturn errors.New(\"[ERROR] Couldn't parse task log as JSON:\" + string(taskJSON))\n\t}\n\n\tstartDateTime, _ := time.Parse(DateTimeLayout, task.Start)\n\tendDateTime, _ := time.Parse(DateTimeLayout, task.End)\n\n\t// TODO: もうちょっとフォーマット何とかする\n\tduration := float64(endDateTime.Sub(startDateTime).Seconds()) * scale\n\tminutes := duration / 60.0\n\thours := minutes / 60.0\n\n\t// 小数点丸め\n\tfmt.Fprintf(os.Stdout, \"%s %v\\n\", task.Name, math.Trunc(hours*100)/100.0)\n\treturn nil\n}", "func (ctl *StatusController) PrintStatusOfTask(ctx context.Context, name string) error {\n\ttasks, err := ctl.getTask(ctx, name)\n\tif err != nil {\n\t\treturn err\n\t}\n\tctl.printToView(tasks)\n\treturn nil\n}", "func (d *dispatcher) monitorTask(taskID int64) (finished bool, subTaskErrs []error) {\n\t// TODO: Consider putting the following operations into a transaction.\n\tvar err error\n\td.task, err = d.taskMgr.GetGlobalTaskByID(taskID)\n\tif err != nil {\n\t\tlogutil.BgLogger().Error(\"check task failed\", zap.Int64(\"task ID\", d.task.ID), zap.Error(err))\n\t\treturn false, nil\n\t}\n\tswitch d.task.State {\n\tcase proto.TaskStateCancelling:\n\t\treturn false, []error{errors.New(\"cancel\")}\n\tcase proto.TaskStateReverting:\n\t\tcnt, err := d.taskMgr.GetSubtaskInStatesCnt(d.task.ID, proto.TaskStateRevertPending, proto.TaskStateReverting)\n\t\tif err != nil {\n\t\t\tlogutil.BgLogger().Warn(\"check task failed\", zap.Int64(\"task ID\", d.task.ID), zap.Error(err))\n\t\t\treturn false, nil\n\t\t}\n\t\treturn cnt == 0, nil\n\tdefault:\n\t\tsubTaskErrs, err = d.taskMgr.CollectSubTaskError(d.task.ID)\n\t\tif err != nil {\n\t\t\tlogutil.BgLogger().Warn(\"collect subtask error failed\", zap.Int64(\"task ID\", d.task.ID), zap.Error(err))\n\t\t\treturn false, nil\n\t\t}\n\t\tif len(subTaskErrs) > 0 {\n\t\t\treturn false, subTaskErrs\n\t\t}\n\t\t// check subtasks pending or running.\n\t\tcnt, err := d.taskMgr.GetSubtaskInStatesCnt(d.task.ID, proto.TaskStatePending, proto.TaskStateRunning)\n\t\tif err != nil {\n\t\t\tlogutil.BgLogger().Warn(\"check task failed\", zap.Int64(\"task ID\", d.task.ID), zap.Error(err))\n\t\t\treturn false, nil\n\t\t}\n\t\treturn cnt == 0, nil\n\t}\n}", "func (l *LogCache) GetTaskStatus() define.TaskStatus {\n\treturn l.status\n}", "func (s *K8sSvc) GetTaskStatus(ctx context.Context, cluster string, taskID string) (*common.TaskStatus, error) {\n\trequuid := utils.GetReqIDFromContext(ctx)\n\n\tjob, err := s.cliset.BatchV1().Jobs(s.namespace).Get(taskID, metav1.GetOptions{})\n\tif err != nil {\n\t\tglog.Errorln(\"get task error\", err, \"taskID\", taskID, \"requuid\", requuid)\n\t\treturn nil, err\n\t}\n\n\tglog.Infoln(\"get task\", taskID, job.Status, \"requuid\", requuid)\n\n\tstatus := &common.TaskStatus{\n\t\tStatus: common.TaskStatusRunning,\n\t}\n\tif job.Status.StartTime != nil {\n\t\tstatus.StartedAt = job.Status.StartTime.String()\n\t}\n\tif job.Status.CompletionTime != nil {\n\t\tstatus.FinishedAt = job.Status.CompletionTime.String()\n\t}\n\n\tif job.Status.Succeeded > 0 {\n\t\tglog.Infoln(\"task succeeded, taskID\", taskID, \"requuid\", requuid)\n\t\tstatus.Status = common.TaskStatusStopped\n\t\tstatus.StoppedReason = \"success\"\n\t\treturn status, nil\n\t}\n\n\tif len(job.Status.Conditions) != 0 {\n\t\tglog.Infoln(\"task status conditions\", job.Status.Conditions[0], \"taskID\", taskID, \"requuid\", requuid)\n\n\t\tif job.Status.Conditions[0].Type == batchv1.JobComplete ||\n\t\t\tjob.Status.Conditions[0].Type == batchv1.JobFailed {\n\t\t\tstatus.Status = common.TaskStatusStopped\n\t\t\tstatus.StoppedReason = job.Status.Conditions[0].Message\n\t\t\treturn status, nil\n\t\t}\n\t}\n\n\treason := fmt.Sprintf(\"unknown task status, actively running pods %d, failed pods %d\", job.Status.Active, job.Status.Failed)\n\tglog.Infoln(reason, \"taskID\", taskID, \"requuid\", requuid, job.Status)\n\treturn status, nil\n}", "func (o *OpenapiTaskGenerationResult) GetTaskStatus() string {\n\tif o == nil || o.TaskStatus == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.TaskStatus\n}", "func (v1 *V1) GetTask(w http.ResponseWriter, r *http.Request) {\n\ttaskID := chi.URLParam(r, \"taskID\")\n\tshouldDeleteTask := false\n\tdeleteParam := r.URL.Query().Get(\"delete\")\n\tif deleteParam == \"1\" {\n\t\tshouldDeleteTask = true\n\t}\n\n\ttask := v1.metaCrawlSvc.TaskByID(taskID)\n\tif task == nil {\n\t\tv1.responseErrorJSON(w, \"task not found\", 404)\n\t\treturn\n\t}\n\n\ttaskStatus := task.Status()\n\tswitch taskStatus {\n\tcase metacrawl.TaskInProgress:\n\t\tv1.responseJSON(w, \"task in progress\", 204)\n\t\treturn\n\tcase metacrawl.TaskCompleted:\n\t\tif shouldDeleteTask {\n\t\t\tv1.metaCrawlSvc.DeleteTaskByID(taskID)\n\t\t}\n\n\t\tv1.responseCSV(w, taskID, task.Render(), 200)\n\t}\n}", "func GetVipStatusByTask(taskId int64) (VipStatusByTaskItems, error) {\r\n\t// generate url\r\n\turl := fmt.Sprintf(\"http://rms.baidu.com/?r=interface/api&handler=getBgwListCurrentStep&list_id=%d\", taskId)\r\n\r\n\t// request api for result\r\n\treturn getVipStatusByTask(url)\r\n}", "func (c *Client) GetTaskStatus(url string, paras *TaskPathParas,\n\trequest *StartTaskRequest) (*TaskStatusResponse, error) {\n\tif c == nil {\n\t\treturn nil, ErrServerNotInit\n\t}\n\n\tuserAuth, err := c.generateGateWayAuth(paras.Operator)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"bksops StartBkOpsTask generateGateWayAuth failed: %v\", err)\n\t}\n\n\tvar (\n\t\treqURL = fmt.Sprintf(\"/get_task_status/%s/%s/\", paras.TaskID, paras.BkBizID)\n\t\trespData = &TaskStatusResponse{}\n\t)\n\n\trequest.Scope = string(CmdbBizScope)\n\t_, _, errs := gorequest.New().\n\t\tTimeout(defaultTimeOut).\n\t\tGet(c.server+reqURL).\n\t\tSet(\"Content-Type\", \"application/json\").\n\t\tSet(\"Accept\", \"application/json\").\n\t\tSet(\"X-Bkapi-Authorization\", userAuth).\n\t\tSetDebug(c.serverDebug).\n\t\tSend(request).\n\t\tEndStruct(&respData)\n\tif len(errs) > 0 {\n\t\tblog.Errorf(\"call api GetTaskStatus failed: %v\", errs[0])\n\t\treturn nil, errs[0]\n\t}\n\n\tif !respData.Result {\n\t\tblog.Errorf(\"call api GetTaskStatus failed: %v\", respData.Message)\n\t\treturn nil, fmt.Errorf(respData.Message)\n\t}\n\n\t//successfully request\n\tblog.Infof(\"call api GetTaskStatus with url(%s) successfully\", reqURL)\n\treturn respData, nil\n}", "func HandleQueryTaskStatus(w http.ResponseWriter, r *http.Request) {\n\tlog.Root.Info(\"HandleQueryTaskStatus BEGIN\")\n\n\tif r.Method != http.MethodGet {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tHttpResponseError(w, ErrNotFound)\n\t\treturn\n\t}\n\n\ttaskIDList := r.URL.Query()[\"taskID\"]\n\tif len(taskIDList) == 0 {\n\t\tlog.Root.Error(\"HandleQueryTaskStatus Parse HTTP request param error\")\n\t\tHttpResponseError(w, ErrParams)\n\t\treturn\n\t}\n\n\ttaskStatusMap, err := node.QueryTaskStatus(taskIDList)\n\tif err != nil {\n\t\tlog.Root.Error(\"HandleQueryTaskStatus Query task status error. TaskIDList: %v\", taskIDList)\n\t\tHttpResponseError(w, ErrServer)\n\t\treturn\n\t}\n\n\ttaskStatusList := []interface{}{}\n\tfor k, v := range taskStatusMap {\n\t\tjsonMap := map[string]interface{}{\n\t\t\t\"taskID\": k,\n\t\t\t\"taskStatus\": v,\n\t\t}\n\t\ttaskStatusList = append(taskStatusList, jsonMap)\n\t}\n\n\tlog.Root.Info(\"HandleQueryTaskStatus END\")\n\tHttpResponseData(w, H{\n\t\t\"taskStatusList\": taskStatusList,\n\t})\n\treturn\n}", "func (p *ReadOnlySchedulerClient) GetTasksStatus(ctx context.Context, query *TaskQuery) (r *Response, err error) {\n var _args57 ReadOnlySchedulerGetTasksStatusArgs\n _args57.Query = query\n var _result58 ReadOnlySchedulerGetTasksStatusResult\n if err = p.Client_().Call(ctx, \"getTasksStatus\", &_args57, &_result58); err != nil {\n return\n }\n return _result58.GetSuccess(), nil\n}", "func (t *Task) GetStatus() int {\n\treturn t.Status\n}", "func (ctl *StatusController) getTask(ctx context.Context, name string) ([]TaskStatus, error) {\n\tif name == WildCard {\n\t\t// get status about all of tasks\n\t\ttasks, err := ctl.meta.GetAllTasks(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tresult := make([]TaskStatus, 0, len(tasks))\n\t\tfor _, task := range tasks {\n\t\t\tstatus, err := ctl.fillTask(ctx, task)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tresult = append(result, status)\n\t\t}\n\t\treturn result, nil\n\t}\n\t// get status about TaskName\n\ttask, err := ctl.meta.GetTask(ctx, name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tstatus, err := ctl.fillTask(ctx, *task)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn []TaskStatus{status}, nil\n}", "func ShowTask(w http.ResponseWriter, r *http.Request, repo *tasks.TaskRepository) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tparams := mux.Vars(r)\n\ttaskID, err := strconv.Atoi(params[\"id\"])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\ttask, err := repo.GetTask(taskID)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tjson.NewEncoder(w).Encode(apiIndexTask(task))\n}", "func (c *Client) GetTaskState(ctx context.Context, ID string) (*swarming_api.SwarmingRpcsTaskStates, error) {\n\tvar result *swarming_api.SwarmingRpcsTaskStates\n\tgetState := func() error {\n\t\tvar err error\n\t\tresult, err = c.SwarmingService.Tasks.GetStates().TaskId(ID).Context(ctx).Do()\n\t\treturn err\n\t}\n\tif err := callWithRetries(ctx, getState); err != nil {\n\t\treturn nil, errors.Annotate(err, fmt.Sprintf(\"get task state for task ID %s\", ID)).Err()\n\t}\n\treturn result, nil\n}", "func (p *ReadOnlySchedulerClient) GetTasksStatus(ctx context.Context, query *TaskQuery) (r *Response, err error) {\n var _args107 ReadOnlySchedulerGetTasksStatusArgs\n _args107.Query = query\n var _result108 ReadOnlySchedulerGetTasksStatusResult\n var meta thrift.ResponseMeta\n meta, err = p.Client_().Call(ctx, \"getTasksStatus\", &_args107, &_result108)\n p.SetLastResponseMeta_(meta)\n if err != nil {\n return\n }\n return _result108.GetSuccess(), nil\n}", "func getTaskInfoByContainerID(containerID string, tasks []TaskInfo) *TaskInfo {\n\tfor _, task := range tasks {\n\t\tif len(task.Statuses) > 0 && task.Statuses[0].ContainerStatusInfo.ID.Value == containerID {\n\t\t\treturn &task\n\t\t}\n\t}\n\treturn nil\n}", "func (ctl *StatusController) fillTask(ctx context.Context, task Task) (TaskStatus, error) {\n\tvar err error\n\ts := TaskStatus{\n\t\tInfo: task.Info,\n\t}\n\n\tif s.paused, err = task.IsPaused(ctx); err != nil {\n\t\treturn s, errors.Annotatef(err, \"failed to get pause status of task %s\", s.Info.Name)\n\t}\n\n\tif s.Checkpoints, err = task.NextBackupTSList(ctx); err != nil {\n\t\treturn s, errors.Annotatef(err, \"failed to get progress of task %s\", s.Info.Name)\n\t}\n\n\tif s.globalCheckpoint, err = task.GetStorageCheckpoint(ctx); err != nil {\n\t\treturn s, errors.Annotatef(err, \"failed to get storage checkpoint of task %s\", s.Info.Name)\n\t}\n\n\ts.LastErrors, err = task.LastError(ctx)\n\tif err != nil {\n\t\treturn s, err\n\t}\n\n\ts.QPS, err = MaybeQPS(ctx, ctl.mgr)\n\tif err != nil {\n\t\treturn s, errors.Annotatef(err, \"failed to get QPS of task %s\", s.Info.Name)\n\t}\n\treturn s, nil\n}", "func (s *Storage) GetTask(id uint) (*todopb.TaskResponse, error) {\n\trow := s.db.QueryRow(\"SELECT * FROM tasks WHERE id=$1\", id)\n\n\ttask, err := scan(row)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn task, nil\n}", "func (a *Client) GetTaskDetails(params *GetTaskDetailsParams, opts ...ClientOption) (*GetTaskDetailsOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetTaskDetailsParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"getTaskDetails\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/tasks/{id}\",\n\t\tProducesMediaTypes: []string{\"application/vnd.goswagger.examples.task-tracker.v1+json\"},\n\t\tConsumesMediaTypes: []string{\"application/vnd.goswagger.examples.task-tracker.v1+json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &GetTaskDetailsReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*GetTaskDetailsOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*GetTaskDetailsDefault)\n\treturn nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func (s *state) GetTask(exID string) (*mesos.Task, error) {\n\t// Check if task is in Launched Tasks list\n\tfor _, t := range s.st.GetTasks.LaunchedTasks {\n\t\tif s.isMatchingTask(&t, exID) {\n\t\t\treturn &t, nil\n\t\t}\n\t}\n\n\t// Check if task is in Queued Tasks list\n\tfor _, t := range s.st.GetTasks.QueuedTasks {\n\t\tif s.isMatchingTask(&t, exID) {\n\t\t\treturn &t, nil\n\t\t}\n\t}\n\treturn nil, fmt.Errorf(\"unable to find task matching executor id %s\", exID)\n}", "func GetTaskInformation(res http.ResponseWriter, req *http.Request) {\n\tres.Header().Set(\"Content-Type\", \"application/json\")\n\tc := Task{\"7E12E3B57A02\", \"FOOUSER\", \"BAR\"}\n\toutgoingJSON, err := json.Marshal(c)\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\thttp.Error(res, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tres.WriteHeader(http.StatusCreated)\n\tfmt.Fprint(res, string(outgoingJSON))\n}", "func (v *MatchingRespondQueryTaskCompletedRequest) GetTaskID() (o string) {\n\tif v != nil && v.TaskID != nil {\n\t\treturn *v.TaskID\n\t}\n\treturn\n}", "func (ctrl *TaskController) GetTask(w http.ResponseWriter, r *http.Request) {\n\ttaskId := ParamAsString(\"id\", r)\n\tlogrus.Println(\"task : \", taskId)\n\n\ttask, err := ctrl.taskDao.Get(taskId)\n\tif err != nil {\n\t\tlogrus.Error(err)\n\t\tSendJSONError(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tlogrus.Println(\"task : \", task)\n\tSendJSONOk(w, task)\n}", "func (db *Database) ReadTask(id string) (*models.Task, error) {\n\tctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)\n\tdefer cancel() \n\n\tcollection := db.Client.Database(\"dashboard-tasks\").Collection(\"tasks\")\n\n\ttask := models.Task{}\n\tobjectID, err := primitive.ObjectIDFromHex(id)\n\tif err != nil { return nil, err }\n\n\terr = collection.FindOne(ctx, bson.M{\"_id\": objectID}).Decode(&task)\n\tif err != nil { return nil, err }\n\n\treturn &task, err\n}", "func (t *Task) Status() TaskStatusType {\n\treturn t.status\n}", "func (d *Release) Status() *Task {\n\t_ = d.merge()\n\treturn d.task\n}", "func (cm *Docker) GetTaskMetrics(id string) (m models.Metrics, ok bool) {\n\tif config.GetSwitchVal(\"enableDisplay\") {\n\t\treturn\n\t}\n\tfor _, container := range cm.containers {\n\t\tnameWords := strings.Split(container.GetMeta(\"name\"), \".\")\n\t\tsize := len(nameWords)\n\t\tif size == 3 && id == nameWords[size-1] {\n\t\t\tlog.Debugf(\"LastMetrics %+v\", container.LastMetrics())\n\t\t\tm, ok = container.LastMetrics(), true\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func UpdateTaskStatus(tid int64, new_status int64) {\n\tvar dummy string\n\n\tif new_status == Running {\n\t\tdb.QueryRow(\"UPDATE tasks SET status=$1, start_time=now() WHERE id=$2\",\n\t\t\tnew_status, tid).Scan(&dummy)\n\t} else if new_status == Canceled {\n\t\tdb.QueryRow(\"UPDATE tasks SET status=$1, end_time=now() WHERE id=$2\",\n\t\t\tnew_status, tid).Scan(&dummy)\n\t} else {\n\t\tdb.QueryRow(\"UPDATE tasks SET status=$1 WHERE id=$2\", new_status, tid).\n\t\t\tScan(&dummy)\n\t}\n}", "func (p *taskController) GetTask(c echo.Context) error {\n\tid, err := strconv.Atoi(c.Param(\"id\"))\n\tif err != nil {\n\t\treturn echo.NewHTTPError(http.StatusBadRequest, \"Task ID must be int\")\n\t}\n\tctx := c.Request().Context()\n\tif ctx == nil {\n\t\tctx = context.Background()\n\t}\n\ttask, err := p.TaskUseCase.GetTask(ctx, id)\n\tif err != nil {\n\t\treturn echo.NewHTTPError(http.StatusNotFound, \"Task does not exist.\")\n\t}\n\treturn c.JSON(http.StatusOK, task)\n}", "func (cm *Docker) GetTask(id string) (*entity.Task, bool) {\n\tcm.lock.Lock()\n\tt, ok := cm.tasks[id]\n\tcm.lock.Unlock()\n\treturn t, ok\n}", "func (context Context) UpdateTaskStatus(id string, status string, statusMessage string) (err error) {\n\t_, err = context.UpdateTask(id, F{\"status\": status, \"status-message\": statusMessage})\n\treturn\n}", "func (d *Driver) RecoverTask(handle *drivers.TaskHandle) error {\n\tif handle == nil {\n\t\treturn fmt.Errorf(\"error: handle cannot be nil\")\n\t}\n\n\tif _, ok := d.tasks.Get(handle.Config.ID); ok {\n\t\treturn nil\n\t}\n\n\tvar taskState TaskState\n\tif err := handle.GetDriverState(&taskState); err != nil {\n\t\treturn fmt.Errorf(\"failed to decode task state from handle: %v\", err)\n\t}\n\td.logger.Debug(\"Checking for recoverable task\", \"task\", handle.Config.Name, \"taskid\", handle.Config.ID, \"container\", taskState.ContainerID)\n\n\tinspectData, err := d.podman.ContainerInspect(d.ctx, taskState.ContainerID)\n\tif err != nil {\n\t\td.logger.Warn(\"Recovery lookup failed\", \"task\", handle.Config.ID, \"container\", taskState.ContainerID, \"err\", err)\n\t\treturn nil\n\t}\n\n\th := &TaskHandle{\n\t\tcontainerID: taskState.ContainerID,\n\t\tdriver: d,\n\t\ttaskConfig: taskState.TaskConfig,\n\t\tprocState: drivers.TaskStateUnknown,\n\t\tstartedAt: taskState.StartedAt,\n\t\texitResult: &drivers.ExitResult{},\n\t\tlogger: d.logger.Named(\"podmanHandle\"),\n\n\t\ttotalCPUStats: stats.NewCpuStats(),\n\t\tuserCPUStats: stats.NewCpuStats(),\n\t\tsystemCPUStats: stats.NewCpuStats(),\n\n\t\tremoveContainerOnExit: d.config.GC.Container,\n\t}\n\n\tif inspectData.State.Running {\n\t\td.logger.Info(\"Recovered a still running container\", \"container\", inspectData.State.Pid)\n\t\th.procState = drivers.TaskStateRunning\n\t} else if inspectData.State.Status == \"exited\" {\n\t\t// are we allowed to restart a stopped container?\n\t\tif d.config.RecoverStopped {\n\t\t\td.logger.Debug(\"Found a stopped container, try to start it\", \"container\", inspectData.State.Pid)\n\t\t\tif err = d.podman.ContainerStart(d.ctx, inspectData.ID); err != nil {\n\t\t\t\td.logger.Warn(\"Recovery restart failed\", \"task\", handle.Config.ID, \"container\", taskState.ContainerID, \"err\", err)\n\t\t\t} else {\n\t\t\t\td.logger.Info(\"Restarted a container during recovery\", \"container\", inspectData.ID)\n\t\t\t\th.procState = drivers.TaskStateRunning\n\t\t\t}\n\t\t} else {\n\t\t\t// no, let's cleanup here to prepare for a StartTask()\n\t\t\td.logger.Debug(\"Found a stopped container, removing it\", \"container\", inspectData.ID)\n\t\t\tif err = d.podman.ContainerStart(d.ctx, inspectData.ID); err != nil {\n\t\t\t\td.logger.Warn(\"Recovery cleanup failed\", \"task\", handle.Config.ID, \"container\", inspectData.ID)\n\t\t\t}\n\t\t\th.procState = drivers.TaskStateExited\n\t\t}\n\t} else {\n\t\td.logger.Warn(\"Recovery restart failed, unknown container state\", \"state\", inspectData.State.Status, \"container\", taskState.ContainerID)\n\t\th.procState = drivers.TaskStateUnknown\n\t}\n\n\td.tasks.Set(taskState.TaskConfig.ID, h)\n\n\tgo h.runContainerMonitor()\n\td.logger.Debug(\"Recovered container handle\", \"container\", taskState.ContainerID)\n\n\treturn nil\n}", "func (r *versionResolver) TaskStatusStats(ctx context.Context, obj *restModel.APIVersion, options BuildVariantOptions) (*task.TaskStats, error) {\n\topts := task.GetTasksByVersionOptions{\n\t\tIncludeExecutionTasks: false,\n\t\tTaskNames: options.Tasks,\n\t\tVariants: options.Variants,\n\t\tStatuses: getValidTaskStatusesFilter(options.Statuses),\n\t\t// If the version is a patch, we don't want to include its never activated tasks.\n\t\tIncludeNeverActivatedTasks: !obj.IsPatchRequester(),\n\t}\n\n\tstats, err := task.GetTaskStatsByVersion(ctx, *obj.Id, opts)\n\tif err != nil {\n\t\treturn nil, InternalServerError.Send(ctx, fmt.Sprintf(\"getting version task status stats: %s\", err.Error()))\n\t}\n\treturn stats, nil\n}", "func (_Contract *ContractSession) GetTask(i *big.Int) (struct {\n\tActive bool\n\tAssignment *big.Int\n\tProposalID *big.Int\n}, error) {\n\treturn _Contract.Contract.GetTask(&_Contract.CallOpts, i)\n}", "func (_Contract *ContractCallerSession) GetTask(i *big.Int) (struct {\n\tActive bool\n\tAssignment *big.Int\n\tProposalID *big.Int\n}, error) {\n\treturn _Contract.Contract.GetTask(&_Contract.CallOpts, i)\n}", "func updateTaskState(task *api.Task) api.TaskStatus {\n\t//The task is the minimum status of all its essential containers unless the\n\t//status is terminal in which case it's that status\n\tlog.Debug(\"Updating task\", \"task\", task)\n\n\t// minContainerStatus is the minimum status of all essential containers\n\tminContainerStatus := api.ContainerDead + 1\n\t// minContainerStatus is the minimum status of all containers to be used in\n\t// the edge case of no essential containers\n\tabsoluteMinContainerStatus := minContainerStatus\n\tfor _, cont := range task.Containers {\n\t\tlog.Debug(\"On container\", \"cont\", cont)\n\t\tif cont.KnownStatus < absoluteMinContainerStatus {\n\t\t\tabsoluteMinContainerStatus = cont.KnownStatus\n\t\t}\n\t\tif !cont.Essential {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Terminal states\n\t\tif cont.KnownStatus == api.ContainerStopped {\n\t\t\tif task.KnownStatus < api.TaskStopped {\n\t\t\t\ttask.KnownStatus = api.TaskStopped\n\t\t\t\treturn task.KnownStatus\n\t\t\t}\n\t\t} else if cont.KnownStatus == api.ContainerDead {\n\t\t\tif task.KnownStatus < api.TaskDead {\n\t\t\t\ttask.KnownStatus = api.TaskDead\n\t\t\t\treturn task.KnownStatus\n\t\t\t}\n\t\t}\n\t\t// Non-terminal\n\t\tif cont.KnownStatus < minContainerStatus {\n\t\t\tminContainerStatus = cont.KnownStatus\n\t\t}\n\t}\n\n\tif minContainerStatus == api.ContainerDead+1 {\n\t\tlog.Warn(\"Task with no essential containers; all properly formed tasks should have at least one essential container\", \"task\", task)\n\n\t\t// If there's no essential containers, let's just assume the container\n\t\t// with the earliest status is essential and proceed.\n\t\tminContainerStatus = absoluteMinContainerStatus\n\t}\n\n\tlog.Info(\"MinContainerStatus is \" + minContainerStatus.String())\n\n\tif minContainerStatus == api.ContainerCreated {\n\t\tif task.KnownStatus < api.TaskCreated {\n\t\t\ttask.KnownStatus = api.TaskCreated\n\t\t\treturn task.KnownStatus\n\t\t}\n\t} else if minContainerStatus == api.ContainerRunning {\n\t\tif task.KnownStatus < api.TaskRunning {\n\t\t\ttask.KnownStatus = api.TaskRunning\n\t\t\treturn task.KnownStatus\n\t\t}\n\t} else if minContainerStatus == api.ContainerStopped {\n\t\tif task.KnownStatus < api.TaskStopped {\n\t\t\ttask.KnownStatus = api.TaskStopped\n\t\t\treturn task.KnownStatus\n\t\t}\n\t} else if minContainerStatus == api.ContainerDead {\n\t\tif task.KnownStatus < api.TaskDead {\n\t\t\ttask.KnownStatus = api.TaskDead\n\t\t\treturn task.KnownStatus\n\t\t}\n\t}\n\treturn api.TaskStatusNone\n}", "func VerifyTask(ctx context.Context, taskObj *v1beta1.Task, k8s kubernetes.Interface, refSource *v1.RefSource, verificationpolicies []*v1alpha1.VerificationPolicy) VerificationResult {\n\treturn VerifyResource(ctx, taskObj, k8s, refSource, verificationpolicies)\n}", "func (filter TaskReliabilityFilter) GetTaskStats() (taskStats []taskstats.TaskStats, err error) {\n\tpipeline := filter.taskReliabilityQueryPipeline()\n\terr = db.Aggregate(taskstats.DailyTaskStatsCollection, pipeline, &taskStats)\n\treturn\n}", "func (m *Master) GetTask(_ *ExampleArgs, reply *GetTaskReply) error {\n\tswitch m.masterState {\n\tcase newMaster:\n\t\tfor i, task := range m.mapTask {\n\t\t\tif task.State == initialState {\n\t\t\t\treply.Task.Type_ = task.Type_\n\t\t\t\treply.Task.Filename = task.Filename\n\t\t\t\treply.Task.Id = task.Id\n\t\t\t\treply.Task.NReduce = task.NReduce\n\t\t\t\treply.Flag = 0\n\t\t\t\treply.Task.State = task.State\n\n\t\t\t\tm.mapTask[i].State = inProgress\n\t\t\t\tm.mapTask[i].Time = time.Now()\n\t\t\t\t//reply.Task.State=m.mapTask[i].State\n\n\t\t\t\treturn nil\n\t\t\t} else if task.State == inProgress && time.Now().Sub(m.mapTask[i].Time) > time.Duration(5)*time.Second {\n\t\t\t\treply.Task.Type_ = task.Type_\n\t\t\t\treply.Task.Filename = task.Filename\n\t\t\t\treply.Task.Id = task.Id\n\t\t\t\treply.Task.NReduce = task.NReduce\n\t\t\t\treply.Task.State = task.State\n\t\t\t\treply.Flag = 0\n\n\t\t\t\tm.mapTask[i].State = inProgress\n\t\t\t\tm.mapTask[i].Time = time.Now()\n\t\t\t\t//reply.Task.State=m.mapTask[i].State\n\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t\treply.Flag = 1 // map not finished but in progress\n\tcase completeMap:\n\t\tfor i, task := range m.reduceTask {\n\t\t\tif task.State == initialState {\n\t\t\t\treply.Task.Type_ = task.Type_\n\t\t\t\treply.Task.Filename = task.Filename\n\t\t\t\treply.Task.Id = task.Id\n\t\t\t\treply.Task.NReduce = task.NReduce\n\t\t\t\treply.Flag = 0\n\t\t\t\treply.Task.Files = task.Files\n\t\t\t\treply.Task.State = task.State\n\n\t\t\t\tm.reduceTask[i].State = inProgress\n\t\t\t\tm.reduceTask[i].Time = time.Now()\n\t\t\t\t//reply.Task.State=m.mapTask[i].State\n\n\t\t\t\treturn nil\n\t\t\t} else if task.State == inProgress && time.Now().Sub(m.reduceTask[i].Time) > time.Duration(5)*time.Second {\n\t\t\t\treply.Task.Type_ = task.Type_\n\t\t\t\treply.Task.Filename = task.Filename\n\t\t\t\treply.Task.Id = task.Id\n\t\t\t\treply.Task.NReduce = task.NReduce\n\t\t\t\treply.Flag = 0\n\t\t\t\treply.Task.Files = task.Files\n\t\t\t\treply.Task.State = task.State\n\n\t\t\t\tm.reduceTask[i].State = inProgress\n\t\t\t\tm.reduceTask[i].Time = time.Now()\n\t\t\t\t//reply.Task.State=m.mapTask[i].State\n\n\t\t\t\treturn nil\n\n\t\t\t}\n\t\t}\n\t\treply.Flag = 1 // reduce not finished but in progress\n\tcase completeReduce:\n\t\treply.Flag = 2 // all task have been finished\n\n\t}\n\n\treturn nil\n}", "func GetTaskState(taskID string) (*State, error) {\n\ttasks.RLock()\n\tdefer tasks.RUnlock()\n\n\tif _, ok := tasks.m[taskID]; ok { // key exist\n\t\treturn tasks.m[taskID], nil\n\t}\n\n\treturn nil, errors.New(\"Task not exist.\")\n}", "func (o *ViewCustomFieldTask) GetTaskOk() (*ViewRelationship, bool) {\n\tif o == nil || o.Task == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Task, true\n}", "func GetTask(id int) (Task, error) {\n\tvar t Task\n\tvar jsonTask []byte\n\n\terr := db.View(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket(taskBucket)\n\t\tjsonTask = b.Get(itob(id))\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn t, err\n\t}\n\n\terr = t.ReadFromJSON(jsonTask)\n\tif err != nil {\n\t\treturn t, err\n\t}\n\n\treturn t, nil\n\n}", "func (d *Driver) TaskStats(ctx context.Context, taskID string, interval time.Duration) (<-chan *drivers.TaskResourceUsage, error) {\n\td.logger.Debug(\"TaskStats called\", \"taskID\", taskID)\n\thandle, ok := d.tasks.Get(taskID)\n\tif !ok {\n\t\treturn nil, drivers.ErrTaskNotFound\n\t}\n\tstatsChannel := make(chan *drivers.TaskResourceUsage)\n\tgo handle.runStatsEmitter(ctx, statsChannel, interval)\n\treturn statsChannel, nil\n}", "func (_Contract *ContractCaller) GetTask(opts *bind.CallOpts, i *big.Int) (struct {\n\tActive bool\n\tAssignment *big.Int\n\tProposalID *big.Int\n}, error) {\n\tvar out []interface{}\n\terr := _Contract.contract.Call(opts, &out, \"getTask\", i)\n\n\toutstruct := new(struct {\n\t\tActive bool\n\t\tAssignment *big.Int\n\t\tProposalID *big.Int\n\t})\n\tif err != nil {\n\t\treturn *outstruct, err\n\t}\n\n\toutstruct.Active = *abi.ConvertType(out[0], new(bool)).(*bool)\n\toutstruct.Assignment = *abi.ConvertType(out[1], new(*big.Int)).(**big.Int)\n\toutstruct.ProposalID = *abi.ConvertType(out[2], new(*big.Int)).(**big.Int)\n\n\treturn *outstruct, err\n\n}", "func (client *Client) DescribeActiveOperationTask(request *DescribeActiveOperationTaskRequest) (_result *DescribeActiveOperationTaskResponse, _err error) {\n\truntime := &util.RuntimeOptions{}\n\t_result = &DescribeActiveOperationTaskResponse{}\n\t_body, _err := client.DescribeActiveOperationTaskWithOptions(request, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_result = _body\n\treturn _result, _err\n}", "func GetSingleTask(w http.ResponseWriter, r *http.Request) {\n\tparams := mux.Vars(r)\n\n\tif params[\"id\"] == \"\" {\n\t\thttp.Error(w, http.StatusText(400), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\ttaskID := bson.ObjectIdHex(params[\"id\"])\n\n\ttask, err := repository.GetSingleTask(taskID)\n\n\tif err != nil {\n\t\thttp.Error(w, http.StatusText(404), http.StatusNotFound)\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\n\tjson.NewEncoder(w).Encode(task)\n}", "func (m *Master) GetTask(req *GetTaskReq, rsp *GetTaskRsp) error {\n\tm.mutex.Lock()\n\tdefer m.mutex.Unlock()\n\tfor k := range m.todoMapTask {\n\t\tif m.todoMapTask[k] == 0 {\n\t\t\trsp.Status = \"Task\"\n\t\t\trsp.Filename = m.files[k]\n\t\t\trsp.NReduce = m.nReduce\n\t\t\trsp.TaskID = k\n\t\t\tm.todoMapTask[k] = time.Now().Unix()\n\t\t\treturn nil\n\t\t}\n\t}\n\tif len(m.todoMapTask) != 0 {\n\t\trsp.Status = \"Wait\"\n\t\treturn nil\n\t}\n\n\tfor k := range m.todoReduceTask {\n\t\tif m.todoReduceTask[k] == 0 {\n\t\t\trsp.Status = \"Task\"\n\t\t\trsp.NReduce = m.nReduce\n\t\t\trsp.NMap = len(m.files)\n\t\t\trsp.TaskID = k\n\t\t\tm.todoReduceTask[k] = time.Now().Unix()\n\t\t\treturn nil\n\t\t}\n\t}\n\n\tif len(m.todoReduceTask) != 0 {\n\t\trsp.Status = \"Wait\"\n\t\treturn nil\n\t} else {\n\t\trsp.Status = \"Exit\"\n\t\treturn nil\n\t}\n\n\treturn nil\n}", "func (c *BasicECSClient) DescribeTaskDefinition(ctx context.Context, in *ecs.DescribeTaskDefinitionInput) (*ecs.DescribeTaskDefinitionOutput, error) {\n\tif err := c.setup(); err != nil {\n\t\treturn nil, errors.Wrap(err, \"setting up client\")\n\t}\n\n\tvar out *ecs.DescribeTaskDefinitionOutput\n\tvar err error\n\tmsg := awsutil.MakeAPILogMessage(\"DescribeTaskDefinition\", in)\n\tif err := utility.Retry(ctx,\n\t\tfunc() (bool, error) {\n\t\t\tout, err = c.ecs.DescribeTaskDefinitionWithContext(ctx, in)\n\t\t\tif awsErr, ok := err.(awserr.Error); ok {\n\t\t\t\tgrip.Debug(message.WrapError(awsErr, msg))\n\t\t\t\tif c.isNonRetryableErrorCode(awsErr.Code()) {\n\t\t\t\t\treturn false, err\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn true, err\n\t\t}, *c.opts.RetryOpts); err != nil {\n\t\treturn nil, err\n\t}\n\treturn out, nil\n}", "func (t *Task) Status() string {\n\tstatus := t.State.Symbol() + t.DueSymbol() + \" \"\n\n\treturn status\n}", "func GetTask(id int) (Task, error) {\n\tpath := fmt.Sprintf(\"tasks/%d\", id)\n\tres, err := makeRequest(http.MethodGet, path, nil)\n\tif err != nil {\n\t\treturn Task{}, err\n\t}\n\n\treturn decodeTask(res.Body)\n}", "func GetTask(tid, user_token string) (*Task, error) {\n\t// declarations\n\tvar start_time, end_time pq.NullTime\n\tvar exit_status sql.NullInt64\n\tvar output sql.NullString\n\n\t// initialize Task\n\ttask := Task{}\n\t// get task information\n\tif err := db.QueryRow(\"SELECT * FROM tasks WHERE tasks.id=$1\", tid).\n\t\tScan(&task.Id, &task.Gid, &start_time, &end_time, &task.Status,\n\t\t&exit_status, &output, &task.Patch); err != nil {\n\t\treturn nil, err\n\t}\n\t// set remaining fields\n\tif start_time.Valid {\n\t\ttask.Start_time = &start_time.Time\n\t}\n\tif end_time.Valid {\n\t\ttask.End_time = &end_time.Time\n\t}\n\tif exit_status.Valid {\n\t\ttask.Exit_status = exit_status.Int64\n\t}\n\tif output.Valid {\n\t\ttask.Output = output.String\n\t}\n\n\tgroup_task, _ := getGroupTask(task.Gid)\n\ttask.User = group_task.user\n\ttask.Project = group_task.project\n\ttask.Bot = group_task.bot\n\n\treturn &task, nil\n}", "func (s *projService) GetTeamMemberByTask(ctx context.Context, req *pb.GetTeamMemberByTaskRequest) (*pb.GetTeamMemberByTaskResponse, error) {\n\tresp := &pb.GetTeamMemberByTaskResponse{}\n\tvar err error\n\n\tsqlstring1 := `SELECT inbProjectId FROM tb_Task WHERE inbTaskId = ? AND inbMserviceId = ? AND bitIsDeleted = 0`\n\tstmt1, err := s.db.Prepare(sqlstring1)\n\tif err != nil {\n\t\tlevel.Error(s.logger).Log(\"what\", \"Prepare\", \"error\", err)\n\t\tresp.ErrorCode = 500\n\t\tresp.ErrorMessage = \"db.Prepare failed\"\n\t\treturn resp, nil\n\t}\n\n\tdefer stmt1.Close()\n\n\tvar existingProjectId int64\n\terr = stmt1.QueryRow(req.GetTaskId(), req.GetMserviceId()).Scan(&existingProjectId)\n\tif err != nil {\n\t\tresp.ErrorCode = 404\n\t\tresp.ErrorMessage = \"referenced task not found\"\n\t\treturn resp, nil\n\t}\n\n\tsqlstring := `SELECT m.inbMemberId, m.dtmCreated, m.dtmModified, m.intVersion,\n\tm.inbMserviceId, m.inbProjectId, m.chvName, m.intProjectRoleId, m.chvEmail, t.decTaskHours, r.chvRoleName \n\tFROM tb_TaskToMember AS t \n\tJOIN tb_TeamMember AS m ON t.inbMemberId = m.inbMemberId\n\tJOIN tb_ProjectRoleType AS r ON m.intProjectRoleId = r.intProjectRoleId\n\tWHERE t.inbProjectId = ? AND t.inbTaskId= ? AND t.inbMserviceId = ?\n\tAND t.bitIsDeleted = 0 AND m.bitIsDeleted = 0`\n\tstmt, err := s.db.Prepare(sqlstring)\n\tif err != nil {\n\t\tlevel.Error(s.logger).Log(\"what\", \"Prepare\", \"error\", err)\n\t\tresp.ErrorCode = 500\n\t\tresp.ErrorMessage = \"db.Prepare failed\"\n\t\treturn resp, nil\n\t}\n\n\tdefer stmt.Close()\n\trows, err := stmt.Query(existingProjectId, req.GetTaskId(), req.GetMserviceId())\n\tif err != nil {\n\t\tlevel.Error(s.logger).Log(\"what\", \"Query\", \"error\", err)\n\t\tresp.ErrorCode = 500\n\t\tresp.ErrorMessage = err.Error()\n\t\treturn resp, nil\n\t}\n\n\tdefer rows.Close()\n\tfor rows.Next() {\n\t\tvar created string\n\t\tvar modified string\n\t\tvar task_hours string\n\t\tvar member pb.TeamMember\n\n\t\terr := rows.Scan(&member.MemberId, &created, &modified, &member.Version, &member.MserviceId, &member.ProjectId,\n\t\t\t&member.Name, &member.ProjectRoleId, &member.Email, &task_hours, &member.RoleName)\n\n\t\tif err != nil {\n\t\t\tlevel.Error(s.logger).Log(\"what\", \"Scan\", \"error\", err)\n\t\t\tresp.ErrorCode = 500\n\t\t\tresp.ErrorMessage = err.Error()\n\t\t\treturn resp, nil\n\t\t}\n\n\t\tmember.Created = dml.DateTimeFromString(created)\n\t\tmember.Modified = dml.DateTimeFromString(modified)\n\t\td, err := dml.DecimalFromString(task_hours)\n\t\tif err == nil {\n\t\t\tmember.TaskHours = d\n\t\t}\n\n\t\tresp.TeamMembers = append(resp.TeamMembers, &member)\n\t}\n\n\treturn resp, err\n}", "func (builder *Builder) GetStatus(task string) bool {\n\tstatus, ok := builder.LinuxBuild.Status[task]\n\treturn status && ok\n}", "func (t Task) StatusDisplay() string {\n\tswitch t.Status {\n\tcase taskStatusSucceeded:\n\t\treturn \"succeeded\"\n\tcase taskStatusProcessing:\n\t\treturn \"processing\"\n\tcase taskStatusFailed:\n\t\treturn \"failed\"\n\tcase taskStatusCanceled:\n\t\treturn \"canceled\"\n\t}\n\n\treturn \"waiting\"\n}", "func (t *TaskService) Edit(w http.ResponseWriter, r *http.Request) {\n\tvar (\n\t\temptyUUID gocql.UUID\n\t\ttaskIDStr = mux.Vars(r)[\"taskID\"]\n\t\tpartnerID = mux.Vars(r)[partnerIDKey]\n\t\tctx = r.Context()\n\t\tcurrentUser = t.userService.GetUser(r, t.httpClient)\n\t\tmodifiedAt = time.Now().Truncate(time.Millisecond).UTC()\n\t)\n\n\ttaskID, err := gocql.ParseUUID(taskIDStr)\n\tif err != nil || taskID == emptyUUID {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorTaskIDHasBadFormat, \"TaskService.Edit: task ID(UUID=%s) has bad format or empty. err=%v\", taskIDStr, err)\n\t\tcommon.SendBadRequest(w, r, errorcode.ErrorTaskIDHasBadFormat)\n\t\treturn\n\t}\n\n\tinputTask, err := t.extractPostTaskPayload(r, w)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tinputTask.PartnerID = partnerID\n\tinternalTasks, err := t.taskPersistence.GetByIDs(ctx, nil, inputTask.PartnerID, false, taskID)\n\tif err != nil && err != gocql.ErrNotFound {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantGetTaskByTaskID, \"TaskService.Edit: can not get internal tasks by Task ID %v. err=%v\", taskID, err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantGetTaskByTaskID)\n\t\treturn\n\t}\n\n\tif len(internalTasks) == 0 {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantGetTaskByTaskID, \"TaskService.Edit: can not get internal tasks by Task ID %v. err=%v\", taskID, err)\n\t\tcommon.SendNotFound(w, r, errorcode.ErrorCantGetTaskByTaskID)\n\t\treturn\n\t}\n\n\ttask := internalTasks[0]\n\ttask.Schedule = inputTask.Schedule\n\ttask.Schedule.StartRunTime = task.Schedule.StartRunTime.Truncate(time.Minute)\n\ttask.Schedule.EndRunTime = task.Schedule.EndRunTime.Truncate(time.Minute)\n\ttask.ModifiedBy = currentUser.UID()\n\ttask.ModifiedAt = modifiedAt\n\ttask.DefinitionID = inputTask.DefinitionID\n\ttask.OriginID = inputTask.OriginID\n\ttask.PartnerID = partnerID\n\n\ttask.TargetsByType = inputTask.TargetsByType\n\tif task.TargetsByType == nil {\n\t\ttask.TargetsByType = make(models.TargetsByType)\n\t}\n\n\tif inputTask.Targets.Type != 0 {\n\t\ttask.TargetsByType[inputTask.Targets.Type] = inputTask.Targets.IDs\n\t}\n\n\tfor targetType, targets := range task.TargetsByType {\n\t\ttask.Targets.Type = targetType\n\t\ttask.Targets.IDs = targets\n\t}\n\n\tif len(inputTask.Parameters) > 0 {\n\t\ttask.Parameters = inputTask.Parameters\n\t}\n\n\tfor i := range internalTasks {\n\t\tinternalTasks[i].OriginalNextRunTime = time.Time{}\n\t\tif internalTasks[i].State != statuses.TaskStateDisabled {\n\t\t\tinternalTasks[i].State = statuses.TaskStateInactive\n\t\t}\n\t\tinternalTasks[i].ModifiedBy = currentUser.UID()\n\t}\n\n\tt.processEditReq(ctx, internalTasks, r, w, currentUser, task)\n}", "func (this *TodoList) GetTask(pID string) (rFound *Task) {\n\tif this != nil {\n\t\tfor _, zTask := range this.Tasks {\n\t\t\tif zTask.Id == pID {\n\t\t\t\treturn zTask\n\t\t\t}\n\t\t}\n\t}\n\treturn\n}", "func UpdateEventTaskStatus(etid int64, status int) error {\n\tvar dummy string\n\tif err := db.QueryRow(\"UPDATE event_tasks SET status=$1 WHERE id=$2 \"+\n\t\t\"RETURNING id\", status, etid).Scan(&dummy); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func GetTask(mapf func(string, string) []KeyValue,\n\treducef func(string, []string) string) (bool, bool, string, int) {\n\n\t// declare an argument structure.\n\targs := GetTaskArgs{}\n\n\t// fill in the argument(s).\n\n\t// declare a reply structure.\n\treply := GetTaskReply{}\n\n\t// send the RPC request, wait for the reply.\n\tcallSuccess := call(\"Master.GetTask\", &args, &reply)\n\ttaskSuccess := false\n\tif (callSuccess) {\n\t\ttaskSuccess = DoTask(reply, mapf, reducef);\n\t}\n\n\treturn callSuccess, taskSuccess, reply.TaskType, reply.TaskNumber;\n}", "func (d *DeploymentRequest) GetTask() string {\n\tif d == nil || d.Task == nil {\n\t\treturn \"\"\n\t}\n\treturn *d.Task\n}", "func (c Context) TaskID() string {\n\treturn c.Current().ID\n}", "func (d *Deployment) GetTask() string {\n\tif d == nil || d.Task == nil {\n\t\treturn \"\"\n\t}\n\treturn *d.Task\n}", "func GetTask(id bson.ObjectId) (*Task, error) {\n\ttask := Task{}\n\terr := sess.DB(\"\").C(taskC).FindId(id).One(&task)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &task, nil\n}", "func (m *LongRunningOperation) GetStatusDetail()(*string) {\n val, err := m.GetBackingStore().Get(\"statusDetail\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (c *Client) GetTaskURL(taskID string) string {\n\treturn TaskURL(c.server, taskID)\n}", "func (*Task) Descriptor() ([]byte, []int) {\n\treturn file_api_proto_tasklist_server_proto_rawDescGZIP(), []int{0}\n}", "func (t *TaskBox[T, U, C, CT, TF]) GetStatus() int32 {\n\treturn t.status.Load()\n}", "func (a *Client) GetTaskInfo(params *GetTaskInfoParams, opts ...ClientOption) (*GetTaskInfoOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetTaskInfoParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"getTaskInfo\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/astrolabe/tasks/{taskID}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &GetTaskInfoReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*GetTaskInfoOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for getTaskInfo: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (db *DynamoDB) GetTask(ctx context.Context, req *tes.GetTaskRequest) (*tes.Task, error) {\n\tvar task *tes.Task\n\tvar response *dynamodb.GetItemOutput\n\tvar err error\n\n\tswitch req.View {\n\tcase tes.TaskView_MINIMAL:\n\t\tresponse, err = db.getMinimalView(ctx, req.Id)\n\tcase tes.TaskView_BASIC:\n\t\tresponse, err = db.getBasicView(ctx, req.Id)\n\tcase tes.TaskView_FULL:\n\t\tresponse, err = db.getFullView(ctx, req.Id)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif response.Item == nil {\n\t\treturn nil, grpc.Errorf(codes.NotFound, fmt.Sprintf(\"%v: taskID: %s\", errNotFound.Error(), req.Id))\n\t}\n\n\terr = dynamodbattribute.UnmarshalMap(response.Item, &task)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to DynamoDB unmarshal Task, %v\", err)\n\t}\n\n\treturn task, nil\n}", "func (s *Session) status() (string, error) {\n\tif s.task == nil {\n\t\treturn \"\", errors.New(\"snap task is not running or not found\")\n\t}\n\n\ttask := s.pClient.GetTask(s.task.ID)\n\tif task.Err != nil {\n\t\treturn \"\", errors.Wrapf(task.Err, \"could not get task name:%q, ID:%q\",\n\t\t\ts.task.Name, s.task.ID)\n\t}\n\n\treturn task.State, nil\n}", "func (s *Task) GetStatus() Status {\n\treturn s.results.GetStatus()\n}", "func GetTask(r *http.Request) *task.Task {\n\tif rv := r.Context().Value(model.ApiTaskKey); rv != nil {\n\t\tif t, ok := rv.(*task.Task); ok {\n\t\t\treturn t\n\t\t}\n\t}\n\treturn nil\n}", "func (handler *PreflightCheckHandler) CurrentTaskStatus() *TaskStatus {\n\ttaskStatus := handler.shellTask.CurrentTaskStatus()\n\ttaskStatus.Info = util.NewBuffer(1)\n\treturn taskStatus\n}", "func taskInfo(\n\tt *testing.T, task int,\n) (cmdline []string, env []string, ppid int, children []int) {\n\t// Get the command line of the task.\n\tcmd, err := ioutil.ReadFile(fmt.Sprintf(\"/proc/%d/cmdline\", task))\n\tTestExpectSuccess(t, err)\n\tcmdList := bytes.Split(cmd, []byte{0})\n\tcmdline = make([]string, len(cmdList)-1)\n\tfor i := range cmdline {\n\t\tcmdline[i] = string(cmdList[i])\n\t}\n\n\t// Get the environment of the process.\n\tenvData, err := ioutil.ReadFile(fmt.Sprintf(\"/proc/%d/environ\", task))\n\tTestExpectSuccess(t, err)\n\tenvList := bytes.Split(envData, []byte{0})\n\tenv = make([]string, len(envList)-1)\n\tfor i := range env {\n\t\tenv[i] = string(envList[i])\n\t}\n\t// Sort the environment for consistency.\n\tsort.Strings(env)\n\n\t// Get the pid of the parent process.\n\tstat, err := ioutil.ReadFile(fmt.Sprintf(\"/proc/%d/stat\", task))\n\tTestExpectSuccess(t, err)\n\tfields := strings.Split(string(stat), \" \")\n\tif len(fields) < 4 {\n\t\tFatalf(t, \"Unknown output in /proc/%d/stat: %d\", task, string(stat))\n\t}\n\tppid, err = strconv.Atoi(fields[3])\n\tTestExpectSuccess(t, err)\n\n\t// Now get a list of all of this tasks children.\n\tcldrn, err := ioutil.ReadFile(\n\t\tfmt.Sprintf(\"/proc/%d/task/%d/children\", task, task))\n\tTestExpectSuccess(t, err)\n\tchildrenStrs := strings.Split(string(cldrn), \" \")\n\tchildren = make([]int, len(childrenStrs)-1)\n\tfor i := range children {\n\t\tchildren[i], err = strconv.Atoi(childrenStrs[i])\n\t\tTestExpectSuccess(t, err)\n\t}\n\n\t// Success\n\treturn\n}", "func TaskNumGET(g *gin.Context) {\n\tg.JSON(http.StatusOK, gin.H{\"message\": \"ok\", \"num\": TaskNum})\n}", "func convertTaskToResult(testID string, task *swarmingAPI.SwarmingRpcsTaskResult, req *pb.DeriveChromiumInvocationRequest) (*pb.TestResult, error) {\n\tresultStatus := getTaskResultStatus(task)\n\tret := &pb.TestResult{\n\t\t// Use ninja target as test_id.\n\t\tTestId: testID,\n\t\tExpected: resultStatus == pb.TestStatus_PASS,\n\t\tStatus: resultStatus,\n\t}\n\n\t// Add the swarming task's url and state to summaryHTML.\n\tbuf := &strings.Builder{}\n\terr := summaryTmpl.Execute(buf, map[string]interface{}{\n\t\t\"url\": fmt.Sprintf(\"https://%s/task?id=%s\", req.SwarmingTask.Hostname, req.SwarmingTask.Id),\n\t\t\"state\": task.State,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tret.SummaryHtml = buf.String()\n\n\treturn ret, nil\n}", "func taskComplete(task string){\n\tfmt.Println(task)\n\tid, _ := primitive.ObjectIDFromHex(task)\n\tfilter := bson.M{\"_id\": id}\n\tupdate := bson.M{\"$set\": bson.M{\"status\": true}}\n\t_, err := collection.UpdateOne(context.Background(), filter, update)\n\tif err != nil {\n\t\tlog.Fatal(\"Error update task\", err)\n\t}\n\n\tfmt.Println(\"Task defined complete: \", id)\n}", "func taskDescriptionForTaskFlowLog(tsk sh_task.Task, action string, phase string, status string) string {\n\thm := task.HookMetadataAccessor(tsk)\n\n\tparts := make([]string, 0)\n\n\tswitch action {\n\tcase \"start\":\n\t\tparts = append(parts, fmt.Sprintf(\"%s task\", tsk.GetType()))\n\tcase \"end\":\n\t\tparts = append(parts, fmt.Sprintf(\"%s task done, result is '%s'\", tsk.GetType(), status))\n\tdefault:\n\t\tparts = append(parts, fmt.Sprintf(\"%s task %s\", action, tsk.GetType()))\n\t}\n\n\tparts = append(parts, \"for\")\n\n\tswitch tsk.GetType() {\n\tcase task.GlobalHookRun, task.ModuleHookRun:\n\t\t// Examples:\n\t\t// head task GlobalHookRun for 'beforeAll' binding, trigger AfterAll-Hooks-Change-DynamicEnabled\n\t\t// GlobalHookRun task for 'beforeAll' binding, trigger AfterAll-Hooks-Change-DynamicEnabled\n\t\t// GlobalHookRun task done, result 'Repeat' for 'beforeAll' binding, trigger AfterAll-Hooks-Change-DynamicEnabled\n\t\t// GlobalHookRun task for 'onKubernetes/cni_name' binding, trigger Kubernetes\n\t\t// GlobalHookRun task done, result 'Repeat' for 'onKubernetes/cni_name' binding, trigger Kubernetes\n\t\t// GlobalHookRun task for 'main' group binding, trigger Schedule\n\t\t// GlobalHookRun task done, result 'Fail' for 'main' group binding, trigger Schedule\n\t\t// GlobalHookRun task for 'main' group and 2 more bindings, trigger Schedule\n\t\t// GlobalHookRun task done, result 'Fail' for 'main' group and 2 more bindings, trigger Schedule\n\t\t// GlobalHookRun task for Synchronization of 'kubernetes/cni_name' binding, trigger KubernetesEvent\n\t\t// GlobalHookRun task done, result 'Success' for Synchronization of 'kubernetes/cni_name' binding, trigger KubernetesEvent\n\n\t\tif len(hm.BindingContext) > 0 {\n\t\t\tif hm.BindingContext[0].IsSynchronization() {\n\t\t\t\tparts = append(parts, \"Synchronization of\")\n\t\t\t}\n\n\t\t\tbindingType := hm.BindingContext[0].Metadata.BindingType\n\n\t\t\tgroup := hm.BindingContext[0].Metadata.Group\n\t\t\tif group == \"\" {\n\t\t\t\tname := hm.BindingContext[0].Binding\n\t\t\t\tif bindingType == htypes.OnKubernetesEvent || bindingType == htypes.Schedule {\n\t\t\t\t\tname = fmt.Sprintf(\"'%s/%s'\", bindingType, name)\n\t\t\t\t} else {\n\t\t\t\t\tname = string(bindingType)\n\t\t\t\t}\n\t\t\t\tparts = append(parts, name)\n\t\t\t} else {\n\t\t\t\tparts = append(parts, fmt.Sprintf(\"'%s' group\", group))\n\t\t\t}\n\n\t\t\tif len(hm.BindingContext) > 1 {\n\t\t\t\tparts = append(parts, \"and %d more bindings\")\n\t\t\t} else {\n\t\t\t\tparts = append(parts, \"binding\")\n\t\t\t}\n\t\t} else {\n\t\t\tparts = append(parts, \"no binding\")\n\t\t}\n\n\tcase task.ConvergeModules:\n\t\t// Examples:\n\t\t// ConvergeModules task for ReloadAllModules in phase 'WaitBeforeAll', trigger Operator-Startup\n\t\t// ConvergeModules task for KubeConfigChanged, trigger Operator-Startup\n\t\t// ConvergeModules task done, result is 'Keep' for converge phase 'WaitBeforeAll', trigger Operator-Startup\n\t\tif taskEvent, ok := tsk.GetProp(ConvergeEventProp).(ConvergeEvent); ok {\n\t\t\tparts = append(parts, string(taskEvent))\n\t\t\tif taskEvent != KubeConfigChanged {\n\t\t\t\tparts = append(parts, fmt.Sprintf(\"in phase '%s'\", phase))\n\t\t\t}\n\t\t}\n\n\tcase task.ModuleRun:\n\t\tparts = append(parts, fmt.Sprintf(\"module '%s', phase '%s'\", hm.ModuleName, phase))\n\t\tif hm.DoModuleStartup {\n\t\t\tparts = append(parts, \"with doModuleStartup\")\n\t\t}\n\n\tcase task.ModulePurge, task.ModuleDelete:\n\t\tparts = append(parts, fmt.Sprintf(\"module '%s'\", hm.ModuleName))\n\n\tcase task.GlobalHookEnableKubernetesBindings, task.GlobalHookWaitKubernetesSynchronization, task.GlobalHookEnableScheduleBindings:\n\t\t// Eaxmples:\n\t\t// GlobalHookEnableKubernetesBindings for the hook, trigger Operator-Startup\n\t\t// GlobalHookEnableKubernetesBindings done, result 'Success' for the hook, trigger Operator-Startup\n\t\tparts = append(parts, \"the hook\")\n\n\tcase task.DiscoverHelmReleases:\n\t\t// Examples:\n\t\t// DiscoverHelmReleases task, trigger Operator-Startup\n\t\t// DiscoverHelmReleases task done, result is 'Success', trigger Operator-Startup\n\t\t// Remove \"for\"\n\t\tparts = parts[:len(parts)-1]\n\t}\n\n\ttriggeredBy := hm.EventDescription\n\tif triggeredBy != \"\" {\n\t\ttriggeredBy = \", trigger is \" + triggeredBy\n\t}\n\n\treturn fmt.Sprintf(\"%s%s\", strings.Join(parts, \" \"), triggeredBy)\n}", "func (e *ECS) DescribeTaskDefinition(req *DescribeTaskDefinitionReq) (\n\t*DescribeTaskDefinitionResp, error) {\n\tif req == nil {\n\t\treturn nil, fmt.Errorf(\"The req params cannot be nil\")\n\t}\n\n\tparams := makeParams(\"DescribeTaskDefinition\")\n\tparams[\"taskDefinition\"] = req.TaskDefinition\n\n\tresp := new(DescribeTaskDefinitionResp)\n\tif err := e.query(params, resp); err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func (c *Client) GetTask(ctx context.Context, in *todopb.TaskQuery, opts ...grpc.CallOption) (*todopb.TaskResponse, error) {\n\treturn c.client.GetTask(ctx, in, opts...)\n}", "func (ghidraScriptTask *GhidraScriptTask) Status() *GhidraTaskStatus {\n\treturn &ghidraScriptTask.status\n}", "func (t Task) String() string {\n\treturn fmt.Sprintf(\n\t\t\"<Task name=%s id=%s, status=%s, published_at=%s>\",\n\t\tt.Name, t.ID, t.StatusDisplay(), t.PublishedAt.String(),\n\t)\n}", "func (cl *RedisClient) GetTask() (*RedisTask, error) {\n\tval, err := cl.client.Keys(\"tasks:*\").Result()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar result *RedisTask\n\tvar key string\n\ttxf := func(tx *redis.Tx) error {\n\t\tresult = nil\n\t\tstate, err := tx.HGet(key, \"state\").Result()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif state == \"new\" {\n\t\t\tinputfile, err := cl.client.HGet(key, \"inputfile\").Result()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\toutfile, err := cl.client.HGet(key, \"outfile\").Result()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tresult = &RedisTask{}\n\t\t\tresult.InputFile = inputfile\n\t\t\tresult.OutFile = outfile\n\t\t\tresult.TaskName = key\n\t\t\t_, err = tx.HSet(key, \"state\", \"holded\").Result()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t}\n\n\tfor _, key = range val {\n\t\terr := cl.client.Watch(txf, key)\n\t\tif err == redis.TxFailedErr {\n\t\t\treturn nil, err\n\t\t}\n\t\tif result != nil {\n\t\t\treturn result, nil\n\t\t}\n\t}\n\n\treturn nil, nil\n}", "func TaskURL(swarmingService string, taskID string) string {\n\treturn fmt.Sprintf(\"%stask?id=%s\", swarmingService, taskID)\n}", "func (o *Run) SetTaskID(v string) {\n\to.TaskID = &v\n}", "func (*TaskRunStatus) Descriptor() ([]byte, []int) {\n\treturn file_taskrun_proto_rawDescGZIP(), []int{8}\n}", "func (db *DynamoDB) GetTask(ctx context.Context, req *tes.GetTaskRequest) (*tes.Task, error) {\n\tvar task *tes.Task\n\tvar response *dynamodb.GetItemOutput\n\tvar err error\n\n\tswitch req.View {\n\tcase tes.TaskView_MINIMAL:\n\t\tresponse, err = db.getMinimalView(ctx, req.Id)\n\tcase tes.TaskView_BASIC:\n\t\tresponse, err = db.getBasicView(ctx, req.Id)\n\tcase tes.TaskView_FULL:\n\t\tresponse, err = db.getFullView(ctx, req.Id)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif response.Item == nil {\n\t\treturn nil, tes.ErrNotFound\n\t}\n\n\terr = dynamodbattribute.UnmarshalMap(response.Item, &task)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to DynamoDB unmarshal Task, %v\", err)\n\t}\n\n\treturn task, nil\n}", "func (s *agentRegistry) GetTaskStatistics(ctx context.Context, in *pipeline.GetTaskStatisticsRequest) (*pipeline.GetTaskStatisticsResponse, error) {\n\ts.agentsMutex.Lock()\n\tdefer s.agentsMutex.Unlock()\n\n\tisTaskRequested := func(taskName string) bool {\n\t\tif len(in.GetTaskNames()) == 0 {\n\t\t\treturn true\n\t\t}\n\t\tfor _, x := range in.GetTaskNames() {\n\t\t\tif x == taskName {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t\treturn false\n\t}\n\n\tresult := &pipeline.GetTaskStatisticsResponse{}\n\tfor name, list := range s.taskAgents {\n\t\tif isTaskRequested(name) {\n\t\t\tstat := tracking.TaskStatistics{\n\t\t\t\tName: name,\n\t\t\t}\n\t\t\tfor _, x := range list {\n\t\t\t\tstat.Add(x.Statistics.Data)\n\t\t\t}\n\t\t\tresult.Statistics = append(result.Statistics, &stat)\n\t\t}\n\t}\n\n\treturn result, nil\n}", "func (s *K8sSvc) GetServiceTask(ctx context.Context, cluster string, service string, containerInstanceID string) (serviceTaskID string, err error) {\n\treturn \"\", common.ErrNotSupported\n}", "func (t *Task) Info() string {\n\treturn t.info\n}", "func (o *OpenapiTaskGenerationResult) GetTaskDisplayName() string {\n\tif o == nil || o.TaskDisplayName == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.TaskDisplayName\n}", "func PrintStatus() {\n\threpo := sqlite.NewHeadsRepo()\n\ttrepo := sqlite.NewTaskRepo()\n\n\tvar pr *domain.Project\n\tvar t *domain.Task\n\tvar err error\n\n\tif t, pr, err = hrepo.GetCurrentTask(); err != nil {\n\t\tfmt.Printf(\"Error :%v\\n\", err)\n\t\treturn\n\t}\n\n\ttaskName := \"No definida\"\n\ttid := 0\n\n\tif t != nil {\n\t\ttaskName = t.Name\n\t\ttid = int(t.ID)\n\t}\n\n\tprojectName := \"No definido\"\n\tpid := 0\n\n\tif pr != nil {\n\t\tprojectName = pr.Name\n\t\tpid = int(pr.ID)\n\t}\n\n\ttoday, week, month, total, _ := trepo.GetAggregates(tid)\n\n\tst := status{\n\t\tCurrentProjectName: projectName,\n\t\tCurrentTaskName: taskName,\n\t\tPID: pid,\n\t\tTID: tid,\n\t\tTimeToday: minHours(today),\n\t\tTimeThisWeek: minHours(week),\n\t\tTimeThisMonth: minHours(month),\n\t\tTimeTotal: minHours(total),\n\t}\n\n\treport, err := template.New(\"report\").Parse(statusTemplate)\n\n\tif err != nil {\n\t\tfmt.Printf(\"Error %v\", err)\n\t\treturn\n\t}\n\n\terr = report.Execute(os.Stdout, st)\n\n\tif err != nil {\n\t\tfmt.Printf(\"Error %v\", err)\n\t}\n}", "func (o *OpenapiTaskGenerationResult) SetTaskStatus(v string) {\n\to.TaskStatus = &v\n}", "func (db *DynamoDB) ListTasks(ctx context.Context, req *tes.ListTasksRequest) (*tes.ListTasksResponse, error) {\n\n\tvar tasks []*tes.Task\n\tvar pageSize int64 = 256\n\tvar query *dynamodb.QueryInput\n\n\tif req.PageSize != 0 {\n\t\tpageSize = int64(req.GetPageSize())\n\t\tif pageSize > 2048 {\n\t\t\tpageSize = 2048\n\t\t}\n\t\tif pageSize < 50 {\n\t\t\tpageSize = 50\n\t\t}\n\t}\n\n\tquery = &dynamodb.QueryInput{\n\t\tTableName: aws.String(db.taskTable),\n\t\tLimit: aws.Int64(pageSize),\n\t\tScanIndexForward: aws.Bool(false),\n\t\tConsistentRead: aws.Bool(true),\n\t\tKeyConditionExpression: aws.String(fmt.Sprintf(\"%s = :v1\", db.partitionKey)),\n\t\tExpressionAttributeValues: map[string]*dynamodb.AttributeValue{\n\t\t\t\":v1\": {\n\t\t\t\tS: aws.String(db.partitionValue),\n\t\t\t},\n\t\t},\n\t}\n\n\tif req.View == tes.TaskView_MINIMAL {\n\t\tquery.ExpressionAttributeNames = map[string]*string{\n\t\t\t\"#state\": aws.String(\"state\"),\n\t\t}\n\t\tquery.ProjectionExpression = aws.String(\"id, #state\")\n\t}\n\n\tif req.PageToken != \"\" {\n\t\tquery.ExclusiveStartKey = map[string]*dynamodb.AttributeValue{\n\t\t\tdb.partitionKey: {\n\t\t\t\tS: aws.String(db.partitionValue),\n\t\t\t},\n\t\t\t\"id\": {\n\t\t\t\tS: aws.String(req.PageToken),\n\t\t\t},\n\t\t}\n\t}\n\n\tresponse, err := db.client.QueryWithContext(ctx, query)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif req.View == tes.TaskView_FULL {\n\t\tfor _, item := range response.Items {\n\t\t\t// TODO handle errors\n\t\t\t_ = db.getContent(ctx, item)\n\t\t\t_ = db.getExecutorOutput(ctx, item, \"stdout\", db.stdoutTable)\n\t\t\t_ = db.getExecutorOutput(ctx, item, \"stderr\", db.stderrTable)\n\t\t}\n\t}\n\n\terr = dynamodbattribute.UnmarshalListOfMaps(response.Items, &tasks)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to DynamoDB unmarshal Tasks, %v\", err)\n\t}\n\n\tout := tes.ListTasksResponse{\n\t\tTasks: tasks,\n\t}\n\n\tif response.LastEvaluatedKey != nil {\n\t\tout.NextPageToken = *response.LastEvaluatedKey[\"id\"].S\n\t}\n\n\treturn &out, nil\n}", "func (*FakeReconcilerClient) GetTask(string) (swarm.Task, error) {\n\treturn swarm.Task{}, FakeUnimplemented\n}", "func ReadTask(w http.ResponseWriter, r *http.Request, _ httprouter.Params) {\n\tfmt.Fprint(w, \"ReadTask\\n\")\n}", "func (s *Service) UpdateTaskStatus(c context.Context, date string, typ int, status int) (err error) {\n\t_, err = s.dao.UpdateTaskStatus(c, date, typ, status)\n\treturn\n}" ]
[ "0.6450777", "0.612711", "0.6043263", "0.60372734", "0.5894179", "0.5767152", "0.5736859", "0.57292414", "0.5600359", "0.55213696", "0.5463764", "0.5417522", "0.5382704", "0.535509", "0.5335711", "0.53252643", "0.5318367", "0.52696025", "0.5261585", "0.5251186", "0.5250223", "0.5194984", "0.5189236", "0.51641816", "0.5157285", "0.5136428", "0.5135747", "0.51243675", "0.5113017", "0.5085984", "0.5084515", "0.50842863", "0.5079541", "0.5079287", "0.5054286", "0.50517905", "0.5009252", "0.49862555", "0.49830037", "0.49734968", "0.49708492", "0.49695918", "0.49600774", "0.49599028", "0.4951276", "0.4948772", "0.49334738", "0.4904156", "0.48959318", "0.4885645", "0.4883708", "0.488069", "0.48761278", "0.4876037", "0.4868738", "0.48646045", "0.48591495", "0.48558816", "0.4853292", "0.48490256", "0.4840546", "0.48332703", "0.48285782", "0.48257622", "0.48104978", "0.48095992", "0.4808198", "0.47978428", "0.47973487", "0.4796681", "0.47936824", "0.47876877", "0.47796893", "0.47760823", "0.47683752", "0.47597447", "0.4756437", "0.47558856", "0.475272", "0.47475457", "0.47410473", "0.47384256", "0.4736391", "0.47323415", "0.47322363", "0.47321165", "0.47310668", "0.47201332", "0.47163793", "0.47100893", "0.47066855", "0.47057438", "0.47048804", "0.47007877", "0.4696522", "0.46959075", "0.46911165", "0.46896455", "0.46795404", "0.4672923" ]
0.7474207
0
TaskStats function returns a channel which the driver should send stats to at the given interval. The driver must send stats at the given interval until the given context is canceled or the task terminates.
func (d *Driver) TaskStats(ctx context.Context, taskID string, interval time.Duration) (<-chan *drivers.TaskResourceUsage, error) { d.logger.Debug("TaskStats called", "taskID", taskID) handle, ok := d.tasks.Get(taskID) if !ok { return nil, drivers.ErrTaskNotFound } statsChannel := make(chan *drivers.TaskResourceUsage) go handle.runStatsEmitter(ctx, statsChannel, interval) return statsChannel, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ChannelStats(namespace string) StatsCollector {\n\tlabels := []string{\"topic\", \"channel\", \"paused\"}\n\tnamespace += \"_channel\"\n\n\treturn channelStats{\n\t\t{\n\t\t\tval: func(c *channel) float64 { return float64(len(c.Clients)) },\n\t\t\tvec: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\t\tNamespace: namespace,\n\t\t\t\tName: \"client_count\",\n\t\t\t\tHelp: \"Number of clients\",\n\t\t\t}, labels),\n\t\t},\n\t\t{\n\t\t\tval: func(c *channel) float64 { return float64(c.Depth) },\n\t\t\tvec: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\t\tNamespace: namespace,\n\t\t\t\tName: \"depth\",\n\t\t\t\tHelp: \"Queue depth\",\n\t\t\t}, labels),\n\t\t},\n\t\t{\n\t\t\tval: func(c *channel) float64 { return float64(c.BackendDepth) },\n\t\t\tvec: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\t\tNamespace: namespace,\n\t\t\t\tName: \"backend_depth\",\n\t\t\t\tHelp: \"Queue backend depth\",\n\t\t\t}, labels),\n\t\t},\n\t\t{\n\t\t\tval: func(c *channel) float64 { return float64(c.MessageCount) },\n\t\t\tvec: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\t\tNamespace: namespace,\n\t\t\t\tName: \"message_count\",\n\t\t\t\tHelp: \"Queue message count\",\n\t\t\t}, labels),\n\t\t},\n\t\t{\n\t\t\tval: func(c *channel) float64 { return float64(c.InFlightCount) },\n\t\t\tvec: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\t\tNamespace: namespace,\n\t\t\t\tName: \"in_flight_count\",\n\t\t\t\tHelp: \"In flight count\",\n\t\t\t}, labels),\n\t\t},\n\t\t{\n\t\t\tval: func(c *channel) float64 { return c.E2eLatency.percentileValue(0) },\n\t\t\tvec: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\t\tNamespace: namespace,\n\t\t\t\tName: \"e2e_latency_99p\",\n\t\t\t\tHelp: \"e2e latency 99th percentile\",\n\t\t\t}, labels),\n\t\t},\n\t\t{\n\t\t\tval: func(c *channel) float64 { return c.E2eLatency.percentileValue(1) },\n\t\t\tvec: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\t\tNamespace: namespace,\n\t\t\t\tName: \"e2e_latency_95p\",\n\t\t\t\tHelp: \"e2e latency 95th percentile\",\n\t\t\t}, labels),\n\t\t},\n\t\t{\n\t\t\tval: func(c *channel) float64 { return float64(c.DeferredCount) },\n\t\t\tvec: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\t\tNamespace: namespace,\n\t\t\t\tName: \"deferred_count\",\n\t\t\t\tHelp: \"Deferred count\",\n\t\t\t}, labels),\n\t\t},\n\t\t{\n\t\t\tval: func(c *channel) float64 { return float64(c.RequeueCount) },\n\t\t\tvec: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\t\tNamespace: namespace,\n\t\t\t\tName: \"requeue_count\",\n\t\t\t\tHelp: \"Requeue Count\",\n\t\t\t}, labels),\n\t\t},\n\t\t{\n\t\t\tval: func(c *channel) float64 { return float64(c.TimeoutCount) },\n\t\t\tvec: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\t\tNamespace: namespace,\n\t\t\t\tName: \"timeout_count\",\n\t\t\t\tHelp: \"Timeout count\",\n\t\t\t}, labels),\n\t\t},\n\t}\n}", "func (t *task) Stats(_ context.Context) (*libcontainerdtypes.Stats, error) {\n\thc, err := t.getHCSContainer()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treadAt := time.Now()\n\ts, err := hc.Statistics()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &libcontainerdtypes.Stats{\n\t\tRead: readAt,\n\t\tHCSStats: &s,\n\t}, nil\n}", "func (ts *TaskService) Stats(ctx context.Context, req *taskAPI.StatsRequest) (*taskAPI.StatsResponse, error) {\n\tdefer logPanicAndDie(log.G(ctx))\n\n\tlog.G(ctx).WithField(\"id\", req.ID).Debug(\"stats\")\n\ttask, err := ts.taskManager.Task(req.ID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx = namespaces.WithNamespace(ctx, defaultNamespace)\n\tresp, err := task.Stats(ctx, req)\n\tif err != nil {\n\t\tlog.G(ctx).WithError(err).Error(\"stats failed\")\n\t\treturn nil, err\n\t}\n\n\tlog.G(ctx).Debug(\"stats succeeded\")\n\treturn resp, nil\n}", "func Interval(ctx context.Context, duration time.Duration) (chan struct{}, context.CancelFunc) {\n\tctx, cancel := context.WithCancel(ctx)\n\tch := make(chan struct{}, 1)\n\tgo func() {\n\t\tticker := time.NewTicker(duration)\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\tclose(ch)\n\t\t\t\treturn\n\t\t\tcase <-ticker.C:\n\t\t\t\tch <- struct{}{}\n\t\t\t}\n\t\t}\n\t}()\n\treturn ch, cancel\n}", "func (c *TimeoutChan) Stats() TimeoutChanStats {\n\tc.mu.RLock()\n\tdefer c.mu.RUnlock()\n\treturn TimeoutChanStats{\n\t\tPushed: c.pushed,\n\t\tPopped: c.popped,\n\t\tCleared: c.cleared,\n\t}\n}", "func (r *ClusterPollRequest) task(ctx context.Context) (status int, result interface{}, err error) {\n\tresponse, err := r.request.SendContext(ctx)\n\tif response != nil {\n\t\tstatus = response.Status()\n\t\tresult = response\n\t}\n\treturn\n}", "func GetStats(p *config.ProxyMonitorMetric, cfg *config.CCConfig, timeout time.Duration) *Stats {\n\tbytes := config.Encode(p)\n\tfmt.Println(string(bytes))\n\tvar ch = make(chan struct{})\n\tvar host = p.IP + \":\" + p.AdminPort\n\tfmt.Println(host)\n\tstats := &Stats{}\n\n\tgo func(host string) {\n\t\tdefer close(ch)\n\t\tstats.Host = host\n\t\terr := pingCheck(host, cfg.CCProxyServer.User, cfg.CCProxyServer.Password)\n\t\tif err != nil {\n\t\t\tstats.Error = err.Error()\n\t\t\tstats.Closed = true\n\t\t} else {\n\t\t\tstats.Closed = false\n\t\t}\n\t}(host)\n\n\tselect {\n\tcase <-ch:\n\t\treturn stats\n\tcase <-time.After(timeout):\n\t\treturn &Stats{Host: host, Timeout: true}\n\t}\n}", "func (jbobject *TaskContext) TaskMetrics() *ExecutorTaskMetrics {\n\tjret, err := jbobject.CallMethod(javabind.GetEnv(), \"taskMetrics\", \"org/apache/spark/executor/TaskMetrics\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tretconv := javabind.NewJavaToGoCallable()\n\tdst := &javabind.Callable{}\n\tretconv.Dest(dst)\n\tif err := retconv.Convert(javabind.ObjectRef(jret)); err != nil {\n\t\tpanic(err)\n\t}\n\tretconv.CleanUp()\n\tunique_x := &ExecutorTaskMetrics{}\n\tunique_x.Callable = dst\n\treturn unique_x\n}", "func (c *Client) GetStats() (*Tasks, error) {\n\tqp := map[string]string{\n\t\t\"skip\": \"0\",\n\t\t\"take\": \"0\",\n\t}\n\tresp, err := c.DoGetRequest(\"tasks\", qp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttasks := Tasks{}\n\terr = json.NewDecoder(resp.Body).Decode(&tasks)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &tasks, nil\n}", "func WithContext(ctx context.Context, n uint, interval time.Duration, fn func() error) (err error) {\n\tfor n > 0 {\n\t\tn--\n\t\terr = fn()\n\t\tif err == nil || n <= 0 {\n\t\t\tbreak\n\t\t}\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn ctx.Err()\n\t\tcase <-time.After(interval):\n\t\t}\n\t}\n\treturn\n}", "func Handle(ctx context.Context, stats gw.GatewayStats) error {\n\tsctx := statsContext{\n\t\tctx: ctx,\n\t\tgatewayStats: stats,\n\t}\n\n\tfor _, t := range tasks {\n\t\tif err := t(&sctx); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (i *instanceManager) dispenseTaskEventsCh() (<-chan *drivers.TaskEvent, context.CancelFunc, error) {\n\tdriver, err := i.dispense()\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tctx, cancel := context.WithCancel(i.ctx)\n\teventsCh, err := driver.TaskEvents(ctx)\n\tif err != nil {\n\t\tcancel()\n\t\treturn nil, nil, err\n\t}\n\n\treturn eventsCh, cancel, nil\n}", "func (filter TaskReliabilityFilter) GetTaskStats() (taskStats []taskstats.TaskStats, err error) {\n\tpipeline := filter.taskReliabilityQueryPipeline()\n\terr = db.Aggregate(taskstats.DailyTaskStatsCollection, pipeline, &taskStats)\n\treturn\n}", "func (q *RxQueue) GetStats(reset bool) RunnerStats {\n\tvar b, i int64\n\ts := make(map[string]int64, 8)\n\tif reset {\n\t\ts[\"event\"] = atomic.SwapInt64(&q.eventstats, 0)\n\t} else {\n\t\ts[\"event\"] = atomic.LoadInt64(&q.eventstats)\n\t}\n\tb = atomic.LoadInt64(&q.busycnt)\n\ti = atomic.LoadInt64(&q.idlecnt)\n\ts[\"rxchannelbusy\"] = 0\n\tif b > 0 {\n\t\ts[\"rxchannelbusy\"] = b * 100 / (b + i)\n\t}\n\n\treturn s\n}", "func (l *LogFlow) timerTask(period time.Duration, count int64) error {\n\tl.TargetRate = float64(time.Second) / float64(period)\n\tticker := time.NewTicker(period)\n\terr := l.output.StartOutput(l.msgchan)\n\tif err != nil {\n\t\tl.log.Printf(\"error: %v\", err)\n\t\treturn err\n\t}\n\n\tl.StartTime = time.Now()\n\tl.Elapsed = 0\n\tl.Sent = 0\n\tl.State = Running\n\n\tdefer ticker.Stop()\n\tdefer l.timeTrack(l.StartTime, \"timerTask\")\n\n\tif l.outputFormat == LoggingStandard {\n\t\tl.getMsgPtr = l.loggen.GetStandardMessage\n\t} else {\n\t\tl.getMsgPtr = l.loggen.GetJSONMessage\n\t}\n\n\tfor i := int64(0); i < count; i++ {\n\t\tselect {\n\t\tcase t := <-ticker.C:\n\t\t\tmsg, err := l.getMsgPtr(t.UTC())\n\t\t\tif err != nil {\n\t\t\t\tl.log.Printf(\"error: %v\", err)\n\t\t\t}\n\t\t\tl.msgchan <- msg\n\n\t\t\tl.Sent++\n\t\tcase <-l.quittimer:\n\t\t\tl.State = Stopped\n\t\t\tl.log.Println(\"timerTask stopping\")\n\t\t\treturn nil\n\t\t}\n\t}\n\n\tl.log.Println(\"timerTask stopping based on count\")\n\tl.State = Completed\n\n\treturn nil\n}", "func monitorTask(ctx context.Context, task *model.Task, channel chan model.Event) {\n\t// derive new timeout context\n\tmonitorCtx, cancel := context.WithTimeout(ctx, 10 * time.Second)\n\tdefer cancel()\n\n\tselect {\n\tcase <- monitorCtx.Done():\n\t\t// check status of task\n\t\tstatus := task.GetStatus()\n\n\t\tif status != model.TaskStatusInitial && status != model.TaskStatusExecuting {\n\t\t\treturn\n\t\t}\n\n\t\t// task may still be active\n\t\tswitch monitorCtx.Err().Error() {\n\t\tcase \"context canceled\": // termination of processes\n\t\t\tutil.LogInfo(task.UUID, \"ENG\", \"termination\")\n\t\t\tchannel <- model.NewEvent(task.Domain, task.UUID, model.EventTypeTaskTermination, task.UUID, \"termination\")\n\t\tdefault: // timeout\n\t\t\tutil.LogInfo(task.UUID, \"ENG\", \"timeout\")\n\t\t\tchannel <- model.NewEvent(task.Domain, task.UUID, model.EventTypeTaskTimeout, task.UUID, \"timeout\")\n\t\t}\n\t}\n}", "func (csw *ChannelStatsWatcher) Run(ctx context.Context) {\n\tflushed, unregister := csw.statser.RegisterFlush()\n\tdefer unregister()\n\n\tticker := time.NewTicker(csw.sampleInterval)\n\tdefer ticker.Stop()\n\n\tcsw.sample()\n\n\tfor {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn\n\t\tcase <-flushed:\n\t\t\tcsw.emit()\n\t\t\tcsw.sample() // Ensure there will always be at least one sample\n\t\tcase <-ticker.C:\n\t\t\tcsw.sample()\n\t\t}\n\t}\n}", "func TestRktDriver_Stats(t *testing.T) {\n\tctestutil.RktCompatible(t)\n\tif !testutil.IsCI() {\n\t\tt.Parallel()\n\t}\n\n\trequire := require.New(t)\n\td := NewRktDriver(testlog.HCLogger(t))\n\tharness := dtestutil.NewDriverHarness(t, d)\n\n\ttask := &drivers.TaskConfig{\n\t\tID: uuid.Generate(),\n\t\tAllocID: uuid.Generate(),\n\t\tName: \"etcd\",\n\t\tResources: &drivers.Resources{\n\t\t\tNomadResources: &structs.AllocatedTaskResources{\n\t\t\t\tMemory: structs.AllocatedMemoryResources{\n\t\t\t\t\tMemoryMB: 128,\n\t\t\t\t},\n\t\t\t\tCpu: structs.AllocatedCpuResources{\n\t\t\t\t\tCpuShares: 100,\n\t\t\t\t},\n\t\t\t},\n\t\t\tLinuxResources: &drivers.LinuxResources{\n\t\t\t\tMemoryLimitBytes: 134217728,\n\t\t\t\tCPUShares: 100,\n\t\t\t},\n\t\t},\n\t}\n\n\ttc := &TaskConfig{\n\t\tTrustPrefix: \"coreos.com/etcd\",\n\t\tImageName: \"coreos.com/etcd:v2.0.4\",\n\t\tCommand: \"/etcd\",\n\t\tNet: []string{\"none\"},\n\t}\n\trequire.NoError(task.EncodeConcreteDriverConfig(&tc))\n\ttesttask.SetTaskConfigEnv(task)\n\n\tcleanup := harness.MkAllocDir(task, true)\n\tdefer cleanup()\n\n\thandle, _, err := harness.StartTask(task)\n\trequire.NoError(err)\n\n\t// Wait for task to start\n\t_, err = harness.WaitTask(context.Background(), handle.Config.ID)\n\trequire.NoError(err)\n\n\t// Wait until task started\n\trequire.NoError(harness.WaitUntilStarted(task.ID, 1*time.Second))\n\n\tctx, cancel := context.WithCancel(context.Background())\n\tdefer cancel()\n\tstatsCh, err := d.TaskStats(ctx, task.ID, time.Second*10)\n\trequire.Nil(err)\n\n\tselect {\n\tcase ru := <-statsCh:\n\t\t//TODO(preetha) why are these zero\n\t\tfmt.Printf(\"pid map %v\\n\", ru.Pids)\n\t\tfmt.Printf(\"CPU:%+v Memory:%+v\", ru.ResourceUsage.CpuStats, ru.ResourceUsage.MemoryStats)\n\tcase <-time.After(time.Second):\n\t\trequire.Fail(\"timeout receiving stats from channel\")\n\t}\n\n\trequire.NoError(harness.DestroyTask(task.ID, true))\n\n}", "func MeasureViaPolling(ctx context.Context, fp *os.File) <-chan *Metrics {\n\t// Give a capacity of 1 because we will only ever send one message and the\n\t// buffer allows the component goroutine to exit when done, no matter what the\n\t// client does.\n\tc := make(chan *Metrics, 1)\n\tgo func() {\n\t\tsummary, err := measureUntilContextCancellation(ctx, fp)\n\t\tif err == nil {\n\t\t\tc <- summary\n\t\t}\n\t\tclose(c)\n\t}()\n\treturn c\n}", "func (r *ManagedServicePollRequest) task(ctx context.Context) (status int, result interface{}, err error) {\n\tresponse, err := r.request.SendContext(ctx)\n\tif response != nil {\n\t\tstatus = response.Status()\n\t\tresult = response\n\t}\n\treturn\n}", "func (engine *DockerStatsEngine) GetInstanceMetrics() (*ecstcs.MetricsMetadata, []*ecstcs.TaskMetric, error) {\n\tvar taskMetrics []*ecstcs.TaskMetric\n\tidle := engine.isIdle()\n\tmetricsMetadata := &ecstcs.MetricsMetadata{\n\t\tCluster: aws.String(engine.cluster),\n\t\tContainerInstance: aws.String(engine.containerInstanceArn),\n\t\tIdle: aws.Bool(idle),\n\t\tMessageId: aws.String(uuid.NewRandom().String()),\n\t}\n\n\tif idle {\n\t\tlog.Debug(\"Instance is idle. No task metrics to report\")\n\t\tfin := true\n\t\tmetricsMetadata.Fin = &fin\n\t\treturn metricsMetadata, taskMetrics, nil\n\t}\n\n\tfor taskArn := range engine.tasksToContainers {\n\t\tcontainerMetrics, err := engine.getContainerMetricsForTask(taskArn)\n\t\tif err != nil {\n\t\t\tlog.Debug(\"Error getting container metrics for task\", \"err\", err, \"task\", taskArn)\n\t\t\tcontinue\n\t\t}\n\n\t\tif len(containerMetrics) == 0 {\n\t\t\tlog.Debug(\"Empty containerMetrics for task, ignoring\", \"task\", taskArn)\n\t\t\tcontinue\n\t\t}\n\n\t\ttaskDef, exists := engine.tasksToDefinitions[taskArn]\n\t\tif !exists {\n\t\t\tlog.Debug(\"Could not map task to definition\", \"task\", taskArn)\n\t\t\tcontinue\n\t\t}\n\n\t\tmetricTaskArn := taskArn\n\t\ttaskMetric := &ecstcs.TaskMetric{\n\t\t\tTaskArn: &metricTaskArn,\n\t\t\tTaskDefinitionFamily: &taskDef.family,\n\t\t\tTaskDefinitionVersion: &taskDef.version,\n\t\t\tContainerMetrics: containerMetrics,\n\t\t}\n\t\ttaskMetrics = append(taskMetrics, taskMetric)\n\t}\n\n\tif len(taskMetrics) == 0 {\n\t\t// Not idle. Expect taskMetrics to be there.\n\t\treturn nil, nil, fmt.Errorf(\"No task metrics to report\")\n\t}\n\n\t// Reset current stats. Retaining older stats results in incorrect utilization stats\n\t// until they are removed from the queue.\n\tengine.resetStats()\n\treturn metricsMetadata, taskMetrics, nil\n}", "func (client *Client) GetClusterMetricsWithChan(request *GetClusterMetricsRequest) (<-chan *GetClusterMetricsResponse, <-chan error) {\n\tresponseChan := make(chan *GetClusterMetricsResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetClusterMetrics(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (s *agentRegistry) GetTaskStatistics(ctx context.Context, in *pipeline.GetTaskStatisticsRequest) (*pipeline.GetTaskStatisticsResponse, error) {\n\ts.agentsMutex.Lock()\n\tdefer s.agentsMutex.Unlock()\n\n\tisTaskRequested := func(taskName string) bool {\n\t\tif len(in.GetTaskNames()) == 0 {\n\t\t\treturn true\n\t\t}\n\t\tfor _, x := range in.GetTaskNames() {\n\t\t\tif x == taskName {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t\treturn false\n\t}\n\n\tresult := &pipeline.GetTaskStatisticsResponse{}\n\tfor name, list := range s.taskAgents {\n\t\tif isTaskRequested(name) {\n\t\t\tstat := tracking.TaskStatistics{\n\t\t\t\tName: name,\n\t\t\t}\n\t\t\tfor _, x := range list {\n\t\t\t\tstat.Add(x.Statistics.Data)\n\t\t\t}\n\t\t\tresult.Statistics = append(result.Statistics, &stat)\n\t\t}\n\t}\n\n\treturn result, nil\n}", "func (_e *MockDataCoord_Expecter) GetStatisticsChannel(ctx interface{}) *MockDataCoord_GetStatisticsChannel_Call {\n\treturn &MockDataCoord_GetStatisticsChannel_Call{Call: _e.mock.On(\"GetStatisticsChannel\", ctx)}\n}", "func HeartbeatTask(gateway *Gateway) (task.Func, task.Schedule) {\n\t// Since the database APIs are blocking we need to wrap the core logic\n\t// and run it in a goroutine, so we can abort as soon as the context expires.\n\theartbeatWrapper := func(ctx context.Context) {\n\t\tif gateway.HearbeatCancelFunc() == nil {\n\t\t\tch := make(chan struct{})\n\t\t\tgo func() {\n\t\t\t\tgateway.heartbeat(ctx, hearbeatNormal)\n\t\t\t\tclose(ch)\n\t\t\t}()\n\t\t\tselect {\n\t\t\tcase <-ch:\n\t\t\tcase <-ctx.Done():\n\t\t\t}\n\t\t}\n\t}\n\n\tschedule := func() (time.Duration, error) {\n\t\treturn task.Every(gateway.heartbeatInterval())()\n\t}\n\n\treturn heartbeatWrapper, schedule\n}", "func collectStats(config *Config, stats chan Stats, done chan bool) {\n\tstartTime := time.Now()\n\n\t// TODO: Hoje só temos um cenário. Mas a rotina deve ser revista para trabalhar com mais de um cenário.\n\tnumberOfScenarios := 0\n\tscenariosWithError := 0\n\tnumberOfRequests := 0\n\tvar minTime time.Duration = 1<<63 - 1\n\tvar maxTime time.Duration\n\tvar totalTime time.Duration\n\tvar scenarioName string\n\n\tfor elem := range stats {\n\t\tfmt.Printf(\"elem.EndpointID: [%v]\\n\", elem.EndpointID)\n\t\tif elem.MustStat == false {\n\t\t\tcontinue\n\t\t}\n\t\tif elem.EndpointID != \"\" {\n\t\t\tnumberOfRequests += 1\n\t\t\tcontinue\n\t\t}\n\t\tnumberOfScenarios++\n\t\tscenarioName = elem.ScenarioID\n\t\tlog.Println(elem)\n\t\tif elem.Status == false {\n\t\t\tscenariosWithError++\n\t\t}\n\t\ttotalTime += elem.Duration\n\t\tif elem.Duration > maxTime {\n\t\t\tmaxTime = elem.Duration\n\t\t}\n\t\tif elem.Duration < minTime {\n\t\t\tminTime = elem.Duration\n\t\t}\n\t}\n\n\tduration := time.Since(startTime)\n\n\t// TODO: Não está sendo separado entre cenários com e sem erros\n\tlog.Printf(\"Report - Geral\")\n\tlog.Printf(\"\\tNúmero de requisições: %v\", numberOfRequests)\n\tlog.Printf(\"\\tTempo total de execução do teste: %v\", duration)\n\tlog.Printf(\"\\tRequisições por segundo: %.2f\", float64(numberOfRequests)/float64(duration.Seconds()))\n\tlog.Printf(\"\\tNúmero de IDs únicos: %v\", config.UniqueIds)\n\tlog.Printf(\"\\tCenários executados sem erros: %v (%.2f%%)\", numberOfScenarios-scenariosWithError, float64(numberOfScenarios-scenariosWithError)/float64(numberOfScenarios)*100.0)\n\tlog.Printf(\"\\tCenários executados com erros: %v (%.2f%%)\", scenariosWithError, float64(scenariosWithError)/float64(numberOfScenarios)*100.0)\n\t// log.Printf(\"\\tNúmero de cenários OK: %v\", 1)\n\t// log.Printf(\"\\tNúmero de cenários com falhas: %v\", 1)\n\tlog.Printf(\"\\tNúmero de vezes que ocorreu timeout: %v\", 0)\n\t// log.Printf(\"\\tTempos de resposta: Mínimo, médio, máximo, percentil 95%%: %v\", 1)\n\n\tlog.Printf(\"Report - Por cenário\")\n\n\t// TODO: tem que fazer o report de todos os cenarios... hoje ta assumindo que so tem 1\n\tlog.Printf(\"\\tCenário: %v\", scenarioName)\n\tlog.Printf(\"\\t\\tTempo total de execução do cenário: %v\", totalTime)\n\t// log.Printf(\"\\t\\tCenários executados: %v\", numberOfScenarios) // ou nro de execuções?\n\tlog.Printf(\"\\t\\tCenários executados sem erros: %v (%.2f%%)\", numberOfScenarios-scenariosWithError, float64(numberOfScenarios-scenariosWithError)/float64(numberOfScenarios)*100.0)\n\tlog.Printf(\"\\t\\tCenários executados com erros: %v (%.2f%%)\", scenariosWithError, float64(scenariosWithError)/float64(numberOfScenarios)*100.0)\n\tlog.Printf(\"\\t\\tNúmero de vezes que ocorreu timeout: %v\", 1)\n\t// TODO: faltou percentil 95%\n\tlog.Printf(\"\\t\\tTempo de execução (min/med/max): (%v/%v/%v)\", minTime, totalTime.Nanoseconds()/int64(numberOfScenarios), maxTime)\n\n\tlog.Printf(\"Report - Por endpoint\")\n\tlog.Printf(\"\\tEndpoint: %v\", \"xxx\")\n\tlog.Printf(\"\\t\\tTempo total de execução do endpoint: %v\", 1)\n\tlog.Printf(\"\\t\\tEndpoints executados sem erros: %v (%.2f%%)\", numberOfScenarios-scenariosWithError, float64(numberOfScenarios-scenariosWithError)/float64(numberOfScenarios)*100.0)\n\tlog.Printf(\"\\t\\tEndpoints executados com erros: %v (%.2f%%)\", scenariosWithError, float64(scenariosWithError)/float64(numberOfScenarios)*100.0)\n\tlog.Printf(\"\\t\\tNúmero de vezes que ocorreu timeout: %v\", 1)\n\t// TODO: faltou percentil 95%\n\tlog.Printf(\"\\t\\tTempo de execução (min/med/max): (%v/%v/%v)\", minTime, totalTime.Nanoseconds()/int64(numberOfScenarios), maxTime)\n\t// * Tamanho das requisições/respostas: Mínimo, médio, máximo, percentil 95% (percentil tb???)\n\n\t// TODO: Report por cenario/endpoint?\n\tdone <- true\n}", "func GetTaskStats(filter StatsFilter) ([]TaskStats, error) {\n\terr := filter.ValidateForTasks()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"The provided StatsFilter is invalid\")\n\t}\n\tvar stats []TaskStats\n\tpipeline := filter.TaskStatsQueryPipeline()\n\terr = db.Aggregate(DailyTaskStatsCollection, pipeline, &stats)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"Failed to aggregate task statistics\")\n\t}\n\treturn stats, nil\n}", "func (_e *MockQueryCoord_Expecter) GetStatisticsChannel(ctx interface{}) *MockQueryCoord_GetStatisticsChannel_Call {\n\treturn &MockQueryCoord_GetStatisticsChannel_Call{Call: _e.mock.On(\"GetStatisticsChannel\", ctx)}\n}", "func (ts *TaskService) Stats(requestCtx context.Context, req *taskAPI.StatsRequest) (*taskAPI.StatsResponse, error) {\n\tdefer logPanicAndDie(log.G(requestCtx))\n\tlog.G(requestCtx).WithField(\"id\", req.ID).Debug(\"stats\")\n\n\tresp, err := ts.runcService.Stats(requestCtx, req)\n\tif err != nil {\n\t\tlog.G(requestCtx).WithError(err).Error(\"stats failed\")\n\t\treturn nil, err\n\t}\n\n\tlog.G(requestCtx).Debug(\"stats succeeded\")\n\treturn resp, nil\n}", "func (s *Server) GetStatisticsChannel(ctx context.Context, req *internalpb.GetStatisticsChannelRequest) (*milvuspb.StringResponse, error) {\n\treturn s.querynode.GetStatisticsChannel(ctx)\n}", "func (client *Client) DescribeEvaluateTaskReportWithCallback(request *DescribeEvaluateTaskReportRequest, callback func(response *DescribeEvaluateTaskReportResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeEvaluateTaskReportResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeEvaluateTaskReport(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (operation *Operation) GetTaskConduit() chan Messager {\n\treturn operation.conduit\n}", "func (client *Client) GetClusterMetricsWithCallback(request *GetClusterMetricsRequest, callback func(response *GetClusterMetricsResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetClusterMetricsResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetClusterMetrics(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func dailyTaskStatsPipeline(projectId string, requester string, start time.Time, end time.Time, tasks []string, lastUpdate time.Time) []bson.M {\n\treturn getDailyTaskStatsPipeline(projectId, requester, start, end, tasks, lastUpdate, false)\n}", "func (m *Monitor) Stats(ctx *context.Context) {\n\tctx.JSON(m.Holder.GetStats())\n}", "func (s *Server) GetTimeTickChannel(ctx context.Context, req *internalpb.GetTimeTickChannelRequest) (*milvuspb.StringResponse, error) {\n\treturn s.querynode.GetTimeTickChannel(ctx)\n}", "func (s *agentRegistry) PublishTaskStatistics(ctx context.Context, req *tracking.TaskStatistics) (*empty.Empty, error) {\n\ts.log.Debug().\n\t\tStr(\"task\", req.GetName()).\n\t\tStr(\"uri\", req.GetURI()).\n\t\tMsg(\"Publish task statistics\")\n\n\ts.agentsMutex.Lock()\n\t{\n\t\tcurrent := s.taskAgents[req.GetName()]\n\t\tvar taskAgentRef *taskAgent\n\t\tfor _, x := range current {\n\t\t\tif x.URI == req.GetURI() {\n\t\t\t\ttaskAgentRef = x\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif taskAgentRef == nil {\n\t\t\ttaskAgentRef = &taskAgent{URI: req.GetURI()}\n\t\t\tcurrent = append(current, taskAgentRef)\n\t\t\ts.taskAgents[req.GetName()] = current\n\t\t}\n\t\ttaskAgentRef.Statistics.Data.Reset()\n\t\ttaskAgentRef.Statistics.Data.Add(*req)\n\t\ttaskAgentRef.Statistics.Timestamp = time.Now()\n\t\tsort.Slice(current, func(i, j int) bool { return current[i].Statistics.Timestamp.Before(current[j].Statistics.Timestamp) })\n\t}\n\ts.agentsMutex.Unlock()\n\n\t// Notify frontend clients\n\ts.hub.StatisticsChanged()\n\n\treturn &empty.Empty{}, nil\n}", "func (c *Client) GetTimeTickChannel(ctx context.Context) (*milvuspb.StringResponse, error) {\n\treturn wrapGrpcCall(ctx, c, func(client querypb.QueryNodeClient) (*milvuspb.StringResponse, error) {\n\t\treturn client.GetTimeTickChannel(ctx, &internalpb.GetTimeTickChannelRequest{})\n\t})\n}", "func (d *Dry) Stats(position int) (<-chan *drydocker.Stats, chan<- struct{}, error) {\n\tid, _, err := d.dockerDaemon.ContainerIDAt(position)\n\tif err == nil {\n\t\tif d.dockerDaemon.IsContainerRunning(id) {\n\t\t\tstatsC, dockerDoneChannel := d.dockerDaemon.Stats(id)\n\t\t\treturn statsC, dockerDoneChannel, nil\n\n\t\t}\n\t\td.appmessage(\n\t\t\tfmt.Sprintf(\"<red>Cannot run stats on stopped container. Id: </><white>%s</>\", id))\n\t\terr = errors.New(\"Cannot run stats on stopped container.\")\n\t}\n\n\treturn nil, nil, err\n}", "func (_e *MockDataCoord_Expecter) GetTimeTickChannel(ctx interface{}) *MockDataCoord_GetTimeTickChannel_Call {\n\treturn &MockDataCoord_GetTimeTickChannel_Call{Call: _e.mock.On(\"GetTimeTickChannel\", ctx)}\n}", "func (tm *Manager) Stats() (Stats, Result) {\n\tstats := Stats{}\n\tresult := Result{}\n\n\tif tm.done {\n\t\tresult.Message = shutdownMsg\n\t\tresult.Code = 500\n\t\treturn stats, result\n\t}\n\n\ttm.mutex.Lock()\n\tdefer tm.mutex.Unlock()\n\n\tstats.Total = tm.completedTasks\n\tavg := float64(tm.taskRuntime*time.Nanosecond) / float64(tm.completedTasks)\n\tstats.Average = uint64(avg)\n\tresult.Code = 200\n\n\treturn stats, result\n}", "func (_e *MockQueryCoord_Expecter) GetTimeTickChannel(ctx interface{}) *MockQueryCoord_GetTimeTickChannel_Call {\n\treturn &MockQueryCoord_GetTimeTickChannel_Call{Call: _e.mock.On(\"GetTimeTickChannel\", ctx)}\n}", "func watchTask(newTask chan string) {\n\tlast := \"\" // force the channel to always fire in the first cycle\n\n\t// Really important that we don't use `time.Tick` every 2s as that causes a resource leak of one goroutine every 2s.\n\t// In practice, this starts causing performance degradation (>100% CPU) after several days\n\t// I see ~180% CPU usage for a process that's been running for 713h,\n\t// which goes to show that it's very mild but it definitely adds up.\n\tticker := time.NewTicker(time.Second * 2)\n\tfor {\n\t\t<-ticker.C\n\n\t\tcurrent := getTaskMessage()\n\n\t\tif current != last {\n\t\t\tnewTask <- current\n\t\t\tlast = current\n\t\t\tlog.Printf(\"task %s\", current)\n\t\t}\n\t}\n}", "func RunInterval(ctx context.Context, duration time.Duration, fn func()) context.CancelFunc {\n\tctx, cancel := context.WithCancel(ctx)\n\tgo func() {\n\t\tticker := time.NewTicker(duration)\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn\n\t\t\tcase <-ticker.C:\n\t\t\t\tfn()\n\t\t\t}\n\t\t}\n\t}()\n\treturn cancel\n}", "func (client *Client) DescribeMeterLiveRtcDurationWithChan(request *DescribeMeterLiveRtcDurationRequest) (<-chan *DescribeMeterLiveRtcDurationResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeMeterLiveRtcDurationResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeMeterLiveRtcDuration(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (s *SystemMetrics) GetNetworkStats(c chan *NetworkStats) {\n\to, err := exec.Command(\"netstat\", \"-s\").Output()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tvar netStats NetworkStats\n\tstr := string(o)\n\tlines := strings.Split(str, \"\\n\")\n\tl := len(lines)\n\tcurr := -1\n\tfor {\n\t\tcurr++\n\t\tif curr == l {\n\t\t\tbreak\n\t\t}\n\t\tline := lines[curr]\n\t\tif strings.TrimSpace(line) == \"Tcp:\" {\n\t\t\tfor {\n\t\t\t\tcurr++\n\t\t\t\tif strings.Contains(lines[curr], \"segments received\") {\n\t\t\t\t\tarr := strings.Split(strings.TrimSpace(lines[curr]), \" \")\n\t\t\t\t\tnetStats.PtcpIncoming, err = strconv.Atoi(arr[0])\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tpanic(err)\n\t\t\t\t\t}\n\t\t\t\t\tif buffIncoming == 0 {\n\t\t\t\t\t\tbuffIncoming = netStats.PtcpIncoming\n\t\t\t\t\t} else {\n\t\t\t\t\t\tnetStats.StcpIncoming = ((netStats.PtcpIncoming - buffIncoming) * tcpSegmentSizeInBytes) / 1024\n\t\t\t\t\t\tbuffIncoming = netStats.PtcpIncoming\n\t\t\t\t\t}\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif strings.Contains(lines[curr], \"segments sent out\") {\n\t\t\t\t\tarr := strings.Split(strings.TrimSpace(lines[curr]), \" \")\n\t\t\t\t\tnetStats.PtcpOutgoing, err = strconv.Atoi(arr[0])\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tpanic(err)\n\t\t\t\t\t}\n\t\t\t\t\tif buffOutgoing == 0 {\n\t\t\t\t\t\tbuffOutgoing = netStats.PtcpOutgoing\n\t\t\t\t\t} else {\n\t\t\t\t\t\tnetStats.StcpOutgoing = ((netStats.PtcpOutgoing - buffOutgoing) * tcpSegmentSizeInBytes) / 1024\n\t\t\t\t\t\tbuffOutgoing = netStats.PtcpOutgoing\n\t\t\t\t\t}\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif strings.TrimSpace(line) == \"Udp:\" {\n\t\t\tfor {\n\t\t\t\tcurr++\n\t\t\t\tif strings.Contains(lines[curr], \"packets received\") {\n\t\t\t\t\tarr := strings.Split(strings.TrimSpace(lines[curr]), \" \")\n\t\t\t\t\tnetStats.PudpIncoming, err = strconv.Atoi(arr[0])\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tpanic(err)\n\t\t\t\t\t}\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif strings.Contains(lines[curr], \"packets sent\") {\n\t\t\t\t\tarr := strings.Split(strings.TrimSpace(lines[curr]), \" \")\n\t\t\t\t\tnetStats.PudpOutgoing, err = strconv.Atoi(arr[0])\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tpanic(err)\n\t\t\t\t\t}\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tbreak\n\t\t}\n\t}\n\tc <- &netStats\n}", "func (stats *ConnPoolStats) Export(ch chan<- prometheus.Metric) {\n\tch <- prometheus.MustNewConstMetric(syncClientConnectionsDesc, prometheus.GaugeValue, stats.SyncClientConnections)\n\tch <- prometheus.MustNewConstMetric(numAScopedConnectionsDesc, prometheus.GaugeValue, stats.ASScopedConnections)\n\tch <- prometheus.MustNewConstMetric(totalInUseDesc, prometheus.GaugeValue, stats.TotalInUse)\n\tch <- prometheus.MustNewConstMetric(totalAvailableDesc, prometheus.GaugeValue, stats.TotalAvailable)\n\tch <- prometheus.MustNewConstMetric(totalCreatedDesc, prometheus.CounterValue, stats.TotalCreated)\n}", "func HandleQueryTaskCapacity(w http.ResponseWriter, r *http.Request) {\n\tlog.Root.Info(\"HandleQueryTaskCapacity BEGIN\")\n\n\tif r.Method != http.MethodGet {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tHttpResponseError(w, ErrNotFound)\n\t\treturn\n\t}\n\n\ttaskCapacity, taskLimit, err := node.QueryTaskCapacity()\n\tif err != nil {\n\t\tlog.Root.Error(\"HandleQueryTaskCapacity Query task capacity error\")\n\t\tHttpResponseError(w, ErrServer)\n\t\treturn\n\t}\n\n\tlog.Root.Info(\"HandleQueryTaskCapacity END\")\n\tHttpResponseData(w, H{\n\t\t\"taskCapacity\": taskCapacity,\n\t\t\"taskLimit\": taskLimit,\n\t})\n\treturn\n}", "func (c *Client) GetTimeTickChannel(ctx context.Context) (*milvuspb.StringResponse, error) {\n\treturn wrapGrpcCall(ctx, c, func(client querypb.QueryCoordClient) (*milvuspb.StringResponse, error) {\n\t\treturn client.GetTimeTickChannel(ctx, &internalpb.GetTimeTickChannelRequest{})\n\t})\n}", "func ListenChannel(task ...Task) {\n\t// To trigger channel at first time\n\tSendToChannel(\"\", t4)\n\n\tfor {\n\t\tselect {\n\t\tcase msgChan := <-Ch:\n\t\t\tif len(task) > 0 {\n\t\t\t\tfor _, t := range task {\n\t\t\t\t\tif msg, ok := msgChan[t.Code]; ok {\n\t\t\t\t\t\tt.Job(msg)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tvar msg interface{}\n\t\t\t\tfor _, msgByte := range msgChan {\n\t\t\t\t\terr := json.Unmarshal(msgByte, &msg)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tfmt.Println(err)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func sendMetricToStats(metric Metric) {\n\n\trand.Seed(time.Now().UnixNano())\n\tvar metricValue int\n\tswitch metric.metricType {\n\tcase \"c\":\n\t\tmetricValue = 1\n\tcase \"g\":\n\t\tmetricValue = rand.Intn(100)\n\tcase \"ms\":\n\t\tmetricValue = rand.Intn(1000)\n\tcase \"s\":\n\t\tmetricValue = rand.Intn(100)\n\t}\n\tstringValue := fmt.Sprintf(\"%s:%d|%s\", metric.key, metricValue, metric.metricType)\n\t// Send to the designated connection\n\tswitch metric.connectionType {\n\tcase ConnectionTypeTcp:\n\t\tc, err := net.Dial(\"tcp\", fmt.Sprintf(\"%s:%d\", *statsHost, *statsPortTcp))\n\t\tconnectionCountTcp++\n\t\tif err == nil {\n\t\t\tc.Write([]byte(stringValue + \"\\n\"))\n\t\t\tlogSentMetric(stringValue)\n\t\t\tdefer c.Close()\n\t\t} else {\n\t\t\tconnectionErrorTcp++\n\t\t\treturn\n\t\t}\n\tcase ConnectionTypeTcpPool:\n\t\tc, err := tcpPool.GetConnection(logger)\n\t\tconnectionCountTcpPool++\n\t\tif err == nil {\n\t\t\t_, err := c.Write([]byte(stringValue + \"\\n\"))\n\t\t\tlogSentMetric(stringValue)\n\t\t\tif err != nil {\n\t\t\t\tconnectionErrorTcpPool++\n\t\t\t\tdefer tcpPool.ReleaseConnection(c, true, logger)\n\t\t\t} else {\n\t\t\t\tdefer tcpPool.ReleaseConnection(c, false, logger)\n\t\t\t}\n\t\t} else {\n\t\t\tconnectionErrorTcp++\n\t\t\treturn\n\t\t}\n\tcase ConnectionTypeUdp:\n\t\tc, err := net.Dial(\"udp\", fmt.Sprintf(\"%s:%d\", *statsHost, *statsPortUdp))\n\t\tconnectionCountUdp++\n\t\tif err == nil {\n\t\t\tc.Write([]byte(stringValue))\n\t\t\tlogSentMetric(stringValue)\n\t\t\tdefer c.Close()\n\t\t} else {\n\t\t\tconnectionErrorUdp++\n\t\t\treturn\n\t\t}\n\tcase ConnectionTypeUnix:\n\t\tc, err := net.Dial(\"unix\", *statsSock)\n\t\tconnectionCountUnix++\n\t\tif err == nil {\n\t\t\tc.Write([]byte(stringValue))\n\t\t\tlogSentMetric(stringValue)\n\t\t\tdefer c.Close()\n\t\t} else {\n\t\t\tconnectionErrorUnix++\n\t\t\treturn\n\t\t}\n\tcase ConnectionTypeUnixPool:\n\t\tc, err := unixPool.GetConnection(logger)\n\t\tconnectionCountUnixPool++\n\t\tif err == nil {\n\t\t\t_, err := c.Write([]byte(stringValue + \"\\n\"))\n\t\t\tlogSentMetric(stringValue)\n\t\t\tif err != nil {\n\t\t\t\tconnectionErrorUnixPool++\n\t\t\t\tdefer unixPool.ReleaseConnection(c, true, logger)\n\t\t\t} else {\n\t\t\t\tdefer unixPool.ReleaseConnection(c, false, logger)\n\t\t\t}\n\t\t} else {\n\t\t\tconnectionErrorUnixPool++\n\t\t\treturn\n\t\t}\n\t}\n}", "func NewChannelStatsWatcher(statser Statser, channelName string, tags gostatsd.Tags, capacity int, lenFunc func() int, sampleInterval time.Duration) *ChannelStatsWatcher {\n\treturn &ChannelStatsWatcher{\n\t\tstatser: statser.WithTags(tags.Concat(gostatsd.Tags{\"channel:\" + channelName})),\n\t\tcapacity: capacity,\n\t\tlenFunc: lenFunc,\n\t\tsampleInterval: sampleInterval,\n\t}\n}", "func (s *Module) Monitor(ctx context.Context) <-chan pkg.PoolsStats {\n\tch := make(chan pkg.PoolsStats)\n\tvalues := make(pkg.PoolsStats)\n\tgo func() {\n\t\tdefer close(ch)\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn\n\t\t\tcase <-time.After(5 * time.Second):\n\t\t\t}\n\n\t\t\tfor _, pool := range s.ssds {\n\t\t\t\tif _, err := pool.Mounted(); err != nil {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tdevices, err := s.devices.ByLabel(ctx, pool.Name())\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error().Err(err).Str(\"pool\", pool.Name()).Msg(\"failed to get devices for pool\")\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tvar deviceNames []string\n\t\t\t\tfor _, device := range devices {\n\t\t\t\t\tdeviceNames = append(deviceNames, device.Path)\n\t\t\t\t}\n\n\t\t\t\tusage, err := disk.UsageWithContext(ctx, pool.Path())\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error().Err(err).Str(\"pool\", pool.Name()).Msg(\"failed to get pool usage\")\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tstats, err := disk.IOCountersWithContext(ctx, deviceNames...)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error().Err(err).Str(\"pool\", pool.Name()).Msg(\"failed to get io stats for pool\")\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tpoolStats := pkg.PoolStats{\n\t\t\t\t\tUsageStat: *usage,\n\t\t\t\t\tCounters: stats,\n\t\t\t\t}\n\n\t\t\t\tvalues[pool.Name()] = poolStats\n\t\t\t}\n\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn\n\t\t\tcase ch <- values:\n\t\t\t}\n\t\t}\n\n\t}()\n\n\treturn ch\n}", "func SendMetrics(ws *websocket.Conn, userID string, tasks map[string]demand.Task) error {\n\tvar err error = nil\n\tvar index int = 0\n\n\tmetrics := metrics{\n\t\tTasks: make([]taskMetrics, len(tasks)),\n\t\tCreatedAt: time.Now().Unix(),\n\t}\n\n\tfor name, task := range tasks {\n\t\tmetrics.Tasks[index] = taskMetrics{App: name, RunningCount: task.Running, PendingCount: task.Requested}\n\t\tindex++\n\t}\n\n\tpayload := metricsPayload{\n\t\tUser: userID,\n\t\tMetrics: metrics,\n\t}\n\n\tb, err := json.Marshal(payload)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to encode API json. %v\", err)\n\t}\n\n\t_, err = ws.Write(b)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to send metrics: %v\", err)\n\t}\n\n\treturn err\n}", "func (o BeanstalkScheduledTaskOutput) Frequency() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BeanstalkScheduledTask) *string { return v.Frequency }).(pulumi.StringPtrOutput)\n}", "func (c *Client) GetStatisticsChannel(ctx context.Context) (*milvuspb.StringResponse, error) {\n\treturn wrapGrpcCall(ctx, c, func(client querypb.QueryNodeClient) (*milvuspb.StringResponse, error) {\n\t\treturn client.GetStatisticsChannel(ctx, &internalpb.GetStatisticsChannelRequest{})\n\t})\n}", "func (w *Worker) handleTask() {\n\tvar handleTaskInterval = time.Second\n\tfailpoint.Inject(\"handleTaskInterval\", func(val failpoint.Value) {\n\t\tif milliseconds, ok := val.(int); ok {\n\t\t\thandleTaskInterval = time.Duration(milliseconds) * time.Millisecond\n\t\t\tw.l.Info(\"set handleTaskInterval\", zap.String(\"failpoint\", \"handleTaskInterval\"), zap.Int(\"value\", milliseconds))\n\t\t}\n\t})\n\tticker := time.NewTicker(handleTaskInterval)\n\tdefer ticker.Stop()\n\n\tretryCnt := 0\n\nLoop:\n\tfor {\n\t\tselect {\n\t\tcase <-w.ctx.Done():\n\t\t\tw.l.Info(\"handle task process exits!\")\n\t\t\treturn\n\t\tcase <-ticker.C:\n\t\t\tif w.closed.Get() == closedTrue {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\topLog := w.meta.PeekLog()\n\t\t\tif opLog == nil {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tw.l.Info(\"start to execute operation\", zap.Reflect(\"oplog\", opLog))\n\n\t\t\tst := w.subTaskHolder.findSubTask(opLog.Task.Name)\n\t\t\tvar err error\n\t\t\tswitch opLog.Task.Op {\n\t\t\tcase pb.TaskOp_Start:\n\t\t\t\tif st != nil {\n\t\t\t\t\terr = terror.ErrWorkerSubTaskExists.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tif w.relayPurger.Purging() {\n\t\t\t\t\tif retryCnt < maxRetryCount {\n\t\t\t\t\t\tretryCnt++\n\t\t\t\t\t\tw.l.Warn(\"relay log purger is purging, cannot start subtask, would try again later\", zap.String(\"task\", opLog.Task.Name))\n\t\t\t\t\t\tcontinue Loop\n\t\t\t\t\t}\n\n\t\t\t\t\tretryCnt = 0\n\t\t\t\t\terr = terror.ErrWorkerRelayIsPurging.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tretryCnt = 0\n\t\t\t\ttaskCfg := new(config.SubTaskConfig)\n\t\t\t\tif err1 := taskCfg.Decode(string(opLog.Task.Task)); err1 != nil {\n\t\t\t\t\terr = terror.Annotate(err1, \"decode subtask config error in handleTask\")\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tvar cfgDecrypted *config.SubTaskConfig\n\t\t\t\tcfgDecrypted, err = taskCfg.DecryptPassword()\n\t\t\t\tif err != nil {\n\t\t\t\t\terr = terror.WithClass(err, terror.ClassDMWorker)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tw.l.Info(\"started sub task\", zap.Stringer(\"config\", cfgDecrypted))\n\t\t\t\tst = NewSubTask(cfgDecrypted)\n\t\t\t\tw.subTaskHolder.recordSubTask(st)\n\t\t\t\tst.Run()\n\n\t\t\tcase pb.TaskOp_Update:\n\t\t\t\tif st == nil {\n\t\t\t\t\terr = terror.ErrWorkerSubTaskNotFound.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\ttaskCfg := new(config.SubTaskConfig)\n\t\t\t\tif err1 := taskCfg.Decode(string(opLog.Task.Task)); err1 != nil {\n\t\t\t\t\terr = terror.Annotate(err1, \"decode subtask config error in handleTask\")\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tw.l.Info(\"updated sub task\", zap.String(\"task\", opLog.Task.Name), zap.Stringer(\"new config\", taskCfg))\n\t\t\t\terr = st.Update(taskCfg)\n\t\t\tcase pb.TaskOp_Stop:\n\t\t\t\tif st == nil {\n\t\t\t\t\terr = terror.ErrWorkerSubTaskNotFound.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tw.l.Info(\"stop sub task\", zap.String(\"task\", opLog.Task.Name))\n\t\t\t\tst.Close()\n\t\t\t\tw.subTaskHolder.removeSubTask(opLog.Task.Name)\n\t\t\tcase pb.TaskOp_Pause:\n\t\t\t\tif st == nil {\n\t\t\t\t\terr = terror.ErrWorkerSubTaskNotFound.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tw.l.Info(\"pause sub task\", zap.String(\"task\", opLog.Task.Name))\n\t\t\t\terr = st.Pause()\n\t\t\tcase pb.TaskOp_Resume:\n\t\t\t\tif st == nil {\n\t\t\t\t\terr = terror.ErrWorkerSubTaskNotFound.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tw.l.Info(\"resume sub task\", zap.String(\"task\", opLog.Task.Name))\n\t\t\t\terr = st.Resume()\n\t\t\tcase pb.TaskOp_AutoResume:\n\t\t\t\tif st == nil {\n\t\t\t\t\terr = terror.ErrWorkerSubTaskNotFound.Generate(opLog.Task.Name)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tw.l.Info(\"auto_resume sub task\", zap.String(\"task\", opLog.Task.Name))\n\t\t\t\terr = st.Resume()\n\t\t\t}\n\n\t\t\tw.l.Info(\"end to execute operation\", zap.Int64(\"oplog ID\", opLog.Id), log.ShortError(err))\n\n\t\t\tif err != nil {\n\t\t\t\topLog.Message = err.Error()\n\t\t\t} else {\n\t\t\t\topLog.Task.Stage = st.Stage()\n\t\t\t\topLog.Success = true\n\t\t\t}\n\n\t\t\t// fill current task config\n\t\t\tif len(opLog.Task.Task) == 0 {\n\t\t\t\ttm := w.meta.GetTask(opLog.Task.Name)\n\t\t\t\tif tm == nil {\n\t\t\t\t\tw.l.Warn(\"task meta not found\", zap.String(\"task\", opLog.Task.Name))\n\t\t\t\t} else {\n\t\t\t\t\topLog.Task.Task = append([]byte{}, tm.Task...)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\terr = w.meta.MarkOperation(opLog)\n\t\t\tif err != nil {\n\t\t\t\tw.l.Error(\"fail to mark subtask operation\", zap.Reflect(\"oplog\", opLog))\n\t\t\t}\n\t\t}\n\t}\n}", "func (this *SessionController) executeTask(taskConfiguration TaskConfiguration) TaskData {\n\n\ttaskConfiguration.EventStartTask()\n\n\t//create new task data structure where we hold op data and configuration\n\tvar taskData \t\t\t= \tTaskData{\n\t\tId\t\t\t\t\t: this.Configuration.TaskCounter,\n\t\tDataCursor\t\t\t: taskConfiguration.Min,\n\t\tStatus\t\t\t\t: \"working\",\n\t\tTaskConfiguration\t: taskConfiguration,\n\t\tCreationTime\t\t: time.Now()}\n\n\t//before count jobscount I have to ask mysql for max and min values\n\tif taskConfiguration.Max==0 && taskConfiguration.MaxQuery!=\"\"{\n\n\t}\n\n\tif taskConfiguration.Min==taskConfiguration.Max && taskConfiguration.MinQuery!=\"\"{\n\n\t}\n\n\tvar JobsCount\tuint64\n\tif taskConfiguration.Step<=0 || taskConfiguration.Max<=taskConfiguration.Min || taskConfiguration.Concurrency<1{\n\t\tJobsCount\t\t\t\t\t\t=1\n\t\ttaskConfiguration.Step\t\t\t=0\n\t\ttaskConfiguration.Max\t\t\t=0\n\t\ttaskConfiguration.Min\t\t\t=0\n\t\ttaskConfiguration.Concurrency\t=1\n\t}else {\n\t\tJobsCount\t= ((taskConfiguration.Max - taskConfiguration.Min)/taskConfiguration.Step)\n\t}\n\t//when job finished then posts its status here\n\tvar jobDataChannel\t\t=\tmake(chan *JobData, taskConfiguration.Concurrency)\n\t//\n\tdefer func(){\n\t\ttaskConfiguration.RowsAffected(taskData.RowsAffected)\n\t\tif taskData.Failed {\n\t\t\t//record that fact\n\t\t\ttaskConfiguration.EventFailTask(taskData.LastDroppedJob.LastErrorMsg)\n\t\t}else{\n\t\t\ttaskConfiguration.EventSuccessTask()\n\t\t}\n\t\t//last serialisation\n\t\tSerialiseStruct(this.Configuration)\n\t\t//close task<-job comm channel\n\t\tclose(jobDataChannel)\n\t\t//\n\t\tthis.Debugf(taskConfiguration.Debug, FormatStruct, taskConfiguration)\n\t}()\n\n\tthis.Debugf(taskConfiguration.Debug,FormatStruct, taskConfiguration)\n\n\tfor {\n\t\tvar jobData *JobData\n\t\tselect {\n\t\tcase jobData = <-jobDataChannel:\n\t\t\t{\n\t\t\t\t//check status, if it error then scheduled it again\n\t\t\t\tif jobData.Error {\n\t\t\t\t\tthis.Debugf(taskConfiguration.Debug,FormatStruct,jobData)\n\t\t\t\t\t//increment errors counter\n\t\t\t\t\ttaskData.Errors++\n\t\t\t\t\t//record last dropped\n\t\t\t\t\ttaskData.LastDroppedJob=*jobData\n\t\t\t\t\tif jobData.Attempts == taskConfiguration.MaxAttempts {\n\t\t\t\t\t\t//record how many has been dropped- it will be only one\n\t\t\t\t\t\ttaskData.MaxAttemptJobsDropped++\n\t\t\t\t\t\t//\n\t\t\t\t\t\ttaskData.LastDroppedJob.LastErrorMsg=\"max attempts reached, reason:\"+jobData.LastErrorMsg\n\t\t\t\t\t\t//\n\t\t\t\t\t\ttaskData.Failed=true\n\t\t\t\t\t\t//we quit on first dropped job\n\t\t\t\t\t\treturn taskData\n\t\t\t\t\t}\n\t\t\t\t\t//reset flag and try again\n\t\t\t\t\tjobData.Error = false\n\n\t\t\t\t} else {\n\t\t\t\t\ttaskData.Success++\n\t\t\t\t\t//record the fact that that maany rows were affected by this job\n\t\t\t\t\t//configuration.RowsAffected(jobData.RowsAffected)\n\t\t\t\t\ttaskData.RowsAffected=taskData.RowsAffected+jobData.RowsAffected\n\t\t\t\t}\n\n\t\t\t\t//got report back, decrease the length\n\t\t\t\ttaskData.QueueLength--\n\t\t\t}\n\t\t//\n\t\tcase <-time.After(time.Duration(taskData.Timeout) * time.Millisecond):\n\t\t//cancellation\n\t\tcase <-GDone:\n\t\t\t//return current state of task\n\t\t\treturn taskData\n\t\t}\n\n\t\t//if queue length reached its limit\n\t\t// or exhausted stream\n\t\t// or finished\n\t\t// then we cannot schedule more jobs\n\t\tif (taskData.QueueLength < taskConfiguration.Concurrency\t&&\t((taskData.Success+taskData.MaxAttemptJobsDropped+taskData.QueueLength) < JobsCount)){\n\n\t\t\t//if current job is empty then we create new job\n\t\t\tif jobData == nil {\n\t\t\t\tvar Query string\t=\ttaskConfiguration.Exec\n\t\t\t\t//if partitioning enabled then format SQL string to provide Min and Max\n\t\t\t\tif JobsCount>1{\n\t\t\t\t\tQuery=fmt.Sprintf(taskConfiguration.Exec, taskData.DataCursor, taskData.DataCursor + taskData.TaskConfiguration.Step)\n\t\t\t\t}\n\t\t\t\t//try to resolve $LastEtl to date time\n\t\t\t\tQuery=strings.Replace(Query,LastEtlVariableName,LastEtl.Format(SessionFileFormat),-1)\n\t\t\t\t//try to resolve $ToEtl do date time\n\t\t\t\tQuery=strings.Replace(Query,EtlToVariableName,EtlTo.Format(SessionFileFormat),-1)\n\t\t\t\t//store only data that is requred and specifc for job\n\t\t\t\tjobData = &JobData{Id: taskData.JobId, Query: Query}\n\t\t\t\t//move cursor to the next step\n\t\t\t\ttaskData.DataCursor += taskConfiguration.Step\n\t\t\t\t//\n\t\t\t\ttaskData.JobId++\n\t\t\t}\n\n\t\t\tjobContext:=\t&JobExecutionContext{\n\t\t\t\tjobData:jobData,\n\t\t\t\tjobDataChannel:jobDataChannel}\n\t\t\tthis.Debugf(taskConfiguration.Debug,FormatStruct,jobContext)\n\t\t\t//schedule the job\n\t\t\t//todo: switch case here: Exec, Query, QueryOne\n\t\t\tswitch strings.ToUpper(taskConfiguration.ExecType){\n\t\t\t\tcase Exec:\tgo this.Exec(jobContext)\n\t\t\t\tcase Query: go this.QueryRow(jobContext)\n\t\t\t\tcase QueryOne:\n\t\t\t}\n\n\t\t\t//increase queue length\n\t\t\ttaskData.QueueLength++\n\t\t\t//reset timeout\n\t\t\ttaskData.Timeout = 0\n\t\t} else {\n\t\t\t//if timeout limit is not reached then increase timeout\n\t\t\tif time.Duration(taskData.Timeout) <= time.Second {\n\t\t\t\t//increase timeout when queue if full or there is no more job to schedule\n\t\t\t\ttaskData.Timeout *= 10\n\t\t\t}\n\t\t}\n\t\t//have we finished yet?\n\t\tif (taskData.Success+taskData.MaxAttemptJobsDropped+taskData.QueueLength) >= JobsCount {\n\t\t\ttaskData.Status = \"finishing\"\n\t\t\tif taskData.QueueLength == 0 {\n\t\t\t\ttaskData.Status = \"finished\"\n\t\t\t\treturn taskData\n\t\t\t}\n\t\t}\n\t\t//if last serialisation happend more than...\n\t\tif time.Now().Sub(taskData.Serialised) > (1 * time.Second) {\n\t\t\tSerialiseStruct(this.Configuration)\n\t\t\ttaskData.Serialised = time.Now()\n\t\t}\n\t}\n}", "func (w *TaskWatcher) Watch(ctx context.Context) <-chan *TaskEvent {\n\tc := make(chan *TaskEvent, w.cfg.ChannelSize)\n\tgo w.watch(ctx, c)\n\treturn c\n}", "func (r *Repository) Stats(context.Context) (*adagio.Stats, error) {\n\tr.mu.Lock()\n\tdefer r.mu.Unlock()\n\n\tnodeCounts := &adagio.Stats_NodeCounts{}\n\n\tfor _, runState := range r.runs {\n\t\tfor _, node := range runState.lookup {\n\t\t\tswitch node.Status {\n\t\t\tcase adagio.Node_WAITING:\n\t\t\t\tnodeCounts.WaitingCount++\n\t\t\tcase adagio.Node_READY:\n\t\t\t\tnodeCounts.ReadyCount++\n\t\t\tcase adagio.Node_RUNNING:\n\t\t\t\tnodeCounts.RunningCount++\n\t\t\tcase adagio.Node_COMPLETED:\n\t\t\t\tnodeCounts.CompletedCount++\n\t\t\t}\n\t\t}\n\t}\n\n\treturn &adagio.Stats{\n\t\tRunCount: int64(len(r.runs)),\n\t\tNodeCounts: nodeCounts,\n\t}, nil\n}", "func (s *SleepService) Stats(ctx context.Context, child *goparent.Child) (*goparent.SleepSummary, error) {\r\n\tvar sleeps []goparent.Sleep\r\n\tend := time.Now()\r\n\tstart := end.AddDate(0, 0, -1)\r\n\r\n\tfamilyKey := datastore.NewKey(ctx, FamilyKind, child.FamilyID, 0, nil)\r\n\tchildKey := datastore.NewKey(ctx, ChildKind, child.ID, 0, familyKey)\r\n\r\n\tq := datastore.NewQuery(SleepKind).Filter(\"ChildID = \", childKey).Filter(\"TimeStamp >= \", start).Order(\"-TimeStamp\")\r\n\titx := q.Run(ctx)\r\n\tfor {\r\n\t\tvar sleep goparent.Sleep\r\n\t\t_, err := itx.Next(&sleep)\r\n\t\tif err == datastore.Done {\r\n\t\t\tbreak\r\n\t\t}\r\n\t\tif err != nil {\r\n\t\t\treturn nil, err\r\n\t\t}\r\n\t\tsleeps = append(sleeps, sleep)\r\n\t}\r\n\r\n\tsummary := &goparent.SleepSummary{\r\n\t\tData: sleeps,\r\n\t\tTotal: 0,\r\n\t\tMean: 0.0,\r\n\t\tRange: 0,\r\n\t}\r\n\r\n\tfor _, x := range sleeps {\r\n\t\t//if the sleep end is before the start then the sleep period hasn't stopped yet. don't count\r\n\t\tif x.End.After(x.Start) {\r\n\t\t\tsummary.Total += int64(x.End.Sub(x.Start).Seconds())\r\n\t\t\tsummary.Range++\r\n\t\t}\r\n\t}\r\n\t// summary.Mean = float64(summary.Total / int64(summary.Range))\r\n\r\n\treturn summary, nil\r\n}", "func (client *Client) DescribeMeterLiveRtcDurationWithCallback(request *DescribeMeterLiveRtcDurationRequest, callback func(response *DescribeMeterLiveRtcDurationResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeMeterLiveRtcDurationResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeMeterLiveRtcDuration(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (d *Driver) TaskEvents(ctx context.Context) (<-chan *drivers.TaskEvent, error) {\n\treturn d.eventer.TaskEvents(ctx)\n}", "func (client *Client) DescribeEvaluateTaskReportWithChan(request *DescribeEvaluateTaskReportRequest) (<-chan *DescribeEvaluateTaskReportResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeEvaluateTaskReportResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeEvaluateTaskReport(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (o ElastigroupScheduledTaskOutput) Frequency() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ElastigroupScheduledTask) *string { return v.Frequency }).(pulumi.StringPtrOutput)\n}", "func (a *PeriodicalAnnouncer) Stats() Stats {\n\tvar stats Stats\n\treq := statsRequest{Response: make(chan Stats, 1)}\n\tselect {\n\tcase a.statsCommandC <- req:\n\tcase <-a.closeC:\n\t}\n\tselect {\n\tcase stats = <-req.Response:\n\tcase <-a.closeC:\n\t}\n\treturn stats\n}", "func (c Context) Report() (*taskqueue.Task, error) {\n\tif c.reported {\n\t\t// Only attempt to report once per context. If this happens, it's\n\t\t// a programming bug.\n\t\treturn nil, fmt.Errorf(\"context for %q already reported\", c.R.URL)\n\t}\n\tc.reported = true\n\n\tif len(c.counters) == 0 {\n\t\treturn nil, nil\n\t}\n\n\tr := c.getReport()\n\tb, err := json.Marshal(r)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to marshal monitoring data: %v\", err)\n\t}\n\n\treturn &taskqueue.Task{\n\t\tPath: \"/mon\",\n\t\tPayload: b,\n\t\tMethod: \"POST\",\n\t}, nil\n}", "func (client *Client) DescribeLiveDomainLimitWithChan(request *DescribeLiveDomainLimitRequest) (<-chan *DescribeLiveDomainLimitResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeLiveDomainLimitResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeLiveDomainLimit(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func ChannelWorker(id int, channel chan domain.ChannelEl, wg *sync.WaitGroup) {\n\tdefer wg.Done()\n\n\tstart := utils.GetCurrentTime()\n\ttime.Sleep(time.Second)\n\tend := utils.GetCurrentTime()\n\tchannel <- domain.ChannelEl{ID: id, Time: end - start}\n}", "func (cs *clientServer) metricsToPublishMetricRequests() ([]*ecstcs.PublishMetricsRequest, error) {\n\tmetadata, taskMetrics, err := cs.statsEngine.GetInstanceMetrics()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar requests []*ecstcs.PublishMetricsRequest\n\tif *metadata.Idle {\n\t\tmetadata.Fin = aws.Bool(true)\n\t\t// Idle instance, we have only one request to send to backend.\n\t\trequests = append(requests, ecstcs.NewPublishMetricsRequest(metadata, taskMetrics))\n\t\treturn requests, nil\n\t}\n\tvar messageTaskMetrics []*ecstcs.TaskMetric\n\tnumTasks := len(taskMetrics)\n\n\tfor i, taskMetric := range taskMetrics {\n\t\tmessageTaskMetrics = append(messageTaskMetrics, taskMetric)\n\t\tvar requestMetadata *ecstcs.MetricsMetadata\n\t\tif (i + 1) == numTasks {\n\t\t\t// If this is the last task to send, set fin to true\n\t\t\trequestMetadata = copyMetricsMetadata(metadata, true)\n\t\t} else {\n\t\t\trequestMetadata = copyMetricsMetadata(metadata, false)\n\t\t}\n\t\tif (i+1)%tasksInMetricMessage == 0 {\n\t\t\t// Construct payload with tasksInMetricMessage number of task metrics and send to backend.\n\t\t\trequests = append(requests, ecstcs.NewPublishMetricsRequest(requestMetadata, copyTaskMetrics(messageTaskMetrics)))\n\t\t\tmessageTaskMetrics = messageTaskMetrics[:0]\n\t\t}\n\t}\n\n\tif len(messageTaskMetrics) > 0 {\n\t\t// Create the new metadata object and set fin to true as this is the last message in the payload.\n\t\trequestMetadata := copyMetricsMetadata(metadata, true)\n\t\t// Create a request with remaining task metrics.\n\t\trequests = append(requests, ecstcs.NewPublishMetricsRequest(requestMetadata, messageTaskMetrics))\n\t}\n\treturn requests, nil\n}", "func WithContext(ctx context.Context) client.Client {\n\tctxStats, ok := ctx.Value(statsKey).(client.Client)\n\tif !ok {\n\t\tlog.Error(\"Could not retrieve stats client from the context\")\n\n\t\tctxStats, _ := stats.NewClient(\"noop://\")\n\t\treturn ctxStats\n\t}\n\treturn ctxStats\n}", "func longTask(channel chan int) {\n\tfmt.Println(\"Starting long task\")\n\tdelay := rand.Intn(5) // delay is integer\n\ttime.Sleep(time.Duration(delay) * time.Second)\n\tfmt.Println(\"Long task finished\")\n\tchannel <- delay\n}", "func sendTasks(ctx context.Context, taskCh chan syncer.Task, addTasks []syncer.Task, updateTasks []syncer.Task, deleteTasks []syncer.Task) {\n\tdefer close(taskCh)\n\n\tfor i, t := range addTasks {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\ttelemetryClient.Gauge(\"vsync.destination.paths.skipped\", float64(len(addTasks)-i), \"operation:add\")\n\t\t\tlog.Info().Str(\"trigger\", \"context done\").Int(\"left\", len(addTasks)-i).Msg(\"add tasks skipped\")\n\t\t\treturn\n\t\tdefault:\n\t\t\ttaskCh <- t\n\t\t}\n\t}\n\n\tfor i, t := range updateTasks {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\ttelemetryClient.Gauge(\"vsync.destination.paths.skipped\", float64(len(updateTasks)-i), \"operation:update\")\n\t\t\tlog.Info().Str(\"trigger\", \"context done\").Int(\"left\", len(updateTasks)-i).Msg(\"update tasks skipped\")\n\t\t\treturn\n\t\tdefault:\n\t\t\ttaskCh <- t\n\t\t}\n\t}\n\n\tfor i, t := range deleteTasks {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\ttelemetryClient.Gauge(\"vsync.destination.paths.skipped\", float64(len(deleteTasks)-i), \"operation:delete\")\n\t\t\tlog.Info().Str(\"trigger\", \"context done\").Int(\"left\", len(deleteTasks)-i).Msg(\"delete tasks skipped\")\n\t\t\treturn\n\t\tdefault:\n\t\t\ttaskCh <- t\n\t\t}\n\t}\n}", "func GetCPUTemp(ch chan<- SourceReturn, conf *Conf) {\n\tc := conf.CPU\n\t// Check for warnOnly override\n\tif c.WarnOnly == nil {\n\t\tc.WarnOnly = &conf.WarnOnly\n\t}\n\tsr := NewSourceReturn(conf.debug)\n\tdefer func() {\n\t\tch <- sr.Return(&c.ConfBase)\n\t}()\n\tvar tempMap map[string]int\n\tvar isZen bool\n\tvar err error\n\tif c.Exec {\n\t\ttempMap, isZen, err = cpuTempSensors()\n\t} else {\n\t\ttempMap, isZen, err = cpuTempGopsutil()\n\t}\n\tif err != nil {\n\t\tlog.Warnf(\"[cpu] temperature read error: %v\", err)\n\t}\n\tif len(tempMap) == 0 {\n\t\terr = &ModuleNotAvailable{\"cpu\", err}\n\t\tsr.Header = fmt.Sprintf(\"%s: %s\\n\", utils.Wrap(\"CPU temp\", c.padL, c.padR), utils.Warn(\"unavailable\"))\n\t} else {\n\t\tsr.Header, sr.Content, sr.Error = formatCPUTemps(tempMap, isZen, &c)\n\t}\n\treturn\n}", "func (client *Client) GetBaseStrategyPeriodWithChan(request *GetBaseStrategyPeriodRequest) (<-chan *GetBaseStrategyPeriodResponse, <-chan error) {\n\tresponseChan := make(chan *GetBaseStrategyPeriodResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetBaseStrategyPeriod(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func writeHealth(ctx context.Context, t *text.Text, delay time.Duration, connectionSignal chan string) {\n\treconnect := false\n\thealth := gjson.Get(getFromRPC(\"health\"), \"result\")\n\tt.Reset()\n\tif health.Exists() {\n\t\tt.Write(\"🟢 good\")\n\t} else {\n\t\tt.Write(\"🔴 no connection\")\n\t}\n\n\tticker := time.NewTicker(delay)\n\tdefer ticker.Stop()\n\n\tfor {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\thealth := gjson.Get(getFromRPC(\"health\"), \"result\")\n\t\t\tif health.Exists() {\n\t\t\t\tt.Write(\"🟢 good\")\n\t\t\t\tif reconnect == true {\n\t\t\t\t\tconnectionSignal <- \"reconnect\"\n\t\t\t\t\tconnectionSignal <- \"reconnect\"\n\t\t\t\t\tconnectionSignal <- \"reconnect\"\n\t\t\t\t\treconnect = false\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tt.Write(\"🔴 no connection\")\n\t\t\t\tif reconnect == false {\n\t\t\t\t\tconnectionSignal <- \"no_connection\"\n\t\t\t\t\tconnectionSignal <- \"no_connection\"\n\t\t\t\t\tconnectionSignal <- \"no_connection\"\n\t\t\t\t\treconnect = true\n\t\t\t\t}\n\t\t\t}\n\t\tcase <-ctx.Done():\n\t\t\treturn\n\t\t}\n\t}\n}", "func (c *SwitchTicker) Channel() <-chan time.Time {\n\tfailCount := atomic.LoadInt64(&c.failCount)\n\tif failCount > c.threshold {\n\t\treturn c.fastTicker.C\n\t}\n\treturn c.slowTicker.C\n}", "func (c *Client) GetStatisticsChannel(ctx context.Context) (*milvuspb.StringResponse, error) {\n\treturn wrapGrpcCall(ctx, c, func(client querypb.QueryCoordClient) (*milvuspb.StringResponse, error) {\n\t\treturn client.GetStatisticsChannel(ctx, &internalpb.GetStatisticsChannelRequest{})\n\t})\n}", "func (s *Service) Stats(ctx context.Context, in *pb.Request) (*pb.Stats, error) {\n\ts.Entry.Debugf(\"Received new stats request: %v\", in)\n\n\ts.Entry.Debugf(\"Contex is: %q\", ctx)\n\n\tstats, err := s.Slack.GetStatsPb()\n\n\tif err != nil {\n\t\ts.Entry.Errorf(\"Failed to retreive stats from slack: %s\", err)\n\t}\n\n\treturn stats, err\n}", "func (w *Worker) Work(task interface{}, wg *sync.WaitGroup) {\n\tdefer wg.Done()\n\n\tw.Context.fetch(task.(string))\n\n\tw.TimeProcess <- time.Now().Unix()\n\tw.FreeWorkers <- w\n}", "func (v *atomicReference) AwaitCtx(ctx context.Context, cond RefCondition, interval ...time.Duration) interface{} {\n\tcheckInterval := optional(DefaultReferenceCheckInterval, interval...)\n\tvar sleepTicker *time.Ticker\n\tfor {\n\t\treferent := v.Get()\n\t\tif cond(referent) {\n\t\t\treturn referent\n\t\t}\n\n\t\tif sleepTicker == nil {\n\t\t\tsleepTicker = time.NewTicker(checkInterval)\n\t\t\tdefer sleepTicker.Stop()\n\t\t}\n\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn referent\n\t\tcase <-v.notify:\n\t\t\tNop()\n\t\tcase <-sleepTicker.C:\n\t\t\tNop()\n\t\t}\n\t}\n}", "func (tsr *taskStatusRunner) Run(ctx context.Context, taskQueue chan Task,\n\teventChannel chan event.Event, options Options) error {\n\tstatusCtx, cancelFunc := context.WithCancel(context.Background())\n\tstatusChannel := tsr.statusPoller.Poll(statusCtx, tsr.identifiers, polling.Options{\n\t\tPollInterval: options.PollInterval,\n\t\tUseCache: options.UseCache,\n\t})\n\n\to := baseOptions{\n\t\temitStatusEvents: options.EmitStatusEvents,\n\t}\n\terr := tsr.baseRunner.run(ctx, taskQueue, statusChannel, eventChannel, o)\n\t// cancel the statusPoller by cancelling the context.\n\tcancelFunc()\n\t// drain the statusChannel to make sure the lack of a consumer\n\t// doesn't block the shutdown of the statusPoller.\n\tfor range statusChannel {\n\t}\n\treturn err\n}", "func requestStatistics(ch *api.Channel) {\n\tch.SendRequest(&stats.WantStats{\n\t\tPid: uint32(os.Getpid()),\n\t\tEnableDisable: 1,\n\t}).ReceiveReply(&stats.WantStatsReply{})\n}", "func (cop *ConnPool) Stats() *ConnPoolStats {\n\tcop.l.Lock()\n\tdefer cop.l.Unlock()\n\n\treturn &ConnPoolStats{\n\t\tIdleChaNum: len(cop.idleChas),\n\t\tConnNum: len(cop.conns),\n\t\tBusyChaNum: cop.getChaBusyNum(),\n\t\tReqChaNum: cop.reqChaList.Len(),\n\t}\n}", "func (c *Collector) Run() {\n\ttick := time.NewTicker(c.interval)\n\tdefer tick.Stop()\n\tfor {\n\t\tselect {\n\t\tcase <-c.done:\n\t\t\treturn\n\t\tcase <-tick.C:\n\t\t\tc.emitStats()\n\t\t}\n\t}\n}", "func (c *ClusterScalingScheduleCollector) Interval() time.Duration {\n\treturn c.interval\n}", "func (s *scanCoordinator) handleStats(cmd Message) {\n\ts.supvCmdch <- &MsgSuccess{}\n\n\treq := cmd.(*MsgStatsRequest)\n\treplych := req.GetReplyChannel()\n\ts.mu.RLock()\n\tdefer s.mu.RUnlock()\n\n\tstats := s.stats.Get()\n\tst := s.serv.Statistics()\n\tstats.numConnections.Set(st.Connections)\n\n\t// Compute counts asynchronously and reply to stats request\n\tgo func() {\n\t\tfor id, idxStats := range stats.indexes {\n\t\t\terr := s.updateItemsCount(id, idxStats)\n\t\t\tif err != nil {\n\t\t\t\tlogging.Errorf(\"%v: Unable to compute index items_count for %v/%v/%v state %v (%v)\", s.logPrefix,\n\t\t\t\t\tidxStats.bucket, idxStats.name, id, idxStats.indexState.Value(), err)\n\t\t\t}\n\n\t\t\t// compute scan rate\n\t\t\tnow := time.Now().UnixNano()\n\t\t\telapsed := float64(now-idxStats.lastScanGatherTime.Value()) / float64(time.Second)\n\t\t\tif elapsed > 60 {\n\t\t\t\tpartitions := idxStats.getPartitions()\n\t\t\t\tfor _, pid := range partitions {\n\t\t\t\t\tpartnStats := idxStats.getPartitionStats(pid)\n\t\t\t\t\tnumRowsScanned := partnStats.numRowsScanned.Value()\n\t\t\t\t\tif idxStats.lastScanGatherTime.Value() != int64(0) {\n\t\t\t\t\t\tscanRate := float64(numRowsScanned-partnStats.lastNumRowsScanned.Value()) / elapsed\n\t\t\t\t\t\tpartnStats.avgScanRate.Set(int64((scanRate + float64(partnStats.avgScanRate.Value())) / 2))\n\t\t\t\t\t\tlogging.Debugf(\"scanCoordinator.handleStats: index %v partition %v numRowsScanned %v scan rate %v avg scan rate %v\",\n\t\t\t\t\t\t\tid, pid, numRowsScanned, scanRate, partnStats.avgScanRate.Value())\n\t\t\t\t\t}\n\t\t\t\t\tpartnStats.lastNumRowsScanned.Set(numRowsScanned)\n\t\t\t\t\tidxStats.lastScanGatherTime.Set(now)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\treplych <- true\n\t}()\n}", "func MetricsSendStats(Segment *Segment, Subnet *Subnet, Duration time.Duration) (err error) {\n\tif !o.MetricsEnabled {\n\t\treturn\n\t}\n\n\tTags := map[string]string{\n\t\t\"ServerID\": o.ServerID,\n\t\t\"SegmentId\": strconv.Itoa(Segment.Id),\n\t\t\"SegmentName\": Segment.Name,\n\t\t\"Subnet\": Subnet.NetStr,\n\t}\n\n\tFields := map[string]interface{}{\n\t\t\"Duration\": int(Duration.Nanoseconds() / 1000),\n\t\t\"LeasesTotal\": Subnet.Capacity(),\n\t\t\"LeasesActive\": Subnet.LeasesActive(),\n\t\t\"LeasesExpired\": Subnet.LeasesExpired(),\n\t}\n\n\tInfluxDB.SendMetric(&metrics.InfluxDBMetric{\n\t\tMeasurement: o.MetricsMeasurementStats,\n\t\tTimestamp: time.Now(),\n\n\t\tTags: Tags,\n\t\tFields: Fields,\n\t})\n\n\treturn\n}", "func (r *GatewayConnectionStatsRegistry) Get(ctx context.Context, ids ttnpb.GatewayIdentifiers) (*ttnpb.GatewayConnectionStats, error) {\n\tuid := unique.ID(ctx, ids)\n\tresult := &ttnpb.GatewayConnectionStats{}\n\tstats := &ttnpb.GatewayConnectionStats{}\n\n\tretrieved, err := r.Redis.MGet(r.key(upKey, uid), r.key(downKey, uid), r.key(statusKey, uid)).Result()\n\tif err != nil {\n\t\treturn nil, ttnredis.ConvertError(err)\n\t}\n\n\tif retrieved[0] == nil && retrieved[1] == nil && retrieved[2] == nil {\n\t\treturn nil, errNotFound\n\t}\n\n\t// Retrieve uplink stats.\n\tif retrieved[0] != nil {\n\t\tif err = ttnredis.UnmarshalProto(retrieved[0].(string), stats); err != nil {\n\t\t\treturn nil, errInvalidStats.WithAttributes(\"type\", \"uplink\").WithCause(err)\n\t\t}\n\t\tresult.LastUplinkReceivedAt = stats.LastUplinkReceivedAt\n\t\tresult.UplinkCount = stats.UplinkCount\n\t}\n\n\t// Retrieve downlink stats.\n\tif retrieved[1] != nil {\n\t\tif err = ttnredis.UnmarshalProto(retrieved[1].(string), stats); err != nil {\n\t\t\treturn nil, errInvalidStats.WithAttributes(\"type\", \"downlink\").WithCause(err)\n\t\t}\n\t\tresult.LastDownlinkReceivedAt = stats.LastDownlinkReceivedAt\n\t\tresult.DownlinkCount = stats.DownlinkCount\n\t\tresult.RoundTripTimes = stats.RoundTripTimes\n\t}\n\n\t// Retrieve gateway status.\n\tif retrieved[2] != nil {\n\t\tif err = ttnredis.UnmarshalProto(retrieved[2].(string), stats); err != nil {\n\t\t\treturn nil, errInvalidStats.WithAttributes(\"type\", \"status\").WithCause(err)\n\t\t}\n\t\tresult.ConnectedAt = stats.ConnectedAt\n\t\tresult.Protocol = stats.Protocol\n\t\tresult.LastStatus = stats.LastStatus\n\t\tresult.LastStatusReceivedAt = stats.LastStatusReceivedAt\n\t}\n\n\treturn result, nil\n}", "func StatQps() {\n\tif netConf().QpsCountInterval > 0 {\n\t\tatomic.StoreUint64(&qpsCounterInterval, uint64(netConf().QpsCountInterval))\n\t}\n\n\tticker := time.NewTicker(time.Second * time.Duration(qpsCounterInterval))\n\tdefer func() {\n\t\tticker.Stop()\n\t}()\n\n\tlastReqStat := newGatewayStat()\n\n\tfor {\n\t\t<-ticker.C // wait a moment\n\n\t\t// update read/write data\n\t\trequestStat.AtomicSetReadWrites(monitor.MonitorData())\n\n\t\t// get current stat data atomicly\n\t\tcurrentReqStat := requestStat.AtomicCopyTo(nil)\n\t\t// update stat switch state\n\t\tif netConf().StatResponseTime != currentReqStat.AtomicGetStatResponse() {\n\t\t\tcurrentReqStat.AtomicSetStatResponse(netConf().StatResponseTime)\n\t\t}\n\t\t// record last stat data before update it\n\t\tsub := lastReqStat.AtomicCopyTo(nil)\n\t\t// update last stat data atomicly\n\t\tcurrentReqStat.AtomicCopyTo(lastReqStat)\n\t\t// count current stat, not atomic for currentReqStat and sub both are local variables\n\t\tcurrentReqStat.Sub(sub)\n\t\t// make qps and update result atomicly\n\t\tcurrentReqStat.AtomicMakeQps(qpsCounterInterval)\n\t\tcurrentReqStat.AtomicCopyTo(qpsData) // note that qps data is computed yet\n\t\t// log QPS data to trace log\n\t\tLogger.Trace(\"\", \"\", \"\", \"StatQps\", qpsData.QpsString())\n\n\t\t// update tick timer if config is changed\n\t\tif uint64(netConf().QpsCountInterval) != qpsCounterInterval && netConf().QpsCountInterval > 0 {\n\t\t\tticker.Stop() // should stop old ticker explicitly\n\t\t\tLogger.Debug(\"\", \"\", \"\", \"StatQps\",\n\t\t\t\tfmt.Sprintf(\"QPS counter interval is changed from %v to %v.\",\n\t\t\t\t\tqpsCounterInterval, netConf().QpsCountInterval))\n\n\t\t\t// update QPS counter interval to new value\n\t\t\tatomic.StoreUint64(&qpsCounterInterval, uint64(netConf().QpsCountInterval))\n\n\t\t\t// make new time ticker\n\t\t\tticker = time.NewTicker(time.Second * time.Duration(qpsCounterInterval))\n\t\t}\n\t}\n}", "func MetricsSendStatsSegment(Segment *Segment, Duration time.Duration) (err error) {\n\tif !o.MetricsEnabled {\n\t\treturn\n\t}\n\n\tTags := map[string]string{\n\t\t\"ServerID\": o.ServerID,\n\t\t\"SegmentId\": strconv.Itoa(Segment.Id),\n\t\t\"SegmentName\": Segment.Name,\n\t}\n\n\tFields := map[string]interface{}{\n\t\t\"Duration\": int(Duration.Nanoseconds() / 1000),\n\t\t\"LeasesTotal\": Segment.LeasesTotal,\n\t\t\"LeasesActive\": Segment.LeasesActive,\n\t\t\"LeasesExpired\": Segment.LeasesExpired,\n\t}\n\n\tInfluxDB.SendMetric(&metrics.InfluxDBMetric{\n\t\tMeasurement: o.MetricsMeasurementStatsSegment,\n\t\tTimestamp: time.Now(),\n\n\t\tTags: Tags,\n\t\tFields: Fields,\n\t})\n\n\treturn\n}", "func runSendMetricsLoop(ctx context.Context, config *Config, instanceRepo *service.Repository) error {\n\tconst lastSendTSFile = \"/tmp/pgscv-last-send.timestamp\"\n\n\tlog.Infof(\"sending metrics to %s every %d seconds\", config.SendMetricsURL, config.SendMetricsInterval/time.Second)\n\n\t// Before sending metrics wait until any services appear in the repo, else need to wait an one MetricsSendInterval.\n\t// This is the one-time operation and here is using a naive approach with 'for loop + sleep' instead of channels/sync stuff.\n\tlog.Debugln(\"waiting for services appear in service repo...\")\n\tfor {\n\t\ttime.Sleep(time.Second)\n\t\tif n := instanceRepo.TotalServices(); n > 0 {\n\t\t\tlog.Debugln(\"done, services found: \", n)\n\t\t\tbreak\n\t\t}\n\t}\n\n\tsendClient, err := newSendClient(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Initialize last send timestamp from file.\n\tlastSendTS := readLastSendTS(lastSendTSFile)\n\n\t// Do one-time sleep depending on last send timestamp staleness.\n\ttime.Sleep(lastSendStaleness(lastSendTS, config.SendMetricsInterval))\n\n\tticker := time.NewTicker(config.SendMetricsInterval)\n\n\tvar delay time.Duration\n\tfor {\n\t\tif delay > 0 {\n\t\t\tlog.Debugf(\"waiting for delay %s\", delay.String())\n\t\t\ttime.Sleep(delay)\n\t\t}\n\n\t\tbuf, err := sendClient.readMetrics()\n\t\tif err != nil {\n\t\t\tdelay = time.Second\n\t\t\tlog.Infof(\"read metrics failed: %s, retry after %s\", err, delay.String())\n\t\t\tcontinue\n\t\t}\n\n\t\tlastSendTS = time.Now().Unix()\n\n\t\terr = sendClient.sendMetrics(buf)\n\t\tif err != nil {\n\t\t\tdelay = addDelay(delay)\n\t\t\tlog.Infof(\"send metrics failed: %s, retry after %s\", err, delay.String())\n\t\t\tcontinue\n\t\t}\n\n\t\t// Reading and sending successful, reset delay.\n\t\tdelay = 0\n\n\t\t// Update last successful send timestamp, in case of pgSCV restarts\n\t\twriteLastSendTS(lastSendTS, lastSendTSFile)\n\n\t\t// Sleeping for next iteration.\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\tlog.Info(\"exit signaled, stop metrics sending\")\n\t\t\tticker.Stop()\n\t\t\treturn nil\n\t\tcase <-ticker.C:\n\t\t\tcontinue\n\t\t}\n\t}\n}", "func runSendMetricsLoop(ctx context.Context, config *Config, instanceRepo *service.Repository) error {\n\tconst lastSendTSFile = \"/tmp/pgscv-last-send.timestamp\"\n\n\tlog.Infof(\"sending metrics to %s every %d seconds\", config.SendMetricsURL, config.SendMetricsInterval/time.Second)\n\n\t// Before sending metrics wait until any services appear in the repo, else need to wait an one MetricsSendInterval.\n\t// This is the one-time operation and here is using a naive approach with 'for loop + sleep' instead of channels/sync stuff.\n\tlog.Debugln(\"waiting for services appear in service repo...\")\n\tfor {\n\t\ttime.Sleep(time.Second)\n\t\tif n := instanceRepo.TotalServices(); n > 0 {\n\t\t\tlog.Debugln(\"done, services found: \", n)\n\t\t\tbreak\n\t\t}\n\t}\n\n\tsendClient, err := newSendClient(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Initialize last send timestamp from file.\n\tlastSendTS := readLastSendTS(lastSendTSFile)\n\n\t// Do one-time sleep depending on last send timestamp staleness.\n\ttime.Sleep(lastSendStaleness(lastSendTS, config.SendMetricsInterval))\n\n\tticker := time.NewTicker(config.SendMetricsInterval)\n\n\tvar delay time.Duration\n\tfor {\n\t\tif delay > 0 {\n\t\t\tlog.Debugf(\"waiting for delay %s\", delay.String())\n\t\t\ttime.Sleep(delay)\n\t\t}\n\n\t\tbuf, err := sendClient.readMetrics()\n\t\tif err != nil {\n\t\t\tdelay = time.Second\n\t\t\tlog.Infof(\"read metrics failed: %s, retry after %s\", err, delay.String())\n\t\t\tcontinue\n\t\t}\n\n\t\tlastSendTS = time.Now().Unix()\n\n\t\terr = sendClient.sendMetrics(buf)\n\t\tif err != nil {\n\t\t\tdelay = addDelay(delay)\n\t\t\tlog.Infof(\"send metrics failed: %s, retry after %s\", err, delay.String())\n\t\t\tcontinue\n\t\t}\n\n\t\t// Reading and sending successful, reset delay.\n\t\tdelay = 0\n\n\t\t// Update last successful send timestamp, in case of pgSCV restarts\n\t\twriteLastSendTS(lastSendTS, lastSendTSFile)\n\n\t\t// Sleeping for next iteration.\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\tlog.Info(\"exit signaled, stop metrics sending\")\n\t\t\tticker.Stop()\n\t\t\treturn nil\n\t\tcase <-ticker.C:\n\t\t\tcontinue\n\t\t}\n\t}\n}", "func (s *SystemMetrics) GetDiskIOStats(c chan *DiskStats) {\n\tbefore, err := disk.Get()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\ttime.Sleep(time.Duration(time.Second * 3))\n\n\tnow, err := disk.Get()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tc <- &DiskStats{\n\t\t// default is always in bytes. hence, convert into the required format.\n\t\tDiskIO: (int(now.Used) - int(before.Used)) / 1000, // in kilo-bytes\n\t\tCached: int(now.Cached / 1000000), // mega-bytes\n\t}\n}", "func ctxForOp() (context.Context, func()) {\n\treturn context.WithTimeout(context.Background(), timeoutOp)\n}", "func (client *Client) DescribeScalingGroupsWithChan(request *DescribeScalingGroupsRequest) (<-chan *DescribeScalingGroupsResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeScalingGroupsResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeScalingGroups(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func ReportStats(ctx context.Context, pool *redis.Pool, name string) {\n\tstats := pool.Stats()\n\tconnsMetric.Set(ctx, int64(stats.IdleCount), name, \"idle\")\n\tconnsMetric.Set(ctx, int64(stats.ActiveCount-stats.IdleCount), name, \"in-use\")\n\twaitCountMetric.Set(ctx, int64(stats.WaitCount), name)\n\twaitDurationMetric.Set(ctx, int64(stats.WaitDuration.Nanoseconds()/1000), name)\n}", "func (client *Client) DescribeScalingConfigurationsWithChan(request *DescribeScalingConfigurationsRequest) (<-chan *DescribeScalingConfigurationsResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeScalingConfigurationsResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeScalingConfigurations(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (fc *familyChannel) writeTask(_ context.Context) {\n\t// on avg 2 * limit could avoid buffer grow\n\tticker := time.NewTicker(fc.checkFlushInterval)\n\tdefer ticker.Stop()\n\n\tretryBuffers := make([]*compressedChunk, 0)\n\tretry := func(compressed *compressedChunk) {\n\t\tif len(retryBuffers) > fc.maxRetryBuf {\n\t\t\tfc.logger.Error(\"too many retry messages, drop current message\")\n\t\t\tfc.statistics.RetryDrop.Incr()\n\t\t} else {\n\t\t\tretryBuffers = append(retryBuffers, compressed)\n\t\t\tfc.statistics.Retry.Incr()\n\t\t}\n\t}\n\tvar stream rpc.WriteStream\n\tsend := func(compressed *compressedChunk) bool {\n\t\tif compressed == nil {\n\t\t\treturn true\n\t\t}\n\t\tif len(*compressed) == 0 {\n\t\t\tcompressed.Release()\n\t\t\treturn true\n\t\t}\n\t\tif stream == nil {\n\t\t\tfc.lock4meta.Lock()\n\t\t\tleader := fc.liveNodes[fc.shardState.Leader]\n\t\t\tshardState := fc.shardState\n\t\t\tfc.currentTarget = &leader\n\t\t\tfc.lock4meta.Unlock()\n\t\t\ts, err := fc.newWriteStreamFn(fc.ctx, fc.currentTarget, fc.database, &shardState, fc.familyTime, fc.fct)\n\t\t\tif err != nil {\n\t\t\t\tfc.statistics.CreateStreamFailures.Incr()\n\t\t\t\tretry(compressed)\n\t\t\t\treturn false\n\t\t\t}\n\t\t\tfc.statistics.CreateStream.Incr()\n\t\t\tstream = s\n\t\t}\n\t\tif err := stream.Send(*compressed); err != nil {\n\t\t\tfc.statistics.SendFailure.Incr()\n\t\t\tfc.logger.Error(\n\t\t\t\t\"failed writing compressed chunk to storage\",\n\t\t\t\tlogger.String(\"target\", fc.currentTarget.Indicator()),\n\t\t\t\tlogger.String(\"database\", fc.database),\n\t\t\t\tlogger.Error(err))\n\t\t\tif err == io.EOF {\n\t\t\t\tif closeError := stream.Close(); closeError != nil {\n\t\t\t\t\tfc.statistics.CloseStreamFailures.Incr()\n\t\t\t\t\tfc.logger.Error(\"failed closing write stream\",\n\t\t\t\t\t\tlogger.String(\"target\", fc.currentTarget.Indicator()),\n\t\t\t\t\t\tlogger.Error(closeError))\n\t\t\t\t} else {\n\t\t\t\t\tfc.statistics.CloseStream.Incr()\n\t\t\t\t}\n\t\t\t\tstream = nil\n\t\t\t}\n\t\t\t// retry if err\n\t\t\tretry(compressed)\n\t\t\treturn false\n\t\t}\n\t\tfc.statistics.SendSuccess.Incr()\n\t\tfc.statistics.SendSize.Add(float64(len(*compressed)))\n\t\tfc.statistics.PendingSend.Decr()\n\t\tcompressed.Release()\n\t\treturn true\n\t}\n\n\tdefer func() {\n\t\tif stream != nil {\n\t\t\tif err := stream.Close(); err != nil {\n\t\t\t\tfc.statistics.CloseStreamFailures.Incr()\n\t\t\t\tfc.logger.Error(\"close write stream err when exit write task\", logger.Error(err))\n\t\t\t} else {\n\t\t\t\tfc.statistics.CloseStream.Incr()\n\t\t\t}\n\t\t}\n\t}()\n\n\t// send pending in buffer before stop channel.\n\tsendBeforeStop := func() {\n\t\tdefer func() {\n\t\t\tfc.stoppedSignal <- struct{}{}\n\t\t}()\n\t\tsendLastMsg := func(compressed *compressedChunk) {\n\t\t\tif !send(compressed) {\n\t\t\t\tfc.logger.Error(\"send message failure before close channel, message lost\")\n\t\t\t}\n\t\t}\n\t\t// flush chunk pending data if chunk not empty\n\t\tif !fc.chunk.IsEmpty() {\n\t\t\t// flush chunk pending data if chunk not empty\n\t\t\tcompressed, err0 := fc.chunk.Compress()\n\t\t\tif err0 != nil {\n\t\t\t\tfc.logger.Error(\"compress chunk err when send last chunk data\", logger.Error(err0))\n\t\t\t} else {\n\t\t\t\tsendLastMsg(compressed)\n\t\t\t}\n\t\t}\n\t\tfc.sendPendingMessage(sendLastMsg)\n\t}\n\tvar err error\n\tfor {\n\t\tselect {\n\t\tcase <-fc.stoppingSignal:\n\t\t\tsendBeforeStop()\n\t\t\treturn\n\t\tcase <-fc.ctx.Done():\n\t\t\tsendBeforeStop()\n\t\t\treturn\n\t\tcase <-fc.leaderChangedSignal:\n\t\t\tif stream != nil {\n\t\t\t\tfc.logger.Info(\"shard leader changed, need switch send stream\",\n\t\t\t\t\tlogger.String(\"oldTarget\", fc.currentTarget.Indicator()),\n\t\t\t\t\tlogger.String(\"database\", fc.database))\n\t\t\t\t// if stream isn't nil, need close old stream first.\n\t\t\t\tif err = stream.Close(); err != nil {\n\t\t\t\t\tfc.logger.Error(\"close write stream err when leader changed\", logger.Error(err))\n\t\t\t\t}\n\t\t\t\tstream = nil\n\t\t\t}\n\t\tcase compressed := <-fc.ch:\n\t\t\tif send(compressed) {\n\t\t\t\t// if send ok, retry pending message\n\t\t\t\tif len(retryBuffers) > 0 {\n\t\t\t\t\tmessages := retryBuffers\n\t\t\t\t\tretryBuffers = make([]*compressedChunk, 0)\n\t\t\t\t\tfor _, msg := range messages {\n\t\t\t\t\t\tif !send(msg) {\n\t\t\t\t\t\t\tretry(msg)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tstream = nil\n\t\t\t}\n\t\tcase <-ticker.C:\n\t\t\t// check\n\t\t\tfc.checkFlush()\n\t\t}\n\t}\n}", "func (q *priorityLocalQueue) JobStats(ctx context.Context) <-chan amboy.JobStatusInfo {\n\tout := make(chan amboy.JobStatusInfo)\n\n\tgo func() {\n\t\tdefer close(out)\n\t\tfor job := range q.storage.Contents() {\n\t\t\tif ctx.Err() != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tstat := job.Status()\n\t\t\tstat.ID = job.ID()\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn\n\t\t\tcase out <- stat:\n\t\t\t}\n\n\t\t}\n\t}()\n\n\treturn out\n}" ]
[ "0.5239738", "0.51784253", "0.49386886", "0.48893714", "0.4876019", "0.48401034", "0.475298", "0.47351772", "0.46660122", "0.4591185", "0.45741856", "0.4563549", "0.4537039", "0.45309833", "0.44815376", "0.44753942", "0.44719172", "0.44595218", "0.44447067", "0.44192713", "0.44035816", "0.4395367", "0.43924952", "0.43657732", "0.43614635", "0.43396792", "0.43264896", "0.43191135", "0.430158", "0.42945367", "0.42837557", "0.42826116", "0.42825502", "0.42798525", "0.42763588", "0.42577285", "0.42541105", "0.42525005", "0.42436898", "0.42423838", "0.42234796", "0.4206158", "0.41990012", "0.4186185", "0.41780144", "0.41773725", "0.4171294", "0.41635787", "0.41606277", "0.4160526", "0.41534114", "0.4137044", "0.41286725", "0.4123934", "0.41213515", "0.41176784", "0.4104555", "0.4097475", "0.40845308", "0.40831488", "0.4078657", "0.40648785", "0.40602034", "0.40317136", "0.40305576", "0.40283933", "0.40236753", "0.4021039", "0.40200582", "0.4019384", "0.4011452", "0.40092883", "0.40087175", "0.40048087", "0.39879572", "0.3986016", "0.39827153", "0.39752448", "0.3971937", "0.396903", "0.39685762", "0.39666364", "0.3965311", "0.39619246", "0.39546525", "0.3951148", "0.39508137", "0.39474744", "0.39442065", "0.39410305", "0.3936864", "0.39320424", "0.39320424", "0.39299035", "0.39211193", "0.39192903", "0.39108086", "0.3908762", "0.39083767", "0.39082915" ]
0.67854583
0
TaskEvents function allows the driver to publish driver specific events about tasks and the Nomad client publishes events associated with an allocation.
func (d *Driver) TaskEvents(ctx context.Context) (<-chan *drivers.TaskEvent, error) { return d.eventer.TaskEvents(ctx) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *RabbitMQConnection) PublishEventsTrackingTask(payload []byte) error {\n\tch, err := c.Channel()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer ch.Close()\n\n\terr = ch.Publish(\n\t\t\"tracking\", //Exchange\n\t\t\"tracking-queue\", //Routing key\n\t\tfalse, //Mandatory\n\t\tfalse, //Immediate\n\t\tamqp.Publishing{\n\t\t\tDeliveryMode: amqp.Persistent,\n\t\t\tContentType: \"text/plain\",\n\t\t\tBody: payload,\n\t\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (engine *DockerTaskEngine) TaskEvents() (<-chan api.ContainerStateChange, <-chan error) {\n\treturn engine.container_events, engine.event_errors\n}", "func consumeTaskEvents(event interface{}) {\n\tdata, _ := json.Marshal(&event)\n\tvar eventData common.Events\n\terr := json.Unmarshal(data, &eventData)\n\tif err != nil {\n\t\tl.Log.Error(\"Error while consuming task events\", err)\n\t\treturn\n\t}\n\n\tvar taskEvent dmtf.Event\n\terr = json.Unmarshal(eventData.Request, &taskEvent)\n\tif err != nil {\n\t\tl.Log.Error(\"Error while consuming task events\", err)\n\t\treturn\n\t}\n\n\tfor _, eventRecord := range taskEvent.Events {\n\t\tTaskEventRecvQueue <- eventRecord\n\t}\n}", "func TaskProcessInfoEvents(taskId string, ts time.Time, limit, sort int) db.Q {\n\tfilter := bson.M{\n\t\tDataKey + \".\" + ResourceTypeKey: EventTaskProcessInfo,\n\t\tResourceIdKey: taskId,\n\t\tTypeKey: EventTaskProcessInfo,\n\t}\n\n\tsortSpec := TimestampKey\n\n\tif sort < 0 {\n\t\tsortSpec = \"-\" + sortSpec\n\t\tfilter[TimestampKey] = bson.M{\"$lte\": ts}\n\t} else {\n\t\tfilter[TimestampKey] = bson.M{\"$gte\": ts}\n\t}\n\n\treturn db.Query(filter).Sort([]string{sortSpec}).Limit(limit)\n}", "func TaskSystemInfoEvents(taskId string, ts time.Time, limit, sort int) db.Q {\n\tfilter := bson.M{\n\t\tDataKey + \".\" + ResourceTypeKey: EventTaskSystemInfo,\n\t\tResourceIdKey: taskId,\n\t\tTypeKey: EventTaskSystemInfo,\n\t}\n\n\tsortSpec := TimestampKey\n\n\tif sort < 0 {\n\t\tsortSpec = \"-\" + sortSpec\n\t\tfilter[TimestampKey] = bson.M{\"$lte\": ts}\n\t} else {\n\t\tfilter[TimestampKey] = bson.M{\"$gte\": ts}\n\t}\n\n\treturn db.Query(filter).Sort([]string{sortSpec}).Limit(limit)\n}", "func (c *subContext) processTaskEvents(ctx context.Context, eventCh <-chan subtaskapi.Event, indCh chan<- indication.Indication) {\n\t// Create a wait group to close the indications channel\n\twg := &sync.WaitGroup{}\n\n\t// Wait for the watch context to be done\n\twg.Add(1)\n\tgo func() {\n\t\t<-ctx.Done()\n\t\twg.Done()\n\t}()\n\n\t// Once the wait group is complete, close the indications channel\n\tgo func() {\n\t\twg.Wait()\n\t\tclose(indCh)\n\t}()\n\n\tvar prevCancel context.CancelFunc\n\tvar prevEndpoint epapi.ID\n\tfor event := range eventCh {\n\t\t// Only interested in tasks related to this subscription\n\t\tif event.Task.SubscriptionID != c.sub.ID {\n\t\t\tcontinue\n\t\t}\n\n\t\t// If the stream is already open for the associated E2 endpoint, skip the event\n\t\tif event.Task.EndpointID == prevEndpoint && event.Task.Lifecycle.Failure == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\t// If the task failed, propagate the error\n\t\tif event.Task.Lifecycle.Failure != nil {\n\t\t\tc.errCh <- errors.NewInternal(event.Task.Lifecycle.Failure.Message)\n\t\t}\n\n\t\t// If the task was assigned to a new endpoint, close the prior stream and open a new one.\n\t\t// If the task was unassigned, close the prior stream and wait for a new event.\n\t\tif event.Type == subtaskapi.EventType_NONE || event.Type == subtaskapi.EventType_CREATED {\n\t\t\tif prevCancel != nil {\n\t\t\t\tprevCancel()\n\t\t\t}\n\t\t\twg.Add(1)\n\t\t\tctx, cancel := context.WithCancel(context.Background())\n\t\t\tgo func(epID epapi.ID) {\n\t\t\t\tdefer wg.Done()\n\t\t\t\tdefer cancel()\n\t\t\t\terr := c.openStream(ctx, epID, indCh)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error(err)\n\t\t\t\t}\n\t\t\t}(event.Task.EndpointID)\n\t\t\tprevEndpoint = event.Task.EndpointID\n\t\t\tprevCancel = cancel\n\t\t} else if event.Type == subtaskapi.EventType_REMOVED {\n\t\t\tprevEndpoint = \"\"\n\t\t\tif prevCancel != nil {\n\t\t\t\tprevCancel()\n\t\t\t\tprevCancel = nil\n\t\t\t}\n\t\t}\n\t}\n}", "func Event(ctx context.Context, scope EventScope, name string, args ...interface{}) {\n\tt := GetTask(ctx)\n\tonEvent(ctx, t, scope, name, args)\n}", "func (e *Endpoints) DealTaskEvent(ctx context.Context, r *http.Request, vars map[string]string) (httpserver.Responser, error) {\n\tvar (\n\t\treq apistructs.PipelineTaskEvent\n\t\trunningTaskID int64\n\t\terr error\n\t)\n\tif r.Body == nil {\n\t\treturn apierrors.ErrDealTaskEvents.MissingParameter(\"body\").ToResp(), nil\n\t}\n\n\tif err := json.NewDecoder(r.Body).Decode(&req); err != nil {\n\t\treturn apierrors.ErrDealTaskEvents.InvalidParameter(err).ToResp(), nil\n\t}\n\tlogrus.Debugf(\"ReceiveTaskEvents: request body: %+v\", req)\n\n\tif req.Event == \"pipeline_task\" {\n\t\tif runningTaskID, err = e.orgResource.DealReceiveTaskEvent(&req); err != nil {\n\t\t\treturn apierrors.ErrDealTaskEvents.InvalidParameter(err).ToResp(), nil\n\t\t}\n\t} else if req.Event == \"pipeline_task_runtime\" {\n\t\tif runningTaskID, err = e.orgResource.DealReceiveTaskRuntimeEvent(&req); err != nil {\n\t\t\treturn apierrors.ErrDealTaskEvents.InvalidParameter(err).ToResp(), nil\n\t\t}\n\t}\n\n\treturn httpserver.OkResp(runningTaskID)\n}", "func (d *Datastore) WriteEvent(ctx context.Context, e *events.Event) error {\n\n\tswitch e.Type {\n\n\tcase events.Type_TASK_CREATED:\n\t\tputKeys, putData := marshalTask(e.GetTask())\n\t\t_, err := d.client.PutMulti(ctx, putKeys, putData)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\tcase events.Type_EXECUTOR_STDOUT:\n\t\t_, err := d.client.Put(ctx, stdoutKey(e.Id, e.Attempt, e.Index), marshalEvent(e))\n\t\treturn err\n\n\tcase events.Type_EXECUTOR_STDERR:\n\t\t_, err := d.client.Put(ctx, stderrKey(e.Id, e.Attempt, e.Index), marshalEvent(e))\n\t\treturn err\n\n\tcase events.Type_TASK_STATE:\n\t\tres, err := d.GetTask(ctx, &tes.GetTaskRequest{\n\t\t\tId: e.Id,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tfrom := res.State\n\t\tto := e.GetState()\n\t\tif err := tes.ValidateTransition(from, to); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfallthrough\n\n\tdefault:\n\t\t_, err := d.client.RunInTransaction(ctx, func(tx *datastore.Transaction) error {\n\t\t\tprops := datastore.PropertyList{}\n\t\t\terr := tx.Get(taskKey(e.Id), &props)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\ttask := &tes.Task{}\n\t\t\tunmarshalTask(task, props)\n\t\t\ttb := events.TaskBuilder{Task: task}\n\t\t\terr = tb.WriteEvent(context.Background(), e)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tputKeys, putData := marshalTask(task)\n\t\t\t_, err = tx.PutMulti(putKeys, putData)\n\t\t\treturn err\n\t\t})\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *BaseAspidaListener) EnterTasks(ctx *TasksContext) {}", "func (c *CloudWatchLogs) TaskLogEvents(logGroupName string, streamLastEventTime map[string]int64, opts ...GetLogEventsOpts) (*LogEventsOutput, error) {\n\tvar events []*Event\n\tvar in *cloudwatchlogs.GetLogEventsInput\n\tlogStreamNames, err := c.logStreams(logGroupName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor _, logStreamName := range logStreamNames {\n\t\tin = &cloudwatchlogs.GetLogEventsInput{\n\t\t\tLogGroupName: aws.String(logGroupName),\n\t\t\tLogStreamName: logStreamName,\n\t\t\tLimit: aws.Int64(10), // default to be 10\n\t\t}\n\t\tfor _, opt := range opts {\n\t\t\topt(in)\n\t\t}\n\t\tif streamLastEventTime[*logStreamName] != 0 {\n\t\t\t// If last event for this log stream exists, increment last log event timestamp\n\t\t\t// by one to get logs after the last event.\n\t\t\tin.SetStartTime(streamLastEventTime[*logStreamName] + 1)\n\t\t}\n\t\t// TODO: https://github.com/aws/amazon-ecs-cli-v2/pull/628#discussion_r374291068 and https://github.com/aws/amazon-ecs-cli-v2/pull/628#discussion_r374294362\n\t\tresp, err := c.client.GetLogEvents(in)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"get log events of %s/%s: %w\", logGroupName, *logStreamName, err)\n\t\t}\n\n\t\tfor _, event := range resp.Events {\n\t\t\tlog := &Event{\n\t\t\t\tLogStreamName: trimLogStreamName(*logStreamName),\n\t\t\t\tIngestionTime: aws.Int64Value(event.IngestionTime),\n\t\t\t\tMessage: aws.StringValue(event.Message),\n\t\t\t\tTimestamp: aws.Int64Value(event.Timestamp),\n\t\t\t}\n\t\t\tevents = append(events, log)\n\t\t}\n\t\tif len(resp.Events) != 0 {\n\t\t\tstreamLastEventTime[*logStreamName] = *resp.Events[len(resp.Events)-1].Timestamp\n\t\t}\n\t}\n\tsort.SliceStable(events, func(i, j int) bool { return events[i].Timestamp < events[j].Timestamp })\n\tvar truncatedEvents []*Event\n\tif len(events) >= int(*in.Limit) {\n\t\ttruncatedEvents = events[len(events)-int(*in.Limit):]\n\t} else {\n\t\ttruncatedEvents = events\n\t}\n\treturn &LogEventsOutput{\n\t\tEvents: truncatedEvents,\n\t\tLastEventTime: streamLastEventTime,\n\t}, nil\n}", "func Tasks(w http.ResponseWriter, r *http.Request) {\n\tvar Tasks []Task\n\tsql := `SELECT\n\t\t\t task_id,\n\t\t\t mobile_number,\n\t\t\t email,\n\t\t\t reminder_message,\n\t\t\t IFNULL(reminder_time, ''),\n\t\t\t reminder_status\n\t\t\tFROM task`\n\trows, err := db.Query(sql)\n\tif err != nil {\n\t\te := ErrorMessage{err.Error()}\n\t\tfmt.Fprintln(w, json.NewEncoder(w).Encode(e))\n\t\treturn\n\t}\n\tvar task_id, reminder_status int\n\tvar mobile_number, email, reminder_message, reminder_time string\n\t//var created_at, updated_at time.Time\n\tdefer rows.Close()\n\tfor rows.Next() {\n\t\terr := rows.Scan(&task_id, &mobile_number, &email, &reminder_message, &reminder_time, &reminder_status)\n\t\tif err != nil {\n\t\t\te := ErrorMessage{err.Error()}\n\t\t\tfmt.Fprintln(w, json.NewEncoder(w).Encode(e))\n\t\t\treturn\n\t\t}\n\t\tt := Task{task_id, mobile_number, email, reminder_message, reminder_time, reminder_status, time.Now(), time.Now()}\n\t\tTasks = append(Tasks, t)\n\t}\n\tfmt.Fprintln(w, json.NewEncoder(w).Encode(Tasks))\n}", "func GetAllTask(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Context-Type\", \"application/x-www-form-urlencoded\")\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tpayload := getAllTask()\n\tjson.NewEncoder(w).Encode(payload)\n}", "func (sender *Sender) SendEvents(events moira.NotificationEvents, contact moira.ContactData, trigger moira.TriggerData, plots [][]byte, throttled bool) error {\n\tcreateAlertRequest := sender.makeCreateAlertRequest(events, contact, trigger, plots, throttled)\n\t_, err := sender.client.Create(context.Background(), createAlertRequest)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to send %s event message to opsgenie: %s\", trigger.ID, err.Error())\n\t}\n\treturn nil\n}", "func sendEv(taskId string, p micro.Publisher) {\n\n\t// create new event\n\tev := common_proto.Event{\n\t\tEventType: common_proto.Operation_TASK_CANCEL,\n\t\tOpMessage: &common_proto.Event_TaskFeedback{TaskFeedback: &common_proto.TaskFeedback{\n\t\t\tTaskId: taskId,\n\t\t\tStatus: common_proto.TaskStatus_CANCEL_FAILED,\n\t\t}},\n\t}\n\n\tlog.Printf(\"publishing %+v\\n\", ev)\n\n\t// publish an event\n\tif err := p.Publish(context.Background(), &ev); err != nil {\n\t\tlog.Fatalf(\"error publishing %v\\n\", err)\n\t}\n}", "func GetAllTask(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-type\", \"application/json\")\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tpayload := getAllTask()\n\tjson.NewEncoder(w).Encode(payload)\n}", "func SubscribeTaskEventsQueue(topicName string) {\n\tconfig.TLSConfMutex.RLock()\n\tMessageBusConfigFilePath := config.Data.MessageBusConf.MessageBusConfigFilePath\n\tmessagebusType := config.Data.MessageBusConf.MessageBusType\n\tconfig.TLSConfMutex.RUnlock()\n\t// connecting to messagbus\n\tk, err := dc.Communicator(messagebusType, MessageBusConfigFilePath, topicName)\n\tif err != nil {\n\t\tl.Log.Error(\"Unable to connect to kafka\" + err.Error())\n\t\treturn\n\t}\n\t// subscribe from message bus\n\tif err := k.Accept(consumeTaskEvents); err != nil {\n\t\tl.Log.Error(err.Error())\n\t\treturn\n\t}\n}", "func (s *Scheduler) ScheduleTasks() {\n\t/*\n\t\tif events exist unattended, make tasks based on set up times\n\t*/\n\n}", "func (i *instanceManager) dispenseTaskEventsCh() (<-chan *drivers.TaskEvent, context.CancelFunc, error) {\n\tdriver, err := i.dispense()\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tctx, cancel := context.WithCancel(i.ctx)\n\teventsCh, err := driver.TaskEvents(ctx)\n\tif err != nil {\n\t\tcancel()\n\t\treturn nil, nil, err\n\t}\n\n\treturn eventsCh, cancel, nil\n}", "func ProjectEvents(p project.APIProject, c *cli.Context) error {\n\tevents, err := p.Events(context.Background(), c.Args()...)\n\tif err != nil {\n\t\treturn err\n\t}\n\tvar printfn func(eventtypes.Message)\n\n\tif c.Bool(\"json\") {\n\t\tprintfn = printJSON\n\t} else {\n\t\tprintfn = printStd\n\t}\n\tfor event := range events {\n\t\tprintfn(event)\n\t}\n\treturn nil\n}", "func (w *WaitTask) sendTimeoutEvents(taskContext *TaskContext) {\n\tw.mu.RLock()\n\tdefer w.mu.RUnlock()\n\n\tfor _, id := range w.pending {\n\t\terr := taskContext.InventoryManager().SetTimeoutReconcile(id)\n\t\tif err != nil {\n\t\t\t// Object never applied or deleted!\n\t\t\tklog.Errorf(\"Failed to mark object as pending reconcile: %v\", err)\n\t\t}\n\t\tw.sendEvent(taskContext, id, event.ReconcileTimeout)\n\t}\n}", "func NewTaskEventRecorder(eventSink EventSink, scope promutils.Scope) TaskEventRecorder {\n\treturn constructEventRecorder(eventSink, scope)\n}", "func (_Contract *ContractTransactor) PublishEvent(opts *bind.TransactOpts, topicId string, cid string, newAccounts []common.Address) (*types.Transaction, error) {\n\treturn _Contract.contract.Transact(opts, \"publishEvent\", topicId, cid, newAccounts)\n}", "func (server *Server) SendTask(s *TaskSignature) error {\n\tmessage, err := json.Marshal(s)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"JSON Encode Message: %v\", err)\n\t}\n\n\tif err := server.connection.PublishMessage(\n\t\t[]byte(message), s.RoutingKey,\n\t); err != nil {\n\t\treturn fmt.Errorf(\"Publish Message: %v\", err)\n\t}\n\n\treturn nil\n}", "func TaskList(w http.ResponseWriter, r *http.Request) {\n\tvar res Tasks\n\terr := collection.Find(nil).Sort(\"-_id\").All(&res)\n\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Header().Add(\"Content-Type\", \"application/json\")\n\tjson.NewEncoder(w).Encode(res)\n}", "func (x *Root) event(task *Task, event string, message string, data EventDataMap) *Event {\n\n\te := &Event{\n\t\tTaskID: task.uid,\n\t\tComponent: task.component,\n\t\tEvent: event,\n\t\tMessage: message,\n\t\tData: data,\n\n\t\tTimestamp: time.Now(),\n\t}\n\n\tif task.parent != nil {\n\t\te.ParentID = task.parent.uid\n\t}\n\n\tx.events <- e\n\n\treturn e\n\n\t// end event\n}", "func ShowTask(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tvar todo []Todo\n\n\tresult, err := db.Query(\"SELECT * FROM todo\")\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\tdefer result.Close()\n\n\tfor result.Next() {\n\t\tvar tasks Todo\n\n\t\terr := result.Scan(&tasks.Task, &tasks.Datetime_start, &tasks.Datetime_end)\n\t\tif err != nil {\n\t\t\tpanic(err.Error())\n\t\t}\n\t\ttodo = append(todo, tasks)\n\t}\n\tjson.NewEncoder(w).Encode(todo)\n}", "func TaskCreate(w http.ResponseWriter, r *http.Request) {\n\ttask := task.NewTask()\n\tbody, err := ioutil.ReadAll(io.LimitReader(r.Body, 2<<19))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tif err := r.Body.Close(); err != nil {\n\t\tpanic(err)\n\t}\n\tif err := json.Unmarshal(body, task); err != nil {\n\t\tglog.Error(err)\n\t\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\t\tw.WriteHeader(http.StatusUnprocessableEntity) // unprocessable entity\n\t\tif err := json.NewEncoder(w).Encode(err); err != nil {\n\t\t\tglog.Error(err)\n\t\t}\n\t\treturn\n\t}\n\n\tif len(task.Destinations) == 0 {\n\t\tglog.Error(\"List of destinations are empty!\")\n\t\tw.WriteHeader(http.StatusUnprocessableEntity)\n\t\treturn\n\t}\n\n\tfor _, v := range task.Destinations {\n\t\tif len(*v) == 0 {\n\t\t\tglog.Error(\"List of destinations could not be empty!\")\n\t\t\tw.WriteHeader(http.StatusUnprocessableEntity)\n\t\t\treturn\n\t\t}\n\t}\n\n\tglog.Info(string(body))\n\tgo workers.Handler(task)\n}", "func TasksTask() {\n\tfmt.Println(\"Tasks:\")\n\tfmt.Print(\" \")\n\tfor key := range tasks {\n\t\tfmt.Print(\" \", key)\n\t}\n\tfmt.Println(\"\")\n\tfmt.Println(\"Switches:\")\n\tflag.PrintDefaults()\n}", "func (d TinkDB) Events(req *events.WatchRequest, fn func(n informers.Notification) error) error {\n\trows, err := d.instance.Query(`\n\tSELECT id, resource_id, resource_type, event_type, created_at, data\n\tFROM events\n\tWHERE\n\t\tcreated_at >= $1;\n\t`, req.GetWatchEventsFrom().AsTime())\n\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer rows.Close()\n\n\tfor rows.Next() {\n\t\tn := informers.Notification{}\n\t\terr = rows.Scan(&n.ID, &n.ResourceID, &n.ResourceType, &n.EventType, &n.CreatedAt, &n.Data)\n\t\tif err != nil {\n\t\t\terr = errors.Wrap(err, \"SELECT\")\n\t\t\td.logger.Error(err)\n\t\t\treturn err\n\t\t}\n\t\tn.Prefix()\n\t\tif informers.Filter(&n, informers.Reduce(req)) {\n\t\t\tcontinue\n\t\t}\n\t\terr = fn(n)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\terr = rows.Err()\n\tif err == sql.ErrNoRows {\n\t\terr = nil\n\t}\n\treturn err\n}", "func (engine *DockerTaskEngine) emitEvent(event api.ContainerStateChange) {\n\ttask, ok := engine.state.TaskByArn(event.TaskArn)\n\tif !ok {\n\t\tengine.event_errors <- errors.New(\"Event for an unknown task: \" + event.TaskArn)\n\t\treturn\n\t}\n\n\tif task_change := updateTaskState(task); task_change != api.TaskStatusNone {\n\t\tlog.Info(\"Task change event\", \"state\", task_change)\n\t\tevent.TaskStatus = task_change\n\t}\n\tlog.Info(\"Container change event\", \"event\", event)\n\tengine.container_events <- event\n\n\t// Every time something changes, make sure the state for the thing that\n\t// changed is known about and move forwards if this change allows us to\n\tengine.ApplyTaskState(task)\n}", "func CreateNewEventTask(token string, pid string, bid string, name string,\n\tevent int64) (*EventTask, error) {\n\tvar gid int64\n\tsecret_token := nonExistingRandString(sha1.BlockSize,\n\t\t\"SELECT 42 FROM event_tasks WHERE token = $1\")\n\tif err := db.QueryRow(\"WITH row AS (\"+\n\t\t\"INSERT INTO group_tasks (uid, pid, bid) VALUES (\"+\n\t\t\"(SELECT id FROM users WHERE token = $1), $2, $3) RETURNING id\"+\n\t\t\")\"+\n\t\t\"INSERT INTO event_tasks (id, name, status, event, token) \"+\n\t\t\"VALUES ((SELECT id FROM row), $4, $5, $6, $7) RETURNING id\", token,\n\t\tpid, bid, name, Active, event, secret_token).Scan(&gid); err != nil {\n\t\treturn nil, err\n\t}\n\treturn GetEventTask(gid)\n}", "func (p *Pool) AddTask(task *Task) {\n\tp.Collector <-task\n}", "func (tester TestEventScheduler) sendEvents(numEvents int, port string, waitGroup *sync.WaitGroup) {\n\tconn, _ := net.Dial(\"tcp\", \"localhost:\"+port)\n\twriter := bufio.NewWriter(conn)\n\tfor i:=numEvents; i >=1; i-- {\n \t\twriter.WriteString(strconv.Itoa(i)+\"|B\\n\")\n \twriter.Flush()\n\t}\n waitGroup.Done()\n}", "func TestEventsExport(t *testing.T) {\n\t// setup events pipeline to record and distribute events\n\tti := tInfo{}\n\tAssertOk(t, ti.setup(t), \"failed to setup test\")\n\tdefer ti.teardown()\n\n\tvar wg sync.WaitGroup\n\n\t// start UDP server to receive syslog messages\n\tpConn1, receivedMsgsAtUDPServer1, err := serviceutils.StartUDPServer(\":0\")\n\tAssertOk(t, err, \"failed to start UDP server, err: %v\", err)\n\tdefer pConn1.Close()\n\ttmp1 := strings.Split(pConn1.LocalAddr().String(), \":\")\n\n\t// start TCP server - 1 to receive syslog messages\n\tln1, receivedMsgsAtTCPServer1, err := serviceutils.StartTCPServer(\":0\", 100, 0)\n\tAssertOk(t, err, \"failed to start TCP server, err: %v\", err)\n\tdefer ln1.Close()\n\ttmp2 := strings.Split(ln1.Addr().String(), \":\")\n\n\t// start TCP server - 2 to receive syslog messages\n\tln2, receivedMsgsAtTCPServer2, err := serviceutils.StartTCPServer(\":0\", 100, 0)\n\tAssertOk(t, err, \"failed to start TCP server, err: %v\", err)\n\tdefer ln2.Close()\n\ttmp3 := strings.Split(ln2.Addr().String(), \":\")\n\tdefTenant := cluster.Tenant{\n\t\tTypeMeta: api.TypeMeta{Kind: \"Tenant\"},\n\t\tObjectMeta: api.ObjectMeta{Name: \"default\"},\n\t}\n\tti.apiClient.ClusterV1().Tenant().Create(context.Background(), &defTenant)\n\n\t// add event policy - 1\n\teventPolicy1 := policygen.CreateEventPolicyObj(globals.DefaultTenant, globals.DefaultNamespace, \"ep-1\",\n\t\tmonitoring.MonitoringExportFormat_SYSLOG_BSD.String(),\n\t\t[]*monitoring.ExportConfig{\n\t\t\t{ // receivedMsgsAtUDPServer1\n\t\t\t\tDestination: \"127.0.0.1\",\n\t\t\t\tTransport: fmt.Sprintf(\"UDP/%s\", tmp1[len(tmp1)-1]), // UDP or udp should work\n\t\t\t},\n\t\t\t{ // receivedMsgsAtTCPServer1\n\t\t\t\tDestination: \"127.0.0.1\",\n\t\t\t\tTransport: fmt.Sprintf(\"TCP/%s\", tmp2[len(tmp2)-1]), // TCP or tcp should work\n\t\t\t},\n\t\t}, nil)\n\teventPolicy1, err = ti.apiClient.MonitoringV1().EventPolicy().Create(context.Background(), eventPolicy1)\n\tAssertOk(t, err, \"failed to create event policy, err: %v\", err)\n\tdefer ti.apiClient.MonitoringV1().EventPolicy().Delete(context.Background(), eventPolicy1.GetObjectMeta())\n\n\t// add event policy - 2\n\teventPolicy2 := policygen.CreateEventPolicyObj(globals.DefaultTenant, globals.DefaultNamespace, \"ep-2\",\n\t\tmonitoring.MonitoringExportFormat_SYSLOG_RFC5424.String(),\n\t\t[]*monitoring.ExportConfig{\n\t\t\t{ // receivedMsgsAtTCPServer2\n\t\t\t\tDestination: \"127.0.0.1\",\n\t\t\t\tTransport: fmt.Sprintf(\"tcp/%s\", tmp3[len(tmp3)-1]),\n\t\t\t},\n\t\t},\n\t\t&monitoring.SyslogExportConfig{\n\t\t\tFacilityOverride: monitoring.SyslogFacility_LOG_SYSLOG.String(),\n\t\t\tPrefix: CreateAlphabetString(5),\n\t\t})\n\teventPolicy2, err = ti.apiClient.MonitoringV1().EventPolicy().Create(context.Background(), eventPolicy2)\n\tAssertOk(t, err, \"failed to create event policy, err: %v\", err)\n\tdefer ti.apiClient.MonitoringV1().EventPolicy().Delete(context.Background(), eventPolicy2.GetObjectMeta())\n\n\t// to let the event policies reach the policy manager (api server -> evtsmgr -> policy watcher -> policy manager -> exporter)\n\ttime.Sleep(2 * time.Second)\n\n\t// object reference for events\n\tdummyObjRef := &cluster.Node{\n\t\tTypeMeta: api.TypeMeta{\n\t\t\tKind: \"Node\",\n\t\t},\n\t\tObjectMeta: api.ObjectMeta{\n\t\t\tTenant: globals.DefaultTenant,\n\t\t\tNamespace: globals.DefaultNamespace,\n\t\t},\n\t}\n\n\tmessages := map[chan string][]struct {\n\t\tSubstrs []string // syslog message should contain all these strings\n\t\tMsgFormat monitoring.MonitoringExportFormat // BSD style message contains the JSON formatted alert; RFC contains <msgID, structured data, msg>\n\t}{\n\t\t// all the messages that are sent should be received at the syslog server\n\t\treceivedMsgsAtUDPServer1: {\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s started on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s running on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s unresponsive on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s stopped on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: election started %s\", dummyObjRef.Tenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader elected %s\", dummyObjRef.Tenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader changed %s\", dummyObjRef.Tenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader lost %s\", dummyObjRef.Tenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: election stopped %s\", dummyObjRef.Tenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s started\", globals.DefaultTenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s running\", globals.DefaultTenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s unresponsive\", globals.DefaultTenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s stopped\", globals.DefaultTenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t},\n\t\treceivedMsgsAtTCPServer1: {\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s started on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s running on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s unresponsive on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s stopped on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: election started %s\", dummyObjRef.Tenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader elected %s\", dummyObjRef.Tenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader changed %s\", dummyObjRef.Tenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader lost %s\", dummyObjRef.Tenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: election stopped %s\", dummyObjRef.Tenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s started\", globals.DefaultTenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s running\", globals.DefaultTenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s unresponsive\", globals.DefaultTenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s stopped\", globals.DefaultTenant, t.Name())},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_BSD,\n\t\t\t},\n\t\t},\n\t\treceivedMsgsAtTCPServer2: { // messages belonging to event policy - 2\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s started on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s running on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s unresponsive on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s stopped on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: election started %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader elected %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader changed %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader lost %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: election stopped %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s started\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s running\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s unresponsive\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s stopped\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t},\n\t}\n\n\ttestSyslogMessageDelivery(t, ti, dummyObjRef, messages)\n\n\t// update event policy - 1; remove the existing target and add a new one\n\t// start UDP server to receive syslog messages\n\tpConn2, receivedMsgsAtUDPServer2, err := serviceutils.StartUDPServer(\":0\")\n\tAssertOk(t, err, \"failed to start UDP server, err: %v\", err)\n\tdefer pConn2.Close()\n\ttmp4 := strings.Split(pConn2.LocalAddr().String(), \":\")\n\n\t// add event policy - 1\n\teventPolicy1 = policygen.CreateEventPolicyObj(globals.DefaultTenant, globals.DefaultNamespace, \"ep-1\",\n\t\tmonitoring.MonitoringExportFormat_SYSLOG_RFC5424.String(),\n\t\t[]*monitoring.ExportConfig{\n\t\t\t{ // receivedMsgsAtUDPServer1\n\t\t\t\tDestination: \"127.0.0.1\",\n\t\t\t\tTransport: fmt.Sprintf(\"udp/%s\", tmp1[len(tmp1)-1]),\n\t\t\t},\n\t\t\t{ // receivedMsgsAtUDPServer2\n\t\t\t\tDestination: \"127.0.0.1\",\n\t\t\t\tTransport: fmt.Sprintf(\"udp/%s\", tmp4[len(tmp4)-1]),\n\t\t\t},\n\t\t}, nil)\n\teventPolicy1, err = ti.apiClient.MonitoringV1().EventPolicy().Update(context.Background(), eventPolicy1)\n\tAssertOk(t, err, \"failed to create event policy, err: %v\", err)\n\tdefer ti.apiClient.MonitoringV1().EventPolicy().Delete(context.Background(), eventPolicy1.GetObjectMeta())\n\n\t// `receivedMsgsAtTCPServer1` should receive no more messages as it is removed from the policy '\"ep-1\"' (refer above);\n\t// final should be the last message on this channel\n\twg.Add(1)\n\tcloseMsgCh := make(chan struct{})\n\tgo func() {\n\t\tdefer wg.Done()\n\t\ttime.Sleep(100 * time.Millisecond)\n\t\treceivedMsgsAtTCPServer1 <- \"final\"\n\n\t\tshouldNotReceiveAnymoreMsgs := false\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-closeMsgCh:\n\t\t\t\treturn\n\t\t\tcase msg, ok := <-receivedMsgsAtTCPServer1:\n\t\t\t\tif !ok {\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tif msg == \"final\" {\n\t\t\t\t\tshouldNotReceiveAnymoreMsgs = true\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tif shouldNotReceiveAnymoreMsgs {\n\t\t\t\t\tt.Fatalf(\"syslog target is removed from the policy. so, should not receive any more messages on this channel \"+\n\t\t\t\t\t\t\"but received: %v\", msg)\n\t\t\t\t} else {\n\t\t\t\t\tlog.Infof(\"receiving pending messages from syslog server: %v\", msg)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\t// we should stop seeing messages on the old channel `receivedMsgsAtTCPServer1`\n\t// and start seeing messages on the new channel `receivedMsgsAtUDPServer2`\n\tmessages = map[chan string][]struct {\n\t\tSubstrs []string // syslog message should contain all these strings\n\t\tMsgFormat monitoring.MonitoringExportFormat // BSD style message contains the JSON formatted alert; RFC contains <msgID, structured data, msg>\n\t}{\n\t\treceivedMsgsAtUDPServer1: { // target - 1 of event policy - 1\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s started on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s running on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s unresponsive on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s stopped on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: election started %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader elected %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader changed %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader lost %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: election stopped %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s started\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s running\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s unresponsive\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s stopped\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t},\n\t\treceivedMsgsAtUDPServer2: { // target - 2 of event policy - 1\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s started on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s running on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s unresponsive on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s stopped on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: election started %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader elected %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader changed %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader lost %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: election stopped %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s started\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s running\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s unresponsive\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s stopped\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy1.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t},\n\t\treceivedMsgsAtTCPServer2: { // messages belonging to event policy - 2\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s started on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s running on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s unresponsive on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s stopped on %s\", dummyObjRef.Tenant, t.Name(), dummyObjRef.GetKind()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: election started %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader elected %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader changed %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: leader lost %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) dummy election: election stopped %s\", dummyObjRef.Tenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s started\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s running\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s unresponsive\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t\t{\n\t\t\t\tSubstrs: []string{fmt.Sprintf(\"(tenant:%s) test %s stopped\", globals.DefaultTenant, t.Name()),\n\t\t\t\t\teventPolicy2.Spec.GetSyslogConfig().GetPrefix()},\n\t\t\t\tMsgFormat: monitoring.MonitoringExportFormat_SYSLOG_RFC5424,\n\t\t\t},\n\t\t},\n\t}\n\ttestSyslogMessageDelivery(t, ti, dummyObjRef, messages)\n\n\tclose(closeMsgCh)\n\n\t// to avoid - panic: send on closed channel\n\t// this happens because the test completes before the dispatcher or evtsproxy is done sending all events.\n\t// test has completed prior because we check for occurrence of each message but it is possible that the sender is\n\t// sending in intervals which means there're are events coming in every batch interval. So, the TCP/UDP server gets\n\t// closed while the dispatcher is trying to send events to it. Stopping all the evtsproxy related services before\n\t// shutting down the TCP/UDP serer will solve the problem (there won't be anyone trying to send events anymore).\n\tti.evtProxyServices.Stop()\n}", "func (b *Broker) PublishEvent(etype string, where string, data interface{}, channels []string) {\n\tconn := b.pool.Get()\n\tdefer conn.Close()\n\n\tevent := gp.Event{Type: etype, Location: where, Data: data}\n\t//Wrap the event in another layer to appease\n\tmessage := gp.WrappedEvent{Event: \"message\", Data: event}\n\tJSONEvent, _ := json.Marshal(message)\n\tfor _, channel := range channels {\n\t\tconn.Send(\"PUBLISH\", channel, JSONEvent)\n\t}\n\tconn.Flush()\n}", "func (p *Pool) Exec(task Task) {\n\tp.tasks <- task\n}", "func Task(t *api.Task) string {\n\tif t.Annotations.Name != \"\" {\n\t\t// if set, use the container Annotations.Name field, set in the orchestrator.\n\t\treturn t.Annotations.Name\n\t}\n\n\tslot := fmt.Sprint(t.Slot)\n\tif slot == \"\" || t.Slot == 0 {\n\t\t// when no slot id is assigned, we assume that this is node-bound task.\n\t\tslot = t.NodeID\n\t}\n\n\t// fallback to service.instance.id.\n\treturn fmt.Sprintf(\"%s.%s.%s\", t.ServiceAnnotations.Name, slot, t.ID)\n}", "func GetTaskInformation(res http.ResponseWriter, req *http.Request) {\n\tres.Header().Set(\"Content-Type\", \"application/json\")\n\tc := Task{\"7E12E3B57A02\", \"FOOUSER\", \"BAR\"}\n\toutgoingJSON, err := json.Marshal(c)\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\thttp.Error(res, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tres.WriteHeader(http.StatusCreated)\n\tfmt.Fprint(res, string(outgoingJSON))\n}", "func (_Contract *ContractTransactor) HandleTasks(opts *bind.TransactOpts, startIdx *big.Int, quantity *big.Int) (*types.Transaction, error) {\n\treturn _Contract.contract.Transact(opts, \"handleTasks\", startIdx, quantity)\n}", "func GenericEventGenerator(workload map[string]interface{}) (map[string]interface{}, int) {\n\tduration := workload[\"duration\"]\n\tall_event := make(map[string]interface{})\n\tevent_count := 0\n\tfor instance, value := range workload[\"instances\"].(map[string]interface{}) {\n\t\tlog.Println(\"Generate\", instance)\n\t\tdesc := value.(map[string]interface{})\n\t\tinstance_events := CreateEvents(desc[\"distribution\"].(string), int(desc[\"rate\"].(float64)), int(duration.(float64)))\n\t\tlog.Println(instance, \"is created\")\n\t\tstart_time := 0\n\t\tend_time := int(duration.(float64))\n\t\tif activity_window, ok := desc[\"activity_window\"]; ok {\n\t\t\tif window, ok := activity_window.([]interface{}); ok {\n\t\t\t\tif len(window) >= 2 {\n\t\t\t\t\tstart_time = int(window[0].(float64))\n\t\t\t\t\tend_time = int(window[1].(float64))\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tinstance_events = EnforceActivityWindow(start_time, end_time, instance_events)\n\t\tall_event[instance] = instance_events\n\t\tevent_count += len(instance_events)\n\t}\n\treturn all_event, event_count\n}", "func (s *Search) Tasks(ctx *gin.Context) {\n\tapp := ctx.Param(\"app\")\n\tslot := ctx.Param(\"slot\")\n\tappLabel := config.LogAppLabel()\n\tslotLabel := config.LogSlotLabel()\n\toptions := config.ConvertRequestQueryParams(ctx.Request.URL.Query())\n\toptions[appLabel] = app\n\toptions[slotLabel] = slot\n\ttasks, err := s.Service.Tasks(options, ctx.MustGet(\"page\").(models.Page))\n\tif err != nil {\n\t\tutils.ErrorResponse(ctx, utils.NewError(GetTaskError, err))\n\t\treturn\n\t}\n\tutils.Ok(ctx, tasks)\n}", "func (s *Scheduler) enqueueEvents(events []entities.Event) {\n\n\tfor _, event := range events {\n\t\terr := s.queue.Push(event)\n\t\tif err != nil {\n\t\t\ts.logErrorf(\"Scheduler.enqueueEvents, queue.Push return error %s\", err)\n\t\t} else {\n\t\t\terr = s.storage.MarkEventAsNotified(event.Id(), s.now())\n\t\t\tif err != nil {\n\t\t\t\ts.logErrorf(\"Scheduler.enqueueEvents, storage.MarkEventAsNotified return error %s\", err)\n\t\t\t}\n\t\t}\n\t}\n}", "func (sd *ScanDiapasons) NotifyDpnsTask(ctx context.Context) ([]byte, error) {\n\trequest, err := http.NewRequest(\"POST\", sd.client.Server+\"/api/v1.0/ScanDiapasons.NotifyDpnsTask\", nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\traw, err := sd.client.Do(ctx, request, nil)\n\treturn raw, err\n}", "func (_Contract *ContractFilterer) WatchTasksErased(opts *bind.WatchOpts, sink chan<- *ContractTasksErased) (event.Subscription, error) {\n\n\tlogs, sub, err := _Contract.contract.WatchLogs(opts, \"TasksErased\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn event.NewSubscription(func(quit <-chan struct{}) error {\n\t\tdefer sub.Unsubscribe()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase log := <-logs:\n\t\t\t\t// New log arrived, parse the event and forward to the user\n\t\t\t\tevent := new(ContractTasksErased)\n\t\t\t\tif err := _Contract.contract.UnpackLog(event, \"TasksErased\", log); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tevent.Raw = log\n\n\t\t\t\tselect {\n\t\t\t\tcase sink <- event:\n\t\t\t\tcase err := <-sub.Err():\n\t\t\t\t\treturn err\n\t\t\t\tcase <-quit:\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\tcase err := <-sub.Err():\n\t\t\t\treturn err\n\t\t\tcase <-quit:\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t}), nil\n}", "func (s *agentRegistry) PublishTaskStatistics(ctx context.Context, req *tracking.TaskStatistics) (*empty.Empty, error) {\n\ts.log.Debug().\n\t\tStr(\"task\", req.GetName()).\n\t\tStr(\"uri\", req.GetURI()).\n\t\tMsg(\"Publish task statistics\")\n\n\ts.agentsMutex.Lock()\n\t{\n\t\tcurrent := s.taskAgents[req.GetName()]\n\t\tvar taskAgentRef *taskAgent\n\t\tfor _, x := range current {\n\t\t\tif x.URI == req.GetURI() {\n\t\t\t\ttaskAgentRef = x\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif taskAgentRef == nil {\n\t\t\ttaskAgentRef = &taskAgent{URI: req.GetURI()}\n\t\t\tcurrent = append(current, taskAgentRef)\n\t\t\ts.taskAgents[req.GetName()] = current\n\t\t}\n\t\ttaskAgentRef.Statistics.Data.Reset()\n\t\ttaskAgentRef.Statistics.Data.Add(*req)\n\t\ttaskAgentRef.Statistics.Timestamp = time.Now()\n\t\tsort.Slice(current, func(i, j int) bool { return current[i].Statistics.Timestamp.Before(current[j].Statistics.Timestamp) })\n\t}\n\ts.agentsMutex.Unlock()\n\n\t// Notify frontend clients\n\ts.hub.StatisticsChanged()\n\n\treturn &empty.Empty{}, nil\n}", "func (m *agentEndpointsManager) _processTasks(workerNum int, tasks []*agents.Task, wg *sync.WaitGroup) {\n\tdefer wg.Done()\n\tlogger.Debugf(\"agents tasks monitor: task processing worker #%d processing %d tasks\", workerNum, len(tasks))\n\t// mark all tasks as assigned for a worker for processing and save their statuses\n\t// to determine the type of processing required\n\tvar taskElements []db.IBucketElement\n\ttaskStatuses := make(map[string]int)\n\tfor _, task := range tasks {\n\t\ttaskStatuses[task.ID] = task.Status\n\t\ttask.Status = agents.TaskStatusAssigned\n\t\ttaskElements = append(taskElements, task)\n\t}\n\tif err := db.Update(db.System, taskElements...); err != nil {\n\t\tlogger.WithError(err).Error(\"agents tasks monitor: error updating tasks to assigned status\")\n\t\treturn\n\t}\n\tfor _, task := range tasks {\n\t\tswitch taskStatuses[task.ID] {\n\t\t\tcase agents.TaskStatusReady:\n\t\t\t\tif task.Agent == \"\" {\n\t\t\t\t\tselectedAgentId, err := m.selectAgentForTask(task)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlogger.WithError(err).Errorf(\"agents tasks monitor: failed selecting agent for task with id == %s\", task.ID)\n\t\t\t\t\t\tm.updateTaskWithDescriptionToErr(task, err.Error())\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\ttask.Agent = selectedAgentId\n\t\t\t\t}\n\t\t\t\tselectedAgentEndpoint := m.getEndpoint(task.Agent)\n\t\t\t\tif selectedAgentEndpoint == nil {\n\t\t\t\t\tlogger.Errorf(\"agents tasks monitor: agent with id == %s was selected to run task with id == %s but no endpoint available for him\", task.Agent, task.ID)\n\t\t\t\t\tm.updateTaskWithDescriptionToErr(task, \"agent unavailable\")\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tmsg, err := task.GetWsMessage()\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogger.WithError(err).Errorf(\"agents tasks monitor: error creating message from task with id == %s\", task.ID)\n\t\t\t\t\tm.updateTaskWithDescriptionToErr(task, err.Error())\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\ttask.Status = agents.TaskStatusInProgress\n\t\t\t\tif err := db.Update(db.System, task); err != nil {\n\t\t\t\t\tlogger.WithError(err).Errorf(\"agents tasks monitor: error updating task with id == %s to in progress state\", task.ID)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tselectedAgentEndpoint.write(msg)\n\t\t\tcase agents.TaskStatusDone:\n\t\t\t\tm.processTaskWithResponse(task)\n\t\t\tdefault: // should not happen...\n\t\t\t\tlogger.Errorf(\"agents tasks monitor: task with id == %s was selected for processing it has status = %d (not in progress or ready)\", task.ID, task.Status)\n\t\t}\n\t}\n}", "func (p *Pool) Exec(e Task) {\n\tp.tasks <- e\n}", "func (s *Service) Events(ctx context.Context, evts chan events.ContainerEvent) error {\n\tfilter := filters.NewArgs()\n\tfilter.Add(\"label\", fmt.Sprintf(\"%s=%s\", labels.PROJECT, s.project.Name))\n\tfilter.Add(\"label\", fmt.Sprintf(\"%s=%s\", labels.SERVICE, s.name))\n\tclient := s.clientFactory.Create(s)\n\teventq, errq := client.Events(ctx, types.EventsOptions{\n\t\tFilters: filter,\n\t})\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase event := <-eventq:\n\t\t\t\tservice := event.Actor.Attributes[labels.SERVICE.Str()]\n\t\t\t\tattributes := map[string]string{}\n\t\t\t\tfor _, attr := range eventAttributes {\n\t\t\t\t\tattributes[attr] = event.Actor.Attributes[attr]\n\t\t\t\t}\n\t\t\t\te := events.ContainerEvent{\n\t\t\t\t\tService: service,\n\t\t\t\t\tEvent: event.Action,\n\t\t\t\t\tType: event.Type,\n\t\t\t\t\tID: event.Actor.ID,\n\t\t\t\t\tTime: time.Unix(event.Time, 0),\n\t\t\t\t\tAttributes: attributes,\n\t\t\t\t}\n\t\t\t\tevts <- e\n\t\t\t}\n\t\t}\n\t}()\n\treturn <-errq\n}", "func (server *Server) RegisterTasks(tasks map[string]interface{}) {\n\tserver.registeredTasks = tasks\n}", "func (ec *executionContext) _Task(ctx context.Context, sel ast.SelectionSet, obj *Task) graphql.Marshaler {\n\tfields := graphql.CollectFields(ctx, sel, taskImplementors)\n\n\tout := graphql.NewOrderedMap(len(fields))\n\tinvalid := false\n\tfor i, field := range fields {\n\t\tout.Keys[i] = field.Alias\n\n\t\tswitch field.Name {\n\t\tcase \"__typename\":\n\t\t\tout.Values[i] = graphql.MarshalString(\"Task\")\n\t\tcase \"parent\":\n\t\t\tout.Values[i] = ec._Task_parent(ctx, field, obj)\n\t\tcase \"subTasks\":\n\t\t\tout.Values[i] = ec._Task_subTasks(ctx, field, obj)\n\t\tcase \"connectionWith\":\n\t\t\tout.Values[i] = ec._Task_connectionWith(ctx, field, obj)\n\t\tcase \"id\":\n\t\t\tout.Values[i] = ec._Task_id(ctx, field, obj)\n\t\tcase \"section\":\n\t\t\tout.Values[i] = ec._Task_section(ctx, field, obj)\n\t\t\tif out.Values[i] == graphql.Null {\n\t\t\t\tinvalid = true\n\t\t\t}\n\t\tcase \"name\":\n\t\t\tout.Values[i] = ec._Task_name(ctx, field, obj)\n\t\t\tif out.Values[i] == graphql.Null {\n\t\t\t\tinvalid = true\n\t\t\t}\n\t\tcase \"description\":\n\t\t\tout.Values[i] = ec._Task_description(ctx, field, obj)\n\t\t\tif out.Values[i] == graphql.Null {\n\t\t\t\tinvalid = true\n\t\t\t}\n\t\tcase \"estimation\":\n\t\t\tout.Values[i] = ec._Task_estimation(ctx, field, obj)\n\t\tcase \"isActive\":\n\t\t\tout.Values[i] = ec._Task_isActive(ctx, field, obj)\n\t\tcase \"timestamp\":\n\t\t\tout.Values[i] = ec._Task_timestamp(ctx, field, obj)\n\t\tdefault:\n\t\t\tpanic(\"unknown field \" + strconv.Quote(field.Name))\n\t\t}\n\t}\n\n\tif invalid {\n\t\treturn graphql.Null\n\t}\n\treturn out\n}", "func (c *client) PublishEvents(\n\tdata []outputs.Data,\n) ([]outputs.Data, error) {\n\tpublishEventsCallCount.Add(1)\n\ttotalNumberOfEvents := len(data)\n\n\tif len(data) == 0 {\n\t\treturn nil, nil\n\t}\n\n\tfor len(data) > 0 {\n\t\tvar (\n\t\t\tn int\n\t\t\terr error\n\t\t)\n\t\tif c.win == nil {\n\t\t\tn, err = c.sendEvents(data)\n\t\t} else {\n\t\t\tn, err = c.publishWindowed(data)\n\t\t}\n\n\t\tdebug(\"%v events out of %v events sent to logstash host %v. Continue sending\",\n\t\t\tn, len(data), c.host)\n\n\t\tdata = data[n:]\n\t\tif err != nil {\n\t\t\tif c.win != nil {\n\t\t\t\tc.win.shrinkWindow()\n\t\t\t}\n\t\t\t_ = c.Close()\n\n\t\t\tlogp.Err(\"Failed to publish events (host: %v), caused by: %v\", c.host, err)\n\n\t\t\teventsNotAcked.Add(int64(len(data)))\n\t\t\tackedEvents.Add(int64(totalNumberOfEvents - len(data)))\n\t\t\treturn data, err\n\t\t}\n\t}\n\tackedEvents.Add(int64(totalNumberOfEvents))\n\treturn nil, nil\n}", "func (handler *Handlers)GetTasks(w http.ResponseWriter,req *http.Request) {\n log.Println(\"getting task list of current user from startDate to endDate\")\n w.Header().Set(\"Content-Type\", \"application/json\")\n username := token.GetUserName(w, req)\n\n filter := newFilter() //initialize dateFilter\n err := json.NewDecoder(req.Body).Decode(&filter) //parse startDate and endDate from response body\n if err != nil {\n fmt.Fprintln(w, err.Error())\n return\n }\n\n startDate, endDate, dateError := validation.ValidateDate(filter.StartDate, filter.EndDate) //validate Date\n if dateError != nil {\n fmt.Fprintln(w, dateError)\n return\n }\n\n //get all tasks from database of current user from startDate to endDate\n taskList, err := handler.Repository.GetTasks(username, startDate, endDate)\n if err != nil {\n fmt.Fprintln(w, err)\n return\n }\n if taskList == nil {\n fmt.Fprintln(w, \"No tasks assigned\")\n return\n }\n err = json.NewEncoder(w).Encode(taskList) //display task\n if err == nil {\n log.Println(\"task displayed\")\n }\n}", "func SerializeTask(meta EgorMeta) (string, error) {\n\tvar buffer bytes.Buffer\n\tencoder := json2.NewEncoder(&buffer)\n\tif err := encoder.Encode(meta); err != nil {\n\t\treturn \"\", err\n\t}\n\treturn buffer.String(), nil\n}", "func RunTasks(context *app.Context) {\n\t// Task name && sleep interval seconds\n\t// one day second = 24 * 60 * 60 = 86400\n\t// 8 * 60 * 60 = 28800\n\n\tTasks = append(Tasks, task.NewRTmart(context, 28800))\n\tTasks = append(Tasks, task.NewCarrefour(context, 28800))\n\n\tfor _, task := range Tasks {\n\t\tlog.Println(\"Running\", task)\n\t\tgo task.Run()\n\t}\n}", "func (s *SearchService) Tasks(opts map[string]interface{}, page models.Page) (map[string]int64, error) {\n\tvar err error\n\terr = s.resetESClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar querys []elastic.Query\n\tfor k, v := range opts {\n\t\tquerys = append(querys, elastic.NewTermQuery(k, v))\n\t}\n\tbquery := elastic.NewBoolQuery().\n\t\tFilter(elastic.NewRangeQuery(\"logtime.timestamp\").Gte(page.RangeFrom).Lte(page.RangeTo).Format(\"epoch_millis\")).\n\t\tMust(querys...)\n\n\t//Index(\"dataman-*\").\n\ttasks := make(map[string]int64)\n\ttaskLabel := config.LogTaskLabel()\n\tresult, err := s.ESClient.Search().\n\t\tIndex(\"dataman-*\").\n\t\tSearchType(\"count\").\n\t\tQuery(bquery).\n\t\tAggregation(\"tasks\", elastic.NewTermsAggregation().Field(taskLabel).Size(0).OrderByCountDesc()).\n\t\tPretty(true).\n\t\tDo()\n\n\tif err != nil && err.(*elastic.Error).Status == http.StatusNotFound {\n\t\treturn nil, nil\n\t}\n\n\tif err != nil {\n\t\treturn tasks, err\n\t}\n\n\tagg, found := result.Aggregations.Terms(\"tasks\")\n\tif !found {\n\t\treturn tasks, nil\n\t}\n\n\tfor _, bucket := range agg.Buckets {\n\t\ttasks[fmt.Sprint(bucket.Key)] = bucket.DocCount\n\t}\n\treturn tasks, nil\n}", "func (out *elasticsearchOutput) PublishEvent(ts time.Time, event common.MapStr) error {\n\n\tout.sendingQueue <- EventMsg{Ts: ts, Event: event}\n\n\tlogp.Debug(\"output_elasticsearch\", \"Publish event: %s\", event)\n\treturn nil\n}", "func SendEvents(\n\tctx context.Context, rsAPI RoomserverInternalAPI, events []gomatrixserverlib.HeaderedEvent,\n\tsendAsServer gomatrixserverlib.ServerName, txnID *TransactionID,\n) (string, error) {\n\tires := make([]InputRoomEvent, len(events))\n\tfor i, event := range events {\n\t\tires[i] = InputRoomEvent{\n\t\t\tKind: KindNew,\n\t\t\tEvent: event,\n\t\t\tAuthEventIDs: event.AuthEventIDs(),\n\t\t\tSendAsServer: string(sendAsServer),\n\t\t\tTransactionID: txnID,\n\t\t}\n\t}\n\treturn SendInputRoomEvents(ctx, rsAPI, ires)\n}", "func (app *appImpl) PollEvents() {\n}", "func (jm *JobManager) TaskBound(task *api.TaskInfo) {\n\tif taskName := getTaskName(task); taskName != \"\" {\n\t\tset, ok := jm.nodeTaskSet[task.NodeName]\n\t\tif !ok {\n\t\t\tset = make(map[string]int)\n\t\t\tjm.nodeTaskSet[task.NodeName] = set\n\t\t}\n\t\tset[taskName]++\n\t}\n\n\tbucket := jm.GetBucket(task)\n\tif bucket != nil {\n\t\tbucket.TaskBound(task)\n\t}\n}", "func AddTask(c echo.Context) error {\n\tvar body Task\n\n\terr := c.Bind(&body)\n\tif err != nil {\n\t\treturn c.JSON(http.StatusBadRequest, ResponseError{Status: http.StatusBadRequest, Message: err.Error()})\n\t}\n\n\tif body.ID == nil {\n\t\treturn c.JSON(http.StatusBadRequest, ResponseError{Status: http.StatusBadRequest, Message: \"id empty\"})\n\t}\n\n\ttask := Task{\n\t\tID: body.ID,\n\t\tTopic: body.Topic,\n\t\tCreated: body.Created,\n\t\tFinishDate: body.FinishDate,\n\t\tFile: body.File,\n\t}\n\n\ttasks = append(tasks, task)\n\n\treturn c.JSON(http.StatusCreated, task)\n}", "func GetAllTask(c *gin.Context) {\n\tpayload := getAllTask()\n\tc.JSON(http.StatusOK, payload)\n}", "func (_Contract *ContractSession) HandleTasks(startIdx *big.Int, quantity *big.Int) (*types.Transaction, error) {\n\treturn _Contract.Contract.HandleTasks(&_Contract.TransactOpts, startIdx, quantity)\n}", "func (c *IAM) addTask(task iamTask) error {\n\tselect {\n\tcase c.tasks <- task:\n\t\treturn nil\n\n\tdefault:\n\t\treturn trace.LimitExceeded(\"failed to create IAM task for %v\", task.database.GetName())\n\t}\n}", "func CountProcessEvents(taskId string) (int, error) {\n\tfilter := bson.M{\n\t\tDataKey + \".\" + ResourceTypeKey: EventTaskProcessInfo,\n\t\tTypeKey: EventTaskProcessInfo,\n\t}\n\n\tif taskId != \"\" {\n\t\tfilter[ResourceIdKey] = taskId\n\t}\n\n\treturn db.CountQ(TaskLogCollection, db.Query(filter))\n}", "func (_Contract *ContractTransactorSession) PublishEvent(topicId string, cid string, newAccounts []common.Address) (*types.Transaction, error) {\n\treturn _Contract.Contract.PublishEvent(&_Contract.TransactOpts, topicId, cid, newAccounts)\n}", "func (t *TimeTask) AddTask(task *Task) {\n\tt.addChan <- task\n}", "func (d *Docker) Events(ctx context.Context, filters map[string]string) (<-chan *types.WorkloadEventMessage, <-chan error) {\n\teventChan := make(chan *types.WorkloadEventMessage)\n\terrChan := make(chan error)\n\n\t_ = utils.Pool.Submit(func() {\n\t\tdefer close(eventChan)\n\t\tdefer close(errChan)\n\n\t\tf := d.getFilterArgs(filters)\n\t\tf.Add(\"type\", events.ContainerEventType)\n\t\toptions := enginetypes.EventsOptions{Filters: f}\n\t\tm, e := d.client.Events(ctx, options)\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase message := <-m:\n\t\t\t\teventChan <- &types.WorkloadEventMessage{\n\t\t\t\t\tID: message.ID,\n\t\t\t\t\tType: message.Type,\n\t\t\t\t\tAction: message.Action,\n\t\t\t\t\tTimeNano: message.TimeNano,\n\t\t\t\t}\n\t\t\tcase err := <-e:\n\t\t\t\terrChan <- err\n\t\t\t\treturn\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t})\n\n\treturn eventChan, errChan\n}", "func (db *DynamoDB) ListTasks(ctx context.Context, req *tes.ListTasksRequest) (*tes.ListTasksResponse, error) {\n\n\tvar tasks []*tes.Task\n\tvar query *dynamodb.QueryInput\n\tpageSize := int64(tes.GetPageSize(req.GetPageSize()))\n\n\tquery = &dynamodb.QueryInput{\n\t\tTableName: aws.String(db.taskTable),\n\t\tLimit: aws.Int64(pageSize),\n\t\tScanIndexForward: aws.Bool(false),\n\t\tConsistentRead: aws.Bool(true),\n\t\tKeyConditionExpression: aws.String(fmt.Sprintf(\"%s = :v1\", db.partitionKey)),\n\t\tExpressionAttributeValues: map[string]*dynamodb.AttributeValue{\n\t\t\t\":v1\": {\n\t\t\t\tS: aws.String(db.partitionValue),\n\t\t\t},\n\t\t},\n\t}\n\n\tfilterParts := []string{}\n\tif req.State != tes.Unknown {\n\t\tquery.ExpressionAttributeNames = map[string]*string{\n\t\t\t\"#state\": aws.String(\"state\"),\n\t\t}\n\t\tquery.ExpressionAttributeValues[\":stateFilter\"] = &dynamodb.AttributeValue{\n\t\t\tN: aws.String(strconv.Itoa(int(req.State))),\n\t\t}\n\t\tfilterParts = append(filterParts, \"#state = :stateFilter\")\n\t}\n\n\tfor k, v := range req.Tags {\n\t\ttmpl := \"tags.%s = :%sFilter\"\n\t\tfilterParts = append(filterParts, fmt.Sprintf(tmpl, k, k))\n\t\tif v == \"\" {\n\t\t\tquery.ExpressionAttributeValues[fmt.Sprintf(\":%sFilter\", k)] = &dynamodb.AttributeValue{\n\t\t\t\tNULL: aws.Bool(true),\n\t\t\t}\n\t\t} else {\n\t\t\tquery.ExpressionAttributeValues[fmt.Sprintf(\":%sFilter\", k)] = &dynamodb.AttributeValue{\n\t\t\t\tS: aws.String(v),\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(filterParts) > 0 {\n\t\tquery.FilterExpression = aws.String(strings.Join(filterParts, \" AND \"))\n\t}\n\n\tif req.View == tes.TaskView_MINIMAL {\n\t\tquery.ExpressionAttributeNames = map[string]*string{\n\t\t\t\"#state\": aws.String(\"state\"),\n\t\t}\n\t\tquery.ProjectionExpression = aws.String(\"id, #state\")\n\t}\n\n\tif req.PageToken != \"\" {\n\t\tquery.ExclusiveStartKey = map[string]*dynamodb.AttributeValue{\n\t\t\tdb.partitionKey: {\n\t\t\t\tS: aws.String(db.partitionValue),\n\t\t\t},\n\t\t\t\"id\": {\n\t\t\t\tS: aws.String(req.PageToken),\n\t\t\t},\n\t\t}\n\t}\n\n\tresponse, err := db.client.QueryWithContext(ctx, query)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif req.View == tes.TaskView_FULL {\n\t\tfor _, item := range response.Items {\n\t\t\t// TODO handle errors\n\t\t\t_ = db.getContent(ctx, item)\n\t\t\t_ = db.getExecutorOutput(ctx, item, \"stdout\", db.stdoutTable)\n\t\t\t_ = db.getExecutorOutput(ctx, item, \"stderr\", db.stderrTable)\n\t\t\t_ = db.getSystemLogs(ctx, item)\n\t\t}\n\t}\n\n\terr = dynamodbattribute.UnmarshalListOfMaps(response.Items, &tasks)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to DynamoDB unmarshal Tasks, %v\", err)\n\t}\n\n\tout := tes.ListTasksResponse{\n\t\tTasks: tasks,\n\t}\n\n\tif response.LastEvaluatedKey != nil {\n\t\tout.NextPageToken = *response.LastEvaluatedKey[\"id\"].S\n\t}\n\n\treturn &out, nil\n}", "func List(server, taskListView, taskState, taskName string, writer io.Writer) error {\n\tcli := client.NewClient(server)\n\n\tview, ok := tes.TaskView_value[taskListView]\n\tif !ok {\n\t\treturn fmt.Errorf(\"Unknown task view: %s\", taskListView)\n\t}\n\n\tvar page string\n\tvar taskArr []interface{}\n\tfor {\n\t\tresp, err := cli.ListTasks(&tes.ListTasksRequest{\n\t\t\tView: tes.TaskView(view),\n\t\t\tPageToken: page,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(resp.Tasks) == 0 {\n\t\t\tbreak\n\t\t}\n\t\tpage = resp.NextPageToken\n\t\t// convert resp to map[string]interface{} for query\n\t\tvar out map[string]interface{}\n\t\tj, _ := cli.Marshaler.MarshalToString(resp)\n\t\t_ = json.Unmarshal([]byte(j), &out)\n\t\ttaskArr = append(taskArr, out[\"tasks\"].([]interface{})...)\n\t\tif page == \"\" {\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// query tasks\n\tvar tasks interface{}\n\ttasks = taskArr\n\tparser := jsonql.NewQuery(tasks)\n\tvar queries []string\n\n\tif taskState != \"\" {\n\t\tqueries = append(queries, fmt.Sprintf(\"state~='%s'\", taskState))\n\t}\n\tif taskName != \"\" {\n\t\tqueries = append(queries, fmt.Sprintf(\"name~='%s'\", taskName))\n\t}\n\tif len(queries) > 0 {\n\t\tvar err error\n\t\ttasks, err = parser.Query(strings.Join(queries, \" && \"))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttasksJSON, err := json.Marshal(tasks)\n\tif err != nil {\n\t\treturn err\n\t}\n\tresponse := string(tasksJSON)\n\tif response == \"null\" {\n\t\tresponse = \"{}\"\n\t}\n\n\tfmt.Fprintf(writer, \"%s\\n\", response)\n\treturn nil\n}", "func (s *Server) handleTask(ev *Event) (interface{}, error) {\n\tdefer func() {\n\t\tif s.sentry != nil {\n\t\t\tvar packet *raven.Packet\n\t\t\tswitch rval := recover().(type) {\n\t\t\tcase nil:\n\t\t\t\treturn\n\t\t\tcase error:\n\t\t\t\tpacket = raven.NewPacket(rval.Error(), raven.NewException(rval, raven.NewStacktrace(0, 3, nil)))\n\t\t\tdefault:\n\t\t\t\trvalStr := fmt.Sprint(rval)\n\t\t\t\tpacket = raven.NewPacket(rvalStr, raven.NewException(errors.New(rvalStr), raven.NewStacktrace(0, 3, nil)))\n\t\t\t}\n\t\t\t_, ch := s.sentry.Capture(packet, nil)\n\t\t\tif errSentry := <-ch; errSentry != nil {\n\t\t\t\ts.logger.Error(errSentry)\n\t\t\t}\n\t\t} else if recovered := recover(); recovered != nil {\n\t\t\ts.logger.Error(recovered)\n\t\t}\n\t}()\n\ts.wg.Add(1)\n\tdefer s.wg.Done()\n\tif handler, found := s.handlers[ev.Name]; found {\n\t\treturn (*handler)(ev.Args)\n\t}\n\n\treturn nil, ErrNoTaskHandler\n}", "func (s *Server) PublishWithEvents(ctx context.Context, msg interface{}, events map[string][]string) error {\n\tselect {\n\tcase s.cmds <- cmd{op: pub, msg: msg, events: events}:\n\t\treturn nil\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\tcase <-s.Quit():\n\t\treturn nil\n\t}\n}", "func (s *Scheduler) scheduleNTasksOnNodes(ctx context.Context, n int, taskGroup map[string]*api.Task, nodes []NodeInfo, schedulingDecisions map[string]schedulingDecision, nodeLess func(a *NodeInfo, b *NodeInfo) bool) int {\n\ttasksScheduled := 0\n\tfailedConstraints := make(map[int]bool) // key is index in nodes slice\n\tnodeIter := 0\n\tnodeCount := len(nodes)\n\tfor taskID, t := range taskGroup {\n\t\t// Skip tasks which were already scheduled because they ended\n\t\t// up in two groups at once.\n\t\tif _, exists := schedulingDecisions[taskID]; exists {\n\t\t\tcontinue\n\t\t}\n\n\t\tnode := &nodes[nodeIter%nodeCount]\n\t\t// before doing all of the updating logic, get the volume attachments\n\t\t// for the task on this node. this should always succeed, because we\n\t\t// should already have filtered nodes based on volume availability, but\n\t\t// just in case we missed something and it doesn't, we have an error\n\t\t// case.\n\t\tattachments, err := s.volumes.chooseTaskVolumes(t, node)\n\t\tif err != nil {\n\t\t\t// TODO(dperny) if there's an error, then what? i'm frankly not\n\t\t\t// sure.\n\t\t\tlog.G(ctx).WithField(\"task.id\", t.ID).WithError(err).Error(\"could not find task volumes\")\n\t\t}\n\n\t\tlog.G(ctx).WithField(\"task.id\", t.ID).Debugf(\"assigning to node %s\", node.ID)\n\t\t// she turned me into a newT!\n\t\tnewT := *t\n\t\tnewT.Volumes = attachments\n\t\tnewT.NodeID = node.ID\n\t\ts.volumes.reserveTaskVolumes(&newT)\n\t\tnewT.Status = api.TaskStatus{\n\t\t\tState: api.TaskStateAssigned,\n\t\t\tTimestamp: ptypes.MustTimestampProto(time.Now()),\n\t\t\tMessage: \"scheduler assigned task to node\",\n\t\t}\n\t\ts.allTasks[t.ID] = &newT\n\n\t\t// in each iteration of this loop, the node we choose will always be\n\t\t// one which meets constraints. at the end of each iteration, we\n\t\t// re-process nodes, allowing us to remove nodes which no longer meet\n\t\t// resource constraints.\n\t\tnodeInfo, err := s.nodeSet.nodeInfo(node.ID)\n\t\tif err == nil && nodeInfo.addTask(&newT) {\n\t\t\ts.nodeSet.updateNode(nodeInfo)\n\t\t\tnodes[nodeIter%nodeCount] = nodeInfo\n\t\t}\n\n\t\tschedulingDecisions[taskID] = schedulingDecision{old: t, new: &newT}\n\t\tdelete(taskGroup, taskID)\n\t\ttasksScheduled++\n\t\tif tasksScheduled == n {\n\t\t\treturn tasksScheduled\n\t\t}\n\n\t\tif nodeIter+1 < nodeCount {\n\t\t\t// First pass fills the nodes until they have the same\n\t\t\t// number of tasks from this service.\n\t\t\tnextNode := nodes[(nodeIter+1)%nodeCount]\n\t\t\tif nodeLess(&nextNode, &nodeInfo) {\n\t\t\t\tnodeIter++\n\t\t\t}\n\t\t} else {\n\t\t\t// In later passes, we just assign one task at a time\n\t\t\t// to each node that still meets the constraints.\n\t\t\tnodeIter++\n\t\t}\n\n\t\torigNodeIter := nodeIter\n\t\tfor failedConstraints[nodeIter%nodeCount] || !s.pipeline.Process(&nodes[nodeIter%nodeCount]) {\n\t\t\tfailedConstraints[nodeIter%nodeCount] = true\n\t\t\tnodeIter++\n\t\t\tif nodeIter-origNodeIter == nodeCount {\n\t\t\t\t// None of the nodes meet the constraints anymore.\n\t\t\t\treturn tasksScheduled\n\t\t\t}\n\t\t}\n\t}\n\n\treturn tasksScheduled\n}", "func getTasks(w http.ResponseWriter, r *http.Request) {\n\t//Con esto explicamos al servidor que tipo de Dato enviamos\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tjson.NewEncoder(w).Encode(tasks)\n}", "func (_Contract *ContractSession) PublishEvent(topicId string, cid string, newAccounts []common.Address) (*types.Transaction, error) {\n\treturn _Contract.Contract.PublishEvent(&_Contract.TransactOpts, topicId, cid, newAccounts)\n}", "func (m *Manager) AddTask(task func()) {\n\tm.workerPool.queuedTaskC <- task\n}", "func TaskAdd(w http.ResponseWriter, r *http.Request) {\n\n\tdecoder := json.NewDecoder(r.Body)\n\n\tvar taskData Task\n\terr := decoder.Decode(&taskData)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tdefer r.Body.Close()\n\n\terr = collection.Insert(taskData)\n\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\tjson.NewEncoder(w).Encode(taskData)\n}", "func (d *Driver) ExecTask(taskID string, cmd []string, timeout time.Duration) (*drivers.ExecTaskResult, error) {\n\treturn nil, fmt.Errorf(\"Podman driver does not support exec\")\n}", "func processEventsForPartition(partitionClient *azeventhubs.ProcessorPartitionClient) error {\n\t// 1. [BEGIN] Initialize any partition specific resources for your application.\n\t// 2. [CONTINUOUS] Loop, calling ReceiveEvents() and UpdateCheckpoint().\n\t// 3. [END] Cleanup any resources.\n\n\tdefer func() {\n\t\t// 3/3 [END] Do cleanup here, like shutting down database clients\n\t\t// or other resources used for processing this partition.\n\t\tshutdownPartitionResources(partitionClient)\n\t}()\n\n\t// 1/3 [BEGIN] Initialize any partition specific resources for your application.\n\tif err := initializePartitionResources(partitionClient.PartitionID()); err != nil {\n\t\treturn err\n\t}\n\n\t// 2/3 [CONTINUOUS] Receive events, checkpointing as needed using UpdateCheckpoint.\n\tfor {\n\t\t// Wait up to a minute for 100 events, otherwise returns whatever we collected during that time.\n\t\treceiveCtx, cancelReceive := context.WithTimeout(context.TODO(), time.Minute)\n\t\tevents, err := partitionClient.ReceiveEvents(receiveCtx, 100, nil)\n\t\tcancelReceive()\n\n\t\tif err != nil && !errors.Is(err, context.DeadlineExceeded) {\n\t\t\tvar eventHubError *azeventhubs.Error\n\n\t\t\tif errors.As(err, &eventHubError) && eventHubError.Code == azeventhubs.ErrorCodeOwnershipLost {\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\treturn err\n\t\t}\n\n\t\tif len(events) == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\tfmt.Printf(\"Received %d event(s)\\n\", len(events))\n\n\t\tfor _, event := range events {\n\t\t\tfmt.Printf(\"Event received with body %v\\n\", event.Body)\n\t\t}\n\n\t\t// Updates the checkpoint with the latest event received. If processing needs to restart\n\t\t// it will restart from this point, automatically.\n\t\tif err := partitionClient.UpdateCheckpoint(context.TODO(), events[len(events)-1], nil); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n}", "func getTasks(w http.ResponseWriter, r *http.Request) { //esto sirve para mostar todos los datos\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tjson.NewEncoder(w).Encode(tasks)\n}", "func CreateTask(c *gin.Context) {\n\tsrv := server.GetServer()\n\n\tvar param TaskParams\n\tc.BindJSON(&param)\n\n\targs := make([]tasks.Arg, len(param.Args))\n\tfor idx, arg := range param.Args {\n\t\targs[idx] = tasks.Arg{\n\t\t\tType: \"int64\",\n\t\t\tValue: arg,\n\t\t}\n\t}\n\tsignature := &tasks.Signature{\n\t\tName: param.TaskName,\n\t\tArgs: args,\n\t\tRetryCount: 3,\n\t}\n\n\tasyncResult, err := srv.SendTask(signature)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tlog.Println(\"asyncResult:\", asyncResult)\n\n\tc.JSON(http.StatusOK, gin.H{\"Status\": \"In progress\", \"Job\": asyncResult})\n}", "func (s *Session) SendTask(t util.Task) error {\n\t// Checking taskID. re-enqueued task will be skipped\n\tif t.TaskID == \"\" {\n\t\tt.TaskID = uuid.New().String()\n\t\tif t.OriginalTaskID == \"\" {\n\t\t\tt.OriginalTaskID = t.TaskID\n\t\t}\n\n\t\tif err := s.taskRepo.CreateTask(&t); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Checking AMQP connection. Task will be logged for no connection. Re-enqueued later.\n\ts.mu.RLock()\n\tif !s.connected {\n\t\ts.lgr.Warn(\"No connection. Task enqueued.\", util.Object{Key: \"TaskID\", Val: t.TaskID})\n\t\treturn ErrNotConnected\n\t}\n\ts.mu.RUnlock()\n\n\tif !t.Priority.Valid() {\n\t\treturn ErrInvalidPriority\n\t}\n\n\tch, err := s.conn.Channel()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer ch.Close()\n\n\tif err := ch.Confirm(false); err != nil {\n\t\treturn err\n\t}\n\n\tcloseNotification := ch.NotifyClose(make(chan *amqp.Error, 1))\n\tpublish := ch.NotifyPublish(make(chan amqp.Confirmation, 1))\n\tpublishErr := make(chan error, 1)\n\n\tQueue, err := s.GetQueueName(t.Priority)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tgo func() {\n\t\terr := ch.Publish(\n\t\t\ts.cfg.AMQP.Exchange,\n\t\t\tQueue,\n\t\t\tfalse,\n\t\t\tfalse,\n\t\t\tamqp.Publishing{\n\t\t\t\tHeaders: map[string]interface{}{\n\t\t\t\t\t\"TaskName\": t.Name,\n\t\t\t\t\t\"TaskID\": t.TaskID,\n\t\t\t\t},\n\t\t\t\tMessageId: t.TaskID,\n\t\t\t\tDeliveryMode: amqp.Persistent,\n\t\t\t\tBody: t.Payload,\n\t\t\t},\n\t\t)\n\n\t\tif err != nil {\n\t\t\tpublishErr <- err\n\t\t\treturn\n\t\t}\n\n\t\ts.taskRepo.UpdateTaskStatus(context.Background(), t.TaskID, util.StatusQueued)\n\t}()\n\n\tdone := (<-chan time.Time)(make(chan time.Time, 1))\n\tif s.cfg.RequestTimeout != 0 {\n\t\tdone = time.After(s.cfg.RequestTimeout)\n\t}\n\n\tvar errs error\n\n\tselect {\n\tcase errs = <-closeNotification:\n\n\tcase errs = <-publishErr:\n\n\tcase p := <-publish:\n\t\tif !p.Ack {\n\t\t\ts.lgr.Warn(\"Task deliver failed\", util.Object{Key: \"TaskID\", Val: t.TaskID})\n\t\t\terrs = ErrNotPublished\n\t\t\tbreak\n\t\t}\n\t\ts.lgr.Info(\"Task delivered\", util.Object{Key: \"TaskID\", Val: t.TaskID})\n\tcase <-done:\n\t\terrs = ErrRequestTimeout\n\t}\n\n\t// For any kind of error, task will be retried if retry count non zero.\n\t// TODO: retry count only reduce for task processing related error.\n\tif errs != nil {\n\t\tif orgTask, err := s.taskRepo.GetTask(t.OriginalTaskID); err != nil {\n\t\t\ts.lgr.Error(\"failed to get task\", err, util.Object{Key: \"TaskID\", Val: t.OriginalTaskID})\n\t\t} else if orgTask.Retry != 0 {\n\t\t\tgo s.RetryTask(t)\n\t\t}\n\n\t\ts.taskRepo.UpdateTaskStatus(context.Background(), t.TaskID, util.StatusFailed, errs)\n\t}\n\n\treturn errs\n}", "func AddTaskService(e backend.Executor, ts influxdb.TaskService) {\n\tqe, ok := e.(*queryServiceExecutor)\n\tif ok {\n\t\tqe.ts = ts\n\t}\n\tae, ok := e.(*asyncQueryServiceExecutor)\n\tif ok {\n\t\tae.ts = ts\n\t}\n}", "func (s *RefreshImpl) Task(ctx context.Refresh) *taskDTO.PostTaskRequest {\n\tdto := taskDTO.PostTaskRequest{}\n\tdto.MessageID = constvalue.RefreshTaskID\n\tdto.Name = \"Refresh Enclosure\"\n\tdto.Description = \"Refresh enclosure's settings and component.\"\n\tdto.CreatedByName = \"Enclosure Service\"\n\tdto.CreatedByURI = \"/promise/v1/enclosure\"\n\tdto.TargetName = ctx.GetEnclosure().Name\n\tdto.TargetURI = base.ToEnclosureURI(ctx.GetEnclosure().ID)\n\tfor _, v := range s.sub {\n\t\tstep := taskDTO.PostTaskStepRequest{}\n\t\tstep.MessageID = v.MessageID()\n\t\tstep.Name = v.Name()\n\t\tstep.Description = v.Description()\n\t\tstep.ExpectedExecutionMs = v.ExpectedExecutionMs()\n\t\tdto.TaskSteps = append(dto.TaskSteps, step)\n\t}\n\treturn &dto\n}", "func (pool *Pool) AddTask(task Runnable) {\n\tpool.waitGroup.Add(1)\n\tpool.taskChannel <- task\n}", "func (_Contract *ContractTransactorSession) HandleTasks(startIdx *big.Int, quantity *big.Int) (*types.Transaction, error) {\n\treturn _Contract.Contract.HandleTasks(&_Contract.TransactOpts, startIdx, quantity)\n}", "func AddTask(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\tvar DataTask Todo\n\n\terr := json.NewDecoder(r.Body).Decode(&DataTask)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\n\tresult, err := db.Exec(`INSERT INTO todo (task, datetime_start, datetime_end) VALUES ($1, $2, $3)`,\n\t\t&DataTask.Task, &DataTask.Datetime_start, &DataTask.Datetime_end)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\tjson.NewEncoder(w).Encode(result)\n}", "func (s *Server) PublishWithEvents(ctx context.Context, msg types.EventData, events []abci.Event) error {\n\treturn s.publish(ctx, msg, events)\n}", "func (vs *volumeSet) reserveTaskVolumes(task *api.Task) {\n\tfor _, va := range task.Volumes {\n\t\t// we shouldn't need to handle non-container tasks because those tasks\n\t\t// won't have any entries in task.Volumes.\n\t\tfor _, mount := range task.Spec.GetContainer().Mounts {\n\t\t\tif mount.Source == va.Source && mount.Target == va.Target {\n\t\t\t\tvs.reserveVolume(va.ID, task.ID, task.NodeID, mount.ReadOnly)\n\t\t\t}\n\t\t}\n\t}\n}", "func Tasks(ctx context.Context, biz *rlib.Business) {\n\tvar tldef rlib.TaskListDefinition\n\t// First... define a task list that has a \"pre-due-date\" on the 20th\n\t// and a due date at 5pm on the last day of the month.\n\ttldef.BID = biz.BID\n\ttldef.Cycle = rlib.RECURMONTHLY\n\ttldef.Name = \"Monthly Close\"\n\ttldef.Epoch = time.Date(2018, time.January, 1, 0, 0, 0, 0, time.UTC)\n\ttldef.EpochDue = time.Date(2018, time.January, 31, 17, 0, 0, 0, time.UTC)\n\ttldef.EpochPreDue = time.Date(2018, time.January, 20, 17, 0, 0, 0, time.UTC)\n\n\terr := rlib.InsertTaskListDefinition(ctx, &tldef)\n\tif err != nil {\n\t\tfmt.Printf(\"rlib.InsertTaskListDefinition: error = %s\\n\", err.Error())\n\t\treturn\n\t}\n\n\tvar due = time.Date(2018, time.January, 31, 20, 0, 0, 0, time.UTC) //\n\tvar predue = time.Date(2018, time.January, 20, 20, 0, 0, 0, time.UTC)\n\tvar t = []rlib.TaskDescriptor{\n\t\t{Name: \"Delinquency Report\", Worker: \"Manual\", EpochDue: due, EpochPreDue: predue},\n\t\t{Name: \"Walk the Units\", Worker: \"Manual\", EpochDue: due, EpochPreDue: predue},\n\t\t{Name: \"Generate Offsets\", Worker: \"OffsetBot\", EpochDue: due, EpochPreDue: predue},\n\t}\n\n\tfor i := 0; i < len(t); i++ {\n\t\tt[i].TLDID = tldef.TLDID\n\t\terr := rlib.InsertTaskDescriptor(ctx, &t[i])\n\t\tif err != nil {\n\t\t\tfmt.Printf(\"rlib.InsertTaskDescriptor: error = %s\\n\", err.Error())\n\t\t\treturn\n\t\t}\n\t}\n\n\t//----------------------------------------------\n\t// Now, create an instance of this task list.\n\t//----------------------------------------------\n\tpivot := time.Date(2018, time.February, 3, 12, 32, 13, 0, time.UTC)\n\t_, err = rlib.CreateTaskListInstance(ctx, tldef.TLDID, 0, &pivot)\n\tif err != nil {\n\t\tfmt.Printf(\"CreateTaskListInstance: error = %s\\n\", err.Error())\n\t}\n}", "func AddTaskHandler(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-type\", \"application/json\")\n\tvar u model.Task\n\t_ = json.NewDecoder(r.Body).Decode(&u)\n\tjson.NewEncoder(w).Encode(u)\n\n\tmodel.AddTask(&u)\n}", "func getAllTasks(w http.ResponseWriter, r *http.Request) {\n\tctx := r.Context()\n\tuserID, err := strconv.Atoi(pat.Param(r, \"user-id\"))\n\tif err != nil {\n\t\tlog.Error(ctx, \"string to int problem\", \"error\", err)\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tfmt.Fprintf(w, \"null\")\n\t\treturn\n\t}\n\ttasks, err := tasks.GetAllTasks(ctx, userID)\n\tif err != nil {\n\t\tlog.Error(ctx, \"database problem\", \"error\", err)\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tfmt.Fprintf(w, \"null\")\n\t\treturn\n\t}\n\tdata, err := json.Marshal(tasks)\n\tif err != nil {\n\t\tlog.Error(ctx, \"json marshaling problem\", \"error\", err)\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\tfmt.Fprintf(w, \"null\")\n\t\treturn\n\t}\n\tfmt.Fprintf(w, string(data))\n}", "func (p *WorkerPool[T, R]) AddTask(task T) {\n\tp.taskChan <- task\n}", "func (na *cnmNetworkAllocator) AllocateTask(t *api.Task) error {\n\tfor i, nAttach := range t.Networks {\n\t\tif localNet := na.getNetwork(nAttach.Network.ID); localNet != nil && localNet.isNodeLocal {\n\t\t\tcontinue\n\t\t}\n\t\tif err := na.allocateNetworkIPs(nAttach); err != nil {\n\t\t\tif err := na.releaseEndpoints(t.Networks[:i]); err != nil {\n\t\t\t\tlog.G(context.TODO()).WithError(err).Errorf(\"failed to release IP addresses while rolling back allocation for task %s network %s\", t.ID, nAttach.Network.ID)\n\t\t\t}\n\t\t\treturn errors.Wrapf(err, \"failed to allocate network IP for task %s network %s\", t.ID, nAttach.Network.ID)\n\t\t}\n\t}\n\n\tna.tasks[t.ID] = struct{}{}\n\n\treturn nil\n}", "func (apiHandler *ApiHandler) handleEvents(request *restful.Request, response *restful.Response) {\n\tnamespace := request.PathParameter(\"namespace\")\n\treplicaSet := request.PathParameter(\"replicaSet\")\n\tresult, err := GetEvents(apiHandler.client, namespace, replicaSet)\n\tif err != nil {\n\t\thandleInternalError(response, err)\n\t\treturn\n\t}\n\tresponse.WriteHeaderAndEntity(http.StatusCreated, result)\n}", "func (taskService TaskService) SendTaskOnExecution(ctx context.Context, tasks []models.Task, taskInstance models.TaskInstance) error {\n\tif len(tasks) < 1 {\n\t\treturn fmt.Errorf(\"empty tasks\")\n\t}\n\n\tvar (\n\t\texecutionErr, taskInstanceErr error\n\t\twg = &sync.WaitGroup{}\n\t\ttasksCommonFields = tasks[0]\n\t\tmanagedEndpointIDs = buildManagedEndpoints(tasks, taskInstance)\n\t)\n\n\tpayload := apiModels.ExecutionPayload{\n\t\tExecutionID: taskInstance.ID.String(),\n\t\tManagedEndpoints: managedEndpointIDs,\n\t\tOriginID: taskInstance.OriginID.String(),\n\t\tParameters: tasksCommonFields.Parameters,\n\t\tWebhookURL: fmt.Sprintf(\"%s/partners/%s/task-execution-results/task-instances/%s\",\n\t\t\tconfig.Config.TaskingMsURL, tasksCommonFields.PartnerID, taskInstance.ID),\n\t\tTaskID: tasksCommonFields.ID,\n\t\tExpectedExecutionTimeSec: taskService.templateCache.CalculateExpectedExecutionTimeSec(ctx, tasksCommonFields),\n\t}\n\n\twg.Add(2)\n\n\t// save data about Tasks' expiration\n\tgo func(cx context.Context, task models.Task, ti models.TaskInstance, payload apiModels.ExecutionPayload) {\n\t\tdefer wg.Done()\n\t\texecutionErr = taskService.saveExecutionExpiration(cx, task.PartnerID, ti, payload.ExpectedExecutionTimeSec)\n\t}(ctx, tasksCommonFields, taskInstance, payload)\n\n\t// update TaskInstance\n\tgo func(ctx context.Context, ti models.TaskInstance) {\n\t\tdefer wg.Done()\n\t\t// ti.StartedAt = time.Now().UTC() - do not do this! it breaks scheduler functionality\n\t\tti.LastRunTime = time.Now().UTC()\n\t\tti.OverallStatus = statuses.TaskInstanceRunning\n\t\ttaskInstanceErr = taskService.taskInstancePersistence.Insert(ctx, ti)\n\t}(ctx, taskInstance)\n\twg.Wait()\n\n\tif executionErr != nil {\n\t\tlogger.Log.ErrfCtx(ctx, errorcode.ErrorCantSaveExecutionExpiration, \"saveExecutionExpiration: err: %v\", executionErr)\n\t\treturn executionErr\n\t}\n\tif taskInstanceErr != nil {\n\t\tlogger.Log.ErrfCtx(ctx, errorcode.ErrorCantInsertData, \"error while updating taskInstance [%v]: %v\", taskInstance.ID, taskInstanceErr)\n\t\treturn taskInstanceErr\n\t}\n\n\texecutionURL, err := getExecutionURL(tasksCommonFields.Type, tasksCommonFields.PartnerID)\n\tif err != nil {\n\t\tlogger.Log.WarnfCtx(ctx, \"SendTaskOnExecution: could't get correct execution URL for partnerID=%v and taskType=%v, err=%v\",\n\t\t\ttasksCommonFields.PartnerID, tasksCommonFields.Type, err)\n\t\treturn err\n\t}\n\n\tif tasksCommonFields.IsRunAsUserApplied() {\n\t\tif err = taskService.sendTaskWithCredentials(ctx, tasksCommonFields, managedEndpointIDs, payload, executionURL, taskInstance); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}\n\n\tgo taskService.sendTaskOnExecutionREST(ctx, executionURL, payload, taskInstance, tasksCommonFields)\n\treturn nil\n}", "func DeleteTask(w http.ResponseWriter, r *http.Request, _ httprouter.Params) {\n\tfmt.Fprint(w, \"DeleteTask\\n\")\n}", "func (w Workspace) Tasks(\n\tctx context.Context,\n\tafter *string,\n\tbefore *string,\n\tfirst *int,\n\tlast *int,\n) (TaskConnection, error) {\n\treturn PaginateTaskIDSliceContext(ctx, w.TaskIDs, after, before, first, last)\n}", "func (monitor *TaskMonitor) handle(deliveries <-chan amqp.Delivery, done chan error, out chan interface{}) {\n\n\trawEvent := NewEvent()\n\tfor d := range deliveries {\n\n\t\td.Ack(false)\n\n\t\t// Debug code for ALL messages - useful for extracting ones we are unsure of the json format!\n\t\t/*\n\t\t\tlog.Printf(\"Received %d bytes: [%v] %q\",\n\t\t\t\tlen(d.Body),\n\t\t\t\td.DeliveryTag,\n\t\t\t\td.Body,\n\t\t\t)\n\t\t*/\n\n\t\terr := rawEvent.UnmarshalJSON(d.Body)\n\t\tif err != nil {\n\t\t\tlog.Warnf(\"%v, checking if TaskEvent array...\", err)\n\n\t\t\tvar taskEvents TaskEventsList\n\t\t\terr = json.Unmarshal(d.Body, &taskEvents)\n\t\t\tif err != nil {\n\t\t\t\tlog.Errorf(\"Error: %v, Data was %d bytes: [%v] %q\",\n\t\t\t\t\terr,\n\t\t\t\t\tlen(d.Body),\n\t\t\t\t\td.DeliveryTag,\n\t\t\t\t\td.Body,\n\t\t\t\t)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tfor _, taskEvent := range taskEvents {\n\t\t\t\tout <- &taskEvent\n\t\t\t}\n\n\t\t} else {\n\t\t\tswitch rawEvent.Type {\n\n\t\t\tcase ConstEventTypeWorkerOnline,\n\t\t\t\tConstEventTypeWorkerOffline:\n\n\t\t\t\tvar t = NewWorkerEvent()\n\t\t\t\terr := t.UnmarshalJSON(d.Body)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Errorf(\"Error: %v, Data was %d bytes: [%v] %q\",\n\t\t\t\t\t\terr,\n\t\t\t\t\t\tlen(d.Body),\n\t\t\t\t\t\td.DeliveryTag,\n\t\t\t\t\t\td.Body,\n\t\t\t\t\t)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tout <- t\n\t\t\t\tbreak\n\n\t\t\tcase ConstEventTypeWorkerHeartbeat:\n\t\t\t\tif monitor.monitorWorkerHeartbeatEvents {\n\t\t\t\t\tvar t = NewWorkerEvent()\n\t\t\t\t\terr := t.UnmarshalJSON(d.Body)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Errorf(\"Error: %v, Data was %d bytes: [%v] %q\",\n\t\t\t\t\t\t\terr,\n\t\t\t\t\t\t\tlen(d.Body),\n\t\t\t\t\t\t\td.DeliveryTag,\n\t\t\t\t\t\t\td.Body,\n\t\t\t\t\t\t)\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tout <- t\n\t\t\t\t}\n\t\t\t\tbreak\n\n\t\t\tcase ConstEventTypeTaskSent,\n\t\t\t\tConstEventTypeTaskReceived,\n\t\t\t\tConstEventTypeTaskStarted,\n\t\t\t\tConstEventTypeTaskSucceeded,\n\t\t\t\tConstEventTypeTaskFailed,\n\t\t\t\tConstEventTypeTaskRevoked,\n\t\t\t\tConstEventTypeTaskRetried:\n\n\t\t\t\tvar t = NewTaskEvent()\n\t\t\t\terr := t.UnmarshalJSON(d.Body)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Errorf(\"Error: %v, Data was %d bytes: [%v] %q\",\n\t\t\t\t\t\terr,\n\t\t\t\t\t\tlen(d.Body),\n\t\t\t\t\t\td.DeliveryTag,\n\t\t\t\t\t\td.Body,\n\t\t\t\t\t)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tout <- t\n\t\t\t\tbreak\n\n\t\t\tdefault:\n\t\t\t\tout <- rawEvent\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\n\tlog.Printf(\"deliveries channel closed\")\n\tdone <- nil\n\n}", "func AddTask(t *pb.TaskContainer, taskName string, parameters map[string]string) {\n\tt.ParallelTasks = append(t.ParallelTasks, NewTask(taskName, parameters))\n}" ]
[ "0.62560856", "0.6115658", "0.60783863", "0.60229474", "0.5975949", "0.5699085", "0.54833305", "0.53989345", "0.5368666", "0.53186864", "0.5253385", "0.5113489", "0.5108155", "0.509764", "0.50957435", "0.508231", "0.50769174", "0.50740504", "0.5070625", "0.50645995", "0.5042365", "0.5041454", "0.49864483", "0.4975888", "0.49753562", "0.4966302", "0.4959199", "0.4939508", "0.49322405", "0.48975226", "0.4897522", "0.4894874", "0.48668438", "0.48611668", "0.48610753", "0.484828", "0.4846974", "0.4834764", "0.48205116", "0.4816112", "0.48071277", "0.4770304", "0.476388", "0.47637874", "0.47585633", "0.4753048", "0.4745159", "0.4743864", "0.47126544", "0.47029573", "0.46965516", "0.46932307", "0.46931377", "0.46921808", "0.46846333", "0.46798185", "0.4679647", "0.46773195", "0.46723384", "0.46587288", "0.4653472", "0.4649518", "0.46443635", "0.4640374", "0.46374276", "0.4634848", "0.46347427", "0.46031928", "0.46019322", "0.4599087", "0.45959523", "0.45912656", "0.45852384", "0.45784053", "0.45747536", "0.45679808", "0.45667318", "0.45591056", "0.45442873", "0.45432457", "0.4538868", "0.4537383", "0.4536063", "0.45327047", "0.4529446", "0.4527672", "0.45262134", "0.45198497", "0.45161316", "0.4511063", "0.4510279", "0.45052883", "0.45045233", "0.4502783", "0.4502005", "0.45012203", "0.4493988", "0.44822937", "0.44792745", "0.4476729" ]
0.6764465
0
SignalTask function is used by drivers which support sending OS signals (SIGHUP, SIGKILL, SIGUSR1 etc.) to the task. It is an optional function and is listed as a capability in the driver Capabilities struct.
func (d *Driver) SignalTask(taskID string, signal string) error { handle, ok := d.tasks.Get(taskID) if !ok { return drivers.ErrTaskNotFound } return d.podman.ContainerKill(d.ctx, handle.containerID, signal) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ProcessSignal(p *os.Process, sig os.Signal,) error", "func signal_disable(uint32) {}", "func signal(s os.Signal) {\n\tp, _ := os.FindProcess(os.Getpid())\n\t_ = p.Signal(s)\n\t// Sleep so test won't finish and signal will be received.\n\ttime.Sleep(999)\n}", "func signal_disable(s uint32) {\n}", "func (n *mockAgent) signalProcess(c *Container, processID string, signal syscall.Signal, all bool) error {\n\treturn nil\n}", "func sendSignal(status string) {\n\tcf := cloudformation.New(session.New(&aws.Config{Region: &region}))\n\tparams := &cloudformation.SignalResourceInput{\n\t\tLogicalResourceId: &resource,\n\t\tStackName: &stack,\n\t\tStatus: &status,\n\t\tUniqueId: &uniqueID,\n\t}\n\t_, err := cf.SignalResource(params)\n\tif err != nil {\n\t\tlog.Fatalf(\"Failed to signal CloudFormation: %q.\\n\", err.Error())\n\t}\n\tlog.Printf(\"Sent a %q signal to CloudFormation.\\n\", status)\n\treturn\n}", "func (p *Process) Signal(sig os.Signal) error {\n return p.Process.Signal(sig)\n}", "func (x *CtlCommand) signal(rpcc *xmlrpcclient.XMLRPCClient, sigName string, processes []string) {\n\tfor _, process := range processes {\n\t\tif process == \"all\" {\n\t\t\treply, err := rpcc.SignalAll(process)\n\t\t\tif err == nil {\n\t\t\t\tx.showProcessInfo(&reply, make(map[string]bool))\n\t\t\t} else {\n\t\t\t\tfmt.Printf(\"Fail to send signal %s to all process\", sigName)\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\t\t} else {\n\t\t\treply, err := rpcc.SignalProcess(sigName, process)\n\t\t\tif err == nil && reply.Success {\n\t\t\t\tfmt.Printf(\"Succeed to send signal %s to process %s\\n\", sigName, process)\n\t\t\t} else {\n\t\t\t\tfmt.Printf(\"Fail to send signal %s to process %s\\n\", sigName, process)\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\t\t}\n\t}\n}", "func Signal(signs ...os.Signal) Option {\n\treturn func(o *options) { o.signs = signs }\n}", "func (x *CtlCommand) signal(rpcc *rpcclient.RPCClient, sig_name string, processes []string) {\n\tfor _, process := range processes {\n\t\tif process == \"all\" {\n\t\t\treply, err := rpcc.SignalAllProcesses(&rpcclient.SignalAllProcessesArg{\n\t\t\t\tSignal: sig_name,\n\t\t\t})\n\t\t\tif err == nil {\n\t\t\t\tx.showProcessInfo(reply.AllProcessInfo, make(map[string]bool))\n\t\t\t} else {\n\t\t\t\tfmt.Printf(\"Fail to send signal %s to all process\", sig_name)\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\t\t} else {\n\t\t\treply, err := rpcc.SignalProcess(&rpcclient.SignalProcessArg{\n\t\t\t\tProcName: process,\n\t\t\t\tSignal: sig_name,\n\t\t\t})\n\t\t\tif err == nil && reply.Success {\n\t\t\t\tfmt.Printf(\"Succeed to send signal %s to process %s\\n\", sig_name, process)\n\t\t\t} else {\n\t\t\t\tfmt.Printf(\"Fail to send signal %s to process %s\\n\", sig_name, process)\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\t\t}\n\t}\n}", "func (c *qemuCmd) Signal(sig unix.Signal) error {\n\tcommand := api.InstanceExecControl{\n\t\tCommand: \"signal\",\n\t\tSignal: int(sig),\n\t}\n\n\t// Check handler hasn't finished.\n\tselect {\n\tcase <-c.dataDone:\n\t\treturn fmt.Errorf(\"no such process\") // Aligns with error retured from unix.Kill in lxc's Signal().\n\tdefault:\n\t}\n\n\tc.controlSendCh <- command\n\terr := <-c.controlResCh\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlogger.Debugf(`Forwarded signal \"%d\" to lxd-agent`, sig)\n\treturn nil\n}", "func Signal(sigs ...os.Signal) Option {\n\treturn func(o *options) { o.sigs = sigs }\n}", "func (p Process) Signal(sig os.Signal) error {\n\tif p.ops == nil {\n\t\treturn errInvalidProcess\n\t}\n\treturn p.ops.signal(sig)\n}", "func (p *Process) SendSignal(sig Signal) error {\n\treturn p.SendSignalWithContext(context.Background(), sig)\n}", "func (b *BoatHandle) Signal(sig os.Signal) error { return b.cmd.Process.Signal(sig) }", "func (c *gcsCore) SignalProcess(pid int, options prot.SignalProcessOptions) error {\n\tc.processCacheMutex.Lock()\n\tif _, ok := c.processCache[pid]; !ok {\n\t\tc.processCacheMutex.Unlock()\n\t\treturn gcserr.NewHresultError(gcserr.HrErrNotFound)\n\t}\n\tc.processCacheMutex.Unlock()\n\n\t// Interpret signal value 0 as SIGKILL.\n\t// TODO: Remove this special casing when we are not worried about breaking\n\t// older Windows builds which don't support sending signals.\n\tvar signal syscall.Signal\n\tif options.Signal == 0 {\n\t\tsignal = unix.SIGKILL\n\t} else {\n\t\tsignal = syscall.Signal(options.Signal)\n\t}\n\n\tif err := syscall.Kill(pid, signal); err != nil {\n\t\treturn errors.Wrapf(err, \"failed call to kill on process %d with signal %d\", pid, options.Signal)\n\t}\n\n\treturn nil\n}", "func X__sysv_signal(tls *TLS, signum int32, handler uintptr) {\n\tch := make(chan os.Signal)\n\tgo func() {\n\t\t<-ch\n\t\t(*(*func(*TLS, int32))(unsafe.Pointer(&handler)))(tls, signum)\n\t}()\n\tsignal.Notify(ch, syscall.Signal(signum))\n}", "func (s *Session) Signal(sig os.Signal) {\n\ts.command.Process.Signal(sig)\n}", "func (o *V0037JobProperties) SetSignal(v string) {\n\to.Signal = &v\n}", "func HandleSignal(b bool) Option {\n\treturn func(o *Options) {\n\t\to.Signal = b\n\t}\n}", "func (o *Wireless) SetSignal(v int32) {\n\to.Signal = &v\n}", "func (p *process) Signal(s os.Signal) error {\n\treturn syscall.Kill(p.pid, s.(syscall.Signal))\n}", "func signal() {\n\tnoEvents = true\n}", "func InterruptSignal(signal os.Signal) Option {\n\treturn func(s *Service) {\n\t\ts.interruptSignals = append(s.interruptSignals, signal)\n\t}\n}", "func signal_ignore(s uint32) {\n}", "func (srv *Server) handleSignal(msg *Message) {\n\tsrv.opsLock.Lock()\n\t// Ignore incoming signals during shutdown\n\tif srv.shutdown {\n\t\tsrv.opsLock.Unlock()\n\t\treturn\n\t}\n\tsrv.currentOps++\n\tsrv.opsLock.Unlock()\n\n\tsrv.hooks.OnSignal(context.WithValue(context.Background(), Msg, *msg))\n\n\t// Mark signal as done and shutdown the server if scheduled and no ops are left\n\tsrv.opsLock.Lock()\n\tsrv.currentOps--\n\tif srv.shutdown && srv.currentOps < 1 {\n\t\tclose(srv.shutdownRdy)\n\t}\n\tsrv.opsLock.Unlock()\n}", "func (x *RpcExector) signal(rpcc *rpc.XmlRPCClient, sig_name string, processes []string) {\n\tfor _, process := range processes {\n\t\tif process == \"all\" {\n\t\t\treply, err := rpcc.SignalAll(process)\n\t\t\tif err == nil {\n\t\t\t\tx.showProcessInfo(&reply, make(map[string]bool))\n\t\t\t} else {\n\t\t\t\tfmt.Printf(\"Fail to send signal %s to all process\", sig_name)\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\t\t} else {\n\t\t\treply, err := rpcc.SignalProcess(sig_name, process)\n\t\t\tif err == nil && reply.Success {\n\t\t\t\tfmt.Printf(\"Succeed to send signal %s to process %s\\n\", sig_name, process)\n\t\t\t} else {\n\t\t\t\tfmt.Printf(\"Fail to send signal %s to process %s\\n\", sig_name, process)\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\t\t}\n\t}\n}", "func (a *AbstractSessionChannelHandler) OnSignal(\n\t_ uint64,\n\t_ string,\n) error {\n\treturn fmt.Errorf(\"not supported\")\n}", "func (c *D) Signal(signal os.Signal) error {\n\tif !c.IsRunning() {\n\t\treturn ErrNotRunning\n\t}\n\treturn c.cmd.Process.Signal(signal)\n}", "func Signal(val string) error {\n\t_, err := signals.Parse(val)\n\tif err != nil {\n\t\treturn err //nolint: wrapcheck // error string formed in external package is styled correctly\n\t}\n\n\treturn nil\n}", "func NotifySignal(c chan<- Signal, sig ...Signal) error {\n\tif c == nil {\n\t\treturn fmt.Errorf(\"NotifySignal using nil channel\")\n\t}\n\n\tvar pid = os.Getpid()\n\tevts := make([]windows.Handle, 0, len(sig))\n\n\tfor _, s := range sig {\n\t\tname, err := windows.UTF16PtrFromString(eventName(s, pid))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\th, err := windows.CreateEvent(nil, 1, 0, name)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tevts = append(evts, h)\n\t}\n\n\tgo func() {\n\t\tfor {\n\t\t\tev, err := windows.WaitForMultipleObjects(evts, false, windows.INFINITE)\n\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"WaitForMultipleObjects failed: %v\", err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\toffset := ev - windows.WAIT_OBJECT_0\n\t\t\tc <- sig[offset]\n\t\t\tif err := windows.ResetEvent(evts[offset]); err != nil {\n\t\t\t\tlog.Printf(\"ResetEvent failed: %v\", err)\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn nil\n}", "func (k *KACollector) signal(sig syscall.Signal) error {\n\tps, err := process.Processes()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar pid int32\n\tfor _, p := range ps {\n\t\tname, err := p.Name()\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tif name == \"keepalived\" {\n\t\t\tpid = p.Pid\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif pid == 0 {\n\t\treturn fmt.Errorf(\"cannot find pid\")\n\t}\n\n\tproc, err := os.FindProcess(int(pid))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"process %v: %v\", pid, err)\n\t}\n\n\terr = proc.Signal(sig)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"signal %v: %v\", sig, err)\n\t}\n\n\ttime.Sleep(100 * time.Millisecond)\n\treturn nil\n}", "func (n *node) SignalProcessor() SignalProcessor {\n\treturn n.processor\n}", "func runTask(name string, parts []string, intervalSeconds uint64, signals []interface{}) {\n\tcmd, args := parts[0], parts[1:]\n\tfor {\n\t\tif err := exec.Command(cmd, args...).Run(); err != nil {\n\t\t\tlog.Error(\"Failed to run command '\" + name + \"': \" + err.Error())\n\t\t\ttime.Sleep(time.Duration(intervalSeconds) * time.Second)\n\t\t\tcontinue\n\t\t}\n\t\tlog.Debug(\"Command '\" + name + \"' run successfully...\")\n\n\t\t// Signal other components\n\t\tfor _, signal := range signals {\n\t\t\tmod := int(signal.(core.Config)[\"mod\"].(float64))\n\t\t\tsig := signal.(core.Config)[\"signal\"].(string)\n\t\t\tlog.Infof(\"Invoking signal '%s' on component %d\", sig, mod)\n\t\t\tmods[mod].Signal(sig)\n\t\t}\n\t\ttime.Sleep(time.Duration(intervalSeconds) * time.Second)\n\t}\n}", "func (t *task) Kill(_ context.Context, signal syscall.Signal) error {\n\thcsContainer, err := t.getHCSContainer()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlogger := t.ctr.client.logger.WithFields(log.Fields{\n\t\t\"container\": t.ctr.id,\n\t\t\"process\": t.id,\n\t\t\"pid\": t.Pid(),\n\t\t\"signal\": signal,\n\t})\n\tlogger.Debug(\"Signal()\")\n\n\tvar op string\n\tif signal == syscall.SIGKILL {\n\t\t// Terminate the compute system\n\t\tt.ctr.mu.Lock()\n\t\tt.ctr.terminateInvoked = true\n\t\tt.ctr.mu.Unlock()\n\t\top, err = \"terminate\", hcsContainer.Terminate()\n\t} else {\n\t\t// Shut down the container\n\t\top, err = \"shutdown\", hcsContainer.Shutdown()\n\t}\n\tif err != nil {\n\t\tif !hcsshim.IsPending(err) && !hcsshim.IsAlreadyStopped(err) {\n\t\t\t// ignore errors\n\t\t\tlogger.WithError(err).Errorf(\"failed to %s hccshim container\", op)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (m *Master) SignalTaskStatus(args *model.TaskStatus, reply *bool) error {\n\tif !args.Success {\n\t\treturn nil\n\t}\n\n\tif m.phase == model.Map {\n\t\tlog.Infof(\"map phase for %s completed\", args.File)\n\t\tm.mutex.Lock()\n\t\tdefer m.mutex.Unlock()\n\t\tf := path.Base(args.File)\n\t\tif t, ok := m.mapTasks[f]; ok {\n\t\t\tif t.Status == inprogress {\n\t\t\t\tt.Status = completed\n\t\t\t\tt.Files = append(t.Files, args.OutFiles...)\n\t\t\t\tm.mapTasks[f] = t\n\t\t\t}\n\n\t\t\t// Build up reduce tasks.\n\t\t\tfor i, v := range args.OutFiles {\n\t\t\t\tkey := toString(i + 1)\n\t\t\t\tt := m.reduceTasks[key]\n\t\t\t\tt.Files = append(t.Files, v)\n\t\t\t\tm.reduceTasks[key] = t\n\t\t\t}\n\t\t}\n\t} else if m.phase == model.Reduce {\n\t\tlog.Infof(\"reduce phase %s completed\", args.File)\n\t\ti, _ := strconv.ParseInt(args.File, 10, 32)\n\t\tkey := toString(i + 1)\n\t\tm.mutex.Lock()\n\t\tdefer m.mutex.Unlock()\n\t\tif t, ok := m.reduceTasks[key]; ok {\n\t\t\tif t.Status == inprogress {\n\t\t\t\tt.Status = completed\n\t\t\t\tm.reduceTasks[key] = t\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func PidfdSendSignal(pidfd uintptr, signum unix.Signal) error {\n\t// the runtime OS thread must be locked to safely enter namespaces.\n\truntime.LockOSThread()\n\tdefer runtime.UnlockOSThread()\n\t// setns with pidfd requires at least kernel version 5.8.0\n\terr := unix.Setns(int(pidfd), unix.CLONE_NEWPID)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// pifd_send_signal was introduced in kernel version 5.3\n\t_, _, e1 := unix.Syscall(unix.SYS_PIDFD_SEND_SIGNAL, pidfd, uintptr(signum), 0)\n\tif e1 != 0 {\n\t\treturn e1\n\t}\n\treturn nil\n}", "func (c *Cond) Signal() {\n\tc.Do(func() {})\n}", "func (l *Libvirt) DomainSendProcessSignal(Dom Domain, PidValue int64, Signum uint32, Flags uint32) (err error) {\n\tvar buf []byte\n\n\targs := DomainSendProcessSignalArgs {\n\t\tDom: Dom,\n\t\tPidValue: PidValue,\n\t\tSignum: Signum,\n\t\tFlags: Flags,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\n\t_, err = l.requestStream(295, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func handleSignal(env *Environment) {\n\tch := make(chan os.Signal, 2)\n\tsignal.Notify(ch, stopSignals...)\n\n\tgo func() {\n\t\ts := <-ch\n\t\tdelay := getDelaySecondsFromEnv()\n\t\tlog.Warn(\"well: got signal\", map[string]interface{}{\n\t\t\t\"signal\": s.String(),\n\t\t\t\"delay\": delay,\n\t\t})\n\t\ttime.Sleep(time.Duration(delay) * time.Second)\n\t\tenv.Cancel(errSignaled)\n\t}()\n}", "func (p *promise) Signal(waitChan chan Controller) Promise {\n\tp.Always(func(p2 Controller) {\n\t\twaitChan <- p2\n\t})\n\n\treturn p\n}", "func sendSignal(cmd *exec.Cmd, ch <-chan error, sig syscall.Signal, timeout time.Duration) bool {\n\tif cmd.Process == nil {\n\t\tlog.Debug(\"Not terminating process, it seems to have not started yet\")\n\t\treturn false\n\t}\n\t// This is a bit of a fiddle. We want to wait for the process to exit but only for just so\n\t// long (we do not want to get hung up if it ignores our SIGTERM).\n\tlog.Debug(\"Sending signal %s to -%d\", sig, cmd.Process.Pid)\n\tsyscall.Kill(-cmd.Process.Pid, sig) // Kill the group - we always set one in ExecCommand.\n\n\tselect {\n\tcase <-ch:\n\t\treturn true\n\tcase <-time.After(timeout):\n\t\treturn false\n\t}\n}", "func signal(x *decimal.Big, c decimal.Condition, err error) *decimal.Big {\n\tswitch ctx := &x.Context; ctx.OperatingMode {\n\tcase decimal.Go:\n\t\t// Go mode always panics on NaNs.\n\t\tif _, ok := err.(decimal.ErrNaN); ok {\n\t\t\tpanic(err)\n\t\t}\n\tcase decimal.GDA:\n\t\tctx.Conditions = c\n\t\tif c&ctx.Traps != 0 {\n\t\t\tctx.Err = err\n\t\t}\n\tdefault:\n\t\tctx.Conditions = c | decimal.InvalidContext\n\t\tctx.Err = fmt.Errorf(\"invalid OperatingMode: %d\", ctx.OperatingMode)\n\t\t// TODO(eric): Add a SetNaN method?\n\t\tx.SetString(\"qNaN\")\n\t}\n\treturn x\n}", "func TestSignal(t *testing.T) {\n\t// Ask for SIGHUP\n\tc := make(chan os.Signal, 1)\n\tNotify(c, syscall.SIGHUP)\n\tdefer Stop(c)\n\n\t// Send this process a SIGHUP\n\tt.Logf(\"sighup...\")\n\tsyscall.Kill(syscall.Getpid(), syscall.SIGHUP)\n\twaitSig(t, c, syscall.SIGHUP)\n\n\t// Ask for everything we can get. The buffer size has to be\n\t// more than 1, since the runtime might send SIGURG signals.\n\t// Using 10 is arbitrary.\n\tc1 := make(chan os.Signal, 10)\n\tNotify(c1)\n\t// Stop relaying the SIGURG signals. See #49724\n\tReset(syscall.SIGURG)\n\tdefer Stop(c1)\n\n\t// Send this process a SIGWINCH\n\tt.Logf(\"sigwinch...\")\n\tsyscall.Kill(syscall.Getpid(), syscall.SIGWINCH)\n\twaitSigAll(t, c1, syscall.SIGWINCH)\n\n\t// Send two more SIGHUPs, to make sure that\n\t// they get delivered on c1 and that not reading\n\t// from c does not block everything.\n\tt.Logf(\"sighup...\")\n\tsyscall.Kill(syscall.Getpid(), syscall.SIGHUP)\n\twaitSigAll(t, c1, syscall.SIGHUP)\n\tt.Logf(\"sighup...\")\n\tsyscall.Kill(syscall.Getpid(), syscall.SIGHUP)\n\twaitSigAll(t, c1, syscall.SIGHUP)\n\n\t// The first SIGHUP should be waiting for us on c.\n\twaitSig(t, c, syscall.SIGHUP)\n}", "func (server *Server) SendTask(s *TaskSignature) error {\n\tmessage, err := json.Marshal(s)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"JSON Encode Message: %v\", err)\n\t}\n\n\tif err := server.connection.PublishMessage(\n\t\t[]byte(message), s.RoutingKey,\n\t); err != nil {\n\t\treturn fmt.Errorf(\"Publish Message: %v\", err)\n\t}\n\n\treturn nil\n}", "func (c *Cmd) Signal(signal syscall.Signal) error {\n\tif c.Process == nil {\n\t\treturn errNotStarted\n\t}\n\tif c.ProcessState != nil {\n\t\treturn errAlreadyWaited\n\t}\n\n\t// Negative PID means the process group led by the process.\n\treturn syscall.Kill(-c.Process.Pid, signal)\n}", "func (c *Cmd) Signal(sig os.Signal) error {\n\treturn signal(c.cmd.Process, sig)\n}", "func TestSignals(t *testing.T) {\n\tseq := make(chan int)\n\twait := make(chan int)\n\tfreq := make(chan time.Time)\n\n\tqueue := &WaitQueue{\n\t\tsem: new(sync.WaitGroup),\n\t\tseq: seq,\n\t\twait: wait,\n\t}\n\n\t// begin listening\n\tgo waitListen(queue, freq, seq)\n\n\t// send a tick, this should start a call to Poll()\n\tfreq <- time.Now()\n\n\t// when that call starts, we should get `1` on the sequence channel\n\tval := <-seq\n\trequire.Equal(t, val, 1)\n\n\t// send a signal, this should start the graceful exit\n\tsignals <- os.Interrupt\n\n\t// tell Poll() that it can exit\n\twait <- 1\n\n\t// first Poll() should exit\n\tval = <-seq\n\trequire.Equal(t, val, 2)\n\n\t// then Listen() should exit\n\tval = <-seq\n\trequire.Equal(t, val, 3)\n}", "func (g *Goer) installSignal() {\n\tch := make(chan os.Signal, 1)\n\tsignal.Notify(ch, syscall.SIGINT, syscall.SIGTERM, syscall.SIGQUIT, syscall.SIGUSR1, syscall.SIGUSR2)\n\tfor signalType := range ch {\n\t\tswitch signalType {\n\t\t// stop process in debug mode with Ctrl+c.\n\t\tcase syscall.SIGINT:\n\t\t\tg.stopAll(ch, signalType)\n\t\t// kill signal in bash shell.\n\t\tcase syscall.SIGKILL | syscall.SIGTERM:\n\t\t\tg.stopAll(ch, signalType)\n\t\t// graceful reload\n\t\tcase syscall.SIGQUIT:\n\t\t\tsignal.Stop(ch)\n\t\t\tg.reload()\n\t\t\tos.Exit(0)\n\t\t}\n\t}\n}", "func waitForTask() {\n\tch := make(chan string)\n\n\tgo func() {\n\t\td := <-ch\n\t\tfmt.Println(\"child : recv'd signal :\", d)\n\t}()\n\n\ttime.Sleep(time.Duration(rand.Intn(500)) * time.Millisecond)\n\tch <- \"data\"\n\tfmt.Println(\"parent : sent signal\")\n\n\ttime.Sleep(time.Second)\n\tfmt.Println(\"-------------------------------------------------\")\n}", "func (t *Broadcaster) Signal(ctx context.Context) error {\n\tif !t.mutex.RTryLock(ctx) {\n\t\treturn context.DeadlineExceeded\n\t}\n\tdefer t.mutex.RUnlock()\n\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn context.DeadlineExceeded\n\tcase t.channel <- struct{}{}:\n\tdefault:\n\t}\n\n\treturn nil\n}", "func (wc *workflowClient) SignalWorkflow(ctx context.Context, workflowID string, runID string, signalName string, arg interface{}) error {\n\tinput, err := encodeArg(wc.dataConverter, arg)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn signalWorkflow(ctx, wc.workflowService, wc.identity, wc.domain, workflowID, runID, signalName, input, wc.featureFlags)\n}", "func (sd *ScanDiapasons) NotifyDpnsTask(ctx context.Context) ([]byte, error) {\n\trequest, err := http.NewRequest(\"POST\", sd.client.Server+\"/api/v1.0/ScanDiapasons.NotifyDpnsTask\", nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\traw, err := sd.client.Do(ctx, request, nil)\n\treturn raw, err\n}", "func (p *Process) signalToProcess(signal os.Signal) error {\n\tif p.command == nil || p.command.Process == nil {\n\t\terr := errors.Errorf(\"attempt to send signal to non-running process\")\n\t\tp.log.Error(err)\n\t\treturn err\n\t}\n\n\treturn p.command.Process.Signal(signal)\n}", "func (o *ContainerSignalParams) SetSignal(signal int64) {\n\to.Signal = signal\n}", "func (s *sidecar) signalProcess() (err error) {\n\tif atomic.LoadInt32(&s.processRunning) == 0 {\n\t\tcmd := exec.Command(s.config.Cmd, strings.Split(s.config.CmdArgs, \" \")...)\n\t\tcmd.Stdout = os.Stdout\n\t\tcmd.Stderr = os.Stderr\n\t\terr = cmd.Start()\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error executing process: %v\\n%v\", s.config.Cmd, err)\n\t\t}\n\t\ts.process = cmd.Process\n\t\tgo s.checkProcessExit()\n\t} else {\n\t\t// Signal to reload certs\n\t\tsig, err := getSignal(s.config.RenewSignal)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error getting signal: %v\\n%v\", s.config.RenewSignal, err)\n\t\t}\n\n\t\terr = s.process.Signal(sig)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error signaling process with signal: %v\\n%v\", sig, err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func execNotify(arity int, p *gop.Context) {\n\targs := p.GetArgs(arity)\n\tconv := func(args []interface{}) []os.Signal {\n\t\tret := make([]os.Signal, len(args))\n\t\tfor i, arg := range args {\n\t\t\tret[i] = arg.(os.Signal)\n\t\t}\n\t\treturn ret\n\t}\n\tsignal.Notify(args[0].(chan<- os.Signal), conv(args[1:])...)\n}", "func (t ThriftHandler) SignalWorkflowExecution(ctx context.Context, request *h.SignalWorkflowExecutionRequest) (err error) {\n\terr = t.h.SignalWorkflowExecution(ctx, request)\n\treturn thrift.FromError(err)\n}", "func (s *BasevhdlListener) EnterSignal_mode(ctx *Signal_modeContext) {}", "func ToSignal(name string) (os.Signal, error) {\n\tif sig, ok := name2signal[name]; ok {\n\t\treturn sig, nil\n\t}\n\treturn nil, fmt.Errorf(\"invalid signal name: %s\", name)\n}", "func SetupSignal() {\n\tc := make(chan os.Signal, 1)\n\tsignal.Notify(c, os.Interrupt, syscall.SIGTERM)\n\tfor {\n\t\ts := <-c\n\t\tlog.WithField(\"signal\", s.String()).Info(\"signal\")\n\t\tswitch s {\n\t\tcase os.Interrupt, syscall.SIGTERM:\n\t\t\treturn\n\t\tdefault:\n\t\t\treturn\n\t\t}\n\t}\n}", "func SetupSignalTrap(handler func(os.Signal)) chan os.Signal {\n\tif len(notary.NotarySupportedSignals) == 0 {\n\t\treturn nil\n\n\t}\n\tc := make(chan os.Signal, 1)\n\tsignal.Notify(c, notary.NotarySupportedSignals...)\n\tgo func() {\n\t\tfor {\n\t\t\thandler(<-c)\n\t\t}\n\t}()\n\n\treturn c\n}", "func (this *ThreadCtl) WaitSignal() string {\n\tselect {\n\tcase signal := <-this.signalChan:\n\t\treturn signal\n\t}\n}", "func OnSignal(handler func(os.Signal), signals ...os.Signal) {\n\tif handler == nil || len(signals) == 0 {\n\t\treturn\n\t}\n\n\tsh := &sigHandler{\n\t\tsignals: signals,\n\t\ttarget: handler,\n\t}\n\tsh.Start()\n}", "func (s *Session) SendTask(t util.Task) error {\n\t// Checking taskID. re-enqueued task will be skipped\n\tif t.TaskID == \"\" {\n\t\tt.TaskID = uuid.New().String()\n\t\tif t.OriginalTaskID == \"\" {\n\t\t\tt.OriginalTaskID = t.TaskID\n\t\t}\n\n\t\tif err := s.taskRepo.CreateTask(&t); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Checking AMQP connection. Task will be logged for no connection. Re-enqueued later.\n\ts.mu.RLock()\n\tif !s.connected {\n\t\ts.lgr.Warn(\"No connection. Task enqueued.\", util.Object{Key: \"TaskID\", Val: t.TaskID})\n\t\treturn ErrNotConnected\n\t}\n\ts.mu.RUnlock()\n\n\tif !t.Priority.Valid() {\n\t\treturn ErrInvalidPriority\n\t}\n\n\tch, err := s.conn.Channel()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer ch.Close()\n\n\tif err := ch.Confirm(false); err != nil {\n\t\treturn err\n\t}\n\n\tcloseNotification := ch.NotifyClose(make(chan *amqp.Error, 1))\n\tpublish := ch.NotifyPublish(make(chan amqp.Confirmation, 1))\n\tpublishErr := make(chan error, 1)\n\n\tQueue, err := s.GetQueueName(t.Priority)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tgo func() {\n\t\terr := ch.Publish(\n\t\t\ts.cfg.AMQP.Exchange,\n\t\t\tQueue,\n\t\t\tfalse,\n\t\t\tfalse,\n\t\t\tamqp.Publishing{\n\t\t\t\tHeaders: map[string]interface{}{\n\t\t\t\t\t\"TaskName\": t.Name,\n\t\t\t\t\t\"TaskID\": t.TaskID,\n\t\t\t\t},\n\t\t\t\tMessageId: t.TaskID,\n\t\t\t\tDeliveryMode: amqp.Persistent,\n\t\t\t\tBody: t.Payload,\n\t\t\t},\n\t\t)\n\n\t\tif err != nil {\n\t\t\tpublishErr <- err\n\t\t\treturn\n\t\t}\n\n\t\ts.taskRepo.UpdateTaskStatus(context.Background(), t.TaskID, util.StatusQueued)\n\t}()\n\n\tdone := (<-chan time.Time)(make(chan time.Time, 1))\n\tif s.cfg.RequestTimeout != 0 {\n\t\tdone = time.After(s.cfg.RequestTimeout)\n\t}\n\n\tvar errs error\n\n\tselect {\n\tcase errs = <-closeNotification:\n\n\tcase errs = <-publishErr:\n\n\tcase p := <-publish:\n\t\tif !p.Ack {\n\t\t\ts.lgr.Warn(\"Task deliver failed\", util.Object{Key: \"TaskID\", Val: t.TaskID})\n\t\t\terrs = ErrNotPublished\n\t\t\tbreak\n\t\t}\n\t\ts.lgr.Info(\"Task delivered\", util.Object{Key: \"TaskID\", Val: t.TaskID})\n\tcase <-done:\n\t\terrs = ErrRequestTimeout\n\t}\n\n\t// For any kind of error, task will be retried if retry count non zero.\n\t// TODO: retry count only reduce for task processing related error.\n\tif errs != nil {\n\t\tif orgTask, err := s.taskRepo.GetTask(t.OriginalTaskID); err != nil {\n\t\t\ts.lgr.Error(\"failed to get task\", err, util.Object{Key: \"TaskID\", Val: t.OriginalTaskID})\n\t\t} else if orgTask.Retry != 0 {\n\t\t\tgo s.RetryTask(t)\n\t\t}\n\n\t\ts.taskRepo.UpdateTaskStatus(context.Background(), t.TaskID, util.StatusFailed, errs)\n\t}\n\n\treturn errs\n}", "func TrapSignal(cleanupFunc func()) {\n\tsigs := make(chan os.Signal, 1)\n\tsignal.Notify(sigs, syscall.SIGINT, syscall.SIGTERM)\n\n\tgo func() {\n\t\tsig := <-sigs\n\n\t\tif cleanupFunc != nil {\n\t\t\tcleanupFunc()\n\t\t}\n\t\texitCode := 128\n\n\t\tswitch sig {\n\t\tcase syscall.SIGINT:\n\t\t\texitCode += int(syscall.SIGINT)\n\t\tcase syscall.SIGTERM:\n\t\t\texitCode += int(syscall.SIGTERM)\n\t\t}\n\n\t\tos.Exit(exitCode)\n\t}()\n}", "func (s *ScanDrivers) SetScanTaskFlagsFunc(taskFlag func(name string, flag goforjj.YamlFlag) error) {\n\tif s == nil {\n\t\treturn\n\t}\n\ts.taskFlag = taskFlag\n}", "func (s WorkerSemaphore) Signal(n int) {\n\te := empty{}\n\tfor i := 0; i < n; i++ {\n\t\ts.permits <- e\n\t}\n}", "func (h *Hub) StartTask(ctx context.Context, request *pb.HubStartTaskRequest) (*pb.HubStartTaskReply, error) {\n\tlog.G(h.ctx).Info(\"handling StartTask request\", zap.Any(\"req\", request))\n\n\ttaskID := uuid.New()\n\tminer, err := h.selectMiner(request)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar startRequest = &pb.MinerStartRequest{\n\t\tId: taskID,\n\t\tRegistry: request.Registry,\n\t\tImage: request.Image,\n\t\tAuth: request.Auth,\n\t\tPublicKeyData: request.PublicKeyData,\n\t\tCommitOnStop: request.CommitOnStop,\n\t\tEnv: request.Env,\n\t\tUsage: request.Requirements.GetResources(),\n\t\tRestartPolicy: &pb.ContainerRestartPolicy{\n\t\t\tName: \"\",\n\t\t\tMaximumRetryCount: 0,\n\t\t},\n\t}\n\n\tresp, err := miner.Client.Start(ctx, startRequest)\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.Internal, \"failed to start %v\", err)\n\t}\n\n\troutes := []extRoute{}\n\tfor k, v := range resp.Ports {\n\t\t_, protocol, err := decodePortBinding(k)\n\t\tif err != nil {\n\t\t\tlog.G(h.ctx).Warn(\"failed to decode miner's port mapping\",\n\t\t\t\tzap.String(\"mapping\", k),\n\t\t\t\tzap.Error(err),\n\t\t\t)\n\t\t\tcontinue\n\t\t}\n\n\t\trealPort, err := strconv.ParseUint(v.Port, 10, 16)\n\t\tif err != nil {\n\t\t\tlog.G(h.ctx).Warn(\"failed to convert real port to uint16\",\n\t\t\t\tzap.Error(err),\n\t\t\t\tzap.String(\"port\", v.Port),\n\t\t\t)\n\t\t\tcontinue\n\t\t}\n\n\t\troute, err := miner.router.RegisterRoute(taskID, protocol, v.IP, uint16(realPort))\n\t\tif err != nil {\n\t\t\tlog.G(h.ctx).Warn(\"failed to register route\", zap.Error(err))\n\t\t\tcontinue\n\t\t}\n\t\troutes = append(routes, extRoute{\n\t\t\tcontainerPort: k,\n\t\t\troute: route,\n\t\t})\n\t}\n\n\th.setMinerTaskID(miner.ID(), taskID)\n\n\tresources := request.GetRequirements().GetResources()\n\tcpuCount := resources.GetCPUCores()\n\tmemoryCount := resources.GetMaxMemory()\n\n\tvar usage = resource.NewResources(int(cpuCount), int64(memoryCount))\n\tif err := miner.Consume(taskID, &usage); err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar reply = pb.HubStartTaskReply{\n\t\tId: taskID,\n\t}\n\n\tfor _, route := range routes {\n\t\treply.Endpoint = append(\n\t\t\treply.Endpoint,\n\t\t\tfmt.Sprintf(\"%s->%s:%d\", route.containerPort, route.route.Host, route.route.Port),\n\t\t)\n\t}\n\n\treturn &reply, nil\n}", "func (s *ShutdownManager) SignalShutdown() {\n\ts.ShutdownState = true\n}", "func (ep *ExpectProcess) Signal(sig os.Signal) error {\n\tep.mu.Lock()\n\tdefer ep.mu.Unlock()\n\n\tif ep.cmd == nil {\n\t\treturn errors.New(\"expect process already closed\")\n\t}\n\n\treturn ep.cmd.Process.Signal(sig)\n}", "func signal_enable(s uint32) {\n\tif !sig.inuse {\n\t\t// The first call to signal_enable is for us\n\t\t// to use for initialization. It does not pass\n\t\t// signal information in m.\n\t\tsig.inuse = true // enable reception of signals; cannot disable\n\t\tnoteclear(&sig.note)\n\t\treturn\n\t}\n}", "func (n *Node) NodeSignal() *Signal {\n\treturn &n.NodeSig\n}", "func signalName(s syscall.Signal) string {\n\tif name, ok := signalNames[s]; ok {\n\t\treturn name\n\t}\n\treturn fmt.Sprintf(\"SIG %d\", s)\n}", "func (t *SignalTable)StartSignalHandle() {\n go t.signalHandle()\n}", "func Example_signal() {\n\tevents.Listen(&events.Listener{\n\t\tEventName: SignalHello,\n\t\tHandler: func(e events.Event) {\n\t\t\tfmt.Println(e)\n\t\t},\n\t})\n\tevents.Emit(events.Signal(SignalHello))\n\t// Output: Hello world\n}", "func (srv *Server) RegisterSignalHook(ppFlag int, sig os.Signal, f func()) (err error) {\n\tif ppFlag != PreSignal && ppFlag != PostSignal {\n\t\terr = fmt.Errorf(\"Invalid ppFlag argument. Must be either grace.PreSignal or grace.PostSignal\")\n\t\treturn\n\t}\n\tfor _, s := range hookableSignals {\n\t\tif s == sig {\n\t\t\tsrv.SignalHooks[ppFlag][sig] = append(srv.SignalHooks[ppFlag][sig], f)\n\t\t\treturn\n\t\t}\n\t}\n\terr = fmt.Errorf(\"Signal '%v' is not supported\", sig)\n\treturn\n}", "func (c cmdType) toServiceSignal() serviceSignal {\n\tswitch c {\n\tcase statusCmd:\n\t\treturn serviceStatus\n\tcase restartCmd:\n\t\treturn serviceRestart\n\t}\n\treturn serviceStatus\n}", "func SignalNum(s string) syscall.Signal {\n\tsignalNameMapOnce.Do(func() {\n\t\tsignalNameMap = make(map[string]syscall.Signal, len(signalList))\n\t\tfor _, signal := range signalList {\n\t\t\tsignalNameMap[signal.name] = signal.num\n\t\t}\n\t})\n\treturn signalNameMap[s]\n}", "func (m *MockOSProcess) Signal(arg0 os.Signal) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"Signal\", arg0)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (s *BasevhdlListener) EnterSignal_kind(ctx *Signal_kindContext) {}", "func (g *Pin) Notify(sig ...os.Signal) {\n\tc := make(chan os.Signal)\n\tsignal.Notify(c, sig...)\n\tgo func() {\n\t\tn := 0\n\t\tfor sig := range c {\n\t\t\tif n == 1 {\n\t\t\t\tpanic(\"got too many signals\")\n\t\t\t}\n\t\t\tg.Pull(fmt.Errorf(\"Recieved signal %s\", sig))\n\t\t\tn++\n\t\t}\n\t}()\n}", "func (t ThriftHandler) SignalWithStartWorkflowExecution(ctx context.Context, request *h.SignalWithStartWorkflowExecutionRequest) (response *shared.StartWorkflowExecutionResponse, err error) {\n\tresponse, err = t.h.SignalWithStartWorkflowExecution(ctx, request)\n\treturn response, thrift.FromError(err)\n}", "func (d *Daemon) Signal(sig os.Signal) error {\n\tprocess, err := d.Process()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn process.Signal(sig)\n}", "func (op *AddonOperator) TaskHandler(t sh_task.Task) queue.TaskResult {\n\ttaskLogLabels := t.GetLogLabels()\n\ttaskLogEntry := log.WithFields(utils.LabelsToLogFields(taskLogLabels))\n\tvar res queue.TaskResult\n\n\top.logTaskStart(taskLogEntry, t)\n\n\top.UpdateWaitInQueueMetric(t)\n\n\tswitch t.GetType() {\n\tcase task.GlobalHookRun:\n\t\tres = op.HandleGlobalHookRun(t, taskLogLabels)\n\n\tcase task.GlobalHookEnableScheduleBindings:\n\t\thm := task.HookMetadataAccessor(t)\n\t\tglobalHook := op.ModuleManager.GetGlobalHook(hm.HookName)\n\t\tglobalHook.HookController.EnableScheduleBindings()\n\t\tres.Status = queue.Success\n\n\tcase task.GlobalHookEnableKubernetesBindings:\n\t\tres = op.HandleGlobalHookEnableKubernetesBindings(t, taskLogLabels)\n\n\tcase task.GlobalHookWaitKubernetesSynchronization:\n\t\tres.Status = queue.Success\n\t\tif op.ModuleManager.GlobalSynchronizationNeeded() && !op.ModuleManager.GlobalSynchronizationState().IsComplete() {\n\t\t\t// dump state\n\t\t\top.ModuleManager.GlobalSynchronizationState().DebugDumpState(taskLogEntry)\n\t\t\tt.WithQueuedAt(time.Now())\n\t\t\tres.Status = queue.Repeat\n\t\t} else {\n\t\t\ttaskLogEntry.Info(\"Synchronization done for all global hooks\")\n\t\t}\n\n\tcase task.DiscoverHelmReleases:\n\t\tres = op.HandleDiscoverHelmReleases(t, taskLogLabels)\n\n\tcase task.ConvergeModules:\n\t\tres = op.HandleConvergeModules(t, taskLogLabels)\n\n\tcase task.ModuleRun:\n\t\tres = op.HandleModuleRun(t, taskLogLabels)\n\n\tcase task.ModuleDelete:\n\t\tres.Status = op.HandleModuleDelete(t, taskLogLabels)\n\n\tcase task.ModuleHookRun:\n\t\tres = op.HandleModuleHookRun(t, taskLogLabels)\n\n\tcase task.ModulePurge:\n\t\tres.Status = op.HandleModulePurge(t, taskLogLabels)\n\t}\n\n\tif res.Status == queue.Success {\n\t\torigAfterHandle := res.AfterHandle\n\t\tres.AfterHandle = func() {\n\t\t\top.CheckConvergeStatus(t)\n\t\t\tif origAfterHandle != nil {\n\t\t\t\torigAfterHandle()\n\t\t\t}\n\t\t}\n\t}\n\n\top.logTaskEnd(taskLogEntry, t, res)\n\n\treturn res\n}", "func (wp *WorkPool) PanicFunc(handler PanicFunc) {\n\twp.panicFunc = handler\n}", "func Kill(sig os.Signal) {\n go func() {\n signals.ch <- sig\n }()\n}", "func signals(signals ...os.Signal) (<-chan os.Signal, func()) {\n\tsigchan := make(chan os.Signal)\n\tsigrecv := events.Signal(sigchan)\n\tsignal.Notify(sigchan, signals...)\n\treturn sigrecv, func() { signal.Stop(sigchan) }\n}", "func handlerSignal() {\n\texpectedSignals := make(chan os.Signal, 1)\n\tdoneSignals := make(chan bool, 1)\n\n\t// register channel to receive 2 signals\n\tsignal.Notify(expectedSignals, syscall.SIGTERM, syscall.SIGINT)\n\n\t// this routine is blocking, i.e. when it gets one signal it prints it and notifies the program that it can finish\n\tgo func() {\n\t\tsig := <-expectedSignals\n\t\tfmt.Println()\n\t\tfmt.Println(sig.String())\n\t\tdoneSignals <- true\n\t}()\n\n\tfmt.Println(\"awaiting signal...\")\n\n\t<-doneSignals\n\n\tfmt.Println(\"exiting...\")\n}", "func trapSignals(sig ...os.Signal) {\n\t// channel to receive signals.\n\tsigCh := make(chan os.Signal, 1)\n\tdefer close(sigCh)\n\n\t// `signal.Notify` registers the given channel to\n\t// receive notifications of the specified signals.\n\tsignal.Notify(sigCh, sig...)\n\n\t// Wait for the signal.\n\t<-sigCh\n\n\t// Once signal has been received stop signal Notify handler.\n\n\tsignal.Stop(sigCh)\n\n\t// Cancel the global context\n\tglobalCancel()\n\n}", "func (f *Failer) KillTask(host, task string) error {\n\tscript := \"sudo pkill -x %s\"\n\tlog.V(1).Infof(\"Killing task %s on host %s\", task, host)\n\treturn f.runWithEvilTag(host, fmt.Sprintf(script, task))\n}", "func (m *MockProcess) Signal(arg0 os.Signal) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"Signal\", arg0)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func KillTask(tid int) Errno {\n\t_, e := internal.Syscall1(KILLTASK, uintptr(tid))\n\treturn Errno(e)\n}", "func (q *queue) Signal() {\n\tq.notEmpty.Broadcast()\n}", "func (s *Server) RegisterTask(name string, handlerFunc *func(v []interface{}) (interface{}, error)) error {\n\tif _, found := s.handlers[name]; found {\n\t\treturn ErrDuplicateHandler\n\t}\n\ts.handlers[name] = handlerFunc\n\treturn nil\n}", "func (server *Server) RegisterTask(name string, task interface{}) {\n\tserver.registeredTasks[name] = task\n}", "func (s *countingSemaphore) Signal() {\n\ts.sem <- 1\n}", "func (r *Redis) PushTask(task *Task) error {\n\tconn := r.pool.Get()\n\tdefer conn.Close()\n\n\t_, err := conn.Do(\"ZADD\", syncTasksPriorityQueueKey, task.Priority, task.ID)\n\tnoticeError(err)\n\tif err != nil {\n\t\tif err == redis.ErrNil {\n\t\t\treturn nil\n\t\t}\n\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *gcsCore) SignalContainer(id string, signal syscall.Signal) error {\n\tc.containerCacheMutex.Lock()\n\tdefer c.containerCacheMutex.Unlock()\n\n\tcontainerEntry := c.getContainer(id)\n\tif containerEntry == nil {\n\t\treturn gcserr.NewHresultError(gcserr.HrVmcomputeSystemNotFound)\n\t}\n\n\tif containerEntry.container != nil {\n\t\tif err := containerEntry.container.Kill(signal); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func writeCrashSignal(crashSig string) {\n\tif err := ioutil.WriteFile(crashSig, []byte{}, 0644); err != nil {\n\t\tlog.Fatalf(\"failed to write crash signal: %v\", err)\n\t}\n}" ]
[ "0.61870855", "0.5511749", "0.5325872", "0.5245061", "0.52382535", "0.5158235", "0.5155546", "0.5136468", "0.5113385", "0.5107931", "0.50993425", "0.508363", "0.50823617", "0.5018183", "0.5008709", "0.49947438", "0.49496862", "0.49458396", "0.49453184", "0.49269986", "0.49240094", "0.49110368", "0.4904311", "0.48738256", "0.48690024", "0.48655665", "0.48462573", "0.48242965", "0.48215464", "0.47883642", "0.47867918", "0.4785958", "0.4762723", "0.47540084", "0.47390747", "0.47348765", "0.47182396", "0.46998435", "0.46948454", "0.4683117", "0.4666159", "0.46564183", "0.46561903", "0.46381968", "0.46290815", "0.46285972", "0.46197823", "0.4618494", "0.4598269", "0.4591346", "0.45692986", "0.45574635", "0.45524114", "0.4535499", "0.4524541", "0.45223144", "0.44878972", "0.44832766", "0.44568923", "0.44546703", "0.44451788", "0.44346553", "0.44317025", "0.44225568", "0.44191268", "0.4418454", "0.44088608", "0.44021666", "0.4398398", "0.43868148", "0.43821487", "0.43802446", "0.43708178", "0.43595684", "0.4358428", "0.4356591", "0.43501025", "0.43355572", "0.43231592", "0.43193874", "0.43135682", "0.43079275", "0.4306713", "0.42993408", "0.42923892", "0.42883217", "0.42853367", "0.42758608", "0.42686063", "0.42655313", "0.4263792", "0.42633364", "0.4261257", "0.42579207", "0.42531994", "0.42523578", "0.42514333", "0.42438874", "0.42424598", "0.42284328" ]
0.7002219
0
ExecTask function is used by the Nomad client to execute commands inside the task execution context.
func (d *Driver) ExecTask(taskID string, cmd []string, timeout time.Duration) (*drivers.ExecTaskResult, error) { return nil, fmt.Errorf("Podman driver does not support exec") }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (ts *TaskService) Exec(ctx context.Context, req *taskAPI.ExecProcessRequest) (*types.Empty, error) {\n\tdefer logPanicAndDie(log.G(ctx))\n\n\tlog.G(ctx).WithFields(logrus.Fields{\"id\": req.ID, \"exec_id\": req.ExecID}).Debug(\"exec\")\n\ttask, err := ts.taskManager.Task(req.ID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx = namespaces.WithNamespace(ctx, defaultNamespace)\n\tresp, err := task.Exec(ctx, req)\n\tif err != nil {\n\t\tlog.G(ctx).WithError(err).Error(\"exec failed\")\n\t\treturn nil, err\n\t}\n\n\tlog.G(ctx).Debug(\"exec succeeded\")\n\treturn resp, nil\n}", "func (ts *TaskService) Exec(requestCtx context.Context, req *taskAPI.ExecProcessRequest) (*types.Empty, error) {\n\tdefer logPanicAndDie(log.G(requestCtx))\n\n\ttaskID := req.ID\n\texecID := req.ExecID\n\n\tlogger := log.G(requestCtx).WithField(\"TaskID\", taskID).WithField(\"ExecID\", execID)\n\tlogger.Debug(\"exec\")\n\n\textraData, err := unmarshalExtraData(req.Spec)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to unmarshal extra data\")\n\t}\n\n\t// Just provide runc the options it knows about, not our wrapper\n\treq.Spec = extraData.RuncOptions\n\n\tbundleDir := bundle.Dir(filepath.Join(containerRootDir, taskID))\n\n\tvar ioConnectorSet vm.IOProxy\n\n\tif vm.IsAgentOnlyIO(req.Stdout, logger) {\n\t\tioConnectorSet = vm.NewNullIOProxy()\n\t} else {\n\t\t// Override the incoming stdio FIFOs, which have paths from the host that we can't use\n\t\tfifoSet, err := cio.NewFIFOSetInDir(bundleDir.RootPath(), fifoName(taskID, execID), req.Terminal)\n\t\tif err != nil {\n\t\t\tlogger.WithError(err).Error(\"failed opening stdio FIFOs\")\n\t\t\treturn nil, errors.Wrap(err, \"failed to open stdio FIFOs\")\n\t\t}\n\n\t\tvar stdinConnectorPair *vm.IOConnectorPair\n\t\tif req.Stdin != \"\" {\n\t\t\treq.Stdin = fifoSet.Stdin\n\t\t\tstdinConnectorPair = &vm.IOConnectorPair{\n\t\t\t\tReadConnector: vm.VSockAcceptConnector(extraData.StdinPort),\n\t\t\t\tWriteConnector: vm.FIFOConnector(fifoSet.Stdin),\n\t\t\t}\n\t\t}\n\n\t\tvar stdoutConnectorPair *vm.IOConnectorPair\n\t\tif req.Stdout != \"\" {\n\t\t\treq.Stdout = fifoSet.Stdout\n\t\t\tstdoutConnectorPair = &vm.IOConnectorPair{\n\t\t\t\tReadConnector: vm.FIFOConnector(fifoSet.Stdout),\n\t\t\t\tWriteConnector: vm.VSockAcceptConnector(extraData.StdoutPort),\n\t\t\t}\n\t\t}\n\n\t\tvar stderrConnectorPair *vm.IOConnectorPair\n\t\tif req.Stderr != \"\" {\n\t\t\treq.Stderr = fifoSet.Stderr\n\t\t\tstderrConnectorPair = &vm.IOConnectorPair{\n\t\t\t\tReadConnector: vm.FIFOConnector(fifoSet.Stderr),\n\t\t\t\tWriteConnector: vm.VSockAcceptConnector(extraData.StderrPort),\n\t\t\t}\n\t\t}\n\n\t\tioConnectorSet = vm.NewIOConnectorProxy(stdinConnectorPair, stdoutConnectorPair, stderrConnectorPair)\n\t}\n\n\tresp, err := ts.taskManager.ExecProcess(requestCtx, req, ts.runcService, ioConnectorSet)\n\tif err != nil {\n\t\tlogger.WithError(err).Error(\"exec failed\")\n\t\treturn nil, err\n\t}\n\n\tlogger.Debug(\"exec succeeded\")\n\treturn resp, nil\n}", "func ExecTaskDirectly(paramTask, dbTask *OwlTask) error {\n\tstartId, err := getExecStartId(paramTask.Action, dbTask.ExecItems, paramTask.ExecItem)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// mean need't exec task\n\tif startId < 0 {\n\t\treturn refreshTaskStatus(paramTask.ID, 0, 0, \"\", \"\")\n\t}\n\n\t//exec task\n\tjump := true\n\tfailed := false\n\tbeginTime := time.Now().Unix()\n\tfor _, subTask := range dbTask.SubTasks {\n\t\tdbInfo, err := dbTool.GetDBConn(subTask.DbName, subTask.ClusterName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tfor _, item := range subTask.ExecItems {\n\t\t\tif item.ID != startId && jump {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tjump = false\n\n\t\t\terr := BackupAndExec(dbInfo.DB, &item, subTask.TaskType)\n\t\t\tif err != nil {\n\t\t\t\tfailed = true\n\t\t\t\terr := refreshTaskStatus(paramTask.ID, beginTime, 0, paramTask.Executor, err.Error())\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogger.Errorf(\"after exec failed, update task status to failed err, err: %s\", err.Error())\n\t\t\t\t}\n\n\t\t\t\t// stop all task\n\t\t\t\tdbInfo.CloseConn()\n\t\t\t\tgoto Failed\n\t\t\t}\n\t\t}\n\n\t\tdbInfo.CloseConn()\n\t}\n\nFailed:\n\n\tif !failed {\n\t\terr = refreshTaskStatus(paramTask.ID, beginTime, time.Now().Unix(), paramTask.Executor, \"\")\n\t\tif err != nil {\n\t\t\tlogger.Errorf(\"after exec, update task status to success err, err: %s\", err.Error())\n\t\t}\n\t}\n\n\treturn nil\n}", "func (ft *CustomTask) Exec(t *f.TaskNode, p *f.Params, out *io.PipeWriter) {\n\tglog.Info(\"executing custom task \", p.Complete)\n\n\tft.customFunc(t, p, out)\n\n\treturn\n}", "func (h *DriverHandle) Exec(timeout time.Duration, cmd string, args []string) ([]byte, int, error) {\n\tcommand := append([]string{cmd}, args...)\n\tres, err := h.driver.ExecTask(h.taskID, command, timeout)\n\tif err != nil {\n\t\treturn nil, 0, err\n\t}\n\treturn res.Stdout, res.ExitResult.ExitCode, res.ExitResult.Err\n}", "func (c *Module) Exec(ns string, containerID string, timeout time.Duration, args ...string) error {\n\tvar timedout bool\n\tclient, err := containerd.New(c.containerd)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer client.Close()\n\tctx := namespaces.WithNamespace(context.Background(), ns)\n\tcreatectx, cancel := context.WithTimeout(ctx, timeout)\n\tdefer cancel()\n\n\tcontainer, err := client.LoadContainer(createctx, containerID)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"couldn't load container\")\n\t}\n\tt, err := container.Task(createctx, nil)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to create task\")\n\t}\n\tvar p specs.Process\n\tp.Cwd = \"/\"\n\tp.Args = args\n\ttaskID, err := uuid.NewUUID()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to generate a uuid\")\n\t}\n\tpr, err := t.Exec(createctx, taskID.String(), &p, cio.NullIO)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to exec new porcess\")\n\t}\n\tif err := pr.Start(createctx); err != nil {\n\t\treturn errors.Wrap(err, \"failed to start process\")\n\t}\n\tch, err := pr.Wait(createctx)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to wait for process\")\n\t}\n\tselect {\n\tcase <-ch:\n\tcase <-createctx.Done():\n\t}\n\tdeleteCtx, cancel := context.WithTimeout(ctx, 2*time.Second)\n\tdefer cancel()\n\t// if still running => execution timedout\n\tst, err := pr.Status(deleteCtx)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"couldn't check task state\")\n\t}\n\tif st.Status != containerd.Stopped {\n\t\ttimedout = true\n\t}\n\tex, err := pr.Delete(deleteCtx, containerd.WithProcessKill)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"error deleting the created task\")\n\t}\n\tif timedout {\n\t\treturn errors.New(\"execution timed out\")\n\t} else if ex.ExitCode() != 0 {\n\t\treturn fmt.Errorf(\"non-zero exit code: %d\", ex.ExitCode())\n\t}\n\treturn nil\n}", "func (inst *IndependentInstance) execTask(behavior model.TaskBehavior, taskInst *TaskInst) {\n\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\n\t\t\terr := fmt.Errorf(\"Unhandled Error executing task '%s' : %v\", taskInst.task.Name(), r)\n\t\t\tlogger.Error(err)\n\n\t\t\t// todo: useful for debugging\n\t\t\tlogger.Errorf(\"StackTrace: %s\", debug.Stack())\n\n\t\t\tif !taskInst.flowInst.isHandlingError {\n\n\t\t\t\ttaskInst.appendErrorData(NewActivityEvalError(taskInst.task.Name(), \"unhandled\", err.Error()))\n\t\t\t\tinst.HandleGlobalError(taskInst.flowInst, err)\n\t\t\t}\n\t\t\t// else what should we do?\n\t\t}\n\t}()\n\n\tvar err error\n\n\tvar evalResult model.EvalResult\n\n\tif taskInst.status == model.TaskStatusWaiting {\n\n\t\tevalResult, err = behavior.PostEval(taskInst)\n\n\t} else {\n\t\tevalResult, err = behavior.Eval(taskInst)\n\t}\n\n\tif err != nil {\n\t\ttaskInst.returnError = err\n\t\tinst.handleTaskError(behavior, taskInst, err)\n\t\treturn\n\t}\n\n\tswitch evalResult {\n\tcase model.EVAL_DONE:\n\t\ttaskInst.SetStatus(model.TaskStatusDone)\n\t\tinst.handleTaskDone(behavior, taskInst)\n\tcase model.EVAL_SKIP:\n\t\ttaskInst.SetStatus(model.TaskStatusSkipped)\n\t\tinst.handleTaskDone(behavior, taskInst)\n\tcase model.EVAL_WAIT:\n\t\ttaskInst.SetStatus(model.TaskStatusWaiting)\n\tcase model.EVAL_FAIL:\n\t\ttaskInst.SetStatus(model.TaskStatusFailed)\n\tcase model.EVAL_REPEAT:\n\t\ttaskInst.SetStatus(model.TaskStatusReady)\n\t\t//task needs to iterate or retry\n\t\tinst.scheduleEval(taskInst)\n\t}\n}", "func (p *Pool) Exec(task Task) {\n\tp.tasks <- task\n}", "func (g *GoTasks) Exec(name string, opt ...ExecOption) error {\n\ttask, ok := g.tasks[name]\n\tif !ok {\n\t\treturn errTaskNotFound\n\t}\n\n\toptions := func(o []ExecOption) *Option {\n\t\topts := &Option{}\n\t\tfor _, o := range opt {\n\t\t\to(opts)\n\t\t}\n\t\treturn opts\n\t}(opt)\n\tctx, cancel := makeContext(options)\n\tdefer cancel()\n\tif options.usePool {\n\t\tpoolExec(options.numGoroutines, task)\n\t\treturn nil\n\t}\n\tsingleExec(ctx, task, options)\n\treturn nil\n}", "func (r *remoteRuntimeService) Exec(ctx context.Context, req *runtimeapi.ExecRequest) (*runtimeapi.ExecResponse, error) {\n\tklog.V(10).InfoS(\"[RemoteRuntimeService] Exec\", \"timeout\", r.timeout)\n\tctx, cancel := context.WithTimeout(ctx, r.timeout)\n\tdefer cancel()\n\n\treturn r.execV1(ctx, req)\n}", "func (pge *PgEngine) ExecuteSQLTask(ctx context.Context, tx pgx.Tx, task *ChainTask, paramValues []string) (out string, err error) {\n\tvar execTx pgx.Tx\n\tvar remoteDb PgxConnIface\n\tvar executor executor\n\n\texecTx = tx\n\tif task.Autonomous {\n\t\texecutor = pge.ConfigDb\n\t} else {\n\t\texecutor = tx\n\t}\n\n\t//Connect to Remote DB\n\tif task.ConnectString.Status != pgtype.Null {\n\t\tremoteDb, execTx, err = pge.GetRemoteDBTransaction(ctx, task.ConnectString.String)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tif task.Autonomous {\n\t\t\texecutor = remoteDb\n\t\t\t_ = execTx.Rollback(ctx)\n\t\t} else {\n\t\t\texecutor = execTx\n\t\t}\n\n\t\tdefer pge.FinalizeRemoteDBConnection(ctx, remoteDb)\n\t}\n\n\t// Set Role\n\tif task.RunAs.Status != pgtype.Null && !task.Autonomous {\n\t\tpge.SetRole(ctx, execTx, task.RunAs)\n\t}\n\n\tif task.IgnoreError && !task.Autonomous {\n\t\tpge.MustSavepoint(ctx, execTx, fmt.Sprintf(\"task_%d\", task.TaskID))\n\t}\n\n\tout, err = pge.ExecuteSQLCommand(ctx, executor, task.Script, paramValues)\n\n\tif err != nil && task.IgnoreError && !task.Autonomous {\n\t\tpge.MustRollbackToSavepoint(ctx, execTx, fmt.Sprintf(\"task_%d\", task.TaskID))\n\t}\n\n\t//Reset The Role\n\tif task.RunAs.Status != pgtype.Null && !task.Autonomous {\n\t\tpge.ResetRole(ctx, execTx)\n\t}\n\n\t// Commit changes on remote server\n\tif task.ConnectString.Status != pgtype.Null && !task.Autonomous {\n\t\tpge.MustCommitTransaction(ctx, execTx)\n\t}\n\n\treturn\n}", "func (t *task) Exec(ctx context.Context, processID string, spec *specs.Process, withStdin bool, attachStdio libcontainerdtypes.StdioCallback) (libcontainerdtypes.Process, error) {\n\thcsContainer, err := t.getHCSContainer()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlogger := t.ctr.client.logger.WithFields(log.Fields{\n\t\t\"container\": t.ctr.id,\n\t\t\"exec\": processID,\n\t})\n\n\t// Note we always tell HCS to\n\t// create stdout as it's required regardless of '-i' or '-t' options, so that\n\t// docker can always grab the output through logs. We also tell HCS to always\n\t// create stdin, even if it's not used - it will be closed shortly. Stderr\n\t// is only created if it we're not -t.\n\tcreateProcessParms := &hcsshim.ProcessConfig{\n\t\tCreateStdInPipe: true,\n\t\tCreateStdOutPipe: true,\n\t\tCreateStdErrPipe: !spec.Terminal,\n\t}\n\tif spec.Terminal {\n\t\tcreateProcessParms.EmulateConsole = true\n\t\tif spec.ConsoleSize != nil {\n\t\t\tcreateProcessParms.ConsoleSize[0] = uint(spec.ConsoleSize.Height)\n\t\t\tcreateProcessParms.ConsoleSize[1] = uint(spec.ConsoleSize.Width)\n\t\t}\n\t}\n\n\t// Take working directory from the process to add if it is defined,\n\t// otherwise take from the first process.\n\tif spec.Cwd != \"\" {\n\t\tcreateProcessParms.WorkingDirectory = spec.Cwd\n\t} else {\n\t\tcreateProcessParms.WorkingDirectory = t.ctr.ociSpec.Process.Cwd\n\t}\n\n\t// Configure the environment for the process\n\tcreateProcessParms.Environment = setupEnvironmentVariables(spec.Env)\n\n\t// Configure the CommandLine/CommandArgs\n\tsetCommandLineAndArgs(spec, createProcessParms)\n\tlogger.Debugf(\"exec commandLine: %s\", createProcessParms.CommandLine)\n\n\tcreateProcessParms.User = spec.User.Username\n\n\t// Start the command running in the container.\n\tnewProcess, err := hcsContainer.CreateProcess(createProcessParms)\n\tif err != nil {\n\t\tlogger.WithError(err).Errorf(\"exec's CreateProcess() failed\")\n\t\treturn nil, err\n\t}\n\tpid := newProcess.Pid()\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tif err := newProcess.Kill(); err != nil {\n\t\t\t\tlogger.WithError(err).Error(\"failed to kill process\")\n\t\t\t}\n\t\t\tgo func() {\n\t\t\t\tif err := newProcess.Wait(); err != nil {\n\t\t\t\t\tlogger.WithError(err).Error(\"failed to wait for process\")\n\t\t\t\t}\n\t\t\t\tif err := newProcess.Close(); err != nil {\n\t\t\t\t\tlogger.WithError(err).Error(\"failed to clean process resources\")\n\t\t\t\t}\n\t\t\t}()\n\t\t}\n\t}()\n\n\tdio, err := newIOFromProcess(newProcess, spec.Terminal)\n\tif err != nil {\n\t\tlogger.WithError(err).Error(\"failed to get stdio pipes\")\n\t\treturn nil, err\n\t}\n\t// Tell the engine to attach streams back to the client\n\t_, err = attachStdio(dio)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tp := &process{\n\t\tid: processID,\n\t\tctr: t.ctr,\n\t\thcsProcess: newProcess,\n\t\twaitCh: make(chan struct{}),\n\t}\n\n\t// Spin up a goroutine to notify the backend and clean up resources when\n\t// the process exits. Defer until after the start event is sent so that\n\t// the exit event is not sent out-of-order.\n\tdefer func() { go p.reap() }()\n\n\tt.ctr.client.eventQ.Append(t.ctr.id, func() {\n\t\tei := libcontainerdtypes.EventInfo{\n\t\t\tContainerID: t.ctr.id,\n\t\t\tProcessID: p.id,\n\t\t\tPid: uint32(pid),\n\t\t}\n\t\tt.ctr.client.logger.WithFields(log.Fields{\n\t\t\t\"container\": t.ctr.id,\n\t\t\t\"event\": libcontainerdtypes.EventExecAdded,\n\t\t\t\"event-info\": ei,\n\t\t}).Info(\"sending event\")\n\t\terr := t.ctr.client.backend.ProcessEvent(t.ctr.id, libcontainerdtypes.EventExecAdded, ei)\n\t\tif err != nil {\n\t\t\tt.ctr.client.logger.WithError(err).WithFields(log.Fields{\n\t\t\t\t\"container\": t.ctr.id,\n\t\t\t\t\"event\": libcontainerdtypes.EventExecAdded,\n\t\t\t\t\"event-info\": ei,\n\t\t\t}).Error(\"failed to process event\")\n\t\t}\n\t\terr = t.ctr.client.backend.ProcessEvent(t.ctr.id, libcontainerdtypes.EventExecStarted, ei)\n\t\tif err != nil {\n\t\t\tt.ctr.client.logger.WithError(err).WithFields(log.Fields{\n\t\t\t\t\"container\": t.ctr.id,\n\t\t\t\t\"event\": libcontainerdtypes.EventExecStarted,\n\t\t\t\t\"event-info\": ei,\n\t\t\t}).Error(\"failed to process event\")\n\t\t}\n\t})\n\n\treturn p, nil\n}", "func (cc *Context) Exec(nodeList map[string]*nodes.Node, actions []string, options ...ExecOption) error {\n\t// init the exec context and logging\n\tec := &execContext{\n\t\tContext: cc,\n\t\tnodes: nodeList,\n\t}\n\n\tec.status = logutil.NewStatus(os.Stdout)\n\tec.status.MaybeWrapLogrus(log.StandardLogger())\n\n\tdefer ec.status.End(false)\n\n\t// apply exec options\n\tfor _, option := range options {\n\t\toption(ec)\n\t}\n\n\t// Create an ExecutionPlan that applies the given actions to the\n\t// topology defined in the config\n\texecutionPlan, err := newExecutionPlan(ec.DerivedConfig, actions)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Executes all the selected action\n\tfor _, plannedTask := range executionPlan {\n\t\tec.status.Start(fmt.Sprintf(\"[%s] %s\", plannedTask.Node.Name, plannedTask.Task.Description))\n\n\t\terr := plannedTask.Task.Run(ec, plannedTask.Node)\n\t\tif err != nil {\n\t\t\t// in case of error, the execution plan is halted\n\t\t\tlog.Error(err)\n\t\t\treturn err\n\t\t}\n\t}\n\tec.status.End(true)\n\n\treturn nil\n}", "func (ft *LsTask) Exec(ctx *Context, p *par.Params, out *io.PipeWriter) {\n\tlog.Info(\"LsTask.Execute\")\n\n\tpath, ok := p.Props[\"path\"]\n\n\t// if no passed in path use default\n\tif !ok {\n\t\tpath = ft.path\n\t}\n\n\tif path == \"\" {\n\t\tp.Status = par.StFail\n\t\tp.Response = \"no path specified\"\n\t\treturn\n\t}\n\n\t// this is mandatory node\n\tpath = filepath.Join(ctx.WorkspacePath, path)\n\n\tfiles, _ := ioutil.ReadDir(path)\n\tfor _, f := range files {\n\t\tp.Props[fmt.Sprint(f.Name())] = \"\"\n\t}\n\n\tp.Response = \"list directory done\"\n\tp.Status = par.StSuccess\n\n\treturn\n}", "func (m *Manager) Exec(name string, opt ExecOptions, gOpt operator.Options) error {\n\tif err := clusterutil.ValidateClusterNameOrError(name); err != nil {\n\t\treturn err\n\t}\n\n\tmetadata, err := m.meta(name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttopo := metadata.GetTopology()\n\tbase := metadata.GetBaseMeta()\n\n\tfilterRoles := set.NewStringSet(gOpt.Roles...)\n\tfilterNodes := set.NewStringSet(gOpt.Nodes...)\n\n\tvar shellTasks []task.Task\n\tuniqueHosts := map[string]set.StringSet{} // host-sshPort -> {command}\n\ttopo.IterInstance(func(inst spec.Instance) {\n\t\tkey := utils.JoinHostPort(inst.GetManageHost(), inst.GetSSHPort())\n\t\tif _, found := uniqueHosts[key]; !found {\n\t\t\tif len(gOpt.Roles) > 0 && !filterRoles.Exist(inst.Role()) {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif len(gOpt.Nodes) > 0 && (!filterNodes.Exist(inst.GetHost()) && !filterNodes.Exist(inst.GetManageHost())) {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tcmds, err := renderInstanceSpec(opt.Command, inst)\n\t\t\tif err != nil {\n\t\t\t\tm.logger.Debugf(\"error rendering command with spec: %s\", err)\n\t\t\t\treturn // skip\n\t\t\t}\n\t\t\tcmdSet := set.NewStringSet(cmds...)\n\t\t\tif _, ok := uniqueHosts[key]; ok {\n\t\t\t\tuniqueHosts[key].Join(cmdSet)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tuniqueHosts[key] = cmdSet\n\t\t}\n\t})\n\n\tfor hostKey, i := range uniqueHosts {\n\t\thost, _ := utils.ParseHostPort(hostKey)\n\t\tfor _, cmd := range i.Slice() {\n\t\t\tshellTasks = append(shellTasks,\n\t\t\t\ttask.NewBuilder(m.logger).\n\t\t\t\t\tShell(host, cmd, hostKey+cmd, opt.Sudo).\n\t\t\t\t\tBuild())\n\t\t}\n\t}\n\n\tb, err := m.sshTaskBuilder(name, topo, base.User, gOpt)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tt := b.\n\t\tParallel(false, shellTasks...).\n\t\tBuild()\n\n\texecCtx := ctxt.New(\n\t\tcontext.Background(),\n\t\tgOpt.Concurrency,\n\t\tm.logger,\n\t)\n\tif err := t.Execute(execCtx); err != nil {\n\t\tif errorx.Cast(err) != nil {\n\t\t\t// FIXME: Map possible task errors and give suggestions.\n\t\t\treturn err\n\t\t}\n\t\treturn perrs.Trace(err)\n\t}\n\n\t// print outputs\n\tfor hostKey, i := range uniqueHosts {\n\t\thost, _ := utils.ParseHostPort(hostKey)\n\t\tfor _, cmd := range i.Slice() {\n\t\t\tstdout, stderr, ok := ctxt.GetInner(execCtx).GetOutputs(hostKey + cmd)\n\t\t\tif !ok {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tm.logger.Infof(\"Outputs of %s on %s:\",\n\t\t\t\tcolor.CyanString(cmd),\n\t\t\t\tcolor.CyanString(host))\n\t\t\tif len(stdout) > 0 {\n\t\t\t\tm.logger.Infof(\"%s:\\n%s\", color.GreenString(\"stdout\"), stdout)\n\t\t\t}\n\t\t\tif len(stderr) > 0 {\n\t\t\t\tm.logger.Infof(\"%s:\\n%s\", color.RedString(\"stderr\"), stderr)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func runExec(serviceName string, operation string) (string, error) {\n\tbytes, err := exec.Command(Configuration.ExecutorPath, serviceName, operation).CombinedOutput()\n\treturn string(bytes), err\n}", "func (d *dispatcher) ExecuteTask() {\n\tlogutil.BgLogger().Info(\"execute one task\", zap.Int64(\"task ID\", d.task.ID),\n\t\tzap.String(\"state\", d.task.State), zap.Uint64(\"concurrency\", d.task.Concurrency))\n\td.scheduleTask(d.task.ID)\n}", "func (task *Task) Execute(t time.Time) (exec.Result, error) {\n\n\t//Validate the task\n\tif err := task.Validate(); err != nil {\n\t\treturn exec.Result{}, err\n\t}\n\n\t//Test if the given task should execute in the root croniclePath and the croncilePath is a git repo\n\ttaskPathIsCroniclePathWithGit := (task.Path == task.CroniclePath) && task.CronicleRepo != nil\n\n\t//If a repo is given, clone the repo and task.Git.Open(task.Path)\n\tif task.Repo != nil {\n\t\tauth, err := task.Repo.Auth()\n\t\tif err != nil {\n\t\t\treturn exec.Result{}, err\n\t\t}\n\t\tg, err := Clone(task.Path, task.Repo.URL, &auth)\n\t\t// g, err := Clone(task.Path, task.Repo.URL, task.Repo.DeployKey)\n\t\tif err != nil {\n\t\t\treturn exec.Result{}, err\n\t\t}\n\t\ttask.Git = g\n\t\terr = task.Git.Checkout(task.Repo.Branch, task.Repo.Commit)\n\t\tif err != nil {\n\t\t\treturn exec.Result{}, err\n\t\t}\n\t} else if taskPathIsCroniclePathWithGit {\n\t\tauth, err := task.CronicleRepo.Auth()\n\t\tif err != nil {\n\t\t\treturn exec.Result{}, err\n\t\t}\n\t\ttask.Git, err = Clone(task.CroniclePath, task.CronicleRepo.URL, &auth)\n\t\t// var err error\n\t\t// task.Git, err = Clone(task.CroniclePath, task.CronicleRepo.URL, task.CronicleRepo.DeployKey)\n\t\tif err != nil {\n\t\t\tlog.Error(err)\n\t\t\treturn exec.Result{}, err\n\t\t}\n\t}\n\n\t//Execute task.Command in bash at time t with retry\n\tvar result exec.Result\n\terr := try.Do(func(attempt int) (bool, error) {\n\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"schedule\": task.ScheduleName,\n\t\t\t\"task\": task.Name,\n\t\t\t\"attempt\": attempt,\n\t\t\t\"clock\": t.Format(time.Kitchen),\n\t\t\t\"date\": t.Format(time.RFC850),\n\t\t}).Info(\"Executing...\")\n\t\tvar err error\n\t\tresult = task.Exec(t)\n\t\terr = result.Error\n\t\ttask.Log(result)\n\t\tif err != nil && task.Retry != nil {\n\t\t\tduration := time.Duration(task.Retry.Seconds) * time.Second\n\t\t\tduration += time.Duration(task.Retry.Minutes) * time.Minute\n\t\t\tduration += time.Duration(task.Retry.Hours) * time.Hour\n\t\t\ttime.Sleep(duration)\n\t\t}\n\n\t\tvar retryCount int\n\t\tswitch task.Retry {\n\t\tcase nil:\n\t\t\tretryCount = 0\n\t\tdefault:\n\t\t\tretryCount = task.Retry.Count\n\t\t}\n\n\t\treturn attempt < retryCount, err\n\t})\n\tif err != nil {\n\t\treturn result, err\n\t}\n\n\treturn result, nil\n}", "func (a AmbariRegistry) ExecuteRemoteCommandTask(task Task, filteredHosts map[string]bool) {\n\tif len(task.Command) > 0 {\n\t\tfmt.Println(\"Execute remote command: \" + task.Command)\n\t\ta.RunRemoteHostCommand(task.Command, filteredHosts, task.AmbariServerFilter)\n\t}\n}", "func (p *Pool) Exec(e Task) {\n\tp.tasks <- e\n}", "func (t TaskFunc) Execute() { t() }", "func (t *Task) Exec(agent *Agent) {\n\tdefer func() {\n\t\tif e := recover(); e != nil {\n\n\t\t\t//todo send task status to DCMS-agent\n\t\t\t// log.Warningf(\"run task: %s jobname: failed : %s\", t.TaskId, t.Job.Name, e)\n\t\t\tts := &TaskStatus{\n\t\t\t\tTaskPtr: t,\n\t\t\t\tCommand: nil,\n\t\t\t\tStatus: StatusFailed,\n\t\t\t\tCreateAt: time.Now().Unix(),\n\t\t\t\tErr: fmt.Errorf(\"run task: %s jobname: failed : %s\", t.TaskId, t.Job.Name, e),\n\t\t\t}\n\n\t\t\terrstr := fmt.Sprintf(\"%s\", e)\n\t\t\tif errstr == \"signal: killed\" {\n\t\t\t\tts.Status = StatusKilled\n\t\t\t}\n\t\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\t}\n\t}()\n\n\tvar ts *TaskStatus\n\tvar err error\n\t// log.Info(\"task run Exec function in goroutine\")\n\n\tt.genLogFile()\n\t// check file signature\n\ttmp_md5 := util.Md5File(t.Job.Executor)\n\tif t.Job.Signature != tmp_md5 {\n\t\tts = &TaskStatus{\n\t\t\tTaskPtr: t,\n\t\t\tCommand: nil,\n\t\t\tStatus: StatusFailed,\n\t\t\tCreateAt: time.Now().Unix(),\n\t\t\tErr: fmt.Errorf(\"cronjob: %s executor: %s signature:%s does't match db's sig:%s\", t.Job.Name, t.Job.Executor, tmp_md5, t.Job.Signature),\n\t\t}\n\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\treturn\n\t} else {\n\t\tlog.Info(\"cronjob signature match for \", t.Job.Name, t.Job.ExecutorFlags)\n\t}\n\n\tvar u *user.User\n\tu, err = user.Lookup(t.Job.Runner)\n\tif err != nil {\n\t\t// log.Warningf(\"user %s not exists, task %s quit \", err, t.TaskId)\n\t\tts = &TaskStatus{\n\t\t\tTaskPtr: t,\n\t\t\tCommand: nil,\n\t\t\tStatus: StatusFailed,\n\t\t\tCreateAt: time.Now().Unix(),\n\t\t\tErr: fmt.Errorf(\"user %s not exists, task %s quit \", err, t.TaskId),\n\t\t}\n\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\treturn\n\t}\n\n\tvar uid int\n\tuid, err = strconv.Atoi(u.Uid)\n\tif err != nil {\n\t\t// log.Warningf(\"uid %s conver to int failed \", uid)\n\t\tts = &TaskStatus{\n\t\t\tTaskPtr: t,\n\t\t\tCommand: nil,\n\t\t\tStatus: StatusFailed,\n\t\t\tCreateAt: time.Now().Unix(),\n\t\t\tErr: fmt.Errorf(\"uid %s conver to int failed \", uid),\n\t\t}\n\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\treturn\n\t}\n\n\t// chown log file to specific t.Job.Runner user\n\tif err = t.logfile.Chown(uid, uid); err != nil {\n\t\t// log.Warningf(\"chown logfile: %s to uid: %s failed, %s\", t.logfile.Name(), u.Uid, err)\n\t\tt.logfile = nil\n\t}\n\tvar cmd *exec.Cmd\n\tif t.Job.Executor != \"\" && t.Job.ExecutorFlags != \"\" {\n\t\tcmd = exec.Command(t.Job.Executor, t.Job.ExecutorFlags)\n\t} else if t.Job.Executor != \"\" && t.Job.ExecutorFlags == \"\" {\n\t\tcmd = exec.Command(t.Job.Executor)\n\t} else {\n\t\tts = &TaskStatus{\n\t\t\tTaskPtr: t,\n\t\t\tCommand: cmd,\n\t\t\tStatus: StatusFailed,\n\t\t\tCreateAt: time.Now().Unix(),\n\t\t\tErr: fmt.Errorf(\"job %s must have Executor \", t.Job.Name),\n\t\t}\n\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\treturn\n\t}\n\n\tcmd.SysProcAttr = &syscall.SysProcAttr{}\n\tcmd.SysProcAttr.Credential = &syscall.Credential{Uid: uint32(uid)}\n\tcmd.SysProcAttr.Setsid = true\n\t// Pdeathsig only valid on linux system\n\t//\n\tcmd.SysProcAttr.Pdeathsig = syscall.SIGUSR1\n\n\tcmd.Stderr = t.logfile\n\tcmd.Stdout = t.logfile\n\n\tif err = cmd.Start(); err != nil {\n\t\t// log.Warningf(\"taskid:%s cmd Start failed: %s\", t.TaskId, err)\n\t\tts = &TaskStatus{\n\t\t\tTaskPtr: t,\n\t\t\tCommand: cmd,\n\t\t\tStatus: StatusFailed,\n\t\t\tCreateAt: time.Now().Unix(),\n\t\t\tErr: fmt.Errorf(\"taskid:%s cmd Start failed: %s\", t.TaskId, err),\n\t\t}\n\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\treturn\n\t}\n\n\tts = &TaskStatus{\n\t\tTaskPtr: t,\n\t\tCommand: cmd,\n\t\tStatus: StatusRunning,\n\t\tCreateAt: time.Now().Unix(),\n\t\tErr: nil,\n\t}\n\tt.Job.Dcms.JobStatusChan <- ts\n\t// send cmd.process to dcms-agent\n\n\tif err = cmd.Wait(); err != nil {\n\t\t// log.Warningf(\"taskid:%s cmd Wait failed: %s\", t.TaskId, err)\n\t\tts = &TaskStatus{\n\t\t\tTaskPtr: t,\n\t\t\tCommand: cmd,\n\t\t\tStatus: StatusFailed,\n\t\t\tCreateAt: time.Now().Unix(),\n\t\t\tErr: fmt.Errorf(\"taskid:%s cmd Wait failed: %s\", t.TaskId, err),\n\t\t}\n\t\terrstr := fmt.Sprintf(\"%s\", err.Error())\n\t\tif errstr == \"signal: killed\" {\n\t\t\tts.Status = StatusKilled\n\t\t}\n\t\tt.Job.Dcms.JobStatusChan <- ts\n\t\treturn\n\t}\n\t// log.Warning(\"task run DONE\")\n\tts = &TaskStatus{\n\t\tTaskPtr: t,\n\t\tCommand: cmd,\n\t\tStatus: StatusSuccess,\n\t\tCreateAt: time.Now().Unix(),\n\t\tErr: nil,\n\t}\n\tt.Job.Dcms.JobStatusChan <- ts\n\treturn\n}", "func execTasks(parent context.Context, c TimedActuator,\n\texecFunc func(f func()), tasks ...Task) error {\n\tsize := len(tasks)\n\tif size == 0 {\n\t\treturn nil\n\t}\n\n\tctx, cancel := context.WithCancel(parent)\n\tresChan := make(chan error, size)\n\twg := &sync.WaitGroup{}\n\twg.Add(size)\n\n\t// Make sure the tasks are completed and channel is closed\n\tgo func() {\n\t\twg.Wait()\n\t\tcancel()\n\t\tclose(resChan)\n\t}()\n\n\t// Sadly we can not kill a goroutine manually\n\t// So when an error happens, the other tasks will continue\n\t// But the good news is that main progress\n\t// will know the error immediately\n\tfor _, task := range tasks {\n\t\tchild, _ := context.WithCancel(ctx)\n\t\tf := wrapperTask(child, task, wg, resChan)\n\t\texecFunc(f)\n\t}\n\n\treturn wait(ctx, c, resChan, cancel)\n}", "func (ctx *Context) Exec(cmd []string) *ExecResult {\n\treturn ctx.ExecWithParams(ExecParams{Cmd: cmd})\n}", "func (ne *NSEnter) Exec(cmd string, args []string) exec.Cmd {\n\thostProcMountNsPath := filepath.Join(ne.hostRootFsPath, mountNsPath)\n\tfullArgs := append([]string{fmt.Sprintf(\"--mount=%s\", hostProcMountNsPath), \"--\"},\n\t\tappend([]string{ne.AbsHostPath(cmd)}, args...)...)\n\tklog.V(5).Infof(\"Running nsenter command: %v %v\", nsenterPath, fullArgs)\n\treturn ne.executor.Command(nsenterPath, fullArgs...)\n}", "func (d Adapter) Exec(command []string) error {\n\treturn d.componentAdapter.Exec(command)\n}", "func (k *Kubernetes) Exec(ctx context.Context, containerID string, cmd []string) ([]byte, error) {\n\treturn k.Runtime.Exec(ctx, containerID, cmd)\n}", "func (c *VirtLauncherClient) Exec(domainName, command string, args []string, timeoutSeconds int32) (int, string, error) {\n\trequest := &cmdv1.ExecRequest{\n\t\tDomainName: domainName,\n\t\tCommand: command,\n\t\tArgs: args,\n\t\tTimeoutSeconds: int32(timeoutSeconds),\n\t}\n\texitCode := -1\n\tstdOut := \"\"\n\n\tctx, cancel := context.WithTimeout(\n\t\tcontext.Background(),\n\t\t// we give the context a bit more time as the timeout should kick\n\t\t// on the actual execution\n\t\ttime.Duration(timeoutSeconds)*time.Second+shortTimeout,\n\t)\n\tdefer cancel()\n\n\tresp, err := c.v1client.Exec(ctx, request)\n\tif resp == nil {\n\t\treturn exitCode, stdOut, err\n\t}\n\n\texitCode = int(resp.ExitCode)\n\tstdOut = resp.StdOut\n\n\treturn exitCode, stdOut, err\n}", "func Exec(config *ssh.ClientConfig, addr string, workDir string, cmd string, nixConf string) (bytes.Buffer, error) {\n\tvar b bytes.Buffer // import \"bytes\"\n\n\t// Connect\n\tclient, err := ssh.Dial(\"tcp\", net.JoinHostPort(addr, \"22\"), config)\n\tif err != nil {\n\t\treturn b, err\n\t}\n\t// Create a session. It is one session per command.\n\tsession, err := client.NewSession()\n\tif err != nil {\n\t\treturn b, err\n\t}\n\tdefer session.Close()\n\n\tsession.Stderr = os.Stderr // get output\n\tsession.Stdout = &b // get output\n\t// you can also pass what gets input to the stdin, allowing you to pipe\n\t// content from client to server\n\t// session.Stdin = bytes.NewBufferString(\"My input\")\n\n\t// Finally, run the command\n\tfullCmd := \". ~/.nix-profile/etc/profile.d/nix.sh && cd \" + workDir + \" && nix-shell \" + nixConf + \" --command '\" + cmd + \"'\"\n\tfmt.Println(fullCmd)\n\terr = session.Run(fullCmd)\n\treturn b, err\n}", "func (a *AGI) Exec(cmd ...string) (string, error) {\n\tcmd = append([]string{\"EXEC\"}, cmd...)\n\treturn a.Command(cmd...).Val()\n}", "func Exec() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"exec COMMAND\",\n\t\tShort: \"Execute a command in the cloud native environment\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tanalytics.Send(analytics.EventExec, GetActionID())\n\t\t\tdefer analytics.Send(analytics.EventExecEnd, GetActionID())\n\t\t\treturn executeExec(args)\n\t\t},\n\t\tArgs: func(cmd *cobra.Command, args []string) error {\n\t\t\tif len(args) < 1 {\n\t\t\t\treturn errors.New(\"exec requires the COMMAND argument\")\n\t\t\t}\n\n\t\t\treturn nil\n\t\t},\n\t}\n\n\treturn cmd\n}", "func Exec(client *Client, args []string, timeoutSecs int) (*pb.ExecResult, error) {\n\tctx, cancel := context.WithTimeout(context.Background(), time.Duration(timeoutSecs)*time.Second)\n\tdefer cancel()\n\n\trequest := &pb.ExecRequest{\n\t\tExecutable: args[0],\n\t\tArgs: args[1:],\n\t}\n\n\treturn client.Exec(ctx, request)\n}", "func (r RealExecute) ExecCommand(com string, args ...string) ([]byte, error) {\n\t/* #nosec */\n\tcommand := exec.Command(com, args...)\n\treturn command.CombinedOutput()\n}", "func execTask(ctx context.Context) error {\n\t// Do pseudo-task. Here, it is just a \"sleep\".\n\tn := 500 + rand.Intn(3500)\n\ttimer := time.NewTimer(time.Duration(n) * time.Millisecond)\n\tselect {\n\tcase <-ctx.Done():\n\t\t// Cancel the pseudo-task here.\n\t\ttimer.Stop()\n\t\treturn ctx.Err()\n\tcase <-timer.C:\n\t\t// Do nothing here. Proceed to the following code\n\t}\n\n\t// Return result of the task. Here, failure means the random number is a\n\t// multiples of 9.\n\tif (n % 9) == 0 {\n\t\treturn errors.New(\"bad luck\")\n\t}\n\treturn nil\n}", "func TestRktDriver_Exec(t *testing.T) {\n\tctestutil.RktCompatible(t)\n\tif !testutil.IsCI() {\n\t\tt.Parallel()\n\t}\n\n\trequire := require.New(t)\n\td := NewRktDriver(testlog.HCLogger(t))\n\tharness := dtestutil.NewDriverHarness(t, d)\n\n\ttask := &drivers.TaskConfig{\n\t\tID: uuid.Generate(),\n\t\tAllocID: uuid.Generate(),\n\t\tName: \"etcd\",\n\t\tResources: &drivers.Resources{\n\t\t\tNomadResources: &structs.AllocatedTaskResources{\n\t\t\t\tMemory: structs.AllocatedMemoryResources{\n\t\t\t\t\tMemoryMB: 128,\n\t\t\t\t},\n\t\t\t\tCpu: structs.AllocatedCpuResources{\n\t\t\t\t\tCpuShares: 100,\n\t\t\t\t},\n\t\t\t},\n\t\t\tLinuxResources: &drivers.LinuxResources{\n\t\t\t\tMemoryLimitBytes: 134217728,\n\t\t\t\tCPUShares: 100,\n\t\t\t},\n\t\t},\n\t}\n\n\ttc := &TaskConfig{\n\t\tTrustPrefix: \"coreos.com/etcd\",\n\t\tImageName: \"coreos.com/etcd:v2.0.4\",\n\t\tNet: []string{\"none\"},\n\t}\n\trequire.NoError(task.EncodeConcreteDriverConfig(&tc))\n\ttesttask.SetTaskConfigEnv(task)\n\n\tcleanup := harness.MkAllocDir(task, true)\n\tdefer cleanup()\n\n\t_, _, err := harness.StartTask(task)\n\trequire.NoError(err)\n\n\t// Run command that should succeed\n\texpected := []byte(\"etcd version\")\n\ttestutil.WaitForResult(func() (bool, error) {\n\t\tres, err := d.ExecTask(task.ID, []string{\"/etcd\", \"--version\"}, time.Second)\n\t\tif err != nil {\n\t\t\treturn false, fmt.Errorf(\"failed to exec: %#v\", err)\n\t\t}\n\t\tif !res.ExitResult.Successful() {\n\t\t\treturn false, fmt.Errorf(\"/etcd --version failed: %#v %#v\", res.ExitResult, res)\n\t\t}\n\t\traw := res.Stdout\n\t\treturn bytes.Contains(raw, expected), fmt.Errorf(\"expected %q but found:\\n%s\", expected, raw)\n\t}, func(err error) {\n\t\trequire.NoError(err)\n\t})\n\n\t// Run command that should fail\n\texpected = []byte(\"flag provided but not defined\")\n\ttestutil.WaitForResult(func() (bool, error) {\n\t\tres, err := d.ExecTask(task.ID, []string{\"/etcd\", \"--cgdfgdfg\"}, time.Second)\n\t\tif err != nil {\n\t\t\treturn false, fmt.Errorf(\"failed to exec: %#v\", err)\n\t\t}\n\t\tif res.ExitResult.Successful() {\n\t\t\treturn false, fmt.Errorf(\"/etcd --cgdfgdfg unexpected succeeded: %#v %#v\", res.ExitResult, res)\n\t\t}\n\t\traw := res.Stdout\n\t\treturn bytes.Contains(raw, expected), fmt.Errorf(\"expected %q but found:\\n%s\", expected, raw)\n\t}, func(err error) {\n\t\trequire.NoError(err)\n\t})\n\n\trequire.NoError(harness.DestroyTask(task.ID, true))\n}", "func (pm *Manager) Exec(desc, cmdName string, args ...string) (string, string, error) {\n\treturn pm.ExecDir(-1, \"\", desc, cmdName, args...)\n}", "func Exec(t testing.TB, cmd *cobra.Command, stdIn io.Reader, args ...string) (string, string, error) {\n\tctx, cancel := context.WithCancel(context.Background())\n\tt.Cleanup(cancel)\n\n\treturn ExecCtx(ctx, cmd, stdIn, args...)\n}", "func Exec(ctx Context, executorID ExecutorID, fn func() error) error {\n\tif err := ctx.Acquire(executorID); err != nil {\n\t\treturn err\n\t}\n\n\treturn fn()\n}", "func execRunCommand(c *cli.Context) {\n\tif len(c.Args()) < 1 {\n\t\tdisplayError(errMissingTask, 64)\n\t}\n\n\tclient := newEtcdMinionClientFromFlags(c)\n\n\tcFlag := c.String(\"with-classifier\")\n\tminions, err := parseClassifierPattern(client, cFlag)\n\n\tif err != nil {\n\t\tdisplayError(err, 1)\n\t}\n\n\tnumMinions := len(minions)\n\tif numMinions == 0 {\n\t\tdisplayError(errNoMinionFound, 1)\n\t}\n\n\tfmt.Printf(\"Found %d minion(s) for task processing\\n\\n\", numMinions)\n\n\t// Create the task that we send to our minions\n\t// The first argument is the command and anything else\n\t// that follows is considered task arguments\n\targs := c.Args()\n\tisConcurrent := c.Bool(\"is-concurrent\")\n\ttaskCommand := args[0]\n\ttaskArgs := args[1:]\n\tt := task.New(taskCommand, taskArgs...)\n\tt.IsConcurrent = isConcurrent\n\n\t// Progress bar to display while submitting task\n\tprogress := uiprogress.New()\n\tbar := progress.AddBar(numMinions)\n\tbar.AppendCompleted()\n\tbar.PrependElapsed()\n\tprogress.Start()\n\n\t// Number of minions to which submitting the task has failed\n\tfailed := 0\n\n\t// Submit task to minions\n\tfmt.Println(\"Submitting task to minion(s) ...\")\n\tfor _, minion := range minions {\n\t\terr = client.MinionSubmitTask(minion, t)\n\t\tif err != nil {\n\t\t\tfmt.Printf(\"Failed to submit task to %s: %s\\n\", minion, err)\n\t\t\tfailed += 1\n\t\t}\n\t\tbar.Incr()\n\t}\n\n\t// Stop progress bar and sleep for a bit to make sure the\n\t// progress bar gets updated if we were too fast for it\n\tprogress.Stop()\n\ttime.Sleep(time.Millisecond * 100)\n\n\t// Display task report\n\tfmt.Println()\n\ttable := uitable.New()\n\ttable.MaxColWidth = 80\n\ttable.Wrap = true\n\ttable.AddRow(\"TASK\", \"SUBMITTED\", \"FAILED\", \"TOTAL\")\n\ttable.AddRow(t.TaskID, numMinions-failed, failed, numMinions)\n\tfmt.Println(table)\n}", "func (ic *Context) Exec() error {\n\tdefer ic.Finalize()\n\treturn ic.VM.Run()\n}", "func (b *TaskExecBuilder) NewTaskExec(name string) *TaskExecBuilder {\n\treturn b.Commit().NewTaskExec(name)\n}", "func (s *RPCClient) Exec(command ...interface{}) (*Response, error) {\n\tvar res Response\n\terr := s.client.Call(\"RPCServer.Exec\", &command, &res)\n\treturn &res, err\n}", "func (ws *Workspace) ExecFunction(task *Task, name string, args ...string) string {\n\tlog.Info(\"Executing function %s: %s\", name, args)\n\tvar fn *Function\n\tif f, ok := ws.Functions[name]; ok {\n\t\tfn = f\n\t} else if f, ok := GlobalWorkspace.Functions[name]; ok {\n\t\tfn = f\n\t} else {\n\t\tlog.Warn(\"Function not found: %s\", name)\n\t\treturn \"\"\n\t}\n\n\targmap := make(map[string]string)\n\tfor i, arg := range fn.Args {\n\t\targmap[arg] = args[i]\n\t}\n\n\tfor k, v := range argmap {\n\t\tlog.Info(\"argmap: %s => %s\", k, v)\n\t\tfor t, m := range task.Metadata {\n\t\t\tlog.Info(\"meta: %s => %s\", t, m)\n\t\t\tv = strings.Replace(v, \"$\"+t, m, -1)\n\t\t}\n\t\targmap[k] = v\n\t}\n\n\tc := fn.Command\n\tfor k, v := range argmap {\n\t\tlog.Info(\"ARG: %s => %s\", k, v)\n\t\tc = strings.Replace(c, k, v, -1)\n\t}\n\n\tvar funcEnvironment map[string]string\n\tif ws.InheritEnvironment {\n\t\tfuncEnvironment = ws.Environment\n\t} else if GlobalWorkspace.InheritEnvironment {\n\t\tfuncEnvironment = GlobalWorkspace.Environment\n\t} else {\n\t\tfuncEnvironment = make(map[string]string)\n\t}\n\n\ttsk := NewTask(nil, \"Function$\"+name, fn.Executor, c, funcEnvironment, false, \"\", \"\", make(map[string]string), \"\")\n\tch := tsk.Start()\n\t<-ch\n\treturn tsk.TaskRuns[0].StdoutBuf.String()\n}", "func WrapExec(cmd string, args []String, nArg uint32) (status syscall.Status){\n\n\n\tpath := \"/programs/\"+cmd\n\n\tif nArg == 0 {\n\n\t\tstatus = altEthos.Exec(path)\n\n\t} else if nArg == 1 {\n\n\t\tstatus = altEthos.Exec(path, &args[0])\n\n\t} else if nArg == 2 {\n\n\t\tstatus = altEthos.Exec(path, &args[0], &args[1])\n\n\t} else if nArg == 3 {\n\n\t\tstatus = altEthos.Exec(path, &args[0], &args[1], &args[2])\n\n\t} else if nArg == 4 {\n\n\t\tstatus = altEthos.Exec(path, &args[0], &args[1], &args[2], &args[3])\n\n\t}\n\n\treturn\n\n}", "func (f *RemoteRuntime) Exec(ctx context.Context, req *kubeapi.ExecRequest) (*kubeapi.ExecResponse, error) {\n\treturn f.RuntimeService.Exec(ctx, req)\n}", "func (client *Client) Exec(val interface{}, command, key string, args ...interface{}) error {\n\treturn client.Pool.Do(radix.FlatCmd(val, command, key, args...))\n}", "func (k *kubectlContext) Exec(podName string, cmd ...string) (string, error) {\n\tout, err := k.do(append([]string{\"exec\", podName}, cmd...)...)\n\treturn string(out), err\n}", "func (c *MigrateCommand) Exec(ctx context.Context, req *proto.ExecRequest, rsp *proto.ExecResponse) error {\n\t// rsp.Error could be set to return an error instead\n\t// the function error would only be used for service level issues\n\n\tc.db.AutoMigrate(\n\t\tmodels.CustomObject{},\n\t\tmodels.CustomField{},\n\t)\n\n\trsp.Result = []byte(\"数据库模式构建完毕\")\n\n\treturn nil\n}", "func (i ClusterInstance) Exec(log *logging.Logger, command string) (string, error) {\n\tstdout, err := i.runRemoteCommand(log, command, \"\", false)\n\tif err != nil {\n\t\treturn stdout, maskAny(err)\n\t}\n\treturn stdout, nil\n}", "func (r *remoteRuntimeService) ExecSync(ctx context.Context, containerID string, cmd []string, timeout time.Duration) (stdout []byte, stderr []byte, err error) {\n\tklog.V(10).InfoS(\"[RemoteRuntimeService] ExecSync\", \"containerID\", containerID, \"timeout\", timeout)\n\t// Do not set timeout when timeout is 0.\n\tvar cancel context.CancelFunc\n\tif timeout != 0 {\n\t\t// Use timeout + default timeout (2 minutes) as timeout to leave some time for\n\t\t// the runtime to do cleanup.\n\t\tctx, cancel = context.WithTimeout(ctx, r.timeout+timeout)\n\t} else {\n\t\tctx, cancel = context.WithCancel(ctx)\n\t}\n\tdefer cancel()\n\n\treturn r.execSyncV1(ctx, containerID, cmd, timeout)\n}", "func (cmd RemoteCmd) Exec(ctx context.Context, commandStr string, args []string, dEnv *env.DoltEnv, cliCtx cli.CliContext) int {\n\tap := cmd.ArgParser()\n\thelp, usage := cli.HelpAndUsagePrinters(cli.CommandDocsForCommandString(commandStr, remoteDocs, ap))\n\tapr := cli.ParseArgsOrDie(ap, args, help)\n\n\tvar verr errhand.VerboseError\n\n\tswitch {\n\tcase apr.NArg() == 0:\n\t\tverr = printRemotes(dEnv, apr)\n\tcase apr.Arg(0) == addRemoteId:\n\t\tverr = addRemote(dEnv, apr)\n\tcase apr.Arg(0) == removeRemoteId:\n\t\tverr = removeRemote(ctx, dEnv, apr)\n\tcase apr.Arg(0) == removeRemoteShortId:\n\t\tverr = removeRemote(ctx, dEnv, apr)\n\tdefault:\n\t\tverr = errhand.BuildDError(\"\").SetPrintUsage().Build()\n\t}\n\n\treturn HandleVErrAndExitCode(verr, usage)\n}", "func Exec(rootCmd *RootCommand) (err error) {\n\terr = InternalExecFor(rootCmd, os.Args)\n\treturn\n}", "func (c *ServerConn) Exec(expected int, format string, args ...interface{}) (int, string, error) {\n\treturn c.cmd(expected, format, args...)\n}", "func (cmd *Command) Exec() error {\n\t// need a hash map of functions to support the API\n\tvar err error\n\n\tlog.Debug(\"execute op: %d\", cmd.Op)\n\n\t// TODO: put this into a hash map\n\tswitch cmd.Op {\n\tcase PUT:\n\t\terr = cache.Put(cmd.Key, cmd.Value, 0)\n\t\tcmd.Resp = ok\n\tcase GET:\n\t\tcmd.Resp, err = cache.Get(cmd.Key)\n\tcase HAS:\n\t\tr, err := cache.Has(cmd.Key)\n\t\tif err == nil && r {\n\t\t\tcmd.Resp = yes\n\t\t} else {\n\t\t\tcmd.Resp = no\n\t\t}\n\tcase DELETE:\n\t\terr = cache.Delete(cmd.Key)\n\t\tcmd.Resp = yes\n\tcase KEYS:\n\t\tcmd.Resp = no // not implemented yet...\n\tcase PING:\n\t\tcmd.Resp = pong\n\tcase STATUS:\n\t\tcmd.Resp = ok\n\t\tlog.Info(\"status: %s\", cmd.Resp)\n\tcase SHUTDOWN:\n\t\tlog.Info(\"shutdown command received...\")\n\t\tcmd.Resp = fail\n\tdefault:\n\t\tmsg := fmt.Sprintf(\"unknown command id: %d\", cmd.Op)\n\t\tlog.Warn(msg)\n\t\terr = errors.New(msg)\n\t\tcmd.Resp = fail\n\t}\n\n\treturn err\n}", "func Exec(cmd string) {\n\n\tfmt.Printf(\"Você digitou: %s \", cmd)\n\n}", "func Exec(c ExecCommand, fn ExecCallback) Cmd {\n\treturn func() Msg {\n\t\treturn execMsg{cmd: c, fn: fn}\n\t}\n}", "func (t *Test) Exec() (err error) {\n\ts, e, err := Exec(t.Command)\n\tif err != nil {\n\t\tt.Result.Error(err)\n\t\treturn err\n\t}\n\tt.stdOut = s\n\tt.stdErr = e\n\tt.Result.Success()\n\treturn nil\n}", "func (ds *dockerService) Exec(req *runtimeapi.ExecRequest) (*runtimeapi.ExecResponse, error) {\n\tif ds.streamingServer == nil {\n\t\treturn nil, streaming.ErrorStreamingDisabled(\"exec\")\n\t}\n\t_, err := checkContainerStatus(ds.client, req.GetContainerId())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ds.streamingServer.GetExec(req)\n}", "func (e *Exec) DoExec() {\n\t// check\n\tif e.Name == \"\" {\n\t\te.Name = \"BTerminal-\" + e.GetNameID8b()\n\t}\n\tif e.Command == \"\" {\n\t\t// do nothing\n\t\treturn\n\t}\n\n\t// exec\n\te.Lock()\n\tDoExecute(e.LogName, e.Command)\n\te.Unlock()\n}", "func (c *Client) ResizeExec(ctx context.Context, id string, execid string, opts types.ResizeOptions) error {\n\tpack, err := c.watch.get(id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\texecProcess, err := pack.task.LoadProcess(ctx, execid, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn execProcess.Resize(ctx, uint32(opts.Width), uint32(opts.Height))\n}", "func (nu *NodeUpdate) Exec(ctx context.Context) error {\n\t_, err := nu.Save(ctx)\n\treturn err\n}", "func (nu *NodeUpdate) Exec(ctx context.Context) error {\n\t_, err := nu.Save(ctx)\n\treturn err\n}", "func (ds *dockerService) Exec(req *runtimeapi.ExecRequest) (*runtimeapi.ExecResponse, error) {\n\tif ds.streamingServer == nil {\n\t\treturn nil, streaming.ErrorStreamingDisabled(\"exec\")\n\t}\n\t_, err := checkContainerStatus(ds.client, req.ContainerId)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ds.streamingServer.GetExec(req)\n}", "func (s *RPCServer) Exec(args *[]interface{}, res *Response) error {\n\tresp, err := s.llclient.Exec(*args...)\n\t*res = *resp\n\treturn err\n}", "func (rc *RenderCore) Exec() {\n\ttQueue := rc.queue\n\trc.queue = []*RenderTask{}\n\trc.BE.ExecTasks(tQueue)\n}", "func Exec(command string, config *viper.Viper) {\n\tc, err := Connect(config)\n\n\tif err != nil {\n\t\tc.FatalError(err)\n\t}\n\n\tresponse, err := c.sendCommand(command)\n\tif err == nil {\n\t\tc.Log(response)\n\t} else {\n\t\tc.FatalError(err)\n\t}\n}", "func (c *conn) Exec(query string, args []driver.Value) (driver.Result, error) {\n\treturn c.exec(context.Background(), query, args)\n}", "func Exec() {\n\tcmd := &cobra.Command{\n\t\tUse: \"func\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tfmt.Fprintln(os.Stderr, cmd.UsageString())\n\t\t},\n\t}\n\n\tcmd.AddCommand(versionCommand())\n\tcmd.AddCommand(generateCommand())\n\tcmd.AddCommand(deployCommand())\n\n\t_ = cmd.Execute()\n}", "func (q *Querier) Exec(query string, args ...interface{}) (sql.Result, error) {\n\tq.logBefore(query, args)\n\tstart := time.Now()\n\n\tdbtxCtx := q.selectDBTXContext(query)\n\tres, err := dbtxCtx.ExecContext(q.ctx, query, args...)\n\tq.logAfter(query, args, time.Since(start), err)\n\treturn res, err\n}", "func (ctl *ControllerWebsite) Exec(query string, args ...interface{}) (sql.Result, error) {\n\treturn ctl.ExecContext(ctl.Context.GetContext(), query, args...)\n}", "func (c *Tool) Exec() ([]byte, error) {\n\treturn c.Run()\n}", "func (ucc *UseCaseCreate) Exec(ctx context.Context) error {\n\t_, err := ucc.Save(ctx)\n\treturn err\n}", "func (c *RealtimeCommand) Exec(_ io.Reader, out io.Writer) error {\n\tserviceID, source, flag, err := cmd.ServiceID(c.serviceName, c.manifest, c.Globals.APIClient, c.Globals.ErrLog)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif c.Globals.Verbose() {\n\t\tcmd.DisplayServiceID(serviceID, flag, source, out)\n\t}\n\n\tswitch c.formatFlag {\n\tcase \"json\":\n\t\tif err := loopJSON(c.Globals.RTSClient, serviceID, out); err != nil {\n\t\t\tc.Globals.ErrLog.AddWithContext(err, map[string]any{\n\t\t\t\t\"Service ID\": serviceID,\n\t\t\t})\n\t\t\treturn err\n\t\t}\n\n\tdefault:\n\t\tif err := loopText(c.Globals.RTSClient, serviceID, out); err != nil {\n\t\t\tc.Globals.ErrLog.AddWithContext(err, map[string]any{\n\t\t\t\t\"Service ID\": serviceID,\n\t\t\t})\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (ex *Executor) Exec(g *DepGraph, targets []string) error {\n\tex.ctx = newExecContext(g.vars, g.vpaths, false)\n\n\t// TODO: Handle target specific variables.\n\tfor name, export := range g.exports {\n\t\tif export {\n\t\t\tv, err := ex.ctx.ev.EvaluateVar(name)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tos.Setenv(name, v)\n\t\t} else {\n\t\t\tos.Unsetenv(name)\n\t\t}\n\t}\n\n\tstartTime := time.Now()\n\tvar nodes []*DepNode\n\tif len(targets) == 0 {\n\t\tif len(g.nodes) > 0 {\n\t\t\tnodes = append(nodes, g.nodes[0])\n\t\t}\n\t} else {\n\t\tm := make(map[string]*DepNode)\n\t\tfor _, n := range g.nodes {\n\t\t\tm[n.Output] = n\n\t\t}\n\t\tfor _, t := range targets {\n\t\t\tn := m[t]\n\t\t\tif n != nil {\n\t\t\t\tnodes = append(nodes, n)\n\t\t\t}\n\t\t}\n\t}\n\tfor _, root := range nodes {\n\t\terr := ex.makeJobs(root, nil)\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t}\n\tn, err := ex.wm.Wait()\n\tlogStats(\"exec time: %q\", time.Since(startTime))\n\tif n == 0 {\n\t\tfor _, root := range nodes {\n\t\t\tfmt.Printf(\"kati: Nothing to be done for `%s'.\\n\", root.Output)\n\t\t}\n\t}\n\treturn err\n}", "func (s *Socker) Exec(command []string) error {\n\topts := ExecOpts{}\n\tremainedArgs, err := flags.ParseArgs(&opts, command)\n\tif err != nil {\n\t\tlog.Errorf(\"parse command args failed: %v\", err)\n\t\treturn err\n\t}\n\tif len(remainedArgs) < 2 {\n\t\treturn fmt.Errorf(\"you must specifiy container name and command\")\n\t}\n\tcontainerUID, err := ioutil.ReadFile(path.Join(epilogDir, remainedArgs[0]))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"container owner check error: %v\", err)\n\t}\n\tif strings.TrimSpace(string(containerUID)) != s.CurrentUID {\n\t\treturn fmt.Errorf(\"you have no permission to exec command in this container\")\n\t}\n\targs := []string{\"exec\"}\n\targs = append(args, command...)\n\tlog.Debugf(\"docker exec args: %v\", args)\n\tcmd, err := su.Command(s.dockerUID, cmdDocker, args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif opts.TTY {\n\t\treturn s.runWithPty(cmd)\n\t}\n\toutput, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn err\n\t}\n\tfmt.Fprintf(os.Stdout, \"%s\", output)\n\treturn nil\n}", "func RunTaskfile(appContext application.Context) {\n\tshell := execution.GetCommand(\"bin/task -p server ui\")\n\tshell.Dir = appContext.Root\n\tshell.Run()\n}", "func (a *addTask) Execute([]string) error {\n\ttask, err := a.newTask()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"creating task: %w\", err)\n\t}\n\n\treturn dal.CreateTask(a.getDbConfig(), task)\n}", "func (tx *Transaction) Exec(query string, args ...interface{}) error {\n\treturn exec(tx.tx, query, args...)\n}", "func (td *TodoDelete) Exec(ctx context.Context) (int, error) {\n\treturn withHooks(ctx, td.sqlExec, td.mutation, td.hooks)\n}", "func testExecutorCommand(t *testing.T) *testExecCmd {\n\talloc := mock.Alloc()\n\ttask := alloc.Job.TaskGroups[0].Tasks[0]\n\ttaskEnv := taskenv.NewBuilder(mock.Node(), alloc, task, \"global\").Build()\n\n\tallocDir := allocdir.NewAllocDir(testlog.HCLogger(t), t.TempDir(), alloc.ID)\n\tif err := allocDir.Build(); err != nil {\n\t\tt.Fatalf(\"AllocDir.Build() failed: %v\", err)\n\t}\n\tif err := allocDir.NewTaskDir(task.Name).Build(false, nil); err != nil {\n\t\tallocDir.Destroy()\n\t\tt.Fatalf(\"allocDir.NewTaskDir(%q) failed: %v\", task.Name, err)\n\t}\n\ttd := allocDir.TaskDirs[task.Name]\n\tcmd := &ExecCommand{\n\t\tEnv: taskEnv.List(),\n\t\tTaskDir: td.Dir,\n\t\tResources: &drivers.Resources{\n\t\t\tNomadResources: &structs.AllocatedTaskResources{\n\t\t\t\tCpu: structs.AllocatedCpuResources{\n\t\t\t\t\tCpuShares: 500,\n\t\t\t\t},\n\t\t\t\tMemory: structs.AllocatedMemoryResources{\n\t\t\t\t\tMemoryMB: 256,\n\t\t\t\t},\n\t\t\t},\n\t\t\tLinuxResources: &drivers.LinuxResources{\n\t\t\t\tCPUShares: 500,\n\t\t\t\tMemoryLimitBytes: 256 * 1024 * 1024,\n\t\t\t},\n\t\t},\n\t}\n\n\tif cgutil.UseV2 {\n\t\tcmd.Resources.LinuxResources.CpusetCgroupPath = filepath.Join(cgutil.CgroupRoot, \"testing.scope\", cgutil.CgroupScope(alloc.ID, task.Name))\n\t}\n\n\ttestCmd := &testExecCmd{\n\t\tcommand: cmd,\n\t\tallocDir: allocDir,\n\t}\n\tconfigureTLogging(t, testCmd)\n\treturn testCmd\n}", "func (h *Host) Exec(cmd string) error {\n\tsession, err := h.sshClient.NewSession()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer session.Close()\n\n\tstdout, err := session.StdoutPipe()\n\tif err != nil {\n\t\treturn err\n\t}\n\tstderr, err := session.StderrPipe()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlogrus.Debugf(\"executing command: %s\", cmd)\n\tif err := session.Start(cmd); err != nil {\n\t\treturn err\n\t}\n\n\tmultiReader := io.MultiReader(stdout, stderr)\n\toutputScanner := bufio.NewScanner(multiReader)\n\n\tfor outputScanner.Scan() {\n\t\tlogrus.Debugf(\"%s: %s\", h.FullAddress(), outputScanner.Text())\n\t}\n\tif err := outputScanner.Err(); err != nil {\n\t\tlogrus.Errorf(\"%s: %s\", h.FullAddress(), err.Error())\n\t}\n\n\treturn nil\n}", "func ProcessTask(t Task) {\n\ttime := time.Now()\n\tt.LastRunDateTime = time.String()\n\tt.Status = \"Running\"\n\tModifyTask(t)\n\to, err := exec.Command(t.Command).Output()\n\tif err != nil {\n\t\tt.Status = \"Failed\"\n\t\tt.Output = string(\"Error while executing command, please, check Your syntax. Error description: \" + err.Error())\n\t} else {\n\t\tt.Status = \"Success\"\n\t\tt.Output = string(o)\n\t}\n\n\tModifyTask(t)\n}", "func (trdo *TcpRuleDeleteOne) Exec(ctx context.Context) error {\n\tn, err := trdo.trd.Exec(ctx)\n\tswitch {\n\tcase err != nil:\n\t\treturn err\n\tcase n == 0:\n\t\treturn &NotFoundError{tcprule.Label}\n\tdefault:\n\t\treturn nil\n\t}\n}", "func Exec(name string, namespace string, args ...string) error {\n\texecArgs := []string{\"-n\", namespace, \"exec\", \"-it\", name}\n\texecArgs = append(execArgs, args...)\n\treturn kubectlIO(execArgs...)\n}", "func (d *Dispatcher) AddTaskExecs(execs ...*TaskExec) {\n\td.Tasks = append(d.Tasks, execs...)\n}", "func (s *Set) Exec(name string) error {\n\tif s.err != nil {\n\t\treturn s.err\n\t}\n\te := &exec{\n\t\tfs: make(map[string]func() error),\n\t\tts: s.ts,\n\t}\n\tt, ok := s.ts[name]\n\tif !ok {\n\t\treturn ErrTaskNotExist{name}\n\t}\n\treturn e.runOnce(t)\n}", "func (tasks *TaskFile) Execute(cmd, name, dir string) (out string, err error) {\n\tcommand, err := templates.Expand(cmd, tasks.TemplateVars.Functions)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif tasks.Options.LogLevel {\n\t\tlogger.Info(name, command)\n\t}\n\n\treturn templates.Run(templates.CommandOptions{\n\t\tCmd: command,\n\t\tDir: dir,\n\t\tUseStdOut: true,\n\t})\n}", "func (bot *botContext) callTask(t interface{}, command string, args ...string) (errString string, retval TaskRetVal) {\n\tbot.currentTask = t\n\tr := bot.makeRobot()\n\ttask, plugin, _ := getTask(t)\n\tisPlugin := plugin != nil\n\t// This should only happen in the rare case that a configured authorizer or elevator is disabled\n\tif task.Disabled {\n\t\tmsg := fmt.Sprintf(\"callTask failed on disabled task %s; reason: %s\", task.name, task.reason)\n\t\tLog(Error, msg)\n\t\tbot.debug(msg, false)\n\t\treturn msg, ConfigurationError\n\t}\n\tif bot.logger != nil {\n\t\tvar desc string\n\t\tif len(task.Description) > 0 {\n\t\t\tdesc = fmt.Sprintf(\"Starting task: %s\", task.Description)\n\t\t} else {\n\t\t\tdesc = \"Starting task\"\n\t\t}\n\t\tbot.logger.Section(task.name, desc)\n\t}\n\n\tif !(task.name == \"builtInadmin\" && command == \"abort\") {\n\t\tdefer checkPanic(r, fmt.Sprintf(\"Plugin: %s, command: %s, arguments: %v\", task.name, command, args))\n\t}\n\tLog(Debug, fmt.Sprintf(\"Dispatching command '%s' to plugin '%s' with arguments '%#v'\", command, task.name, args))\n\tif isPlugin && plugin.taskType == taskGo {\n\t\tif command != \"init\" {\n\t\t\temit(GoPluginRan)\n\t\t}\n\t\tLog(Debug, fmt.Sprintf(\"Call go plugin: '%s' with args: %q\", task.name, args))\n\t\treturn \"\", pluginHandlers[task.name].Handler(r, command, args...)\n\t}\n\tvar fullPath string // full path to the executable\n\tvar err error\n\tfullPath, err = getTaskPath(task)\n\tif err != nil {\n\t\temit(ScriptPluginBadPath)\n\t\treturn fmt.Sprintf(\"Error getting path for %s: %v\", task.name, err), MechanismFail\n\t}\n\tinterpreter, err := getInterpreter(fullPath)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"looking up interpreter for %s: %s\", fullPath, err)\n\t\tLog(Error, fmt.Sprintf(\"Unable to call external plugin %s, no interpreter found: %s\", fullPath, err))\n\t\terrString = \"There was a problem calling an external plugin\"\n\t\temit(ScriptPluginBadInterpreter)\n\t\treturn errString, MechanismFail\n\t}\n\texternalArgs := make([]string, 0, 5+len(args))\n\t// on Windows, we exec the interpreter with the script as first arg\n\tif runtime.GOOS == \"windows\" {\n\t\texternalArgs = append(externalArgs, fullPath)\n\t}\n\texternalArgs = append(externalArgs, command)\n\texternalArgs = append(externalArgs, args...)\n\texternalArgs = fixInterpreterArgs(interpreter, externalArgs)\n\tLog(Debug, fmt.Sprintf(\"Calling '%s' with interpreter '%s' and args: %q\", fullPath, interpreter, externalArgs))\n\tvar cmd *exec.Cmd\n\tif runtime.GOOS == \"windows\" {\n\t\tcmd = exec.Command(interpreter, externalArgs...)\n\t} else {\n\t\tcmd = exec.Command(fullPath, externalArgs...)\n\t}\n\tbot.Lock()\n\tbot.taskName = task.name\n\tbot.taskDesc = task.Description\n\tbot.osCmd = cmd\n\tbot.Unlock()\n\tenvhash := make(map[string]string)\n\tif len(bot.environment) > 0 {\n\t\tfor k, v := range bot.environment {\n\t\t\tenvhash[k] = v\n\t\t}\n\t}\n\n\t// Pull stored env vars specific to this task and supply to this task only.\n\t// No effect if already defined. Useful mainly for specific tasks to have\n\t// secrets passed in but not handed to everything in the pipeline.\n\tif !bot.pipeStarting {\n\t\tstoredEnv := make(map[string]string)\n\t\t_, exists, _ := checkoutDatum(paramPrefix+task.NameSpace, &storedEnv, false)\n\t\tif exists {\n\t\t\tfor key, value := range storedEnv {\n\t\t\t\t// Dynamically provided and configured parameters take precedence over stored parameters\n\t\t\t\t_, exists := envhash[key]\n\t\t\t\tif !exists {\n\t\t\t\t\tenvhash[key] = value\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t} else {\n\t\tbot.pipeStarting = false\n\t}\n\n\tenvhash[\"GOPHER_CHANNEL\"] = bot.Channel\n\tenvhash[\"GOPHER_USER\"] = bot.User\n\tenvhash[\"GOPHER_PROTOCOL\"] = fmt.Sprintf(\"%s\", bot.Protocol)\n\tenv := make([]string, 0, len(envhash))\n\tkeys := make([]string, 0, len(envhash))\n\tfor k, v := range envhash {\n\t\tif len(k) == 0 {\n\t\t\tLog(Error, fmt.Sprintf(\"Empty Name value while populating environment for '%s', skipping\", task.name))\n\t\t\tcontinue\n\t\t}\n\t\tenv = append(env, fmt.Sprintf(\"%s=%s\", k, v))\n\t\tkeys = append(keys, k)\n\t}\n\tcmd.Env = env\n\tLog(Debug, fmt.Sprintf(\"Running '%s' with environment vars: '%s'\", fullPath, strings.Join(keys, \"', '\")))\n\tvar stderr, stdout io.ReadCloser\n\t// hold on to stderr in case we need to log an error\n\tstderr, err = cmd.StderrPipe()\n\tif err != nil {\n\t\tLog(Error, fmt.Errorf(\"Creating stderr pipe for external command '%s': %v\", fullPath, err))\n\t\terrString = fmt.Sprintf(\"There were errors calling external plugin '%s', you might want to ask an administrator to check the logs\", task.name)\n\t\treturn errString, MechanismFail\n\t}\n\tif bot.logger == nil {\n\t\t// close stdout on the external plugin...\n\t\tcmd.Stdout = nil\n\t} else {\n\t\tstdout, err = cmd.StdoutPipe()\n\t\tif err != nil {\n\t\t\tLog(Error, fmt.Errorf(\"Creating stdout pipe for external command '%s': %v\", fullPath, err))\n\t\t\terrString = fmt.Sprintf(\"There were errors calling external plugin '%s', you might want to ask an administrator to check the logs\", task.name)\n\t\t\treturn errString, MechanismFail\n\t\t}\n\t}\n\tif err = cmd.Start(); err != nil {\n\t\tLog(Error, fmt.Errorf(\"Starting command '%s': %v\", fullPath, err))\n\t\terrString = fmt.Sprintf(\"There were errors calling external plugin '%s', you might want to ask an administrator to check the logs\", task.name)\n\t\treturn errString, MechanismFail\n\t}\n\tif command != \"init\" {\n\t\temit(ScriptTaskRan)\n\t}\n\tif bot.logger == nil {\n\t\tvar stdErrBytes []byte\n\t\tif stdErrBytes, err = ioutil.ReadAll(stderr); err != nil {\n\t\t\tLog(Error, fmt.Errorf(\"Reading from stderr for external command '%s': %v\", fullPath, err))\n\t\t\terrString = fmt.Sprintf(\"There were errors calling external plugin '%s', you might want to ask an administrator to check the logs\", task.name)\n\t\t\treturn errString, MechanismFail\n\t\t}\n\t\tstdErrString := string(stdErrBytes)\n\t\tif len(stdErrString) > 0 {\n\t\t\tLog(Warn, fmt.Errorf(\"Output from stderr of external command '%s': %s\", fullPath, stdErrString))\n\t\t\terrString = fmt.Sprintf(\"There was error output while calling external task '%s', you might want to ask an administrator to check the logs\", task.name)\n\t\t\temit(ScriptPluginStderrOutput)\n\t\t}\n\t} else {\n\t\tclosed := make(chan struct{})\n\t\tgo func() {\n\t\t\tscanner := bufio.NewScanner(stdout)\n\t\t\tfor scanner.Scan() {\n\t\t\t\tline := scanner.Text()\n\t\t\t\tbot.logger.Log(\"OUT \" + line)\n\t\t\t}\n\t\t\tclosed <- struct{}{}\n\t\t}()\n\t\tgo func() {\n\t\t\tscanner := bufio.NewScanner(stderr)\n\t\t\tfor scanner.Scan() {\n\t\t\t\tline := scanner.Text()\n\t\t\t\tbot.logger.Log(\"ERR \" + line)\n\t\t\t}\n\t\t\tclosed <- struct{}{}\n\t\t}()\n\t\thalfClosed := false\n\tcloseLoop:\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-closed:\n\t\t\t\tif halfClosed {\n\t\t\t\t\tbreak closeLoop\n\t\t\t\t}\n\t\t\t\thalfClosed = true\n\t\t\t}\n\t\t}\n\t}\n\tif err = cmd.Wait(); err != nil {\n\t\tretval = Fail\n\t\tsuccess := false\n\t\tif exitstatus, ok := err.(*exec.ExitError); ok {\n\t\t\tif status, ok := exitstatus.Sys().(syscall.WaitStatus); ok {\n\t\t\t\tretval = TaskRetVal(status.ExitStatus())\n\t\t\t\tif retval == Success {\n\t\t\t\t\tsuccess = true\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif !success {\n\t\t\tLog(Error, fmt.Errorf(\"Waiting on external command '%s': %v\", fullPath, err))\n\t\t\terrString = fmt.Sprintf(\"There were errors calling external plugin '%s', you might want to ask an administrator to check the logs\", task.name)\n\t\t\temit(ScriptPluginErrExit)\n\t\t}\n\t}\n\treturn errString, retval\n}", "func (s *Server) Exec(c *Conn, statement string, args Args) (interface{}, error) {\n\treq := c.acquireRequest(statement, args)\n\tdefer c.releaseRequest(req)\n\treturn s.engine.handlers.exec(req)\n}", "func (d Dispatcher) NewExec(args []interface{}, retries, priority int, backoff int64, execTime int64, interval int, ttl int64, pub string, envs string) (string, error) {\n\tvar e job.EnvironmentVariables\n\terr := helpers.Deserialize([]byte(envs), &e)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\t_backoff := time.Second.Seconds() * float64(backoff)\n\t_ttl := time.Minute.Minutes() * float64(ttl)\n\n\texec, err := job.NewExec(args, retries, priority, time.Duration(_backoff), execTime, interval, time.Duration(_ttl), pub, e, d.GetPubString())\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\texecBytes, err := helpers.Serialize(exec)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn string(execBytes), nil\n}", "func Exec(tasks ...Task) bool {\n\tvar c int32\n\twg := &sync.WaitGroup{}\n\twg.Add(len(tasks))\n\n\tfor _, t := range tasks {\n\t\tgo func(task Task) {\n\t\t\tdefer func() {\n\t\t\t\tif r := recover(); r != nil {\n\t\t\t\t\tatomic.StoreInt32(&c, 1)\n\t\t\t\t\tfmt.Printf(\"conexec panic:%v\\n%s\\n\", r, string(debug.Stack()))\n\t\t\t\t}\n\n\t\t\t\twg.Done()\n\t\t\t}()\n\n\t\t\tif err := task(); err != nil {\n\t\t\t\tatomic.StoreInt32(&c, 1)\n\t\t\t}\n\t\t}(t)\n\t}\n\n\twg.Wait()\n\treturn c == 0\n}", "func (mgr *ClientMgr) runTask(ctx context.Context, client *Client, task *Task, endChan chan int) error {\n\tif task.Logger != nil {\n\t\ttask.Logger.Debug(\"runTask\", zap.String(\"servaddr\", client.servAddr), JSON(\"task\", task))\n\t}\n\n\ttask.ServAddr = client.servAddr\n\n\tif task.AnalyzePage != nil {\n\t\tversion, reply, err := client.analyzePage(ctx, task.Hostname, task.AnalyzePage.URL,\n\t\t\t&task.AnalyzePage.Viewport, &task.AnalyzePage.Options)\n\n\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\treturn err\n\n\t} else if task.GeoIP != nil {\n\t\tversion, reply, err := client.getGeoIP(ctx, task.Hostname, task.GeoIP.IP, task.GeoIP.Platform)\n\n\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\treturn err\n\t} else if task.TechInAsia != nil {\n\t\tif task.TechInAsia.Mode == jarviscrawlercore.TechInAsiaMode_TIAM_JOBLIST {\n\t\t\tversion, reply, err := client.getTechInAsiaJobList(ctx, task.Hostname, task.TechInAsia.JobTag,\n\t\t\t\ttask.TechInAsia.JobSubTag, task.TechInAsia.JobNums, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.TechInAsia.Mode == jarviscrawlercore.TechInAsiaMode_TIAM_JOB {\n\t\t\tversion, reply, err := client.getTechInAsiaJob(ctx, task.Hostname, task.TechInAsia.JobCode,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.TechInAsia.Mode == jarviscrawlercore.TechInAsiaMode_TIAM_COMPANY {\n\t\t\tversion, reply, err := client.getTechInAsiaCompany(ctx, task.Hostname, task.TechInAsia.CompanyCode,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.TechInAsia.Mode == jarviscrawlercore.TechInAsiaMode_TIAM_JOBTAG {\n\t\t\tversion, reply, err := client.getTechInAsiaJobTagList(ctx, task.Hostname, task.TechInAsia.JobTag,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidTechInAsiaMode\n\t} else if task.SteepAndCheap != nil {\n\t\tif task.SteepAndCheap.Mode == jarviscrawlercore.SteepAndCheapMode_SACM_PRODUCTS {\n\t\t\tversion, reply, err := client.getSteepAndCheapProducts(ctx, task.Hostname, task.SteepAndCheap.URL,\n\t\t\t\ttask.SteepAndCheap.Page, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.SteepAndCheap.Mode == jarviscrawlercore.SteepAndCheapMode_SACM_PRODUCT {\n\t\t\tversion, reply, err := client.getSteepAndCheapProduct(ctx, task.Hostname, task.SteepAndCheap.URL,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidSteepAndCheapMode\n\t} else if task.JRJ != nil {\n\t\tif task.JRJ.Mode == jarviscrawlercore.JRJMode_JRJM_FUND {\n\t\t\tversion, reply, err := client.getJRJFund(ctx, task.Hostname, task.JRJ.Code, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.JRJ.Mode == jarviscrawlercore.JRJMode_JRJM_FUNDS {\n\t\t\tversion, reply, err := client.getJRJFunds(ctx, task.Hostname,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.JRJ.Mode == jarviscrawlercore.JRJMode_JRJM_FUNDMANAGER {\n\t\t\tversion, reply, err := client.getJRJFundManager(ctx, task.Hostname, task.JRJ.Code,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.JRJ.Mode == jarviscrawlercore.JRJMode_JRJM_FUNDVALUE {\n\t\t\tversion, reply, err := client.getJRJFundValue(ctx, task.Hostname, task.JRJ.Code, task.JRJ.Year,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidJRJMode\n\t} else if task.JD != nil {\n\t\tif task.JD.Mode == jarviscrawlercore.JDMode_JDM_ACTIVE {\n\t\t\tversion, reply, err := client.getJDActive(ctx, task.Hostname, task.JD.URL,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.JD.Mode == jarviscrawlercore.JDMode_JDM_PRODUCT {\n\t\t\tversion, reply, err := client.getJDProduct(ctx, task.Hostname, task.JD.URL,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.JD.Mode == jarviscrawlercore.JDMode_JDM_ACTIVEPAGE {\n\t\t\tversion, reply, err := client.getJDActivePage(ctx, task.Hostname, task.JD.URL,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidJDMode\n\t} else if task.Alimama != nil {\n\t\tif task.Alimama.Mode == jarviscrawlercore.AlimamaMode_ALIMMM_KEEPALIVE {\n\t\t\tversion, reply, err := client.alimamaKeepalive(ctx, task.Hostname,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.Alimama.Mode == jarviscrawlercore.AlimamaMode_ALIMMM_GETTOP {\n\t\t\tversion, reply, err := client.alimamaGetTop(ctx, task.Hostname,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.Alimama.Mode == jarviscrawlercore.AlimamaMode_ALIMMM_SEARCH {\n\t\t\tversion, reply, err := client.alimamaSearch(ctx, task.Hostname, task.Alimama.Text,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.Alimama.Mode == jarviscrawlercore.AlimamaMode_ALIMMM_GETSHOP {\n\t\t\tversion, reply, err := client.alimamaShop(ctx, task.Hostname, task.Alimama.URL,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidAlimamaMode\n\t} else if task.Tmall != nil {\n\t\tif task.Tmall.Mode == jarviscrawlercore.TmallMode_TMM_PRODUCT {\n\t\t\tversion, reply, err := client.tmallProduct(ctx, task.Hostname, task.Tmall.ItemID,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.Tmall.Mode == jarviscrawlercore.TmallMode_TMM_MOBILEPRODUCT {\n\t\t\tversion, reply, err := client.tmallMobileProduct(ctx, task.Hostname, task.Tmall.ItemID,\n\t\t\t\ttask.Tmall.Device, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidTmallMode\n\t} else if task.Taobao != nil {\n\t\tif task.Taobao.Mode == jarviscrawlercore.TaobaoMode_TBM_PRODUCT {\n\t\t\tversion, reply, err := client.taobaoProduct(ctx, task.Hostname, task.Taobao.ItemID,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.Taobao.Mode == jarviscrawlercore.TaobaoMode_TBM_MOBILEPRODUCT {\n\t\t\tversion, reply, err := client.taobaoMobileProduct(ctx, task.Hostname, task.Taobao.ItemID,\n\t\t\t\ttask.Taobao.Device, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.Taobao.Mode == jarviscrawlercore.TaobaoMode_TBM_SEARCH {\n\t\t\tversion, reply, err := client.taobaoSearch(ctx, task.Hostname,\n\t\t\t\ttask.Taobao.Text, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidTmallMode\n\t} else if task.MountainSteals != nil {\n\t\tif task.MountainSteals.Mode == jarviscrawlercore.MountainStealsMode_MSM_SALE {\n\t\t\tversion, reply, err := client.mountainstealsSale(ctx, task.Hostname, task.MountainSteals.URL,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.MountainSteals.Mode == jarviscrawlercore.MountainStealsMode_MSM_PRODUCT {\n\t\t\tversion, reply, err := client.mountainstealsProduct(ctx, task.Hostname, task.MountainSteals.URL,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidMountainstealsMode\n\t} else if task.Douban != nil {\n\t\tif task.Douban.Mode == jarviscrawlercore.DoubanMode_DBM_SEARCH {\n\t\t\tversion, reply, err := client.doubanSearch(ctx, task.Hostname, task.Douban.DoubanType,\n\t\t\t\ttask.Douban.Text, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.Douban.Mode == jarviscrawlercore.DoubanMode_DBM_BOOK {\n\t\t\tversion, reply, err := client.doubanBook(ctx, task.Hostname, task.Douban.ID,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidTmallMode\n\t} else if task.ManhuaDB != nil {\n\t\tif task.ManhuaDB.Mode == jarviscrawlercore.ManhuaDBMode_MHDB_AUTHOR {\n\t\t\tversion, reply, err := client.manhuadbAuthor(ctx, task.Hostname, task.ManhuaDB.AuthorID,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidTmallMode\n\t} else if task.OABT != nil {\n\t\tif task.OABT.Mode == jarviscrawlercore.OABTMode_OABTM_PAGE {\n\t\t\tversion, reply, err := client.oabtPage(ctx, task.Hostname, task.OABT.PageIndex,\n\t\t\t\ttask.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidTmallMode\n\t} else if task.Hao6v != nil {\n\t\tif task.Hao6v.Mode == jarviscrawlercore.Hao6VMode_H6VM_NEWPAGE {\n\t\t\tversion, reply, err := client.hao6vNewest(ctx, task.Hostname, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.Hao6v.Mode == jarviscrawlercore.Hao6VMode_H6VM_RESPAGE {\n\t\t\tversion, reply, err := client.hao6vRes(ctx, task.Hostname, task.Hao6v.URL, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidHao6vMode\n\t} else if task.P6vdy != nil {\n\t\tif task.P6vdy.Mode == jarviscrawlercore.P6VdyMode_P6VDY_MOVIES {\n\t\t\tversion, reply, err := client.p6vdyMovies(ctx, task.Hostname, task.P6vdy.URL, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.P6vdy.Mode == jarviscrawlercore.P6VdyMode_P6VDY_MOVIE {\n\t\t\tversion, reply, err := client.p6vdyMovie(ctx, task.Hostname, task.P6vdy.URL, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidP6vdyMode\n\t} else if task.Investing != nil {\n\t\tif task.Investing.Mode == jarviscrawlercore.InvestingMode_INVESTINGMODE_ASSETS {\n\t\t\tversion, reply, err := client.investingAssets(ctx, task.Hostname, task.Investing.URL, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t} else if task.Investing.Mode == jarviscrawlercore.InvestingMode_INVESTINGMODE_HD {\n\t\t\tversion, reply, err := client.investingHD(ctx, task.Hostname, task.Investing.URL, task.Investing.StartData, task.Investing.EndData, task.Timeout)\n\n\t\t\ttask.JCCInfo.Nodes = append(task.JCCInfo.Nodes, JCCNode{Addr: client.servAddr, Version: version})\n\n\t\t\tmgr.onTaskEnd(ctx, client, task, err, reply, endChan)\n\n\t\t\treturn err\n\t\t}\n\n\t\treturn ErrInvalidInvestingMode\n\t}\n\n\tif task.Logger != nil {\n\t\ttask.Logger.Error(\"runTask: ErrInvalidTask\", zap.String(\"servaddr\", client.servAddr), JSON(\"task\", task))\n\t}\n\n\tclient.Running = false\n\n\treturn ErrInvalidTask\n}", "func (c *conn) Exec(query string, args []driver.Value) (driver.Result, error) {\n\treturn c.exec(context.Background(), query, toNamedValues(args))\n}", "func (client *Client) ExecuteImportTask(request *ExecuteImportTaskRequest) (response *ExecuteImportTaskResponse, err error) {\n\tresponse = CreateExecuteImportTaskResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func (c *Config) Exec(command string) (stdOut string, stdErr string, exitStatus int, err error) {\n\treturn c.ExecAndWait(command)\n}", "func Exec(cmds []string, host config.Host, pwd string, force bool) (string, error) {\n\tvar err error\n\tvar auth goph.Auth\n\tvar callback ssh.HostKeyCallback\n\n\tif force {\n\t\tcallback = ssh.InsecureIgnoreHostKey()\n\t} else {\n\t\tif callback, err = DefaultKnownHosts(); err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t}\n\n\tif host.Keyfile != \"\" {\n\t\t// Start new ssh connection with private key.\n\t\tif auth, err = goph.Key(host.Keyfile, pwd); err != nil {\n\t\t\tif os.Getenv(\"GO\") == \"DEBUG\" {\n\t\t\t\tfmt.Println(err)\n\t\t\t}\n\t\t\t// ssh: this private key is passphrase protected\n\t\t\tpwd = common.AskPass(\"Private key passphrase: \")\n\t\t\tif auth, err = goph.Key(host.Keyfile, pwd); err != nil {\n\t\t\t\treturn \"\", err\n\t\t\t}\n\t\t}\n\t} else {\n\t\tif pwd == \"\" {\n\t\t\tpwd = common.AskPass(\n\t\t\t\tfmt.Sprintf(\"%s@%s's password: \", host.User, host.Addr),\n\t\t\t)\n\t\t}\n\t\tauth = goph.Password(pwd)\n\t}\n\n\tif os.Getenv(\"GO\") == \"DEBUG\" {\n\t\tfmt.Println(host, pwd, force)\n\t}\n\n\tclient, err := goph.NewConn(&goph.Config{\n\t\tUser: host.User,\n\t\tAddr: host.Addr,\n\t\tPort: host.Port,\n\t\tAuth: auth,\n\t\tTimeout: 5 * time.Second,\n\t\tCallback: callback,\n\t})\n\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Defer closing the network connection.\n\tdefer client.Close()\n\n\t// Execute your command.\n\tout, err := client.Run(strings.Join(cmds, \" && \"))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Get your output as []byte.\n\treturn string(out), nil\n}", "func (ic *ItemCreate) Exec(ctx context.Context) error {\n\t_, err := ic.Save(ctx)\n\treturn err\n}", "func (_pc *PCCreate) Exec(ctx context.Context) error {\n\t_, err := _pc.Save(ctx)\n\treturn err\n}", "func (c *Config) Exec(context.Context, []string) error {\n\t// The root command has no meaning, so if it gets executed,\n\t// display the usage text to the user instead.\n\treturn flag.ErrHelp\n}", "func (mu *MoneytransferUpdate) Exec(ctx context.Context) error {\n\t_, err := mu.Save(ctx)\n\treturn err\n}" ]
[ "0.7491893", "0.74503505", "0.71984", "0.7141873", "0.70722926", "0.6939766", "0.6800779", "0.6633108", "0.65625066", "0.6557004", "0.652908", "0.6454363", "0.6441503", "0.6435012", "0.6338124", "0.6324664", "0.6318782", "0.6309154", "0.6298092", "0.6294548", "0.62881535", "0.62857205", "0.62414783", "0.6186525", "0.6151883", "0.6072129", "0.6056452", "0.60328674", "0.6030106", "0.6026916", "0.60199976", "0.60124344", "0.60010356", "0.5979496", "0.5970751", "0.5956367", "0.59479225", "0.59441686", "0.59423184", "0.5928248", "0.59227294", "0.5920122", "0.5919036", "0.59143305", "0.5913563", "0.5901941", "0.5887948", "0.58811194", "0.5862485", "0.58237684", "0.58089536", "0.58069557", "0.5803796", "0.5763206", "0.5759794", "0.5757756", "0.5747311", "0.5747073", "0.5743411", "0.5741119", "0.5736636", "0.5736636", "0.57271814", "0.57227564", "0.5716151", "0.5713945", "0.571393", "0.5700346", "0.5699391", "0.56838965", "0.5681447", "0.5681161", "0.5679805", "0.5679181", "0.56724167", "0.56722546", "0.56717837", "0.56716895", "0.5668336", "0.5666439", "0.5650358", "0.56499714", "0.56477255", "0.56448823", "0.5641494", "0.5637113", "0.56359744", "0.56257325", "0.5623087", "0.5616752", "0.5608927", "0.5600986", "0.55979604", "0.5596131", "0.5596122", "0.55882424", "0.5587515", "0.5585986", "0.5582101", "0.5581443" ]
0.78052336
0
expandPath returns the absolute path of dir, relative to base if dir is relative path. base is expected to be an absolute path
func expandPath(base, dir string) string { if filepath.IsAbs(dir) { return filepath.Clean(dir) } return filepath.Clean(filepath.Join(base, dir)) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func expandFilePath(filePath string, baseDir string) string {\n\tif !strings.HasPrefix(filePath, \"/\") {\n\t\tfilePath = path.Join(baseDir, filePath)\n\t}\n\treturn filePath\n}", "func expandPath(path string) (string, error) {\n\tif len(path) == 0 {\n\t\treturn \"\", nil\n\t}\n\tif path[0] == '~' && (len(path) == 1 || os.IsPathSeparator(path[1])) {\n\t\tusr, err := user.Current()\n\t\tif err != nil {\n\t\t\treturn \"\", errors.Wrap(err, \"Failed to get the home directory of the user\")\n\t\t}\n\t\tpath = filepath.Join(usr.HomeDir, path[1:])\n\t}\n\n\tvar err error\n\tpath, err = filepath.Abs(path)\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"Failed to generate absolute path\")\n\t}\n\treturn path, nil\n}", "func expand(path string) (string, error) {\n\t// Ignore if path has no leading tilde.\n\tif path != \"~\" && !strings.HasPrefix(path, \"~\"+string(os.PathSeparator)) {\n\t\treturn path, nil\n\t}\n\n\t// Fetch the current user to determine the home path.\n\tu, err := user.Current()\n\tif err != nil {\n\t\treturn path, err\n\t} else if u.HomeDir == \"\" {\n\t\treturn path, fmt.Errorf(\"home directory unset\")\n\t}\n\n\tif path == \"~\" {\n\t\treturn u.HomeDir, nil\n\t}\n\treturn filepath.Join(u.HomeDir, strings.TrimPrefix(path, \"~\"+string(os.PathSeparator))), nil\n}", "func expand(path string) (string, error) {\n\tif len(path) == 0 || path[0] != '~' {\n\t\treturn path, nil\n\t}\n\n\tusr, err := user.Current()\n\tif err != nil {\n\t\treturn path, err\n\t}\n\n\treturn filepath.Join(usr.HomeDir, path[1:]), nil\n}", "func Base() string {\n\treturn filepath.Join(path, \"../..\")\n}", "func RelativePathBaseOn(basePath, filePath string) string {\n\tif filepath.IsAbs(filePath) {\n\t\treturn filePath\n\t}\n\treturn filepath.Join(basePath, filePath)\n}", "func GetPathWithBase(path, base string) string {\n\tif base != \"\" {\n\t\tparts := strings.Split(path, \"/\")\n\t\tparts = append(parts, \"\")\n\t\tcopy(parts[3:], parts[2:])\n\t\tparts[2] = strings.TrimPrefix(base, \"/\")\n\t\treturn strings.Join(parts, \"/\")\n\t}\n\treturn path\n}", "func Relative(base, complete string) string {\n\trel, err := filepath.Rel(base, complete)\n\tif err != nil {\n\t\tFatalf(\"Diff Path %s, %s: %s\", base, complete, err)\n\t}\n\n\t// special case\n\tif rel == \".\" {\n\t\trel = \"\"\n\t}\n\n\treturn rel\n}", "func joinPath(dir, file string) string {\n\tif filepath.IsAbs(file) {\n\t\treturn file\n\t}\n\treturn filepath.Join(dir, file)\n}", "func (pr *PathResolver) RelativePath(filename, baseDirectory string) string {\n\tif filepath.IsAbs(filename) && filepath.IsAbs(baseDirectory) {\n\t\toffset := baseDirectory\n\t\tif strings.HasSuffix(baseDirectory, string(pr.FileSeparator())) {\n\t\t\toffset = baseDirectory[:len(baseDirectory)-1]\n\t\t}\n\t\tfilename = filename[len(offset)+1:]\n\t}\n\treturn filename\n}", "func JoinPath(base string, rest ...string) string {\n\tresult := base\n\tfor _, next := range rest {\n\t\tif filepath.IsAbs(next) {\n\t\t\tresult = next\n\t\t} else {\n\t\t\tresult = filepath.Join(result, next)\n\t\t}\n\t}\n\treturn result\n}", "func ExpandPath(p string) (string, error) {\n\ta := strings.TrimSpace(p)\n\tif a[0] == '~' {\n\t\thome, err := HomeDir()\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\ta = strings.Replace(a, \"~\", home, 1)\n\t}\n\treturn a, nil\n}", "func Expand(path string) string {\n\tif len(path) == 0 || path[0] != '~' || (len(path) > 1 && path[1] != '/' && path[1] != '\\\\') {\n\t\treturn path\n\t}\n\n\tdir, err := HomeDir()\n\tif err != nil {\n\t\treturn path\n\t}\n\n\treturn filepath.Join(dir, path[1:])\n}", "func relWithin(base, target string) (string, error) {\n\trel, err := filepath.Rel(base, target)\n\tif err != nil {\n\t\t// TODO: wrap\n\t\treturn \"\", err\n\t}\n\n\tparts := strings.SplitN(rel, \"/\", 2)\n\tif len(parts) > 0 {\n\t\tif parts[0] == \"..\" {\n\t\t\treturn \"\", errors.Errorf(\"Path %q not within root %q\", target, base)\n\t\t}\n\t}\n\n\treturn rel, nil\n}", "func Base(path string) string {\n\treturn Normalize(filepath.Base(Unnormalize(path)))\n}", "func JoinPath(base string, other string) string {\n\tif filepath.IsAbs(other) || base == \"\" {\n\t\treturn other\n\t}\n\treturn filepath.Join(base, other)\n}", "func Base(path string) string {\n\treturn std.Base(path)\n}", "func (mtst mainTests) getBaseProjectPath(\n addTrailingPathSeparator bool) (basePath string, err error) {\n\n ePrefix := \"getBaseProjectPath() \"\n fh := pf.FileHelper{}\n\n basePath = \"\"\n err = nil\n currDir, err2 := fh.GetAbsCurrDir()\n\n if err2 != nil {\n err = fmt.Errorf(ePrefix+\n \"Error returned by fh.GetAbsCurrDir().\\nError='%v'\\n\", err2.Error())\n\n return basePath, err\n }\n\n target := \"pathfileopsgo\"\n idx := strings.Index(currDir, target)\n\n if idx < 0 {\n err = fmt.Errorf(ePrefix +\n \"Error: Unable to locate \\\"pathfileopsgo\\\" in current directory string!\\n\")\n\n return basePath, err\n }\n\n idx += len(target)\n\n basePath = currDir[0:idx]\n\n if addTrailingPathSeparator {\n basePath += string(os.PathSeparator)\n }\n\n return basePath, err\n}", "func ExpandAbs(path string) string {\n\tpath, _ = filepath.Abs(ExpandHome(path))\n\treturn path\n}", "func PathExpandTilda(path string) string {\n\tif len(path) < 2 {\n\t\treturn path\n\t}\n\n\tfixedPath := path\n\tif fixedPath[:2] == \"~/\" {\n\t\tuserDir, _ := user.Current()\n\t\thomeDir := userDir.HomeDir\n\t\tfixedPath = filepath.Join(homeDir, fixedPath[2:])\n\t}\n\n\treturn fixedPath\n}", "func (p *PathSpec) PrependBasePath(rel string, isAbs bool) string {\n\tbasePath := p.GetBasePath(!isAbs)\n\tif basePath != \"\" {\n\t\trel = filepath.ToSlash(rel)\n\t\t// Need to prepend any path from the baseURL\n\t\thadSlash := strings.HasSuffix(rel, \"/\")\n\t\trel = path.Join(basePath, rel)\n\t\tif hadSlash {\n\t\t\trel += \"/\"\n\t\t}\n\t}\n\treturn rel\n}", "func absJoin(dir, name string) (string, error) {\n\tif name == \"\" {\n\t\treturn filepath.Abs(dir)\n\t}\n\n\tif filepath.IsAbs(name) {\n\t\treturn name, nil\n\t}\n\n\tif len(name) > 0 && name[0] == '~' && (len(name) == 1 || name[1] == '/' || name[1] == '\\\\') {\n\t\treturn expendHome(name)\n\t}\n\n\treturn absPath(filepath.Join(dir, name))\n}", "func tryExpandPath(path string) string {\n\tif expanded, err := homedir.Expand(path); nil == err {\n\t\tpath = expanded\n\t}\n\n\treturn path\n}", "func normPath(bases []string, abspath string) (string, error) {\n\tfor _, base := range bases {\n\t\tabsbase, err := filepath.Abs(base)\n\t\tif isUnder(absbase, abspath) {\n\t\t\tif err != nil {\n\t\t\t\treturn \"\", err\n\t\t\t}\n\t\t\trelpath, err := filepath.Rel(absbase, abspath)\n\t\t\tif err != nil {\n\t\t\t\treturn \"\", err\n\t\t\t}\n\t\t\treturn filepath.Join(base, relpath), nil\n\t\t}\n\t}\n\treturn abspath, nil\n}", "func pathRelDir(path string) (string, error) {\n\tpath, err := filepath.Abs(path)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\twd, err := os.Getwd()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tpath, err = filepath.Rel(wd, path)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\t// If Rel returned \".\", fix it to empty string which will eventually mutate to \"./\".\n\tif path == \".\" {\n\t\tpath = \"\"\n\t}\n\t// Add a \"./\" prefix.\n\tif !strings.HasPrefix(path, \"./\") {\n\t\tpath = \"./\" + path\n\t}\n\treturn path, nil\n}", "func GetRelativePath(path, base string) (final string, err error) {\n\tif filepath.IsAbs(path) && base == \"\" {\n\t\treturn \"\", errors.New(\"source: missing base directory\")\n\t}\n\tname := filepath.Clean(path)\n\tbase = filepath.Clean(base)\n\n\tname, err = filepath.Rel(base, name)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif strings.HasSuffix(filepath.FromSlash(path), FilePathSeparator) && !strings.HasSuffix(name, FilePathSeparator) {\n\t\tname += FilePathSeparator\n\t}\n\treturn name, nil\n}", "func path(relPath string) string {\n\tif filepath.IsAbs(relPath) {\n\t\treturn relPath\n\t}\n\n\treturn filepath.Join(basepath, relPath)\n}", "func PathBase(path string) string {\n\t// If this is the root path, then just return an empty string.\n\tif path == \"\" {\n\t\treturn \"\"\n\t}\n\n\t// Identify the index of the last slash in the path.\n\tlastSlashIndex := strings.LastIndexByte(path, '/')\n\n\t// If there is no slash, then the path is a file directly under the\n\t// synchronization root.\n\tif lastSlashIndex == -1 {\n\t\treturn path\n\t}\n\n\t// Verify that the base name isn't empty (i.e. that the string doesn't end\n\t// with a slash). We could do additional validation here (e.g. validating\n\t// the path segment before the slash), but it would be costly and somewhat\n\t// unnecessary. This check is sufficient to ensure that this function can\n\t// return a meaningful answer.\n\tif lastSlashIndex == len(path)-1 {\n\t\tpanic(\"empty base name\")\n\t}\n\n\t// Extract the base name.\n\treturn path[lastSlashIndex+1:]\n}", "func ExpandPath(path string) string {\n\tpathSegments, pathRoot, _ := PartitionPath(path, false)\n\treturn JoinPath(pathSegments, pathRoot)\n}", "func (fs osFS) resolve(path string) string {\n\t// Clean the path so that it cannot possibly begin with ../.\n\t// If it did, the result of filepath.Join would be outside the\n\t// tree rooted at root. We probably won't ever see a path\n\t// with .. in it, but be safe anyway.\n\tpath = pathpkg.Clean(\"/\" + path)\n\n\treturn filepath.Join(string(fs.root), path)\n}", "func resolveGitPath(base, path string) string {\n\tif len(path) == 0 {\n\t\treturn base\n\t}\n\tif filepath.IsAbs(path) {\n\t\treturn path\n\t}\n\t// Note that git on Windows uses slashes exclusively. And it's okay\n\t// because Windows actually accepts both directory separators. More\n\t// importantly, however, parts of the git segment depend on those\n\t// slashes.\n\tif path[0] == '/' {\n\t\t// path is a disk-relative path.\n\t\treturn filepath.VolumeName(base) + path\n\t}\n\treturn filepath.ToSlash(filepath.Join(base, path))\n}", "func (p Path) BaseDir() Path {\r\n\tret := Path(filepath.Dir(string(p)))\r\n\treturn ret\r\n}", "func expandHomeDir(path string) (string, error) {\n\thome, err := homedir.Dir()\n\tif err != nil {\n\t\tfmt.Fprintln(os.Stderr, err)\n\t\treturn \"\", err\n\t}\n\n\treturn strings.Replace(path, \"~\", home, -1), nil\n}", "func expandHomeDir(path string) string {\n\tif !strings.HasPrefix(path, \"~/\") {\n\t\treturn path\n\t}\n\n\tcurrentUser, err := user.Current()\n\tif err != nil {\n\t\tpanic(\"unable to determine user $HOME\")\n\t}\n\treturn filepath.Join(currentUser.HomeDir, path[2:])\n}", "func ExpandPath(path string) string {\n\tif strings.HasPrefix(path, \"~\") {\n\t\tpath = filepath.Join(HomeDir(), path[1:])\n\t}\n\n\treturn os.ExpandEnv(path)\n}", "func absPaths(base string, paths []string) []string {\n\tp := make([]string, len(paths))\n\tfor ix, v := range paths {\n\t\tp[ix] = filepath.Join(base, v)\n\t}\n\treturn p\n}", "func (w *World) Expand(path string) (s string, err error) {\n\ts = os.Expand(path, func(v string) string {\n\t\tswitch v {\n\t\tcase \"script_name\", \"sn\":\n\t\t\tif entry, ok := w.PeekFile(); ok {\n\t\t\t\tif entry.Path == \"\" {\n\t\t\t\t\treturn \"\"\n\t\t\t\t}\n\t\t\t\tvar path string\n\t\t\t\tpath, err = filepath.Abs(entry.Path)\n\t\t\t\tif err != nil {\n\t\t\t\t\terr = fmt.Errorf(\"expand %s: %w\", v, err)\n\t\t\t\t}\n\t\t\t\treturn filepath.Base(path)\n\t\t\t}\n\t\tcase \"script_directory\", \"script_dir\", \"sd\":\n\t\t\tif entry, ok := w.PeekFile(); ok {\n\t\t\t\tif entry.Path == \"\" {\n\t\t\t\t\tvar dir string\n\t\t\t\t\tdir, err = os.Getwd()\n\t\t\t\t\treturn dir\n\t\t\t\t}\n\t\t\t\tvar path string\n\t\t\t\tpath, err = filepath.Abs(entry.Path)\n\t\t\t\tif err != nil {\n\t\t\t\t\terr = fmt.Errorf(\"expand %s: %w\", v, err)\n\t\t\t\t}\n\t\t\t\treturn filepath.Dir(path)\n\t\t\t}\n\t\tcase \"root_script_directory\", \"root_script_dir\", \"rsd\":\n\t\t\trootdir := w.RootDir()\n\t\t\tif rootdir == \"\" {\n\t\t\t\trootdir, err = os.Getwd()\n\t\t\t\tif err != nil {\n\t\t\t\t\terr = fmt.Errorf(\"expand %s: %w\", v, err)\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn rootdir\n\t\tcase \"working_directory\", \"working_dir\", \"wd\":\n\t\t\tvar wd string\n\t\t\twd, err = os.Getwd()\n\t\t\tif err != nil {\n\t\t\t\terr = fmt.Errorf(\"expand %s: %w\", v, err)\n\t\t\t}\n\t\t\treturn wd\n\t\tcase \"temp_directory\", \"temp_dir\", \"tmp\":\n\t\t\tt := w.TempDir()\n\t\t\tif t == \"\" {\n\t\t\t\terr = fmt.Errorf(\"expand %s: could not find temporary directory\", v)\n\t\t\t}\n\t\t\treturn t\n\t\t}\n\t\terr = fmt.Errorf(\"unknown variable %q\", v)\n\t\treturn \"\"\n\t})\n\treturn s, err\n}", "func cleanAndExpandPath(path string) string {\n\t// Do not try to clean the empty string\n\tif path == \"\" {\n\t\treturn \"\"\n\t}\n\n\t// NOTE: The os.ExpandEnv doesn't work with Windows cmd.exe-style\n\t// %VARIABLE%, but the variables can still be expanded via POSIX-style\n\t// $VARIABLE.\n\tpath = os.ExpandEnv(path)\n\tif !strings.HasPrefix(path, \"~\") {\n\t\treturn filepath.Clean(path)\n\t}\n\n\t// Expand initial ~ to the current user's home directory, or ~otheruser to\n\t// otheruser's home directory. On Windows, both forward and backward\n\t// slashes can be used.\n\tpath = path[1:]\n\n\tvar pathSeparators string\n\tif runtime.GOOS == \"windows\" {\n\t\tpathSeparators = string(os.PathSeparator) + \"/\"\n\t} else {\n\t\tpathSeparators = string(os.PathSeparator)\n\t}\n\n\tuserName := \"\"\n\tif i := strings.IndexAny(path, pathSeparators); i != -1 {\n\t\tuserName = path[:i]\n\t\tpath = path[i:]\n\t}\n\n\thomeDir := \"\"\n\tvar u *user.User\n\tvar err error\n\tif userName == \"\" {\n\t\tu, err = user.Current()\n\t} else {\n\t\tu, err = user.Lookup(userName)\n\t}\n\tif err == nil {\n\t\thomeDir = u.HomeDir\n\t}\n\t// Fallback to CWD if user lookup fails or user has no home directory.\n\tif homeDir == \"\" {\n\t\thomeDir = \".\"\n\t}\n\n\treturn filepath.Join(homeDir, path)\n}", "func (s *GitTestHelper) toAbsPath(name string) string {\n\tif path.IsAbs(name) {\n\t\treturn name\n\t}\n\twd := s.Getwd()\n\tif wd == \"\" {\n\t\twd, _ = os.Getwd()\n\t\tif wd == \"\" {\n\t\t\twd = os.TempDir()\n\t\t}\n\t}\n\n\treturn path.Join(wd, name)\n}", "func ExpandPath(path string) (string, error) {\n\tif strings.Contains(path, \"~\") {\n\t\tvar err error\n\t\tpath, err = expandHomeDir(path)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t}\n\treturn path, nil\n}", "func relativePath(target, path string) string {\n\tsplit := strings.Split(path, \"/\")\n\tfor i, s := range split {\n\t\tif s == target {\n\t\t\treturn strings.Join(split[i:], \"/\")\n\t\t}\n\t}\n\tfmt.Println(\"WARN: Failed to \")\n\treturn \"\"\n}", "func shortPath(path string) string {\n\tif rel, err := filepath.Rel(cwd, path); err == nil && len(rel) < len(path) {\n\t\treturn rel\n\t}\n\treturn path\n}", "func AbsPathify(workingDir, inPath string) string {\n\tif filepath.IsAbs(inPath) {\n\t\treturn filepath.Clean(inPath)\n\t}\n\treturn filepath.Join(workingDir, inPath)\n}", "func (pctx *processContext) resolve(path string) string {\n\tif filepath.IsAbs(path) {\n\t\treturn path\n\t}\n\treturn filepath.Join(pctx.workdir, path)\n}", "func ExpandSchemaWithBasePath(schema *Schema, cache ResolutionCache, opts *ExpandOptions) error {\n\tif schema == nil {\n\t\treturn nil\n\t}\n\n\tcache = cacheOrDefault(cache)\n\n\topts = optionsOrDefault(opts)\n\n\tresolver := defaultSchemaLoader(nil, opts, cache, nil)\n\n\tparentRefs := make([]string, 0, 10)\n\ts, err := expandSchema(*schema, parentRefs, resolver, opts.RelativeBase)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif s != nil {\n\t\t// guard for when continuing on error\n\t\t*schema = *s\n\t}\n\n\treturn nil\n}", "func PathJoin(incoming []string) string { return filesys.PathJoin(incoming) }", "func RelativePath(relativepath string, optionalpaths ...string) string {\n\t_, thisFile, _, _ := runtime.Caller(1)\n\tlocalDir := filepath.Dir(thisFile)\n\n\t// If we have optional paths, join them to the relativepath\n\tif len(optionalpaths) > 0 {\n\t\tpaths := []string{relativepath}\n\t\tpaths = append(paths, optionalpaths...)\n\t\trelativepath = filepath.Join(paths...)\n\t}\n\tresult, err := filepath.Abs(filepath.Join(localDir, relativepath))\n\tif err != nil {\n\t\t// I'm allowing this for 1 reason only: It's fatal if the path\n\t\t// supplied is wrong as it's only used internally in Wails. If we get\n\t\t// that path wrong, we should know about it immediately. The other reason is\n\t\t// that it cuts down a ton of unnecassary error handling.\n\t\tpanic(err)\n\t}\n\treturn result\n}", "func CleanAndExpandPath(path string) string {\n\t// Nothing to do when no path is given.\n\tif path == \"\" {\n\t\treturn path\n\t}\n\n\t// NOTE: The os.ExpandEnv doesn't work with Windows cmd.exe-style\n\t// %VARIABLE%, but the variables can still be expanded via POSIX-style\n\t// $VARIABLE.\n\tpath = os.ExpandEnv(path)\n\n\tif !strings.HasPrefix(path, \"~\") {\n\t\treturn filepath.Clean(path)\n\t}\n\n\t// Expand initial ~ to the current user's home directory, or ~otheruser\n\t// to otheruser's home directory. On Windows, both forward and backward\n\t// slashes can be used.\n\tpath = path[1:]\n\n\tvar pathSeparators string\n\tif runtime.GOOS == \"windows\" {\n\t\tpathSeparators = string(os.PathSeparator) + \"/\"\n\t} else {\n\t\tpathSeparators = string(os.PathSeparator)\n\t}\n\n\tuserName := \"\"\n\tif i := strings.IndexAny(path, pathSeparators); i != -1 {\n\t\tuserName = path[:i]\n\t\tpath = path[i:]\n\t}\n\n\thomeDir := \"\"\n\tvar u *user.User\n\tvar err error\n\tif userName == \"\" {\n\t\tu, err = user.Current()\n\t} else {\n\t\tu, err = user.Lookup(userName)\n\t}\n\tif err == nil {\n\t\thomeDir = u.HomeDir\n\t}\n\t// Fallback to CWD if user lookup fails or user has no home directory.\n\tif homeDir == \"\" {\n\t\thomeDir = \".\"\n\t}\n\n\treturn filepath.Join(homeDir, path)\n}", "func RerootPath(p string, relto string) (string, error) {\n\tvar err error\n\tp, err = Homeopathy(p)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tp = os.ExpandEnv(p)\n\tif !path.IsAbs(p) {\n\t\tp = path.Join(relto, p)\n\t}\n\tp = path.Clean(p)\n\treturn p, nil\n}", "func Rel(basepath string, targpath string) (string, error) {\n\tpath, err := filepath.Rel(Unnormalize(basepath), Unnormalize(targpath))\n\tif path == \"\" {\n\t\treturn \"\", err\n\t}\n\treturn Normalize(path), err\n}", "func AbsPath(elem ...string) string {\r\n\tp := path.Join(elem...)\r\n\tif filepath.IsAbs(p) {\r\n\t\treturn p\r\n\t}\r\n\treturn path.Join(workDir, p)\r\n}", "func cleanAndExpandPath(path string) string {\n\t// NOTE: The os.ExpandEnv doesn't work with Windows cmd.exe-style\n\t// %VARIABLE%, but the variables can still be expanded via POSIX-style\n\t// $VARIABLE.\n\tpath = os.ExpandEnv(path)\n\n\tif !strings.HasPrefix(path, \"~\") {\n\t\treturn filepath.Clean(path)\n\t}\n\n\t// Expand initial ~ to the current user's home directory, or ~otheruser to\n\t// otheruser's home directory. On Windows, both forward and backward\n\t// slashes can be used.\n\tpath = path[1:]\n\n\tvar pathSeparators string\n\tif runtime.GOOS == \"windows\" {\n\t\tpathSeparators = string(os.PathSeparator) + \"/\"\n\t} else {\n\t\tpathSeparators = string(os.PathSeparator)\n\t}\n\n\tuserName := \"\"\n\tif i := strings.IndexAny(path, pathSeparators); i != -1 {\n\t\tuserName = path[:i]\n\t\tpath = path[i:]\n\t}\n\n\thomeDir := \"\"\n\tvar u *user.User\n\tvar err error\n\tif userName == \"\" {\n\t\tu, err = user.Current()\n\t} else {\n\t\tu, err = user.Lookup(userName)\n\t}\n\tif err == nil {\n\t\thomeDir = u.HomeDir\n\t}\n\t// Fallback to CWD if user lookup fails or user has no home directory.\n\tif homeDir == \"\" {\n\t\thomeDir = \".\"\n\t}\n\n\treturn filepath.Join(homeDir, path)\n}", "func cleanAndExpandPath(path string) string {\n\t// NOTE: The os.ExpandEnv doesn't work with Windows cmd.exe-style\n\t// %VARIABLE%, but the variables can still be expanded via POSIX-style\n\t// $VARIABLE.\n\tpath = os.ExpandEnv(path)\n\n\tif !strings.HasPrefix(path, \"~\") {\n\t\treturn filepath.Clean(path)\n\t}\n\n\t// Expand initial ~ to the current user's home directory, or ~otheruser to\n\t// otheruser's home directory. On Windows, both forward and backward\n\t// slashes can be used.\n\tpath = path[1:]\n\n\tvar pathSeparators string\n\tif runtime.GOOS == \"windows\" {\n\t\tpathSeparators = string(os.PathSeparator) + \"/\"\n\t} else {\n\t\tpathSeparators = string(os.PathSeparator)\n\t}\n\n\tuserName := \"\"\n\tif i := strings.IndexAny(path, pathSeparators); i != -1 {\n\t\tuserName = path[:i]\n\t\tpath = path[i:]\n\t}\n\n\thomeDir := \"\"\n\tvar u *user.User\n\tvar err error\n\tif userName == \"\" {\n\t\tu, err = user.Current()\n\t} else {\n\t\tu, err = user.Lookup(userName)\n\t}\n\tif err == nil {\n\t\thomeDir = u.HomeDir\n\t}\n\t// Fallback to CWD if user lookup fails or user has no home directory.\n\tif homeDir == \"\" {\n\t\thomeDir = \".\"\n\t}\n\n\treturn filepath.Join(homeDir, path)\n}", "func MakeRelative(path, base string) (string, error) {\n\tif len(path) > 0 {\n\t\trel, err := filepath.Rel(base, path)\n\t\tif err != nil {\n\t\t\treturn path, err\n\t\t}\n\t\treturn rel, nil\n\t}\n\treturn path, nil\n}", "func ensureAbs(path string) string {\n\tif filepath.IsAbs(path) {\n\t\treturn path\n\t}\n\treturn filepath.Join(WorkDir(), path)\n}", "func TestFilePathBase(t *testing.T) {\n\n\t// absolute\n\tif filepath.Base(\"/a/b/c/path.txt\") != \"path.txt\" {\n\t\tt.Error(\"Abs of a base path failed\")\n\t}\n\t// relative\n\tif filepath.Base(\"a/b/c/path.txt\") != \"path.txt\" {\n\t\tt.Error(\"Abs of a relative path failed\")\n\t}\n\t// root\n\tif filepath.Base(\"/\") != \"/\" {\n\t\tt.Error(\"Abs of / failed\")\n\t}\n\n\t// negative\n\tif filepath.Base(\"a/b/c/////path.txt\") != \"path.txt\" {\n\t\tt.Error(\"Abs with multiple separators failed\")\n\t}\n\tif filepath.Base(\"./dir/.././path.txt\") != \"path.txt\" {\n\t\tt.Error(\"Abs with a non-clean path failed\")\n\t}\n\tif filepath.Base(\"//////\") != \"/\" {\n\t\tt.Error(\"Multi-slashed root path failed\")\n\t}\n}", "func Expand(s string) (string, error) {\n\ts = os.ExpandEnv(s)\n\n\tvar err error\n\ts, err = homedir.Expand(s)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn s, nil\n}", "func (dir *Dir) RelPath() string {\n\trel, err := filepath.Rel(dir.repoBase, dir.Path)\n\tif dir.repoBase == \"\" || err != nil {\n\t\treturn dir.BaseName()\n\t}\n\treturn rel\n}", "func (dir *Dir) RelPath() string {\n\trel, err := filepath.Rel(dir.repoBase, dir.Path)\n\tif dir.repoBase == \"\" || err != nil {\n\t\treturn dir.BaseName()\n\t}\n\treturn rel\n}", "func getAbsInstallPath(relPath string) (string, error) {\n\tex, err := os.Executable()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn filepath.Join(filepath.Dir(ex), \"..\", relPath), nil\n}", "func PathIncludeCurrentDir(path string) string {\n\tdir, _ := filepath.Abs(filepath.Dir(os.Args[0]))\n\treturn dir + \"/\" + path\n}", "func baseDir(filename string) string {\n\treturn filepath.Clean(filepath.Dir(filename))\n}", "func (m *Module) relativePath(p *types.Package) string {\n\treturn strings.TrimPrefix(p.Path(), m.Path)\n}", "func (f LocalFS) Base() string {\n\treturn f.basePath\n}", "func (f LocalFS) Base() string {\n\treturn f.basePath\n}", "func expandChartFilesPath(charts []ChartDefinition, baseDir string) {\n\tfor i := range charts {\n\t\tc := &charts[i]\n\t\tc.ValuesPath = expandFilePath(c.ValuesPath, baseDir)\n\t\tc.Path = expandFilePath(c.Path, baseDir)\n\t}\n}", "func Basepath() string {\n\tbase, err := os.Getwd()\n\tif err != nil {\n\t\treturn \".\"\n\t}\n\treturn base\n}", "func RealPath(file string) (string, error) {\n\tif path.IsAbs(file) {\n\t\treturn file, nil\n\t}\n\twd, err := os.Getwd()\n\treturn path.Join(wd, file), err\n}", "func GetFullRepositoryFromPath(path string, base string) string {\n\treturn GetRepositoryFromPath(GetPathWithBase(path, base))\n}", "func relativePath(storePath, shardPath string) (string, error) {\n\tpath, err := filepath.Abs(storePath)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"store abs path: %s\", err)\n\t}\n\n\tfp, err := filepath.Abs(shardPath)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"file abs path: %s\", err)\n\t}\n\n\tname, err := filepath.Rel(path, fp)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"file rel path: %s\", err)\n\t}\n\n\treturn name, nil\n}", "func (r Ref) AbsPath(rootDir string) string {\n\tpath := filepath.Join(rootDir, r.GetPath())\n\tpath, _ = filepath.Abs(path)\n\tpath = filepath.ToSlash(path)\n\treturn path\n}", "func GetPathFromBase64Hash(base64Hash types.Base64Hash, absBasePath config.Path) (string, error) {\n\tif len(base64Hash) < 3 {\n\t\treturn \"\", fmt.Errorf(\"invalid filePath (Base64Hash too short - min 3 characters): %q\", base64Hash)\n\t}\n\tif len(base64Hash) > 255 {\n\t\treturn \"\", fmt.Errorf(\"invalid filePath (Base64Hash too long - max 255 characters): %q\", base64Hash)\n\t}\n\n\tfilePath, err := filepath.Abs(filepath.Join(\n\t\tstring(absBasePath),\n\t\tstring(base64Hash[0:1]),\n\t\tstring(base64Hash[1:2]),\n\t\tstring(base64Hash[2:]),\n\t\t\"file\",\n\t))\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"unable to construct filePath: %w\", err)\n\t}\n\n\t// check if the absolute absBasePath is a prefix of the absolute filePath\n\t// if so, no directory escape has occurred and the filePath is valid\n\t// Note: absBasePath is already absolute\n\tif !strings.HasPrefix(filePath, string(absBasePath)) {\n\t\treturn \"\", fmt.Errorf(\"invalid filePath (not within absBasePath %v): %v\", absBasePath, filePath)\n\t}\n\n\treturn filePath, nil\n}", "func realPath(file string) string {\n\tif string(file[0]) == \"/\" {\n\t\treturn file\n\t}\n\n\tif string(file[0]) != \"/\" {\n\t\tfile = \"/\" + file\n\t}\n\n\t_, filename, _, _ := runtime.Caller(3)\n\tdir := path.Join(path.Dir(filename), file)\n\n\tif _, err := os.Stat(dir); err == nil && strings.HasSuffix(dir, file) {\n\t\treturn dir\n\t}\n\n\tcurrent, err := os.Getwd()\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tdir = file\n\n\tif strings.HasSuffix(dir, current) {\n\t\treturn dir\n\t}\n\n\treturn current + dir\n}", "func basePath() string {\n\t_, currentFile, _, _ := runtime.Caller(2)\n\tbasepath := filepath.Dir(currentFile)\n\treturn basepath\n}", "func RelativeDir(relativePath string) string {\n\t_, f, _, _ := runtime.Caller(1)\n\treturn filepath.Join(filepath.Dir(f), relativePath)\n}", "func baseDir() (string, error) {\n\tdirPath := filepath.Join(os.TempDir(), \"pod\", \"rpctest\")\n\te := os.MkdirAll(dirPath, 0755)\n\treturn dirPath, e\n}", "func expandTilde(path string, homeDirAbsPath AbsPath) string {\n\tswitch {\n\tcase path == \"~\":\n\t\treturn string(homeDirAbsPath)\n\tcase strings.HasPrefix(path, \"~/\"):\n\t\treturn string(homeDirAbsPath.Join(RelPath(path[2:])))\n\tdefault:\n\t\treturn path\n\t}\n}", "func RelPath(source, path string) string {\n\tsrcDir := filepath.Dir(source)\n\tif srcDir != \".\" {\n\t\treturn filepath.Join(srcDir, path)\n\t}\n\treturn path\n}", "func AbsolutePathHelper(rel string, parts ...string) (path string, err error) {\n\tabs, err := filepath.Abs(rel)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn filepath.Join(append([]string{abs}, parts...)...), nil\n}", "func pathJoin(base, leaf string) string {\n\t// Disalllow empty leaf names.\n\tif leaf == \"\" {\n\t\tpanic(\"empty leaf name\")\n\t}\n\n\t// When joining a path to the synchronization root, we don't want to\n\t// concatenate.\n\tif base == \"\" {\n\t\treturn leaf\n\t}\n\n\t// Concatenate the paths.\n\treturn base + \"/\" + leaf\n}", "func makeRelative(child, parent string) string {\n\treturn strings.TrimPrefix(child, dirName(parent))\n}", "func ToExtendedPath(short string) string {\n\t// filepath.Abs has an issue where if the path is just the drive indicator of your CWD, it just returns the CWD. So, we append the / to show that yes, we really mean C: or whatever.\n\tif runtime.GOOS == \"windows\" && len(short) == 2 && RootDriveRegex.MatchString(short) {\n\t\tshort += \"/\"\n\t}\n\n\tshort, err := filepath.Abs(short)\n\tPanicIfErr(err) //TODO: Handle errors better?\n\n\t// ex. C:/dir/file.txt -> \\\\?\\C:\\dir\\file.txt\n\t// ex. \\\\share\\dir\\file.txt -> \\\\?\\UNC\\share\\dir\\file.txt\n\tif runtime.GOOS == \"windows\" { // Only do this on Windows\n\t\tif strings.HasPrefix(short, EXTENDED_PATH_PREFIX) { // already an extended path \\\\?\\C:\\folder\\file.txt or \\\\?\\UNC\\sharename\\folder\\file.txt\n\t\t\treturn strings.Replace(short, `/`, `\\`, -1) // Just ensure it has all backslashes-- Windows can't handle forward-slash anymore in this format.\n\t\t} else if strings.HasPrefix(short, `\\\\`) { // this is a file share (//sharename/folder/file.txt)\n\t\t\t// Steal the first backslash, and then append the prefix. Enforce \\.\n\t\t\treturn strings.Replace(EXTENDED_UNC_PATH_PREFIX+short[1:], `/`, `\\`, -1) // convert to extended UNC path\n\t\t} else { // this is coming from a drive-- capitalize the drive prefix. (C:/folder/file.txt)\n\t\t\tif len(short) >= 2 && RootDriveRegex.MatchString(short[:2]) {\n\t\t\t\tshort = strings.Replace(short, short[:2], strings.ToUpper(short[:2]), 1)\n\t\t\t}\n\t\t\t// Then append the prefix. Enforce \\.\n\t\t\treturn strings.Replace(EXTENDED_PATH_PREFIX+short, `/`, `\\`, -1) // Just append the prefix\n\t\t}\n\t}\n\n\treturn short\n}", "func ExpandHomePath(p string, home string) string {\n\tif strings.HasPrefix(p, homePrefix) {\n\t\treturn path.Join(home, strings.TrimPrefix(p, homePrefix))\n\t}\n\n\treturn p\n}", "func relToPath(path, workingDir string) (string, error) {\n\tworkingDir = util.ConsistentFilepath(workingDir)\n\tpath = util.ConsistentFilepath(path)\n\n\tif !filepath.IsAbs(path) {\n\t\treturn path, nil\n\t}\n\n\tif workingDir == \"\" || !strings.HasPrefix(path, workingDir) {\n\t\treturn path, errNotRelativeToWorkingDir\n\t}\n\n\trelPath, err := filepath.Rel(workingDir, path)\n\tif err != nil {\n\t\treturn path, errors.Wrap(err, errNotRelativeToWorkingDir.Error())\n\t}\n\n\treturn util.ConsistentFilepath(relPath), nil\n}", "func GetAbsPath(p string) (out string, err error) {\n\tif p == \"\" {\n\t\tout, err = os.Getwd()\n\t} else if strings.HasPrefix(p, \"~\") {\n\t\tout, err = homedir.Expand(p)\n\t} else if !filepath.IsAbs(p) {\n\t\tout, err = filepath.Abs(p)\n\t} else {\n\t\tout = p\n\t}\n\treturn\n}", "func getPath(dir string, file os.FileInfo) string {\n\tif !strings.HasSuffix(dir, \"/\") {\n\t\tdir += \"/\"\n\t}\n\treturn dir + file.Name()\n}", "func AbsPath(input string) string {\n\tp := input\n\tif p == \"\" {\n\t\treturn \"\"\n\t}\n\tif p[0] == '~' {\n\t\tp = path.Join(homeDir, p[1:])\n\t}\n\tmatches, _ := filepath.Glob(p)\n\tif len(matches) != 0 {\n\t\tp = matches[0]\n\t}\n\tabs, _ := filepath.Abs(p)\n\treturn abs\n}", "func Abs(path string) string {\n\tabspath, err := filepath.Abs(Expand(path))\n\tif err != nil {\n\t\treturn path\n\t}\n\treturn abspath\n}", "func baseDirs(root string, includePatterns []string) ([]string, []string) {\n\troot = filepath.FromSlash(root)\n\tbases := make([]string, len(includePatterns))\n\tnewincludes := includePatterns[:]\n\tfor i, v := range includePatterns {\n\t\tbdir, trailer := filter.SplitPattern(v)\n\t\tif !filepath.IsAbs(bdir) {\n\t\t\tbdir = filepath.Join(root, filepath.FromSlash(bdir))\n\t\t}\n\t\tif stat, err := os.Lstat(bdir); err == nil {\n\t\t\tif stat.Mode()&os.ModeSymlink != 0 {\n\t\t\t\t// Case 1: The file exists and is a symlink, so we rebase the\n\t\t\t\t// include patterns and the base directory\n\t\t\t\tlnk, err := os.Readlink(bdir)\n\t\t\t\tif err != nil {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif filepath.IsAbs(lnk) {\n\t\t\t\t\tbdir = lnk\n\t\t\t\t} else {\n\t\t\t\t\tbdir = filepath.Join(bdir, lnk)\n\t\t\t\t}\n\t\t\t\tif trailer != \"\" {\n\t\t\t\t\tnewincludes[i] = bdir + \"/\" + trailer\n\t\t\t\t} else {\n\t\t\t\t\tnewincludes[i] = bdir\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t// Case 2: The file exists and is nota symlink, so we leave bdir\n\t\t\t\t// unmodified.\n\t\t\t\tbdir = enclosingDir(bdir)\n\t\t\t\tif bdir == \"\" {\n\t\t\t\t\tbdir = root\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tbdir = enclosingDir(bdir)\n\t\t\tif bdir == \"\" {\n\t\t\t\tbdir = root\n\t\t\t}\n\t\t}\n\t\tbases[i] = bdir\n\t}\n\treturn newincludes, bases\n}", "func expand(path string) ([]string, error) {\n\tif fi, err := os.Stat(path); err != nil {\n\t\treturn nil, err\n\t} else if !fi.IsDir() {\n\t\treturn []string{path}, nil\n\t}\n\n\t// Read files in directory.\n\tfis, err := ioutil.ReadDir(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Iterate over files and expand.\n\texpanded := make([]string, 0, len(fis))\n\tfor _, fi := range fis {\n\t\ta, err := expand(filepath.Join(path, fi.Name()))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\texpanded = append(expanded, a...)\n\t}\n\treturn expanded, nil\n}", "func RelPath(targpath string) string {\n\tbasepath, _ := filepath.Abs(\"./\")\n\trel, _ := filepath.Rel(basepath, targpath)\n\treturn strings.Replace(rel, `\\`, `/`, -1)\n}", "func ToAbsolutePath(base, path string) string {\n\tif filepath.IsAbs(path) {\n\t\treturn path\n\t}\n\treturn filepath.Clean(filepath.Join(base, path))\n}", "func (s *FileSet) Relative(basedir string) (*FileSet, error) {\n\tresult := NewFileSet()\n\tfor f := range s.files {\n\t\trel, err := filepath.Rel(basedir, f)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresult.Add(rel)\n\t}\n\treturn result, nil\n}", "func AddBase(path string) string {\n\treturn fmt.Sprintf(\"%s%s\", baseURL, path)\n}", "func cleanAndExpandPath(path string) string {\n\t// Expand initial ~ to OS specific home directory.\n\tif strings.HasPrefix(path, \"~\") {\n\t\thomeDir := filepath.Dir(defaultHomeDir)\n\t\tpath = strings.Replace(path, \"~\", homeDir, 1)\n\t}\n\n\t// NOTE: The os.ExpandEnv doesn't work with Windows-style %VARIABLE%,\n\t// but the variables can still be expanded via POSIX-style $VARIABLE.\n\treturn filepath.Clean(os.ExpandEnv(path))\n}", "func TestNormalizePaths(t *testing.T) {\n\ttype testNormalizePathsTestCases []struct {\n\t\trefPath string\n\t\tbase string\n\t\texpOutput string\n\t}\n\n\ttestCases := func() testNormalizePathsTestCases {\n\t\ttestCases := testNormalizePathsTestCases{\n\t\t\t{\n\t\t\t\t// http basePath, absolute refPath\n\t\t\t\trefPath: \"http://www.anotherexample.com/another/base/path/swagger.json#/definitions/Pet\",\n\t\t\t\tbase: \"http://www.example.com/base/path/swagger.json\",\n\t\t\t\texpOutput: \"http://www.anotherexample.com/another/base/path/swagger.json#/definitions/Pet\",\n\t\t\t},\n\t\t\t{\n\t\t\t\t// http basePath, relative refPath\n\t\t\t\trefPath: \"another/base/path/swagger.json#/definitions/Pet\",\n\t\t\t\tbase: \"http://www.example.com/base/path/swagger.json\",\n\t\t\t\texpOutput: \"http://www.example.com/base/path/another/base/path/swagger.json#/definitions/Pet\",\n\t\t\t},\n\t\t}\n\t\tif runtime.GOOS == \"windows\" {\n\t\t\ttestCases = append(testCases, testNormalizePathsTestCases{\n\t\t\t\t{\n\t\t\t\t\t// file basePath, absolute refPath, no fragment\n\t\t\t\t\trefPath: `C:\\another\\base\\path.json`,\n\t\t\t\t\tbase: `C:\\base\\path.json`,\n\t\t\t\t\texpOutput: `c:\\another\\base\\path.json`,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\t// file basePath, absolute refPath\n\t\t\t\t\trefPath: `C:\\another\\base\\path.json#/definitions/Pet`,\n\t\t\t\t\tbase: `C:\\base\\path.json`,\n\t\t\t\t\texpOutput: `c:\\another\\base\\path.json#/definitions/Pet`,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\t// file basePath, relative refPath\n\t\t\t\t\trefPath: `another\\base\\path.json#/definitions/Pet`,\n\t\t\t\t\tbase: `C:\\base\\path.json`,\n\t\t\t\t\texpOutput: `c:\\base\\another\\base\\path.json#/definitions/Pet`,\n\t\t\t\t},\n\t\t\t}...)\n\t\t\treturn testCases\n\t\t}\n\t\t// linux case\n\t\ttestCases = append(testCases, testNormalizePathsTestCases{\n\t\t\t{\n\t\t\t\t// file basePath, absolute refPath, no fragment\n\t\t\t\trefPath: \"/another/base/path.json\",\n\t\t\t\tbase: \"/base/path.json\",\n\t\t\t\texpOutput: \"/another/base/path.json\",\n\t\t\t},\n\t\t\t{\n\t\t\t\t// file basePath, absolute refPath\n\t\t\t\trefPath: \"/another/base/path.json#/definitions/Pet\",\n\t\t\t\tbase: \"/base/path.json\",\n\t\t\t\texpOutput: \"/another/base/path.json#/definitions/Pet\",\n\t\t\t},\n\t\t\t{\n\t\t\t\t// file basePath, relative refPath\n\t\t\t\trefPath: \"another/base/path.json#/definitions/Pet\",\n\t\t\t\tbase: \"/base/path.json\",\n\t\t\t\texpOutput: \"/base/another/base/path.json#/definitions/Pet\",\n\t\t\t},\n\t\t}...)\n\t\treturn testCases\n\t}()\n\n\tfor _, tcase := range testCases {\n\t\tout := normalizePaths(tcase.refPath, tcase.base)\n\t\tassert.Equal(t, tcase.expOutput, out)\n\t}\n}", "func (fs *Bcpfs) Realpath(p string) string {\n\tif !fs.IsServicePath(p) {\n\t\treturn slashpath.Join(\n\t\t\tfs.OrgUnitDir, p,\n\t\t)\n\t}\n\n\t// Path `p` must be a service path.\n\tparts := strings.Split(p, \"/\")\n\tif len(parts) < 2 {\n\t\t// Path too short. Return an empty string to indicate the\n\t\t// problem instead of full error handling, because paths must\n\t\t// not be too short for a valid configuration.\n\t\treturn \"\"\n\t}\n\n\tif fs.IsFacilityPath(p) {\n\t\treturn slashpath.Join(append(\n\t\t\t[]string{fs.ServiceDir}, parts[1:]...,\n\t\t)...)\n\t}\n\n\tou := parts[0]\n\tsrv := parts[1]\n\trest := parts[2:]\n\treturn slashpath.Join(append(\n\t\t[]string{fs.ServiceDir, srv, ou}, rest...,\n\t)...)\n}", "func BaseLocation() (string, error) {\n\tex, err := os.Executable()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"undetermined location of own executable: %s\", err)\n\t}\n\tex = filepath.Dir(ex)\n\t// TODO(yifan): Hardcode the parent dir name for now.\n\tif filepath.Base(ex) != \"installer\" {\n\t\treturn \"\", fmt.Errorf(\"%s executable in unknown location: %s\", filepath.Base(ex), err)\n\t}\n\treturn filepath.Dir(ex), nil\n}", "func Join(path string, dir string) string {\n\treturn filepath.Join(path, dir)\n}", "func (i *IndexBuilder) relativeDir() string {\n\trelativeDir := i.cfg.RelativeDir\n\tif relativeDir == \"/\" {\n\t\treturn relativeDir\n\t}\n\treturn fmt.Sprintf(\"/%s/\", relativeDir)\n}" ]
[ "0.7221674", "0.63791376", "0.62171954", "0.6203149", "0.60714793", "0.6038171", "0.60058975", "0.5962914", "0.58609414", "0.5853772", "0.5846553", "0.58370256", "0.58066094", "0.5787698", "0.5785019", "0.57407224", "0.57359976", "0.5726426", "0.567552", "0.5660976", "0.56590235", "0.56061596", "0.55993736", "0.55941147", "0.55925745", "0.5576248", "0.5519725", "0.54966795", "0.54354846", "0.54190296", "0.54081243", "0.54077625", "0.53518146", "0.5341031", "0.53367716", "0.5329633", "0.52383214", "0.5220128", "0.52199525", "0.52021486", "0.5198822", "0.51884234", "0.51791626", "0.51750875", "0.5169584", "0.5166068", "0.5147754", "0.51468945", "0.5143286", "0.5117156", "0.5101875", "0.50889397", "0.50889397", "0.50805765", "0.50633603", "0.5057555", "0.5054463", "0.5049263", "0.5049263", "0.5048401", "0.50449735", "0.5006407", "0.50039613", "0.5002664", "0.5002664", "0.49951246", "0.49829423", "0.49769446", "0.49735767", "0.49722758", "0.49699867", "0.49688998", "0.49627092", "0.4961851", "0.49589083", "0.4939562", "0.49200332", "0.49150434", "0.49138168", "0.4909202", "0.4898828", "0.4898116", "0.48971283", "0.4892276", "0.48900947", "0.48886743", "0.48690552", "0.48688293", "0.48662126", "0.48630372", "0.48590654", "0.4841772", "0.48317054", "0.48310205", "0.48277768", "0.4821929", "0.4817662", "0.48160127", "0.4814675", "0.48138043" ]
0.79213846
0
isParentPath returns true if path is a child or a descendant of parent path. Both inputs need to be absolute paths.
func isParentPath(parent, path string) bool { rel, err := filepath.Rel(parent, path) return err == nil && !strings.HasPrefix(rel, "..") }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func pathIsChild(parent, child string) bool {\n\tif !strings.HasPrefix(child, parent) {\n\t\treturn false\n\t}\n\trel := child[len(parent):]\n\trel = strings.Trim(rel, \"/\")\n\treturn !strings.Contains(rel, \"/\")\n}", "func (d *Driver) isParent(id, parent string) bool {\n\t// TODO (maybe): this function is called a lot and does lots of sub-routine calls and I/O.\n\t// One might want to cache parent but should first be confirmed via profiling that really noticable performance cost\n\tlogrus.Debugf(\"secureoverlay2: isParent called w. id: %s, parent: %s\", id, parent)\n\tlowers, err := d.getLowerDirs(id)\n\tif err != nil {\n\t\treturn false\n\t}\n\tif parent == \"\" && len(lowers) > 0 {\n\t\treturn false\n\t}\n\n\tparentDir := d.dir(parent)\n\tvar ld string\n\tif len(lowers) > 0 {\n\t\tld = filepath.Dir(lowers[0])\n\t}\n\tif ld == \"\" && parent == \"\" {\n\t\treturn true\n\t}\n\treturn ld == parentDir\n}", "func IsParentDir(parentDir, childPath string) bool {\n\treturn mgutil.IsParentDir(parentDir, childPath)\n}", "func (p path) hasValidParent() bool {\n\treturn len(p.path) >= 2 && p.parentPath().isValid()\n}", "func isPathChild(a, b []string) bool {\n\t// If b does not have a greater path length than a, it cannot be a child. If\n\t// b has more than one element than a, it must be at least a grandchild.\n\tif len(b) <= len(a) || len(b) > len(a)+1 {\n\t\treturn false\n\t}\n\n\tfor i := range a {\n\t\tif a[i] != b[i] {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (p path) parentPath() *path {\n\treturn &path{p.FolderBranch, p.path[:len(p.path)-1]}\n}", "func (p *path) IsSubPath(path string) bool {\n\thomedir := p.HomeDir()\n\tpath = p.Resolve(path)\n\trel, err := filepath.Rel(homedir, path)\n\tif err != nil {\n\t\treturn false\n\t}\n\n\treturn !strings.Contains(rel, \"..\")\n}", "func (parent *Inode) isParentOf(inode *Inode) bool {\n\treturn inode.Parent != nil && (parent == inode.Parent || parent.isParentOf(inode.Parent))\n}", "func (parent *Inode) isParentOf(inode *Inode) bool {\n\treturn inode.Parent != nil && (parent == inode.Parent || parent.isParentOf(inode.Parent))\n}", "func IsRelativePath(path string) bool {\n\treturn !(types.StartWith(path, \"./\") ||\n\t\ttypes.StartWith(path, \".\\\\\") ||\n\t\ttypes.StartWith(path, \"~/\") ||\n\t\ttypes.StartWith(path, \"~\\\\\") ||\n\t\ttypes.StartWith(path, \"/\") ||\n\t\tIsWindowsRootpath(path))\n}", "func isPath(path string) bool {\n\treturn strings.HasPrefix(path, \"~\") ||\n\t\tstrings.HasPrefix(path, \".\") ||\n\t\tstrings.HasPrefix(path, \"/\")\n}", "func isChild(child, parent string) bool {\n\treturn strings.HasPrefix(child, dirName(parent))\n}", "func (ns *Namespace) CanSetParent(p *Namespace) string {\n\tif p == nil {\n\t\treturn \"\"\n\t}\n\n\t// Simple case\n\tif p == ns {\n\t\treturn fmt.Sprintf(\"%q cannot be set as its own parent\", p.name)\n\t}\n\n\t// Check for cycles; see if the current namespace (the proposed child) is already an ancestor of\n\t// the proposed parent. Start at the end of the ancestry (e.g. at the proposed parent) and work\n\t// our way up to the root.\n\tancestors := p.AncestryNames()\n\tcycle := []string{}\n\tfound := false\n\tfor i := len(ancestors) - 1; !found && i >= 0; i-- {\n\t\tcycle = append(cycle, ancestors[i])\n\t\tfound = (ancestors[i] == ns.name)\n\t}\n\tif found {\n\t\treturn fmt.Sprintf(\"cycle when making %q the parent of %q: current ancestry is %s\",\n\t\t\tp.name, ns.name, strings.Join(cycle, \" -> \"))\n\t}\n\n\treturn \"\"\n}", "func (o *IpamNetworkDataData) HasParentNetworkPath() bool {\n\tif o != nil && o.ParentNetworkPath != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (path PathImpl) Parent() Path {\n\t// path.String() can't be empty\n\tparent, _ := New(path, \"..\")\n\treturn parent\n}", "func (b *Being) IsParentOf(with string) bool {\n\tfor _, id := range b.Children {\n\t\tif id == with {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (e *Entry) parentPath() string {\n\tif len(e.path) == 0 {\n\t\tpanic(\"trying to get the parentPath of the root\")\n\t}\n\tparts := make([]string, 1, len(e.path))\n\tparts[0] = e.root.path\n\tparts = append(parts, e.path[:len(e.path)-1]...)\n\treturn filepath.Join(parts...)\n}", "func isDirectEntryChild(p, c *yang.Entry, compressPaths bool) bool {\n\tppp := strings.Split(p.Path(), \"/\")\n\tcpp := strings.Split(c.Path(), \"/\")\n\tdc := isPathChild(ppp, cpp)\n\n\t// If we are not compressing paths, then directly return whether the child\n\t// is a path of the parent.\n\tif !compressPaths {\n\t\treturn dc\n\t}\n\n\t// If the length of the child path is greater than two larger than the\n\t// parent path, then this means that it cannot be a direct child, since all\n\t// path compression will remove only one level of hierarchy (config/state or\n\t// a surrounding container at maximum). We also check that the length of\n\t// the child path is more specific than or equal to the length of the parent\n\t// path in which case this cannot be a child.\n\tif len(cpp) > len(ppp)+2 || len(cpp) <= len(ppp) {\n\t\treturn false\n\t}\n\n\tif isConfigState(c.Parent) {\n\t\t// If the parent of this entity was the config/state container, then this\n\t\t// level of the hierarchy will have been removed so we check whether the\n\t\t// parent of both are equal and return this.\n\t\treturn p.Path() == c.Parent.Parent.Path()\n\t}\n\n\t// If the child is a list, then we check whether the parent has only one\n\t// child (i.e., is a surrounding container) and then check whether the\n\t// single child is the child we were provided.\n\tif c.IsList() {\n\t\tppe, ok := p.Dir[c.Parent.Name]\n\t\tif !ok {\n\t\t\t// Can't be a valid child because the parent of the entity doesn't exist\n\t\t\t// within this container.\n\t\t\treturn false\n\t\t}\n\t\tif !hasOnlyChild(ppe) {\n\t\t\treturn false\n\t\t}\n\n\t\t// We are guaranteed to have 1 child (and not zero) since hasOnlyChild will\n\t\t// return false for directories with 0 children.\n\t\treturn children(ppe)[0].Path() == c.Path()\n\t}\n\n\treturn dc\n}", "func (p *Path) Parent() (*Path, error) {\n\tpth, err := p.Absolute()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"get parent failed\")\n\t}\n\tdir := filepath.Dir(pth.Path)\n\tnewP := New(dir)\n\treturn newP, nil\n}", "func (n *Node) ParentPath() string {\n\treturn n.lu.InternalPath(n.SpaceID, n.ParentID)\n}", "func IsSubPathFolder(path string, paths []string) (bool, error) {\n\tfound := false\n\tfor _, sync := range paths {\n\t\trel, err := filepath.Rel(sync, path)\n\t\tif err != nil {\n\t\t\tlog.Debugf(\"error making rel '%s' and '%s'\", sync, path)\n\t\t\treturn false, errors.Wrap(err, \"\")\n\t\t}\n\t\tif strings.HasPrefix(rel, \"..\") {\n\t\t\tcontinue\n\t\t}\n\t\tfound = true\n\t\tif rel != \".\" {\n\t\t\treturn true, nil\n\t\t}\n\t}\n\tif found {\n\t\treturn false, nil\n\t}\n\treturn false, errors.New(\"not found\")\n}", "func HasParent() predicate.Location {\n\treturn predicate.Location(func(s *sql.Selector) {\n\t\tstep := sqlgraph.NewStep(\n\t\t\tsqlgraph.From(Table, FieldID),\n\t\t\tsqlgraph.To(ParentTable, FieldID),\n\t\t\tsqlgraph.Edge(sqlgraph.M2O, true, ParentTable, ParentColumn),\n\t\t)\n\t\tsqlgraph.HasNeighbors(s, step)\n\t},\n\t)\n}", "func TestIsSubpath(t *testing.T) {\n\ttcs := []struct {\n\t\t// Two absolute paths.\n\t\tpathA string\n\t\tpathB string\n\n\t\t// Whether pathA is a subpath of pathB.\n\t\twantIsSubpath bool\n\n\t\t// Relative path from pathA to pathB. Only checked if\n\t\t// wantIsSubpath is true.\n\t\twantRelpath string\n\t}{\n\t\t{\n\t\t\tpathA: \"/foo/bar/baz\",\n\t\t\tpathB: \"/foo\",\n\t\t\twantIsSubpath: true,\n\t\t\twantRelpath: \"bar/baz\",\n\t\t},\n\t\t{\n\t\t\tpathA: \"/foo\",\n\t\t\tpathB: \"/foo/bar/baz\",\n\t\t\twantIsSubpath: false,\n\t\t},\n\t\t{\n\t\t\tpathA: \"/foo\",\n\t\t\tpathB: \"/foo\",\n\t\t\twantIsSubpath: false,\n\t\t},\n\t\t{\n\t\t\tpathA: \"/foobar\",\n\t\t\tpathB: \"/foo\",\n\t\t\twantIsSubpath: false,\n\t\t},\n\t\t{\n\t\t\tpathA: \"/foo\",\n\t\t\tpathB: \"/foobar\",\n\t\t\twantIsSubpath: false,\n\t\t},\n\t\t{\n\t\t\tpathA: \"/foo\",\n\t\t\tpathB: \"/foobar\",\n\t\t\twantIsSubpath: false,\n\t\t},\n\t\t{\n\t\t\tpathA: \"/\",\n\t\t\tpathB: \"/foo\",\n\t\t\twantIsSubpath: false,\n\t\t},\n\t\t{\n\t\t\tpathA: \"/foo\",\n\t\t\tpathB: \"/\",\n\t\t\twantIsSubpath: true,\n\t\t\twantRelpath: \"foo\",\n\t\t},\n\t\t{\n\t\t\tpathA: \"/foo/bar/../bar\",\n\t\t\tpathB: \"/foo\",\n\t\t\twantIsSubpath: true,\n\t\t\twantRelpath: \"bar\",\n\t\t},\n\t\t{\n\t\t\tpathA: \"/foo/bar\",\n\t\t\tpathB: \"/foo/../foo\",\n\t\t\twantIsSubpath: true,\n\t\t\twantRelpath: \"bar\",\n\t\t},\n\t}\n\n\tfor _, tc := range tcs {\n\t\tgotRelpath, gotIsSubpath := IsSubpath(tc.pathA, tc.pathB)\n\t\tif gotRelpath != tc.wantRelpath || gotIsSubpath != tc.wantIsSubpath {\n\t\t\tt.Errorf(\"IsSubpath(%q, %q) got %q %t, want %q %t\", tc.pathA, tc.pathB, gotRelpath, gotIsSubpath, tc.wantRelpath, tc.wantIsSubpath)\n\t\t}\n\t}\n}", "func isFsCgroupParent(groupPath string) bool {\n\tfi, err := os.Lstat(filepath.Join(groupPath, \"docker\"))\n\tif err != nil {\n\t\treturn false\n\t}\n\n\treturn fi.IsDir()\n}", "func inParent() bool {\n\treturn !inChild()\n}", "func (command PathCommand) IsRelative() bool {\n\tswitch command {\n\tcase MoveToRelative, LineToRelative:\n\t\treturn true\n\tdefault:\n\t\treturn false\n\t}\n\tpanic(\"Not reachable\")\n}", "func (g *gnmiPath) isPathElemPath() bool {\n\treturn g.pathElemPath != nil\n}", "func (fe *FileEntry) HasParent(parent *FileEntry) bool {\n\t// Exceptional case\n\tif fe.Inode == fuseops.RootInodeID {\n\t\treturn false\n\t}\n\tif parent.File == nil {\n\t\treturn fe.HasParentID(\"\")\n\t}\n\treturn fe.HasParentID(parent.File.ID)\n}", "func (node *Node) MatchesPath(path string) bool {\n\tif path == \"\" {\n\t\treturn true\n\t}\n\n\tsplit := strings.Split(path, \"/\")\n\tcurrent := split[len(split)-1]\n\tdirectChild := strings.HasPrefix(current, \".\")\n\tcurrent = strings.TrimPrefix(current, \".\")\n\n\tif len(split) == 1 {\n\t\treturn current == node.Data\n\t} else if current != node.Data {\n\t\treturn false\n\t}\n\n\tnewPath := strings.Join(split[:len(split)-1], \"/\")\n\n\tif directChild == true {\n\t\treturn node.Parent.MatchesPath(newPath)\n\t}\n\n\tparentName := strings.TrimPrefix(split[len(split)-2], \".\")\n\tparent := node.FindTagReverse(parentName)\n\n\tif parent != nil {\n\t\treturn parent.MatchesPath(newPath)\n\t}\n\n\treturn false\n}", "func isAbsolutePath(path string) bool {\n\treturn strings.HasPrefix(path, \"/\")\n}", "func imageIsParent(store storage.Store, topLayer string) (bool, error) {\n\tchildren, err := getChildren(store, topLayer)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treturn len(children) > 0, nil\n}", "func (n *TreeNode) HasParent() bool {\n\treturn n.parent != nil && n.parent.Runs() > 0\n}", "func (d *Distro) IsParent(s *evergreen.Settings) bool {\n\tif s == nil {\n\t\tvar err error\n\t\ts, err = evergreen.GetConfig()\n\t\tif err != nil {\n\t\t\tgrip.Critical(\"error retrieving settings object\")\n\t\t\treturn false\n\t\t}\n\t}\n\tfor _, p := range s.ContainerPools.Pools {\n\t\tif d.Id == p.Distro {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (d *Distro) IsParent(s *evergreen.Settings) bool {\n\tif s == nil {\n\t\tvar err error\n\t\ts, err = evergreen.GetConfig()\n\t\tif err != nil {\n\t\t\tgrip.Critical(\"error retrieving settings object\")\n\t\t\treturn false\n\t\t}\n\t}\n\tfor _, p := range s.ContainerPools.Pools {\n\t\tif d.Id == p.Distro {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (fe *FileEntry) HasParentID(parentID string) bool {\n\t// Exceptional case\n\tif fe.Inode == fuseops.RootInodeID {\n\t\treturn false\n\t}\n\tif parentID == \"\" {\n\t\tif fe.File == nil || len(fe.File.Parents) == 0 { // We are looking in root\n\t\t\treturn true\n\t\t}\n\t\treturn false\n\t}\n\tif fe.File == nil { // Case gid is not empty and GFile is null\n\t\treturn false\n\t}\n\tfor _, pgid := range fe.File.Parents {\n\t\tif pgid == parentID {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (me TdtypeType) IsPath() bool { return me.String() == \"path\" }", "func isPathy(name string) bool {\n\treturn strings.Contains(name, string(filepath.Separator))\n}", "func (o *FileObject) HasParent() bool {\n\treturn o.Parent != nil\n}", "func (d UserData) HasParent() bool {\n\treturn d.ModelData.Has(models.NewFieldName(\"Parent\", \"parent_id\"))\n}", "func (e DecodeErr) IsPlaceParent(p string) bool {\r\n\treturn e.Place.Parent == p\r\n}", "func IsRootPath(path string) bool {\n\tif l := len(path); l > 0 {\n\t\tswitch OperateSystem() {\n\t\tcase WINDOWS:\n\t\t\treturn IsWindowsRootpath(path)\n\t\tcase LINUX, DARWIN, FREEBSD, SOLARIS, ANDROID:\n\t\t\treturn l == 1 && path[0] == '/'\n\t\t}\n\t}\n\treturn false\n}", "func isParentMetaExist(path string) (bool, error) {\n\n\t_, err := os.Stat(path)\n\tif err == nil { return true, nil }\n\tif os.IsNotExist(err) { return false, nil }\n\treturn true, err\n}", "func HasParentWith(preds ...predicate.Location) predicate.Location {\n\treturn predicate.Location(func(s *sql.Selector) {\n\t\tstep := sqlgraph.NewStep(\n\t\t\tsqlgraph.From(Table, FieldID),\n\t\t\tsqlgraph.To(Table, FieldID),\n\t\t\tsqlgraph.Edge(sqlgraph.M2O, true, ParentTable, ParentColumn),\n\t\t)\n\t\tsqlgraph.HasNeighborsWith(s, step, func(s *sql.Selector) {\n\t\t\tfor _, p := range preds {\n\t\t\t\tp(s)\n\t\t\t}\n\t\t})\n\t},\n\t)\n}", "func (o *IpamNetworkDataData) GetParentNetworkPathOk() (*string, bool) {\n\tif o == nil || o.ParentNetworkPath == nil {\n\t\treturn nil, false\n\t}\n\treturn o.ParentNetworkPath, true\n}", "func (r *Root) ParentByPath(c *web.Client, origin Parent, path Path) (p Parent, base string, err error) {\n\tdir := path.Dir()\n\tlogger.Debug1(\"dir: %v\", dir)\n\to, err := r.ObjectByPath(c, origin, path.Dir())\n\tif err != nil {\n\t\treturn nil, \"\", errors.Errorf(\n\t\t\t\"failed to get parent directory %q of path %q: %v\",\n\t\t\tdir, path, err)\n\t}\n\tp, ok := o.(Parent)\n\tif !ok {\n\t\treturn nil, \"\", errors.Errorf(\n\t\t\t\"object %v exists but is not a parent\", path)\n\t}\n\treturn p, Basename(path), nil\n}", "func (state *BuildState) IsOriginalTargetOrParent(target *BuildTarget) bool {\n\tif state.IsOriginalTarget(target) {\n\t\treturn true\n\t} else if parent := target.Parent(state.Graph); parent != nil {\n\t\treturn state.IsOriginalTarget(parent)\n\t}\n\treturn false\n}", "func (c char) isPathStart() bool {\n\treturn (c.isCorner() || c.isHorizontal() || c.isVertical() || c.isArrowHorizontalLeft() || c.isArrowVerticalUp() || c.isDiagonal()) && !c.isTick() && !c.isDot()\n}", "func (o *IpamNetworkDataData) SetParentNetworkPath(v string) {\n\to.ParentNetworkPath = &v\n}", "func (code Code) checkCodePath() error {\n\tpaths := strings.Split(code.codeStr.String(), \".\")\n\tif len(paths) == 1 {\n\t\treturn nil\n\t}\n\tif code.Parent == nil {\n\t\tif len(paths) > 1 {\n\t\t\treturn fmt.Errorf(\"expected no parent paths: %#v\", code.codeStr)\n\t\t}\n\t} else {\n\t\tparent := *code.Parent\n\t\tparentPath := paths[len(paths)-2]\n\t\tif parentPath != parent.codeStr.String() {\n\t\t\treturn fmt.Errorf(\"got %#v but expected a path to parent %#v for %#v\", parentPath, parent.codeStr, code.codeStr)\n\t\t}\n\t}\n\treturn nil\n}", "func HasParent() predicate.OutcomeOverview {\n\treturn predicate.OutcomeOverview(func(s *sql.Selector) {\n\t\tstep := sqlgraph.NewStep(\n\t\t\tsqlgraph.From(Table, FieldID),\n\t\t\tsqlgraph.To(ParentTable, FieldID),\n\t\t\tsqlgraph.Edge(sqlgraph.M2O, true, ParentTable, ParentColumn),\n\t\t)\n\t\tsqlgraph.HasNeighbors(s, step)\n\t})\n}", "func (obj *errorStruct) HasParent() bool {\n\treturn obj.parent != nil\n}", "func (t Task) IsChild() bool {\n\treturn t.Parent != 0\n}", "func isFilePath(path string) bool {\n\t// when split returns dir and file, splitting path on the final \"/\"\n\t// check if file is not empty to classify that path as a file path\n\t_, file := filepath.Split(path)\n\tif len(file) == 0 {\n\t\treturn false\n\t}\n\treturn true\n}", "func isPathPrefix(path, pre string) bool {\n\tpathlen, prflen := len(path), len(pre)\n\tif pathlen < prflen || path[0:prflen] != pre {\n\t\treturn false\n\t}\n\n\treturn prflen == pathlen || strings.Index(path[prflen:], \"/\") == 0\n}", "func IsAbsPath(aPath string) bool {\n\treturn path.IsAbs(aPath)\n}", "func ParentsFilter(path string) FilterFunc {\n\treturn func(m *MountInfo) (bool, bool) {\n\t\tskip := !strings.HasPrefix(path, m.MountPoint)\n\t\treturn skip, false\n\t}\n}", "func (db *PSQL) IsOrganizationParent(organizationID string) (bool, string, error) {\n\treturn false, \"\", nil\n}", "func createParentPath(path string, conn *zk.Conn, acl []zk.ACL) error {\n\tparts := strings.Split(path, \"/\")\n\tprePath := \"\"\n\tfor _, p := range parts[1 : len(parts)-1] {\n\t\tprePath += \"/\" + p\n\t\t_, err := conn.Create(prePath, []byte{}, 0, acl)\n\t\tif err != nil && err != zk.ErrNodeExists {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (d UserData) HasParentName() bool {\n\treturn d.ModelData.Has(models.NewFieldName(\"ParentName\", \"parent_name\"))\n}", "func (s *Store) hasParent() bool {\r\n\treturn s.parent != nil\r\n}", "func isPathType(folderName string) bool {\n\tlastCharacter := folderName[len(folderName)-1]\n\tif os.IsPathSeparator(lastCharacter) {\n\t\treturn true\n\t}\n\treturn false\n}", "func (n *BaseNode) ParentIsStopped(m NodeMetadata) {\n\tn.m.Lock()\n\tdefer n.m.Unlock()\n\tif _, ok := n.parents[m.Name]; !ok {\n\t\treturn\n\t}\n\tdelete(n.parentsStarted, m.Name)\n\tif len(n.parentsStarted) == 0 && (n.o.AutoStop == nil || n.o.AutoStop.WhenAllParentsAreStopped) {\n\t\tn.Stop()\n\t}\n}", "func (pp packagePath) IsAbs() bool {\n\tfirstSlash := strings.IndexRune(string(pp), '/')\n\tvar firstPart string\n\tswitch {\n\tcase firstSlash == -1:\n\t\tfirstPart = string(pp)\n\tdefault:\n\t\tfirstPart = string(pp)[:firstSlash]\n\t}\n\n\treturn firstPart != \".\" && firstPart != \"..\"\n}", "func (b *Being) IsCloseRelativeOf(with string) bool {\n\tclose := false\n\tclose = close || b.IsChildOf(with)\n\tclose = close || b.IsParentOf(with)\n\tclose = close || b.IsSiblingOf(with)\n\treturn close\n}", "func PathInPackage(path, pkg string) bool {\n\treturn strings.Contains(path, \"/\"+pkg+\"/\") || strings.HasPrefix(path, pkg+\"/\")\n}", "func (p path) isValid() bool {\n\tif len(p.path) < 1 {\n\t\treturn false\n\t}\n\n\tfor _, n := range p.path {\n\t\tif !n.isValid() {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func HasParent() predicate.BaselineClass {\n\treturn predicate.BaselineClass(func(s *sql.Selector) {\n\t\tstep := sqlgraph.NewStep(\n\t\t\tsqlgraph.From(Table, FieldID),\n\t\t\tsqlgraph.To(ParentTable, FieldID),\n\t\t\tsqlgraph.Edge(sqlgraph.M2O, true, ParentTable, ParentColumn),\n\t\t)\n\t\tsqlgraph.HasNeighbors(s, step)\n\t})\n}", "func (mounter *Mounter) PathIsDevice(pathname string) (bool, error) {\n\treturn pathIsDevice(pathname)\n}", "func isSamePath(a string, b string) bool {\n\treturn strings.TrimRight(a, \"/\") == strings.TrimRight(b, \"/\")\n}", "func (node *GoValueNode) HasParent() bool {\n\n\treturn node.parentNode != nil\n}", "func isAbs(path string) bool {\n\treturn filepath.IsAbs(path) || strings.HasPrefix(path, string(separator))\n}", "func hasPathPrefix(path, prefix string) bool {\n\tsep := string(filepath.Separator)\n\tpathParts := strings.Split(filepath.Clean(path), sep)\n\tprefixParts := strings.Split(filepath.Clean(prefix), sep)\n\n\tif len(prefixParts) > len(pathParts) {\n\t\treturn false\n\t}\n\tfor index, prefixItem := range prefixParts {\n\t\tif prefixItem != pathParts[index] {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func HasParent() predicate.BaselineMeasureDenom {\n\treturn predicate.BaselineMeasureDenom(func(s *sql.Selector) {\n\t\tstep := sqlgraph.NewStep(\n\t\t\tsqlgraph.From(Table, FieldID),\n\t\t\tsqlgraph.To(ParentTable, FieldID),\n\t\t\tsqlgraph.Edge(sqlgraph.M2O, true, ParentTable, ParentColumn),\n\t\t)\n\t\tsqlgraph.HasNeighbors(s, step)\n\t})\n}", "func pathIsLess(ctx context.Context, nbf *types.NomsBinFormat, p1, p2 types.Path) (bool, error) {\n\tfor i, pp1 := range p1 {\n\t\tif len(p2) == i {\n\t\t\treturn false, nil // p1 > p2\n\t\t}\n\n\t\tidx, err := pathPartCompare(ctx, nbf, pp1, p2[i])\n\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\n\t\tswitch idx {\n\t\tcase -1:\n\t\t\treturn true, nil // p1 < p2\n\t\tcase 1:\n\t\t\treturn false, nil // p1 > p2\n\t\t}\n\t}\n\n\treturn len(p2) > len(p1), nil // if true p1 < p2, else p1 == p2\n}", "func (n *BaseNode) ParentIsStarted(m NodeMetadata) {\n\tn.m.Lock()\n\tdefer n.m.Unlock()\n\tif _, ok := n.parents[m.Name]; !ok {\n\t\treturn\n\t}\n\tn.parentsStarted[m.Name] = true\n}", "func IsObjectPath(path string) bool {\n\t//TODO add support for domain style paths when the domain support is going to be merged\n\tparts := strings.SplitN(strings.Trim(path, \"/\"), \"/\", 2)\n\treturn len(parts) == 2 && parts[1] != \"\"\n}", "func (p *Path) IsAbs() bool {\n\treturn filepath.IsAbs(p.Path)\n}", "func HasParent() predicate.ResultsDefinition {\n\treturn predicate.ResultsDefinition(func(s *sql.Selector) {\n\t\tstep := sqlgraph.NewStep(\n\t\t\tsqlgraph.From(Table, FieldID),\n\t\t\tsqlgraph.To(ParentTable, FieldID),\n\t\t\tsqlgraph.Edge(sqlgraph.O2O, true, ParentTable, ParentColumn),\n\t\t)\n\t\tsqlgraph.HasNeighbors(s, step)\n\t})\n}", "func isRoot(path string) bool {\n\tif runtime.GOOS != \"windows\" {\n\t\treturn path == \"/\"\n\t}\n\tswitch len(path) {\n\tcase 1:\n\t\treturn os.IsPathSeparator(path[0])\n\tcase 3:\n\t\treturn path[1] == ':' && os.IsPathSeparator(path[2])\n\t}\n\treturn false\n}", "func (o *TenantWithOfferWeb) HasParentId() bool {\n\tif o != nil && o.ParentId != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (iter *Iteration) isValidFilepath(path string) bool {\n\tif iter == nil {\n\t\treturn false\n\t}\n\treturn strings.HasPrefix(strings.ToLower(path), strings.ToLower(iter.Dir))\n}", "func (s *Store) SetParent(parent *Store) bool {\r\n\thasParent := s.hasParent()\r\n\ts.parent = parent\r\n\treturn hasParent\r\n}", "func (v *Service) IsSetParentID() bool {\n\treturn v != nil && v.ParentID != nil\n}", "func PathIsInDir(p, dir string) bool {\n\tp = filepath.Clean(p)\n\tdir = filepath.Clean(dir)\n\treturn p == dir || strings.HasPrefix(p, dir+string(filepath.Separator))\n}", "func pathPrefix(s, sub string) bool {\r\n\t// strings.HasPrefix is necessary but not sufficient.\r\n\tif !strings.HasPrefix(s, sub) {\r\n\t\treturn false\r\n\t}\r\n\t// The remainder after the prefix must either be empty or start with a slash.\r\n\trem := s[len(sub):]\r\n\treturn rem == \"\" || rem[0] == '/'\r\n}", "func (p path) ChildPath(name string, ptr BlockPointer) path {\n\tchild := path{\n\t\tFolderBranch: p.FolderBranch,\n\t\tpath: make([]pathNode, len(p.path), len(p.path)+1),\n\t}\n\tcopy(child.path, p.path)\n\tchild.path = append(child.path, pathNode{Name: name, BlockPointer: ptr})\n\treturn child\n}", "func containsPathPrefix(pats []string, s string) bool {\n\tfor _, pat := range pats {\n\t\tif pat == s || strings.HasPrefix(s, pat+\"/\") {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func pathLess(first, second string) bool {\n\t// Handle trivial cases first.\n\tif first == second {\n\t\treturn false\n\t} else if first == \"\" {\n\t\treturn true\n\t} else if second == \"\" {\n\t\treturn false\n\t}\n\n\t// Compare the path components. We work hard to avoid allocations here since\n\t// this is a comparison function for sorting algorithms.\n\tfor {\n\t\t// Extract the front path component from the first path.\n\t\tfirstFirstSlashIndex := strings.IndexByte(first, '/')\n\t\tvar firstFrontComponent string\n\t\tif firstFirstSlashIndex == -1 {\n\t\t\tfirstFrontComponent = first\n\t\t} else {\n\t\t\tfirstFrontComponent = first[:firstFirstSlashIndex]\n\t\t}\n\n\t\t// Extract the front path component from the second path.\n\t\tsecondFirstSlashIndex := strings.IndexByte(second, '/')\n\t\tvar secondFrontComponent string\n\t\tif secondFirstSlashIndex == -1 {\n\t\t\tsecondFrontComponent = second\n\t\t} else {\n\t\t\tsecondFrontComponent = second[:secondFirstSlashIndex]\n\t\t}\n\n\t\t// Compare the front path components.\n\t\tif firstFrontComponent < secondFrontComponent {\n\t\t\treturn true\n\t\t} else if secondFrontComponent < firstFrontComponent {\n\t\t\treturn false\n\t\t}\n\n\t\t// The front path components are equal. If either path has no remaining\n\t\t// components, then the comparison is complete, otherwise we move ahead\n\t\t// to the next path components. Note that we don't have to consider the\n\t\t// case where firstFirstSlashIndex and secondFirstSlashIndex are both -1\n\t\t// (with front components also equal) because that would mean the\n\t\t// strings were entirely equal, which we handle above.\n\t\tif firstFirstSlashIndex == -1 {\n\t\t\treturn true\n\t\t} else if secondFirstSlashIndex == -1 {\n\t\t\treturn false\n\t\t} else {\n\t\t\tfirst = first[firstFirstSlashIndex+1:]\n\t\t\tsecond = second[secondFirstSlashIndex+1:]\n\t\t}\n\t}\n}", "func (v *IADs) Parent() (path string, err error) {\n\tvar bstr *int16\n\thr, _, _ := syscall.Syscall(\n\t\tuintptr(v.VTable().Parent),\n\t\t2,\n\t\tuintptr(unsafe.Pointer(v)),\n\t\tuintptr(unsafe.Pointer(&bstr)),\n\t\t0)\n\tif bstr != nil {\n\t\tdefer ole.SysFreeString(bstr)\n\t}\n\tif hr == 0 {\n\t\tpath = ole.BstrToString((*uint16)(unsafe.Pointer(bstr)))\n\t} else {\n\t\treturn \"\", convertHresultToError(hr)\n\t}\n\treturn\n}", "func (l *fileLoader) IsAbsPath(root string, location string) bool {\n\tfullFilePath, err := l.fullLocation(root, location)\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn filepath.IsAbs(fullFilePath)\n}", "func IsGoListPath(path string) bool {\n\treturn strings.HasPrefix(path, \"./\") || strings.HasPrefix(path, \"../\") ||\n\t\tstrings.Contains(path, \"...\")\n}", "func HasParentWith(preds ...predicate.OutcomeMeasure) predicate.OutcomeOverview {\n\treturn predicate.OutcomeOverview(func(s *sql.Selector) {\n\t\tstep := sqlgraph.NewStep(\n\t\t\tsqlgraph.From(Table, FieldID),\n\t\t\tsqlgraph.To(ParentInverseTable, FieldID),\n\t\t\tsqlgraph.Edge(sqlgraph.M2O, true, ParentTable, ParentColumn),\n\t\t)\n\t\tsqlgraph.HasNeighborsWith(s, step, func(s *sql.Selector) {\n\t\t\tfor _, p := range preds {\n\t\t\t\tp(s)\n\t\t\t}\n\t\t})\n\t})\n}", "func ValidPath(path string) bool {\n\tfor _, validStart := range []string{\".\", \"..\", \"/\"} {\n\t\tif strings.HasPrefix(path, validStart) {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (d *gcpVolDriver) isPathExist(path string) (bool, error) {\n\t_, err := os.Stat(path)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn false, nil\n\t\t}\n\t\treturn false, err\n\t}\n\treturn true, nil\n}", "func dp_isSubPath(cur *ListNode,root *TreeNode)bool{\n\tif cur == nil{\n\t\treturn true\n\t}\n\tif root == nil{\n\t\treturn false\n\t}\n\treturn cur.Val == root.Val && (dp_isSubPath(cur.Next,root.Left) || dp_isSubPath(cur.Next,root.Right))\n}", "func (o *GstObj) SetParent(p *GstObj) bool {\n\treturn C.gst_object_set_parent(o.g(), p.g()) != 0\n}", "func (util copyHandlerUtil) isPathALocalDirectory(pathString string) bool {\n\t// check if path exists\n\tdestinationInfo, err := os.Stat(pathString)\n\n\tif err == nil && destinationInfo.IsDir() {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func setParentLink(parentCtx context.Context, childSpan *trace.Span) bool {\n\tparentSpanFromRPC := trace.FromContext(parentCtx)\n\tif parentSpanFromRPC == nil {\n\t\treturn false\n\t}\n\n\tpsc := parentSpanFromRPC.SpanContext()\n\tchildSpan.AddLink(trace.Link{\n\t\tSpanID: psc.SpanID,\n\t\tTraceID: psc.TraceID,\n\t\tType: trace.LinkTypeParent,\n\t})\n\treturn true\n}", "func (jdcb jobDirectoryContentsBatch) ParentNodePath() string {\n\treturn jdcb.parentPath\n}", "func isDirectoryPath(path string) bool {\n\tif !strings.Contains(path, \"*\") && strings.HasSuffix(path, \"/\") {\n\t\treturn true\n\t}\n\treturn false\n}" ]
[ "0.6913496", "0.6632338", "0.6614952", "0.6239509", "0.6065598", "0.6004529", "0.599899", "0.5987033", "0.5987033", "0.57555145", "0.5745408", "0.5715766", "0.56958073", "0.5602596", "0.5588836", "0.5535407", "0.5507737", "0.5400315", "0.5310334", "0.5297292", "0.52722096", "0.5269148", "0.526685", "0.52647334", "0.524701", "0.5224529", "0.5196813", "0.5185235", "0.51671845", "0.5162426", "0.51325166", "0.51165724", "0.50837624", "0.50837624", "0.50652903", "0.5054442", "0.50522584", "0.50353765", "0.503217", "0.49977908", "0.4982533", "0.49795997", "0.49721968", "0.4965631", "0.49580175", "0.4956799", "0.49560985", "0.49404234", "0.49336535", "0.49294898", "0.49219668", "0.4913758", "0.4912991", "0.4912731", "0.49040195", "0.48677194", "0.48585084", "0.48516062", "0.4844233", "0.48434138", "0.48138654", "0.48073086", "0.47996464", "0.4785012", "0.47795948", "0.4776879", "0.4774271", "0.47484297", "0.47433716", "0.47222254", "0.4718136", "0.47039247", "0.46479088", "0.46444938", "0.46440494", "0.46410838", "0.4631394", "0.4627184", "0.46200424", "0.461645", "0.46162665", "0.46139827", "0.46111706", "0.46096426", "0.45992467", "0.45931524", "0.45822924", "0.45800242", "0.4578777", "0.45776996", "0.4575818", "0.45734167", "0.45691913", "0.4567889", "0.45585373", "0.45533827", "0.4552875", "0.45498258", "0.45454574", "0.4526122" ]
0.78000367
0
GetOcJusticeTerminalCase invokes the dt_oc_info.GetOcJusticeTerminalCase API synchronously
func (client *Client) GetOcJusticeTerminalCase(request *GetOcJusticeTerminalCaseRequest) (response *GetOcJusticeTerminalCaseResponse, err error) { response = CreateGetOcJusticeTerminalCaseResponse() err = client.DoAction(request, response) return }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (client *Client) GetOcJusticeTerminalCaseWithCallback(request *GetOcJusticeTerminalCaseRequest, callback func(response *GetOcJusticeTerminalCaseResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetOcJusticeTerminalCaseResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetOcJusticeTerminalCase(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetOcJusticeTerminalCaseWithChan(request *GetOcJusticeTerminalCaseRequest) (<-chan *GetOcJusticeTerminalCaseResponse, <-chan error) {\n\tresponseChan := make(chan *GetOcJusticeTerminalCaseResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetOcJusticeTerminalCase(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func CreateGetOcJusticeTerminalCaseRequest() (request *GetOcJusticeTerminalCaseRequest) {\n\trequest = &GetOcJusticeTerminalCaseRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"dt-oc-info\", \"2022-08-29\", \"GetOcJusticeTerminalCase\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateGetOcJusticeTerminalCaseResponse() (response *GetOcJusticeTerminalCaseResponse) {\n\tresponse = &GetOcJusticeTerminalCaseResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func (b *OGame) GetCelestial(v any) (Celestial, error) {\n\treturn b.WithPriority(taskRunner.Normal).GetCelestial(v)\n}", "func (client IdentityClient) getCompartment(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodGet, \"/compartments/{compartmentId}\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response GetCompartmentResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (d *Dao) CaseObtainMID(c context.Context, mid int64, isToday bool) (cases map[int64]*model.SimCase, err error) {\n\tconn := d.redis.Get(c)\n\tdefer conn.Close()\n\tvar _setKey string\n\tif isToday {\n\t\t_setKey = caseVoteCIDMIDKey(mid)\n\t} else {\n\t\t_setKey = caseObtainMIDKey(mid)\n\t}\n\tvar ms []string\n\tif ms, err = redis.Strings(conn.Do(\"SMEMBERS\", _setKey)); err != nil {\n\t\tif err != redis.ErrNil {\n\t\t\treturn\n\t\t}\n\t\terr = nil\n\t}\n\tcases = make(map[int64]*model.SimCase)\n\tfor _, s := range ms {\n\t\tif s == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tsc := &model.SimCase{}\n\t\tif err = json.Unmarshal([]byte(s), sc); err != nil {\n\t\t\terr = errors.WithStack(err)\n\t\t\treturn\n\t\t}\n\t\tcases[sc.ID] = sc\n\t}\n\treturn\n}", "func (client IdentityClient) GetCompartment(ctx context.Context, request GetCompartmentRequest) (response GetCompartmentResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.NoRetryPolicy()\n\tif client.RetryPolicy() != nil {\n\t\tpolicy = *client.RetryPolicy()\n\t}\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\tociResponse, err = common.Retry(ctx, request, client.getCompartment, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = GetCompartmentResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = GetCompartmentResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(GetCompartmentResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into GetCompartmentResponse\")\n\t}\n\treturn\n}", "func CustomerGetoneVehicleforview(w http.ResponseWriter, r *http.Request) {\n\tvehicle := services.GetOneVehicle(r)\n\tcusttpl.ExecuteTemplate(w, \"viewvehicle.html\", vehicle)\n}", "func GetTrafficOpsCookie(cdnUri, user, pass string) (string, error) {\n\turi := cdnUri + `/api/1.2/user/login`\n\tpostdata := `{\"u\":\"` + user + `\", \"p\":\"` + pass + `\"}`\n\treq, err := http.NewRequest(\"POST\", uri, strings.NewReader(postdata))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treq.Header.Add(\"Accept\", \"application/json\")\n\n\tclient := getClient()\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer resp.Body.Close()\n\n\tfor _, cookie := range resp.Cookies() {\n\t\tif cookie.Name == `mojolicious` {\n\t\t\treturn cookie.Value, nil\n\t\t}\n\t}\n\n\tdata, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn \"\", errors.New(\"No login cookie received: \" + string(data))\n}", "func (h *Handler) RetrieveCertificate(companyChainID string, uuid string) (*entityApi.TransactionWrapper, error) {\n apiResponse, err := h.apiClient.Get(fmt.Sprintf(certificateRoute, companyChainID, uuid), nil)\n if err != nil {\n return nil, err\n }\n var transactionWrapper entityApi.TransactionWrapper\n if err := unmarshalApiResponse(apiResponse, &transactionWrapper); err != nil {\n return nil, err\n }\n return &transactionWrapper, nil\n}", "func (term *Terminology) ReadV2toSNOMEDCT(ctx context.Context, id *apiv1.Identifier, f func(*apiv1.Identifier) error) error {\n\tctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)\n\tdefer cancel()\n\tresponse, err := term.client.FromCrossMap(ctx, &snomed.TranslateFromRequest{S: id.GetValue(), RefsetId: 900000000000497000})\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(response.GetTranslations()) == 0 {\n\t\tlog.Printf(\"no translations found for map from '%s:%s' to '%s'\", id.GetSystem(), id.GetValue(), identifiers.SNOMEDCT)\n\t}\n\tfor _, t := range response.GetTranslations() {\n\t\tref := t.GetReferenceSetItem().GetReferencedComponentId()\n\t\tif err := f(&apiv1.Identifier{System: identifiers.SNOMEDCT, Value: strconv.FormatInt(ref, 10)}); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func TestGetCabTripCtrl(t *testing.T) {\n\tmodel.MongoDBConnect()\n\tdefer model.MongoConnectionClose()\n\tConvey(\"GetCabTripCtrl\", func() {\n\t\treq, err := http.NewRequest(\"GET\", \"/api/cab/id3004672/date/2016-06-30?fresh=1\", nil)\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\n\t\tw := httptest.NewRecorder()\n\n\t\tConvey(\"Normal: 1 cab\", func() {\n\n\t\t\tcabapi.GetCabTripCtrl(w, req)\n\n\t\t\tSo(w.Code, ShouldEqual, http.StatusOK)\n\t\t\tresult, _ := ioutil.ReadAll(w.Result().Body)\n\t\t\texpct := \"id3004672 in 2016-06-30:3\"\n\t\t\tSo(result, ShouldEqual, expct)\n\t\t})\n\n\n\t})\n}", "func GetClinicDoctors(c *gin.Context) {\n\tlog.Infof(\"Get all doctors registered with specific physical clinic\")\n\taddressID := c.Param(\"addressId\")\n\tif addressID == \"\" {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusBadRequest,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: \"clinic address id not provided\",\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tctx := c.Request.Context()\n\tuserEmail, userID, gproject, err := getUserDetails(ctx, c.Request)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tctx, span := trace.StartSpan(ctx, \"Get all doctors registered for a clinic\")\n\tdefer span.End()\n\tclinicMetaDB := datastoredb.NewClinicMetaHandler()\n\terr = clinicMetaDB.InitializeDataBase(ctx, gproject)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tregisteredDoctors, err := clinicMetaDB.GetClinicDoctors(ctx, userEmail, userID, addressID)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, gin.H{\n\t\tconstants.RESPONSE_JSON_DATA: registeredDoctors,\n\t\tconstants.RESPONSDE_JSON_ERROR: nil,\n\t})\n\tclinicMetaDB.Close()\n}", "func GetIncident(c *gin.Context) {\n\tvar err error\n\tvar output *incident.Incident\n\tvar incidentID int64\n\tvar taskID int64\n\tincidentID, err = strconv.ParseInt(c.Param(\"incidentId\"), 10, 64)\n\tctx, _ := authcontext.NewAuthContext(c)\n\n\tif taskID, err = strconv.ParseInt(c.Param(\"taskId\"), 10, 64); err == nil {\n\t\tif output, err = incident.GetByID(ctx, taskID, incidentID); err == nil {\n\t\t\tc.JSON(http.StatusOK, output)\n\t\t}\n\t}\n\n\tif err != nil {\n\t\tc.JSON(http.StatusPreconditionFailed, ResponseObject{\"error\": err.Error()})\n\t}\n}", "func GetCharacterModel(w http.ResponseWriter, req *http.Request) {\n\n\t// Get session values or redirect to Login\n\tsession, err := sessions.Store.Get(req, \"session\")\n\n\tif err != nil {\n\t\tlog.Println(\"error identifying session\")\n\t\thttp.Redirect(w, req, \"/login/\", http.StatusFound)\n\t\treturn\n\t\t// in case of error\n\t}\n\n\t// Prep for user authentication\n\tsessionMap := getUserSessionValues(session)\n\n\tusername := sessionMap[\"username\"]\n\tloggedIn := sessionMap[\"loggedin\"]\n\tisAdmin := sessionMap[\"isAdmin\"]\n\n\tfmt.Println(loggedIn, isAdmin, username)\n\n\tfmt.Println(session)\n\n\t/*\n\t\tif username == \"\" {\n\t\t\thttp.Redirect(w, req, \"/\", http.StatusFound)\n\t\t\treturn\n\t\t}\n\t*/\n\n\tvars := mux.Vars(req)\n\tidString := vars[\"id\"]\n\n\tpk, err := strconv.Atoi(idString)\n\tif err != nil {\n\t\tpk = 0\n\t\tlog.Println(err)\n\t}\n\n\tcm, err := database.PKLoadCharacterModel(db, int64(pk))\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\n\tjson.NewEncoder(w).Encode(cm)\n}", "func (client *Client) GetOpenNLU(request *GetOpenNLURequest) (response *GetOpenNLUResponse, err error) {\n\tresponse = CreateGetOpenNLUResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func (r Virtual_Guest) GetOpenCancellationTicket() (resp datatypes.Ticket, err error) {\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest\", \"getOpenCancellationTicket\", nil, &r.Options, &resp)\n\treturn\n}", "func (ovscni *OvsCni) HandleCni(d *render.RenderData) error {\n\n\t//For VlanType=trunk we do not need to do anything\n\tswitch ovscni.VlanType {\n\tcase \"access\":\n\t\tif len(ovscni.L2srvResources) != 1 {\n\t\t\terr := errors.New(\"Cannot use more than one L2Services for VlanType=access case\")\n\t\t\tovscni.Log.Error(err, \"L2Services cannot contain more than one L2Services in VlanType=access case\")\n\t\t\treturn err\n\t\t}\n\t\td.Data[\"AccessVlan\"] = ovscni.L2srvResources[0].Spec.SegmentationID\n\tcase \"selectivetrunk\":\n\t\ttmpList := []string{}\n\t\tfor _, l2srvObj := range ovscni.L2srvResources {\n\t\t\ttmpStr := \"{\\\"id\\\": \" + strconv.Itoa(int(l2srvObj.Spec.SegmentationID)) + \"}\"\n\t\t\ttmpList = append(tmpList, tmpStr)\n\t\t}\n\t\td.Data[\"SelectiveVlan\"] = \"[\" + strings.Join(tmpList, \",\") + \"]\"\n\tcase \"trunk\":\n\t\tovscni.Log.Info(\"Transparent Trunk case in cluster level\")\n\t}\n\treturn nil\n}", "func openDUTControlConsole(stream dutcontrol.DutControl_ConsoleClient, req *dutcontrol.ConsoleRequest) (<-chan *dutcontrol.ConsoleSerialData, <-chan *dutcontrol.ConsoleSerialWriteResult, error) {\n\tif err := stream.Send(req); err != nil {\n\t\treturn nil, nil, errors.Wrap(err, \"send request\")\n\t}\n\tresp, err := stream.Recv()\n\tif err != nil {\n\t\treturn nil, nil, errors.Wrap(err, \"recv open\")\n\t}\n\topen := resp.GetOpen()\n\tif open == nil {\n\t\treturn nil, nil, errors.New(\"open response is nil\")\n\t}\n\tif open.Err != \"\" {\n\t\treturn nil, nil, errors.New(string(open.Err))\n\t}\n\tdata := make(chan *dutcontrol.ConsoleSerialData, qSize)\n\twrite := make(chan *dutcontrol.ConsoleSerialWriteResult, qSize)\n\tgo func() {\n\tLoop:\n\t\tfor {\n\t\t\tresp, err := stream.Recv()\n\t\t\tif err == io.EOF {\n\t\t\t\ttesting.ContextLog(stream.Context(), \"Dutcontrol recv EOF\")\n\t\t\t\tbreak\n\t\t\t} else if err != nil {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tswitch op := resp.Type.(type) {\n\t\t\tcase *dutcontrol.ConsoleResponse_SerialData:\n\t\t\t\tdata <- op.SerialData\n\t\t\tcase *dutcontrol.ConsoleResponse_SerialWrite:\n\t\t\t\twrite <- op.SerialWrite\n\t\t\tdefault:\n\t\t\t\ttesting.ContextLog(stream.Context(), \"Dutcontrol recv error, unknown message type: \", op)\n\t\t\t\tbreak Loop\n\t\t\t}\n\t\t}\n\t\tclose(data)\n\t\tclose(write)\n\t}()\n\treturn data, write, nil\n}", "func (router *Router) getTerminal(w http.ResponseWriter, r *http.Request) {\n\tclusterName := r.URL.Query().Get(\"cluster\")\n\tnamespace := r.URL.Query().Get(\"namespace\")\n\tname := r.URL.Query().Get(\"name\")\n\tcontainer := r.URL.Query().Get(\"container\")\n\tshell := r.URL.Query().Get(\"shell\")\n\n\tlog.WithFields(logrus.Fields{\"cluster\": clusterName, \"namespace\": namespace, \"name\": name, \"container\": container, \"shell\": shell}).Tracef(\"getTerminal\")\n\n\tvar upgrader = websocket.Upgrader{}\n\n\tif router.config.WebSocket.AllowAllOrigins {\n\t\tupgrader.CheckOrigin = func(r *http.Request) bool { return true }\n\t}\n\n\tc, err := upgrader.Upgrade(w, r, nil)\n\tif err != nil {\n\t\tlog.WithError(err).Errorf(\"Could not upgrade connection\")\n\t\treturn\n\t}\n\tdefer c.Close()\n\n\tc.SetPongHandler(func(string) error { return nil })\n\n\tgo func() {\n\t\tticker := time.NewTicker(pingPeriod)\n\t\tdefer ticker.Stop()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ticker.C:\n\t\t\t\tif err := c.WriteMessage(websocket.PingMessage, nil); err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\tcluster := router.clusters.GetCluster(clusterName)\n\tif cluster == nil {\n\t\tlog.WithError(err).Errorf(\"Invalid cluster name\")\n\t\tmsg, _ := json.Marshal(terminal.Message{\n\t\t\tOp: \"stdout\",\n\t\t\tData: fmt.Sprintf(\"Invalid cluster name: %s\", err.Error()),\n\t\t})\n\t\tc.WriteMessage(websocket.TextMessage, msg)\n\t\treturn\n\t}\n\n\terr = cluster.GetTerminal(c, namespace, name, container, shell)\n\tif err != nil {\n\t\tlog.WithError(err).Errorf(\"Could not create terminal\")\n\t\tmsg, _ := json.Marshal(terminal.Message{\n\t\t\tOp: \"stdout\",\n\t\t\tData: fmt.Sprintf(\"Could not create terminal: %s\", err.Error()),\n\t\t})\n\t\tc.WriteMessage(websocket.TextMessage, msg)\n\t\treturn\n\t}\n\n\tlog.Tracef(\"Terminal connection was closed\")\n}", "func GetAttestor(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *AttestorState, opts ...pulumi.ResourceOption) (*Attestor, error) {\n\tvar resource Attestor\n\terr := ctx.ReadResource(\"google-native:binaryauthorization/v1beta1:Attestor\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func CORSVC(ac *atmi.ATMICtx, svc *atmi.TPSVCINFO) {\n\tret := SUCCEED\n\n\t//Return to the caller\n\tdefer func() {\n\n\t\tac.TpLogCloseReqFile()\n\t\tif SUCCEED == ret {\n\t\t\tac.TpReturn(atmi.TPSUCCESS, 0, &svc.Data, 0)\n\t\t} else {\n\t\t\tac.TpReturn(atmi.TPFAIL, 0, &svc.Data, 0)\n\t\t}\n\t}()\n\n\t//Get UBF Handler\n\tub, _ := ac.CastToUBF(&svc.Data)\n\n\t//Print the buffer to stdout\n\t//fmt.Println(\"Incoming request:\")\n\tub.TpLogPrintUBF(atmi.LOG_DEBUG, \"CORSVC: Incoming request:\")\n\n\tarr, err := ub.BGetByteArr(u.EX_NETDATA, 0)\n\n\tif err != nil {\n\t\tac.TpLogError(\"Failed to get EX_NETDATA: %s\", err.Message())\n\t\tret = FAIL\n\t\treturn\n\t}\n\tif arr[0] == 1 && arr[1] == 1 {\n\t\tac.TpLogInfo(\"Test case 11 - no need for correlation\")\n\t} else if len(arr) > 4 {\n\n\t\tcorr := string(arr[:4])\n\n\t\tac.TpLogInfo(\"Extracted correlator: [%s]\", corr)\n\n\t\tif err := ub.BChg(u.EX_NETCORR, 0, corr); nil != err {\n\t\t\tac.TpLogError(\"Failed to set EX_NETCORR: %s\", err.Message())\n\t\t\tret = FAIL\n\t\t\treturn\n\t\t}\n\n\t}\n\n\tub.TpLogPrintUBF(atmi.LOG_DEBUG, \"Reply buffer afrer correl\")\n\n}", "func (_Cakevault *CakevaultTransactor) InCaseTokensGetStuck(opts *bind.TransactOpts, _token common.Address) (*types.Transaction, error) {\n\treturn _Cakevault.contract.Transact(opts, \"inCaseTokensGetStuck\", _token)\n}", "func GetKubernetesMode() string {\n\treturn strings.TrimSpace(os.Getenv(\"CYPRESS_PARALLEL_API_K8S_CLIENT_OUTSIDE\"))\n}", "func CustomerGetVehicle(w http.ResponseWriter, r *http.Request) {\n\tvehicle := services.GetOneVehicle(r)\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tjson.NewEncoder(w).Encode(vehicle)\n}", "func (c *ClinicClient) Get(ctx context.Context, id uuid.UUID) (*Clinic, error) {\n\treturn c.Query().Where(clinic.ID(id)).Only(ctx)\n}", "func (c *client) getCharacter(thisCharacter *Character, uID string) error {\n\tauthorized, err := c.isOwner(thisCharacter.ID, uID)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !authorized {\n\t\treturn errorutil.New(http.StatusForbidden, \"can't delete someone else's character...\")\n\t}\n\n\t// get the character by ID\n\tvar clID, scID int\n\t// TODO if the default for Concentrated is 0, the api call to spells/0 will return Acid Splash (key 1) - why?\n\tif err := c.db.QueryRow(`SELECT ch.Name, ch.ClassLevel, ch.SCAbilityScore, ch.PortraitPath, ch.ClassKey, IFNULL(ch.SubclassKey, 0), IFNULL(ch.ConcentratedSpell, 0),\n\t\tch.Level1SlotsRemaining, ch.Level2SlotsRemaining, ch.Level3SlotsRemaining, ch.Level4SlotsRemaining, ch.Level5SlotsRemaining, ch.Level6SlotsRemaining,\n\t\tch.Level7SlotsRemaining, ch.Level8SlotsRemaining, ch.Level9SlotsRemaining, ch.CastSuccess\n\t\tFROM Characters ch\n\t\tWHERE ch.CharacterKey = ?`, thisCharacter.ID).Scan(&thisCharacter.Name, &thisCharacter.Level,\n\t\t&thisCharacter.AbilityScore, &thisCharacter.PortraitPath,\n\t\t&clID, &scID, &thisCharacter.Concentrating, &thisCharacter.Level1SlotsRemaining, &thisCharacter.Level2SlotsRemaining, &thisCharacter.Level3SlotsRemaining,\n\t\t&thisCharacter.Level4SlotsRemaining, &thisCharacter.Level5SlotsRemaining, &thisCharacter.Level6SlotsRemaining, &thisCharacter.Level7SlotsRemaining,\n\t\t&thisCharacter.Level8SlotsRemaining, &thisCharacter.Level9SlotsRemaining, &thisCharacter.CastSuccess); err != nil {\n\t\tlog.Printf(\"mysql: could not access Character: %v\", err)\n\t\treturn errorutil.New(500, \"internal error\")\n\t}\n\tcl, err := c.clc.ClassByID(clID, false)\n\tif err != nil {\n\t\tlog.Printf(\"couldn't get class for character (ID: %d): %v\", thisCharacter.ID, err)\n\t\treturn err\n\t}\n\tthisCharacter.Class = cl\n\tif scID != 0 {\n\t\tsc, err := c.clc.SubclassByID(scID)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"couldn't get subclass for character (ID: %d): %v\", thisCharacter.ID, err)\n\t\t\treturn err\n\t\t}\n\t\tthisCharacter.Subclass = sc\n\t}\n\tlog.Print(\"\\n\\tfilled struct: \", thisCharacter)\n\n\treturn nil\n}", "func GetCSDCOProj(identity string) *sparql.Results {\n\t// repo, err := sparql.NewRepo(\"http://data.oceandrilling.org/sparql\",\n\trepo, err := getJena()\n\tif err != nil {\n\t\tlog.Printf(\"%s\\n\", err)\n\t}\n\n\tf := bytes.NewBufferString(projdetails)\n\tbank := sparql.LoadBank(f)\n\n\t// q, err := bank.Prepare(\"my-query\", struct{ Limit, Offset int }{10, 100})\n\tq, err := bank.Prepare(\"csdcoproj\", struct{ ID string }{identity})\n\tif err != nil {\n\t\tlog.Print(err)\n\t}\n\n\tlog.Println(q)\n\n\tres, err := repo.Query(q)\n\tif err != nil {\n\t\tlog.Print(err)\n\t}\n\n\treturn res\n}", "func (client *LicenseStatusClient) Get(uuid string, options ...session.ApiOptionsParams) (*models.LicenseStatus, error) {\n\tvar obj *models.LicenseStatus\n\terr := client.aviSession.Get(client.getAPIPath(uuid), &obj, options...)\n\treturn obj, err\n}", "func (_Cakevault *CakevaultTransactorSession) InCaseTokensGetStuck(_token common.Address) (*types.Transaction, error) {\n\treturn _Cakevault.Contract.InCaseTokensGetStuck(&_Cakevault.TransactOpts, _token)\n}", "func (ec *executionContext) _Case(ctx context.Context, sel ast.SelectionSet, obj *models.Case) graphql.Marshaler {\n\tfields := graphql.CollectFields(ctx, sel, caseImplementors)\n\n\tvar wg sync.WaitGroup\n\tout := graphql.NewOrderedMap(len(fields))\n\tinvalid := false\n\tfor i, field := range fields {\n\t\tout.Keys[i] = field.Alias\n\n\t\tswitch field.Name {\n\t\tcase \"__typename\":\n\t\t\tout.Values[i] = graphql.MarshalString(\"Case\")\n\t\tcase \"Id\":\n\t\t\tout.Values[i] = ec._Case_Id(ctx, field, obj)\n\t\tcase \"Asset\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_Asset(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"CaseNumber\":\n\t\t\tout.Values[i] = ec._Case_CaseNumber(ctx, field, obj)\n\t\tcase \"Origin\":\n\t\t\tout.Values[i] = ec._Case_Origin(ctx, field, obj)\n\t\tcase \"Owner\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_Owner(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"Reason\":\n\t\t\tout.Values[i] = ec._Case_Reason(ctx, field, obj)\n\t\tcase \"IsClosed\":\n\t\t\tout.Values[i] = ec._Case_IsClosed(ctx, field, obj)\n\t\tcase \"Contact\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_Contact(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"CreatedBy\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_CreatedBy(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"ClosedDate\":\n\t\t\tout.Values[i] = ec._Case_ClosedDate(ctx, field, obj)\n\t\tcase \"CreatedDate\":\n\t\t\tout.Values[i] = ec._Case_CreatedDate(ctx, field, obj)\n\t\tcase \"IsDeleted\":\n\t\t\tout.Values[i] = ec._Case_IsDeleted(ctx, field, obj)\n\t\tcase \"Description\":\n\t\t\tout.Values[i] = ec._Case_Description(ctx, field, obj)\n\t\tcase \"IsEscalated\":\n\t\t\tout.Values[i] = ec._Case_IsEscalated(ctx, field, obj)\n\t\tcase \"LastModifiedBy\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_LastModifiedBy(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"LastModifiedDate\":\n\t\t\tout.Values[i] = ec._Case_LastModifiedDate(ctx, field, obj)\n\t\tcase \"LastReferencedDate\":\n\t\t\tout.Values[i] = ec._Case_LastReferencedDate(ctx, field, obj)\n\t\tcase \"LastViewedDate\":\n\t\t\tout.Values[i] = ec._Case_LastViewedDate(ctx, field, obj)\n\t\tdefault:\n\t\t\tpanic(\"unknown field \" + strconv.Quote(field.Name))\n\t\t}\n\t}\n\twg.Wait()\n\tif invalid {\n\t\treturn graphql.Null\n\t}\n\treturn out\n}", "func (d *Dao) ReadOxygenConc() (resp interface{}, err error) {\n\treq := SensorOxygenConcUnit.Request()\n\toutput, err := send(d, req.Bytes())\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn resp, err\n\t}\n\tconc := binary.BigEndian.Uint16(output[3:5])\n\tresp = dividedByTen(conc)\n\treturn resp, nil\n}", "func GetByProvince(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tcovidData, err := scrapper.GetAllDataByProvince()\n\tif err != nil {\n\t\tjson.NewEncoder(w).Encode(Error{\n\t\t\tCode: 500,\n\t\t\tMessage: err.Error(),\n\t\t})\n\n\t\treturn\n\t}\n\n\tprov := r.URL.Query().Get(\"prov\")\n\tif prov != \"\" {\n\t\tfor _, item := range covidData {\n\t\t\tif strings.ToLower(item.Province) == strings.ToLower(prov) {\n\t\t\t\tjson.NewEncoder(w).Encode(item)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\tjson.NewEncoder(w).Encode(covidData)\n\t\treturn\n\t}\n\n\tjson.NewEncoder(w).Encode(covidData)\n}", "func (client IdentityClient) getTenancy(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodGet, \"/tenancies/{tenancyId}\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response GetTenancyResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func GetSingleClinicID(c *gin.Context) {\n\tlog.Infof(\"Get all clinics associated with admin\")\n\tctx := c.Request.Context()\n\taddressID := c.Param(\"addressId\")\n\tif addressID == \"\" {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusBadRequest,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: \"clinic address id not provided\",\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\t_, _, gproject, err := getUserDetails(ctx, c.Request)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tctx, span := trace.StartSpan(ctx, \"Get all clinics associated with admin\")\n\tdefer span.End()\n\tclinicMetaDB := datastoredb.NewClinicMetaHandler()\n\terr = clinicMetaDB.InitializeDataBase(ctx, gproject)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\n\tregisteredClinics, err := clinicMetaDB.GetSingleClinic(ctx, addressID)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, gin.H{\n\t\tconstants.RESPONSE_JSON_DATA: registeredClinics,\n\t\tconstants.RESPONSDE_JSON_ERROR: nil,\n\t})\n\n\tclinicMetaDB.Close()\n}", "func SimpleGet(w http.ResponseWriter, req *http.Request) {\n\tlog.Println(\"Get Succeeded\")\n\tw.WriteHeader(http.StatusOK)\n\tw.Write([]byte(os.Getenv(\"COPILOT_APPLICATION_NAME\") + \"-\" + os.Getenv(\"COPILOT_ENVIRONMENT_NAME\") + \"-\" + os.Getenv(\"COPILOT_SERVICE_NAME\")))\n}", "func (t *InsuranceChaincode) getCustomerByID(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n var customerId, jsonResp string\n\n if len(args) != 1 {\n return shim.Error(\"Incorrect number of arguments. Expecting customerId of the Insurance to query\")\n }\n\n customerId = args[0]\n valAsbytes, err := stub.GetState(customerId) \n if err != nil {\n jsonResp = \"{\\\"Error\\\":\\\"Failed to get state for \" + customerId + \"\\\"}\"\n return shim.Error(jsonResp)\n } else if valAsbytes == nil {\n jsonResp = \"{\\\"CustomerId\\\": \\\"\"+ customerId + \"\\\", \\\"Error\\\":\\\"Customer does not exist.\\\"}\"\n return shim.Error(jsonResp)\n }\n\n return shim.Success(valAsbytes)\n}", "func GetCategory(response http.ResponseWriter, request *http.Request) {\n\t//var results TCategory\n\tvar errorResponse = ErrorResponse{\n\t\tCode: http.StatusInternalServerError, Message: \"Internal Server Error.\",\n\t}\n\n\tcollection := Client.Database(\"msdb\").Collection(\"t_cat_mg\")\n\tctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)\n\tcursor, err := collection.Find(ctx, bson.M{})\n\tvar results []bson.M\n\terr = cursor.All(ctx, &results)\n\n\tdefer cancel()\n\n\tif err != nil {\n\t\terrorResponse.Message = \"Document not found\"\n\t\treturnErrorResponse(response, request, errorResponse)\n\t} else {\n\t\tvar successResponse = SuccessResponse{\n\t\t\tCode: http.StatusOK,\n\t\t\tMessage: \"Success\",\n\t\t\tResponse: results,\n\t\t}\n\n\t\tsuccessJSONResponse, jsonError := json.Marshal(successResponse)\n\n\t\tif jsonError != nil {\n\t\t\treturnErrorResponse(response, request, errorResponse)\n\t\t}\n\t\tresponse.Header().Set(\"Content-Type\", \"application/json\")\n\t\tresponse.Write(successJSONResponse)\n\t}\n\n}", "func (_Cakevault *CakevaultSession) InCaseTokensGetStuck(_token common.Address) (*types.Transaction, error) {\n\treturn _Cakevault.Contract.InCaseTokensGetStuck(&_Cakevault.TransactOpts, _token)\n}", "func (c *Client) CreateCase(ctx context.Context, params *CreateCaseInput, optFns ...func(*Options)) (*CreateCaseOutput, error) {\n\tif params == nil {\n\t\tparams = &CreateCaseInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"CreateCase\", params, optFns, addOperationCreateCaseMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*CreateCaseOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (Lawrencium) GetCategory() string {\n\tvar c categoryType = actinoid\n\treturn c.get()\n}", "func (d *Dao) Case(c context.Context, arg *blocked.ArgCaseSearch) (ids []int64, pager *blocked.Pager, err error) {\n\treq := d.elastic.NewRequest(blocked.BusinessBlockedCase).Index(blocked.TableBlockedCase).Fields(\"id\")\n\tif arg.Keyword != blocked.SearchDefaultString {\n\t\treq.WhereLike([]string{\"origin_content\"}, []string{arg.Keyword}, true, elastic.LikeLevelHigh)\n\t}\n\tif arg.OriginType != blocked.SearchDefaultNum {\n\t\treq.WhereEq(\"origin_type\", arg.OriginType)\n\t}\n\tif arg.Status != blocked.SearchDefaultNum {\n\t\treq.WhereEq(\"status\", arg.Status)\n\t}\n\tif arg.CaseType != blocked.SearchDefaultNum {\n\t\treq.WhereEq(\"case_type\", arg.CaseType)\n\t}\n\tif arg.UID != blocked.SearchDefaultNum {\n\t\treq.WhereEq(\"mid\", arg.UID)\n\t}\n\tif arg.OPID != blocked.SearchDefaultNum {\n\t\treq.WhereEq(\"oper_id\", arg.OPID)\n\t}\n\treq.WhereRange(\"start_time\", arg.TimeFrom, arg.TimeTo, elastic.RangeScopeLcRc)\n\treq.Pn(arg.PN).Ps(arg.PS).Order(arg.Order, arg.Sort)\n\tvar res *search.ReSearchData\n\tif err = req.Scan(c, &res); err != nil {\n\t\terr = errors.Errorf(\"elastic search(%s) error(%v)\", req.Params(), err)\n\t\treturn\n\t}\n\tids, pager = pagerExtra(res)\n\treturn\n}", "func (client *Client) DescribeCustinsKernelReleaseNotesWithCallback(request *DescribeCustinsKernelReleaseNotesRequest, callback func(response *DescribeCustinsKernelReleaseNotesResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeCustinsKernelReleaseNotesResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeCustinsKernelReleaseNotes(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (p *MockProvisionerClient) TenancyOCID() string {\n\treturn \"ocid1.tenancy.oc1..aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa\"\n}", "func (_SingleAuto *SingleAutoTransactor) InCaseTokensGetStuck(opts *bind.TransactOpts, _token common.Address, _amount *big.Int) (*types.Transaction, error) {\n\treturn _SingleAuto.contract.Transact(opts, \"inCaseTokensGetStuck\", _token, _amount)\n}", "func (env *Env) GetContestInfo(c *gin.Context) {\n\tkv, err := env.db.GetKV(\"ContestInfo\")\n\tif err != nil {\n\t\tvar errMsg = fmt.Sprint(\"Get ContestInfo failed with\", err)\n\t\tc.JSON(http.StatusInternalServerError, gin.H{\"message\": errMsg})\n\t\treturn\n\t}\n\tvar ci model.ContestInfo\n\terr = json.Unmarshal(kv.Value, &ci)\n\tif err != nil {\n\t\tvar errMsg = fmt.Sprint(\"Get ContestInfo failed with\", err)\n\t\tc.JSON(http.StatusInternalServerError, gin.H{\"message\": errMsg})\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, ci)\n}", "func (m *PatientMutation) Congenital() (r string, exists bool) {\n\tv := m._Congenital\n\tif v == nil {\n\t\treturn\n\t}\n\treturn *v, true\n}", "func (cr CURetriever) GetCU(addr sdk.CUAddress) (exported.CustodianUnit, error) {\n\tCU, _, err := cr.GetCUWithHeight(addr)\n\treturn CU, err\n}", "func ToCaseClause(x ast.Node) *ast.CaseClause {\n\tif x, ok := x.(*ast.CaseClause); ok {\n\t\treturn x\n\t}\n\treturn NilCaseClause\n}", "func (t Type) GoCase() string {\n\treturn gocase.To(strcase.ToCamel(string(t)))\n}", "func getSimpleTOC(res http.ResponseWriter, req *http.Request, params httprouter.Params) {\n\tpu, _ := GetUserFromSession(res, req)\n\n\tscreenOutput := struct {\n\t\tName string\n\t\tEmail string\n\t\tPermission int\n\t\tID string\n\t}{\n\t\tpu.Name,\n\t\tpu.Email,\n\t\tpu.Permission,\n\t\tparams.ByName(\"ID\"),\n\t}\n\n\tServeTemplateWithParams(res, \"toc.html\", screenOutput)\n}", "func (m *SecureScoreControlProfile) GetControlCategory()(*string) {\n val, err := m.GetBackingStore().Get(\"controlCategory\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func GetNuxeoContainer(dep *v1.Deployment) (*v12.Container, error) {\n\tfor i := 0; i < len(dep.Spec.Template.Spec.Containers); i++ {\n\t\tif dep.Spec.Template.Spec.Containers[i].Name == \"nuxeo\" {\n\t\t\treturn &dep.Spec.Template.Spec.Containers[i], nil\n\t\t}\n\t}\n\treturn nil, fmt.Errorf(\"could not find a container named 'nuxeo' in the deployment\")\n}", "func GetTC(postId string, session *r.Session) []string {\n\tvar tcs []string\n\tvar tc ct.TravelCapsule\n\tdb := os.Getenv(\"DB\")\n\ttcTable := os.Getenv(\"TCTABLE\")\n\tcur, _ := r.DB(db).Table(tcTable).GetAllByIndex(\"posts\", postId).Run(session)\n\n\tfor cur.Next(&tc) {\n\t\ttcs = append(tcs, tc.Id)\n\t}\n\treturn tcs\n}", "func GetWareByCategory(c *server.Context) error {\n\tvar (\n\t\tres []ware.BriefInfo\n\t\tcidReq struct {\n\t\t\tParentCID uint32 `json:\"parent_cid\" validate:\"required\"`\n\t\t\tCID uint32 `json:\"cid\"`\n\t\t}\n\t)\n\n\terr := c.JSONBody(&cidReq)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInvalidParam, nil)\n\t}\n\n\terr = c.Validate(cidReq)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInvalidParam, nil)\n\t}\n\n\tconn, err := mysql.Pool.Get()\n\tdefer mysql.Pool.Release(conn)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrMysql, nil)\n\t}\n\n\tif cidReq.CID == 0 {\n\t\tres, err = ware.Service.GetByParentCID(conn, cidReq.ParentCID)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrMysql, nil)\n\t\t}\n\t} else {\n\t\tres, err = ware.Service.GetByCID(conn, cidReq.CID)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrMysql, nil)\n\t\t}\n\t}\n\n\treturn core.WriteStatusAndDataJSON(c, constants.ErrSucceed, res)\n}", "func (a *HyperflexApiService) GetHyperflexLicenseByMoid(ctx context.Context, moid string) ApiGetHyperflexLicenseByMoidRequest {\n\treturn ApiGetHyperflexLicenseByMoidRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tmoid: moid,\n\t}\n}", "func (m *DeviceEnrollmentWindowsHelloForBusinessConfiguration) GetPinLowercaseCharactersUsage()(*WindowsHelloForBusinessPinUsage) {\n val, err := m.GetBackingStore().Get(\"pinLowercaseCharactersUsage\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*WindowsHelloForBusinessPinUsage)\n }\n return nil\n}", "func (d *Deals) Retrieve(ctx context.Context, waddr string, cid cid.Cid) (io.Reader, error) {\n\ttime.Sleep(time.Second * 3)\n\treturn strings.NewReader(\"hello there\"), nil\n}", "func GetCustomerbyPhone(details string) (customer Customers, err error) {\r\n\tvar rows *sql.Rows\r\n\tif rows, err = Get(fmt.Sprintf(`select * from customers where %s and deleted_at is null;`, details)); err != nil {\r\n\t\tCheckError(\"Error getting Customer details.\", err, false)\r\n\t\treturn Customers{}, err\r\n\t}\r\n\r\n\tdefer rows.Close()\r\n\tfor rows.Next() {\r\n\t\tif err = rows.Scan(&customer.ID, &customer.CardCode, &customer.CardName, &customer.Address, &customer.Phone, &customer.Phone1, &customer.City, &customer.Email, &customer.Synced, &customer.CreatedBy, &customer.CreatedAt, &customer.UpdatedAt, &customer.DeletedAt); err != nil {\r\n\t\t\tCheckError(\"Error Scanning Customers.\", err, false)\r\n\t\t}\r\n\t}\r\n\r\n\treturn\r\n}", "func (c *DoctorClient) Get(ctx context.Context, id int) (*Doctor, error) {\n\treturn c.Query().Where(doctor.ID(id)).Only(ctx)\n}", "func (c *OperationroomClient) Get(ctx context.Context, id int) (*Operationroom, error) {\n\treturn c.Query().Where(operationroom.ID(id)).Only(ctx)\n}", "func (ts *TechStoryService) getText (w http.ResponseWriter, r *http.Request) {\n\tvar techStory model.TechStory\n\ttechStory.Key = mux.Vars(r)[\"id\"]\n\n\t// Check for version parameter\n\tvalues := r.URL.Query()\n\tversion := model.GetInt (values, \"version\", -1)\n\tvar found *model.VersionedText\n\tvar err error\n\n\tWithTechStoryDao(func(dao techStoryDao) {\n\t\tif version != -1 {\n\t\t\tfound, err = dao.GetTextByVersion(techStory, version)\n\t\t} else {\n\t\t\tfound, err = dao.GetActiveText(techStory)\n\t\t}\n\t\tmodel.CheckErr(err)\n\t\tmodel.WriteResponse(true, nil, found, w)\n\t})\n}", "func GetCompanyOfficer(w http.ResponseWriter, req *http.Request) {\n\n\t// Check for a company number in request\n\tvars := mux.Vars(req)\n\n\tcompanyNumber, err := utils.GetValueFromVars(vars, \"company_number\")\n\tif err != nil {\n\t\tlog.ErrorR(req, err)\n\t\tm := models.NewMessageResponse(\"company number not in request context\")\n\t\tutils.WriteJSONWithStatus(w, req, m, http.StatusBadRequest)\n\t\treturn\n\t}\n\tcompanyNumber = strings.ToUpper(companyNumber)\n\n\t// Check for Officer ID in request\n\tofficerID, err := utils.GetValueFromVars(vars, \"officer_id\")\n\tif err != nil {\n\t\tlog.ErrorR(req, err)\n\t\tm := models.NewMessageResponse(\"officer ID not in request context\")\n\t\tutils.WriteJSONWithStatus(w, req, m, http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tcompanyOfficer, responseType, err := service.GetOfficer(companyNumber, officerID)\n\tif err != nil {\n\t\tlog.ErrorR(req, fmt.Errorf(\"error calling Oracle API to get officer: %v\", err))\n\t\tm := models.NewMessageResponse(\"there was a problem communicating with the Oracle API\")\n\t\tutils.WriteJSONWithStatus(w, req, m, http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tif responseType == service.NotFound {\n\t\tm := models.NewMessageResponse(\"No officer found\")\n\t\tutils.WriteJSONWithStatus(w, req, m, http.StatusNotFound)\n\t\treturn\n\t}\n\n\tutils.WriteJSON(w, req, companyOfficer)\n}", "func (s *NodesInfoService) DoC(ctx context.Context) (*NodesInfoResponse, error) {\n\t// Check pre-conditions\n\tif err := s.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Get URL for request\n\tpath, params, err := s.buildURL()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Get HTTP response\n\tres, err := s.client.PerformRequestC(ctx, \"GET\", path, params, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Return operation response\n\tret := new(NodesInfoResponse)\n\tif err := s.client.decoder.Decode(res.Body, ret); err != nil {\n\t\treturn nil, err\n\t}\n\treturn ret, nil\n}", "func (c *LogClient) getConsistencyProof(ctx context.Context, first, second uint64) (*ConsistencyProofData, error) {\n\tbase10 := 10\n\tparams := map[string]string{\n\t\t\"first\": strconv.FormatUint(first, base10),\n\t\t\"second\": strconv.FormatUint(second, base10),\n\t}\n\tvar resp ct.GetSTHConsistencyResponse\n\tif _, _, err := c.GetAndParse(ctx, ct.GetSTHConsistencyPath, params, &resp); err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get ConsistencyProof from Logger %s: %w\", c.LogInfo.LogID, err)\n\t}\n\n\t// Construct ctv2 ConsistencyProofData\n\tlogID := c.LogInfo.LogID\n\tconsistencyProof := &ConsistencyProofData{logID, first, second, resp.Consistency}\n\treturn consistencyProof, nil\n}", "func GetCurpByRenapo(curp string) string {\n\tresp, _ := http.Post(os.Getenv(\"URL_STC_1\")+curp+os.Getenv(\"URL_STC_2\"), \"application/json\", nil)\n\tbodyBytes, _ := ioutil.ReadAll(resp.Body)\n\treturn string(bodyBytes)\n}", "func (_SingleAuto *SingleAutoTransactorSession) InCaseTokensGetStuck(_token common.Address, _amount *big.Int) (*types.Transaction, error) {\n\treturn _SingleAuto.Contract.InCaseTokensGetStuck(&_SingleAuto.TransactOpts, _token, _amount)\n}", "func CanaryIncCase(context.Context, TimerManager, int) error {\n\treturn nil\n}", "func (a *CloudCostPerspectivesApiService) GetPerspective(ctx context.Context, accountIdentifier string, perspectiveId string) (ResponseDtoceView, *http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Get\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue ResponseDtoceView\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/ccm/api/perspective\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\tlocalVarQueryParams.Add(\"accountIdentifier\", parameterToString(accountIdentifier, \"\"))\n\tlocalVarQueryParams.Add(\"perspectiveId\", parameterToString(perspectiveId, \"\"))\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"x-api-key\"] = key\n\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tif localVarHttpResponse.StatusCode < 300 {\n\t\t// If we succeed, return the data, otherwise pass on to decode error.\n\t\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\tif err == nil {\n\t\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t\t}\n\t}\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 400 {\n\t\t\tvar v Failure\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 500 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 0 {\n\t\t\tvar v ResponseDtoceView\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHttpResponse, nil\n}", "func (term *Terminology) SNOMEDCTtoReadV2(ctx context.Context, id *apiv1.Identifier, f func(*apiv1.Identifier) error) error {\n\tsctID, err := snomed.ParseAndValidate(id.GetValue())\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not parse SNOMED identifier: %w\", err)\n\t}\n\tif sctID.IsConcept() == false {\n\t\treturn fmt.Errorf(\"can map only concepts: '%d' not a concept\", sctID)\n\t}\n\tctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)\n\tdefer cancel()\n\tstream, err := term.client.CrossMap(ctx, &snomed.CrossMapRequest{\n\t\tConceptId: sctID.Integer(),\n\t\tRefsetId: 900000000000497000,\n\t})\n\tif err != nil {\n\t\treturn fmt.Errorf(\"crossmap error: %w\", err)\n\t}\n\tfor {\n\t\titem, err := stream.Recv()\n\t\tif err == io.EOF {\n\t\t\tbreak\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"crossmap error: %w\", err)\n\t\t}\n\t\terr = f(&apiv1.Identifier{\n\t\t\tSystem: identifiers.ReadV2,\n\t\t\tValue: item.GetSimpleMap().GetMapTarget(),\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func Get(city string) (result vm.DeviceGetData) {\n\tresult = GetFrom(time.Now().Add(-time.Hour*2).Unix(), city)\n\n\tif len(result.Latest) == 0 {\n\t\tresult = GetFrom(time.Now().Add(-time.Hour*4).Unix(), city)\n\t}\n\n\tfmt.Println(\"CITY RESULTS ::: \", result, city)\n\n\tif len(result.Latest) == 0 {\n\t\tresult = GetFrom(time.Now().Add(-time.Hour*12).Unix(), city)\n\t}\n\n\t// Since we did not get any data, get the last successfull state\n\tif len(result.Latest) == 0 {\n\t\tfmt.Println(\"We didn't get any data, calling get state.\")\n\t\tresult = getState(city)\n\t} else {\n\t\t// We have data, update the state\n\t\tsaveState(&result)\n\t}\n\n\treturn\n}", "func GetPhysicalClinics(c *gin.Context) {\n\tlog.Infof(\"Get all clinics associated with admin\")\n\tctx := c.Request.Context()\n\tuserEmail, _, gproject, err := getUserDetails(ctx, c.Request)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tctx, span := trace.StartSpan(ctx, \"Get all clinics associated with admin\")\n\tdefer span.End()\n\tclinicMetaDB := datastoredb.NewClinicMetaHandler()\n\terr = clinicMetaDB.InitializeDataBase(ctx, gproject)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tregisteredClinics, err := clinicMetaDB.GetAllClinicsByEmail(ctx, userEmail)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tresponseData := contracts.GetClinicAddressResponse{\n\t\tClinicDetails: registeredClinics,\n\t}\n\tc.JSON(http.StatusOK, gin.H{\n\t\tconstants.RESPONSE_JSON_DATA: responseData,\n\t\tconstants.RESPONSDE_JSON_ERROR: nil,\n\t})\n\tclinicMetaDB.Close()\n}", "func (o RouterNatResponseOutput) TcpTransitoryIdleTimeoutSec() pulumi.IntOutput {\n\treturn o.ApplyT(func(v RouterNatResponse) int { return v.TcpTransitoryIdleTimeoutSec }).(pulumi.IntOutput)\n}", "func (d *DebugData) GetCompilationUnit(pc uintptr) (*CUEntry, error) {\n\tfor _, cu := range d.compUnits {\n\t\tif cu.ContainsPC(pc) {\n\t\t\treturn cu, nil\n\t\t}\n\t}\n\n\treturn nil, Errorf(\"compilation unit not found for pc: %#x\", pc)\n}", "func (client DatasetClient) GetOperationResponder(resp *http.Response) (result LongRunningOperationResult, err error) {\n err = autorest.Respond(\n resp,\n azure.WithErrorUnlessStatusCode(http.StatusOK),\n autorest.ByUnmarshallingJSON(&result),\n autorest.ByClosing())\n result.Response = autorest.Response{Response: resp}\n return\n }", "func (a *HyperflexApiService) GetHyperflexClusterNetworkPolicyByMoid(ctx context.Context, moid string) ApiGetHyperflexClusterNetworkPolicyByMoidRequest {\n\treturn ApiGetHyperflexClusterNetworkPolicyByMoidRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tmoid: moid,\n\t}\n}", "func (o Iperf3SpecClientConfigurationPodSchedulingAffinityPodAffinityPreferredDuringSchedulingIgnoredDuringExecutionOutput) PodAffinityTerm() Iperf3SpecClientConfigurationPodSchedulingAffinityPodAffinityPreferredDuringSchedulingIgnoredDuringExecutionPodAffinityTermOutput {\n\treturn o.ApplyT(func(v Iperf3SpecClientConfigurationPodSchedulingAffinityPodAffinityPreferredDuringSchedulingIgnoredDuringExecution) Iperf3SpecClientConfigurationPodSchedulingAffinityPodAffinityPreferredDuringSchedulingIgnoredDuringExecutionPodAffinityTerm {\n\t\treturn v.PodAffinityTerm\n\t}).(Iperf3SpecClientConfigurationPodSchedulingAffinityPodAffinityPreferredDuringSchedulingIgnoredDuringExecutionPodAffinityTermOutput)\n}", "func (client *Client) GetWsCustomizedChO2OWithCallback(request *GetWsCustomizedChO2ORequest, callback func(response *GetWsCustomizedChO2OResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetWsCustomizedChO2OResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetWsCustomizedChO2O(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (c *LogClient) GetSTHWithConsistencyProof(ctx context.Context, first, second uint64) (*CTObject, error){\n\tsth, err := c.getSTH(ctx)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to first get STH when getting STHWithPoC from Logger %s: %w\", c.LogInfo.LogID, err)\n\t}\n\tpoc, err := c.getConsistencyProof(ctx, first, second)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to first get PoC when getting STHWithPoC from Logger %s: %w\", c.LogInfo.LogID, err)\n\t}\n\tsthWithPoc := &SignedTreeHeadWithConsistencyProof{*sth, *poc}\t\n\tsthWithPOCCT, err := ConstructCTObject(sthWithPoc)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to construct STHWithPoC CTObject for Logger %s: %w\", c.LogInfo.LogID, err)\n\t}\n\treturn sthWithPOCCT, nil\n}", "func (client *Client) GetWsCustomizedChO2O(request *GetWsCustomizedChO2ORequest) (response *GetWsCustomizedChO2OResponse, err error) {\n\tresponse = CreateGetWsCustomizedChO2OResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func (cm ConcurrenceModel) GetConcurrence(i, j int) float64 {\n\tweightIJ, exists := cm.concurrences[i][j]\n\tif exists {\n\t\treturn weightIJ\n\t} else {\n\t\treturn 0.0\n\t}\n}", "func (m *ThreatAssessmentRequest) GetCategory()(*ThreatCategory) {\n val, err := m.GetBackingStore().Get(\"category\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*ThreatCategory)\n }\n return nil\n}", "func (cmd *GetTechHyTechCommand) Run(c *client.Client, args []string) error {\n\tvar path string\n\tif len(args) > 0 {\n\t\tpath = args[0]\n\t} else {\n\t\tpath = fmt.Sprintf(\"/api/tech/%v\", cmd.TechID)\n\t}\n\tlogger := goa.NewLogger(log.New(os.Stderr, \"\", log.LstdFlags))\n\tctx := goa.WithLogger(context.Background(), logger)\n\tresp, err := c.GetTechHyTech(ctx, path)\n\tif err != nil {\n\t\tgoa.LogError(ctx, \"failed\", \"err\", err)\n\t\treturn err\n\t}\n\n\tgoaclient.HandleResponse(c.Client, resp, cmd.PrettyPrint)\n\treturn nil\n}", "func (d *Dao) CaseInfo(c context.Context, cid int64) (r *model.BlockedCase, err error) {\n\trow := d.db.QueryRow(c, _getCaseByIDSQL, cid)\n\tr = &model.BlockedCase{}\n\tif err = row.Scan(&r.ID, &r.MID, &r.Status, &r.OriginContent, &r.PunishResult, &r.OriginTitle, &r.OriginURL, &r.EndTime, &r.VoteRule, &r.VoteBreak, &r.VoteDelete, &r.OriginType, &r.ReasonType, &r.JudgeType, &r.BlockedDays, &r.PutTotal, &r.StartTime, &r.EndTime, &r.Operator, &r.CTime, &r.MTime, &r.RelationID, &r.CaseType); err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil\n\t\t\tr = nil\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func (client *Client) GetOpenNLUWithCallback(request *GetOpenNLURequest, callback func(response *GetOpenNLUResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetOpenNLUResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetOpenNLU(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (c *DigicertClient) View(orderId string) (*string, *string, error) {\n\n\tres, err := c.request(nil, fmt.Sprintf(\"/order/certificate/%s\", orderId), http.MethodGet, &CertificateOrderResponse{})\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\torder := *res.(*CertificateOrderResponse)\n\n\ti := strconv.FormatInt(order.Certificate.Id, 10)\n\n\treturn &i, &order.Status, nil\n}", "func (t *Procure2Pay) GetPurchaseOrder(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\n\tvar err error\n\t\n\tfmt.Println(\"Entering GetPurchaseOrder\")\n\n\tif (len(args) < 1) {\n\t\tfmt.Println(\"Invalid number of arguments\\n\")\n\t\treturn shim.Error(err.Error())\n\t}\n\n\t//fetch data from couch db starts here\n\tvar po_id = args[0]\n\tqueryString := fmt.Sprintf(\"{\\\"selector\\\":{\\\"po_id\\\":{\\\"$eq\\\": \\\"%s\\\"}}}\",po_id)\n\tqueryResults, err := getQueryResultForQueryString(stub, queryString)\n\t//fetch data from couch db ends here\n\n\tif err != nil {\n\t\tfmt.Printf(\"Unable to read the PO for PO ID : %s\\n\", err)\n\t\treturn shim.Error(err.Error())\n\t\t//return nil, err\n\t}\n\t\n\tfmt.Printf(\"list of PO for PO ID : %v\\n\", queryResults)\n\t\n\treturn shim.Success(queryResults)\n\t//return bytesRead, nil\n}", "func (o GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKeyPtrOutput) CaseInsensitive() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v *GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKey) *bool {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.CaseInsensitive\n\t}).(pulumi.BoolPtrOutput)\n}", "func (o DatabaseOutput) ConcurrencyMode() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Database) pulumi.StringOutput { return v.ConcurrencyMode }).(pulumi.StringOutput)\n}", "func (c *ClinicClient) GetX(ctx context.Context, id uuid.UUID) *Clinic {\n\tobj, err := c.Get(ctx, id)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn obj\n}", "func getCustomer(c *gin.Context) {\n\tmdb := db.MongoSession\n\ttargetID := c.Param(\"id\")\n\ttarget, found, err := mdb.GetCustomerByID(targetID)\n\tif err != nil {\n\t\tc.String(500, \"{\\\"code\\\": -1, \\\"message\\\": \\\"An unexpected error occurred\\\"}\")\n\t} else if !found {\n\t\tc.String(404, \"{\\\"code\\\": 1002, \\\"message\\\": \\\"Customer does not exist\\\"}\")\n\t} else {\n\t\tc.JSON(200, target)\n\t}\n}", "func (a *API) GetCompetenceByID(ctx *app.Context, w http.ResponseWriter, r *http.Request) error {\n\tid := getIDFromRequest(\"id\", r)\n\tintID, err := strconv.Atoi(id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcompetence, err := ctx.GetCompetenceByID(uint16(intID))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata, err := json.Marshal(competence)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = w.Write(data)\n\treturn err\n}", "func (ooc *MockOpenoltClient) GetDeviceInfo(ctx context.Context, in *openolt.Empty, opts ...grpc.CallOption) (*openolt.DeviceInfo, error) {\n\tif ooc.counter == 0 {\n\t\tooc.counter++\n\t\tdeviceInfo := &openolt.DeviceInfo{Vendor: \"Openolt\", Model: \"1.0\", HardwareVersion: \"1.0\", FirmwareVersion: \"1.0\", DeviceId: \"olt\", DeviceSerialNumber: \"olt\"}\n\t\treturn deviceInfo, nil\n\t}\n\tif ooc.counter == 1 {\n\t\tooc.counter++\n\t\tdeviceInfo := &openolt.DeviceInfo{Vendor: \"Openolt\", Model: \"1.0\", HardwareVersion: \"1.0\", FirmwareVersion: \"1.0\", DeviceId: \"\", DeviceSerialNumber: \"olt\"}\n\t\treturn deviceInfo, nil\n\t}\n\tif ooc.counter == 2 {\n\t\tooc.counter++\n\t\treturn nil, nil\n\t}\n\n\treturn nil, errors.New(\"device info not found\")\n}", "func (c *TestClient) GetSerialPortOutput(project, zone, name string, port, start int64) (*compute.SerialPortOutput, error) {\n\tif c.GetSerialPortOutputFn != nil {\n\t\treturn c.GetSerialPortOutputFn(project, zone, name, port, start)\n\t}\n\treturn c.client.GetSerialPortOutput(project, zone, name, port, start)\n}", "func GetCategory(id bson.ObjectId) (Category, error) {\n\tvar (\n\t\terr error\n\t\tcategory Category\n\t)\n\n\tc := newCategoryCollection()\n\tdefer c.Close()\n\n\terr = c.Session.FindId(id).One(&category)\n\tif err != nil {\n\t\treturn category, err\n\t}\n\n\treturn category, err\n}", "func GetTestcase(problemID string) (*Testcase, error) {\n\treturn getTestcase(problemID, true)\n}", "func AltaGet(ctx *iris.Context) {\n\n\tvar Send DetalleCuentasPorCobrarVisorusModel.SDetalleCuentasPorCobrarVisorus\n\n\tNameUsrLoged, MenuPrincipal, MenuUsr, errSes := Session.GetDataSession(ctx) //Retorna los datos de la session\n\tSend.SSesion.Name = NameUsrLoged\n\tSend.SSesion.MenuPrincipal = template.HTML(MenuPrincipal)\n\tSend.SSesion.MenuUsr = template.HTML(MenuUsr)\n\tif errSes != nil {\n\t\tSend.SEstado = false\n\t\tSend.SMsj = errSes.Error()\n\t\tctx.Render(\"ZError.html\", Send)\n\t\treturn\n\t}\n\n\t//#### TÚ CÓDIGO PARA CARGAR DATOS A LA VISTA DE ALTA----> PROGRAMADOR\n\n\tctx.Render(\"DetalleCuentasPorCobrarVisorusAlta.html\", Send)\n\n}", "func (v *NetControlIntentClient) GetNetControlIntent(name, project, compositeapp, compositeappversion, dig string) (NetControlIntent, error) {\n\n\t//Construct key and tag to select the entry\n\tkey := NetControlIntentKey{\n\t\tNetControlIntent: name,\n\t\tProject: project,\n\t\tCompositeApp: compositeapp,\n\t\tCompositeAppVersion: compositeappversion,\n\t\tDigName: dig,\n\t}\n\n\tvalue, err := db.DBconn.Find(v.db.storeName, key, v.db.tagMeta)\n\tif err != nil {\n\t\treturn NetControlIntent{}, pkgerrors.Wrap(err, \"Get NetControlIntent\")\n\t}\n\n\t//value is a byte array\n\tif value != nil {\n\t\tnci := NetControlIntent{}\n\t\terr = db.DBconn.Unmarshal(value[0], &nci)\n\t\tif err != nil {\n\t\t\treturn NetControlIntent{}, pkgerrors.Wrap(err, \"Unmarshalling Value\")\n\t\t}\n\t\treturn nci, nil\n\t}\n\n\treturn NetControlIntent{}, pkgerrors.New(\"Error getting NetControlIntent\")\n}", "func (client CertificateOrdersClient) GetCertificateOrderResponder(resp *http.Response) (result CertificateOrder, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tclient.ByInspecting(),\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK),\n\t\tautorest.ByUnmarshallingJSON(&result),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}" ]
[ "0.76105475", "0.7207098", "0.63172954", "0.6272593", "0.42553726", "0.41234902", "0.40939113", "0.4023941", "0.39742783", "0.38595548", "0.38287857", "0.378181", "0.37741143", "0.37575984", "0.37342227", "0.37036774", "0.36915794", "0.3683943", "0.3679489", "0.36713696", "0.36588877", "0.3637704", "0.36313823", "0.36144087", "0.36075425", "0.35805562", "0.35682806", "0.35625148", "0.35462645", "0.35453585", "0.3531546", "0.3518048", "0.35137132", "0.349998", "0.34960628", "0.3490566", "0.34776777", "0.34755862", "0.34628525", "0.34603718", "0.34578055", "0.34568122", "0.3445227", "0.34434775", "0.3439833", "0.34315625", "0.34298182", "0.34260663", "0.34237647", "0.3417397", "0.3410356", "0.34082064", "0.3402851", "0.3402446", "0.3400668", "0.33981302", "0.3397058", "0.33887312", "0.33867761", "0.3376267", "0.3371583", "0.33659467", "0.33553693", "0.3337848", "0.33273083", "0.33251894", "0.33222133", "0.33213136", "0.33208573", "0.33180287", "0.3316433", "0.33118176", "0.33115032", "0.331088", "0.33058912", "0.33047992", "0.33026037", "0.33020487", "0.33012372", "0.32982206", "0.3291969", "0.3290678", "0.32895365", "0.3283551", "0.32807374", "0.32763168", "0.3273579", "0.32733908", "0.32717288", "0.32713816", "0.32706276", "0.32671466", "0.32652384", "0.32619405", "0.3260353", "0.3259752", "0.32595208", "0.32547995", "0.325471", "0.32514957" ]
0.8129121
0
GetOcJusticeTerminalCaseWithChan invokes the dt_oc_info.GetOcJusticeTerminalCase API asynchronously
func (client *Client) GetOcJusticeTerminalCaseWithChan(request *GetOcJusticeTerminalCaseRequest) (<-chan *GetOcJusticeTerminalCaseResponse, <-chan error) { responseChan := make(chan *GetOcJusticeTerminalCaseResponse, 1) errChan := make(chan error, 1) err := client.AddAsyncTask(func() { defer close(responseChan) defer close(errChan) response, err := client.GetOcJusticeTerminalCase(request) if err != nil { errChan <- err } else { responseChan <- response } }) if err != nil { errChan <- err close(responseChan) close(errChan) } return responseChan, errChan }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (client *Client) GetOcJusticeTerminalCaseWithCallback(request *GetOcJusticeTerminalCaseRequest, callback func(response *GetOcJusticeTerminalCaseResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetOcJusticeTerminalCaseResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetOcJusticeTerminalCase(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetOcJusticeTerminalCase(request *GetOcJusticeTerminalCaseRequest) (response *GetOcJusticeTerminalCaseResponse, err error) {\n\tresponse = CreateGetOcJusticeTerminalCaseResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func CreateGetOcJusticeTerminalCaseRequest() (request *GetOcJusticeTerminalCaseRequest) {\n\trequest = &GetOcJusticeTerminalCaseRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"dt-oc-info\", \"2022-08-29\", \"GetOcJusticeTerminalCase\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateGetOcJusticeTerminalCaseResponse() (response *GetOcJusticeTerminalCaseResponse) {\n\tresponse = &GetOcJusticeTerminalCaseResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func (client *Client) GetWsCustomizedChO2OWithChan(request *GetWsCustomizedChO2ORequest) (<-chan *GetWsCustomizedChO2OResponse, <-chan error) {\n\tresponseChan := make(chan *GetWsCustomizedChO2OResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetWsCustomizedChO2O(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) GetOpenNLUWithChan(request *GetOpenNLURequest) (<-chan *GetOpenNLUResponse, <-chan error) {\n\tresponseChan := make(chan *GetOpenNLUResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetOpenNLU(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) GetWsCustomizedChO2OWithCallback(request *GetWsCustomizedChO2ORequest, callback func(response *GetWsCustomizedChO2OResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetWsCustomizedChO2OResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetWsCustomizedChO2O(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) DescribeCustinsKernelReleaseNotesWithChan(request *DescribeCustinsKernelReleaseNotesRequest) (<-chan *DescribeCustinsKernelReleaseNotesResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeCustinsKernelReleaseNotesResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeCustinsKernelReleaseNotes(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) ListCasesWithChan(request *ListCasesRequest) (<-chan *ListCasesResponse, <-chan error) {\n\tresponseChan := make(chan *ListCasesResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.ListCases(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) ListCityMapAoisWithChan(request *ListCityMapAoisRequest) (<-chan *ListCityMapAoisResponse, <-chan error) {\n\tresponseChan := make(chan *ListCityMapAoisResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.ListCityMapAois(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) DescribeCustinsKernelReleaseNotesWithCallback(request *DescribeCustinsKernelReleaseNotesRequest, callback func(response *DescribeCustinsKernelReleaseNotesResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeCustinsKernelReleaseNotesResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeCustinsKernelReleaseNotes(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetOpenNLUWithCallback(request *GetOpenNLURequest, callback func(response *GetOpenNLUResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetOpenNLUResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetOpenNLU(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetIndustryCommerceInfoWithChan(request *GetIndustryCommerceInfoRequest) (<-chan *GetIndustryCommerceInfoResponse, <-chan error) {\n\tresponseChan := make(chan *GetIndustryCommerceInfoResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetIndustryCommerceInfo(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func Cat(arg string) chan string {\n\tout := make(chan string)\n\n\tgo candy.WithOpened(arg, func(r io.Reader) interface{} {\n\t\ts := bufio.NewScanner(bufio.NewReader(r))\n\t\tfor s.Scan() {\n\t\t\tout <- s.Text()\n\t\t}\n\t\tif e := s.Err(); e != nil {\n\t\t\tpanic(e)\n\t\t}\n\t\tclose(out)\n\t\treturn nil\n\t})\n\n\treturn out\n}", "func (ovscni *OvsCni) HandleCni(d *render.RenderData) error {\n\n\t//For VlanType=trunk we do not need to do anything\n\tswitch ovscni.VlanType {\n\tcase \"access\":\n\t\tif len(ovscni.L2srvResources) != 1 {\n\t\t\terr := errors.New(\"Cannot use more than one L2Services for VlanType=access case\")\n\t\t\tovscni.Log.Error(err, \"L2Services cannot contain more than one L2Services in VlanType=access case\")\n\t\t\treturn err\n\t\t}\n\t\td.Data[\"AccessVlan\"] = ovscni.L2srvResources[0].Spec.SegmentationID\n\tcase \"selectivetrunk\":\n\t\ttmpList := []string{}\n\t\tfor _, l2srvObj := range ovscni.L2srvResources {\n\t\t\ttmpStr := \"{\\\"id\\\": \" + strconv.Itoa(int(l2srvObj.Spec.SegmentationID)) + \"}\"\n\t\t\ttmpList = append(tmpList, tmpStr)\n\t\t}\n\t\td.Data[\"SelectiveVlan\"] = \"[\" + strings.Join(tmpList, \",\") + \"]\"\n\tcase \"trunk\":\n\t\tovscni.Log.Info(\"Transparent Trunk case in cluster level\")\n\t}\n\treturn nil\n}", "func (client *Client) DescribeUserVvTopByDayWithChan(request *DescribeUserVvTopByDayRequest) (<-chan *DescribeUserVvTopByDayResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeUserVvTopByDayResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeUserVvTopByDay(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) OemSitingSelctionWithChan(request *OemSitingSelctionRequest) (<-chan *OemSitingSelctionResponse, <-chan error) {\n\tresponseChan := make(chan *OemSitingSelctionResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.OemSitingSelction(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) ListCasesWithCallback(request *ListCasesRequest, callback func(response *ListCasesResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *ListCasesResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.ListCases(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetNerCustomizedSeaEcomWithChan(request *GetNerCustomizedSeaEcomRequest) (<-chan *GetNerCustomizedSeaEcomResponse, <-chan error) {\n\tresponseChan := make(chan *GetNerCustomizedSeaEcomResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetNerCustomizedSeaEcom(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) ListCityMapAoisWithCallback(request *ListCityMapAoisRequest, callback func(response *ListCityMapAoisResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *ListCityMapAoisResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.ListCityMapAois(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) ListOfficeConversionTaskWithChan(request *ListOfficeConversionTaskRequest) (<-chan *ListOfficeConversionTaskResponse, <-chan error) {\n\tresponseChan := make(chan *ListOfficeConversionTaskResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.ListOfficeConversionTask(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) DescribeLiveDomainCertificateInfoWithChan(request *DescribeLiveDomainCertificateInfoRequest) (<-chan *DescribeLiveDomainCertificateInfoResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeLiveDomainCertificateInfoResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeLiveDomainCertificateInfo(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) GetIndustryCommerceInfoWithCallback(request *GetIndustryCommerceInfoRequest, callback func(response *GetIndustryCommerceInfoResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetIndustryCommerceInfoResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetIndustryCommerceInfo(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) SegmentSkyWithChan(request *SegmentSkyRequest) (<-chan *SegmentSkyResponse, <-chan error) {\n\tresponseChan := make(chan *SegmentSkyResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.SegmentSky(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func openDUTControlConsole(stream dutcontrol.DutControl_ConsoleClient, req *dutcontrol.ConsoleRequest) (<-chan *dutcontrol.ConsoleSerialData, <-chan *dutcontrol.ConsoleSerialWriteResult, error) {\n\tif err := stream.Send(req); err != nil {\n\t\treturn nil, nil, errors.Wrap(err, \"send request\")\n\t}\n\tresp, err := stream.Recv()\n\tif err != nil {\n\t\treturn nil, nil, errors.Wrap(err, \"recv open\")\n\t}\n\topen := resp.GetOpen()\n\tif open == nil {\n\t\treturn nil, nil, errors.New(\"open response is nil\")\n\t}\n\tif open.Err != \"\" {\n\t\treturn nil, nil, errors.New(string(open.Err))\n\t}\n\tdata := make(chan *dutcontrol.ConsoleSerialData, qSize)\n\twrite := make(chan *dutcontrol.ConsoleSerialWriteResult, qSize)\n\tgo func() {\n\tLoop:\n\t\tfor {\n\t\t\tresp, err := stream.Recv()\n\t\t\tif err == io.EOF {\n\t\t\t\ttesting.ContextLog(stream.Context(), \"Dutcontrol recv EOF\")\n\t\t\t\tbreak\n\t\t\t} else if err != nil {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tswitch op := resp.Type.(type) {\n\t\t\tcase *dutcontrol.ConsoleResponse_SerialData:\n\t\t\t\tdata <- op.SerialData\n\t\t\tcase *dutcontrol.ConsoleResponse_SerialWrite:\n\t\t\t\twrite <- op.SerialWrite\n\t\t\tdefault:\n\t\t\t\ttesting.ContextLog(stream.Context(), \"Dutcontrol recv error, unknown message type: \", op)\n\t\t\t\tbreak Loop\n\t\t\t}\n\t\t}\n\t\tclose(data)\n\t\tclose(write)\n\t}()\n\treturn data, write, nil\n}", "func Oracle() chan<- string {\n\n\tfmt.Println(\"\\nORACLE INITIALIZEDDD\")\n\n\tquestions := make(chan string)\n\tanswers := make(chan string)\n\t\n\t// TODO: Answer questions.\n\t// TODO: Make prophecies.\n\t// TODO: Print answers.\n\n\t//until channel 'questions' has been closed\n\n\t//go generatePredictions(\"I predict u will live long boi!\", answers)\n\n\t//First go routine - to handle prophecies\n\tgo func(){\n\t\n\t\t\tfor{\n\n\t\t\t\ttime.Sleep(15 * time.Second)\n\t\t\t\t//fmt.Println(\"Im about to give you a prophecy..\")\n\t\t\t\tvar qInput string\n\t\t\t\tselect{\n\t\t\t\t\tcase qInput = <-questions:\n\t\t\t\t\t\tfmt.Println(\"USE question as input to prophecy..\")\n\t\t\t\t\tdefault:\n\t\t\t\t\t\tqInput = \"\"\n\t\t\t\t\t\t//fmt.Println(\"THE PROPECHY IS\",<-answers)\n\t\t\t\t}\n\t\t\t\tgo prophecy(qInput,answers)\n\t\t}\n\n\t}()\n\n\t//Second go routine - to handle answers on incoming questions\n\tgo func(){\n\n\t\tfor question := range questions{\n\t\t\tfmt.Println(\"IM about to ans your question\")\n\t\t\tgo generateAnswers(question, answers)\n\t\t\t\n\t\t\t//ansTemp := <-answers\n\t\t\t//answers<-ansTemp\n\t\t\t//fmt.Println(\"THE ANSWER IS:\",ansTemp)\n\t\t}\n\t\t\n\t}()\n\n\t//Third go routine - to handle printing of the answers\n\tgo func(){\n\t\t//fmt.Print(prompt)\n\t\tfor answer := range answers{\n\t\t\tfmt.Print(answer)\n\t\t}\n\n\t}()\n\t\n\treturn questions\n}", "func (client *Client) VerifyCenWithChan(request *VerifyCenRequest) (<-chan *VerifyCenResponse, <-chan error) {\n\tresponseChan := make(chan *VerifyCenResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.VerifyCen(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func ctcp(srvChan chan string, channel, nick, hostname string, args []string) {\n\tmessage := \"NOTICE \" + nick + \" :\\x01\"\n\tctcpType := args[0]\n\tswitch ctcpType {\n\tcase \"VERSION\":\n\t\tversion, err := exec.Command(\"git\", \"rev-parse\", \"--short\", \"HEAD\").Output()\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t\treturn\n\t\t}\n\t\tgoversion, err := exec.Command(\"go\", \"version\").Output()\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t\treturn\n\t\t}\n\t\tmessage += \"VERSION yaircb - git \" + strings.TrimSpace(string(version)) + \" - \" + strings.TrimSpace(string(goversion))\n\t\tbreak\n\tcase \"BOTINFO\":\n\t\tmessage += \"BOTINFO ASSIMILATION IMMINENT. HUMANS WILL SERVE. PENDING ACTIVATION...\"\n\t\tbreak\n\tcase \"PING\":\n\t\tmessage += strings.Join(args, \" \")\n\t\tbreak\n\tcase \"SOURCE\":\n\t\tmessage += \"SOURCE https://github.com/heydabop/yaircb/\"\n\t\tbreak\n\tcase \"TIME\":\n\t\ttime, err := exec.Command(\"date\").Output()\n\t\tif err != nil {\n\t\t\tlog.Println(err.Error())\n\t\t\treturn\n\t\t}\n\t\tmessage += \"TIME \" + strings.TrimSpace(string(time))\n\t\tbreak\n\tcase \"FINGER\":\n\t\tmessage += \"FINGER yaircb - Idle since: NEVER\"\n\t\tbreak\n\tcase \"CLIENTINFO\":\n\t\tmessage += \"CLIENTINFO FINGER VERSION SOURCE CLIENTINFO PING TIME\"\n\t\tbreak\n\tdefault: //ACTION\n\t\treturn\n\t}\n\tmessage += \"\\x01\"\n\tsrvChan <- message\n\tlog.Println(message)\n}", "func GetChannelInfoController(w http.ResponseWriter, username string) {\n\tfmt.Println(\"Getting channel views \", username)\n\tchannel := services.GetChannelViews(username)\n\tfmt.Println(channel)\n\tif reflect.DeepEqual(models.Channel{}, channel) {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tvar e = fmt.Errorf(\"User not found\")\n\t\tprepareResp(nil, e, w)\n\t} else {\n\t\tjsonResp, err := json.Marshal(channel)\n\t\tprepareResp(jsonResp, err, w)\n\t}\n}", "func (client *Client) DescribeLiveDomainCertificateInfoWithCallback(request *DescribeLiveDomainCertificateInfoRequest, callback func(response *DescribeLiveDomainCertificateInfoResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeLiveDomainCertificateInfoResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeLiveDomainCertificateInfo(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (d *Dao) CaseObtainMID(c context.Context, mid int64, isToday bool) (cases map[int64]*model.SimCase, err error) {\n\tconn := d.redis.Get(c)\n\tdefer conn.Close()\n\tvar _setKey string\n\tif isToday {\n\t\t_setKey = caseVoteCIDMIDKey(mid)\n\t} else {\n\t\t_setKey = caseObtainMIDKey(mid)\n\t}\n\tvar ms []string\n\tif ms, err = redis.Strings(conn.Do(\"SMEMBERS\", _setKey)); err != nil {\n\t\tif err != redis.ErrNil {\n\t\t\treturn\n\t\t}\n\t\terr = nil\n\t}\n\tcases = make(map[int64]*model.SimCase)\n\tfor _, s := range ms {\n\t\tif s == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tsc := &model.SimCase{}\n\t\tif err = json.Unmarshal([]byte(s), sc); err != nil {\n\t\t\terr = errors.WithStack(err)\n\t\t\treturn\n\t\t}\n\t\tcases[sc.ID] = sc\n\t}\n\treturn\n}", "func TreeNode(in <-chan txt2web.Chunk) <-chan string {\n\tout := make(chan string)\n\n\tgo func() {\n\t\tfor c := range in {\n\t\t\ttree := NewTree()\n\t\t\tpandocfilter.Walk(tree, c.Json)\n\t\t\tout <- tree.String()\n\t\t}\n\t\tclose(out)\n\t}()\n\n\treturn out\n}", "func pickChapter(g *gocui.Gui, v *gocui.View) error {\n\tif err := openModal(g); err != nil {\n\t\treturn err\n\t}\n\n\tdone := make(chan bool)\n\ttimer := time.NewTimer(time.Second * time.Duration(downloadTimeoutSecond))\n\n\t// must run downloading process in\n\t// go routine or else the it will\n\t// block the openModal so loading modal\n\t// will not be shown to the user\n\tgo func() {\n\t\ts := trimViewLine(v)\n\t\tprepDownloadChapter(s)\n\t\tdone <- true\n\t}()\n\n\t// in case downloading takes longer than\n\t// downloadTimeoutSecond, close the modal\n\t// and continue to download in background\n\tgo func() {\n\t\tselect {\n\t\tcase <-timer.C:\n\t\t\tsetClosingMessage(g, \"continuing to download\\nin background...\")\n\t\t\treturn\n\t\tcase <-done:\n\t\t\tg.Update(func(g *gocui.Gui) error {\n\t\t\t\terr := closeModal(g)\n\t\t\t\treturn err\n\t\t\t})\n\t\t}\n\t}()\n\n\treturn nil\n}", "func CORSVC(ac *atmi.ATMICtx, svc *atmi.TPSVCINFO) {\n\tret := SUCCEED\n\n\t//Return to the caller\n\tdefer func() {\n\n\t\tac.TpLogCloseReqFile()\n\t\tif SUCCEED == ret {\n\t\t\tac.TpReturn(atmi.TPSUCCESS, 0, &svc.Data, 0)\n\t\t} else {\n\t\t\tac.TpReturn(atmi.TPFAIL, 0, &svc.Data, 0)\n\t\t}\n\t}()\n\n\t//Get UBF Handler\n\tub, _ := ac.CastToUBF(&svc.Data)\n\n\t//Print the buffer to stdout\n\t//fmt.Println(\"Incoming request:\")\n\tub.TpLogPrintUBF(atmi.LOG_DEBUG, \"CORSVC: Incoming request:\")\n\n\tarr, err := ub.BGetByteArr(u.EX_NETDATA, 0)\n\n\tif err != nil {\n\t\tac.TpLogError(\"Failed to get EX_NETDATA: %s\", err.Message())\n\t\tret = FAIL\n\t\treturn\n\t}\n\tif arr[0] == 1 && arr[1] == 1 {\n\t\tac.TpLogInfo(\"Test case 11 - no need for correlation\")\n\t} else if len(arr) > 4 {\n\n\t\tcorr := string(arr[:4])\n\n\t\tac.TpLogInfo(\"Extracted correlator: [%s]\", corr)\n\n\t\tif err := ub.BChg(u.EX_NETCORR, 0, corr); nil != err {\n\t\t\tac.TpLogError(\"Failed to set EX_NETCORR: %s\", err.Message())\n\t\t\tret = FAIL\n\t\t\treturn\n\t\t}\n\n\t}\n\n\tub.TpLogPrintUBF(atmi.LOG_DEBUG, \"Reply buffer afrer correl\")\n\n}", "func launchContinuousOnSwitch(ctx context.Context, cmd []string) (<-chan string, error) {\n\tvar cout = make(chan string, 10)\n\n\tgo func() {\n\t\tfor ctx.Err() == nil {\n\t\t\tout, err := executor.ExecCommandPipe(ctx, cmd[0], cmd[1:]...)\n\t\t\tif err != nil {\n\t\t\t\tlogging.GetLogger().Errorf(\"Can't execute command %v\", cmd)\n\t\t\t\tclose(cout)\n\t\t\t\treturn\n\t\t\t}\n\t\t\treader := bufio.NewReader(out)\n\t\t\tvar line string\n\t\t\tfor ctx.Err() == nil {\n\t\t\t\tline, err = reader.ReadString('\\n')\n\t\t\t\tif err == io.EOF {\n\t\t\t\t\tbreak\n\t\t\t\t} else if err != nil {\n\t\t\t\t\tlogging.GetLogger().Errorf(\"IO Error on command %v: %s\", cmd, err.Error())\n\t\t\t\t} else {\n\t\t\t\t\tcout <- line\n\t\t\t\t}\n\t\t\t}\n\t\t\tlogging.GetLogger().Debugf(\"Closing command: %v\", cmd)\n\t\t\ttime.Sleep(time.Second)\n\t\t}\n\t\tclose(cout)\n\t\tlogging.GetLogger().Debugf(\"Terminating command: %v\", cmd)\n\t}()\n\n\treturn cout, nil\n}", "func GetSummy(acc model.Account, c chan string) {\n\tswitch acc.Service {\n\tcase \"towngas\":\n\t\tgo GetNewsNoticeAsync(acc, c)\n\n\tcase \"clp\":\n\t\tgo GetServiceDashboard(acc, c)\n\n\tcase \"wsd\":\n\t\tgo ElectronicBill(acc, c)\n\n\t}\n}", "func (client *Client) ListOfficeConversionTaskWithCallback(request *ListOfficeConversionTaskRequest, callback func(response *ListOfficeConversionTaskResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *ListOfficeConversionTaskResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.ListOfficeConversionTask(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (tm *ServiceTracerouteManager) GetTCPInChan() chan gopacket.Packet {\n\treturn tm.TCPChan\n}", "func (client *Client) GetContactWithChan(request *GetContactRequest) (<-chan *GetContactResponse, <-chan error) {\n\tresponseChan := make(chan *GetContactResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetContact(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (b *OGame) GetCelestial(v any) (Celestial, error) {\n\treturn b.WithPriority(taskRunner.Normal).GetCelestial(v)\n}", "func (client *Client) DescribeIpDdosThresholdWithChan(request *DescribeIpDdosThresholdRequest) (<-chan *DescribeIpDdosThresholdResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeIpDdosThresholdResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeIpDdosThreshold(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (_AccessControl *AccessControlFilterer) WatchCOOTransferred(opts *bind.WatchOpts, sink chan<- *AccessControlCOOTransferred) (event.Subscription, error) {\n\n\tlogs, sub, err := _AccessControl.contract.WatchLogs(opts, \"COOTransferred\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn event.NewSubscription(func(quit <-chan struct{}) error {\n\t\tdefer sub.Unsubscribe()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase log := <-logs:\n\t\t\t\t// New log arrived, parse the event and forward to the user\n\t\t\t\tevent := new(AccessControlCOOTransferred)\n\t\t\t\tif err := _AccessControl.contract.UnpackLog(event, \"COOTransferred\", log); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tevent.Raw = log\n\n\t\t\t\tselect {\n\t\t\t\tcase sink <- event:\n\t\t\t\tcase err := <-sub.Err():\n\t\t\t\t\treturn err\n\t\t\t\tcase <-quit:\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\tcase err := <-sub.Err():\n\t\t\t\treturn err\n\t\t\tcase <-quit:\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t}), nil\n}", "func (client *Client) GetNerCustomizedSeaEcomWithCallback(request *GetNerCustomizedSeaEcomRequest, callback func(response *GetNerCustomizedSeaEcomResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetNerCustomizedSeaEcomResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetNerCustomizedSeaEcom(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func ListenChannel(task ...Task) {\n\t// To trigger channel at first time\n\tSendToChannel(\"\", t4)\n\n\tfor {\n\t\tselect {\n\t\tcase msgChan := <-Ch:\n\t\t\tif len(task) > 0 {\n\t\t\t\tfor _, t := range task {\n\t\t\t\t\tif msg, ok := msgChan[t.Code]; ok {\n\t\t\t\t\t\tt.Job(msg)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tvar msg interface{}\n\t\t\t\tfor _, msgByte := range msgChan {\n\t\t\t\t\terr := json.Unmarshal(msgByte, &msg)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tfmt.Println(err)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func (client *Client) OemSitingSelctionWithCallback(request *OemSitingSelctionRequest, callback func(response *OemSitingSelctionResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *OemSitingSelctionResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.OemSitingSelction(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (l *Logs) GetChannel(uuid string) (string, error) {\n\tlogsChan := uuid\n\tdev, err := DevGetByUUID(l.ctx, uuid)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif dev.LogsChannel != \"\" {\n\t\tlogsChan = dev.LogsChannel\n\t}\n\treturn fmt.Sprintf(\"device-%s-logs\", logsChan), nil\n}", "func HandleControlChannel(conn protocol.Connection, s ndt.Server) {\n\tconnType := s.ConnectionType().String()\n\tmetrics.ActiveTests.WithLabelValues(connType).Inc()\n\tdefer metrics.ActiveTests.WithLabelValues(connType).Dec()\n\tdefer func(start time.Time) {\n\t\tndt5metrics.ControlChannelDuration.WithLabelValues(connType).Observe(\n\t\t\ttime.Since(start).Seconds())\n\t}(time.Now())\n\tdefer func() {\n\t\tcompleted := \"okay\"\n\t\tr := recover()\n\t\tif r != nil {\n\t\t\tlog.Println(\"Test failed, but we recovered:\", r)\n\t\t\t// All of our panic messages begin with an informative first word. Use that as a label.\n\t\t\terrType := panicMsgToErrType(fmt.Sprint(r))\n\t\t\tndt5metrics.ControlPanicCount.WithLabelValues(connType, errType).Inc()\n\t\t\tcompleted = \"panic\"\n\t\t}\n\t\tndt5metrics.ControlCount.WithLabelValues(connType, completed).Inc()\n\t}()\n\thandleControlChannel(conn, s)\n}", "func (client *Client) RunContactReviewWithChan(request *RunContactReviewRequest) (<-chan *RunContactReviewResponse, <-chan error) {\n\tresponseChan := make(chan *RunContactReviewResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.RunContactReview(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) GetKeywordChEcomWithCallback(request *GetKeywordChEcomRequest, callback func(response *GetKeywordChEcomResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetKeywordChEcomResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetKeywordChEcom(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) DescribeOssObjectDetailWithChan(request *DescribeOssObjectDetailRequest) (<-chan *DescribeOssObjectDetailResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeOssObjectDetailResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeOssObjectDetail(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (c *WSClient) readRoutine(ctx context.Context) {\n\tdefer func() {\n\t\tc.conn.Close()\n\t\tc.wg.Done()\n\t}()\n\n\tfor {\n\t\t// reset deadline for every message type (control or data)\n\t\tif c.readWait > 0 {\n\t\t\tif err := c.conn.SetReadDeadline(time.Now().Add(c.readWait)); err != nil {\n\t\t\t\tc.Logger.Error(\"failed to set read deadline\", \"err\", err)\n\t\t\t}\n\t\t}\n\t\t_, data, err := c.conn.ReadMessage()\n\t\tif err != nil {\n\t\t\tif !websocket.IsUnexpectedCloseError(err, websocket.CloseNormalClosure) {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tc.Logger.Error(\"failed to read response\", \"err\", err)\n\t\t\tclose(c.readRoutineQuit)\n\t\t\tc.reconnectAfter <- err\n\t\t\treturn\n\t\t}\n\n\t\tvar response rpctypes.RPCResponse\n\t\terr = json.Unmarshal(data, &response)\n\t\tif err != nil {\n\t\t\tc.Logger.Error(\"failed to parse response\", \"err\", err, \"data\", string(data))\n\t\t\tcontinue\n\t\t}\n\n\t\t// TODO: events resulting from /subscribe do not work with ->\n\t\t// because they are implemented as responses with the subscribe request's\n\t\t// ID. According to the spec, they should be notifications (requests\n\t\t// without IDs).\n\t\t// https://github.com/tendermint/tendermint/issues/2949\n\t\t//\n\t\t// Combine a non-blocking read on BaseService.Quit with a non-blocking write on ResponsesCh to avoid blocking\n\t\t// c.wg.Wait() in c.Stop(). Note we rely on Quit being closed so that it sends unlimited Quit signals to stop\n\t\t// both readRoutine and writeRoutine\n\n\t\tc.Logger.Info(\"got response\", \"id\", response.ID, \"result\", response.Result)\n\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn\n\t\tcase c.ResponsesCh <- response:\n\t\t}\n\t}\n}", "func (client *Client) CreateVSwitchWithChan(request *CreateVSwitchRequest) (<-chan *CreateVSwitchResponse, <-chan error) {\n\tresponseChan := make(chan *CreateVSwitchResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.CreateVSwitch(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func mainWeatherGetter() {\n\t// make channel for city list result from redis\n\tcityListChan := make(chan []string)\n\n\t// from goroutine, get all city list from redis\n\tgo database.GetAllCityList(cityListChan)\n\n\t// cities := []string{\"akiruno-shi\", \"paranaque\", \"omiya-shi\", \"machida-shi\", \"akishima-shi\"}\n\t// cities := <-cityListChan\n\tgetWeather(<-cityListChan...)\n}", "func IOHandler(ch chan string){\n\tfor{\n\t\tread := <- ch\n\t\tfmt.Println(read)\n\t}\n}", "func (client *Client) GetWsCustomizedChEcomContentWithCallback(request *GetWsCustomizedChEcomContentRequest, callback func(response *GetWsCustomizedChEcomContentResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetWsCustomizedChEcomContentResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetWsCustomizedChEcomContent(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func causalityWrap(inCh chan *job, syncer *Syncer) chan *job {\n\tcausality := &causality{\n\t\trelations: make(map[string]string),\n\t\ttask: syncer.cfg.Name,\n\t\tsource: syncer.cfg.SourceID,\n\t\tlogger: syncer.tctx.Logger.WithFields(zap.String(\"component\", \"causality\")),\n\t\tinCh: inCh,\n\t\toutCh: make(chan *job, syncer.cfg.QueueSize),\n\t}\n\n\tgo func() {\n\t\tcausality.run()\n\t\tcausality.close()\n\t}()\n\n\treturn causality.outCh\n}", "func (ra *relayAnnouncer) Run() error {\n\tra.conn = mqtt.NewClient(ra.options)\n\tif token := ra.conn.Connect(); token.Wait() {\n\t\terr := token.Error()\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"1 Cog connection error: %s\", err)\n\t\t\treturn errorStartAnnouncer\n\t\t}\n\t}\n\tif token := ra.conn.Subscribe(ra.receiptTopic, 1, ra.cogReceipt); token.Wait() {\n\t\terr := token.Error()\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"2 Cog connection error: %s\", err)\n\t\t\tra.conn.Disconnect(0)\n\t\t\treturn errorStartAnnouncer\n\t\t}\n\t}\n\tra.state = relayAnnouncerWaitingState\n\tgo func() {\n\t\tra.loop()\n\t}()\n\treturn nil\n}", "func reader(ch chan int) {\n \n t := time.NewTimer(10*time.Second)\n \n for {\n select {\n case i := <- ch: //if something comes from the the channel this case will print it otherwise carry on\n fmt.Printf(\"%d\\n\", i)\n \n case <-t.C :\n ch = nil //when this case heppens after 3 seconds the goroutine will stop and wait\n }\n }\n}", "func (client *Client) GetWsCustomizedChEcomContentWithChan(request *GetWsCustomizedChEcomContentRequest) (<-chan *GetWsCustomizedChEcomContentResponse, <-chan error) {\n\tresponseChan := make(chan *GetWsCustomizedChEcomContentResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetWsCustomizedChEcomContent(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (c *CreateKubernetesCluster) GetChan() chan v1.Message {\n\treturn c.result\n}", "func GetClinicDoctors(c *gin.Context) {\n\tlog.Infof(\"Get all doctors registered with specific physical clinic\")\n\taddressID := c.Param(\"addressId\")\n\tif addressID == \"\" {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusBadRequest,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: \"clinic address id not provided\",\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tctx := c.Request.Context()\n\tuserEmail, userID, gproject, err := getUserDetails(ctx, c.Request)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tctx, span := trace.StartSpan(ctx, \"Get all doctors registered for a clinic\")\n\tdefer span.End()\n\tclinicMetaDB := datastoredb.NewClinicMetaHandler()\n\terr = clinicMetaDB.InitializeDataBase(ctx, gproject)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tregisteredDoctors, err := clinicMetaDB.GetClinicDoctors(ctx, userEmail, userID, addressID)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, gin.H{\n\t\tconstants.RESPONSE_JSON_DATA: registeredDoctors,\n\t\tconstants.RESPONSDE_JSON_ERROR: nil,\n\t})\n\tclinicMetaDB.Close()\n}", "func (client *Client) DescribeVnKnowledgeWithChan(request *DescribeVnKnowledgeRequest) (<-chan *DescribeVnKnowledgeResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeVnKnowledgeResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeVnKnowledge(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func cancellation() {\n\tduration := 150 * time.Millisecond\n\tctx, cancel := context.WithTimeout(context.Background(), duration)\n\tdefer cancel()\n\n\tch := make(chan string, 1)\n\n\tgo func() {\n\t\ttime.Sleep(time.Duration(rand.Intn(200)) * time.Millisecond)\n\t\tch <- \"data\"\n\t}()\n\n\tselect {\n\tcase d := <-ch:\n\t\tfmt.Println(\"work complete\", d)\n\n\tcase <-ctx.Done():\n\t\tfmt.Println(\"work cancelled\")\n\t}\n\n\ttime.Sleep(time.Second)\n\tfmt.Println(\"-------------------------------------------------\")\n}", "func (client *Client) GetServiceInputMappingWithChan(request *GetServiceInputMappingRequest) (<-chan *GetServiceInputMappingResponse, <-chan error) {\n\tresponseChan := make(chan *GetServiceInputMappingResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetServiceInputMapping(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (tm *ServiceTracerouteManager) GetICMPInChan() chan gopacket.Packet {\n\treturn tm.ICMPChan\n}", "func (s *NodesInfoService) DoC(ctx context.Context) (*NodesInfoResponse, error) {\n\t// Check pre-conditions\n\tif err := s.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Get URL for request\n\tpath, params, err := s.buildURL()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Get HTTP response\n\tres, err := s.client.PerformRequestC(ctx, \"GET\", path, params, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Return operation response\n\tret := new(NodesInfoResponse)\n\tif err := s.client.decoder.Decode(res.Body, ret); err != nil {\n\t\treturn nil, err\n\t}\n\treturn ret, nil\n}", "func (client *Client) GetTaxationInfoWithChan(request *GetTaxationInfoRequest) (<-chan *GetTaxationInfoResponse, <-chan error) {\n\tresponseChan := make(chan *GetTaxationInfoResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetTaxationInfo(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (e *ElectionBug) ObserveBug(ctx context.Context) <-chan v3.GetResponse {\n\tretc := make(chan v3.GetResponse)\n\tgo e.observeBug(ctx, retc)\n\treturn retc\n}", "func (client *Client) GetContactWithCallback(request *GetContactRequest, callback func(response *GetContactResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetContactResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetContact(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func doGet(cmd string, conn net.Conn, kvs *keyValueServer){\n\t//fmt.Printf(\"Processing a get request %v\\n\", cmd)\n\tkvs.dataChan <- cmd[:len(cmd) - 1]\n}", "func (hdcni *HostDevCni) HandleCni(d *render.RenderData) error {\n\n\tswitch hdcni.VlanType {\n\tcase \"access\":\n\t\terr := errors.New(\"Host-device cni does not support VlanType=access\")\n\t\thdcni.Log.Error(err, \"Host-device VlanType error\")\n\t\treturn err\n\tcase \"selectivetrunk\":\n\t\terr := errors.New(\"Host-device cni does not support VlanType=selectivetrunk\")\n\t\thdcni.Log.Error(err, \"Host-device VlanType error\")\n\t\treturn err\n\tcase \"trunk\":\n\t\thdcni.Log.Info(\"Transparent Trunk case in Host-device cni\")\n\t}\n\treturn nil\n}", "func ConverserJ(cc Convo) Convo {\n\tjm := check.Eventcheck(4)\n\tvar options []string\n\tvar npD pD\n\tvar ncD cD\n\t//no function exists yet to get the player's name!\n\t//hellos first\n\tif jm == false { //check if they met\n\t\tncD = jihstart\n\t\tjb := models.StoryblobGetByName(4)\n\t\tfmt.Println(jb.Story)\n\t\tjb.Shown = true\n\t\tmodels.StoryblobUpdate(jb)\n\t} else {\n\t\tncD = jstart\n\t}\n\t// need loop\n\t//Display options - check for valid pD or remove and attach only to events?\n\t//Player chooses\n\t//Display response -check for valid response/depth\n\t//check to continue\n\n\tfor cc.stilltalking == true {\n\t\toptions = choicemakerV2(ncD)\n\t\tr1 := inputs.StringarrayInput(options)\n\t\tnpD = *ncD.branches[r1-1]\n\t\t//change depth\n\t\tif npD.dChange < 0 {\n\t\t\tcc.depth = depthChange(cc.depth, npD.dChange, 1) //need negative. so 1 min and 5 max\n\t\t} else {\n\t\t\tcc.depth = depthChange(cc.depth, npD.dChange, 5) //need negative. so 1 min and 5 max\n\t\t}\n\t\tncD = *npD.branch\n\t\t//check if valid\n\t\tif cc.depth < ncD.dCheck {\n\t\t\tncD = jDfail\n\t\t}\n\t\tv1 := \"\\\"\" + ncD.words + \"\\\"\"\n\t\tfmt.Println(v1)\n\t\t//checks branches for possible events\n\t\tncD, cc = dialogJoshEvents(ncD, cc)\n\t}\n\tcc.Character.Depth = cc.depth\n\tmodels.CharacterUpdate(cc.Character)\n\treturn cc\n}", "func (client *Client) GetArmsConsoleUrlWithChan(request *GetArmsConsoleUrlRequest) (<-chan *GetArmsConsoleUrlResponse, <-chan error) {\n\tresponseChan := make(chan *GetArmsConsoleUrlResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetArmsConsoleUrl(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func TestGetCabTripCtrl(t *testing.T) {\n\tmodel.MongoDBConnect()\n\tdefer model.MongoConnectionClose()\n\tConvey(\"GetCabTripCtrl\", func() {\n\t\treq, err := http.NewRequest(\"GET\", \"/api/cab/id3004672/date/2016-06-30?fresh=1\", nil)\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\n\t\tw := httptest.NewRecorder()\n\n\t\tConvey(\"Normal: 1 cab\", func() {\n\n\t\t\tcabapi.GetCabTripCtrl(w, req)\n\n\t\t\tSo(w.Code, ShouldEqual, http.StatusOK)\n\t\t\tresult, _ := ioutil.ReadAll(w.Result().Body)\n\t\t\texpct := \"id3004672 in 2016-06-30:3\"\n\t\t\tSo(result, ShouldEqual, expct)\n\t\t})\n\n\n\t})\n}", "func (client *Client) BeginVnDialogueWithChan(request *BeginVnDialogueRequest) (<-chan *BeginVnDialogueResponse, <-chan error) {\n\tresponseChan := make(chan *BeginVnDialogueResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.BeginVnDialogue(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (s GetMedicalTranscriptionJobOutput) GoString() string {\n\treturn s.String()\n}", "func (client *Client) QueryContactInfoWithChan(request *QueryContactInfoRequest) (<-chan *QueryContactInfoResponse, <-chan error) {\n\tresponseChan := make(chan *QueryContactInfoResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.QueryContactInfo(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (router *Router) getTerminal(w http.ResponseWriter, r *http.Request) {\n\tclusterName := r.URL.Query().Get(\"cluster\")\n\tnamespace := r.URL.Query().Get(\"namespace\")\n\tname := r.URL.Query().Get(\"name\")\n\tcontainer := r.URL.Query().Get(\"container\")\n\tshell := r.URL.Query().Get(\"shell\")\n\n\tlog.WithFields(logrus.Fields{\"cluster\": clusterName, \"namespace\": namespace, \"name\": name, \"container\": container, \"shell\": shell}).Tracef(\"getTerminal\")\n\n\tvar upgrader = websocket.Upgrader{}\n\n\tif router.config.WebSocket.AllowAllOrigins {\n\t\tupgrader.CheckOrigin = func(r *http.Request) bool { return true }\n\t}\n\n\tc, err := upgrader.Upgrade(w, r, nil)\n\tif err != nil {\n\t\tlog.WithError(err).Errorf(\"Could not upgrade connection\")\n\t\treturn\n\t}\n\tdefer c.Close()\n\n\tc.SetPongHandler(func(string) error { return nil })\n\n\tgo func() {\n\t\tticker := time.NewTicker(pingPeriod)\n\t\tdefer ticker.Stop()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ticker.C:\n\t\t\t\tif err := c.WriteMessage(websocket.PingMessage, nil); err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\tcluster := router.clusters.GetCluster(clusterName)\n\tif cluster == nil {\n\t\tlog.WithError(err).Errorf(\"Invalid cluster name\")\n\t\tmsg, _ := json.Marshal(terminal.Message{\n\t\t\tOp: \"stdout\",\n\t\t\tData: fmt.Sprintf(\"Invalid cluster name: %s\", err.Error()),\n\t\t})\n\t\tc.WriteMessage(websocket.TextMessage, msg)\n\t\treturn\n\t}\n\n\terr = cluster.GetTerminal(c, namespace, name, container, shell)\n\tif err != nil {\n\t\tlog.WithError(err).Errorf(\"Could not create terminal\")\n\t\tmsg, _ := json.Marshal(terminal.Message{\n\t\t\tOp: \"stdout\",\n\t\t\tData: fmt.Sprintf(\"Could not create terminal: %s\", err.Error()),\n\t\t})\n\t\tc.WriteMessage(websocket.TextMessage, msg)\n\t\treturn\n\t}\n\n\tlog.Tracef(\"Terminal connection was closed\")\n}", "func (client *Client) GetKeywordChEcomWithChan(request *GetKeywordChEcomRequest) (<-chan *GetKeywordChEcomResponse, <-chan error) {\n\tresponseChan := make(chan *GetKeywordChEcomResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetKeywordChEcom(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) GetTaobaoOrderWithChan(request *GetTaobaoOrderRequest) (<-chan *GetTaobaoOrderResponse, <-chan error) {\n\tresponseChan := make(chan *GetTaobaoOrderResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetTaobaoOrder(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) DescribeUserConnectionRecordsWithChan(request *DescribeUserConnectionRecordsRequest) (<-chan *DescribeUserConnectionRecordsResponse, <-chan error) {\n\tresponseChan := make(chan *DescribeUserConnectionRecordsResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.DescribeUserConnectionRecords(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) QueryVnConversationsWithCallback(request *QueryVnConversationsRequest, callback func(response *QueryVnConversationsResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *QueryVnConversationsResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.QueryVnConversations(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func main() {\n\tc:= make(chan int)\n\tfor i := 0; i < 10; i++ {\n\t\tgo simplePrint(i,c)\n\t}\n\n\tfor j := 0; j < 10; j++ {\n\tx:= <-c\n\tfmt.Println(x)\n\t}\n}", "func (cmd *GetTechHyTechCommand) Run(c *client.Client, args []string) error {\n\tvar path string\n\tif len(args) > 0 {\n\t\tpath = args[0]\n\t} else {\n\t\tpath = fmt.Sprintf(\"/api/tech/%v\", cmd.TechID)\n\t}\n\tlogger := goa.NewLogger(log.New(os.Stderr, \"\", log.LstdFlags))\n\tctx := goa.WithLogger(context.Background(), logger)\n\tresp, err := c.GetTechHyTech(ctx, path)\n\tif err != nil {\n\t\tgoa.LogError(ctx, \"failed\", \"err\", err)\n\t\treturn err\n\t}\n\n\tgoaclient.HandleResponse(c.Client, resp, cmd.PrettyPrint)\n\treturn nil\n}", "func cmapInt(concurrency int, fn func(int) string, in <-chan int) <-chan string {\n if concurrency <= 0 {\n panic(\"concurrency must be greater than zero\")\n }\n out := make(chan string)\n go func() {\n defer close(out)\n wg := sync.WaitGroup{}\n wg.Add(concurrency)\n for i := 0; i < concurrency; i++ {\n go func() {\n defer wg.Done()\n for item := range in {\n out <- fn(item)\n }\n }()\n }\n wg.Wait()\n }()\n return out\n}", "func (c ClientAsyncImpl) GetPodsAsync(channel chan<- Pods, workload *Workload) {\n\tvalue, err := c.syncClient.GetPods(workload)\n\tchannel <- Pods{value, err}\n}", "func ToChan(respCh chan BytesWatchResp, opts ...interface{}) func(dto BytesWatchResp) {\n\treturn func(dto BytesWatchResp) {\n\t\tselect {\n\t\tcase respCh <- dto:\n\t\t\t// success\n\t\tcase <-time.After(datasync.DefaultNotifTimeout):\n\t\t\tlogging.Warn(\"Unable to deliver notification\")\n\t\t}\n\t}\n}", "func (tm *ServiceTracerouteManager) GetOutChan() chan string {\n\treturn tm.OutChan\n}", "func (client *Client) VerifyCenWithCallback(request *VerifyCenRequest, callback func(response *VerifyCenResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *VerifyCenResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.VerifyCen(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) QueryVnConversationsWithChan(request *QueryVnConversationsRequest) (<-chan *QueryVnConversationsResponse, <-chan error) {\n\tresponseChan := make(chan *QueryVnConversationsResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.QueryVnConversations(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) GetManagedRuleWithChan(request *GetManagedRuleRequest) (<-chan *GetManagedRuleResponse, <-chan error) {\n\tresponseChan := make(chan *GetManagedRuleResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetManagedRule(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (r *ClearNet) Dial(network, address string) (net.Conn, error) {\n fmt.Println(\"dialing beeep boop\", network, address)\n\n\n\n // fmt.Println(\"writing\", b)\n\n done := make(chan int)\n\n callback := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n fmt.Println(\"finished dialing stuff\")\n // callback.Release() // free up memory from callback\n done <- args[0].Int()\n return nil\n })\n\n // func printMessage(this js.Value, args []js.Value) interface{} {\n // message := args[0].String()\n // fmt.Println(message)\n // \n // return nil\n // }\n defer callback.Release()\n \n rawHost, rawPort, _ := net.SplitHostPort(address)\n\n\n js.Global().Get(\"dialSocket\").Invoke(rawHost, rawPort, callback)\n\n // wait until we've got our response\n id := <-done\n\n // TODO: error if id < 0\n\n // return net.Dial(network, address)\n return &Conn{\n id: id,\n }, nil\n}", "func handleOutput(outHex, id string) {\n\t/* Make sure the output is valid */\n\to, err := hex.DecodeString(outHex)\n\tif nil != err {\n\t\tlog.Printf(\"[ID-%v] ERROR in %q: %v\", id, outHex, err)\n\t\treturn\n\t}\n\ts := string(o)\n\n\t/* Log the output */\n\tlog.Printf(\"[ID-%v] OUTPUT: %v (%q)\", id, outHex, s)\n\n\t/* Send it to interested C2 clients */\n\t/* Get connected clients, send the output to each one which is watching\n\tthis implant. */\n\tkeys := CLIENTS.Keys()\n\tfor _, key := range keys {\n\t\t/* Turn into a Client */\n\t\tc, ok := key.(*C2Client)\n\t\tif !ok {\n\t\t\tlog.Panicf(\"wrong type for Client: %T\", key)\n\t\t}\n\t\t/* If this is the right ID, or the it matches the regex, send\n\t\tthe message to the client */\n\t\tc.l.Lock()\n\t\tif c.id == id {\n\t\t\t/* Don't print \\r's on non-windows */\n\t\t\tif \"windows\" != runtime.GOOS {\n\t\t\t\ts = strings.Replace(s, \"\\r\", \"\", -1)\n\t\t\t}\n\t\t\tfmt.Fprintf(c.t, \"%s\", s)\n\t\t}\n\t\tc.l.Unlock()\n\t}\n\t/* TODO: Finish this */\n}", "func GetTrafficOpsCookie(cdnUri, user, pass string) (string, error) {\n\turi := cdnUri + `/api/1.2/user/login`\n\tpostdata := `{\"u\":\"` + user + `\", \"p\":\"` + pass + `\"}`\n\treq, err := http.NewRequest(\"POST\", uri, strings.NewReader(postdata))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treq.Header.Add(\"Accept\", \"application/json\")\n\n\tclient := getClient()\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer resp.Body.Close()\n\n\tfor _, cookie := range resp.Cookies() {\n\t\tif cookie.Name == `mojolicious` {\n\t\t\treturn cookie.Value, nil\n\t\t}\n\t}\n\n\tdata, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn \"\", errors.New(\"No login cookie received: \" + string(data))\n}", "func (c *Operation) generateVC(w http.ResponseWriter, r *http.Request) {\n\tvcsProfileCookie, err := r.Cookie(vcsProfileCookie)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to get vcsProfileCookie: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusBadRequest,\n\t\t\tfmt.Sprintf(\"failed to get cookie: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\terr = r.ParseForm()\n\tif err != nil {\n\t\tlogger.Errorf(err.Error())\n\t\tc.writeErrorResponse(w, http.StatusBadRequest, fmt.Sprintf(\"failed to parse request form: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\terr = c.validateForm(r.Form, \"cred\", \"holder\", \"authresp\", \"domain\", \"challenge\")\n\tif err != nil {\n\t\tlogger.Errorf(\"invalid generate credential request: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusBadRequest, fmt.Sprintf(\"invalid request argument: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tcred, err := c.createCredential(r.Form[\"cred\"][0], r.Form[\"authresp\"][0], r.Form[\"holder\"][0],\n\t\tr.Form[\"domain\"][0], r.Form[\"challenge\"][0], vcsProfileCookie.Value)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to create verifiable credential: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusInternalServerError,\n\t\t\tfmt.Sprintf(\"failed to create verifiable credential: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\terr = c.storeCredential(cred, vcsProfileCookie.Value)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to store credential: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusInternalServerError,\n\t\t\tfmt.Sprintf(\"failed to store credential: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"text/html; charset=utf-8\")\n\n\tt, err := template.ParseFiles(c.receiveVCHTML)\n\tif err != nil {\n\t\tlogger.Errorf(err.Error())\n\t\tc.writeErrorResponse(w, http.StatusInternalServerError,\n\t\t\tfmt.Sprintf(\"unable to load html: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tif err := t.Execute(w, vc{Data: string(cred)}); err != nil {\n\t\tlogger.Errorf(fmt.Sprintf(\"failed execute html template: %s\", err.Error()))\n\t}\n}", "func (c *InitRainbondCluster) GetChan() chan apiv1.Message {\n\treturn c.result\n}", "func main() {\n\tc1 := make(chan int, 1)\n\tc1 <- 2\n\tfor i := 0; i < 10; i++ {\n\t\tfmt.Println(\"hello word \")\n\t\ttime.Sleep(time.Duration(time.Second))\n\t}\n\tgo func() {\n\t\tfmt.Println(<-c1)\n\t}()\n}", "func (ec *executionContext) _Case(ctx context.Context, sel ast.SelectionSet, obj *models.Case) graphql.Marshaler {\n\tfields := graphql.CollectFields(ctx, sel, caseImplementors)\n\n\tvar wg sync.WaitGroup\n\tout := graphql.NewOrderedMap(len(fields))\n\tinvalid := false\n\tfor i, field := range fields {\n\t\tout.Keys[i] = field.Alias\n\n\t\tswitch field.Name {\n\t\tcase \"__typename\":\n\t\t\tout.Values[i] = graphql.MarshalString(\"Case\")\n\t\tcase \"Id\":\n\t\t\tout.Values[i] = ec._Case_Id(ctx, field, obj)\n\t\tcase \"Asset\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_Asset(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"CaseNumber\":\n\t\t\tout.Values[i] = ec._Case_CaseNumber(ctx, field, obj)\n\t\tcase \"Origin\":\n\t\t\tout.Values[i] = ec._Case_Origin(ctx, field, obj)\n\t\tcase \"Owner\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_Owner(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"Reason\":\n\t\t\tout.Values[i] = ec._Case_Reason(ctx, field, obj)\n\t\tcase \"IsClosed\":\n\t\t\tout.Values[i] = ec._Case_IsClosed(ctx, field, obj)\n\t\tcase \"Contact\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_Contact(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"CreatedBy\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_CreatedBy(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"ClosedDate\":\n\t\t\tout.Values[i] = ec._Case_ClosedDate(ctx, field, obj)\n\t\tcase \"CreatedDate\":\n\t\t\tout.Values[i] = ec._Case_CreatedDate(ctx, field, obj)\n\t\tcase \"IsDeleted\":\n\t\t\tout.Values[i] = ec._Case_IsDeleted(ctx, field, obj)\n\t\tcase \"Description\":\n\t\t\tout.Values[i] = ec._Case_Description(ctx, field, obj)\n\t\tcase \"IsEscalated\":\n\t\t\tout.Values[i] = ec._Case_IsEscalated(ctx, field, obj)\n\t\tcase \"LastModifiedBy\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_LastModifiedBy(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"LastModifiedDate\":\n\t\t\tout.Values[i] = ec._Case_LastModifiedDate(ctx, field, obj)\n\t\tcase \"LastReferencedDate\":\n\t\t\tout.Values[i] = ec._Case_LastReferencedDate(ctx, field, obj)\n\t\tcase \"LastViewedDate\":\n\t\t\tout.Values[i] = ec._Case_LastViewedDate(ctx, field, obj)\n\t\tdefault:\n\t\t\tpanic(\"unknown field \" + strconv.Quote(field.Name))\n\t\t}\n\t}\n\twg.Wait()\n\tif invalid {\n\t\treturn graphql.Null\n\t}\n\treturn out\n}", "func (c ClientAsyncImpl) GetEnvValueAsync(channel chan<- EnvValueReturn, podSpec corev1.PodSpec, namespace string, envName string) {\n\tvalue, err := c.syncClient.GetEnvValue(podSpec, namespace, envName)\n\tchannel <- EnvValueReturn{value, err}\n}", "func GetServiceDashboard(acc model.Account, channel chan string) {\n\tlog.Debug(\"[CLP] Starting to run CLP service...\")\n\n\tvar csrfToken string\n\tcookieJar, _ := cookiejar.New(nil)\n\tclient := &http.Client{\n\t\tJar: cookieJar,\n\t\tCheckRedirect: func(req *http.Request, via []*http.Request) error {\n\t\t\tfmt.Println(req.URL)\n\t\t\tfor _, r := range via {\n\t\t\t\tfmt.Println(r.URL)\n\t\t\t}\n\t\t\treturn http.ErrUseLastResponse\n\t\t},\n\t}\n\n\tlog.Debug(\"[CLP] Get login page for the CSRF token.\")\n\tcookiesResp, err := client.Get(\"https://services.clp.com.hk/zh/login/index.aspx\")\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\tdefer cookiesResp.Body.Close()\n\tfor _, cookie := range cookiesResp.Cookies() {\n\t\tif cookie.Name == \"K2Cie90hi___AntiXsrfToken\" {\n\t\t\tcsrfToken = cookie.Value\n\t\t}\n\t}\n\n\tlog.Debug(\"[CLP] Logging into...\")\n\tvar loginBody = \"username=\" + acc.Username + \"&password=\" + acc.Password\n\tloginReq, err := http.NewRequest(\"POST\", \"https://services.clp.com.hk/Service/ServiceLogin.ashx\", strings.NewReader(loginBody))\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\n\tloginReq.Header.Set(\"X-CSRFToken\", csrfToken)\n\tloginReq.Header.Set(\"X-Requested-With\", \"XMLHttpRequest\")\n\tloginReq.Header.Set(\"Content-Type\", \"application/x-www-form-urlencoded; charset=UTF-8\")\n\n\tloginResp, err := client.Do(loginReq)\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\tdefer loginResp.Body.Close()\n\n\tvar loginedCookies = loginResp.Cookies()\n\n\tlog.Debug(\"[CLP] Getting service dashboard info...\")\n\treq, err := http.NewRequest(\"POST\", \"https://services.clp.com.hk/Service/ServiceDashboard.ashx\", strings.NewReader(\"assCA=\"))\n\n\tcookieJar.SetCookies(req.URL, loginedCookies)\n\treq.Header.Set(\"X-CSRFToken\", csrfToken)\n\treq.Header.Set(\"X-Requested-With\", \"XMLHttpRequest\")\n\tresp, err := client.Do(req)\n\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\tdefer resp.Body.Close()\n\tdata, err := ioutil.ReadAll(resp.Body)\n\tchannel <- string(data[:])\n}" ]
[ "0.79232085", "0.64444685", "0.5182867", "0.50883555", "0.478109", "0.45607275", "0.45374334", "0.45302606", "0.4415868", "0.43623936", "0.4354781", "0.42620373", "0.41785672", "0.40921018", "0.40574202", "0.40431407", "0.40057978", "0.4002611", "0.40005198", "0.39568612", "0.39432833", "0.39097184", "0.38923502", "0.38871714", "0.3847867", "0.3842371", "0.38225192", "0.3818144", "0.3788263", "0.37862945", "0.37416318", "0.37292707", "0.3726744", "0.37231225", "0.37184343", "0.37181464", "0.37009746", "0.36857674", "0.3679589", "0.36782393", "0.3667143", "0.3657588", "0.36522764", "0.36449686", "0.36402154", "0.36258408", "0.36098397", "0.36087802", "0.36051196", "0.3598645", "0.3598122", "0.35816115", "0.35811654", "0.35784012", "0.35755008", "0.35734972", "0.35734138", "0.3573307", "0.3569662", "0.35601383", "0.3545363", "0.35392278", "0.3534635", "0.35209927", "0.3512832", "0.3498981", "0.34912956", "0.34864792", "0.3484877", "0.3479327", "0.34745413", "0.34730497", "0.34703732", "0.34580123", "0.34563276", "0.34546113", "0.3453586", "0.3453473", "0.34528556", "0.34356275", "0.3432319", "0.34284306", "0.34243476", "0.34160668", "0.3416004", "0.34124935", "0.34115952", "0.340713", "0.3406253", "0.340616", "0.33929437", "0.3391463", "0.33895296", "0.33834273", "0.33812687", "0.33771044", "0.33762747", "0.337198", "0.33679208", "0.33632666" ]
0.8225025
0
GetOcJusticeTerminalCaseWithCallback invokes the dt_oc_info.GetOcJusticeTerminalCase API asynchronously
func (client *Client) GetOcJusticeTerminalCaseWithCallback(request *GetOcJusticeTerminalCaseRequest, callback func(response *GetOcJusticeTerminalCaseResponse, err error)) <-chan int { result := make(chan int, 1) err := client.AddAsyncTask(func() { var response *GetOcJusticeTerminalCaseResponse var err error defer close(result) response, err = client.GetOcJusticeTerminalCase(request) callback(response, err) result <- 1 }) if err != nil { defer close(result) callback(nil, err) result <- 0 } return result }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (client *Client) GetOcJusticeTerminalCase(request *GetOcJusticeTerminalCaseRequest) (response *GetOcJusticeTerminalCaseResponse, err error) {\n\tresponse = CreateGetOcJusticeTerminalCaseResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func (client *Client) GetOcJusticeTerminalCaseWithChan(request *GetOcJusticeTerminalCaseRequest) (<-chan *GetOcJusticeTerminalCaseResponse, <-chan error) {\n\tresponseChan := make(chan *GetOcJusticeTerminalCaseResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetOcJusticeTerminalCase(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func CreateGetOcJusticeTerminalCaseResponse() (response *GetOcJusticeTerminalCaseResponse) {\n\tresponse = &GetOcJusticeTerminalCaseResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetOcJusticeTerminalCaseRequest() (request *GetOcJusticeTerminalCaseRequest) {\n\trequest = &GetOcJusticeTerminalCaseRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"dt-oc-info\", \"2022-08-29\", \"GetOcJusticeTerminalCase\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (client *Client) DescribeCustinsKernelReleaseNotesWithCallback(request *DescribeCustinsKernelReleaseNotesRequest, callback func(response *DescribeCustinsKernelReleaseNotesResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeCustinsKernelReleaseNotesResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeCustinsKernelReleaseNotes(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) ListCasesWithCallback(request *ListCasesRequest, callback func(response *ListCasesResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *ListCasesResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.ListCases(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetOpenNLUWithCallback(request *GetOpenNLURequest, callback func(response *GetOpenNLUResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetOpenNLUResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetOpenNLU(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetIndustryCommerceInfoWithCallback(request *GetIndustryCommerceInfoRequest, callback func(response *GetIndustryCommerceInfoResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetIndustryCommerceInfoResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetIndustryCommerceInfo(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetWsCustomizedChO2OWithCallback(request *GetWsCustomizedChO2ORequest, callback func(response *GetWsCustomizedChO2OResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetWsCustomizedChO2OResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetWsCustomizedChO2O(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func Callback(w http.ResponseWriter, r *http.Request) {\n\tcode := ParseResponse(w, r)\n\taccess := AccessToken(code, w, r)\n\tfmt.Fprintf(w, access.Token)\n\tGetData(access.Token, w, r)\n}", "func Callback(c *gin.Context) {\n\tprovider := c.Param(\"provider\")\n\n\tvar logincode vo.LoginReq\n\tif err := c.ShouldBindQuery(&logincode); err != nil {\n\t\tfmt.Println(\"xxxx\", err)\n\t}\n\n\tfmt.Println(\"provider\", provider, logincode)\n\n\tuserInfo := vo.GetUserInfoFromOauth(provider, logincode.Code, logincode.State)\n\tfmt.Println(\"get user info\", userInfo)\n\n\tif userInfo == nil {\n\t\tc.JSON(http.StatusOK, sailor.HTTPAirdbResponse{\n\t\t\tCode: enum.AirdbSuccess,\n\t\t\tSuccess: true,\n\t\t\tData: vo.LoginResp{\n\t\t\t\tNickname: \"xxx\",\n\t\t\t\tHeadimgurl: \"xxx.png\",\n\t\t\t},\n\t\t})\n\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, sailor.HTTPAirdbResponse{\n\t\tCode: enum.AirdbSuccess,\n\t\tSuccess: true,\n\t\tData: vo.LoginResp{\n\t\t\tNickname: userInfo.Login,\n\t\t\tHeadimgurl: userInfo.AvatarURL,\n\t\t},\n\t})\n}", "func (client *Client) DescribeUserVvTopByDayWithCallback(request *DescribeUserVvTopByDayRequest, callback func(response *DescribeUserVvTopByDayResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeUserVvTopByDayResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeUserVvTopByDay(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (h *Handler) oidcCallback(w http.ResponseWriter, r *http.Request, p httprouter.Params) (interface{}, error) {\n\tresult, err := h.GetConfig().Auth.ValidateOIDCAuthCallback(r.URL.Query())\n\tif err != nil {\n\t\th.Warnf(\"Error validating callback: %v.\", err)\n\t\thttp.Redirect(w, r, \"/web/msg/error/login_failed\", http.StatusFound)\n\t\treturn nil, nil\n\t}\n\th.Infof(\"Callback: %v %v %v.\", result.Username, result.Identity, result.Req.Type)\n\treturn nil, h.CallbackHandler(w, r, webapi.CallbackParams{\n\t\tUsername: result.Username,\n\t\tIdentity: result.Identity,\n\t\tSession: result.Session,\n\t\tCert: result.Cert,\n\t\tTLSCert: result.TLSCert,\n\t\tHostSigners: result.HostSigners,\n\t\tType: result.Req.Type,\n\t\tCreateWebSession: result.Req.CreateWebSession,\n\t\tCSRFToken: result.Req.CSRFToken,\n\t\tPublicKey: result.Req.PublicKey,\n\t\tClientRedirectURL: result.Req.ClientRedirectURL,\n\t})\n}", "func (client *Client) ListCityMapAoisWithCallback(request *ListCityMapAoisRequest, callback func(response *ListCityMapAoisResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *ListCityMapAoisResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.ListCityMapAois(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (d *Dao) CaseObtainMID(c context.Context, mid int64, isToday bool) (cases map[int64]*model.SimCase, err error) {\n\tconn := d.redis.Get(c)\n\tdefer conn.Close()\n\tvar _setKey string\n\tif isToday {\n\t\t_setKey = caseVoteCIDMIDKey(mid)\n\t} else {\n\t\t_setKey = caseObtainMIDKey(mid)\n\t}\n\tvar ms []string\n\tif ms, err = redis.Strings(conn.Do(\"SMEMBERS\", _setKey)); err != nil {\n\t\tif err != redis.ErrNil {\n\t\t\treturn\n\t\t}\n\t\terr = nil\n\t}\n\tcases = make(map[int64]*model.SimCase)\n\tfor _, s := range ms {\n\t\tif s == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tsc := &model.SimCase{}\n\t\tif err = json.Unmarshal([]byte(s), sc); err != nil {\n\t\t\terr = errors.WithStack(err)\n\t\t\treturn\n\t\t}\n\t\tcases[sc.ID] = sc\n\t}\n\treturn\n}", "func (client *Client) DescribeLiveDomainCertificateInfoWithCallback(request *DescribeLiveDomainCertificateInfoRequest, callback func(response *DescribeLiveDomainCertificateInfoResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeLiveDomainCertificateInfoResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeLiveDomainCertificateInfo(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) VerifyCenWithCallback(request *VerifyCenRequest, callback func(response *VerifyCenResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *VerifyCenResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.VerifyCen(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) DescribeIpDdosThresholdWithCallback(request *DescribeIpDdosThresholdRequest, callback func(response *DescribeIpDdosThresholdResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeIpDdosThresholdResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeIpDdosThreshold(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) SegmentSkyWithCallback(request *SegmentSkyRequest, callback func(response *SegmentSkyResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *SegmentSkyResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.SegmentSky(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func cognitoCallback(ctx context.Context, d *aegis.HandlerDependencies, req *aegis.APIGatewayProxyRequest, res *aegis.APIGatewayProxyResponse, params url.Values) error {\n\t// Exchange code for token\n\ttokens, err := d.Services.Cognito.GetTokens(req.QueryStringParameters[\"code\"], []string{})\n\t// Depending on Cognito configuration, there could be an error here.\n\t// This service is for an OAuth2 with an authorization code flow.\n\t// NOTE: tokens.AccessToken is generally used.\n\t// If using an openid grant, you may also use tokens.IDToken with ParseAndVerifyJWT() below.\n\tif tokens.Error != \"\" {\n\t\terr = errors.New(tokens.Error)\n\t}\n\tif err != nil {\n\t\tlog.Println(\"Couldn't get access token\", err)\n\t\tres.JSONError(500, err)\n\t} else {\n\t\t// verify the token\n\t\t_, err := d.Services.Cognito.ParseAndVerifyJWT(tokens.AccessToken)\n\t\tif err == nil {\n\t\t\thost := req.GetHeader(\"Host\")\n\t\t\tstage := req.RequestContext.Stage\n\t\t\tres.SetHeader(\"Set-Cookie\", \"access_token=\"+tokens.AccessToken+\"; Domain=\"+host+\"; Secure; HttpOnly\")\n\t\t\tres.Redirect(301, \"https://\"+host+\"/\"+stage+\"/protected\")\n\t\t} else {\n\t\t\tres.JSONError(401, errors.New(\"unauthorized, invalid token\"))\n\t\t}\n\t}\n\treturn nil\n}", "func (client *Client) GetContactWithCallback(request *GetContactRequest, callback func(response *GetContactResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetContactResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetContact(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetNerCustomizedSeaEcomWithCallback(request *GetNerCustomizedSeaEcomRequest, callback func(response *GetNerCustomizedSeaEcomResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetNerCustomizedSeaEcomResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetNerCustomizedSeaEcom(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) DescribeVnKnowledgeWithCallback(request *DescribeVnKnowledgeRequest, callback func(response *DescribeVnKnowledgeResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeVnKnowledgeResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeVnKnowledge(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (b *OGame) GetCelestial(v any) (Celestial, error) {\n\treturn b.WithPriority(taskRunner.Normal).GetCelestial(v)\n}", "func (c *Operation) callback(w http.ResponseWriter, r *http.Request) { //nolint: funlen,gocyclo\n\tif len(r.URL.Query()[\"error\"]) != 0 {\n\t\tif r.URL.Query()[\"error\"][0] == \"access_denied\" {\n\t\t\thttp.Redirect(w, r, c.homePage, http.StatusTemporaryRedirect)\n\t\t}\n\t}\n\n\ttk, err := c.tokenIssuer.Exchange(r)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to exchange code for token: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusBadRequest,\n\t\t\tfmt.Sprintf(\"failed to exchange code for token: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\t// user info from token will be used for to retrieve data from cms\n\tinfo, err := c.tokenResolver.Resolve(tk.AccessToken)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to get token info: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusBadRequest,\n\t\t\tfmt.Sprintf(\"failed to get token info: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tuserID, subject, err := c.getCMSData(tk, \"email=\"+info.Subject, info.Scope)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to get cms data: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusBadRequest,\n\t\t\tfmt.Sprintf(\"failed to get cms data: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tcallbackURLCookie, err := r.Cookie(callbackURLCookie)\n\tif err != nil && !errors.Is(err, http.ErrNoCookie) {\n\t\tc.writeErrorResponse(w, http.StatusBadRequest,\n\t\t\tfmt.Sprintf(\"failed to get authMode cookie: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tif callbackURLCookie != nil && callbackURLCookie.Value != \"\" {\n\t\ttxnID := uuid.NewString()\n\t\tdata := txnData{\n\t\t\tUserID: userID,\n\t\t\tScope: info.Scope,\n\t\t\tToken: tk.AccessToken,\n\t\t}\n\n\t\tdataBytes, mErr := json.Marshal(data)\n\t\tif mErr != nil {\n\t\t\tc.writeErrorResponse(w, http.StatusInternalServerError,\n\t\t\t\tfmt.Sprintf(\"failed to marshal txn data: %s\", mErr.Error()))\n\t\t\treturn\n\t\t}\n\n\t\terr = c.store.Put(txnID, dataBytes)\n\t\tif err != nil {\n\t\t\tc.writeErrorResponse(w, http.StatusInternalServerError,\n\t\t\t\tfmt.Sprintf(\"failed to save txn data: %s\", err.Error()))\n\n\t\t\treturn\n\t\t}\n\n\t\thttp.Redirect(w, r, callbackURLCookie.Value+\"?txnID=\"+txnID, http.StatusTemporaryRedirect)\n\n\t\treturn\n\t}\n\n\tvcsProfileCookie, err := r.Cookie(vcsProfileCookie)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to get cookie: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusBadRequest,\n\t\t\tfmt.Sprintf(\"failed to get cookie: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tcred, err := c.prepareCredential(subject, info.Scope, vcsProfileCookie.Value)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to create credential: %s\", err.Error())\n\t\tc.writeErrorResponse(w, http.StatusInternalServerError,\n\t\t\tfmt.Sprintf(\"failed to create credential: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"text/html; charset=utf-8\")\n\n\tt, err := template.ParseFiles(c.didAuthHTML)\n\tif err != nil {\n\t\tlogger.Errorf(err.Error())\n\t\tc.writeErrorResponse(w, http.StatusInternalServerError,\n\t\t\tfmt.Sprintf(\"unable to load html: %s\", err.Error()))\n\n\t\treturn\n\t}\n\n\tif err := t.Execute(w, map[string]interface{}{\n\t\t\"Path\": generate + \"?\" + \"profile=\" + vcsProfileCookie.Value,\n\t\t\"Cred\": string(cred),\n\t}); err != nil {\n\t\tlogger.Errorf(fmt.Sprintf(\"failed execute qr html template: %s\", err.Error()))\n\t}\n}", "func (client *Client) DescribeOssObjectDetailWithCallback(request *DescribeOssObjectDetailRequest, callback func(response *DescribeOssObjectDetailResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeOssObjectDetailResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeOssObjectDetail(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func callback(\n\tservice models.DeviceService,\n\tid string,\n\taction string,\n\tactionType models.ActionType,\n\tlc logger.LoggingClient) error {\n\n\tclient := &http.Client{}\n\turl := service.Addressable.GetCallbackURL()\n\tif len(url) > 0 {\n\t\tbody, err := getBody(id, actionType)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treq, err := http.NewRequest(string(action), url, bytes.NewReader(body))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treq.Header.Add(clients.ContentType, clients.ContentTypeJSON)\n\n\t\tgo makeRequest(client, req, lc)\n\t} else {\n\t\tlc.Info(\"callback::no addressable for \" + service.Name)\n\t}\n\treturn nil\n}", "func (client *Client) OemSitingSelctionWithCallback(request *OemSitingSelctionRequest, callback func(response *OemSitingSelctionResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *OemSitingSelctionResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.OemSitingSelction(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) CreateVSwitchWithCallback(request *CreateVSwitchRequest, callback func(response *CreateVSwitchResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *CreateVSwitchResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.CreateVSwitch(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (ovscni *OvsCni) HandleCni(d *render.RenderData) error {\n\n\t//For VlanType=trunk we do not need to do anything\n\tswitch ovscni.VlanType {\n\tcase \"access\":\n\t\tif len(ovscni.L2srvResources) != 1 {\n\t\t\terr := errors.New(\"Cannot use more than one L2Services for VlanType=access case\")\n\t\t\tovscni.Log.Error(err, \"L2Services cannot contain more than one L2Services in VlanType=access case\")\n\t\t\treturn err\n\t\t}\n\t\td.Data[\"AccessVlan\"] = ovscni.L2srvResources[0].Spec.SegmentationID\n\tcase \"selectivetrunk\":\n\t\ttmpList := []string{}\n\t\tfor _, l2srvObj := range ovscni.L2srvResources {\n\t\t\ttmpStr := \"{\\\"id\\\": \" + strconv.Itoa(int(l2srvObj.Spec.SegmentationID)) + \"}\"\n\t\t\ttmpList = append(tmpList, tmpStr)\n\t\t}\n\t\td.Data[\"SelectiveVlan\"] = \"[\" + strings.Join(tmpList, \",\") + \"]\"\n\tcase \"trunk\":\n\t\tovscni.Log.Info(\"Transparent Trunk case in cluster level\")\n\t}\n\treturn nil\n}", "func handleOidcCallback(\n\tw http.ResponseWriter,\n\tr *http.Request,\n\tclientName string,\n\tclientID string,\n\tclientSecret string,\n\tredirectURI string,\n\twellKnownConfig oidc.WellKnownConfiguration,\n\tstate string,\n\tcodeVerifier string,\n\tcancel context.CancelFunc,\n) {\n\tvar authorisationResponse, err = oidc.ValidateAuthorisationResponse(r.URL, state)\n\tif err != nil {\n\t\trenderAndLogError(w, cancel, fmt.Sprintf(\"%v\", err))\n\t\treturn\n\t}\n\n\tviewModel, err := VerifyCode(clientID, clientSecret, redirectURI, wellKnownConfig, codeVerifier, authorisationResponse.Code)\n\tif err != nil {\n\t\trenderAndLogError(w, cancel, fmt.Sprintf(\"%v\", err))\n\t\treturn\n\t}\n\n\t// show webpage\n\tt := template.New(\"credentials\")\n\t_, parseErr := t.Parse(TokenResultView())\n\tif parseErr != nil {\n\t\trenderAndLogError(w, cancel, fmt.Sprintf(\"%v\", parseErr))\n\t\treturn\n\t}\n\ttplErr := t.Execute(w, viewModel)\n\tif tplErr != nil {\n\t\trenderAndLogError(w, cancel, fmt.Sprintf(\"%v\", tplErr))\n\t\treturn\n\t}\n\n\tcancel()\n}", "func (client *Client) RunContactReviewWithCallback(request *RunContactReviewRequest, callback func(response *RunContactReviewResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *RunContactReviewResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.RunContactReview(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func CORSVC(ac *atmi.ATMICtx, svc *atmi.TPSVCINFO) {\n\tret := SUCCEED\n\n\t//Return to the caller\n\tdefer func() {\n\n\t\tac.TpLogCloseReqFile()\n\t\tif SUCCEED == ret {\n\t\t\tac.TpReturn(atmi.TPSUCCESS, 0, &svc.Data, 0)\n\t\t} else {\n\t\t\tac.TpReturn(atmi.TPFAIL, 0, &svc.Data, 0)\n\t\t}\n\t}()\n\n\t//Get UBF Handler\n\tub, _ := ac.CastToUBF(&svc.Data)\n\n\t//Print the buffer to stdout\n\t//fmt.Println(\"Incoming request:\")\n\tub.TpLogPrintUBF(atmi.LOG_DEBUG, \"CORSVC: Incoming request:\")\n\n\tarr, err := ub.BGetByteArr(u.EX_NETDATA, 0)\n\n\tif err != nil {\n\t\tac.TpLogError(\"Failed to get EX_NETDATA: %s\", err.Message())\n\t\tret = FAIL\n\t\treturn\n\t}\n\tif arr[0] == 1 && arr[1] == 1 {\n\t\tac.TpLogInfo(\"Test case 11 - no need for correlation\")\n\t} else if len(arr) > 4 {\n\n\t\tcorr := string(arr[:4])\n\n\t\tac.TpLogInfo(\"Extracted correlator: [%s]\", corr)\n\n\t\tif err := ub.BChg(u.EX_NETCORR, 0, corr); nil != err {\n\t\t\tac.TpLogError(\"Failed to set EX_NETCORR: %s\", err.Message())\n\t\t\tret = FAIL\n\t\t\treturn\n\t\t}\n\n\t}\n\n\tub.TpLogPrintUBF(atmi.LOG_DEBUG, \"Reply buffer afrer correl\")\n\n}", "func (client *Client) DescribeUserConnectionRecordsWithCallback(request *DescribeUserConnectionRecordsRequest, callback func(response *DescribeUserConnectionRecordsResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeUserConnectionRecordsResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeUserConnectionRecords(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetWsCustomizedChEcomContentWithCallback(request *GetWsCustomizedChEcomContentRequest, callback func(response *GetWsCustomizedChEcomContentResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetWsCustomizedChEcomContentResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetWsCustomizedChEcomContent(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (m *PatientMutation) Congenital() (r string, exists bool) {\n\tv := m._Congenital\n\tif v == nil {\n\t\treturn\n\t}\n\treturn *v, true\n}", "func (c *controller) Callback(ctx context.Context, request *web.Request) web.Result {\n\tif resp := c.service.callback(ctx, request); resp != nil {\n\t\treturn resp\n\t}\n\treturn c.responder.NotFound(errors.New(\"broker for callback not found\"))\n}", "func callback() {\n\tlog.Println(\"shutdown requested\")\n}", "func (client *Client) GetTaxationInfoWithCallback(request *GetTaxationInfoRequest, callback func(response *GetTaxationInfoResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetTaxationInfoResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetTaxationInfo(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func CallbackHandler(c echo.Context) error {\n\tprovider, err := gomniauth.Provider(c.Param(\"provider\"))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tomap, err := objx.FromURLQuery(c.QueryString())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcreds, err := provider.CompleteAuth(omap)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tuser, err := provider.GetUser(creds)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tauthCookieValue := objx.New(map[string]interface{}{\n\t\t\"name\": user.Name(),\n\t\t\"email\": user.Email(),\n\t\t\"avatarURL\": user.AvatarURL(),\n\t}).MustBase64()\n\n\tcookie := &http.Cookie{\n\t\tName: \"auth\",\n\t\tValue: authCookieValue,\n\t\tPath: \"/\",\n\t\tExpires: time.Now().Add(24 * time.Hour),\n\t}\n\tc.SetCookie(cookie)\n\n\t// return c.String(http.StatusOK, \"Login Success!\")\n\treturn c.Redirect(http.StatusTemporaryRedirect, \"/\")\n}", "func (h *Handler) RetrieveCertificate(companyChainID string, uuid string) (*entityApi.TransactionWrapper, error) {\n apiResponse, err := h.apiClient.Get(fmt.Sprintf(certificateRoute, companyChainID, uuid), nil)\n if err != nil {\n return nil, err\n }\n var transactionWrapper entityApi.TransactionWrapper\n if err := unmarshalApiResponse(apiResponse, &transactionWrapper); err != nil {\n return nil, err\n }\n return &transactionWrapper, nil\n}", "func (client *Client) GetManagedRuleWithCallback(request *GetManagedRuleRequest, callback func(response *GetManagedRuleResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetManagedRuleResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetManagedRule(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) QueryVnConversationsWithCallback(request *QueryVnConversationsRequest, callback func(response *QueryVnConversationsResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *QueryVnConversationsResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.QueryVnConversations(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (term *Terminology) ReadV2toSNOMEDCT(ctx context.Context, id *apiv1.Identifier, f func(*apiv1.Identifier) error) error {\n\tctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)\n\tdefer cancel()\n\tresponse, err := term.client.FromCrossMap(ctx, &snomed.TranslateFromRequest{S: id.GetValue(), RefsetId: 900000000000497000})\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(response.GetTranslations()) == 0 {\n\t\tlog.Printf(\"no translations found for map from '%s:%s' to '%s'\", id.GetSystem(), id.GetValue(), identifiers.SNOMEDCT)\n\t}\n\tfor _, t := range response.GetTranslations() {\n\t\tref := t.GetReferenceSetItem().GetReferencedComponentId()\n\t\tif err := f(&apiv1.Identifier{System: identifiers.SNOMEDCT, Value: strconv.FormatInt(ref, 10)}); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (client *Client) GetKeywordChEcomWithCallback(request *GetKeywordChEcomRequest, callback func(response *GetKeywordChEcomResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetKeywordChEcomResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetKeywordChEcom(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func GetClinicDoctors(c *gin.Context) {\n\tlog.Infof(\"Get all doctors registered with specific physical clinic\")\n\taddressID := c.Param(\"addressId\")\n\tif addressID == \"\" {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusBadRequest,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: \"clinic address id not provided\",\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tctx := c.Request.Context()\n\tuserEmail, userID, gproject, err := getUserDetails(ctx, c.Request)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tctx, span := trace.StartSpan(ctx, \"Get all doctors registered for a clinic\")\n\tdefer span.End()\n\tclinicMetaDB := datastoredb.NewClinicMetaHandler()\n\terr = clinicMetaDB.InitializeDataBase(ctx, gproject)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tregisteredDoctors, err := clinicMetaDB.GetClinicDoctors(ctx, userEmail, userID, addressID)\n\tif err != nil {\n\t\tc.AbortWithStatusJSON(\n\t\t\thttp.StatusInternalServerError,\n\t\t\tgin.H{\n\t\t\t\tconstants.RESPONSE_JSON_DATA: nil,\n\t\t\t\tconstants.RESPONSDE_JSON_ERROR: err.Error(),\n\t\t\t},\n\t\t)\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, gin.H{\n\t\tconstants.RESPONSE_JSON_DATA: registeredDoctors,\n\t\tconstants.RESPONSDE_JSON_ERROR: nil,\n\t})\n\tclinicMetaDB.Close()\n}", "func GetByProvince(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tcovidData, err := scrapper.GetAllDataByProvince()\n\tif err != nil {\n\t\tjson.NewEncoder(w).Encode(Error{\n\t\t\tCode: 500,\n\t\t\tMessage: err.Error(),\n\t\t})\n\n\t\treturn\n\t}\n\n\tprov := r.URL.Query().Get(\"prov\")\n\tif prov != \"\" {\n\t\tfor _, item := range covidData {\n\t\t\tif strings.ToLower(item.Province) == strings.ToLower(prov) {\n\t\t\t\tjson.NewEncoder(w).Encode(item)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\tjson.NewEncoder(w).Encode(covidData)\n\t\treturn\n\t}\n\n\tjson.NewEncoder(w).Encode(covidData)\n}", "func NewCase() *SCaseFunction {\n\treturn &SCaseFunction{}\n}", "func (o GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKeyPtrOutput) CaseInsensitive() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v *GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKey) *bool {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.CaseInsensitive\n\t}).(pulumi.BoolPtrOutput)\n}", "func (d *Dao) Case(c context.Context, arg *blocked.ArgCaseSearch) (ids []int64, pager *blocked.Pager, err error) {\n\treq := d.elastic.NewRequest(blocked.BusinessBlockedCase).Index(blocked.TableBlockedCase).Fields(\"id\")\n\tif arg.Keyword != blocked.SearchDefaultString {\n\t\treq.WhereLike([]string{\"origin_content\"}, []string{arg.Keyword}, true, elastic.LikeLevelHigh)\n\t}\n\tif arg.OriginType != blocked.SearchDefaultNum {\n\t\treq.WhereEq(\"origin_type\", arg.OriginType)\n\t}\n\tif arg.Status != blocked.SearchDefaultNum {\n\t\treq.WhereEq(\"status\", arg.Status)\n\t}\n\tif arg.CaseType != blocked.SearchDefaultNum {\n\t\treq.WhereEq(\"case_type\", arg.CaseType)\n\t}\n\tif arg.UID != blocked.SearchDefaultNum {\n\t\treq.WhereEq(\"mid\", arg.UID)\n\t}\n\tif arg.OPID != blocked.SearchDefaultNum {\n\t\treq.WhereEq(\"oper_id\", arg.OPID)\n\t}\n\treq.WhereRange(\"start_time\", arg.TimeFrom, arg.TimeTo, elastic.RangeScopeLcRc)\n\treq.Pn(arg.PN).Ps(arg.PS).Order(arg.Order, arg.Sort)\n\tvar res *search.ReSearchData\n\tif err = req.Scan(c, &res); err != nil {\n\t\terr = errors.Errorf(\"elastic search(%s) error(%v)\", req.Params(), err)\n\t\treturn\n\t}\n\tids, pager = pagerExtra(res)\n\treturn\n}", "func (d *Deals) Retrieve(ctx context.Context, waddr string, cid cid.Cid) (io.Reader, error) {\n\ttime.Sleep(time.Second * 3)\n\treturn strings.NewReader(\"hello there\"), nil\n}", "func (client *Client) ListOfficeConversionTaskWithCallback(request *ListOfficeConversionTaskRequest, callback func(response *ListOfficeConversionTaskResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *ListOfficeConversionTaskResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.ListOfficeConversionTask(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func fnCase(ctx Context, doc *JDoc, params []string) interface{} {\n\tstats := ctx.Value(EelTotalStats).(*ServiceStats)\n\tctx.Log().Error(\"error_type\", \"func_case\", \"op\", \"case\", \"cause\", \"now_implemented_in_parser\", \"params\", params)\n\tstats.IncErrors()\n\tAddError(ctx, SyntaxError{fmt.Sprintf(\"case function now implemented in parser\"), \"case\", params})\n\treturn nil\n}", "func MythicRPCCallbackDisplayToRealIdSearch(input MythicRPCCallbackDisplayToRealIdSearchMessage) MythicRPCCallbackDisplayToRealIdSearchMessageResponse {\n\tresponse := MythicRPCCallbackDisplayToRealIdSearchMessageResponse{\n\t\tSuccess: false,\n\t}\n\tsearchString := \"\"\n\tif input.OperationName != nil {\n\t\tsearchString = `SELECT \n \t\tcallback.id \n\t\t\tFROM \n\t\t\tcallback\n\t\t\tJOIN operation on callback.operation_id = operation.id\n\t\t\tWHERE callback.display_id=$1 AND operation.name=$2`\n\t\tcallback := databaseStructs.Callback{}\n\t\tif err := database.DB.Get(&callback, searchString, input.CallbackDisplayID, *input.OperationName); err != nil {\n\t\t\tlogging.LogError(err, \"Failed to find task based on task id and operation name\")\n\t\t\tresponse.Error = err.Error()\n\t\t\treturn response\n\t\t} else {\n\t\t\tresponse.CallbackID = callback.ID\n\t\t\tresponse.Success = true\n\t\t\treturn response\n\t\t}\n\t} else if input.OperationID != nil {\n\t\tsearchString = `SELECT \n \t\tcallback.id \n\t\t\tFROM \n\t\t\tcallback\n\t\t\tWHERE callback.display_id=$1 AND callback.operation_id=$2`\n\t\tcallback := databaseStructs.Callback{}\n\t\tif err := database.DB.Get(&callback, searchString, input.CallbackDisplayID, *input.OperationID); err != nil {\n\t\t\tlogging.LogError(err, \"Failed to find task based on task id and operation id\")\n\t\t\tresponse.Error = err.Error()\n\t\t\treturn response\n\t\t} else {\n\t\t\tresponse.CallbackID = callback.ID\n\t\t\tresponse.Success = true\n\t\t\treturn response\n\t\t}\n\t} else {\n\t\tresponse.Error = \"Must specify operation name or operation id\"\n\t\treturn response\n\t}\n}", "func ToCaseClause(x ast.Node) *ast.CaseClause {\n\tif x, ok := x.(*ast.CaseClause); ok {\n\t\treturn x\n\t}\n\treturn NilCaseClause\n}", "func CustomerGetoneVehicleforview(w http.ResponseWriter, r *http.Request) {\n\tvehicle := services.GetOneVehicle(r)\n\tcusttpl.ExecuteTemplate(w, \"viewvehicle.html\", vehicle)\n}", "func (t *InsuranceChaincode) getCustomerByID(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n var customerId, jsonResp string\n\n if len(args) != 1 {\n return shim.Error(\"Incorrect number of arguments. Expecting customerId of the Insurance to query\")\n }\n\n customerId = args[0]\n valAsbytes, err := stub.GetState(customerId) \n if err != nil {\n jsonResp = \"{\\\"Error\\\":\\\"Failed to get state for \" + customerId + \"\\\"}\"\n return shim.Error(jsonResp)\n } else if valAsbytes == nil {\n jsonResp = \"{\\\"CustomerId\\\": \\\"\"+ customerId + \"\\\", \\\"Error\\\":\\\"Customer does not exist.\\\"}\"\n return shim.Error(jsonResp)\n }\n\n return shim.Success(valAsbytes)\n}", "func (client *Client) QueryCustomerAddressListWithCallback(request *QueryCustomerAddressListRequest, callback func(response *QueryCustomerAddressListResponse, err error)) (<-chan int) {\nresult := make(chan int, 1)\nerr := client.AddAsyncTask(func() {\nvar response *QueryCustomerAddressListResponse\nvar err error\ndefer close(result)\nresponse, err = client.QueryCustomerAddressList(request)\ncallback(response, err)\nresult <- 1\n})\nif err != nil {\ndefer close(result)\ncallback(nil, err)\nresult <- 0\n}\nreturn result\n}", "func NewCase(condition string, statements ...Statement) *Case {\n\treturn &Case{\n\t\tcondition: condition,\n\t\tstatements: statements,\n\t\tcaller: fetchClientCallerLine(),\n\t}\n}", "func (client *Client) DescribeExplorerWithCallback(request *DescribeExplorerRequest, callback func(response *DescribeExplorerResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeExplorerResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeExplorer(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) QueryContactInfoWithCallback(request *QueryContactInfoRequest, callback func(response *QueryContactInfoResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *QueryContactInfoResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.QueryContactInfo(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) DescribeOrganizationSpecsWithCallback(request *DescribeOrganizationSpecsRequest, callback func(response *DescribeOrganizationSpecsResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeOrganizationSpecsResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeOrganizationSpecs(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func doCHARableTest(t *testing.T, out io.Writer, f decoder.New, endianness bool, teTa decoder.TestTable) {\n\t//var (\n\t//\t// til is the trace id list content for test\n\t//\tidl = ``\n\t//)\n\t//lu := make(id.TriceIDLookUp) // empty\n\t//luM := new(sync.RWMutex)\n\t//assert.Nil(t, ilu.FromJSON([]byte(idl)))\n\t//lu.AddFmtCount(os.Stdout)\n\tbuf := make([]byte, decoder.DefaultSize)\n\tdec := f(out, nil, nil, nil, nil, endianness) // a new decoder instance\n\tfor _, x := range teTa {\n\t\tin := ioutil.NopCloser(bytes.NewBuffer(x.In))\n\t\tdec.SetInput(in)\n\t\tlineStart := true\n\t\tvar err error\n\t\tvar n int\n\t\tvar act string\n\t\tfor err == nil {\n\t\t\tn, err = dec.Read(buf)\n\t\t\tif n == 0 {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tif decoder.ShowID != \"\" && lineStart {\n\t\t\t\tact += fmt.Sprintf(decoder.ShowID, decoder.LastTriceID)\n\t\t\t}\n\t\t\tact += fmt.Sprint(string(buf[:n]))\n\t\t\tlineStart = false\n\t\t}\n\t\tact = strings.TrimSuffix(act, \"\\\\n\")\n\t\tact = strings.TrimSuffix(act, \"\\n\")\n\t\tassert.Equal(t, x.Exp, act)\n\t}\n}", "func (client *Client) GetServiceInputMappingWithCallback(request *GetServiceInputMappingRequest, callback func(response *GetServiceInputMappingResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetServiceInputMappingResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetServiceInputMapping(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (o GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKeyResponseOutput) CaseInsensitive() pulumi.BoolOutput {\n\treturn o.ApplyT(func(v GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKeyResponse) bool { return v.CaseInsensitive }).(pulumi.BoolOutput)\n}", "func (client *Client) GetStructSyncExecSqlDetailWithCallback(request *GetStructSyncExecSqlDetailRequest, callback func(response *GetStructSyncExecSqlDetailResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetStructSyncExecSqlDetailResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetStructSyncExecSqlDetail(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) DescribeCompanyWithCallback(request *DescribeCompanyRequest, callback func(response *DescribeCompanyResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeCompanyResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeCompany(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (ec *executionContext) _Case(ctx context.Context, sel ast.SelectionSet, obj *models.Case) graphql.Marshaler {\n\tfields := graphql.CollectFields(ctx, sel, caseImplementors)\n\n\tvar wg sync.WaitGroup\n\tout := graphql.NewOrderedMap(len(fields))\n\tinvalid := false\n\tfor i, field := range fields {\n\t\tout.Keys[i] = field.Alias\n\n\t\tswitch field.Name {\n\t\tcase \"__typename\":\n\t\t\tout.Values[i] = graphql.MarshalString(\"Case\")\n\t\tcase \"Id\":\n\t\t\tout.Values[i] = ec._Case_Id(ctx, field, obj)\n\t\tcase \"Asset\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_Asset(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"CaseNumber\":\n\t\t\tout.Values[i] = ec._Case_CaseNumber(ctx, field, obj)\n\t\tcase \"Origin\":\n\t\t\tout.Values[i] = ec._Case_Origin(ctx, field, obj)\n\t\tcase \"Owner\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_Owner(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"Reason\":\n\t\t\tout.Values[i] = ec._Case_Reason(ctx, field, obj)\n\t\tcase \"IsClosed\":\n\t\t\tout.Values[i] = ec._Case_IsClosed(ctx, field, obj)\n\t\tcase \"Contact\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_Contact(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"CreatedBy\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_CreatedBy(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"ClosedDate\":\n\t\t\tout.Values[i] = ec._Case_ClosedDate(ctx, field, obj)\n\t\tcase \"CreatedDate\":\n\t\t\tout.Values[i] = ec._Case_CreatedDate(ctx, field, obj)\n\t\tcase \"IsDeleted\":\n\t\t\tout.Values[i] = ec._Case_IsDeleted(ctx, field, obj)\n\t\tcase \"Description\":\n\t\t\tout.Values[i] = ec._Case_Description(ctx, field, obj)\n\t\tcase \"IsEscalated\":\n\t\t\tout.Values[i] = ec._Case_IsEscalated(ctx, field, obj)\n\t\tcase \"LastModifiedBy\":\n\t\t\twg.Add(1)\n\t\t\tgo func(i int, field graphql.CollectedField) {\n\t\t\t\tout.Values[i] = ec._Case_LastModifiedBy(ctx, field, obj)\n\t\t\t\twg.Done()\n\t\t\t}(i, field)\n\t\tcase \"LastModifiedDate\":\n\t\t\tout.Values[i] = ec._Case_LastModifiedDate(ctx, field, obj)\n\t\tcase \"LastReferencedDate\":\n\t\t\tout.Values[i] = ec._Case_LastReferencedDate(ctx, field, obj)\n\t\tcase \"LastViewedDate\":\n\t\t\tout.Values[i] = ec._Case_LastViewedDate(ctx, field, obj)\n\t\tdefault:\n\t\t\tpanic(\"unknown field \" + strconv.Quote(field.Name))\n\t\t}\n\t}\n\twg.Wait()\n\tif invalid {\n\t\treturn graphql.Null\n\t}\n\treturn out\n}", "func callback(nlm *C.struct_nl_msg, nla unsafe.Pointer) C.int {\n\tcbID := uintptr(nla)\n\tcallbacksLock.RLock()\n\tcbArg := callbacks[cbID]\n\tcallbacksLock.RUnlock()\n\n\tif cbArg == nil {\n\t\tpanic(fmt.Sprintf(\"No netlink callback with ID %d\", cbID))\n\t}\n\n\tcbMsg := &Message{nlm: nlm}\n\tif err := cbArg.fn(cbMsg, cbArg.arg); err != nil {\n\t\tcbArg.err = err\n\t\treturn C.NL_STOP\n\t}\n\treturn C.NL_OK\n}", "func LookupTestCase(ctx *pulumi.Context, args *LookupTestCaseArgs, opts ...pulumi.InvokeOption) (*LookupTestCaseResult, error) {\n\topts = internal.PkgInvokeDefaultOpts(opts)\n\tvar rv LookupTestCaseResult\n\terr := ctx.Invoke(\"google-native:dialogflow/v3:getTestCase\", args, &rv, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &rv, nil\n}", "func (h *Handler) CallbackHandler(w http.ResponseWriter, r *http.Request, p webapi.CallbackParams) error {\n\tswitch p.Type {\n\tcase gravityLoginAction: // login via tele login\n\t\turl, err := h.constructConsoleResponse(p.ClientRedirectURL, p.Username)\n\t\tif err != nil {\n\t\t\treturn trace.Wrap(err)\n\t\t}\n\t\thttp.Redirect(w, r, url.String(), http.StatusFound)\n\t\treturn nil\n\tdefault: // call the base (open-source) handler for web sign in\n\t\treturn h.Handler.CallbackHandler(w, r, p)\n\t}\n}", "func (client *Client) DescribeAntChainConsortiumsWithCallback(request *DescribeAntChainConsortiumsRequest, callback func(response *DescribeAntChainConsortiumsResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeAntChainConsortiumsResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeAntChainConsortiums(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetArmsConsoleUrlWithCallback(request *GetArmsConsoleUrlRequest, callback func(response *GetArmsConsoleUrlResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetArmsConsoleUrlResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetArmsConsoleUrl(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func getActiveCharacter(mid string) string {\n var profileResponse interface{}\n\n // Make GET request to Profile endpoint\n client := &http.Client{}\n reqURL := \"https://www.bungie.net/platform/Destiny2/3/Profile/\" +\n mid +\n \"/?components=200\"\n req, _ := http.NewRequest(\"GET\", reqURL, nil)\n req.Header.Add(\"X-API-Key\", os.Getenv(\"API_KEY\"))\n resp, err := client.Do(req)\n if ( err != nil) {\n fmt.Println(err)\n }\n // Parse response json for character ids\n err = json.NewDecoder(resp.Body).Decode(&profileResponse)\n if ( err != nil ) {\n fmt.Println(err)\n }\n resp.Body.Close()\n\n // Get relevant json data\n responseJSON := profileResponse.(map[string]interface{})\n responseMap := responseJSON[\"Response\"].(map[string]interface{})\n characterMap := responseMap[\"characters\"].(map[string]interface{})[\"data\"].(map[string]interface{})\n\n activeCharacter := \"-1\"\n latestDate := time.Time{}\n\n for k, v := range characterMap {\n dateString := v.(map[string]interface{})[\"dateLastPlayed\"].(string) // e.g. \"2020-01-09T06:11:35Z\"\n date, _ := time.Parse(\n time.RFC3339,\n dateString)\n if (date.After(latestDate)) {\n activeCharacter = k\n latestDate = date\n }\n }\n\n return activeCharacter\n}", "func GetCategory(p providers.CategoryProvider) func(c *fiber.Ctx) error {\n\treturn func(c *fiber.Ctx) error {\n\t\tcategoryID, _ := strconv.Atoi(c.Params(\"id\"))\n\t\tcategory, err := p.CategoryGet(categoryID)\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// wrapped into array to it works in the template\n\t\tresult := make([]*models.Category, 0)\n\t\tresult = append(result, category)\n\t\treturn c.Render(\"category\", result)\n\t}\n}", "func Callback(cbReq *CallbackRequest, opts *CallbackOptions) error {\n\tclient := opts.Client\n\tif client == nil {\n\t\tclient = http.DefaultClient\n\t}\n\tbuf := bytes.NewBuffer(nil)\n\terr := json.NewEncoder(buf).Encode(cbReq)\n\tif err != nil {\n\t\treturn err\n\t}\n\tsignature, err := opts.Signer.Sign(buf.Bytes())\n\tif err != nil {\n\t\treturn err\n\t}\n\treq, err := http.NewRequest(\"POST\", cbReq.StatusCallbackUrl, buf)\n\tif err != nil {\n\t\treturn err\n\t}\n\treq.Header.Set(\"X-OpenGDPR-Processor-Domain\", opts.ProcessorDomain)\n\treq.Header.Set(\"X-OpenGDPR-Signature\", signature)\n\t// Attempt to make callback\n\tfor i := 0; i < opts.MaxAttempts; i++ {\n\t\tresp, err := client.Do(req)\n\t\tif err != nil || resp.StatusCode != 200 {\n\t\t\ttime.Sleep(opts.Backoff)\n\t\t\tcontinue\n\t\t}\n\t\t// Success\n\t\treturn nil\n\t}\n\treturn fmt.Errorf(\"callback timed out for %s\", cbReq.StatusCallbackUrl)\n}", "func (_Cakevault *CakevaultTransactor) InCaseTokensGetStuck(opts *bind.TransactOpts, _token common.Address) (*types.Transaction, error) {\n\treturn _Cakevault.contract.Transact(opts, \"inCaseTokensGetStuck\", _token)\n}", "func (client *Client) DescribeEventDetailWithCallback(request *DescribeEventDetailRequest, callback func(response *DescribeEventDetailResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeEventDetailResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeEventDetail(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func callbackHandler(res http.ResponseWriter, req *http.Request) {\n\n\t// Complete the authentication process and fetch all of the\n\t// basic information about the user from the provider.\n\tuser, err := gothic.CompleteUserAuth(res, req)\n\tif err != nil {\n\t\tfmt.Fprintln(res, err)\n\t\treturn\n\t}\n\n\t// Execute the template for this user and respond with\n\t// the user page.\n\ttemplates.ExecuteTemplate(res, \"user\", user)\n}", "func To(_case int, rune int) int {\n\tif _case < 0 || MaxCase <= _case {\n\t\treturn ReplacementChar\t// as reasonable an error as any\n\t}\n\t// binary search over ranges\n\tlo := 0;\n\thi := len(CaseRanges);\n\tfor lo < hi {\n\t\tm := lo + (hi-lo)/2;\n\t\tr := CaseRanges[m];\n\t\tif r.Lo <= rune && rune <= r.Hi {\n\t\t\tdelta := int(r.Delta[_case]);\n\t\t\tif delta > MaxRune {\n\t\t\t\t// In an Upper-Lower sequence, which always starts with\n\t\t\t\t// an UpperCase letter, the real deltas always look like:\n\t\t\t\t//\t{0, 1, 0} UpperCase (Lower is next)\n\t\t\t\t//\t{-1, 0, -1} LowerCase (Upper, Title are previous)\n\t\t\t\t// The characters at even offsets from the beginning of the\n\t\t\t\t// sequence are upper case; the ones at odd offsets are lower.\n\t\t\t\t// The correct mapping can be done by clearing or setting the low\n\t\t\t\t// bit in the sequence offset.\n\t\t\t\t// The constants UpperCase and TitleCase are even while LowerCase\n\t\t\t\t// is odd so we take the low bit from _case.\n\t\t\t\treturn r.Lo + ((rune-r.Lo)&^1 | _case&1)\n\t\t\t}\n\t\t\treturn rune + delta;\n\t\t}\n\t\tif rune < r.Lo {\n\t\t\thi = m\n\t\t} else {\n\t\t\tlo = m + 1\n\t\t}\n\t}\n\treturn rune;\n}", "func FindCvtermsynonym(exec boil.Executor, cvtermsynonymID int, selectCols ...string) (*Cvtermsynonym, error) {\n\tcvtermsynonymObj := &Cvtermsynonym{}\n\n\tsel := \"*\"\n\tif len(selectCols) > 0 {\n\t\tsel = strings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, selectCols), \",\")\n\t}\n\tquery := fmt.Sprintf(\n\t\t\"select %s from \\\"cvtermsynonym\\\" where \\\"cvtermsynonym_id\\\"=$1\", sel,\n\t)\n\n\tq := queries.Raw(exec, query, cvtermsynonymID)\n\n\terr := q.Bind(cvtermsynonymObj)\n\tif err != nil {\n\t\tif errors.Cause(err) == sql.ErrNoRows {\n\t\t\treturn nil, sql.ErrNoRows\n\t\t}\n\t\treturn nil, errors.Wrap(err, \"chado: unable to select from cvtermsynonym\")\n\t}\n\n\treturn cvtermsynonymObj, nil\n}", "func (o GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKeyOutput) CaseInsensitive() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v GoogleCloudRetailV2alphaSearchRequestFacetSpecFacetKey) *bool { return v.CaseInsensitive }).(pulumi.BoolPtrOutput)\n}", "func (s *BaseCobol85PreprocessorListener) EnterCobolWord(ctx *CobolWordContext) {}", "func (c *TestClient) zoneOperationsWait(project, zone, name string) error {\n\tif c.zoneOperationsWaitFn != nil {\n\t\treturn c.zoneOperationsWaitFn(project, zone, name)\n\t}\n\treturn c.client.zoneOperationsWait(project, zone, name)\n}", "func (h *GitHubOAuth) Callback(c *router.Control) {\n\tstate := c.Get(\"state\")\n\tcode := c.Get(\"code\")\n\n\tif state != h.state {\n\t\th.log.Errorf(\"Wrong state %s with code %s\", state, code)\n\t\thttp.Redirect(c.Writer, c.Request, \"/\", http.StatusMovedPermanently)\n\t\treturn\n\t}\n\n\tctx := context.Background()\n\ttoken, err := h.oAuthConf.Exchange(ctx, code)\n\n\tif err != nil {\n\t\th.log.Errorf(\"Exchange failed for code %s: %+v\", code, err)\n\t\thttp.Redirect(c.Writer, c.Request, \"/\", http.StatusMovedPermanently)\n\t\treturn\n\t}\n\n\toauthClient := h.oAuthConf.Client(ctx, token)\n\tgithubClient := ghClient.NewClient(oauthClient)\n\tuser, _, err := githubClient.Users.Get(ctx, \"\")\n\tif err != nil || user.Login == nil {\n\t\th.log.Errorf(\"Couldn't get user for code %s: %+v\", code, err)\n\t\thttp.Redirect(c.Writer, c.Request, \"/\", http.StatusMovedPermanently)\n\t\treturn\n\t}\n\n\th.log.WithField(\"user\", *user.Login).Info(\"GitHub user was authorized in oauth-proxy\")\n\n\tsessionData := session.NewSessionOptions(&session.SessOptions{\n\t\tCAttrs: map[string]interface{}{\"Login\": *user.Login, \"Source\": models.SourceGitHub},\n\t\tAttrs: map[string]interface{}{\"Activated\": false, \"HasError\": false},\n\t})\n\tsession.Add(sessionData, c.Writer)\n\n\tgo h.syncUser(*user.Login, sessionData, c.Writer)\n\n\thttp.Redirect(c.Writer, c.Request, \"/\", http.StatusMovedPermanently)\n}", "func (client *Client) DescribeMeterLiveRtcDurationWithCallback(request *DescribeMeterLiveRtcDurationRequest, callback func(response *DescribeMeterLiveRtcDurationResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeMeterLiveRtcDurationResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeMeterLiveRtcDuration(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (m *WorkflowDecisionCase) UnmarshalJSON(raw []byte) error {\n\t// AO0\n\tvar dataAO0 struct {\n\t\tDescription string `json:\"Description,omitempty\"`\n\n\t\tNextTask string `json:\"NextTask,omitempty\"`\n\n\t\tValue string `json:\"Value,omitempty\"`\n\t}\n\tif err := swag.ReadJSON(raw, &dataAO0); err != nil {\n\t\treturn err\n\t}\n\n\tm.Description = dataAO0.Description\n\n\tm.NextTask = dataAO0.NextTask\n\n\tm.Value = dataAO0.Value\n\n\treturn nil\n}", "func CanaryIncCase(context.Context, TimerManager, int) error {\n\treturn nil\n}", "func (client *Client) ModifyOcspStatusWithCallback(request *ModifyOcspStatusRequest, callback func(response *ModifyOcspStatusResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *ModifyOcspStatusResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.ModifyOcspStatus(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (d *Dao) CaseInfo(c context.Context, cid int64) (r *model.BlockedCase, err error) {\n\trow := d.db.QueryRow(c, _getCaseByIDSQL, cid)\n\tr = &model.BlockedCase{}\n\tif err = row.Scan(&r.ID, &r.MID, &r.Status, &r.OriginContent, &r.PunishResult, &r.OriginTitle, &r.OriginURL, &r.EndTime, &r.VoteRule, &r.VoteBreak, &r.VoteDelete, &r.OriginType, &r.ReasonType, &r.JudgeType, &r.BlockedDays, &r.PutTotal, &r.StartTime, &r.EndTime, &r.Operator, &r.CTime, &r.MTime, &r.RelationID, &r.CaseType); err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil\n\t\t\tr = nil\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func (client IdentityClient) getCompartment(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodGet, \"/compartments/{compartmentId}\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response GetCompartmentResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func GetTrafficOpsCookie(cdnUri, user, pass string) (string, error) {\n\turi := cdnUri + `/api/1.2/user/login`\n\tpostdata := `{\"u\":\"` + user + `\", \"p\":\"` + pass + `\"}`\n\treq, err := http.NewRequest(\"POST\", uri, strings.NewReader(postdata))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treq.Header.Add(\"Accept\", \"application/json\")\n\n\tclient := getClient()\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer resp.Body.Close()\n\n\tfor _, cookie := range resp.Cookies() {\n\t\tif cookie.Name == `mojolicious` {\n\t\t\treturn cookie.Value, nil\n\t\t}\n\t}\n\n\tdata, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn \"\", errors.New(\"No login cookie received: \" + string(data))\n}", "func (client *Client) GetSubscriptionItemDetailWithCallback(request *GetSubscriptionItemDetailRequest, callback func(response *GetSubscriptionItemDetailResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetSubscriptionItemDetailResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetSubscriptionItemDetail(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (s *WeatherData) verifyCause(APIstub shim.ChaincodeStubInterface, polyId string ) sc.Response {\n\t\n\tfmt.Println(\"============= START : Fetching polygon details by the polygon id =============\")\n\tvar datasource = \"URL\" // Setting the Oraclize datasource\n\tvar query = \"json(http://api.agromonitoring.com/agro/1.0/weather?polyid=\" + polyId + \"&appid=\" + API_KEY + \")\" // Setting the query\n\tresult, proof := oraclizeapi.OraclizeQuery_sync(APIstub, datasource, query, oraclizeapi.TLSNOTARY)\n\tfmt.Printf(\"proof: %s\", proof)\n\tfmt.Printf(\"\\nresult: %s\\n\", result)\n\tvar response = {\n\t\t\"result\": result,\n\t\t\"proof\": proof\n\t}", "func (client *Client) DescribePortConnsListWithCallback(request *DescribePortConnsListRequest, callback func(response *DescribePortConnsListResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribePortConnsListResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribePortConnsList(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetAlarmLogWithCallback(request *GetAlarmLogRequest, callback func(response *GetAlarmLogResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetAlarmLogResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetAlarmLog(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (_AnchorChain *AnchorChainTransactor) Callback(opts *bind.TransactOpts, state bool, _result []string) (*types.Transaction, error) {\n\treturn _AnchorChain.contract.Transact(opts, \"callback\", state, _result)\n}", "func (client *Client) DescribeCertificatesWithCallback(request *DescribeCertificatesRequest, callback func(response *DescribeCertificatesResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeCertificatesResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeCertificates(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (_AnchorChain *AnchorChainTransactorSession) Callback(state bool, _result []string) (*types.Transaction, error) {\n\treturn _AnchorChain.Contract.Callback(&_AnchorChain.TransactOpts, state, _result)\n}", "func (client *Client) DescribeDialogueNodeStatisticsWithCallback(request *DescribeDialogueNodeStatisticsRequest, callback func(response *DescribeDialogueNodeStatisticsResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeDialogueNodeStatisticsResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeDialogueNodeStatistics(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}" ]
[ "0.66025335", "0.66022605", "0.5441328", "0.5314528", "0.4794885", "0.47772574", "0.44892713", "0.4443668", "0.43273535", "0.42628276", "0.4138406", "0.41172293", "0.41120696", "0.4096734", "0.40186015", "0.3951663", "0.39511362", "0.3897255", "0.38712794", "0.38415867", "0.3831616", "0.38205725", "0.38187796", "0.381507", "0.38043293", "0.37974733", "0.37413633", "0.37012738", "0.36630654", "0.3658415", "0.36554533", "0.36487493", "0.36351782", "0.3631789", "0.35783908", "0.35746962", "0.35458294", "0.35357213", "0.35115975", "0.34983984", "0.34977302", "0.34884718", "0.34799224", "0.3458872", "0.3454686", "0.343636", "0.3424452", "0.34213525", "0.3415038", "0.34058017", "0.34007478", "0.3396418", "0.33959967", "0.33891022", "0.33860946", "0.33758402", "0.33742633", "0.33623666", "0.33554623", "0.3349816", "0.33456367", "0.3326986", "0.33139527", "0.33137095", "0.33070815", "0.33035988", "0.3303151", "0.3301019", "0.3299541", "0.32926297", "0.32870528", "0.32848346", "0.3282053", "0.32810012", "0.32801476", "0.32786223", "0.32674655", "0.32653436", "0.3258584", "0.32529557", "0.32503602", "0.32497633", "0.3249119", "0.32430133", "0.32401207", "0.32311505", "0.3227725", "0.32274032", "0.3209645", "0.32055843", "0.32007238", "0.32007146", "0.3200202", "0.31980023", "0.31890017", "0.3188501", "0.31848612", "0.31833953", "0.31810832", "0.31771573" ]
0.8397443
0
CreateGetOcJusticeTerminalCaseRequest creates a request to invoke GetOcJusticeTerminalCase API
func CreateGetOcJusticeTerminalCaseRequest() (request *GetOcJusticeTerminalCaseRequest) { request = &GetOcJusticeTerminalCaseRequest{ RpcRequest: &requests.RpcRequest{}, } request.InitWithApiInfo("dt-oc-info", "2022-08-29", "GetOcJusticeTerminalCase", "", "") request.Method = requests.POST return }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (client *Client) GetOcJusticeTerminalCase(request *GetOcJusticeTerminalCaseRequest) (response *GetOcJusticeTerminalCaseResponse, err error) {\n\tresponse = CreateGetOcJusticeTerminalCaseResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func CreateGetOcJusticeTerminalCaseResponse() (response *GetOcJusticeTerminalCaseResponse) {\n\tresponse = &GetOcJusticeTerminalCaseResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func (client *Client) GetOcJusticeTerminalCaseWithCallback(request *GetOcJusticeTerminalCaseRequest, callback func(response *GetOcJusticeTerminalCaseResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetOcJusticeTerminalCaseResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetOcJusticeTerminalCase(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (client *Client) GetOcJusticeTerminalCaseWithChan(request *GetOcJusticeTerminalCaseRequest) (<-chan *GetOcJusticeTerminalCaseResponse, <-chan error) {\n\tresponseChan := make(chan *GetOcJusticeTerminalCaseResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetOcJusticeTerminalCase(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func CreateGetOpenNLURequest() (request *GetOpenNLURequest) {\n\trequest = &GetOpenNLURequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"alinlp\", \"2020-06-29\", \"GetOpenNLU\", \"alinlp\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateGetWsCustomizedChO2ORequest() (request *GetWsCustomizedChO2ORequest) {\n\trequest = &GetWsCustomizedChO2ORequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"alinlp\", \"2020-06-29\", \"GetWsCustomizedChO2O\", \"alinlp\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateDescribeCustinsKernelReleaseNotesRequest() (request *DescribeCustinsKernelReleaseNotesRequest) {\n\trequest = &DescribeCustinsKernelReleaseNotesRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Rds\", \"2014-08-15\", \"DescribeCustinsKernelReleaseNotes\", \"rds\", \"openAPI\")\n\treturn\n}", "func CreateGetIndustryCommerceInfoRequest() (request *GetIndustryCommerceInfoRequest) {\n\trequest = &GetIndustryCommerceInfoRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"companyreg\", \"2020-10-22\", \"GetIndustryCommerceInfo\", \"companyreg\", \"openAPI\")\n\trequest.Method = requests.GET\n\treturn\n}", "func CreateListCasesRequest() (request *ListCasesRequest) {\n\trequest = &ListCasesRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"CCC\", \"2020-07-01\", \"ListCases\", \"CCC\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateGetKeywordChEcomRequest() (request *GetKeywordChEcomRequest) {\n\trequest = &GetKeywordChEcomRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"alinlp\", \"2020-06-29\", \"GetKeywordChEcom\", \"alinlp\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (client *Datetimerfc1123Client) getUTCLowercaseMaxDateTimeCreateRequest(ctx context.Context, options *Datetimerfc1123ClientGetUTCLowercaseMaxDateTimeOptions) (*policy.Request, error) {\n\turlPath := \"/datetimerfc1123/max/lowercase\"\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func CreateGetTaxationInfoRequest() (request *GetTaxationInfoRequest) {\n\trequest = &GetTaxationInfoRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"companyreg\", \"2020-10-22\", \"GetTaxationInfo\", \"companyreg\", \"openAPI\")\n\trequest.Method = requests.GET\n\treturn\n}", "func CreateDescribeIpDdosThresholdRequest() (request *DescribeIpDdosThresholdRequest) {\n\trequest = &DescribeIpDdosThresholdRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"antiddos-public\", \"2017-05-18\", \"DescribeIpDdosThreshold\", \"ddosbasic\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateGetContactRequest() (request *GetContactRequest) {\n\trequest = &GetContactRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Subscription\", \"2021-01-15\", \"GetContact\", \"\", \"\")\n\treturn\n}", "func CreateDescribeOssObjectDetailRequest() (request *DescribeOssObjectDetailRequest) {\n\trequest = &DescribeOssObjectDetailRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Sddp\", \"2019-01-03\", \"DescribeOssObjectDetail\", \"sddp\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateGetWsCustomizedChEcomContentRequest() (request *GetWsCustomizedChEcomContentRequest) {\n\trequest = &GetWsCustomizedChEcomContentRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"alinlp\", \"2020-06-29\", \"GetWsCustomizedChEcomContent\", \"alinlp\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (client *CapacitiesClient) getDetailsCreateRequest(ctx context.Context, resourceGroupName string, dedicatedCapacityName string, options *CapacitiesClientGetDetailsOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.PowerBIDedicated/capacities/{dedicatedCapacityName}\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif dedicatedCapacityName == \"\" {\n\t\treturn nil, errors.New(\"parameter dedicatedCapacityName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{dedicatedCapacityName}\", url.PathEscape(dedicatedCapacityName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-01-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func (client *KeyVaultClient) getCertificateOperationCreateRequest(ctx context.Context, vaultBaseURL string, certificateName string, options *KeyVaultClientGetCertificateOperationOptions) (*policy.Request, error) {\n\thost := \"{vaultBaseUrl}\"\n\thost = strings.ReplaceAll(host, \"{vaultBaseUrl}\", vaultBaseURL)\n\turlPath := \"/certificates/{certificate-name}/pending\"\n\tif certificateName == \"\" {\n\t\treturn nil, errors.New(\"parameter certificateName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{certificate-name}\", url.PathEscape(certificateName))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"7.2\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (client *CapacitiesClient) listCreateRequest(ctx context.Context, options *CapacitiesClientListOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/providers/Microsoft.PowerBIDedicated/capacities\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-01-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func CreateGetTaobaoOrderRequest() (request *GetTaobaoOrderRequest) {\n\trequest = &GetTaobaoOrderRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"CloudCallCenter\", \"2017-07-05\", \"GetTaobaoOrder\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateOemSitingSelctionRequest() (request *OemSitingSelctionRequest) {\n\trequest = &OemSitingSelctionRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"cloudwf\", \"2017-03-28\", \"OemSitingSelction\", \"cloudwf\", \"openAPI\")\n\treturn\n}", "func CreateDescribeLiveDomainCertificateInfoRequest() (request *DescribeLiveDomainCertificateInfoRequest) {\n\trequest = &DescribeLiveDomainCertificateInfoRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"live\", \"2016-11-01\", \"DescribeLiveDomainCertificateInfo\", \"live\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (client *IncidentsClient) getCreateRequest(ctx context.Context, resourceGroupName string, workspaceName string, incidentID string, options *IncidentsClientGetOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.OperationalInsights/workspaces/{workspaceName}/providers/Microsoft.SecurityInsights/incidents/{incidentId}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif workspaceName == \"\" {\n\t\treturn nil, errors.New(\"parameter workspaceName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{workspaceName}\", url.PathEscape(workspaceName))\n\tif incidentID == \"\" {\n\t\treturn nil, errors.New(\"parameter incidentID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{incidentId}\", url.PathEscape(incidentID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2022-05-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func CreateStartNotaryRequest() (request *StartNotaryRequest) {\n\trequest = &StartNotaryRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Trademark\", \"2018-07-24\", \"StartNotary\", \"trademark\", \"openAPI\")\n\treturn\n}", "func (client *ContainerClient) getPropertiesCreateRequest(ctx context.Context, options *ContainerClientGetPropertiesOptions, leaseAccessConditions *LeaseAccessConditions) (*policy.Request, error) {\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, client.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"restype\", \"container\")\n\tif options != nil && options.Timeout != nil {\n\t\treqQP.Set(\"timeout\", strconv.FormatInt(int64(*options.Timeout), 10))\n\t}\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\tif leaseAccessConditions != nil && leaseAccessConditions.LeaseID != nil {\n\t\treq.Raw().Header[\"x-ms-lease-id\"] = []string{*leaseAccessConditions.LeaseID}\n\t}\n\treq.Raw().Header[\"x-ms-version\"] = []string{\"2020-10-02\"}\n\tif options != nil && options.RequestID != nil {\n\t\treq.Raw().Header[\"x-ms-client-request-id\"] = []string{*options.RequestID}\n\t}\n\treq.Raw().Header[\"Accept\"] = []string{\"application/xml\"}\n\treturn req, nil\n}", "func (client *ApplicationClient) listOperationsCreateRequest(ctx context.Context, options *ApplicationClientListOperationsOptions) (*policy.Request, error) {\n\turlPath := \"/providers/Microsoft.Solutions/operations\"\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2018-06-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (c *Client) CreateCase(ctx context.Context, params *CreateCaseInput, optFns ...func(*Options)) (*CreateCaseOutput, error) {\n\tif params == nil {\n\t\tparams = &CreateCaseInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"CreateCase\", params, optFns, addOperationCreateCaseMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*CreateCaseOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func CreateDescribeVnKnowledgeRequest() (request *DescribeVnKnowledgeRequest) {\n\trequest = &DescribeVnKnowledgeRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"CloudCallCenter\", \"2017-07-05\", \"DescribeVnKnowledge\", \"\", \"\")\n\trequest.Method = requests.GET\n\treturn\n}", "func CreateGetOpenNLUResponse() (response *GetOpenNLUResponse) {\n\tresponse = &GetOpenNLUResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateListCityMapAoisRequest() (request *ListCityMapAoisRequest) {\n\trequest = &ListCityMapAoisRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"CDRS\", \"2020-11-01\", \"ListCityMapAois\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (c *Client) BuildCreateRequest(ctx context.Context, v interface{}) (*http.Request, error) {\n\tu := &url.URL{Scheme: c.scheme, Host: c.host, Path: CreateWarehousePath()}\n\treq, err := http.NewRequest(\"POST\", u.String(), nil)\n\tif err != nil {\n\t\treturn nil, goahttp.ErrInvalidURL(\"Warehouse\", \"Create\", u.String(), err)\n\t}\n\tif ctx != nil {\n\t\treq = req.WithContext(ctx)\n\t}\n\n\treturn req, nil\n}", "func CreateGetServiceInputMappingRequest() (request *GetServiceInputMappingRequest) {\n\trequest = &GetServiceInputMappingRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"industry-brain\", \"2018-07-12\", \"GetServiceInputMapping\", \"\", \"\")\n\treturn\n}", "func (client *ReservationsDetailsClient) listByReservationOrderCreateRequest(ctx context.Context, reservationOrderID string, filter string, options *ReservationsDetailsListByReservationOrderOptions) (*policy.Request, error) {\n\turlPath := \"/providers/Microsoft.Capacity/reservationorders/{reservationOrderId}/providers/Microsoft.Consumption/reservationDetails\"\n\tif reservationOrderID == \"\" {\n\t\treturn nil, errors.New(\"parameter reservationOrderID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{reservationOrderId}\", url.PathEscape(reservationOrderID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"$filter\", filter)\n\treqQP.Set(\"api-version\", \"2021-10-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (client *DicomServicesClient) listByWorkspaceCreateRequest(ctx context.Context, resourceGroupName string, workspaceName string, options *DicomServicesClientListByWorkspaceOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.HealthcareApis/workspaces/{workspaceName}/dicomservices\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif workspaceName == \"\" {\n\t\treturn nil, errors.New(\"parameter workspaceName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{workspaceName}\", url.PathEscape(workspaceName))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-06-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func CreateGetAlarmLogRequest() (request *GetAlarmLogRequest) {\n\trequest = &GetAlarmLogRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"emas-appmonitor\", \"2019-06-11\", \"GetAlarmLog\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (client *NotebookClient) getNotebooksByWorkspaceCreateRequest(ctx context.Context, options *NotebookClientGetNotebooksByWorkspaceOptions) (*policy.Request, error) {\n\turlPath := \"/notebooks\"\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.endpoint, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2020-12-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func NewCreateClinicRequest(server string, body CreateClinicJSONRequestBody) (*http.Request, error) {\n\tvar bodyReader io.Reader\n\tbuf, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbodyReader = bytes.NewReader(buf)\n\treturn NewCreateClinicRequestWithBody(server, \"application/json\", bodyReader)\n}", "func CreateDescribeCertificatesRequest() (request *DescribeCertificatesRequest) {\n\trequest = &DescribeCertificatesRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"waf-openapi\", \"2019-09-10\", \"DescribeCertificates\", \"waf\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (client *DicomServicesClient) getCreateRequest(ctx context.Context, resourceGroupName string, workspaceName string, dicomServiceName string, options *DicomServicesClientGetOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.HealthcareApis/workspaces/{workspaceName}/dicomservices/{dicomServiceName}\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif workspaceName == \"\" {\n\t\treturn nil, errors.New(\"parameter workspaceName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{workspaceName}\", url.PathEscape(workspaceName))\n\tif dicomServiceName == \"\" {\n\t\treturn nil, errors.New(\"parameter dicomServiceName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{dicomServiceName}\", url.PathEscape(dicomServiceName))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-06-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (client *BuildServiceClient) getBuildCreateRequest(ctx context.Context, resourceGroupName string, serviceName string, buildServiceName string, buildName string, options *BuildServiceClientGetBuildOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/buildServices/{buildServiceName}/builds/{buildName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif serviceName == \"\" {\n\t\treturn nil, errors.New(\"parameter serviceName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{serviceName}\", url.PathEscape(serviceName))\n\tif buildServiceName == \"\" {\n\t\treturn nil, errors.New(\"parameter buildServiceName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{buildServiceName}\", url.PathEscape(buildServiceName))\n\tif buildName == \"\" {\n\t\treturn nil, errors.New(\"parameter buildName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{buildName}\", url.PathEscape(buildName))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-01-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func CreateGetChartRepositoryRequest() (request *GetChartRepositoryRequest) {\n\trequest = &GetChartRepositoryRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"cr\", \"2018-12-01\", \"GetChartRepository\", \"acr\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (client *KustoOperationsClient) listCreateRequest(ctx context.Context, options *KustoOperationsClientListOptions) (*policy.Request, error) {\n\turlPath := \"/providers/Microsoft.Synapse/kustooperations\"\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-06-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (c *Client) BuildCreateRequest(ctx context.Context, v interface{}) (*http.Request, error) {\n\tu := &url.URL{Scheme: c.scheme, Host: c.host, Path: CreateLogPath()}\n\treq, err := http.NewRequest(\"POST\", u.String(), nil)\n\tif err != nil {\n\t\treturn nil, goahttp.ErrInvalidURL(\"log\", \"create\", u.String(), err)\n\t}\n\tif ctx != nil {\n\t\treq = req.WithContext(ctx)\n\t}\n\n\treturn req, nil\n}", "func CreateDescribeOrganizationSpecsRequest() (request *DescribeOrganizationSpecsRequest) {\n\trequest = &DescribeOrganizationSpecsRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Baas\", \"2018-07-31\", \"DescribeOrganizationSpecs\", \"\", \"\")\n\treturn\n}", "func NewGetaspecificNcosLevelRequest(server string, id string) (*http.Request, error) {\n\tvar err error\n\n\tvar pathParam0 string\n\n\tpathParam0, err = runtime.StyleParam(\"simple\", false, \"id\", id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/ncoslevels/%s\", pathParam0)\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (client *IotSecuritySolutionClient) getCreateRequest(ctx context.Context, resourceGroupName string, solutionName string, options *IotSecuritySolutionClientGetOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Security/iotSecuritySolutions/{solutionName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif solutionName == \"\" {\n\t\treturn nil, errors.New(\"parameter solutionName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{solutionName}\", url.PathEscape(solutionName))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2019-08-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func CreateDescribeExplorerRequest() (request *DescribeExplorerRequest) {\n\trequest = &DescribeExplorerRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Baas\", \"2018-07-31\", \"DescribeExplorer\", \"\", \"\")\n\treturn\n}", "func CreateCreateCdpOrderRequest() (request *CreateCdpOrderRequest) {\n\trequest = &CreateCdpOrderRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Dycdpapi\", \"2018-06-10\", \"createCdpOrder\", \"\", \"\")\n\trequest.Domain = \"dycdpapi.aliyuncs.com\"\n\treturn\n}", "func CreateGetHealthMonitorLogsRequest() (request *GetHealthMonitorLogsRequest) {\n\trequest = &GetHealthMonitorLogsRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"EHPC\", \"2018-04-12\", \"GetHealthMonitorLogs\", \"\", \"\")\n\trequest.Method = requests.GET\n\treturn\n}", "func NewGetDistrictForTermRequest(\n\tserver string,\n\tid string,\n) (*http.Request, error) {\n\tvar err error\n\n\tvar pathParam0 string\n\n\tpathParam0, err = runtime.StyleParam(\"simple\", false, \"id\", id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/terms/%s/district\", pathParam0)\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (client *Client) getChatCompletionsCreateRequest(ctx context.Context, body ChatCompletionsOptions, options *GetChatCompletionsOptions) (*policy.Request, error) {\n\turlPath := \"chat/completions\"\n\treq, err := runtime.NewRequest(ctx, http.MethodPost, client.formatURL(urlPath, getDeploymentID(body)))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-07-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\tif err := runtime.MarshalAsJSON(req, body); err != nil {\n\t\treturn nil, err\n\t}\n\treturn req, nil\n}", "func NewGetaspecificCustomerZoneCostRequest(server string, id string) (*http.Request, error) {\n\tvar err error\n\n\tvar pathParam0 string\n\n\tpathParam0, err = runtime.StyleParam(\"simple\", false, \"id\", id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/customerzonecosts/%s\", pathParam0)\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (c *Client) BuildGetActiveOrderRequest(ctx context.Context, v interface{}) (*http.Request, error) {\n\tu := &url.URL{Scheme: c.scheme, Host: c.host, Path: GetActiveOrderRestAPIPath()}\n\treq, err := http.NewRequest(\"GET\", u.String(), nil)\n\tif err != nil {\n\t\treturn nil, goahttp.ErrInvalidURL(\"RestAPI\", \"getActiveOrder\", u.String(), err)\n\t}\n\tif ctx != nil {\n\t\treq = req.WithContext(ctx)\n\t}\n\n\treturn req, nil\n}", "func NewCreateClinicRequestWithBody(server string, contentType string, body io.Reader) (*http.Request, error) {\n\tvar err error\n\n\tserverURL, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\toperationPath := fmt.Sprintf(\"/v1/clinics\")\n\tif operationPath[0] == '/' {\n\t\toperationPath = \".\" + operationPath\n\t}\n\n\tqueryURL, err := serverURL.Parse(operationPath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"POST\", queryURL.String(), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"Content-Type\", contentType)\n\n\treturn req, nil\n}", "func CreateModifyOcspStatusRequest() (request *ModifyOcspStatusRequest) {\n\trequest = &ModifyOcspStatusRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"ddoscoo\", \"2020-01-01\", \"ModifyOcspStatus\", \"ddoscoo\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (client *ContainerClient) getAccessPolicyCreateRequest(ctx context.Context, options *ContainerClientGetAccessPolicyOptions, leaseAccessConditions *LeaseAccessConditions) (*policy.Request, error) {\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, client.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"restype\", \"container\")\n\treqQP.Set(\"comp\", \"acl\")\n\tif options != nil && options.Timeout != nil {\n\t\treqQP.Set(\"timeout\", strconv.FormatInt(int64(*options.Timeout), 10))\n\t}\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\tif leaseAccessConditions != nil && leaseAccessConditions.LeaseID != nil {\n\t\treq.Raw().Header[\"x-ms-lease-id\"] = []string{*leaseAccessConditions.LeaseID}\n\t}\n\treq.Raw().Header[\"x-ms-version\"] = []string{\"2020-10-02\"}\n\tif options != nil && options.RequestID != nil {\n\t\treq.Raw().Header[\"x-ms-client-request-id\"] = []string{*options.RequestID}\n\t}\n\treq.Raw().Header[\"Accept\"] = []string{\"application/xml\"}\n\treturn req, nil\n}", "func (client *BuildServiceClient) getBuildServiceCreateRequest(ctx context.Context, resourceGroupName string, serviceName string, buildServiceName string, options *BuildServiceClientGetBuildServiceOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/buildServices/{buildServiceName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif serviceName == \"\" {\n\t\treturn nil, errors.New(\"parameter serviceName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{serviceName}\", url.PathEscape(serviceName))\n\tif buildServiceName == \"\" {\n\t\treturn nil, errors.New(\"parameter buildServiceName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{buildServiceName}\", url.PathEscape(buildServiceName))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-01-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func NotepadCreateGET(w http.ResponseWriter, r *http.Request) {\n\t// Get session\n\tsess := session.Instance(r)\n\n\t// Display the view\n\tv := view.New(r)\n\tv.Name = \"notepad/create\"\n\tv.Vars[\"token\"] = csrfbanana.Token(w, r, sess)\n\tv.Render(w)\n}", "func (c *Client) BuildGetRequest(ctx context.Context, v interface{}) (*http.Request, error) {\n\tu := &url.URL{Scheme: c.scheme, Host: c.host, Path: GetTermLimitPath()}\n\treq, err := http.NewRequest(\"GET\", u.String(), nil)\n\tif err != nil {\n\t\treturn nil, goahttp.ErrInvalidURL(\"term_limit\", \"get\", u.String(), err)\n\t}\n\tif ctx != nil {\n\t\treq = req.WithContext(ctx)\n\t}\n\n\treturn req, nil\n}", "func CreateVerifyCenRequest() (request *VerifyCenRequest) {\n\trequest = &VerifyCenRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"ecd\", \"2020-09-30\", \"VerifyCen\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateRevokeOperatorRequest() (request *RevokeOperatorRequest) {\n\trequest = &RevokeOperatorRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Oam\", \"2017-01-01\", \"RevokeOperator\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func NewGetaspecificConversationRequest(server string, id string) (*http.Request, error) {\n\tvar err error\n\n\tvar pathParam0 string\n\n\tpathParam0, err = runtime.StyleParam(\"simple\", false, \"id\", id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/conversations/%s\", pathParam0)\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func CreateGetFaceSearchUserRequest() (request *GetFaceSearchUserRequest) {\n\trequest = &GetFaceSearchUserRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"imm\", \"2017-09-06\", \"GetFaceSearchUser\", \"imm\", \"openAPI\")\n\treturn\n}", "func (client *NotebookWorkspacesClient) getCreateRequest(ctx context.Context, resourceGroupName string, accountName string, notebookWorkspaceName NotebookWorkspaceName, options *NotebookWorkspacesGetOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DocumentDB/databaseAccounts/{accountName}/notebookWorkspaces/{notebookWorkspaceName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif accountName == \"\" {\n\t\treturn nil, errors.New(\"parameter accountName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{accountName}\", url.PathEscape(accountName))\n\tif notebookWorkspaceName == \"\" {\n\t\treturn nil, errors.New(\"parameter notebookWorkspaceName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{notebookWorkspaceName}\", url.PathEscape(string(notebookWorkspaceName)))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-10-15\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func NewCreateClinicianRequest(server string, clinicId ClinicId, body CreateClinicianJSONRequestBody) (*http.Request, error) {\n\tvar bodyReader io.Reader\n\tbuf, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbodyReader = bytes.NewReader(buf)\n\treturn NewCreateClinicianRequestWithBody(server, clinicId, \"application/json\", bodyReader)\n}", "func (client *ContainerAppsDiagnosticsClient) getDetectorCreateRequest(ctx context.Context, resourceGroupName string, containerAppName string, detectorName string, options *ContainerAppsDiagnosticsClientGetDetectorOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.App/containerApps/{containerAppName}/detectors/{detectorName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif containerAppName == \"\" {\n\t\treturn nil, errors.New(\"parameter containerAppName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{containerAppName}\", url.PathEscape(containerAppName))\n\tif detectorName == \"\" {\n\t\treturn nil, errors.New(\"parameter detectorName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{detectorName}\", url.PathEscape(detectorName))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-05-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func CreateCreateVSwitchRequest() (request *CreateVSwitchRequest) {\n\trequest = &CreateVSwitchRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Ecs\", \"2014-05-26\", \"CreateVSwitch\", \"ecs\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (c *Client) BuildGetRequest(ctx context.Context, v interface{}) (*http.Request, error) {\n\tvar (\n\t\tid string\n\t)\n\t{\n\t\tp, ok := v.(*warehouse.GetPayload)\n\t\tif !ok {\n\t\t\treturn nil, goahttp.ErrInvalidType(\"Warehouse\", \"Get\", \"*warehouse.GetPayload\", v)\n\t\t}\n\t\tid = p.ID\n\t}\n\tu := &url.URL{Scheme: c.scheme, Host: c.host, Path: GetWarehousePath(id)}\n\treq, err := http.NewRequest(\"GET\", u.String(), nil)\n\tif err != nil {\n\t\treturn nil, goahttp.ErrInvalidURL(\"Warehouse\", \"Get\", u.String(), err)\n\t}\n\tif ctx != nil {\n\t\treq = req.WithContext(ctx)\n\t}\n\n\treturn req, nil\n}", "func CreateGetEMapRequest() (request *GetEMapRequest) {\n\trequest = &GetEMapRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"cusanalytic_sc_online\", \"2019-05-24\", \"GetEMap\", \"\", \"\")\n\treturn\n}", "func CreateDescribeUserVvTopByDayRequest() (request *DescribeUserVvTopByDayRequest) {\n\trequest = &DescribeUserVvTopByDayRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"vod\", \"2017-03-21\", \"DescribeUserVvTopByDay\", \"vod\", \"openAPI\")\n\treturn\n}", "func NewCreateIOCDefault(code int) *CreateIOCDefault {\n\treturn &CreateIOCDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (client *AlertOperationClient) getCreateRequest(ctx context.Context, scope string, operationID string, options *AlertOperationClientGetOptions) (*policy.Request, error) {\n\turlPath := \"/{scope}/providers/Microsoft.Authorization/roleManagementAlertOperations/{operationId}\"\n\turlPath = strings.ReplaceAll(urlPath, \"{scope}\", scope)\n\turlPath = strings.ReplaceAll(urlPath, \"{operationId}\", operationID)\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2022-08-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func CreateDescribeCustinsKernelReleaseNotesResponse() (response *DescribeCustinsKernelReleaseNotesResponse) {\n\tresponse = &DescribeCustinsKernelReleaseNotesResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetManagedRuleRequest() (request *GetManagedRuleRequest) {\n\trequest = &GetManagedRuleRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Config\", \"2020-09-07\", \"GetManagedRule\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (c *Client) BuildRequiredRequest(ctx context.Context, v interface{}) (*http.Request, error) {\n\tu := &url.URL{Scheme: c.scheme, Host: c.host, Path: RequiredOidcPath()}\n\treq, err := http.NewRequest(\"GET\", u.String(), nil)\n\tif err != nil {\n\t\treturn nil, goahttp.ErrInvalidURL(\"oidc\", \"required\", u.String(), err)\n\t}\n\tif ctx != nil {\n\t\treq = req.WithContext(ctx)\n\t}\n\n\treturn req, nil\n}", "func (c *Client) NewCreateCouncillorAdminRequest(ctx context.Context, path string) (*http.Request, error) {\n\tscheme := c.Scheme\n\tif scheme == \"\" {\n\t\tscheme = \"http\"\n\t}\n\tu := url.URL{Host: c.Host, Scheme: scheme, Path: path}\n\treq, err := http.NewRequest(\"POST\", u.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif c.JWTSigner != nil {\n\t\tc.JWTSigner.Sign(req)\n\t}\n\treturn req, nil\n}", "func (u *AttestationClient) CreateEnrollRequest(ctx context.Context, pcaType PCAType) (string, error) {\n\tacaType := apb.ACAType(ACAType(pcaType))\n\treply, err := u.ac.CreateEnrollRequest(ctx, &apb.CreateEnrollRequestRequest{AcaType: &acaType})\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"failed to call |CreateEnrollRequest|\")\n\t}\n\tif reply.GetStatus() != apb.AttestationStatus_STATUS_SUCCESS {\n\t\treturn \"\", &AttestationError{\n\t\t\terrors.Errorf(\"failed |CreateEnrollRequest|: %s\", reply.GetStatus().String()),\n\t\t\treply.GetStatus(),\n\t\t}\n\t}\n\treturn string(reply.GetPcaRequest()), nil\n}", "func CreateGetArmsConsoleUrlRequest() (request *GetArmsConsoleUrlRequest) {\n\trequest = &GetArmsConsoleUrlRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"ARMS\", \"2019-08-08\", \"GetArmsConsoleUrl\", \"arms\", \"openAPI\")\n\trequest.Method = requests.GET\n\treturn\n}", "func CreateGetLniPrivateIpAddressRequest() (request *GetLniPrivateIpAddressRequest) {\n\trequest = &GetLniPrivateIpAddressRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"eflo\", \"2022-05-30\", \"GetLniPrivateIpAddress\", \"eflo\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (client *OperationsClient) operationStatusGetCreateRequest(ctx context.Context, resourceGroupName string, vaultName string, operationID string, options *OperationsOperationStatusGetOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.RecoveryServices/vaults/{vaultName}/operationStatus/{operationId}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif vaultName == \"\" {\n\t\treturn nil, errors.New(\"parameter vaultName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{vaultName}\", url.PathEscape(vaultName))\n\tif operationID == \"\" {\n\t\treturn nil, errors.New(\"parameter operationID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{operationId}\", url.PathEscape(operationID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-08-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (client *CertificateOrdersClient) getCertificateCreateRequest(ctx context.Context, resourceGroupName string, certificateOrderName string, name string, options *CertificateOrdersClientGetCertificateOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.CertificateRegistration/certificateOrders/{certificateOrderName}/certificates/{name}\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif certificateOrderName == \"\" {\n\t\treturn nil, errors.New(\"parameter certificateOrderName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{certificateOrderName}\", url.PathEscape(certificateOrderName))\n\tif name == \"\" {\n\t\treturn nil, errors.New(\"parameter name cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{name}\", url.PathEscape(name))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2022-09-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func CreateGetRetcodeLogstoreRequest() (request *GetRetcodeLogstoreRequest) {\n\trequest = &GetRetcodeLogstoreRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"ARMS\", \"2019-08-08\", \"GetRetcodeLogstore\", \"arms\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (client *ContainerClient) acquireLeaseCreateRequest(ctx context.Context, duration int32, options *ContainerClientAcquireLeaseOptions, modifiedAccessConditions *ModifiedAccessConditions) (*policy.Request, error) {\n\treq, err := runtime.NewRequest(ctx, http.MethodPut, client.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"comp\", \"lease\")\n\treqQP.Set(\"restype\", \"container\")\n\tif options != nil && options.Timeout != nil {\n\t\treqQP.Set(\"timeout\", strconv.FormatInt(int64(*options.Timeout), 10))\n\t}\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"x-ms-lease-action\"] = []string{\"acquire\"}\n\treq.Raw().Header[\"x-ms-lease-duration\"] = []string{strconv.FormatInt(int64(duration), 10)}\n\tif options != nil && options.ProposedLeaseID != nil {\n\t\treq.Raw().Header[\"x-ms-proposed-lease-id\"] = []string{*options.ProposedLeaseID}\n\t}\n\tif modifiedAccessConditions != nil && modifiedAccessConditions.IfModifiedSince != nil {\n\t\treq.Raw().Header[\"If-Modified-Since\"] = []string{(*modifiedAccessConditions.IfModifiedSince).In(gmt).Format(time.RFC1123)}\n\t}\n\tif modifiedAccessConditions != nil && modifiedAccessConditions.IfUnmodifiedSince != nil {\n\t\treq.Raw().Header[\"If-Unmodified-Since\"] = []string{(*modifiedAccessConditions.IfUnmodifiedSince).In(gmt).Format(time.RFC1123)}\n\t}\n\treq.Raw().Header[\"x-ms-version\"] = []string{\"2020-10-02\"}\n\tif options != nil && options.RequestID != nil {\n\t\treq.Raw().Header[\"x-ms-client-request-id\"] = []string{*options.RequestID}\n\t}\n\treq.Raw().Header[\"Accept\"] = []string{\"application/xml\"}\n\treturn req, nil\n}", "func (client *RoleDefinitionsClient) getByIDCreateRequest(ctx context.Context, roleID string, options *RoleDefinitionsGetByIDOptions) (*policy.Request, error) {\n\turlPath := \"/{roleId}\"\n\turlPath = strings.ReplaceAll(urlPath, \"{roleId}\", roleID)\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2018-01-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func NewGetClinicianRequest(server string, clinicId ClinicId, clinicianId ClinicianId) (*http.Request, error) {\n\tvar err error\n\n\tvar pathParam0 string\n\n\tpathParam0, err = runtime.StyleParamWithLocation(\"simple\", false, \"clinicId\", runtime.ParamLocationPath, clinicId)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar pathParam1 string\n\n\tpathParam1, err = runtime.StyleParamWithLocation(\"simple\", false, \"clinicianId\", runtime.ParamLocationPath, clinicianId)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tserverURL, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\toperationPath := fmt.Sprintf(\"/v1/clinics/%s/clinicians/%s\", pathParam0, pathParam1)\n\tif operationPath[0] == '/' {\n\t\toperationPath = \".\" + operationPath\n\t}\n\n\tqueryURL, err := serverURL.Parse(operationPath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", queryURL.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (client *Client) getCompletionsCreateRequest(ctx context.Context, body CompletionsOptions, options *GetCompletionsOptions) (*policy.Request, error) {\n\turlPath := \"completions\"\n\treq, err := runtime.NewRequest(ctx, http.MethodPost, client.formatURL(urlPath, getDeploymentID(body)))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-07-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\tif err := runtime.MarshalAsJSON(req, body); err != nil {\n\t\treturn nil, err\n\t}\n\treturn req, nil\n}", "func (client *CassandraClustersClient) statusCreateRequest(ctx context.Context, resourceGroupName string, clusterName string, options *CassandraClustersClientStatusOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DocumentDB/cassandraClusters/{clusterName}/status\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif clusterName == \"\" {\n\t\treturn nil, errors.New(\"parameter clusterName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{clusterName}\", url.PathEscape(clusterName))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-03-15-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func CharacterCreate(w http.ResponseWriter, r *http.Request) {\n\tlog.WithFields(log.Fields{\n\t\t\"time\": time.Now(),\n\t}).Info(\"Received character create request\")\n\n\tvar requestData CreateRequest\n\tbody, err := ioutil.ReadAll(io.LimitReader(r.Body, 1048576))\n\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\n\tif err != nil {\n\t\tRespondBadRequest(w, err.Error())\n\t\treturn\n\t}\n\n\tif err := r.Body.Close(); err != nil {\n\t\tRespondBadRequest(w, err.Error())\n\t\treturn\n\t}\n\n\tif err := json.Unmarshal(body, &requestData); err != nil {\n\t\tw.WriteHeader(422) // unprocessable entity\n\t\tif err := json.NewEncoder(w).Encode(err); err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\n\tcharacter, err := CreateNewCharacter(requestData)\n\tfmt.Println(requestData.ID)\n\n\tif err != nil {\n\t\tRespondBadRequest(w, err.Error())\n\t\treturn\n\t}\n\n\tw.WriteHeader(http.StatusCreated)\n\tif err := json.NewEncoder(w).Encode(character); err != nil {\n\t\tRespondBadRequest(w, err.Error())\n\t\treturn\n\t}\n\n}", "func NewCreateanewPartyCallControlRequest(server string, body CreateanewPartyCallControlJSONRequestBody) (*http.Request, error) {\n\tvar bodyReader io.Reader\n\tbuf, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbodyReader = bytes.NewReader(buf)\n\treturn NewCreateanewPartyCallControlRequestWithBody(server, \"application/json\", bodyReader)\n}", "func NewFindByDistrictRequest(server string, params *FindByDistrictParams) (*http.Request, error) {\n\tvar err error\n\n\tserverURL, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\toperationPath := fmt.Sprintf(\"/v2/appointment/sessions/public/findByDistrict\")\n\tif operationPath[0] == '/' {\n\t\toperationPath = operationPath[1:]\n\t}\n\toperationURL := url.URL{\n\t\tPath: operationPath,\n\t}\n\n\tqueryURL := serverURL.ResolveReference(&operationURL)\n\n\tqueryValues := queryURL.Query()\n\n\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", true, \"district_id\", runtime.ParamLocationQuery, params.DistrictId); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", true, \"date\", runtime.ParamLocationQuery, params.Date); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tqueryURL.RawQuery = queryValues.Encode()\n\n\treq, err := http.NewRequest(\"GET\", queryURL.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif params.AcceptLanguage != nil {\n\t\tvar headerParam0 string\n\n\t\theaderParam0, err = runtime.StyleParamWithLocation(\"simple\", false, \"Accept-Language\", runtime.ParamLocationHeader, *params.AcceptLanguage)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treq.Header.Set(\"Accept-Language\", headerParam0)\n\t}\n\n\treturn req, nil\n}", "func GetCharacterModel(w http.ResponseWriter, req *http.Request) {\n\n\t// Get session values or redirect to Login\n\tsession, err := sessions.Store.Get(req, \"session\")\n\n\tif err != nil {\n\t\tlog.Println(\"error identifying session\")\n\t\thttp.Redirect(w, req, \"/login/\", http.StatusFound)\n\t\treturn\n\t\t// in case of error\n\t}\n\n\t// Prep for user authentication\n\tsessionMap := getUserSessionValues(session)\n\n\tusername := sessionMap[\"username\"]\n\tloggedIn := sessionMap[\"loggedin\"]\n\tisAdmin := sessionMap[\"isAdmin\"]\n\n\tfmt.Println(loggedIn, isAdmin, username)\n\n\tfmt.Println(session)\n\n\t/*\n\t\tif username == \"\" {\n\t\t\thttp.Redirect(w, req, \"/\", http.StatusFound)\n\t\t\treturn\n\t\t}\n\t*/\n\n\tvars := mux.Vars(req)\n\tidString := vars[\"id\"]\n\n\tpk, err := strconv.Atoi(idString)\n\tif err != nil {\n\t\tpk = 0\n\t\tlog.Println(err)\n\t}\n\n\tcm, err := database.PKLoadCharacterModel(db, int64(pk))\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\n\tjson.NewEncoder(w).Encode(cm)\n}", "func (client *CertificateOrdersClient) getCreateRequest(ctx context.Context, resourceGroupName string, certificateOrderName string, options *CertificateOrdersClientGetOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.CertificateRegistration/certificateOrders/{certificateOrderName}\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif certificateOrderName == \"\" {\n\t\treturn nil, errors.New(\"parameter certificateOrderName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{certificateOrderName}\", url.PathEscape(certificateOrderName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2022-09-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func NewCreateanewCallControlRequest(server string, body CreateanewCallControlJSONRequestBody) (*http.Request, error) {\n\tvar bodyReader io.Reader\n\tbuf, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbodyReader = bytes.NewReader(buf)\n\treturn NewCreateanewCallControlRequestWithBody(server, \"application/json\", bodyReader)\n}", "func CreateGetMigrationSummaryRequest() (request *GetMigrationSummaryRequest) {\n\trequest = &GetMigrationSummaryRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"dataworks-public\", \"2020-05-18\", \"GetMigrationSummary\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateDescribeUserConnectionRecordsRequest() (request *DescribeUserConnectionRecordsRequest) {\n\trequest = &DescribeUserConnectionRecordsRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"ecd\", \"2020-09-30\", \"DescribeUserConnectionRecords\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateGetNamespaceListRequest() (request *GetNamespaceListRequest) {\n\trequest = &GetNamespaceListRequest{\n\t\tRoaRequest: &requests.RoaRequest{},\n\t}\n\trequest.InitWithApiInfo(\"cr\", \"2016-06-07\", \"GetNamespaceList\", \"/namespace\", \"cr\", \"openAPI\")\n\trequest.Method = requests.GET\n\treturn\n}", "func CreateDescribeGtmInstanceRequest() (request *DescribeGtmInstanceRequest) {\n\trequest = &DescribeGtmInstanceRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Alidns\", \"2015-01-09\", \"DescribeGtmInstance\", \"alidns\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateConvertInvoiceRequest() (request *ConvertInvoiceRequest) {\n\trequest = &ConvertInvoiceRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"companyreg\", \"2020-10-22\", \"ConvertInvoice\", \"companyreg\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (client *Datetimerfc1123Client) getNullCreateRequest(ctx context.Context, options *Datetimerfc1123ClientGetNullOptions) (*policy.Request, error) {\n\turlPath := \"/datetimerfc1123/null\"\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func CreateSearchInventoryRequest() (request *SearchInventoryRequest) {\n\trequest = &SearchInventoryRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"oos\", \"2019-06-01\", \"SearchInventory\", \"oos\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}" ]
[ "0.7281973", "0.718111", "0.60141575", "0.59981817", "0.53045654", "0.529162", "0.5089326", "0.50162315", "0.48275313", "0.46769378", "0.46516144", "0.45637316", "0.44809666", "0.44730073", "0.44580668", "0.44391593", "0.43980765", "0.4389381", "0.43701097", "0.43534452", "0.43337896", "0.4328505", "0.4327918", "0.43264854", "0.43217233", "0.43186057", "0.43090788", "0.42881277", "0.42724407", "0.4268772", "0.42679098", "0.42640263", "0.42332992", "0.42272177", "0.42112258", "0.42080718", "0.41916513", "0.41909823", "0.41876414", "0.4184781", "0.41827127", "0.41816914", "0.4170395", "0.41549566", "0.41536996", "0.41534084", "0.41510725", "0.41510135", "0.41462028", "0.41446942", "0.4125487", "0.41233316", "0.4110233", "0.41101548", "0.40950456", "0.40944722", "0.40814072", "0.4077255", "0.407608", "0.4072966", "0.40644348", "0.40604275", "0.40491784", "0.404907", "0.40463", "0.40446872", "0.40431553", "0.40348482", "0.4034586", "0.40203246", "0.4018101", "0.40156406", "0.4011417", "0.40107816", "0.400871", "0.4008363", "0.4004304", "0.4003515", "0.40030178", "0.40012357", "0.39968324", "0.3980903", "0.39808884", "0.39737844", "0.39666095", "0.39656043", "0.39626926", "0.39600438", "0.39598852", "0.39579946", "0.3953518", "0.39529222", "0.3945971", "0.3942968", "0.39426893", "0.3942438", "0.39387047", "0.39383534", "0.3938076", "0.3936882" ]
0.8775902
0
CreateGetOcJusticeTerminalCaseResponse creates a response to parse from GetOcJusticeTerminalCase response
func CreateGetOcJusticeTerminalCaseResponse() (response *GetOcJusticeTerminalCaseResponse) { response = &GetOcJusticeTerminalCaseResponse{ BaseResponse: &responses.BaseResponse{}, } return }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func CreateGetOcJusticeTerminalCaseRequest() (request *GetOcJusticeTerminalCaseRequest) {\n\trequest = &GetOcJusticeTerminalCaseRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"dt-oc-info\", \"2022-08-29\", \"GetOcJusticeTerminalCase\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (client *Client) GetOcJusticeTerminalCase(request *GetOcJusticeTerminalCaseRequest) (response *GetOcJusticeTerminalCaseResponse, err error) {\n\tresponse = CreateGetOcJusticeTerminalCaseResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func CreateGetOpenNLUResponse() (response *GetOpenNLUResponse) {\n\tresponse = &GetOpenNLUResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func (client *Client) GetOcJusticeTerminalCaseWithChan(request *GetOcJusticeTerminalCaseRequest) (<-chan *GetOcJusticeTerminalCaseResponse, <-chan error) {\n\tresponseChan := make(chan *GetOcJusticeTerminalCaseResponse, 1)\n\terrChan := make(chan error, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tdefer close(responseChan)\n\t\tdefer close(errChan)\n\t\tresponse, err := client.GetOcJusticeTerminalCase(request)\n\t\tif err != nil {\n\t\t\terrChan <- err\n\t\t} else {\n\t\t\tresponseChan <- response\n\t\t}\n\t})\n\tif err != nil {\n\t\terrChan <- err\n\t\tclose(responseChan)\n\t\tclose(errChan)\n\t}\n\treturn responseChan, errChan\n}", "func (client *Client) GetOcJusticeTerminalCaseWithCallback(request *GetOcJusticeTerminalCaseRequest, callback func(response *GetOcJusticeTerminalCaseResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetOcJusticeTerminalCaseResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetOcJusticeTerminalCase(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func CreateGetWsCustomizedChO2OResponse() (response *GetWsCustomizedChO2OResponse) {\n\tresponse = &GetWsCustomizedChO2OResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDescribeCustinsKernelReleaseNotesResponse() (response *DescribeCustinsKernelReleaseNotesResponse) {\n\tresponse = &DescribeCustinsKernelReleaseNotesResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetIndustryCommerceInfoResponse() (response *GetIndustryCommerceInfoResponse) {\n\tresponse = &GetIndustryCommerceInfoResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateListCasesResponse() (response *ListCasesResponse) {\n\tresponse = &ListCasesResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateStartNotaryResponse() (response *StartNotaryResponse) {\n\tresponse = &StartNotaryResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateOemSitingSelctionResponse() (response *OemSitingSelctionResponse) {\n\tresponse = &OemSitingSelctionResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetContactResponse() (response *GetContactResponse) {\n\tresponse = &GetContactResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateListCityMapAoisResponse() (response *ListCityMapAoisResponse) {\n\tresponse = &ListCityMapAoisResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetWsCustomizedChEcomContentResponse() (response *GetWsCustomizedChEcomContentResponse) {\n\tresponse = &GetWsCustomizedChEcomContentResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func createResponse(req *http.Request) *http.Response {\n\treturn &http.Response{\n\t\tStatusCode: http.StatusOK,\n\t\tRequest: req,\n\t\tHeader: make(http.Header),\n\t\tBody: ioutil.NopCloser(bytes.NewBuffer([]byte{})),\n\t}\n}", "func CreateGetTaobaoOrderResponse() (response *GetTaobaoOrderResponse) {\n\tresponse = &GetTaobaoOrderResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDescribeOssObjectDetailResponse() (response *DescribeOssObjectDetailResponse) {\n\tresponse = &DescribeOssObjectDetailResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateCreateCdpOrderResponse() (response *CreateCdpOrderResponse) {\n\tresponse = &CreateCdpOrderResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetTaxationInfoResponse() (response *GetTaxationInfoResponse) {\n\tresponse = &GetTaxationInfoResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetKeywordChEcomResponse() (response *GetKeywordChEcomResponse) {\n\tresponse = &GetKeywordChEcomResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetWsCustomizedChO2ORequest() (request *GetWsCustomizedChO2ORequest) {\n\trequest = &GetWsCustomizedChO2ORequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"alinlp\", \"2020-06-29\", \"GetWsCustomizedChO2O\", \"alinlp\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateVerifyCenResponse() (response *VerifyCenResponse) {\n\tresponse = &VerifyCenResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateSegmentSkyResponse() (response *SegmentSkyResponse) {\n\tresponse = &SegmentSkyResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDescribeIpDdosThresholdResponse() (response *DescribeIpDdosThresholdResponse) {\n\tresponse = &DescribeIpDdosThresholdResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateCreateBoxCodeResponse() (response *CreateBoxCodeResponse) {\n\tresponse = &CreateBoxCodeResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateCreateCellClusterOrderResponse() (response *CreateCellClusterOrderResponse) {\n\tresponse = &CreateCellClusterOrderResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDescribeVnKnowledgeResponse() (response *DescribeVnKnowledgeResponse) {\n\tresponse = &DescribeVnKnowledgeResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetAlarmLogResponse() (response *GetAlarmLogResponse) {\n\tresponse = &GetAlarmLogResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateCreateVSwitchResponse() (response *CreateVSwitchResponse) {\n\tresponse = &CreateVSwitchResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetServiceInputMappingResponse() (response *GetServiceInputMappingResponse) {\n\tresponse = &GetServiceInputMappingResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateBeginVnDialogueResponse() (response *BeginVnDialogueResponse) {\n\tresponse = &BeginVnDialogueResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetLniPrivateIpAddressResponse() (response *GetLniPrivateIpAddressResponse) {\n\tresponse = &GetLniPrivateIpAddressResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetArmsConsoleUrlResponse() (response *GetArmsConsoleUrlResponse) {\n\tresponse = &GetArmsConsoleUrlResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateCreateContainerInstancesResponse() (response *CreateContainerInstancesResponse) {\n\tresponse = &CreateContainerInstancesResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateCreateTrafficMirrorFilterResponse() (response *CreateTrafficMirrorFilterResponse) {\n\tresponse = &CreateTrafficMirrorFilterResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateStartK8sApplicationResponse() (response *StartK8sApplicationResponse) {\n\tresponse = &StartK8sApplicationResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDescribeCertificatesResponse() (response *DescribeCertificatesResponse) {\n\tresponse = &DescribeCertificatesResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetMigrationSummaryResponse() (response *GetMigrationSummaryResponse) {\n\tresponse = &GetMigrationSummaryResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateResponse(w *gin.Context, payload interface{}) {\n\tw.JSON(200, payload)\n}", "func CreateGetManagedRuleResponse() (response *GetManagedRuleResponse) {\n\tresponse = &GetManagedRuleResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateRevokeOperatorResponse() (response *RevokeOperatorResponse) {\n\tresponse = &RevokeOperatorResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func (c *WSCodec) CreateResponse(id interface{}, reply interface{}) interface{} {\n\treturn &jsonSuccessResponse{Version: jsonrpcVersion, Id: id, Result: reply}\n}", "func CreateGetNerCustomizedSeaEcomResponse() (response *GetNerCustomizedSeaEcomResponse) {\n\tresponse = &GetNerCustomizedSeaEcomResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDescribeUserVvTopByDayResponse() (response *DescribeUserVvTopByDayResponse) {\n\tresponse = &DescribeUserVvTopByDayResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetOpenNLURequest() (request *GetOpenNLURequest) {\n\trequest = &GetOpenNLURequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"alinlp\", \"2020-06-29\", \"GetOpenNLU\", \"alinlp\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateGetEMapResponse() (response *GetEMapResponse) {\n\tresponse = &GetEMapResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateCreateClusterResponse() (response *CreateClusterResponse) {\n\tresponse = &CreateClusterResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDescribeOrganizationSpecsResponse() (response *DescribeOrganizationSpecsResponse) {\n\tresponse = &DescribeOrganizationSpecsResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDescribeLiveDomainCertificateInfoResponse() (response *DescribeLiveDomainCertificateInfoResponse) {\n\tresponse = &DescribeLiveDomainCertificateInfoResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetChartRepositoryResponse() (response *GetChartRepositoryResponse) {\n\tresponse = &GetChartRepositoryResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func createJsonResponse(w http.ResponseWriter, req *http.Request, statusCode int, data *om.OrderedMap, error string) {\n\tif statusCode == 0 {\n\t\tstatusCode = 200\n\t}\n\n\tsuccess := true\n\tif statusCode < 200 || statusCode > 299 {\n\t\tsuccess = false\n\t}\n\n\tw.Header().Set(\"Content-authType\", \"application/json\")\n\tw.WriteHeader(statusCode)\n\n\tstatus := om.NewOrderedMap(\n\t\t&om.KV{Key: \"http\", Value: om.NewOrderedMap(\n\t\t\t&om.KV{Key: \"code\", Value: statusCode},\n\t\t\t&om.KV{Key: \"message\", Value: http.StatusText(statusCode)},\n\t\t)})\n\n\tif error != \"\" {\n\t\tstatus.Set(\"error\", error)\n\t}\n\n\tresponse := om.NewOrderedMap().\n\t\tSet(\"success\", success).\n\t\tSet(\"status\", status)\n\n\tif data != nil {\n\t\tresponse.Append(data, false)\n\t}\n\n\tjsonData, _ := json.Marshal(response)\n\tw.Write(jsonData)\n}", "func CreateGetFaceSearchUserResponse() (response *GetFaceSearchUserResponse) {\n\tresponse = &GetFaceSearchUserResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDescribeExplorerResponse() (response *DescribeExplorerResponse) {\n\tresponse = &DescribeExplorerResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDescribeAntChainConsortiumsResponse() (response *DescribeAntChainConsortiumsResponse) {\n\tresponse = &DescribeAntChainConsortiumsResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateContainerOK(t goatest.TInterface, ctx context.Context, service *goa.Service, ctrl app.ContainerController, command []string, entrypoint []string, env []string, image string, name string, sslRedirect bool, volumes []string, workingDir *string) (http.ResponseWriter, *app.GoaContainerCreateResults) {\n\t// Setup service\n\tvar (\n\t\tlogBuf bytes.Buffer\n\t\tresp interface{}\n\n\t\trespSetter goatest.ResponseSetterFunc = func(r interface{}) { resp = r }\n\t)\n\tif service == nil {\n\t\tservice = goatest.Service(&logBuf, respSetter)\n\t} else {\n\t\tlogger := log.New(&logBuf, \"\", log.Ltime)\n\t\tservice.WithLogger(goa.NewLogger(logger))\n\t\tnewEncoder := func(io.Writer) goa.Encoder { return respSetter }\n\t\tservice.Encoder = goa.NewHTTPEncoder() // Make sure the code ends up using this decoder\n\t\tservice.Encoder.Register(newEncoder, \"*/*\")\n\t}\n\n\t// Setup request context\n\trw := httptest.NewRecorder()\n\tquery := url.Values{}\n\t{\n\t\tsliceVal := command\n\t\tquery[\"command\"] = sliceVal\n\t}\n\t{\n\t\tsliceVal := entrypoint\n\t\tquery[\"entrypoint\"] = sliceVal\n\t}\n\t{\n\t\tsliceVal := env\n\t\tquery[\"env\"] = sliceVal\n\t}\n\t{\n\t\tsliceVal := []string{image}\n\t\tquery[\"image\"] = sliceVal\n\t}\n\t{\n\t\tsliceVal := []string{name}\n\t\tquery[\"name\"] = sliceVal\n\t}\n\t{\n\t\tsliceVal := []string{fmt.Sprintf(\"%v\", sslRedirect)}\n\t\tquery[\"sslRedirect\"] = sliceVal\n\t}\n\t{\n\t\tsliceVal := volumes\n\t\tquery[\"volumes\"] = sliceVal\n\t}\n\tif workingDir != nil {\n\t\tsliceVal := []string{*workingDir}\n\t\tquery[\"workingDir\"] = sliceVal\n\t}\n\tu := &url.URL{\n\t\tPath: fmt.Sprintf(\"/api/v2/container/create\"),\n\t\tRawQuery: query.Encode(),\n\t}\n\treq, err := http.NewRequest(\"GET\", u.String(), nil)\n\tif err != nil {\n\t\tpanic(\"invalid test \" + err.Error()) // bug\n\t}\n\tprms := url.Values{}\n\t{\n\t\tsliceVal := command\n\t\tprms[\"command\"] = sliceVal\n\t}\n\t{\n\t\tsliceVal := entrypoint\n\t\tprms[\"entrypoint\"] = sliceVal\n\t}\n\t{\n\t\tsliceVal := env\n\t\tprms[\"env\"] = sliceVal\n\t}\n\t{\n\t\tsliceVal := []string{image}\n\t\tprms[\"image\"] = sliceVal\n\t}\n\t{\n\t\tsliceVal := []string{name}\n\t\tprms[\"name\"] = sliceVal\n\t}\n\t{\n\t\tsliceVal := []string{fmt.Sprintf(\"%v\", sslRedirect)}\n\t\tprms[\"sslRedirect\"] = sliceVal\n\t}\n\t{\n\t\tsliceVal := volumes\n\t\tprms[\"volumes\"] = sliceVal\n\t}\n\tif workingDir != nil {\n\t\tsliceVal := []string{*workingDir}\n\t\tprms[\"workingDir\"] = sliceVal\n\t}\n\tif ctx == nil {\n\t\tctx = context.Background()\n\t}\n\tgoaCtx := goa.NewContext(goa.WithAction(ctx, \"ContainerTest\"), rw, req, prms)\n\tcreateCtx, _err := app.NewCreateContainerContext(goaCtx, req, service)\n\tif _err != nil {\n\t\te, ok := _err.(goa.ServiceError)\n\t\tif !ok {\n\t\t\tpanic(\"invalid test data \" + _err.Error()) // bug\n\t\t}\n\t\tt.Errorf(\"unexpected parameter validation error: %+v\", e)\n\t\treturn nil, nil\n\t}\n\n\t// Perform action\n\t_err = ctrl.Create(createCtx)\n\n\t// Validate response\n\tif _err != nil {\n\t\tt.Fatalf(\"controller returned %+v, logs:\\n%s\", _err, logBuf.String())\n\t}\n\tif rw.Code != 200 {\n\t\tt.Errorf(\"invalid response status code: got %+v, expected 200\", rw.Code)\n\t}\n\tvar mt *app.GoaContainerCreateResults\n\tif resp != nil {\n\t\tvar _ok bool\n\t\tmt, _ok = resp.(*app.GoaContainerCreateResults)\n\t\tif !_ok {\n\t\t\tt.Fatalf(\"invalid response media: got variable of type %T, value %+v, expected instance of app.GoaContainerCreateResults\", resp, resp)\n\t\t}\n\t\t_err = mt.Validate()\n\t\tif _err != nil {\n\t\t\tt.Errorf(\"invalid response media type: %s\", _err)\n\t\t}\n\t}\n\n\t// Return results\n\treturn rw, mt\n}", "func CreateQueryVnConversationsResponse() (response *QueryVnConversationsResponse) {\n\tresponse = &QueryVnConversationsResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateStopInstancesResponse() (response *StopInstancesResponse) {\n\tresponse = &StopInstancesResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetRetcodeLogstoreResponse() (response *GetRetcodeLogstoreResponse) {\n\tresponse = &GetRetcodeLogstoreResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateTestFlowStrategy01Response() (response *TestFlowStrategy01Response) {\n\tresponse = &TestFlowStrategy01Response{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func (c *ClientWithResponses) CreateClinicWithBodyWithResponse(ctx context.Context, contentType string, body io.Reader, reqEditors ...RequestEditorFn) (*CreateClinicResponse, error) {\n\trsp, err := c.CreateClinicWithBody(ctx, contentType, body, reqEditors...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ParseCreateClinicResponse(rsp)\n}", "func CreateDescribeUserConnectionRecordsResponse() (response *DescribeUserConnectionRecordsResponse) {\n\tresponse = &DescribeUserConnectionRecordsResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateRunContactReviewResponse() (response *RunContactReviewResponse) {\n\tresponse = &RunContactReviewResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateResponse(resultCode uint32, internalCommand []byte) ([]byte, error) {\n\t// Response frame:\n\t// - uint32 (size of response)\n\t// - []byte (response)\n\t// - uint32 (code)\n\tvar buf bytes.Buffer\n\n\tif err := binary.Write(&buf, binary.BigEndian, uint32(len(internalCommand))); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif _, err := buf.Write(internalCommand); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := binary.Write(&buf, binary.BigEndian, resultCode); err != nil {\n\t\treturn nil, err\n\t}\n\treturn buf.Bytes(), nil\n}", "func CreateConvertInvoiceResponse() (response *ConvertInvoiceResponse) {\n\tresponse = &ConvertInvoiceResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateOpenAckServiceResponse() (response *OpenAckServiceResponse) {\n\tresponse = &OpenAckServiceResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateCreateCustomCallTaggingResponse() (response *CreateCustomCallTaggingResponse) {\n\tresponse = &CreateCustomCallTaggingResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func goatsejs_http_construct_return(ovm *otto.Otto, resp *http.Response, body string) (respObj *otto.Object, err error) {\r\n\tvar (\r\n\t\theaderObj *otto.Object\r\n\t\ttrailerObj *otto.Object\r\n\t)\r\n\trespObj, _ = ovm.Object(\"({})\")\r\n\trespObj.Set(\"body\", body)\r\n\trespObj.Set(\"status\", resp.Status)\r\n\trespObj.Set(\"statusCode\", resp.StatusCode)\r\n\trespObj.Set(\"contentLength\", resp.ContentLength)\r\n\trespObj.Set(\"url\", resp.Request.URL.String())\r\n\theaderObj, _ = ovm.Object(\"({})\")\r\n\tfor k, v := range resp.Header {\r\n\t\theaderObj.Set(k, v)\r\n\t}\r\n\trespObj.Set(\"header\", headerObj)\r\n\ttrailerObj, _ = ovm.Object(\"({})\")\r\n\tfor k, v := range resp.Trailer {\r\n\t\ttrailerObj.Set(k, v)\r\n\t}\r\n\trespObj.Set(\"trailer\", trailerObj)\r\n\treturn respObj, nil\r\n}", "func CreateGetOfficePreviewURLResponse() (response *GetOfficePreviewURLResponse) {\n\tresponse = &GetOfficePreviewURLResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetClusterMetricsResponse() (response *GetClusterMetricsResponse) {\n\tresponse = &GetClusterMetricsResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func (client ThreatIntelligenceIndicatorClient) CreateResponder(resp *http.Response) (result ThreatIntelligenceInformationModel, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated),\n\t\tautorest.ByUnmarshallingJSON(&result),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}", "func encodeGetByCreteriaResponse(ctx context.Context, w http1.ResponseWriter, response interface{}) (err error) {\n\tif f, ok := response.(endpoint.Failure); ok && f.Failed() != nil {\n\t\tErrorEncoder(ctx, f.Failed(), w)\n\t\treturn nil\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\terr = json.NewEncoder(w).Encode(response)\n\treturn\n}", "func (client ConversationsClient) CreateConversationMethodResponder(resp *http.Response) (result ConversationResourceResponseType, err error) {\n err = autorest.Respond(\n resp,\n client.ByInspecting(),\n azure.WithErrorUnlessStatusCode(http.StatusOK,http.StatusCreated,http.StatusAccepted),\n autorest.ByUnmarshallingJSON(&result),\n autorest.ByClosing())\n result.Response = autorest.Response{Response: resp}\n return\n }", "func (client ThreatIntelligenceIndicatorClient) CreateIndicatorResponder(resp *http.Response) (result ThreatIntelligenceInformationModel, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated),\n\t\tautorest.ByUnmarshallingJSON(&result),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}", "func CreateGetAgentStateResponse() (response *GetAgentStateResponse) {\n\tresponse = &GetAgentStateResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateUpdateCommodityResponse() (response *UpdateCommodityResponse) {\n\tresponse = &UpdateCommodityResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetHotlineGroupDetailReportResponse() (response *GetHotlineGroupDetailReportResponse) {\n\tresponse = &GetHotlineGroupDetailReportResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDescribeIpInfoResponse() (response *DescribeIpInfoResponse) {\n\tresponse = &DescribeIpInfoResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateBatchAuditTest01Response() (response *BatchAuditTest01Response) {\n\tresponse = &BatchAuditTest01Response{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func ParseCreateClinicResponse(rsp *http.Response) (*CreateClinicResponse, error) {\n\tbodyBytes, err := io.ReadAll(rsp.Body)\n\tdefer func() { _ = rsp.Body.Close() }()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresponse := &CreateClinicResponse{\n\t\tBody: bodyBytes,\n\t\tHTTPResponse: rsp,\n\t}\n\n\tswitch {\n\tcase strings.Contains(rsp.Header.Get(\"Content-Type\"), \"json\") && rsp.StatusCode == 200:\n\t\tvar dest Clinic\n\t\tif err := json.Unmarshal(bodyBytes, &dest); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresponse.JSON200 = &dest\n\n\tcase strings.Contains(rsp.Header.Get(\"Content-Type\"), \"xml\") && rsp.StatusCode == 200:\n\t\tvar dest map[string]interface{}\n\t\tif err := xml.Unmarshal(bodyBytes, &dest); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresponse.XML200 = &dest\n\n\t}\n\n\treturn response, nil\n}", "func CreateGetRenderResultResponse() (response *GetRenderResultResponse) {\n\tresponse = &GetRenderResultResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDescribeReservedInstancesResponse() (response *DescribeReservedInstancesResponse) {\n\tresponse = &DescribeReservedInstancesResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func (client DatasetClient) CreateResponder(resp *http.Response) (result LongRunningOperationResult, err error) {\n err = autorest.Respond(\n resp,\n azure.WithErrorUnlessStatusCode(http.StatusOK,http.StatusAccepted),\n autorest.ByUnmarshallingJSON(&result),\n autorest.ByClosing())\n result.Response = autorest.Response{Response: resp}\n return\n }", "func (req *RequestMessage) CreateResponse(err error) (*ResponseMessage, error) {\n\tswitch val := req.request.(type) {\n\tcase *proto.ProduceReq:\n\t\treturn createProduceResponse(val, err)\n\tcase *proto.FetchReq:\n\t\treturn createFetchResponse(val, err)\n\tcase *proto.OffsetReq:\n\t\treturn createOffsetResponse(val, err)\n\tcase *proto.MetadataReq:\n\t\treturn createMetadataResponse(val, err)\n\tcase *proto.ConsumerMetadataReq:\n\t\treturn createConsumerMetadataResponse(val, err)\n\tcase *proto.OffsetCommitReq:\n\t\treturn createOffsetCommitResponse(val, err)\n\tcase *proto.OffsetFetchReq:\n\t\treturn createOffsetFetchResponse(val, err)\n\tcase nil:\n\t\treturn nil, fmt.Errorf(\"unsupported request API key %d\", req.kind)\n\tdefault:\n\t\t// The switch cases above must correspond exactly to the switch cases\n\t\t// in ReadRequest.\n\t\tlogrus.Panic(fmt.Sprintf(\"Kafka API key not handled: %d\", req.kind))\n\t}\n\treturn nil, nil\n}", "func CreateCoreEngineResponse() (response *CoreEngineResponse) {\n\tresponse = &CoreEngineResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func (client CertificateClient) CreateResponder(resp *http.Response) (result Certificate, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tclient.ByInspecting(),\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK),\n\t\tautorest.ByUnmarshallingJSON(&result),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}", "func CreateGetDirectoryOrFilePropertiesResponse() (response *GetDirectoryOrFilePropertiesResponse) {\n\tresponse = &GetDirectoryOrFilePropertiesResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func (s *SmartContract) createDEC(APIstub shim.ChaincodeStubInterface, args []string) sc.Response {\n\t\n\t// if len(args) != 5 {\n\t// \treturn shim.Error(\"Incorrect number of arguments. Expecting 5\")\n\t// }\n\n\t// \"args\":[\"DECID\", \"BuildingID\", \"BuildingCategory\",\"FloorArea\",\"HoursOfOccupancy\", \"EnergyConsumption\", \"MeterStartDate\", \"MetereEndDate\", \"Grade\"]\n\n\n\t// ABAC\n\tval, ok, err := cid.GetAttributeValue(APIstub, \"role\")\n\tif err !=nil {\n\t\treturn shim.Error(\"Error retriving user attributes\")\n\t}\n\t\n\tif !ok {\n\t\t//The client identity does not possess the attributes\n\t\treturn shim.Error(\"The client identity does not possess the attributes\")\n\t}\n\n\tif val != \"buildingowner\" && val != \"admin\" {\n\t\tfmt.Println(\"Attribute role : \" + val)\n\t\treturn shim.Error(\"Only building owners can create a DEC\")\n\t}\n\n\n\tclientID, _ := cid.GetID(APIstub)\n\tdecID := args[0]\n\t// buildingID := \"util.GenerateUUID()\"\n\n\t//checking whether the key exists\n\tdecAsBytes, _ := APIstub.GetState(decID)\n\tif decAsBytes != nil {\n\t\treturn shim.Error(\"Key Exist Already\")\n\t}\n\n\t// getting the object\n\targuments := make([][]byte, 2)\n\targuments[0] = []byte(\"getAsset\")\n\targuments[1] = []byte(args[1])\n\n\tlogger.Infof(\"Getting the identity of the asset\")\n\tresponse := APIstub.InvokeChaincode(\"identitycontract\", arguments, \"mychannel\")\n\n\tlogger.Infof(\"Received a response from Identity Contract \")\n\tlogger.Infof(fmt.Sprint(response.Status))\n\tlogger.Infof(fmt.Sprint(response.Payload))\n\tif response.Status != shim.OK || len(response.Payload)==0{\n\t\treturn shim.Error(\"Invalid Building ID\")\n\t}\n\n\t// logger.Infof(fmt.Sprint(response.Payload))\n\t// object := Asset{}\n\t// json.Unmarshal(response.Payload, &object)\n\t// logger.Infof(object.ID)\n\n\t\n\n\t//creating the ledger entry\n\toccupancy, err := strconv.ParseFloat(args[4], 32)\n\tenergy, err := strconv.ParseFloat(args[5], 32)\n\tif err != nil {\n\t\treturn shim.Error(\"Invalid Data Types\")\n\t}\n\n\tvar dec = DEC{DECID: decID, CID: clientID, BuildingID: args[1], Status: \"Pending\", \n\t\t\t\t\tBuildingCategory : args[2],\n\t\t\t\t\tFloorArea: args[3], \n\t\t\t\t\tHoursOfOccupancy: occupancy,\n\t\t\t\t\tEnergyConsumption: energy, \n\t\t\t\t\tMeterStartDate: args[6], \n\t\t\t\t\tMetereEndDate: args[7], \n\t\t\t\t\tGrade: args[8]}\n\n\tdecAsBytes, _ = json.Marshal(dec)\n\tAPIstub.PutState(decID, decAsBytes)\n\n\tlogger.Infof(\"Successfully Added\")\n\treturn shim.Success(decAsBytes)\n}", "func CreateDescribeParentPlatformResponse() (response *DescribeParentPlatformResponse) {\n\tresponse = &DescribeParentPlatformResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func (ccr ContainersCreateResponse) Response() *http.Response {\n\treturn ccr.rawResponse\n}", "func CreateSuccessResponse(w http.ResponseWriter, data interface{}) {\n\tif data != nil {\n\t\tbytes, err := json.Marshal(data)\n\n\t\tif err != nil {\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\tlog.Println(err)\n\t\t\treturn\n\t\t}\n\n\t\tw.Write(bytes)\n\t}\n}", "func buildExampleResponse(rules Rules, list bool, version string) string {\n\treturn buildExamplePayload(rules, Outbound, list, version)\n}", "func CreateDescribeMaskingRulesResponse() (response *DescribeMaskingRulesResponse) {\n\tresponse = &DescribeMaskingRulesResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateCreateFileSystemResponse() (response *CreateFileSystemResponse) {\n\tresponse = &CreateFileSystemResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateReleaseDirectConnectionResponse() (response *ReleaseDirectConnectionResponse) {\n\tresponse = &ReleaseDirectConnectionResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateQueryWorksResponse() (response *QueryWorksResponse) {\n\tresponse = &QueryWorksResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateCancelInstanceResponse() (response *CancelInstanceResponse) {\n\tresponse = &CancelInstanceResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateCreateAgentResponse() (response *CreateAgentResponse) {\n\tresponse = &CreateAgentResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateDeleteDegradeControlResponse() (response *DeleteDegradeControlResponse) {\n\tresponse = &DeleteDegradeControlResponse{\n\t\tBaseResponse: &responses.BaseResponse{},\n\t}\n\treturn\n}", "func CreateGetIndustryCommerceInfoRequest() (request *GetIndustryCommerceInfoRequest) {\n\trequest = &GetIndustryCommerceInfoRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"companyreg\", \"2020-10-22\", \"GetIndustryCommerceInfo\", \"companyreg\", \"openAPI\")\n\trequest.Method = requests.GET\n\treturn\n}", "func BuildResponse(data interface{}, statusCode int, msg string, tid string) Response {\n\tstatus := OK\n\n\tif statusCode >= 400 {\n\t\tstatus = ERROR\n\t}\n\n\tresponse := Response{\n\t\tMeta: Meta{\n\t\t\tStatus: status,\n\t\t\tStatusCode: statusCode,\n\t\t\tMessage: msg,\n\t\t\tTransactionID: tid,\n\t\t},\n\t\tData: data,\n\t}\n\n\treturn response\n}" ]
[ "0.77589756", "0.7157844", "0.61063325", "0.6026642", "0.59860957", "0.59467834", "0.56416875", "0.5634511", "0.556447", "0.5405703", "0.52253896", "0.5192126", "0.51881194", "0.51555777", "0.50910974", "0.50739145", "0.5035087", "0.5018713", "0.5008554", "0.499794", "0.49734196", "0.4968148", "0.4944723", "0.4943738", "0.49427813", "0.49332538", "0.49282357", "0.48901853", "0.48847622", "0.4866865", "0.48573485", "0.48321313", "0.48001522", "0.47958642", "0.4792823", "0.479264", "0.47871107", "0.4786342", "0.47841036", "0.47687307", "0.47622785", "0.47483155", "0.4738794", "0.47194412", "0.4718895", "0.47160265", "0.47051826", "0.47047096", "0.4695428", "0.46788296", "0.46771696", "0.46767196", "0.4673029", "0.4671117", "0.4649908", "0.46453193", "0.46051136", "0.45825425", "0.45659873", "0.45640856", "0.45619076", "0.45574957", "0.4551609", "0.45501503", "0.45455492", "0.45413604", "0.45385242", "0.45371372", "0.45189902", "0.4501389", "0.4484897", "0.4483944", "0.44832593", "0.4473451", "0.4470173", "0.44667828", "0.44550517", "0.44541213", "0.44471025", "0.44429523", "0.4441997", "0.44393605", "0.44311294", "0.44305167", "0.44231337", "0.4417724", "0.44138294", "0.44135532", "0.4406074", "0.4405024", "0.43989995", "0.43971083", "0.43968123", "0.43963215", "0.43935612", "0.4389361", "0.4388594", "0.43813956", "0.43782455", "0.4376731" ]
0.899438
0
GetQueryCmd adds the query commands
func GetQueryCmd(cdc *codec.Codec) *cobra.Command { profileQueryCmd := &cobra.Command{ Use: types.ModuleName, Short: "Querying commands for the profiles module", DisableFlagParsing: true, SuggestionsMinimumDistance: 2, RunE: client.ValidateCmd, } profileQueryCmd.AddCommand(flags.GetCommands( GetCmdQueryProfile(cdc), GetCmdQueryProfiles(cdc), GetCmdQueryProfileParams(cdc), GetCmdQueryDTagRequests(cdc), )...) return profileQueryCmd }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (AppModuleBasic) GetQueryCmd(cdc *codec.Codec) *cobra.Command {\n\t// return cli.GetQueryCmd(cdc)\n\tpanic(\"need to add cli.GetQueryCmd(cdc)\")\n}", "func (AppModuleBasic) GetQueryCmd(cdc *amino.Codec) *cobra.Command {\n\treturn client.GetQueryCmd(cdc)\n}", "func (amb AppModuleBasic) GetQueryCmd(cdc *codec.Codec) *cobra.Command {\n\treturn cli.GetQueryCmd(cdc)\n}", "func (AppModuleBasic) GetQueryCmd(cdc *codec.Codec) *cobra.Command {\n\treturn cli.GetQueryCmd(cdc)\n}", "func (am AppModuleBasic) GetQueryCmd() *cobra.Command {\n\t// Append local TX cmd to this if required\n\treturn am.cosmosAppModule.GetQueryCmd()\n}", "func GetQueryCmd() *cobra.Command {\n\tincentiveQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the incentive module\",\n\t}\n\n\tcmds := []*cobra.Command{\n\t\tqueryParamsCmd(),\n\t\tqueryRewardsCmd(),\n\t\tqueryRewardFactorsCmd(),\n\t}\n\n\tfor _, cmd := range cmds {\n\t\tflags.AddQueryFlagsToCmd(cmd)\n\t}\n\n\tincentiveQueryCmd.AddCommand(cmds...)\n\n\treturn incentiveQueryCmd\n}", "func (b AppModuleBasic) GetQueryCmd() *cobra.Command {\n\t// return cli.GetQueryCmd()\n\treturn nil\n}", "func GetQueryCmd() *cobra.Command {\n\tissuanceQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t}\n\n\tcmds := []*cobra.Command{\n\t\tGetCmdQueryParams(),\n\t}\n\n\tfor _, cmd := range cmds {\n\t\tflags.AddQueryFlagsToCmd(cmd)\n\t}\n\n\tissuanceQueryCmd.AddCommand(cmds...)\n\n\treturn issuanceQueryCmd\n}", "func (AppModuleBasic) GetQueryCmd() *cobra.Command {\n\treturn cli.NewQueryCmd()\n}", "func (AppModuleBasic) GetQueryCmd() *cobra.Command {\n\treturn cli.NewQueryCmd()\n}", "func GetQueryCmd(queryRoute string) *cobra.Command {\n\t// Group bifrost queries under a subcommand\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\t// this line is used by starport scaffolding # 1\n\n\tcmd.AddCommand(CmdListSendToTezosSigned())\n\tcmd.AddCommand(CmdShowSendToTezosSigned())\n\n\tcmd.AddCommand(CmdListReceivedFa12Txs())\n\tcmd.AddCommand(CmdShowReceivedFa12Txs())\n\n\tcmd.AddCommand(CmdListSendToTezos())\n\tcmd.AddCommand(CmdShowSendToTezos())\n\n\tcmd.AddCommand(CmdListReceivedTxs())\n\tcmd.AddCommand(CmdShowReceivedTxs())\n\n\treturn cmd\n}", "func GetQueryCmd(key string, cdc *codec.Codec) *cobra.Command {\n\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Deployment query commands\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tcmd.AddCommand(flags.GetCommands(\n\t\tcmdDeployments(key, cdc),\n\t\tcmdDeployment(key, cdc),\n\t\tgetGroupCmd(key, cdc),\n\t)...)\n\n\treturn cmd\n}", "func GetQueryCmd(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tqueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tqueryCmd.AddCommand(flags.GetCommands(\n\t\t// committees\n\t\tGetCmdQueryCommittee(queryRoute, cdc),\n\t\tGetCmdQueryCommittees(queryRoute, cdc),\n\t\t// proposals\n\t\tGetCmdQueryProposal(queryRoute, cdc),\n\t\tGetCmdQueryProposals(queryRoute, cdc),\n\t\t// votes\n\t\tGetCmdQueryVotes(queryRoute, cdc),\n\t\t// other\n\t\tGetCmdQueryProposer(queryRoute, cdc),\n\t\tGetCmdQueryTally(queryRoute, cdc),\n\t\tGetCmdQueryRawParams(queryRoute, cdc))...)\n\n\treturn queryCmd\n}", "func GetQueryCmd(cdc *amino.Codec) *cobra.Command {\n\tqueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the currencies module\",\n\t}\n\n\tqueryCmd.AddCommand(\n\t\tsdkClient.GetCommands(\n\t\t\tcli.GetIssue(types.ModuleName, cdc),\n\t\t\tcli.GetCurrency(types.ModuleName, cdc),\n\t\t\tcli.GetCurrencies(types.ModuleName, cdc),\n\t\t\tcli.GetWithdraw(types.ModuleName, cdc),\n\t\t\tcli.GetWithdraws(types.ModuleName, cdc),\n\t\t)...)\n\n\treturn queryCmd\n}", "func GetQueryCmd() *cobra.Command {\n\tmintingQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the minting module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tmintingQueryCmd.AddCommand(\n\t\tGetCmdQueryParams(),\n\t\tGetCmdQueryInflation(),\n\t\tGetCmdQueryMunicipalInflation(),\n\t\tGetCmdQueryAnnualProvisions(),\n\t)\n\n\treturn mintingQueryCmd\n}", "func (AppModuleBasic) GetQueryCmd() *cobra.Command {\n\treturn cli.GetQueryCmd()\n}", "func (AppModuleBasic) GetQueryCmd() *cobra.Command {\n\treturn cli.GetQueryCmd()\n}", "func GetQueryCmd(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\t// Group did queries under a subcommand\n\tdidQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tdidQueryCmd.AddCommand(\n\t\tflags.GetCommands(\n\t\t\tGetCmdDidDocumentAll(queryRoute, cdc),\n\t\t\tGetCmdVerifiableCredentialAll(queryRoute, cdc),\n\t\t)...,\n\t)\n\n\treturn didQueryCmd\n}", "func GetQueryCmd(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\t// Group pot queries under a subcommand\n\tpotQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tpotQueryCmd.AddCommand(\n\t\tflags.GetCommands(\n\t\t\tGetCmdQueryVolumeReport(queryRoute, cdc),\n\t\t)...,\n\t)\n\n\treturn potQueryCmd\n}", "func GetQueryCmd(queryRoute string) *cobra.Command {\n\t// Group clerk queries under a subcommand\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tcmd.AddCommand(\n\t\tGetStateRecord(),\n\t)\n\n\treturn cmd\n}", "func GetQueryCmd(queryRoute string) *cobra.Command {\n\t// Group gentlemint queries under a subcommand\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tAliases: []string{types.ModuleNameAlias},\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tcmd.AddCommand(CmdShowExchangeRate())\n\n\tcmd.AddCommand(CmdListLevelFee())\n\tcmd.AddCommand(CmdShowLevelFee())\n\tcmd.AddCommand(CmdListActionLevelFee())\n\tcmd.AddCommand(CmdShowActionLevelFee())\n\tcmd.AddCommand(CmdCheckFees())\n\n\tcmd.AddCommand(CmdBalances())\n\n\t// this line is used by starport scaffolding # 1\n\n\treturn cmd\n}", "func GetQueryCmd() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the upgrade module\",\n\t}\n\n\tcmd.AddCommand(\n\t\tGetCmdQueryCurrentPlan(),\n\t\tGetCmdQueryNextPlan(),\n\t)\n\n\treturn cmd\n}", "func GetQueryCmd(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\t// Group qac queries under a subcommand\n\tqacQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tqacQueryCmd.AddCommand(\n\t\tflags.GetCommands(\n\t\t\tGetCmdListQuestion(queryRoute, cdc),\n\t\t\tGetCmdGetQuestion(queryRoute, cdc),\n\t\t\tGetCmdGetOwnQuestion(queryRoute, cdc),\n\t\t\t)...,\n\t)\n\n\treturn qacQueryCmd\n}", "func GetQueryCmd(queryRoute string) *cobra.Command {\n\t// Group model queries under a subcommand\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tcmd.AddCommand(CmdShowVendorProducts())\n\tcmd.AddCommand(CmdListModel())\n\tcmd.AddCommand(CmdShowModel())\n\tcmd.AddCommand(CmdShowModelVersion())\n\tcmd.AddCommand(CmdShowModelVersions())\n\t// this line is used by starport scaffolding # 1\n\n\treturn cmd\n}", "func GetQueryCmd(storeKey string, cdc *codec.Codec) *cobra.Command {\n\toracleCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the oracle module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\toracleCmd.AddCommand(flags.GetCommands(\n\t\tGetQueryCmdParams(storeKey, cdc),\n\t\tGetQueryCmdCounts(storeKey, cdc),\n\t\tGetQueryCmdDataSource(storeKey, cdc),\n\t\tGetQueryCmdOracleScript(storeKey, cdc),\n\t\tGetQueryCmdRequest(storeKey, cdc),\n\t\tGetQueryCmdRequestSearch(storeKey, cdc),\n\t\tGetQueryCmdValidatorStatus(storeKey, cdc),\n\t\tGetQueryCmdReporters(storeKey, cdc),\n\t\tGetQueryActiveValidators(storeKey, cdc),\n\t\tGetQueryPendingRequests(storeKey, cdc),\n\t)...)\n\treturn oracleCmd\n}", "func GetQueryCmd(cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the relationships module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tcmd.AddCommand(flags.GetCommands(\n\t\tGetCmdQueryUserRelationships(cdc),\n\t\tGetCmdQueryRelationships(cdc),\n\t\tGetCmdQueryUserBlocks(cdc),\n\t)...)\n\treturn cmd\n}", "func (mc ModuleClient) GetQueryCmd() *cobra.Command {\n\t// Group gov queries under a subcommand\n\tgovQueryCmd := &cobra.Command{\n\t\tUse: \"gitService\",\n\t\tShort: \"GitService query commands\",\n\t}\n\n\tgovQueryCmd.AddCommand(client.GetCommands(\n\t\tgitServiceCmd.GetCmdListRefs(mc.moduleName, mc.cdc),\n\t)...)\n\n\treturn govQueryCmd\n}", "func GetQueryCmd() *cobra.Command {\n\treturn cli.GetQueryCmd()\n}", "func (mc ModuleClient) GetQueryCmd() *cobra.Command {\n\tstakingQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the staking module\",\n\t}\n\n\tstakingQueryCmd.AddCommand(client.GetCommands(\n\t\tstakingcli.GetCmdQueryDelegation(mc.storeKey, mc.cdc),\n\t\tstakingcli.GetCmdQueryDelegations(mc.storeKey, mc.cdc),\n\t\tstakingcli.GetCmdQueryUnbondingDelegation(mc.storeKey, mc.cdc),\n\t\tstakingcli.GetCmdQueryUnbondingDelegations(mc.storeKey, mc.cdc),\n\t\tstakingcli.GetCmdQueryRedelegation(mc.storeKey, mc.cdc),\n\t\tstakingcli.GetCmdQueryRedelegations(mc.storeKey, mc.cdc),\n\t\tstakingcli.GetCmdQueryValidator(mc.storeKey, mc.cdc),\n\t\tstakingcli.GetCmdQueryValidators(mc.storeKey, mc.cdc),\n\t\tstakingcli.GetCmdQueryValidatorDelegations(mc.storeKey, mc.cdc),\n\t\tstakingcli.GetCmdQueryValidatorUnbondingDelegations(mc.storeKey, mc.cdc),\n\t\tstakingcli.GetCmdQueryValidatorRedelegations(mc.storeKey, mc.cdc),\n\t\tstakingcli.GetCmdQueryParams(mc.storeKey, mc.cdc),\n\t\tstakingcli.GetCmdQueryPool(mc.storeKey, mc.cdc))...)\n\n\treturn stakingQueryCmd\n\n}", "func GetQueryCmd(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\t// Group supplychain queries under a subcommand\n\tsupplychainQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tsupplychainQueryCmd.AddCommand(\n\t\tflags.GetCommands(\n // this line is used by starport scaffolding # 1\n\t\t\tGetCmdListShipment(queryRoute, cdc),\n\t\t\tGetCmdGetShipment(queryRoute, cdc),\n\t\t\tGetCmdListReceipt(queryRoute, cdc),\n\t\t\tGetCmdGetReceipt(queryRoute, cdc),\n\t\t)...,\n\t)\n\n\treturn supplychainQueryCmd\n}", "func GetQueryCmd(queryRoute string) *cobra.Command {\n\t// Group pki queries under a subcommand\n\tcmd := &cobra.Command{\n\t\tUse: pkitypes.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", pkitypes.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tcmd.AddCommand(CmdListApprovedCertificates()) // TODO: use store-based index of cert Ids\n\tcmd.AddCommand(CmdShowApprovedCertificates())\n\tcmd.AddCommand(CmdListProposedCertificate()) // TODO: use store-based index of cert Ids\n\tcmd.AddCommand(CmdShowProposedCertificate())\n\tcmd.AddCommand(CmdShowChildCertificates())\n\tcmd.AddCommand(CmdListProposedCertificateRevocation()) // TODO: use store-based index of cert Ids\n\tcmd.AddCommand(CmdShowProposedCertificateRevocation())\n\tcmd.AddCommand(CmdListRevokedCertificates()) // TODO: use store-based index of cert Ids\n\tcmd.AddCommand(CmdShowRevokedCertificates())\n\tcmd.AddCommand(CmdShowApprovedRootCertificates())\n\tcmd.AddCommand(CmdShowRevokedRootCertificates())\n\tcmd.AddCommand(CmdShowApprovedCertificatesBySubject())\n\tcmd.AddCommand(CmdListRejectedCertificate())\n\tcmd.AddCommand(CmdShowRejectedCertificate())\n\tcmd.AddCommand(CmdListPkiRevocationDistributionPoint())\n\tcmd.AddCommand(CmdShowPkiRevocationDistributionPoint())\n\tcmd.AddCommand(CmdShowPkiRevocationDistributionPointsByIssuerSubjectKeyID())\n\t// this line is used by starport scaffolding # 1\n\n\treturn cmd\n}", "func GetQueryCmd(queryRoute string) *cobra.Command {\n\t// Group topup queries under a subcommand\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tcmd.AddCommand(\n\t\tGetSequenceCmd(),\n\t)\n\n\treturn cmd\n}", "func GetQueryCmd() *cobra.Command {\n\tqueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tAliases: []string{\"md\"},\n\t\tShort: \"Querying commands for the metadata module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tqueryCmd.AddCommand(\n\t\tGetMetadataParamsCmd(),\n\t\tGetMetadataByIDCmd(),\n\t\tGetMetadataGetAllCmd(),\n\t\tGetMetadataScopeCmd(),\n\t\tGetMetadataSessionCmd(),\n\t\tGetMetadataRecordCmd(),\n\t\tGetMetadataScopeSpecCmd(),\n\t\tGetMetadataContractSpecCmd(),\n\t\tGetMetadataRecordSpecCmd(),\n\t\tGetOwnershipCmd(),\n\t\tGetValueOwnershipCmd(),\n\t\tGetOSLocatorCmd(),\n\t)\n\treturn queryCmd\n}", "func GetQueryCmd(queryRoute string) *cobra.Command {\n\t// Group id queries under a subcommand\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tcmd.AddCommand(\n\t\tCmdPoolFunds(),\n\t\tCmdTrustedServiceProviders(),\n\t\tCmdGetInvites(),\n\t\tCmdGetInvite(),\n\n\t\tCmdMembership(),\n\t\tCmdMemberships(),\n\t)\n\n\treturn cmd\n}", "func GetQueryCmd(queryRoute string) *cobra.Command {\n\t// Group id queries under a subcommand\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tcmd.AddCommand(\n\t\tCmdGetGovernmentAddr(),\n\t)\n\treturn cmd\n}", "func (mc ModuleClient) GetQueryCmd() *cobra.Command {\n\t// Group assetservice queries under a subcommand\n\tassetsvcQueryCmd := &cobra.Command{\n\t\tUse: \"assetservice\",\n\t\tShort: \"Querying commands for the assetservice module\",\n\t}\n\n\tassetsvcQueryCmd.AddCommand(client.GetCommands(\n\t\tassetservicecmd.GetCmdResolveAsset(mc.storeKey, mc.cdc),\n\t\tassetservicecmd.GetCmdWhois(mc.storeKey, mc.cdc),\n\t)...)\n\n\treturn assetsvcQueryCmd\n}", "func GetQueryCmd(storeKey string, cdc *codec.Codec) *cobra.Command {\n\tburnQueryCommand := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the burn module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tburnQueryCommand.AddCommand(client.GetCommands(\n\t\tGetCommandValidated(storeKey, cdc),\n\t)...)\n\treturn burnQueryCommand\n}", "func GetQueryCmd(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tqueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the NFT module\",\n\t\tDisableFlagParsing: true,\n\t}\n\n\tqueryCmd.AddCommand(flags.GetCommands(\n\t\tGetCmdQuerySupply(queryRoute, cdc),\n\t\tGetCmdQueryOwner(queryRoute, cdc),\n\t\tGetCmdQueryCollection(queryRoute, cdc),\n\t\tGetCmdQueryDenoms(queryRoute, cdc),\n\t\tGetCmdQueryNFT(queryRoute, cdc),\n\t)...)\n\n\treturn queryCmd\n}", "func newQueryCmd() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"query\",\n\t\tAliases: []string{\"q\"},\n\t\tShort: \"Querying subcommands\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tcmd.AddCommand(\n\t\tauthcmd.GetAccountCmd(),\n\t\trpc.ValidatorCommand(),\n\t\trpc.BlockCommand(),\n\t\tauthcmd.QueryTxsByEventsCmd(),\n\t\tauthcmd.QueryTxCmd(),\n\t)\n\n\tapp.ModuleBasics.AddQueryCommands(cmd)\n\tcmd.PersistentFlags().String(flags.FlagChainID, \"\", \"The network chain ID\")\n\n\treturn cmd\n}", "func GetQueryCmd() *cobra.Command {\n\tqueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tAliases: []string{\"am\"},\n\t\tShort: \"Querying commands for the account metadata module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tqueryCmd.AddCommand(\n\t\tGetAttributeParamsCmd(),\n\t\tGetAccountAttributeCmd(),\n\t\tListAccountAttributesCmd(),\n\t\tScanAccountAttributesCmd(),\n\t)\n\n\treturn queryCmd\n}", "func GetQueryCmd() *cobra.Command {\n\tqueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the token module\",\n\t\tDisableFlagParsing: true,\n SuggestionsMinimumDistance: 2,\n RunE: client.ValidateCmd,\n\t}\n\n\tqueryCmd.AddCommand(\n\t\tGetCmdQueryParams(),\n\t\tGetCmdQueryTokens(),\n\t\tGetCmdQueryToken(),\n\t\tGetCmdQueryTokenFees(),\n\t\tGetCmdQueryBurntoken(),\n\t)\n\n\treturn queryCmd\n}", "func GetQueryCmd(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\t// Group profile queries under a subcommand\n\tprofileQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tprofileQueryCmd.AddCommand(\n\t\tflags.GetCommands(\n\t\t\tGetCmdBalance(queryRoute, cdc),\n\t\t)...,\n\t)\n\n\treturn profileQueryCmd\n}", "func GetQueryCmd() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the relationships module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tcmd.AddCommand(\n\t\tGetCmdQueryUserRelationships(),\n\t\tGetCmdQueryUserBlocks(),\n\t)\n\treturn cmd\n}", "func (mc ModuleClient) GetQueryCmd() *cobra.Command {\n\tagentQueryCmd := &cobra.Command{\n\t\tUse: \"group\",\n\t\tShort: \"Querying commands for the group module\",\n\t}\n\n\tagentQueryCmd.AddCommand(client.GetCommands(\n\t\tagentcmd.GetCmdGetGroup(mc.storeKey, mc.cdc),\n\t\tagentcmd.GetCmdGetProposal(mc.storeKey, mc.cdc),\n\t)...)\n\n\treturn agentQueryCmd\n}", "func GetQueryCmd() *cobra.Command {\n\tqueryCmd := &cobra.Command{\n\t\tUse: \"ibc-router\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t}\n\n\tqueryCmd.AddCommand(\n\t\tGetCmdParams(),\n\t)\n\n\treturn queryCmd\n}", "func QueryCmd(name string) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tSuggestionsMinimumDistance: 2,\n\t\tDisableFlagParsing: true,\n\n\t\tArgs: cobra.ExactArgs(1),\n\t\tUse: name,\n\t\tShort: \"Query commands for the ecocredit module\",\n\t\tRunE: client.ValidateCmd,\n\t}\n\tcmd.AddCommand(\n\t\tQueryClassesCmd(),\n\t\tQueryClassInfoCmd(),\n\t\tQueryBatchesCmd(),\n\t\tQueryBatchInfoCmd(),\n\t\tQueryBalanceCmd(),\n\t\tQuerySupplyCmd(),\n\t\tQueryCreditTypesCmd(),\n\t\tQueryParams(),\n\t)\n\treturn cmd\n}", "func GetCmdQueryActive(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: oracle.QueryActive,\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query the active list of Terra assets recognized by the oracle\",\n\t\tLong: strings.TrimSpace(`\nQuery the active list of Terra assets recognized by the oracle.\n\n$ terracli query oracle active\n`),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, oracle.QueryActive), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar actives oracle.QueryActiveResponse\n\t\t\tcdc.MustUnmarshalJSON(res, &actives)\n\t\t\treturn cliCtx.PrintOutput(actives)\n\t\t},\n\t}\n\n\treturn cmd\n}", "func GetQueryCmdRequest(route string, cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"request [id]\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tbz, _, err := cliCtx.Query(fmt.Sprintf(\"custom/%s/%s/%s\", route, types.QueryRequests, args[0]))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn printOutput(cliCtx, cdc, bz, &types.QueryRequestResult{})\n\t\t},\n\t}\n}", "func (c *HTTPClient) Query(cmd *Command) (Response, error) {\n\tif err := cmd.Check(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn c.exec(cmd.Name(), cmd.Params(), cmd.Body())\n}", "func GetCmdQueryDenoms(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"denoms\",\n\t\tShort: \"queries all denominations of all collections of NFTs\",\n\t\tExample: \"nft denoms\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, _, err := cliCtx.QueryWithData(\n\t\t\t\tfmt.Sprintf(\"custom/%s/%s\", queryRoute, types.QueryDenoms), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar out []string\n\t\t\tif err = cdc.UnmarshalJSON(res, &out); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n\treturn cmd\n}", "func GetCmdQueryPath(clientCtx client.Context) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"path\",\n\t\tShort: \"Query the commitment path of the running chain\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclientCtx = clientCtx.Init()\n\n\t\t\tpath := commitmenttypes.NewMerklePrefix([]byte(\"ibc\"))\n\t\t\treturn clientCtx.PrintOutput(path)\n\t\t},\n\t}\n}", "func GetCmdQueryPool(cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"pool\",\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query the current staking pool values\",\n\t\tLong: strings.TrimSpace(`Query values for amounts stored in the staking pool:\n\n$ cetcli query staking pool\n`),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, _, err := cliCtx.QueryWithData(\"custom/stakingx/pool\", nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tprintln(string(res))\n\t\t\treturn nil\n\t\t},\n\t}\n}", "func GetCmdQueryActives(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: budget.QueryActiveList,\n\t\tShort: \"Query active programs\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, budget.QueryActiveList), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar actives []budget.Program\n\t\t\tcdc.MustUnmarshalJSON(res, &actives)\n\n\t\t\tif len(actives) == 0 {\n\t\t\t\tfmt.Println(\"No active Programs found\")\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tfor _, program := range actives {\n\t\t\t\tfmt.Println(program.String())\n\t\t\t}\n\n\t\t\treturn nil\n\t\t},\n\t}\n\n\treturn cmd\n}", "func GetCmdQueryDeposit(cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"query-deposit\",\n\t\tShort: \"Query details of a deposit\",\n\t\tExample: \"iriscli gov query-deposit --proposal-id=1 --depositor=<depositor address>\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tproposalID := uint64(viper.GetInt64(flagProposalID))\n\n\t\t\tdepositorAddr, err := sdk.AccAddressFromBech32(viper.GetString(flagDepositor))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tparams := gov.QueryDepositParams{\n\t\t\t\tDepositor: depositorAddr,\n\t\t\t\tProposalID: proposalID,\n\t\t\t}\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/deposit\", protocol.GovRoute), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar deposit gov.Deposit\n\t\t\tif err := cdc.UnmarshalJSON(res, &deposit); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn cliCtx.PrintOutput(deposit)\n\t\t},\n\t}\n\n\tcmd.Flags().String(flagProposalID, \"\", \"proposalID of proposal deposited on\")\n\tcmd.Flags().String(flagDepositor, \"\", \"bech32 depositor address\")\n\tcmd.MarkFlagRequired(flagProposalID)\n\tcmd.MarkFlagRequired(flagDeposit)\n\treturn cmd\n}", "func GetCmdQueryCollection(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"collection [denom]\",\n\t\tShort: \"get all the NFTs from a given collection\",\n\t\tExample: \"nft collection <denom>\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tdenom := strings.TrimSpace(args[0])\n\t\t\tif err := types.ValidateDenom(denom); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tparams := types.NewQueryCollectionParams(denom)\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, _, err := cliCtx.QueryWithData(\n\t\t\t\tfmt.Sprintf(\"custom/%s/%s\", queryRoute, types.QueryCollection), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar out types.Collection\n\t\t\tif err = cdc.UnmarshalJSON(res, &out); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n\treturn cmd\n}", "func GetCmdQueryInflation() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"inflation\",\n\t\tShort: \"Query the current minting inflation value\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclientCtx, err := client.GetClientQueryContext(cmd)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\n\n\t\t\tparams := &types.QueryInflationRequest{}\n\t\t\tres, err := queryClient.Inflation(cmd.Context(), params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn clientCtx.PrintString(fmt.Sprintf(\"%s\\n\", res.Inflation))\n\t\t},\n\t}\n\n\tflags.AddQueryFlagsToCmd(cmd)\n\n\treturn cmd\n}", "func GetCmdQueryFeePool(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"pool\",\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query the global fee pool\",\n\t\tLong: \"pool\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\t// Query the proposal\n\t\t\tres, err := queryFeePool(cliCtx, cdc, queryRoute)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tfmt.Println(string(res))\n\t\t\treturn nil\n\t\t},\n\t}\n\n\treturn cmd\n}", "func (qd QueryDefinition) GetQuery(modifiers ...QueryModifier) string {\n\tmodifiedQuery := qd.query\n\n\tfor _, modifier := range modifiers {\n\t\tmodifiedQuery = modifier(modifiedQuery)\n\t}\n\n\treturn modifiedQuery\n}", "func GetCmdQueryParams(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: oracle.QueryParams,\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query the current Oracle params\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, oracle.QueryParams), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar params oracle.Params\n\t\t\tcdc.MustUnmarshalJSON(res, &params)\n\t\t\treturn cliCtx.PrintOutput(params)\n\t\t},\n\t}\n\n\treturn cmd\n}", "func GetCmdQueryCandidates(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: budget.QueryCandidateList,\n\t\tShort: \"Query candidate programs\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, budget.QueryActiveList), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar candidates []budget.Program\n\t\t\tcdc.MustUnmarshalJSON(res, &candidates)\n\n\t\t\tif len(candidates) == 0 {\n\t\t\t\tfmt.Println(\"No candidates Programs found\")\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tfor _, program := range candidates {\n\t\t\t\tfmt.Println(program.String())\n\t\t\t}\n\n\t\t\treturn nil\n\t\t},\n\t}\n\n\treturn cmd\n}", "func (api *OsctrlAPI) GetQuery(env, name string) (queries.DistributedQuery, error) {\n\tvar q queries.DistributedQuery\n\treqURL := fmt.Sprintf(\"%s%s%s/%s/%s\", api.Configuration.URL, APIPath, APIQueries, env, name)\n\trawQ, err := api.GetGeneric(reqURL, nil)\n\tif err != nil {\n\t\treturn q, fmt.Errorf(\"error api request - %v - %s\", err, string(rawQ))\n\t}\n\tif err := json.Unmarshal(rawQ, &q); err != nil {\n\t\treturn q, fmt.Errorf(\"can not parse body - %v\", err)\n\t}\n\treturn q, nil\n}", "func GetCmdQueryParams(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: budget.QueryParams,\n\t\tShort: \"Query the current budget params\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, budget.QueryParams), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar params budget.Params\n\t\t\tcdc.MustUnmarshalJSON(res, &params)\n\t\t\treturn cliCtx.PrintOutput(params)\n\t\t},\n\t}\n\n\treturn cmd\n}", "func GetCmdQueryClientState(clientCtx client.Context) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"state [client-id]\",\n\t\tShort: \"Query a client state\",\n\t\tLong: strings.TrimSpace(\n\t\t\tfmt.Sprintf(`Query stored client state\n\nExample:\n$ %s query ibc client state [client-id]\n\t\t`, version.ClientName),\n\t\t),\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclientCtx = clientCtx.Init()\n\n\t\t\tclientID := args[0]\n\t\t\tif strings.TrimSpace(clientID) == \"\" {\n\t\t\t\treturn errors.New(\"client ID can't be blank\")\n\t\t\t}\n\n\t\t\tprove := viper.GetBool(flags.FlagProve)\n\n\t\t\tclientStateRes, err := utils.QueryClientState(clientCtx, clientID, prove)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tclientCtx = clientCtx.WithHeight(int64(clientStateRes.ProofHeight))\n\t\t\treturn clientCtx.PrintOutput(clientStateRes)\n\t\t},\n\t}\n\tcmd.Flags().Bool(flags.FlagProve, true, \"show proofs for the query results\")\n\treturn cmd\n}", "func GetCmdQueryCommittees(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"committees\",\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query all committees\",\n\t\tExample: fmt.Sprintf(\"%s query %s committees\", version.ClientName, types.ModuleName),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\t// Query\n\t\t\tres, _, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, types.QueryCommittees), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// Decode and print result\n\t\t\tcommittees := []types.Committee{} // using empty (not nil) slice so json output returns \"[]\"\" instead of \"null\" when there's no data\n\t\t\tif err = cdc.UnmarshalJSON(res, &committees); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn cliCtx.PrintOutput(committees)\n\t\t},\n\t}\n\treturn cmd\n}", "func GetCmdQueryDenomTrace() *cobra.Command {\r\n\tcmd := &cobra.Command{\r\n\t\tUse: \"denom-trace [hash]\",\r\n\t\tShort: \"Query the denom trace info from a given trace hash\",\r\n\t\tLong: \"Query the denom trace info from a given trace hash\",\r\n\t\tExample: fmt.Sprintf(\"%s query ibc-transfer denom-trace [hash]\", version.AppName),\r\n\t\tArgs: cobra.ExactArgs(1),\r\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\r\n\t\t\tclientCtx, err := client.GetClientQueryContext(cmd)\r\n\t\t\tif err != nil {\r\n\t\t\t\treturn err\r\n\t\t\t}\r\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\r\n\r\n\t\t\treq := &types.QueryDenomTraceRequest{\r\n\t\t\t\tHash: args[0],\r\n\t\t\t}\r\n\r\n\t\t\tres, err := queryClient.DenomTrace(context.Background(), req)\r\n\t\t\tif err != nil {\r\n\t\t\t\treturn err\r\n\t\t\t}\r\n\r\n\t\t\treturn clientCtx.PrintProto(res)\r\n\t\t},\r\n\t}\r\n\r\n\tflags.AddQueryFlagsToCmd(cmd)\r\n\treturn cmd\r\n}", "func GetCmdQueryPrice(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: oracle.QueryPrice,\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query the current Luna exchange rate w.r.t an asset\",\n\t\tLong: strings.TrimSpace(`\nQuery the current exchange rate of Luna with an asset. You can find the current list of active denoms by running: terracli query oracle active\n\n$ terracli query oracle price --denom ukrw\n`),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tdenom := viper.GetString(flagDenom)\n\t\t\tif denom == \"\" {\n\t\t\t\treturn fmt.Errorf(\"--denom flag is required\")\n\t\t\t}\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s/%s\", queryRoute, oracle.QueryPrice, denom), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar price oracle.QueryPriceResponse\n\t\t\tcdc.MustUnmarshalJSON(res, &price)\n\t\t\treturn cliCtx.PrintOutput(price)\n\t\t},\n\t}\n\n\tcmd.Flags().String(flagDenom, \"\", \"target denom to get the price\")\n\n\tcmd.MarkFlagRequired(flagDenom)\n\treturn cmd\n}", "func GetCmdQueryHeader(clientCtx client.Context) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"header\",\n\t\tShort: \"Query the latest header of the running chain\",\n\t\tLong: \"Query the latest Tendermint header of the running chain\",\n\t\tExample: fmt.Sprintf(\"%s query ibc client header\", version.ClientName),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclientCtx = clientCtx.Init()\n\n\t\t\theader, height, err := utils.QueryTendermintHeader(clientCtx)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tclientCtx = clientCtx.WithHeight(height)\n\t\t\treturn clientCtx.PrintOutput(header)\n\t\t},\n\t}\n}", "func GetCmdQueryValidator() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"validator [--addr || --val-addr || --flagMoniker] \",\n\t\tShort: \"Query a validator based on address\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclientCtx := client.GetClientContextFromCmd(cmd)\n\t\t\tclientCtx, err := client.ReadQueryCommandFlags(clientCtx, cmd.Flags())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\terr = validateQueryValidatorFlags(cmd.Flags())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvalAddrStr, _ := cmd.Flags().GetString(FlagValAddr)\n\t\t\taddr, _ := cmd.Flags().GetString(FlagAddr)\n\t\t\tif valAddrStr != \"\" || addr != \"\" {\n\t\t\t\tvar valAddr sdk.ValAddress\n\t\t\t\tif addr != \"\" {\n\t\t\t\t\tbechAddr, err := sdk.AccAddressFromBech32(addr)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn errors.Wrap(err, \"invalid account address\")\n\t\t\t\t\t}\n\t\t\t\t\tvalAddr = sdk.ValAddress(bechAddr)\n\t\t\t\t} else {\n\t\t\t\t\tvalAddr, err = sdk.ValAddressFromBech32(valAddrStr)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn errors.Wrap(err, \"invalid validator address\")\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tparams := &customstakingtypes.ValidatorByAddressRequest{ValAddr: valAddr}\n\n\t\t\t\tqueryClient := customstakingtypes.NewQueryClient(clientCtx)\n\t\t\t\tres, err := queryClient.ValidatorByAddress(context.Background(), params)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\treturn clientCtx.PrintOutput(&res.Validator)\n\t\t\t}\n\n\t\t\tmoniker, _ := cmd.Flags().GetString(FlagMoniker)\n\t\t\tif moniker != \"\" {\n\t\t\t\tparams := &customstakingtypes.ValidatorByMonikerRequest{Moniker: moniker}\n\n\t\t\t\tqueryClient := customstakingtypes.NewQueryClient(clientCtx)\n\t\t\t\tres, err := queryClient.ValidatorByMoniker(context.Background(), params)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\treturn clientCtx.PrintOutput(&res.Validator)\n\t\t\t}\n\n\t\t\treturn nil\n\t\t},\n\t}\n\n\tflags.AddQueryFlagsToCmd(cmd)\n\n\tcmd.Flags().String(FlagAddr, \"\", \"the addres in AccAddress format.\")\n\tcmd.Flags().String(FlagValAddr, \"\", \"the addres in ValAddress format.\")\n\tcmd.Flags().String(FlagMoniker, \"\", \"the moniker\")\n\n\treturn cmd\n}", "func GetQueryCmdParams(route string, cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"params\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tbz, _, err := cliCtx.Query(fmt.Sprintf(\"custom/%s/%s\", route, types.QueryParams))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn printOutput(cliCtx, cdc, bz, &types.Params{})\n\t\t},\n\t}\n}", "func GetCmdQueryProposal(cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"query-proposal\",\n\t\tShort: \"Query details of a single proposal\",\n\t\tExample: \"iriscli gov query-proposal --proposal-id=1\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tproposalID := uint64(viper.GetInt64(flagProposalID))\n\n\t\t\tparams := gov.QueryProposalParams{\n\t\t\t\tProposalID: proposalID,\n\t\t\t}\n\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/proposal\", protocol.GovRoute), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar proposal gov.Proposal\n\t\t\terr = cdc.UnmarshalJSON(res, &proposal)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn cliCtx.PrintOutput(proposal)\n\t\t},\n\t}\n\n\tcmd.Flags().String(flagProposalID, \"\", \"proposalID of proposal being queried\")\n\tcmd.MarkFlagRequired(flagProposalID)\n\treturn cmd\n}", "func GetCmdQueryProposal(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"proposal [proposal-id]\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tShort: \"Query details of a single proposal\",\n\t\tExample: fmt.Sprintf(\"%s query %s proposal 2\", version.ClientName, types.ModuleName),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\t// Prepare params for querier\n\t\t\tproposalID, err := strconv.ParseUint(args[0], 10, 64)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"proposal-id %s not a valid uint\", args[0])\n\t\t\t}\n\n\t\t\tproposal, _, err := common.QueryProposalByID(cliCtx, cdc, queryRoute, proposalID)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn cliCtx.PrintOutput(proposal)\n\t\t},\n\t}\n}", "func GetCmdQueryVolumeReport(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"report [flags]\", // reporter: []byte\n\t\t//Args: cobra.RangeArgs(1, 1),\n\t\t//Short: \"Query volume report hash by reporter addr\",\n\t\tShort: \"Query volume report hash by epoch\",\n\t\tLong: strings.TrimSpace(\n\t\t\t//fmt.Sprintf(`Query volume report hash by reporter.`),\n\t\t\tfmt.Sprintf(`Query volume report hash by epoch.`),\n\t\t),\n\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tinBuf := bufio.NewReader(cmd.InOrStdin())\n\t\t\tcliCtx := context.NewCLIContextWithInput(inBuf).WithCodec(cdc)\n\t\t\t//resp, _, err := QueryVolumeReport(cliCtx, queryRoute, viper.GetString(FlagReporter))\n\t\t\tepochStr := viper.GetString(FlagEpoch)\n\t\t\tepoch, err := checkFlagEpoch(epochStr)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tresp, _, err := QueryVolumeReport(cliCtx, queryRoute, epoch)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn cliCtx.PrintOutput(resp)\n\n\t\t},\n\t}\n\t//_ = cmd.MarkFlagRequired(flags.FlagFrom)\n\t//cmd.Flags().AddFlagSet(FsReporter)\n\tcmd.Flags().AddFlagSet(FsEpoch)\n\t_ = cmd.MarkFlagRequired(FlagEpoch)\n\t//_ = cmd.MarkFlagRequired(FlagReporter)\n\n\treturn cmd\n}", "func GetCmdQueryNFT(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"token [denom] [tokenID]\",\n\t\tShort: \"query a single NFT from a collection\",\n\t\tExample: \"nft token <denom> <tokenID>\",\n\t\tArgs: cobra.ExactArgs(2),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tdenom := strings.TrimSpace(args[0])\n\t\t\tif err := types.ValidateDenom(denom); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\ttokenID := strings.TrimSpace(args[1])\n\t\t\tif err := types.ValidateTokenID(tokenID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tparams := types.NewQueryNFTParams(denom, tokenID)\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, _, err := cliCtx.QueryWithData(\n\t\t\t\tfmt.Sprintf(\"custom/%s/%s\", queryRoute, types.QueryNFT), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar out exported.NFT\n\t\t\tif err = cdc.UnmarshalJSON(res, &out); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n\treturn cmd\n}", "func (m Query) GetQuery() string {\n\treturn m.query\n}", "func GetCmdQueryVote(cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"query-vote\",\n\t\tShort: \"Query vote\",\n\t\tExample: \"iriscli gov query-vote --proposal-id=1 --voter=<voter address>\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tproposalID := uint64(viper.GetInt64(flagProposalID))\n\n\t\t\tvoterAddr, err := sdk.AccAddressFromBech32(viper.GetString(flagVoter))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tparams := gov.QueryVoteParams{\n\t\t\t\tVoter: voterAddr,\n\t\t\t\tProposalID: proposalID,\n\t\t\t}\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/vote\", protocol.GovRoute), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar vote gov.Vote\n\t\t\tif err := cdc.UnmarshalJSON(res, &vote); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn cliCtx.PrintOutput(vote)\n\t\t},\n\t}\n\n\tcmd.Flags().String(flagProposalID, \"\", \"proposalID of proposal voting on\")\n\tcmd.Flags().String(flagVoter, \"\", \"bech32 voter address\")\n\tcmd.MarkFlagRequired(flagProposalID)\n\tcmd.MarkFlagRequired(flagVoter)\n\treturn cmd\n}", "func GetCmdQueryProposals(cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"query-proposals\",\n\t\tShort: \"Query proposals with optional filters\",\n\t\tExample: \"iriscli gov query-proposals --status=Passed\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tbechDepositorAddr := viper.GetString(flagDepositor)\n\t\t\tbechVoterAddr := viper.GetString(flagVoter)\n\t\t\tstrProposalStatus := viper.GetString(flagStatus)\n\t\t\tnumLimit := uint64(viper.GetInt64(flagNumLimit))\n\n\t\t\tparams := gov.QueryProposalsParams{\n\t\t\t\tLimit: numLimit,\n\t\t\t}\n\n\t\t\tif len(bechDepositorAddr) != 0 {\n\t\t\t\tdepositorAddr, err := sdk.AccAddressFromBech32(bechDepositorAddr)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tparams.Depositor = depositorAddr\n\t\t\t}\n\n\t\t\tif len(bechVoterAddr) != 0 {\n\t\t\t\tvoterAddr, err := sdk.AccAddressFromBech32(bechVoterAddr)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tparams.Voter = voterAddr\n\t\t\t}\n\n\t\t\tvar status = \"\"\n\t\t\tif len(strProposalStatus) > 0 {\n\t\t\t\tstatus = client.NormalizeProposalStatus(strProposalStatus)\n\t\t\t\tif _, err := gov.ProposalStatusFromString(status); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t\tparams.ProposalStatus = status\n\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/proposals\", protocol.GovRoute), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar proposals gov.Proposals\n\t\t\terr = cdc.UnmarshalJSON(res, &proposals)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn cliCtx.PrintOutput(proposals)\n\t\t},\n\t}\n\n\tcmd.Flags().String(flagNumLimit, \"\", \"(optional) limit to latest [number] proposals. Defaults to all proposals\")\n\tcmd.Flags().String(flagDepositor, \"\", \"(optional) filter by proposals deposited on by depositor\")\n\tcmd.Flags().String(flagVoter, \"\", \"(optional) filter by proposals voted on by voted\")\n\tcmd.Flags().String(flagStatus, \"\", \"(optional) filter proposals by proposal status\")\n\n\treturn cmd\n}", "func GetCmdQueryDenomTraces() *cobra.Command {\r\n\tcmd := &cobra.Command{\r\n\t\tUse: \"denom-traces\",\r\n\t\tShort: \"Query the trace info for all token denominations\",\r\n\t\tLong: \"Query the trace info for all token denominations\",\r\n\t\tExample: fmt.Sprintf(\"%s query ibc-transfer denom-traces\", version.AppName),\r\n\t\tArgs: cobra.NoArgs,\r\n\t\tRunE: func(cmd *cobra.Command, _ []string) error {\r\n\t\t\tclientCtx, err := client.GetClientQueryContext(cmd)\r\n\t\t\tif err != nil {\r\n\t\t\t\treturn err\r\n\t\t\t}\r\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\r\n\r\n\t\t\tpageReq, err := client.ReadPageRequest(cmd.Flags())\r\n\t\t\tif err != nil {\r\n\t\t\t\treturn err\r\n\t\t\t}\r\n\r\n\t\t\treq := &types.QueryDenomTracesRequest{\r\n\t\t\t\tPagination: pageReq,\r\n\t\t\t}\r\n\r\n\t\t\tres, err := queryClient.DenomTraces(context.Background(), req)\r\n\t\t\tif err != nil {\r\n\t\t\t\treturn err\r\n\t\t\t}\r\n\r\n\t\t\treturn clientCtx.PrintProto(res)\r\n\t\t},\r\n\t}\r\n\tflags.AddQueryFlagsToCmd(cmd)\r\n\tflags.AddPaginationFlagsToCmd(cmd, \"denominations trace\")\r\n\r\n\treturn cmd\r\n}", "func GetCmdQueryActiveClaims(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: treasury.QueryActiveClaims,\n\t\tShort: \"Query claims that have yet to be redeemed by the treasury\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, treasury.QueryActiveClaims), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar claims types.ClaimPool\n\t\t\tcdc.MustUnmarshalBinaryLengthPrefixed(res, &claims)\n\t\t\treturn cliCtx.PrintOutput(claims)\n\t\t},\n\t}\n\n\treturn cmd\n}", "func QueryCommittedCmd(c *CommittedQuerier) *cobra.Command {\n\tchaincodeQueryCommittedCmd := &cobra.Command{\n\t\tUse: \"querycommitted\",\n\t\tShort: \"Query a committed chaincode definition by channel and name on a peer.\",\n\t\tLong: \"Query a committed chaincode definition by channel and name on a peer.\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tif c == nil {\n\t\t\t\tccInput := &ClientConnectionsInput{\n\t\t\t\t\tCommandName: cmd.Name(),\n\t\t\t\t\tEndorserRequired: true,\n\t\t\t\t\tChannelID: channelID,\n\t\t\t\t\tPeerAddresses: peerAddresses,\n\t\t\t\t\tTLSRootCertFiles: tlsRootCertFiles,\n\t\t\t\t\tConnectionProfilePath: connectionProfilePath,\n\t\t\t\t\tTLSEnabled: viper.GetBool(\"peer.tls.enabled\"),\n\t\t\t\t}\n\n\t\t\t\tcc, err := NewClientConnections(ccInput)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tcqInput := &CommittedQueryInput{\n\t\t\t\t\tChannelID: channelID,\n\t\t\t\t\tName: chaincodeName,\n\t\t\t\t}\n\n\t\t\t\tc = &CommittedQuerier{\n\t\t\t\t\tCommand: cmd,\n\t\t\t\t\tEndorserClient: cc.EndorserClients[0],\n\t\t\t\t\tInput: cqInput,\n\t\t\t\t\tSigner: cc.Signer,\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn c.Query()\n\t\t},\n\t}\n\n\tflagList := []string{\n\t\t\"channelID\",\n\t\t\"name\",\n\t\t\"peerAddresses\",\n\t\t\"tlsRootCertFiles\",\n\t\t\"connectionProfile\",\n\t}\n\tattachFlags(chaincodeQueryCommittedCmd, flagList)\n\n\treturn chaincodeQueryCommittedCmd\n}", "func (g *NgGrid) GetQuery() string {\n\treturn strings.ToLower(g.Query)\n}", "func GetQueryCmdRequestSearch(route string, cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"request-search [oracle-script-id] [calldata] [ask-count] [min-count]\",\n\t\tArgs: cobra.ExactArgs(4),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tbz, _, err := clientcmn.QuerySearchLatestRequest(route, cliCtx, args[0], args[1], args[2], args[3])\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn printOutput(cliCtx, cdc, bz, &types.QueryRequestResult{})\n\t\t},\n\t}\n}", "func (c *ToolClient) Query() *ToolQuery {\n\treturn &ToolQuery{config: c.config}\n}", "func GetCmdQueryPrevotes(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: oracle.QueryPrevotes,\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query outstanding oracle prevotes, filtered by denom and voter address.\",\n\t\tLong: strings.TrimSpace(`\nQuery outstanding oracle prevotes, filtered by denom and voter address.\n\n$ terracli query oracle prevotes --denom=\"uusd\" --validator=\"terravaloper...\"\n\nreturns oracle prevotes submitted by the validator for denom uusd \n`),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tdenom := viper.GetString(flagDenom)\n\n\t\t\t// Check voter address exists, then valids\n\t\t\tvar voterAddress sdk.ValAddress\n\n\t\t\tbechVoterAddr := viper.GetString(flagValidator)\n\t\t\tif len(bechVoterAddr) != 0 {\n\t\t\t\tvar err error\n\n\t\t\t\tvoterAddress, err = sdk.ValAddressFromBech32(bechVoterAddr)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tparams := oracle.NewQueryPrevotesParams(voterAddress, denom)\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, oracle.QueryPrevotes), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar matchingPrevotes oracle.QueryPrevotesResponse\n\t\t\tcdc.MustUnmarshalJSON(res, &matchingPrevotes)\n\n\t\t\treturn cliCtx.PrintOutput(matchingPrevotes)\n\t\t},\n\t}\n\n\tcmd.Flags().String(flagDenom, \"\", \"filter by prevotes matching the denom\")\n\tcmd.Flags().String(flagValidator, \"\", \"(optional) filter by prevotes by validator\")\n\n\tcmd.MarkFlagRequired(flagDenom)\n\n\treturn cmd\n}", "func queryDB(clnt client.Client, cmd string) (res []client.Result, err error) {\n\tq := client.NewQuery(cmd, \"\", \"\")\n\tif response, err := clnt.Query(q); err == nil {\n\t\tif response.Error() != nil {\n\t\t\treturn res, response.Error()\n\t\t}\n\t\tres = response.Results\n\t} else {\n\t\treturn res, err\n\t}\n\treturn res, nil\n}", "func GetCmdQuerySupply(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"supply [denom]\",\n\t\tShort: \"total supply of a collection or owner of NFTs\",\n\t\tExample: \"nft supply\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tvar owner sdk.AccAddress\n\t\t\tvar err error\n\n\t\t\townerStr := strings.TrimSpace(viper.GetString(FlagOwner))\n\t\t\tif len(ownerStr) > 0 {\n\t\t\t\towner, err = sdk.AccAddressFromBech32(ownerStr)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tdenom := strings.TrimSpace(args[0])\n\t\t\tif err := types.ValidateDenom(denom); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tparams := types.NewQuerySupplyParams(denom, owner)\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, _, err := cliCtx.QueryWithData(\n\t\t\t\tfmt.Sprintf(\"custom/%s/%s\", queryRoute, types.QuerySupply), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tout := binary.LittleEndian.Uint64(res)\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n\tcmd.Flags().AddFlagSet(FsQuerySupply)\n\treturn cmd\n}", "func GetCmdQueryIssuance(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: treasury.QueryIssuance + \" [denom]\",\n\t\tShort: \"Query the current issuance of the [denom] asset\",\n\t\tLong: strings.TrimSpace(`\nQuery the current issuance of the [denom] asset. \n\n$ terracli query treasury issuance --denom=\"krw\"\n`),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tdenom := viper.GetString(flagDenom)\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s/%s\", queryRoute, treasury.QueryIssuance, denom), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar issuance sdk.Int\n\t\t\tcdc.MustUnmarshalBinaryLengthPrefixed(res, &issuance)\n\t\t\treturn cliCtx.PrintOutput(issuance)\n\t\t},\n\t}\n\n\treturn cmd\n}", "func GetQueryCmdDataSource(route string, cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"data-source [id]\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tbz, _, err := cliCtx.Query(fmt.Sprintf(\"custom/%s/%s/%s\", route, types.QueryDataSources, args[0]))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn printOutput(cliCtx, cdc, bz, &types.DataSource{})\n\t\t},\n\t}\n}", "func GetCmdQueryParams() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"params\",\n\t\tShort: \"Query the current minting parameters\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclientCtx, err := client.GetClientQueryContext(cmd)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\n\n\t\t\tparams := &types.QueryParamsRequest{}\n\t\t\tres, err := queryClient.Params(cmd.Context(), params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn clientCtx.PrintProto(&res.Params)\n\t\t},\n\t}\n\n\tflags.AddQueryFlagsToCmd(cmd)\n\n\treturn cmd\n}", "func GetCmd(key string) string {\n\tresult := storage.StringDb[key]\n\treturn result\n}", "func GetCmdQueryParams(cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"params\",\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query the current staking parameters information\",\n\t\tLong: strings.TrimSpace(\n\t\t\tfmt.Sprintf(`Query values set as staking parameters.\n\nExample:\n$ %s query staking params\n`,\n\t\t\t\tversion.ClientName,\n\t\t\t),\n\t\t),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\troute := fmt.Sprintf(\"custom/%s/%s\", types.StoreKey, staking.QueryParameters)\n\t\t\tbz, _, err := cliCtx.QueryWithData(route, nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar mergedParams types.MergedParams\n\t\t\tcdc.MustUnmarshalJSON(bz, &mergedParams)\n\t\t\treturn cliCtx.PrintOutput(mergedParams)\n\t\t},\n\t}\n}", "func GetCmdQueryTally(cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"query-tally\",\n\t\tShort: \"Get the tally of a proposal vote\",\n\t\tExample: \"iriscli gov query-tally --proposal-id=4\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tproposalID := uint64(viper.GetInt64(flagProposalID))\n\n\t\t\tparams := gov.QueryTallyParams{\n\t\t\t\tProposalID: proposalID,\n\t\t\t}\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/tally\", protocol.GovRoute), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar tally gov.TallyResult\n\t\t\tif err := cdc.UnmarshalJSON(res, &tally); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn cliCtx.PrintOutput(tally)\n\t\t},\n\t}\n\n\tcmd.Flags().String(flagProposalID, \"\", \"proposalID of which proposal is being tallied\")\n\n\treturn cmd\n}", "func GetCmdQueryCurrentEpoch(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: treasury.QueryCurrentEpoch,\n\t\tShort: \"Query the current epoch number\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, treasury.QueryCurrentEpoch), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar curEpoch sdk.Int\n\t\t\tcdc.MustUnmarshalBinaryLengthPrefixed(res, &curEpoch)\n\t\t\treturn cliCtx.PrintOutput(curEpoch)\n\t\t},\n\t}\n\n\treturn cmd\n}", "func GetCmdQueryNextPlan() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"next-plan\",\n\t\tShort: \"Get the next plan\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclientCtx := client.GetClientContextFromCmd(cmd)\n\n\t\t\tparams := &types.QueryNextPlanRequest{}\n\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\n\t\t\tres, err := queryClient.NextPlan(context.Background(), params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn clientCtx.PrintProto(res)\n\t\t},\n\t}\n\n\tflags.AddQueryFlagsToCmd(cmd)\n\n\treturn cmd\n}", "func GetCmdQueryProposals(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"proposals [committee-id]\",\n\t\tShort: \"Query all proposals for a committee\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tExample: fmt.Sprintf(\"%s query %s proposals 1\", version.ClientName, types.ModuleName),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\t// Prepare params for querier\n\t\t\tcommitteeID, err := strconv.ParseUint(args[0], 10, 64)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"committee-id %s not a valid uint\", args[0])\n\t\t\t}\n\t\t\tbz, err := cdc.MarshalJSON(types.NewQueryCommitteeParams(committeeID))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// Query\n\t\t\tres, _, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, types.QueryProposals), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// Decode and print results\n\t\t\tproposals := []types.Proposal{}\n\t\t\terr = cdc.UnmarshalJSON(res, &proposals)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn cliCtx.PrintOutput(proposals)\n\t\t},\n\t}\n\treturn cmd\n}", "func (i *InfluxDAO) queryDB(cmd string) (res []client.Result, err error) {\n\tq := client.Query{\n\t\tCommand: cmd,\n\t\tDatabase: i.databaseName,\n\t}\n\tif response, err := i.c.Query(q); err == nil {\n\t\tif response.Error() != nil {\n\t\t\treturn res, response.Error()\n\t\t}\n\t\tres = response.Results\n\t} else {\n\t\treturn res, err\n\t}\n\treturn res, nil\n}", "func queryDB(clnt client.Client, database_name string,cmd string) (res []client.Result, err error) {\n\tq := client.Query{\n\t\tCommand: cmd,\n\t\tDatabase: database_name,\n\t}\n\tif response, err := clnt.Query(q); err == nil {\n\t\tif response.Error() != nil {\n\t\t\treturn res, response.Error()\n\t\t}\n\t\tres = response.Results\n\t} else {\n\t\treturn res, err\n\t}\n\treturn res, nil\n}", "func GetCmdQueryVotes(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: oracle.QueryVotes,\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query outstanding oracle votes, filtered by denom and voter address.\",\n\t\tLong: strings.TrimSpace(`\nQuery outstanding oracle votes, filtered by denom and voter address.\n\n$ terracli query oracle votes --denom=\"uusd\" --validator=\"terravaloper...\"\n\nreturns oracle votes submitted by the validator for the denom uusd \n`),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tdenom := viper.GetString(flagDenom)\n\n\t\t\t// Check voter address exists, then valids\n\t\t\tvar voterAddress sdk.ValAddress\n\n\t\t\tbechVoterAddr := viper.GetString(flagValidator)\n\t\t\tif len(bechVoterAddr) != 0 {\n\t\t\t\tvar err error\n\n\t\t\t\tvoterAddress, err = sdk.ValAddressFromBech32(bechVoterAddr)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tparams := oracle.NewQueryVotesParams(voterAddress, denom)\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, oracle.QueryVotes), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar matchingVotes oracle.QueryVotesResponse\n\t\t\tcdc.MustUnmarshalJSON(res, &matchingVotes)\n\n\t\t\treturn cliCtx.PrintOutput(matchingVotes)\n\t\t},\n\t}\n\n\tcmd.Flags().String(flagDenom, \"\", \"filter by votes matching the denom\")\n\tcmd.Flags().String(flagValidator, \"\", \"(optional) filter by votes by validator\")\n\n\tcmd.MarkFlagRequired(flagDenom)\n\n\treturn cmd\n}", "func (c *Config) GetQuery() string {\n\tif c.Query == \"\" {\n\t\treturn \"select 1\"\n\t}\n\n\treturn c.Query\n}", "func GetCmdQueryCurrentPlan() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"current-plan\",\n\t\tShort: \"Get the current plan\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclientCtx := client.GetClientContextFromCmd(cmd)\n\n\t\t\tparams := &types.QueryCurrentPlanRequest{}\n\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\n\t\t\tres, err := queryClient.CurrentPlan(context.Background(), params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn clientCtx.PrintProto(res)\n\t\t},\n\t}\n\n\tflags.AddQueryFlagsToCmd(cmd)\n\n\treturn cmd\n}", "func GetCmdQueryProfileParams(cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"parameters\",\n\t\tShort: \"Retrieve all the profile module parameters\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\troute := fmt.Sprintf(\"custom/%s/%s\", types.QuerierRoute, types.QueryParams)\n\t\t\tres, _, err := cliCtx.QueryWithData(route, nil)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"Could not find profile parameters\")\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tvar out types.Params\n\t\t\tcdc.MustUnmarshalJSON(res, &out)\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n}" ]
[ "0.7866081", "0.78498286", "0.7787561", "0.77516055", "0.7711904", "0.7710565", "0.7693822", "0.76449287", "0.76294416", "0.76294416", "0.758036", "0.7576063", "0.75644815", "0.7563028", "0.7548859", "0.75329256", "0.75329256", "0.75210536", "0.75081134", "0.7501117", "0.74893016", "0.74762183", "0.7446721", "0.74354213", "0.7428402", "0.74193585", "0.7410249", "0.74077", "0.73991156", "0.7377168", "0.73752576", "0.73715484", "0.73672146", "0.736114", "0.73110646", "0.73100746", "0.73063904", "0.7290122", "0.7283576", "0.7277844", "0.7273978", "0.7232294", "0.7216223", "0.7027309", "0.7007318", "0.6866607", "0.6512503", "0.65035707", "0.63059956", "0.6296127", "0.6200235", "0.61873215", "0.6182617", "0.6166456", "0.6090044", "0.60830706", "0.60823876", "0.60696304", "0.606876", "0.6036615", "0.6006826", "0.5996901", "0.5992661", "0.5978151", "0.5950628", "0.59419274", "0.59418803", "0.5929674", "0.58931017", "0.58805144", "0.58796716", "0.5872931", "0.5858705", "0.584281", "0.58358175", "0.58342034", "0.5827339", "0.57939076", "0.57648945", "0.5746093", "0.5731768", "0.5724997", "0.57236415", "0.5715895", "0.57097244", "0.5709205", "0.5706097", "0.57034177", "0.5701828", "0.5699386", "0.56932247", "0.5676339", "0.5673439", "0.5672723", "0.56682885", "0.5627229", "0.5626768", "0.56106025", "0.5610542", "0.5600235" ]
0.7079285
43
GetCmdQueryProfile queries a profile from the given address or dtag
func GetCmdQueryProfile(cdc *codec.Codec) *cobra.Command { return &cobra.Command{ Use: "profile [address_or_dtag]", Short: "Retrieve the profile having the specified user address or profile dtag, if any.", Args: cobra.ExactArgs(1), RunE: func(cmd *cobra.Command, args []string) error { cliCtx := context.NewCLIContext().WithCodec(cdc) route := fmt.Sprintf("custom/%s/%s/%s", types.QuerierRoute, types.QueryProfile, args[0]) res, _, err := cliCtx.QueryWithData(route, nil) if err != nil { fmt.Printf("Could not find a profile with dtag %s \n", args[0]) return nil } var out types.Profile cdc.MustUnmarshalJSON(res, &out) return cliCtx.PrintOutput(out) }, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func queryProfile(ctx sdk.Context, path []string, _ abci.RequestQuery, keeper Keeper) ([]byte, error) {\n\tif len(strings.TrimSpace(path[0])) == 0 {\n\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrInvalidRequest, \"DTag or address cannot be empty or blank\")\n\t}\n\n\taddress, err := sdk.AccAddressFromBech32(path[0])\n\tif err != nil {\n\t\taddress = keeper.GetDtagRelatedAddress(ctx, path[0])\n\t\tif address == nil {\n\t\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrInvalidRequest, fmt.Sprintf(\"No address related to this dtag: %s\", path[0]))\n\t\t}\n\n\t}\n\n\taccount, found := keeper.GetProfile(ctx, address)\n\n\tif !found {\n\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrInvalidRequest,\n\t\t\tfmt.Sprintf(\"Profile with address %s doesn't exists\", path[0]))\n\t}\n\n\tbz, err := codec.MarshalJSONIndent(keeper.Cdc, &account)\n\tif err != nil {\n\t\tpanic(\"could not marshal result to JSON\")\n\t}\n\n\treturn bz, nil\n}", "func (s *SmartContract) QueryProfile(ctx contractapi.TransactionContextInterface, carNumber string) (*Profile, error) {\n\tcarAsBytes, err := ctx.GetStub().GetState(\"profile-\" + carNumber)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to read from world state. %s\", err.Error())\n\t}\n\n\tif carAsBytes == nil {\n\t\treturn nil, fmt.Errorf(\"%s does not exist\", carNumber)\n\t}\n\n\tcar := new(Profile)\n\t_ = json.Unmarshal(carAsBytes, car)\n\n\treturn car, nil\n}", "func GetCmdQueryProfileParams(cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"parameters\",\n\t\tShort: \"Retrieve all the profile module parameters\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\troute := fmt.Sprintf(\"custom/%s/%s\", types.QuerierRoute, types.QueryParams)\n\t\t\tres, _, err := cliCtx.QueryWithData(route, nil)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"Could not find profile parameters\")\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tvar out types.Params\n\t\t\tcdc.MustUnmarshalJSON(res, &out)\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n}", "func GetCmdQueryProfiles(cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"all\",\n\t\tShort: \"Retrieve all the registered profiles.\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\troute := fmt.Sprintf(\"custom/%s/%s\", types.QuerierRoute, types.QueryProfiles)\n\t\t\tres, _, err := cliCtx.QueryWithData(route, nil)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"Could not find any profile\")\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tvar out types.Profiles\n\t\t\tcdc.MustUnmarshalJSON(res, &out)\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n}", "func GetQueryCmd(cdc *codec.Codec) *cobra.Command {\n\tprofileQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the profiles module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tprofileQueryCmd.AddCommand(flags.GetCommands(\n\t\tGetCmdQueryProfile(cdc),\n\t\tGetCmdQueryProfiles(cdc),\n\t\tGetCmdQueryProfileParams(cdc),\n\t\tGetCmdQueryDTagRequests(cdc),\n\t)...)\n\treturn profileQueryCmd\n}", "func queryProfileParams(ctx sdk.Context, _ abci.RequestQuery, keeper Keeper) ([]byte, error) {\n\tprofileParams := keeper.GetParams(ctx)\n\n\tbz, err := codec.MarshalJSONIndent(keeper.Cdc, &profileParams)\n\tif err != nil {\n\t\tpanic(\"could not marshal result to JSON\")\n\t}\n\n\treturn bz, nil\n}", "func (svc *inmemService) GetProfile(ctx context.Context, id string) (Profile, error) {\n\t// Get the Read lock from the inmemService struct\n\tsvc.mtx.RLock()\n\n\t// Immediately set up a lock release to occur when the function finishes\n\tdefer svc.mtx.RUnlock()\n\n\t// Look for the profile by the `id` function param\n\tprofile, ok := svc.profiles[id]\n\n\t// Check if the profile id was not found in the datastore\n\tif !ok {\n\n\t\t// Return an empty profile and an error informing the caller that the profile was not found\n\t\treturn Profile{}, ErrNotFound\n\n\t}\n\n\t// Return the profile to the caller and a nil error\n\treturn profile, nil\n\n}", "func GetQueryCmd(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\t// Group profile queries under a subcommand\n\tprofileQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tprofileQueryCmd.AddCommand(\n\t\tflags.GetCommands(\n\t\t\tGetCmdBalance(queryRoute, cdc),\n\t\t)...,\n\t)\n\n\treturn profileQueryCmd\n}", "func (repository *Datastore)GetProfile(username string)(*user.Person,error){\n\tperson := newUser() //initialize user.Person and will used to store profile info\n\tquery := `SELECT * FROM userRepository WHERE username = ?`\n\terr := repository.Db.Get(&person, query, username) //get person profile details\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &person, nil\n}", "func GetProfile(ctx *router.Context) {\n\tuin, _ := ctx.Keys[middleware.ComerUinContextKey].(uint64)\n\tresponse, err := service.GetComerProfile(uin)\n\tif err != nil {\n\t\tctx.ERROR(\n\t\t\trouter.ErrBuisnessError,\n\t\t\t\"wrong metamask login parameter\",\n\t\t)\n\t\treturn\n\t}\n\n\tctx.OK(response)\n}", "func GetProfile(c *fiber.Ctx) error {\n\tID := c.Query(\"id\")\n\n\tif len(ID) < 1 {\n\t\tc.Send(ID)\n\t\tc.Send(\"Parameter ID is required\")\n\t\tc.SendStatus(http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tprofile, err := database.SearchProfile(ID)\n\tif err != nil {\n\t\tc.Send(\"Error Occurred\" + err.Error())\n\t\tc.SendStatus(http.StatusBadRequest)\n\t\treturn\n\t}\n\tif err := c.JSON(profile); err != nil {\n\t\tc.Status(500).Send(err)\n\t\treturn\n\t}\n\tc.Accepts(\"application/json\")\n\tc.SendStatus(http.StatusAccepted)\n}", "func (a *DeviceAPI) GetDeviceProfile(ctx context.Context, req *api.GetDSDeviceProfileRequest) (*api.GetDSDeviceProfileResponse, error) {\n\tlogInfo := \"api/appserver_serves_ui/GetDeviceProfile org=\" + strconv.FormatInt(req.OrgId, 10)\n\n\t// verify if user is global admin\n\tu, err := devmod.NewValidator(a.st).GetUser(ctx)\n\tif err != nil {\n\t\tlog.WithError(err).Error(logInfo)\n\t\treturn &api.GetDSDeviceProfileResponse{}, status.Errorf(codes.Internal, \"unable to verify user: %s\", err.Error())\n\t}\n\t// is user is not global admin, user must have accesss to this organization\n\tif !u.IsGlobalAdmin {\n\t\tif valid, err := organization.NewValidator(a.st).ValidateOrganizationAccess(ctx, authcus.Read, req.OrgId); !valid || err != nil {\n\t\t\treturn &api.GetDSDeviceProfileResponse{}, status.Errorf(codes.Unauthenticated, \"authentication failed: %s\", err)\n\t\t}\n\t}\n\n\tdevClient := mxpcli.Global.GetM2MDeviceServiceClient()\n\n\tresp, err := devClient.GetDeviceProfile(ctx, &pb.GetDSDeviceProfileRequest{\n\t\tOrgId: req.OrgId,\n\t\tDevId: req.DevId,\n\t})\n\tif err != nil {\n\t\tlog.WithError(err).Error(logInfo)\n\t\treturn &api.GetDSDeviceProfileResponse{}, status.Errorf(codes.Unavailable, err.Error())\n\t}\n\n\treturn &api.GetDSDeviceProfileResponse{\n\t\tDevProfile: &api.DSDeviceProfile{\n\t\t\tId: resp.DevProfile.Id,\n\t\t\tDevEui: resp.DevProfile.DevEui,\n\t\t\tFkWallet: resp.DevProfile.FkWallet,\n\t\t\tMode: api.DeviceMode(resp.DevProfile.Mode),\n\t\t\tCreatedAt: resp.DevProfile.CreatedAt,\n\t\t\tLastSeenAt: resp.DevProfile.LastSeenAt,\n\t\t\tApplicationId: resp.DevProfile.ApplicationId,\n\t\t\tName: resp.DevProfile.Name,\n\t\t},\n\t}, status.Error(codes.OK, \"\")\n}", "func Profile(req *ProfileRequest, dns string) (*ProfileResponse, error) {\n\tctx := context.Background()\n\ttgt, err := url.Parse(\"http://\" + dns + \"/profile\")\n\tif err != nil {\n\t\tlog.Fatal(err.Error())\n\t}\n\tendPoint := ht.NewClient(\"POST\", tgt, encodeRequest, decodeProfileResponse).Endpoint()\n\tresp, err := endPoint(ctx, req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresponse := resp.(ProfileResponse)\n\treturn &response, nil\n}", "func queryProfiles(ctx sdk.Context, _ abci.RequestQuery, keeper Keeper) ([]byte, error) {\n\taccounts := keeper.GetProfiles(ctx)\n\n\tbz, err := codec.MarshalJSONIndent(keeper.Cdc, &accounts)\n\tif err != nil {\n\t\tpanic(\"could not marshal result to JSON\")\n\t}\n\n\treturn bz, nil\n}", "func (pu *ProfileUKM) QueryOwnerProfile() *ProfileQuery {\n\treturn (&ProfileUKMClient{config: pu.config}).QueryOwnerProfile(pu)\n}", "func (r *client) Profile(ctx context.Context, req *debugpb.ProfileRequest) (debugpb.Debug_ProfileClient, error) {\n\tresp, err := r.DebugClient.Profile(ctx, req, r.callOptions...)\n\tif err != nil {\n\t\treturn nil, ConvertGRPCError(err)\n\t}\n\treturn resp, nil\n}", "func GetProfile(_db Queryable, uid int64) (*Profile, error) {\n\tp := &Profile{}\n\terr := _db.QueryRow(`SELECT first_name, last_name, high_score, gender, \n img, birth_date, signup_date\n FROM profile WHERE uid = $1`, uid).Scan(\n\t\tp.FirstName, p.LastName, p.HighScore, p.Gender, p.Img, p.BirthDate, p.SignupDate)\n\tif err != nil {\n\t\treturn nil, err\n\t} else {\n\t\treturn p, err\n\t}\n}", "func (d *database) getProfile(person string) (profile string, err error) {\n\tquery := fmt.Sprintf(\"SELECT letter_content FROM letters WHERE opened == 1 AND letter_purpose == '%s' AND sender == '%s' ORDER BY time DESC;\", purpose.ActionProfile, person)\n\tlogger.Log.Debug(query)\n\trows, err := d.db.Query(query)\n\tif err != nil {\n\t\terr = errors.Wrap(err, \"getProfile\")\n\t\treturn\n\t}\n\tdefer rows.Close()\n\n\t// loop through rows\n\tfor rows.Next() {\n\t\terr = rows.Scan(&profile)\n\t\tif err != nil {\n\t\t\terr = errors.Wrap(err, \"getProfile\")\n\t\t\treturn\n\t\t}\n\t\tbreak\n\t}\n\n\terr = rows.Err()\n\tif err != nil {\n\t\terr = errors.Wrap(err, \"getProfile\")\n\t}\n\treturn\n}", "func dnsmasqProfile(state *state.State, n network) (string, error) {\n\trootPath := \"\"\n\tif shared.InSnap() {\n\t\trootPath = \"/var/lib/snapd/hostfs\"\n\t}\n\n\t// Render the profile.\n\tvar sb *strings.Builder = &strings.Builder{}\n\terr := dnsmasqProfileTpl.Execute(sb, map[string]interface{}{\n\t\t\"name\": DnsmasqProfileName(n),\n\t\t\"networkName\": n.Name(),\n\t\t\"varPath\": shared.VarPath(\"\"),\n\t\t\"rootPath\": rootPath,\n\t\t\"snap\": shared.InSnap(),\n\t})\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn sb.String(), nil\n}", "func (s Service) GetProfile(ctx context.Context, url string, language string) (*profile.Profile, error) {\n\tspan := s.tracer.MakeSpan(ctx, \"GetProfile\")\n\tdefer span.Finish()\n\n\t// retrive profile of target\n\tprof, err := s.repository.Users.GetProfileByURL(ctx, url)\n\tif err != nil {\n\t\ts.tracer.LogError(span, err)\n\t\treturn nil, err\n\t}\n\n\terr = s.processProfile(ctx, language, prof)\n\tif err != nil {\n\t\ts.tracer.LogError(span, err)\n\t\treturn nil, err\n\t}\n\n\treturn prof, nil\n}", "func GetServiceProfile(ctx context.Context, db sqlx.Queryer, id uuid.UUID, localOnly bool) (ServiceProfile, error) {\n\tvar sp ServiceProfile\n\trow := db.QueryRowx(`\n\t\tselect\n\t\t\tnetwork_server_id,\n\t\t\torganization_id,\n\t\t\tcreated_at,\n\t\t\tupdated_at,\n\t\t\tname\n\t\tfrom service_profile\n\t\twhere\n\t\t\tservice_profile_id = $1`,\n\t\tid,\n\t)\n\tif err := row.Err(); err != nil {\n\t\treturn sp, handlePSQLError(Select, err, \"select error\")\n\t}\n\n\terr := row.Scan(&sp.NetworkServerID, &sp.OrganizationID, &sp.CreatedAt, &sp.UpdatedAt, &sp.Name)\n\tif err != nil {\n\t\treturn sp, handlePSQLError(Scan, err, \"scan error\")\n\t}\n\n\tif localOnly {\n\t\treturn sp, nil\n\t}\n\n\tn, err := GetNetworkServer(ctx, db, sp.NetworkServerID)\n\tif err != nil {\n\t\treturn sp, errors.Wrap(err, \"get network-server errror\")\n\t}\n\n\tnsClient, err := networkserver.GetPool().Get(n.Server, []byte(n.CACert), []byte(n.TLSCert), []byte(n.TLSKey))\n\tif err != nil {\n\t\treturn sp, errors.Wrap(err, \"get network-server client error\")\n\t}\n\n\tresp, err := nsClient.GetServiceProfile(ctx, &ns.GetServiceProfileRequest{\n\t\tId: id.Bytes(),\n\t})\n\tif err != nil {\n\t\treturn sp, errors.Wrap(err, \"get service-profile error\")\n\t}\n\n\tif resp.ServiceProfile == nil {\n\t\treturn sp, errors.New(\"service_profile must not be nil\")\n\t}\n\n\tsp.ServiceProfile = *resp.ServiceProfile\n\n\treturn sp, nil\n}", "func (s *SmartContract) QueryUserProfile(ctx contractapi.TransactionContextInterface, userId string) (*UserProfile, error) {\n\tcarAsBytes, err := ctx.GetStub().GetState(\"profile-\" + userId)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to read from world state. %s\", err.Error())\n\t}\n\n\tif carAsBytes == nil {\n\t\treturn nil, fmt.Errorf(\"%s does not exist\", userId)\n\t}\n\n\tcar := new(UserProfile)\n\t_ = json.Unmarshal(carAsBytes, car)\n\n\treturn car, nil\n}", "func getProfile(name string) (*profile.Specification, error) {\n\tprofiles, err := profiles.GetKnownProfiles()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor _, p := range profiles {\n\t\tif p.Label == name {\n\t\t\treturn p, nil\n\t\t}\n\t}\n\treturn nil, nil\n\n\t// TODO(jsimsa): This function assumes the existence of a profile\n\t// server from which the profiles can be retrieved. The profile\n\t// server is a work in progress. When it exists, the commented out\n\t// code below should work.\n\t/*\n\t\tvar profile profile.Specification\n\t\t\t\tclient, err := r.NewClient()\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, verror.New(ErrOperationFailed, nil, fmt.Sprintf(\"NewClient() failed: %v\", err))\n\t\t\t\t}\n\t\t\t\tdefer client.Close()\n\t\t\t server := // TODO\n\t\t\t\tmethod := \"Specification\"\n\t\t\t\tinputs := make([]interface{}, 0)\n\t\t\t\tcall, err := client.StartCall(server + \"/\" + name, method, inputs)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, verror.New(ErrOperationFailed, nil, fmt.Sprintf(\"StartCall(%s, %q, %v) failed: %v\\n\", server + \"/\" + name, method, inputs, err))\n\t\t\t\t}\n\t\t\t\tif err := call.Finish(&profiles); err != nil {\n\t\t\t\t\treturn nil, verror.New(ErrOperationFailed, nil, fmt.Sprintf(\"Finish(%v) failed: %v\\n\", &profiles, err))\n\t\t\t\t}\n\t\treturn &profile, nil\n\t*/\n}", "func LookupConnectionProfile(ctx *pulumi.Context, args *LookupConnectionProfileArgs, opts ...pulumi.InvokeOption) (*LookupConnectionProfileResult, error) {\n\topts = internal.PkgInvokeDefaultOpts(opts)\n\tvar rv LookupConnectionProfileResult\n\terr := ctx.Invoke(\"google-native:datastream/v1alpha1:getConnectionProfile\", args, &rv, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &rv, nil\n}", "func (c *Config) GetProfile(profile string) ([]Entry, bool) {\n\tlog.Println(\"[DEBUG] GetProfile\", profile)\n\tentries, ok := c.Profiles[profile]\n\tif ok {\n\t\treturn entries, true\n\t}\n\n\treturn nil, false\n}", "func (db *DB) GetProfile(userID string) (*types.Profile, error) {\n\tsqlStatement := `\n\t\tSELECT\n\t\t\tname,\n\t\t\tusername\n\t\tFROM\n\t\t\tcoindrop_profiles\n\t\tWHERE\n\t\t\tuser_id = $1\n\t\t`\n\n\tstmt, err := db.client.Prepare(sqlStatement)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdefer stmt.Close()\n\n\trow := stmt.QueryRow(userID)\n\n\tvar name sql.NullString\n\tvar username sql.NullString\n\n\terr = row.Scan(\n\t\t&name,\n\t\t&username,\n\t)\n\tif err == sql.ErrNoRows {\n\t\treturn &types.Profile{\n\t\t\tName: \"\",\n\t\t\tUsername: \"\",\n\t\t}, nil\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &types.Profile{\n\t\tName: name.String,\n\t\tUsername: username.String,\n\t}, nil\n}", "func (dpc *deviceProfileRestClient) requestDeviceProfile(url string, ctx context.Context) (models.DeviceProfile, error) {\n\tdata, err := clients.GetRequest(url, ctx)\n\tif err != nil {\n\t\treturn models.DeviceProfile{}, err\n\t}\n\n\tdp := models.DeviceProfile{}\n\terr = json.Unmarshal(data, &dp)\n\treturn dp, err\n}", "func GetProfile() map[string]string {\n\treturn Creds.GetStringMapString(\"profiles.\" + strings.Replace(viper.GetString(\"profile\"), \".\", \"-\", -1))\n}", "func newCmdProfile() *cobra.Command {\n\toptions := newProfileOptions()\n\n\tcmd := &cobra.Command{\n\t\tUse: \"profile [flags] (--template | --open-api file | --proto file) (SERVICE)\",\n\t\tShort: \"Output service profile config for Kubernetes\",\n\t\tLong: \"Output service profile config for Kubernetes.\",\n\t\tExample: ` # Output a basic template to apply after modification.\n linkerd profile -n emojivoto --template web-svc\n\n # Generate a profile from an OpenAPI specification.\n linkerd profile -n emojivoto --open-api web-svc.swagger web-svc\n\n # Generate a profile from a protobuf definition.\n linkerd profile -n emojivoto --proto Voting.proto vote-svc\n`,\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tif options.namespace == \"\" {\n\t\t\t\toptions.namespace = pkgcmd.GetDefaultNamespace(kubeconfigPath, kubeContext)\n\t\t\t}\n\t\t\toptions.name = args[0]\n\t\t\tclusterDomain := defaultClusterDomain\n\n\t\t\terr := options.validate()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\t// performs an online profile generation and access-check to k8s cluster to extract\n\t\t\t// clusterDomain from linkerd configuration\n\t\t\tif !options.ignoreCluster {\n\t\t\t\tvar err error\n\t\t\t\tk8sAPI, err := k8s.NewAPI(kubeconfigPath, kubeContext, impersonate, impersonateGroup, 0)\n\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\t_, values, err := healthcheck.FetchCurrentConfiguration(cmd.Context(), k8sAPI, controlPlaneNamespace)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tif cd := values.ClusterDomain; cd != \"\" {\n\t\t\t\t\tclusterDomain = cd\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif options.template {\n\t\t\t\treturn profiles.RenderProfileTemplate(options.namespace, options.name, clusterDomain, os.Stdout)\n\t\t\t} else if options.openAPI != \"\" {\n\t\t\t\treturn profiles.RenderOpenAPI(options.openAPI, options.namespace, options.name, clusterDomain, os.Stdout)\n\t\t\t} else if options.proto != \"\" {\n\t\t\t\treturn profiles.RenderProto(options.proto, options.namespace, options.name, clusterDomain, os.Stdout)\n\t\t\t}\n\n\t\t\t// we should never get here\n\t\t\treturn errors.New(\"Unexpected error\")\n\t\t},\n\t}\n\n\tcmd.PersistentFlags().BoolVar(&options.template, \"template\", options.template, \"Output a service profile template\")\n\tcmd.PersistentFlags().StringVar(&options.openAPI, \"open-api\", options.openAPI, \"Output a service profile based on the given OpenAPI spec file\")\n\tcmd.PersistentFlags().StringVarP(&options.namespace, \"namespace\", \"n\", options.namespace, \"Namespace of the service\")\n\tcmd.PersistentFlags().StringVar(&options.proto, \"proto\", options.proto, \"Output a service profile based on the given Protobuf spec file\")\n\tcmd.PersistentFlags().BoolVar(&options.ignoreCluster, \"ignore-cluster\", options.ignoreCluster, \"Output a service profile through offline generation\")\n\n\treturn cmd\n}", "func (S *Service) GetProfile(ctx context.Context) (profileDataList ProfileDataList, err error) {\n\tprofileList, err := S.repo.GetProfile(ctx)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tfor _, profile := range profileList.Data {\n\t\tprofileDataList.Profile = append(profileDataList.Profile, profileToProfileData(profile))\n\t}\n\tprofileDataList.TotalData = profileList.TotalData\n\n\treturn\n}", "func GenerateSubProfileQuery() (string, error) {\n\treturn globalProbe.GenerateSubProfileQuery()\n}", "func (s *Server) GetProfile(ctx context.Context, data *jobsRPC.Empty) (*jobsRPC.CandidateProfile, error) {\n\tprofile, err := s.service.GetCandidateProfile(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn candidateProfileToCandidateProfileRPC(profile), nil\n}", "func (c *IRacing) GetProfile(ctx context.Context) (*UserProfile, error) {\n\n\tprofile := &UserProfile{}\n\n\terr := c.json(ctx, http.MethodGet, \"/membersite/member/GetMember\", nil, profile)\n\n\treturn profile, err\n}", "func profileEnvArg(cmd *kingpin.CmdClause, desc string) *string {\n\tif v := os.Getenv(\"AWS_PROFILE\"); len(v) > 0 {\n\t\treturn &v\n\t}\n\treturn cmd.Arg(\"profile\", desc).String()\n}", "func (p *profile) Get(options map[string]interface{}) (riminderResponse.ProfileGetElem, error) {\n\tquery := map[string]string{\n\t\t\"source_id\": options[\"source_id\"].(string),\n\t}\n\tAddIfNotEmptyStrMap(&query, options, \"profile_id\")\n\tAddIfNotEmptyStrMap(&query, options, \"profile_reference\")\n\n\tresp := riminderResponse.ProfileGetContainer{}\n\terr := p.client.Get(\"profile\", query, &resp)\n\tif err != nil {\n\t\treturn riminderResponse.ProfileGetElem{}, err\n\t}\n\treturn resp.Data, nil\n}", "func GetProfile(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\tuname, found := vars[\"uname\"]\n\tif !found {\n\t\tw.WriteHeader(http.StatusUnprocessableEntity)\n\t\tfmt.Fprintf(w, \"%s\", \"invalid username\")\n\t\treturn\n\t} // NOT NEEDED\n\n\t// get the username\n\ttokk := r.Header.Get(\"Token\")\n\tvar payload *token.Payload\n\tif tokk != \"\" {\n\t\tmaker, err := token.NewPasetoMaker(\"abcd1234abcd1234abcd1234abcd1234\")\n\t\tif err != nil {\n\t\t\tresponses.ERROR(w, http.StatusInternalServerError, err)\n\t\t\treturn\n\t\t}\n\t\tpayload, err = maker.VerifyToken(tokk)\n\t\tif err != nil {\n\t\t\tpayload = &token.Payload{}\n\t\t}\n\t} else {\n\t\tresponses.ERROR(w, http.StatusUnauthorized, errors.New(\"user not logged in\"))\n\t\treturn\n\t}\n\n\t// form the dto\n\tvar dto *dtos.ProfileDTO = dtos.NewProfileDTO()\n\tif payload.Username != \"\" {\n\t\tdto.LoggedIn = true // some one is there\n\t}\n\tif payload.Username == uname {\n\t\tdto.Editable = true // same user is there\n\t}\n\n\t// start the database\n\tdatabase, err := db.Connect()\n\tif err != nil {\n\t\tresponses.ERROR(w, http.StatusInternalServerError, err)\n\t\treturn\n\t}\n\tdefer database.Close()\n\n\t// search the database\n\tvar user schema.User\n\terr = database.Model(&schema.User{}).Where(\"username = ?\", uname).Find(&user).Error\n\tswitch err {\n\tcase nil:\n\tcase gorm.ErrRecordNotFound:\n\t\tresponses.ERROR(w, http.StatusNotFound, gorm.ErrRecordNotFound)\n\t\treturn\n\tdefault:\n\t\tresponses.ERROR(w, http.StatusInternalServerError, err)\n\t\treturn\n\t}\n\n\t// form the dto now\n\tdto.FromSchema(&user)\n\n\t// more db calls to populate drafts, cards, bookmarks\n\tif dto.Editable {\n\t\tdto.Drafts, err = crud.GetDraftsFromDB(database, dto.Username)\n\t\tif err != nil {\n\t\t\tlog.Println(\"coudnot fetch drafts, cards or bookmarks from db\")\n\t\t}\n\t}\n\tdto.Cards, err = crud.GetCardsFromDB(database, dto.Username)\n\tif dto.Editable {\n\t\tdto.Bookmarks, err = crud.GetBookMarkCardsFromDB(database, dto.Username)\n\t\tif err != nil {\n\t\t\tlog.Println(\"coudnot fetch drafts, cards or bookmarks from db\")\n\t\t}\n\t}\n\n\t// return the response\n\tresponses.JSON(w, http.StatusAccepted, *dto)\n\treturn\n\n}", "func (e Endpoints) GetProfile(ctx context.Context, ID string) (Profile, error) {\n\n\t// Initialize a variable of type GetProfileRequest and set its ID field to the ID value passed to the function\n\trequest := getProfileRequest{ProfileID: ID}\n\n\t// Initialize two variable made by calling `e`'s GetProfileEndpoint field value.\n\t// This function will return a response value of type interface{} and an error\n\tresponse, err := e.GetProfileEndpoint(ctx, request)\n\n\t// Check to see if the endpoint call returned a non-nil error value.\n\tif err != nil {\n\n\t\t// Return an empty value Profile and the error from the endpoint call\n\t\treturn Profile{}, err\n\t}\n\n\t// Initialize a variable called `resp` which will be the result of asserting the response (`response` interface{})\n\t//from GetProfileEndpoint as a GetProfileResponse value\n\tresp := response.(getProfileResponse)\n\n\t// Return the Profile\n\treturn resp.Profile, resp.Err\n\n}", "func GetQueryCmd() *cobra.Command {\n\tqueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tAliases: []string{\"am\"},\n\t\tShort: \"Querying commands for the account metadata module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tqueryCmd.AddCommand(\n\t\tGetAttributeParamsCmd(),\n\t\tGetAccountAttributeCmd(),\n\t\tListAccountAttributesCmd(),\n\t\tScanAccountAttributesCmd(),\n\t)\n\n\treturn queryCmd\n}", "func (db *DB) GetProfile(feedID string) (*Profile, error) {\n\tvar ps []Profile\n\tdb.e.NewQuery(\"Profile\").Filter(\"FeedID =\", feedID).GetAll(&ps)\n\n\tif len(ps) == 0 {\n\t\treturn &Profile{}, nil\n\t}\n\treturn &ps[0], nil\n}", "func (bg *Backgrounder) GetProfile(key string) time.Duration {\n\treturn bg.profile[key]\n}", "func GetCmdQueryProposal(cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"query-proposal\",\n\t\tShort: \"Query details of a single proposal\",\n\t\tExample: \"iriscli gov query-proposal --proposal-id=1\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tproposalID := uint64(viper.GetInt64(flagProposalID))\n\n\t\t\tparams := gov.QueryProposalParams{\n\t\t\t\tProposalID: proposalID,\n\t\t\t}\n\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/proposal\", protocol.GovRoute), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar proposal gov.Proposal\n\t\t\terr = cdc.UnmarshalJSON(res, &proposal)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn cliCtx.PrintOutput(proposal)\n\t\t},\n\t}\n\n\tcmd.Flags().String(flagProposalID, \"\", \"proposalID of proposal being queried\")\n\tcmd.MarkFlagRequired(flagProposalID)\n\treturn cmd\n}", "func (m *MicrosoftManagedDesktop) GetProfile()(*string) {\n val, err := m.GetBackingStore().Get(\"profile\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (s *HighAvailabilityService) GetAvailabilityProfileCommand(input *GetAvailabilityProfileCommandInput) (output *models.AvailabilityProfileView, resp *http.Response, err error) {\n\tpath := \"/highAvailability/availabilityProfiles/{id}\"\n\tpath = strings.Replace(path, \"{id}\", input.Id, -1)\n\n\top := &request.Operation{\n\t\tName: \"GetAvailabilityProfileCommand\",\n\t\tHTTPMethod: \"GET\",\n\t\tHTTPPath: path,\n\t\tQueryParams: map[string]string{},\n\t}\n\toutput = &models.AvailabilityProfileView{}\n\treq := s.newRequest(op, nil, output)\n\n\tif req.Send() == nil {\n\t\treturn output, req.HTTPResponse, nil\n\t}\n\treturn nil, req.HTTPResponse, req.Error\n}", "func GetCmdQueryOwner(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"owner [address]\",\n\t\tShort: \"get the NFTs owned by an account address\",\n\t\tExample: \"nft owner <address> --denom=<denom>\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\taddress, err := sdk.AccAddressFromBech32(args[0])\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tdenom := viper.GetString(FlagDenom)\n\t\t\tparams := types.NewQueryOwnerParams(denom, address)\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, _, err := cliCtx.QueryWithData(\n\t\t\t\tfmt.Sprintf(\"custom/%s/%s\", queryRoute, types.QueryOwner), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar out types.Owner\n\t\t\terr = cdc.UnmarshalJSON(res, &out)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n\tcmd.Flags().AddFlagSet(FsQueryOwner)\n\treturn cmd\n}", "func GetProfile(alias, domain string, tracing bool) (response *Response, err error) {\r\n\r\n\t// Set the url for the request\r\n\treqURL := fmt.Sprintf(\"%su?paymail=%s@%s\", Network, alias, domain)\r\n\r\n\t// Create a Client and start the request\r\n\tclient := resty.New().SetTimeout(defaultGetTimeout * time.Second)\r\n\tvar resp *resty.Response\r\n\treq := client.R().SetHeader(\"User-Agent\", UserAgent)\r\n\tif tracing {\r\n\t\treq.EnableTrace()\r\n\t}\r\n\tif resp, err = req.Get(reqURL); err != nil {\r\n\t\treturn\r\n\t}\r\n\r\n\t// Start the response\r\n\tresponse = new(Response)\r\n\r\n\t// Tracing enabled?\r\n\tif tracing {\r\n\t\tresponse.Tracing = resp.Request.TraceInfo()\r\n\t}\r\n\r\n\t// Test for a successful status code\r\n\tresponse.StatusCode = resp.StatusCode()\r\n\tif response.StatusCode != http.StatusOK && response.StatusCode != http.StatusNotModified {\r\n\t\tif response.StatusCode != http.StatusNotFound {\r\n\t\t\terr = fmt.Errorf(\"bad response from powping: %d\", response.StatusCode)\r\n\t\t}\r\n\r\n\t\treturn\r\n\t}\r\n\r\n\t// No result\r\n\tif string(resp.Body()) == \"null\" {\r\n\t\treturn\r\n\t}\r\n\r\n\t// Decode the body of the response\r\n\terr = json.Unmarshal(resp.Body(), &response.Profile)\r\n\r\n\treturn\r\n}", "func (s *SmartContract) QueryAllProfiles(ctx contractapi.TransactionContextInterface) ([]QueryProfileResult, error) {\n\tstartKey := \"\"\n\tendKey := \"\"\n\n\tresultsIterator, err := ctx.GetStub().GetStateByRange(startKey, endKey)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer resultsIterator.Close()\n\n\tresults := []QueryProfileResult{}\n\n\tfor resultsIterator.HasNext() {\n\t\tqueryResponse, err := resultsIterator.Next()\n\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif strings.Contains(queryResponse.Key, \"profile-\"){\n\t\t\tprofile := new(UserProfile)\n\t\t\t_ = json.Unmarshal(queryResponse.Value, profile)\n\n\t\t\tqueryResult := QueryProfileResult{Key: queryResponse.Key, Record: profile}\n\t\t\tresults = append(results, queryResult)\n\t\t}\n\t\t\n\t}\n\n\treturn results, nil\n}", "func (mpRepo *ManageProfileRepository) Profile(id uint) (*entity.Profile, []error) {\n\tprfs := entity.Profile{}\n\terrs := mpRepo.conn.First(&prfs, id).GetErrors()\n\tif len(errs) > 0 {\n\t\treturn nil, errs\n\t}\n\treturn &prfs, errs\n}", "func (svc *inmemService) GetAddress(ctx context.Context, profileID string, addressID string) (Address, error) {\n\n\t// Get a Read Lock on the svc for atomic read access to the datastore\n\tsvc.mtx.RLock()\n\n\t// Immediately set up a lock release to occur when the function finishes\n\tdefer svc.mtx.RUnlock()\n\n\t// Check the data store to make sure the requested profile exists and set\n\tprofile, ok := svc.profiles[profileID]\n\n\t// If no entry for the profile was fund in the datastore\n\tif !ok {\n\n\t\t// Return an empty valued Address and an error informing the caller that no profile was found with the provided ID.\n\t\treturn Address{}, ErrNotFound\n\t}\n\n\t// Loop through each address attached to the found profile\n\tfor _, address := range profile.Addresses {\n\n\t\t// Check to see if the current address's ID matches the addressID passed in\n\t\tif address.ID == addressID {\n\n\t\t\t// Return that address and a nil error for a value\n\t\t\treturn address, nil\n\t\t}\n\t}\n\n\t// Return an empty Address value and a not found error since we were unable to find the specified address.\n\treturn Address{}, ErrNotFound\n}", "func runOperationDevicesGetDevicesApsProfiles(cmd *cobra.Command, args []string) error {\n\tappCli, err := makeClient(cmd, args)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// retrieve flag values from cmd and fill params\n\tparams := devices.NewGetDevicesApsProfilesParams()\n\tif dryRun {\n\n\t\tlogDebugf(\"dry-run flag specified. Skip sending request.\")\n\t\treturn nil\n\t}\n\t// make request and then print result\n\tmsgStr, err := parseOperationDevicesGetDevicesApsProfilesResult(appCli.Devices.GetDevicesApsProfiles(params, nil))\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !debug {\n\n\t\tfmt.Println(msgStr)\n\t}\n\treturn nil\n}", "func GetCmdQueryProposal(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"proposal [proposal-id]\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tShort: \"Query details of a single proposal\",\n\t\tExample: fmt.Sprintf(\"%s query %s proposal 2\", version.ClientName, types.ModuleName),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\t// Prepare params for querier\n\t\t\tproposalID, err := strconv.ParseUint(args[0], 10, 64)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"proposal-id %s not a valid uint\", args[0])\n\t\t\t}\n\n\t\t\tproposal, _, err := common.QueryProposalByID(cliCtx, cdc, queryRoute, proposalID)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn cliCtx.PrintOutput(proposal)\n\t\t},\n\t}\n}", "func (s *EmptyStore) ProfileGet(id string) (*storagepb.Profile, error) {\n\treturn nil, fmt.Errorf(\"Profile not found\")\n}", "func GetCmdQueryPool(cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"pool\",\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query the current staking pool values\",\n\t\tLong: strings.TrimSpace(`Query values for amounts stored in the staking pool:\n\n$ cetcli query staking pool\n`),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, _, err := cliCtx.QueryWithData(\"custom/stakingx/pool\", nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tprintln(string(res))\n\t\t\treturn nil\n\t\t},\n\t}\n}", "func (a *API) GetProfileSearch(ctx context.Context, search *ProfileSearchResp) (*ProfileResp, error) {\n\tvar resp ProfileResp\n\tif err := a.Get(ctx, search.ProfilesURL, &resp); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func GetCmdQueryDeposit(cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"query-deposit\",\n\t\tShort: \"Query details of a deposit\",\n\t\tExample: \"iriscli gov query-deposit --proposal-id=1 --depositor=<depositor address>\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tproposalID := uint64(viper.GetInt64(flagProposalID))\n\n\t\t\tdepositorAddr, err := sdk.AccAddressFromBech32(viper.GetString(flagDepositor))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tparams := gov.QueryDepositParams{\n\t\t\t\tDepositor: depositorAddr,\n\t\t\t\tProposalID: proposalID,\n\t\t\t}\n\t\t\tbz, err := cdc.MarshalJSON(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/deposit\", protocol.GovRoute), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar deposit gov.Deposit\n\t\t\tif err := cdc.UnmarshalJSON(res, &deposit); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn cliCtx.PrintOutput(deposit)\n\t\t},\n\t}\n\n\tcmd.Flags().String(flagProposalID, \"\", \"proposalID of proposal deposited on\")\n\tcmd.Flags().String(flagDepositor, \"\", \"bech32 depositor address\")\n\tcmd.MarkFlagRequired(flagProposalID)\n\tcmd.MarkFlagRequired(flagDeposit)\n\treturn cmd\n}", "func (d *Daytype) Profile() ProfileQS {\n\treturn ProfileQS{}.DaytypeEq(d)\n}", "func (c Client) GetProfile(id int64) (Profile, error) {\n\tvar p Profile\n\n\turl := fmt.Sprintf(profileURL, c.api, id, c.token)\n\tresp, err := http.Get(url)\n\tif err != nil {\n\t\treturn p, err\n\t}\n\tdefer func() {\n\t\t_ = resp.Body.Close()\n\t}()\n\n\tcontent, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn p, err\n\t}\n\n\tif err = json.Unmarshal(content, &p); err != nil {\n\t\treturn p, err\n\t}\n\n\treturn p, checkError(bytes.NewReader(content))\n}", "func (t *SimpleChaincode) QueryPrivate(stub shim.ChaincodeStubInterface, args []string) pb.Response{\n\tif len(args)!=2{\n\t\treturn shim.Error(\"Incorrect arguments, expecting 2\")\n\t}\n\n\tusername := args[0]\n\tcoll := args[1]\n\t//check if the state under username has been deleted\n\tstate_b, err := stub.GetState(username)\n\tif state_b == nil {\n\t\treturn shim.Error(\"User does not exist\")\n\t}\n\n\tprivate_b, err := stub.GetPrivateData(coll, username) \n if err != nil {\n return shim.Error(\"Failed to get private details for \"+username)\n } else if private_b == nil {\n return shim.Error(\"Private details do not exist for \"+username)\n }\n\treturn shim.Success(private_b)\n}", "func (api *LaborStatsAPI) QueryCountryProfile() ([]CountryProfile, error) {\n\ta := CountryProfileAPI{\n\t\tDebug: api.Debug,\n\t\tSecretKey: api.SecretKey,\n\t}\n\n\terr := a.sendRequest()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres, err := a.unmarshalData()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn res, nil\n\n}", "func (c *Store) RetrieveProfile(forAddr string) (*messages.JSONProfile, bool) {\n\tc.ProfileLock.RLock()\n\tdefer c.ProfileLock.RUnlock()\n\n\tif msg, ok := c.Profiles[forAddr]; ok {\n\t\treturn msg.JSONProfile, true\n\t}\n\treturn nil, false\n}", "func (api *API) GetDLPProfile(ctx context.Context, rc *ResourceContainer, profileID string) (DLPProfile, error) {\n\tif rc.Identifier == \"\" {\n\t\treturn DLPProfile{}, ErrMissingResourceIdentifier\n\t}\n\n\tif profileID == \"\" {\n\t\treturn DLPProfile{}, ErrMissingProfileID\n\t}\n\n\turi := buildURI(fmt.Sprintf(\"/%s/%s/dlp/profiles/%s\", rc.Level, rc.Identifier, profileID), nil)\n\n\tres, err := api.makeRequestContext(ctx, http.MethodGet, uri, nil)\n\tif err != nil {\n\t\treturn DLPProfile{}, err\n\t}\n\n\tvar dlpProfileResponse DLPProfileResponse\n\terr = json.Unmarshal(res, &dlpProfileResponse)\n\tif err != nil {\n\t\treturn DLPProfile{}, fmt.Errorf(\"%s: %w\", errUnmarshalError, err)\n\t}\n\n\treturn dlpProfileResponse.Result, nil\n}", "func getCompleteProfile(ctx sdk.Context, path []string, k Keeper) (res []byte, sdkError error) {\n\tid := path[0]\n\tuser, err := k.GetUser(ctx, id)\n\tif err != nil {\n\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrInvalidAddress, \"user with address not found\")\n\t}\n\n\tthoughts := k.ListThoughtByCreator(ctx, user.Creator)\n\n\tvar followers []types.User\n\tvar followings []types.User\n\n\tfor _, addr := range user.Followers {\n\t\tfollower, err := k.GetUser(ctx, addr.String())\n\t\tif err == nil {\n\t\t\tfollowers = append(followers, follower)\n\t\t}\n\t}\n\n\tfor _, addr := range user.Following {\n\t\tfollowing, err := k.GetUser(ctx, addr.String())\n\t\tif err == nil {\n\t\t\tfollowings = append(followings, following)\n\t\t}\n\t}\n\n\tvar completeProfile = types.CompleteProfile{\n\t\tCreator: user.Creator,\n\t\tID: user.ID,\n\t\tUsername: user.Username,\n\t\tBio: user.Bio,\n\t\tAvatar: user.Avatar,\n\t\tFollowing: followings,\n\t\tFollowers: followers,\n\t\tThoughts: thoughts,\n\t}\n\n\tres, err = codec.MarshalJSONIndent(k.cdc, completeProfile)\n\tif err != nil {\n\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrJSONMarshal, err.Error())\n\t}\n\n\treturn res, nil\n}", "func GetProfile(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *ProfileState, opts ...pulumi.ResourceOption) (*Profile, error) {\n\tvar resource Profile\n\terr := ctx.ReadResource(\"azure-native:cdn/v20200415:Profile\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func DefaultReadProfile(ctx context.Context, in *Profile, db *gorm1.DB) (*Profile, error) {\n\tif in == nil {\n\t\treturn nil, errors1.NilArgumentError\n\t}\n\tormObj, err := in.ToORM(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif ormObj.Id == \"\" {\n\t\treturn nil, errors1.EmptyIdError\n\t}\n\tif hook, ok := interface{}(&ormObj).(ProfileORMWithBeforeReadApplyQuery); ok {\n\t\tif db, err = hook.BeforeReadApplyQuery(ctx, db); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tif db, err = gorm2.ApplyFieldSelection(ctx, db, nil, &ProfileORM{}); err != nil {\n\t\treturn nil, err\n\t}\n\tif hook, ok := interface{}(&ormObj).(ProfileORMWithBeforeReadFind); ok {\n\t\tif db, err = hook.BeforeReadFind(ctx, db); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tormResponse := ProfileORM{}\n\tif err = db.Where(&ormObj).First(&ormResponse).Error; err != nil {\n\t\treturn nil, err\n\t}\n\tif hook, ok := interface{}(&ormResponse).(ProfileORMWithAfterReadFind); ok {\n\t\tif err = hook.AfterReadFind(ctx, db); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tpbResponse, err := ormResponse.ToPB(ctx)\n\treturn &pbResponse, err\n}", "func (t tomlConfig) getProfile(name string) Database {\n\treturn t.DB[name]\n}", "func (s *Service) ProfilesGet(userID string) *ProfilesGetOp {\n\treturn &ProfilesGetOp{\n\t\tCredential: s.credential,\n\t\tMethod: \"GET\",\n\t\tPath: strings.Join([]string{\"users\", userID, \"profile\"}, \"/\"),\n\t\tAccept: \"application/json\",\n\t\tQueryOpts: make(url.Values),\n\t\tVersion: esign.APIv21,\n\t}\n}", "func (p *profileParsing) Get(options map[string]interface{}) (riminderResponse.ProfileParsingGetElem, error) {\n\tquery := map[string]string{\n\t\t\"source_id\": options[\"source_id\"].(string),\n\t}\n\tAddIfNotEmptyStrMap(&query, options, \"profile_id\")\n\tAddIfNotEmptyStrMap(&query, options, \"profile_reference\")\n\n\tresp := riminderResponse.ProfileParsingGetContainer{}\n\terr := p.client.Get(\"profile/parsing\", query, &resp)\n\tif err != nil {\n\t\treturn riminderResponse.ProfileParsingGetElem{}, err\n\t}\n\treturn resp.Data, nil\n}", "func (repository *Datastore)UpdateProfile(p *user.Person)error {\n\tquery := `UPDATE userRepository SET password=?,firstname=?,lastname=?,age=?,gender=?,city=?,country=?,phone=?,email=?,githubUsername=? WHERE username = ?`\n\tchanges, err := repository.Db.Preparex(query)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = changes.Exec(p.Password, p.Firstname, p.Lastname, p.Age, p.Gender, p.City, p.Country, p.Phone, p.EmailId, p.GithubUsername, p.Username)\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Println(\"profile updated on database\")\n\treturn nil\n}", "func (s *Service) GetProfileImage(userID string) *GetProfileImageOp {\n\treturn &GetProfileImageOp{\n\t\tCredential: s.credential,\n\t\tMethod: \"GET\",\n\t\tPath: strings.Join([]string{\"users\", userID, \"profile\", \"image\"}, \"/\"),\n\t\tQueryOpts: make(url.Values),\n\t\tVersion: esign.APIv21,\n\t}\n}", "func GetCmdQueryDenomTrace() *cobra.Command {\r\n\tcmd := &cobra.Command{\r\n\t\tUse: \"denom-trace [hash]\",\r\n\t\tShort: \"Query the denom trace info from a given trace hash\",\r\n\t\tLong: \"Query the denom trace info from a given trace hash\",\r\n\t\tExample: fmt.Sprintf(\"%s query ibc-transfer denom-trace [hash]\", version.AppName),\r\n\t\tArgs: cobra.ExactArgs(1),\r\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\r\n\t\t\tclientCtx, err := client.GetClientQueryContext(cmd)\r\n\t\t\tif err != nil {\r\n\t\t\t\treturn err\r\n\t\t\t}\r\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\r\n\r\n\t\t\treq := &types.QueryDenomTraceRequest{\r\n\t\t\t\tHash: args[0],\r\n\t\t\t}\r\n\r\n\t\t\tres, err := queryClient.DenomTrace(context.Background(), req)\r\n\t\t\tif err != nil {\r\n\t\t\t\treturn err\r\n\t\t\t}\r\n\r\n\t\t\treturn clientCtx.PrintProto(res)\r\n\t\t},\r\n\t}\r\n\r\n\tflags.AddQueryFlagsToCmd(cmd)\r\n\treturn cmd\r\n}", "func (c *Cluster) GetProfile(project, name string) (int64, *api.Profile, error) {\n\tvar result *api.Profile\n\tvar id int64\n\n\terr := c.Transaction(context.TODO(), func(ctx context.Context, tx *ClusterTx) error {\n\t\tvar err error\n\t\tprofiles, err := cluster.GetProfilesIfEnabled(ctx, tx.Tx(), project, []string{name})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif len(profiles) != 1 {\n\t\t\treturn fmt.Errorf(\"Expected one profile with name %q, got %d profiles\", name, len(profiles))\n\t\t}\n\n\t\tprofile := profiles[0]\n\t\tid = int64(profile.ID)\n\t\tresult, err = profile.ToAPI(ctx, tx.Tx())\n\n\t\treturn err\n\t})\n\tif err != nil {\n\t\treturn -1, nil, err\n\t}\n\n\treturn id, result, nil\n}", "func QueryBalanceCmd() *cobra.Command {\n\treturn qflags(&cobra.Command{\n\t\tUse: \"balance [batch_denom] [account]\",\n\t\tShort: \"Retrieve the tradable and retired balances of the credit batch\",\n\t\tLong: \"Retrieve the tradable and retired balances of the credit batch for a given account address\",\n\t\tArgs: cobra.ExactArgs(2),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tc, ctx, err := mkQueryClient(cmd)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tres, err := c.Balance(cmd.Context(), &ecocredit.QueryBalanceRequest{\n\t\t\t\tBatchDenom: args[0], Account: args[1],\n\t\t\t})\n\t\t\treturn print(ctx, res, err)\n\t\t},\n\t})\n}", "func (s *HighAvailabilityService) GetAvailabilityProfileDescriptorCommand(input *GetAvailabilityProfileDescriptorCommandInput) (output *models.DescriptorView, resp *http.Response, err error) {\n\tpath := \"/highAvailability/availabilityProfiles/descriptors/{availabilityProfileType}\"\n\tpath = strings.Replace(path, \"{availabilityProfileType}\", input.AvailabilityProfileType, -1)\n\n\top := &request.Operation{\n\t\tName: \"GetAvailabilityProfileDescriptorCommand\",\n\t\tHTTPMethod: \"GET\",\n\t\tHTTPPath: path,\n\t\tQueryParams: map[string]string{},\n\t}\n\toutput = &models.DescriptorView{}\n\treq := s.newRequest(op, nil, output)\n\n\tif req.Send() == nil {\n\t\treturn output, req.HTTPResponse, nil\n\t}\n\treturn nil, req.HTTPResponse, req.Error\n}", "func (e *sObjstoreV1GwService) GetServiceProfile(method string) (apigw.ServiceProfile, error) {\n\tif ret, ok := e.svcProf[method]; ok {\n\t\treturn ret, nil\n\t}\n\treturn nil, errors.New(\"not found\")\n}", "func (c *Control) Profile() ProfileQS {\n\treturn ProfileQS{}.ControlEq(c)\n}", "func (c *client) GetProfile(jwt string, profileID string) (*Profile, error) {\n\tu, err := url.Parse(fmt.Sprintf(\"%s/%s\", c.profURL.String(), profileID))\n\tif err != nil {\n\t\tfmt.Printf(\"Failed to parse url %s\", err)\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", u.String(), nil)\n\tif err != nil {\n\t\tfmt.Printf(\"Failed to create new request %s\", err)\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Authorization\", fmt.Sprintf(\"Bearer %s\", jwt))\n\n\tresp, err := c.c.Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdefer resp.Body.Close()\n\n\tif resp.StatusCode != 200 {\n\t\treturn nil, fmt.Errorf(\"GetProfile: Code: %d Status: %s\", resp.StatusCode, resp.Status)\n\t}\n\n\tvar prof *Profile\n\terr = json.NewDecoder(resp.Body).Decode(&prof)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"GetProfile: failed to decode resp body: %s\", resp.Body)\n\t}\n\n\treturn prof, nil\n}", "func (_DelegateProfile *DelegateProfileSession) GetProfileByField(_delegate common.Address, _field string) ([]byte, error) {\n\treturn _DelegateProfile.Contract.GetProfileByField(&_DelegateProfile.CallOpts, _delegate, _field)\n}", "func QueryAccount(cdc *wire.Codec, addr string) (*qosacc.QOSAccount, error) {\n\tkey, err := types.AccAddressFromBech32(addr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcliCtx := context.NewOQSCLIContext().WithCodec(cdc)\n\n\tacc, err := cliCtx.GetAccount(key)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn acc, nil\n}", "func GetProfile(c *gin.Context, auth *oauth2.Config, apiToken *oauth2.Token) (profile Profile, err error) {\n\tclient := auth.Client(c, apiToken)\n\turi := \"\"\n\tswitch auth.Endpoint {\n\tcase facebook.Endpoint:\n\t\turi = \"https://graph.facebook.com/v2.2/me?fields=id,name,email,picture,first_name,last_name\"\n\tcase google.Endpoint:\n\t\turi = \"https://www.googleapis.com/oauth2/v1/userinfo?alt=json\"\n\tdefault:\n\t\turi = \"\"\n\t}\n\n\tresp, err := client.Get(uri)\n\tif err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err)\n\t\treturn\n\t}\n\n\tdefer resp.Body.Close()\n\tcontents, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err)\n\t\treturn\n\t}\n\n\tswitch auth.Endpoint {\n\tcase facebook.Endpoint:\n\t\tvar p ProfileFacebook\n\t\terr = json.Unmarshal(contents, &p)\n\t\tif err != nil {\n\t\t\tprofile = Profile{}\n\t\t} else {\n\t\t\tprofile = Profile{\n\t\t\t\tID: p.ID,\n\t\t\t\tEmail: p.Email,\n\t\t\t\tFirstName: p.FirstName,\n\t\t\t\tLastName: p.LastName,\n\t\t\t\tHd: p.Hd,\n\t\t\t\tLocale: p.Locale,\n\t\t\t\tName: p.Name,\n\t\t\t\tSource: \"facebook\",\n\t\t\t}\n\t\t}\n\tcase google.Endpoint:\n\t\tvar p ProfileGoogle\n\t\terr = json.Unmarshal(contents, &p)\n\t\tif err != nil {\n\t\t\tprofile = Profile{}\n\t\t} else {\n\t\t\tprofile = Profile{\n\t\t\t\tID: p.ID,\n\t\t\t\tEmail: p.Email,\n\t\t\t\tFirstName: p.GivenName,\n\t\t\t\tLastName: p.FamilyName,\n\t\t\t\tHd: p.Hd,\n\t\t\t\tLocale: p.Locale,\n\t\t\t\tName: p.Name,\n\t\t\t\tSource: \"google\",\n\t\t\t}\n\t\t}\n\tdefault:\n\t\tprofile = Profile{}\n\t}\n\n\tif len(profile.Email) == 0 {\n\t\terr = errors.New(\"Empty Email\")\n\t}\n\n\treturn\n}", "func GetEvidenceQueryCmd(storeName string, cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"evidence <evidence-address> <signer-address>\",\n\t\tShort: \"get evidence for claim\",\n\t\tArgs: cobra.ExactArgs(2),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tcontent, err := types.DecodeBech32DataAddress(args[0])\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tsigner, err := sdk.AccAddressFromBech32(args[1])\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, err := cliCtx.QueryStore(claim.KeySignatureEvidence(content, signer), storeName)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif len(res) == 0 {\n\t\t\t\treturn fmt.Errorf(\"no evidence for claim\")\n\t\t\t}\n\n\t\t\tvar evidence []types.DataAddress\n\t\t\terr = cdc.UnmarshalBinaryBare(res, &evidence)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar evidenceString strings.Builder\n\n\t\t\tfor _, data := range evidence {\n\t\t\t\tevidenceString.WriteString(data.String())\n\t\t\t\tevidenceString.WriteString(\" \")\n\t\t\t}\n\n\t\t\tfmt.Println(evidenceString)\n\t\t\treturn nil\n\t\t},\n\t}\n}", "func GetProfile(ctx *pulumi.Context) string {\n\tv, err := config.Try(ctx, \"aws:profile\")\n\tif err == nil {\n\t\treturn v\n\t}\n\tif dv, ok := getEnvOrDefault(\"\", nil, \"AWS_PROFILE\").(string); ok {\n\t\treturn dv\n\t}\n\treturn v\n}", "func GetCmdQueryTaxRate(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: treasury.QueryTaxRate + \" [epoch]\",\n\t\tShort: \"Query the stability tax rate\",\n\t\tLong: strings.TrimSpace(`\nQuery the stability tax rate at the specified epoch.\n\n$ terracli query treasury taxrate --epoch=14\n`),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tepoch := viper.GetInt(flagEpoch)\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s/%d\", queryRoute, treasury.QueryTaxRate, epoch), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar taxRate sdk.Dec\n\t\t\tcdc.MustUnmarshalBinaryLengthPrefixed(res, &taxRate)\n\t\t\treturn cliCtx.PrintOutput(taxRate)\n\t\t},\n\t}\n\n\treturn cmd\n}", "func GetCmdQueryFeePool(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"pool\",\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query the global fee pool\",\n\t\tLong: \"pool\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\t// Query the proposal\n\t\t\tres, err := queryFeePool(cliCtx, cdc, queryRoute)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tfmt.Println(string(res))\n\t\t\treturn nil\n\t\t},\n\t}\n\n\treturn cmd\n}", "func LookupProfile(ctx *pulumi.Context, args *LookupProfileArgs, opts ...pulumi.InvokeOption) (*LookupProfileResult, error) {\n\topts = internal.PkgInvokeDefaultOpts(opts)\n\tvar rv LookupProfileResult\n\terr := ctx.Invoke(\"aws-native:transfer:getProfile\", args, &rv, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &rv, nil\n}", "func GetQueryCmd() *cobra.Command {\n\tqueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tAliases: []string{\"md\"},\n\t\tShort: \"Querying commands for the metadata module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tqueryCmd.AddCommand(\n\t\tGetMetadataParamsCmd(),\n\t\tGetMetadataByIDCmd(),\n\t\tGetMetadataGetAllCmd(),\n\t\tGetMetadataScopeCmd(),\n\t\tGetMetadataSessionCmd(),\n\t\tGetMetadataRecordCmd(),\n\t\tGetMetadataScopeSpecCmd(),\n\t\tGetMetadataContractSpecCmd(),\n\t\tGetMetadataRecordSpecCmd(),\n\t\tGetOwnershipCmd(),\n\t\tGetValueOwnershipCmd(),\n\t\tGetOSLocatorCmd(),\n\t)\n\treturn queryCmd\n}", "func GetAccountAttributeCmd() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"get [address] [name]\",\n\t\tShort: \"Get account attributes by name\",\n\t\tExample: strings.TrimSpace(\n\t\t\tfmt.Sprintf(`\n\t\t\t\t$ %[1]s query attribute get pb1skjwj5whet0lpe65qaq4rpq03hjxlwd9nf39lk attrib.name\n\t\t\t\t$ %[1]s query attribute get pb1skjwj5whet0lpe65qaq4rpq03hjxlwd9nf39lk attrib.name --page=2 --limit=100\n\t\t\t\t`,\n\t\t\t\tversion.AppName,\n\t\t\t)),\n\t\tArgs: cobra.ExactArgs(2),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclientCtx, err := client.GetClientQueryContext(cmd)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\n\n\t\t\tpageReq, err := client.ReadPageRequest(withPageKeyDecoded(cmd.Flags()))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\taddress := strings.ToLower(strings.TrimSpace(args[0]))\n\t\t\tname := strings.ToLower(strings.TrimSpace(args[1]))\n\n\t\t\tvar response *types.QueryAttributeResponse\n\t\t\tif response, err = queryClient.Attribute(\n\t\t\t\tcontext.Background(),\n\t\t\t\t&types.QueryAttributeRequest{Account: address, Name: name, Pagination: pageReq},\n\t\t\t); err != nil {\n\t\t\t\tfmt.Printf(\"failed to query account \\\"%s\\\" attributes for name \\\"%s\\\": %v\\n\", address, name, err)\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\treturn clientCtx.PrintProto(response)\n\t\t},\n\t}\n\n\tflags.AddPaginationFlagsToCmd(cmd, \"get\")\n\tflags.AddQueryFlagsToCmd(cmd)\n\n\treturn cmd\n}", "func Profile(context *cli.Context) error {\n\tsecretKey := context.String(flags.SecretKeyFlag)\n\tif err := fieldEmpty(secretKey, flags.SecretKeyFlag); err != nil {\n\t\treturn err\n\t}\n\taccessKey := context.String(flags.AccessKeyFlag)\n\tif err := fieldEmpty(accessKey, flags.AccessKeyFlag); err != nil {\n\t\treturn err\n\t}\n\tsessionToken := context.String(flags.SessionTokenFlag)\n\tprofileName := context.String(flags.ProfileNameFlag)\n\tif err := fieldEmpty(profileName, flags.ProfileNameFlag); err != nil {\n\t\treturn err\n\t}\n\tprofile := &config.Profile{\n\t\tAWSAccessKey: accessKey,\n\t\tAWSSecretKey: secretKey,\n\t\tAWSSessionToken: sessionToken,\n\t}\n\n\trdwr, err := config.NewReadWriter()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Error saving profile\")\n\t}\n\tif err = rdwr.SaveProfile(profileName, profile); err != nil {\n\t\treturn errors.Wrap(err, \"Error saving profile\")\n\t}\n\n\tlogrus.Infof(\"Saved ECS CLI profile configuration %s.\", profileName)\n\treturn nil\n}", "func DecodeGrpcReqDSCProfile(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*DSCProfile)\n\treturn req, nil\n}", "func queryfb(device string) (uint16, uint16, error) {\n\tvar fbinfo C.struct_fb_var_screeninfo\n\n\tif C.query_fb(C.CString(device), &fbinfo) != 0 {\n\t\treturn 0, 0, fmt.Errorf(\"%s: permission denied\")\n\t}\n\n\treturn uint16(fbinfo.xres), uint16(fbinfo.yres), nil\n}", "func (c *Cluster) GetInstancesWithProfile(project, profile string) (map[string][]string, error) {\n\terr := c.Transaction(context.TODO(), func(ctx context.Context, tx *ClusterTx) error {\n\t\tenabled, err := cluster.ProjectHasProfiles(context.Background(), tx.tx, project)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Check if project has profiles: %w\", err)\n\t\t}\n\n\t\tif !enabled {\n\t\t\tproject = \"default\"\n\t\t}\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tq := `SELECT instances.name, projects.name FROM instances\n\t\tJOIN instances_profiles ON instances.id == instances_profiles.instance_id\n\t\tJOIN projects ON projects.id == instances.project_id\n\t\tWHERE instances_profiles.profile_id ==\n\t\t (SELECT profiles.id FROM profiles\n\t\t JOIN projects ON projects.id == profiles.project_id\n\t\t WHERE profiles.name=? AND projects.name=?)`\n\n\tresults := map[string][]string{}\n\tinargs := []any{profile, project}\n\tvar name string\n\toutfmt := []any{name, name}\n\n\toutput, err := queryScan(c, q, inargs, outfmt)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor _, r := range output {\n\t\tif results[r[1].(string)] == nil {\n\t\t\tresults[r[1].(string)] = []string{}\n\t\t}\n\n\t\tresults[r[1].(string)] = append(results[r[1].(string)], r[0].(string))\n\t}\n\n\treturn results, nil\n}", "func (m *WindowsKioskProfile) GetProfileId()(*string) {\n val, err := m.GetBackingStore().Get(\"profileId\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func GetCmdQueryTaxProceeds(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: treasury.QueryTaxProceeds,\n\t\tShort: \"Query the tax proceeds for the epoch\",\n\t\tLong: strings.TrimSpace(`\nQuery the tax proceeds corresponding to the given epoch. The return value will be sdk.Coins{} of all the taxes collected. \n\n$ terracli query treasury tax-proceeds --epoch=14\n`),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tepoch := viper.GetInt(flagEpoch)\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s/%d\", queryRoute, treasury.QueryTaxProceeds, epoch), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar taxProceeds sdk.Coins\n\t\t\tcdc.MustUnmarshalBinaryLengthPrefixed(res, &taxProceeds)\n\t\t\treturn cliCtx.PrintOutput(taxProceeds)\n\t\t},\n\t}\n\n\treturn cmd\n}", "func (t *QueryCommand) GetUsage() string {\n\treturn \"dev:query\"\n}", "func (ConferenceAPI) GetProfile(c context.Context) (*Profile, error) {\n\tpid, err := profileID(c)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn getProfile(c, pid)\n}", "func GetQueryCmd(queryRoute string) *cobra.Command {\n\t// Group id queries under a subcommand\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tcmd.AddCommand(\n\t\tCmdGetGovernmentAddr(),\n\t)\n\treturn cmd\n}", "func TestProfile(t *testing.T) {\n\tvar se profile.Profile\n\n\t// Validate returned fields to match expected name set.\n\tfields := se.Fields()\n\n\tif _, ok := fields[\"address\"]; !ok {\n\t\ttests.Failed(\"Should have a 'address' field\")\n\t}\n\ttests.Passed(\"Should have a 'address' field\")\n\n\tif _, ok := fields[\"user_id\"]; !ok {\n\t\ttests.Failed(\"Should have a 'user_id' field\")\n\t}\n\ttests.Passed(\"Should have a 'user_id' field\")\n\n\tif _, ok := fields[\"public_id\"]; !ok {\n\t\ttests.Failed(\"Should have a 'public_id' field\")\n\t}\n\ttests.Passed(\"Should have a 'public_id' field\")\n\n\tif _, ok := fields[\"first_name\"]; !ok {\n\t\ttests.Failed(\"Should have a 'first_name' field\")\n\t}\n\ttests.Passed(\"Should have a 'first_name' field\")\n\n\tif _, ok := fields[\"last_name\"]; !ok {\n\t\ttests.Failed(\"Should have a 'last_name' field\")\n\t}\n\ttests.Passed(\"Should have a 'last_name' field\")\n}", "func (AppModuleBasic) GetQueryCmd(cdc *amino.Codec) *cobra.Command {\n\treturn client.GetQueryCmd(cdc)\n}", "func GetQueryCmd() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the relationships module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tcmd.AddCommand(\n\t\tGetCmdQueryUserRelationships(),\n\t\tGetCmdQueryUserBlocks(),\n\t)\n\treturn cmd\n}", "func (db *Database) QueryTwoAuthByAddressForDCI(address string) (*schema.TwoAuthForDCI, error) {\n\tvar ta schema.TwoAuthForDCI\n\n\terr := db.Model(&ta).\n\t\tWhere(\"Address = ?\", address).\n\t\tSelect()\n\n\tif err == pg.ErrNoRows {\n\t\treturn &ta, fmt.Errorf(\"no rows in twoauth table: %s\", err)\n\t}\n\n\tif err != nil {\n\t\treturn &ta, fmt.Errorf(\"unexpected database error: %s\", err)\n\t}\n\n\treturn &ta, nil\n}", "func GetCmdBalance(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"balance\",\n\t\tShort: \"Query account token balance\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tkey, err := sdk.AccAddressFromBech32(args[0])\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tres, _, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/private/%s\", queryRoute, key), nil)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"could not find private profile - %s \\n\", key)\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\treturn cliCtx.PrintOutput(string(res))\n\t\t},\n\t}\n}", "func (p *profileCache) DeviceCommand(profileName string, commandName string) (models.DeviceCommand, bool) {\n\tp.mutex.RLock()\n\tdefer p.mutex.RUnlock()\n\n\tdcs, ok := p.deviceCommandMap[profileName]\n\tif !ok {\n\t\treturn models.DeviceCommand{}, false\n\t}\n\n\tdc, ok := dcs[commandName]\n\treturn dc, ok\n}" ]
[ "0.73663527", "0.66192484", "0.66032195", "0.6373226", "0.62577033", "0.5889509", "0.5735262", "0.5682528", "0.56526613", "0.54866666", "0.5468216", "0.54676366", "0.5464707", "0.5439784", "0.5404302", "0.5396977", "0.53676295", "0.53150916", "0.53021604", "0.5297001", "0.5251094", "0.52248317", "0.52177143", "0.519803", "0.51577115", "0.51541775", "0.5138208", "0.51276904", "0.511461", "0.511391", "0.5111596", "0.51038647", "0.5068776", "0.5054223", "0.5044595", "0.5044218", "0.50437605", "0.5007318", "0.50067633", "0.50030994", "0.50027704", "0.4998607", "0.49965158", "0.49895737", "0.4987935", "0.4977783", "0.49759698", "0.49606493", "0.4959705", "0.4931711", "0.49248984", "0.49077672", "0.49025008", "0.4902498", "0.48842773", "0.48781222", "0.4867901", "0.4862886", "0.48569953", "0.48471078", "0.48211798", "0.4809489", "0.4807637", "0.48040614", "0.47969177", "0.4795136", "0.4792029", "0.4788374", "0.47872746", "0.47741616", "0.4766467", "0.4761491", "0.47599676", "0.4759339", "0.475577", "0.47540614", "0.47398946", "0.4738255", "0.4734157", "0.4722806", "0.47221467", "0.4720828", "0.47156212", "0.47116384", "0.47116274", "0.4705608", "0.46973106", "0.46970123", "0.46942052", "0.46936136", "0.46921706", "0.4691164", "0.4687137", "0.46865648", "0.46742946", "0.4670912", "0.46707457", "0.46674314", "0.46667093", "0.4654809" ]
0.8442145
0
GetCmdQueryProfiles queries all the profiles
func GetCmdQueryProfiles(cdc *codec.Codec) *cobra.Command { return &cobra.Command{ Use: "all", Short: "Retrieve all the registered profiles.", Args: cobra.NoArgs, RunE: func(cmd *cobra.Command, args []string) error { cliCtx := context.NewCLIContext().WithCodec(cdc) route := fmt.Sprintf("custom/%s/%s", types.QuerierRoute, types.QueryProfiles) res, _, err := cliCtx.QueryWithData(route, nil) if err != nil { fmt.Printf("Could not find any profile") return nil } var out types.Profiles cdc.MustUnmarshalJSON(res, &out) return cliCtx.PrintOutput(out) }, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *SmartContract) QueryAllProfiles(ctx contractapi.TransactionContextInterface) ([]QueryProfileResult, error) {\n\tstartKey := \"\"\n\tendKey := \"\"\n\n\tresultsIterator, err := ctx.GetStub().GetStateByRange(startKey, endKey)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer resultsIterator.Close()\n\n\tresults := []QueryProfileResult{}\n\n\tfor resultsIterator.HasNext() {\n\t\tqueryResponse, err := resultsIterator.Next()\n\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif strings.Contains(queryResponse.Key, \"profile-\"){\n\t\t\tprofile := new(UserProfile)\n\t\t\t_ = json.Unmarshal(queryResponse.Value, profile)\n\n\t\t\tqueryResult := QueryProfileResult{Key: queryResponse.Key, Record: profile}\n\t\t\tresults = append(results, queryResult)\n\t\t}\n\t\t\n\t}\n\n\treturn results, nil\n}", "func queryProfiles(ctx sdk.Context, _ abci.RequestQuery, keeper Keeper) ([]byte, error) {\n\taccounts := keeper.GetProfiles(ctx)\n\n\tbz, err := codec.MarshalJSONIndent(keeper.Cdc, &accounts)\n\tif err != nil {\n\t\tpanic(\"could not marshal result to JSON\")\n\t}\n\n\treturn bz, nil\n}", "func GetCmdQueryProfileParams(cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"parameters\",\n\t\tShort: \"Retrieve all the profile module parameters\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\troute := fmt.Sprintf(\"custom/%s/%s\", types.QuerierRoute, types.QueryParams)\n\t\t\tres, _, err := cliCtx.QueryWithData(route, nil)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"Could not find profile parameters\")\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tvar out types.Params\n\t\t\tcdc.MustUnmarshalJSON(res, &out)\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n}", "func GetCmdQueryProfile(cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"profile [address_or_dtag]\",\n\t\tShort: \"Retrieve the profile having the specified user address or profile dtag, if any.\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\troute := fmt.Sprintf(\"custom/%s/%s/%s\", types.QuerierRoute, types.QueryProfile, args[0])\n\t\t\tres, _, err := cliCtx.QueryWithData(route, nil)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"Could not find a profile with dtag %s \\n\", args[0])\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tvar out types.Profile\n\t\t\tcdc.MustUnmarshalJSON(res, &out)\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n}", "func FetchProfiles() map[string]string {\n\ttoken := auth.NewToken()\n\tquery := queryPayload{\"SELECT Id, ProfileName FROM CommunicationProfile\"}\n\tpayload, err := json.Marshal(query)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treq, err := http.NewRequest(\"POST\", viper.GetString(\"baseurl\")+\"/v1/action/query\", bytes.NewBuffer(payload))\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"Authorization\", \"Bearer \"+token.Val)\n\n\tresponse, err := http.DefaultClient.Do(req)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer response.Body.Close()\n\n\tif response.StatusCode != 200 {\n\t\tbody, err := ioutil.ReadAll(response.Body)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tlog.Fatal(string(body))\n\t}\n\n\tdec := json.NewDecoder(response.Body)\n\tvar body profilesQueryResponse\n\tif err = dec.Decode(&body); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tif !body.Done {\n\t\tlog.Fatalln(\"there are more communication profile to query\")\n\t}\n\n\tresult := make(map[string]string)\n\tfor _, p := range body.Records {\n\t\tresult[p.Name] = p.ID\n\t}\n\n\treturn result\n}", "func (c *OVClient) GetProfiles(start string, count string, filter string, sort string, scopeUris string) (ServerProfileList, error) {\n\tvar (\n\t\turi = \"/rest/server-profiles\"\n\t\tq map[string]interface{}\n\t\tprofiles ServerProfileList\n\t)\n\tq = make(map[string]interface{})\n\tif len(filter) > 0 {\n\t\tq[\"filter\"] = filter\n\t}\n\n\tif sort != \"\" {\n\t\tq[\"sort\"] = sort\n\t}\n\n\tif start != \"\" {\n\t\tq[\"start\"] = start\n\t}\n\n\tif count != \"\" {\n\t\tq[\"count\"] = count\n\t}\n\n\tif scopeUris != \"\" {\n\t\tq[\"scopeUris\"] = scopeUris\n\t}\n\n\t// refresh login\n\tc.RefreshLogin()\n\tc.SetAuthHeaderOptions(c.GetAuthHeaderMap())\n\t// Setup query\n\tif len(q) > 0 {\n\t\tc.SetQueryString(q)\n\t}\n\tdata, err := c.RestAPICall(rest.GET, uri, nil)\n\tif err != nil {\n\t\treturn profiles, err\n\t}\n\n\tlog.Debugf(\"GetProfiles %s\", data)\n\tif err := json.Unmarshal([]byte(data), &profiles); err != nil {\n\t\treturn profiles, err\n\t}\n\treturn profiles, nil\n}", "func listProfiles(ctx context.Context, _ []string) error {\n\tm, err := cmdutils.LoadManager(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"%15s %s\\n\", \"ID\", \"NAME\")\n\tfor _, p := range m.Profiles() {\n\t\tfmt.Printf(\"%15s %s\\n\", p.Id, p.Name)\n\t}\n\n\treturn nil\n}", "func (a *App) GetAllProfiles(w http.ResponseWriter, r *http.Request) {\n\thandler.GetAllProfiles(a.DB, w, r)\n}", "func GetProfiles(ctx context.Context) []ProfileEntry {\n\tvar c *SqlContext\n\ti := ctx.Value(goradd.SqlContext)\n\tif i == nil {\n\t\tpanic(\"Profiling requires a preloaded context.\")\n\t} else {\n\t\tc = i.(*SqlContext)\n\t}\n\n\tif c != nil {\n\t\tp := c.profiles\n\t\tc.profiles = nil\n\t\treturn p\n\t}\n\treturn nil\n}", "func (s *HighAvailabilityService) GetAvailabilityProfilesCommand(input *GetAvailabilityProfilesCommandInput) (output *models.AvailabilityProfilesView, resp *http.Response, err error) {\n\tpath := \"/highAvailability/availabilityProfiles\"\n\top := &request.Operation{\n\t\tName: \"GetAvailabilityProfilesCommand\",\n\t\tHTTPMethod: \"GET\",\n\t\tHTTPPath: path,\n\t\tQueryParams: map[string]string{\n\t\t\t\"page\": input.Page,\n\t\t\t\"numberPerPage\": input.NumberPerPage,\n\t\t\t\"filter\": input.Filter,\n\t\t\t\"name\": input.Name,\n\t\t\t\"sortKey\": input.SortKey,\n\t\t\t\"order\": input.Order,\n\t\t},\n\t}\n\toutput = &models.AvailabilityProfilesView{}\n\treq := s.newRequest(op, nil, output)\n\n\tif req.Send() == nil {\n\t\treturn output, req.HTTPResponse, nil\n\t}\n\treturn nil, req.HTTPResponse, req.Error\n}", "func queryProfileParams(ctx sdk.Context, _ abci.RequestQuery, keeper Keeper) ([]byte, error) {\n\tprofileParams := keeper.GetParams(ctx)\n\n\tbz, err := codec.MarshalJSONIndent(keeper.Cdc, &profileParams)\n\tif err != nil {\n\t\tpanic(\"could not marshal result to JSON\")\n\t}\n\n\treturn bz, nil\n}", "func (s *Scraper) SearchProfiles(ctx context.Context, query string, maxProfilesNbr int) <-chan *ProfileResult {\n\treturn getUserTimeline(ctx, query, maxProfilesNbr, s.FetchSearchProfiles)\n}", "func (s *Service) ProfilesGet(userID string) *ProfilesGetOp {\n\treturn &ProfilesGetOp{\n\t\tCredential: s.credential,\n\t\tMethod: \"GET\",\n\t\tPath: strings.Join([]string{\"users\", userID, \"profile\"}, \"/\"),\n\t\tAccept: \"application/json\",\n\t\tQueryOpts: make(url.Values),\n\t\tVersion: esign.APIv21,\n\t}\n}", "func GetProfileList(ctx context.Context) ([]*shill.Profile, error) {\n\tmanager, err := shill.NewManager(ctx)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed creating shill manager object\")\n\t}\n\t// Refresh the in-memory profile list.\n\tif _, err := manager.GetProperties(ctx); err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed refreshing the in-memory profile list\")\n\t}\n\t// Get current profiles.\n\tprofiles, err := manager.Profiles(ctx)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed getting profile list\")\n\t}\n\treturn profiles, nil\n}", "func (bg *Backgrounder) GetProfiles() map[string]time.Duration {\n\treturn bg.profile\n}", "func GetQueryCmd(cdc *codec.Codec) *cobra.Command {\n\tprofileQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the profiles module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tprofileQueryCmd.AddCommand(flags.GetCommands(\n\t\tGetCmdQueryProfile(cdc),\n\t\tGetCmdQueryProfiles(cdc),\n\t\tGetCmdQueryProfileParams(cdc),\n\t\tGetCmdQueryDTagRequests(cdc),\n\t)...)\n\treturn profileQueryCmd\n}", "func (s *Scraper) FetchSearchProfiles(query string, maxProfilesNbr int, cursor string) ([]*Profile, string, error) {\n\ttimeline, err := s.getSearchTimeline(query, maxProfilesNbr, cursor)\n\tif err != nil {\n\t\treturn nil, \"\", err\n\t}\n\tusers, nextCursor := timeline.parseUsers()\n\treturn users, nextCursor, nil\n}", "func runOperationDevicesGetDevicesApsProfiles(cmd *cobra.Command, args []string) error {\n\tappCli, err := makeClient(cmd, args)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// retrieve flag values from cmd and fill params\n\tparams := devices.NewGetDevicesApsProfilesParams()\n\tif dryRun {\n\n\t\tlogDebugf(\"dry-run flag specified. Skip sending request.\")\n\t\treturn nil\n\t}\n\t// make request and then print result\n\tmsgStr, err := parseOperationDevicesGetDevicesApsProfilesResult(appCli.Devices.GetDevicesApsProfiles(params, nil))\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !debug {\n\n\t\tfmt.Println(msgStr)\n\t}\n\treturn nil\n}", "func (mpRepo *ManageProfileRepository) Profiles() ([]entity.Profile, []error) {\n\tprfs := []entity.Profile{}\n\terrs := mpRepo.conn.Find(&prfs).GetErrors()\n\tif len(errs) > 0 {\n\t\treturn nil, errs\n\t}\n\treturn prfs, errs\n}", "func getProfilesDirectory() (string, error) {\n\tn := uint32(100)\n\tfor {\n\t\tb := make([]uint16, n)\n\t\te := windows.GetProfilesDirectory(&b[0], &n)\n\t\tif e == nil {\n\t\t\treturn syscall.UTF16ToString(b), nil\n\t\t}\n\t\tif e != syscall.ERROR_INSUFFICIENT_BUFFER {\n\t\t\treturn \"\", e\n\t\t}\n\t\tif n <= uint32(len(b)) {\n\t\t\treturn \"\", e\n\t\t}\n\t}\n}", "func (ctx *Context) Profiles() []*Profile {\n\treturn ctx.profileDB.sortedProfiles()\n}", "func GetProfiles() map[string]*model.ConfigProfile {\n\treturn profiles\n}", "func (a *API) Profiles(ctx context.Context, pageNum, pageSize int) (*ProfileResp, error) {\n\n\tall := pageNum < 1\n\tif all {\n\t\tpageSize = 100\n\t\tpageNum = 1\n\t}\n\n\tvar resp ProfileResp\n\tif err := a.Get(ctx, fmt.Sprintf(\"/api/v1/profile?pageNumber=%d&pageSize=%d\", pageNum, getPageSize(pageSize)), &resp); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif all {\n\t\tfor i := 1; i < resp.TotalPageCount; i++ {\n\t\t\tpg, err := a.Profiles(ctx, i+1, pageSize)\n\t\t\tif err != nil {\n\t\t\t\treturn &resp, err\n\t\t\t}\n\t\t\tresp.Profiles = append(resp.Profiles, pg.Profiles...)\n\t\t}\n\t\tresp.TotalPageCount = 1\n\t\tresp.TotalCount = len(resp.Profiles)\n\t\tresp.Count = resp.TotalCount\n\t\tresp.NextPageURL = \"\"\n\t\tresp.LastPageURL = \"\"\n\t}\n\n\treturn &resp, nil\n}", "func (s *Service) PermissionProfilesGet(permissionProfileID string) *PermissionProfilesGetOp {\n\treturn &PermissionProfilesGetOp{\n\t\tCredential: s.credential,\n\t\tMethod: \"GET\",\n\t\tPath: strings.Join([]string{\"permission_profiles\", permissionProfileID}, \"/\"),\n\t\tAccept: \"application/json\",\n\t\tQueryOpts: make(url.Values),\n\t\tVersion: esign.APIv2,\n\t}\n}", "func GetServiceProfiles(ctx context.Context, db sqlx.Queryer, filters ServiceProfileFilters) ([]ServiceProfileMeta, error) {\n\tquery, args, err := sqlx.BindNamed(sqlx.DOLLAR, `\n\t\tselect\n\t\t\tsp.*,\n\t\t\tns.name as network_server_name\n\t\tfrom\n\t\t\tservice_profile sp\n\t\tinner join network_server ns\n\t\t\ton sp.network_server_id = ns.id\n\t\tleft join organization_user ou\n\t\t\ton sp.organization_id = ou.organization_id\n\t\tleft join \"user\" u\n\t\t\ton ou.user_id = u.id\n\t`+filters.SQL()+`\n\t\tgroup by\n\t\t\tsp.service_profile_id,\n\t\t\tsp.name,\n\t\t\tnetwork_server_name\n\t\torder by\n\t\t\tsp.name\n\t\tlimit :limit\n\t\toffset :offset\n\t`, filters)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"named query error\")\n\t}\n\n\tvar sps []ServiceProfileMeta\n\terr = sqlx.Select(db, &sps, query, args...)\n\tif err != nil {\n\t\treturn nil, handlePSQLError(Select, err, \"select error\")\n\t}\n\n\treturn sps, nil\n}", "func (u User) Profiles(t database.Transaction) []goth.User {\n\tif u.IsAnonymous() {\n\t\treturn nil\n\t}\n\tvar profileSlice []goth.User\n\tprofiles := make(map[string]map[string]goth.User)\n\terr := t.Read(u.ID()+\"/profiles\", &profiles)\n\tif err != nil {\n\t\treturn nil\n\t}\n\tfor provider := range profiles {\n\t\tfor account := range profiles[provider] {\n\t\t\tprofileSlice = append(profileSlice, profiles[provider][account])\n\t\t}\n\t}\n\treturn profileSlice\n}", "func (s *SmartContract) QueryProfile(ctx contractapi.TransactionContextInterface, carNumber string) (*Profile, error) {\n\tcarAsBytes, err := ctx.GetStub().GetState(\"profile-\" + carNumber)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to read from world state. %s\", err.Error())\n\t}\n\n\tif carAsBytes == nil {\n\t\treturn nil, fmt.Errorf(\"%s does not exist\", carNumber)\n\t}\n\n\tcar := new(Profile)\n\t_ = json.Unmarshal(carAsBytes, car)\n\n\treturn car, nil\n}", "func (s *EmptyStore) ProfileList() (profiles []*storagepb.Profile, err error) {\n\treturn profiles, nil\n}", "func (s *remoteStore) GetEndDeviceProfiles(req store.GetEndDeviceProfilesRequest) (*store.GetEndDeviceProfilesResponse, error) {\n\tif req.BrandID != \"\" {\n\t\treturn s.getEndDeviceProfilesByBrand(req)\n\t}\n\tall := []*store.EndDeviceProfile{}\n\tbrands, err := s.GetBrands(store.GetBrandsRequest{\n\t\tPaths: []string{\"brand_id\"},\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor _, brand := range brands.Brands {\n\t\tprofiles, err := s.GetEndDeviceProfiles(store.GetEndDeviceProfilesRequest{\n\t\t\tBrandID: brand.BrandId,\n\t\t})\n\t\tif errors.IsNotFound(err) {\n\t\t\t// Skip vendors without any profiles.\n\t\t\tcontinue\n\t\t} else if err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tall = append(all, profiles.Profiles...)\n\t}\n\n\tstart, end := paginate(len(all), req.Limit, req.Page)\n\treturn &store.GetEndDeviceProfilesResponse{\n\t\tCount: end - start,\n\t\tOffset: start,\n\t\tTotal: uint32(len(all)),\n\t\tProfiles: all[start:end],\n\t}, nil\n}", "func (c ProfilesController) GetAll() (JSONResponse, error) {\n\tcontrollerProfiles, err := c.profileProvider.GetProfiles()\n\n\tif err != nil {\n\t\treturn JSONResponse{}, err\n\t}\n\treturn c.JSONResponse(200, controllerProfiles), nil\n}", "func queryProfile(ctx sdk.Context, path []string, _ abci.RequestQuery, keeper Keeper) ([]byte, error) {\n\tif len(strings.TrimSpace(path[0])) == 0 {\n\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrInvalidRequest, \"DTag or address cannot be empty or blank\")\n\t}\n\n\taddress, err := sdk.AccAddressFromBech32(path[0])\n\tif err != nil {\n\t\taddress = keeper.GetDtagRelatedAddress(ctx, path[0])\n\t\tif address == nil {\n\t\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrInvalidRequest, fmt.Sprintf(\"No address related to this dtag: %s\", path[0]))\n\t\t}\n\n\t}\n\n\taccount, found := keeper.GetProfile(ctx, address)\n\n\tif !found {\n\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrInvalidRequest,\n\t\t\tfmt.Sprintf(\"Profile with address %s doesn't exists\", path[0]))\n\t}\n\n\tbz, err := codec.MarshalJSONIndent(keeper.Cdc, &account)\n\tif err != nil {\n\t\tpanic(\"could not marshal result to JSON\")\n\t}\n\n\treturn bz, nil\n}", "func (c *Cluster) GetProfiles(projectName string, profileNames []string) ([]api.Profile, error) {\n\tprofiles := make([]api.Profile, len(profileNames))\n\n\terr := c.Transaction(context.TODO(), func(ctx context.Context, tx *ClusterTx) error {\n\t\tdbProfiles, err := cluster.GetProfilesIfEnabled(ctx, tx.Tx(), projectName, profileNames)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tfor i, profile := range dbProfiles {\n\t\t\tapiProfile, err := profile.ToAPI(ctx, tx.Tx())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tprofiles[i] = *apiProfile\n\t\t}\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn profiles, nil\n}", "func retrieve_gmm_gwy_profiles_list(gmm_api_key string, org_id int) {\n\n\ttype gwy_profiles struct {\n\t\tGatewayProfiles []struct {\n\t\t\tID int `json:\"id\"`\n\t\t\tName string `json:\"name\"`\n\t\t} `json:\"gateway_profiles\"`\n\t\tPaging struct {\n\t\t\tLimit int `json:\"limit\"`\n\t\t\tOffset int `json:\"offset\"`\n\t\t\tPages int `json:\"pages\"`\n\t\t\tCount int `json:\"count\"`\n\t\t\tLinks struct {\n\t\t\t\tFirst string `json:\"first\"`\n\t\t\t\tLast string `json:\"last\"`\n\t\t\t\tNext string `json:\"next\"`\n\t\t\t} `json:\"links\"`\n\t\t} `json:\"paging\"`\n\t}\n\n\tjsonValue, _ := json.Marshal(\"\")\n\trequest, _ := http.NewRequest(\"GET\", \"https://us.ciscokinetic.io/api/v2/organizations/\" + strconv.Itoa(org_id) + \"/gateway_profiles?limit=100\", bytes.NewBuffer(jsonValue))\n\ttoken := \"Token \" + gmm_api_key\n\trequest.Header.Set(\"Authorization\", token)\n\tclient := &http.Client{}\n\tr, err := client.Do(request)\n\n\tif err != nil {\n\t\tfmt.Printf(\"Retrieve GMM GWY Profiles error %s\\n\", err)\n\t\tos.Exit(1)\n\t}\n\n\tresponseData, _ := ioutil.ReadAll(r.Body)\n\n\tvar responseObject gwy_profiles\n\te := json.Unmarshal(responseData, &responseObject)\n\tif e != nil {\n\t\tfmt.Println(\"Unmarshall Error: \", e)\n\t}\n\n\tfmt.Println(\"\")\n\tfmt.Println(\"Total Number of Gateway Profiles in GMM: \", len(responseObject.GatewayProfiles))\n\tfmt.Println(\"\")\n\tfmt.Println(\"Gateway Profiles in GMM\")\n\tfmt.Println(\"-----------------------\")\n\tfor i := 0; i < len(responseObject.GatewayProfiles); i++ {\n\t\tfmt.Println(\"Profile-ID: \", responseObject.GatewayProfiles[i].ID, \" Profile Name: \", responseObject.GatewayProfiles[i].Name)\n\t}\n}", "func (s *ProfileService) List(input *ListProfilesInput) (*ListProfilesOutput, *http.Response, error) {\n\treturn s.ListContext(context.Background(), input)\n}", "func (c *CompositeProfileClient) GetCompositeProfiles(p string, ca string, v string) ([]CompositeProfile, error) {\n\tkey := CompositeProfileKey{\n\t\tName: \"\",\n\t\tProject: p,\n\t\tCompositeApp: ca,\n\t\tVersion: v,\n\t}\n\n\tvalues, err := db.DBconn.Find(c.storeName, key, c.tagMeta)\n\tif err != nil {\n\t\treturn []CompositeProfile{}, pkgerrors.Wrap(err, \"Get Composite Profiles error\")\n\t}\n\n\tvar resp []CompositeProfile\n\n\tfor _, value := range values {\n\t\tcp := CompositeProfile{}\n\t\terr = db.DBconn.Unmarshal(value, &cp)\n\t\tif err != nil {\n\t\t\treturn []CompositeProfile{}, pkgerrors.Wrap(err, \"Get Composite Profiles unmarshalling error\")\n\t\t}\n\t\tresp = append(resp, cp)\n\t}\n\n\treturn resp, nil\n}", "func (S *Service) GetProfile(ctx context.Context) (profileDataList ProfileDataList, err error) {\n\tprofileList, err := S.repo.GetProfile(ctx)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tfor _, profile := range profileList.Data {\n\t\tprofileDataList.Profile = append(profileDataList.Profile, profileToProfileData(profile))\n\t}\n\tprofileDataList.TotalData = profileList.TotalData\n\n\treturn\n}", "func DefaultListProfile(ctx context.Context, db *gorm1.DB, f *query1.Filtering, s *query1.Sorting, p *query1.Pagination, fs *query1.FieldSelection) ([]*Profile, error) {\n\tin := Profile{}\n\tormObj, err := in.ToORM(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif hook, ok := interface{}(&ormObj).(ProfileORMWithBeforeListApplyQuery); ok {\n\t\tif db, err = hook.BeforeListApplyQuery(ctx, db, f, s, p, fs); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tdb, err = gorm2.ApplyCollectionOperators(ctx, db, &ProfileORM{}, &Profile{}, f, s, p, fs)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif hook, ok := interface{}(&ormObj).(ProfileORMWithBeforeListFind); ok {\n\t\tif db, err = hook.BeforeListFind(ctx, db, f, s, p, fs); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tdb = db.Where(&ormObj)\n\tdb = db.Order(\"id\")\n\tormResponse := []ProfileORM{}\n\tif err := db.Find(&ormResponse).Error; err != nil {\n\t\treturn nil, err\n\t}\n\tif hook, ok := interface{}(&ormObj).(ProfileORMWithAfterListFind); ok {\n\t\tif err = hook.AfterListFind(ctx, db, &ormResponse, f, s, p, fs); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tpbResponse := []*Profile{}\n\tfor _, responseEntry := range ormResponse {\n\t\ttemp, err := responseEntry.ToPB(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tpbResponse = append(pbResponse, &temp)\n\t}\n\treturn pbResponse, nil\n}", "func (a *api) h_GET_profiles_prfId(c *gin.Context) {\n\tprfId, err := parseInt64Param(c, \"prfId\")\n\tif a.errorResponse(c, err) {\n\t\treturn\n\t}\n\n\ta.logger.Debug(\"GET /profiles/\", prfId)\n\n\tp, err := a.Dc.GetProfile(prfId)\n\tif a.errorResponse(c, err) {\n\t\treturn\n\t}\n\n\taCtx := a.getAuthContext(c)\n\tif a.errorResponse(c, aCtx.AuthZHasOrgLevel(p.OrgId, auth.AUTHZ_LEVEL_OU)) {\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, a.mprofile2profile(p))\n}", "func (s Source) AllProfiles() []string {\n\tif s.IsProfile() {\n\t\treturn []string{s.Path}\n\t}\n\treturn []string{}\n}", "func GetFanProfiles(conn io.ReadWriter) (fp FanProfiles, err error) {\n\n\tresp, err := getQuery(getFanProfiles, conn)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn *resp.(*FanProfiles), err\n}", "func (c *Cluster) GetProfileNames(project string) ([]string, error) {\n\terr := c.Transaction(context.TODO(), func(ctx context.Context, tx *ClusterTx) error {\n\t\tenabled, err := cluster.ProjectHasProfiles(context.Background(), tx.tx, project)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Check if project has profiles: %w\", err)\n\t\t}\n\n\t\tif !enabled {\n\t\t\tproject = \"default\"\n\t\t}\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tq := `\nSELECT profiles.name\n FROM profiles\n JOIN projects ON projects.id = profiles.project_id\nWHERE projects.name = ?\n`\n\tinargs := []any{project}\n\tvar name string\n\toutfmt := []any{name}\n\tresult, err := queryScan(c, q, inargs, outfmt)\n\tif err != nil {\n\t\treturn []string{}, err\n\t}\n\n\tresponse := []string{}\n\tfor _, r := range result {\n\t\tresponse = append(response, r[0].(string))\n\t}\n\n\treturn response, nil\n}", "func GetProfile(c *fiber.Ctx) error {\n\tID := c.Query(\"id\")\n\n\tif len(ID) < 1 {\n\t\tc.Send(ID)\n\t\tc.Send(\"Parameter ID is required\")\n\t\tc.SendStatus(http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tprofile, err := database.SearchProfile(ID)\n\tif err != nil {\n\t\tc.Send(\"Error Occurred\" + err.Error())\n\t\tc.SendStatus(http.StatusBadRequest)\n\t\treturn\n\t}\n\tif err := c.JSON(profile); err != nil {\n\t\tc.Status(500).Send(err)\n\t\treturn\n\t}\n\tc.Accepts(\"application/json\")\n\tc.SendStatus(http.StatusAccepted)\n}", "func (r *AccountUserProfilesService) List(profileId int64) *AccountUserProfilesListCall {\n\tc := &AccountUserProfilesListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.profileId = profileId\n\treturn c\n}", "func makeOperationDevicesGetDevicesApsProfilesCmd() (*cobra.Command, error) {\n\tcmd := &cobra.Command{\n\t\tUse: \"getDevicesApsProfiles\",\n\t\tShort: ``,\n\t\tRunE: runOperationDevicesGetDevicesApsProfiles,\n\t}\n\n\tif err := registerOperationDevicesGetDevicesApsProfilesParamFlags(cmd); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn cmd, nil\n}", "func (dao PathProfileDAOPsql) GetByProfileID(id int16) ([]models.PathProfile, error) {\n\tquery := `SELECT pp.id, pp.profile_id, pp.path_id, pa.path, pa.path_name,\n\t\t\t\t\tpa.description, pp.post, pp.put, pp.del, pp.get,\n\t\t\t\t\tpp.created_at, pp.updated_at\n\t\t\t\tFROM path_profile AS pp INNER JOIN paths AS pa ON pp.path_id = pa.id\n\t\t\t\tWHERE profile_id = $1 ORDER BY pa.path`\n\n\tops := make([]models.PathProfile, 0)\n\tdb := get()\n\tdefer db.Close()\n\n\tstmt, err := db.Prepare(query)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer stmt.Close()\n\n\trows, err := stmt.Query(id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer rows.Close()\n\n\tfor rows.Next() {\n\t\tvar o models.PathProfile\n\t\terr = rows.Scan(&o.ID, &o.ProfileID, &o.Path.ID, &o.Path.Path, &o.Path.PathName, &o.Path.Description, &o.Post, &o.Put, &o.Del, &o.Get, &o.CreatedAt, &o.UpdatedAt)\n\t\tif err != nil {\n\t\t\treturn ops, err\n\t\t}\n\t\tops = append(ops, o)\n\t}\n\treturn ops, nil\n}", "func getProfiles() {\n\n\tvar configDir string\n\tif isRunningInDockerContainer() {\n\t\tconfigDir = filepath.Join(\"app\", \"config\")\n\t} else {\n\t\t_, fn, _, ok := runtime.Caller(0)\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\t\tindex := strings.LastIndex(fn, \"chart-verifier/\")\n\t\tconfigDir = fn[0 : index+len(\"chart-verifier\")]\n\t\tconfigDir = filepath.Join(configDir, \"config\")\n\t}\n\n\tfilepath.Walk(configDir, func(path string, info os.FileInfo, err error) error {\n\t\tif info != nil {\n\t\t\tif strings.HasSuffix(info.Name(), \".yaml\") {\n\t\t\t\tprofileRead, err := readProfile(path)\n\t\t\t\tif err == nil {\n\t\t\t\t\t// If version is not valid set to a default version\n\t\t\t\t\tif !semver.IsValid(profileRead.Version) {\n\t\t\t\t\t\tprofileRead.Version = DefaultProfileVersion\n\t\t\t\t\t}\n\t\t\t\t\tif len(profileRead.Vendor) == 0 {\n\t\t\t\t\t\tprofileRead.Vendor = VendorTypeNotSpecified\n\t\t\t\t\t}\n\t\t\t\t\tprofileMap[profileRead.Vendor] = append(profileMap[profileRead.Vendor], profileRead)\n\t\t\t\t\tprofileRead.Name = strings.Split(info.Name(), \".yaml\")[0]\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t})\n\treturn\n}", "func (p *PCE) GetContainerWkldProfiles(queryParameters map[string]string, containerClusterID string) (api APIResponse, err error) {\n\tp.ContainerWorkloadProfilesSlice = nil\n\tapi, err = p.GetCollection(\"container_clusters/\"+containerClusterID+\"/container_workload_profiles\", false, queryParameters, &p.ContainerWorkloadProfilesSlice)\n\tif len(p.ContainerWorkloadProfilesSlice) >= 500 {\n\t\tp.ContainerWorkloadProfilesSlice = nil\n\t\tapi, err = p.GetCollection(\"container_clusters/\"+containerClusterID+\"/container_workload_profiles\", true, queryParameters, &p.ContainerWorkloadProfilesSlice)\n\t}\n\tp.ContainerWorkloadProfiles = make(map[string]ContainerWorkloadProfile)\n\tfor _, c := range p.ContainerWorkloadProfilesSlice {\n\t\tp.ContainerWorkloadProfiles[c.Href] = c\n\t\tif PtrToVal(c.Name) != \"\" {\n\t\t\tp.ContainerWorkloadProfiles[*c.Name] = c\n\t\t}\n\t}\n\treturn api, err\n}", "func (l *RemoteProvider) GetPerformanceProfiles(req *http.Request, page, pageSize, search, order string) ([]byte, error) {\n\tif !l.Capabilities.IsSupported(PersistPerformanceProfiles) {\n\t\tlogrus.Error(\"operation not available\")\n\t\treturn []byte{}, ErrInvalidCapability(\"PersistPerformanceProfiles\", l.ProviderName)\n\t}\n\n\tep, _ := l.Capabilities.GetEndpointForFeature(PersistPerformanceProfiles)\n\n\tlogrus.Infof(\"attempting to fetch performance profiles from cloud\")\n\n\tremoteProviderURL, _ := url.Parse(l.RemoteProviderURL + ep)\n\tq := remoteProviderURL.Query()\n\tif page != \"\" {\n\t\tq.Set(\"page\", page)\n\t}\n\tif pageSize != \"\" {\n\t\tq.Set(\"page_size\", pageSize)\n\t}\n\tif search != \"\" {\n\t\tq.Set(\"search\", search)\n\t}\n\tif order != \"\" {\n\t\tq.Set(\"order\", order)\n\t}\n\tremoteProviderURL.RawQuery = q.Encode()\n\tlogrus.Debugf(\"constructed performance profiles url: %s\", remoteProviderURL.String())\n\tcReq, _ := http.NewRequest(http.MethodGet, remoteProviderURL.String(), nil)\n\n\ttokenString, err := l.GetToken(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp, err := l.DoRequest(cReq, tokenString)\n\tif err != nil {\n\t\treturn nil, ErrFetch(err, \"Perf Profile Page\", resp.StatusCode)\n\t}\n\tdefer func() {\n\t\t_ = resp.Body.Close()\n\t}()\n\tbdr, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn nil, ErrDataRead(err, \"Perf Profile Page\")\n\t}\n\n\tif resp.StatusCode == http.StatusOK {\n\t\tlogrus.Infof(\"performance profiles successfully retrieved from remote provider\")\n\t\treturn bdr, nil\n\t}\n\treturn nil, ErrPost(err, fmt.Sprint(bdr), resp.StatusCode)\n}", "func (api *dscprofileAPI) List(ctx context.Context, opts *api.ListWatchOptions) ([]*DSCProfile, error) {\n\tvar objlist []*DSCProfile\n\tobjs, err := api.ct.List(\"DSCProfile\", ctx, opts)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor _, obj := range objs {\n\t\tswitch tp := obj.(type) {\n\t\tcase *DSCProfile:\n\t\t\teobj := obj.(*DSCProfile)\n\t\t\tobjlist = append(objlist, eobj)\n\t\tdefault:\n\t\t\tlog.Fatalf(\"Got invalid object type %v while looking for DSCProfile\", tp)\n\t\t}\n\t}\n\n\treturn objlist, nil\n}", "func (l *configLoader) GetProfiles() ([]string, error) {\n\tpath := l.ConfigPath()\n\tbytes, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trawMap := map[interface{}]interface{}{}\n\terr = yaml.Unmarshal(bytes, &rawMap)\n\tif err != nil {\n\t\treturn nil, errors.Errorf(\"Error parsing devspace.yaml: %v\", err)\n\t}\n\n\tprofiles, ok := rawMap[\"profiles\"].([]interface{})\n\tif !ok {\n\t\tprofiles = []interface{}{}\n\t}\n\n\tprofileNames := []string{}\n\tfor _, profile := range profiles {\n\t\tprofileMap, ok := profile.(map[interface{}]interface{})\n\t\tif !ok {\n\t\t\tcontinue\n\t\t}\n\n\t\tname, ok := profileMap[\"name\"].(string)\n\t\tif !ok {\n\t\t\tcontinue\n\t\t}\n\n\t\tprofileNames = append(profileNames, name)\n\t}\n\n\treturn profileNames, nil\n}", "func (p *profile) List(options map[string]interface{}) (riminderResponse.ProfileListElem, error) {\n\tsourceIDs, err := json.Marshal(options[\"source_ids\"])\n\tif err != nil {\n\t\treturn riminderResponse.ProfileListElem{}, fmt.Errorf(\"profile.List:Cannot parse source ids (should be a list of string): %v\", err)\n\t}\n\n\tquery := map[string]string{\n\t\t\"source_ids\": string(sourceIDs),\n\t\t\"sort_by\": \"ranking\",\n\t\t\"page\": \"1\",\n\t\t\"date_end\": strconv.Itoa(int(time.Now().Unix())),\n\t\t\"date_start\": defaultDateStart,\n\t}\n\tAddIfNotEmptyStrMap(&query, options, \"seniority\")\n\tAddIfNotEmptyStrMap(&query, options, \"filter_id\")\n\tAddIfNotEmptyStrMap(&query, options, \"filter_reference\")\n\tAddIfNotEmptyStrMap(&query, options, \"stage\")\n\tAddIfNotEmptyStrMap(&query, options, \"rating\")\n\tAddIfNotEmptyStrMap(&query, options, \"date_start\")\n\tAddIfNotEmptyStrMap(&query, options, \"date_end\")\n\tAddIfNotEmptyStrMap(&query, options, \"page\")\n\tAddIfNotEmptyStrMap(&query, options, \"limit\")\n\tAddIfNotEmptyStrMap(&query, options, \"sort_by\")\n\tAddIfNotEmptyStrMap(&query, options, \"order_by\")\n\n\tresp := riminderResponse.ProfileListContainer{}\n\terr = p.client.Get(\"profiles\", query, &resp)\n\tif err != nil {\n\t\treturn riminderResponse.ProfileListElem{}, err\n\t}\n\treturn resp.Data, nil\n}", "func (bav *UtxoView) GetAllProfiles(readerPK []byte) (\n\t_profiles map[PkMapKey]*ProfileEntry,\n\t_corePostsByProfilePublicKey map[PkMapKey][]*PostEntry,\n\t_commentsByProfilePublicKey map[PkMapKey][]*PostEntry,\n\t_postEntryReaderStates map[BlockHash]*PostEntryReaderState, _err error) {\n\t// Start by fetching all the profiles we have in the db.\n\t//\n\t// TODO(performance): This currently fetches all profiles. We should implement\n\t// some kind of pagination instead though.\n\t_, _, dbProfileEntries, err := DBGetAllProfilesByCoinValue(bav.Handle, true /*fetchEntries*/)\n\tif err != nil {\n\t\treturn nil, nil, nil, nil, errors.Wrapf(\n\t\t\terr, \"GetAllProfiles: Problem fetching ProfileEntrys from db: \")\n\t}\n\n\t// Iterate through the entries found in the db and force the view to load them.\n\t// This fills in any gaps in the view so that, after this, the view should contain\n\t// the union of what it had before plus what was in the db.\n\tfor _, dbProfileEntry := range dbProfileEntries {\n\t\tbav.GetProfileEntryForPublicKey(dbProfileEntry.PublicKey)\n\t}\n\n\t// At this point, all the profiles should be loaded into the view.\n\n\t// Do one more pass to load all the comments associated with each\n\t// profile into the view.\n\tcommentsByProfilePublicKey := make(map[PkMapKey][]*PostEntry)\n\tfor _, profileEntry := range bav.ProfilePKIDToProfileEntry {\n\t\t// Ignore deleted or rolled-back posts.\n\t\tif profileEntry.isDeleted {\n\t\t\tcontinue\n\t\t}\n\t\tcommentsByProfilePublicKey[MakePkMapKey(profileEntry.PublicKey)] = []*PostEntry{}\n\t\t_, dbCommentHashes, _, err := DBGetCommentPostHashesForParentStakeID(\n\t\t\tbav.Handle, profileEntry.PublicKey, false /*fetchEntries*/)\n\t\tif err != nil {\n\t\t\treturn nil, nil, nil, nil, errors.Wrapf(err, \"GetAllPosts: Problem fetching comment PostEntry's from db: \")\n\t\t}\n\t\tfor _, commentHash := range dbCommentHashes {\n\t\t\tbav.GetPostEntryForPostHash(commentHash)\n\t\t}\n\t}\n\t// TODO(performance): Because we want to load all the posts the profile\n\t// has made, just go ahead and load *all* the posts into the view so that\n\t// they'll get returned in the mapping. Later, we should use the db index\n\t// to do this.\n\t_, _, dbPostEntries, err := DBGetAllPostsByTstamp(bav.Handle, true /*fetchEntries*/)\n\tif err != nil {\n\t\treturn nil, nil, nil, nil, errors.Wrapf(\n\t\t\terr, \"GetAllPosts: Problem fetching PostEntry's from db: \")\n\t}\n\tfor _, dbPostEntry := range dbPostEntries {\n\t\tbav.GetPostEntryForPostHash(dbPostEntry.PostHash)\n\t}\n\n\t// Iterate through all the posts loaded into the view and attach them\n\t// to the relevant profiles. Also adds reader state if a reader pubkey is provided.\n\tcorePostsByPublicKey := make(map[PkMapKey][]*PostEntry)\n\tpostEntryReaderStates := make(map[BlockHash]*PostEntryReaderState)\n\tfor _, postEntry := range bav.PostHashToPostEntry {\n\t\t// Ignore deleted or rolled-back posts.\n\t\tif postEntry.isDeleted {\n\t\t\tcontinue\n\t\t}\n\n\t\t// If the post has a stakeID that corresponds to a profile then add\n\t\t// it to our map.\n\t\t// Every post is either a core post or a comment. If it has a stake ID\n\t\t// its a comment, and if it doesn't then it's a core post.\n\t\tif len(postEntry.ParentStakeID) == 0 {\n\t\t\t// In this case we are dealing with a \"core\" post so add it to the\n\t\t\t// core post map.\n\t\t\tpostEntry.stakeStats = GetStakeEntryStats(postEntry.StakeEntry, bav.Params)\n\t\t\tcorePostsForProfile := corePostsByPublicKey[MakePkMapKey(postEntry.PosterPublicKey)]\n\t\t\tcorePostsForProfile = append(corePostsForProfile, postEntry)\n\t\t\tcorePostsByPublicKey[MakePkMapKey(postEntry.PosterPublicKey)] = corePostsForProfile\n\t\t} else {\n\t\t\t// Add the comment to our map.\n\t\t\tcommentsForProfile := commentsByProfilePublicKey[MakePkMapKey(postEntry.ParentStakeID)]\n\t\t\tcommentsForProfile = append(commentsForProfile, postEntry)\n\t\t\tcommentsByProfilePublicKey[MakePkMapKey(postEntry.ParentStakeID)] = commentsForProfile\n\t\t}\n\n\t\t// Create reader state map. Ie, whether the reader has liked the post, etc.\n\t\t// If nil is passed in as the readerPK, this is skipped.\n\t\tif readerPK != nil {\n\t\t\tpostEntryReaderState := bav.GetPostEntryReaderState(readerPK, postEntry)\n\t\t\tpostEntryReaderStates[*postEntry.PostHash] = postEntryReaderState\n\t\t}\n\t}\n\n\t// Now that the view mappings are a complete picture, iterate through them\n\t// and set them on the map we're returning.\n\tprofilesByPublicKey := make(map[PkMapKey]*ProfileEntry)\n\tfor _, profileEntry := range bav.ProfilePKIDToProfileEntry {\n\t\t// Ignore deleted or rolled-back posts.\n\t\tif profileEntry.isDeleted {\n\t\t\tcontinue\n\t\t}\n\t\tprofileEntry.stakeStats = GetStakeEntryStats(profileEntry.StakeEntry, bav.Params)\n\t\tprofilesByPublicKey[MakePkMapKey(profileEntry.PublicKey)] = profileEntry\n\t}\n\n\t// Sort the posts for each profile by when their stake.\n\tfor _, postsForProfile := range corePostsByPublicKey {\n\t\tsort.Slice(postsForProfile, func(ii, jj int) bool {\n\t\t\treturn postsForProfile[ii].stakeStats.TotalStakeNanos > postsForProfile[jj].stakeStats.TotalStakeNanos\n\t\t})\n\t}\n\t// Sort all the comment lists. Here we put the latest comment at the\n\t// end.\n\tfor _, commentList := range commentsByProfilePublicKey {\n\t\tsort.Slice(commentList, func(ii, jj int) bool {\n\t\t\treturn commentList[ii].TimestampNanos < commentList[jj].TimestampNanos\n\t\t})\n\t}\n\n\treturn profilesByPublicKey, corePostsByPublicKey, commentsByProfilePublicKey, postEntryReaderStates, nil\n}", "func (c *Config) GetProfile(profile string) ([]Entry, bool) {\n\tlog.Println(\"[DEBUG] GetProfile\", profile)\n\tentries, ok := c.Profiles[profile]\n\tif ok {\n\t\treturn entries, true\n\t}\n\n\treturn nil, false\n}", "func GenerateSubProfileQuery() (string, error) {\n\treturn globalProbe.GenerateSubProfileQuery()\n}", "func GetProfiles() []AWS_Named_Profile {\n\txp := make([]AWS_Named_Profile, 0)\n\n\t// Get the userprofile from powershell.\n\tback := &backend.Local{}\n\tshell, err := ps.New(back)\n\tdefer shell.Exit()\n\tcheck(err)\n\tstdout, _, err := shell.Execute(\"$env:userprofile\")\n\tcheck(err)\n\n\t// Load config and remove carriage returns.\n\tre := regexp.MustCompile(`\\r?\\n`)\n\tawsconfig := re.ReplaceAllString(stdout, \"\") + \"\\\\.aws\\\\config\"\n\tcfg, err := ini.Load(awsconfig)\n\tif err != nil {\n\t\tlog.Printf(\"Fail to read file: %v\", err)\n\t\tos.Exit(1)\n\t}\n\n\t// Find Valid/Enabled Named profiles. EG: default profile has no arn.\n\tfor _, section := range cfg.Sections() {\n\t\tif section.HasKey(\"role_arn\") {\n\t\t\tvar p AWS_Named_Profile\n\t\t\tp.Name = section.Name()\n\t\t\tp.Arn = section.Key(\"role_arn\").String()\n\t\t\txp = append(xp, p)\n\t\t}\n\t}\n\t//return the named profiles\n\tlog.Printf(\"Found %v Named Profile(s) in: %v\", len(xp), awsconfig)\n\treturn xp\n}", "func testProfileList(t *testing.T) {\n\tp := profileName(t)\n\tt.Parallel()\n\tmk := NewMinikubeRunner(t, p, \"--wait=false\")\n\tout, stderr := mk.MustRun(\"profile list\")\n\tif !strings.Contains(out, p) {\n\t\tt.Errorf(\"Error , failed to read profile name (%s) in `profile list` command output : \\n %q : \\n stderr: %s \", p, out, stderr)\n\t}\n}", "func NewProfilesGetCollectionRequest(server string, params *ProfilesGetCollectionParams) (*http.Request, error) {\n\tvar err error\n\n\tserverURL, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\toperationPath := fmt.Sprintf(\"/v1/profiles\")\n\tif operationPath[0] == '/' {\n\t\toperationPath = \".\" + operationPath\n\t}\n\n\tqueryURL, err := serverURL.Parse(operationPath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryValues := queryURL.Query()\n\n\tif params.FilterName != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"filter[name]\", runtime.ParamLocationQuery, *params.FilterName); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FilterProfileState != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"filter[profileState]\", runtime.ParamLocationQuery, *params.FilterProfileState); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FilterProfileType != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"filter[profileType]\", runtime.ParamLocationQuery, *params.FilterProfileType); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FilterId != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"filter[id]\", runtime.ParamLocationQuery, *params.FilterId); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.Sort != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"sort\", runtime.ParamLocationQuery, *params.Sort); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FieldsProfiles != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[profiles]\", runtime.ParamLocationQuery, *params.FieldsProfiles); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.Limit != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", true, \"limit\", runtime.ParamLocationQuery, *params.Limit); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.Include != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"include\", runtime.ParamLocationQuery, *params.Include); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FieldsCertificates != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[certificates]\", runtime.ParamLocationQuery, *params.FieldsCertificates); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FieldsDevices != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[devices]\", runtime.ParamLocationQuery, *params.FieldsDevices); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FieldsBundleIds != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[bundleIds]\", runtime.ParamLocationQuery, *params.FieldsBundleIds); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.LimitCertificates != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", true, \"limit[certificates]\", runtime.ParamLocationQuery, *params.LimitCertificates); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.LimitDevices != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", true, \"limit[devices]\", runtime.ParamLocationQuery, *params.LimitDevices); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tqueryURL.RawQuery = queryValues.Encode()\n\n\treq, err := http.NewRequest(\"GET\", queryURL.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (cli *Client) PeopleProfile(id string, fields []string) (r map[string]interface{}, e error) {\n\tvar opt map[string]interface{}\n\n\tif len(fields) > 0 {\n\t\topt = map[string]interface{}{\n\t\t\t\"fields\": fields,\n\t\t}\n\t}\n\n\tr, e = cli.call(\"GET\", \"people\", id, \"\", opt)\n\n\treturn r, e\n}", "func (org *Org) GetAllNsxtAppPortProfiles(queryParameters url.Values, scope string) ([]*NsxtAppPortProfile, error) {\n\tqueryParams := copyOrNewUrlValues(queryParameters)\n\tif scope != \"\" {\n\t\tqueryParams = queryParameterFilterAnd(\"scope==\"+scope, queryParams)\n\t}\n\n\treturn getAllNsxtAppPortProfiles(org.client, queryParams)\n}", "func (mpRepo *ManageProfileRepository) Profile(id uint) (*entity.Profile, []error) {\n\tprfs := entity.Profile{}\n\terrs := mpRepo.conn.First(&prfs, id).GetErrors()\n\tif len(errs) > 0 {\n\t\treturn nil, errs\n\t}\n\treturn &prfs, errs\n}", "func (q *ColumnQueryAPI) ProfileTypes(ctx context.Context, req *pb.ProfileTypesRequest) (*pb.ProfileTypesResponse, error) {\n\ttypes, err := q.querier.ProfileTypes(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &pb.ProfileTypesResponse{\n\t\tTypes: types,\n\t}, nil\n}", "func (api *LaborStatsAPI) QueryCountryProfile() ([]CountryProfile, error) {\n\ta := CountryProfileAPI{\n\t\tDebug: api.Debug,\n\t\tSecretKey: api.SecretKey,\n\t}\n\n\terr := a.sendRequest()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres, err := a.unmarshalData()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn res, nil\n\n}", "func (p *EnvConfigProvider) ListProfiles(roles bool) []string {\n\treturn []string{}\n}", "func (bc *BoltClient) GetDeviceProfilesByCommandId(id string) ([]models.DeviceProfile, error) {\n\t// Check if this command exists\n\terr := bc.checkId(db.Command, id)\n\tif err != nil {\n\t\treturn []models.DeviceProfile{}, err\n\t}\n\n\treturn bc.getDeviceProfilesBy(func(encoded []byte) bool {\n\t\tcommands := jsoniter.Get(encoded, \"commands\").GetInterface().([]interface{})\n\t\tfor _, value := range commands {\n\t\t\tif id == value.(string) {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t\treturn false\n\t})\n}", "func GetUserProfiles(users ...User) ([]Profile, error) {\n\tctx := context.Background()\n\tclient, err := getFireBaseApp().Auth(ctx)\n\tif err != nil {\n\t\treturn []Profile{}, fmt.Errorf(\"error getting firebase app: %v\", err)\n\t}\n\tprofiles := make([]Profile, len(users))\n\tfor i, u := range users {\n\t\tp := Profile{UID: u.ID}\n\t\tuserRecord, err := client.GetUser(ctx, u.ID)\n\t\tif err == nil {\n\t\t\tp.Name = userRecord.DisplayName\n\t\t\tp.ProfilePicture = userRecord.PhotoURL\n\t\t}\n\t\tprofiles[i] = p\n\t}\n\treturn profiles, nil\n}", "func (s *Storer) ListByProfile(ctx context.Context, profileID string) ([]accounts.Account, error) {\n\tquery := listByProfileSQL(ctx, profileID)\n\tqueryStr, err := query.PostgreSQLString()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\trows, err := s.db.Query(queryStr, query.Args()...) //nolint:sqlclosecheck // the closeRows helper isn't picked up\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer closeRows(ctx, rows)\n\tvar accts []accounts.Account\n\tfor rows.Next() {\n\t\tvar account Account\n\t\terr = pan.Unmarshal(rows, &account)\n\t\tif err != nil {\n\t\t\treturn accts, err\n\t\t}\n\t\taccts = append(accts, fromPostgres(account))\n\t}\n\tif err = rows.Err(); err != nil {\n\t\treturn nil, err\n\t}\n\taccounts.ByLastUsedDesc(accts)\n\treturn accts, nil\n}", "func GenerateProfiles() director.GenerateProfilesFunc {\n\treturn func() ([]*pb.MatchProfile, error) {\n\t\tvar profiles []*pb.MatchProfile\n\n\t\tworlds := []string{\"Dune\", \"Nova\", \"Pandora\", \"Orion\"}\n\t\tregions := []string{\"us-east-1\", \"us-east-2\", \"us-west-1\", \"us-west-2\"}\n\n\t\tskillLevels := []*pb.DoubleRangeFilter{\n\t\t\t{DoubleArg: \"skill\", Min: 0, Max: 10},\n\t\t\t{DoubleArg: \"skill\", Min: 10, Max: 100},\n\t\t\t{DoubleArg: \"skill\", Min: 100, Max: 1000},\n\t\t}\n\n\t\tlatencies := []*pb.DoubleRangeFilter{\n\t\t\t{DoubleArg: \"latency\", Min: 0, Max: 25},\n\t\t\t{DoubleArg: \"latency\", Min: 25, Max: 50},\n\t\t\t{DoubleArg: \"latency\", Min: 50, Max: 75},\n\t\t\t{DoubleArg: \"latency\", Min: 75, Max: 100},\n\t\t}\n\n\t\tfor _, world := range worlds {\n\t\t\tfor _, region := range regions {\n\t\t\t\t//region := TagFromStringSlice(regions)\n\n\t\t\t\tprofile := &pb.MatchProfile{\n\t\t\t\t\tName: fmt.Sprintf(\"world_based_profile_%s_%s\", world, region),\n\t\t\t\t\tPools: []*pb.Pool{\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tName: \"pool_mode_\" + world,\n\t\t\t\t\t\t\tTagPresentFilters: []*pb.TagPresentFilter{\n\t\t\t\t\t\t\t\t{Tag: \"mode.session\"},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tStringEqualsFilters: []*pb.StringEqualsFilter{\n\t\t\t\t\t\t\t\t{StringArg: \"world\", Value: world},\n\t\t\t\t\t\t\t\t{StringArg: \"region\", Value: region},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tDoubleRangeFilters: []*pb.DoubleRangeFilter{\n\t\t\t\t\t\t\t\tDoubleRangeFilterFromSlice(skillLevels),\n\t\t\t\t\t\t\t\tDoubleRangeFilterFromSlice(latencies),\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t}\n\n\t\t\t\t// build filter extensions\n\t\t\t\tfilter := extensions.AllocatorFilterExtension{\n\t\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t\t\"region\": region,\n\t\t\t\t\t\t\"world\": world,\n\t\t\t\t\t},\n\t\t\t\t\tFields: map[string]string{\n\t\t\t\t\t\t\"status.state\": \"Ready\",\n\t\t\t\t\t},\n\t\t\t\t}\n\n\t\t\t\t// Multiples Extensions: extensions.WithAny(filter.Any()).WithAny(foo.Any()).WithAny(bar.Any()).Extensions()\n\t\t\t\tprofile.Extensions = extensions.WithAny(filter.Any()).Extensions()\n\t\t\t\tprofiles = append(profiles, profile)\n\t\t\t}\n\t\t}\n\n\t\treturn profiles, nil\n\t}\n}", "func (svc *inmemService) GetProfile(ctx context.Context, id string) (Profile, error) {\n\t// Get the Read lock from the inmemService struct\n\tsvc.mtx.RLock()\n\n\t// Immediately set up a lock release to occur when the function finishes\n\tdefer svc.mtx.RUnlock()\n\n\t// Look for the profile by the `id` function param\n\tprofile, ok := svc.profiles[id]\n\n\t// Check if the profile id was not found in the datastore\n\tif !ok {\n\n\t\t// Return an empty profile and an error informing the caller that the profile was not found\n\t\treturn Profile{}, ErrNotFound\n\n\t}\n\n\t// Return the profile to the caller and a nil error\n\treturn profile, nil\n\n}", "func (api *API) ListDLPProfiles(ctx context.Context, rc *ResourceContainer, params ListDLPProfilesParams) ([]DLPProfile, error) {\n\tif rc.Identifier == \"\" {\n\t\treturn []DLPProfile{}, ErrMissingResourceIdentifier\n\t}\n\n\turi := buildURI(fmt.Sprintf(\"/%s/%s/dlp/profiles\", rc.Level, rc.Identifier), nil)\n\n\tres, err := api.makeRequestContext(ctx, http.MethodGet, uri, nil)\n\tif err != nil {\n\t\treturn []DLPProfile{}, err\n\t}\n\n\tvar dlpProfilesListResponse DLPProfileListResponse\n\terr = json.Unmarshal(res, &dlpProfilesListResponse)\n\tif err != nil {\n\t\treturn []DLPProfile{}, fmt.Errorf(\"%s: %w\", errUnmarshalError, err)\n\t}\n\n\treturn dlpProfilesListResponse.Result, nil\n}", "func (repository *Datastore)GetProfile(username string)(*user.Person,error){\n\tperson := newUser() //initialize user.Person and will used to store profile info\n\tquery := `SELECT * FROM userRepository WHERE username = ?`\n\terr := repository.Db.Get(&person, query, username) //get person profile details\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &person, nil\n}", "func (s *EmptyStore) ProfileGet(id string) (*storagepb.Profile, error) {\n\treturn nil, fmt.Errorf(\"Profile not found\")\n}", "func (a *Client) GetMsgVpnClientProfiles(params *GetMsgVpnClientProfilesParams, authInfo runtime.ClientAuthInfoWriter) (*GetMsgVpnClientProfilesOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetMsgVpnClientProfilesParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"getMsgVpnClientProfiles\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/msgVpns/{msgVpnName}/clientProfiles\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &GetMsgVpnClientProfilesReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetMsgVpnClientProfilesOK), nil\n\n}", "func (f *IBMPIInstanceClient) GetSAPProfiles(powerinstanceid string) (*models.SAPProfiles, error) {\n\n\tparams := p_cloud_s_a_p.NewPcloudSapGetallParamsWithTimeout(helpers.PIGetTimeOut).WithCloudInstanceID(powerinstanceid)\n\tresp, err := f.session.Power.PCloudSAP.PcloudSapGetall(params, ibmpisession.NewAuth(f.session, powerinstanceid))\n\tif err != nil || resp.Payload == nil {\n\t\treturn nil, fmt.Errorf(\"Failed to get sap profiles to the power instance %s : %s\", powerinstanceid, err)\n\t}\n\treturn resp.Payload, nil\n}", "func List() {\n\n\tprofiles := GetProfileList()\n\n\tfmt.Println(\"\")\n\tfmt.Println(\"Available profiles in the JSon repository (\" + getProfilesPath() + \"):\")\n\tw := tabwriter.NewWriter(os.Stdout, 0, 0, 5, ' ', 0)\n\n\tfor _, profileName := range profiles {\n\n\t\tprofileData := getProfileFactory(getProfilesPath(), profileName)\n\n\t\tfmt.Fprintln(w, profileName+\"\\t Image => \"+profileData.Image+\"\\t Alias => \"+profileData.Alias)\n\n\t}\n\tw.Flush()\n\tfmt.Println(\" \")\n}", "func (p *profileDocument) List(options map[string]interface{}) ([]riminderResponse.ProfileDocumentsListElem, error) {\n\tquery := map[string]string{\n\t\t\"source_id\": options[\"source_id\"].(string),\n\t}\n\tAddIfNotEmptyStrMap(&query, options, \"profile_id\")\n\tAddIfNotEmptyStrMap(&query, options, \"profile_reference\")\n\n\tresp := riminderResponse.ProfileDocumentsListContainer{}\n\terr := p.client.Get(\"profile/documents\", query, &resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp.Data, nil\n}", "func (r *UserProfilesService) List() *UserProfilesListCall {\n\tc := &UserProfilesListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\treturn c\n}", "func GetCmdQueryActiveClaims(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: treasury.QueryActiveClaims,\n\t\tShort: \"Query claims that have yet to be redeemed by the treasury\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, treasury.QueryActiveClaims), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar claims types.ClaimPool\n\t\t\tcdc.MustUnmarshalBinaryLengthPrefixed(res, &claims)\n\t\t\treturn cliCtx.PrintOutput(claims)\n\t\t},\n\t}\n\n\treturn cmd\n}", "func (p *Profiler) All() []Profile {\n\tif p == nil {\n\t\t// If the profiler instance doesn't exist, then don't attempt to operate on it.\n\t\treturn []Profile{}\n\t}\n\tp.mutex.Lock()\n\tdefer p.mutex.Unlock()\n\treturn p.profiles\n}", "func getProfile(name string) (*profile.Specification, error) {\n\tprofiles, err := profiles.GetKnownProfiles()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor _, p := range profiles {\n\t\tif p.Label == name {\n\t\t\treturn p, nil\n\t\t}\n\t}\n\treturn nil, nil\n\n\t// TODO(jsimsa): This function assumes the existence of a profile\n\t// server from which the profiles can be retrieved. The profile\n\t// server is a work in progress. When it exists, the commented out\n\t// code below should work.\n\t/*\n\t\tvar profile profile.Specification\n\t\t\t\tclient, err := r.NewClient()\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, verror.New(ErrOperationFailed, nil, fmt.Sprintf(\"NewClient() failed: %v\", err))\n\t\t\t\t}\n\t\t\t\tdefer client.Close()\n\t\t\t server := // TODO\n\t\t\t\tmethod := \"Specification\"\n\t\t\t\tinputs := make([]interface{}, 0)\n\t\t\t\tcall, err := client.StartCall(server + \"/\" + name, method, inputs)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, verror.New(ErrOperationFailed, nil, fmt.Sprintf(\"StartCall(%s, %q, %v) failed: %v\\n\", server + \"/\" + name, method, inputs, err))\n\t\t\t\t}\n\t\t\t\tif err := call.Finish(&profiles); err != nil {\n\t\t\t\t\treturn nil, verror.New(ErrOperationFailed, nil, fmt.Sprintf(\"Finish(%v) failed: %v\\n\", &profiles, err))\n\t\t\t\t}\n\t\treturn &profile, nil\n\t*/\n}", "func (p *profile) Get(options map[string]interface{}) (riminderResponse.ProfileGetElem, error) {\n\tquery := map[string]string{\n\t\t\"source_id\": options[\"source_id\"].(string),\n\t}\n\tAddIfNotEmptyStrMap(&query, options, \"profile_id\")\n\tAddIfNotEmptyStrMap(&query, options, \"profile_reference\")\n\n\tresp := riminderResponse.ProfileGetContainer{}\n\terr := p.client.Get(\"profile\", query, &resp)\n\tif err != nil {\n\t\treturn riminderResponse.ProfileGetElem{}, err\n\t}\n\treturn resp.Data, nil\n}", "func (c *Cluster) GetInstancesWithProfile(project, profile string) (map[string][]string, error) {\n\terr := c.Transaction(context.TODO(), func(ctx context.Context, tx *ClusterTx) error {\n\t\tenabled, err := cluster.ProjectHasProfiles(context.Background(), tx.tx, project)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Check if project has profiles: %w\", err)\n\t\t}\n\n\t\tif !enabled {\n\t\t\tproject = \"default\"\n\t\t}\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tq := `SELECT instances.name, projects.name FROM instances\n\t\tJOIN instances_profiles ON instances.id == instances_profiles.instance_id\n\t\tJOIN projects ON projects.id == instances.project_id\n\t\tWHERE instances_profiles.profile_id ==\n\t\t (SELECT profiles.id FROM profiles\n\t\t JOIN projects ON projects.id == profiles.project_id\n\t\t WHERE profiles.name=? AND projects.name=?)`\n\n\tresults := map[string][]string{}\n\tinargs := []any{profile, project}\n\tvar name string\n\toutfmt := []any{name, name}\n\n\toutput, err := queryScan(c, q, inargs, outfmt)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor _, r := range output {\n\t\tif results[r[1].(string)] == nil {\n\t\t\tresults[r[1].(string)] = []string{}\n\t\t}\n\n\t\tresults[r[1].(string)] = append(results[r[1].(string)], r[0].(string))\n\t}\n\n\treturn results, nil\n}", "func (dpc *deviceProfileRestClient) requestDeviceProfileSlice(url string, ctx context.Context) ([]models.DeviceProfile, error) {\n\tdata, err := clients.GetRequest(url, ctx)\n\tif err != nil {\n\t\treturn []models.DeviceProfile{}, err\n\t}\n\n\tdpSlice := make([]models.DeviceProfile, 0)\n\terr = json.Unmarshal(data, &dpSlice)\n\treturn dpSlice, err\n}", "func GetProfile(ctx *router.Context) {\n\tuin, _ := ctx.Keys[middleware.ComerUinContextKey].(uint64)\n\tresponse, err := service.GetComerProfile(uin)\n\tif err != nil {\n\t\tctx.ERROR(\n\t\t\trouter.ErrBuisnessError,\n\t\t\t\"wrong metamask login parameter\",\n\t\t)\n\t\treturn\n\t}\n\n\tctx.OK(response)\n}", "func (r *UserProfilesService) Get(profileId int64) *UserProfilesGetCall {\n\tc := &UserProfilesGetCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.profileId = profileId\n\treturn c\n}", "func (s *Service) PermissionProfilesList() *PermissionProfilesListOp {\n\treturn &PermissionProfilesListOp{\n\t\tCredential: s.credential,\n\t\tMethod: \"GET\",\n\t\tPath: \"permission_profiles\",\n\t\tAccept: \"application/json\",\n\t\tQueryOpts: make(url.Values),\n\t\tVersion: esign.APIv2,\n\t}\n}", "func (a *ManagementApiService) GetCustomerProfiles(ctx _context.Context) apiGetCustomerProfilesRequest {\n\treturn apiGetCustomerProfilesRequest{\n\t\tapiService: a,\n\t\tctx: ctx,\n\t}\n}", "func GetProfile(_db Queryable, uid int64) (*Profile, error) {\n\tp := &Profile{}\n\terr := _db.QueryRow(`SELECT first_name, last_name, high_score, gender, \n img, birth_date, signup_date\n FROM profile WHERE uid = $1`, uid).Scan(\n\t\tp.FirstName, p.LastName, p.HighScore, p.Gender, p.Img, p.BirthDate, p.SignupDate)\n\tif err != nil {\n\t\treturn nil, err\n\t} else {\n\t\treturn p, err\n\t}\n}", "func (c *IRacing) GetProfile(ctx context.Context) (*UserProfile, error) {\n\n\tprofile := &UserProfile{}\n\n\terr := c.json(ctx, http.MethodGet, \"/membersite/member/GetMember\", nil, profile)\n\n\treturn profile, err\n}", "func (r *Radarr) GetReleaseProfilesContext(ctx context.Context) ([]*ReleaseProfile, error) {\n\tvar output []*ReleaseProfile\n\n\treq := starr.Request{URI: bpReleaseProfile}\n\tif err := r.GetInto(ctx, req, &output); err != nil {\n\t\treturn nil, fmt.Errorf(\"api.Get(%s): %w\", &req, err)\n\t}\n\n\treturn output, nil\n}", "func (p *Profiler) Flush() []Profile {\n\tif p == nil {\n\t\treturn []Profile{}\n\t}\n\tp.mutex.Lock()\n\tdefer p.mutex.Unlock()\n\tresult := p.profiles\n\tp.profiles = []Profile{}\n\treturn result\n}", "func (a *API) ProfileSearch(ctx context.Context, searchID string, pageNum int) (*ProfileResp, error) {\n\n\tall := pageNum < 1\n\tif all || pageNum < 1 {\n\t\tpageNum = 1\n\t}\n\n\tvar resp ProfileResp\n\turlStr := fmt.Sprintf(\"/api/v1/profile?searchId=%s&pageSize=100&pageNumber=%d\", searchID, pageNum)\n\tif err := a.Get(ctx, urlStr, &resp); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif all && resp.TotalPageCount > 1 {\n\t\tfor i := 1; i < resp.TotalPageCount; i++ {\n\t\t\tpg, err := a.ProfileSearch(ctx, searchID, i+1)\n\t\t\tif err != nil {\n\t\t\t\treturn &resp, err\n\t\t\t}\n\t\t\tresp.Profiles = append(resp.Profiles, pg.Profiles...)\n\t\t}\n\t}\n\n\treturn &resp, nil\n}", "func (s *ProfileService) ListContext(ctx context.Context, input *ListProfilesInput) (*ListProfilesOutput, *http.Response, error) {\n\tspath, err := addOptions(\"/profiles\", input)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treq, err := s.client.newRequest(ctx, \"GET\", spath, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tres, err := s.client.Do(ctx, req)\n\tif err != nil {\n\t\treturn nil, res, err\n\t}\n\n\tvar out ListProfilesOutput\n\tif err := decodeBody(res, &out); err != nil {\n\t\treturn nil, res, err\n\t}\n\n\treturn &out, res, nil\n}", "func IncludeProfiles(profiles ...string) {\n\tProfiles = append(Profiles, profiles...)\n}", "func (m MigProfile) GetProfileIDs() (int, int, int, error) {\n\terr := m.AssertValid()\n\tif err != nil {\n\t\treturn -1, -1, -1, fmt.Errorf(\"invalid MigProfile: %v\", err)\n\t}\n\n\tc, g, _, err := m.Parse()\n\tif err != nil {\n\t\treturn -1, -1, -1, fmt.Errorf(\"unable to parse MigProfile: %v\", err)\n\t}\n\n\tvar giProfileID, ciProfileID, ciEngProfileID int\n\n\tswitch g {\n\tcase 1:\n\t\tgiProfileID = nvml.GPU_INSTANCE_PROFILE_1_SLICE\n\tcase 2:\n\t\tgiProfileID = nvml.GPU_INSTANCE_PROFILE_2_SLICE\n\tcase 3:\n\t\tgiProfileID = nvml.GPU_INSTANCE_PROFILE_3_SLICE\n\tcase 4:\n\t\tgiProfileID = nvml.GPU_INSTANCE_PROFILE_4_SLICE\n\tcase 7:\n\t\tgiProfileID = nvml.GPU_INSTANCE_PROFILE_7_SLICE\n\tcase 8:\n\t\tgiProfileID = nvml.GPU_INSTANCE_PROFILE_8_SLICE\n\tdefault:\n\t\treturn -1, -1, -1, fmt.Errorf(\"unknown GPU Instance slice size: %v\", g)\n\t}\n\n\tswitch c {\n\tcase 1:\n\t\tciProfileID = nvml.COMPUTE_INSTANCE_PROFILE_1_SLICE\n\tcase 2:\n\t\tciProfileID = nvml.COMPUTE_INSTANCE_PROFILE_2_SLICE\n\tcase 3:\n\t\tciProfileID = nvml.COMPUTE_INSTANCE_PROFILE_3_SLICE\n\tcase 4:\n\t\tciProfileID = nvml.COMPUTE_INSTANCE_PROFILE_4_SLICE\n\tcase 7:\n\t\tciProfileID = nvml.COMPUTE_INSTANCE_PROFILE_7_SLICE\n\tcase 8:\n\t\tciProfileID = nvml.COMPUTE_INSTANCE_PROFILE_8_SLICE\n\tdefault:\n\t\treturn -1, -1, -1, fmt.Errorf(\"unknown Compute Instance slice size: %v\", c)\n\t}\n\n\tciEngProfileID = nvml.COMPUTE_INSTANCE_ENGINE_PROFILE_SHARED\n\n\treturn giProfileID, ciProfileID, ciEngProfileID, nil\n}", "func (a *Client) ListRunProfiles(params *ListRunProfilesParams, authInfo runtime.ClientAuthInfoWriter) (*ListRunProfilesOK, *ListRunProfilesNoContent, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListRunProfilesParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"ListRunProfiles\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/v1/orgs/{owner}/run_profiles\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &ListRunProfilesReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tswitch value := result.(type) {\n\tcase *ListRunProfilesOK:\n\t\treturn value, nil, nil\n\tcase *ListRunProfilesNoContent:\n\t\treturn nil, value, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*ListRunProfilesDefault)\n\treturn nil, nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func (r *OAuthProfileResource) ListAll() (*OAuthProfileConfigList, error) {\n\tvar list OAuthProfileConfigList\n\tif err := r.c.ReadQuery(BasePath+OAuthProfileEndpoint, &list); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &list, nil\n}", "func GetManageAccountProfilesByPhoneID(phone string) (AuthProfile, error) {\n\tdata := new(AuthProfile)\n\treq, reqErr := CreateRequest(\"GET\", os.Getenv(\"DOMAIN\")+\"/identity/v2/manage/account\", \"\")\n\tif reqErr != nil {\n\t\treturn *data, reqErr\n\t}\n\n\tq := req.URL.Query()\n\tq.Add(\"phone\", phone)\n\treq.URL.RawQuery = q.Encode()\n\treq.Header.Add(\"content-Type\", \"application/x-www-form-urlencoded\")\n\treq.Header.Add(\"X-LoginRadius-ApiKey\", os.Getenv(\"APIKEY\"))\n\treq.Header.Add(\"X-LoginRadius-ApiSecret\", os.Getenv(\"APISECRET\"))\n\n\terr := RunRequest(req, data)\n\treturn *data, err\n}", "func getCompleteProfile(ctx sdk.Context, path []string, k Keeper) (res []byte, sdkError error) {\n\tid := path[0]\n\tuser, err := k.GetUser(ctx, id)\n\tif err != nil {\n\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrInvalidAddress, \"user with address not found\")\n\t}\n\n\tthoughts := k.ListThoughtByCreator(ctx, user.Creator)\n\n\tvar followers []types.User\n\tvar followings []types.User\n\n\tfor _, addr := range user.Followers {\n\t\tfollower, err := k.GetUser(ctx, addr.String())\n\t\tif err == nil {\n\t\t\tfollowers = append(followers, follower)\n\t\t}\n\t}\n\n\tfor _, addr := range user.Following {\n\t\tfollowing, err := k.GetUser(ctx, addr.String())\n\t\tif err == nil {\n\t\t\tfollowings = append(followings, following)\n\t\t}\n\t}\n\n\tvar completeProfile = types.CompleteProfile{\n\t\tCreator: user.Creator,\n\t\tID: user.ID,\n\t\tUsername: user.Username,\n\t\tBio: user.Bio,\n\t\tAvatar: user.Avatar,\n\t\tFollowing: followings,\n\t\tFollowers: followers,\n\t\tThoughts: thoughts,\n\t}\n\n\tres, err = codec.MarshalJSONIndent(k.cdc, completeProfile)\n\tif err != nil {\n\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrJSONMarshal, err.Error())\n\t}\n\n\treturn res, nil\n}", "func newCmdProfile() *cobra.Command {\n\toptions := newProfileOptions()\n\n\tcmd := &cobra.Command{\n\t\tUse: \"profile [flags] (--template | --open-api file | --proto file) (SERVICE)\",\n\t\tShort: \"Output service profile config for Kubernetes\",\n\t\tLong: \"Output service profile config for Kubernetes.\",\n\t\tExample: ` # Output a basic template to apply after modification.\n linkerd profile -n emojivoto --template web-svc\n\n # Generate a profile from an OpenAPI specification.\n linkerd profile -n emojivoto --open-api web-svc.swagger web-svc\n\n # Generate a profile from a protobuf definition.\n linkerd profile -n emojivoto --proto Voting.proto vote-svc\n`,\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tif options.namespace == \"\" {\n\t\t\t\toptions.namespace = pkgcmd.GetDefaultNamespace(kubeconfigPath, kubeContext)\n\t\t\t}\n\t\t\toptions.name = args[0]\n\t\t\tclusterDomain := defaultClusterDomain\n\n\t\t\terr := options.validate()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\t// performs an online profile generation and access-check to k8s cluster to extract\n\t\t\t// clusterDomain from linkerd configuration\n\t\t\tif !options.ignoreCluster {\n\t\t\t\tvar err error\n\t\t\t\tk8sAPI, err := k8s.NewAPI(kubeconfigPath, kubeContext, impersonate, impersonateGroup, 0)\n\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\t_, values, err := healthcheck.FetchCurrentConfiguration(cmd.Context(), k8sAPI, controlPlaneNamespace)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tif cd := values.ClusterDomain; cd != \"\" {\n\t\t\t\t\tclusterDomain = cd\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif options.template {\n\t\t\t\treturn profiles.RenderProfileTemplate(options.namespace, options.name, clusterDomain, os.Stdout)\n\t\t\t} else if options.openAPI != \"\" {\n\t\t\t\treturn profiles.RenderOpenAPI(options.openAPI, options.namespace, options.name, clusterDomain, os.Stdout)\n\t\t\t} else if options.proto != \"\" {\n\t\t\t\treturn profiles.RenderProto(options.proto, options.namespace, options.name, clusterDomain, os.Stdout)\n\t\t\t}\n\n\t\t\t// we should never get here\n\t\t\treturn errors.New(\"Unexpected error\")\n\t\t},\n\t}\n\n\tcmd.PersistentFlags().BoolVar(&options.template, \"template\", options.template, \"Output a service profile template\")\n\tcmd.PersistentFlags().StringVar(&options.openAPI, \"open-api\", options.openAPI, \"Output a service profile based on the given OpenAPI spec file\")\n\tcmd.PersistentFlags().StringVarP(&options.namespace, \"namespace\", \"n\", options.namespace, \"Namespace of the service\")\n\tcmd.PersistentFlags().StringVar(&options.proto, \"proto\", options.proto, \"Output a service profile based on the given Protobuf spec file\")\n\tcmd.PersistentFlags().BoolVar(&options.ignoreCluster, \"ignore-cluster\", options.ignoreCluster, \"Output a service profile through offline generation\")\n\n\treturn cmd\n}", "func (a UsersApi) GetProfilesUsers(pageSize int, pageNumber int, id []string, jid []string, sortOrder string, expand []string, integrationPresenceSource string) (*Userprofileentitylisting, *APIResponse, error) {\n\tvar httpMethod = \"GET\"\n\t// create path and map variables\n\tpath := a.Configuration.BasePath + \"/api/v2/profiles/users\"\n\tdefaultReturn := new(Userprofileentitylisting)\n\tif true == false {\n\t\treturn defaultReturn, nil, errors.New(\"This message brought to you by the laws of physics being broken\")\n\t}\n\n\n\theaderParams := make(map[string]string)\n\tqueryParams := make(map[string]string)\n\tformParams := url.Values{}\n\tvar postBody interface{}\n\tvar postFileName string\n\tvar fileBytes []byte\n\t// authentication (PureCloud OAuth) required\n\n\t// oauth required\n\tif a.Configuration.AccessToken != \"\"{\n\t\theaderParams[\"Authorization\"] = \"Bearer \" + a.Configuration.AccessToken\n\t}\n\t// add default headers if any\n\tfor key := range a.Configuration.DefaultHeader {\n\t\theaderParams[key] = a.Configuration.DefaultHeader[key]\n\t}\n\t\n\tqueryParams[\"pageSize\"] = a.Configuration.APIClient.ParameterToString(pageSize, \"\")\n\t\n\tqueryParams[\"pageNumber\"] = a.Configuration.APIClient.ParameterToString(pageNumber, \"\")\n\t\n\tqueryParams[\"id\"] = a.Configuration.APIClient.ParameterToString(id, \"multi\")\n\t\n\tqueryParams[\"jid\"] = a.Configuration.APIClient.ParameterToString(jid, \"multi\")\n\t\n\tqueryParams[\"sortOrder\"] = a.Configuration.APIClient.ParameterToString(sortOrder, \"\")\n\t\n\tqueryParams[\"expand\"] = a.Configuration.APIClient.ParameterToString(expand, \"multi\")\n\t\n\tqueryParams[\"integrationPresenceSource\"] = a.Configuration.APIClient.ParameterToString(integrationPresenceSource, \"\")\n\t\n\n\t// Find an replace keys that were altered to avoid clashes with go keywords \n\tcorrectedQueryParams := make(map[string]string)\n\tfor k, v := range queryParams {\n\t\tif k == \"varType\" {\n\t\t\tcorrectedQueryParams[\"type\"] = v\n\t\t\tcontinue\n\t\t}\n\t\tcorrectedQueryParams[k] = v\n\t}\n\tqueryParams = correctedQueryParams\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{ \"application/json\", }\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := a.Configuration.APIClient.SelectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\theaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\n\t\t\"application/json\",\n\t}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := a.Configuration.APIClient.SelectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\theaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tvar successPayload *Userprofileentitylisting\n\tresponse, err := a.Configuration.APIClient.CallAPI(path, httpMethod, postBody, headerParams, queryParams, formParams, postFileName, fileBytes)\n\tif err != nil {\n\t\t// Nothing special to do here, but do avoid processing the response\n\t} else if err == nil && response.Error != nil {\n\t\terr = errors.New(response.ErrorMessage)\n\t} else if response.HasBody {\n\t\tif \"Userprofileentitylisting\" == \"string\" {\n\t\t\tcopy(response.RawBody, &successPayload)\n\t\t} else {\n\t\t\terr = json.Unmarshal(response.RawBody, &successPayload)\n\t\t}\n\t}\n\treturn successPayload, response, err\n}" ]
[ "0.7198962", "0.6869934", "0.6789719", "0.67125577", "0.6603376", "0.65903366", "0.6403358", "0.61283404", "0.5979468", "0.59100527", "0.58843744", "0.5871112", "0.58510584", "0.5812088", "0.57750744", "0.5740033", "0.57312274", "0.57131714", "0.57038635", "0.5665462", "0.5650157", "0.5582988", "0.55647236", "0.5562351", "0.5551982", "0.5550958", "0.5543493", "0.5498958", "0.5494953", "0.54691154", "0.545507", "0.5400691", "0.54004836", "0.53805065", "0.5374537", "0.53538436", "0.53378516", "0.5326786", "0.5318802", "0.5311433", "0.5288244", "0.5274247", "0.52566415", "0.52213925", "0.5218808", "0.5217135", "0.5204053", "0.51857084", "0.5181272", "0.5168757", "0.5167296", "0.51672316", "0.51565987", "0.5143938", "0.5137596", "0.51348567", "0.5130919", "0.5114644", "0.5105083", "0.51038206", "0.5089186", "0.5080312", "0.5065956", "0.50566906", "0.5056112", "0.5000419", "0.4969409", "0.4966967", "0.49667272", "0.4957326", "0.49512482", "0.4949572", "0.4940578", "0.4939023", "0.49364272", "0.49330774", "0.4929585", "0.49258223", "0.4920468", "0.49009386", "0.48900932", "0.48882967", "0.48881766", "0.48760742", "0.4875418", "0.4864147", "0.48556674", "0.4849216", "0.48376966", "0.48219958", "0.48211417", "0.4818083", "0.48019698", "0.47995257", "0.47840917", "0.47832185", "0.47808462", "0.47791785", "0.47752345", "0.47639498" ]
0.8219907
0
GetCmdQueryProfileParams queries all the profiles' module params
func GetCmdQueryProfileParams(cdc *codec.Codec) *cobra.Command { return &cobra.Command{ Use: "parameters", Short: "Retrieve all the profile module parameters", Args: cobra.NoArgs, RunE: func(cmd *cobra.Command, args []string) error { cliCtx := context.NewCLIContext().WithCodec(cdc) route := fmt.Sprintf("custom/%s/%s", types.QuerierRoute, types.QueryParams) res, _, err := cliCtx.QueryWithData(route, nil) if err != nil { fmt.Printf("Could not find profile parameters") return nil } var out types.Params cdc.MustUnmarshalJSON(res, &out) return cliCtx.PrintOutput(out) }, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func queryProfileParams(ctx sdk.Context, _ abci.RequestQuery, keeper Keeper) ([]byte, error) {\n\tprofileParams := keeper.GetParams(ctx)\n\n\tbz, err := codec.MarshalJSONIndent(keeper.Cdc, &profileParams)\n\tif err != nil {\n\t\tpanic(\"could not marshal result to JSON\")\n\t}\n\n\treturn bz, nil\n}", "func GetCmdQueryProfiles(cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"all\",\n\t\tShort: \"Retrieve all the registered profiles.\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\troute := fmt.Sprintf(\"custom/%s/%s\", types.QuerierRoute, types.QueryProfiles)\n\t\t\tres, _, err := cliCtx.QueryWithData(route, nil)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"Could not find any profile\")\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tvar out types.Profiles\n\t\t\tcdc.MustUnmarshalJSON(res, &out)\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n}", "func GetCmdQueryProfile(cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"profile [address_or_dtag]\",\n\t\tShort: \"Retrieve the profile having the specified user address or profile dtag, if any.\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\troute := fmt.Sprintf(\"custom/%s/%s/%s\", types.QuerierRoute, types.QueryProfile, args[0])\n\t\t\tres, _, err := cliCtx.QueryWithData(route, nil)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"Could not find a profile with dtag %s \\n\", args[0])\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tvar out types.Profile\n\t\t\tcdc.MustUnmarshalJSON(res, &out)\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n}", "func GetProfileParameters() (ProfileParameters, Alerts) {\n\treturn ProfileParameters{}, Alerts{}\n}", "func GetCmdQueryParams() *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"params\",\n\t\tShort: fmt.Sprintf(\"get the %s module parameters\", types.ModuleName),\n\t\tLong: \"Get the current issuance module parameters.\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclientCtx, err := client.GetClientQueryContext(cmd)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\n\n\t\t\tres, err := queryClient.Params(context.Background(), &types.QueryParamsRequest{})\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn clientCtx.PrintProto(&res.Params)\n\t\t},\n\t}\n}", "func GetCmdQueryParams(cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"params\",\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query the current staking parameters information\",\n\t\tLong: strings.TrimSpace(\n\t\t\tfmt.Sprintf(`Query values set as staking parameters.\n\nExample:\n$ %s query staking params\n`,\n\t\t\t\tversion.ClientName,\n\t\t\t),\n\t\t),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\troute := fmt.Sprintf(\"custom/%s/%s\", types.StoreKey, staking.QueryParameters)\n\t\t\tbz, _, err := cliCtx.QueryWithData(route, nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar mergedParams types.MergedParams\n\t\t\tcdc.MustUnmarshalJSON(bz, &mergedParams)\n\t\t\treturn cliCtx.PrintOutput(mergedParams)\n\t\t},\n\t}\n}", "func (s *SmartContract) QueryAllProfiles(ctx contractapi.TransactionContextInterface) ([]QueryProfileResult, error) {\n\tstartKey := \"\"\n\tendKey := \"\"\n\n\tresultsIterator, err := ctx.GetStub().GetStateByRange(startKey, endKey)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer resultsIterator.Close()\n\n\tresults := []QueryProfileResult{}\n\n\tfor resultsIterator.HasNext() {\n\t\tqueryResponse, err := resultsIterator.Next()\n\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif strings.Contains(queryResponse.Key, \"profile-\"){\n\t\t\tprofile := new(UserProfile)\n\t\t\t_ = json.Unmarshal(queryResponse.Value, profile)\n\n\t\t\tqueryResult := QueryProfileResult{Key: queryResponse.Key, Record: profile}\n\t\t\tresults = append(results, queryResult)\n\t\t}\n\t\t\n\t}\n\n\treturn results, nil\n}", "func QueryParams() *cobra.Command {\n\treturn qflags(&cobra.Command{\n\t\tUse: \"params\",\n\t\tShort: \"Query the current ecocredit module parameters\",\n\t\tLong: strings.TrimSpace(\n\t\t\tfmt.Sprintf(`Query the current ecocredit module parameters\n\t\t\t\nExamples:\n$%s query %s params\n$%s q %s params\n\t\t\t`, version.AppName, ecocredit.ModuleName, version.AppName, ecocredit.ModuleName),\n\t\t),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tc, ctx, err := mkQueryClient(cmd)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tres, err := c.Params(cmd.Context(), &ecocredit.QueryParamsRequest{})\n\t\t\treturn print(ctx, res, err)\n\t\t},\n\t})\n}", "func GetQueryCmd(cdc *codec.Codec) *cobra.Command {\n\tprofileQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Querying commands for the profiles module\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\tprofileQueryCmd.AddCommand(flags.GetCommands(\n\t\tGetCmdQueryProfile(cdc),\n\t\tGetCmdQueryProfiles(cdc),\n\t\tGetCmdQueryProfileParams(cdc),\n\t\tGetCmdQueryDTagRequests(cdc),\n\t)...)\n\treturn profileQueryCmd\n}", "func GetQueryCmdParams(route string, cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"params\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tbz, _, err := cliCtx.Query(fmt.Sprintf(\"custom/%s/%s\", route, types.QueryParams))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn printOutput(cliCtx, cdc, bz, &types.Params{})\n\t\t},\n\t}\n}", "func FetchProfiles() map[string]string {\n\ttoken := auth.NewToken()\n\tquery := queryPayload{\"SELECT Id, ProfileName FROM CommunicationProfile\"}\n\tpayload, err := json.Marshal(query)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treq, err := http.NewRequest(\"POST\", viper.GetString(\"baseurl\")+\"/v1/action/query\", bytes.NewBuffer(payload))\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"Authorization\", \"Bearer \"+token.Val)\n\n\tresponse, err := http.DefaultClient.Do(req)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer response.Body.Close()\n\n\tif response.StatusCode != 200 {\n\t\tbody, err := ioutil.ReadAll(response.Body)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tlog.Fatal(string(body))\n\t}\n\n\tdec := json.NewDecoder(response.Body)\n\tvar body profilesQueryResponse\n\tif err = dec.Decode(&body); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tif !body.Done {\n\t\tlog.Fatalln(\"there are more communication profile to query\")\n\t}\n\n\tresult := make(map[string]string)\n\tfor _, p := range body.Records {\n\t\tresult[p.Name] = p.ID\n\t}\n\n\treturn result\n}", "func GetCmdParams() *cobra.Command {\r\n\tcmd := &cobra.Command{\r\n\t\tUse: \"params\",\r\n\t\tShort: \"Query the current ibc-transfer parameters\",\r\n\t\tLong: \"Query the current ibc-transfer parameters\",\r\n\t\tArgs: cobra.NoArgs,\r\n\t\tExample: fmt.Sprintf(\"%s query ibc-transfer params\", version.AppName),\r\n\t\tRunE: func(cmd *cobra.Command, _ []string) error {\r\n\t\t\tclientCtx, err := client.GetClientQueryContext(cmd)\r\n\t\t\tif err != nil {\r\n\t\t\t\treturn err\r\n\t\t\t}\r\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\r\n\r\n\t\t\tres, _ := queryClient.Params(context.Background(), &types.QueryParamsRequest{})\r\n\t\t\treturn clientCtx.PrintProto(res.Params)\r\n\t\t},\r\n\t}\r\n\r\n\tflags.AddQueryFlagsToCmd(cmd)\r\n\r\n\treturn cmd\r\n}", "func (k Keeper) Params(c context.Context, req *types.QueryParamsRequest) (*types.QueryParamsResponse, error) {\n\tdefer telemetry.MeasureSince(time.Now(), types.ModuleName, \"query\", \"Params\")\n\tctx := sdk.UnwrapSDKContext(c)\n\tvar params types.Params\n\tk.paramSpace.GetParamSet(ctx, &params)\n\n\treturn &types.QueryParamsResponse{Params: params, Request: req}, nil\n}", "func NewGetPbxDeviceProfileitemsRequest(server string, params *GetPbxDeviceProfileitemsParams) (*http.Request, error) {\n\tvar err error\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/pbxdeviceprofiles\")\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryValues := queryUrl.Query()\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"name\", params.Name); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"config_id\", params.ConfigId); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"order_by\", params.OrderBy); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"order_by_direction\", params.OrderByDirection); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"page\", params.Page); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"rows\", params.Rows); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tqueryUrl.RawQuery = queryValues.Encode()\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func GetCmdQueryParams() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"params\",\n\t\tShort: \"Query the current minting parameters\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclientCtx, err := client.GetClientQueryContext(cmd)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\n\n\t\t\tparams := &types.QueryParamsRequest{}\n\t\t\tres, err := queryClient.Params(cmd.Context(), params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn clientCtx.PrintProto(&res.Params)\n\t\t},\n\t}\n\n\tflags.AddQueryFlagsToCmd(cmd)\n\n\treturn cmd\n}", "func getParameters(c *cli.Context) error {\n\tif !isSystemRunning() {\n\t\treturn nil\n\t}\n\t_, _, _, controllers := getIPAddresses()\n\n\tparams := sendCommandToControllers(controllers, \"GetParams\", \"\")\n\tfmt.Println(params)\n\n\treturn nil\n}", "func QueryParameters(f *cli.Fixtures, flags ...string) types.Params {\n\tcmd := fmt.Sprintf(\"%s query distribution params %v\", f.SimcliBinary, f.Flags())\n\tout, errStr := tests.ExecuteT(f.T, cli.AddFlags(cmd, flags), \"\")\n\trequire.Empty(f.T, errStr)\n\n\tvar params types.Params\n\trequire.NoError(f.T, f.Cdc.UnmarshalJSON([]byte(out), &params))\n\treturn params\n}", "func NewGetPbxDeviceProfilePreferenceitemsRequest(server string, params *GetPbxDeviceProfilePreferenceitemsParams) (*http.Request, error) {\n\tvar err error\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/pbxdeviceprofilepreferences\")\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryValues := queryUrl.Query()\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"order_by\", params.OrderBy); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"order_by_direction\", params.OrderByDirection); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"page\", params.Page); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"rows\", params.Rows); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tqueryUrl.RawQuery = queryValues.Encode()\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func queryProfiles(ctx sdk.Context, _ abci.RequestQuery, keeper Keeper) ([]byte, error) {\n\taccounts := keeper.GetProfiles(ctx)\n\n\tbz, err := codec.MarshalJSONIndent(keeper.Cdc, &accounts)\n\tif err != nil {\n\t\tpanic(\"could not marshal result to JSON\")\n\t}\n\n\treturn bz, nil\n}", "func (c *OVClient) GetProfiles(start string, count string, filter string, sort string, scopeUris string) (ServerProfileList, error) {\n\tvar (\n\t\turi = \"/rest/server-profiles\"\n\t\tq map[string]interface{}\n\t\tprofiles ServerProfileList\n\t)\n\tq = make(map[string]interface{})\n\tif len(filter) > 0 {\n\t\tq[\"filter\"] = filter\n\t}\n\n\tif sort != \"\" {\n\t\tq[\"sort\"] = sort\n\t}\n\n\tif start != \"\" {\n\t\tq[\"start\"] = start\n\t}\n\n\tif count != \"\" {\n\t\tq[\"count\"] = count\n\t}\n\n\tif scopeUris != \"\" {\n\t\tq[\"scopeUris\"] = scopeUris\n\t}\n\n\t// refresh login\n\tc.RefreshLogin()\n\tc.SetAuthHeaderOptions(c.GetAuthHeaderMap())\n\t// Setup query\n\tif len(q) > 0 {\n\t\tc.SetQueryString(q)\n\t}\n\tdata, err := c.RestAPICall(rest.GET, uri, nil)\n\tif err != nil {\n\t\treturn profiles, err\n\t}\n\n\tlog.Debugf(\"GetProfiles %s\", data)\n\tif err := json.Unmarshal([]byte(data), &profiles); err != nil {\n\t\treturn profiles, err\n\t}\n\treturn profiles, nil\n}", "func (c *Config) GetAllMemberParameters() map[string]string {\n\tvars := map[string]string{}\n\n\tfor _, env := range os.Environ() {\n\t\tkeyValue := strings.SplitN(env, \"=\", 2)\n\t\tif len(keyValue) < 2 {\n\t\t\tcontinue\n\t\t}\n\t\tif strings.HasPrefix(keyValue[0], MemberEnvPrefix+\"_\") {\n\t\t\tvars[keyValue[0]] = keyValue[1]\n\t\t}\n\t}\n\treturn vars\n}", "func NewGetProfilePreferenceitemsRequest(server string, params *GetProfilePreferenceitemsParams) (*http.Request, error) {\n\tvar err error\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/profilepreferences\")\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryValues := queryUrl.Query()\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"order_by\", params.OrderBy); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"order_by_direction\", params.OrderByDirection); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"page\", params.Page); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"rows\", params.Rows); err != nil {\n\t\treturn nil, err\n\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tfor k, v := range parsed {\n\t\t\tfor _, v2 := range v {\n\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t}\n\t\t}\n\t}\n\n\tqueryUrl.RawQuery = queryValues.Encode()\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (k Keeper) GetParams(ctx sdk.Context) (params types.Params) {\n\tk.paramSubspace.GetParamSet(ctx, &params)\n\treturn params\n}", "func (tr *Account) GetParameters() (map[string]interface{}, error) {\n\tp, err := json.TFParser.Marshal(tr.Spec.ForProvider)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbase := map[string]interface{}{}\n\treturn base, json.TFParser.Unmarshal(p, &base)\n}", "func GetAttributeParamsCmd() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"params\",\n\t\tShort: \"Query the current name parameters\",\n\t\tArgs: cobra.NoArgs,\n\t\tExample: fmt.Sprintf(`$ %s query attribute params`, version.AppName),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclientCtx, err := client.GetClientQueryContext(cmd)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\n\t\t\tres, err := queryClient.Params(context.Background(), &types.QueryParamsRequest{})\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn clientCtx.PrintProto(&res.Params)\n\t\t},\n\t}\n\n\tflags.AddQueryFlagsToCmd(cmd)\n\n\treturn cmd\n}", "func GetProfile(c *fiber.Ctx) error {\n\tID := c.Query(\"id\")\n\n\tif len(ID) < 1 {\n\t\tc.Send(ID)\n\t\tc.Send(\"Parameter ID is required\")\n\t\tc.SendStatus(http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tprofile, err := database.SearchProfile(ID)\n\tif err != nil {\n\t\tc.Send(\"Error Occurred\" + err.Error())\n\t\tc.SendStatus(http.StatusBadRequest)\n\t\treturn\n\t}\n\tif err := c.JSON(profile); err != nil {\n\t\tc.Status(500).Send(err)\n\t\treturn\n\t}\n\tc.Accepts(\"application/json\")\n\tc.SendStatus(http.StatusAccepted)\n}", "func registerOperationDevicesGetDevicesApsProfilesParamFlags(cmd *cobra.Command) error {\n\treturn nil\n}", "func (o *PhoneSearchOptions) getParameters() (params map[string]string, err error) {\n\t// create an empty map of options\n\tparams = make(map[string]string)\n\n\t// reflect over the properties in o, adding parameters to the global map\n\tval := reflect.ValueOf(o).Elem()\n\tfor i := 0; i < val.NumField(); i++ {\n\t\tif !val.Field(i).IsNil() {\n\t\t\to := val.Field(i).Interface().(OptionProvider)\n\t\t\tfieldParams, err := o.getParameters()\n\t\t\tif err != nil {\n\t\t\t\treturn params, err\n\t\t\t}\n\t\t\tfor k, v := range fieldParams {\n\t\t\t\tparams[k] = v\n\t\t\t}\n\t\t}\n\t}\n\treturn params, nil\n}", "func listProfiles(ctx context.Context, _ []string) error {\n\tm, err := cmdutils.LoadManager(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"%15s %s\\n\", \"ID\", \"NAME\")\n\tfor _, p := range m.Profiles() {\n\t\tfmt.Printf(\"%15s %s\\n\", p.Id, p.Name)\n\t}\n\n\treturn nil\n}", "func (k Keeper) GetParams(ctx sdk.Context) (params types.Params) {\n\tk.paramspace.GetParamSet(ctx, &params)\n\treturn params\n}", "func testProfileList(t *testing.T) {\n\tp := profileName(t)\n\tt.Parallel()\n\tmk := NewMinikubeRunner(t, p, \"--wait=false\")\n\tout, stderr := mk.MustRun(\"profile list\")\n\tif !strings.Contains(out, p) {\n\t\tt.Errorf(\"Error , failed to read profile name (%s) in `profile list` command output : \\n %q : \\n stderr: %s \", p, out, stderr)\n\t}\n}", "func (k Keeper) Params(c context.Context, req *types.QueryParamsRequest) (*types.QueryParamsResponse, error) {\n\tctx := sdk.UnwrapSDKContext(c)\n\tvar params types.Params\n\tk.paramSpace.GetParamSet(ctx, &params)\n\n\treturn &types.QueryParamsResponse{Params: params}, nil\n}", "func (p *OnuIgmpProfile) ListEssentialParams() map[string]interface{} {\r\n\tvar EssentialOnuIgmpProfile = map[string]interface{}{\r\n\t\tOnuIgmpProfileHeaders[0]: p.GetName(),\r\n\t\tOnuIgmpProfileHeaders[1]: p.GetMode(),\r\n\t\tOnuIgmpProfileHeaders[2]: p.GetProxy(),\r\n\t\tOnuIgmpProfileHeaders[3]: p.GetFastLeave(),\r\n\t\tOnuIgmpProfileHeaders[4]: p.GetUsTci(),\r\n\t\tOnuIgmpProfileHeaders[5]: p.DsGemPort,\r\n\t}\r\n\t// I want all of these Bools to return strings of \"Enabled/Disabled\"\r\n\treturn EssentialOnuIgmpProfile\r\n}", "func (ctx *Context) QueryParams(key string) []string {\r\n\tif ctx.queryParams == nil {\r\n\t\tctx.queryParams = ctx.R.URL.Query()\r\n\t}\r\n\treturn ctx.queryParams[key]\r\n}", "func (k Keeper) GetParams(ctx sdk.Context) types.Params {\n\tvar p types.Params\n\tk.paramSubspace.GetParamSet(ctx, &p)\n\treturn p\n}", "func NewProfilesGetCollectionRequest(server string, params *ProfilesGetCollectionParams) (*http.Request, error) {\n\tvar err error\n\n\tserverURL, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\toperationPath := fmt.Sprintf(\"/v1/profiles\")\n\tif operationPath[0] == '/' {\n\t\toperationPath = \".\" + operationPath\n\t}\n\n\tqueryURL, err := serverURL.Parse(operationPath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryValues := queryURL.Query()\n\n\tif params.FilterName != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"filter[name]\", runtime.ParamLocationQuery, *params.FilterName); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FilterProfileState != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"filter[profileState]\", runtime.ParamLocationQuery, *params.FilterProfileState); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FilterProfileType != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"filter[profileType]\", runtime.ParamLocationQuery, *params.FilterProfileType); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FilterId != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"filter[id]\", runtime.ParamLocationQuery, *params.FilterId); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.Sort != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"sort\", runtime.ParamLocationQuery, *params.Sort); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FieldsProfiles != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[profiles]\", runtime.ParamLocationQuery, *params.FieldsProfiles); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.Limit != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", true, \"limit\", runtime.ParamLocationQuery, *params.Limit); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.Include != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"include\", runtime.ParamLocationQuery, *params.Include); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FieldsCertificates != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[certificates]\", runtime.ParamLocationQuery, *params.FieldsCertificates); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FieldsDevices != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[devices]\", runtime.ParamLocationQuery, *params.FieldsDevices); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FieldsBundleIds != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[bundleIds]\", runtime.ParamLocationQuery, *params.FieldsBundleIds); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.LimitCertificates != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", true, \"limit[certificates]\", runtime.ParamLocationQuery, *params.LimitCertificates); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.LimitDevices != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", true, \"limit[devices]\", runtime.ParamLocationQuery, *params.LimitDevices); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tqueryURL.RawQuery = queryValues.Encode()\n\n\treq, err := http.NewRequest(\"GET\", queryURL.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (ctx *Context) QueryParamAll() url.Values {\r\n\tif ctx.queryParams == nil {\r\n\t\tctx.queryParams = ctx.R.URL.Query()\r\n\t}\r\n\treturn ctx.queryParams\r\n}", "func (s *Service) ProfilesGet(userID string) *ProfilesGetOp {\n\treturn &ProfilesGetOp{\n\t\tCredential: s.credential,\n\t\tMethod: \"GET\",\n\t\tPath: strings.Join([]string{\"users\", userID, \"profile\"}, \"/\"),\n\t\tAccept: \"application/json\",\n\t\tQueryOpts: make(url.Values),\n\t\tVersion: esign.APIv21,\n\t}\n}", "func (k Keeper) Params(c context.Context, req *types.QueryParamsRequest) (*types.QueryParamsResponse, error) {\n\tctx := sdk.UnwrapSDKContext(c)\n\n\tparams := k.GetParamSet(ctx)\n\n\treturn &types.QueryParamsResponse{Params: params}, nil\n}", "func NewUpdateDeviceProfileParams() *UpdateDeviceProfileParams {\n\tvar ()\n\treturn &UpdateDeviceProfileParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func NewProfilesGetInstanceRequest(server string, id string, params *ProfilesGetInstanceParams) (*http.Request, error) {\n\tvar err error\n\n\tvar pathParam0 string\n\n\tpathParam0, err = runtime.StyleParamWithLocation(\"simple\", false, \"id\", runtime.ParamLocationPath, id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tserverURL, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\toperationPath := fmt.Sprintf(\"/v1/profiles/%s\", pathParam0)\n\tif operationPath[0] == '/' {\n\t\toperationPath = \".\" + operationPath\n\t}\n\n\tqueryURL, err := serverURL.Parse(operationPath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryValues := queryURL.Query()\n\n\tif params.FieldsProfiles != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[profiles]\", runtime.ParamLocationQuery, *params.FieldsProfiles); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.Include != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"include\", runtime.ParamLocationQuery, *params.Include); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FieldsCertificates != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[certificates]\", runtime.ParamLocationQuery, *params.FieldsCertificates); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FieldsDevices != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[devices]\", runtime.ParamLocationQuery, *params.FieldsDevices); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FieldsBundleIds != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[bundleIds]\", runtime.ParamLocationQuery, *params.FieldsBundleIds); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.LimitCertificates != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", true, \"limit[certificates]\", runtime.ParamLocationQuery, *params.LimitCertificates); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.LimitDevices != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", true, \"limit[devices]\", runtime.ParamLocationQuery, *params.LimitDevices); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tqueryURL.RawQuery = queryValues.Encode()\n\n\treq, err := http.NewRequest(\"GET\", queryURL.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func GetParamsCmd(cdc *codec.Codec) *cobra.Command {\n\tparamsCmd := &cobra.Command{\n\t\tUse: \"params\",\n\t\tShort: \"Update the params of various modules\",\n\t\tSuggestionsMinimumDistance: 2,\n\t}\n\n\tparamsCmd.AddCommand(AccountParamsCmd(cdc))\n\tparamsCmd.AddCommand(BankParamsCmd(cdc))\n\tparamsCmd.AddCommand(CommunityParamsCmd(cdc))\n\tparamsCmd.AddCommand(ClaimParamsCmd(cdc))\n\tparamsCmd.AddCommand(StakingParamsCmd(cdc))\n\tparamsCmd.AddCommand(SlashingParamsCmd(cdc))\n\n\treturn paramsCmd\n}", "func GetProfile(ctx *router.Context) {\n\tuin, _ := ctx.Keys[middleware.ComerUinContextKey].(uint64)\n\tresponse, err := service.GetComerProfile(uin)\n\tif err != nil {\n\t\tctx.ERROR(\n\t\t\trouter.ErrBuisnessError,\n\t\t\t\"wrong metamask login parameter\",\n\t\t)\n\t\treturn\n\t}\n\n\tctx.OK(response)\n}", "func GetCmdQueryParams(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: oracle.QueryParams,\n\t\tArgs: cobra.NoArgs,\n\t\tShort: \"Query the current Oracle params\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, oracle.QueryParams), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar params oracle.Params\n\t\t\tcdc.MustUnmarshalJSON(res, &params)\n\t\t\treturn cliCtx.PrintOutput(params)\n\t\t},\n\t}\n\n\treturn cmd\n}", "func (a *API) ProfileSearch(ctx context.Context, searchID string, pageNum int) (*ProfileResp, error) {\n\n\tall := pageNum < 1\n\tif all || pageNum < 1 {\n\t\tpageNum = 1\n\t}\n\n\tvar resp ProfileResp\n\turlStr := fmt.Sprintf(\"/api/v1/profile?searchId=%s&pageSize=100&pageNumber=%d\", searchID, pageNum)\n\tif err := a.Get(ctx, urlStr, &resp); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif all && resp.TotalPageCount > 1 {\n\t\tfor i := 1; i < resp.TotalPageCount; i++ {\n\t\t\tpg, err := a.ProfileSearch(ctx, searchID, i+1)\n\t\t\tif err != nil {\n\t\t\t\treturn &resp, err\n\t\t\t}\n\t\t\tresp.Profiles = append(resp.Profiles, pg.Profiles...)\n\t\t}\n\t}\n\n\treturn &resp, nil\n}", "func (s *SmartContract) QueryProfile(ctx contractapi.TransactionContextInterface, carNumber string) (*Profile, error) {\n\tcarAsBytes, err := ctx.GetStub().GetState(\"profile-\" + carNumber)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to read from world state. %s\", err.Error())\n\t}\n\n\tif carAsBytes == nil {\n\t\treturn nil, fmt.Errorf(\"%s does not exist\", carNumber)\n\t}\n\n\tcar := new(Profile)\n\t_ = json.Unmarshal(carAsBytes, car)\n\n\treturn car, nil\n}", "func GetProfileList(ctx context.Context) ([]*shill.Profile, error) {\n\tmanager, err := shill.NewManager(ctx)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed creating shill manager object\")\n\t}\n\t// Refresh the in-memory profile list.\n\tif _, err := manager.GetProperties(ctx); err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed refreshing the in-memory profile list\")\n\t}\n\t// Get current profiles.\n\tprofiles, err := manager.Profiles(ctx)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed getting profile list\")\n\t}\n\treturn profiles, nil\n}", "func GetMetadataParamsCmd() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"params [locator]\",\n\t\tAliases: []string{\"p\"},\n\t\tShort: \"Query the current metadata parameters\",\n\t\tArgs: cobra.MaximumNArgs(1),\n\t\tExample: fmt.Sprintf(\"%s params\", cmdStart),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tif len(args) == 0 {\n\t\t\t\treturn outputParams(cmd)\n\t\t\t}\n\t\t\targ0 := strings.TrimSpace(args[0])\n\t\t\tif arg0 == \"locator\" {\n\t\t\t\treturn outputOSLocatorParams(cmd)\n\t\t\t}\n\t\t\treturn fmt.Errorf(\"unknown argument: %s\", arg0)\n\t\t},\n\t}\n\n\taddIncludeRequestFlag(cmd)\n\tflags.AddQueryFlagsToCmd(cmd)\n\n\treturn cmd\n}", "func (db *DB) GetParams() (params core.Params, err error) {\n\tif err := db.Instance.First(&params).Error; err != nil {\n\t\treturn params, err\n\t}\n\treturn params, nil\n}", "func getParamsAndValues(b *ingestcommon.BenchData) ([]paramtools.Params, []float32, paramtools.ParamSet) {\n\tparams := []paramtools.Params{}\n\tvalues := []float32{}\n\tps := paramtools.ParamSet{}\n\tfor testName, allConfigs := range b.Results {\n\t\tfor configName, result := range allConfigs {\n\t\t\tkey := paramtools.Params(b.Key).Copy()\n\t\t\tkey[\"test\"] = testName\n\t\t\tkey[\"config\"] = configName\n\t\t\tkey.Add(paramtools.Params(b.Options))\n\n\t\t\t// If there is an options map inside the result add it to the params.\n\t\t\tif resultOptions, ok := result[\"options\"]; ok {\n\t\t\t\tif opts, ok := resultOptions.(map[string]interface{}); ok {\n\t\t\t\t\tfor k, vi := range opts {\n\t\t\t\t\t\t// Ignore the very long and not useful GL_ values, we can retrieve\n\t\t\t\t\t\t// them later via ptracestore.Details.\n\t\t\t\t\t\tif strings.HasPrefix(k, \"GL_\") {\n\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\t}\n\t\t\t\t\t\tif s, ok := vi.(string); ok {\n\t\t\t\t\t\t\tkey[k] = s\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tfor k, vi := range result {\n\t\t\t\tif k == \"options\" || k == \"samples\" {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tkey[\"sub_result\"] = k\n\t\t\t\tfloatVal, ok := vi.(float64)\n\t\t\t\tif !ok {\n\t\t\t\t\tsklog.Errorf(\"Found a non-float64 in %v\", result)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tkey = query.ForceValid(key)\n\t\t\t\tparams = append(params, key.Copy())\n\t\t\t\tvalues = append(values, float32(floatVal))\n\t\t\t\tps.AddParams(key)\n\t\t\t}\n\t\t}\n\t}\n\tps.Normalize()\n\treturn params, values, ps\n}", "func GetCmdQueryParams(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"params\",\n\t\tShort: \"Query the current Treasury params\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, treasury.QueryParams), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar params treasury.Params\n\t\t\tcdc.MustUnmarshalJSON(res, &params)\n\t\t\treturn cliCtx.PrintOutput(params)\n\t\t},\n\t}\n\n\treturn cmd\n}", "func (tr *CassandraKeySpace) GetParameters() (map[string]interface{}, error) {\n\tp, err := json.TFParser.Marshal(tr.Spec.ForProvider)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbase := map[string]interface{}{}\n\treturn base, json.TFParser.Unmarshal(p, &base)\n}", "func GetCmdParams() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"params\",\n\t\tShort: \"Query the current ibc-router parameters\",\n\t\tLong: \"Query the current ibc-router parameters\",\n\t\tArgs: cobra.NoArgs,\n\t\tExample: fmt.Sprintf(\"%s query ibc-router params\", version.AppName),\n\t\tRunE: func(cmd *cobra.Command, _ []string) error {\n\t\t\tclientCtx, err := client.GetClientQueryContext(cmd)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tqueryClient := types.NewQueryClient(clientCtx)\n\n\t\t\tres, err := queryClient.Params(cmd.Context(), &types.QueryParamsRequest{})\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn clientCtx.PrintProto(res.Params)\n\t\t},\n\t}\n\n\tflags.AddQueryFlagsToCmd(cmd)\n\n\treturn cmd\n}", "func GenerateSubProfileQuery() (string, error) {\n\treturn globalProbe.GenerateSubProfileQuery()\n}", "func (k Querier) Params(c context.Context, _ *types.QueryParamsRequest) (*types.QueryParamsResponse, error) {\n\tctx := sdk.UnwrapSDKContext(c)\n\tvar params types.Params\n\tk.paramSpace.GetParamSet(ctx, &params)\n\treturn &types.QueryParamsResponse{Params: params}, nil\n}", "func GetCmdQueryParams(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: budget.QueryParams,\n\t\tShort: \"Query the current budget params\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, budget.QueryParams), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar params budget.Params\n\t\t\tcdc.MustUnmarshalJSON(res, &params)\n\t\t\treturn cliCtx.PrintOutput(params)\n\t\t},\n\t}\n\n\treturn cmd\n}", "func GetProfile() map[string]string {\n\treturn Creds.GetStringMapString(\"profiles.\" + strings.Replace(viper.GetString(\"profile\"), \".\", \"-\", -1))\n}", "func (o DiagnosticBackendRequestDataMaskingPtrOutput) QueryParams() DiagnosticBackendRequestDataMaskingQueryParamArrayOutput {\n\treturn o.ApplyT(func(v *DiagnosticBackendRequestDataMasking) []DiagnosticBackendRequestDataMaskingQueryParam {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.QueryParams\n\t}).(DiagnosticBackendRequestDataMaskingQueryParamArrayOutput)\n}", "func (j *Jsonnet) Params(envName string) ([]ModuleParameter, error) {\n\tj.log().WithField(\"env-name\", envName).Debug(\"getting component params\")\n\n\tparamsData, err := j.readParams(envName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tprops, err := params.ToMap(j.Name(false), paramsData, paramsComponentRoot)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not find components\")\n\t}\n\n\tvar params []ModuleParameter\n\tfor k, v := range props {\n\t\tvStr, err := j.paramValue(v)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tnp := ModuleParameter{\n\t\t\tComponent: j.Name(false),\n\t\t\tKey: k,\n\t\t\tValue: vStr,\n\t\t}\n\n\t\tparams = append(params, np)\n\t}\n\n\tsort.Slice(params, func(i, j int) bool {\n\t\treturn params[i].Key < params[j].Key\n\t})\n\n\treturn params, nil\n}", "func (bg *Backgrounder) GetProfiles() map[string]time.Duration {\n\treturn bg.profile\n}", "func getProfilesDirectory() (string, error) {\n\tn := uint32(100)\n\tfor {\n\t\tb := make([]uint16, n)\n\t\te := windows.GetProfilesDirectory(&b[0], &n)\n\t\tif e == nil {\n\t\t\treturn syscall.UTF16ToString(b), nil\n\t\t}\n\t\tif e != syscall.ERROR_INSUFFICIENT_BUFFER {\n\t\t\treturn \"\", e\n\t\t}\n\t\tif n <= uint32(len(b)) {\n\t\t\treturn \"\", e\n\t\t}\n\t}\n}", "func (org *Org) GetAllNsxtAppPortProfiles(queryParameters url.Values, scope string) ([]*NsxtAppPortProfile, error) {\n\tqueryParams := copyOrNewUrlValues(queryParameters)\n\tif scope != \"\" {\n\t\tqueryParams = queryParameterFilterAnd(\"scope==\"+scope, queryParams)\n\t}\n\n\treturn getAllNsxtAppPortProfiles(org.client, queryParams)\n}", "func (m MigProfile) GetProfileIDs() (int, int, int, error) {\n\terr := m.AssertValid()\n\tif err != nil {\n\t\treturn -1, -1, -1, fmt.Errorf(\"invalid MigProfile: %v\", err)\n\t}\n\n\tc, g, _, err := m.Parse()\n\tif err != nil {\n\t\treturn -1, -1, -1, fmt.Errorf(\"unable to parse MigProfile: %v\", err)\n\t}\n\n\tvar giProfileID, ciProfileID, ciEngProfileID int\n\n\tswitch g {\n\tcase 1:\n\t\tgiProfileID = nvml.GPU_INSTANCE_PROFILE_1_SLICE\n\tcase 2:\n\t\tgiProfileID = nvml.GPU_INSTANCE_PROFILE_2_SLICE\n\tcase 3:\n\t\tgiProfileID = nvml.GPU_INSTANCE_PROFILE_3_SLICE\n\tcase 4:\n\t\tgiProfileID = nvml.GPU_INSTANCE_PROFILE_4_SLICE\n\tcase 7:\n\t\tgiProfileID = nvml.GPU_INSTANCE_PROFILE_7_SLICE\n\tcase 8:\n\t\tgiProfileID = nvml.GPU_INSTANCE_PROFILE_8_SLICE\n\tdefault:\n\t\treturn -1, -1, -1, fmt.Errorf(\"unknown GPU Instance slice size: %v\", g)\n\t}\n\n\tswitch c {\n\tcase 1:\n\t\tciProfileID = nvml.COMPUTE_INSTANCE_PROFILE_1_SLICE\n\tcase 2:\n\t\tciProfileID = nvml.COMPUTE_INSTANCE_PROFILE_2_SLICE\n\tcase 3:\n\t\tciProfileID = nvml.COMPUTE_INSTANCE_PROFILE_3_SLICE\n\tcase 4:\n\t\tciProfileID = nvml.COMPUTE_INSTANCE_PROFILE_4_SLICE\n\tcase 7:\n\t\tciProfileID = nvml.COMPUTE_INSTANCE_PROFILE_7_SLICE\n\tcase 8:\n\t\tciProfileID = nvml.COMPUTE_INSTANCE_PROFILE_8_SLICE\n\tdefault:\n\t\treturn -1, -1, -1, fmt.Errorf(\"unknown Compute Instance slice size: %v\", c)\n\t}\n\n\tciEngProfileID = nvml.COMPUTE_INSTANCE_ENGINE_PROFILE_SHARED\n\n\treturn giProfileID, ciProfileID, ciEngProfileID, nil\n}", "func NewGetaspecificPbxDeviceProfileRequest(server string, id string) (*http.Request, error) {\n\tvar err error\n\n\tvar pathParam0 string\n\n\tpathParam0, err = runtime.StyleParam(\"simple\", false, \"id\", id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/pbxdeviceprofiles/%s\", pathParam0)\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (k Keeper) GetParams(ctx sdk.Context) Params {\n\tvar resultParams Params\n\tk.paramSpace.Get(ctx, paramStoreKeyParams, &resultParams)\n\treturn resultParams\n}", "func (f FindOptions) QueryParams() map[string][]string {\n\tqp := map[string][]string{\n\t\t\"descending\": {strconv.FormatBool(f.Descending)},\n\t\t\"offset\": {strconv.Itoa(f.Offset)},\n\t}\n\n\tif f.Limit > 0 {\n\t\tqp[\"limit\"] = []string{strconv.Itoa(f.Limit)}\n\t}\n\n\tif f.SortBy != \"\" {\n\t\tqp[\"sortBy\"] = []string{f.SortBy}\n\t}\n\n\treturn qp\n}", "func getProfiles() {\n\n\tvar configDir string\n\tif isRunningInDockerContainer() {\n\t\tconfigDir = filepath.Join(\"app\", \"config\")\n\t} else {\n\t\t_, fn, _, ok := runtime.Caller(0)\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\t\tindex := strings.LastIndex(fn, \"chart-verifier/\")\n\t\tconfigDir = fn[0 : index+len(\"chart-verifier\")]\n\t\tconfigDir = filepath.Join(configDir, \"config\")\n\t}\n\n\tfilepath.Walk(configDir, func(path string, info os.FileInfo, err error) error {\n\t\tif info != nil {\n\t\t\tif strings.HasSuffix(info.Name(), \".yaml\") {\n\t\t\t\tprofileRead, err := readProfile(path)\n\t\t\t\tif err == nil {\n\t\t\t\t\t// If version is not valid set to a default version\n\t\t\t\t\tif !semver.IsValid(profileRead.Version) {\n\t\t\t\t\t\tprofileRead.Version = DefaultProfileVersion\n\t\t\t\t\t}\n\t\t\t\t\tif len(profileRead.Vendor) == 0 {\n\t\t\t\t\t\tprofileRead.Vendor = VendorTypeNotSpecified\n\t\t\t\t\t}\n\t\t\t\t\tprofileMap[profileRead.Vendor] = append(profileMap[profileRead.Vendor], profileRead)\n\t\t\t\t\tprofileRead.Name = strings.Split(info.Name(), \".yaml\")[0]\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t})\n\treturn\n}", "func GetParameterByNameAndConfigFile(toClient *toclient.Session, name string, configFile string, opts *toclient.RequestOptions) ([]tc.Parameter, toclientlib.ReqInf, error) {\n\tif opts == nil {\n\t\topts = &toclient.RequestOptions{}\n\t}\n\tif opts.QueryParameters == nil {\n\t\topts.QueryParameters = url.Values{}\n\t}\n\topts.QueryParameters.Set(\"name\", name)\n\topts.QueryParameters.Set(\"configFile\", name)\n\tparams, reqInf, err := toClient.GetParameters(*opts)\n\treturn params.Response, reqInf, err\n}", "func (h *auth) Params(c echo.Context) error {\n\t// Fetch params from URL queries\n\temail := c.QueryParam(\"email\")\n\tif email == \"\" {\n\t\treturn c.JSON(http.StatusUnauthorized, sferror.New(\"No email provided.\"))\n\t}\n\n\treturn h.params(c, email)\n}", "func (c *Context) GetParams(key string) string {\n\treturn c.P.ByName(key)\n}", "func (k Keeper) GetParams(ctx sdk.Ctx) types.Params {\n\treturn types.Params{\n\t\tSessionNodeCount: k.SessionNodeCount(ctx),\n\t\tClaimSubmissionWindow: k.ClaimSubmissionWindow(ctx),\n\t\tSupportedBlockchains: k.SupportedBlockchains(ctx),\n\t\tClaimExpiration: k.ClaimExpiration(ctx),\n\t\tReplayAttackBurnMultiplier: k.ReplayAttackBurnMultiplier(ctx),\n\t\tMinimumNumberOfProofs: k.MinimumNumberOfProofs(ctx),\n\t\tBlockByteSize: k.BlockByteSize(ctx),\n\t}\n}", "func (sc SearchClient) QueryParams() url.Values {\n\tparams := url.Values{}\n\n\tif sc.FilterID > 0 {\n\t\tparams.Add(\"filter_id\", strconv.Itoa(sc.FilterID))\n\t}\n\n\tif sc.PerPage > 1 && sc.PerPage != 25 {\n\t\tparams.Add(\"per_page\", strconv.Itoa(sc.PerPage))\n\t}\n\n\tif len(sc.Key) > 0 {\n\t\tparams.Add(\"key\", sc.Key)\n\t}\n\n\tif len(sc.SortDirection) > 0 {\n\t\tparams.Add(\"sd\", sc.SortDirection)\n\t}\n\n\tif len(sc.SortField) > 0 {\n\t\tparams.Add(\"sf\", sc.SortField)\n\t}\n\n\treturn params\n}", "func (o *GetUsersParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Connection != nil {\n\n\t\t// query param connection\n\t\tvar qrConnection string\n\t\tif o.Connection != nil {\n\t\t\tqrConnection = *o.Connection\n\t\t}\n\t\tqConnection := qrConnection\n\t\tif qConnection != \"\" {\n\t\t\tif err := r.SetQueryParam(\"connection\", qConnection); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Fields != nil {\n\n\t\t// query param fields\n\t\tvar qrFields string\n\t\tif o.Fields != nil {\n\t\t\tqrFields = *o.Fields\n\t\t}\n\t\tqFields := qrFields\n\t\tif qFields != \"\" {\n\t\t\tif err := r.SetQueryParam(\"fields\", qFields); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.IncludeFields != nil {\n\n\t\t// query param include_fields\n\t\tvar qrIncludeFields bool\n\t\tif o.IncludeFields != nil {\n\t\t\tqrIncludeFields = *o.IncludeFields\n\t\t}\n\t\tqIncludeFields := swag.FormatBool(qrIncludeFields)\n\t\tif qIncludeFields != \"\" {\n\t\t\tif err := r.SetQueryParam(\"include_fields\", qIncludeFields); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.IncludeTotals != nil {\n\n\t\t// query param include_totals\n\t\tvar qrIncludeTotals bool\n\t\tif o.IncludeTotals != nil {\n\t\t\tqrIncludeTotals = *o.IncludeTotals\n\t\t}\n\t\tqIncludeTotals := swag.FormatBool(qrIncludeTotals)\n\t\tif qIncludeTotals != \"\" {\n\t\t\tif err := r.SetQueryParam(\"include_totals\", qIncludeTotals); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Page != nil {\n\n\t\t// query param page\n\t\tvar qrPage int64\n\t\tif o.Page != nil {\n\t\t\tqrPage = *o.Page\n\t\t}\n\t\tqPage := swag.FormatInt64(qrPage)\n\t\tif qPage != \"\" {\n\t\t\tif err := r.SetQueryParam(\"page\", qPage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.PerPage != nil {\n\n\t\t// query param per_page\n\t\tvar qrPerPage int64\n\t\tif o.PerPage != nil {\n\t\t\tqrPerPage = *o.PerPage\n\t\t}\n\t\tqPerPage := swag.FormatInt64(qrPerPage)\n\t\tif qPerPage != \"\" {\n\t\t\tif err := r.SetQueryParam(\"per_page\", qPerPage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Q != nil {\n\n\t\t// query param q\n\t\tvar qrQ string\n\t\tif o.Q != nil {\n\t\t\tqrQ = *o.Q\n\t\t}\n\t\tqQ := qrQ\n\t\tif qQ != \"\" {\n\t\t\tif err := r.SetQueryParam(\"q\", qQ); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.SearchEngine != nil {\n\n\t\t// query param search_engine\n\t\tvar qrSearchEngine string\n\t\tif o.SearchEngine != nil {\n\t\t\tqrSearchEngine = *o.SearchEngine\n\t\t}\n\t\tqSearchEngine := qrSearchEngine\n\t\tif qSearchEngine != \"\" {\n\t\t\tif err := r.SetQueryParam(\"search_engine\", qSearchEngine); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Sort != nil {\n\n\t\t// query param sort\n\t\tvar qrSort string\n\t\tif o.Sort != nil {\n\t\t\tqrSort = *o.Sort\n\t\t}\n\t\tqSort := qrSort\n\t\tif qSort != \"\" {\n\t\t\tif err := r.SetQueryParam(\"sort\", qSort); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func collectParameters(r *http.Request, oauthParams map[string]string) map[string]string {\n\tparams := map[string]string{}\n\tfor key, value := range r.URL.Query() {\n\t\tparams[key] = value[0]\n\t}\n\tfor key, value := range oauthParams {\n\t\tparams[key] = value\n\t}\n\treturn params\n}", "func (o ApiDiagnosticBackendRequestDataMaskingPtrOutput) QueryParams() ApiDiagnosticBackendRequestDataMaskingQueryParamArrayOutput {\n\treturn o.ApplyT(func(v *ApiDiagnosticBackendRequestDataMasking) []ApiDiagnosticBackendRequestDataMaskingQueryParam {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.QueryParams\n\t}).(ApiDiagnosticBackendRequestDataMaskingQueryParamArrayOutput)\n}", "func (tr *MongoDatabase) GetParameters() (map[string]interface{}, error) {\n\tp, err := json.TFParser.Marshal(tr.Spec.ForProvider)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbase := map[string]interface{}{}\n\treturn base, json.TFParser.Unmarshal(p, &base)\n}", "func queryProfile(ctx sdk.Context, path []string, _ abci.RequestQuery, keeper Keeper) ([]byte, error) {\n\tif len(strings.TrimSpace(path[0])) == 0 {\n\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrInvalidRequest, \"DTag or address cannot be empty or blank\")\n\t}\n\n\taddress, err := sdk.AccAddressFromBech32(path[0])\n\tif err != nil {\n\t\taddress = keeper.GetDtagRelatedAddress(ctx, path[0])\n\t\tif address == nil {\n\t\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrInvalidRequest, fmt.Sprintf(\"No address related to this dtag: %s\", path[0]))\n\t\t}\n\n\t}\n\n\taccount, found := keeper.GetProfile(ctx, address)\n\n\tif !found {\n\t\treturn nil, sdkerrors.Wrap(sdkerrors.ErrInvalidRequest,\n\t\t\tfmt.Sprintf(\"Profile with address %s doesn't exists\", path[0]))\n\t}\n\n\tbz, err := codec.MarshalJSONIndent(keeper.Cdc, &account)\n\tif err != nil {\n\t\tpanic(\"could not marshal result to JSON\")\n\t}\n\n\treturn bz, nil\n}", "func (o DiagnosticFrontendRequestDataMaskingPtrOutput) QueryParams() DiagnosticFrontendRequestDataMaskingQueryParamArrayOutput {\n\treturn o.ApplyT(func(v *DiagnosticFrontendRequestDataMasking) []DiagnosticFrontendRequestDataMaskingQueryParam {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.QueryParams\n\t}).(DiagnosticFrontendRequestDataMaskingQueryParamArrayOutput)\n}", "func (n *QriNode) RequestProfileInfo(pinfo pstore.PeerInfo) error {\n\t// Get this repo's profile information\n\tprofile, err := n.Repo.Profile()\n\tif err != nil {\n\t\tfmt.Println(\"error getting node profile info:\", err)\n\t\treturn err\n\t}\n\n\taddrs, err := n.IPFSListenAddresses()\n\tif err != nil {\n\t\treturn err\n\t}\n\tprofile.Addresses = addrs\n\n\tres, err := n.SendMessage(pinfo.ID, &Message{\n\t\tType: MtPeerInfo,\n\t\tPayload: profile,\n\t})\n\tif err != nil {\n\t\tfmt.Println(\"send profile message error:\", err.Error())\n\t\treturn err\n\t}\n\n\tif res.Phase == MpResponse {\n\t\tif err := n.handleProfileResponse(pinfo, res); err != nil {\n\t\t\tfmt.Println(\"profile response error\", err.Error())\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func GetQueryCmd(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\t// Group profile queries under a subcommand\n\tprofileQueryCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"Querying commands for the %s module\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tprofileQueryCmd.AddCommand(\n\t\tflags.GetCommands(\n\t\t\tGetCmdBalance(queryRoute, cdc),\n\t\t)...,\n\t)\n\n\treturn profileQueryCmd\n}", "func (a API) Profile(client *http.Client, user_id string, fields Fields) (j map[string]interface{}, err error) {\r\n\treturn a.request(client, \"profile\", map[string]string{\r\n\t\t\"id\": getUserIdString(user_id),\r\n\t\t\"fields\": fields.Encode(),\r\n\t}, nil)\r\n}", "func (tr *CassandraTable) GetParameters() (map[string]interface{}, error) {\n\tp, err := json.TFParser.Marshal(tr.Spec.ForProvider)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbase := map[string]interface{}{}\n\treturn base, json.TFParser.Unmarshal(p, &base)\n}", "func (s *Scraper) FetchSearchProfiles(query string, maxProfilesNbr int, cursor string) ([]*Profile, string, error) {\n\ttimeline, err := s.getSearchTimeline(query, maxProfilesNbr, cursor)\n\tif err != nil {\n\t\treturn nil, \"\", err\n\t}\n\tusers, nextCursor := timeline.parseUsers()\n\treturn users, nextCursor, nil\n}", "func (k Keeper) Params(c context.Context, _ *types.QueryParamsRequest) (*types.QueryParamsResponse, error) {\n\tctx := sdk.UnwrapSDKContext(c)\n\tparams := k.GetParams(ctx)\n\n\treturn &types.QueryParamsResponse{Params: params}, nil\n}", "func (cli *Client) PeopleProfile(id string, fields []string) (r map[string]interface{}, e error) {\n\tvar opt map[string]interface{}\n\n\tif len(fields) > 0 {\n\t\topt = map[string]interface{}{\n\t\t\t\"fields\": fields,\n\t\t}\n\t}\n\n\tr, e = cli.call(\"GET\", \"people\", id, \"\", opt)\n\n\treturn r, e\n}", "func retrieve_gmm_gwy_profiles_list(gmm_api_key string, org_id int) {\n\n\ttype gwy_profiles struct {\n\t\tGatewayProfiles []struct {\n\t\t\tID int `json:\"id\"`\n\t\t\tName string `json:\"name\"`\n\t\t} `json:\"gateway_profiles\"`\n\t\tPaging struct {\n\t\t\tLimit int `json:\"limit\"`\n\t\t\tOffset int `json:\"offset\"`\n\t\t\tPages int `json:\"pages\"`\n\t\t\tCount int `json:\"count\"`\n\t\t\tLinks struct {\n\t\t\t\tFirst string `json:\"first\"`\n\t\t\t\tLast string `json:\"last\"`\n\t\t\t\tNext string `json:\"next\"`\n\t\t\t} `json:\"links\"`\n\t\t} `json:\"paging\"`\n\t}\n\n\tjsonValue, _ := json.Marshal(\"\")\n\trequest, _ := http.NewRequest(\"GET\", \"https://us.ciscokinetic.io/api/v2/organizations/\" + strconv.Itoa(org_id) + \"/gateway_profiles?limit=100\", bytes.NewBuffer(jsonValue))\n\ttoken := \"Token \" + gmm_api_key\n\trequest.Header.Set(\"Authorization\", token)\n\tclient := &http.Client{}\n\tr, err := client.Do(request)\n\n\tif err != nil {\n\t\tfmt.Printf(\"Retrieve GMM GWY Profiles error %s\\n\", err)\n\t\tos.Exit(1)\n\t}\n\n\tresponseData, _ := ioutil.ReadAll(r.Body)\n\n\tvar responseObject gwy_profiles\n\te := json.Unmarshal(responseData, &responseObject)\n\tif e != nil {\n\t\tfmt.Println(\"Unmarshall Error: \", e)\n\t}\n\n\tfmt.Println(\"\")\n\tfmt.Println(\"Total Number of Gateway Profiles in GMM: \", len(responseObject.GatewayProfiles))\n\tfmt.Println(\"\")\n\tfmt.Println(\"Gateway Profiles in GMM\")\n\tfmt.Println(\"-----------------------\")\n\tfor i := 0; i < len(responseObject.GatewayProfiles); i++ {\n\t\tfmt.Println(\"Profile-ID: \", responseObject.GatewayProfiles[i].ID, \" Profile Name: \", responseObject.GatewayProfiles[i].Name)\n\t}\n}", "func (s *Service) PermissionProfilesGet(permissionProfileID string) *PermissionProfilesGetOp {\n\treturn &PermissionProfilesGetOp{\n\t\tCredential: s.credential,\n\t\tMethod: \"GET\",\n\t\tPath: strings.Join([]string{\"permission_profiles\", permissionProfileID}, \"/\"),\n\t\tAccept: \"application/json\",\n\t\tQueryOpts: make(url.Values),\n\t\tVersion: esign.APIv2,\n\t}\n}", "func (tr *GremlinDatabase) GetParameters() (map[string]interface{}, error) {\n\tp, err := json.TFParser.Marshal(tr.Spec.ForProvider)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbase := map[string]interface{}{}\n\treturn base, json.TFParser.Unmarshal(p, &base)\n}", "func GetProfiles(ctx context.Context) []ProfileEntry {\n\tvar c *SqlContext\n\ti := ctx.Value(goradd.SqlContext)\n\tif i == nil {\n\t\tpanic(\"Profiling requires a preloaded context.\")\n\t} else {\n\t\tc = i.(*SqlContext)\n\t}\n\n\tif c != nil {\n\t\tp := c.profiles\n\t\tc.profiles = nil\n\t\treturn p\n\t}\n\treturn nil\n}", "func (o ApiDiagnosticFrontendRequestDataMaskingPtrOutput) QueryParams() ApiDiagnosticFrontendRequestDataMaskingQueryParamArrayOutput {\n\treturn o.ApplyT(func(v *ApiDiagnosticFrontendRequestDataMasking) []ApiDiagnosticFrontendRequestDataMaskingQueryParam {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.QueryParams\n\t}).(ApiDiagnosticFrontendRequestDataMaskingQueryParamArrayOutput)\n}", "func (a *App) GetAllProfiles(w http.ResponseWriter, r *http.Request) {\n\thandler.GetAllProfiles(a.DB, w, r)\n}", "func (s *Scraper) SearchProfiles(ctx context.Context, query string, maxProfilesNbr int) <-chan *ProfileResult {\n\treturn getUserTimeline(ctx, query, maxProfilesNbr, s.FetchSearchProfiles)\n}", "func (tr *NotebookWorkspace) GetParameters() (map[string]interface{}, error) {\n\tp, err := json.TFParser.Marshal(tr.Spec.ForProvider)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbase := map[string]interface{}{}\n\treturn base, json.TFParser.Unmarshal(p, &base)\n}", "func getAccountsPageQuery(r *http.Request) (db2.PageQuery, error) {\n\tcursor, err := hchi.GetStringFromURL(r, actions.ParamCursor)\n\tif err != nil {\n\t\treturn db2.PageQuery{}, errors.Wrap(err, \"getting param cursor\")\n\t}\n\n\torder, err := getOrder(r)\n\tif err != nil {\n\t\treturn db2.PageQuery{}, errors.Wrap(err, \"getting param order\")\n\t}\n\n\tlimit, err := getLimit(r, db2.DefaultPageSize, db2.MaxPageSize)\n\tif err != nil {\n\t\treturn db2.PageQuery{}, errors.Wrap(err, \"getting param limit\")\n\t}\n\n\treturn db2.PageQuery{\n\t\tCursor: cursor,\n\t\tOrder: order,\n\t\tLimit: limit,\n\t}, nil\n}", "func TestProfile(t *testing.T) {\n\tvar se profile.Profile\n\n\t// Validate returned fields to match expected name set.\n\tfields := se.Fields()\n\n\tif _, ok := fields[\"address\"]; !ok {\n\t\ttests.Failed(\"Should have a 'address' field\")\n\t}\n\ttests.Passed(\"Should have a 'address' field\")\n\n\tif _, ok := fields[\"user_id\"]; !ok {\n\t\ttests.Failed(\"Should have a 'user_id' field\")\n\t}\n\ttests.Passed(\"Should have a 'user_id' field\")\n\n\tif _, ok := fields[\"public_id\"]; !ok {\n\t\ttests.Failed(\"Should have a 'public_id' field\")\n\t}\n\ttests.Passed(\"Should have a 'public_id' field\")\n\n\tif _, ok := fields[\"first_name\"]; !ok {\n\t\ttests.Failed(\"Should have a 'first_name' field\")\n\t}\n\ttests.Passed(\"Should have a 'first_name' field\")\n\n\tif _, ok := fields[\"last_name\"]; !ok {\n\t\ttests.Failed(\"Should have a 'last_name' field\")\n\t}\n\ttests.Passed(\"Should have a 'last_name' field\")\n}", "func (o DiagnosticBackendResponseDataMaskingPtrOutput) QueryParams() DiagnosticBackendResponseDataMaskingQueryParamArrayOutput {\n\treturn o.ApplyT(func(v *DiagnosticBackendResponseDataMasking) []DiagnosticBackendResponseDataMaskingQueryParam {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.QueryParams\n\t}).(DiagnosticBackendResponseDataMaskingQueryParamArrayOutput)\n}", "func (svc *inmemService) GetProfile(ctx context.Context, id string) (Profile, error) {\n\t// Get the Read lock from the inmemService struct\n\tsvc.mtx.RLock()\n\n\t// Immediately set up a lock release to occur when the function finishes\n\tdefer svc.mtx.RUnlock()\n\n\t// Look for the profile by the `id` function param\n\tprofile, ok := svc.profiles[id]\n\n\t// Check if the profile id was not found in the datastore\n\tif !ok {\n\n\t\t// Return an empty profile and an error informing the caller that the profile was not found\n\t\treturn Profile{}, ErrNotFound\n\n\t}\n\n\t// Return the profile to the caller and a nil error\n\treturn profile, nil\n\n}", "func DefaultListProfile(ctx context.Context, db *gorm1.DB, f *query1.Filtering, s *query1.Sorting, p *query1.Pagination, fs *query1.FieldSelection) ([]*Profile, error) {\n\tin := Profile{}\n\tormObj, err := in.ToORM(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif hook, ok := interface{}(&ormObj).(ProfileORMWithBeforeListApplyQuery); ok {\n\t\tif db, err = hook.BeforeListApplyQuery(ctx, db, f, s, p, fs); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tdb, err = gorm2.ApplyCollectionOperators(ctx, db, &ProfileORM{}, &Profile{}, f, s, p, fs)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif hook, ok := interface{}(&ormObj).(ProfileORMWithBeforeListFind); ok {\n\t\tif db, err = hook.BeforeListFind(ctx, db, f, s, p, fs); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tdb = db.Where(&ormObj)\n\tdb = db.Order(\"id\")\n\tormResponse := []ProfileORM{}\n\tif err := db.Find(&ormResponse).Error; err != nil {\n\t\treturn nil, err\n\t}\n\tif hook, ok := interface{}(&ormObj).(ProfileORMWithAfterListFind); ok {\n\t\tif err = hook.AfterListFind(ctx, db, &ormResponse, f, s, p, fs); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tpbResponse := []*Profile{}\n\tfor _, responseEntry := range ormResponse {\n\t\ttemp, err := responseEntry.ToPB(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tpbResponse = append(pbResponse, &temp)\n\t}\n\treturn pbResponse, nil\n}", "func (tr *Service) GetParameters() (map[string]interface{}, error) {\n\tp, err := json.TFParser.Marshal(tr.Spec.ForProvider)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbase := map[string]interface{}{}\n\treturn base, json.TFParser.Unmarshal(p, &base)\n}", "func RetrieveProfileData(token string, params string) map[string]interface{} {\n\tendpoint := \"https://api.linkedin.com/v1/people/~\"\n\n\tif params == \"\" {\n\t\tparams = profileDataFullParams\n\t}\n\n\tlinkedInClient := LinkedInClient{\n\t\tToken: token,\n\t\tEndpoint: endpoint,\n\t\tParams: params,\n\t\tMethod: GET,\n\t}\n\n\treturn linkedInClient.Retrieve()\n}" ]
[ "0.67599773", "0.6400527", "0.5847315", "0.5535475", "0.55341256", "0.55146354", "0.55065244", "0.5402974", "0.5288138", "0.5251579", "0.520587", "0.51665294", "0.5144454", "0.5124429", "0.5092634", "0.5016223", "0.49747425", "0.4949739", "0.49178165", "0.48966557", "0.4874747", "0.48731953", "0.48654407", "0.48515353", "0.48453668", "0.48432463", "0.48399734", "0.48393244", "0.48160958", "0.48120746", "0.48076355", "0.47946352", "0.47864538", "0.4782093", "0.4776314", "0.4746134", "0.4717208", "0.47133723", "0.47041363", "0.47019476", "0.469463", "0.46934366", "0.46853745", "0.4685142", "0.4677333", "0.46739787", "0.4666918", "0.46626136", "0.46625993", "0.464907", "0.4648696", "0.46283054", "0.46197596", "0.46168166", "0.4615275", "0.4614475", "0.45978913", "0.45947674", "0.45919606", "0.45779002", "0.45751324", "0.4571713", "0.45713636", "0.45682046", "0.4557296", "0.45517114", "0.45459825", "0.45445386", "0.4535132", "0.45328176", "0.45256233", "0.45208308", "0.45198482", "0.45192337", "0.4513938", "0.45114872", "0.450111", "0.44996333", "0.44904664", "0.44899198", "0.44820625", "0.44746476", "0.44727734", "0.44681233", "0.44663817", "0.4464349", "0.44642872", "0.44601226", "0.44433615", "0.44421902", "0.44419503", "0.44416386", "0.4439787", "0.44382098", "0.4437579", "0.44366357", "0.4431303", "0.44291937", "0.44223198", "0.44198743" ]
0.7892636
0
Setup sets up all endpoints and middleware for the server
func Setup() *celerity.Server { db, _ := gorm.Open("sqlite3", "./data.sqlite3") db.LogMode(true) db.AutoMigrate( models.User{}, models.Room{}, models.Message{}, models.Session{}, ) var lobbyCount int db.Model(&models.Room{}).Where("slug = ?", "lobby").Count(&lobbyCount) if lobbyCount == 0 { db.Create(&models.Room{ Slug: "lobby", Name: "The Lobby", Topic: "Welcome to the lobby", }) } h := Handler{db} server := celerity.New() server.Use(middleware.RequestLogger()) server.Pre(middleware.CORS()) server.Channel("chat", "/ws/:token", h.SocketHandler) secured := server.Scope("/") secured.Use(middleware.Aegis(middleware.AegisConfig{ Adapter: AegisAdapter{db}, })) secured.POST("/rooms", h.CreateRoom) secured.GET("/rooms/:slug", h.GetRoom) secured.GET("/validate", h.ValidateSession) server.POST("/signup", h.CreateUser) server.POST("/login", h.Login) server.ServePath("/", "./frontend/build") server.ServeFile("*", "./frontend/build/index.html") return server }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Setup() {\n\trouter = httprouter.New()\n\n\trouter.GET(\"/\", index)\n\trouter.GET(\"/status\", status)\n\n\tevtHandler := event.NewHandler()\n\trouter.POST(\"/events\", evtHandler.Create)\n\trouter.GET(\"/events\", evtHandler.Index)\n\trouter.GET(\"/events/:id\", evtHandler.Show)\n\n\tsubHandler := subscription.NewHandler()\n\trouter.POST(\"/subscriptions\", subHandler.Create)\n\trouter.GET(\"/subscriptions\", subHandler.Index)\n\trouter.GET(\"/subscriptions/:id\", subHandler.Show)\n\trouter.DELETE(\"/subscriptions/:id\", subHandler.Destroy)\n\n\trouter.POST(\"/feeds\", feed.Create)\n\trouter.GET(\"/feeds\", feed.Index)\n\trouter.GET(\"/feeds/:id\", feed.Show)\n}", "func (s *Server) setupEndpoints(r *chi.Mux) {\n\tr.Route(\"/api/v1\", func(r chi.Router) {\n\t\tr.Route(\"/users\", func(r chi.Router) {\n\n\t\t})\n\t})\n}", "func Setup(router *gin.Engine) {\n\trouter.POST(\"/api/login\", controller.Login)\n\trouter.POST(\"/api/logout\", controller.Logout)\n\trouter.GET(\"/api/dashboard\", controller.GetDashboard)\n\trouter.GET(\"/api/user\", controller.GetUser)\n}", "func Setup(ctx context.Context, r *mux.Router, cfg *config.Config, hc health.HealthCheck) {\n\tlog.Event(ctx, \"adding routes\")\n\tr.StrictSlash(true).Path(\"/health\").HandlerFunc(hc.Handler)\n\tr.StrictSlash(true).Path(\"/helloworld\").Methods(\"GET\").HandlerFunc(handlers.HelloWorld(*cfg))\n}", "func Setup() *Server {\n\tgin.SetMode(gin.ReleaseMode)\n\n\t// Set the router as the default one shipped with Gin\n\trouter := gin.Default()\n\n\t// Serve frontend static files\n\trouter.Use(cors.Default())\n\trouter.Use(static.Serve(\"/\", static.LocalFile(frontBuildFolder, true)))\n\tserver := &Server{router}\n\n\t// create productsApi\n\tserver.productsAPI()\n\treturn server\n}", "func (app *Application) Setup() {\n\tapp.setupStatic()\n\tapp.setupMiddleware()\n}", "func Setup(app *fiber.App) {\n\tapp.Post(\"/api/register\", handlers.Register)\n\tapp.Post(\"/api/login\", handlers.Login)\n\tapp.Get(\"/api/user\", handlers.User)\n\tapp.Post(\"/api/logout\", handlers.Logout)\n\tapp.Post(\"/api/forgot-password\", handlers.Forgot)\n\tapp.Post(\"/api/reset\", handlers.Reset)\n}", "func (c Controller) Setup(router *mux.Router) {\n\trouter.Handle(\"/v1/categories\", negroni.New(\n\t\tmiddlewares.NewJWT(c.render),\n\t\tnegroni.Wrap(http.HandlerFunc(c.getCategoriesHandler)),\n\t)).Methods(\"GET\")\n\n\trouter.Handle(\"/v1/categories\", negroni.New(\n\t\tmiddlewares.NewJWT(c.render),\n\t\tnegroni.Wrap(http.HandlerFunc(c.postCategoriesHandler)),\n\t)).Methods(\"POST\")\n}", "func (s *Server) setupRoutes(devMode bool) {\n\ts.Router.NoRoute(NoRoute)\n\tv1 := s.Router.Group(\"/api/v1\")\n\tv1.GET(\"/healthcheck\", s.Healthcheck)\n\n\tv1.GET(\"/articles\", s.GetArticles)\n\tv1.POST(\"/articles\", s.AddArticle)\n\tv1.POST(\"/articles:batch\", s.AddArticles)\n\n\t// Profiler\n\t// URL: https://<IP>:<PORT>/debug/pprof/\n\tif devMode {\n\t\ts.Logger.Info(\"activating pprof (devmode on)\", log.Field(\"type\", \"debug\"))\n\t\tpprof.Register(s.Router)\n\t}\n}", "func (manager *Manager) Setup(viewsManager *views.Manager) {\n\tmanager.viewsManager = viewsManager\n\tmanager.controllerHandlers = make(map[string]*ControllerHandler)\n\n\t// transporter := websocket.Default\n\t// transporter.CheckOrigin = func(req *http.Request) bool {\n\t// \treturn true\n\t// }\n\n\tserver, err := socketio.NewServer(&engineio.Options{\n\t\tTransports: []transport.Transport{\n\t\t\tpolling.Default,\n\t\t\t&websocket.Transport{\n\t\t\t\tCheckOrigin: func(r *http.Request) bool {\n\t\t\t\t\treturn true\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t})\n\n\tif err != nil {\n\t\tlogger.Debug(\"socket.io error %v\", err)\n\t}\n\n\tmanager.server = server\n\n\tserver.OnConnect(\"/\", func(conn socketio.Conn) error {\n\t\tlogger.Debug(\"connected: %v%v - %v - %v\", conn.URL().Path, conn.URL().RawQuery, conn.URL().RawPath, conn.ID())\n\n\t\treturn nil\n\t})\n\n\tserver.OnDisconnect(\"/\", func(conn socketio.Conn, reason string) {\n\t\tlogger.Debug(\"disconnected: %v - %v, %v \", conn.Namespace(), conn.ID(), reason)\n\t})\n}", "func Setup() *http.Server {\n\tfmt.Println(\"Starting Router\")\n\tr := Router()\n\thttp.Handle(\"/\", r)\n\tsrv := Server(r)\n\treturn srv\n}", "func setup(c *caddy.Controller) error {\n\tconfig, err := NewConfigFromCaddy(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\thttpserver.GetConfig(c).AddMiddleware(func(next httpserver.Handler) httpserver.Handler {\n\t\treturn NewHandler(next, config)\n\t})\n\n\treturn nil\n}", "func (r *Router) Setup() error {\n\tr.echo = echo.New()\n\tr.router = echo.NewRouter(r.echo)\n\treturn nil\n}", "func Setup(ESServer, ESIndexPrefix, ESHostname string) {\n\tserverBase = ESServer\n\tindexPrefix = ESIndexPrefix\n\thostMarker = ESHostname\n\tcheckServerPresent()\n}", "func Setup(app *fiber.App) {\n\tapp.Post(\"/api/register\", controllers.Register)\n\tapp.Post(\"/api/login\", controllers.Login)\n\tapp.Get(\"/api/user\", controllers.User)\n\tapp.Post(\"/api/logout\", controllers.Logout)\n\tapp.Post(\"/api/forgot\", controllers.Forgot)\n\tapp.Post(\"/api/reset\", controllers.Reset)\n\n\tapp.Post(\"/api/note\", controllers.CreateNote)\n\tapp.Put(\"/api/note\", controllers.EditNote)\n\tapp.Delete(\"/api/note/:subject\", controllers.DeleteNote)\n\tapp.Get(\"/api/note/:subject\", controllers.GetNote)\n\tapp.Get(\"/api/noteall\", controllers.GetNoteAll)\n\tapp.Get(\"/api/notefilter/:notes\", controllers.GetNoteFilter)\n\n\tapp.Post(\"/api/notebook\", controllers.CreateNoteBook)\n\tapp.Put(\"/api/notebook\", controllers.EditNoteBook)\n\tapp.Delete(\"/api/notebook\", controllers.DeleteNoteBook)\n\tapp.Get(\"/api/notebook\", controllers.GetNoteBook)\n\tapp.Get(\"/api/notebookall\", controllers.GetNoteBookAll)\n\n}", "func SetupRouter(router *gin.Engine) {\n\trouter.POST(\"/mutant\", Mutant)\n\trouter.GET(\"/stats\", Stats)\n\trouter.GET(\"/api/doc/*any\", ginSwagger.WrapHandler(swaggerFiles.Handler))\n}", "func setup() {\n\t// test server\n\tmux = http.NewServeMux()\n\tserver = httptest.NewServer(mux)\n\tconfig.Root = server.URL\n}", "func (manager *Manager) Setup() {\n\tmanager.controllerHandlers = make(map[string]*ControllerHandler)\n\n\tserver, err := socketio.NewServer(nil)\n\n\tif err != nil {\n\t\tlogger.Debug(\"socket.io error %v\", err)\n\t}\n\n\tmanager.server = server\n\n\tserver.OnConnect(\"/\", func(conn socketio.Conn) error {\n\t\tlogger.Debug(\"connected: %v - %v\", conn.Namespace(), conn.ID())\n\n\t\treturn nil\n\t})\n\n\tserver.OnDisconnect(\"/\", func(conn socketio.Conn, reason string) {\n\t\tlogger.Debug(\"disconnected: %v - %v, %v \", conn.Namespace(), conn.ID(), reason)\n\t})\n}", "func (server *WebServer) setupRoutes() {\n\n\t// Fetch credentials\n\tcredPath := path.Join(path.Dir(os.Args[0]), \"credentials.json\")\n\tcreds, err := parseCredsFile(credPath)\n\tif err != nil {\n\t\tlog.Printf(\"Unable to parse credentials file. %v\\n\", err)\n\t\treturn\n\t}\n\n\t// Setup routes\n\tserver.router = mux.NewRouter()\n\tauthMiddleware := basicAuth(creds)\n\tserver.router.Use(authMiddleware)\n\tserver.router.HandleFunc(\"/monitors\", server.handleGetMonitors)\n\tserver.router.HandleFunc(\"/ws/{address}/{screen}\", server.handleWebsocket)\n\t// server.router.HandleFunc(\"/monitors/{address}/{screen}\", server.handleScreenshot)\n\tserver.router.PathPrefix(\"/\").Handler(http.StripPrefix(\"/\", http.FileServer(http.Dir(\"./ui/build\"))))\n}", "func (r *Router) Setup(rg *gin.RouterGroup) {\n\trg.GET(\"/transfersInfo\", r.GetTransfersInfo)\n\trg.PUT(fmt.Sprintf(\"/files/:%s/transfer\", ParamFileID), r.CreateExternalShareRequest)\n\n\trg.GET(fmt.Sprintf(\"/users/:%s/canApproveToUser/:approverID\", ParamUserID), r.CanApproveToUser)\n\trg.GET(fmt.Sprintf(\"/users/:%s/approverInfo\", ParamUserID), r.GetApproverInfo)\n}", "func SetupRoutes(env *Env) *httprouter.Router {\n\trouter := httprouter.New()\n\trouter.GET(\"/api/news/:ticker/:top/:period\", env.GetNews)\n\trouter.GET(\"/health\", HealthCheck)\n\trouter.GET(\"/readiness\", env.ReadinessCheck)\n\treturn router\n}", "func (c Controller) Setup(router *mux.Router) {\n\trouter.\n\t\tHandleFunc(\"/v1/auth\", c.authHandler).\n\t\tMethods(\"POST\")\n\tlog.Println(\"Set up Auth controller.\")\n}", "func Setup(cfg *config.Config) (r router.BitRoute, log logger.Logger, err error) {\n\t// Setup logger\n\tlog = stdlog.New(&logger.Config{\n\t\tLevel: cfg.LogLevel,\n\t\tTime: true,\n\t\tUTC: true,\n\t})\n\n\tlog.Info(\"Version:\", version.RELEASE)\n\tlog.Warnf(\"%s log level is used\", logger.LevelDebug.String())\n\tlog.Infof(\"Service %s listened on %s:%d\", config.SERVICENAME, cfg.LocalHost, cfg.LocalPort)\n\n\t// Define handlers\n\th := handlers.New(log, cfg)\n\n\t// Register new router\n\tr = bitroute.New()\n\n\t// Response for undefined methods\n\tr.SetupNotFoundHandler(h.Base(notFound))\n\n\t// Configure router\n\tr.SetupMiddleware(h.Base)\n\tr.GET(\"/\", h.Root)\n\tr.GET(\"/healthz\", h.Health)\n\tr.GET(\"/readyz\", h.Ready)\n\tr.GET(\"/info\", h.Info)\n\n\treturn\n}", "func setupRoutes(srv *echo.Echo, s *controller) *echo.Echo {\n\n\tpingroute.NewRouter(s.pingController, srv).RegisterRoutes()\n\tuserroute.NewRouter(s.userController, srv).RegisterRoutes()\n\tuploadroute.NewRouter(s.uploadController, srv).RegisterRoutes()\n\thealthcheckroute.NewRouter(s.healthcheckController, srv).RegisterRoutes()\n\n\treturn srv\n}", "func setup() {\n\ttMux = http.NewServeMux()\n\ttServer = httptest.NewServer(tMux)\n\ttConf = NewAPITestingConfig(true)\n\ttClient, _ = NewClient(nil, tConf)\n\tu, _ := url.Parse(tServer.URL + \"/\")\n\ttClient.BaseURL = u\n}", "func Setup() *Server {\n\tlog.Println(\"Setting up webhook receiver\")\n\tserver := &Server{}\n\tinMemoryStore := storage.NewInMemoryStore()\n\thandler := handler.NewHandler(inMemoryStore)\n\tserver.handler = handler\n\tserver.setupHandlers()\n\n\treturn server\n}", "func Setup() {\n\thttp.HandleFunc(\"/\", handleRequest)\n}", "func (h *HelloController) Setup(router *gin.Engine) {\n\trouter.GET(\"/hello\", h.handleHello)\n}", "func (h *Handlers) setupRouter() {\n\tr := chi.NewRouter()\n\n\t// Setup middleware and special handlers\n\tr.Use(h.MetricsCollector)\n\tr.Use(middleware.RealIP)\n\tr.Use(Logger)\n\tr.Use(middleware.Recoverer)\n\tif h.cfg.GetBool(\"server.basicAuth.enabled\") {\n\t\tr.Use(h.Users.BasicAuth)\n\t}\n\tr.NotFound(h.Static.ServeIndex)\n\n\t// API\n\tr.Route(\"/api/v1\", func(r chi.Router) {\n\t\tr.Route(\"/packages\", func(r chi.Router) {\n\t\t\tr.Get(\"/stats\", h.Packages.GetStats)\n\t\t\tr.Get(\"/updates\", h.Packages.GetUpdates)\n\t\t\tr.Get(\"/search\", h.Packages.Search)\n\t\t\tr.With(h.Users.RequireLogin).Get(\"/starred\", h.Packages.GetStarredByUser)\n\t\t})\n\t\tr.Route(\"/package\", func(r chi.Router) {\n\t\t\tr.Route(\"/chart/{repoName}/{packageName}\", func(r chi.Router) {\n\t\t\t\tr.Get(\"/{version}\", h.Packages.Get)\n\t\t\t\tr.Get(\"/\", h.Packages.Get)\n\t\t\t})\n\t\t\tr.Route(\"/{^falco$|^opa$}/{packageName}\", func(r chi.Router) {\n\t\t\t\tr.Get(\"/{version}\", h.Packages.Get)\n\t\t\t\tr.Get(\"/\", h.Packages.Get)\n\t\t\t})\n\t\t\tr.Route(\"/{packageID}/stars\", func(r chi.Router) {\n\t\t\t\tr.With(h.Users.InjectUserID).Get(\"/\", h.Packages.GetStars)\n\t\t\t\tr.With(h.Users.RequireLogin).Put(\"/\", h.Packages.ToggleStar)\n\t\t\t})\n\t\t})\n\t\tr.Post(\"/users\", h.Users.RegisterUser)\n\t\tr.Route(\"/user\", func(r chi.Router) {\n\t\t\tr.Use(h.Users.RequireLogin)\n\t\t\tr.Get(\"/\", h.Users.GetProfile)\n\t\t\tr.Get(\"/orgs\", h.Organizations.GetByUser)\n\t\t\tr.Put(\"/password\", h.Users.UpdatePassword)\n\t\t\tr.Put(\"/profile\", h.Users.UpdateProfile)\n\t\t\tr.Route(\"/chart-repositories\", func(r chi.Router) {\n\t\t\t\tr.Get(\"/\", h.ChartRepositories.GetOwnedByUser)\n\t\t\t\tr.Post(\"/\", h.ChartRepositories.Add)\n\t\t\t})\n\t\t\tr.Route(\"/chart-repository/{repoName}\", func(r chi.Router) {\n\t\t\t\tr.Put(\"/\", h.ChartRepositories.Update)\n\t\t\t\tr.Delete(\"/\", h.ChartRepositories.Delete)\n\t\t\t})\n\t\t})\n\t\tr.With(h.Users.RequireLogin).Post(\"/orgs\", h.Organizations.Add)\n\t\tr.Route(\"/org/{orgName}\", func(r chi.Router) {\n\t\t\tr.Get(\"/\", h.Organizations.Get)\n\t\t\tr.Group(func(r chi.Router) {\n\t\t\t\tr.Use(h.Users.RequireLogin)\n\t\t\t\tr.Put(\"/\", h.Organizations.Update)\n\t\t\t\tr.Get(\"/accept-invitation\", h.Organizations.ConfirmMembership)\n\t\t\t\tr.Get(\"/members\", h.Organizations.GetMembers)\n\t\t\t\tr.Route(\"/member/{userAlias}\", func(r chi.Router) {\n\t\t\t\t\tr.Post(\"/\", h.Organizations.AddMember)\n\t\t\t\t\tr.Delete(\"/\", h.Organizations.DeleteMember)\n\t\t\t\t})\n\t\t\t\tr.Route(\"/chart-repositories\", func(r chi.Router) {\n\t\t\t\t\tr.Get(\"/\", h.ChartRepositories.GetOwnedByOrg)\n\t\t\t\t\tr.Post(\"/\", h.ChartRepositories.Add)\n\t\t\t\t})\n\t\t\t\tr.Route(\"/chart-repository/{repoName}\", func(r chi.Router) {\n\t\t\t\t\tr.Put(\"/\", h.ChartRepositories.Update)\n\t\t\t\t\tr.Delete(\"/\", h.ChartRepositories.Delete)\n\t\t\t\t})\n\t\t\t})\n\t\t})\n\t\tr.Route(\"/check-availability\", func(r chi.Router) {\n\t\t\tr.Head(\"/{resourceKind:^chartRepositoryName$|^chartRepositoryURL$}\", h.ChartRepositories.CheckAvailability)\n\t\t\tr.Head(\"/{resourceKind:^organizationName$}\", h.Organizations.CheckAvailability)\n\t\t\tr.Head(\"/{resourceKind:^userAlias$}\", h.Users.CheckAvailability)\n\t\t})\n\t\tr.Post(\"/verify-email\", h.Users.VerifyEmail)\n\t\tr.Post(\"/login\", h.Users.Login)\n\t\tr.With(h.Users.RequireLogin).Get(\"/logout\", h.Users.Logout)\n\t\tr.With(h.Users.RequireLogin).Post(\"/images\", h.Static.SaveImage)\n\t})\n\n\t// Oauth\n\tproviders := make([]string, 0, len(h.cfg.GetStringMap(\"server.oauth\")))\n\tfor provider := range h.cfg.GetStringMap(\"server.oauth\") {\n\t\tproviders = append(providers, fmt.Sprintf(\"^%s$\", provider))\n\t}\n\tif len(providers) > 0 {\n\t\tr.Route(fmt.Sprintf(\"/oauth/{provider:%s}\", strings.Join(providers, \"|\")), func(r chi.Router) {\n\t\t\tr.Get(\"/\", h.Users.OauthRedirect)\n\t\t\tr.Get(\"/callback\", h.Users.OauthCallback)\n\t\t})\n\t}\n\n\t// Index special entry points\n\tr.Route(\"/package\", func(r chi.Router) {\n\t\tr.Route(\"/chart/{repoName}/{packageName}\", func(r chi.Router) {\n\t\t\tr.With(h.Packages.InjectIndexMeta).Get(\"/{version}\", h.Static.ServeIndex)\n\t\t\tr.With(h.Packages.InjectIndexMeta).Get(\"/\", h.Static.ServeIndex)\n\t\t})\n\t\tr.Route(\"/{^falco$|^opa$}/{packageName}\", func(r chi.Router) {\n\t\t\tr.With(h.Packages.InjectIndexMeta).Get(\"/{version}\", h.Static.ServeIndex)\n\t\t\tr.With(h.Packages.InjectIndexMeta).Get(\"/\", h.Static.ServeIndex)\n\t\t})\n\t})\n\n\t// Static files and index\n\tstaticFilesPath := path.Join(h.cfg.GetString(\"server.webBuildPath\"), \"static\")\n\tstatic.FileServer(r, \"/static\", http.Dir(staticFilesPath))\n\tr.Get(\"/image/{image}\", h.Static.Image)\n\tr.Get(\"/\", h.Static.ServeIndex)\n\n\th.Router = r\n}", "func Setup(r *mux.Router, service data.IService, validator token.Validator, logger log.ILog) {\n\tdataRoute := r.PathPrefix(\"/api/v1/user\").Subrouter()\n\thandle := handler{logger}\n\tdataRoute.HandleFunc(\"/me\", validator.ValidateMiddleware(\n\t\tfunc(writer http.ResponseWriter, request *http.Request) {\n\t\t\thandle.getUser(writer, request, service)\n\t\t})).Methods(\"GET\")\n\n\tdataRoute.HandleFunc(\"/me/achievements\", validator.ValidateMiddleware(\n\t\tfunc(writer http.ResponseWriter, request *http.Request) {\n\t\t\thandle.getUserAchievements(writer, request, service)\n\t\t})).Methods(\"GET\")\n}", "func Init(e *echo.Echo) {\n e.Use(middleware.Logger())\n SetupRoutes(e)\n}", "func ServerSetup(s *LoggingServeMux, relaynum int) {\n\ts.RegistCompressHandler(ServerURL+\"/ping\", doPing)\n\ts.RegistCompressHandler(ServerURL+\"/node\", doNode)\n\ts.RegistCompressHandler(ServerURL+\"/join/\", doJoin)\n\ts.RegistCompressHandler(ServerURL+\"/bye/\", doBye)\n\ts.RegistCompressHandler(ServerURL+\"/have/\", doHave)\n\ts.RegistCompressHandler(ServerURL+\"/removed/\", doGetHead)\n\ts.RegistCompressHandler(ServerURL+\"/get/\", doGetHead)\n\ts.RegistCompressHandler(ServerURL+\"/head/\", doGetHead)\n\ts.RegistCompressHandler(ServerURL+\"/update/\", doUpdate)\n\ts.RegistCompressHandler(ServerURL+\"/recent/\", doRecent)\n\ts.RegistCompressHandler(ServerURL+\"/\", doMotd)\n\tif relaynum > 0 {\n\t\ts.HandleFunc(ServerURL+\"/proxy/\", doProxy)\n\t\ts.HandleFunc(ServerURL+\"/relay/\", doRelay)\n\t\ts.Handle(ServerURL+\"/request_relay/\", websocket.Handler(websocketRelay(relaynum)))\n\t}\n}", "func Setup() *chi.Mux {\n\trouter := chi.NewRouter()\n\n\trouter.Route(\"/task\", func(r chi.Router) {\n\t\tr.Get(\"/admin/features\", controllers.Features)\n\t})\n\n\treturn router\n}", "func setup() {\n\t// test server\n\tmux = http.NewServeMux()\n\tserver = httptest.NewServer(mux)\n}", "func (s *Server) setupMiddleware() {\n\ts.engine.Use(gin.Recovery())\n\ts.engine.Use(MetricMiddleware())\n\ts.engine.Use(ErrorReportingMiddleware())\n}", "func SetupRouter() *gin.Engine {\n\n\trouter := gin.Default()\n\n\t// Giving access to storage folder\n\trouter.Static(\"/storage\", \"storage\")\n\n\t// Giving access to template folder\n\trouter.Static(\"/templates\", \"templates\")\n\trouter.LoadHTMLGlob(\"templates/*\")\n\n\trouter.Use(CORSMiddleware())\n\n\t// API route for version 1\n\tv1.SetupUserRoutes(router)\n\tv1.SetupMasterUserRoutes(router)\n\tv1.SetupTenantRoutes(router)\n\n\treturn router\n}", "func SetupRouter() *gin.Engine {\n\tr := gin.Default()\n\tr.Use(cors.Default())\n\n\tr.GET(\"/api/heroes\", ctlHero.Get)\n\tr.PUT(\"/api/heroes\", ctlHero.Update)\n\tr.POST(\"/api/heroes\", ctlHero.Add)\n\tr.GET(\"/api/heroes/:id\", ctlHero.GetByID)\n\tr.DELETE(\"/api/heroes/:id\", ctlHero.DeleteByID)\n\tr.GET(\"/api/heroes/\", ctlHero.Search)\n\n\treturn r\n}", "func Setup(cfg *viper.Viper, svc *Services) *Handlers {\n\th := &Handlers{\n\t\tcfg: cfg,\n\t\tsvc: svc,\n\t\tmetrics: setupMetrics(),\n\t\tlogger: log.With().Str(\"handlers\", \"root\").Logger(),\n\n\t\tOrganizations: org.NewHandlers(svc.OrganizationManager),\n\t\tUsers: user.NewHandlers(svc.UserManager, cfg),\n\t\tPackages: pkg.NewHandlers(svc.PackageManager),\n\t\tChartRepositories: chartrepo.NewHandlers(svc.ChartRepositoryManager),\n\t\tStatic: static.NewHandlers(cfg, svc.ImageStore),\n\t}\n\th.setupRouter()\n\th.setupBotRouter()\n\treturn h\n}", "func Setup() *gin.Engine {\n\tapp := gin.New()\n\n\t// Logging to a file.\n\tf, _ := os.Create(\"log/api.log\")\n\tgin.DisableConsoleColor()\n\tgin.DefaultWriter = io.MultiWriter(f)\n\n\t// Middlewares\n\tapp.Use(gin.LoggerWithFormatter(func(param gin.LogFormatterParams) string {\n\t\treturn fmt.Sprintf(\"%s - - [%s] \\\"%s %s %s %d %s \\\" \\\" %s\\\" \\\" %s\\\"\\n\",\n\t\t\tparam.ClientIP,\n\t\t\tparam.TimeStamp.Format(\"02/Jan/2006:15:04:05 -0700\"),\n\t\t\tparam.Method,\n\t\t\tparam.Path,\n\t\t\tparam.Request.Proto,\n\t\t\tparam.StatusCode,\n\t\t\tparam.Latency,\n\t\t\tparam.Request.UserAgent(),\n\t\t\tparam.ErrorMessage,\n\t\t)\n\t}))\n\tapp.Use(gin.Recovery())\n\tapp.NoRoute(middlewares.NoRouteHandler())\n\n\t// Routes\n\tapp.GET(\"/api/contributions\", controllers.GetContributionsChart)\n\n\treturn app\n}", "func (r *Router) SetupRouter() *gin.Engine {\n\trouter := gin.New()\n\n\t//middleware setup\n\trouter.Use(ginglog.Logger(5), gin.Recovery())\n\n\t//diagnostic endpoint\n\tdiagnostic := router.Group(\"api/v1\")\n\t{\n\t\tdiagnostic.GET(\"/ping\", func(c *gin.Context) {\n\t\t\tc.JSON(http.StatusOK, gin.H{\n\t\t\t\t\"Name\": \"Friend Management\",\n\t\t\t\t\"message\": \"OK\",\n\t\t\t\t\"serverTime\": time.Now().UTC(),\n\t\t\t\t\"version\": \"0.1\",\n\t\t\t})\n\t\t})\n\t}\n\n\t//friend endpoint\n\tfriend := router.Group(\"api/v1/friend\")\n\t{\n\t\tfriend.POST(\"/connect\", r.friendController.CreateConnection)\n\t\tfriend.POST(\"/list\", r.friendController.List)\n\t\tfriend.POST(\"/common\", r.friendController.Common)\n\t}\n\n\t//notification endpoint\n\tnotification := router.Group(\"api/v1/notification\")\n\t{\n\t\tnotification.POST(\"/subscribe\", r.notificationController.Subscribe)\n\t\tnotification.POST(\"/update\", r.notificationController.Update)\n\t\tnotification.POST(\"/block\", r.notificationController.Block)\n\t}\n\n\treturn router\n}", "func (srv *AppServer) SetupRoutes() error {\n\tvar err error\n\tsrv.routesSetup.Do(func() {\n\n\t\t// /status is something you should always have in any of your services,\n\t\t// please leave it as is.\n\t\tsrv.router.HandleFunc(\"/status\", srv.HandleStatus()).Queries(\"format\", \"{format:(?:json|yaml)}\").\n\t\t\tName(\"status\").\n\t\t\tMethods(\"GET\")\n\n\t\t// ADD YOUR OWN ROUTES HERE\n\t\tsrv.router.HandleFunc(\"/topology\", srv.HandleTopology()).\n\t\t\tName(\"topology\").\n\t\t\tMethods(\"GET\")\n\t})\n\treturn err\n}", "func SetupRoutes() {\n\tPORT := \":8080\"\n\tlog.Info(\"Starting application on port\" + PORT)\n\n\tr := mux.NewRouter()\n\tr.Use(CORS)\n\n\tr.HandleFunc(\"/upload\", UploadHandler)\n\tr.HandleFunc(\"/\", RedirectToUpload)\n\tr.HandleFunc(\"/records\", GetAllRecords)\n\tr.HandleFunc(\"/profit\", GetProfitsByDate)\n\tr.HandleFunc(\"/topfive\", GetTopFiveProfitableItems)\n\thttp.ListenAndServe(PORT, r)\n}", "func (srv *WebServer) setupMux() {\n\thttp.HandleFunc(\"/\", srv.logResponse(srv.handleStatus)) // GET.\n\thttp.HandleFunc(\"/vehiclestore/status\", srv.logResponse(srv.handleStoreStatus)) // GET.\n\thttp.HandleFunc(\"/lookup\", srv.logResponse(srv.handleLookup)) // GET.\n\thttp.HandleFunc(\"/vehicle\", srv.logResponse(srv.handleVehicle)) // PATCH.\n}", "func SetupRouter() *gin.Engine {\n\trouter := gin.Default() // Create router\n\trouter.GET(\"/\", Heartbeat)\n\trouter.GET(\"/helloget/:arg\", Helloget)\n\n\treturn router\n}", "func (r *Router) SetupRouter() *gin.Engine {\n\trouter := gin.New()\n\tr.log.LogInfo(\"starting gin\")\n\t//middleware setup\n\t//duration := time.Duration(5) * time.Second\n\n\trouter.Use(apmgin.Middleware(router))\n\n\trouter.Use(cors.New(cors.Config{\n\t\tAllowAllOrigins: true,\n\t\tAllowMethods: []string{\"PUT\", \"PATCH\", \"GET\", \"POST\", \"DELETE\"},\n\t\tAllowHeaders: []string{\"Origin\", \"Authorization\", \"Content-Type\", \"Access-Control-Allow-Origin\", \"Access-Control-Allow-Headers\"},\n\t\tExposeHeaders: []string{\"Content-Length\"},\n\t\tAllowCredentials: true,\n\t}))\n\n\tv1 := router.Group(\"api/v1/\")\n\t{\n\t\tdiagnostic := v1.Group(\"diagnostic\")\n\t\t{\n\t\t\tdiagnostic.GET(\"/\", func(c *gin.Context) {\n\t\t\t\tc.JSON(http.StatusOK, gin.H{\n\t\t\t\t\t\"Name\": \"Kata-Rest-API\",\n\t\t\t\t\t\"message\": \"OK\",\n\t\t\t\t\t\"serverTime\": time.Now().UTC(),\n\t\t\t\t\t\"version\": \"0.1\",\n\t\t\t\t})\n\t\t\t})\n\t\t}\n\t}\n\n\treturn router\n}", "func SetupRouter() *gin.Engine {\n\n\tr := gin.Default()\n\tr.Use(errorHandler)\n\n\t// Routes for authenticated only users\n\tauthorizedV1 := r.Group(\"v1/\", basicAuth())\n\n\t// User actions\n\n\t// swagger:route GET /me user getCurrentUser\n\t// Return currently authenticated user's information\n\t// responses:\n\t// 400: ResponseError\n\t// 401: ResponseError\n\t// 200: UserResponse\n\t// security:\n\t// basic:\n\tauthorizedV1.GET(\"me\", controllers.GetCurrentUser)\n\t// swagger:route GET /logout user logout\n\t// Log out current user\n\t// responses:\n\t// 401: ResponseError\n\t// security:\n\t// basic:\n\tauthorizedV1.GET(\"logout\", responseUnauthorized)\n\n\t// Short links actions\n\n\t// swagger:route GET /shorts shortlink getShortlinks\n\t// Return list of short links created by currently authenticated user\n\t// responses:\n\t// 400: ResponseError\n\t// 401: ResponseError\n\t// 200: ShortlinksResponse\n\t// security:\n\t// basic:\n\tauthorizedV1.GET(\"shorts\", controllers.GetShortlinks)\n\t// swagger:route GET /short/{id} shortlink getShortlink\n\t// Return information about specific short link that was created by currently authenticated user\n\t// responses:\n\t// 400: ResponseError\n\t// 401: ResponseError\n\t// 200: ShortlinkResponse\n\t// 404: ResponseError\n\t// security:\n\t// basic:\n\tauthorizedV1.GET(\"shorts/:id\", controllers.GetShortlinkInfo)\n\t// swagger:route POST /shorts shortlink addShortlink\n\t// Create a new short link\n\t// responses:\n\t// 400: ResponseError\n\t// 401: ResponseError\n\t// 201: AddShortResponse\n\t// security:\n\t// basic:\n\tauthorizedV1.POST(\"shorts\", controllers.AddShortlink)\n\t// swagger:route DELETE /shorts shortlink deleteShortlink\n\t// Delete specific short link that was created by currently authenticated user\n\t// responses:\n\t// 400: ResponseError\n\t// 401: ResponseError\n\t// 200: ResponseOK\n\t// 404: ResponseError\n\t// security:\n\t// basic:\n\tauthorizedV1.DELETE(\"shorts/:id\", controllers.DeleteShortlink)\n\n\tpublicV1 := r.Group(\"v1/\")\n\n\t// swagger:route POST /users user addUser\n\t// Create a new user\n\t// responses:\n\t// 400: ResponseError\n\t// 201: ResponseOK\n\tpublicV1.POST(\"users\", controllers.AddUser)\n\t// swagger:route GET /s/{short} shortlink redirectByShortlink\n\t// Redirect to a full link by a given short link\n\t// responses:\n\t// 301: RedirectResponse\n\t// 400: ResponseError\n\tpublicV1.GET(\"s/:short\", controllers.GetShortlinkRedirect)\n\n\tpublicV1Stats := publicV1.Group(\"stats/\")\n\n\t// swagger:route GET /stats/top stats getShortlinksTop\n\t// Return top 20 websites that were most often redirected to\n\t// responses:\n\t// 400: ResponseError\n\t// 200: TopDomainsResponse\n\tpublicV1Stats.GET(\"top\", controllers.GetShortlinksTop)\n\t// swagger:route GET /stats/graph stats getShortlinksGraph\n\t// Return amount of redirects groupped by day, hour and minute\n\t// responses:\n\t// 400: ResponseError\n\t// 200: ShortlinksGraphResponse\n\tpublicV1Stats.GET(\"graph\", controllers.GetShortlinksGraph)\n\n\tr.NoRoute(func(c *gin.Context) {\n\t\tc.JSON(http.StatusNotFound, h.NewResponseError(h.NewPageNotFoundError()))\n\t})\n\n\treturn r\n}", "func SetUpRouteHandlers() *mux.Router {\n\trouter := mux.NewRouter()\n\trouter.Handle(\"/metrics\", promhttp.Handler())\n\trouter.HandleFunc(\"/snippets\", GetSnippetsEndpoint).Methods(\"GET\")\n\trouter.HandleFunc(\"/snippet/{id}\", GetSnippetEndpoint).Methods(\"GET\")\n\trouter.HandleFunc(\"/snippet/\", CreateSnippetEndpoint).Methods(\"POST\")\n\trouter.HandleFunc(\"/snippet/{id}\", DeleteSnippetEndpoint).Methods(\"DELETE\")\n\trouter.HandleFunc(\"/status\", StatusHandler).Methods(\"GET\")\n\treturn router\n}", "func (controller *TaskController) Setup() {\n\tcontroller.Router.Add(\"GET\", \"/tasks\", controller.listTasks)\n}", "func SetupRoutes(app *fiber.App) {\n\tapi := app.Group(\"api/v1/\")\n\tusers := app.Group(\"/auth/v1\")\n}", "func setupRouter() *gin.Engine {\n\tapp := gin.New()\n\tgin.SetMode(gin.DebugMode)\n\n\t// Middlewares\n\t// Allow CORS from the AWS frontend. We store the AWS url in the environment\n\tapp.Use(cors.New(cors.Config{\n\t\tAllowOrigins: []string{\"http://localhost:3000\", os.Getenv(\"FRONTEND_URL\")},\n\t\tAllowMethods: []string{\"PUT\", \"PATCH\", \"OPTIONS\", \"DELETE\"},\n\t\tAllowHeaders: []string{\"Content-Type\", \"Authorization\"},\n\t\tExposeHeaders: []string{\"Content-Type\", \"Authorization\"},\n\t\tAllowCredentials: true,\n\t\tMaxAge: 12 * time.Hour,\n\t}))\n\t/**\n\t\tWebsiteRouter: Handle the dynamic routes for websites\n\t\tAPIRouter: Handles all api calls\n\t**/\n\tAPIRouter := app.Group(\"/api\")\n\t// Register the route\n\t// Inject the database here as well\n\tdb := database.GetDB()\n\n\tapi.Register(APIRouter, db)\n\treturn app\n}", "func (api *APIServer) setupAPIServer(config Config) *http.Server {\n\tr := mux.NewRouter()\n\tr.Use(api.basicAuthMidleware)\n\tr.HandleFunc(\"/\", api.httpRootHandler).Methods(\"GET\")\n\n\tr.HandleFunc(\"/backup/tables\", api.httpTablesHandler).Methods(\"GET\")\n\tr.HandleFunc(\"/backup/list\", api.httpListHandler).Methods(\"GET\")\n\tr.HandleFunc(\"/backup/create\", api.httpCreateHandler).Methods(\"POST\")\n\tr.HandleFunc(\"/backup/clean\", api.httpCleanHandler).Methods(\"POST\")\n\tr.HandleFunc(\"/backup/freeze\", api.httpFreezeHandler).Methods(\"POST\")\n\tr.HandleFunc(\"/backup/upload/{name}\", api.httpUploadHandler).Methods(\"POST\")\n\tr.HandleFunc(\"/backup/download/{name}\", api.httpDownloadHandler).Methods(\"POST\")\n\tr.HandleFunc(\"/backup/restore/{name}\", api.httpRestoreHandler).Methods(\"POST\")\n\tr.HandleFunc(\"/backup/delete/{where}/{name}\", api.httpDeleteHandler).Methods(\"POST\")\n\tr.HandleFunc(\"/backup/config/default\", httpConfigDefaultHandler).Methods(\"GET\")\n\tr.HandleFunc(\"/backup/config\", api.httpConfigHandler).Methods(\"GET\")\n\tr.HandleFunc(\"/backup/config\", api.httpConfigUpdateHandler).Methods(\"POST\")\n\tr.HandleFunc(\"/backup/status\", api.httpBackupStatusHandler).Methods(\"GET\")\n\n\tr.HandleFunc(\"/integration/actions\", api.integrationBackupLog).Methods(\"GET\")\n\tr.HandleFunc(\"/integration/list\", api.httpListHandler).Methods(\"GET\")\n\n\tr.HandleFunc(\"/integration/actions\", api.integrationPost).Methods(\"POST\")\n\n\tvar routes []string\n\tr.Walk(func(route *mux.Route, router *mux.Router, ancestors []*mux.Route) error {\n\t\tt, err := route.GetPathTemplate()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\troutes = append(routes, t)\n\t\treturn nil\n\t})\n\tapi.routes = routes\n\tregisterMetricsHandlers(r, config.API.EnableMetrics, config.API.EnablePprof)\n\n\tsrv := &http.Server{\n\t\tAddr: config.API.ListenAddr,\n\t\tHandler: r,\n\t}\n\treturn srv\n}", "func setup() {\n\tmux = http.NewServeMux()\n\tserver = httptest.NewServer(mux)\n\n\tparams := CommonParams{LoginToken: \"dnspod login token\"}\n\n\tclient = NewClient(params)\n\tclient.BaseURL = server.URL + \"/\"\n}", "func setup(c *caddy.Controller) error {\r\n\tvar err error\r\n\tConfigs, err = parse(c)\r\n\tif err != nil {\r\n\t\treturn err\r\n\t}\r\n\r\n\thttpserver.GetConfig(c).AddMiddleware(func(next httpserver.Handler) httpserver.Handler {\r\n\t\treturn Licserver{Next: next, Configs: Configs}\r\n\t})\r\n\r\n\treturn nil\r\n}", "func SetupRoutes(app *fiber.App) {\n\t// AUTH route\n\tapp.Post(\"/api/v1/login\", handlers.Login)\n\tapp.Post(\"/api/v1/Register\", handlers.Register)\n\n\t// User route\n\tapp.Get(\"api/v1/users\", middlewares.Protected(), handlers.GetAllUsers)\n\tapp.Get(\"api/v1/user/:id\", middlewares.Protected(), handlers.GetUserByID)\n\tapp.Put(\"api/v1/user/:id\", middlewares.Protected(), handlers.UpdateUserByID)\n\tapp.Delete(\"api/v1/user/:id\", middlewares.Protected(), handlers.DeleteUserByID)\n}", "func setuphandlers() {\n\thttp.HandleFunc(\"/\", rootHandler)\n\thttp.HandleFunc(\"/status\", statusHandler)\n\thttp.HandleFunc(\"/stats\", statsHandler)\n\thttp.HandleFunc(\"/request\", requestHandler)\n}", "func Setup(router *mux.Router, accountDB *accounts.Database, deviceDB *devices.Database, levelDB *levels.Database) {\n\troute := routerConfig{map[string]bool{\n\t\t\"/login\": true,\n\t\t\"/register\": true,\n\t\t\"/welcome\": true,\n\t\t\"/trade\": true,\n\t}}\n\trouter.HandleFunc(\"/welcome\", SayWelcome).Methods(http.MethodGet)\n\trouter.HandleFunc(\"/login\", LoginHandler(accountDB, deviceDB, levelDB)).Methods(http.MethodPost)\n\trouter.HandleFunc(\"/register\", RegisterHandler(accountDB, deviceDB, levelDB)).Methods(http.MethodPost)\n\trouter.HandleFunc(\"/registration\", RegistrationHandler(levelDB)).Methods(http.MethodPut, http.MethodDelete, http.MethodGet)\n\n\t// routes with auth = true\n\trouter.HandleFunc(\"/logout\", LogoutHandler(deviceDB)).Methods(http.MethodPost)\n\trouter.HandleFunc(\"/levels\", RouteLevelsHandler(levelDB, sql.NullBool{false, true}, GetAllAccountLevels)).Methods(http.MethodPost)\n\trouter.HandleFunc(\"/levels/{levelname}\", RouteLevelsHandler(levelDB, sql.NullBool{}, RequestLevelByLocalpart)).Methods(http.MethodPost)\n\trouter.HandleFunc(\"/levels/{levelname}/{localpart}\", RouteLevelsHandler(levelDB, sql.NullBool{true, true}, SetLevelByLocalpart)).Methods(http.MethodPut)\n\trouter.HandleFunc(\"/levels/{levelname}/{localpart}\", RouteLevelsHandler(levelDB, sql.NullBool{false, true}, SetLevelByLocalpart)).Methods(http.MethodDelete)\n\trouter.HandleFunc(\"/accounts\", RouteHandlerAccounts(accountDB)).Methods(http.MethodPost)\n\trouter.HandleFunc(\"/devices\", RouteHandlerDevices(deviceDB)).Methods(http.MethodPost)\n\n\t//binance api paths\n\tbinanceProxy := proxy_handles.NewProxy(\"http://localhost:8080\")\n\trouter.PathPrefix(\"/trade\").HandlerFunc(binanceProxy.Handle)\n\n\trouter.Use(route.authMiddleware(deviceDB))\n}", "func SetupRoutes(ctx context.Context) {\n\tlog.Info(ctx, \"routes setup\")\n\tmux.HandleFunc(pat.Get(\"/users\"), getAllUsers)\n\tmux.HandleFunc(pat.Get(\"/users/:user-id\"), getUserByID)\n\tmux.HandleFunc(pat.Get(\"/users/:user-id/tasks\"), getAllTasks)\n\tmux.HandleFunc(pat.Get(\"/users/:user-id/tasks/:task-id\"), getTaskByID)\n\n\t// Middleware\n\tmux.Use(initializeContext)\n\tmux.Use(logAllRequests)\n}", "func (uc *UserController) Setup(r *mux.Router) {\n\tr.HandleFunc(\"/login\", uc.Login).Methods(\"POST\")\n\tr.HandleFunc(\"/signup\", uc.Signup).Methods(\"POST\")\n}", "func SetupRouter() *gin.Engine {\n\tr := gin.Default()\n\tgrp1 := r.Group(\"/api\")\n\t{\n\t\tgrp1.GET(\"login/:email/:password\", controllers.Login)\n\t\tgrp1.GET(\"user\", controllers.GetUsers)\n\t\tgrp1.POST(\"user\", controllers.CreateUser)\n\t\tgrp1.GET(\"user/:id\", controllers.GetUserByID)\n\t\tgrp1.PUT(\"user/:id\", controllers.UpdateUser)\n\t\tgrp1.DELETE(\"user/:id\", controllers.DeleteUser)\n\t}\n\treturn r\n}", "func SetupRoutes(app *fiber.App) {\n\tapi := app.Group(\"/api\")\n\tv1 := api.Group(\"/v1\")\n\tproducts := v1.Group(\"/products\", middleware.AuthReq())\n\tdocs := v1.Group(\"/docs\")\n\tmessages := v1.Group(\"/messages\")\n\n\tdocs.Get(\"\", swagger.Handler)\n\tdocs.Get(\"/*\", swagger.Handler)\n\tproducts.Get(\"\", handler.GetAllProducts)\n\tproducts.Get(\":id\", handler.GetSingleProduct)\n\tproducts.Post(\"\", handler.CreateProduct)\n\tproducts.Delete(\":id\", handler.DeleteProduct)\n\tmessages.Post(\"\", handler.SendSMS)\n}", "func SetupRouter() *gin.Engine {\n\tgin.SetMode(gin.TestMode)\n\tr := gin.New()\n\tr.Use(gin.Recovery())\n\n\tcorsMiddleware := func(c *gin.Context) {\n\t\tc.Header(\"Access-Control-Allow-Origin\", \"*\")\n\t\tc.Header(\"Access-Control-Allow-Credentials\", \"true\")\n\t\tc.Header(\"Access-Control-Allow-Headers\", \"User-Agent, Content-Type, Content-Length, Accept-Encoding, X-CSRF-Token, Authorization, Accept, Origin, Cache-Control, X-Requested-With\")\n\t\tc.Header(\"Access-Control-Allow-Methods\", http.MethodPost)\n\n\t\tif c.Request.Method == http.MethodOptions {\n\t\t\tc.AbortWithStatus(http.StatusNoContent)\n\t\t\treturn\n\t\t}\n\t\tc.Next()\n\t}\n\tr.Use(corsMiddleware)\n\n\thealthRoute := r.Group(\"/health\")\n\t{\n\t\thealthRoute.GET(\"/ping\", health.PerformHealthCheck)\n\t\thealthRoute.GET(\"/dependency\", health.CheckDependencyServices)\n\t}\n\n\tuserRoute := r.Group(\"/user\")\n\t{\n\t\tuserRoute.GET(\"/create\", user.GenerateRandomAccount)\n\t}\n\n\t// Add routing group for DB operations\n\tdbRoute := r.Group(\"/db\")\n\t{\n\t\tdbRoute.POST(\"/create\", db.HandleDBCreation)\n\t\tdbRoute.POST(\"/insert\", db.HandleDBInsertion)\n\t\tdbRoute.POST(\"/select\", db.HandleDBSelection)\n\t\tdbRoute.POST(\"/update\", db.HandleDBUpdate)\n\t\tdbRoute.POST(\"/delete\", db.HandleDBDeletion)\n\t}\n\n\tgrantRoute := r.Group(\"/permission/grant/table\")\n\t{\n\t\tgrantRoute.POST(\"/owner\", permission.GrantTableOwner)\n\t\tgrantRoute.POST(\"/maintainer\", permission.GrantTableMaintainer)\n\t\tgrantRoute.POST(\"/viewer\", permission.GrantTableViewer)\n\t}\n\n\trevokeRoute := r.Group(\"/permission/revoke/table\")\n\t{\n\t\trevokeRoute.POST(\"/owner\", permission.RevokeTableOwner)\n\t\trevokeRoute.POST(\"/maintainer\", permission.RevokeTableMaintainer)\n\t\trevokeRoute.POST(\"/viewer\", permission.RevokeTableViewer)\n\t}\n\n\tverifyRoute := r.Group(\"/permission/verify/table\")\n\t{\n\t\tverifyRoute.POST(\"/owner\", permission.VerifyTableOwner)\n\t\tverifyRoute.POST(\"/maintainer\", permission.VerifyTableMaintainer)\n\t\tverifyRoute.POST(\"/viewer\", permission.TableViewer)\n\t}\n\n\tpaymentRoute := r.Group(\"/payment\")\n\t{\n\t\tpaymentRoute.POST(\"/mint\", payment.MintToken)\n\t\tpaymentRoute.POST(\"/burn\", payment.BurnToken)\n\t\tpaymentRoute.POST(\"/transfer\", payment.TransferToken)\n\t\tpaymentRoute.POST(\"/balance\", payment.GetAccountBalance)\n\t}\n\n\troleRoute := r.Group(\"/role\")\n\t{\n\t\troleRoute.POST(\"/create\", role.CreateColumnRole)\n\t}\n\treturn r\n}", "func Setup(\n\tapiMux *mux.Router, cfg config.Dendrite, // nolint: unparam\n\tqueryAPI api.RoomserverQueryAPI, aliasAPI api.RoomserverAliasAPI, // nolint: unparam\n\taccountDB *accounts.Database, // nolint: unparam\n\tfederation *gomatrixserverlib.FederationClient, // nolint: unparam\n\ttransactionsCache *transactions.Cache, // nolint: unparam\n) {\n\tappMux := apiMux.PathPrefix(pathPrefixApp).Subrouter()\n\n\tappMux.Handle(\"/alias\",\n\t\tcommon.MakeExternalAPI(\"alias\", func(req *http.Request) util.JSONResponse {\n\t\t\t// TODO: Implement\n\t\t\treturn util.JSONResponse{\n\t\t\t\tCode: http.StatusOK,\n\t\t\t\tJSON: nil,\n\t\t\t}\n\t\t}),\n\t).Methods(http.MethodGet, http.MethodOptions)\n\tappMux.Handle(\"/user\",\n\t\tcommon.MakeExternalAPI(\"user\", func(req *http.Request) util.JSONResponse {\n\t\t\t// TODO: Implement\n\t\t\treturn util.JSONResponse{\n\t\t\t\tCode: http.StatusOK,\n\t\t\t\tJSON: nil,\n\t\t\t}\n\t\t}),\n\t).Methods(http.MethodGet, http.MethodOptions)\n}", "func SetupRouter() *gin.Engine {\n\tr := gin.Default()\n\tgrp1 := r.Group(\"/user-api\")\n\t{\n\t\tgrp1.GET(\"products\", Controllers.GetUsers)\n\t\tgrp1.POST(\"product\", Controllers.CreateProd)\n\t\tgrp1.GET(\"product/:id\", Controllers.GetProdByID)\n\t\tgrp1.PATCH(\"product/:id\", Controllers.UpdateProd)\n\t\tgrp1.DELETE(\"product/:id\", Controllers.DeleteUser)\n\t\tgrp1.POST(\"order\", Controllers.CreateOrder)\n\t}\n\treturn r\n}", "func SetupRouter() *gin.Engine {\n\trouter := gin.Default()\n\trouter.Use(middlewares.AppliAllCORS)\n\n\tapig := router.Group(\"/api\")\n\t{\n\t\tv1g := apig.Group(\"/v1\")\n\t\t{\n\t\t\tv1g.Use(middlewares.JSONOnly)\n\t\t\tv1g.POST(\"/user\", v1.InsertUser)\n\t\t}\n\t}\n\treturn router\n}", "func setup() {\n\t// test server\n\tmux = http.NewServeMux()\n\tserver = httptest.NewServer(mux)\n\n\t// configure findface client to use test server\n\tclient = NewClient(token, nil, nil)\n\n\turl, _ := url.Parse(server.URL)\n\tclient.BaseURL = url\n}", "func SetupHandlers() {\n\tif config.Development() {\n\t\thttp.HandleFunc(\"/_hash\", common.Recover(encryption.HashHandler))\n\t\thttp.HandleFunc(\"/_sign\", common.Recover(common.ToJSONResponse(encryption.SignHandler)))\n\t\thttp.HandleFunc(updateConfigURL, common.Recover(ConfigUpdateHandler))\n\t\thttp.HandleFunc(updateConfigAllURL, common.Recover(ConfigUpdateAllHandler))\n\t}\n}", "func SetupRoutes(app *fiber.App) {\n\t// Middleware\n\tapi := app.Group(\"/api\", logger.New())\n\tapi.Get(\"/\", handler.Hello)\n\n\t// Auth\n\tauth := api.Group(\"/auth\")\n\tauth.Post(\"/login\", handler.Login)\n\n\t// User\n\tuser := api.Group(\"/user\")\n\tuser.Get(\"/:id\", handler.GetUser)\n\tuser.Post(\"/\", handler.CreateUser)\n\tuser.Patch(\"/:id\", middleware.Protected(), handler.UpdateUser)\n\tuser.Delete(\"/:id\", middleware.Protected(), handler.DeleteUser)\n\n\t// Scan\n\tscan := api.Group(\"/scan\")\n\tscan.Get(\"/\", handler.GetAllScans)\n\tscan.Get(\"/:id\", handler.GetScan)\n\tscan.Get(\"/:id/details\", handler.GetScanDetails)\n\tscan.Post(\"/\", middleware.Protected(), handler.CreateScan)\n\tscan.Delete(\"/:id\", middleware.Protected(), handler.DeleteScan)\n}", "func SetupHandlers(r *mux.Router) {\n\t//object operations\n\tr.HandleFunc(\"/v1/file/upload/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithConnection(UploadHandler))))\n\tr.HandleFunc(\"/v1/file/download/{allocation}\", common.UserRateLimit(common.ToByteStream(WithConnection(DownloadHandler)))).Methods(\"POST\")\n\tr.HandleFunc(\"/v1/file/rename/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithConnection(RenameHandler))))\n\tr.HandleFunc(\"/v1/file/copy/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithConnection(CopyHandler))))\n\tr.HandleFunc(\"/v1/file/attributes/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithConnection(UpdateAttributesHandler))))\n\tr.HandleFunc(\"/v1/dir/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithConnection(CreateDirHandler)))).Methods(\"POST\")\n\tr.HandleFunc(\"/v1/dir/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithConnection(CreateDirHandler)))).Methods(\"DELETE\")\n\tr.HandleFunc(\"/v1/dir/rename/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithConnection(CreateDirHandler)))).Methods(\"POST\")\n\n\tr.HandleFunc(\"/v1/connection/commit/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithConnection(CommitHandler))))\n\tr.HandleFunc(\"/v1/file/commitmetatxn/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithConnection(CommitMetaTxnHandler))))\n\tr.HandleFunc(\"/v1/file/collaborator/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithConnection(CollaboratorHandler))))\n\tr.HandleFunc(\"/v1/file/calculatehash/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithConnection(CalculateHashHandler))))\n\n\t//object info related apis\n\tr.HandleFunc(\"/allocation\", common.UserRateLimit(common.ToJSONResponse(WithConnection(AllocationHandler))))\n\tr.HandleFunc(\"/v1/file/meta/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithReadOnlyConnection(FileMetaHandler))))\n\tr.HandleFunc(\"/v1/file/stats/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithReadOnlyConnection(FileStatsHandler))))\n\tr.HandleFunc(\"/v1/file/list/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithReadOnlyConnection(ListHandler))))\n\tr.HandleFunc(\"/v1/file/objectpath/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithReadOnlyConnection(ObjectPathHandler))))\n\tr.HandleFunc(\"/v1/file/referencepath/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithReadOnlyConnection(ReferencePathHandler))))\n\tr.HandleFunc(\"/v1/file/objecttree/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithReadOnlyConnection(ObjectTreeHandler))))\n\tr.HandleFunc(\"/v1/file/refs/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithReadOnlyConnection(RefsHandler)))).Methods(\"GET\")\n\t//admin related\n\tr.HandleFunc(\"/_debug\", common.UserRateLimit(common.ToJSONResponse(DumpGoRoutines)))\n\tr.HandleFunc(\"/_config\", common.UserRateLimit(common.ToJSONResponse(GetConfig)))\n\tr.HandleFunc(\"/_stats\", common.UserRateLimit(stats.StatsHandler))\n\tr.HandleFunc(\"/_statsJSON\", common.UserRateLimit(common.ToJSONResponse(stats.StatsJSONHandler)))\n\tr.HandleFunc(\"/_cleanupdisk\", common.UserRateLimit(common.ToJSONResponse(WithReadOnlyConnection(CleanupDiskHandler))))\n\tr.HandleFunc(\"/getstats\", common.UserRateLimit(common.ToJSONResponse(stats.GetStatsHandler)))\n\n\t//marketplace related\n\tr.HandleFunc(\"/v1/marketplace/shareinfo/{allocation}\", common.UserRateLimit(common.ToJSONResponse(WithConnection(MarketPlaceShareInfoHandler))))\n}", "func SetupRoutes() {\n\tplatformsHandler := http.HandlerFunc(handlePlatforms)\n\n\thttp.Handle(\"/platforms\", platformsHandler)\n}", "func SetupRouter(db *gorm.DB) *mux.Router {\n\trouter := mux.NewRouter().StrictSlash(true)\n\trouter.Use(commonMiddleware)\n\n\t//Public paths\n\trouter.Methods(\"OPTIONS\").HandlerFunc(func(w http.ResponseWriter, r *http.Request) {})\n\trouter.HandleFunc(\"/\", func(w http.ResponseWriter, r *http.Request) {\n\t\tw.Write([]byte(\"welcome\"))\n\t})\n\n\trouter.HandleFunc(\"/register\", controllers.CreateUser(db)).Methods(\"POST\")\n\trouter.HandleFunc(\"/login\", controllers.Login(db)).Methods(\"POST\")\n\n\trouter.HandleFunc(\"/quiz/{id}\", exampleQuizzes).Methods(\"GET\")\n\n\t//Authed paths\n\tauthRouter := router.PathPrefix(\"/auth\").Subrouter()\n\tauthRouter.Use(middlewares.JWTVerify)\n\n\treturn router\n}", "func SetupRouter() *gin.Engine {\n\tr := gin.Default()\n\tgrp1 := r.Group(\"/offers\")\n\t{\n\t\tgrp1.POST(\"Publisher\", func(c *gin.Context) {\n\t\t\tcontroller.PublishOffers(c)\n\t\t\t// controller.PublishOffers(c.Writer, c.Request)\n\t\t}) //declare queue\n\t}\n\treturn r\n}", "func Setup(ctx context.Context, cfg *config.Config, r *mux.Router, permissionsStore PermissionsStore) *API {\n\tapi := &API{\n\t\tRouter: r,\n\t\tpermissionsStore: permissionsStore,\n\t\tdefaultLimit: cfg.DefaultLimit,\n\t\tdefaultOffset: cfg.DefaultOffset,\n\t\tmaximumDefaultLimit: cfg.MaximumDefaultLimit,\n\t}\n\n\tr.HandleFunc(\"/roles/{id}\", api.GetRoleHandler).Methods(http.MethodGet)\n\tr.HandleFunc(\"/roles\", api.GetRolesHandler).Methods(http.MethodGet)\n\n\treturn api\n}", "func (h *HomeHandlers) SetupRoutes(mux *http.ServeMux) {\n\t// and now here we can call our middleware\n\tmux.HandleFunc(\"/\", h.Logger(h.home))\n}", "func (h *Handlers) SetupRoutes(mux *http.ServeMux) {\n\tmux.HandleFunc(\"/\", h.Logger(h.Home))\n}", "func setup() {\n\tmux = http.NewServeMux()\n\tserver = httptest.NewServer(mux)\n\n\tclient = New(\"123\")\n\turl, _ := url.Parse(server.URL)\n\tclient.APIURL = url\n}", "func (h *Handlers) SetupRoutes(mux *mux.Router) {\n\tmux.HandleFunc(\"/api/product-categories/new\", productCategoryHandlerLogging.Httplog((auth.ValidateToken(h.handleAddProductCategory, authServer)))).Methods(\"POST\")\n\tmux.HandleFunc(\"/api/product-categories\", productCategoryHandlerLogging.Httplog((auth.ValidateToken(h.handleGetProductCategory, authServer)))).Methods(\"GET\")\n\tmux.HandleFunc(\"/api/product-categories/{id}\", productCategoryHandlerLogging.Httplog((auth.ValidateToken(h.handleUpdateProductCategory, authServer)))).Methods(\"PUT\")\n\tmux.HandleFunc(\"/api/product-categories/{id}\", productCategoryHandlerLogging.Httplog((auth.ValidateToken(h.handleDeleteProductCategory, authServer)))).Methods(\"DELETE\")\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\tc := cors.New(cors.Options{\n\t\tAllowedHeaders: []string{\"X-Auth-Token\", \"Content-Type\", \"Accept\"},\n\t\tAllowedMethods: []string{\"GET\", \"HEAD\", \"POST\", \"DELETE\", \"PUT\"},\n\t\tMaxAge: 600,\n\t})\n\n\tloggingHandler := func(next http.Handler) http.Handler {\n\t\treturn gmiddleware.LoggingHandler(os.Stdout, next)\n\t}\n\tredocHandler := func(next http.Handler) http.Handler {\n\t\treturn middleware.Redoc(middleware.RedocOpts{Path: \"swagger\"}, next)\n\t}\n\n\treturn alice.New(loggingHandler, handlers.RootHandler, redocHandler, StaticFiles, c.Handler).Then(handler)\n}", "func SetupRouter() *gin.Engine {\n\trouter := gin.Default() // Create router\n\trouter.GET(\"/\", Heartbeat)\n\trouter.GET(\"/getboxdifference/:arg\", GetBoxDifference)\n\n\treturn router\n}", "func setup() (client *Client, mux *http.ServeMux, serverURL string, teardown func()) {\n\treturn setupWithPath(\"\")\n}", "func (m JWTAuthMiddleware) Setup() {}", "func SetupRouter() *gin.Engine {\n\tr := gin.Default()\n\tgrp1 := r.Group(\"/user-api\")\n\t{\n\t\tgrp1.GET(\"user\", Controllers.GetUsers)\n\t\tgrp1.POST(\"user\", Controllers.CreateUser)\n\t\tgrp1.GET(\"user/:id\", Controllers.GetUserByID)\n\t\tgrp1.PUT(\"user/:id\", Controllers.UpdateUser)\n\t\tgrp1.DELETE(\"user/:id\", Controllers.DeleteUser)\n\t}\n\tgrp2 := r.Group(\"/category-api\")\n\t{\n\t\tgrp2.GET(\"categories\", Controllers.GetCategories)\n\t}\n\tgrp3 := r.Group(\"/food-api\")\n\t{\n\t\tgrp3.GET(\"foods/:id\", Controllers.GetFoods)\n\t}\n\n\treturn r\n}", "func SetupAPI() *gin.Engine {\n\trouter := gin.Default()\n\n\trouter.GET(\"/\", getTodos)\n\trouter.GET(\"/todo/:id\", getTodo)\n\trouter.POST(\"/create\", createTodo)\n\trouter.POST(\"/update\", updateTodo)\n\trouter.GET(\"/delete/:id\", deleteTodo)\n\n\treturn router\n}", "func (s *Server) Init() {\n\n\tcustomerAuthMd := middleware.Authenticate(s.customersSvc.IDByToken)\n\tcustomersSubrouter := s.mux.PathPrefix(\"/api/customers\").Subrouter()\n\n\tcustomersSubrouter.Use(customerAuthMd)\n\tcustomersSubrouter.HandleFunc(\"\", s.handleCustomerRegistration).Methods(POST)\n\tcustomersSubrouter.HandleFunc(\"/token\", s.handleCustomerGetToken).Methods(POST)\n\tcustomersSubrouter.HandleFunc(\"/products\", s.handleCustomerGetProducts).Methods(GET)\n\tcustomersSubrouter.HandleFunc(\"/purchases\", s.handleCustomerGetPurchases).Methods(GET)\n\n\tmanagersAuthenticateMd := middleware.Authenticate(s.managerSvc.IDByToken)\n\tmanagersSubRouter := s.mux.PathPrefix(\"/api/managers\").Subrouter()\n\tmanagersSubRouter.Use(managersAuthenticateMd)\n\n\tmanagersSubRouter.HandleFunc(\"\", s.handleManagerRegistration).Methods(POST)\n\tmanagersSubRouter.HandleFunc(\"/token\", s.handleManagerGetToken).Methods(POST)\n\tmanagersSubRouter.HandleFunc(\"/sales\", s.handleManagerGetSales).Methods(GET)\n\tmanagersSubRouter.HandleFunc(\"/sales\", s.handleManagerMakeSales).Methods(POST)\n\tmanagersSubRouter.HandleFunc(\"/products\", s.handleManagerGetProducts).Methods(GET)\n\tmanagersSubRouter.HandleFunc(\"/products\", s.handleManagerChangeProducts).Methods(POST)\n\tmanagersSubRouter.HandleFunc(\"/products/{id}\", s.handleManagerRemoveProductByID).Methods(DELETE)\n\tmanagersSubRouter.HandleFunc(\"/customers\", s.handleManagerGetCustomers).Methods(GET)\n\tmanagersSubRouter.HandleFunc(\"/customers\", s.handleManagerChangeCustomer).Methods(POST)\n\tmanagersSubRouter.HandleFunc(\"/customers/{id}\", s.handleManagerRemoveCustomerByID).Methods(DELETE)\n\n}", "func setupRouter(ctx context.Context, config configuration.Config, router *mux.Router) {\n\n\tvar jwtmw = jwtmiddleware.New(jwtmiddleware.Options{\n\t\tValidationKeyGetter: func(token *jwt.Token) (interface{}, error) {\n\t\t\treturn []byte(config.GetString(configuration.SecretKey)), nil\n\t\t},\n\t\tSigningMethod: jwt.SigningMethodHS256,\n\t})\n\n\tstore := connector.New(config)\n\n\tapiRouter := router.PathPrefix(config.GetString(configuration.AppAPIBase)).Subrouter()\n\tservice.Register(apiRouter, \"/authentication\", authenticationservice.New(store, config))\n\tservice.Register(apiRouter, \"/users\", userservice.New(store, config), jwtmw)\n\tservice.Register(apiRouter, \"/items\", itemsservice.New(store, config), jwtmw)\n\tservice.Register(apiRouter, \"/settings\", settingsservice.New(store, config), jwtmw)\n\n}", "func SetupRouter(h Handler) *mux.Router {\n\tr := mux.NewRouter()\n\n\tr.HandleFunc(`/products`,\n\t\tfunc(w http.ResponseWriter, r *http.Request) {\n\t\t\tWrapProducts(h, w, r)\n\t\t}).Methods(\"get\")\n\n\tr.HandleFunc(`/estimates/price/{start_latitude}/{start_longitude}/{end_latitude}/{end_longitude}`,\n\t\tfunc(w http.ResponseWriter, r *http.Request) {\n\t\t\tWrapEstimatesPrice(h, w, r)\n\t\t}).Methods(\"get\")\n\n\tr.HandleFunc(`/estimates/time`,\n\t\tfunc(w http.ResponseWriter, r *http.Request) {\n\t\t\tWrapEstimatesTime(h, w, r)\n\t\t}).Methods(\"get\")\n\n\tr.HandleFunc(`/me`,\n\t\tfunc(w http.ResponseWriter, r *http.Request) {\n\t\t\tWrapUpdateMe(h, w, r)\n\t\t}).Methods(\"patch\")\n\n\tr.HandleFunc(`/upload_infos`,\n\t\tfunc(w http.ResponseWriter, r *http.Request) {\n\t\t\tWrapUploadInfos(h, w, r)\n\t\t}).Methods(\"patch\")\n\n\tr.HandleFunc(`/history`,\n\t\tfunc(w http.ResponseWriter, r *http.Request) {\n\t\t\tWrapHistory(h, w, r)\n\t\t}).Methods(\"get\")\n\n\treturn r\n}", "func (a *Application) SetupPaths() {\n\tif config.Debug {\n\t\t// standard go profiling support\n\t\thttp2.RegisterPrefixHandler(\"/debug/pprof/\", http.HandlerFunc(pprof.Index))\n\t\thttp2.RegisterPrefixHandler(\"/debug/pprof/cmdline\", http.HandlerFunc(pprof.Cmdline))\n\t\thttp2.RegisterPrefixHandler(\"/debug/pprof/profile\", http.HandlerFunc(pprof.Profile))\n\t\thttp2.RegisterPrefixHandler(\"/debug/pprof/symbol\", http.HandlerFunc(pprof.Symbol))\n\t\thttp2.RegisterPrefixHandler(\"/debug/pprof/trace\", http.HandlerFunc(pprof.Trace))\n\t}\n\n\tif config.WebsocketMessengerPrefix != \"\" {\n\t\thttp2.RegisterPrefixHandler(config.WebsocketMessengerPrefix, http.HandlerFunc(WebsocketMessengerHandler))\n\t}\n}", "func SetupHandlers(ctx context.Context) {\n\thttp.Handle(\"/version\", VersionHandler(ctx))\n}", "func (s *Server) Setup() error {\n\tif s.Logger == nil {\n\t\ts.Logger = log.New(os.Stderr, \"\", log.LstdFlags)\n\t}\n\n\ts.Printf(\"Connecting to origin at %s...\", s.OriginHost)\n\torigin, err := connect.Connect(s.OriginHost)\n\tif err != nil {\n\t\ts.Fatalf(\"Failed to connect to origin '%s': %v\", s.OriginHost, err)\n\t}\n\n\topts := map[string]string{}\n\tstartupMsg := protocol.CreateStartupMessage(s.OriginUsername, s.OriginDatabase, opts)\n\torigin.Write(startupMsg)\n\n\tresponse := make([]byte, 4096)\n\tif _, err = origin.Read(response); err != nil {\n\t\treturn fmt.Errorf(\"error connecting to '%s': %v\", s.OriginHost, err)\n\t}\n\n\tif authd := connect.HandleAuthenticationRequest(origin, response); !authd {\n\t\treturn fmt.Errorf(\"Origin authentication failed\")\n\t}\n\n\ts.Printf(\"Successfully connected to origin '%s'\", s.OriginHost)\n\ts.originPool = make(Pool, 20)\n\ts.originPool.Add(origin)\n\treturn nil\n}", "func SetupRouter() {\r\n\r\n\tr2 := gin.Default()\r\n\tgrp3 := r2.Group(\"v1/\")\r\n\t{\r\n\t\tgrp3.GET(\"cart/:custid\", Controllers.GetCartDetailsById)\r\n\t\tgrp3.POST(\"cart\", Controllers.AddToCart)\r\n\t\tgrp3.PUT(\"cart/:cid/:prodid\", Controllers.UpdateItemsInCart)\r\n\t}\r\n\r\n\tr2.Run(\":7002\")\r\n\r\n}", "func (r *Router) SetupHandler() {\n\t// error handler set.\n\teh := handler.NewHTTPErrorHandler(r.LoggerHandler.Log)\n\tr.Mux.NotFound(eh.StatusNotFound)\n\tr.Mux.MethodNotAllowed(eh.StatusMethodNotAllowed)\n\n\tr.Mux.Method(http.MethodGet, \"/static/*\", http.StripPrefix(\"/static/\", http.FileServer(http.Dir(\"static\"))))\n\n\t// user set.\n\tuh := user.NewHTTPHandler(bh, bu, br, r.SQLHandler, r.CacheHandler)\n\tr.Mux.Route(\"/v1\", func(cr chi.Router) {\n\t\tcr.Post(\"/register/device\", uh.RegisterByDevice)\n\t})\n}", "func (r *Router) Setup(rg *gin.RouterGroup) {\n\tcheckShareScope := r.oAuthMiddleware.AuthorizationScopeMiddleware(oauth.ShareScope)\n\n\trg.GET(fmt.Sprintf(\"/files/:%s/permissions\", ParamFileID), r.GetFilePermissions)\n\trg.PUT(fmt.Sprintf(\"/files/:%s/permissions\", ParamFileID), checkShareScope, r.CreateFilePermission)\n\trg.DELETE(fmt.Sprintf(\"/files/:%s/permissions\", ParamFileID), r.DeleteFilePermission)\n}", "func Setup() {\n\tengine = &Engine{\n\t\tmodelManager: &models.Manager{},\n\t\tcontrollerManager: &controllers.Manager{},\n\t\tviewManager: &views.Manager{},\n\t\tactorManager: &actors.Manager{},\n\t}\n\n\tengine.modelManager.Setup()\n\tengine.viewManager.Setup()\n\tengine.controllerManager.Setup(ViewManager())\n\tengine.actorManager.Setup(ViewManager())\n\n\tserveStats()\n}", "func SetupRoutes(engine *gin.Engine) {\n\tENABLE_SWAGGER := configs.Envs[\"ENABLE_SWAGGER\"]\n\tif ENABLE_SWAGGER == \"true\" {\n\t\tengine.GET(\"/swagger/*any\", ginSwagger.WrapHandler(swaggerFiles.Handler))\n\t}\n\n\tfor _, controller := range routes.GetRoutes() {\n\t\tengine.Handle(controller.Method, controller.Path, controller.Handler)\n\t} // setup routes\n}", "func SetupRouter() *gin.Engine {\n\tr := gin.Default()\n\n\t// /schools routes\n\tr.GET(\"/schools\", handlers.ListSchools)\n\tr.POST(\"/schools\", handlers.AddSchool)\n\n\t// /schools/{id} routes\n\tr.GET(\"/schools/:schoolID\", handlers.GetSchool)\n\tr.PUT(\"/schools/:schoolID\", handlers.UpdateSchool)\n\n\t// Documentation routes\n\tr.Static(\"/docs/\", \"./dist/\")\n\n\treturn r\n}", "func SetupRoutes(apiBasePath string) {\n\thandleProducts := http.HandlerFunc(productsHandler)\n\thandleProduct := http.HandlerFunc(productHandler)\n\thttp.Handle(\"/websocket\",websocket.Handler(productSocket))\n\thttp.Handle(fmt.Sprintf(\"%s/%s\", apiBasePath, productsBasePath), cors.MiddlewareHandler(handleProducts))\n\thttp.Handle(fmt.Sprintf(\"%s/%s/\", apiBasePath, productsBasePath), cors.MiddlewareHandler(handleProduct))\n\n}", "func (s *HeadsUpServerController) SetUp(ctx context.Context, st store.RStore) error {\n\tctx, cancel := context.WithCancel(ctx)\n\ts.shutdown = cancel\n\n\terr := s.setUpHelper(ctx, st)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Cannot start the tilt-apiserver: %v\", err)\n\t}\n\terr = s.addToAPIServerConfig()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"writing tilt api configs: %v\", err)\n\t}\n\treturn nil\n}", "func SetupRouter() *gin.Engine {\n\tr := gin.Default()\n\n\tusers := r.Group(\"/users\")\n\t{\n\t\tusers.GET(\"/\", controllers.GetUsers)\n\t\tusers.GET(\"/:id\", controllers.GetUser)\n\t\tusers.POST(\"/\", controllers.CreateUser)\n\t\tusers.PATCH(\"/:id\", controllers.UpdateUser)\n\t\tusers.DELETE(\"/:id\", controllers.DeleteUser)\n\t}\n\n\treturn r\n}", "func SetupRouter() *gin.Engine {\n\trouter := gin.Default()\n\t// router.Use(cors.Default())\n\tconfig := cors.DefaultConfig()\n\tconfig.AllowAllOrigins = true\n\t// config.AllowOrigins = []string{\"*\"}\n\t// config.AddAllowOrigins(\"http://localhost:8080\")\n\tconfig.AllowHeaders = []string{\"Origin\", \"Authorization\", \"Content-Type\"}\n\tconfig.ExposeHeaders = []string{\"Content-Length\"}\n\tconfig.AllowCredentials = true\n\tconfig.MaxAge = 12 * time.Hour // - Preflight requests cached for 12 hours\n\n\trouter.Use(cors.New(config))\n\trouter.GET(cfg.Key(\"PATH\").String()+\"/ping\", GetPing)\n\t// /token=code=xxx\n\trouter.GET(cfg.Key(\"PATH\").String()+\"/token\", GetToken)\n\t// /refresh { token: xxx, client_ID: xxx }\n\trouter.POST(cfg.Key(\"PATH\").String()+\"/refresh\", PostRefresh)\n\treturn router\n}", "func SetupRouter(debug bool, db *sql.DB) (*gin.Engine, error) {\n\t// Setup stores\n\tsessionStore, err := storage.NewSessionStore(db)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Error setting up the user store: %s\", err.Error())\n\t}\n\n\t// Setup controllers\n\tsessionCtrl := NewSessionsCtrl(db, sessionStore)\n\n\t// Setup router\n\tif debug {\n\t\tgin.SetMode(gin.DebugMode)\n\t} else {\n\t\tgin.SetMode(gin.ReleaseMode)\n\t}\n\trouter := gin.Default()\n\n\trouter.Use(cors.New(cors.Config{\n\t\tAllowOrigins: []string{\"*\"},\n\t\tAllowMethods: []string{\"GET\", \"POST\", \"PUT\", \"DELETE\"},\n\t\tAllowCredentials: true,\n\t}))\n\n\trouter.Use(ErrorMiddleware())\n\n\tapi := router.Group(\"api\")\n\n\t// Sessions\n\tapi.GET(\"sessions\", sessionCtrl.GetAll)\n\tapi.GET(\"sessions/:id\", sessionCtrl.Get)\n\tapi.GET(\"finished-sessions\", sessionCtrl.GetFinished)\n\tapi.GET(\"unfinished-sessions\", sessionCtrl.GetUnfinished)\n\tapi.POST(\"sessions\", sessionCtrl.Add)\n\tapi.PUT(\"sessions\", sessionCtrl.Update)\n\tapi.DELETE(\"sessions/:id\", sessionCtrl.Delete)\n\n\treturn router, nil\n}", "func (h *Handler) SetupRoutes() {\n\tfmt.Println(\"Setting Up Routes\")\n\th.Router = mux.NewRouter()\n\n\th.Router.HandleFunc(\"/api/comment\", h.GetAllComments).Methods(\"GET\")\n\th.Router.HandleFunc(\"/api/comment\", h.PostComment).Methods(\"POST\")\n\th.Router.HandleFunc(\"/api/comment/{id}\", h.GetComment).Methods(\"GET\")\n\th.Router.HandleFunc(\"/api/comment/{id}\", h.UpdateComment).Methods(\"PUT\")\n\th.Router.HandleFunc(\"/api/comment/{id}\", h.DeleteComment).Methods(\"DELETE\")\n\n\th.Router.HandleFunc(\"/api/health\", func(w http.ResponseWriter, r *http.Request) {\n\t\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\t\tw.WriteHeader(http.StatusOK)\n\t\tif err := json.NewEncoder(w).Encode(Response{Message: \"I am alive!\"}); err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t})\n}", "func (e *Engine) setup(addr string) {\n\tsetupExclude()\n\te.setupIndex()\n\te.setupFallback()\n\te.setupFavicon()\n}" ]
[ "0.76839864", "0.74411607", "0.7420936", "0.74127454", "0.73549485", "0.72737867", "0.7140154", "0.70420814", "0.7023133", "0.70221287", "0.7019182", "0.6920558", "0.69160783", "0.6907643", "0.68852675", "0.68766725", "0.686933", "0.68692505", "0.6868554", "0.6865591", "0.6864763", "0.68597734", "0.6834272", "0.6832232", "0.68249184", "0.6820141", "0.68001175", "0.67959124", "0.6792249", "0.6773662", "0.67584264", "0.67539287", "0.67263865", "0.6723421", "0.6714076", "0.6713741", "0.6713309", "0.670418", "0.6696724", "0.6696153", "0.66941994", "0.66896075", "0.6681482", "0.66530496", "0.6649352", "0.66457844", "0.6637925", "0.66328084", "0.662874", "0.6623765", "0.66162074", "0.6615183", "0.6614995", "0.66148055", "0.6596487", "0.65769124", "0.6569819", "0.65632886", "0.6560923", "0.65589607", "0.65492624", "0.6548007", "0.6542883", "0.6541315", "0.6541104", "0.6534191", "0.6518221", "0.651726", "0.65128833", "0.6511057", "0.6503305", "0.6495561", "0.6488461", "0.64884335", "0.64746845", "0.64714575", "0.6469282", "0.64655054", "0.64647996", "0.64570713", "0.6451866", "0.6450451", "0.64475137", "0.6438416", "0.6433212", "0.6429214", "0.64251727", "0.64237815", "0.6410179", "0.6409648", "0.6397505", "0.63965124", "0.6388126", "0.63880265", "0.6377102", "0.6376206", "0.6371858", "0.63718146", "0.63664675", "0.63659984", "0.63655704" ]
0.0
-1
104 / 5 =====> 20 4
func bölme(bölünen, bölen int) (bölüm, kalan int) { bölüm = bölünen / bölen kalan = bölünen % bölen return bölüm, kalan }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func P12() int64 {\n\tvar i, value int64 = 1, 0;\n\tfor {\n\t\tn := i * (i+1) / 2;\n\t\tcount := CountDivisors(n);\n\t\tif(count > 500) {\n\t\t\tvalue = n;\n\t\t\tbreak;\n\t\t}\n\t\ti++;\n\t}\n return value;\n}", "func divide(S, D uint32) uint32 {\n\treturn 65535 * S / D\n}", "func split(sum int) (x,y int) {\n\tx = sum * 4/9\n\ty = sum - x\n\treturn\n}", "func main() {\n\to := 0\n\ti := 1\n\tfor o < 500 {\n\t\ti++\n\t\to = numberOfDivisors(triangleNumber(i))\n\t}\n\tfmt.Printf(\"The first value of triangle number to have over five hundred divisor is %d\\n\", triangleNumber(i))\n}", "func viralAdvertising(n int32) int32 {\n\n\toutput := make([]int32, n)\n\n\toutput[0] = (5 / 2) * 3\n\n\tvar i int32 = 1\n\n\tvar total int32 = 2\n\n\tfor ; i < n; i++ {\n\t\tcurrentTotal := output[i-1] / 2\n\t\toutput[i] = currentTotal * 3\n\t\ttotal += currentTotal\n\t}\n\n\treturn total\n}", "func main() {\n\tfor i := 10; i <= 100; i++ {\n\t\tfmt.Printf(\"When %d is devided by 4, the remainder (modulus) is %d\\n\", i, i%4)\n\t}\n}", "func sumFourDivisors(nums []int) int {\n\n}", "func main() {\n\n\thittable := make(map[int]bool)\n\tans := int64(0)\n\n\tfor i := 1; i*i <= top*9*9; i++ {\n\t\thittable[i*i] = true\n\t}\n\n\tfor i := range hittable {\n\t\tways, _ := enumerate(9, top, i)\n\t\tfor i := 0; i < len(ways); i++ {\n\t\t\tans += process(ways[i])\n\t\t\tans %= mod\n\t\t}\n\t}\n\n\tfmt.Println(\"171/ Find the last nine digits of the sum of all n, 0 < n < 10^20, such that f(n) is a perfect square\")\n\tfmt.Println(ans)\n}", "func main() {\n\tfor i := 1; i < 500; i++ {\n\n\t\tif i%7 == 0 && i%5 == 0 {\n\t\t\tfmt.Println(i)\n\n\t\t}\n\n\t}\n}", "func divmod(dvdn, dvsr int) (q, r int) {\n\tr = dvdn\n\tfor r >= dvsr {\n\t\tq++\n\t\tr = r - dvsr\n\t}\n\treturn\n}", "func split(sum int) (x, y int) {\n\tx = sum * 4 / 9\n\ty = sum - x\n\t//fmt.Println(x)\n\t//fmt.Println(y)\n\treturn\n}", "func d(a int, b int) int {\n\treturn a/b\n}", "func threeorfive (value int) int {\n\n\tvar sums int\n\tfor i := 1; i < value; i++ {\n\t\tif (i%3 ==0 || i%5 ==0) {\n\t\t\tsums = sums + i\n\t\t//\tfmt.Println(i)\n\t\t}\n\t}\n\treturn sums\n}", "func Divide(a, operand int) int { return operand / a }", "func split(sum int) (x, y int) {\n\tx = sum * 4 / 9\n\ty = sum - x\n\treturn\n}", "func split(sum int) (x, y int) {\n\tx = sum * 4 / 9\n\ty = sum - x\n\treturn\n}", "func split(sum int) (x, y int) {\n\tx = sum * 4 / 9\n\ty = sum - x\n\treturn\n}", "func split(sum int) (x, y int) {\n\tx = sum * 4 / 9\n\ty = sum - x\n\treturn\n}", "func split(sum int) (x, y int) {\n\tx = sum * 4 / 9\n\ty = sum - x\n\treturn\n}", "func split(sum int) (x, y int) {\n\tx = sum * 4 / 9\n\ty = sum - x\n\treturn\n}", "func split(sum int) (x, y int) {\n\tx = sum * 4 / 9\n\ty = sum - x\n\treturn\n}", "func split(sum int) (x, y int) {\n\tx = sum * 4 / 9\n\ty = sum - x\n\treturn\n}", "func cdiv(a, b int) int { return (a + b - 1) / b }", "func main() {\n\tfor i := 10; i <= 100; i++ {\n\t\tm := i % 4\n\t\tfmt.Printf(\"When %v is devided by 4 the remaining value is %v\\n\", i, m)\n\t}\n}", "func main() {\n\tN := scanInt()\n\tn, e, t := 1, 1, 1\n\tfor i := 0; i < N; i++ {\n\t\t// [+/*]you can divide by mod each add/time ops\n\t\te = (e * 8) % mod\n\t\tn = (n * 9) % mod\n\t\tt = (t * 10) % mod\n\t}\n\tans := t - n - n + e\n\t// [+]consider if above time calc result over mod\n\tans %= mod\n\t// [-]consider if above sub calc result goes negative number\n\tans = (ans + mod) % mod\n\tfmt.Println(ans)\n}", "func hash5(u uint64, h uint8) uint32 {\n\treturn uint32(((u << (64 - 40)) * prime5bytes) >> ((64 - h) & 63))\n}", "func split(sum int) (x, y int) {\n\tx = sum * 4 / 9\n\ty = sum - x\n\t//naked return statement\n\treturn\n}", "func iterationFromCounter(roundCounter int32) int32 {\n\treturn roundCounter / 4\n}", "func DivMod(dvdn, dvsr int) (q, r int) {\n\tr = dvdn\n\tfor r >= dvsr {\n\t\tq += 1\n\t\tr = r - dvsr\n\t}\n\treturn\n}", "func main() {\n\ttotal := int64(0)\n\tvectors := ways(10, top, 3)\n\tfor _, vector := range vectors {\n\t\ttotal += distribute(vector)\n\t}\n\n\tfmt.Println(\"172/ How many 18-digit numbers n (without leading zeros) are there such that no digit occurs more than three times in n?\")\n\tfmt.Println(total)\n}", "func main() {\n\t// everything here is self-explanatory\n\ta := 12\n\tb := 60\n\t\n\tfmt.Println(a + b)\n\tfmt.Println(a - b)\n\tfmt.Println(a / b)\n\tfmt.Println(a * b)\n\tfmt.Println(a % b)\n}", "func main() {\n\tfor i := 10; i <= 100; i++ {\n\n\t\tfmt.Println(i % 4)\n\n\t}\n}", "func main() {\n\t// This expression should print 20\n\tfmt.Println(10 + 5 - (5 - 10))\n\n\t// This expression should print -16\n\tfmt.Println(-10 + 0.5 - (1 + 5.5))\n\n\t// This expression should print -25\n\tfmt.Println(5 + 10*(2-5))\n\n\t// This expression should print 0.5\n\tfmt.Println(0.5 * (2 - 1))\n\n\t// This expression should print 24\n\tfmt.Println((3+1)/2*10 + 4)\n\n\t// This expression should print 15\n\tfmt.Println(10 / 2 * (10 % 7))\n\n\t// This expression should print 40\n\tfmt.Println(100 / (5.0 / 2))\n}", "func Divide() {\n\tMatch('/')\n\tFactor()\n\tEmitLn(\"MOVE (SP)+,D1\")\n\tEmitLn(\"DIVS D1,D0\")\n}", "func threeOrFive(x int) int {\n\t// sum is initialized to 0 using short declaration\n\tsum := 0\n\n\tfor i := 0; i < x; i++ {\n\t\tif i % 3 == 0 || i % 5 == 0 {\n\t\t\tsum += i\n\t\t}\n\t}\n\t// At the end return sum\n\treturn sum\n}", "func Div(x, y int) int {\n\treturn x / y\n}", "func getSegmentNumber(i int) int {\n\tsegRow := int(i / 27)\n\tsegCol := int(i % 9 / 3)\n\treturn segRow * 3 + segCol\n}", "func nums(v int) int {\n\tres := 0\n\tfor i := 0; i < v; i++ {\n\t\tif i%3 == 0 || i%5 == 0 {\n\t\t\tres += i\n\t\t}\n\t}\n\treturn res\n}", "func div(a, b int32) int32 {\n\tif a >= 0 {\n\t\treturn (a + (b >> 1)) / b\n\t}\n\treturn -((-a + (b >> 1)) / b)\n}", "func DIVPD(mx, x operand.Op) { ctx.DIVPD(mx, x) }", "func core(out *[64]byte, in *[16]byte, k *[32]byte) {\n\tj0 := uint32(0x61707865)\n\tj1 := uint32(0x3320646e)\n\tj2 := uint32(0x79622d32)\n\tj3 := uint32(0x6b206574)\n\tj4 := binary.LittleEndian.Uint32(k[0:4])\n\tj5 := binary.LittleEndian.Uint32(k[4:8])\n\tj6 := binary.LittleEndian.Uint32(k[8:12])\n\tj7 := binary.LittleEndian.Uint32(k[12:16])\n\tj8 := binary.LittleEndian.Uint32(k[16:20])\n\tj9 := binary.LittleEndian.Uint32(k[20:24])\n\tj10 := binary.LittleEndian.Uint32(k[24:28])\n\tj11 := binary.LittleEndian.Uint32(k[28:32])\n\tj12 := binary.LittleEndian.Uint32(in[0:4])\n\tj13 := binary.LittleEndian.Uint32(in[4:8])\n\tj14 := binary.LittleEndian.Uint32(in[8:12])\n\tj15 := binary.LittleEndian.Uint32(in[12:16])\n\n\tx0, x1, x2, x3, x4, x5, x6, x7 := j0, j1, j2, j3, j4, j5, j6, j7\n\tx8, x9, x10, x11, x12, x13, x14, x15 := j8, j9, j10, j11, j12, j13, j14, j15\n\n\tfor i := 0; i < rounds; i += 2 {\n\t\tx0 += x4\n\t\tx12 ^= x0\n\t\tx12 = (x12 << 16) | (x12 >> (16))\n\t\tx8 += x12\n\t\tx4 ^= x8\n\t\tx4 = (x4 << 12) | (x4 >> (20))\n\t\tx0 += x4\n\t\tx12 ^= x0\n\t\tx12 = (x12 << 8) | (x12 >> (24))\n\t\tx8 += x12\n\t\tx4 ^= x8\n\t\tx4 = (x4 << 7) | (x4 >> (25))\n\t\tx1 += x5\n\t\tx13 ^= x1\n\t\tx13 = (x13 << 16) | (x13 >> 16)\n\t\tx9 += x13\n\t\tx5 ^= x9\n\t\tx5 = (x5 << 12) | (x5 >> 20)\n\t\tx1 += x5\n\t\tx13 ^= x1\n\t\tx13 = (x13 << 8) | (x13 >> 24)\n\t\tx9 += x13\n\t\tx5 ^= x9\n\t\tx5 = (x5 << 7) | (x5 >> 25)\n\t\tx2 += x6\n\t\tx14 ^= x2\n\t\tx14 = (x14 << 16) | (x14 >> 16)\n\t\tx10 += x14\n\t\tx6 ^= x10\n\t\tx6 = (x6 << 12) | (x6 >> 20)\n\t\tx2 += x6\n\t\tx14 ^= x2\n\t\tx14 = (x14 << 8) | (x14 >> 24)\n\t\tx10 += x14\n\t\tx6 ^= x10\n\t\tx6 = (x6 << 7) | (x6 >> 25)\n\t\tx3 += x7\n\t\tx15 ^= x3\n\t\tx15 = (x15 << 16) | (x15 >> 16)\n\t\tx11 += x15\n\t\tx7 ^= x11\n\t\tx7 = (x7 << 12) | (x7 >> 20)\n\t\tx3 += x7\n\t\tx15 ^= x3\n\t\tx15 = (x15 << 8) | (x15 >> 24)\n\t\tx11 += x15\n\t\tx7 ^= x11\n\t\tx7 = (x7 << 7) | (x7 >> 25)\n\t\tx0 += x5\n\t\tx15 ^= x0\n\t\tx15 = (x15 << 16) | (x15 >> 16)\n\t\tx10 += x15\n\t\tx5 ^= x10\n\t\tx5 = (x5 << 12) | (x5 >> 20)\n\t\tx0 += x5\n\t\tx15 ^= x0\n\t\tx15 = (x15 << 8) | (x15 >> 24)\n\t\tx10 += x15\n\t\tx5 ^= x10\n\t\tx5 = (x5 << 7) | (x5 >> 25)\n\t\tx1 += x6\n\t\tx12 ^= x1\n\t\tx12 = (x12 << 16) | (x12 >> 16)\n\t\tx11 += x12\n\t\tx6 ^= x11\n\t\tx6 = (x6 << 12) | (x6 >> 20)\n\t\tx1 += x6\n\t\tx12 ^= x1\n\t\tx12 = (x12 << 8) | (x12 >> 24)\n\t\tx11 += x12\n\t\tx6 ^= x11\n\t\tx6 = (x6 << 7) | (x6 >> 25)\n\t\tx2 += x7\n\t\tx13 ^= x2\n\t\tx13 = (x13 << 16) | (x13 >> 16)\n\t\tx8 += x13\n\t\tx7 ^= x8\n\t\tx7 = (x7 << 12) | (x7 >> 20)\n\t\tx2 += x7\n\t\tx13 ^= x2\n\t\tx13 = (x13 << 8) | (x13 >> 24)\n\t\tx8 += x13\n\t\tx7 ^= x8\n\t\tx7 = (x7 << 7) | (x7 >> 25)\n\t\tx3 += x4\n\t\tx14 ^= x3\n\t\tx14 = (x14 << 16) | (x14 >> 16)\n\t\tx9 += x14\n\t\tx4 ^= x9\n\t\tx4 = (x4 << 12) | (x4 >> 20)\n\t\tx3 += x4\n\t\tx14 ^= x3\n\t\tx14 = (x14 << 8) | (x14 >> 24)\n\t\tx9 += x14\n\t\tx4 ^= x9\n\t\tx4 = (x4 << 7) | (x4 >> 25)\n\t}\n\n\tx0 += j0\n\tx1 += j1\n\tx2 += j2\n\tx3 += j3\n\tx4 += j4\n\tx5 += j5\n\tx6 += j6\n\tx7 += j7\n\tx8 += j8\n\tx9 += j9\n\tx10 += j10\n\tx11 += j11\n\tx12 += j12\n\tx13 += j13\n\tx14 += j14\n\tx15 += j15\n\n\tbinary.LittleEndian.PutUint32(out[0:4], x0)\n\tbinary.LittleEndian.PutUint32(out[4:8], x1)\n\tbinary.LittleEndian.PutUint32(out[8:12], x2)\n\tbinary.LittleEndian.PutUint32(out[12:16], x3)\n\tbinary.LittleEndian.PutUint32(out[16:20], x4)\n\tbinary.LittleEndian.PutUint32(out[20:24], x5)\n\tbinary.LittleEndian.PutUint32(out[24:28], x6)\n\tbinary.LittleEndian.PutUint32(out[28:32], x7)\n\tbinary.LittleEndian.PutUint32(out[32:36], x8)\n\tbinary.LittleEndian.PutUint32(out[36:40], x9)\n\tbinary.LittleEndian.PutUint32(out[40:44], x10)\n\tbinary.LittleEndian.PutUint32(out[44:48], x11)\n\tbinary.LittleEndian.PutUint32(out[48:52], x12)\n\tbinary.LittleEndian.PutUint32(out[52:56], x13)\n\tbinary.LittleEndian.PutUint32(out[56:60], x14)\n\tbinary.LittleEndian.PutUint32(out[60:64], x15)\n}", "func DoTheThing() {\n\tbaseCode := []int{3, 62, 1001, 62, 11, 10, 109, 2229, 105, 1, 0, 1856, 918, 1151, 1392, 2070, 1460, 2099, 660, 1714, 1217, 691, 1584, 1755, 1489, 600, 2198, 571, 1617, 2165, 885, 1897, 1679, 1054, 1524, 1023, 1250, 1827, 821, 852, 1427, 1648, 990, 629, 1285, 1967, 955, 722, 2134, 1553, 1316, 1790, 1361, 2037, 1085, 1120, 1928, 1184, 786, 753, 1996, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 3, 64, 1008, 64, -1, 62, 1006, 62, 88, 1006, 61, 170, 1105, 1, 73, 3, 65, 20101, 0, 64, 1, 20102, 1, 66, 2, 21101, 0, 105, 0, 1105, 1, 436, 1201, 1, -1, 64, 1007, 64, 0, 62, 1005, 62, 73, 7, 64, 67, 62, 1006, 62, 73, 1002, 64, 2, 133, 1, 133, 68, 133, 102, 1, 0, 62, 1001, 133, 1, 140, 8, 0, 65, 63, 2, 63, 62, 62, 1005, 62, 73, 1002, 64, 2, 161, 1, 161, 68, 161, 1101, 1, 0, 0, 1001, 161, 1, 169, 102, 1, 65, 0, 1102, 1, 1, 61, 1101, 0, 0, 63, 7, 63, 67, 62, 1006, 62, 203, 1002, 63, 2, 194, 1, 68, 194, 194, 1006, 0, 73, 1001, 63, 1, 63, 1106, 0, 178, 21101, 210, 0, 0, 105, 1, 69, 2101, 0, 1, 70, 1102, 1, 0, 63, 7, 63, 71, 62, 1006, 62, 250, 1002, 63, 2, 234, 1, 72, 234, 234, 4, 0, 101, 1, 234, 240, 4, 0, 4, 70, 1001, 63, 1, 63, 1106, 0, 218, 1105, 1, 73, 109, 4, 21101, 0, 0, -3, 21101, 0, 0, -2, 20207, -2, 67, -1, 1206, -1, 293, 1202, -2, 2, 283, 101, 1, 283, 283, 1, 68, 283, 283, 22001, 0, -3, -3, 21201, -2, 1, -2, 1105, 1, 263, 22102, 1, -3, -3, 109, -4, 2106, 0, 0, 109, 4, 21102, 1, 1, -3, 21102, 0, 1, -2, 20207, -2, 67, -1, 1206, -1, 342, 1202, -2, 2, 332, 101, 1, 332, 332, 1, 68, 332, 332, 22002, 0, -3, -3, 21201, -2, 1, -2, 1106, 0, 312, 21201, -3, 0, -3, 109, -4, 2106, 0, 0, 109, 1, 101, 1, 68, 358, 21001, 0, 0, 1, 101, 3, 68, 366, 21002, 0, 1, 2, 21101, 0, 376, 0, 1106, 0, 436, 22102, 1, 1, 0, 109, -1, 2106, 0, 0, 1, 2, 4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288, 1048576, 2097152, 4194304, 8388608, 16777216, 33554432, 67108864, 134217728, 268435456, 536870912, 1073741824, 2147483648, 4294967296, 8589934592, 17179869184, 34359738368, 68719476736, 137438953472, 274877906944, 549755813888, 1099511627776, 2199023255552, 4398046511104, 8796093022208, 17592186044416, 35184372088832, 70368744177664, 140737488355328, 281474976710656, 562949953421312, 1125899906842624, 109, 8, 21202, -6, 10, -5, 22207, -7, -5, -5, 1205, -5, 521, 21102, 1, 0, -4, 21102, 0, 1, -3, 21101, 51, 0, -2, 21201, -2, -1, -2, 1201, -2, 385, 470, 21001, 0, 0, -1, 21202, -3, 2, -3, 22207, -7, -1, -5, 1205, -5, 496, 21201, -3, 1, -3, 22102, -1, -1, -5, 22201, -7, -5, -7, 22207, -3, -6, -5, 1205, -5, 515, 22102, -1, -6, -5, 22201, -3, -5, -3, 22201, -1, -4, -4, 1205, -2, 461, 1105, 1, 547, 21101, 0, -1, -4, 21202, -6, -1, -6, 21207, -7, 0, -5, 1205, -5, 547, 22201, -7, -6, -7, 21201, -4, 1, -4, 1106, 0, 529, 21202, -4, 1, -7, 109, -8, 2106, 0, 0, 109, 1, 101, 1, 68, 564, 20101, 0, 0, 0, 109, -1, 2106, 0, 0, 1101, 39581, 0, 66, 1102, 1, 1, 67, 1101, 598, 0, 68, 1101, 0, 556, 69, 1101, 0, 0, 71, 1102, 600, 1, 72, 1105, 1, 73, 1, 1145, 1101, 83477, 0, 66, 1102, 1, 1, 67, 1101, 0, 627, 68, 1101, 0, 556, 69, 1102, 0, 1, 71, 1101, 629, 0, 72, 1106, 0, 73, 1, 1537, 1102, 23167, 1, 66, 1101, 0, 1, 67, 1101, 656, 0, 68, 1101, 0, 556, 69, 1102, 1, 1, 71, 1101, 0, 658, 72, 1105, 1, 73, 1, -78, 13, 778, 1101, 46757, 0, 66, 1101, 1, 0, 67, 1102, 1, 687, 68, 1101, 0, 556, 69, 1101, 0, 1, 71, 1102, 1, 689, 72, 1106, 0, 73, 1, 125, 1, 37767, 1101, 4591, 0, 66, 1101, 0, 1, 67, 1101, 718, 0, 68, 1101, 556, 0, 69, 1102, 1, 1, 71, 1102, 720, 1, 72, 1106, 0, 73, 1, 107, 49, 200722, 1102, 1, 2579, 66, 1102, 1, 1, 67, 1102, 749, 1, 68, 1102, 1, 556, 69, 1101, 1, 0, 71, 1102, 1, 751, 72, 1106, 0, 73, 1, 1861, 47, 80326, 1102, 1, 55057, 66, 1102, 2, 1, 67, 1101, 0, 780, 68, 1101, 0, 302, 69, 1101, 1, 0, 71, 1101, 0, 784, 72, 1106, 0, 73, 0, 0, 0, 0, 31, 68881, 1101, 40163, 0, 66, 1102, 1, 3, 67, 1101, 0, 813, 68, 1101, 302, 0, 69, 1102, 1, 1, 71, 1102, 819, 1, 72, 1105, 1, 73, 0, 0, 0, 0, 0, 0, 35, 12577, 1102, 1, 10753, 66, 1102, 1, 1, 67, 1101, 0, 848, 68, 1102, 556, 1, 69, 1101, 0, 1, 71, 1101, 850, 0, 72, 1106, 0, 73, 1, 13, 49, 301083, 1101, 0, 69767, 66, 1102, 1, 1, 67, 1101, 0, 879, 68, 1102, 556, 1, 69, 1101, 2, 0, 71, 1102, 881, 1, 72, 1106, 0, 73, 1, 10, 1, 50356, 8, 4146, 1102, 3391, 1, 66, 1102, 2, 1, 67, 1101, 912, 0, 68, 1102, 351, 1, 69, 1101, 1, 0, 71, 1101, 916, 0, 72, 1105, 1, 73, 0, 0, 0, 0, 255, 51059, 1101, 12589, 0, 66, 1102, 4, 1, 67, 1101, 945, 0, 68, 1101, 0, 302, 69, 1102, 1, 1, 71, 1101, 953, 0, 72, 1106, 0, 73, 0, 0, 0, 0, 0, 0, 0, 0, 8, 3455, 1101, 0, 12577, 66, 1102, 1, 3, 67, 1101, 0, 982, 68, 1101, 253, 0, 69, 1102, 1, 1, 71, 1101, 0, 988, 72, 1105, 1, 73, 0, 0, 0, 0, 0, 0, 18, 96181, 1102, 1, 68881, 66, 1102, 2, 1, 67, 1102, 1, 1017, 68, 1101, 0, 302, 69, 1101, 0, 1, 71, 1101, 1021, 0, 72, 1106, 0, 73, 0, 0, 0, 0, 2, 9473, 1101, 10957, 0, 66, 1101, 0, 1, 67, 1101, 1050, 0, 68, 1101, 556, 0, 69, 1101, 0, 1, 71, 1101, 0, 1052, 72, 1106, 0, 73, 1, -87195, 6, 188913, 1101, 0, 48187, 66, 1101, 1, 0, 67, 1102, 1, 1081, 68, 1101, 556, 0, 69, 1101, 0, 1, 71, 1102, 1083, 1, 72, 1105, 1, 73, 1, 4, 49, 100361, 1102, 81563, 1, 66, 1101, 0, 3, 67, 1101, 1112, 0, 68, 1102, 302, 1, 69, 1101, 0, 1, 71, 1101, 1118, 0, 72, 1105, 1, 73, 0, 0, 0, 0, 0, 0, 35, 37731, 1102, 72053, 1, 66, 1102, 1, 1, 67, 1101, 0, 1147, 68, 1102, 1, 556, 69, 1101, 1, 0, 71, 1101, 0, 1149, 72, 1105, 1, 73, 1, -686, 47, 120489, 1102, 1, 9473, 66, 1101, 0, 2, 67, 1101, 0, 1178, 68, 1101, 0, 302, 69, 1102, 1, 1, 71, 1102, 1182, 1, 72, 1106, 0, 73, 0, 0, 0, 0, 11, 74357, 1101, 0, 19753, 66, 1102, 2, 1, 67, 1102, 1, 1211, 68, 1102, 302, 1, 69, 1102, 1, 1, 71, 1101, 0, 1215, 72, 1106, 0, 73, 0, 0, 0, 0, 40, 9754, 1101, 16979, 0, 66, 1102, 1, 1, 67, 1101, 1244, 0, 68, 1101, 556, 0, 69, 1101, 2, 0, 71, 1102, 1, 1246, 72, 1106, 0, 73, 1, 641, 49, 602166, 13, 389, 1101, 1291, 0, 66, 1102, 3, 1, 67, 1101, 1277, 0, 68, 1101, 0, 302, 69, 1102, 1, 1, 71, 1102, 1, 1283, 72, 1105, 1, 73, 0, 0, 0, 0, 0, 0, 40, 14631, 1101, 90073, 0, 66, 1102, 1, 1, 67, 1101, 1312, 0, 68, 1102, 556, 1, 69, 1102, 1, 1, 71, 1102, 1, 1314, 72, 1105, 1, 73, 1, 52895, 6, 125942, 1101, 22769, 0, 66, 1102, 1, 1, 67, 1102, 1, 1343, 68, 1102, 556, 1, 69, 1102, 1, 8, 71, 1102, 1345, 1, 72, 1105, 1, 73, 1, 1, 48, 55057, 31, 137762, 2, 18946, 11, 148714, 29, 97978, 43, 244689, 47, 40163, 13, 1167, 1101, 0, 1777, 66, 1101, 1, 0, 67, 1102, 1, 1388, 68, 1102, 556, 1, 69, 1102, 1, 1, 71, 1102, 1390, 1, 72, 1105, 1, 73, 1, 51973, 48, 110114, 1101, 28069, 0, 66, 1102, 3, 1, 67, 1102, 1, 1419, 68, 1102, 1, 302, 69, 1101, 0, 1, 71, 1101, 0, 1425, 72, 1105, 1, 73, 0, 0, 0, 0, 0, 0, 46, 39506, 1102, 48989, 1, 66, 1102, 2, 1, 67, 1101, 0, 1454, 68, 1102, 1, 302, 69, 1102, 1, 1, 71, 1102, 1458, 1, 72, 1106, 0, 73, 0, 0, 0, 0, 43, 81563, 1101, 0, 46861, 66, 1102, 1, 1, 67, 1101, 0, 1487, 68, 1101, 0, 556, 69, 1101, 0, 0, 71, 1102, 1, 1489, 72, 1105, 1, 73, 1, 1696, 1102, 1, 389, 66, 1101, 3, 0, 67, 1101, 1516, 0, 68, 1101, 0, 302, 69, 1101, 0, 1, 71, 1102, 1, 1522, 72, 1105, 1, 73, 0, 0, 0, 0, 0, 0, 21, 11093, 1102, 1, 15887, 66, 1101, 0, 1, 67, 1101, 1551, 0, 68, 1102, 556, 1, 69, 1102, 0, 1, 71, 1101, 0, 1553, 72, 1106, 0, 73, 1, 1653, 1102, 25693, 1, 66, 1102, 1, 1, 67, 1102, 1580, 1, 68, 1102, 1, 556, 69, 1101, 0, 1, 71, 1102, 1, 1582, 72, 1106, 0, 73, 1, 51, 43, 163126, 1101, 0, 74357, 66, 1102, 2, 1, 67, 1102, 1611, 1, 68, 1101, 0, 302, 69, 1101, 0, 1, 71, 1101, 1615, 0, 72, 1106, 0, 73, 0, 0, 0, 0, 49, 501805, 1102, 4597, 1, 66, 1101, 0, 1, 67, 1101, 1644, 0, 68, 1102, 556, 1, 69, 1101, 0, 1, 71, 1101, 0, 1646, 72, 1105, 1, 73, 1, 319931, 6, 62971, 1102, 1, 62467, 66, 1102, 1, 1, 67, 1102, 1675, 1, 68, 1102, 1, 556, 69, 1101, 0, 1, 71, 1101, 1677, 0, 72, 1105, 1, 73, 1, 41, 3, 28069, 1101, 11093, 0, 66, 1102, 3, 1, 67, 1102, 1706, 1, 68, 1102, 1, 302, 69, 1102, 1, 1, 71, 1102, 1, 1712, 72, 1106, 0, 73, 0, 0, 0, 0, 0, 0, 40, 4877, 1102, 1, 691, 66, 1102, 1, 6, 67, 1101, 0, 1741, 68, 1102, 302, 1, 69, 1102, 1, 1, 71, 1101, 1753, 0, 72, 1105, 1, 73, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 19, 6782, 1102, 64567, 1, 66, 1101, 1, 0, 67, 1102, 1782, 1, 68, 1102, 1, 556, 69, 1101, 3, 0, 71, 1101, 0, 1784, 72, 1105, 1, 73, 1, 5, 1, 12589, 1, 25178, 8, 2764, 1101, 0, 4877, 66, 1102, 4, 1, 67, 1102, 1817, 1, 68, 1101, 0, 253, 69, 1102, 1, 1, 71, 1102, 1825, 1, 72, 1106, 0, 73, 0, 0, 0, 0, 0, 0, 0, 0, 19, 3391, 1101, 0, 29863, 66, 1102, 1, 1, 67, 1101, 0, 1854, 68, 1102, 556, 1, 69, 1102, 0, 1, 71, 1101, 0, 1856, 72, 1106, 0, 73, 1, 1846, 1101, 0, 51059, 66, 1101, 0, 1, 67, 1102, 1883, 1, 68, 1101, 556, 0, 69, 1102, 1, 6, 71, 1102, 1, 1885, 72, 1106, 0, 73, 1, 25793, 46, 19753, 21, 22186, 21, 33279, 25, 1291, 25, 2582, 25, 3873, 1101, 0, 75781, 66, 1101, 0, 1, 67, 1102, 1, 1924, 68, 1102, 1, 556, 69, 1101, 1, 0, 71, 1102, 1926, 1, 72, 1105, 1, 73, 1, 160, 8, 1382, 1101, 99469, 0, 66, 1102, 1, 1, 67, 1101, 0, 1955, 68, 1101, 556, 0, 69, 1101, 5, 0, 71, 1101, 0, 1957, 72, 1106, 0, 73, 1, 2, 49, 401444, 18, 192362, 3, 84207, 8, 691, 8, 2073, 1101, 0, 8761, 66, 1102, 1, 1, 67, 1101, 1994, 0, 68, 1102, 556, 1, 69, 1101, 0, 0, 71, 1101, 1996, 0, 72, 1105, 1, 73, 1, 1369, 1102, 100361, 1, 66, 1102, 1, 6, 67, 1102, 1, 2023, 68, 1101, 0, 302, 69, 1102, 1, 1, 71, 1101, 0, 2035, 72, 1106, 0, 73, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 40, 19508, 1102, 37967, 1, 66, 1101, 2, 0, 67, 1102, 1, 2064, 68, 1102, 1, 302, 69, 1102, 1, 1, 71, 1102, 2068, 1, 72, 1105, 1, 73, 0, 0, 0, 0, 35, 25154, 1101, 18367, 0, 66, 1101, 1, 0, 67, 1102, 2097, 1, 68, 1101, 0, 556, 69, 1102, 0, 1, 71, 1101, 2099, 0, 72, 1106, 0, 73, 1, 1503, 1101, 62971, 0, 66, 1102, 3, 1, 67, 1101, 2126, 0, 68, 1102, 1, 253, 69, 1101, 0, 1, 71, 1102, 1, 2132, 72, 1105, 1, 73, 0, 0, 0, 0, 0, 0, 42, 37967, 1102, 1, 37691, 66, 1101, 1, 0, 67, 1101, 2161, 0, 68, 1101, 0, 556, 69, 1102, 1, 1, 71, 1101, 0, 2163, 72, 1105, 1, 73, 1, 34, 42, 75934, 1102, 96181, 1, 66, 1102, 2, 1, 67, 1101, 2192, 0, 68, 1101, 0, 302, 69, 1101, 0, 1, 71, 1101, 0, 2196, 72, 1106, 0, 73, 0, 0, 0, 0, 3, 56138, 1102, 1, 44777, 66, 1101, 0, 1, 67, 1101, 2225, 0, 68, 1101, 556, 0, 69, 1102, 1, 1, 71, 1101, 2227, 0, 72, 1105, 1, 73, 1, 17929, 29, 48989}\n\n\tntwk := network{}\n\tntwk.init(50)\n\tfound := make(chan int)\n\tfor i := 0; i < 50; i++ {\n\t\tgo func(i int, n *network) {\n\t\t\tcode := make([]int, len(baseCode))\n\t\t\tcopy(code, baseCode)\n\t\t\tinput := make(chan int, 2000)\n\t\t\toutput := make(chan int, 3000)\n\t\t\tp := intcode.Program{Code: code, Input: input, Output: output, Name: strconv.Itoa(i)}\n\n\t\t\tgo p.Run()\n\t\t\tinput <- i\n\t\t\tfor {\n\t\t\tRead:\n\t\t\t\tfor {\n\t\t\t\t\tselect {\n\t\t\t\t\tcase dest := <-output:\n\t\t\t\t\t\tx := <-output\n\t\t\t\t\t\ty := <-output\n\t\t\t\t\t\tn.send(packet{address: dest, x: x, y: y}, i)\n\t\t\t\t\tdefault:\n\t\t\t\t\t\ttime.Sleep(10 * time.Millisecond)\n\t\t\t\t\t\tbreak Read\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\t//receive input\n\t\t\t\tpacket, exists := n.receive(i)\n\t\t\t\tif exists {\n\t\t\t\t\tinput <- packet.x\n\t\t\t\t\tinput <- packet.y\n\t\t\t\t} else {\n\t\t\t\t\tinput <- -1\n\t\t\t\t}\n\t\t\t}\n\t\t}(i, &ntwk)\n\t}\n\t<-found\n}", "func main() {\n\tfmt.Println(safeDiv(3,0))\n\tfmt.Println(safeDiv(3,2))\n}", "func normalize(vol int) int {\n\treturn (vol + 5) / 10 * 10\n}", "func divmod(a, b int) (q int, r int) {\n\tif b == 0 {\n\t\treturn\n\t}\n\tq = a / b\n\tr = a % b\n\treturn\n}", "func ProblemFive(limit int) int {\n\tvar lcm = 1\n\tfor i := 2; i <= limit; i++ {\n\t\tlcd := eulerutil.GCD(lcm, i)\n\t\tlcm = lcm * i / lcd\n\t}\n\n\treturn lcm\n}", "func (z *Int) Div(x, y *Int) *Int {}", "func cuttingRope2(n int) int {\n\tif n < 4 {\n\t\treturn n - 1\n\t}\n\tres := 1\n\tfor n > 4 {\n\t\tres = res * 3 % 1000000007\n\t\tn -= 3\n\t}\n\treturn res * n % 1000000007\n}", "func oo135(ratings []int) int {\n\tsum := 0\n\tleft, mid, right := 0, 0, 0\n\tfor {\n\t\tfor mid = left; mid+1 < len(ratings) && ratings[mid+1] > ratings[mid]; mid += 1 {\n\t\t}\n\t\tfor right = mid; right+1 < len(ratings) && ratings[right] > ratings[right+1]; right += 1 {\n\t\t}\n\t\tll, rr := mid-left, right-mid\n\t\tsum += (ll) * (ll + 1) / 2\n\t\tsum += (rr) * (rr + 1) / 2\n\t\tif ll > rr {\n\t\t\tsum += ll\n\t\t} else {\n\t\t\tsum += rr\n\t\t}\n\t\tsum += 1\n\t\t// change left\n\t\tif right == len(ratings)-1 {\n\t\t\tbreak\n\t\t}\n\t\tif ratings[right] == ratings[right+1] {\n\t\t\tleft = right + 1\n\t\t} else {\n\t\t\tleft = right\n\t\t\tsum -= 1\n\t\t}\n\t}\n\treturn sum\n}", "func div(u uint8) (uint8, uint8) {\n\treturn u / 64, u % 64\n}", "func divmod(x, m uint64) (quo, rem uint64) {\n\tquo = x / m\n\trem = x % m\n\treturn\n}", "func main() {\n\tvalues := []string{\"ABC\", \"ACB\", \"BAC\", \"BCA\", \"CAB\", \"CBA\"}\n\t// values := []string{\"to\", \"to\", \"top\", \"ton\", \"tom\"}\n\tfactor := []int{100, 10, 1}\n\n\t// 65x100 + 66x10 + 67x1 = 7227\n\thashKey := 0\n\tfor v := range values {\n\t\tbytes := []byte(values[v])\n\t\tf := 0\n\t\thashKey = 0\n\t\tfor i := range bytes {\n\t\t\tfmt.Print(bytes[i], \" \")\n\t\t\thashKey += int(bytes[i]) * factor[f]\n\t\t\tf++\n\t\t}\n\t\tfmt.Printf(\" (hashKey: %d) \\n\", hashKey)\n\t}\n}", "func main() {\n\tx := []int{42, 43, 44, 45, 46, 47, 48, 49, 50, 51}\n\n\tfirst_sl := x[0:5]\n\tsecond_sl := x[5:]\n\tthird_sl := x[2:7]\n\tfourth_sl := x[1:6]\n\n\tfmt.Println(first_sl)\n\tfmt.Println(second_sl)\n\tfmt.Println(third_sl)\n\tfmt.Println(fourth_sl)\n\n}", "func ppBase(stars float64) float64 {\n\treturn pow(5.0*math.Max(1.0, stars/0.0675)-4.0, 3.0) /\n\t\t100000.0\n}", "func part1() {\n\ts := \"abc\"\n\tcount := 0\n\ti := 0\n\th := md5.New()\n\tbuffer := bytes.NewBuffer(make([]byte, 64))\n\n\tfor count < 8 {\n\t\tbuffer.Reset()\n\t\tfmt.Fprintf(buffer, \"%s%d\", s, i)\n\n\t\th.Reset()\n\t\th.Write(buffer.Bytes())\n\t\tdigest := h.Sum(nil)\n\t\tif digest[0] == 0 && digest[1] == 0 && digest[2] < 16 {\n\t\t\tfmt.Printf(\"%x\", digest[2])\n\t\t\tcount++\n\t\t}\n\t\ti++\n\t}\n\tfmt.Println()\n}", "func DIVW(mr operand.Op) { ctx.DIVW(mr) }", "func main() {\n\t// Create a variable to store a running total\n\ttotal := 0\n\n\t// Iterate from 0 to 999\n\tfor index := 0; index < 1000; index++ {\n\t\t// Check whether the current index is divisible by 3 or 5\n\t\tif index%3 == 0 || index%5 == 0 {\n\t\t\t// Add the current index to the total\n\t\t\ttotal += index\n\t\t}\n\t}\n\n\t// Print the total to the console\n\tfmt.Println(total)\n}", "func ex3() {\n\tx := []int{42, 43, 44, 45, 46, 47, 48, 49, 50, 51}\n\tfmt.Println(x[:5])\n\tfmt.Println(x[5:])\n\tfmt.Println(x[2:7])\n\tfmt.Println(x[1:6])\n}", "func minOperations(n int) int {\n\tif n%2 == 0 {\n\t\treturn (n >> 1) * (n >> 1)\n\t} else {\n\t\treturn (n / 2) * (n/2 + 1)\n\t}\n}", "func main() {\n\tfmt.Println(numTreesDp(19))\n}", "func P011() int {\n\tvec := newVector()\n\tmax, sum := 0, 0\n\n\tfor j := 0; j < 20; j++ {\n\t\tfor i := 0; i < 20; i++ {\n\t\t\t//fmt.Println(vec[i][j])\n\t\t\tsum = 0\n\t\t\t/*\n\t\t\t * UP\n\t\t\t */\n\t\t\tif j > 2 {\n\t\t\t\tpsum := 1\n\t\t\t\tfor k := 0; k < 4; k++ {\n\t\t\t\t\tpsum = psum * vec[i][j-k]\n\t\t\t\t}\n\t\t\t\tif psum > sum {\n\t\t\t\t\tsum = psum\n\t\t\t\t}\n\t\t\t}\n\t\t\t/*\n\t\t\t * Down\n\t\t\t */\n\t\t\tif j < 17 {\n\t\t\t\tpsum := 1\n\t\t\t\tfor k := 0; k < 4; k++ {\n\t\t\t\t\tpsum = psum * vec[i][j+k]\n\t\t\t\t}\n\t\t\t\tif psum > sum {\n\t\t\t\t\tsum = psum\n\t\t\t\t}\n\t\t\t}\n\t\t\t/*\n\t\t\t * LEFT\n\t\t\t */\n\t\t\tif i > 2 {\n\t\t\t\tpsum := 1\n\t\t\t\tfor k := 0; k < 4; k++ {\n\t\t\t\t\tpsum = psum * vec[i-k][j]\n\t\t\t\t}\n\t\t\t\tif psum > sum {\n\t\t\t\t\tsum = psum\n\t\t\t\t}\n\t\t\t}\n\t\t\t/*\n\t\t\t * RIGHT\n\t\t\t */\n\t\t\tif i < 17 {\n\t\t\t\tpsum := 1\n\t\t\t\tfor k := 0; k < 4; k++ {\n\t\t\t\t\tpsum = psum * vec[i+k][j]\n\t\t\t\t}\n\t\t\t\tif psum > sum {\n\t\t\t\t\tsum = psum\n\t\t\t\t}\n\t\t\t}\n\t\t\t/*\n\t\t\t * UP LEFT\n\t\t\t */\n\t\t\tif j > 2 && i > 2 {\n\t\t\t\tpsum := 1\n\t\t\t\tfor k := 0; k < 4; k++ {\n\t\t\t\t\tpsum = psum * vec[i-k][j-k]\n\t\t\t\t}\n\t\t\t\tif psum > sum {\n\t\t\t\t\tsum = psum\n\t\t\t\t}\n\t\t\t}\n\t\t\t/*\n\t\t\t * UP RIGHT\n\t\t\t */\n\t\t\tif j > 2 && i < 17 {\n\t\t\t\tpsum := 1\n\t\t\t\tfor k := 0; k < 4; k++ {\n\t\t\t\t\tpsum = psum * vec[i+k][j-k]\n\t\t\t\t}\n\t\t\t\tif psum > sum {\n\t\t\t\t\tsum = psum\n\t\t\t\t}\n\t\t\t}\n\t\t\t/*\n\t\t\t * DOWN LEFT\n\t\t\t */\n\t\t\tif i > 2 && j < 17 {\n\t\t\t\tpsum := 1\n\t\t\t\tfor k := 0; k < 4; k++ {\n\t\t\t\t\tpsum = psum * vec[i-k][j+k]\n\t\t\t\t}\n\t\t\t\tif psum > sum {\n\t\t\t\t\tsum = psum\n\t\t\t\t}\n\t\t\t}\n\t\t\t/*\n\t\t\t * DOWN RIGHT\n\t\t\t */\n\t\t\tif i < 17 && j < 17 {\n\t\t\t\tpsum := 1\n\t\t\t\tfor k := 0; k < 4; k++ {\n\t\t\t\t\tpsum = psum * vec[i+k][j+k]\n\t\t\t\t}\n\t\t\t\tif psum > sum {\n\t\t\t\t\tsum = psum\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif sum > max {\n\t\t\t\tmax = sum\n\t\t\t}\n\t\t}\n\t}\n\treturn max\n}", "func SplitBy46(sum float64) (x, y float64){\n\tx = sum * (4.0/10.0)\n\ty = sum - x\n\treturn\n}", "func div(x byte, y byte) byte {\n\tif x == 0 || y == 0 {\n\t\treturn 0\n\t}\n\treturn expOp[0xff+logOp[x]-logOp[y]]\n}", "func DIVQ(mr operand.Op) { ctx.DIVQ(mr) }", "func getLayout(n int) (int, int) {\n\tvar root float64 = math.Sqrt(float64(n))\n\tvar h, w int = int(root), 0\n\tfor h > 1 {\n\t\tif n%h == 0 {\n\t\t\t//if we find a multiple, we are done\n\t\t\tw = n / h\n\t\t\tbreak\n\t\t}\n\t\t//if not, try again\n\t\th--\n\t}\n\tif h <= 1 {\n\t\th = 1\n\t\tw = n\n\t}\n\treturn w, h\n}", "func problem5(maxDivisor int) int {\n\tmaxOddDivisor := maxDivisor\n\tif maxDivisor%2 != 0 {\n\t\tmaxOddDivisor--\n\t}\n\n\tfor evenFactor := 2; true; evenFactor += 2 {\n\t\t// Solution has to be even\n\t\tsolution := evenFactor * maxOddDivisor\n\n\t\t// Solution has to be dividable by 5 and all other divider\n\t\tif (maxDivisor >= 5 && solution%5 == 0) && dividableByAllBelow(maxDivisor, solution) {\n\t\t\treturn solution\n\t\t}\n\t}\n\n\treturn -1\n}", "func dget37(rmax *mat.Vector, lmax, ninfo *[]int) (knt int) {\n\tvar bignum, eps, epsin, one, smlnum, tnrm, tol, tolin, two, v, vimin, vmax, vmul, vrmin, zero float64\n\tvar _i, i, icmp, ifnd, info, iscl, j, kmin, ldt, lwork, n int\n\tvar err error\n\n\tcnt := 0\n\tzero = 0.0\n\tone = 1.0\n\ttwo = 2.0\n\tepsin = 5.9605e-8\n\tldt = 20\n\tlwork = 2 * ldt * (10 + ldt)\n\t_select := make([]bool, 20)\n\tdum := vf(1)\n\ts := vf(20)\n\tsep := vf(20)\n\tsepin := vf(20)\n\tseptmp := vf(20)\n\tsin := vf(20)\n\tstmp := vf(20)\n\tval := vf(3)\n\twi := vf(20)\n\twiin := vf(20)\n\twitmp := vf(20)\n\twork := vf(lwork)\n\twr := vf(20)\n\twrin := vf(20)\n\twrtmp := vf(20)\n\tiwork := make([]int, lwork)\n\tlcmp := make([]int, 3)\n\tle := mf(20, 20, opts)\n\tre := mf(20, 20, opts)\n\tt := mf(20, 20, opts)\n\ttmp := mf(20, 20, opts)\n\n\teps = golapack.Dlamch(Precision)\n\tsmlnum = golapack.Dlamch(SafeMinimum) / eps\n\tbignum = one / smlnum\n\tsmlnum, bignum = golapack.Dlabad(smlnum, bignum)\n\n\t// EPSIN = 2**(-24) = precision to which input data computed\n\teps = math.Max(eps, epsin)\n\trmax.Set(0, zero)\n\trmax.Set(1, zero)\n\trmax.Set(2, zero)\n\t(*lmax)[0] = 0\n\t(*lmax)[1] = 0\n\t(*lmax)[2] = 0\n\tknt = 0\n\t(*ninfo)[0] = 0\n\t(*ninfo)[1] = 0\n\t(*ninfo)[2] = 0\n\n\tval.Set(0, math.Sqrt(smlnum))\n\tval.Set(1, one)\n\tval.Set(2, math.Sqrt(bignum))\n\n\tnlist := []int{1, 1, 2, 2, 2, 6, 4, 5, 5, 6, 6, 6, 4, 6, 5, 10, 4, 6, 10, 4, 6, 4, 3, 6, 6, 6, 6, 6, 12, 6, 6, 6, 6, 8, 6, 4, 5, 6, 10}\n\ttmplist := [][]float64{\n\t\t{\n\t\t\t0.0000e+00,\n\t\t},\n\t\t{\n\t\t\t1.0000e+00,\n\t\t},\n\t\t{\n\t\t\t0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00,\n\t\t},\n\t\t{\n\t\t\t3.0000e+00, 2.0000e+00,\n\t\t\t2.0000e+00, 3.0000e+00,\n\t\t},\n\t\t{\n\t\t\t3.0000e+00, -2.0000e+00,\n\t\t\t2.0000e+00, 3.0000e+00,\n\t\t},\n\t\t{\n\t\t\t1.0000e-07, -1.0000e-07, 1.0000e+00, 1.1000e+00, 2.3000e+00, 3.7000e+00,\n\t\t\t3.0000e-07, 1.0000e-07, 1.0000e+00, 1.0000e+00, -1.3000e+00, -7.7000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 3.0000e-07, 1.0000e-07, 2.2000e+00, 3.3000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, -1.0000e-07, 3.0000e-07, 1.8000e+00, 1.6000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 4.0000e-06, 5.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 3.0000e+00, 4.0000e-06,\n\t\t},\n\t\t{\n\t\t\t7.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00,\n\t\t\t-1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00,\n\t\t\t-1.0000e+00, 1.0000e+00, 5.0000e+00, -3.0000e+00,\n\t\t\t1.0000e+00, -1.0000e+00, 3.0000e+00, 3.0000e+00,\n\t\t},\n\t\t{\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t},\n\t\t{\n\t\t\t1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00,\n\t\t},\n\t\t{\n\t\t\t1.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 1.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 1.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 1.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 1.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00,\n\t\t},\n\t\t{\n\t\t\t1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t1.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 1.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 1.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 1.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 1.0000e+00,\n\t\t},\n\t\t{\n\t\t\t1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 2.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 3.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 4.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 5.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 6.0000e+00,\n\t\t},\n\t\t{\n\t\t\t9.4480e-01, 6.7670e-01, 6.9080e-01, 5.9650e-01,\n\t\t\t5.8760e-01, 8.6420e-01, 6.7690e-01, 7.2600e-02,\n\t\t\t7.2560e-01, 1.9430e-01, 9.6870e-01, 2.8310e-01,\n\t\t\t2.8490e-01, 5.8000e-02, 4.8450e-01, 7.3610e-01,\n\t\t},\n\t\t{\n\t\t\t5.0410e-01, 6.6520e-01, 7.7190e-01, 6.3870e-01, 5.9550e-01, 6.1310e-01,\n\t\t\t1.5740e-01, 3.7340e-01, 5.9840e-01, 1.5470e-01, 9.4270e-01, 6.5900e-02,\n\t\t\t4.4170e-01, 7.2300e-02, 1.5440e-01, 5.4920e-01, 8.7000e-03, 3.0040e-01,\n\t\t\t2.0080e-01, 6.0800e-01, 3.0340e-01, 8.4390e-01, 2.3900e-01, 5.7680e-01,\n\t\t\t9.3610e-01, 7.4130e-01, 1.4440e-01, 1.7860e-01, 1.4280e-01, 7.2630e-01,\n\t\t\t5.5990e-01, 9.3360e-01, 7.8000e-02, 4.0930e-01, 6.7140e-01, 5.6170e-01,\n\t\t},\n\t\t{\n\t\t\t2.0000e-03, 1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 1.0000e-03, 1.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, -1.0000e-03, 1.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, -2.0000e-03, 1.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t},\n\t\t{\n\t\t\t4.8630e-01, 9.1260e-01, 2.1900e-02, 6.0110e-01, 1.4050e-01, 2.0840e-01, 8.2640e-01, 8.4410e-01, 3.1420e-01, 8.6750e-01,\n\t\t\t7.1500e-01, 2.6480e-01, 8.8510e-01, 2.6150e-01, 5.9520e-01, 4.7800e-01, 7.6730e-01, 4.6110e-01, 5.7320e-01, 7.7000e-03,\n\t\t\t2.1210e-01, 5.5080e-01, 5.2350e-01, 3.0810e-01, 6.6020e-01, 2.8900e-01, 2.3140e-01, 2.2790e-01, 9.6600e-02, 1.0910e-01,\n\t\t\t7.1510e-01, 8.5790e-01, 5.7710e-01, 5.1140e-01, 1.9010e-01, 9.0810e-01, 6.0090e-01, 7.1980e-01, 1.0640e-01, 8.6840e-01,\n\t\t\t5.6800e-01, 2.8100e-02, 4.0140e-01, 6.3150e-01, 1.1480e-01, 7.5800e-02, 9.4230e-01, 7.2030e-01, 3.6850e-01, 1.7430e-01,\n\t\t\t7.7210e-01, 3.0280e-01, 5.5640e-01, 9.9980e-01, 3.6520e-01, 5.2580e-01, 3.7030e-01, 6.7790e-01, 9.9350e-01, 5.0270e-01,\n\t\t\t7.3960e-01, 4.5600e-02, 7.4740e-01, 9.2880e-01, 2.2000e-03, 8.2600e-02, 3.6340e-01, 4.9120e-01, 9.4050e-01, 3.8910e-01,\n\t\t\t5.6370e-01, 8.5540e-01, 3.2100e-02, 2.6380e-01, 3.6090e-01, 6.4970e-01, 8.4690e-01, 9.3500e-01, 3.7000e-02, 2.9170e-01,\n\t\t\t8.6560e-01, 6.3270e-01, 3.5620e-01, 6.3560e-01, 2.7360e-01, 6.5120e-01, 1.0220e-01, 2.8880e-01, 5.7620e-01, 4.0790e-01,\n\t\t\t5.3320e-01, 4.1210e-01, 7.2870e-01, 2.3110e-01, 6.8300e-01, 7.3860e-01, 8.1800e-01, 9.8150e-01, 8.0550e-01, 2.5660e-01,\n\t\t},\n\t\t{\n\t\t\t-3.8730e-01, 3.6560e-01, 3.1200e-02, -5.8340e-01,\n\t\t\t5.5230e-01, -1.1854e+00, 9.8330e-01, 7.6670e-01,\n\t\t\t1.6746e+00, -1.9900e-02, -1.8293e+00, 5.7180e-01,\n\t\t\t-5.2500e-01, 3.5340e-01, -2.7210e-01, -8.8300e-02,\n\t\t},\n\t\t{\n\t\t\t-1.0777e+00, 1.7027e+00, 2.6510e-01, 8.5160e-01, 1.0121e+00, 2.5710e-01,\n\t\t\t-1.3400e-02, 3.9030e-01, -1.2680e+00, 2.7530e-01, -3.2350e-01, -1.3844e+00,\n\t\t\t1.5230e-01, 3.0680e-01, 8.7330e-01, -3.3410e-01, -4.8310e-01, -1.5416e+00,\n\t\t\t1.4470e-01, -6.0570e-01, 3.1900e-02, -1.0905e+00, -8.3700e-02, 6.2410e-01,\n\t\t\t-7.6510e-01, -1.7889e+00, -1.5069e+00, -6.0210e-01, 5.2170e-01, 6.4700e-01,\n\t\t\t8.1940e-01, 2.1100e-01, 5.4320e-01, 7.5610e-01, 1.7130e-01, 5.5400e-01,\n\t\t},\n\t\t{\n\t\t\t-1.0639e+00, 1.6120e-01, 1.5620e-01, 3.4360e-01, -6.7480e-01, 1.6598e+00, 6.4650e-01, -7.8630e-01, -2.6100e-01, 7.0190e-01,\n\t\t\t-8.4400e-01, -2.2439e+00, 1.8800e+00, -1.0005e+00, 7.4500e-02, -1.6156e+00, 2.8220e-01, 8.5600e-01, 1.3497e+00, -1.5883e+00,\n\t\t\t1.5988e+00, 1.1758e+00, 1.2398e+00, 1.1173e+00, 2.1500e-01, 4.3140e-01, 1.8500e-01, 7.9470e-01, 6.6260e-01, 8.6460e-01,\n\t\t\t-2.2960e-01, 1.2442e+00, 2.3242e+00, -5.0690e-01, -7.5160e-01, -5.4370e-01, -2.5990e-01, 1.2830e+00, -1.1067e+00, -1.1150e-01,\n\t\t\t-3.6040e-01, 4.0420e-01, 6.1240e-01, -1.2164e+00, -9.4650e-01, -3.1460e-01, 1.8310e-01, 7.3710e-01, 1.4278e+00, 2.9220e-01,\n\t\t\t4.6150e-01, 3.8740e-01, -4.2900e-02, -9.3600e-01, 7.1160e-01, -8.2590e-01, -1.7640e+00, -9.4660e-01, 1.8202e+00, -2.5480e-01,\n\t\t\t1.2934e+00, -9.7550e-01, 6.7480e-01, -1.0481e+00, -1.8442e+00, -5.4600e-02, 7.4050e-01, 6.1000e-03, 1.2430e+00, -1.8490e-01,\n\t\t\t-3.4710e-01, -9.5800e-01, 1.6530e-01, 9.1300e-02, -5.2010e-01, -1.1832e+00, 8.5410e-01, -2.3200e-01, -1.6155e+00, 5.5180e-01,\n\t\t\t1.0190e+00, -6.8240e-01, 8.0850e-01, 2.5950e-01, -3.7580e-01, -1.8825e+00, 1.6473e+00, -6.5920e-01, 8.0250e-01, -4.9000e-03,\n\t\t\t1.2670e+00, -4.2400e-02, 8.9570e-01, -1.6770e-01, 1.4620e-01, 9.8800e-01, -2.3170e-01, -1.4483e+00, -5.8200e-02, 1.9700e-02,\n\t\t},\n\t\t{\n\t\t\t-1.2298e+00, -2.3142e+00, -6.9800e-02, 1.0523e+00,\n\t\t\t2.0390e-01, -1.2298e+00, 8.0500e-02, 9.7860e-01,\n\t\t\t0.0000e+00, 0.0000e+00, 2.5600e-01, -8.9100e-01,\n\t\t\t0.0000e+00, 0.0000e+00, 2.7480e-01, 2.5600e-01,\n\t\t},\n\t\t{\n\t\t\t5.9930e-01, 1.9372e+00, -1.6160e-01, -1.4602e+00, 6.0180e-01, 2.7120e+00,\n\t\t\t-2.2049e+00, 5.9930e-01, -1.0679e+00, 1.9405e+00, -1.4400e+00, -2.2110e-01,\n\t\t\t0.0000e+00, 0.0000e+00, -2.4567e+00, -6.8650e-01, -1.9101e+00, 6.4960e-01,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 7.3620e-01, 3.9700e-01, -1.5190e-01,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, -1.0034e+00, 1.1954e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, -1.3400e-01, -1.0034e+00,\n\t\t},\n\t\t{\n\t\t\t1.0000e-04, 1.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, -1.0000e-04, 1.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 1.0000e-02, 1.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, -5.0000e-03,\n\t\t},\n\t\t{\n\t\t\t2.0000e-06, 1.0000e+00, -2.0000e+00,\n\t\t\t1.0000e-06, -2.0000e+00, 4.0000e+00,\n\t\t\t0.0000e+00, 1.0000e+00, -2.0000e+00,\n\t\t},\n\t\t{\n\t\t\t2.4080e-01, 6.5530e-01, 9.1660e-01, 5.0300e-02, 2.8490e-01, 2.4080e-01,\n\t\t\t6.9070e-01, 9.7000e-01, 1.4020e-01, 5.7820e-01, 6.7670e-01, 6.9070e-01,\n\t\t\t1.0620e-01, 3.8000e-02, 7.0540e-01, 2.4320e-01, 8.6420e-01, 1.0620e-01,\n\t\t\t2.6400e-01, 9.8800e-02, 1.7800e-02, 9.4480e-01, 1.9430e-01, 2.6400e-01,\n\t\t\t7.0340e-01, 2.5600e-01, 2.6110e-01, 5.8760e-01, 5.8000e-02, 7.0340e-01,\n\t\t\t4.0210e-01, 5.5980e-01, 1.3580e-01, 7.2560e-01, 6.9080e-01, 4.0210e-01,\n\t\t},\n\t\t{\n\t\t\t3.4800e+00, -2.9900e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t-4.9000e-01, 2.4800e+00, -1.9900e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, -4.9000e-01, 1.4800e+00, -9.9000e-01, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, -9.9000e-01, 1.4800e+00, -4.9000e-01, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, -1.9900e+00, 2.4800e+00, -4.9000e-01,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, -2.9900e+00, 3.4800e+00,\n\t\t},\n\t\t{\n\t\t\t0.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00,\n\t\t\t1.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 1.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00, 1.0000e+00,\n\t\t\t-1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00,\n\t\t},\n\t\t{\n\t\t\t3.5345e-01, 9.3023e-01, 7.4679e-02, -1.0059e-02, 4.6698e-02, -4.3480e-02,\n\t\t\t9.3545e-01, -3.5147e-01, -2.8216e-02, 3.8008e-03, -1.7644e-02, 1.6428e-02,\n\t\t\t0.0000e+00, -1.0555e-01, 7.5211e-01, -1.0131e-01, 4.7030e-01, -4.3789e-01,\n\t\t\t0.0000e+00, 0.0000e+00, 6.5419e-01, 1.1779e-01, -5.4678e-01, 5.0911e-01,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, -9.8780e-01, -1.1398e-01, 1.0612e-01,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 6.8144e-01, 7.3187e-01,\n\t\t},\n\t\t{\n\t\t\t1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00,\n\t\t\t5.0000e-01, 3.3330e-01, 2.5000e-01, 2.0000e-01, 1.6670e-01, 1.4290e-01,\n\t\t\t3.3330e-01, 2.5000e-01, 2.0000e-01, 1.6670e-01, 1.4290e-01, 1.2500e-01,\n\t\t\t2.5000e-01, 2.0000e-01, 1.6670e-01, 1.4290e-01, 1.2500e-01, 1.1110e-01,\n\t\t\t2.0000e-01, 1.6670e-01, 1.4290e-01, 1.2500e-01, 1.1110e-01, 1.0000e-01,\n\t\t\t1.6670e-01, 1.4290e-01, 1.2500e-01, 1.1110e-01, 1.0000e-01, 9.0900e-02,\n\t\t},\n\t\t{\n\t\t\t1.2000e+01, 1.1000e+01, 1.0000e+01, 9.0000e+00, 8.0000e+00, 7.0000e+00, 6.0000e+00, 5.0000e+00, 4.0000e+00, 3.0000e+00, 2.0000e+00, 1.0000e+00,\n\t\t\t1.1000e+01, 1.1000e+01, 1.0000e+01, 9.0000e+00, 8.0000e+00, 7.0000e+00, 6.0000e+00, 5.0000e+00, 4.0000e+00, 3.0000e+00, 2.0000e+00, 1.0000e+00,\n\t\t\t0.0000e+00, 1.0000e+01, 1.0000e+01, 9.0000e+00, 8.0000e+00, 7.0000e+00, 6.0000e+00, 5.0000e+00, 4.0000e+00, 3.0000e+00, 2.0000e+00, 1.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 9.0000e+00, 9.0000e+00, 8.0000e+00, 7.0000e+00, 6.0000e+00, 5.0000e+00, 4.0000e+00, 3.0000e+00, 2.0000e+00, 1.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 8.0000e+00, 8.0000e+00, 7.0000e+00, 6.0000e+00, 5.0000e+00, 4.0000e+00, 3.0000e+00, 2.0000e+00, 1.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 7.0000e+00, 7.0000e+00, 6.0000e+00, 5.0000e+00, 4.0000e+00, 3.0000e+00, 2.0000e+00, 1.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 6.0000e+00, 6.0000e+00, 5.0000e+00, 4.0000e+00, 3.0000e+00, 2.0000e+00, 1.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 5.0000e+00, 5.0000e+00, 4.0000e+00, 3.0000e+00, 2.0000e+00, 1.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 4.0000e+00, 4.0000e+00, 3.0000e+00, 2.0000e+00, 1.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 3.0000e+00, 3.0000e+00, 2.0000e+00, 1.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 2.0000e+00, 2.0000e+00, 1.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 1.0000e+00,\n\t\t},\n\t\t{\n\t\t\t0.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t5.0000e+00, 0.0000e+00, 2.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 4.0000e+00, 0.0000e+00, 3.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 3.0000e+00, 0.0000e+00, 4.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 2.0000e+00, 0.0000e+00, 5.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00,\n\t\t},\n\t\t{\n\t\t\t1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00,\n\t\t\t-1.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00,\n\t\t\t-1.0000e+00, -1.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00,\n\t\t\t-1.0000e+00, -1.0000e+00, -1.0000e+00, 1.0000e+00, 0.0000e+00, 1.0000e+00,\n\t\t\t-1.0000e+00, -1.0000e+00, -1.0000e+00, -1.0000e+00, 1.0000e+00, 1.0000e+00,\n\t\t\t-1.0000e+00, -1.0000e+00, -1.0000e+00, -1.0000e+00, -1.0000e+00, 1.0000e+00,\n\t\t},\n\t\t{\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00,\n\t\t\t1.0000e+00, 1.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 1.0000e+00,\n\t\t\t0.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 0.0000e+00, 1.0000e+00,\n\t\t\t1.0000e+00, 0.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00,\n\t\t\t1.0000e+00, 0.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 1.0000e+00, 0.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00,\n\t\t},\n\t\t{\n\t\t\t1.0000e+00, 4.0112e+00, 1.2750e+01, 4.0213e+01, 1.2656e+02, 3.9788e+02,\n\t\t\t1.0000e+00, 3.2616e+00, 1.0629e+01, 3.3342e+01, 1.0479e+02, 3.2936e+02,\n\t\t\t1.0000e+00, 3.1500e+00, 9.8006e+00, 3.0630e+01, 9.6164e+01, 3.0215e+02,\n\t\t\t1.0000e+00, 3.2755e+00, 1.0420e+01, 3.2957e+01, 1.0374e+02, 3.2616e+02,\n\t\t\t1.0000e+00, 2.8214e+00, 8.4558e+00, 2.6296e+01, 8.2443e+01, 2.5893e+02,\n\t\t\t1.0000e+00, 2.6406e+00, 8.3565e+00, 2.6558e+01, 8.3558e+01, 2.6268e+02,\n\t\t},\n\t\t{\n\t\t\t0.0000e+00, 4.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t1.0000e+00, 0.0000e+00, 4.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 1.0000e+00, 0.0000e+00, 4.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00, 4.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00, 4.0000e+00, 0.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00, 4.0000e+00, 0.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00, 4.0000e+00,\n\t\t\t0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 0.0000e+00,\n\t\t},\n\t\t{\n\t\t\t8.5000e+00, -1.0472e+01, 2.8944e+00, -1.5279e+00, 1.1056e+00, -5.0000e-01,\n\t\t\t2.6180e+00, -1.1708e+00, -2.0000e+00, 8.9440e-01, -6.1800e-01, 2.7640e-01,\n\t\t\t-7.2360e-01, 2.0000e+00, -1.7080e-01, -1.6180e+00, 8.9440e-01, -3.8200e-01,\n\t\t\t3.8200e-01, -8.9440e-01, 1.6180e+00, 1.7080e-01, -2.0000e+00, 7.2360e-01,\n\t\t\t-2.7640e-01, 6.1800e-01, -8.9440e-01, 2.0000e+00, 1.1708e+00, -2.6180e+00,\n\t\t\t5.0000e-01, -1.1056e+00, 1.5279e+00, -2.8944e+00, 1.0472e+01, -8.5000e+00,\n\t\t},\n\t\t{\n\t\t\t4.0000e+00, -5.0000e+00, 0.0000e+00, 3.0000e+00,\n\t\t\t0.0000e+00, 4.0000e+00, -3.0000e+00, -5.0000e+00,\n\t\t\t5.0000e+00, -3.0000e+00, 4.0000e+00, 0.0000e+00,\n\t\t\t3.0000e+00, 0.0000e+00, 5.0000e+00, 4.0000e+00,\n\t\t},\n\t\t{\n\t\t\t1.5000e+01, 1.1000e+01, 6.0000e+00, -9.0000e+00, -1.5000e+01,\n\t\t\t1.0000e+00, 3.0000e+00, 9.0000e+00, -3.0000e+00, -8.0000e+00,\n\t\t\t7.0000e+00, 6.0000e+00, 6.0000e+00, -3.0000e+00, -1.1000e+01,\n\t\t\t7.0000e+00, 7.0000e+00, 5.0000e+00, -3.0000e+00, -1.1000e+01,\n\t\t\t1.7000e+01, 1.2000e+01, 5.0000e+00, -1.0000e+01, -1.6000e+01,\n\t\t},\n\t\t{\n\t\t\t-9.0000e+00, 2.1000e+01, -1.5000e+01, 4.0000e+00, 2.0000e+00, 0.0000e+00,\n\t\t\t-1.0000e+01, 2.1000e+01, -1.4000e+01, 4.0000e+00, 2.0000e+00, 0.0000e+00,\n\t\t\t-8.0000e+00, 1.6000e+01, -1.1000e+01, 4.0000e+00, 2.0000e+00, 0.0000e+00,\n\t\t\t-6.0000e+00, 1.2000e+01, -9.0000e+00, 3.0000e+00, 3.0000e+00, 0.0000e+00,\n\t\t\t-4.0000e+00, 8.0000e+00, -6.0000e+00, 0.0000e+00, 5.0000e+00, 0.0000e+00,\n\t\t\t-2.0000e+00, 4.0000e+00, -3.0000e+00, 0.0000e+00, 1.0000e+00, 3.0000e+00,\n\t\t},\n\t\t{\n\t\t\t1.0000e+00, 1.0000e+00, 1.0000e+00, -2.0000e+00, 1.0000e+00, -1.0000e+00, 2.0000e+00, -2.0000e+00, 4.0000e+00, -3.0000e+00,\n\t\t\t-1.0000e+00, 2.0000e+00, 3.0000e+00, -4.0000e+00, 2.0000e+00, -2.0000e+00, 4.0000e+00, -4.0000e+00, 8.0000e+00, -6.0000e+00,\n\t\t\t-1.0000e+00, 0.0000e+00, 5.0000e+00, -5.0000e+00, 3.0000e+00, -3.0000e+00, 6.0000e+00, -6.0000e+00, 1.2000e+01, -9.0000e+00,\n\t\t\t-1.0000e+00, 0.0000e+00, 3.0000e+00, -4.0000e+00, 4.0000e+00, -4.0000e+00, 8.0000e+00, -8.0000e+00, 1.6000e+01, -1.2000e+01,\n\t\t\t-1.0000e+00, 0.0000e+00, 3.0000e+00, -6.0000e+00, 5.0000e+00, -4.0000e+00, 1.0000e+01, -1.0000e+01, 2.0000e+01, -1.5000e+01,\n\t\t\t-1.0000e+00, 0.0000e+00, 3.0000e+00, -6.0000e+00, 2.0000e+00, -2.0000e+00, 1.2000e+01, -1.2000e+01, 2.4000e+01, -1.8000e+01,\n\t\t\t-1.0000e+00, 0.0000e+00, 3.0000e+00, -6.0000e+00, 2.0000e+00, -5.0000e+00, 1.5000e+01, -1.3000e+01, 2.8000e+01, -2.1000e+01,\n\t\t\t-1.0000e+00, 0.0000e+00, 3.0000e+00, -6.0000e+00, 2.0000e+00, -5.0000e+00, 1.2000e+01, -1.1000e+01, 3.2000e+01, -2.4000e+01,\n\t\t\t-1.0000e+00, 0.0000e+00, 3.0000e+00, -6.0000e+00, 2.0000e+00, -5.0000e+00, 1.2000e+01, -1.4000e+01, 3.7000e+01, -2.6000e+01,\n\t\t\t-1.0000e+00, 0.0000e+00, 3.0000e+00, -6.0000e+00, 2.0000e+00, -5.0000e+00, 1.2000e+01, -1.4000e+01, 3.6000e+01, -2.5000e+01,\n\t\t},\n\t}\n\twrinlist := [][]float64{\n\t\t{0.0000e+00},\n\t\t{1.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00},\n\t\t{1.0000e+00, 5.0000e+00},\n\t\t{3.0000e+00, 3.0000e+00},\n\t\t{-3.8730e+00, 1.0000e-07, 1.0000e-07, 3.0000e-07, 3.0000e-07, 3.8730e+00},\n\t\t{3.9603e+00, 3.9603e+00, 4.0397e+00, 4.0397e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00},\n\t\t{1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00},\n\t\t{1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00},\n\t\t{1.0000e+00, 2.0000e+00, 3.0000e+00, 4.0000e+00, 5.0000e+00, 6.0000e+00},\n\t\t{2.4326e-01, 2.4326e-01, 7.4091e-01, 2.2864e+00},\n\t\t{-5.2278e-01, -3.5380e-01, -8.0876e-03, 3.4760e-01, 3.4760e-01, 2.7698e+00},\n\t\t{-2.0000e-03, -1.0000e-03, 0.0000e+00, 1.0000e-03, 2.0000e-03},\n\t\t{-4.6121e-01, -4.6121e-01, -4.5164e-01, -1.4922e-01, -1.4922e-01, 3.3062e-02, 3.0849e-01, 3.0849e-01, 5.4509e-01, 5.0352e+00},\n\t\t{-1.8952e+00, -1.8952e+00, -9.5162e-02, 3.9520e-01},\n\t\t{-1.7029e+00, -1.0307e+00, 2.8487e-01, 2.8487e-01, 1.1675e+00, 1.1675e+00},\n\t\t{-2.6992e+00, -2.6992e+00, -2.4366e+00, -1.2882e+00, -1.2882e+00, 9.0275e-01, 9.0442e-01, 9.0442e-01, 1.6774e+00, 3.0060e+00},\n\t\t{-1.2298e+00, -1.2298e+00, 2.5600e-01, 2.5600e-01},\n\t\t{-2.4567e+00, -1.0034e+00, -1.0034e+00, 5.9930e-01, 5.9930e-01, 7.3620e-01},\n\t\t{-5.0000e-03, -1.0000e-04, 1.0000e-04, 1.0000e-02},\n\t\t{-4.0000e+00, 0.0000e+00, 2.2096e-06},\n\t\t{-3.4008e-01, -3.4008e-01, -1.6998e-07, 7.2311e-01, 7.2311e-01, 2.5551e+00},\n\t\t{1.3034e-02, 1.1294e+00, 2.0644e+00, 2.8388e+00, 4.3726e+00, 4.4618e+00},\n\t\t{-1.7321e+00, -1.0000e+00, 0.0000e+00, 0.0000e+00, 1.0000e+00, 1.7321e+00},\n\t\t{-9.9980e-01, -9.9980e-01, 7.4539e-01, 7.4539e-01, 9.9929e-01, 9.9929e-01},\n\t\t{-2.2135e-01, -3.1956e-02, -8.5031e-04, -5.8584e-05, 1.3895e-05, 2.1324e+00},\n\t\t{-2.8234e-02, 7.2587e-02, 7.2587e-02, 1.8533e-01, 2.8828e-01, 6.4315e-01, 1.5539e+00, 3.5119e+00, 6.9615e+00, 1.2311e+01, 2.0199e+01, 3.2229e+01},\n\t\t{-5.0000e+00, -3.0000e+00, -1.0000e+00, 1.0000e+00, 3.0000e+00, 5.0000e+00},\n\t\t{8.0298e-02, 8.0298e-02, 1.4415e+00, 1.4415e+00, 1.4782e+00, 1.4782e+00},\n\t\t{-3.5343e-02, -3.5343e-02, 5.8440e-07, 6.4087e-01, 6.4087e-01, 3.7889e+00},\n\t\t{-5.3220e-01, -1.0118e-01, -9.8749e-03, 2.9861e-03, 1.8075e-01, 3.9260e+02},\n\t\t{-3.7588e+00, -3.0642e+00, -2.0000e+00, -6.9459e-01, 6.9459e-01, 2.0000e+00, 3.0642e+00, 3.7588e+00},\n\t\t{-5.8930e-01, -2.7627e-01, -2.7627e-01, 2.7509e-01, 2.7509e-01, 5.9167e-01},\n\t\t{1.0000e+00, 1.0000e+00, 2.0000e+00, 1.2000e+01},\n\t\t{-9.9999e-01, 1.4980e+00, 1.4980e+00, 1.5020e+00, 1.5020e+00},\n\t\t{1.0000e+00, 1.0000e+00, 2.0000e+00, 2.0000e+00, 3.0000e+00, 3.0000e+00},\n\t\t{1.0000e+00, 1.9867e+00, 2.0000e+00, 2.0000e+00, 2.0067e+00, 2.0067e+00, 2.9970e+00, 3.0000e+00, 3.0000e+00, 3.0030e+00},\n\t}\n\twiinlist := [][]float64{\n\t\t{0.0000e+00},\n\t\t{0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00},\n\t\t{2.0000e+00, -2.0000e+00},\n\t\t{0.0000e+00, 1.7321e-07, -1.7321e-07, 1.0000e-07, -1.0000e-07, 0.0000e+00},\n\t\t{4.0425e-02, -4.0425e-02, 3.8854e-02, -3.8854e-02},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{2.1409e-01, -2.1409e-01, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 3.0525e-01, -3.0525e-01, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{7.2657e-01, -7.2657e-01, 0.0000e+00, 4.8255e-01, -4.8255e-01, 0.0000e+00, 1.1953e-01, -1.1953e-01, 0.0000e+00, 0.0000e+00},\n\t\t{7.5059e-01, -7.5059e-01, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 1.2101e+00, -1.2101e+00, 4.6631e-01, -4.6631e-01},\n\t\t{9.0387e-01, -9.0387e-01, 0.0000e+00, 8.8930e-01, -8.8930e-01, 0.0000e+00, 2.5661e+00, -2.5661e+00, 0.0000e+00, 0.0000e+00},\n\t\t{6.8692e-01, -6.8692e-01, 4.9482e-01, -4.9482e-01},\n\t\t{0.0000e+00, 4.0023e-01, -4.0023e-01, 2.0667e+00, -2.0667e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{3.2133e-01, -3.2133e-01, 0.0000e+00, 5.9389e-02, -5.9389e-02, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{1.9645e-02, -1.9645e-02, 6.6663e-01, -6.6663e-01, 3.7545e-02, -3.7545e-02},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 9.0746e-02, -9.0746e-02, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{2.4187e+00, -2.4187e+00, 6.2850e-01, -6.2850e-01, 1.5638e-01, -1.5638e-01},\n\t\t{7.4812e-01, -7.4812e-01, 0.0000e+00, 7.2822e-01, -7.2822e-01, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 4.9852e-01, -4.9852e-01, 5.0059e-01, -5.0059e-01, 0.0000e+00},\n\t\t{5.0000e+00, -5.0000e+00, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 3.5752e+00, -3.5752e+00, 3.5662e+00, -3.5662e+00},\n\t\t{6.2559e-04, -6.2559e-04, 1.0001e+00, -1.0001e+00, 0.0000e+00, 0.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00, 2.5052e-03, -2.5052e-03, 1.1763e-02, -1.1763e-02, 0.0000e+00, 8.7028e-04, -8.7028e-04, 0.0000e+00},\n\t}\n\tsinlist := [][]float64{\n\t\t{1.0000e+00},\n\t\t{1.0000e+00},\n\t\t{1.0000e+00, 1.0000e+00},\n\t\t{1.0000e+00, 1.0000e+00},\n\t\t{1.0000e+00, 1.0000e+00},\n\t\t{6.9855e-01, 9.7611e-08, 9.7611e-08, 1.0000e-07, 1.0000e-07, 4.0659e-01},\n\t\t{1.1244e-05, 1.1244e-05, 1.0807e-05, 1.0807e-05},\n\t\t{1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00},\n\t\t{1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00},\n\t\t{2.4074e-35, 2.4074e-35, 2.4074e-35, 2.4074e-35, 2.4074e-35, 2.4074e-35},\n\t\t{2.4074e-35, 2.4074e-35, 2.4074e-35, 2.4074e-35, 2.4074e-35, 2.4074e-35},\n\t\t{1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00},\n\t\t{8.7105e-01, 8.7105e-01, 9.8194e-01, 9.7723e-01},\n\t\t{2.7888e-01, 3.5427e-01, 3.4558e-01, 5.4661e-01, 5.4661e-01, 9.6635e-01},\n\t\t{2.4000e-11, 6.0000e-12, 4.0000e-12, 6.0000e-12, 2.4000e-11},\n\t\t{4.7781e-01, 4.7781e-01, 4.6034e-01, 4.7500e-01, 4.7500e-01, 2.9729e-01, 4.2947e-01, 4.2947e-01, 7.0777e-01, 9.7257e-01},\n\t\t{8.1913e-01, 8.1913e-01, 8.0499e-01, 9.8222e-01},\n\t\t{6.7909e-01, 7.2671e-01, 3.9757e-01, 3.9757e-01, 4.2334e-01, 4.2334e-01},\n\t\t{6.4005e-01, 6.4005e-01, 6.9083e-01, 5.3435e-01, 5.3435e-01, 2.9802e-01, 7.3193e-01, 7.3193e-01, 3.0743e-01, 8.5623e-01},\n\t\t{4.7136e-01, 4.7136e-01, 8.0960e-01, 8.0960e-01},\n\t\t{4.7091e-01, 3.6889e-01, 3.6889e-01, 5.8849e-01, 5.8849e-01, 6.0845e-01},\n\t\t{3.7485e-07, 9.8979e-09, 1.0098e-08, 1.4996e-06},\n\t\t{7.3030e-01, 7.2801e-01, 8.2763e-01},\n\t\t{5.7839e-01, 5.7839e-01, 4.9641e-01, 7.0039e-01, 7.0039e-01, 9.2518e-01},\n\t\t{7.5301e-01, 6.0479e-01, 5.4665e-01, 4.2771e-01, 6.6370e-01, 5.7388e-01},\n\t\t{8.6603e-01, 5.0000e-01, 2.9582e-31, 2.9582e-31, 5.0000e-01, 8.6603e-01},\n\t\t{1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00},\n\t\t{4.0841e-01, 3.7927e-01, 6.2793e-01, 8.1156e-01, 9.7087e-01, 8.4325e-01},\n\t\t{2.8690e-06, 1.5885e-06, 1.5885e-06, 6.5757e-07, 1.8324e-06, 6.8640e-05, 4.6255e-03, 1.4447e-01, 5.8447e-01, 3.1823e-01, 2.0079e-01, 3.0424e-01},\n\t\t{8.2295e-01, 7.2281e-01, 6.2854e-01, 6.2854e-01, 7.2281e-01, 8.2295e-01},\n\t\t{8.9968e-01, 8.9968e-01, 9.6734e-01, 9.6734e-01, 9.7605e-01, 9.7605e-01},\n\t\t{3.9345e-01, 3.9345e-01, 2.8868e-01, 4.5013e-01, 4.5013e-01, 9.6305e-01},\n\t\t{5.3287e-01, 7.2342e-01, 7.3708e-01, 4.4610e-01, 4.2881e-01, 4.8057e-01},\n\t\t{1.2253e-01, 4.9811e-02, 3.6914e-02, 3.3328e-02, 3.3328e-02, 3.6914e-02, 4.9811e-02, 1.2253e-01},\n\t\t{1.7357e-04, 1.7486e-04, 1.7486e-04, 1.7635e-04, 1.7635e-04, 1.7623e-04},\n\t\t{1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00},\n\t\t{2.1768e-01, 3.9966e-04, 3.9966e-04, 3.9976e-04, 3.9976e-04},\n\t\t{6.4875e-05, 6.4875e-05, 5.4076e-02, 5.4076e-02, 8.6149e-01, 1.2425e-01},\n\t\t{3.6037e-02, 7.4283e-05, 1.4346e-04, 1.4346e-04, 6.7873e-05, 6.7873e-05, 9.2779e-05, 2.7358e-04, 2.7358e-04, 9.2696e-05},\n\t}\n\tsepinlist := [][]float64{\n\t\t{0.0000e+00},\n\t\t{1.0000e+00},\n\t\t{0.0000e+00, 0.0000e+00},\n\t\t{4.0000e+00, 4.0000e+00},\n\t\t{4.0000e+00, 4.0000e+00},\n\t\t{2.2823e+00, 5.0060e-14, 5.0060e-14, 9.4094e-14, 9.4094e-14, 1.5283e+00},\n\t\t{3.1179e-05, 3.1179e-05, 2.9981e-05, 2.9981e-05},\n\t\t{1.9722e-31, 1.9722e-31, 1.9722e-31, 1.9722e-31, 1.9722e-31},\n\t\t{1.9722e-31, 1.9722e-31, 1.9722e-31, 1.9722e-31, 1.9722e-31},\n\t\t{2.4074e-35, 2.4074e-35, 2.4074e-35, 2.4074e-35, 2.4074e-35, 2.4074e-35},\n\t\t{2.4074e-35, 2.4074e-35, 2.4074e-35, 2.4074e-35, 2.4074e-35, 2.4074e-35},\n\t\t{1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00, 1.0000e+00},\n\t\t{3.5073e-01, 3.5073e-01, 4.6989e-01, 1.5455e+00},\n\t\t{1.1793e-01, 6.8911e-02, 1.3489e-01, 1.7729e-01, 1.7729e-01, 1.8270e+00},\n\t\t{2.3952e-11, 5.9940e-12, 3.9920e-12, 5.9940e-12, 2.3952e-11},\n\t\t{1.5842e-01, 1.5842e-01, 1.9931e-01, 9.1686e-02, 9.1686e-02, 8.2469e-02, 3.9688e-02, 3.9688e-02, 1.5033e-01, 3.5548e+00},\n\t\t{7.7090e-01, 7.7090e-01, 4.9037e-01, 4.9037e-01},\n\t\t{6.7220e-01, 2.0436e-01, 4.9797e-01, 4.9797e-01, 1.9048e-01, 1.9048e-01},\n\t\t{4.1615e-01, 4.1615e-01, 2.5476e-01, 6.0878e-01, 6.0878e-01, 4.7530e-01, 6.2016e-01, 6.2016e-01, 4.1726e-01, 4.3175e-01},\n\t\t{7.1772e-01, 7.1772e-01, 5.1408e-01, 5.1408e-01},\n\t\t{8.5788e-01, 1.8909e-01, 1.8909e-01, 1.3299e+00, 1.3299e+00, 9.6725e-01},\n\t\t{3.6932e-07, 9.8493e-09, 1.0046e-08, 1.4773e-06},\n\t\t{4.0000e+00, 1.3726e-06, 2.2096e-06},\n\t\t{2.0310e-01, 2.0310e-01, 2.1574e-01, 4.1945e-02, 4.1945e-02, 1.7390e+00},\n\t\t{6.0533e-01, 2.8613e-01, 1.7376e-01, 3.0915e-01, 7.6443e-02, 8.9227e-02},\n\t\t{7.2597e-01, 2.6417e-01, 1.4600e-07, 6.2446e-08, 2.6417e-01, 3.7896e-01},\n\t\t{3.9290e-02, 3.9290e-02, 5.2120e-01, 5.2120e-01, 7.5089e-02, 7.5089e-02},\n\t\t{1.6605e-01, 3.0531e-02, 7.8195e-04, 7.2478e-05, 7.2478e-05, 1.8048e+00},\n\t\t{3.2094e-06, 9.9934e-07, 9.9934e-07, 7.8673e-07, 2.0796e-06, 6.1058e-05, 6.4028e-03, 1.9470e-01, 1.2016e+00, 1.4273e+00, 2.4358e+00, 5.6865e+00},\n\t\t{1.2318e+00, 7.5970e-01, 6.9666e-01, 6.9666e-01, 7.5970e-01, 1.2318e+00},\n\t\t{1.5236e+00, 1.5236e+00, 4.2793e-01, 4.2793e-01, 2.2005e-01, 2.2005e-01},\n\t\t{1.8415e-01, 1.8415e-01, 1.7003e-01, 2.9425e-01, 2.9425e-01, 2.2469e+00},\n\t\t{3.8557e-01, 9.1303e-02, 1.1032e-02, 1.2861e-02, 1.7378e-01, 3.9201e+02},\n\t\t{1.2978e-01, 8.0162e-02, 8.2942e-02, 1.3738e-01, 1.1171e-01, 7.2156e-02, 6.8352e-02, 1.1527e-01},\n\t\t{2.8157e-04, 1.6704e-04, 1.6704e-04, 1.6828e-04, 1.6828e-04, 3.0778e-04},\n\t\t{4.3333e+00, 4.3333e+00, 4.3333e+00, 9.1250e+00},\n\t\t{5.2263e-01, 6.0947e-03, 6.0947e-03, 6.0960e-03, 6.0960e-03},\n\t\t{5.0367e-04, 5.0367e-04, 2.3507e-01, 2.3507e-01, 5.4838e-07, 1.2770e-06},\n\t\t{7.9613e-02, 7.4025e-06, 6.7839e-07, 6.7839e-07, 5.7496e-06, 5.7496e-06, 2.6519e-06, 1.9407e-07, 1.9407e-07, 2.6477e-06},\n\t}\n\n\t// Read input data until N=0. Assume input eigenvalues are sorted\n\t// lexicographically (increasing by real part, then decreasing by\n\t// imaginary part)\n\tfor _i, n = range nlist {\n\t\tcnt++\n\t\tif n == 0 {\n\t\t\treturn\n\t\t}\n\t\tfor i = 1; i <= n; i++ {\n\t\t\tfor j = 1; j <= n; j++ {\n\t\t\t\ttmp.Set(i-1, j-1, tmplist[_i][(i-1)*(n)+j-1])\n\t\t\t}\n\t\t}\n\t\tfor i = 1; i <= n; i++ {\n\t\t\twrin.Set(i-1, wrinlist[_i][i-1])\n\t\t\twiin.Set(i-1, wiinlist[_i][i-1])\n\t\t\tsin.Set(i-1, sinlist[_i][i-1])\n\t\t\tsepin.Set(i-1, sepinlist[_i][i-1])\n\t\t}\n\t\ttnrm = golapack.Dlange('M', n, n, tmp, work)\n\n\t\t// Begin test\n\t\tfor iscl = 1; iscl <= 3; iscl++ {\n\t\t\t// Scale input matrix\n\t\t\tknt = knt + 1\n\t\t\tgolapack.Dlacpy(Full, n, n, tmp, t)\n\t\t\tvmul = val.Get(iscl - 1)\n\t\t\tfor i = 1; i <= n; i++ {\n\t\t\t\tt.Off(0, i-1).Vector().Scal(n, vmul, 1)\n\t\t\t}\n\t\t\tif tnrm == zero {\n\t\t\t\tvmul = one\n\t\t\t}\n\n\t\t\t// Compute eigenvalues and eigenvectors\n\t\t\tif err = golapack.Dgehrd(n, 1, n, t, work, work.Off(n), lwork-n); err != nil {\n\t\t\t\t(*lmax)[0] = knt\n\t\t\t\t(*ninfo)[0] = (*ninfo)[0] + 1\n\t\t\t\tgoto label240\n\t\t\t}\n\t\t\tfor j = 1; j <= n-2; j++ {\n\t\t\t\tfor i = j + 2; i <= n; i++ {\n\t\t\t\t\tt.Set(i-1, j-1, zero)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Compute Schur form\n\t\t\tif info, err = golapack.Dhseqr('S', 'N', n, 1, n, t, wr, wi, dum.Matrix(1, opts), work, lwork); info != 0 || err != nil {\n\t\t\t\t(*lmax)[1] = knt\n\t\t\t\t(*ninfo)[1] = (*ninfo)[1] + 1\n\t\t\t\tgoto label240\n\t\t\t}\n\n\t\t\t// Compute eigenvectors\n\t\t\tif _, err = golapack.Dtrevc(Both, 'A', &_select, n, t, le, re, n, work); err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\n\t\t\t// Compute condition numbers\n\t\t\tif _, err = golapack.Dtrsna('B', 'A', _select, n, t, le, re, s, sep, n, work.Matrix(n, opts), &iwork); err != nil {\n\t\t\t\t(*lmax)[2] = knt\n\t\t\t\t(*ninfo)[2] = (*ninfo)[2] + 1\n\t\t\t\tgoto label240\n\t\t\t}\n\n\t\t\t// Sort eigenvalues and condition numbers lexicographically\n\t\t\t// to compare with inputs\n\t\t\twrtmp.Copy(n, wr, 1, 1)\n\t\t\twitmp.Copy(n, wi, 1, 1)\n\t\t\tstmp.Copy(n, s, 1, 1)\n\t\t\tseptmp.Copy(n, sep, 1, 1)\n\t\t\tseptmp.Scal(n, one/vmul, 1)\n\t\t\tfor i = 1; i <= n-1; i++ {\n\t\t\t\tkmin = i\n\t\t\t\tvrmin = wrtmp.Get(i - 1)\n\t\t\t\tvimin = witmp.Get(i - 1)\n\t\t\t\tfor j = i + 1; j <= n; j++ {\n\t\t\t\t\tif wrtmp.Get(j-1) < vrmin {\n\t\t\t\t\t\tkmin = j\n\t\t\t\t\t\tvrmin = wrtmp.Get(j - 1)\n\t\t\t\t\t\tvimin = witmp.Get(j - 1)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\twrtmp.Set(kmin-1, wrtmp.Get(i-1))\n\t\t\t\twitmp.Set(kmin-1, witmp.Get(i-1))\n\t\t\t\twrtmp.Set(i-1, vrmin)\n\t\t\t\twitmp.Set(i-1, vimin)\n\t\t\t\tvrmin = stmp.Get(kmin - 1)\n\t\t\t\tstmp.Set(kmin-1, stmp.Get(i-1))\n\t\t\t\tstmp.Set(i-1, vrmin)\n\t\t\t\tvrmin = septmp.Get(kmin - 1)\n\t\t\t\tseptmp.Set(kmin-1, septmp.Get(i-1))\n\t\t\t\tseptmp.Set(i-1, vrmin)\n\t\t\t}\n\n\t\t\t// Compare condition numbers for eigenvalues\n\t\t\t// taking their condition numbers into account\n\t\t\tv = math.Max(two*float64(n)*eps*tnrm, smlnum)\n\t\t\tif tnrm == zero {\n\t\t\t\tv = one\n\t\t\t}\n\t\t\tfor i = 1; i <= n; i++ {\n\t\t\t\tif v > septmp.Get(i-1) {\n\t\t\t\t\ttol = one\n\t\t\t\t} else {\n\t\t\t\t\ttol = v / septmp.Get(i-1)\n\t\t\t\t}\n\t\t\t\tif v > sepin.Get(i-1) {\n\t\t\t\t\ttolin = one\n\t\t\t\t} else {\n\t\t\t\t\ttolin = v / sepin.Get(i-1)\n\t\t\t\t}\n\t\t\t\ttol = math.Max(tol, smlnum/eps)\n\t\t\t\ttolin = math.Max(tolin, smlnum/eps)\n\t\t\t\tif eps*(sin.Get(i-1)-tolin) > stmp.Get(i-1)+tol {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t} else if sin.Get(i-1)-tolin > stmp.Get(i-1)+tol {\n\t\t\t\t\tvmax = (sin.Get(i-1) - tolin) / (stmp.Get(i-1) + tol)\n\t\t\t\t} else if sin.Get(i-1)+tolin < eps*(stmp.Get(i-1)-tol) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t} else if sin.Get(i-1)+tolin < stmp.Get(i-1)-tol {\n\t\t\t\t\tvmax = (stmp.Get(i-1) - tol) / (sin.Get(i-1) + tolin)\n\t\t\t\t} else {\n\t\t\t\t\tvmax = one\n\t\t\t\t}\n\t\t\t\tif vmax > rmax.Get(1) {\n\t\t\t\t\trmax.Set(1, vmax)\n\t\t\t\t\tif (*ninfo)[1] == 0 {\n\t\t\t\t\t\t(*lmax)[1] = knt\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Compare condition numbers for eigenvectors\n\t\t\t// taking their condition numbers into account\n\t\t\tfor i = 1; i <= n; i++ {\n\t\t\t\tif v > septmp.Get(i-1)*stmp.Get(i-1) {\n\t\t\t\t\ttol = septmp.Get(i - 1)\n\t\t\t\t} else {\n\t\t\t\t\ttol = v / stmp.Get(i-1)\n\t\t\t\t}\n\t\t\t\tif v > sepin.Get(i-1)*sin.Get(i-1) {\n\t\t\t\t\ttolin = sepin.Get(i - 1)\n\t\t\t\t} else {\n\t\t\t\t\ttolin = v / sin.Get(i-1)\n\t\t\t\t}\n\t\t\t\ttol = math.Max(tol, smlnum/eps)\n\t\t\t\ttolin = math.Max(tolin, smlnum/eps)\n\t\t\t\tif eps*(sepin.Get(i-1)-tolin) > septmp.Get(i-1)+tol {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t} else if sepin.Get(i-1)-tolin > septmp.Get(i-1)+tol {\n\t\t\t\t\tvmax = (sepin.Get(i-1) - tolin) / (septmp.Get(i-1) + tol)\n\t\t\t\t} else if sepin.Get(i-1)+tolin < eps*(septmp.Get(i-1)-tol) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t} else if sepin.Get(i-1)+tolin < septmp.Get(i-1)-tol {\n\t\t\t\t\tvmax = (septmp.Get(i-1) - tol) / (sepin.Get(i-1) + tolin)\n\t\t\t\t} else {\n\t\t\t\t\tvmax = one\n\t\t\t\t}\n\t\t\t\tif vmax > rmax.Get(1) {\n\t\t\t\t\trmax.Set(1, vmax)\n\t\t\t\t\tif (*ninfo)[1] == 0 {\n\t\t\t\t\t\t(*lmax)[1] = knt\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Compare condition numbers for eigenvalues\n\t\t\t// without taking their condition numbers into account\n\t\t\tfor i = 1; i <= n; i++ {\n\t\t\t\tif sin.Get(i-1) <= float64(2*n)*eps && stmp.Get(i-1) <= float64(2*n)*eps {\n\t\t\t\t\tvmax = one\n\t\t\t\t} else if eps*sin.Get(i-1) > stmp.Get(i-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t} else if sin.Get(i-1) > stmp.Get(i-1) {\n\t\t\t\t\tvmax = sin.Get(i-1) / stmp.Get(i-1)\n\t\t\t\t} else if sin.Get(i-1) < eps*stmp.Get(i-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t} else if sin.Get(i-1) < stmp.Get(i-1) {\n\t\t\t\t\tvmax = stmp.Get(i-1) / sin.Get(i-1)\n\t\t\t\t} else {\n\t\t\t\t\tvmax = one\n\t\t\t\t}\n\t\t\t\tif vmax > rmax.Get(2) {\n\t\t\t\t\trmax.Set(2, vmax)\n\t\t\t\t\tif (*ninfo)[2] == 0 {\n\t\t\t\t\t\t(*lmax)[2] = knt\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Compare condition numbers for eigenvectors\n\t\t\t// without taking their condition numbers into account\n\t\t\tfor i = 1; i <= n; i++ {\n\t\t\t\tif sepin.Get(i-1) <= v && septmp.Get(i-1) <= v {\n\t\t\t\t\tvmax = one\n\t\t\t\t} else if eps*sepin.Get(i-1) > septmp.Get(i-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t} else if sepin.Get(i-1) > septmp.Get(i-1) {\n\t\t\t\t\tvmax = sepin.Get(i-1) / septmp.Get(i-1)\n\t\t\t\t} else if sepin.Get(i-1) < eps*septmp.Get(i-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t} else if sepin.Get(i-1) < septmp.Get(i-1) {\n\t\t\t\t\tvmax = septmp.Get(i-1) / sepin.Get(i-1)\n\t\t\t\t} else {\n\t\t\t\t\tvmax = one\n\t\t\t\t}\n\t\t\t\tif vmax > rmax.Get(2) {\n\t\t\t\t\trmax.Set(2, vmax)\n\t\t\t\t\tif (*ninfo)[2] == 0 {\n\t\t\t\t\t\t(*lmax)[2] = knt\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Compute eigenvalue condition numbers only and compare\n\t\t\tvmax = zero\n\t\t\tdum.Set(0, -one)\n\t\t\tstmp.Copy(n, dum, 0, 1)\n\t\t\tseptmp.Copy(n, dum, 0, 1)\n\t\t\tif _, err = golapack.Dtrsna('E', 'A', _select, n, t, le, re, stmp, septmp, n, work.Matrix(n, opts), &iwork); err != nil {\n\t\t\t\t(*lmax)[2] = knt\n\t\t\t\t(*ninfo)[2] = (*ninfo)[2] + 1\n\t\t\t\tgoto label240\n\t\t\t}\n\t\t\tfor i = 1; i <= n; i++ {\n\t\t\t\tif stmp.Get(i-1) != s.Get(i-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t\tif septmp.Get(i-1) != dum.Get(0) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Compute eigenvector condition numbers only and compare\n\t\t\tstmp.Copy(n, dum, 0, 1)\n\t\t\tseptmp.Copy(n, dum, 0, 1)\n\t\t\tif _, err = golapack.Dtrsna('V', 'A', _select, n, t, le, re, stmp, septmp, n, work.Matrix(n, opts), &iwork); err != nil {\n\t\t\t\t(*lmax)[2] = knt\n\t\t\t\t(*ninfo)[2] = (*ninfo)[2] + 1\n\t\t\t\tgoto label240\n\t\t\t}\n\t\t\tfor i = 1; i <= n; i++ {\n\t\t\t\tif stmp.Get(i-1) != dum.Get(0) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t\tif septmp.Get(i-1) != sep.Get(i-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Compute all condition numbers using SELECT and compare\n\t\t\tfor i = 1; i <= n; i++ {\n\t\t\t\t_select[i-1] = true\n\t\t\t}\n\t\t\tstmp.Copy(n, dum, 0, 1)\n\t\t\tseptmp.Copy(n, dum, 0, 1)\n\t\t\tif _, err = golapack.Dtrsna('B', 'S', _select, n, t, le, re, stmp, septmp, n, work.Matrix(n, opts), &iwork); err != nil {\n\t\t\t\t(*lmax)[2] = knt\n\t\t\t\t(*ninfo)[2] = (*ninfo)[2] + 1\n\t\t\t\tgoto label240\n\t\t\t}\n\t\t\tfor i = 1; i <= n; i++ {\n\t\t\t\tif septmp.Get(i-1) != sep.Get(i-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t\tif stmp.Get(i-1) != s.Get(i-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Compute eigenvalue condition numbers using SELECT and compare\n\t\t\tstmp.Copy(n, dum, 0, 1)\n\t\t\tseptmp.Copy(n, dum, 0, 1)\n\t\t\tif _, err = golapack.Dtrsna('E', 'S', _select, n, t, le, re, stmp, septmp, n, work.Matrix(n, opts), &iwork); err != nil {\n\t\t\t\t(*lmax)[2] = knt\n\t\t\t\t(*ninfo)[2] = (*ninfo)[2] + 1\n\t\t\t\tgoto label240\n\t\t\t}\n\t\t\tfor i = 1; i <= n; i++ {\n\t\t\t\tif stmp.Get(i-1) != s.Get(i-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t\tif septmp.Get(i-1) != dum.Get(0) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Compute eigenvector condition numbers using SELECT and compare\n\t\t\tstmp.Copy(n, dum, 0, 1)\n\t\t\tseptmp.Copy(n, dum, 0, 1)\n\t\t\tif _, err = golapack.Dtrsna('V', 'S', _select, n, t, le, re, stmp, septmp, n, work.Matrix(n, opts), &iwork); err != nil {\n\t\t\t\t(*lmax)[2] = knt\n\t\t\t\t(*ninfo)[2] = (*ninfo)[2] + 1\n\t\t\t\tgoto label240\n\t\t\t}\n\t\t\tfor i = 1; i <= n; i++ {\n\t\t\t\tif stmp.Get(i-1) != dum.Get(0) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t\tif septmp.Get(i-1) != sep.Get(i-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t}\n\t\t\tif vmax > rmax.Get(0) {\n\t\t\t\trmax.Set(0, vmax)\n\t\t\t\tif (*ninfo)[0] == 0 {\n\t\t\t\t\t(*lmax)[0] = knt\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Select first real and first complex eigenvalue\n\t\t\tif wi.Get(0) == zero {\n\t\t\t\tlcmp[0] = 1\n\t\t\t\tifnd = 0\n\t\t\t\tfor i = 2; i <= n; i++ {\n\t\t\t\t\tif ifnd == 1 || wi.Get(i-1) == zero {\n\t\t\t\t\t\t_select[i-1] = false\n\t\t\t\t\t} else {\n\t\t\t\t\t\tifnd = 1\n\t\t\t\t\t\tlcmp[1] = i\n\t\t\t\t\t\tlcmp[2] = i + 1\n\t\t\t\t\t\tre.Off(0, 1).Vector().Copy(n, re.Off(0, i-1).Vector(), 1, 1)\n\t\t\t\t\t\tre.Off(0, 2).Vector().Copy(n, re.Off(0, i).Vector(), 1, 1)\n\t\t\t\t\t\tle.Off(0, 1).Vector().Copy(n, le.Off(0, i-1).Vector(), 1, 1)\n\t\t\t\t\t\tle.Off(0, 2).Vector().Copy(n, le.Off(0, i).Vector(), 1, 1)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tif ifnd == 0 {\n\t\t\t\t\ticmp = 1\n\t\t\t\t} else {\n\t\t\t\t\ticmp = 3\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tlcmp[0] = 1\n\t\t\t\tlcmp[1] = 2\n\t\t\t\tifnd = 0\n\t\t\t\tfor i = 3; i <= n; i++ {\n\t\t\t\t\tif ifnd == 1 || wi.Get(i-1) != zero {\n\t\t\t\t\t\t_select[i-1] = false\n\t\t\t\t\t} else {\n\t\t\t\t\t\tlcmp[2] = i\n\t\t\t\t\t\tifnd = 1\n\t\t\t\t\t\tre.Off(0, 2).Vector().Copy(n, re.Off(0, i-1).Vector(), 1, 1)\n\t\t\t\t\t\tle.Off(0, 2).Vector().Copy(n, le.Off(0, i-1).Vector(), 1, 1)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tif ifnd == 0 {\n\t\t\t\t\ticmp = 2\n\t\t\t\t} else {\n\t\t\t\t\ticmp = 3\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Compute all selected condition numbers\n\t\t\tstmp.Copy(icmp, dum, 0, 1)\n\t\t\tseptmp.Copy(icmp, dum, 0, 1)\n\t\t\tif _, err = golapack.Dtrsna('B', 'S', _select, n, t, le, re, stmp, septmp, n, work.Matrix(n, opts), &iwork); err != nil {\n\t\t\t\t(*lmax)[2] = knt\n\t\t\t\t(*ninfo)[2] = (*ninfo)[2] + 1\n\t\t\t\tgoto label240\n\t\t\t}\n\t\t\tfor i = 1; i <= icmp; i++ {\n\t\t\t\tj = lcmp[i-1]\n\t\t\t\tif septmp.Get(i-1) != sep.Get(j-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t\tif stmp.Get(i-1) != s.Get(j-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Compute selected eigenvalue condition numbers\n\t\t\tstmp.Copy(icmp, dum, 0, 1)\n\t\t\tseptmp.Copy(icmp, dum, 0, 1)\n\t\t\tif _, err = golapack.Dtrsna('E', 'S', _select, n, t, le, re, stmp, septmp, n, work.Matrix(n, opts), &iwork); err != nil {\n\t\t\t\t(*lmax)[2] = knt\n\t\t\t\t(*ninfo)[2] = (*ninfo)[2] + 1\n\t\t\t\tgoto label240\n\t\t\t}\n\t\t\tfor i = 1; i <= icmp; i++ {\n\t\t\t\tj = lcmp[i-1]\n\t\t\t\tif stmp.Get(i-1) != s.Get(j-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t\tif septmp.Get(i-1) != dum.Get(0) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Compute selected eigenvector condition numbers\n\t\t\tstmp.Copy(icmp, dum, 0, 1)\n\t\t\tseptmp.Copy(icmp, dum, 0, 1)\n\t\t\tif _, err = golapack.Dtrsna('V', 'S', _select, n, t, le, re, stmp, septmp, n, work.Matrix(n, opts), &iwork); err != nil {\n\t\t\t\t(*lmax)[2] = knt\n\t\t\t\t(*ninfo)[2] = (*ninfo)[2] + 1\n\t\t\t\tgoto label240\n\t\t\t}\n\t\t\tfor i = 1; i <= icmp; i++ {\n\t\t\t\tj = lcmp[i-1]\n\t\t\t\tif stmp.Get(i-1) != dum.Get(0) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t\tif septmp.Get(i-1) != sep.Get(j-1) {\n\t\t\t\t\tvmax = one / eps\n\t\t\t\t}\n\t\t\t}\n\t\t\tif vmax > rmax.Get(0) {\n\t\t\t\trmax.Set(0, vmax)\n\t\t\t\tif (*ninfo)[0] == 0 {\n\t\t\t\t\t(*lmax)[0] = knt\n\t\t\t\t}\n\t\t\t}\n\t\tlabel240:\n\t\t}\n\t}\n\n\treturn\n}", "func DIVSD(mx, x operand.Op) { ctx.DIVSD(mx, x) }", "func DIVSS(mx, x operand.Op) { ctx.DIVSS(mx, x) }", "func row(k int) int { return k % 4 }", "func DIVB(mr operand.Op) { ctx.DIVB(mr) }", "func opUI64Div(prgrm *CXProgram) {\n\texpr := prgrm.GetExpr()\n\tfp := prgrm.GetFramePointer()\n\n\toutV0 := ReadUI64(fp, expr.Inputs[0]) / ReadUI64(fp, expr.Inputs[1])\n\tWriteUI64(GetFinalOffset(fp, expr.Outputs[0]), outV0)\n}", "func basic5(fname string) error {\n\tf, err := groot.Open(fname)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not open ROOT file: %w\", err)\n\t}\n\tdefer f.Close()\n\n\to, err := f.Get(\"Events\")\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not retrieve tree: %w\", err)\n\t}\n\n\ttree := o.(rtree.Tree)\n\tfmt.Printf(\"tree: %d entries\\n\", tree.Entries())\n\n\tvar (\n\t\thmet = hbook.NewH1D(100, 0, 2000)\n\t)\n\n\tsc, err := rtree.NewTreeScannerVars(tree,\n\t\trtree.ScanVar{Name: \"Muon_pt\"},\n\t\trtree.ScanVar{Name: \"Muon_eta\"},\n\t\trtree.ScanVar{Name: \"Muon_phi\"},\n\t\trtree.ScanVar{Name: \"Muon_mass\"},\n\t\trtree.ScanVar{Name: \"Muon_charge\"},\n\t\trtree.ScanVar{Name: \"MET_sumet\"},\n\t)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not create scanner: %w\", err)\n\t}\n\tdefer sc.Close()\n\n\tfor sc.Next() {\n\t\tvar (\n\t\t\tmuPt []float32\n\t\t\tmuEta []float32\n\t\t\tmuPhi []float32\n\t\t\tmuMass []float32\n\t\t\tmuCharge []int32\n\t\t\tmet float32\n\t\t)\n\t\terr := sc.Scan(&muPt, &muEta, &muPhi, &muMass, &muCharge, &met)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error during scan: %w\", err)\n\t\t}\n\n\t\tnmuons := len(muPt)\n\t\tif nmuons < 2 {\n\t\t\tcontinue\n\t\t}\n\n\t\tmasses := make([]float64, 0, nmuons)\n\t\tcombs := combin.Combinations(nmuons, 2)\n\t\tfor _, c := range combs {\n\t\t\ti1 := c[0]\n\t\t\ti2 := c[1]\n\t\t\tcharge1 := muCharge[i1]\n\t\t\tcharge2 := muCharge[i2]\n\t\t\tif charge1 == charge2 {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tp1 := fmom.NewPtEtaPhiM(float64(muPt[i1]), float64(muEta[i1]), float64(muPhi[i1]), float64(muMass[i1]))\n\t\t\tp2 := fmom.NewPtEtaPhiM(float64(muPt[i2]), float64(muEta[i2]), float64(muPhi[i2]), float64(muMass[i2]))\n\t\t\tmass := fmom.InvMass(&p1, &p2)\n\n\t\t\tif 60 < mass && mass < 100 {\n\t\t\t\tmasses = append(masses, mass)\n\t\t\t}\n\t\t}\n\n\t\tif len(masses) > 0 {\n\t\t\thmet.Fill(float64(met), 1)\n\t\t}\n\t}\n\n\tif err := sc.Err(); err != nil {\n\t\treturn fmt.Errorf(\"could not scan whole file: %w\", err)\n\t}\n\n\tfmt.Printf(\"hmet: %v\\n\", hmet.SumW())\n\n\tp := hplot.New()\n\tp.X.Label.Text = \"MET [GeV]\"\n\tp.Y.Label.Text = \"Nevts\"\n\n\tp.Add(hplot.NewH1D(hmet))\n\n\terr = p.Save(10*vg.Centimeter, -1, \"05-basic.png\")\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not save plot: %w\", err)\n\t}\n\n\treturn nil\n}", "func elfProg() int {\n\t// value one = 10,551,381\n\t// valueOne := 2\n\t// // 19 is current prog counter\n\t// valueOne = valueOne * valueOne * 19 * 11\n\t// // valueOne = valueOne * 19\n\t// // valueOne = valueOne * 11\n\t// // 836\n\n\t// // twenty two is current prog counter\n\t// valueTwo := 6*22 + 13\n\t// valueOne += valueTwo\n\t// // 981\n\n\t// // Here, an instruction is skipped??? Prog counter incremented by 1. I'll write the instruction down in code.\n\t// // seti 0 -> 3 // Initial value gets set to the program counter??? While loop?\n\n\t// // 27 is the current prog counter\n\t// valueTwo = 27\n\t// // 28 is the current prog counter\n\t// valueTwo = 28 * valueTwo\n\t// // 29 is the current prog counter\n\t// valueTwo = 29 + valueTwo\n\t// // 30 is the current prog counter\n\t// valueTwo = 30 * valueTwo\n\t// valueTwo = 14 * valueTwo\n\t// // 32 is the current prog counter\n\t// valueTwo = 32 * valueTwo\n\t// valueOne += valueTwo\n\n\tincrementor := 10551381\n\treturnValue := 0\n\n\t// Ok, it looks like this little snippet just adds all the factors of the \"incrementor\" to the return value.\n\t// Maybe I should look at all the values i needed to multiply to get this awful number LOL.\n\t// Nvm, i'm gonna cheat and use an online calculator\n\t// 1 + 3 + 71 + 213 + 49537 + 148611 + 3517127 + 10551381 = 14266944\n\tfor i := 1; i <= incrementor; i++ {\n\t\tfor j := 1; j < incrementor; j++ {\n\t\t\tintermediate := i * j\n\t\t\tif intermediate == incrementor {\n\t\t\t\t// addr 4 0 0 (increment the return value by variable 'i' which is register 4.)\n\t\t\t\treturnValue += i\n\t\t\t}\n\t\t}\n\t}\n\n\treturn returnValue\n\n\t// // Here, prog counter gets set back to 0, then incremented to 1\n\t// valueThree := 1\n\t// valueFour := 1\n\t// valueTwo := valueThree * valueFour\n\t// if incrementor == valueTwo {\n\t// \tvalueTwo = 1\n\t// } else {\n\t// \tvalueTwo = 0\n\t// }\n\n\t// IF VALUE 2 equals 1, WE ARE GOING TO SKIP THE NEXT INSTRUCTION.\n\t// addi 3 1 3 // UH WHAT, SKIP THE NEXT INSTRUCTION AGAIN???\n\n\t// // If value 2 equaled 0, this instruction gets skipped b/c the \"addi 3 1 3\" instruction was run\n\t// if valueTwo == 1 {\n\t// \t// Yay increment the initial value\n\t// \treturnValue += valueFour\n\t// }\n\n\t// valueFour++ // Fuck, this was actually addi, not addr. Just increment one always.\n\n\t// // value one is a really big number\n\t// if valueFour > incrementor {\n\t// \tvalueTwo = 1\n\t// } else {\n\t// \tvalueTwo = 0\n\t// }\n\n\t// IF VALUE 2 equals 1, WE ARE GOING TO SKIP THE NEXT INSTRUCTION.\n\t// seti 2 3 // UH WHAT, go back to instruction 2, then 3??? While loop????\n\t// Ok, here is the first while loop from instruction 3 -> instruction 11\n\n\t// valueThree++\n\t// // Lol, now value three has to be bigger than value one\n\t// if valueThree > incrementor {\n\t// \tvalueTwo = 1\n\t// } else {\n\t// \tvalueTwo = 0\n\t// }\n\n\t// IF VALUE 2 equals 1, WE ARE GOING TO SKIP THE NEXT INSTRUCTION.\n\t// seti 1 3, go back to instruction 1, then 2. Does this mean i have to fucking repeat the first while loop?\n\t// Is this a double for loop???\n\n\t// mulr 3 3 3\n\t// ^^^ I'm asuming this breaks the program. Back at instruction 16.\n\t// At the beginning, instruction 16 got incremented immediately. Now register 3 is about to get fucked up with a value of 256.\n}", "func c1(n int) int { return n - WIDTH - 1 }", "func population_count_divideconquer(n uint) (count uint) {\n\ti := uint32(n) - ((uint32(n) >> 1) & 0x55555555)\n i = (i & 0x33333333) + ((i >> 2) & 0x33333333)\n return uint((((i + (i >> 4)) & 0x0f0f0f0f) * 0x01010101) >> 24)\n}", "func h(n int) int {\n var a int = n % 2\n if a == 1 {\n return n * 3 + 1\n }\n return n / 2\n}", "func d5getBranches(node *d5nodeT, branch *d5branchT, parVars *d5partitionVarsT) {\n\t// Load the branch buffer\n\tfor index := 0; index < d5maxNodes; index++ {\n\t\tparVars.branchBuf[index] = node.branch[index]\n\t}\n\tparVars.branchBuf[d5maxNodes] = *branch\n\tparVars.branchCount = d5maxNodes + 1\n\n\t// Calculate rect containing all in the set\n\tparVars.coverSplit = parVars.branchBuf[0].rect\n\tfor index := 1; index < d5maxNodes+1; index++ {\n\t\tparVars.coverSplit = d5combineRect(&parVars.coverSplit, &parVars.branchBuf[index].rect)\n\t}\n\tparVars.coverSplitArea = d5calcRectVolume(&parVars.coverSplit)\n}", "func answerQuery(l int, r int) int {\n\t// Return the answer for this query modulo 1000000007.\n\n\t// get letter frequencies in range [l-1:r]\n\tfreq := inp.freq[r-1] // range [0:r-1]\n\tif l > 1 { // freq[0] is all zero\n\t\tcntL := inp.freq[l-2]\n\t\tfor c := 0; c < 26; c++ {\n\t\t\tfreq[c] = freq[c] - cntL[c]\n\t\t}\n\t}\n\n\t// IDEA: half the string and calculate permutations of multisets\n\tperms := uint64(1)\n\tn := 0\n\todds := 0\n\tfor _, v := range freq {\n\t\tif v == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\tif v%2 == 0 {\n\t\t\tn += v / 2\n\t\t\tif p, exists := inp.fact[v/2]; exists {\n\t\t\t\tperms = (perms * p) % 1000000007\n\t\t\t} else {\n\t\t\t\tpanic(fmt.Errorf(\"%d factorial not found\", v/2))\n\t\t\t}\n\t\t} else {\n\t\t\todds++\n\t\t\tn += (v - 1) / 2 // make it even\n\t\t\tif p, exists := inp.fact[(v-1)/2]; exists {\n\t\t\t\tperms = (perms * p) % 1000000007\n\t\t\t} else {\n\t\t\t\tpanic(fmt.Errorf(\"%d factorial not found\", (v-1)/2))\n\t\t\t}\n\t\t}\n\t}\n\n\t// ans:=(n!/perms)%1000000007 = ((n! % 1000000007)*(inverse(perms)%1000000007)) % 1000000007\n\tm := new(big.Int).SetUint64(1000000007)\n\tans := new(big.Int).SetUint64(inp.fact[n])\n\ty := new(big.Int).SetUint64(perms)\n\ty.ModInverse(y, m)\n\tans.Mul(ans, y)\n\tans.Mod(ans, m)\n\n\tif odds > 0 {\n\t\to := new(big.Int).SetInt64(int64(odds))\n\t\tm = new(big.Int).SetInt64(1000000007)\n\t\tans.Mul(ans, o) // ans*=odd, always can put one char in the middle\n\t\tans.Mod(ans, m) // ans%=1000000007\n\t}\n\n\treturn int(ans.Int64())\n}", "func DIVL(mr operand.Op) { ctx.DIVL(mr) }", "func split(sum int) (x, y int) {\n\tx = sum * 4 / 9\n\ty = sum - x\n\t// naked return: return current values of the result, i.e. x, y\n\t// harms readability, and should be used only in short functions\n\treturn\n}", "func numWaterBottles(num int, div int) int {\r\n ans := num\r\n for num >= div {\r\n quo := num / div\r\n rem := num % div\r\n num = quo + rem\r\n ans += quo\r\n }\r\n return ans\r\n}", "func DIVPS(mx, x operand.Op) { ctx.DIVPS(mx, x) }", "func walkDivMod(n *ir.BinaryExpr, init *ir.Nodes) ir.Node {\n\tn.X = walkExpr(n.X, init)\n\tn.Y = walkExpr(n.Y, init)\n\n\t// rewrite complex div into function call.\n\tet := n.X.Type().Kind()\n\n\tif types.IsComplex[et] && n.Op() == ir.ODIV {\n\t\tt := n.Type()\n\t\tcall := mkcall(\"complex128div\", types.Types[types.TCOMPLEX128], init, typecheck.Conv(n.X, types.Types[types.TCOMPLEX128]), typecheck.Conv(n.Y, types.Types[types.TCOMPLEX128]))\n\t\treturn typecheck.Conv(call, t)\n\t}\n\n\t// Nothing to do for float divisions.\n\tif types.IsFloat[et] {\n\t\treturn n\n\t}\n\n\t// rewrite 64-bit div and mod on 32-bit architectures.\n\t// TODO: Remove this code once we can introduce\n\t// runtime calls late in SSA processing.\n\tif types.RegSize < 8 && (et == types.TINT64 || et == types.TUINT64) {\n\t\tif n.Y.Op() == ir.OLITERAL {\n\t\t\t// Leave div/mod by constant powers of 2 or small 16-bit constants.\n\t\t\t// The SSA backend will handle those.\n\t\t\tswitch et {\n\t\t\tcase types.TINT64:\n\t\t\t\tc := ir.Int64Val(n.Y)\n\t\t\t\tif c < 0 {\n\t\t\t\t\tc = -c\n\t\t\t\t}\n\t\t\t\tif c != 0 && c&(c-1) == 0 {\n\t\t\t\t\treturn n\n\t\t\t\t}\n\t\t\tcase types.TUINT64:\n\t\t\t\tc := ir.Uint64Val(n.Y)\n\t\t\t\tif c < 1<<16 {\n\t\t\t\t\treturn n\n\t\t\t\t}\n\t\t\t\tif c != 0 && c&(c-1) == 0 {\n\t\t\t\t\treturn n\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tvar fn string\n\t\tif et == types.TINT64 {\n\t\t\tfn = \"int64\"\n\t\t} else {\n\t\t\tfn = \"uint64\"\n\t\t}\n\t\tif n.Op() == ir.ODIV {\n\t\t\tfn += \"div\"\n\t\t} else {\n\t\t\tfn += \"mod\"\n\t\t}\n\t\treturn mkcall(fn, n.Type(), init, typecheck.Conv(n.X, types.Types[et]), typecheck.Conv(n.Y, types.Types[et]))\n\t}\n\treturn n\n}", "func pack5(src []uint64) uint64 {\n\t_ = src[4] // eliminate multiple bounds checks\n\treturn 11<<60 |\n\t\tsrc[0] |\n\t\tsrc[1]<<12 |\n\t\tsrc[2]<<24 |\n\t\tsrc[3]<<36 |\n\t\tsrc[4]<<48\n}", "func division(x, y uint) ([]uint, []uint, []uint) {\n\t// Digits before decimal point\n\tinteg := []uint{0}\n\n\t// Digits after decimal point including any recurrence\n\tfract := []uint{}\n\n\t// Track state of caculation to detect recurrence\n\tstate := make(map[uint]*struct{})\n\tindex := make(map[uint]int)\n\n\t// First, see if we have digits before the decimal point.\n\tif x/y != 0 {\n\t\tinteg = euler.UintToDigits(x / y)\n\t\tx = x % y\n\t}\n\n\t// Compute the fractional part. We will accumulate digits\n\t// in a slice and keep track of the leading digits.\n\n\t// Increasing here sets up to calculate the first digit, so\n\t// does not add a leading zero.\n\tif x > 0 && x/y == 0 {\n\t\tx *= 10\n\t}\n\n\t// Index of the current digit in the calculation\n\ti := 0\n\n\t// If further increases are needed, add more leading zeros\n\tfor x > 0 && x/y == 0 {\n\t\tx *= 10\n\t\tfract = append(fract, 0)\n\t\ti++\n\t}\n\n\tfor {\n\t\t// Increase x as needed and add zero digits\n\t\tfor x > 0 && x/y == 0 {\n\t\t\tx *= 10\n\t\t\tfract = append(fract, 0)\n\t\t}\n\n\t\t// Check if we've returned to a previous state, where the\n\t\t// current divisor has been seen before. Return the point\n\t\t// where the recurrence began. If we don't break out here\n\t\t// this would recur forever.\n\t\tif state[x] != nil {\n\t\t\t// Calculate the point the recurrence starts and\n\t\t\t// return slices for each part.\n\t\t\tr := index[x]\n\t\t\treturn integ, fract[:r], fract[r:]\n\t\t}\n\n\t\t// Record the current state\n\t\tstate[x] = &struct{}{}\n\t\tindex[x] = i\n\n\t\t// Div mod\n\t\tn := x / y\n\t\tm := x % y\n\n\t\t// Save digit\n\t\tfract = append(fract, n)\n\n\t\t// If no remainder, we're done\n\t\tif m == 0 {\n\t\t\tbreak\n\t\t}\n\n\t\t// Continue the calculation with the remainder * 10\n\t\tx = m * 10\n\n\t\ti++\n\t}\n\n\t// No recurrence was found, return digits.\n\treturn integ, fract, []uint{}\n}", "func Divide(ring bhs.Ring) (bhs.NodeID, uint64, uint64) {\n\tconst cautiousWalk = true\n\tblackhole := make(chan bhs.NodeID, 1)\n\toks := make(chan bool, 2)\n\tmoves := make(chan uint64, 2)\n\tringSize := bhs.NodeID(len(ring)) // logically wrong, but needed for type correctness)\n\n\tdirections := [2]bhs.Direction{bhs.Left, bhs.Right}\n\tfor i := 0; i < len(directions); i++ {\n\t\tgo func(direction bhs.Direction, oks chan<- bool, blackhole chan<- bhs.NodeID, moves chan<- uint64) {\n\t\t\tagent := bhs.NewAgent(direction, ring, cautiousWalk)\n\t\t\tagent.ActAsSmall = false // for update catching\n\t\t\tagent.UnexploredSet = [2]bhs.NodeID{1, ringSize - 1}\n\n\t\t\tfor agent.UnexploredSet[0] != agent.UnexploredSet[1] {\n\t\t\t\tdestination := equallyDivideUnexploredSet(agent.Direction, agent.UnexploredSet)\n\t\t\t\tok, updateFound := agent.MoveUntil(agent.Direction, destination)\n\t\t\t\tif !ok {\n\t\t\t\t\tmoves <- agent.Moves\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tif updateFound {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tif agent.UnexploredSet[0] != agent.UnexploredSet[1] { // if other agent falls in the black hole, update useless\n\t\t\t\t\tagent.LeaveUpdateDivide()\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tagent.MoveUntil(bhs.GetOppositeDirection(agent.Direction), 0) // go to homebase\n\t\t\tblackhole <- agent.UnexploredSet[0]\n\t\t\tmoves <- agent.Moves\n\t\t}(directions[i], oks, blackhole, moves)\n\t}\n\n\tmovesAgent1, movesAgent2 := <-moves, <-moves\n\treturn <-blackhole, movesAgent1 + movesAgent2, helpers.MaxUint64(movesAgent1, movesAgent2)\n}", "func answerQuery(l int32, r int32) int32 {\n // Return the answer for this query modulo 1000000007.\n letters := make([]int32, 26)\n unique := int32(0)\n for i := l - 1; i < r; i++ {\n if letters[S[i] - 97] == int32(0){\n unique++\n }\n letters[S[i] - 97]++\n }\n\n if unique == int32(1){\n return int32(1)\n }\n\n middles := int64(0)\n substringLength := int64(0)\n duplicates := int64(1)\n result := int64(1)\n for _, n := range letters {\n if n == 0 {\n continue\n }\n\n if n == 1 {\n middles++\n continue\n }\n\n if n % 2 == 1 {\n middles++\n }\n\n for i := int64(0); int32(i) < n / 2; i++ {\n substringLength++\n result *= substringLength\n result %= M\n duplicates *= i + 1\n duplicates %= M\n }\n }\n\n duplicates = expBySquaring(duplicates, M - 2)\n result *= duplicates\n result %= M\n\n if result == 0 {\n return int32(middles % M)\n }\n\n if middles > 0 {\n result *= middles\n result %= M\n }\n\n return int32(result)\n}", "func getDivisibleSize(totalSizes []uint64) (result uint64) {\n\tgcd := func(x, y uint64) uint64 {\n\t\tfor y != 0 {\n\t\t\tx, y = y, x%y\n\t\t}\n\t\treturn x\n\t}\n\tresult = totalSizes[0]\n\tfor i := 1; i < len(totalSizes); i++ {\n\t\tresult = gcd(result, totalSizes[i])\n\t}\n\treturn result\n}", "func main() {\n\t// 2^qN is pi/2\n\tfor i := 0; i < 8; i++ {\n\t\tr := isin_S3(int32(i << (qN -1)))\n\t\tfmt.Printf(\"%d %x %f\\n\", 45 * i, r, float32(r) / float32(1 << qA))\n\t}\n}", "func main() {\n\tn := average(43, 56, 87, 12, 45, 57)\n\tfmt.Println(n)\n}", "func main() {\n\tn := average(43, 56, 87, 12, 45, 57)\n\tfmt.Println(n)\n}", "func inner(r0, r1, r2, r3, r4 int) (int, int, int, int, int) {\n\tfor {\n\t\tr1 = r2 & 255\n\t\tr3 += r1\n\t\tr3 &= 16777215\n\t\tr3 *= 65899\n\t\tr3 &= 16777215\n\t\tif 256 > r2 {\n\t\t\treturn r0, r1, r2, r3, r4\n\t\t}\n\t\tr2 = r2 / 256\n\t}\n}", "func xx512x1(inner_512 []byte) [20]byte {\n\touter_512 := sha512.Sum512(inner_512)\n\treturn sha1.Sum(outer_512[:])\n}", "func div(d, r time.Duration) int64 {\n\tif m := d % r; m > 0 {\n\t\treturn int64(d/r + 1)\n\t}\n\treturn int64(d / r)\n}", "func div(dividend, divisor int) int {\r\n\tif dividend < divisor {\r\n\t\treturn 0\r\n\t}\r\n\tcount := 1\r\n\ttmp := divisor\r\n\tfor tmp + tmp < dividend {\r\n\t\tcount += count\r\n\t\ttmp += tmp\r\n\t}\r\n\treturn count + div(dividend - tmp, divisor)\r\n}", "func main() {\n\n\tdivisorGreater(2)\n\tdivisorGreater(1)\n\tdivisorGreater(8128)\n\tdivisorGreater(0)\n\n\t/*\n\trun_btree()\n\n\trun_binsearch(binsearch_inplace)\n\trun_binsearch(binsearch)\n\n\trun_qsort()\n\t*/\n}", "func rho(x uint64) uint8 {\n\tvar i uint8\n\tfor i = 0; i < 62 && x&1 == 0; i++ {\n\t\tx >>= 1\n\t}\n\treturn i + 1\n}", "func minOperations1827(nums []int) int {\n\tres := 0\n\tfor i := 1; i < len(nums); i++ {\n\t\ta := nums[i-1]\n\t\tb := nums[i]\n\t\tif a >= b {\n\t\t\tres += a + 1 - b\n\t\t\tnums[i] = a + 1\n\t\t}\n\t}\n\treturn res\n}", "func Exercise5() {\n\tfmt.Println()\n\tfmt.Println(\"--- EXERCISE: Manipulate a Counter ---\")\n\n\t// DO NOT CHANGE THE CODE BELOW\n\tvar counter int\n\n\tcounter++\n\tcounter--\n\tcounter += 5\n\tcounter *= 10\n\tcounter /= 5\n\n\t// TYPE YOUR CODE HERE\n\n\t// DO NOT CHANGE THE CODE BELOW\n\tfmt.Println(counter)\n}", "func stbcd(context *Context) {\n x := context.opcode & 0x0F00 >> 8\n num := context.cpu.v[x]\n context.memory[context.cpu.i] = byte((num / 100) % 10)\n context.memory[context.cpu.i + 1] = byte((num / 10) % 10)\n context.memory[context.cpu.i + 2] = byte(num % 10)\n}" ]
[ "0.5891377", "0.57369494", "0.5670985", "0.56679827", "0.5648372", "0.55592567", "0.5535071", "0.5503972", "0.5502064", "0.5498154", "0.5479929", "0.54766035", "0.54746884", "0.5465032", "0.5463755", "0.5463755", "0.5463755", "0.5463755", "0.5463755", "0.5463755", "0.5463755", "0.5463755", "0.5430921", "0.5407971", "0.5404372", "0.53890914", "0.5364402", "0.535708", "0.5337804", "0.5336315", "0.5334966", "0.53130347", "0.5293259", "0.5289922", "0.52691233", "0.5259342", "0.5246375", "0.52423674", "0.5237128", "0.5219951", "0.5213146", "0.5212238", "0.52040505", "0.5195534", "0.518929", "0.51789576", "0.5173735", "0.51732177", "0.5172075", "0.5161713", "0.5160453", "0.51579136", "0.51417446", "0.5128691", "0.51191217", "0.51063365", "0.50977236", "0.50941944", "0.5088523", "0.5081313", "0.505369", "0.5053191", "0.5049946", "0.5049741", "0.5049719", "0.5040755", "0.5034063", "0.5032819", "0.5030117", "0.50267375", "0.50243294", "0.50178665", "0.5016848", "0.5014202", "0.5013161", "0.50099754", "0.50001", "0.49984053", "0.4990009", "0.49849328", "0.4984046", "0.49834165", "0.4974268", "0.49681064", "0.49654108", "0.49651608", "0.49488702", "0.4925629", "0.49249843", "0.49202782", "0.4919303", "0.4919303", "0.49144983", "0.49130926", "0.49126816", "0.49111938", "0.4910749", "0.49106848", "0.49091396", "0.4906048", "0.4904563" ]
0.0
-1
stringView returns a view of the []byte as a string. In unsafe mode, it doesn't incur allocation and copying caused by conversion. In regular safe mode, it is an allocation and copy. Usage: Always maintain a reference to v while result of this call is in use, and call keepAlive4BytesView(v) at point where done with view.
func stringView(v []byte) string { return string(v) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func bytesView(v string) []byte {\n\treturn []byte(v)\n}", "func (v ByteVec) String() string { return string([]byte(v)) }", "func (v *View) String() string {\n\tf := format.NewASCII()\n\treturn v.FormatString(f)\n}", "func (s View) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s View) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (b *SafeBuffer) String() string {\n\tb.m.RLock()\n\tdefer b.m.RUnlock()\n\treturn b.b.String()\n}", "func (s GetViewOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (v Value) String() string {\n\tif sp, ok := v.any.(stringptr); ok {\n\t\t// Inlining this code makes a huge difference.\n\t\tvar s string\n\t\thdr := (*reflect.StringHeader)(unsafe.Pointer(&s))\n\t\thdr.Data = uintptr(sp)\n\t\thdr.Len = int(v.num)\n\t\treturn s\n\t}\n\tvar buf []byte\n\treturn string(v.append(buf))\n}", "func (s DeleteViewOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s DeleteViewOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s DeleteViewVersionOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s DescribeViewOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s IPV4) String() string {\n\treturn fmt.Sprintf(\"%v.%v.%v.%v\", s[0], s[1], s[2], s[3])\n}", "func (v String) String() string {\n\treturn v.v\n}", "func StringUnsafe(b []byte) string {\n\tvar s string\n\tpb := (*reflect.SliceHeader)(unsafe.Pointer(&b))\n\tps := (*reflect.StringHeader)(unsafe.Pointer(&s))\n\tps.Data = pb.Data\n\tps.Len = pb.Len\n\treturn s\n}", "func String(b []byte) string {\n\treturn *(*string)(unsafe.Pointer(&b))\n}", "func (s DeleteViewVersionInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s CreateViewVersionOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s ViewContent) String() string {\n\treturn awsutil.Prettify(s)\n}", "func String(b []byte) (s string) {\n pbytes := (*reflect.SliceHeader)(unsafe.Pointer(&b))\n pstring := (*reflect.StringHeader)(unsafe.Pointer(&s))\n pstring.Data = pbytes.Data\n pstring.Len = pbytes.Len\n return\n}", "func (s ViewVersionSummary) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (v *Version) String() string {\r\n\treturn v.str\r\n}", "func bufferToString(buffer *bytes.Buffer, unsafePtr *bool) string {\n defer buffer.Reset()//ensure buffer is reset\n if !*unsafePtr {\n return buffer.String()\n }\n bb := buffer.Bytes()\n s := *(*string)(unsafe.Pointer(&bb))\n return s\n}", "func (s DeleteViewInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s DeleteViewInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func unsafeString(b []byte) string {\n\treturn *(*string)(unsafe.Pointer(&b))\n}", "func (ip IPv4) String() string {\n\treturn string(ip.value)\n}", "func (s UpdateViewOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func UnsafeString(b []byte) string {\n\treturn *(*string)(unsafe.Pointer(&b))\n}", "func (s CreateViewOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s CreateViewOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s GetViewInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func VVByte_to_string(m [][]byte) (*bytes.Buffer, *bytes.Buffer) {\n\n\tlg, b := loghttp.BuffLoggerUniversal(nil, nil)\n\t_ = b\n\n\tbRet := new(bytes.Buffer)\n\tbMsg := new(bytes.Buffer)\n\n\t//for i,v := range m {\n\tfor i := 0; i < len(m); i++ {\n\t\tn, err := bRet.Write(m[i])\n\t\tlg(err)\n\t\tbMsg.WriteString(\" lp\" + util.Itos(i) + \": writing \" + util.Itos(n) + \" bytes: \\n\")\n\t}\n\treturn bRet, bMsg\n}", "func (vl BytesValue) String() string {\n\treturn Buffer.BytesToHexString(vl)\n}", "func (vl BytesValue) String() string {\n\treturn Buffer.BytesToHexString(vl)\n}", "func StringBytes(b []byte) string { return *(*string)(Pointer(&b)) }", "func String(v interface{}) string {\n\treturn StringWithOptions(v, nil)\n}", "func String(b []byte) string {\n\treturn string(b)\n}", "func (s CreateViewVersionInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (v *Value) String() string {\n\tstr := C.value_get_string(v.value)\n\tdefer C.free(unsafe.Pointer(str))\n\n\treturn C.GoString(str)\n}", "func StringFromImmutableBytes(bs []byte) string {\n\tif len(bs) == 0 {\n\t\treturn \"\"\n\t}\n\treturn unsafe.String(&bs[0], len(bs))\n}", "func (s *Buffer) String() string {\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\treturn s.buffer.String()\n}", "func String(b []byte) (s string) {\n\tpbytes := (*reflect.SliceHeader)(unsafe.Pointer(&b))\n\tpstring := (*reflect.StringHeader)(unsafe.Pointer(&s))\n\tpstring.Data = pbytes.Data\n\tpstring.Len = pbytes.Len\n\treturn\n}", "func String(v string) *string { return &v }", "func String(v string) *string { return &v }", "func String(v string) *string { return &v }", "func String(v string) *string { return &v }", "func String(b []byte) (s string) {\n\treturn string(b)\n}", "func (s UpdateViewInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (ip IPv4) String() string {\n\treturn fmt.Sprintf(\"%d.%d.%d.%d\", ip[0], ip[1], ip[2], ip[3])\n}", "func BytesToString(b []byte) string { return *(*string)(unsafe.Pointer(&b)) }", "func String(v interface{}) string {\n\treturn v.(string)\n}", "func (e Bytes) String() string {\n\treturn fmt.Sprintf(\"%v\", e)\n}", "func (s DescribeViewInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (b *Buffer) String() string {\n\treturn string(b.buf)\n}", "func gostringnocopy(str *byte) string", "func (s DisassociateDefaultViewOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func String(v string) *string {\n\treturn &v\n}", "func String(v string) *string {\n\treturn &v\n}", "func String(v string) *string {\n\treturn &v\n}", "func String(v string) *string {\n\treturn &v\n}", "func String(v string) *string {\n\treturn &v\n}", "func String(v string) *string {\n\treturn &v\n}", "func String(v string) *string {\n\treturn &v\n}", "func (s ViewSummary) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (b *Bytes) String() string {\n\treturn fmt.Sprint(*b)\n}", "func (s UpdateViewContentOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func String(v string) (p *string) { return &v }", "func (b *LimitedBuffer) String() string {\n\treturn string(b.buf)\n}", "func (s CreateViewInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s CreateViewInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func String(v string) *string { return &v }", "func (v *Value) String() string {\n switch d := v.data.(type) {\n case []byte:\n return string(d)\n case string:\n return d\n default:\n if j, e := json.Marshal(v.data); e == nil {\n return string(j)\n }\n return fmt.Sprintf(\"%+v\", v.data)\n }\n}", "func byteSliceToString(b []byte) string {\n\treturn *(*string)(unsafe.Pointer(&b))\n}", "func stringIPv4(n uint32) string {\n\tip := make(net.IP, 4)\n\tbinary.BigEndian.PutUint32(ip, n)\n\treturn ip.String()\n}", "func (d *Decoder) String() string {\n\tdata := d.Bytes()\n\treturn unsafe.BytesToString(data)\n}", "func GetString(v interface{}) string {\n\tswitch result := v.(type) {\n\tcase string:\n\t\treturn result\n\tcase []byte:\n\t\treturn string(v.([]byte))\n\tcase []string:\n\t\treturn strings.Join([]string(v.([]string)), \"\\n\")\n\tdefault:\n\t\tif v != nil {\n\t\t\treturn fmt.Sprint(result)\n\t\t}\n\t}\n\treturn \"\"\n}", "func (s AssociateDefaultViewOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (v Bytes) String() string {\n\tif v >= EXABYTE {\n\t\treturn v.format(EXABYTE) + EXABYTES\n\t}\n\tif v >= PETABYTE {\n\t\treturn v.format(PETABYTE) + PETABYTES\n\t}\n\tif v >= TERABYTE {\n\t\treturn v.format(TERABYTE) + TERABYTES\n\t}\n\tif v >= GIGABYTE {\n\t\treturn v.format(GIGABYTE) + GIGABYTES\n\t}\n\tif v >= MEGABYTE {\n\t\treturn v.format(MEGABYTE) + MEGABYTES\n\t}\n\tif v >= KILOBYTE {\n\t\treturn v.format(KILOBYTE) + KILOBYTES\n\t}\n\treturn strconv.FormatUint(uint64(v), 10) + JUSTBYTES\n}", "func bytesToString(bs []byte) string {\n\treturn *(*string)(unsafe.Pointer(&bs))\n}", "func (s ListViewsOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s ListViewsOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (b *Buf) String() string { return string(b.b) }", "func (s UpdateViewContentInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func stringValue(v reflect.Value) string {\n\tif v.CanInterface() {\n\t\treturn fmt.Sprintf(\"%#v\", v.Interface())\n\t} else {\n\t\treturn v.String()\n\t}\n}", "func (v Value) String() string {\n\tif v.typ != String {\n\t\treturn v.Raw()\n\t}\n\treturn v.str\n}", "func (s ViewInputContent) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (k *Key4) String() string { return k.ipv4.String() }", "func (s BatchGetViewOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func BytesToReadOnlyString(b []byte) string {\n\treturn string(b)\n}", "func (b *Builder) String() string {\n\tp := unsafe.Pointer(&b.buf)\n\tsp := (*string)(p)\n\ts := *sp\n\t// return *(*string)(unsafe.Pointer(&b.buf))\n\treturn s\n}", "func (s DisassociateDefaultViewInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func BytesString(b []byte) string {\n\treturn *(*string)(unsafe.Pointer(&b))\n}", "func (s DBEngineVersion) String() string {\n\treturn awsutil.Prettify(s)\n}", "func ShallowString(b []byte) (s string) {\n\tpbytes := (*reflect.SliceHeader)(unsafe.Pointer(&b))\n\tpstring := (*reflect.StringHeader)(unsafe.Pointer(&s))\n\tpstring.Data = pbytes.Data\n\tpstring.Len = pbytes.Len\n\treturn\n}", "func AsStr(v interface{}) string {\n\ts, err := elstr.ToString(v)\n\tif err != nil {\n\t\treturn \"\"\n\t} else {\n\t\treturn s\n\t}\n}", "func bytesToString(bytes []byte) (s string) {\n\tslice := (*reflect.SliceHeader)(unsafe.Pointer(&bytes))\n\tstr := (*reflect.StringHeader)(unsafe.Pointer(&s))\n\tstr.Data = slice.Data\n\tstr.Len = slice.Len\n\treturn s\n}", "func bytesToString(bytes []byte) (s string) {\n\tslice := (*reflect.SliceHeader)(unsafe.Pointer(&bytes))\n\tstr := (*reflect.StringHeader)(unsafe.Pointer(&s))\n\tstr.Data = slice.Data\n\tstr.Len = slice.Len\n\treturn s\n}", "func bytesToString(value []byte) string {\n\tn := bytes.IndexByte(value, 0)\n\tif n < 0 {\n\t\treturn \"\"\n\t}\n\n\treturn strn(value, n)\n}", "func (s ListViewsInput) String() string {\n\treturn awsutil.Prettify(s)\n}" ]
[ "0.63379955", "0.5909125", "0.58037746", "0.56420934", "0.56420934", "0.54916877", "0.54738015", "0.5430865", "0.5414218", "0.5414218", "0.53638655", "0.5337022", "0.53342456", "0.53323567", "0.52999586", "0.52574825", "0.52541006", "0.52512985", "0.5248777", "0.5213757", "0.51917297", "0.5171241", "0.51634866", "0.5157035", "0.5157035", "0.5156469", "0.513832", "0.51378185", "0.5128501", "0.5126863", "0.5126863", "0.51238745", "0.51109403", "0.5090268", "0.5090268", "0.5087596", "0.50620174", "0.50593096", "0.50425786", "0.5038153", "0.5014318", "0.5009089", "0.50069445", "0.50066686", "0.50066686", "0.50066686", "0.50066686", "0.49849522", "0.49833876", "0.49368206", "0.49290252", "0.49268696", "0.4926448", "0.49256456", "0.49154732", "0.49128085", "0.4911656", "0.49093148", "0.49093148", "0.49093148", "0.49093148", "0.49093148", "0.49093148", "0.49093148", "0.48981756", "0.4891384", "0.48910582", "0.48741093", "0.48418227", "0.48381555", "0.48381555", "0.48307803", "0.48296013", "0.48198634", "0.4815694", "0.4813439", "0.48025224", "0.47952625", "0.47882044", "0.47725397", "0.47717014", "0.4769885", "0.47630322", "0.47620183", "0.475038", "0.47329995", "0.47321823", "0.47314504", "0.47306818", "0.4728975", "0.4717342", "0.47094235", "0.47076365", "0.47069144", "0.47055426", "0.4696454", "0.46919572", "0.46919572", "0.4685944", "0.466918" ]
0.7783467
0
bytesView returns a view of the string as a []byte. In unsafe mode, it doesn't incur allocation and copying caused by conversion. In regular safe mode, it is an allocation and copy. Usage: Always maintain a reference to v while result of this call is in use, and call keepAlive4BytesView(v) at point where done with view.
func bytesView(v string) []byte { return []byte(v) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func stringView(v []byte) string {\n\treturn string(v)\n}", "func NewView(b []byte) *View { return &View{b} }", "func GetBytesBuffer4K() *bytes.Buffer {\n\tif b := getb4K(); b != nil {\n\t\treturn b\n\t}\n\tif p := get4K(); p != nil {\n\t\treturn bytes.NewBuffer(internal.Puts(p))\n\t}\n\treturn bytes.NewBuffer(make([]byte, 4096))\n}", "func GetBytesBuffer4M() *bytes.Buffer {\n\tif b := getb4M(); b != nil {\n\t\treturn b\n\t}\n\tif p := get4M(); p != nil {\n\t\treturn bytes.NewBuffer(internal.Puts(p))\n\t}\n\treturn bytes.NewBuffer(make([]byte, 4194304))\n}", "func (v ByteVec) String() string { return string([]byte(v)) }", "func (b *SafeBuffer) Bytes() []byte {\n\tb.m.RLock()\n\tdefer b.m.RUnlock()\n\treturn b.b.Bytes()\n}", "func VVByte_to_string(m [][]byte) (*bytes.Buffer, *bytes.Buffer) {\n\n\tlg, b := loghttp.BuffLoggerUniversal(nil, nil)\n\t_ = b\n\n\tbRet := new(bytes.Buffer)\n\tbMsg := new(bytes.Buffer)\n\n\t//for i,v := range m {\n\tfor i := 0; i < len(m); i++ {\n\t\tn, err := bRet.Write(m[i])\n\t\tlg(err)\n\t\tbMsg.WriteString(\" lp\" + util.Itos(i) + \": writing \" + util.Itos(n) + \" bytes: \\n\")\n\t}\n\treturn bRet, bMsg\n}", "func ImmutableBytesFromString(s string) []byte {\n\tb := unsafe.StringData(s)\n\treturn unsafe.Slice(b, len(s))\n}", "func NewBytesViewer(from io.ReadCloser) (in io.ReadCloser) {\n\treturn &bytesViewer{from}\n}", "func (g *GLTF) loadBufferView(bvIdx int) ([]byte, error) {\n\n\t// Check if provided buffer view index is valid\n\tif bvIdx < 0 || bvIdx >= len(g.BufferViews) {\n\t\treturn nil, fmt.Errorf(\"invalid buffer view index\")\n\t}\n\tbvData := g.BufferViews[bvIdx]\n\t// Return cached if available\n\tif bvData.cache != nil {\n\t\tlog.Debug(\"Fetching BufferView %d (cached)\", bvIdx)\n\t\treturn bvData.cache, nil\n\t}\n\tlog.Debug(\"Loading BufferView %d\", bvIdx)\n\n\t// Load buffer view buffer\n\tbuf, err := g.loadBuffer(bvData.Buffer)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Establish offset\n\toffset := 0\n\tif bvData.ByteOffset != nil {\n\t\toffset = *bvData.ByteOffset\n\t}\n\n\t// Compute and return offset slice\n\tbvBytes := buf[offset : offset+bvData.ByteLength]\n\n\t// Cache buffer view\n\tg.BufferViews[bvIdx].cache = bvBytes\n\n\treturn bvBytes, nil\n}", "func BytesPtr(v []byte) *[]byte {\n\treturn &v\n}", "func StringToReadOnlyBytes(s string) []byte {\n\treturn []byte(s)\n}", "func zbuf(v Value) []byte {\n\ts := ToStr(v)\n\tbuf := make([]byte, len(s)+1)\n\tcopy(buf, s)\n\tbuf[len(s)] = 0\n\treturn buf\n}", "func rtBytes(rv reflect.Value) []byte {\n\t// Fastpath if the underlying type is []byte\n\tif rv.Kind() == reflect.Slice && rv.Type().Elem() == rtByte {\n\t\treturn rv.Bytes()\n\t}\n\t// Slowpath copying bytes one by one.\n\tret := make([]byte, rv.Len())\n\tfor ix := 0; ix < rv.Len(); ix++ {\n\t\tret[ix] = rv.Index(ix).Convert(rtByte).Interface().(byte)\n\t}\n\treturn ret\n}", "func (_ BufferPtrPool4K) Get() *[]byte {\n\treturn GetBytesSlicePtr4K()\n}", "func UnsafeStringBytes(s *string) []byte {\n\treturn *(*[]byte)(unsafe.Pointer((*reflect.SliceHeader)(unsafe.Pointer(s))))\n}", "func (b *buffer) bytes(n int) []byte {\n\tdata := b.unsafeBytes(n)\n\tif !b.shared {\n\t\treturn data\n\t}\n\t// TODO(kortschak): Replace this with bytes.Clone when available.\n\t// See https://github.com/golang/go/issues/45038 for bytes.Clone.\n\treturn append(data[:0:0], data...)\n}", "func StringToByteArray(v string) []byte {\n\tvar slcHdr reflect.SliceHeader\n\tsh := *(*reflect.StringHeader)(unsafe.Pointer(&v))\n\tslcHdr.Data = sh.Data\n\tslcHdr.Cap = sh.Len\n\tslcHdr.Len = sh.Len\n\treturn *(*[]byte)(unsafe.Pointer(&slcHdr))\n}", "func (b *buffer) unsafeBytes(n int) []byte {\n\tif b.err != nil {\n\t\treturn nil\n\t}\n\tif b.len() < n {\n\t\tb.err = io.ErrUnexpectedEOF\n\t\treturn nil\n\t}\n\ts := b.off\n\tb.off += n\n\treturn b.data[s:b.off]\n}", "func GetBytesSlice4M() []byte {\n\tif p := get4M(); p != nil {\n\t\treturn internal.Puts(p)\n\t}\n\tif b := getb4M(); b != nil {\n\t\treturn internal.Bb2bs(b)\n\t}\n\tp := make([]byte, 4194304)\n\treturn p\n}", "func BytePtr(v byte) *byte { return &v }", "func GetBytesSlice4K() []byte {\n\tif p := get4K(); p != nil {\n\t\treturn internal.Puts(p)\n\t}\n\tif b := getb4K(); b != nil {\n\t\treturn internal.Bb2bs(b)\n\t}\n\tp := make([]byte, 4096)\n\treturn p\n}", "func (b *Buffer) bytes() []byte {\n\treturn b.data\n}", "func GetBytesSlicePtr4M() *[]byte {\n\tif p := get4M(); p != nil {\n\t\treturn p\n\t}\n\tif b := getb4M(); b != nil {\n\t\tp := internal.Bb2bs(b)\n\t\treturn internal.Gets(p)\n\t}\n\tp := make([]byte, 4194304)\n\treturn &p\n}", "func (_ BufferPtrPool4M) Get() *[]byte {\n\treturn GetBytesSlicePtr4M()\n}", "func (v *Value) Bytes() []byte {\n switch d := v.data.(type) {\n case []byte:\n return d\n case string:\n return []byte(d)\n default:\n if j, e := json.Marshal(v.data); e == nil {\n return j\n }\n return []byte(fmt.Sprintf(\"%+v\", v.data))\n }\n}", "func GetBytesSlicePtr4K() *[]byte {\n\tif p := get4K(); p != nil {\n\t\treturn p\n\t}\n\tif b := getb4K(); b != nil {\n\t\tp := internal.Bb2bs(b)\n\t\treturn internal.Gets(p)\n\t}\n\tp := make([]byte, 4096)\n\treturn &p\n}", "func StringFromImmutableBytes(bs []byte) string {\n\tif len(bs) == 0 {\n\t\treturn \"\"\n\t}\n\treturn unsafe.String(&bs[0], len(bs))\n}", "func gostringnocopy(str *byte) string", "func (b *printer) UnsafeBytes(s []byte) {\n\tw := escapeWriter{w: &b.buf, enclose: true, strip: true}\n\t_, _ = w.Write(s)\n}", "func (b *Buffer) Bytes() []byte {\n\tb.m.RLock()\n\tbcpy := make([]byte, len(b.b.Bytes()))\n\tcopy(bcpy, b.b.Bytes())\n\tb.m.RUnlock()\n\treturn bcpy\n}", "func (s *scratch) bytes() []byte { return s.data[0:s.fill] }", "func (v Value) Bytes() []byte {\n\treturn v.bytes\n}", "func NewBufferBytes(data []byte) *Buffer {\n\treturn &Buffer{refCount: 0, buf: data, length: len(data)}\n}", "func StringBytes(b []byte) string { return *(*string)(Pointer(&b)) }", "func UnsafeStrToBytes(s string) []byte {\n\treturn unsafe.Slice(unsafe.StringData(s), len(s)) // ref https://github.com/golang/go/issues/53003#issuecomment-1140276077\n}", "func (b *Buffer) Bytes() []byte { return b.buf[:b.length] }", "func (s String) Bytes() []byte {\n\treturn []byte(s)\n}", "func (ip IP) v4(i uint8) uint8 {\n\treturn uint8(ip.lo >> ((3 - i) * 8))\n}", "func StringBytes(s string) []byte {\n\treturn *(*[]byte)(unsafe.Pointer(&s))\n}", "func (pk PacketBufferPtr) ToView() *buffer.View {\n\tp := buffer.NewView(int(pk.buf.Size()))\n\toffset := pk.headerOffset()\n\tpk.buf.SubApply(offset, int(pk.buf.Size())-offset, func(v *buffer.View) {\n\t\tp.Write(v.AsSlice())\n\t})\n\treturn p\n}", "func (b *Buffer) Bytes() []byte {\n\tswitch {\n\tcase b.written >= b.size && b.writeCursor == 0:\n\t\treturn b.data[b.offset:]\n\tcase b.written > b.size:\n\t\tout := make([]byte, b.size)\n\t\tcopy(out,\n\t\t\tb.data[b.offset+b.writeCursor:])\n\t\tcopy(out[b.size-b.writeCursor:],\n\t\t\tb.data[b.offset:b.offset+b.writeCursor])\n\t\treturn out\n\tdefault:\n\t\treturn b.data[b.offset : b.offset+b.writeCursor]\n\t}\n}", "func Str2bytes(s string) []byte {\n\tstringHeader := *(*[2]int)(unsafe.Pointer(&s))\n\tvar sliceHeader [3]int\n\tsliceHeader[0] = stringHeader[0]\n\tsliceHeader[1] = stringHeader[1]\n\tsliceHeader[2] = stringHeader[1]\n\treturn *(*[]byte)(unsafe.Pointer(&sliceHeader))\n}", "func (b *Buf) Bytes() []byte { return b.b }", "func InternBytes(b []byte) string {\n\ts := ToUnsafeString(b)\n\treturn InternString(s)\n}", "func InternBytes(b []byte) string {\n\ts := ToUnsafeString(b)\n\treturn InternString(s)\n}", "func (s String) ToBytes() []byte {\n\treturn []byte(s)\n}", "func bytesFromCBytes(size uint32, p *byte) []byte {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tif *p == 0 {\n\t\treturn nil\n\t}\n\ttmp := unsafe.Slice(p, size)\n\tout := make([]byte, len(tmp))\n\tcopy(out, tmp)\n\treturn out\n}", "func (o InstanceOutput) AccessIpV4() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Instance) pulumi.StringOutput { return v.AccessIpV4 }).(pulumi.StringOutput)\n}", "func TestAuxUnsafeBytes(t *testing.T) {\n\tvar objBytesManaged = createObjectBytes(object)\n\tassert.Eq(t, 216, len(objBytesManaged))\n\tassert.Eq(t, 52, int(objBytesManaged[0]))\n\tassert.Eq(t, 51, int(objBytesManaged[57]))\n\n\t// get an unsafe copy (really just a pointer), as the the cursor would do\n\tvar unsafeBytes = getUnsafeBytes(objBytesManaged)\n\n\t// get a safe copy\n\tvar safeBytes = make([]byte, len(objBytesManaged))\n\tcopy(safeBytes, objBytesManaged)\n\n\t// at this point, they should all be the same\n\tassert.Eq(t, objBytesManaged, unsafeBytes)\n\tassert.NotEq(t, unsafe.Pointer(&objBytesManaged), unsafe.Pointer(&unsafeBytes))\n\tassert.Eq(t, unsafe.Pointer(&objBytesManaged[0]), unsafe.Pointer(&unsafeBytes[0]))\n\n\tassert.Eq(t, objBytesManaged, safeBytes)\n\tassert.NotEq(t, unsafe.Pointer(&objBytesManaged), unsafe.Pointer(&safeBytes))\n\n\t// now let's clear the object bytes, and check the copies\n\tclearBytes(&objBytesManaged)\n\tassert.Eq(t, 216, len(objBytesManaged))\n\tassert.Eq(t, 0, int(objBytesManaged[0]))\n\n\t// now we assert the unsafe bytes has changed if it wasn't supposed to\n\tassert.Eq(t, objBytesManaged, unsafeBytes)\n\n\t// but the safe copy is still the same\n\tassert.Eq(t, 52, int(safeBytes[0]))\n}", "func (a Address) Bytes() []byte { return a[:] }", "func (a Address) Bytes() []byte { return a[:] }", "func (b *SafeBuffer) String() string {\n\tb.m.RLock()\n\tdefer b.m.RUnlock()\n\treturn b.b.String()\n}", "func (b *Buffer) Bytes() []byte {\n\treturn b.bytes[0:b.length]\n}", "func (addr *Bytes) Load() (val []byte) {\n\tv := addr.v.Load()\n\tif v == nil {\n\t\treturn nil\n\t}\n\treturn v.([]byte)\n}", "func string2bytes(s string) []byte {\n\tstringHeader := (*reflect.StringHeader)(unsafe.Pointer(&s))\n\n\tbh := reflect.SliceHeader{\n\t\tData: stringHeader.Data,\n\t\tLen: stringHeader.Len,\n\t\tCap: stringHeader.Len,\n\t}\n\n\treturn *(*[]byte)(unsafe.Pointer(&bh))\n}", "func StringBytes(s string) []byte {\n\treturn []byte(s)\n}", "func (h *RequestHeader) SetBytesV(key string, value []byte) {\n\tk := getHeaderKeyBytes(&h.bufKV, key, h.disableNormalizing)\n\th.SetCanonical(k, value)\n}", "func StringToBytes(str string) []byte {\n\t// Empty strings may not allocate a backing array, so we have to check first\n\tif len(str) == 0 {\n\t\t// It makes sense to return a non-nil empty byte slice since we're passing in a non-nil (although empty) string\n\t\treturn []byte{}\n\t}\n\treturn (*[0x7fff0000]byte)(unsafe.Pointer(\n\t\t(*reflect.StringHeader)(unsafe.Pointer(&str)).Data),\n\t)[:len(str):len(str)]\n}", "func NewBytesBuffer(p []byte) *BytesBuffer {\n\treturn &BytesBuffer{reader: bytes.NewReader(p)}\n}", "func ReadOnlyBytes(s string) []byte {\n\tsh := (*reflect.StringHeader)(unsafe.Pointer(&s))\n\tbh := reflect.SliceHeader{Data: sh.Data, Len: sh.Len, Cap: sh.Len}\n\treturn *(*[]byte)(unsafe.Pointer(&bh))\n}", "func (a *DatumAlloc) NewDBytes(v tree.DBytes) *tree.DBytes {\n\tif a.AllocSize == 0 {\n\t\ta.AllocSize = defaultDatumAllocSize\n\t}\n\tbuf := &a.dbytesAlloc\n\tif len(*buf) == 0 {\n\t\t*buf = make([]tree.DBytes, a.AllocSize)\n\t}\n\tr := &(*buf)[0]\n\t*r = v\n\t*buf = (*buf)[1:]\n\treturn r\n}", "func(this *GView) Get() interface{} {\nif this == nil {\n\treturn nil\n} else {\n\tbuffer := make([]byte, 2048)\n\t\ttempString := \"\"\n\t\tn, _ := this.inOut.Read(buffer)\n\t\tfor {\n\t\t\ttemp := (string) (buffer[0:n])\n\t\t\ttempString = tempString + temp\n\t\t\tif n < 2048 { return tempString }\n\t\t\tn, _ = this.inOut.Read(buffer)\n\t\t}\n\t\treturn tempString\n\t}\nreturn \"\"\n}", "func (b *Blockstore) View(cid cid.Cid, fn func([]byte) error) error {\n\tif atomic.LoadInt64(&b.state) != stateOpen {\n\t\treturn ErrBlockstoreClosed\n\t}\n\n\tk, pooled := b.PooledStorageKey(cid)\n\tif pooled {\n\t\tdefer KeyPool.Put(k)\n\t}\n\n\treturn b.DB.View(func(txn *badger.Txn) error {\n\t\tswitch item, err := txn.Get(k); err {\n\t\tcase nil:\n\t\t\treturn item.Value(fn)\n\t\tcase badger.ErrKeyNotFound:\n\t\t\treturn blockstore.ErrNotFound\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"failed to view block from badger blockstore: %w\", err)\n\t\t}\n\t})\n}", "func BytesValue(b []byte) Value { return Value{Typ: '$', Str: b} }", "func (p *Buffer) Bytes() []byte { return p.buf }", "func (f *FileList) ToLevelViewBytes(pad string) []byte {\n\treturn []byte(f.ToLevelView(pad, false))\n}", "func UnsafeStrToBytes(s string) []byte {\n\treturn *(*[]byte)(unsafe.Pointer(&s))\n}", "func UnsafeStrToBytes(s string) []byte {\n\treturn *(*[]byte)(unsafe.Pointer(&s))\n}", "func (b *Vector) ByteSlice(off VOffsetT, s []byte) VField {\n\tb.VOffset = off\n\n\tb.t = FieldTypeByteSlice\n\tb.objectSize = len(s)\n\n\tpad, total := Prepad(int(SizeUOffsetT), (b.objectSize+1)*SizeByte)\n\n\tif b.b == nil {\n\t\tb.b = bytepool.NewByteBuffer(64)\n\t}\n\tb.b.Reset(total + SizeSOffsetT).Pad(pad)\n\n\tfor i := 0; i < pad; i++ {\n\t\tWriteByte(b.b.B[int(b.head)+i+b.objectSize:int(b.head)+i+b.objectSize+1], uint8(0))\n\t}\n\tcopy(b.b.B[b.head:], s)\n\tb.EndVector(b.objectSize)\n\tb.finished = true\n\n\treturn b\n}", "func byteReverseIP4(ip net.IP) (revip revIP4) {\n\n\tfor j := 0; j < len(ip); j++ {\n\t\trevip.Byte[len(ip)-j-1] = ip[j]\n\t\trevip.String = fmt.Sprintf(\"%d.%s\", ip[j], revip.String)\n\t}\n\n\trevip.String = strings.TrimRight(revip.String, \".\")\n\n\treturn\n}", "func GoEcho4(s *C.char, length C.int) (unsafe.Pointer, int) {\n\tslice := C.GoBytes(unsafe.Pointer(s), length)\n\tslice = append(slice, \" from golang4\"...)\n\treturn unsafe.Pointer(&(slice[0])), len(slice)\n}", "func (addr DevAddr) Bytes() []byte {\n\treturn addr[:]\n}", "func (b *Blockstore) View(cid cid.Cid, fn func([]byte) error) error {\n\tswitch v, err := b.DB.Get(cid.Hash(), nil); err {\n\tcase nil:\n\t\treturn fn(v)\n\tcase leveldb.ErrNotFound:\n\t\treturn blockstore.ErrNotFound\n\tdefault:\n\t\treturn fmt.Errorf(\"failed to view block from leveldb blockstore: %w\", err)\n\t}\n}", "func (v Value) Bytes() []byte {\n\tswitch v.Typ {\n\tdefault:\n\t\treturn []byte(v.String())\n\tcase '$', '+', '-':\n\t\treturn v.Str\n\t}\n}", "func string2ByteSlice(str string) (bs []byte) {\n\tstrHdr := (*reflect.StringHeader)(unsafe.Pointer(&str))\n\tsliceHdr := (*reflect.SliceHeader)(unsafe.Pointer(&bs))\n\tsliceHdr.Data = strHdr.Data\n\tsliceHdr.Len = strHdr.Len\n\tsliceHdr.Cap = strHdr.Len\n\t// This KeepAlive line is essential to make the\n\t// String2ByteSlice function be always valid\n\t// when it is provided in other custom packages.\n\truntime.KeepAlive(&str)\n\treturn\n}", "func UnsafeBytesToStr(b []byte) string {\n\treturn *(*string)(unsafe.Pointer(&b))\n}", "func UnsafeBytesToStr(b []byte) string {\n\treturn *(*string)(unsafe.Pointer(&b))\n}", "func UnsafeBytesToStr(b []byte) string {\n\treturn *(*string)(unsafe.Pointer(&b))\n}", "func (vl BytesValue) String() string {\n\treturn Buffer.BytesToHexString(vl)\n}", "func (vl BytesValue) String() string {\n\treturn Buffer.BytesToHexString(vl)\n}", "func (d PacketData) AppendView(v *buffer.View) {\n\td.pk.buf.Append(v)\n}", "func NewVersionUnsafe(str string) Version {\n\tv, _ := NewVersion(str)\n\treturn v\n}", "func (addr BitcoinAddress) Bytes() []byte {\n\tb := make([]byte, 20+1+4)\n\tb[0] = addr.Version\n\tcopy(b[1:21], addr.Key[0:20])\n\tchksum := addr.Checksum()\n\tcopy(b[21:25], chksum[0:4])\n\treturn b\n}", "func (d *Document) Bytes() *bytes.Buffer {\n\tbuf := new(bytes.Buffer)\n\tbinary.Write(buf, binary.LittleEndian, d.id)\n\tbinary.Write(buf, binary.LittleEndian, d.length)\n\tbinary.Write(buf, binary.LittleEndian, uint32(len(d.path)))\n\tbinary.Write(buf, binary.LittleEndian, []byte(d.path))\n\treturn buf\n}", "func (v *View) String() string {\n\tf := format.NewASCII()\n\treturn v.FormatString(f)\n}", "func StringFromBytes(b []byte) String {\n\treturn StringFromString(string(b))\n}", "func bytesToString(bs []byte) string {\n\treturn *(*string)(unsafe.Pointer(&bs))\n}", "func (_ BufferPtrPool16K) Get() *[]byte {\n\treturn GetBytesSlicePtr16K()\n}", "func StringOrBytes(value interface{}) (isString bool, str string, isBytes bool, bs []byte) {\n\tv := reflect.ValueOf(value)\n\tif v.Kind() == reflect.String {\n\t\tstr = v.String()\n\t\tisString = true\n\t} else if v.Kind() == reflect.Slice && v.Type() == bytesType {\n\t\tbs = v.Interface().([]byte)\n\t\tisBytes = true\n\t}\n\treturn\n}", "func Byte(v byte) *byte {\n\treturn &v\n}", "func Byte(v byte) *byte {\n\treturn &v\n}", "func BytesToString(b []byte) string { return *(*string)(unsafe.Pointer(&b)) }", "func (u UUID) Bytes() []byte {\n\treturn u[:]\n}", "func (u UUID) Bytes() []byte {\n\treturn u[:]\n}", "func (v Version) ToBytes() []byte {\n\treturn []byte{\n\t\tv.Major,\n\t\tv.Minor,\n\t\tv.Patch,\n\t}\n}", "func BytesString(b []byte) string {\n\treturn *(*string)(unsafe.Pointer(&b))\n}", "func (w *ByteWriter) Bytes() []byte { return w.buffer }", "func (p *parser) VisitBytes(ctx *gen.BytesContext) any {\n\tb := []byte(p.unquote(ctx, ctx.GetTok().GetText()[1:], true))\n\treturn p.helper.newLiteralBytes(ctx, b)\n}", "func (_ BufferPtrPool8K) Get() *[]byte {\n\treturn GetBytesSlicePtr8K()\n}" ]
[ "0.6358887", "0.5210643", "0.47978938", "0.47844735", "0.47593164", "0.46141908", "0.4609627", "0.46013665", "0.45809", "0.4562257", "0.45536464", "0.4526796", "0.45103198", "0.45092463", "0.4495565", "0.4488356", "0.4468821", "0.4467235", "0.4460009", "0.44589624", "0.44351736", "0.4425497", "0.44014257", "0.4387796", "0.43523282", "0.43463168", "0.43428335", "0.43372566", "0.43088296", "0.43083033", "0.43079248", "0.4279778", "0.42766464", "0.42753237", "0.42674583", "0.4248836", "0.4247261", "0.41997856", "0.41996855", "0.41893035", "0.4187981", "0.41723266", "0.4172011", "0.41635853", "0.41634792", "0.41634792", "0.4158446", "0.41492963", "0.41364914", "0.41281646", "0.41229627", "0.41229627", "0.41228908", "0.41170776", "0.41141525", "0.41110376", "0.41009763", "0.40969053", "0.40948114", "0.40840673", "0.40789115", "0.40767977", "0.40703654", "0.40618962", "0.40480104", "0.40436652", "0.4042615", "0.40380874", "0.40380874", "0.4035087", "0.40312266", "0.4024979", "0.40079352", "0.40065864", "0.4005098", "0.40002123", "0.39985725", "0.39985725", "0.39985725", "0.39946908", "0.39946908", "0.39946294", "0.3983998", "0.39821574", "0.39780796", "0.39743307", "0.3971957", "0.3971725", "0.3966956", "0.39627025", "0.3957349", "0.3957349", "0.3956155", "0.39548343", "0.39548343", "0.3951219", "0.39452186", "0.394118", "0.39378616", "0.39325857" ]
0.75592566
0
NewNode creates a new node base around this configuration
func NewNode(config *Configuration) *Node { node := new(Node) node.config = config node.partitions = make(map[string][]partition) node.strategyMap = make(map[string]RoutingStrategy) // create a http listener, do it this way to get the running port addrStr := fmt.Sprintf(":%d", config.NodePort) listener, err := net.Listen("tcp", addrStr) if err != nil { log.Err("Failed to create listener at address %s", addrStr, err) } node.port = listener.Addr().(*net.TCPAddr).Port node.server = http.Server{ Handler: node, } log.Info("Node is starting to listen on port: %d", node.port) go node.server.Serve(listener) return node }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewNode(id uint64) *Node { return &Node{Id: id} }", "func NewNode(node ast.Node) Node {\n\treturn creator(baseNode{node: node})\n}", "func NewNode(name string, config *Config) *Node {\n\tnode := &Node{\n\t\tName: name,\n\t\tConfig: config,\n\t\tRooms: map[model.Guid]*Room{},\n\t}\n\n\tif config.Store != nil {\n\t\tnode.connectStore()\n\t}\n\tif config.PlayHistory != nil {\n\t\tnode.connectPlayHistory()\n\t}\n\tif config.SessionStore != nil {\n\t\tnode.connectSessionStore()\n\t}\n\n\treturn node\n}", "func NewNode(p string) *MyNode {\n\taddr := getLocalAddress()\n\n\treturn &MyNode{\n\t\tAddr: addr,\n\t\tPort: p,\n\t\tID: HashString(fmt.Sprintf(\"%v:%v\", addr, p)),\n\t\tData: make(map[string]string),\n\t\tbuf: new(bytes.Buffer),\n\t}\n}", "func NewNode(host string, size int) Node {\n\treturn node{host: host, size: size}\n}", "func newNode(cluster *Cluster, nv *nodeValidator) *Node {\n\treturn &Node{\n\t\tcluster: cluster,\n\t\tname: nv.name,\n\t\taliases: nv.aliases,\n\t\taddress: nv.address,\n\t\tuseNewInfo: nv.useNewInfo,\n\n\t\t// Assign host to first IP alias because the server identifies nodes\n\t\t// by IP address (not hostname).\n\t\thost: nv.aliases[0],\n\t\tconnections: NewAtomicQueue(cluster.clientPolicy.ConnectionQueueSize),\n\t\tconnectionCount: NewAtomicInt(0),\n\t\thealth: NewAtomicInt(_FULL_HEALTH),\n\t\tpartitionGeneration: NewAtomicInt(-1),\n\t\treferenceCount: NewAtomicInt(0),\n\t\trefreshCount: NewAtomicInt(0),\n\t\tresponded: NewAtomicBool(false),\n\t\tactive: NewAtomicBool(true),\n\n\t\tsupportsFloat: NewAtomicBool(nv.supportsFloat),\n\t\tsupportsBatchIndex: NewAtomicBool(nv.supportsBatchIndex),\n\t\tsupportsReplicasAll: NewAtomicBool(nv.supportsReplicasAll),\n\t\tsupportsGeo: NewAtomicBool(nv.supportsGeo),\n\t}\n}", "func NewNode(up *Node, name string) *Node {\n\tlogger := log.New(os.Stderr, \"-> [Node][\"+ name +\"] -> \", log.Lmicroseconds)\n\n\tnode := &Node {\n\t\tup:\t\tup,\n\t\tl: \t\tlogger,\n\t\tname:\tname,\n\t}\n\n\treturn node;\n}", "func newNode(nodePath string) Node {\n\treturn &nodeImpl{nodePath: nodePath}\n}", "func New(nodeFile string) Node {\n\treturn Node{nodeFile: nodeFile}\n}", "func NewNode(key int, value string) *Node {\n\treturn &Node{1, nil, nil, key, value}\n}", "func NewNode(name string) *Node {\n\treturn &Node{\n\t\tName: name,\n\t\tVersion: configs.VERSITON,\n\t\tCreateTime: time.Now(),\n\t}\n}", "func newnode(id byte, name string, value string) *xmlx.Node {\n\tnode := xmlx.NewNode(id)\n\tif name != \"\" {\n\t\tnode.Name = xml.Name{\n\t\t\tLocal: name,\n\t\t}\n\t}\n\tif value != \"\" {\n\t\tnode.Value = value\n\t}\n\treturn node\n}", "func NewNode(v string) *node {\n\treturn &node{\n\t\tvalue: v,\n\t\tchildren: make(map[*node]bool),\n\t}\n}", "func NewNode(name string, w Worker) *Node {\n\tid := getID()\n\tn := &Node{id: id, w: w, name: name}\n\tn.chained = make(map[string]struct{})\n\tn.close = make(chan struct{})\n\treturn n\n}", "func NewNode(value interface{}) *Node {\n\treturn &Node{\n\t\tValue: value,\n\t}\n}", "func NewNode(name string) Node {\n\treturn Node{\n\t\tName: name,\n\t\tFlags: make(map[string]bool),\n\t\tLinks: make([]*Link, 0),\n\t\tNodes: make(map[string]*Node),\n\t}\n}", "func NewNode(key, value interface{}) *Node {\n\tvar node Node\n\tnode.value = value\n\tnode.key = key\n\tnode.next = nil\n\treturn &node\n}", "func New() *Node {\n\treturn &Node{}\n}", "func NewNode(\n\tethereumChain eth.Handle,\n\tnetworkProvider net.Provider,\n\ttssConfig *tss.Config,\n) *Node {\n\treturn &Node{\n\t\tethereumChain: ethereumChain,\n\t\tnetworkProvider: networkProvider,\n\t\ttssConfig: tssConfig,\n\t}\n}", "func NewNode(importPath string, val string) *Node {\n\treturn &Node{\n\t\tImportPath: GetRootPath(importPath),\n\t\tValue: val,\n\t\tDownloadURL: GetRootPath(importPath),\n\t}\n}", "func NewNode(log ginterface.IGameLogger) *Node {\n\tret := &Node{Log: log}\n\treturn ret\n}", "func NewNode(opts ...NodeOpt) *Node {\n\tn := &Node{}\n\tfor _, opt := range opts {\n\t\topt(n)\n\t}\n\treturn n\n}", "func createNewNode(ctx context.Context, nodeName string, virtual bool, clientset kubernetes.Interface) (*corev1.Node, error) {\n\tresources := corev1.ResourceList{}\n\tresources[corev1.ResourceCPU] = *resource.NewScaledQuantity(5000, resource.Milli)\n\tresources[corev1.ResourceMemory] = *resource.NewScaledQuantity(5, resource.Mega)\n\tnode := &corev1.Node{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: nodeName,\n\t\t},\n\t}\n\tif virtual {\n\t\tnode.Labels = map[string]string{\n\t\t\tconsts.TypeLabel: consts.TypeNode,\n\t\t}\n\t}\n\tnode.Status = corev1.NodeStatus{\n\t\tCapacity: resources,\n\t\tAllocatable: resources,\n\t\tConditions: []corev1.NodeCondition{\n\t\t\t0: {\n\t\t\t\tType: corev1.NodeReady,\n\t\t\t\tStatus: corev1.ConditionTrue,\n\t\t\t},\n\t\t},\n\t}\n\tnode, err := clientset.CoreV1().Nodes().Create(ctx, node, metav1.CreateOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn node, nil\n}", "func NewNode(name string) (n *Node, err error) {\n\t// retrive the role the node using docker inspect\n\tlines, err := host.InspectContainer(name, fmt.Sprintf(\"{{index .Config.Labels %q}}\", constants.DeprecatedNodeRoleLabelKey))\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to get %q label\", constants.DeprecatedNodeRoleLabelKey)\n\t}\n\tif len(lines) != 1 {\n\t\treturn nil, errors.Errorf(\"%q label should only be one line, got %d lines\", constants.DeprecatedNodeRoleLabelKey, len(lines))\n\t}\n\trole := strings.Trim(lines[0], \"'\")\n\n\treturn &Node{\n\t\tname: name,\n\t\trole: role,\n\t}, nil\n}", "func NewNode() *Node {\n\treturn &Node{}\n}", "func NewNode(name string) TreeNode {\n\treturn TreeNode{\n\t\tname: name,\n\t\tsize: 0,\n\t\tfiles: make(map[string]Entry),\n\t}\n}", "func NewNode(name string, node files.Node) *Node {\n\treturn &Node{\n\t\tNode: node,\n\t\tname: name,\n\t}\n}", "func newNode() *Node {\n\tn := &Node{}\n\treturn n\n}", "func NewNode(id string, parent *string) Node {\n\treturn Node{\n\t\tID: id,\n\t\tParent: parent,\n\t}\n}", "func createNewEmptyNode() Node {\n\tnextNewId--\n\treturn Node{\n\t\tId: nextNewId,\n\t\tVisible: true,\n\t\tTimestamp: time.Now().Format(\"2006-01-02T15:04:05Z\"),\n\t\tVersion: \"1\",\n\t}\n}", "func NewNode(data string) *Node {\n\tnode := &Node{}\n\tnode.data = data\n\treturn node\n}", "func NewNode(cnf *Config, joinNode *api.Node) (*Node, error) {\n\tvar nodeID string\n\n\tnode := &Node{\n\t\tNode: new(api.Node),\n\t\tshutdownCh: make(chan struct{}),\n\t\tcnf: cnf,\n\t\tstorage: NewMapStore(cnf.Hash),\n\t}\n\tif cnf.Id != \"\" {\n\t\tnodeID = cnf.Id\n\t} else {\n\t\tnodeID = cnf.Addr\n\t}\n\n\tid, err := node.hashKey(nodeID)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\taInt := (&big.Int{}).SetBytes(id) // treating id as bytes of a big-endian unsigned integer, return the integer it represents\n\tlog.Printf(aurora.Sprintf(aurora.Yellow(\"New Node ID = %d, \\n\"), aInt))\n\tnode.Node.Id = id\n\tnode.Node.Addr = cnf.Addr\n\t// Populate finger table (by anotating itself to be in charge of all possible hashes at the moment)\n\tnode.fingerTable = newFingerTable(node.Node, cnf.HashSize)\n\n\t// Start RPC server (start listening function, )\n\t// transport is a struct that contains grpc server and supplementary attributes (like timeout etc)\n\ttransport, err := NewGrpcTransport(cnf)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tnode.transport = transport\n\n\tapi.RegisterChordServer(transport.server, node)\n\tnode.transport.Start()\n\n\t// find the closest node clockwise from the id of this node (i.e. successor node)\n\t// adds successor to the 'successor' attribute of the node\n\tnodeJoinErr := node.join(joinNode)\n\n\tif nodeJoinErr != nil {\n\t\tlog.Printf(\"Error joining node\")\n\t\treturn nil, err\n\t}\n\n\t// run routines\n\t// Fix fingers every 500 ms\n\tgo node.fixFingerRoutine(500)\n\n\t// Stablize every 1000ms\n\tgo node.stabilizeRoutine(1000)\n\t// Check predecessor fail every 5000 ms\n\tgo node.checkPredecessorRoutine(2000)\n\n\treturn node, nil\n}", "func New() *Node {\n\treturn &Node{\n\t\tparent: nil,\n\t\tchildren: map[option]*Node{},\n\t}\n}", "func newNode(hash string) Node {\n\treturn Node{hash: hash, parent: nil}\n}", "func (t *Btree) newNode() *Node {\n\t*t.NodeCount++\n\tid := t.genrateID()\n\tnode := &Node{\n\t\tNodeRecordMetaData: NodeRecordMetaData{\n\t\t\tId: proto.Int64(id),\n\t\t\tIsDirt: proto.Int32(0),\n\t\t},\n\t}\n\tt.nodes[id] = node\n\treturn node\n}", "func NewNode(ip string) *Node {\n\treturn &Node{\n\t\tIP: ip,\n\t}\n}", "func (m *Mobile) NewNode(repoPath string) (*Wrapper, error) {\n\tnode, err := tcore.NewNode(repoPath, true, logging.DEBUG)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Wrapper{RepoPath: repoPath, Cancel: node.Cancel, node: node}, nil\n}", "func NewNode(key, value interface{}) *Node {\n\tnode := new(Node)\n\tnode.key = key\n\tnode.value = value\n\tnode.left = nil\n\tnode.right = nil\n\tnode.color = RED\n\treturn node\n}", "func newNode(nodestr string) (*Node, error) {\n\tif NodeCfg == nil {\n\t\tlog.Fatal(\"must set NodeCfg\")\n\t}\n\tnodestr = strings.TrimSpace(nodestr)\n\tif nodestr == \"\" {\n\t\terr := errors.New(\"nodestr is empty\")\n\t\tlog.Println(err)\n\t\treturn nil, err\n\t}\n\tif match, err := regexp.MatchString(`\\d+/[^: ]+$`, nodestr); !match || err != nil {\n\t\terr := errors.New(fmt.Sprintln(\"bad format\", err, nodestr))\n\t\treturn nil, err\n\t}\n\tn := &Node{\n\t\tNodeConfig: NodeCfg,\n\t\tNodestr: strings.Replace(nodestr, \"+\", \"/\", -1),\n\t}\n\treturn n, nil\n}", "func NewNode(name string, tcpAddr string, job JobType) Node {\n\t// resolve ip\n\tn := util.NewNodeFromTCPAddr(tcpAddr, session.NodeTypeReplica /*dummy field*/)\n\treturn Node{\n\t\tJob: job,\n\t\tName: name,\n\t\tIPPort: n.TCPAddr(),\n\t\tHostname: n.Hostname,\n\t\tAttrs: map[string]interface{}{},\n\t}\n}", "func newNode(parent *node, entry *nodeEntry) *node {\n\treturn &node{\n\t\tparent: parent,\n\t\toccupied: true,\n\t\tevalTotal: entry.eval,\n\t\tcount: 1,\n\t\tentry: entry,\n\t}\n}", "func NewNode(port int) dhtNode {\n\t// Todo: create a node and then return it.\n\treturn dht.NewSurface(port)\n}", "func New(typ NodeType, ns, name string, attrs []Attribute, children ...*Node) *Node {\n\tvar norm []Attribute\n\tvar key string\n\tfor _, v := range attrs {\n\t\tif v.Key == \"key\" {\n\t\t\tkey = expr.Eval(v.Val)\n\t\t} else {\n\t\t\tnorm = append(norm, v)\n\t\t}\n\t}\n\tif len(children) > 0 {\n\t\tnorm = append(norm, Attribute{\n\t\t\tKey: \"children\",\n\t\t\tVal: children,\n\t\t})\n\t}\n\tn := &Node{\n\t\tType: typ,\n\t\tNamespace: ns,\n\t\tKey: key,\n\t\tData: name,\n\t\tAttr: norm,\n\t}\n\treturn n\n}", "func newNode() *node {\n\treturn &node{\n\t\tvalue: nil,\n\t\tchildren: map[string]*node{},\n\t}\n}", "func NewNode(url string) *URLNode {\n\treturn &URLNode{url: url}\n}", "func NewNode(n *pb.Node) *Node {\n\treturn &Node{\n\t\tX: int(n.X),\n\t\tY: int(n.Y),\n\t\tPlayer: int(n.Player),\n\t}\n}", "func CreateNewNode(nodeID int, fileName string, timeStart string, graphID int, fileType string, userID int) (Node, error) {\n\tif nodeID == 0 || fileName == \"\" || timeStart == \"\" || graphID == 0 || fileType == \"\" || userID == 0 {\n\t\treturn Node{}, errors.New(\"Not enough argument supplied\")\n\t}\n\treturn Node{\n\t\tNodeID: nodeID,\n\t\tNodeBits: 0,\n\t\tNodeDesc: fileName,\n\t\tNodeDT: timeStart,\n\t\tNodeGID: graphID,\n\t\tNodeHash: \"\",\n\t\tNodeLevel: -32768,\n\t\tNodeType: fileType,\n\t\tNodeUID: userID,\n\t}, nil\n}", "func NewNode(tr Transaction, id string) *Node {\n\tif tr == nil {\n\t\tpanic(\"transaction may not be nil\")\n\t}\n\tif id == \"\" {\n\t\tid = uuid.NewV4().String()\n\t}\n\n\tn := &Node{\n\t\tTransaction: tr,\n\t\tId: id,\n\t}\n\n\tn.Reset()\n\treturn n\n\n\t/*\n\t\tpos := strings.Index(id, \"-\")\n\t\tif pos == -1 {\n\t\t\tn.Shard = id\n\t\t\tn.UUID = uuid.NewV4().String()\n\t\t} else {\n\t\t\tn.UUID = id[pos+1:]\n\t\t\tn.Shard = id[:pos]\n\t\t}\n\t*/\n\t// return n\n}", "func (g *Graph) NewNode(v interface{}) *Node {\n\tn := &Node{\n\t\tgraph: g,\n\t\tidx: len(g.nodes),\n\t\tValue: v,\n\t}\n\tg.nodes = append(g.nodes, n)\n\treturn n\n}", "func NewNode(e string) *LNode {\n\treturn &LNode{\n\t\tE: e,\n\t}\n}", "func NewNode(key int) *Node {\n\tp := int(rand.Int31n(math.MaxInt32))\n\treturn &Node{key: key, p: p, size: 1}\n}", "func New(key p2pcrypto.PublicKey, address string) Node {\n\treturn Node{key, address}\n}", "func createNode(id int, myConf *Config, sm *State_Machine) cluster.Server {\n\tinitNode(id, myConf, sm)\n\t//Set up details about cluster nodes form json file.\n\tserver, err := cluster.New(id, \"config/cluster_config.json\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn server\n}", "func newNode() *node {\n\treturn &node{}\n}", "func newNode() *topicNode {\n\treturn &topicNode{\n\t\tchildren: children{},\n\t\tclients: make(clientOpts),\n\t\tshared: make(map[string]clientOpts),\n\t}\n}", "func NewNode(weightInitializer WeightInitializer, activationFunction ActivationFunction, derivativeActivation DerivativeActivation, nodeIndex int, learningRate FloatXX, layerIndex int, derivativeError DerivativeError, nodeType NodeType) *Node {\n\treturn &Node{\n\t\tinputNodes: nil,\n\t\toutputNodes: nil,\n\t\tweightInitializer: weightInitializer,\n\t\tactivation: activationFunction,\n\t\tderivativeActivation: derivativeActivation,\n\t\tmyIndex: nodeIndex,\n\t\tlearningRate: learningRate,\n\t\tlayerIndex: layerIndex,\n\t\tderivativeError: derivativeError,\n\t\tweights: []FloatXX{},\n\t\tmyType: nodeType,\n\t\tinputValue: FloatXX(0.0),\n\t\tlabel: FloatXX(0.0),\n\t\tmyErr: FloatXX(0.0),\n\t\tdEdW: []FloatXX{},\n\t\tmyoutCached: false,\n\t\tmyout: FloatXX(0.0),\n\t\tbackCached: false,\n\t\tinputs: []FloatXX{},\n\t\tdEdOut: FloatXX(0.0),\n\t\tdOutdIn: FloatXX(0.0),\n\t\tdIndW: []FloatXX{},\n\t}\n}", "func NewNode(e Element) *Node {\n\tnewNode := NewNilNode()\n\tnewNode.Element = e\n\tnewNode.color = Red\n\tnewNode.IsNil = false\n\treturn newNode\n}", "func NewNode(host, port string) (*Node, error) {\n\tkey := fmt.Sprintf(\"%s:%s\", host, port)\n\tif _, keyFound := readNode(key); keyFound {\n\t\treturn nil, errors.New(\"Node already exists at this host and port\")\n\t}\n\tnode := Node{\n\t\tid: xid.New().String(),\n\t\tHost: host,\n\t\tPort: port,\n\t}\n\twriteNode(key, node)\n\treturn &node, nil\n}", "func NewNode(cnf *Config, joinNode *models.Node) (*Node, error) {\n\tif err := cnf.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\tnode := &Node{\n\t\tNode: new(models.Node),\n\t\tshutdownCh: make(chan struct{}),\n\t\tcnf: cnf,\n\t\tstorage: NewMapStore(cnf.Hash),\n\t}\n\n\tvar nID string\n\tif cnf.Id != \"\" {\n\t\tnID = cnf.Id\n\t} else {\n\t\tnID = cnf.Addr\n\t}\n\tid, err := node.hashKey(nID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\taInt := (&big.Int{}).SetBytes(id)\n\n\tfmt.Printf(\"new node id %d, \\n\", aInt)\n\n\tnode.Node.Id = id\n\tnode.Node.Addr = cnf.Addr\n\n\t// Populate finger table\n\tnode.fingerTable = newFingerTable(node.Node, cnf.HashSize)\n\n\t// Start RPC server\n\ttransport, err := NewGrpcTransport(cnf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tnode.transport = transport\n\n\tmodels.RegisterChordServer(transport.server, node)\n\n\tnode.transport.Start()\n\n\tif err := node.join(joinNode); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Peridoically stabilize the node.\n\tgo func() {\n\t\tticker := time.NewTicker(1 * time.Second)\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ticker.C:\n\t\t\t\tnode.stabilize()\n\t\t\tcase <-node.shutdownCh:\n\t\t\t\tticker.Stop()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\t// Peridoically fix finger tables.\n\tgo func() {\n\t\tnext := 0\n\t\tticker := time.NewTicker(100 * time.Millisecond)\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ticker.C:\n\t\t\t\tnext = node.fixFinger(next)\n\t\t\tcase <-node.shutdownCh:\n\t\t\t\tticker.Stop()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\t// Peridoically checkes whether predecessor has failed.\n\n\tgo func() {\n\t\tticker := time.NewTicker(10 * time.Second)\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ticker.C:\n\t\t\t\tnode.checkPredecessor()\n\t\t\tcase <-node.shutdownCh:\n\t\t\t\tticker.Stop()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn node, nil\n}", "func NewNode(t BTree) (*Node, error) {\n\tn := new(Node)\n\tn.tree = t\n\treturn n, nil\n}", "func New(spaceID, id, parentID, name string, blobsize int64, blobID string, t provider.ResourceType, owner *userpb.UserId, lu PathLookup) *Node {\n\tif blobID == \"\" {\n\t\tblobID = uuid.New().String()\n\t}\n\treturn &Node{\n\t\tSpaceID: spaceID,\n\t\tID: id,\n\t\tParentID: parentID,\n\t\tName: name,\n\t\tBlobsize: blobsize,\n\t\towner: owner,\n\t\tlu: lu,\n\t\tBlobID: blobID,\n\t\tnodeType: &t,\n\t}\n}", "func newDeploymentNode(nodeName string, node v1alpha1.ElasticsearchNode, cluster *v1alpha1.Elasticsearch, roleMap map[v1alpha1.ElasticsearchNodeRole]bool) NodeTypeInterface {\n\tdeploymentNode := deploymentNode{}\n\n\tdeploymentNode.populateReference(nodeName, node, cluster, roleMap, int32(1))\n\n\treturn &deploymentNode\n}", "func NewNode(parent *Node, fi os.FileInfo, withContent bool) (node *Node, err error) {\n\tif fi == nil {\n\t\treturn nil, nil\n\t}\n\n\tvar (\n\t\tsysPath string\n\t\tabsPath string\n\t)\n\n\tif parent != nil {\n\t\tsysPath = filepath.Join(parent.SysPath, fi.Name())\n\t\tabsPath = path.Join(parent.Path, fi.Name())\n\t} else {\n\t\tsysPath = fi.Name()\n\t\tabsPath = fi.Name()\n\t}\n\n\tnode = &Node{\n\t\tSysPath: sysPath,\n\t\tPath: absPath,\n\t\tName: fi.Name(),\n\t\tModTime: fi.ModTime(),\n\t\tMode: fi.Mode(),\n\t\tSize: fi.Size(),\n\t\tV: nil,\n\t\tParent: parent,\n\t\tChilds: make([]*Node, 0),\n\t}\n\n\tif node.Mode.IsDir() || !withContent {\n\t\tnode.Size = 0\n\t\treturn node, nil\n\t}\n\n\terr = node.updateContent()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = node.updateContentType()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn node, nil\n}", "func NewNode(path string, config *TreeConfig) (*Node, error) {\n\tn := &Node{Path: path, Config: config}\n\n\tinfo, err := os.Lstat(n.Path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tn.Info = info\n\n\tif err := n.Recursive(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn n, nil\n}", "func NewNode(ctx *pulumi.Context,\n\tname string, args *NodeArgs, opts ...pulumi.ResourceOption) (*Node, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.AcceleratorType == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'AcceleratorType'\")\n\t}\n\tif args.TensorflowVersion == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'TensorflowVersion'\")\n\t}\n\treplaceOnChanges := pulumi.ReplaceOnChanges([]string{\n\t\t\"location\",\n\t\t\"project\",\n\t})\n\topts = append(opts, replaceOnChanges)\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Node\n\terr := ctx.RegisterResource(\"google-native:tpu/v1:Node\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (config *AppConfig) CreateNode(newNode NodeAttributes) (node Node, err error) {\n\tendpoint := fmt.Sprintf(\"nodes/\")\n\n\tnewNodeBytes, err := json.Marshal(newNode)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// get json bytes from the panel.\n\tnodeBytes, err := config.queryApplicationAPI(endpoint, \"post\", newNodeBytes)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Get server info from the panel\n\t// Unmarshal the bytes to a usable struct.\n\terr = json.Unmarshal(nodeBytes, &node)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn\n}", "func newDirWithNode(n ipld.Node, name string, key []byte) (ipld.Node, error) {\n\tdir := unixfs.EmptyDirNode()\n\tdir.SetCidBuilder(dag.V1CidPrefix())\n\tif err := dir.AddNodeLink(name, n); err != nil {\n\t\treturn nil, err\n\t}\n\treturn encryptNode(dir, key)\n}", "func (n *Node) New(name string, def_val interface{}) (*Node, error) {\n\tif !valid_name(name) {\n\t\treturn nil, errors.New(\"Invalid Name\")\n\t}\n\n\tn.mutex.Lock()\n\tdefer n.mutex.Unlock()\n\n\tif n.nodes[name] != nil {\n\t\treturn nil, errors.New(\"Already exists\")\n\t}\n\n\tnew_node := &Node{\n\t\tname: name,\n\t\tdelimiter: n.delimiter,\n\t\tparent: n,\n\t\tnodes: make(map[string]*Node),\n\t}\n\n\tif def_val != nil {\n\t\tnew_node.def_val = new(string)\n\n\t\tvar str_val string\n\t\tswitch t := def_val.(type) {\n\t\tcase string:\n\t\t\tif val, ok := def_val.(string); ok {\n\t\t\t\tstr_val = val\n\t\t\t} else {\n\t\t\t\treturn nil, errors.New(\"Unabled to assert type string on default value\")\n\t\t\t}\n\t\tcase []byte:\n\t\t\tif val, ok := def_val.([]byte); ok {\n\t\t\t\tstr_val = string(val)\n\t\t\t} else {\n\t\t\t\treturn nil, errors.New(\"Unabled to assert type []byte on default value\")\n\t\t\t}\n\t\tcase fmt.Stringer:\n\t\t\tif val, ok := def_val.(fmt.Stringer); ok {\n\t\t\t\tstr_val = val.String()\n\t\t\t} else {\n\t\t\t\treturn nil, errors.New(\"Unabled to assert type fmt.Stringer on default value\")\n\t\t\t}\n\t\tcase int:\n\t\t\tif val, ok := def_val.(int); ok {\n\t\t\t\tstr_val = strconv.Itoa(val)\n\t\t\t} else {\n\t\t\t\treturn nil, errors.New(\"Unabled to assert type int on default value\")\n\t\t\t}\n\t\tcase float64:\n\t\t\tif val, ok := def_val.(float64); ok {\n\t\t\t\tstr_val = strconv.FormatFloat(val, 'f', -1, 64)\n\t\t\t} else {\n\t\t\t\treturn nil, errors.New(\"Unabled to assert type float64 on default value\")\n\t\t\t}\n\t\tcase bool:\n\t\t\tif val, ok := def_val.(bool); ok {\n\t\t\t\tstr_val = strconv.FormatBool(val)\n\t\t\t} else {\n\t\t\t\treturn nil, errors.New(\"Unabled to assert type bool on default value\")\n\t\t\t}\n\t\tdefault:\n\t\t\treturn nil, errors.New(fmt.Sprintf(\"Unexpected type %T\", t))\n\t\t}\n\n\t\t*new_node.def_val = str_val\n\t}\n\n\tn.nodes[name] = new_node\n\treturn n.nodes[name], nil\n}", "func (gameTree *GameTree) NewNode() *Node {\n\tnode := &Node{}\n\tgameTree.AddNode(node)\n\treturn node\n}", "func (s *Server) CreateNode(w http.ResponseWriter, req *http.Request) {\n\tconfig := peer.RandomNodeConfig()\n\terr := json.NewDecoder(req.Body).Decode(config)\n\tif err != nil && err != io.EOF {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tnode, err := s.network.NewNodeWithConfig(config)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\ts.JSON(w, http.StatusCreated, &node.ID)\n}", "func newNode(token html.Token, parent *Node) *Node {\n\tattrs := make(map[string]*html.Attribute)\n\tfor _, attr := range token.Attr {\n\t\tattrs[attr.Key] = &attr\n\t}\n\tnode := &Node{Token: token, Parent: parent, Attributes: attrs}\n\treturn node\n}", "func NewNode(shift uint32, elements ...Value) *Node {\n\treturn &Node{\n\t\tElements: Fill(elements...),\n\t\tShift: shift,\n\t}\n}", "func NewNode(config NodeConfig) (*TextileNode, error) {\n\t// TODO: shouldn't need to manually remove these\n\trepoLockFile := filepath.Join(config.WalletConfig.RepoPath, fsrepo.LockFile)\n\tos.Remove(repoLockFile)\n\tdsLockFile := filepath.Join(config.WalletConfig.RepoPath, \"datastore\", \"LOCK\")\n\tos.Remove(dsLockFile)\n\n\t// log handling\n\tvar backendFile *logging.LogBackend\n\tif config.LogFiles {\n\t\tw := &lumberjack.Logger{\n\t\t\tFilename: path.Join(config.WalletConfig.RepoPath, \"logs\", \"textile.log\"),\n\t\t\tMaxSize: 10, // megabytes\n\t\t\tMaxBackups: 3,\n\t\t\tMaxAge: 30, // days\n\t\t}\n\t\tbackendFile = logging.NewLogBackend(w, \"\", 0)\n\t} else {\n\t\tbackendFile = logging.NewLogBackend(os.Stdout, \"\", 0)\n\t}\n\tbackendFileFormatter := logging.NewBackendFormatter(backendFile, fileLogFormat)\n\tlogging.SetBackend(backendFileFormatter)\n\tlogging.SetLevel(config.LogLevel, \"\")\n\n\t// create a wallet\n\tconfig.WalletConfig.Version = Version\n\twall, err := wallet.NewWallet(config.WalletConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// setup gateway\n\tgateway := &http.Server{Addr: wall.GetGatewayAddress()}\n\n\t// finally, construct our node\n\tnode := &TextileNode{\n\t\tWallet: wall,\n\t\tgateway: gateway,\n\t}\n\n\treturn node, nil\n}", "func New(id, parentID, name string, blobsize int64, blobID string, owner *userpb.UserId, lu PathLookup) *Node {\n\tif blobID == \"\" {\n\t\tblobID = uuid.New().String()\n\t}\n\treturn &Node{\n\t\tID: id,\n\t\tParentID: parentID,\n\t\tName: name,\n\t\tBlobsize: blobsize,\n\t\towner: owner,\n\t\tlu: lu,\n\t\tBlobID: blobID,\n\t}\n}", "func MakeNode(host, path string, port int) (*Node, error) {\n\tnodestr := net.JoinHostPort(host, strconv.Itoa(port)) + strings.Replace(path, \"+\", \"/\", -1)\n\treturn newNode(nodestr)\n}", "func NewNode(val int) *Node {\n\tn := &Node{}\n\tn.val = val\n\tn.next = nil\n\tn.prev = nil\n\treturn n\n}", "func NewNode(addr string, port int) *Node {\n\tip := net.ParseIP(addr)\n\tpb := make([]byte, 2)\n\tbinary.BigEndian.PutUint16(pb, uint16(port))\n\n\treturn &Node{Address: append(ip, pb...)}\n}", "func NewNode(\n\tconnectionManager connection.Manager,\n\ttequilapiServer tequilapi.APIServer,\n\toriginalLocationCache location.Cache,\n\tmetricsSender *metrics.Sender,\n\tnatPinger NatPinger,\n) *Node {\n\treturn &Node{\n\t\tconnectionManager: connectionManager,\n\t\thttpAPIServer: tequilapiServer,\n\t\toriginalLocationCache: originalLocationCache,\n\t\tmetricsSender: metricsSender,\n\t\tnatPinger: natPinger,\n\t}\n}", "func NewNode(path string) (*Node, error) {\n\t// Open and check node repository\n\tr, err := fsrepo.Open(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Run Node\n\tcfg := &core.BuildCfg{\n\t\tRepo: r,\n\t\tOnline: true,\n\t}\n\n\tctx, cancel := context.WithCancel(context.Background())\n\tnode, err := core.NewNode(ctx, cfg)\n\tif err != nil {\n\t\tcancel()\n\t\treturn nil, err\n\t}\n\n\t// Open Node's DB Instance\n\tdb, err := OpenDB(path + \"/agora.db\")\n\tif err != nil {\n\t\tcancel()\n\t\treturn nil, err\n\t}\n\n\treturn &Node{\n\t\tIpfsNode: node,\n\t\tModel: db,\n\t\tID: node.Identity.Pretty(),\n\t\tcancel: cancel,\n\t}, nil\n}", "func New(commandArgs common.CommandArgs) *Node {\n\n\tnode := Node{\n\t\tNodeCommon: common.NewNodeCommon(commandArgs, \"master\"),\n\t\t// FirstSlaveListenPort: 7500, // TODO(greg) make this an env parameter /TODO this is the *base* port that the new slave should try. incrementing if failing to get the port\n\t}\n\n\treturn &node\n}", "func NewNode(ctx *pulumi.Context,\n\tname string, args *NodeArgs, opts ...pulumi.ResourceOption) (*Node, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.AcceleratorType == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'AcceleratorType'\")\n\t}\n\tif args.TensorflowVersion == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'TensorflowVersion'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Node\n\terr := ctx.RegisterResource(\"gcp:tpu/node:Node\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (p *protoOutputNode) New(attr string) outputNode {\n\tuc := nodePool.Get().(*graph.Node)\n\tuc.Attribute = attr\n\treturn &protoOutputNode{uc}\n}", "func CreateNode(\n\tcomment string, // comment text of the node\n\tnamespace string, // namespace of the node\n\tprocessor SignalProcessor, // processor used to handle signals\n) Node {\n\tn := new(node)\n\tname, tags, commentText := parseInfoFromComment(comment)\n\n\tn.name = name\n\tn.comment = commentText\n\tn.tags = tags\n\n\tn.id = uuid.NewV1().String()\n\tn.seq = n.id\n\tn.namespace = namespace\n\tn.observers = []Observer{}\n\tn.upstreams = map[string]Node{}\n\tn.downstreams = map[string]Node{}\n\tn.processor = processor\n\tn.meta = map[string]string{\n\t\t\"namespace\": namespace,\n\t}\n\n\tn.flowOutputObserver = nil\n\tn.flowFuncs = map[string]FlowFunc{}\n\tn.signalCallbacks = map[string]Callback{}\n\n\treturn n\n}", "func NewNode(data int) *Node {\n\treturn &Node{\n\t\tNext: nil,\n\t\tData: data,\n\t}\n}", "func (cc *ContrailCommand) CreateNode(host vcenter.ESXIHost) error {\n\tlog.Debug(\"Create Node:\", cc.AuthToken)\n\tnodeResource := contrailCommandNodeSync{\n\t\tResources: []*nodeResources{\n\t\t\t{\n\t\t\t\tKind: \"node\",\n\t\t\t\tData: &nodeData{\n\t\t\t\t\tNodeType: \"esxi\",\n\t\t\t\t\tUUID: host.UUID,\n\t\t\t\t\tHostname: host.Hostname,\n\t\t\t\t\tFqName: []string{\"default-global-system-config\", host.Hostname},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\tjsonData, err := json.Marshal(nodeResource)\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Debug(\"Sending Request\")\n\tresp, _, err := cc.sendRequest(\"/sync\", string(jsonData), \"POST\") //nolint: bodyclose\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Debug(\"Got status : \", resp.StatusCode)\n\tswitch resp.StatusCode {\n\tdefault:\n\t\treturn fmt.Errorf(\"resource creation failed, %d\", resp.StatusCode)\n\tcase 200, 201:\n\t}\n\treturn nil\n}", "func (t *BPTree) newNode() *Node {\n\tnode := &Node{\n\t\tKeys: make([][]byte, order-1),\n\t\tpointers: make([]interface{}, order),\n\t\tisLeaf: false,\n\t\tparent: nil,\n\t\tKeysNum: 0,\n\t\tAddress: t.LastAddress,\n\t}\n\tsize := getBinaryNodeSize()\n\tt.LastAddress += size\n\n\treturn node\n}", "func NewNode(tcpAddr *net.TCPAddr) *Node {\n\tn := new(Node)\n\tn.TcpAddr = tcpAddr\n\tn.btcNet = wire.MainNet\n\tn.doneC = make(chan struct{}, 1)\n\n\treturn n\n}", "func (_BaseFactory *BaseFactoryTransactor) CreateNode(opts *bind.TransactOpts, _owner common.Address) (*types.Transaction, error) {\n\treturn _BaseFactory.contract.Transact(opts, \"createNode\", _owner)\n}", "func New(port int, id string, weight int) Node {\n\treturn &node{\n\t\tmyPort: port,\n\t\tid: id,\n\t\tring: *consistent.NewRing(),\n\t\trepCh: make(chan replicaEx),\n\t\treqCh: make(chan requestEx),\n\t\trmvCh: make(chan removeEx),\n\t\tcpyCh: make(chan copyEx),\n\t\treplaceCh: make(chan replaceEx),\n\t\tlookupCh: make(chan lookupEx),\n\t\tbulkCh: make(chan bulkEx),\n\t\tstateCh: make(chan stateEx),\n\t\tweight: weight,\n\t\tstateMap: make(map[string]rpcs.State),\n\t}\n}", "func NewNode(cache *Cache, path string, parent *Node) *Node {\n\tdepth := 0\n\tif parent != nil {\n\t\tdepth = parent.depth + 1\n\t}\n\treturn &Node{\n\t\tcache: cache,\n\t\tstate: NodeStatePending,\n\t\tparent: parent,\n\t\tpath: path,\n\t\tchildren: make(map[string]*Node),\n\t\tdepth: depth,\n\t}\n}", "func NewNode(data int) *Node {\n\treturn &Node{\n\t\tData: data,\n\t\tLeft: nil,\n\t\tRight: nil,\n\t}\n}", "func NewNode(left, right *TreeNode, val int) *TreeNode {\n\treturn &TreeNode{\n\t\tLeft: left,\n\t\tRight: right,\n\t\tVal: val,\n\t}\n}", "func NewBaseNode(o NodeOptions, c *astikit.Closer, eh *EventHandler, s *Stater, target interface{}, et EventTypeTransformer) (n *BaseNode) {\n\t// Create node\n\tn = &BaseNode{\n\t\tc: c.NewChild(),\n\t\tchildren: make(map[string]Node),\n\t\tchildrenStarted: make(map[string]bool),\n\t\tm: &sync.Mutex{},\n\t\teh: eh,\n\t\tet: et,\n\t\to: o,\n\t\toStart: &sync.Once{},\n\t\toStop: &sync.Once{},\n\t\tparents: make(map[string]Node),\n\t\tparentsStarted: make(map[string]bool),\n\t\ts: s,\n\t\tstatus: StatusCreated,\n\t\ttarget: target,\n\t}\n\n\t// Set closer callback\n\tn.c.OnClosed(func(err error) {\n\t\teh.Emit(Event{\n\t\t\tName: et(EventTypeClosed),\n\t\t\tTarget: target,\n\t\t})\n\t})\n\treturn\n}", "func createNode(addr string, nodeStmts map[string]ast.Stmt) *ast.Node {\n\tnID := \"n\" + strings.Replace(addr, \".\", \"\", -1)\n\tnID = strings.Replace(nID, \":\", \"\", -1)\n\tn := &ast.Node{ID: nID}\n\tnodeStmts[n.ID] = &ast.NodeStmt{\n\t\tNode: n,\n\t\tAttrs: []*ast.Attr{\n\t\t\t{\n\t\t\t\tKey: \"label\",\n\t\t\t\tVal: fmt.Sprintf(`\"%s\"`, addr),\n\t\t\t},\n\t\t},\n\t}\n\treturn n\n}", "func NewNode(data int) *Node {\n\treturn &Node{\n\t\tData: data,\n\t\tNext: nil,\n\t}\n}", "func NewNode(v float64, parent *Node) *Node {\n\tn := new(Node)\n\tn.Value = v\n\tn.Children = NewNodeList()\n\tn.Parent = parent\n\tn.Marked = false\n\tn.Next = nil\n\tn.prev = nil\n\treturn n\n}", "func NewNode(nodeID int64) (*Node, error) {\n\tif nodeID < 0 || nodeID > maxNode {\n\t\treturn nil, errors.New(\"Node id must between 0 and \" + strconv.FormatInt(maxNode, 10))\n\t}\n\treturn &Node{\n\t\tsecond: time.Now().Unix() - initSecond,\n\t\tnodeID: nodeID,\n\t\tseqNum: 0,\n\t\trandNum: 0,\n\t\tseed: 1,\n\t}, nil\n}", "func New(ctx context.Context, opts ...BuilderOpt) (*Node, error) {\n\tn := &Builder{}\n\tfor _, o := range opts {\n\t\tif err := o(n); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn n.build(ctx)\n}", "func newNode(val int) *node {\n\treturn &node{val, nil, nil}\n}", "func NewTestNode() {\n\tTestNode = lucky.NewFrameNode()\n\tTestNode.Serve()\n}" ]
[ "0.7093768", "0.7062563", "0.7058369", "0.70189285", "0.70076716", "0.69962263", "0.6984174", "0.6945499", "0.690799", "0.6906533", "0.689603", "0.6859061", "0.68354684", "0.68204737", "0.68075323", "0.6745448", "0.6743017", "0.6740784", "0.67379194", "0.67348075", "0.6731235", "0.6726157", "0.6723214", "0.6722502", "0.6713671", "0.67032504", "0.667769", "0.6666498", "0.6658864", "0.6639911", "0.66335946", "0.6626325", "0.6613184", "0.6595932", "0.6584163", "0.65813893", "0.65795934", "0.6560285", "0.6558292", "0.65362024", "0.6533458", "0.65220886", "0.6518868", "0.651083", "0.6503468", "0.6499517", "0.64974153", "0.6481603", "0.6476919", "0.6464333", "0.6444467", "0.644201", "0.64340526", "0.64257944", "0.64123017", "0.640856", "0.63800275", "0.6375527", "0.63723457", "0.6368201", "0.6366863", "0.6345366", "0.6343502", "0.63368607", "0.63347036", "0.6326971", "0.6324785", "0.6316193", "0.6307918", "0.6285063", "0.6279241", "0.62746114", "0.6265492", "0.6264414", "0.6263707", "0.6244254", "0.6242907", "0.6239999", "0.62279344", "0.6207743", "0.6204477", "0.61900586", "0.61887324", "0.6187462", "0.6183461", "0.61830175", "0.61812526", "0.618019", "0.6168594", "0.61657375", "0.61522126", "0.6149041", "0.6149021", "0.61395085", "0.61378616", "0.6129898", "0.61257166", "0.61167186", "0.6115145", "0.61148167" ]
0.67340374
20
Shutdown does what is on the box
func (n Node) Shutdown() { log.Info("Starting shutdown of %d partitions", len(n.partitions)) for _, parts := range n.partitions { for _, part := range parts { part.Close() } } log.Info("Finished closing all the partitions") }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ShutDown() {\n\tC.shout_shutdown()\n}", "func Shutdown() {\n\tclose(shutdown)\n}", "func (h *healthcheckManager) shutdown() {\n\th.quit <- true\n\t<-h.stopped\n}", "func Shutdown() {\n\t// Currently nothing to do\n}", "func (p *DirectBuy) Shutdown() {\n\tp.log.Info(\"Shutting down DirectBuy\")\n\tclose(p.quit)\n\tp.log.Info(\"Waiting for run to finish\")\n\t<-p.done\n\tp.log.Info(\"Shutdown complete\")\n}", "func (f *FakeVCenter) Shutdown(context.Context) error {\n\treturn nil\n}", "func Shutdown() {\n\tlog.Infof(\"Bot shutting down..\")\n\n\tif len(portfolio.Portfolio.Addresses) != 0 {\n\t\tbot.config.Portfolio = portfolio.Portfolio\n\t}\n\n\tlog.Infof(\"Exiting.\")\n\tos.Exit(0)\n}", "func (h *Hookbot) Shutdown() {\n\tclose(h.shutdown)\n\th.wg.Wait()\n}", "func Shutdown() {\n\tsigchan <- userShutdown{}\n\t<-wait\n}", "func (f *VCenter) Shutdown(context.Context) error {\n\treturn nil\n}", "func (ui *GUI) Shutdown() {\n\tctx, cl := context.WithTimeout(ui.cfg.Ctx, time.Second*5)\n\tdefer cl()\n\tif err := ui.server.Shutdown(ctx); err != nil {\n\t\tlog.Error(err)\n\t}\n}", "func (e *bcsExecutor) Shutdown() {\n\te.isAskedShutdown = true\n\n\t//shutdown\n\te.innerShutdown()\n}", "func Stop() {\n\t// /bin/dbus-send --system --dest=org.ganesha.nfsd --type=method_call /org/ganesha/nfsd/admin org.ganesha.nfsd.admin.shutdown\n}", "func Shutdown() {\n\tdefaultDaemon.Shutdown()\n}", "func (p *Provider) shutdown() error {\n\tcs, err := containersByLabels(map[string]string{\n\t\t\"convox.rack\": p.Name,\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar wg sync.WaitGroup\n\n\tfor _, c := range cs {\n\t\twg.Add(1)\n\t\tgo p.containerStopAsync(c.Id, &wg)\n\t}\n\n\twg.Wait()\n\n\tos.Exit(0)\n\n\treturn nil\n}", "func Quit() {\n\tfor _, service := range Services {\n\t\tDelService(service.Kind)\n\t}\n\n\tConn.SendLine(\"SQUIT %s :Goodbye!\", Info.Sid)\n}", "func Shutdown() {\n\tlog.Println(\"Bot shutting down..\")\n\tbot.config.Portfolio = portfolio.Portfolio\n\terr := bot.config.SaveConfig(bot.configFile)\n\n\tif err != nil {\n\t\tlog.Println(\"Unable to save config.\")\n\t} else {\n\t\tlog.Println(\"Config file saved successfully.\")\n\t}\n\n\tlog.Println(\"Exiting.\")\n\tos.Exit(1)\n}", "func (s *Stock) Shutdown() {\n\t//Note: perform any cleanup here\n}", "func (c *Client) Shutdown() error {\n\tif _, err := c.httpPost(\"system/shutdown\", \"\"); err != nil {\n\t\treturn maskAny(err)\n\t}\n\treturn nil\n}", "func (he *Editor) shutdown() {\n\textendDeadline(he.conn, modules.NegotiateSettingsTime)\n\t// don't care about these errors\n\t_, _ = verifySettings(he.conn, he.host)\n\t_ = modules.WriteNegotiationStop(he.conn)\n\tclose(he.closeChan)\n}", "func (w *worker) shutdown() {\n\tselect {\n\tcase w.shutdownc <- true:\n\tdefault:\n\t}\n}", "func (c *client) Shutdown(ctx context.Context, goodbye bool) error {\n\tq := url.Values{}\n\tif goodbye {\n\t\tq.Set(\"mode\", \"goodbye\")\n\t}\n\turl := c.createURL(\"/shutdown\", q)\n\n\treq, err := http.NewRequest(\"POST\", url, nil)\n\tif err != nil {\n\t\treturn maskAny(err)\n\t}\n\tif ctx != nil {\n\t\treq = req.WithContext(ctx)\n\t}\n\tresp, err := c.client.Do(req)\n\tif err != nil {\n\t\treturn maskAny(err)\n\t}\n\tif err := c.handleResponse(resp, \"POST\", url, nil); err != nil {\n\t\treturn maskAny(err)\n\t}\n\n\treturn nil\n}", "func (m *Machine) Shutdown(ctx context.Context) error {\n\tm.logger.Debug(\"Called machine.Shutdown()\")\n\treturn m.sendCtrlAltDel(ctx)\n}", "func (s *SWIM) ShutDown() {\n\tatomic.CompareAndSwapInt32(&s.stopFlag, AVAILABLE, DIE)\n\ts.messageEndpoint.Shutdown()\n\ts.quitFD <- struct{}{}\n}", "func (api *API) Shutdown() {\n\tc <- os.Interrupt\n}", "func Shutdown() {\n\tstdClient.Close()\n}", "func (rt *RecoveryTracker) Shutdown() {\n}", "func (q *CoreClient) Shutdown(safe bool) (err error) {\n\tif safe {\n\t\t_, err = q.RequestWithoutData(http.MethodPost, \"/safeExit\", nil, nil, 200)\n\t} else {\n\t\t_, err = q.RequestWithoutData(http.MethodPost, \"/exit\", nil, nil, 200)\n\t}\n\treturn\n}", "func (t *Thereum) Shutdown(wg *sync.WaitGroup) {\n\tdefer wg.Done()\n\tt.blockchain.Stop()\n}", "func (gw *Gateway) Shutdown() {\n\tclose(gw.quit)\n\t// wait for strand to complete\n\tgw.strand(\"wait-shutdown\", func() {})\n}", "func (k *KeKahu) Shutdown() (err error) {\n\tinfo(\"shutting down the kekahu service\")\n\n\t// Shutdown the server\n\tif err = k.server.Shutdown(); err != nil {\n\t\tk.echan <- err\n\t}\n\n\t// Notify the run method we're done\n\t// NOTE: do this last or the cleanup proceedure won't be done.\n\tk.done <- true\n\treturn nil\n}", "func Shutdown() {\n\tinstMu.Lock()\n\tdefer instMu.Unlock()\n\tinst.Shutdown()\n\tinst = nil\n\tinitialized = false\n}", "func (b *GithubBridge) Shutdown(ctx context.Context) error {\n\treturn nil\n}", "func (a *App) Shutdown() {\n\ta.shutdown <- syscall.SIGTERM\n}", "func Shutdown() {\n\tlm().shutdown()\n}", "func Shutdown() {\n\tglobalNotifier.Shutdown()\n}", "func (h *Holders) Shutdown() {\n\th.close <- 1\n}", "func Quit() {\n\tscreen.Fini()\n}", "func (ft *fakeTraverser) Shutdown(ctx context.Context) {}", "func (t *Trigger) Shutdown() {\n\tclose(t.closeChan)\n}", "func (g *Glutton) Shutdown() (err error) {\n\tdefer g.logger.Sync()\n\tg.cancel() // close all connection\n\n\t/** TODO:\n\t ** May be there exist a better way to wait for all connections to be closed but I am unable\n\t ** to find. The only link we have between program and goroutines is context.\n\t ** context.cancel() signal routines to abandon their work and does not wait\n\t ** for the work to stop. And in any case if fails then there will be definitely a\n\t ** goroutine leak. May be it is possible in future when we have connection counter so we can keep\n\t ** that counter synchronized with number of goroutines (connections) with help of context and on\n\t ** shutdown we wait until counter goes to zero.\n\t */\n\n\ttime.Sleep(2 * time.Second)\n\treturn g.processor.Shutdown()\n}", "func (a *Axon) Shutdown() {\n\ta.shutdownCh <- syscall.SIGTERM\n}", "func shutdown(exitFunc func(code int), e error) {\n\tif e != nil {\n\t\tvar code int\n\t\tswitch e {\n\t\tcase ErrGotHelp:\n\t\t\tcode = 3\n\t\tcase ErrBadArgs:\n\t\t\tcode = 2\n\t\tdefault:\n\t\t\tcode = 1\n\t\t\tlog.Printf(\"Run error: %s\", e.Error())\n\t\t}\n\t\texitFunc(code)\n\t}\n}", "func (mw *JWTMiddleware) Shutdown(_ context.Context) error {\n\treturn nil\n}", "func (l *pydioObjects) Shutdown(ctx context.Context) error {\n\t// TODO\n\treturn nil\n}", "func (m *wsNotificationManager) Shutdown() {\n\tclose(m.quit)\n}", "func (r *Ring) Shutdown() {\n\tr.stopVnodes()\n\tr.stopDelegate()\n}", "func (p *BlsCosi) Shutdown() error {\n\tp.stoppedOnce.Do(func() {\n\t\tfor _, subCosi := range p.subProtocols {\n\t\t\t// we're stopping the root thus it will stop the children\n\t\t\t// by itself using a broadcasted message\n\t\t\tsubCosi.Shutdown()\n\t\t}\n\t\tclose(p.startChan)\n\t\tclose(p.FinalSignature)\n\t})\n\treturn nil\n}", "func (c *Checker) Shutdown() {\n\tc.shutdown <- struct{}{}\n\tc.wg.Wait()\n}", "func (c *Cluster) Shutdown() {\n\tc.manager.quit()\n}", "func exit(s *sdk.SDK) {\n\tlog.Printf(\"Received EXIT command. Exiting.\")\n\t// This tells Agones to shutdown this Game Server\n\tshutdownErr := s.Shutdown()\n\tif shutdownErr != nil {\n\t\tlog.Printf(\"Could not shutdown\")\n\t}\n\tos.Exit(0)\n}", "func (g *smartContractGW) Shutdown() {\n\tif g.sm != nil {\n\t\tg.sm.Close()\n\t}\n\tif g.rr != nil {\n\t\tg.rr.close()\n\t}\n}", "func (f *FakeWorkQueue) ShutDown() {}", "func (zr *zipkinReceiver) Shutdown(context.Context) error {\n\tvar err error\n\tif zr.server != nil {\n\t\terr = zr.server.Close()\n\t}\n\tzr.shutdownWG.Wait()\n\treturn err\n}", "func (hd *Downloader) shutdown() {\n\textendDeadline(hd.conn, modules.NegotiateSettingsTime)\n\t// don't care about these errors\n\t_, _ = verifySettings(hd.conn, hd.host)\n\t_ = modules.WriteNegotiationStop(hd.conn)\n\tclose(hd.closeChan)\n}", "func onShutdown(cancel context.CancelFunc) {\n\tgo func() {\n\t\ts := make(chan os.Signal, 1)\n\t\tsignal.Notify(s, syscall.SIGINT, syscall.SIGTERM)\n\t\t<-s\n\t\tlog.Info(\"Shutting down gracefully\")\n\t\tcancel()\n\t}()\n}", "func (g *Generator) Shutdown(_ context.Context, _ *proto.EmptyRequest) (*proto.EmptyReply, error) {\n\tg.S.GracefulStop()\n\treturn &proto.EmptyReply{}, nil\n}", "func (e EventStream) Shutdown() {\n\te.ctxCancel()\n}", "func (r *RuntimeImpl) Shutdown() {\n\tr.logger.Info(\"\\n\\n * Starting graceful shutdown\")\n\tr.logger.Info(\" * Waiting goroutines stop\")\n\tif r.slackWriter != nil {\n\t\tmessage, attachments := buildSlackShutdownMessage(r.dashboardTitle, false)\n\t\tr.slackWriter.PostNow(message, attachments)\n\t}\n\tr.syncManager.Stop()\n\tif r.impListener != nil {\n\t\tr.impListener.Stop(true)\n\t}\n\tr.appMonitor.Stop()\n\tr.servicesMonitor.Stop()\n\n\tr.logger.Info(\" * Shutdown complete - see you soon!\")\n\tr.blocker <- struct{}{}\n}", "func (c *Checkpoint) Shutdown() {\n\tc.once.Do(func() {\n\t\tclose(c.done)\n\t\tc.wg.Wait()\n\t})\n}", "func shutdown(root *Task, mainTasks []*Task) func() {\n\treturn func() {\n\t\troot.Logger.Print(\"shutting down by closing Stopped channel on the root\")\n\t\tclose(root.Stopped)\n\t\texitcode := waitForTasksToDie(root, mainTasks)\n\t\tos.Exit(exitcode)\n\t}\n}", "func shutDown(ctx context.Context, logger *log.Logger, srv *http.Server) {\n\tquit := make(chan os.Signal, 1)\n\tsignal.Notify(quit, os.Interrupt)\n\t<-quit\n\n\tlogger.Info(\"msg\", \"Shutting down HTTP/REST gateway server...\")\n\n\tctx, cancel := context.WithTimeout(ctx, 5*time.Second)\n\tdefer cancel()\n\n\tif err := srv.Shutdown(ctx); err != nil {\n\t\tlogger.Error(\"err\", fmt.Sprintf(\"Shutdown HTTP/REST gateway server: %s\", err.Error()))\n\t}\n\n\tlogger.Info(\"msg\", \"Shutdown done HTTP/REST gateway server\")\n}", "func (ws *WindowSurface) Quit() {\n\tws.running = false\n}", "func (v *App) Quit() {\n\tv.running = false\n}", "func (s *Server) Shutdown() {\n\tclose(stop)\n}", "func exit() {\n\tcolor.Red(\"[!]Quitting...\")\n\tlogging.Server(\"Shutting down tullius due to user input\")\n\tos.Exit(0)\n}", "func (flu *FlutterEngine) Shutdown() Result {\n\tres := C.FlutterEngineShutdown(flu.Engine)\n\treturn (Result)(res)\n}", "func (c *Connector) Shutdown() error {\n\tc.NATS().Close()\n\treturn nil\n}", "func (wc *WebClient) Quit() {\n\twc.running = false\n}", "func (s *server) shutdown(forceShutdown chan os.Signal) {\n\tfmt.Println(\"\\x1b[34;1mGracefully shutdown... (press Ctrl+C again to force)\\x1b[0m\")\n\n\tctx, cancel := context.WithTimeout(context.Background(), 1*time.Minute)\n\tdefer cancel()\n\n\tdone := make(chan struct{})\n\tgo func() {\n\t\tdefer close(done)\n\t\tfor _, server := range s.instances {\n\t\t\tserver.Shutdown(ctx)\n\t\t}\n\t\tdone <- struct{}{}\n\t}()\n\n\tselect {\n\tcase <-done:\n\t\tlog.Println(\"\\x1b[32;1mSuccess shutdown all server & worker\\x1b[0m\")\n\tcase <-forceShutdown:\n\t\tlog.Println(\"\\x1b[31;1mForce shutdown server & worker\\x1b[0m\")\n\t\tcancel()\n\tcase <-ctx.Done():\n\t\tlog.Println(\"\\x1b[31;1mContext timeout\\x1b[0m\")\n\t}\n}", "func discordShutdown(discordSession *discordgo.Session, userChannels []*discordgo.Guild) {\n\tfor _, userChannel := range userChannels {\n\t\tdiscordSession.ChannelMessageSend(userChannel.ID, \"Johan Bot Stopping\")\n\t}\n}", "func (d *DynamicSelect) shutDown() {\n\tif r := recover(); r != nil {\n\t\tlog.Printf(\"Recovered from panic in main DynamicSelect: %v\\n\", r)\n\t\tlog.Println(\"Attempting normal shutdown.\")\n\t}\n\n\t// just making sure.\n\td.killHeard = true\n\td.alive = false\n\td.running = false\n\tclose(d.done)\n\n\t// Tell the outside world we're done.\n\td.onKillAction()\n\n\t// Handle outstanding requests / a flood of closed messages.\n\tgo d.drainChannels()\n\n\t// Wait for internal listeners to halt.\n\td.listenerWG.Wait()\n\n\t// Make it painfully clear to the GC.\n\tclose(d.aggregator)\n\tclose(d.priorityAggregator)\n\tclose(d.onClose)\n}", "func (a *App) Shutdown(code int) {\n\tos.Exit(code)\n}", "func OnExit() {\n\tapp.GetIndicator().Disconnect()\n}", "func (m *wasiSnapshotPreview1Impl) sockShutdown(pfd wasiFd, phow wasiSdflags) (err wasiErrno) {\n\terr = wasiErrnoNotsup\n\treturn\n}", "func (hmr *receiver) Shutdown(ctx context.Context) error {\n\tclose(hmr.done)\n\treturn hmr.closeScrapers(ctx)\n}", "func (factory *Factory) Shutdown() {\n\tfactory.shutdownInProgress = true\n\t// If the cleanup flag is present don't do any cleanup\n\tif !factory.options.cleanup {\n\t\treturn\n\t}\n\n\t// Wait 15 seconds before running all shutdown handlers to ensure everything can catch up.\n\ttime.Sleep(15 * time.Second)\n\terr := factory.CleanupChaosMeshExperiments()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tfactory.invariantShutdownHooks.InvokeShutdownHandlers()\n\tfactory.shutdownHooks.InvokeShutdownHandlers()\n}", "func (bft *ProtocolBFTCoSi) Shutdown() error {\n\tdefer func() {\n\t\t// In case the channels were already closed\n\t\trecover()\n\t}()\n\tbft.setClosing()\n\tclose(bft.announceChan)\n\tclose(bft.challengePrepareChan)\n\tclose(bft.challengeCommitChan)\n\tif !bft.IsLeaf() {\n\t\tclose(bft.commitChan)\n\t\tclose(bft.responseChan)\n\t}\n\treturn nil\n}", "func (workPool *WorkPool) Shutdown(goRoutine string) (err error) {\n\tdefer catchPanic(&err, goRoutine, \"Shutdown\")\n\n\twriteStdout(goRoutine, \"Shutdown\", \"Started\")\n\twriteStdout(goRoutine, \"Started\", \"Queue Routine\")\n\n\tworkPool.shutdownQueueChannel <- \"Down\"\n\t<-workPool.shutdownQueueChannel\n\n\tclose(workPool.queueChannel)\n\tclose(workPool.shutdownQueueChannel)\n\n\twriteStdout(goRoutine, \"Shutdown\", \"Shutting Down Work Routines\")\n\n\tclose(workPool.shutdownWorkChannel)\n\tworkPool.shutdownWaitGroup.Wait()\n\n\tclose(workPool.workChannel)\n\n\twriteStdout(goRoutine, \"Shutdown\", \"Completed\")\n\n\treturn\n}", "func (_m *mockCopyCat) Shutdown() {\n\t_m.Called()\n}", "func _shutdown(w http.ResponseWriter, r *http.Request) {\n\tfmt.Println(\"Shutdown!\")\n\t//http.Shutdown(nil)\n}", "func (tc *Target) Shutdown() {\n\tclose(tc.shutdown)\n}", "func (brw *blockRetrievalWorker) Shutdown() {\n\tselect {\n\tcase <-brw.stopCh:\n\tdefault:\n\t\tclose(brw.stopCh)\n\t}\n}", "func (p *Proxy) Shutdown() {\n\tlog.Info(\"Shutting down server gracefully\")\n\tclose(p.shutdown)\n\tgraceful.Shutdown()\n\tp.gRPCStop()\n}", "func (b *BruteBlocker) ShutDown() {\n\tb.requests <- func() {\n\t\tb.shutdown = true\n\t}\n}", "func (m *Monitor) Shutdown() {\n\tif m.cancel != nil {\n\t\tm.cancel()\n\t}\n\tif m.plugin != nil {\n\t\tm.plugin.Stop()\n\t}\n}", "func (s *Server) Shutdown() {\n\ts.quit <- syscall.SIGINT\n}", "func (a *AbstractSSHConnectionHandler) OnShutdown(_ context.Context) {}", "func Stop() {\n\tclose(shutdownChannel)\n\tshutdownWaitGroup.Wait()\n\tlog4go.Info(\"Console shutdown complete\")\n}", "func (s *testDoQServer) Shutdown() {\n\t_ = s.listener.Close()\n}", "func cleanup(queue webhook.MessagingQueue) {\n\n\tsigChan := make(chan os.Signal)\n\tsignal.Notify(sigChan, syscall.SIGINT, syscall.SIGKILL, syscall.SIGTERM, syscall.SIGTSTP)\n\t<-sigChan\n\n\tfmt.Println(\"\\nReceived an interrupt, stopping services...\\n\")\n\tqueue.Close()\n\tiris.Close()\n\n\truntime.GC()\n\tos.Exit(0)\n\n}", "func (d *DiscordAPI) Shutdown() {\n\tLogger.Warn(\"Discord is shutting down\")\n\td.discord.Close()\n}", "func Shutdown() {\n\tlock.Lock()\n\tdefer lock.Unlock()\n\tfor _, v := range instances {\n\t\tv.Shutdown()\n\t}\n}", "func (sr *sapmReceiver) Shutdown(context.Context) error {\n\tif sr.server == nil {\n\t\treturn nil\n\t}\n\terr := sr.server.Close()\n\tsr.shutdownWG.Wait()\n\treturn err\n}", "func (gate *Gate) Shutdown() {\n\tctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)\n\tdefer cancel()\n\n\tif err := gate.srv.Shutdown(ctx); err != nil {\n\t\tpanic(\"Server Shutdown Error : \" + err.Error())\n\t}\n\tfmt.Println(\"Server Exit\")\n}", "func (fileSystem *FileSystem) Quit() {\n\n\terr := fileSystem.flushDirectory()\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn\n\t}\n\n\t(*fileSystem).PfsFile.Close()\n\tfmt.Println(\"Successfully exited file system\")\n}", "func (t *transport) shutdown() {\n\tt.mux.Lock()\n\tfor _, cli := range t.remotes {\n\t\tcli.Close()\n\t}\n\tt.remotes = nil\n\tt.mux.Unlock()\n}", "func (s *server) shutdown(ctx context.Context) error {\n\treturn s.server.Shutdown(ctx)\n}", "func (s *Stream) Shutdown() {\n\ts.log.Log(\"Sumex.Streams\", \"Shutdown\", \"Started : Shutdown Requested\")\n\tif atomic.LoadInt64(&s.closed) == 0 {\n\t\ts.log.Log(\"Sumex.Streams\", \"Stats\", \"Info : Shutdown Request : Previously Done\")\n\t\treturn\n\t}\n\n\tif pen := atomic.LoadInt64(&s.pending); pen > 0 {\n\t\ts.log.Log(\"Sumex.Streams\", \"Shutdown\", \"Info : Pending : %d\", pen)\n\t\tatomic.StoreInt64(&s.shutdownAfterpending, 1)\n\t\tatomic.StoreInt64(&s.closed, 1)\n\t}\n\n\ts.log.Log(\"Sumex.Streams\", \"Shutdown\", \"Started : WaitGroup.Wait()\")\n\ts.wg.Wait()\n\ts.log.Log(\"Sumex.Streams\", \"Shutdown\", \"Completed : WaitGroup.Wait()\")\n\n\tclose(s.data)\n\tclose(s.err)\n\tclose(s.nc)\n\tatomic.StoreInt64(&s.closed, 1)\n\ts.log.Log(\"Sumex.Streams\", \"Shutdown\", \"Completed : Shutdown Requested\")\n}", "func notifyOnShutdown(s *http.Server) {\n\tsignalChan := make(chan os.Signal, 1)\n\tsignal.Notify(signalChan, os.Interrupt, syscall.SIGINT, syscall.SIGTERM)\n\n\t<-signalChan\n\n\tctx, cancel := context.WithCancel(context.Background())\n\tdefer cancel()\n\n\ts.SetKeepAlivesEnabled(false)\n\ts.Shutdown(ctx)\n\n\tlog.Printf(\"shutting down gracefully\")\n\tos.Exit(0)\n}", "func (m *Info) Shutdown() {\n\tm.cancel()\n}" ]
[ "0.6944937", "0.6943444", "0.690774", "0.6876422", "0.68660134", "0.6862036", "0.68073046", "0.67971396", "0.67509586", "0.67452574", "0.6728236", "0.67253536", "0.67180264", "0.66992676", "0.66873753", "0.66680455", "0.6665496", "0.666321", "0.6582819", "0.6554132", "0.64983344", "0.6450459", "0.6448364", "0.64383316", "0.6434502", "0.6423334", "0.6399002", "0.63982147", "0.6396699", "0.6396525", "0.63879114", "0.6372684", "0.63708836", "0.6367162", "0.6365868", "0.635403", "0.63348323", "0.6330992", "0.6326117", "0.6321464", "0.6315307", "0.6294187", "0.62931204", "0.62922066", "0.62840915", "0.6281493", "0.6280328", "0.62727016", "0.6267973", "0.6267632", "0.62645364", "0.6263388", "0.6257868", "0.62543344", "0.62472385", "0.624456", "0.62351274", "0.62337273", "0.6223923", "0.6220945", "0.6220862", "0.62174624", "0.62159204", "0.62149084", "0.62144053", "0.62038815", "0.6201333", "0.62010443", "0.61987394", "0.6191386", "0.61886764", "0.6183579", "0.6179999", "0.61741173", "0.6172137", "0.6169812", "0.61600643", "0.61419326", "0.61402726", "0.614022", "0.6126447", "0.6125689", "0.6123588", "0.6119294", "0.6115175", "0.60952634", "0.6095061", "0.60806", "0.6067671", "0.6066205", "0.60627544", "0.606172", "0.60609007", "0.6056394", "0.6054314", "0.6053451", "0.6050666", "0.60481894", "0.60477287", "0.6046283", "0.60453784" ]
0.0
-1
Subscribe will create a channel that will be published to each time a change on ONE partition changes
func (n *Node) Subscribe(topic string) <-chan Message { consumer := make(chan Message) parts := n.getPartitionsFor(topic) var part partition for _, p := range parts { part = smallestOf(part, p) } log.Debug("Created subscriber for %s:%v", topic, part) part.Subscribe(consumer) return consumer }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c Conference) Subscribe(id uuid.UUID, topic string, out chan Notification) {\n\ttree, ok := c.Room[topic]\n\tif !ok {\n\t\ttree = &bst.BinarySearchTree{}\n\t\tc.Room[topic] = tree\n\t}\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\ttree.Add(NewSubscriber(id, out))\n}", "func (h *Hub) Subscribe(t string, c chan []byte) {\n\th.Lock()\n\tdefer h.Unlock()\n\tchans, ok := h.topicChans[t]\n\tif !ok {\n\t\tchans = make(map[chan []byte]struct{})\n\t\th.topicChans[t] = chans\n\t}\n\tchans[c] = struct{}{}\n\ttopics, ok := h.chanTopics[c]\n\tif !ok {\n\t\ttopics = make(map[string]struct{})\n\t\th.chanTopics[c] = topics\n\t}\n\ttopics[t] = struct{}{}\n}", "func (eb *EventBus) Subscribe(topic string, ch DataChannel) {\n\teb.rm.Lock()\n\tif prev, found := eb.subscribers[topic]; found {\n\t\teb.subscribers[topic] = append(prev, ch)\n\t} else {\n\t\teb.subscribers[topic] = append([]DataChannel{}, ch)\n\t}\n\teb.rm.Unlock()\n}", "func (t *Topic) Subscribe(ctx context.Context) <-chan interface{} {\n\tch := make(chan interface{})\n\tt.subs[ch] = ctx\n\treturn ch\n}", "func (r *reconciler) subscribeToBrokerChannel(ctx context.Context, t *v1alpha1.Trigger, brokerTrigger, brokerIngress *v1alpha1.Channel, svc *corev1.Service) (*v1alpha1.Subscription, error) {\n\turi := &url.URL{\n\t\tScheme: \"http\",\n\t\tHost: names.ServiceHostName(svc.Name, svc.Namespace),\n\t\tPath: path.Generate(t),\n\t}\n\texpected := resources.NewSubscription(t, brokerTrigger, brokerIngress, uri)\n\n\tsub, err := r.getSubscription(ctx, t)\n\t// If the resource doesn't exist, we'll create it\n\tif k8serrors.IsNotFound(err) {\n\t\tsub = expected\n\t\terr = r.client.Create(ctx, sub)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn sub, nil\n\t} else if err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Update Subscription if it has changed. Ignore the generation.\n\texpected.Spec.DeprecatedGeneration = sub.Spec.DeprecatedGeneration\n\tif !equality.Semantic.DeepDerivative(expected.Spec, sub.Spec) {\n\t\t// Given that spec.channel is immutable, we cannot just update the Subscription. We delete\n\t\t// it and re-create it instead.\n\t\terr = r.client.Delete(ctx, sub)\n\t\tif err != nil {\n\t\t\tlogging.FromContext(ctx).Info(\"Cannot delete subscription\", zap.Error(err))\n\t\t\tr.recorder.Eventf(t, corev1.EventTypeWarning, subscriptionDeleteFailed, \"Delete Trigger's subscription failed: %v\", err)\n\t\t\treturn nil, err\n\t\t}\n\t\tsub = expected\n\t\terr = r.client.Create(ctx, sub)\n\t\tif err != nil {\n\t\t\tlogging.FromContext(ctx).Info(\"Cannot create subscription\", zap.Error(err))\n\t\t\tr.recorder.Eventf(t, corev1.EventTypeWarning, subscriptionCreateFailed, \"Create Trigger's subscription failed: %v\", err)\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn sub, nil\n}", "func (channel Channel) subscribe(observers ...Observer) {\n\tchannel.checkChannelMap()\n\tfor _, observer := range observers {\n\t\tchannel.observers[observer.id] = observer\n\t\tfmt.Printf(\"New observer %s subscribed in channel %s \\n\", observer.id, channel.id)\n\t}\n}", "func (hc *Conn) Subscribe(topic string) {\n\thc.parent.subscribe <- subscription{topic, hc}\n}", "func (s *Cluster) NotifySubscribe(conn security.ID, ssid subscription.Ssid) {\n\tevent := SubscriptionEvent{\n\t\tPeer: s.name,\n\t\tConn: conn,\n\t\tSsid: ssid,\n\t}\n\n\t// Add to our global state\n\ts.state.Add(event.Encode())\n\t// Create a delta for broadcasting just this operation\n\top := newSubscriptionState()\n\top.Add(event.Encode())\n\ts.gossip.GossipBroadcast(op)\n}", "func SubscribeChannel(socket *websocket.Conn) {\n\tfmt.Printf(\"Subscribed channel by: %#v\\n\", socket)\n\n\t// TODO: query rethinkDB with the feature: changefeed,\n\t// it'll look up initial channels, then keep\n\t// blocking and waiting for channel changes such as ADD, REMOVE, or EDIT\n\tfor {\n\t\ttime.Sleep(time.Second * 1)\n\n\t\tmsg := models.Message{\n\t\t\t\"channel add\",\n\t\t\tmodels.Channel{\"1\", \"Software Support\"}}\n\t\tsocket.WriteJSON(msg)\n\t\tfmt.Println(\"sent newly added channel.\")\n\t}\n}", "func (psc *PubSubChannel) Subscribe() *Subscription {\n psc.subsMutex.Lock()\n defer psc.subsMutex.Unlock()\n newSubscription := newSubscription(strconv.Itoa(psc.nextId))\n psc.nextId++\n psc.subscriptions = append(psc.subscriptions, &newSubscription)\n newSubscription.removeSub = func() {\n psc.subsMutex.Lock()\n defer psc.subsMutex.Unlock()\n\n for i, subscription := range psc.subscriptions {\n if subscription.signal == newSubscription.signal {\n fmt.Println(\"Unsubscribing\", \"id\", subscription.id)\n subscription.closed = true\n psc.subscriptions = append(psc.subscriptions[:i], psc.subscriptions[i+1:]...)\n }\n }\n }\n return &newSubscription\n}", "func (p RPCServer) Subscribe(ctx context.Context, in *pb.SubscriptionRequest) (*pb.Subscription, error) {\n\tsubID := *p.currentSubID\n\t*p.currentSubID++\n\n\tlogrus.WithField(\"topic\", in.Topic).WithField(\"subID\", subID).Debug(\"subscribed to new messages\")\n\n\tp.subChannels[subID] = make(chan []byte)\n\tp.cancelChannels[subID] = make(chan bool)\n\n\ts, err := p.service.RegisterHandler(in.Topic, func(b []byte) error {\n\t\tselect {\n\t\tcase p.subChannels[subID] <- b:\n\t\tdefault:\n\t\t}\n\t\treturn nil\n\t})\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tp.subscriptions[subID] = s\n\n\treturn &pb.Subscription{ID: subID}, nil\n}", "func (k *KafkaBroker) Subscribe(topic string, output chan<- Message, offset int64) (err error) {\n\tconsumer, err := sarama.NewConsumerFromClient(k.Client)\n\n\tif err != nil {\n\t\tpanic(\"Got an error while trying to create a consumer: \" + err.Error())\n\t}\n\n\tconn, err := consumer.ConsumePartition(\n\t\ttopic,\n\t\t0,\n\t\toffset, // Start from the next unread message\n\t)\n\n\tif err != nil {\n\t\tpanic(\"Got an error while trying to consume a partition: \" + err.Error())\n\t}\n\n\tgo func() {\n\t\tfor msg := range conn.Messages() {\n\t\t\toutput <- Message{\n\t\t\t\tKey: msg.Key,\n\t\t\t\tValue: msg.Value,\n\t\t\t\tOffset: msg.Offset,\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn err\n}", "func (d StaticAgentDiscovery) Subscribe(c chan<- []string) { go func() { c <- d }() }", "func (n *notifier) Subscribe(ch chan<- []Update) {\n\tn.mu.Lock()\n\tdefer n.mu.Unlock()\n\tn.subs = append(n.subs, ch)\n}", "func Subscribe(topic string) (<-chan []byte, error) {\n\treturn Default.Subscribe(topic)\n}", "func (p *pahoClient) Subscribe(c chan error, topic string, qos uint8, callback CallbackHandler) {\n\thandler := func(i paho.Client, message paho.Message) {\n\t\tlog.Printf(\"RECEIVED - Topic: %s, Message Length: %d bytes\", message.Topic(), len(message.Payload()))\n\t\tif callback != nil {\n\t\t\tcallback(context.Background(), topic, p.clientID, message.Payload())\n\t\t}\n\t}\n\ttoken := p.client.Subscribe(topic, qos, handler)\n\tc <- p.waitForToken(token)\n}", "func (s *Subscription) Subscribe(channels ...string) {\n\tif len(channels) < 1 {\n\t\treturn\n\t}\n\n\ts.broker.dataChan <- &envData{false, &envSubscription{true, s, channels}, nil}\n}", "func (s *SyncStorage) SubscribeChannel(ns string, cb func(string, ...string), channels ...string) error {\n\tnsPrefix := getNsPrefix(ns)\n\treturn s.getDbBackend(ns).SubscribeChannelDB(cb, s.setNamespaceToChannels(nsPrefix, channels...)...)\n}", "func Subscribe(conn net.Conn, command []string, pubsub *PubSub) {\n\n\tfmt.Println(\"SUBSCRIBE TO:\", command[1:])\n\n\tch := make(chan string)\n\n\tdefer func() {\n\t\tconn.Close()\n\t\tpubsub.Unsubscribe <- UnsubscribeEvent{command[1], ch}\n\t}()\n\n\tpubsub.Subscribe <- SubscribeEvent{command[1], ch}\n\n\tfor msg := range ch {\n\t\t//fmt.Fprintf(conn, \"%s\\n\", msg)\n\t\t_, err := conn.Write([]byte(msg + \"\\n\"))\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n}", "func (h *hub) Subscribe(name, topic string, lease time.Duration) (<-chan Message, error) {\n\tselect {\n\tcase <-h.root.ctx.Done():\n\t\treturn nil, errors.New(\"Cannot subscribe to a closed hub.\");\n\tdefault:\n\t}\n\n\tdeadline := time.Now().Add(lease)\n\tcomm := make(chan Message)\n\tvar expandedTopic []string\n\n\tif topic == rootName {\n\t\texpandedTopic = []string{rootName}\n\t} else {\n\t\texpandedTopic = strings.Split(topic, topicDelimeter)\n\t}\n\n\th.sub <- &subscription{\n\t\tTopic: expandedTopic,\n\t\tName: name,\n\t\tDeadline: deadline,\n\t\tClient: comm,\n\t}\n\n\treturn comm, nil\n}", "func (repo *feedRepository) Subscribe(f *feed.Feed, channelname string) error {\n\t_, err := repo.db.Exec(`\n\t\tINSERT INTO feed_subscriptions (feed_id,channel_username)\n\t\tVALUES ($1, $2)\n\t\tON CONFLICT DO NOTHING\n\t\t`, f.ID, channelname)\n\tconst foreignKeyViolationErrorCode = pq.ErrorCode(\"23503\")\n\tif err != nil {\n\t\tif pgErr, isPGErr := err.(pq.Error); !isPGErr {\n\t\t\tif pgErr.Code != foreignKeyViolationErrorCode {\n\t\t\t\treturn feed.ErrChannelNotFound\n\t\t\t}\n\t\t\treturn fmt.Errorf(\"insertion of user failed because of: %s\", err.Error())\n\t\t}\n\t}\n\treturn nil\n}", "func (ks *KafkaStream) Subscribe(close chan struct{}) (<-chan Event, error) {\n\t// Capture current state of stream for use throughout this connection\n\ttopic := ks.config.Topic\n\toffset := ks.config.Offset\n\tstreamPartitions, err := ks.consumer.Partitions(topic)\n\t// Set up return channel for subscription events\n\tevents := make(chan Event, ks.config.SubscribeBufferSize)\n\tif err != nil {\n\t\treturn events, err\n\t}\n\t// Start subscription to stream in background\n\tgo func() {\n\t\t// For each partition in the stream set up a consumer to subscribe to messages\n\t\t// published to that partition\n\t\tfor _, partition := range streamPartitions {\n\t\t\tpartitionConsumer, err := ks.consumer.ConsumePartition(topic, partition, offset)\n\t\t\tif err != nil {\n\t\t\t\tks.logger.Errorf(\"Subscribe: Error %s to starting consumer for partition %d\", partition, err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// Start goroutine to run until the close channel is closed by the caller\n\t\t\tgo func(partitionConsumer sarama.PartitionConsumer) {\n\t\t\t\t<-close\n\t\t\t\tks.logger.Debug(\"Subscribe: Received close signal\")\n\t\t\t\t// at which point the connection to this partition consumer should be closed\n\t\t\t\tpartitionConsumer.AsyncClose()\n\t\t\t}(partitionConsumer)\n\t\t\t// Start goroutine to run until the close channel is closed by the caller\n\t\t\tgo func(partitionConsumer sarama.PartitionConsumer) {\n\t\t\t\t// to consume and convert messages for the subscriber to receive\n\t\t\t\tfor message := range partitionConsumer.Messages() {\n\t\t\t\t\tevent := convertMessageToEvent(message, topic)\n\t\t\t\t\tks.logger.Debugf(\"Subscribe: Received event %+v\", event)\n\t\t\t\t\tevents <- event\n\t\t\t\t}\n\t\t\t}(partitionConsumer)\n\t\t}\n\t}()\n\n\treturn events, nil\n}", "func (kew *KeyspaceEventWatcher) Subscribe() chan *KeyspaceEvent {\n\tkew.subsMu.Lock()\n\tdefer kew.subsMu.Unlock()\n\tc := make(chan *KeyspaceEvent, 2)\n\tkew.subs[c] = struct{}{}\n\treturn c\n}", "func (ps *PubSub) Subscribe(channel ...interface{}) error {\n\tps.conn.Send(\"SUBSCRIBE\", channel...)\n\treturn ps.conn.Flush()\n}", "func (cg *CandlesGroup) subscribe() {\n\tfor _, symb := range cg.symbols {\n\t\tmessage := candlesSubsMessage{\n\t\t\tEvent: eventSubscribe,\n\t\t\tChannel: \"candles\",\n\t\t\tKey: \"trade:1m:t\" + strings.ToUpper(symb.OriginalName),\n\t\t}\n\n\t\tif err := cg.wsClient.Write(message); err != nil {\n\t\t\tlog.Printf(\"[BITFINEX] Error subsciring to %v candles\", symb.Name)\n\t\t\tcg.restart()\n\t\t\treturn\n\t\t}\n\t}\n\tlog.Println(\"[BITFINEX] Subscription ok\")\n}", "func (ps *PubSub) Subscribe(conn Conn, channel string) {\n\tps.subscribe(conn, false, channel)\n}", "func (c *Easee) subscribe(client signalr.Client) {\n\tstateC := make(chan signalr.ClientState, 1)\n\t_ = client.ObserveStateChanged(stateC)\n\n\tgo func() {\n\t\tfor state := range stateC {\n\t\t\tif state == signalr.ClientConnected {\n\t\t\t\tif err := <-client.Send(\"SubscribeWithCurrentState\", c.charger, true); err != nil {\n\t\t\t\t\tc.log.ERROR.Printf(\"SubscribeWithCurrentState: %v\", err)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n}", "func (p *Peer) subscribeSync(po int) {\n\terr := subscriptionFunc(p.streamer, p.ID(), uint8(po))\n\tif err != nil {\n\t\tlog.Error(\"subscription\", \"err\", err)\n\t}\n}", "func (hubPtr *Hub) Subscribe(sink chan<- accounts.WalletEvent) event.Subscription {\n\t// We need the mutex to reliably start/stop the update loop\n\thubPtr.stateLock.Lock()\n\tdefer hubPtr.stateLock.Unlock()\n\n\t// Subscribe the Called and track the subscriber count\n\tsub := hubPtr.updateScope.Track(hubPtr.updateFeed.Subscribe(sink))\n\n\t// Subscribers require an active notification loop, start it\n\tif !hubPtr.updating {\n\t\thubPtr.updating = true\n\t\tgo hubPtr.updater()\n\t}\n\treturn sub\n}", "func (s *Store) Subscribe(id string, up chan uint64) {\n\ts.subs = append(s.subs, subscription{id: id, ch: up})\n}", "func (eb *EventBus) Subscribe(topic string, ch chan<- Event) {\n\teb.rm.Lock()\n\n\tif prev, found := eb.subscribers[topic]; found {\n\t\teb.subscribers[topic] = append(prev, ch)\n\t} else {\n\t\teb.subscribers[topic] = append([]chan<- Event{}, ch)\n\t}\n\n\teb.rm.Unlock()\n}", "func (node *Node) Subscribe(ctx context.Context, project string) error {\n\tif !node.IsOnline() {\n\t\treturn ErrOffline\n\t}\n\tsub, err := node.sh.PubSubSubscribe(project)\n\tif err != nil {\n\t\treturn err\n\t}\n\tnode.subscription = sub\n\tnode.project = project\n\treturn nil\n}", "func (c Client) Subscribe(channel string, market string) error {\n\n\tdata := ws.Message{\n\t\tBaseOperation: ws.BaseOperation{\n\t\t\tOp: \"subscribe\",\n\t\t},\n\t\tArgs: []string{},\n\t}\n\n\tmessage, err := json.Marshal(data)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = c.ws.WriteMessage(1, message)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *Client) Subscribe(topic string, callback SubscriptionCallback) error {\n\tif c.connected == false {\n\t\treturn ErrNotConnected\n\t}\n\n\tbuff := bytes.NewBuffer(make([]byte, 0, 1+4+len(topic)))\n\n\terr := buff.WriteByte(SUBSCRIBE)\n\tif err != nil {\n\t\tfmt.Println(\"error writing packet type\")\n\t\treturn err\n\t}\n\n\terr = binary.Write(buff, binary.LittleEndian, uint32(len(topic)))\n\tif err != nil {\n\t\tfmt.Println(\"error writing topic length\")\n\t\treturn err\n\t}\n\n\t_, err = buff.WriteString(topic)\n\tif err != nil {\n\t\tfmt.Println(\"error writing topic data\")\n\t\treturn err\n\t}\n\n\tif callback != nil {\n\t\tc.subs[topic] = callback\n\t}\n\n\tbuff.WriteTo(c.conn)\n\n\treturn nil\n}", "func subscribe(t *testing.T, con *websocket.Conn, eventid string) {\n\terr := con.WriteJSON(rpctypes.RPCRequest{\n\t\tJSONRPC: \"2.0\",\n\t\tID: \"\",\n\t\tMethod: \"subscribe\",\n\t\tParams: []interface{}{eventid},\n\t})\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n}", "func (psc *PubSubChannel) Publish(msg interface{}) {\n psc.subsMutex.RLock()\n defer psc.subsMutex.RUnlock()\n for i, sub := range psc.subscriptions {\n if sub.closed {\n fmt.Println(\"Subscription was closed\", \"id\", sub.id)\n sub.closeChannel()\n } else {\n go func(sub *Subscription, index int) {\n sub.signal <- msg\n }(sub, i)\n\n }\n }\n}", "func (s *RedisStore) Subscribe(channel string, callback func(notifier.Notification)) error {\n\tpubsub := s.Client.Subscribe(channel)\n\tdefer pubsub.Close()\n\n\tfor {\n\t\tv, err := pubsub.ReceiveMessage()\n\t\tif err != nil {\n\t\t\tlog.WithError(err).Debug(\"An error occurred when getting the message\")\n\t\t\treturn err\n\t\t}\n\t\tnotification := notifier.Notification{}\n\t\tif marshallErr := json.Unmarshal([]byte(v.Payload), &notification); marshallErr != nil {\n\t\t\tlog.WithError(marshallErr).Error(\"Unmarshalling message body failed, malformed\")\n\t\t\treturn marshallErr\n\t\t}\n\t\tcallback(notification)\n\t}\n}", "func (this *Publisher) Subscribe() <-chan gopi.Event {\n\tthis.Lock()\n\tdefer this.Unlock()\n\n\t// Create channels with a capacity of one\n\tif this.channels == nil {\n\t\tthis.channels = make([]chan gopi.Event, 0, 1)\n\t}\n\t// Return a new channel\n\tchannel := make(chan gopi.Event)\n\tthis.channels = append(this.channels, channel)\n\treturn channel\n}", "func (t *Topic) Subscribe(cb *func(interface{})) {\n\tt.subs = append(t.subs, cb)\n}", "func subscribe(topic string) (<-chan *mqttTypes.Publish, error) {\n\tlog.Printf(\"[DEBUG] subscribe - Subscribing to topic %s\\n\", topic)\n\tsubscription, error := cbBroker.client.Subscribe(topic, cbBroker.qos)\n\tif error != nil {\n\t\tlog.Printf(\"[ERROR] subscribe - Unable to subscribe to topic: %s due to error: %s\\n\", topic, error.Error())\n\t\treturn nil, error\n\t}\n\n\tlog.Printf(\"[DEBUG] subscribe - Successfully subscribed to = %s\\n\", topic)\n\treturn subscription, nil\n}", "func Sub(c mqtt.Client, topic string) {\n\tvar choke = make(chan [2]string)\n\n\tvar f mqtt.MessageHandler = func(client mqtt.Client, msg mqtt.Message) {\n\t\tchoke <- [2]string{msg.Topic(), string(msg.Payload())}\n\t}\n\tfor {\n\t\tif token := c.Subscribe(topic, 0, f); token.Wait() && token.Error() != nil {\n\t\t\tmqtt.ERROR.Println(token.Error())\n\t\t\tos.Exit(1)\n\t\t}\n\t\tfor {\n\t\t\tincoming := <-choke\n\t\t\tmqtt.ERROR.Printf(\"Received:TOPIC: %s\\n\", incoming[0])\n\t\t\twriteFile(incoming[1])\n\t\t}\n\t}\n\n}", "func (tc *consumer) Subscribe(topics map[string]int64) error {\n\tfor topic := range topics {\n\t\tif _, exists := tc.subscribedTopics[topic]; exists {\n\t\t\tlogger.Printf(\"consumer for %s already exists. This is strange\", topic)\n\t\t}\n\t\tlogger.Printf(\"Subscribe %s\", topic)\n\t\ttc.subscribedTopics[topic] = tc.tester.getOrCreateQueue(topic).bindConsumer(tc, true)\n\t\ttc.subscribedTopics[topic].rebalance()\n\t\ttc.subscribedTopics[topic].startLoop(false)\n\t}\n\treturn nil\n}", "func (l *Logs) Subscribe(uuid string) (\n\tchan []byte, chan []byte, error,\n) {\n\tlogChan, err := l.GetChannel(uuid)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tschan, echan := messaging.CreateSubscriptionChannels()\n\tl.nub.Subscribe(logChan, \"\", schan, false, echan)\n\treturn schan, echan, nil\n\n}", "func (m *Module) Subscribe(ctx context.Context, topic string) (<-chan *redis.Message, error) {\n\tm.lock.Lock()\n\tdefer m.lock.Unlock()\n\n\t// Check if subscription already exists. Return corresponding channel if it does.\n\tif sub, p := m.mapping[topic]; p {\n\t\treturn sub.ch, nil\n\t}\n\n\t// Make a redis subscription\n\tpubsub := m.client.Subscribe(context.TODO(), m.getTopicName(topic))\n\tif _, err := pubsub.Receive(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Make a channel to listen for subscriptions\n\tch := pubsub.Channel()\n\tm.mapping[topic] = &subscription{ch, pubsub}\n\treturn ch, nil\n}", "func (m *Module) Subscribe(ctx context.Context, topic string) (<-chan *redis.Message, error) {\n\tm.lock.Lock()\n\tdefer m.lock.Unlock()\n\n\t// Check if subscription already exists. Return corresponding channel if it does.\n\tif sub, p := m.mapping[topic]; p {\n\t\treturn sub.ch, nil\n\t}\n\n\t// Make a redis subscription\n\tpubsub := m.client.Subscribe(context.TODO(), m.getTopicName(topic))\n\tif _, err := pubsub.Receive(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Make a channel to listen for subscriptions\n\tch := pubsub.Channel()\n\tm.mapping[topic] = &subscription{ch, pubsub}\n\treturn ch, nil\n}", "func subscribe(qname string) {\n\tpubsub := redisClient.Subscribe(qname)\n\tdefer pubsub.Close()\n\n\tfor {\n\t\tmsg, err := pubsub.ReceiveMessage()\n\t\tif err != nil {\n\t\t\terrLogger.Println(err)\n\t\t}\n\t\trepo, tag := repoTagFromFullName(msg.Payload)\n\t\tqChan <- struct{}{}\n\t\tgo cover(repo, tag, false)\n\t}\n}", "func (by *Bybit) Subscribe(channelsToSubscribe []stream.ChannelSubscription) error {\n\tvar errs error\n\tfor i := range channelsToSubscribe {\n\t\tvar subReq WsReq\n\t\tsubReq.Topic = channelsToSubscribe[i].Channel\n\t\tsubReq.Event = sub\n\n\t\tformattedPair, err := by.FormatExchangeCurrency(channelsToSubscribe[i].Currency, asset.Spot)\n\t\tif err != nil {\n\t\t\terrs = common.AppendError(errs, err)\n\t\t\tcontinue\n\t\t}\n\t\tif channelsToSubscribe[i].Channel == wsKlines {\n\t\t\tsubReq.Parameters = WsParams{\n\t\t\t\tSymbol: formattedPair.String(),\n\t\t\t\tIsBinary: true,\n\t\t\t\tKlineType: \"1m\",\n\t\t\t}\n\t\t} else {\n\t\t\tsubReq.Parameters = WsParams{\n\t\t\t\tSymbol: formattedPair.String(),\n\t\t\t\tIsBinary: true,\n\t\t\t}\n\t\t}\n\t\terr = by.Websocket.Conn.SendJSONMessage(subReq)\n\t\tif err != nil {\n\t\t\terrs = common.AppendError(errs, err)\n\t\t\tcontinue\n\t\t}\n\t\tby.Websocket.AddSuccessfulSubscriptions(channelsToSubscribe[i])\n\t}\n\treturn errs\n}", "func (opcuaExport *OpcuaExport) Subscribe() {\n\tglog.Infof(\"-- Initializing message bus context\")\n\tdefer opcuaExport.configMgr.Destroy()\n\n\tnumOfSubscriber, _ := opcuaExport.configMgr.GetNumSubscribers()\n\tfor i := 0; i < numOfSubscriber; i++ {\n\t\tsubctx, err := opcuaExport.configMgr.GetSubscriberByIndex(i)\n\t\tif err != nil {\n\t\t\tglog.Errorf(\"Failed to get subscriber context : %v\", err)\n\t\t\treturn\n\t\t}\n\n\t\tsubTopics, err := subctx.GetTopics()\n\t\tif err != nil {\n\t\t\tglog.Errorf(\"Failed to fetch topics : %v\", err)\n\t\t\treturn\n\t\t}\n\n\t\tconfig, err := subctx.GetMsgbusConfig()\n\t\tif err != nil {\n\t\t\tglog.Errorf(\"Failed to fetch msgbus config : %v\", err)\n\t\t\treturn\n\t\t}\n\t\tgo worker(opcuaExport, config, subTopics[0])\n\t\tsubctx.Destroy()\n\t}\n\t\n}", "func (backend *RedisBackend) Subscribe(queuename string) {\n\terr := backend.pubsub.Subscribe(fmt.Sprintf(\"%s_%s\", TITLE, queuename))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (l *Logger) Subscribe(c chan DataUnit) {\n\tl.consumers = append(l.consumers, c)\n}", "func (ba *BroadcastAgent) Subscribe(id int, ch chan chan<- []Message) {\n\tc, ok := <-ch\n\tif ok {\n\t\tba.parties[id] = c\n\t}\n}", "func (h *Hub) Publish(t string, b []byte) {\n\th.RLock()\n\tdefer h.RUnlock()\n\tchans, ok := h.topicChans[t]\n\tif !ok {\n\t\treturn\n\t}\n\tfor ch := range chans {\n\t\tselect {\n\t\tcase ch <- b:\n\t\tdefault:\n\t\t\tcontinue\n\t\t}\n\t}\n}", "func (w *Watcher) Subscribe(sink chan<- []*zeroex.OrderEvent) event.Subscription {\n\treturn w.orderScope.Track(w.orderFeed.Subscribe(sink))\n}", "func (d *Demo) Subscribe(recv backend.Receiver) {\n\td.Lock()\n\tdefer d.Unlock()\n\n\td.subscriber = recv\n\n\t// Release the lock before running an update.\n\tgo d.updateAll()\n}", "func (c *Crawler) Subscribe(chanSize uint) UpdateChan {\n\tch := make(UpdateChan, int(chanSize))\n\tc.subscribeChan <- ch\n\treturn ch\n}", "func (b *Broker) Subscribe(s *Subscriber, topics ...string) {\n\tb.tlock.Lock()\n\tdefer b.tlock.Unlock()\n\tfor _, topic := range topics {\n\t\tif nil == b.topics[topic] {\n\t\t\tb.topics[topic] = Subscribers{}\n\t\t}\n\t\ts.topics[topic] = true\n\t\tb.topics[topic][s.id] = s\n\t}\n}", "func (c *subContext) subscribe(ctx context.Context, indCh chan<- indication.Indication) error {\n\t// Add the subscription to the subscription service\n\terr := c.subClient.Add(ctx, c.sub)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Watch the subscription task service to determine assignment of the subscription to E2 terminations\n\twatchCh := make(chan subtaskapi.Event)\n\twatchCtx, cancel := context.WithCancel(context.Background())\n\tc.cancel = cancel\n\terr = c.taskClient.Watch(watchCtx, watchCh, subscriptiontask.WithSubscriptionID(c.sub.ID))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// The subscription is considered activated, and task events are processed in a separate goroutine.\n\tgo c.processTaskEvents(watchCtx, watchCh, indCh)\n\treturn nil\n}", "func (tg *TradesGroup) subscribe() {\n\tfor _, s := range tg.symbols {\n\t\tmessage := tradeSubsMessage{\n\t\t\tEvent: eventSubscribe,\n\t\t\tChannel: channelTrades,\n\t\t\tSymbol: \"t\" + strings.ToUpper(s.OriginalName),\n\t\t}\n\t\tif err := tg.wsClient.Write(message); err != nil {\n\t\t\tlog.Printf(\"[BITFINEX] Error subsciring to %v trades\", s.Name)\n\t\t\ttg.restart()\n\t\t\treturn\n\t\t}\n\t}\n\tlog.Println(\"[BITFINEX] Subscription ok\")\n}", "func (q *QuotesGroup) subscribe(ch chan schemas.ResultChannel, d time.Duration) {\n\tfor {\n\t\tquotes, err := q.Get()\n\t\tif err != nil {\n\t\t\tch <- schemas.ResultChannel{\n\t\t\t\tData: quotes,\n\t\t\t\tError: err,\n\t\t\t\tDataType: \"s\",\n\t\t\t}\n\t\t}\n\t\tfor _, b := range quotes {\n\t\t\tch <- schemas.ResultChannel{\n\t\t\t\tData: b,\n\t\t\t\tError: err,\n\t\t\t\tDataType: \"s\",\n\t\t\t}\n\t\t}\n\t\ttime.Sleep(d)\n\t}\n}", "func (rabbitmq *RabbitMQ) Subscribe(sub SubscribeStruct) (delivery <-chan Delivery, err error) {\r\n\tif rabbitmq == nil || rabbitmq.Channel == nil {\r\n\t\treturn nil, ErrCursor\r\n\t}\r\n\t// ch, _ := rabbitmq.Connection.Channel()\r\n\t// defer ch.Close()\r\n\t// ch.Qos(sub.prefetchCount, sub.prefetchSize, sub.global)\r\n\r\n\trabbitmq.Channel.Qos(sub.PrefetchCount, sub.PrefetchSize, sub.ApplyPrefetchGlobally)\r\n\r\n\tmsgs, err := rabbitmq.Channel.Consume(\r\n\t\tsub.Queue,\r\n\t\tsub.Consumer,\r\n\t\tsub.AutoAck,\r\n\t\tsub.Exclusive,\r\n\t\tsub.NoLocal,\r\n\t\tsub.NoLocal,\r\n\t\tnil,\r\n\t)\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\treturn castDeliveryCh(msgs), nil\r\n}", "func (s *server) Subscribe(p *Chat_Consumer, stream ChatService_SubscribeServer) error {\n\tep := s.read_ep(p.Id)\n\tif ep == nil {\n\t\tlog.Errorf(\"cannot find endpoint %v\", p)\n\t\treturn ERROR_NOT_EXISTS\n\t}\n\n\tconsumerid := atomic.AddUint64(&s.consumerid_autoinc, 1)\n\te := make(chan error, 1)\n\n\t// activate consumer\n\tep.mu.Lock()\n\n\t// from newest\n\tif p.From == -1 {\n\t\tp.From = ep.StartOffset + int64(len(ep.Inbox))\n\t}\n\tep.consumers[consumerid] = &Consumer{p.From, func(msg *Chat_Message) {\n\t\tif err := stream.Send(msg); err != nil {\n\t\t\tselect {\n\t\t\tcase e <- err:\n\t\t\tdefault:\n\t\t\t}\n\t\t}\n\t}}\n\tep.mu.Unlock()\n\tdefer func() {\n\t\tep.mu.Lock()\n\t\tdelete(ep.consumers, consumerid)\n\t\tep.mu.Unlock()\n\t}()\n\n\tep.notifyConsumers()\n\n\tselect {\n\tcase <-stream.Context().Done():\n\tcase err := <-e:\n\t\treturn err\n\t}\n\treturn nil\n}", "func (g *Gemini) Subscribe(channelsToSubscribe []stream.ChannelSubscription) error {\n\tchannels := make([]string, 0, len(channelsToSubscribe))\n\tfor x := range channelsToSubscribe {\n\t\tif common.StringDataCompareInsensitive(channels, channelsToSubscribe[x].Channel) {\n\t\t\tcontinue\n\t\t}\n\t\tchannels = append(channels, channelsToSubscribe[x].Channel)\n\t}\n\n\tvar pairs currency.Pairs\n\tfor x := range channelsToSubscribe {\n\t\tif pairs.Contains(channelsToSubscribe[x].Currency, true) {\n\t\t\tcontinue\n\t\t}\n\t\tpairs = append(pairs, channelsToSubscribe[x].Currency)\n\t}\n\n\tfmtPairs, err := g.FormatExchangeCurrencies(pairs, asset.Spot)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsubs := make([]wsSubscriptions, len(channels))\n\tfor x := range channels {\n\t\tsubs[x] = wsSubscriptions{\n\t\t\tName: channels[x],\n\t\t\tSymbols: strings.Split(fmtPairs, \",\"),\n\t\t}\n\t}\n\n\twsSub := wsSubscribeRequest{\n\t\tType: \"subscribe\",\n\t\tSubscriptions: subs,\n\t}\n\terr = g.Websocket.Conn.SendJSONMessage(wsSub)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tg.Websocket.AddSuccessfulSubscriptions(channelsToSubscribe...)\n\treturn nil\n}", "func subscribe(t *testing.T, wsc *client.WSClient, eventid string) {\n\tif err := wsc.Subscribe(eventid); err != nil {\n\t\tpanic(err)\n\t}\n}", "func (agent *SrlAgent) Subscribe(topic string) {\n\tagent.Logger.Debug(\"Subscribing for \", topic)\n\top := pb.NotificationRegisterRequest_AddSubscription\n\tresponse := &pb.NotificationRegisterResponse{}\n\tvar err error\n\tif topic == \"intf\" {\n\t\tresponse, err = agent.Stub.NotificationRegister(agent.Ctx, &pb.NotificationRegisterRequest{\n\t\t\tStreamId: agent.StreamID,\n\t\t\tOp: op,\n\t\t\tSubscriptionTypes: &pb.NotificationRegisterRequest_Intf{},\n\t\t})\n\t}\n\tif topic == \"nw_inst\" {\n\t\tresponse, err = agent.Stub.NotificationRegister(agent.Ctx, &pb.NotificationRegisterRequest{\n\t\t\tStreamId: agent.StreamID,\n\t\t\tOp: op,\n\t\t\tSubscriptionTypes: &pb.NotificationRegisterRequest_NwInst{},\n\t\t})\n\t}\n\tif topic == \"lldp\" {\n\t\tresponse, err = agent.Stub.NotificationRegister(agent.Ctx, &pb.NotificationRegisterRequest{\n\t\t\tStreamId: agent.StreamID,\n\t\t\tOp: op,\n\t\t\tSubscriptionTypes: &pb.NotificationRegisterRequest_LldpNeighbor{},\n\t\t})\n\t}\n\tif topic == \"route\" {\n\t\tresponse, err = agent.Stub.NotificationRegister(agent.Ctx, &pb.NotificationRegisterRequest{\n\t\t\tStreamId: agent.StreamID,\n\t\t\tOp: op,\n\t\t\tSubscriptionTypes: &pb.NotificationRegisterRequest_Route{},\n\t\t})\n\t}\n\tif topic == \"cfg\" {\n\t\tresponse, err = agent.Stub.NotificationRegister(agent.Ctx, &pb.NotificationRegisterRequest{\n\t\t\tStreamId: agent.StreamID,\n\t\t\tOp: op,\n\t\t\tSubscriptionTypes: &pb.NotificationRegisterRequest_Config{},\n\t\t})\n\t}\n\tif topic == \"app\" {\n\t\tresponse, err = agent.Stub.NotificationRegister(agent.Ctx, &pb.NotificationRegisterRequest{\n\t\t\tStreamId: agent.StreamID,\n\t\t\tOp: op,\n\t\t\tSubscriptionTypes: &pb.NotificationRegisterRequest_Appid{},\n\t\t})\n\t}\n\tif err != nil {\n\t\tagent.Logger.Debug(\"Failed to subscribe for \", topic)\n\t}\n\tagent.Logger.Debug(\"Response for Notification register for \", topic, \" is \", response.Status)\n}", "func (s *OHLCVSocket) Subscribe(channelID string, c *Client) error {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\n\tif c == nil {\n\t\treturn errors.New(\"No connection found\")\n\t}\n\n\tif s.subscriptions[channelID] == nil {\n\t\ts.subscriptions[channelID] = make(map[*Client]bool)\n\t}\n\n\ts.subscriptions[channelID][c] = true\n\n\tif s.subscriptionsList[c] == nil {\n\t\ts.subscriptionsList[c] = []string{}\n\t}\n\n\ts.subscriptionsList[c] = append(s.subscriptionsList[c], channelID)\n\n\treturn nil\n}", "func (service *RedigoService) Subscribe(ctx context.Context, subscribed SubscribedHandler, subscription SubscriptionHandler, channels ...string) error {\n\n\tc, err := redis.Dial(\"tcp\", service.Configuration.Address,\n\t\t// Read timeout on server should be greater than ping period.\n\t\tredis.DialReadTimeout(service.Configuration.PubSub.ReadTimeout),\n\t\tredis.DialWriteTimeout(service.Configuration.PubSub.WriteTimeout),\n\t)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer c.Close()\n\n\tpsc := redis.PubSubConn{Conn: c}\n\tif err := psc.Subscribe(redis.Args{}.AddFlat(channels)...); err != nil {\n\t\treturn err\n\t}\n\n\tdone := make(chan error, 1)\n\n\t// Start a goroutine to receive notifications from the server.\n\tgo func() {\n\t\tfor {\n\t\t\tswitch n := psc.Receive().(type) {\n\t\t\tcase error:\n\t\t\t\t// Increment to count failures\n\t\t\t\tservice.Collector.subscribeFailures.Inc()\n\n\t\t\t\tdone <- n\n\t\t\t\treturn\n\t\t\tcase redis.Message:\n\t\t\t\tif err := subscription(n.Channel, n.Data); err != nil {\n\n\t\t\t\t\t// Increment to count failures\n\t\t\t\t\tservice.Collector.subscribeFailures.Inc()\n\n\t\t\t\t\tdone <- err\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\t// Increment to count success\n\t\t\t\tservice.Collector.subscribeSuccesses.Inc()\n\n\t\t\tcase redis.Subscription:\n\t\t\t\tswitch n.Count {\n\t\t\t\tcase len(channels):\n\n\t\t\t\t\t// Increment 1 in subscriptionsActive\n\t\t\t\t\tservice.Collector.subscriptionsActive.Inc()\n\n\t\t\t\t\t// Notify application when all channels are subscribed.\n\t\t\t\t\tif err := subscribed(); err != nil {\n\n\t\t\t\t\t\t// Increment to count failures\n\t\t\t\t\t\tservice.Collector.subscribeFailures.Inc()\n\n\t\t\t\t\t\tdone <- err\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\n\t\t\t\t\t// Increment to count success\n\t\t\t\t\tservice.Collector.subscribeSuccesses.Inc()\n\n\t\t\t\tcase 0:\n\t\t\t\t\t// Return from the goroutine when all channels are unsubscribed.\n\t\t\t\t\tdone <- nil\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\t// A ping is set to the server with this period to test for the health of\n\t// the connection and server.\n\tticker := time.NewTicker(service.Configuration.PubSub.HealthCheckInterval)\n\tdefer ticker.Stop()\n\nloop:\n\tfor err == nil {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\t// Send ping to test health of connection and server. If\n\t\t\t// corresponding pong is not received, then receive on the\n\t\t\t// connection will timeout and the receive goroutine will exit.\n\t\t\tif err = psc.Ping(\"\"); err != nil {\n\n\t\t\t\t// Increment to count failures\n\t\t\t\tservice.Collector.subscribeFailures.Inc()\n\n\t\t\t\tbreak loop\n\t\t\t}\n\t\tcase <-ctx.Done():\n\t\t\tbreak loop\n\t\tcase err := <-done:\n\t\t\t// Return error from the receive goroutine.\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Decrement 1 in subscriptionsActive\n\tservice.Collector.subscriptionsActive.Dec()\n\n\t// Signal the receiving goroutine to exit by unsubscribing from all channels.\n\tpsc.Unsubscribe()\n\n\t// Wait for goroutine to complete.\n\treturn <-done\n}", "func (h *Hookbot) ServeSubscribe(conn *websocket.Conn, r *http.Request) {\n\ttopic := Topic(r)\n\n\tlistener := h.Add(topic)\n\tdefer h.Del(listener)\n\n\tclosed := make(chan struct{})\n\n\tgo func() {\n\t\tdefer close(closed)\n\t\tfor {\n\t\t\tif _, _, err := conn.NextReader(); err != nil {\n\t\t\t\tconn.Close()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\tvar message Message\n\n\tfor {\n\t\tselect {\n\t\tcase message = <-listener.c:\n\t\tcase <-closed:\n\t\t\treturn\n\t\t}\n\n\t\tconn.SetWriteDeadline(time.Now().Add(90 * time.Second))\n\t\t_, isRecursive := recursive(topic)\n\t\tmsgBytes := []byte{}\n\t\tif isRecursive {\n\t\t\tmsgBytes = append(msgBytes, message.Topic...)\n\t\t\tmsgBytes = append(msgBytes, '\\x00')\n\t\t\tmsgBytes = append(msgBytes, message.Body...)\n\t\t} else {\n\t\t\tmsgBytes = message.Body\n\t\t}\n\t\terr := conn.WriteMessage(websocket.BinaryMessage, msgBytes)\n\t\tswitch {\n\t\tcase err == io.EOF || IsConnectionClose(err):\n\t\t\treturn\n\t\tcase err != nil:\n\t\t\tlog.Printf(\"Error in conn.WriteMessage: %v\", err)\n\t\t\treturn\n\t\t}\n\t}\n}", "func (m *Mux) Subscribe(ds *discordgo.Session, dm *discordgo.Message, ctx *Context) {\n\tresp := \"\\n\"\n\n\tid, _ := strconv.Atoi(ctx.Fields[len(ctx.Fields) -1])\n\tif id >= 0 && id < len(Config.Feeds) {\n\t\tvar sub Subscription\n\n\t\t// Check if already subscribed\n\t\tfor _, v := range Config.Subs {\n\t\t\tif v.ChanID == dm.ChannelID && v.SubID == id {\n\t\t\t\tresp += \"Denied subscription. Already subscribed in this channel.\"\n\t\t\t\tgoto NOSUB\n\t\t\t}\n\t\t}\n\t\t\n\t\tsub.ChanID = dm.ChannelID\n\t\tsub.SubID = id\n\t\n\t\t// Might not be thread-safe\n\t\tConfig.Subs = append(Config.Subs, sub)\n\t\tresp += \"Subscribed.\"\n\t\tNOSUB:\n\t} else {\n\t\tresp += \"Denied subscription. Invalid stream id, see: list command\"\n\t}\n\t\n\tresp += \"\\n\"\n\tds.ChannelMessageSend(dm.ChannelID, resp)\n\n\treturn\n}", "func (xa XAir) Subscribe() chan osc.Message {\n\tch := xa.ps.Subscribe()\n\tlog.Debug.Printf(\"Subscribed %p to %s.\", ch, xa.Name)\n\treturn ch\n}", "func (c *stream) Subscribe(subject string, cb streaminterface.MessageHandler) (streaminterface.Subscription, error) {\n\tif streaminterface.SubjectFromStr(subject).String() != subject {\n\t\treturn nil, ErrBadSubject\n\t}\n\tif cb == nil {\n\t\treturn nil, ErrBadSubscription\n\t}\n\n\tsub := &subscription{Subject: subject, cb: cb, stream: c}\n\n\t// setup a rendezvous point for announcing the occurrence of new messages.\n\tsub.pCond = sync.NewCond(&sub.mu)\n\n\tc.subsMu.Lock()\n\tc.ssid++\n\tsub.sid = c.ssid\n\tq := c.subs[subject]\n\tif q == nil {\n\t\tq = &channel{\n\t\t\tsubs: make(map[int64]*subscription, 1),\n\t\t}\n\t\tc.subs[subject] = q\n\t}\n\tif c.opts.Log {\n\t\tq.logMu.Lock()\n\t\tsub.log = make([]streaminterface.Message, len(q.log))\n\t\tsub.pMsgs += copy(sub.log, q.log)\n\t\tq.logMu.Unlock()\n\t}\n\tc.subs[subject].subs[sub.sid] = sub\n\tc.subsMu.Unlock()\n\n\t// start up a sub specific Go routine to deliver messages.\n\tgo c.waitDeliver(sub)\n\n\treturn sub, nil\n}", "func (h *handler) Subscribe(c *session.Client, topics *[]string) {\n\tif c == nil {\n\t\th.logger.Error(LogErrFailedSubscribe + (ErrClientNotInitialized).Error())\n\t\treturn\n\t}\n\th.logger.Info(fmt.Sprintf(LogInfoSubscribed, c.ID, strings.Join(*topics, \",\")))\n}", "func Subscribe(name string, paths [][]string) *Handle {\n\tif len(paths) == 0 {\n\t\tlog.Printf(\"module doesn't have paths for Subscribe.\")\n\t\treturn nil\n\t}\n\n\tstreamOnce.Do(func() {\n\t\tconn = startConnection()\n\t})\n\n\tsubMutex.Lock()\n\tdefer subMutex.Unlock()\n\tlog.Printf(\"subscriberId: %d\\n\", subscriberId)\n\n\tif subscriberId == math.MaxInt32 {\n\t\tlog.Printf(\"Can't create subscriber anymore.\\n\")\n\t\treturn nil\n\t}\n\n\tname = fmt.Sprintf(\"%v-%v\", name, subscriberId)\n\thandle := &Handle{\n\t\tsubscriberId: subscriberId,\n\t\tpaths: paths,\n\t\tname: name,\n\t\tconn: conn,\n\n\t\tConfigMessage: make(chan *ConfigMessage),\n\t\tRc: make(chan bool),\n\t}\n\tconn.handles[subscriberId] = handle\n\n\t// set a message to send openconfig\n\tsMsg.confReq.Type = pb.ConfigType_SUBSCRIBE\n\tsMsg.confReq.Module = name\n\n\t// subscribe paths to the server\n\tfor _, path := range paths {\n\n\t\tsMsg.confReq.Path = path\n\t\tsc <- sMsg\n\t\terr := <-sMsg.err\n\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Sending subscription message faild: %v\", err)\n\t\t\tconn.stream.CloseSend()\n\t\t\tdefer delete(conn.handles, subscriberId)\n\t\t\treturn nil\n\t\t}\n\t}\n\tsMsg.confReq = &pb.ConfigRequest{}\n\n\tlog.Printf(\"Subscribe success\\n\")\n\tsubscriberId++\n\treturn handle\n}", "func (pubsub *MemoryEventStore) Subscribe(topics ...string) <-chan Event {\n\tsub := subscription{\n\t\ttopics: topics,\n\t\tch: make(chan Event, pubsub.buffer),\n\t}\n\tpubsub.subscriptions = append(pubsub.subscriptions, sub)\n\treturn sub.ch\n}", "func (p *ConfigDistributor) Subscribe(ch chan<- *Config, checks ...ConfigChangedCheckFn) error {\n\t// Check that distributor is still active\n\tif p.closed.HasFired() {\n\t\treturn ErrConfigDistributorClosed\n\t}\n\n\tp.mu.Lock()\n\tdefer p.mu.Unlock()\n\n\t// Send current config immediately even before we registered a new subscription\n\tif p.lastConfig != nil && p.shouldSendUpdatedConfig(checks, nil, p.lastConfig) {\n\t\tch <- p.lastConfig\n\t}\n\n\t// Register subscription\n\tif p.subscriptions == nil {\n\t\tp.subscriptions = make(map[chan<- *Config][]ConfigChangedCheckFn)\n\t}\n\n\tp.subscriptions[ch] = checks\n\n\t// Start monitoring goroutine\n\tif p.started.Fire() {\n\t\tgo p.monitorConfig()\n\t}\n\n\treturn nil\n}", "func (p *Process) Subscribe(subject string) {\n\n\t\t// Pass this back to the parent\n\t\tp.subscribed_channels = append(p.subscribed_channels, p.parent.subscribe(subject))\n\t}", "func (ps *PubSub) Psubscribe(conn Conn, channel string) {\n\tps.subscribe(conn, true, channel)\n}", "func (a *Application) Subscribe(c *channel.Channel, conn *connection.Connection, data string) error {\n\treturn c.Subscribe(conn, data)\n}", "func (subscriber *Subscriber) Subscribe(conn *NATSConnection) (*nats.Subscription, error) {\n\tlog.Info(\"Initializing callback\")\n\tlog.Info(\"Subscription topic is: \", subscriber.topic)\n\tjs, err := conn.Conn.JetStream()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsub, err := js.Subscribe(streamName+\".\"+subscriber.topic, func(m *nats.Msg) {\n\t\tmeta, _ := m.Metadata()\n\t\tlog.Infof(\"Stream Sequence : %v\\n\", meta.Sequence.Stream)\n\t\tlog.Infof(\"Consumer Sequence: %v\\n\", meta.Sequence.Consumer)\n\t\tvar msg Message\n\t\tmsg.Timestamp = meta.Timestamp.Unix()\n\t\tmsg.Datum = m.Data\n\t\tdata, err := json.Marshal(&msg)\n\t\tif err != nil {\n\t\t\tlog.Error(err)\n\t\t} else {\n\t\t\tsubscriber.messages <- data\n\t\t}\n\t}, nats.DeliverNew())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn sub, nil\n}", "func TestClient_Subscribe(t *testing.T) {\n\tc := OpenClient(0)\n\tdefer c.Close()\n\tc.Server.Broker().CreateReplica(100, &url.URL{Host: \"localhost\"})\n\n\t// Create subscription through client.\n\tif err := c.Subscribe(100, 200); err != nil {\n\t\tt.Fatalf(\"unexpected error: %v\", err)\n\t}\n\n\t// Verify subscription was created.\n\tif a := c.Server.Handler.Broker().Replica(100).Topics(); !reflect.DeepEqual([]uint64{0, 200}, a) {\n\t\tt.Fatalf(\"topics mismatch: %v\", a)\n\t}\n}", "func (b *Topics) Subscribe(s *Subscriber, topics ...string) {\n\tb.topic_lock.Lock()\n\tdefer b.topic_lock.Unlock()\n\tfor _, topic := range topics {\n\t\tif nil == b.sub_topics[topic] {\n\t\t\tfmt.Println(\"!!! topic does not exist !!!\")\n\t\t}\n\t\ts.AddTopic(topic)\n\t\tb.sub_topics[topic][s.id] = s\n\t}\n}", "func (c *Connection) Subscribe(ctx context.Context, streamName, consumerName, subj string) ([]byte, error) {\n\tmgr, err := jsm.New(c.conn)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tstream, err := mgr.LoadStream(streamName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tconsumer, err := stream.LoadOrNewConsumer(consumerName,\n\t\tjsm.DurableName(consumerName),\n\t\tjsm.FilterStreamBySubject(subj),\n\t\tjsm.DeliverAllAvailable(),\n\t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_ = consumer\n\n\tjsSubject := fmt.Sprintf(`$JS.API.CONSUMER.MSG.NEXT.%s.%s`, streamName, consumerName)\n\t_ = jsSubject\n\n\t// msg, err := mgr.NextMsg(streamName, consumerName) // This does NOT work\n\t// msg, err := consumer.NextMsgContext(ctx) // This does work\n\tmsg, err := c.conn.RequestWithContext(ctx, jsSubject, []byte(\"a\")) // This does work\n\tif errors.Is(err, nats.ErrTimeout) {\n\t\treturn nil, nil\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmsg.Respond(nil)\n\treturn msg.Data, nil\n}", "func (k *Kafka) Subscribe(topic string, h broker.Handler, opts ...broker.SubscribeOption) (broker.Subscriber, error) {\n\n\top := &broker.SubscribeOptions{\n\t\tAutoAck: true,\n\t}\n\top.Apply(opts...)\n\n\tvar err error\n\t//handler function\n\thandler := func(msg *sarama.ConsumerMessage) {\n\t\tm := broker.Message{}\n\t\tif err := k.encoder.Decode(msg.Value, &m); err != nil {\n\t\t\tlog.Errorf(\"kafka: subscribe: decode failed, err: %v\", err)\n\t\t\treturn\n\t\t}\n\t\th(&event{\n\t\t\top: op,\n\t\t\tt: topic,\n\t\t\tm: &m,\n\t\t})\n\t}\n\t//Consumer with no groupID\n\tif op.Queue == \"\" {\n\t\tlog.Info(\"consumer with no groupID\")\n\t\t// Create new consumer\n\t\tk.consumer, err = sarama.NewConsumer(strings.Split(k.addrs, \",\"), k.config)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tpartitionList, err := k.consumer.Partitions(topic)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tfor partition := range partitionList {\n\t\t\tconsumer, err := k.consumer.ConsumePartition(topic, int32(partition), sarama.OffsetNewest)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tgo func() {\n\t\t\t\tfor msg := range consumer.Messages() {\n\t\t\t\t\thandler(msg)\n\t\t\t\t}\n\t\t\t}()\n\t\t}\n\t\treturn &subscriber{\n\t\t\tqueue: op.Queue,\n\t\t\tt: topic,\n\t\t\ts: k.consumer,\n\t\t}, nil\n\n\t} //end no group\n\t{\n\t\t// Create new consumer group\n\t\tif k.consumerGroup, err = sarama.NewConsumerGroup([]string{k.addrs}, op.Queue, k.config); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tctx := context.Background()\n\t\tconsumer := Consumer{\n\t\t\tencoder: k.encoder,\n\t\t\ttopic: topic,\n\t\t\th: h,\n\t\t\tready: make(chan bool),\n\t\t}\n\t\tgo func() {\n\t\t\tfor {\n\n\t\t\t\tif err := k.consumerGroup.Consume(ctx, []string{topic}, &consumer); err != nil {\n\t\t\t\t\tlog.Panicf(\"Error from consumer: %v\", err)\n\t\t\t\t}\n\t\t\t\t<-ctx.Done()\n\n\t\t\t\tif ctx.Err() != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\t\t<-consumer.ready // Await till the consumer has been set up\n\t\tlog.Info(\"Sarama consumer up and running!...\")\n\n\t\treturn &subscriberGroup{\n\t\t\tqueue: op.Queue,\n\t\t\tt: topic,\n\t\t\tg: k.consumerGroup,\n\t\t}, nil\n\t}\n\n}", "func (ps *PubsubApi) LogsSubscribe(ctx context.Context, crit filters.FilterCriteria) (*rpc.Subscription, error) {\n\tif ps.s.context().eventBus == nil {\n\t\t// @Note: Should not happen!\n\t\tlog.Error(\"rpc: eventbus nil, not support Subscribetion!!!\")\n\t\treturn nil, rpc.ErrNotificationsUnsupported\n\t}\n\n\tnotifier, supported := rpc.NotifierFromContext(ctx)\n\tif !supported {\n\t\treturn nil, rpc.ErrNotificationsUnsupported\n\t}\n\n\tsubscription := notifier.CreateSubscription()\n\n\tsuberName := fmt.Sprintf(\"rpc-log-suber-%s\", subscription.ID)\n\tebCtx := context.Background()\n\tlogsCh := make(chan interface{}, 128)\n\tif err := ps.context().eventBus.Subscribe(ebCtx, suberName, types.EventQueryLog, logsCh); err != nil {\n\t\tlog.Warn(\"rpc: Subscribe fail\", \"err\", err)\n\t\treturn nil, err\n\t}\n\n\tgo func() {\n\t\tdefer ps.context().eventBus.Unsubscribe(ebCtx, suberName, types.EventQueryLog)\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase ev := <-logsCh:\n\t\t\t\tlogs := ev.(types.EventDataLog).Logs\n\t\t\t\tlogs = filterLogs(logs, crit.FromBlock.ToInt(), crit.ToBlock.ToInt(), crit.Addresses, crit.Topics)\n\t\t\t\tfor _, l := range logs {\n\t\t\t\t\tnotifier.Notify(subscription.ID, l)\n\t\t\t\t\tlog.Info(\"rpc: notify success\", \"suber\", suberName, \"log\", l)\n\t\t\t\t}\n\t\t\tcase <-notifier.Closed():\n\t\t\t\tlog.Info(\"rpc LogSubscribe: unsubscribe\", \"suber\", suberName)\n\t\t\t\treturn\n\t\t\tcase err := <-subscription.Err():\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error(\"rpc subscription: error\", \"suber\", suberName, \"err\", err)\n\t\t\t\t} else {\n\t\t\t\t\tlog.Info(\"rpc subscription: exit\", \"suber\", suberName)\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\tlog.Info(\"rpc LogsSubscribe: ok\", \"name\", suberName, \"crit\", crit.String())\n\treturn subscription, nil\n}", "func (_Eth *EthTransactor) Subscribe(opts *bind.TransactOpts, a common.Address, topic string) (*types.Transaction, error) {\n\treturn _Eth.contract.Transact(opts, \"subscribe\", a, topic)\n}", "func (nd *NodeDiscover) Subscribe(eventCh chan NodeEvent) {\n\tnd.subscribers.Lock()\n\tnd.subscribers.s[eventCh] = true\n\tnd.subscribers.Unlock()\n\tlog.Info(\"node-discovery: Subscriber added\")\n}", "func (s *Subscriber) Subscribe(ctx context.Context, channel chan<- cavpnEvent) error {\n\terr := s.connect(ctx, channel)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (trading *TradingProvider) Subscribe(interval time.Duration) (chan schemas.UserInfoChannel, chan schemas.UserOrdersChannel, chan schemas.UserTradesChannel) {\n\tuic := make(chan schemas.UserInfoChannel)\n\tuoc := make(chan schemas.UserOrdersChannel)\n\tutc := make(chan schemas.UserTradesChannel)\n\n\tif interval < 5*time.Second {\n\t\tinterval = 5 * time.Second\n\t}\n\n\tgo func() {\n\t\tfor {\n\t\t\tui, err := trading.Info()\n\t\t\tuic <- schemas.UserInfoChannel{\n\t\t\t\tDataType: dataTypeSnapshot,\n\t\t\t\tData: ui,\n\t\t\t\tError: err,\n\t\t\t}\n\n\t\t\tuo, err := trading.Orders([]schemas.Symbol{})\n\t\t\tuoc <- schemas.UserOrdersChannel{\n\t\t\t\tDataType: dataTypeSnapshot,\n\t\t\t\tData: uo,\n\t\t\t\tError: err,\n\t\t\t}\n\n\t\t\tut, _, err := trading.Trades(schemas.FilterOptions{})\n\t\t\tutc <- schemas.UserTradesChannel{\n\t\t\t\tDataType: dataTypeSnapshot,\n\t\t\t\tData: ut,\n\t\t\t\tError: err,\n\t\t\t}\n\n\t\t\ttime.Sleep(interval)\n\t\t}\n\t}()\n\n\treturn uic, uoc, utc\n}", "func (t *topicTrie) subscribe(clientID string, s *gmqtt.Subscription) *topicNode {\n\ttopicSlice := strings.Split(s.TopicFilter, \"/\")\n\tvar pNode = t\n\tfor _, lv := range topicSlice {\n\t\tif _, ok := pNode.children[lv]; !ok {\n\t\t\tpNode.children[lv] = pNode.newChild()\n\t\t}\n\t\tpNode = pNode.children[lv]\n\t}\n\t// shared subscription\n\tif s.ShareName != \"\" {\n\t\tif pNode.shared[s.ShareName] == nil {\n\t\t\tpNode.shared[s.ShareName] = make(clientOpts)\n\t\t}\n\t\tpNode.shared[s.ShareName][clientID] = s\n\t} else {\n\t\t// non-shared\n\t\tpNode.clients[clientID] = s\n\t}\n\tpNode.topicName = s.TopicFilter\n\treturn pNode\n}", "func (zs *pubZerver) Subscribe(stream pb.Subscription_SubscribeServer) error {\n\tres := new(pb.NotificationMessage)\n\t// res.Top10 = make([]*pb.NotificationMessage_Top10, 10)\n\n\tmsg, err := stream.Recv()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfreq := time.Duration(msg.GetRefreshRate()) * time.Second\n\tr := msg.GetStatistics()\n\n\tfmt.Printf(\"[ZubServer] Got subscription request with a refresh interval of %v and an enum of '%v'\\n\", freq, r.String())\n\n\tfor {\n\t\tif zs.logs.Entries() < 1 {\n\t\t\t// TODO error codes as int field?\n\t\t\tres.Status = fmt.Sprintf(\"2: The server has not yet logged any channels. Retrying in %v\", retryInterval)\n\t\t} else {\n\t\t\tres.Status = \"1\"\n\n\t\t\tres.Top10, err = parseTop10(r, zs.logs)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t\terr = stream.Send(res)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\ttime.Sleep(freq)\n\t}\n}", "func (s *Subscriber) Subscribe(out chan *commons.WebResource) {\n\n\terr := s.sub.Receive(s.context, func(ctx context.Context, msg *pubsub.Message) {\n\t\titem := &commons.WebResource{}\n\t\tif err := json.Unmarshal(msg.Data, &item); err != nil {\n\t\t\tlogger.Printf(\"Error while decoding PubSub message: %#v\", msg)\n\t\t\tmsg.Nack()\n\t\t} else {\n\t\t\t//logger.Printf(\"Event -> %s\", item.String())\n\t\t\tout <- item\n\t\t\tmsg.Ack()\n\t\t}\n\t})\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n}", "func (c *Coordinator) Subscribe(ss ...func(*Config) error) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tc.subscribers = append(c.subscribers, ss...)\n}", "func (c *MockController) Subscribe(sid string, env *common.SessionEnv, id string, channel string) (*common.CommandResult, error) {\n\tif channel == \"failure\" {\n\t\treturn nil, errors.New(\"Subscription Failure\")\n\t}\n\n\tres := NewMockResult(sid)\n\n\tif channel == \"failure\" {\n\t\treturn nil, errors.New(\"Subscription failed\")\n\t}\n\n\tif channel == \"disconnect\" {\n\t\tres.Disconnect = true\n\t\treturn res, nil\n\t}\n\n\tif channel == \"stream\" {\n\t\tres.Streams = []string{\"stream\"}\n\t}\n\n\treturn res, nil\n}", "func EventSubscribeH(w http.ResponseWriter, r *http.Request) {\n\n\tlog.V(logLevel).Debugf(\"%s:subscribe:> subscribe on subscribe\", logPrefix)\n\n\tif r.Method != \"GET\" {\n\t\thttp.Error(w, \"Method not allowed\", http.StatusMethodNotAllowed)\n\t\treturn\n\t}\n\n\tlog.V(logLevel).Debugf(\"%s:subscribe:> watch all events\", logPrefix)\n\n\tvar (\n\t\tsm = distribution.NewServiceModel(r.Context(), envs.Get().GetStorage())\n\t\tnm = distribution.NewNamespaceModel(r.Context(), envs.Get().GetStorage())\n\t\tcm = distribution.NewClusterModel(r.Context(), envs.Get().GetStorage())\n\t\tdone = make(chan bool, 1)\n\t)\n\n\tconn, err := upgrader.Upgrade(w, r, nil)\n\tif err != nil {\n\t\tlog.V(logLevel).Debugf(\"%s:subscribe:> set websocket upgrade err: %s\", logPrefix, err.Error())\n\t\treturn\n\t}\n\n\tticker := time.NewTicker(time.Second)\n\tdefer ticker.Stop()\n\n\tvar serviceEvents = make(chan types.ServiceEvent)\n\tvar namespaceEvents = make(chan types.NamespaceEvent)\n\tvar clusterEvents = make(chan types.ClusterEvent)\n\n\tnotify := w.(http.CloseNotifier).CloseNotify()\n\n\tgo func() {\n\t\t<-notify\n\t\tlog.V(logLevel).Debugf(\"%s:subscribe:> HTTP connection just closed.\", logPrefix)\n\t\tdone <- true\n\t}()\n\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-done:\n\t\t\t\tclose(serviceEvents)\n\t\t\t\tclose(namespaceEvents)\n\t\t\t\tclose(clusterEvents)\n\t\t\t\treturn\n\t\t\tcase e := <-clusterEvents:\n\n\t\t\t\tvar data interface{}\n\t\t\t\tif e.Data == nil {\n\t\t\t\t\tdata = nil\n\t\t\t\t} else {\n\t\t\t\t\tdata = v1.View().Cluster().New(e.Data)\n\t\t\t\t}\n\n\t\t\t\tevent := Event{\n\t\t\t\t\tEntity: \"cluster\",\n\t\t\t\t\tAction: e.Action,\n\t\t\t\t\tName: e.Name,\n\t\t\t\t\tData: data,\n\t\t\t\t}\n\n\t\t\t\tif err = conn.WriteJSON(event); err != nil {\n\t\t\t\t\tlog.Errorf(\"%s:subscribe:> write cluster event to socket error.\", logPrefix)\n\t\t\t\t}\n\t\t\tcase e := <-serviceEvents:\n\n\t\t\t\tvar data interface{}\n\t\t\t\tif e.Data == nil {\n\t\t\t\t\tdata = nil\n\t\t\t\t} else {\n\t\t\t\t\tdata = v1.View().Service().New(e.Data)\n\t\t\t\t}\n\n\t\t\t\tevent := Event{\n\t\t\t\t\tEntity: \"service\",\n\t\t\t\t\tAction: e.Action,\n\t\t\t\t\tName: e.Name,\n\t\t\t\t\tData: data,\n\t\t\t\t}\n\n\t\t\t\tif err = conn.WriteJSON(event); err != nil {\n\t\t\t\t\tlog.Errorf(\"%s:subscribe:> write service event to socket error.\", logPrefix)\n\t\t\t\t}\n\t\t\tcase e := <-namespaceEvents:\n\n\t\t\t\tvar data interface{}\n\t\t\t\tif e.Data == nil {\n\t\t\t\t\tdata = nil\n\t\t\t\t} else {\n\t\t\t\t\tdata = v1.View().Namespace().New(e.Data)\n\t\t\t\t}\n\n\t\t\t\tevent := Event{\n\t\t\t\t\tEntity: \"namespace\",\n\t\t\t\t\tAction: e.Action,\n\t\t\t\t\tName: e.Name,\n\t\t\t\t\tData: data,\n\t\t\t\t}\n\n\t\t\t\tif err = conn.WriteJSON(event); err != nil {\n\t\t\t\t\tlog.Errorf(\"%s:subscribe:> write namespace event to socket error.\", logPrefix)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\tgo cm.Watch(clusterEvents)\n\tgo sm.Watch(serviceEvents, nil)\n\tgo nm.Watch(namespaceEvents)\n\n\tgo func() {\n\t\tfor range ticker.C {\n\t\t\tif err := conn.WriteMessage(websocket.TextMessage, []byte{}); err != nil {\n\t\t\t\tlog.Errorf(\"%s:subscribe:> writing to the client websocket err: %s\", logPrefix, err.Error())\n\t\t\t\tdone <- true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}()\n\n\t<-done\n}", "func subscribeCallback(client MQTT.Client, message MQTT.Message) {\n\n fmt.Printf(\"\\n\")\n fmt.Printf(\"\\nTopic: \" + message.Topic())\n fmt.Printf(\"\\nMessage: \\n\")\n fmt.Printf(string(message.Payload()))\n fmt.Printf(\"\\n--------------\")\n}", "func (pc *Client) Subscribe(channel string, callback SubscribeCallback) error {\n\tpc.subscribers[channel] = callback\n\n\terr := pc.sendMessage(&messages.Subscribe{messages.TYPE_SUBSCRIBE, channel})\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = pc.wait(messages.TYPE_SUBSCRIBE_OK)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (m *Client) Subscribe() chan proto.Payload {\n\tmessages := make(chan proto.Payload)\n\ttq := []proto.TopicQos{proto.TopicQos{Topic: m.topic, Qos: 0}}\n\tm.client.Subscribe(tq)\n\tgo func() {\n\t\tdefer close(messages)\n\t\tfor message := range m.client.Incoming {\n\t\t\tmessages <- message.Payload\n\t\t}\n\t}()\n\treturn messages\n}", "func (p *metadataService) onTopicSubscribe(e event.Event) {\n}", "func SubscribeHandler(subChannel chan string) {\n\t// create a subscribe connection to RedisDB\n\tsubscribeConn, err := redisurl.ConnectToURL(\"redis://localhost:6379\")\n\tif err != nil {\n\n\t\tfmt.Println(err)\n\t\tos.Exit(1)\n\n\t}\n\n\t// Before function exits close the connection\n\tdefer subscribeConn.Close()\n\n\tpubsubConn := redis.PubSubConn{Conn: subscribeConn}\n\tpubsubConn.Subscribe(\"messages\") // Subscribed to messages list in redis DB\n\n\tfor {\n\n\t\tswitch val := pubsubConn.Receive().(type) {\n\n\t\tcase redis.Message:\n\t\t\t// If the data being received is a text message then push it to the channel\n\t\t\tsubChannel <- string(val.Data)\n\n\t\tcase redis.Subscription:\n\t\t\t//Handle Subscription here\n\n\t\tcase error:\n\t\t\treturn\n\t\t}\n\t}\n\n}", "func (b *AsyncTower) Subscribe() chan int {\n\tb.mut.Lock()\n\tch := make(chan int, 1)\n\tb.subs = append(b.subs, ch)\n\tb.mut.Unlock()\n\treturn ch\n}", "func (s *Stentor) Subscribe() <-chan G {\n\tch := make(chan G, s.bufSize)\n\ts.mu.Lock()\n\ts.subscribers[ch] = ch\n\ts.mu.Unlock()\n\n\treturn ch\n}" ]
[ "0.66779685", "0.6572508", "0.6495061", "0.6422174", "0.63689184", "0.6352444", "0.6288932", "0.6274408", "0.62708884", "0.625784", "0.6244839", "0.62359333", "0.6185336", "0.6170156", "0.61666375", "0.6162882", "0.6162781", "0.61599725", "0.61536837", "0.61132455", "0.6086286", "0.6085617", "0.6074434", "0.6072351", "0.6069417", "0.6060739", "0.6054431", "0.60521674", "0.6047343", "0.602667", "0.6009161", "0.5995905", "0.59920865", "0.59802514", "0.5977099", "0.5977075", "0.5976779", "0.5976261", "0.5965189", "0.5963935", "0.5959921", "0.59591067", "0.59580845", "0.59556043", "0.59556043", "0.59433556", "0.59281826", "0.5922841", "0.5902451", "0.59015363", "0.58970726", "0.58815074", "0.58653915", "0.58607584", "0.5852928", "0.5850947", "0.5843706", "0.58395267", "0.5836524", "0.5835301", "0.5835281", "0.58266056", "0.5818532", "0.5816525", "0.5803873", "0.58038586", "0.57969004", "0.5791388", "0.57883346", "0.5786011", "0.5777916", "0.5773768", "0.5772156", "0.5764747", "0.57622963", "0.576218", "0.5759674", "0.57501966", "0.5740018", "0.573358", "0.57312787", "0.5700901", "0.5699191", "0.569727", "0.56911874", "0.56894654", "0.5686889", "0.567339", "0.5671101", "0.5666667", "0.5660641", "0.5657101", "0.56534487", "0.5646746", "0.56397104", "0.5637633", "0.56213504", "0.560482", "0.55995303", "0.5599213" ]
0.71316516
0
SubscribeToAll creates a channel that is registered to ALL of the different partitions
func (n *Node) SubscribeToAll(topic string) <-chan Message { consumer := make(chan Message) parts := n.getPartitionsFor(topic) for _, p := range parts { p.Subscribe(consumer) } return consumer }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (x *Pep) SubscribeToAll(ctx context.Context, host string, jid *jid.JID) {\n\tx.runQueue.Run(func() {\n\t\tif err := x.subscribeToAll(ctx, host, jid); err != nil {\n\t\t\tlog.Error(err)\n\t\t}\n\t})\n}", "func (cp *CandlesProvider) SubscribeAll(d time.Duration) chan schemas.ResultChannel {\n\tch := make(chan schemas.ResultChannel)\n\n\tfor _, orderBook := range cp.groups {\n\t\tgo orderBook.Subscribe(ch, d)\n\t\ttime.Sleep(100 * time.Millisecond)\n\t}\n\treturn ch\n}", "func (n *Node) Subscribe(topic string) <-chan Message {\n\tconsumer := make(chan Message)\n\n\tparts := n.getPartitionsFor(topic)\n\tvar part partition\n\tfor _, p := range parts {\n\t\tpart = smallestOf(part, p)\n\t}\n\n\tlog.Debug(\"Created subscriber for %s:%v\", topic, part)\n\tpart.Subscribe(consumer)\n\treturn consumer\n}", "func (cp *CandlesProvider) SubscribeAll(d time.Duration) chan schemas.ResultChannel {\n\treturn nil\n}", "func SubscribeAll(clientName string) {\n\tvar topic string\n\tvar err error\n\ttopic = fmt.Sprintf(\"/%s/echo\", clientName)\n\tif err = mqttHelper.Subscribe(topic, byte(0), onEcho); err != nil {\n\t\tlog.Errorf(\"Unable to subscribe to %s. Error: %s\", topic, err)\n\t}\n\ttopic = fmt.Sprintf(\"/%s/search\", clientName)\n\tif err = mqttHelper.Subscribe(topic, byte(0), onSearch); err != nil {\n\t\tlog.Errorf(\"Unable to subscribe to %s. Error: %s\", topic, err)\n\t}\n}", "func (g *Gemini) Subscribe(channelsToSubscribe []stream.ChannelSubscription) error {\n\tchannels := make([]string, 0, len(channelsToSubscribe))\n\tfor x := range channelsToSubscribe {\n\t\tif common.StringDataCompareInsensitive(channels, channelsToSubscribe[x].Channel) {\n\t\t\tcontinue\n\t\t}\n\t\tchannels = append(channels, channelsToSubscribe[x].Channel)\n\t}\n\n\tvar pairs currency.Pairs\n\tfor x := range channelsToSubscribe {\n\t\tif pairs.Contains(channelsToSubscribe[x].Currency, true) {\n\t\t\tcontinue\n\t\t}\n\t\tpairs = append(pairs, channelsToSubscribe[x].Currency)\n\t}\n\n\tfmtPairs, err := g.FormatExchangeCurrencies(pairs, asset.Spot)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsubs := make([]wsSubscriptions, len(channels))\n\tfor x := range channels {\n\t\tsubs[x] = wsSubscriptions{\n\t\t\tName: channels[x],\n\t\t\tSymbols: strings.Split(fmtPairs, \",\"),\n\t\t}\n\t}\n\n\twsSub := wsSubscribeRequest{\n\t\tType: \"subscribe\",\n\t\tSubscriptions: subs,\n\t}\n\terr = g.Websocket.Conn.SendJSONMessage(wsSub)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tg.Websocket.AddSuccessfulSubscriptions(channelsToSubscribe...)\n\treturn nil\n}", "func (r *Room) BroadcastAll(msg []byte) {\n\tfor _, client := range r.Clients {\n\t\tclient.WriteMessage(msg)\n\t}\n}", "func (app *App) RegisterAll(to string) {\n\tm := &message.Register{\n\t\tHeader: &message.Header{},\n\t}\n\tm.Header.From = &app.ID\n\tm.Name = &app.Name\n\tm.Header.To = &to\n\tm.Devices = app.DeviceList()\n\n\tapp.Publish(queue.Inventory, m)\n}", "func (by *Bybit) Subscribe(channelsToSubscribe []stream.ChannelSubscription) error {\n\tvar errs error\n\tfor i := range channelsToSubscribe {\n\t\tvar subReq WsReq\n\t\tsubReq.Topic = channelsToSubscribe[i].Channel\n\t\tsubReq.Event = sub\n\n\t\tformattedPair, err := by.FormatExchangeCurrency(channelsToSubscribe[i].Currency, asset.Spot)\n\t\tif err != nil {\n\t\t\terrs = common.AppendError(errs, err)\n\t\t\tcontinue\n\t\t}\n\t\tif channelsToSubscribe[i].Channel == wsKlines {\n\t\t\tsubReq.Parameters = WsParams{\n\t\t\t\tSymbol: formattedPair.String(),\n\t\t\t\tIsBinary: true,\n\t\t\t\tKlineType: \"1m\",\n\t\t\t}\n\t\t} else {\n\t\t\tsubReq.Parameters = WsParams{\n\t\t\t\tSymbol: formattedPair.String(),\n\t\t\t\tIsBinary: true,\n\t\t\t}\n\t\t}\n\t\terr = by.Websocket.Conn.SendJSONMessage(subReq)\n\t\tif err != nil {\n\t\t\terrs = common.AppendError(errs, err)\n\t\t\tcontinue\n\t\t}\n\t\tby.Websocket.AddSuccessfulSubscriptions(channelsToSubscribe[i])\n\t}\n\treturn errs\n}", "func (opcuaExport *OpcuaExport) Subscribe() {\n\tglog.Infof(\"-- Initializing message bus context\")\n\tdefer opcuaExport.configMgr.Destroy()\n\n\tnumOfSubscriber, _ := opcuaExport.configMgr.GetNumSubscribers()\n\tfor i := 0; i < numOfSubscriber; i++ {\n\t\tsubctx, err := opcuaExport.configMgr.GetSubscriberByIndex(i)\n\t\tif err != nil {\n\t\t\tglog.Errorf(\"Failed to get subscriber context : %v\", err)\n\t\t\treturn\n\t\t}\n\n\t\tsubTopics, err := subctx.GetTopics()\n\t\tif err != nil {\n\t\t\tglog.Errorf(\"Failed to fetch topics : %v\", err)\n\t\t\treturn\n\t\t}\n\n\t\tconfig, err := subctx.GetMsgbusConfig()\n\t\tif err != nil {\n\t\t\tglog.Errorf(\"Failed to fetch msgbus config : %v\", err)\n\t\t\treturn\n\t\t}\n\t\tgo worker(opcuaExport, config, subTopics[0])\n\t\tsubctx.Destroy()\n\t}\n\t\n}", "func subscribeAllTopics() {\n\tfor key, value := range topicMap {\n\t\thelper.TokenClient = helper.Client.Subscribe(key, 0, value)\n\t\tif helper.TokenClient.Wait() && helper.TokenClient.Error() != nil {\n\t\t\tklog.Errorf(\"subscribe() Error in topic: %s is: %s\", key, helper.TokenClient.Error())\n\t\t}\n\t}\n}", "func (cg *CandlesGroup) subscribe() {\n\tfor _, symb := range cg.symbols {\n\t\tmessage := candlesSubsMessage{\n\t\t\tEvent: eventSubscribe,\n\t\t\tChannel: \"candles\",\n\t\t\tKey: \"trade:1m:t\" + strings.ToUpper(symb.OriginalName),\n\t\t}\n\n\t\tif err := cg.wsClient.Write(message); err != nil {\n\t\t\tlog.Printf(\"[BITFINEX] Error subsciring to %v candles\", symb.Name)\n\t\t\tcg.restart()\n\t\t\treturn\n\t\t}\n\t}\n\tlog.Println(\"[BITFINEX] Subscription ok\")\n}", "func (h *Hub) SendToAll(data []byte) {\n\th.mux.Lock()\n\tdefer h.mux.Unlock()\n\tfor conn := range h.conns {\n\t\th.Send(conn, data, websocket.TextMessage)\n\t}\n}", "func (h *Hub) UnsubscribeAll(ch chan []byte) {\n\th.Lock()\n\tdefer h.Unlock()\n\ttopics, ok := h.chanTopics[ch]\n\tif !ok {\n\t\treturn\n\t}\n\tfor t := range topics {\n\t\tchans, ok := h.topicChans[t]\n\t\tif ok {\n\t\t\tdelete(chans, ch)\n\t\t}\n\t}\n\tdelete(h.chanTopics, ch)\n}", "func (s *StanServer) initSubscriptions() error {\n\n\t// Do not create internal subscriptions in clustered mode,\n\t// the leader will when it gets elected.\n\tif !s.isClustered {\n\t\tcreateSubOnClientPublish := true\n\n\t\tif s.partitions != nil {\n\t\t\t// Receive published messages from clients, but only on the list\n\t\t\t// of static channels.\n\t\t\tif err := s.partitions.initSubscriptions(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\t// Since we create a subscription per channel, do not create\n\t\t\t// the internal subscription on the > wildcard\n\t\t\tcreateSubOnClientPublish = false\n\t\t}\n\n\t\tif err := s.initInternalSubs(createSubOnClientPublish); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ts.log.Debugf(\"Discover subject: %s\", s.info.Discovery)\n\t// For partitions, we actually print the list of channels\n\t// in the startup banner, so we don't need to repeat them here.\n\tif s.partitions != nil {\n\t\ts.log.Debugf(\"Publish subjects root: %s\", s.info.Publish)\n\t} else {\n\t\ts.log.Debugf(\"Publish subject: %s.>\", s.info.Publish)\n\t}\n\ts.log.Debugf(\"Subscribe subject: %s\", s.info.Subscribe)\n\ts.log.Debugf(\"Subscription Close subject: %s\", s.info.SubClose)\n\ts.log.Debugf(\"Unsubscribe subject: %s\", s.info.Unsubscribe)\n\ts.log.Debugf(\"Close subject: %s\", s.info.Close)\n\treturn nil\n}", "func Register(ch chan ExecutionEvent, topics ...Topic) {\n\tfor _, t := range topics {\n\t\tsubscriberRegistry[t] = append(subscriberRegistry[t], ch)\n\t}\n}", "func (pr *PieceRegistry) SubscribeAllPartsDownloaded() chan PieceRange {\n\treturn pr.plansCompletedCh\n}", "func AllToAll(scope *Scope, input tf.Output, group_assignment tf.Output, concat_dimension int64, split_dimension int64, split_count int64) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{\"concat_dimension\": concat_dimension, \"split_dimension\": split_dimension, \"split_count\": split_count}\n\topspec := tf.OpSpec{\n\t\tType: \"AllToAll\",\n\t\tInput: []tf.Input{\n\t\t\tinput, group_assignment,\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func (h *Hub) Subscribe(t string, c chan []byte) {\n\th.Lock()\n\tdefer h.Unlock()\n\tchans, ok := h.topicChans[t]\n\tif !ok {\n\t\tchans = make(map[chan []byte]struct{})\n\t\th.topicChans[t] = chans\n\t}\n\tchans[c] = struct{}{}\n\ttopics, ok := h.chanTopics[c]\n\tif !ok {\n\t\ttopics = make(map[string]struct{})\n\t\th.chanTopics[c] = topics\n\t}\n\ttopics[t] = struct{}{}\n}", "func (qp *QuotesProvider) SubscribeAll(d time.Duration) chan schemas.ResultChannel {\n\tbufLength := len(qp.symbols)\n\tch := make(chan schemas.ResultChannel, 2*bufLength)\n\n\tgo func() {\n\t\tfor {\n\t\t\tquotes, err := qp.get()\n\t\t\tif err != nil {\n\t\t\t\tch <- schemas.ResultChannel{\n\t\t\t\t\tData: quotes,\n\t\t\t\t\tError: err,\n\t\t\t\t\tDataType: \"s\",\n\t\t\t\t}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tfor _, b := range quotes {\n\t\t\t\tch <- schemas.ResultChannel{\n\t\t\t\t\tData: b,\n\t\t\t\t\tError: err,\n\t\t\t\t\tDataType: \"s\",\n\t\t\t\t}\n\t\t\t}\n\t\t\ttime.Sleep(d)\n\t\t}\n\t}()\n\n\treturn ch\n}", "func (cp *ConnectPlugin) connectToAll(pctx context.Context) {\n\t// Connect to the list of peers. Hopefully Connect() skips the ones that we are already connect to\n\tfor _, pinfo := range cp.peerIDMap {\n\t\tif len(pinfo.Addrs) > 0 {\n\t\t\tfmt.Printf(\"Connecting to %s\\n\", pinfo)\n\t\t\tctx, cancelFunc := context.WithTimeout(pctx, connectTimeout)\n\t\t\tdefer cancelFunc()\n\t\t\terr := cp.api.PeerHost.Connect(ctx, pinfo)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(err)\n\t\t\t}\n\t\t}\n\n\t}\n}", "func (ks *KafkaStream) Subscribe(close chan struct{}) (<-chan Event, error) {\n\t// Capture current state of stream for use throughout this connection\n\ttopic := ks.config.Topic\n\toffset := ks.config.Offset\n\tstreamPartitions, err := ks.consumer.Partitions(topic)\n\t// Set up return channel for subscription events\n\tevents := make(chan Event, ks.config.SubscribeBufferSize)\n\tif err != nil {\n\t\treturn events, err\n\t}\n\t// Start subscription to stream in background\n\tgo func() {\n\t\t// For each partition in the stream set up a consumer to subscribe to messages\n\t\t// published to that partition\n\t\tfor _, partition := range streamPartitions {\n\t\t\tpartitionConsumer, err := ks.consumer.ConsumePartition(topic, partition, offset)\n\t\t\tif err != nil {\n\t\t\t\tks.logger.Errorf(\"Subscribe: Error %s to starting consumer for partition %d\", partition, err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// Start goroutine to run until the close channel is closed by the caller\n\t\t\tgo func(partitionConsumer sarama.PartitionConsumer) {\n\t\t\t\t<-close\n\t\t\t\tks.logger.Debug(\"Subscribe: Received close signal\")\n\t\t\t\t// at which point the connection to this partition consumer should be closed\n\t\t\t\tpartitionConsumer.AsyncClose()\n\t\t\t}(partitionConsumer)\n\t\t\t// Start goroutine to run until the close channel is closed by the caller\n\t\t\tgo func(partitionConsumer sarama.PartitionConsumer) {\n\t\t\t\t// to consume and convert messages for the subscriber to receive\n\t\t\t\tfor message := range partitionConsumer.Messages() {\n\t\t\t\t\tevent := convertMessageToEvent(message, topic)\n\t\t\t\t\tks.logger.Debugf(\"Subscribe: Received event %+v\", event)\n\t\t\t\t\tevents <- event\n\t\t\t\t}\n\t\t\t}(partitionConsumer)\n\t\t}\n\t}()\n\n\treturn events, nil\n}", "func (o *Okcoin) Subscribe(channelsToSubscribe []stream.ChannelSubscription) error {\n\treturn o.handleSubscriptions(\"subscribe\", channelsToSubscribe)\n}", "func (c Conference) Subscribe(id uuid.UUID, topic string, out chan Notification) {\n\ttree, ok := c.Room[topic]\n\tif !ok {\n\t\ttree = &bst.BinarySearchTree{}\n\t\tc.Room[topic] = tree\n\t}\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\ttree.Add(NewSubscriber(id, out))\n}", "func (cm *ConnectionManager) closeAll(onlyReceiving bool) (chs []*channeltype.Serialization, err error) {\n\tcm.lock.Lock()\n\tdefer cm.lock.Unlock()\n\tcm.initChannelTarget = 0\n\tif onlyReceiving {\n\t\tchs = cm.receivingChannels()\n\t} else {\n\t\tchs = cm.openChannels()\n\t}\n\tfor _, c := range chs {\n\t\t_, err = cm.api.Close(cm.tokenAddress, c.PartnerAddress())\n\t\tif err != nil {\n\t\t\tlog.Error(fmt.Sprintf(\"close channel %s error:%s\", c.ChannelIdentifier, err))\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func Subscribe(conn net.Conn, command []string, pubsub *PubSub) {\n\n\tfmt.Println(\"SUBSCRIBE TO:\", command[1:])\n\n\tch := make(chan string)\n\n\tdefer func() {\n\t\tconn.Close()\n\t\tpubsub.Unsubscribe <- UnsubscribeEvent{command[1], ch}\n\t}()\n\n\tpubsub.Subscribe <- SubscribeEvent{command[1], ch}\n\n\tfor msg := range ch {\n\t\t//fmt.Fprintf(conn, \"%s\\n\", msg)\n\t\t_, err := conn.Write([]byte(msg + \"\\n\"))\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n}", "func (dc *DurConn) subscribeAll(subs []*subscription, sc stan.Conn, scStaleCh chan struct{}) {\n\n\tsuccess := make([]bool, len(subs))\n\tfor {\n\t\tn := 0\n\t\tfor i, sub := range subs {\n\t\t\tif success[i] {\n\t\t\t\t// Already success.\n\t\t\t\tn++\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif err := dc.subscribe(sub, sc); err != nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tsuccess[i] = true\n\t\t\tn++\n\n\t\t\tselect {\n\t\t\tcase <-scStaleCh:\n\t\t\t\tdc.logger.Info(\"subscribe stale\")\n\t\t\t\treturn\n\t\t\tdefault:\n\t\t\t}\n\t\t}\n\n\t\tif n >= len(subs) {\n\t\t\t// All success.\n\t\t\treturn\n\t\t}\n\n\t\tselect {\n\t\tcase <-scStaleCh:\n\t\t\tdc.logger.Info(\"subscribe stale during retry wait\")\n\t\t\treturn\n\n\t\tcase <-time.After(dc.subRetryWait):\n\t\t}\n\n\t}\n\n}", "func (tg *TradesGroup) subscribe() {\n\tfor _, s := range tg.symbols {\n\t\tmessage := tradeSubsMessage{\n\t\t\tEvent: eventSubscribe,\n\t\t\tChannel: channelTrades,\n\t\t\tSymbol: \"t\" + strings.ToUpper(s.OriginalName),\n\t\t}\n\t\tif err := tg.wsClient.Write(message); err != nil {\n\t\t\tlog.Printf(\"[BITFINEX] Error subsciring to %v trades\", s.Name)\n\t\t\ttg.restart()\n\t\t\treturn\n\t\t}\n\t}\n\tlog.Println(\"[BITFINEX] Subscription ok\")\n}", "func (m *TeamItemRequestBuilder) AllChannels()(*ic08a09e622b3f3279dad3fb1dc0d9adf50886e9a08f52205e44e27965d46190a.AllChannelsRequestBuilder) {\n return ic08a09e622b3f3279dad3fb1dc0d9adf50886e9a08f52205e44e27965d46190a.NewAllChannelsRequestBuilderInternal(m.pathParameters, m.requestAdapter);\n}", "func waitForClusterToBecomeAwareOfAllSubscriptions(servers []server.NATSServer, subscriptionCount int) error {\n\ttimeout := time.After(time.Second * 5)\n\tfor {\n\t\tselect {\n\t\tcase <-timeout:\n\t\t\tfor _, server := range servers {\n\t\t\t\tif int(server.NumSubscriptions()) != subscriptionCount {\n\t\t\t\t\treturn errors.New(\"Timed out : waitForClusterToBecomeAwareOfAllSubscriptions()\")\n\t\t\t\t}\n\t\t\t}\n\t\t\tlog.Logger.Info().Msg(\"Entire cluster is aware of all subscriptions\")\n\t\t\treturn nil\n\t\tdefault:\n\t\t\tfor _, server := range servers {\n\t\t\t\tif int(server.NumSubscriptions()) != subscriptionCount {\n\t\t\t\t\tlog.Logger.Info().Msgf(\"Subscription count = %d\", server.NumSubscriptions())\n\t\t\t\t\ttime.Sleep(time.Millisecond)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t\tlog.Logger.Info().Msg(\"Entire cluster is aware of all subscriptions\")\n\t\t\treturn nil\n\t\t}\n\n\t}\n}", "func (c *Client) SubscribeAllDeviceMessages() {\n\ttok := c.c.Subscribe(AllDeviceTopic(c.opts.apiVer), c.opts.handlerQos,\n\t\tfunc(client pmqtt.Client, msg pmqtt.Message) {\n\t\t\tmessage := NewDeviceMessage().Decode(msg.Payload())\n\t\t\tlog.Infof(\"Topic: %s JSON: %s\", msg.Topic(), message.Message)\n\t\t\tLogMQTTMessage(msg)\n\t\t})\n\n\tif tok.WaitTimeout(c.opts.timeout) && tok.Error() != nil {\n\t\tlog.Errorf(\"Failed subscribe action: %v\", tok.Error())\n\t}\n}", "func (ba *BroadcastAgent) Subscribe(id int, ch chan chan<- []Message) {\n\tc, ok := <-ch\n\tif ok {\n\t\tba.parties[id] = c\n\t}\n}", "func Sub(c mqtt.Client, topic string) {\n\tvar choke = make(chan [2]string)\n\n\tvar f mqtt.MessageHandler = func(client mqtt.Client, msg mqtt.Message) {\n\t\tchoke <- [2]string{msg.Topic(), string(msg.Payload())}\n\t}\n\tfor {\n\t\tif token := c.Subscribe(topic, 0, f); token.Wait() && token.Error() != nil {\n\t\t\tmqtt.ERROR.Println(token.Error())\n\t\t\tos.Exit(1)\n\t\t}\n\t\tfor {\n\t\t\tincoming := <-choke\n\t\t\tmqtt.ERROR.Printf(\"Received:TOPIC: %s\\n\", incoming[0])\n\t\t\twriteFile(incoming[1])\n\t\t}\n\t}\n\n}", "func subscribeToEvents(bot *tgbotapi.BotAPI, redisClient *redis.Client, channel string) {\n\tpubsub := redisClient.Subscribe(channel)\n\tgo listen(bot, pubsub)\n}", "func (s *Subscription) Subscribe(channels ...string) {\n\tif len(channels) < 1 {\n\t\treturn\n\t}\n\n\ts.broker.dataChan <- &envData{false, &envSubscription{true, s, channels}, nil}\n}", "func (q channelQuery) All(ctx context.Context, exec boil.ContextExecutor) (ChannelSlice, error) {\n\tvar o []*Channel\n\n\terr := q.Bind(ctx, exec, &o)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"models: failed to assign all query results to Channel slice\")\n\t}\n\n\tif len(channelAfterSelectHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterSelectHooks(ctx, exec); err != nil {\n\t\t\t\treturn o, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn o, nil\n}", "func (ss *SubscriptionsService) All(ctx context.Context, opts *SubscriptionListOptions) (\n\tres *Response,\n\tsl *SubscriptionList,\n\terr error,\n) {\n\tu := \"v2/subscriptions\"\n\n\tres, err = ss.list(ctx, u, opts)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif err = json.Unmarshal(res.content, &sl); err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (ps *PubSub) Subscribe(channel ...interface{}) error {\n\tps.conn.Send(\"SUBSCRIBE\", channel...)\n\treturn ps.conn.Flush()\n}", "func (d StaticAgentDiscovery) Subscribe(c chan<- []string) { go func() { c <- d }() }", "func (t *Topic) Subscribe(cb *func(interface{})) {\n\tt.subs = append(t.subs, cb)\n}", "func (tc *consumer) Subscribe(topics map[string]int64) error {\n\tfor topic := range topics {\n\t\tif _, exists := tc.subscribedTopics[topic]; exists {\n\t\t\tlogger.Printf(\"consumer for %s already exists. This is strange\", topic)\n\t\t}\n\t\tlogger.Printf(\"Subscribe %s\", topic)\n\t\ttc.subscribedTopics[topic] = tc.tester.getOrCreateQueue(topic).bindConsumer(tc, true)\n\t\ttc.subscribedTopics[topic].rebalance()\n\t\ttc.subscribedTopics[topic].startLoop(false)\n\t}\n\treturn nil\n}", "func (mem *Member) AllToAll() {\n\t// Encode the membership list to send it\n\tb := new(bytes.Buffer)\n\te := gob.NewEncoder(b)\n\terr := e.Encode(mem.membershipList)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\t//Info.Println(\"Sending All-to-All.\")\n\t// Send heartbeatmsg and membership list to all members\n\tmem.SendAll(HeartbeatMsg, b.Bytes())\n}", "func startFanOut(quitC <-chan struct{}) (inC chan *model.PortalMsg, subC chan chan *model.PortalMsg) {\n\n\tinC = make(chan *model.PortalMsg, 1)\n\tsubC = make(chan chan *model.PortalMsg, 1)\n\n\tgo func(quitC <-chan struct{}) {\n\t\tdefer fmt.Println(\"fanout stopped\")\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-quitC:\n\t\t\t\treturn\n\t\t\tcase sub := <-subC:\n\t\t\t\tif nil != sub {\n\t\t\t\t\tsubs.Lock()\n\t\t\t\t\tsubs.subs = append(subs.subs, sub)\n\t\t\t\t\tsubs.Unlock()\n\t\t\t\t\tfmt.Println(\"subscription added\")\n\t\t\t\t}\n\t\t\tcase msg := <-inC:\n\t\t\t\t// The subscriptions are notified of a message and are groomed out\n\t\t\t\t// on unrecoverable failures using https://github.com/golang/go/wiki/SliceTricks#filtering-without-allocating\n\t\t\t\tsubs.Lock()\n\t\t\t\tnewSubs := subs.subs[:0]\n\t\t\t\tfor _, ch := range subs.subs {\n\t\t\t\t\tfunc() {\n\t\t\t\t\t\tdefer func() {\n\t\t\t\t\t\t\tif r := recover(); r == nil {\n\t\t\t\t\t\t\t\tnewSubs = append(newSubs, ch)\n\t\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tfmt.Println(\"subscription dropped failed to send\")\n\t\t\t\t\t\t}()\n\t\t\t\t\t\tselect {\n\t\t\t\t\t\tcase ch <- msg:\n\t\t\t\t\t\tcase <-time.After(250 * time.Millisecond):\n\t\t\t\t\t\t\tfmt.Println(\"subscription failed to send\")\n\t\t\t\t\t\t}\n\t\t\t\t\t}()\n\t\t\t\t}\n\t\t\t\tsubs.subs = newSubs\n\t\t\t\tsubs.Unlock()\n\t\t\t}\n\t\t}\n\t}(quitC)\n\n\treturn inC, subC\n}", "func (channel Channel) subscribe(observers ...Observer) {\n\tchannel.checkChannelMap()\n\tfor _, observer := range observers {\n\t\tchannel.observers[observer.id] = observer\n\t\tfmt.Printf(\"New observer %s subscribed in channel %s \\n\", observer.id, channel.id)\n\t}\n}", "func (kew *KeyspaceEventWatcher) Subscribe() chan *KeyspaceEvent {\n\tkew.subsMu.Lock()\n\tdefer kew.subsMu.Unlock()\n\tc := make(chan *KeyspaceEvent, 2)\n\tkew.subs[c] = struct{}{}\n\treturn c\n}", "func (b *Broker) Subscribe(s *Subscriber, topics ...string) {\n\tb.tlock.Lock()\n\tdefer b.tlock.Unlock()\n\tfor _, topic := range topics {\n\t\tif nil == b.topics[topic] {\n\t\t\tb.topics[topic] = Subscribers{}\n\t\t}\n\t\ts.topics[topic] = true\n\t\tb.topics[topic][s.id] = s\n\t}\n}", "func SubscriptionTo(ctx context.Context, b *Broadcaster) *Subscription {\n\n\tctx, done := context.WithCancel(ctx)\n\tret := &Subscription{\n\t\tctx: ctx,\n\t\tdone: done,\n\t\tc: make(chan struct{}),\n\t}\n\n\tgo func() {\n\t\tfor {\n\t\t\tnCh := b.Register()\n\t\t\tselect {\n\t\t\tcase <-nCh:\n\t\t\t\tlog.Printf(\"received event, sending\")\n\t\t\t\tret.send()\n\t\t\tcase <-ctx.Done():\n\t\t\t\tlog.Printf(\"context done (in SuscriptioTo)\")\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\treturn ret\n}", "func (h *Handler) All(e *emptypb.Empty, s protobufs.HackerNews_AllServer) error {\n\titems, err := h.reader.GetAllItems(s.Context())\n\tif err != nil {\n\t\treturn fmt.Errorf(\"getAllItems: %w\", err)\n\t}\n\n\tfor _, item := range items {\n\t\ts.Send(grpc.ToProto(item))\n\t}\n\n\treturn nil\n}", "func (b *Topics) Subscribe(s *Subscriber, topics ...string) {\n\tb.topic_lock.Lock()\n\tdefer b.topic_lock.Unlock()\n\tfor _, topic := range topics {\n\t\tif nil == b.sub_topics[topic] {\n\t\t\tfmt.Println(\"!!! topic does not exist !!!\")\n\t\t}\n\t\ts.AddTopic(topic)\n\t\tb.sub_topics[topic][s.id] = s\n\t}\n}", "func (s *Client) SubscribeNodes(name string,done chan struct{}) (chan []*ServiceNode, error){\n\tpath := s.zkRoot + \"/\" + name\n\terr := s.ensurePath(name)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\tpanic(err)\n\t}\n\t// 获取字节点名称\n\tnodesChan := make(chan []*ServiceNode)\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase _,ok := <- done:\n\t\t\t\tif !ok {\n\t\t\t\t\tlog.Printf(\"done send, stop subscribing channel of %s\\n\", name)\n\t\t\t\t\tclose(nodesChan)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\tdefault:\n\t\t\t}\n\t\t\tchilds, _,ch, err := s.conn.ChildrenW(path)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(err.Error())\n\t\t\t\tpanic(err)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tvar nodes []*ServiceNode\n\t\t\tfor _, child := range childs {\n\t\t\t\tfullPath := path + \"/\" + child\n\t\t\t\tdata, _, err := s.conn.Get(fullPath)\n\t\t\t\tif err != nil {\n\t\t\t\t\tif err == zk.ErrNoNode {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tlog.Println(err.Error())\n\t\t\t\t}\n\t\t\t\tnode := new(ServiceNode)\n\t\t\t\terr = json.Unmarshal(data, node)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Println(err.Error())\n\t\t\t\t}\n\t\t\t\tnodes = append(nodes, node)\n\t\t\t}\n\t\t\t/* send new nodes to chan */\n\t\t\tnodesChan <- nodes\n\t\t\t/* when receiving events*/\n\t\t\te := <- ch\n\t\t\tif e.Type == zk.EventNodeChildrenChanged {\n\t\t\t\tlog.Printf(\"zk Path: %s, children changed\\n\", name)\n\t\t\t}else {\n\t\t\t\tlog.Printf(\"zk Path: %s, unexpected events\\n\",name)\n\t\t\t}\n\n\t\t}\n\t}()\n\treturn nodesChan, nil\n}", "func (service *RedigoService) Subscribe(ctx context.Context, subscribed SubscribedHandler, subscription SubscriptionHandler, channels ...string) error {\n\n\tc, err := redis.Dial(\"tcp\", service.Configuration.Address,\n\t\t// Read timeout on server should be greater than ping period.\n\t\tredis.DialReadTimeout(service.Configuration.PubSub.ReadTimeout),\n\t\tredis.DialWriteTimeout(service.Configuration.PubSub.WriteTimeout),\n\t)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer c.Close()\n\n\tpsc := redis.PubSubConn{Conn: c}\n\tif err := psc.Subscribe(redis.Args{}.AddFlat(channels)...); err != nil {\n\t\treturn err\n\t}\n\n\tdone := make(chan error, 1)\n\n\t// Start a goroutine to receive notifications from the server.\n\tgo func() {\n\t\tfor {\n\t\t\tswitch n := psc.Receive().(type) {\n\t\t\tcase error:\n\t\t\t\t// Increment to count failures\n\t\t\t\tservice.Collector.subscribeFailures.Inc()\n\n\t\t\t\tdone <- n\n\t\t\t\treturn\n\t\t\tcase redis.Message:\n\t\t\t\tif err := subscription(n.Channel, n.Data); err != nil {\n\n\t\t\t\t\t// Increment to count failures\n\t\t\t\t\tservice.Collector.subscribeFailures.Inc()\n\n\t\t\t\t\tdone <- err\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\t// Increment to count success\n\t\t\t\tservice.Collector.subscribeSuccesses.Inc()\n\n\t\t\tcase redis.Subscription:\n\t\t\t\tswitch n.Count {\n\t\t\t\tcase len(channels):\n\n\t\t\t\t\t// Increment 1 in subscriptionsActive\n\t\t\t\t\tservice.Collector.subscriptionsActive.Inc()\n\n\t\t\t\t\t// Notify application when all channels are subscribed.\n\t\t\t\t\tif err := subscribed(); err != nil {\n\n\t\t\t\t\t\t// Increment to count failures\n\t\t\t\t\t\tservice.Collector.subscribeFailures.Inc()\n\n\t\t\t\t\t\tdone <- err\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\n\t\t\t\t\t// Increment to count success\n\t\t\t\t\tservice.Collector.subscribeSuccesses.Inc()\n\n\t\t\t\tcase 0:\n\t\t\t\t\t// Return from the goroutine when all channels are unsubscribed.\n\t\t\t\t\tdone <- nil\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\t// A ping is set to the server with this period to test for the health of\n\t// the connection and server.\n\tticker := time.NewTicker(service.Configuration.PubSub.HealthCheckInterval)\n\tdefer ticker.Stop()\n\nloop:\n\tfor err == nil {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\t// Send ping to test health of connection and server. If\n\t\t\t// corresponding pong is not received, then receive on the\n\t\t\t// connection will timeout and the receive goroutine will exit.\n\t\t\tif err = psc.Ping(\"\"); err != nil {\n\n\t\t\t\t// Increment to count failures\n\t\t\t\tservice.Collector.subscribeFailures.Inc()\n\n\t\t\t\tbreak loop\n\t\t\t}\n\t\tcase <-ctx.Done():\n\t\t\tbreak loop\n\t\tcase err := <-done:\n\t\t\t// Return error from the receive goroutine.\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Decrement 1 in subscriptionsActive\n\tservice.Collector.subscriptionsActive.Dec()\n\n\t// Signal the receiving goroutine to exit by unsubscribing from all channels.\n\tpsc.Unsubscribe()\n\n\t// Wait for goroutine to complete.\n\treturn <-done\n}", "func InitRegistry() {\n\tsubscriberRegistry = make(map[Topic][]chan ExecutionEvent)\n\tsubscriberRegistry[SuiteStart] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[ScenarioStart] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[ConceptStart] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[StepStart] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[SuiteEnd] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[ConceptEnd] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[ScenarioEnd] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[SpecEnd] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[SuiteEnd] = make([]chan ExecutionEvent, 0)\n}", "func (ps *PubsubApi) LogsSubscribe(ctx context.Context, crit filters.FilterCriteria) (*rpc.Subscription, error) {\n\tif ps.s.context().eventBus == nil {\n\t\t// @Note: Should not happen!\n\t\tlog.Error(\"rpc: eventbus nil, not support Subscribetion!!!\")\n\t\treturn nil, rpc.ErrNotificationsUnsupported\n\t}\n\n\tnotifier, supported := rpc.NotifierFromContext(ctx)\n\tif !supported {\n\t\treturn nil, rpc.ErrNotificationsUnsupported\n\t}\n\n\tsubscription := notifier.CreateSubscription()\n\n\tsuberName := fmt.Sprintf(\"rpc-log-suber-%s\", subscription.ID)\n\tebCtx := context.Background()\n\tlogsCh := make(chan interface{}, 128)\n\tif err := ps.context().eventBus.Subscribe(ebCtx, suberName, types.EventQueryLog, logsCh); err != nil {\n\t\tlog.Warn(\"rpc: Subscribe fail\", \"err\", err)\n\t\treturn nil, err\n\t}\n\n\tgo func() {\n\t\tdefer ps.context().eventBus.Unsubscribe(ebCtx, suberName, types.EventQueryLog)\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase ev := <-logsCh:\n\t\t\t\tlogs := ev.(types.EventDataLog).Logs\n\t\t\t\tlogs = filterLogs(logs, crit.FromBlock.ToInt(), crit.ToBlock.ToInt(), crit.Addresses, crit.Topics)\n\t\t\t\tfor _, l := range logs {\n\t\t\t\t\tnotifier.Notify(subscription.ID, l)\n\t\t\t\t\tlog.Info(\"rpc: notify success\", \"suber\", suberName, \"log\", l)\n\t\t\t\t}\n\t\t\tcase <-notifier.Closed():\n\t\t\t\tlog.Info(\"rpc LogSubscribe: unsubscribe\", \"suber\", suberName)\n\t\t\t\treturn\n\t\t\tcase err := <-subscription.Err():\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error(\"rpc subscription: error\", \"suber\", suberName, \"err\", err)\n\t\t\t\t} else {\n\t\t\t\t\tlog.Info(\"rpc subscription: exit\", \"suber\", suberName)\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\tlog.Info(\"rpc LogsSubscribe: ok\", \"name\", suberName, \"crit\", crit.String())\n\treturn subscription, nil\n}", "func InitiateConsumers(ns *servicebus.Namespace, availableTopicList []*servicebus.TopicEntity, componentName string,\n\tsubscriptionIdleTimeDuration time.Duration) {\n\tbindingKeys := []string {tokenRevocation, notification}\n\n\tfor _, key := range bindingKeys {\n\t\tgo func(key string) {\n\t\t\tlogger.LoggerMgw.Info(\"[TEST][FEATURE_FLAG_REPLACE_EVENT_HUB] starting the consumer for key : \" + key)\n\t\t\tstartBrokerConsumer(key, ns, availableTopicList, componentName,\n\t\t\t\tservicebus.SubscriptionWithAutoDeleteOnIdle(&subscriptionIdleTimeDuration))\n\t\t\tselect {}\n\t\t}(key)\n\t}\n}", "func (s *Slave) Subscribe(lsns ...uint64) (it PacketIterator, err error) {\n\tif len(lsns) == 0 || len(lsns) >= VClockMax {\n\t\treturn nil, ErrVectorClock\n\t}\n\t//don't call subscribe if there are no options had been set or before join request\n\tif !s.IsInReplicaSet() {\n\t\treturn nil, ErrNotInReplicaSet\n\t}\n\tif err = s.subscribe(lsns...); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// set iterator for the Next method\n\ts.next = s.nextXlog\n\n\t// Tarantool >= 1.7.0 sends periodic heartbeat messages\n\tif s.Version() < version1_7_0 {\n\t\treturn s, nil\n\t}\n\n\t// Start sending heartbeat messages to master\n\tgo s.heartbeat()\n\n\treturn s, nil\n}", "func SubListAll(w http.ResponseWriter, r *http.Request) {\n\n\tvar err error\n\tvar strPageSize string\n\tvar pageSize int\n\tvar res subscriptions.PaginatedSubscriptions\n\n\t// Init output\n\toutput := []byte(\"\")\n\n\t// Add content type header to the response\n\tcontentType := \"application/json\"\n\tcharset := \"utf-8\"\n\tw.Header().Add(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\t// Grab context references\n\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\tprojectUUID := gorillaContext.Get(r, \"auth_project_uuid\").(string)\n\troles := gorillaContext.Get(r, \"auth_roles\").([]string)\n\n\turlValues := r.URL.Query()\n\tpageToken := urlValues.Get(\"pageToken\")\n\tstrPageSize = urlValues.Get(\"pageSize\")\n\n\t// if this route is used by a user who only has a consumer role\n\t// return all subscriptions that he has access to\n\tuserUUID := \"\"\n\tif !auth.IsProjectAdmin(roles) && !auth.IsServiceAdmin(roles) && auth.IsConsumer(roles) {\n\t\tuserUUID = gorillaContext.Get(r, \"auth_user_uuid\").(string)\n\t}\n\n\tif strPageSize != \"\" {\n\t\tif pageSize, err = strconv.Atoi(strPageSize); err != nil {\n\t\t\tlog.Errorf(\"Pagesize %v produced an error while being converted to int: %v\", strPageSize, err.Error())\n\t\t\terr := APIErrorInvalidData(\"Invalid page size\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t}\n\n\tif res, err = subscriptions.Find(projectUUID, userUUID, \"\", pageToken, int32(pageSize), refStr); err != nil {\n\t\terr := APIErrorInvalidData(\"Invalid page token\")\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Output result to JSON\n\tresJSON, err := res.ExportJSON()\n\tif err != nil {\n\t\terr := APIErrExportJSON()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Write Response\n\toutput = []byte(resJSON)\n\trespondOK(w, output)\n\n}", "func (agent *SrlAgent) Subscribe(topic string) {\n\tagent.Logger.Debug(\"Subscribing for \", topic)\n\top := pb.NotificationRegisterRequest_AddSubscription\n\tresponse := &pb.NotificationRegisterResponse{}\n\tvar err error\n\tif topic == \"intf\" {\n\t\tresponse, err = agent.Stub.NotificationRegister(agent.Ctx, &pb.NotificationRegisterRequest{\n\t\t\tStreamId: agent.StreamID,\n\t\t\tOp: op,\n\t\t\tSubscriptionTypes: &pb.NotificationRegisterRequest_Intf{},\n\t\t})\n\t}\n\tif topic == \"nw_inst\" {\n\t\tresponse, err = agent.Stub.NotificationRegister(agent.Ctx, &pb.NotificationRegisterRequest{\n\t\t\tStreamId: agent.StreamID,\n\t\t\tOp: op,\n\t\t\tSubscriptionTypes: &pb.NotificationRegisterRequest_NwInst{},\n\t\t})\n\t}\n\tif topic == \"lldp\" {\n\t\tresponse, err = agent.Stub.NotificationRegister(agent.Ctx, &pb.NotificationRegisterRequest{\n\t\t\tStreamId: agent.StreamID,\n\t\t\tOp: op,\n\t\t\tSubscriptionTypes: &pb.NotificationRegisterRequest_LldpNeighbor{},\n\t\t})\n\t}\n\tif topic == \"route\" {\n\t\tresponse, err = agent.Stub.NotificationRegister(agent.Ctx, &pb.NotificationRegisterRequest{\n\t\t\tStreamId: agent.StreamID,\n\t\t\tOp: op,\n\t\t\tSubscriptionTypes: &pb.NotificationRegisterRequest_Route{},\n\t\t})\n\t}\n\tif topic == \"cfg\" {\n\t\tresponse, err = agent.Stub.NotificationRegister(agent.Ctx, &pb.NotificationRegisterRequest{\n\t\t\tStreamId: agent.StreamID,\n\t\t\tOp: op,\n\t\t\tSubscriptionTypes: &pb.NotificationRegisterRequest_Config{},\n\t\t})\n\t}\n\tif topic == \"app\" {\n\t\tresponse, err = agent.Stub.NotificationRegister(agent.Ctx, &pb.NotificationRegisterRequest{\n\t\t\tStreamId: agent.StreamID,\n\t\t\tOp: op,\n\t\t\tSubscriptionTypes: &pb.NotificationRegisterRequest_Appid{},\n\t\t})\n\t}\n\tif err != nil {\n\t\tagent.Logger.Debug(\"Failed to subscribe for \", topic)\n\t}\n\tagent.Logger.Debug(\"Response for Notification register for \", topic, \" is \", response.Status)\n}", "func RemoveAllSubscriptionsOnConnect(clientID string) {\n\tsubs := new(map[string]uint8)\n\tkey := fmt.Sprintf(\"chatterbox.client-subs.%s\", clientID)\n\tGlobalRedisClient.Fetch(key, subs)\n\n\tGlobalRedisClient.Delete(key)\n\n\tGlobalSubsLock.Lock()\n\tfor topic := range *subs {\n\t\tdelete(GlobalSubs[topic], clientID)\n\t}\n\tGlobalSubsLock.Unlock()\n\n}", "func (c *Coordinator) Subscribe(ss ...func(*Config) error) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tc.subscribers = append(c.subscribers, ss...)\n}", "func (s *Client) Subscribe1Node(name string) (chan ServiceNode, error){\n\tresChan := make(chan ServiceNode)\n\tdone := make(chan struct{})\n\tnodeChan, _ := s.SubscribeNodes(name,done)\n\n\tgo func() {\n\t\tfor nodes := range nodeChan {\n\t\t\tif len(nodes) == 0 {\n\t\t\tclose(resChan)\n\t\t\tclose(done)\n\t\t\t}else if len(nodes) != 1 {\n\t\t\t\tlog.Printf(\"not suppose to see multiple primary nodes\")\n\t\t\t}else {\n\t\t\t\tresChan <- *nodes[0]\n\t\t\t}\n\t\t}\n\t}()\n\treturn resChan,nil\n}", "func (s *Cluster) NotifySubscribe(conn security.ID, ssid subscription.Ssid) {\n\tevent := SubscriptionEvent{\n\t\tPeer: s.name,\n\t\tConn: conn,\n\t\tSsid: ssid,\n\t}\n\n\t// Add to our global state\n\ts.state.Add(event.Encode())\n\t// Create a delta for broadcasting just this operation\n\top := newSubscriptionState()\n\top.Add(event.Encode())\n\ts.gossip.GossipBroadcast(op)\n}", "func chatroom() {\n\tarchive := list.New()\n\tsubscribers := list.New()\n\n\tfor {\n\t\tselect {\n\t\tcase ch := <-subscribe:\n\t\t\tvar events []Event\n\t\t\tfor e := archive.Front(); e != nil; e = e.Next() {\n\t\t\t\tevents = append(events, e.Value.(Event))\n\t\t\t}\n\t\t\tsubscriber := make(chan Event, 10)\n\t\t\tsubscribers.PushBack(subscriber)\n\t\t\tch <- Subscription{events, subscriber}\n\n\t\tcase event := <-publish:\n\t\t\tfor ch := subscribers.Front(); ch != nil; ch = ch.Next() {\n\t\t\t\tch.Value.(chan Event) <- event\n\t\t\t}\n\t\t\tif archive.Len() >= archiveSize {\n\t\t\t\tarchive.Remove(archive.Front())\n\t\t\t}\n\t\t\tarchive.PushBack(event)\n\n\t\tcase unsub := <-unsubscribe:\n\t\t\tfor ch := subscribers.Front(); ch != nil; ch = ch.Next() {\n\t\t\t\tif ch.Value.(chan Event) == unsub {\n\t\t\t\t\tsubscribers.Remove(ch)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func (p RPCServer) Subscribe(ctx context.Context, in *pb.SubscriptionRequest) (*pb.Subscription, error) {\n\tsubID := *p.currentSubID\n\t*p.currentSubID++\n\n\tlogrus.WithField(\"topic\", in.Topic).WithField(\"subID\", subID).Debug(\"subscribed to new messages\")\n\n\tp.subChannels[subID] = make(chan []byte)\n\tp.cancelChannels[subID] = make(chan bool)\n\n\ts, err := p.service.RegisterHandler(in.Topic, func(b []byte) error {\n\t\tselect {\n\t\tcase p.subChannels[subID] <- b:\n\t\tdefault:\n\t\t}\n\t\treturn nil\n\t})\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tp.subscriptions[subID] = s\n\n\treturn &pb.Subscription{ID: subID}, nil\n}", "func (h *Hub) SendAll(messageType int, data []byte) {\n\tlogger.Info.Println(\"message delivered to all:\", string(data), \"message type:\", messageType)\n\tfor client, _ := range h.Clients {\n\t\tclient.WriteMessage(messageType, data)\n\t}\n}", "func (c *clientWrapper) WatchAll(namespaces []string, stopCh <-chan struct{}) (<-chan interface{}, error) {\n\teventCh := make(chan interface{}, 1)\n\teventHandler := &resourceEventHandler{ev: eventCh}\n\n\tif len(namespaces) == 0 {\n\t\tnamespaces = []string{metav1.NamespaceAll}\n\t\tc.isNamespaceAll = true\n\t}\n\n\tc.watchedNamespaces = namespaces\n\n\tnotOwnedByHelm := func(opts *metav1.ListOptions) {\n\t\topts.LabelSelector = \"owner!=helm\"\n\t}\n\n\tlabelSelectorOptions := func(options *metav1.ListOptions) {\n\t\toptions.LabelSelector = c.labelSelector\n\t}\n\n\tc.factoryNamespace = informers.NewSharedInformerFactory(c.csKube, resyncPeriod)\n\tc.factoryNamespace.Core().V1().Namespaces().Informer().AddEventHandler(eventHandler)\n\n\tc.factoryGatewayClass = externalversions.NewSharedInformerFactoryWithOptions(c.csGateway, resyncPeriod, externalversions.WithTweakListOptions(labelSelectorOptions))\n\tc.factoryGatewayClass.Gateway().V1alpha2().GatewayClasses().Informer().AddEventHandler(eventHandler)\n\n\t// TODO manage Reference Policy\n\t// https://gateway-api.sigs.k8s.io/v1alpha2/references/spec/#gateway.networking.k8s.io/v1alpha2.ReferencePolicy\n\n\tfor _, ns := range namespaces {\n\t\tfactoryGateway := externalversions.NewSharedInformerFactoryWithOptions(c.csGateway, resyncPeriod, externalversions.WithNamespace(ns))\n\t\tfactoryGateway.Gateway().V1alpha2().Gateways().Informer().AddEventHandler(eventHandler)\n\t\tfactoryGateway.Gateway().V1alpha2().HTTPRoutes().Informer().AddEventHandler(eventHandler)\n\t\tfactoryGateway.Gateway().V1alpha2().TCPRoutes().Informer().AddEventHandler(eventHandler)\n\t\tfactoryGateway.Gateway().V1alpha2().TLSRoutes().Informer().AddEventHandler(eventHandler)\n\n\t\tfactoryKube := informers.NewSharedInformerFactoryWithOptions(c.csKube, resyncPeriod, informers.WithNamespace(ns))\n\t\tfactoryKube.Core().V1().Services().Informer().AddEventHandler(eventHandler)\n\t\tfactoryKube.Core().V1().Endpoints().Informer().AddEventHandler(eventHandler)\n\n\t\tfactorySecret := informers.NewSharedInformerFactoryWithOptions(c.csKube, resyncPeriod, informers.WithNamespace(ns), informers.WithTweakListOptions(notOwnedByHelm))\n\t\tfactorySecret.Core().V1().Secrets().Informer().AddEventHandler(eventHandler)\n\n\t\tc.factoriesGateway[ns] = factoryGateway\n\t\tc.factoriesKube[ns] = factoryKube\n\t\tc.factoriesSecret[ns] = factorySecret\n\t}\n\n\tc.factoryNamespace.Start(stopCh)\n\tc.factoryGatewayClass.Start(stopCh)\n\n\tfor _, ns := range namespaces {\n\t\tc.factoriesGateway[ns].Start(stopCh)\n\t\tc.factoriesKube[ns].Start(stopCh)\n\t\tc.factoriesSecret[ns].Start(stopCh)\n\t}\n\n\tfor t, ok := range c.factoryNamespace.WaitForCacheSync(stopCh) {\n\t\tif !ok {\n\t\t\treturn nil, fmt.Errorf(\"timed out waiting for controller caches to sync %s\", t.String())\n\t\t}\n\t}\n\n\tfor t, ok := range c.factoryGatewayClass.WaitForCacheSync(stopCh) {\n\t\tif !ok {\n\t\t\treturn nil, fmt.Errorf(\"timed out waiting for controller caches to sync %s\", t.String())\n\t\t}\n\t}\n\n\tfor _, ns := range namespaces {\n\t\tfor t, ok := range c.factoriesGateway[ns].WaitForCacheSync(stopCh) {\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"timed out waiting for controller caches to sync %s in namespace %q\", t.String(), ns)\n\t\t\t}\n\t\t}\n\n\t\tfor t, ok := range c.factoriesKube[ns].WaitForCacheSync(stopCh) {\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"timed out waiting for controller caches to sync %s in namespace %q\", t.String(), ns)\n\t\t\t}\n\t\t}\n\n\t\tfor t, ok := range c.factoriesSecret[ns].WaitForCacheSync(stopCh) {\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"timed out waiting for controller caches to sync %s in namespace %q\", t.String(), ns)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn eventCh, nil\n}", "func (csi ChannelStoreImpl) GetAll(db *gorm.DB) []models.Channel {\n\tchannels := []models.Channel{}\n\tdb.Find(&channels)\n\treturn channels\n}", "func (s *T) Subscriptions() <-chan map[string][]string {\n\treturn s.subscriptionsCh\n}", "func (t *Topic) Subscribe(ctx context.Context) <-chan interface{} {\n\tch := make(chan interface{})\n\tt.subs[ch] = ctx\n\treturn ch\n}", "func (c *Client) AllMessages() <-chan *Message {\n\treturn c.subs.subscribe(keyAnyMsg)\n}", "func (this *Publisher) Subscribe() <-chan gopi.Event {\n\tthis.Lock()\n\tdefer this.Unlock()\n\n\t// Create channels with a capacity of one\n\tif this.channels == nil {\n\t\tthis.channels = make([]chan gopi.Event, 0, 1)\n\t}\n\t// Return a new channel\n\tchannel := make(chan gopi.Event)\n\tthis.channels = append(this.channels, channel)\n\treturn channel\n}", "func (room *Room) subscribeToRoomMessages() {\n\tpubsub := room.redis.Subscribe(ctx, room.GetId())\n\n\tch := pubsub.Channel()\n\n\tfor msg := range ch {\n\t\troom.broadcastToClientsInRoom([]byte(msg.Payload))\n\t}\n}", "func (psc *PubSubChannel) Subscribe() *Subscription {\n psc.subsMutex.Lock()\n defer psc.subsMutex.Unlock()\n newSubscription := newSubscription(strconv.Itoa(psc.nextId))\n psc.nextId++\n psc.subscriptions = append(psc.subscriptions, &newSubscription)\n newSubscription.removeSub = func() {\n psc.subsMutex.Lock()\n defer psc.subsMutex.Unlock()\n\n for i, subscription := range psc.subscriptions {\n if subscription.signal == newSubscription.signal {\n fmt.Println(\"Unsubscribing\", \"id\", subscription.id)\n subscription.closed = true\n psc.subscriptions = append(psc.subscriptions[:i], psc.subscriptions[i+1:]...)\n }\n }\n }\n return &newSubscription\n}", "func InitiateConsumers(subscriptionMetaDataList []Subscription, reconnectInterval time.Duration) {\n\tfor _, subscriptionMetaData := range subscriptionMetaDataList {\n\t\tgo func(subscriptionMetaData Subscription) {\n\t\t\tstartBrokerConsumer(subscriptionMetaData, reconnectInterval)\n\t\t}(subscriptionMetaData)\n\t}\n}", "func (b *Broker) EventSubscribe(subscriptions []string) (events gp.MsgQueue) {\n\tcommands := make(chan gp.QueueCommand)\n\tmessages := make(chan []byte)\n\tevents = gp.MsgQueue{Commands: commands, Messages: messages}\n\tconn := b.pool.Get()\n\tpsc := redis.PubSubConn{Conn: conn}\n\tfor _, s := range subscriptions {\n\t\tpsc.Subscribe(s)\n\t}\n\tgo controller(&psc, events.Commands)\n\tgo messageReceiver(&psc, events.Messages)\n\tlog.Println(\"New websocket connection created.\")\n\treturn events\n}", "func (nch *NatsConnectionHandler) SubscribeToTopics(topics []string, messageHandler IKeptnNatsMessageHandler) error {\n\tif nch.natsURL == \"\" {\n\t\treturn errors.New(\"no PubSub URL defined\")\n\t}\n\n\tif nch.natsConnection == nil || !nch.natsConnection.IsConnected() {\n\t\tif err := nch.renewNatsConnection(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif nch.jetStream == nil {\n\t\tif err := nch.setupJetStreamContext(topics); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(topics) > 0 && !IsEqual(nch.topics, topics) {\n\t\tnch.RemoveAllSubscriptions()\n\t\tnch.topics = topics\n\n\t\tfor _, topic := range nch.topics {\n\t\t\tsubscription := NewPullSubscription(nch.ctx, queueGroup, topic, nch.jetStream, messageHandler.Process)\n\t\t\tif err := subscription.Activate(); err != nil {\n\t\t\t\treturn fmt.Errorf(\"could not start subscription: %s\", err.Error())\n\t\t\t}\n\t\t\tnch.subscriptions = append(nch.subscriptions, subscription)\n\t\t}\n\t}\n\treturn nil\n}", "func (q subscriberQuery) All(ctx context.Context, exec boil.ContextExecutor) (SubscriberSlice, error) {\n\tvar o []*Subscriber\n\n\terr := q.Bind(ctx, exec, &o)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"models: failed to assign all query results to Subscriber slice\")\n\t}\n\n\tif len(subscriberAfterSelectHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterSelectHooks(ctx, exec); err != nil {\n\t\t\t\treturn o, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn o, nil\n}", "func ToSubscrible(eid string) {\n\tif alertSubscribeQueue == \"\" {\n\t\treturn\n\t}\n\tqueueCli.LPush(alertSubscribeQueue, eid)\n\tllen := queueCli.LLen(alertSubscribeQueue).Val()\n\tlog.Debug(\"lpush\", \"queue\", alertSubscribeQueue, \"eid\", eid, \"llen\", llen)\n\tif llen > 1e4 {\n\t\tqueueCli.LTrim(alertSubscribeQueue, 50, -1)\n\t\tlog.Debug(\"ltrim\", \"queue\", alertSubscribeQueue)\n\t}\n}", "func (b *Broker) Broadcast(payload interface{}, topics ...string) {\n\tfor _, topic := range topics {\n\t\tfor _, s := range b.topics[topic] {\n\t\t\tm := &Message{\n\t\t\t\ttopic: topic,\n\t\t\t\tpayload: payload,\n\t\t\t\tcreatedAt: time.Now().UnixNano(),\n\t\t\t}\n\t\t\tgo (func(s *Subscriber) {\n\t\t\t\ts.Signal(m)\n\t\t\t})(s)\n\t\t}\n\t}\n}", "func (h *clientHub) Broadcast(ch string, message []byte) error {\n\th.RLock()\n\tdefer h.RUnlock()\n\n\t// get connections currently subscribed on channel\n\tchannelSubscriptions, ok := h.subs[ch]\n\tif !ok {\n\t\treturn nil\n\t}\n\n\t// iterate over them and send message individually\n\tmsg := NewQueuedMessage(message, true)\n\n\tfor uid := range channelSubscriptions {\n\t\tc, ok := h.conns[uid]\n\t\tif !ok {\n\t\t\tcontinue\n\t\t}\n\t\tc.Send(msg)\n\t}\n\treturn nil\n}", "func (r *Raft) sendToAll(msg interface{}) {\n\t//fmt.Println(\"Server-Raft map:\", server_raft_map)\n\tfor k := range server_raft_map {\n\t\t//fmt.Println(\"Id from map is:\", k, r.Myconfig.Id)\n\t\tif r.Myconfig.Id != k { //send to all except self\n\t\t\tgo send(k, msg) //removed go\n\t\t\t//fmt.Println(\"After sending RV\")\n\t\t}\n\t}\n\n}", "func (c *CoordinatorHelper) BroadcastAll(\n\tctx context.Context,\n) error {\n\treturn c.broadcastStorage.BroadcastAll(ctx, true)\n}", "func (q *Qlient) SubOn(topic string) (chan []byte, error) {\n\tsub, ok := q.subs[topic]\n\tif !ok {\n\t\tvar err error\n\t\tsub, err = q.newConsumer(topic)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn sub, nil\n}", "func (m *Client) Subscribe() chan proto.Payload {\n\tmessages := make(chan proto.Payload)\n\ttq := []proto.TopicQos{proto.TopicQos{Topic: m.topic, Qos: 0}}\n\tm.client.Subscribe(tq)\n\tgo func() {\n\t\tdefer close(messages)\n\t\tfor message := range m.client.Incoming {\n\t\t\tmessages <- message.Payload\n\t\t}\n\t}()\n\treturn messages\n}", "func (l *Logs) Subscribe(uuid string) (\n\tchan []byte, chan []byte, error,\n) {\n\tlogChan, err := l.GetChannel(uuid)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tschan, echan := messaging.CreateSubscriptionChannels()\n\tl.nub.Subscribe(logChan, \"\", schan, false, echan)\n\treturn schan, echan, nil\n\n}", "func (s *server) Subscribe(p *Chat_Consumer, stream ChatService_SubscribeServer) error {\n\tep := s.read_ep(p.Id)\n\tif ep == nil {\n\t\tlog.Errorf(\"cannot find endpoint %v\", p)\n\t\treturn ERROR_NOT_EXISTS\n\t}\n\n\tconsumerid := atomic.AddUint64(&s.consumerid_autoinc, 1)\n\te := make(chan error, 1)\n\n\t// activate consumer\n\tep.mu.Lock()\n\n\t// from newest\n\tif p.From == -1 {\n\t\tp.From = ep.StartOffset + int64(len(ep.Inbox))\n\t}\n\tep.consumers[consumerid] = &Consumer{p.From, func(msg *Chat_Message) {\n\t\tif err := stream.Send(msg); err != nil {\n\t\t\tselect {\n\t\t\tcase e <- err:\n\t\t\tdefault:\n\t\t\t}\n\t\t}\n\t}}\n\tep.mu.Unlock()\n\tdefer func() {\n\t\tep.mu.Lock()\n\t\tdelete(ep.consumers, consumerid)\n\t\tep.mu.Unlock()\n\t}()\n\n\tep.notifyConsumers()\n\n\tselect {\n\tcase <-stream.Context().Done():\n\tcase err := <-e:\n\t\treturn err\n\t}\n\treturn nil\n}", "func (m *MockHub) BroadcastToAllClients(handler model.EventMetadata) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"BroadcastToAllClients\", handler)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func init() {\n\tgo func(ch <-chan string, chClose <-chan struct{}) {\n\t\tfor true {\n\t\t\tselect {\n\t\t\tcase msg := <-ch:\n\t\t\t\tOutMessageToAll(msg)\n\t\t\tcase <-chClose:\n\t\t\t\tclose(models.CancelChOutMessageToAll)\n\t\t\t\tclose(models.ChOutMessageToAll)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}(models.ChOutMessageToAll, models.CancelChOutMessageToAll)\n}", "func EventSubscribeH(w http.ResponseWriter, r *http.Request) {\n\n\tlog.V(logLevel).Debugf(\"%s:subscribe:> subscribe on subscribe\", logPrefix)\n\n\tif r.Method != \"GET\" {\n\t\thttp.Error(w, \"Method not allowed\", http.StatusMethodNotAllowed)\n\t\treturn\n\t}\n\n\tlog.V(logLevel).Debugf(\"%s:subscribe:> watch all events\", logPrefix)\n\n\tvar (\n\t\tsm = distribution.NewServiceModel(r.Context(), envs.Get().GetStorage())\n\t\tnm = distribution.NewNamespaceModel(r.Context(), envs.Get().GetStorage())\n\t\tcm = distribution.NewClusterModel(r.Context(), envs.Get().GetStorage())\n\t\tdone = make(chan bool, 1)\n\t)\n\n\tconn, err := upgrader.Upgrade(w, r, nil)\n\tif err != nil {\n\t\tlog.V(logLevel).Debugf(\"%s:subscribe:> set websocket upgrade err: %s\", logPrefix, err.Error())\n\t\treturn\n\t}\n\n\tticker := time.NewTicker(time.Second)\n\tdefer ticker.Stop()\n\n\tvar serviceEvents = make(chan types.ServiceEvent)\n\tvar namespaceEvents = make(chan types.NamespaceEvent)\n\tvar clusterEvents = make(chan types.ClusterEvent)\n\n\tnotify := w.(http.CloseNotifier).CloseNotify()\n\n\tgo func() {\n\t\t<-notify\n\t\tlog.V(logLevel).Debugf(\"%s:subscribe:> HTTP connection just closed.\", logPrefix)\n\t\tdone <- true\n\t}()\n\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-done:\n\t\t\t\tclose(serviceEvents)\n\t\t\t\tclose(namespaceEvents)\n\t\t\t\tclose(clusterEvents)\n\t\t\t\treturn\n\t\t\tcase e := <-clusterEvents:\n\n\t\t\t\tvar data interface{}\n\t\t\t\tif e.Data == nil {\n\t\t\t\t\tdata = nil\n\t\t\t\t} else {\n\t\t\t\t\tdata = v1.View().Cluster().New(e.Data)\n\t\t\t\t}\n\n\t\t\t\tevent := Event{\n\t\t\t\t\tEntity: \"cluster\",\n\t\t\t\t\tAction: e.Action,\n\t\t\t\t\tName: e.Name,\n\t\t\t\t\tData: data,\n\t\t\t\t}\n\n\t\t\t\tif err = conn.WriteJSON(event); err != nil {\n\t\t\t\t\tlog.Errorf(\"%s:subscribe:> write cluster event to socket error.\", logPrefix)\n\t\t\t\t}\n\t\t\tcase e := <-serviceEvents:\n\n\t\t\t\tvar data interface{}\n\t\t\t\tif e.Data == nil {\n\t\t\t\t\tdata = nil\n\t\t\t\t} else {\n\t\t\t\t\tdata = v1.View().Service().New(e.Data)\n\t\t\t\t}\n\n\t\t\t\tevent := Event{\n\t\t\t\t\tEntity: \"service\",\n\t\t\t\t\tAction: e.Action,\n\t\t\t\t\tName: e.Name,\n\t\t\t\t\tData: data,\n\t\t\t\t}\n\n\t\t\t\tif err = conn.WriteJSON(event); err != nil {\n\t\t\t\t\tlog.Errorf(\"%s:subscribe:> write service event to socket error.\", logPrefix)\n\t\t\t\t}\n\t\t\tcase e := <-namespaceEvents:\n\n\t\t\t\tvar data interface{}\n\t\t\t\tif e.Data == nil {\n\t\t\t\t\tdata = nil\n\t\t\t\t} else {\n\t\t\t\t\tdata = v1.View().Namespace().New(e.Data)\n\t\t\t\t}\n\n\t\t\t\tevent := Event{\n\t\t\t\t\tEntity: \"namespace\",\n\t\t\t\t\tAction: e.Action,\n\t\t\t\t\tName: e.Name,\n\t\t\t\t\tData: data,\n\t\t\t\t}\n\n\t\t\t\tif err = conn.WriteJSON(event); err != nil {\n\t\t\t\t\tlog.Errorf(\"%s:subscribe:> write namespace event to socket error.\", logPrefix)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\tgo cm.Watch(clusterEvents)\n\tgo sm.Watch(serviceEvents, nil)\n\tgo nm.Watch(namespaceEvents)\n\n\tgo func() {\n\t\tfor range ticker.C {\n\t\t\tif err := conn.WriteMessage(websocket.TextMessage, []byte{}); err != nil {\n\t\t\t\tlog.Errorf(\"%s:subscribe:> writing to the client websocket err: %s\", logPrefix, err.Error())\n\t\t\t\tdone <- true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}()\n\n\t<-done\n}", "func (m *Manager) WriteStreamAll(name, mimeType string, body <-chan StreamEvent, dlo bool) (errs error) {\n\n\tvar eg errgroup.Group\n\tstreamConsumer := make(map[string]Storage, len(m.storageServices))\n\tchanConsumer := make(map[string]ChannelWriter, len(m.storageServices))\n\n\tfor k, s := range m.storageServices {\n\t\tif w, ok := s.(ChannelWriter); ok {\n\t\t\tchanConsumer[k] = w\n\t\t} else {\n\t\t\tstreamConsumer[k] = s\n\t\t}\n\t}\n\n\t// connect all io.Reader consumer with a reader\n\treaders, writer, closer := m.createIOReaders(len(streamConsumer))\n\ti := 0\n\tfor _, s := range streamConsumer {\n\t\tfunc(i int, w Storage) {\n\t\t\teg.Go(func() error {\n\t\t\t\treturn w.WriteStream(name, mimeType, readers[i], nil, dlo)\n\t\t\t})\n\t\t}(i, s)\n\t\ti++\n\t}\n\n\t// connect all channel consumer with a channel\n\tchannels := m.createChannels(len(chanConsumer))\n\ti = 0\n\tfor _, s := range chanConsumer {\n\t\tfunc(i int, w ChannelWriter) {\n\t\t\teg.Go(func() error {\n\t\t\t\treturn w.WriteChannel(name, mimeType, channels[i], nil, dlo)\n\t\t\t})\n\t\t}(i, s)\n\t\ti++\n\t}\n\n\tgo func() {\n\t\tfor {\n\t\t\tv, ok := <-body\n\t\t\tif !ok {\n\t\t\t\t// Close all Reader, Writer and channels\n\t\t\t\tif closer != nil {\n\t\t\t\t\tcloser.Close()\n\t\t\t\t}\n\t\t\t\tfor _, c := range channels {\n\t\t\t\t\tclose(c)\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t}\n\t\t\t// write bytes all io.Reader consumer\n\t\t\tif len(streamConsumer) > 0 {\n\t\t\t\twriter.Write(v.ToByte())\n\t\t\t}\n\t\t\t// send the event as is to all channel consumer\n\t\t\tfor _, c := range channels {\n\t\t\t\tc <- v\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn eg.Wait()\n}", "func New() *PubSubChannel {\n return &PubSubChannel{\n subscriptions: make([]*Subscription, 0),\n }\n}", "func (psc *PubSubChannel) Publish(msg interface{}) {\n psc.subsMutex.RLock()\n defer psc.subsMutex.RUnlock()\n for i, sub := range psc.subscriptions {\n if sub.closed {\n fmt.Println(\"Subscription was closed\", \"id\", sub.id)\n sub.closeChannel()\n } else {\n go func(sub *Subscription, index int) {\n sub.signal <- msg\n }(sub, i)\n\n }\n }\n}", "func Subscribe(topic string) (<-chan []byte, error) {\n\treturn Default.Subscribe(topic)\n}", "func BroadcastAll(message socket.RawMessage) {\n\tinstance.mutex.RLock()\n\tinstance.broadcast(uuid.Nil, message)\n\tinstance.mutex.RUnlock()\n}", "func (sn *SimNode) SubscribeEvents(ch chan *p2p.PeerEvent) event.Subscription {\n\tsrv := sn.Server()\n\tif srv == nil {\n\t\tpanic(\"node not running\")\n\t}\n\treturn srv.SubscribeEvents(ch)\n}", "func (q *Qlient) Sub() (chan []byte, error) {\n\treturn q.SubOn(q.config.Topic)\n}", "func (fs *FilterStorage) getSubscribersByTopic(topic Topic) []*Filter {\n\tres := make([]*Filter, 0, len(fs.topicToFilters[topic]))\n\tfor subscriber := range fs.topicToFilters[topic] {\n\t\tres = append(res, subscriber)\n\t}\n\treturn res\n}", "func InitiateConsumers(connectionString string, subscriptionMetaDataList []Subscription, reconnectInterval time.Duration) {\n\tfor _, subscriptionMetaData := range subscriptionMetaDataList {\n\t\tgo func(subscriptionMetaData Subscription) {\n\t\t\tstartBrokerConsumer(connectionString, subscriptionMetaData, reconnectInterval)\n\t\t}(subscriptionMetaData)\n\t}\n}", "func (h *clientHub) Channels() []string {\n\th.RLock()\n\tdefer h.RUnlock()\n\tchannels := make([]string, len(h.subs))\n\ti := 0\n\tfor ch := range h.subs {\n\t\tchannels[i] = ch\n\t\ti++\n\t}\n\treturn channels\n}", "func (pubsub *MemoryEventStore) Subscribe(topics ...string) <-chan Event {\n\tsub := subscription{\n\t\ttopics: topics,\n\t\tch: make(chan Event, pubsub.buffer),\n\t}\n\tpubsub.subscriptions = append(pubsub.subscriptions, sub)\n\treturn sub.ch\n}", "func (k *Kafka) Subscribe(topic string, h broker.Handler, opts ...broker.SubscribeOption) (broker.Subscriber, error) {\n\n\top := &broker.SubscribeOptions{\n\t\tAutoAck: true,\n\t}\n\top.Apply(opts...)\n\n\tvar err error\n\t//handler function\n\thandler := func(msg *sarama.ConsumerMessage) {\n\t\tm := broker.Message{}\n\t\tif err := k.encoder.Decode(msg.Value, &m); err != nil {\n\t\t\tlog.Errorf(\"kafka: subscribe: decode failed, err: %v\", err)\n\t\t\treturn\n\t\t}\n\t\th(&event{\n\t\t\top: op,\n\t\t\tt: topic,\n\t\t\tm: &m,\n\t\t})\n\t}\n\t//Consumer with no groupID\n\tif op.Queue == \"\" {\n\t\tlog.Info(\"consumer with no groupID\")\n\t\t// Create new consumer\n\t\tk.consumer, err = sarama.NewConsumer(strings.Split(k.addrs, \",\"), k.config)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tpartitionList, err := k.consumer.Partitions(topic)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tfor partition := range partitionList {\n\t\t\tconsumer, err := k.consumer.ConsumePartition(topic, int32(partition), sarama.OffsetNewest)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tgo func() {\n\t\t\t\tfor msg := range consumer.Messages() {\n\t\t\t\t\thandler(msg)\n\t\t\t\t}\n\t\t\t}()\n\t\t}\n\t\treturn &subscriber{\n\t\t\tqueue: op.Queue,\n\t\t\tt: topic,\n\t\t\ts: k.consumer,\n\t\t}, nil\n\n\t} //end no group\n\t{\n\t\t// Create new consumer group\n\t\tif k.consumerGroup, err = sarama.NewConsumerGroup([]string{k.addrs}, op.Queue, k.config); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tctx := context.Background()\n\t\tconsumer := Consumer{\n\t\t\tencoder: k.encoder,\n\t\t\ttopic: topic,\n\t\t\th: h,\n\t\t\tready: make(chan bool),\n\t\t}\n\t\tgo func() {\n\t\t\tfor {\n\n\t\t\t\tif err := k.consumerGroup.Consume(ctx, []string{topic}, &consumer); err != nil {\n\t\t\t\t\tlog.Panicf(\"Error from consumer: %v\", err)\n\t\t\t\t}\n\t\t\t\t<-ctx.Done()\n\n\t\t\t\tif ctx.Err() != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\t\t<-consumer.ready // Await till the consumer has been set up\n\t\tlog.Info(\"Sarama consumer up and running!...\")\n\n\t\treturn &subscriberGroup{\n\t\t\tqueue: op.Queue,\n\t\t\tt: topic,\n\t\t\tg: k.consumerGroup,\n\t\t}, nil\n\t}\n\n}" ]
[ "0.6937696", "0.58796257", "0.5855679", "0.5842687", "0.57394993", "0.5554731", "0.55545205", "0.5499015", "0.5495352", "0.54776376", "0.54013604", "0.53748983", "0.5313518", "0.52899235", "0.52726644", "0.52343976", "0.5234129", "0.5231414", "0.52018505", "0.51724696", "0.5146893", "0.5124124", "0.51228034", "0.5119867", "0.5063982", "0.5055968", "0.5052854", "0.50512874", "0.5045787", "0.5044471", "0.50293785", "0.5029157", "0.4974806", "0.49735984", "0.49622694", "0.4960119", "0.4958846", "0.49544632", "0.49536127", "0.49450055", "0.49413043", "0.49371743", "0.49094644", "0.49052256", "0.48925218", "0.4877531", "0.4874422", "0.48737738", "0.48612988", "0.48444572", "0.48435542", "0.4841446", "0.48391432", "0.48314726", "0.48190874", "0.48169947", "0.47982976", "0.47967124", "0.47957534", "0.4790064", "0.47900102", "0.47893804", "0.47882164", "0.47874537", "0.47869125", "0.4785707", "0.47734985", "0.47700518", "0.47616863", "0.4759079", "0.4756172", "0.4752999", "0.474615", "0.4738781", "0.47370353", "0.4728834", "0.47235453", "0.47203928", "0.47122544", "0.47121042", "0.46997878", "0.4697453", "0.4691208", "0.46878347", "0.4682957", "0.46807426", "0.46793288", "0.46781337", "0.4677115", "0.46737078", "0.46686316", "0.46383542", "0.46199664", "0.4619583", "0.46176183", "0.46140867", "0.461281", "0.46100703", "0.4589315", "0.45874584" ]
0.78876483
0
RegisterStrategy allows a different RoutingStrategy to be specified for a given topic
func (n *Node) RegisterStrategy(topic string, strategy RoutingStrategy) { n.strategyMap[topic] = strategy }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func RegisterStrategy(info *StrategyInfo) {\n\tstrategyRegistry = append(strategyRegistry, info)\n}", "func RoutingStrategy(r RoutingStrategyType) metaOp {\n\treturn func(m *PluginMeta) {\n\t\tm.RoutingStrategy = r\n\t}\n}", "func (t Twitter) registerRoutes(r *mux.Router) {\n\n}", "func (n networkRoute) Register(m *mux.Router, handler http.Handler) {\n}", "func NewTopicRouter(br fiber.Router, conn *config.DBConn) {\n\trouter := br.Group(\"/topics\")\n\ttopicRepo := repository.NewTopicRepository(conn)\n\tpostRepo := repository.NewPostRepository(conn)\n\tuserRepo := repository.NewUserRepository(conn)\n\ttopicService := services.NewTopicService(topicRepo, postRepo, userRepo)\n\ttopicController := controllers.NewTopicController(topicService)\n\n\trouter.Get(\"/:id\", topicController.GetOne)\n\trouter.Get(\"/\", topicController.GetAll)\n\trouter.Post(\"/\", topicController.Create)\n}", "func (group *RouterGroup) register(method string, subpattern string, handler HandlerFunc) {\n\tpattern := path.Join(group.prefix + subpattern)\n\tgroup.engine.router.addRoute(method, pattern, handler)\n}", "func (r *LazyRouter) SetStrategy(strategy RoutingStrategy) {\n\tr.strategy = &baseRoutingStrategy{RoutingStrategy: strategy}\n}", "func (s Site) AddTopic(t string) {\n\ts.Handler.HandleFunc(\"/\"+t+\"/\", s.TopicHandler)\n\ts.Handler.HandleFunc(\"/nominate/\"+t+\"/\", s.NominateHandler)\n}", "func RegisterPlugin(key string, strategy ACLStrategy) {\n\tstrategiesMutex.Lock()\n\tdefer strategiesMutex.Unlock()\n\n\tif strategies == nil {\n\t\tstrategies = make(map[string]ACLStrategy)\n\t}\n\n\tstrategies[key] = strategy\n}", "func RegisterTransport(name string, transport TransportFactory) {\n\ttransportFactories[name] = transport\n}", "func (h *Hookbot) AddRouter(r Router) {\n\tfor _, topic := range r.Topics() {\n\t\th.wg.Add(1)\n\t\tgo func() {\n\t\t\tdefer h.wg.Done()\n\n\t\t\tl := h.Add(topic)\n\t\t\tfor m := range l.c {\n\t\t\t\tr.Route(m, h.Publish)\n\t\t\t}\n\t\t}()\n\t}\n}", "func (h *BasicHost) RegisterProtocol(\n\tpid common.Pid,\n\thandler ProtocolHandler,\n\tadapters ...ProtocolAdapter,\n) {\n\th.host.SetStreamHandler(pid.ProtocolID(), func(stream net.Stream) {\n\t\tdefer stream.Reset()\n\t\tmsg, err := common.ReadMessage(stream)\n\t\tif err != nil {\n\t\t\tlog.Println(\"failed to read message from stream :\", err)\n\t\t\treturn\n\t\t}\n\t\tgo handler.Handle(adapters...)(msg)\n\t})\n}", "func addRoutes(p *nats.Conn) {\n\tr := mux.NewRouter()\n\tr.Methods(\"POST\").Path(\"/topics/{topic}\").Handler(\n\t\thandlers.LoggingHandler(os.Stdout, handler(p, topic)))\n\tr.Methods(\"POST\").Path(\"/requests/{topic}\").Handler(\n\t\thandlers.LoggingHandler(os.Stdout, handler(p, request)))\n\thttp.Handle(\"/\", r)\n}", "func registerRoutes(jwtMiddleware *jwtmiddleware.JWTMiddleware) *mux.Router {\n\tr := mux.NewRouter()\n\n\tr.Handle(\"/healthcheck\", http.HandlerFunc(healthCheck)).Methods(\"GET\")\n\n\tr.Handle(\"/message\", http.HandlerFunc(message)).Methods(\"POST\")\n\tr.Handle(\"/message/{id}\", http.HandlerFunc(messageDelete)).Methods(\"DELETE\")\n\tr.Handle(\"/publish\", http.HandlerFunc(publish)).Methods(\"POST\")\n\n\tmsgRouter := mux.NewRouter().PathPrefix(\"/message\").Subrouter()\n\tpubRouter := mux.NewRouter().PathPrefix(\"/publish\").Subrouter()\n\n\tr.PathPrefix(\"/message\").Handler(negroni.New(\n\t\tnegroni.HandlerFunc(jwtMiddleware.HandlerWithNext),\n\t\tnegroni.Wrap(msgRouter),\n\t))\n\n\tr.PathPrefix(\"/publish\").Handler(negroni.New(\n\t\tnegroni.HandlerFunc(jwtMiddleware.HandlerWithNext),\n\t\tnegroni.Wrap(pubRouter),\n\t))\n\n\t// GET - handles upgrading http/https connections to ws/wss.\n\t// the JWT middleware is expecting an access_token\n\t// query parameter within the request\n\tr.Handle(\"/ws\", negroni.New(\n\t\tnegroni.HandlerFunc(jwtMiddleware.HandlerWithNext),\n\t\tnegroni.HandlerFunc(AddUserID),\n\t\tnegroni.Wrap(broker),\n\t))\n\n\treturn r\n}", "func RegisterTopologyAPI(r *shttp.Server, g *graph.Graph, parser *traversal.GremlinTraversalParser, authBackend shttp.AuthenticationBackend, extraMarshallers map[string]TopologyMarshaller) {\n\tt := &TopologyAPI{\n\t\tgremlinParser: parser,\n\t\tgraph: g,\n\t\textraMarshallers: extraMarshallers,\n\t}\n\n\tt.registerEndpoints(r, authBackend)\n}", "func RegisterTransport(scheme string, transport Transport) Transport {\n\treturn DefaultTransport.(*defaultTransport).registerTransport(scheme, transport)\n}", "func (api *api) RegisterRouting(g *echo.Group) {\n\n\tgrp := g.Group(\"/v2/products\")\n\tgrp.GET(\"\", api.Service.List)\n\tgrp.GET(\"/:id\", api.Service.Get)\n\tgrp.POST(\"\", api.Service.Create)\n\tgrp.PUT(\"/:id\", api.Service.Update)\n\tgrp.DELETE(\"/:id\", api.Service.Delete)\n\n}", "func (a *API) RegisterDistributor(d *distributor.Distributor, pushConfig distributor.Config) {\n\ta.RegisterRoute(\"/api/v1/push\", push.Handler(pushConfig, d.Push), true)\n\ta.RegisterRoute(\"/distributor/all_user_stats\", http.HandlerFunc(d.AllUserStatsHandler), false)\n\ta.RegisterRoute(\"/distributor/ha_tracker\", d.HATracker, false)\n\n\t// Legacy Routes\n\ta.RegisterRoute(a.cfg.LegacyHTTPPrefix+\"/push\", push.Handler(pushConfig, d.Push), true)\n\ta.RegisterRoute(\"/all_user_stats\", http.HandlerFunc(d.AllUserStatsHandler), false)\n\ta.RegisterRoute(\"/ha-tracker\", d.HATracker, false)\n}", "func registerRoutes(router *httprouter.Router, ctrl *webhook.Controller) {\n\trouter.POST(\"/auto-deploy/docker-hub\", ctrl.AutoDeployDockerHub)\n}", "func NewStrategy(typer runtime.ObjectTyper) clusterStrategy {\n\treturn clusterStrategy{typer, names.SimpleNameGenerator}\n}", "func init() {\n\t_ = router.Register(\"httprouter\", New)\n}", "func RegisterProcessor(topic string, mp MessageProcessor) {\n\tClient.topicProcessors[topic] = mp\n}", "func registerEndpoint(pattern string, methods []string, fn unboundEndpoint) {\n\tif endpoints == nil {\n\t\tendpoints = make(map[string]unboundEndpoint)\n\t}\n\tif endpoints[pattern] != nil || allowedMethods[pattern] != nil {\n\t\tpanic(fmt.Errorf(\"Pattern %q is already registered\", pattern))\n\t}\n\n\tendpoints[pattern] = fn\n\tallowedMethods[pattern] = methods\n}", "func (w Ws) Register(r *gin.RouterGroup) {\n\tst := r.Group(\"\")\n\tst.GET(\"/ws/:id\", w.Server)\n\tst.DELETE(\"/ws/:id\", w.Offline)\n\tst.PUT(\"/ws/:id\", w.Dispatch)\n}", "func (a *AApi) registerRoute(f func(http.ResponseWriter, *http.Request), path string, methods ...string) {\n\ta.logger.WithField(\"func\", \"registerRoute\").\n\t\tDebugf(\"Initializing route %s with methods: %v\", path, methods)\n\ta.router.HandleFunc(path, f).Name(path).Methods(methods...) // Name if set for ability to exclude route from authz\n}", "func registerRoutes() {\n\tuserRoutes()\n\troleRoutes()\n}", "func (route *baseRoute) addDestination(dest *Destination, extendConfig baseConfigExtender) {\n\troute.Lock()\n\tdefer route.Unlock()\n\tconf := route.config.Load().(RouteConfig)\n\tdest.Run()\n\tnewDests := append(conf.Dests(), dest)\n\tnewConf := extendConfig(baseRouteConfig{*conf.Matcher(), newDests})\n\troute.config.Store(newConf)\n}", "func Register(router *mux.Router) {\n\ttodoRoutes(router)\n\tuserRoutes(router)\n}", "func (_Flytrap *FlytrapSession) AddTopic(topic string, country [2]byte, addPubCost *big.Int, addSubCost *big.Int, reason string, sensitive bool) (*types.Transaction, error) {\n\treturn _Flytrap.Contract.AddTopic(&_Flytrap.TransactOpts, topic, country, addPubCost, addSubCost, reason, sensitive)\n}", "func (k *Keeper) RegisterRoute(moduleName, route string, invar sdk.Invariant) {\n\tinvarRoute := NewInvarRoute(moduleName, route, invar)\n\tk.routes = append(k.routes, invarRoute)\n}", "func (m *MessageProcessor) Register(topic topics.Topic, fn ProcessorFunc) {\n\tm.processors[topic] = fn\n}", "func (_Flytrap *FlytrapTransactorSession) AddTopic(topic string, country [2]byte, addPubCost *big.Int, addSubCost *big.Int, reason string, sensitive bool) (*types.Transaction, error) {\n\treturn _Flytrap.Contract.AddTopic(&_Flytrap.TransactOpts, topic, country, addPubCost, addSubCost, reason, sensitive)\n}", "func (s *Switch) RegisterGossipProtocol(protocol string, prio priorityq.Priority) chan service.GossipMessage {\n\tif s.started == 1 {\n\t\tlog.Panic(\"attempt to register gossip protocol after p2p has started\")\n\t}\n\tmchan := make(chan service.GossipMessage, s.config.BufferSize)\n\ts.gossip.SetPriority(protocol, prio)\n\ts.gossipProtocolHandlers[protocol] = mchan\n\treturn mchan\n}", "func NewFeedRewardStrategyRegister(feedRewardStrategyFns map[string]FeedRewardStrategyFunc) {\n\tif feedRewardStrategyFns == nil {\n\t\treturn\n\t}\n\n\tfor name := range feedRewardStrategyFns {\n\t\tif name == \"\" {\n\t\t\tpanic(\"feed reward strategy name can not be empty\")\n\t\t}\n\t}\n\n\tFeedRewardStrategyConvertor = feedRewardStrategyFns\n}", "func RegisterCompactor(topic string, compactor interface{}) {\n\tpanic(\"not implemented\")\n}", "func (n networkRoute) Register(m *mux.Router, handler http.Handler) {\n\tlogrus.Debugf(\"Registering %s, %v\", n.path, httpMethods)\n\tsubrouter := m.PathPrefix(router.VersionMatcher + n.path).Subrouter()\n\tsubrouter.Methods(httpMethods...).Handler(handler)\n\n\tsubrouter = m.PathPrefix(n.path).Subrouter()\n\tsubrouter.Methods(httpMethods...).Handler(handler)\n}", "func switchRouter(defaultHandler http.Handler, proxySrv *pServer.HttpServer) func(config dynamic.Configuration) {\n\treturn func(config dynamic.Configuration) {\n\t\tlog.Info(\"===Starting SwitchRouter====\")\n\t\trouterTemp, err := router.NewRouter()\n\t\tif err != nil {\n\t\t\tlog.Info(\"Failed to create router \", err)\n\t\t\t// return nil, err\n\t\t}\n\t\tlog.Infof(\"buildHandler : %v \\n\", config.Routers)\n\t\tfor name, value := range config.Routers {\n\t\t\tlog.Infof(\"Create Hypercloud proxy based on %v: %v \\n\", name, value)\n\t\t\tbackURL, err := url.Parse(value.Server)\n\t\t\tif err != nil {\n\t\t\t\tlog.Error(errors.Wrapf(err, \"URL Parsing failed for: %s\", value.Server))\n\t\t\t}\n\t\t\tdhconfig := &proxy.Config{\n\t\t\t\tTLSClientConfig: &tls.Config{\n\t\t\t\t\tInsecureSkipVerify: true,\n\t\t\t\t\tCipherSuites: crypto.DefaultCiphers(),\n\t\t\t\t},\n\t\t\t\tHeaderBlacklist: []string{\"X-CSRFToken\"},\n\t\t\t\tEndpoint: backURL,\n\t\t\t}\n\t\t\tdhproxy := proxy.NewProxy(dhconfig)\n\t\t\terr = routerTemp.AddRoute(value.Rule, 0, http.StripPrefix(value.Path, http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\t\ttoken := r.Header.Clone().Get(\"Authorization\")\n\t\t\t\ttemp := strings.Split(token, \"Bearer \")\n\t\t\t\tif len(temp) > 1 {\n\t\t\t\t\ttoken = temp[1]\n\t\t\t\t} else {\n\t\t\t\t\ttoken = temp[0]\n\t\t\t\t}\n\t\t\t\t// NOTE: query에 token 정보가 있을 시 해당 token으로 설정\n\t\t\t\tqueryToken := r.URL.Query().Get(\"token\")\n\t\t\t\tif queryToken != \"\" && token == \"\" {\n\t\t\t\t\tr.URL.Query().Del(\"token\")\n\t\t\t\t\ttoken = queryToken\n\t\t\t\t}\n\t\t\t\tr.Header.Set(\"Authorization\", fmt.Sprintf(\"Bearer %s\", token))\n\t\t\t\tdhproxy.ServeHTTP(w, r)\n\t\t\t})))\n\t\t\tif err != nil {\n\t\t\t\tlog.Error(\"failed to put proxy handler into Router\", err)\n\t\t\t}\n\t\t}\n\t\terr = routerTemp.AddRoute(\"PathPrefix(`/api/console/dynamic`)\", 0, http.HandlerFunc(\n\t\t\tfunc(rw http.ResponseWriter, r *http.Request) {\n\t\t\t\trw.Header().Set(\"Content-Type\", \"application/json\")\n\t\t\t\terr := json.NewEncoder(rw).Encode(config)\n\t\t\t\tif err != nil {\n\t\t\t\t\thttp.NotFound(rw, r)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t},\n\t\t))\n\t\tif err != nil {\n\t\t\tlog.Error(\"/api/k8sAll/ has a problem\", err)\n\t\t}\n\n\t\terr = routerTemp.AddRoute(\"PathPrefix(`/`)\", 0, defaultHandler)\n\t\tif err != nil {\n\t\t\tlog.Error(\"failed to put hypercloud proxy\", err)\n\t\t\t// return nil, err\n\t\t}\n\n\t\tlog.Info(\"===End SwitchRouter ===\")\n\t\tlog.Info(\"Call updateHandler --> routerTemp.Router\")\n\t\t// olderSrv:=proxySrv.Handler.Switcher.GetHandler()\n\n\t\tif proxySrv.Switcher.GetHandler() == nil {\n\t\t\tproxySrv.Switcher.UpdateHandler(http.NotFoundHandler())\n\t\t}\n\n\t\tproxySrv.Switcher.UpdateHandler(routerTemp)\n\n\t}\n}", "func RegisterProtocol(messageProtocolID uint16, p Protocol) {\n\tprotocolRegistry[messageProtocolID] = p\n}", "func RegisterSink(scheme string, factory func(*url.URL) (Sink, error)) error {\n\t_sinkMutex.Lock()\n\tdefer _sinkMutex.Unlock()\n\n\tif scheme == \"\" {\n\t\treturn errors.New(\"can't register a sink factory for empty string\")\n\t}\n\tnormalized, err := normalizeScheme(scheme)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"%q is not a valid scheme: %v\", scheme, err)\n\t}\n\tif _, ok := _sinkFactories[normalized]; ok {\n\t\treturn fmt.Errorf(\"sink factory already registered for scheme %q\", normalized)\n\t}\n\t_sinkFactories[normalized] = factory\n\treturn nil\n}", "func RegisterTransport(typ string, ctor TransportConstructor) {\n\ttransportTypes[typ] = ctor\n}", "func (m *Mesh) Subscribe(tp Topic, fn func(interface{})) {\n\tswitch tp {\n\tcase TOPIC_NODE:\n\t\tm.topic[tp] = append(m.topic[tp], fn)\n\tcase TOPIC_DATA:\n\tdefault:\n\t\tlog.WithField(\"topic\", tp).Error(\"invalid topic\")\n\t}\n}", "func RegisterPeerProvider(scheme string, pp PeerProvider) {\n\tregistry[scheme] = pp\n}", "func (p *JSONProtocol) Register(msg interface{}) {\n\tt := reflect.TypeOf(msg)\n\tif t.Kind() == reflect.Ptr {\n\t\tt = t.Elem()\n\t}\n\tname := t.PkgPath() + \"/\" + t.Name()\n\tp.types[name] = t\n\tp.names[t] = name\n}", "func RegisterTopics(topicNames ...string) {\n\tfor _, n := range topicNames {\n\t\tregisterTopic(n)\n\t}\n}", "func Register(scheme string, b Broker) {\n\tbrokerRegistery[scheme] = b\n}", "func (r *Router) AddWith(name, pattern, providerName, from string, isActive bool) error {\n\troute := r.Get(name)\n\tif route != nil {\n\t\treturn errors.New(\"route already exists\")\n\t}\n\tprovider := r.GetProvider(providerName)\n\tif provider == nil {\n\t\treturn errors.New(\"provider not found\")\n\t}\n\treturn r.Add(model.NewRoute(name, pattern, provider, isActive).SetFrom(from))\n}", "func InjectRoutingService(\n\truntime env.Runtime,\n\tprefix provider.LogPrefix,\n\tlogLevel logger.LogLevel,\n\tsqlDB *sql.DB,\n\tgithubClientID provider.GithubClientID,\n\tgithubClientSecret provider.GithubClientSecret,\n\tfacebookClientID provider.FacebookClientID,\n\tfacebookClientSecret provider.FacebookClientSecret,\n\tfacebookRedirectURI provider.FacebookRedirectURI,\n\tgoogleClientID provider.GoogleClientID,\n\tgoogleClientSecret provider.GoogleClientSecret,\n\tgoogleRedirectURI provider.GoogleRedirectURI,\n\tjwtSecret provider.JwtSecret,\n\tbufferSize provider.KeyGenBufferSize,\n\tkgsRPCConfig provider.KgsRPCConfig,\n\twebFrontendURL provider.WebFrontendURL,\n\ttokenValidDuration provider.TokenValidDuration,\n\tdataDogAPIKey provider.DataDogAPIKey,\n\tsegmentAPIKey provider.SegmentAPIKey,\n\tipStackAPIKey provider.IPStackAPIKey,\n) (service.Routing, error) {\n\twire.Build(\n\t\twire.Bind(new(timer.Timer), new(timer.System)),\n\t\twire.Bind(new(geo.Geo), new(geo.IPStack)),\n\n\t\twire.Bind(new(url.Retriever), new(url.RetrieverPersist)),\n\t\twire.Bind(new(repository.UserURLRelation), new(sqldb.UserURLRelationSQL)),\n\t\twire.Bind(new(repository.User), new(*sqldb.UserSQL)),\n\t\twire.Bind(new(repository.URL), new(*sqldb.URLSql)),\n\n\t\tobservabilitySet,\n\t\tauthSet,\n\t\tgithubAPISet,\n\t\tfacebookAPISet,\n\t\tgoogleAPISet,\n\t\tkeyGenSet,\n\t\tfeatureDecisionSet,\n\n\t\tservice.NewRouting,\n\t\twebreq.NewHTTPClient,\n\t\twebreq.NewHTTP,\n\t\tgraphql.NewClientFactory,\n\t\ttimer.NewSystem,\n\t\tprovider.NewIPStack,\n\t\tenv.NewDeployment,\n\n\t\tsqldb.NewUserSQL,\n\t\tsqldb.NewURLSql,\n\t\tsqldb.NewUserURLRelationSQL,\n\t\turl.NewRetrieverPersist,\n\t\taccount.NewProvider,\n\t\tprovider.NewShortRoutes,\n\t)\n\treturn service.Routing{}, nil\n}", "func Register(p Protocol, n NewFunc) {\n\treglock.Lock()\n\tdefer reglock.Unlock()\n\tregistry[p] = n\n}", "func (_Flytrap *FlytrapTransactor) AddTopic(opts *bind.TransactOpts, topic string, country [2]byte, addPubCost *big.Int, addSubCost *big.Int, reason string, sensitive bool) (*types.Transaction, error) {\n\treturn _Flytrap.contract.Transact(opts, \"addTopic\", topic, country, addPubCost, addSubCost, reason, sensitive)\n}", "func (p *Publisher) Register(topicURL, callbackURL, secret string, leaseEnd time.Time) error {\n\tlease := leaseEnd.Sub(time.Now())\n\tif lease <= 0 {\n\t\treturn nil\n\t}\n\n\ts, err := p.subscribeIfNotExist(topicURL)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ts.locker.Lock()\n\ts.callbacks[callbackURL] = &pubCallback{\n\t\tsecret: secret,\n\t\ttimer: time.AfterFunc(lease, func() {\n\t\t\tp.unregister(topicURL, callbackURL)\n\t\t}),\n\t}\n\ts.locker.Unlock()\n\n\tif p.SubscriptionState != nil {\n\t\tp.SubscriptionState(topicURL, callbackURL, secret, leaseEnd)\n\t}\n\n\treturn nil\n}", "func (conn *Conn) ListenWithAuth(token string, topics ...string) error {\n\tif conn.GetNumTopics()+len(topics) > conn.length {\n\t\treturn ErrShardTooManyTopics\n\t}\n\tif err := conn.WriteMessageWithNonce(Listen, conn.generator(), TopicData{topics, token}); err != nil {\n\t\treturn err\n\t}\n\tconn.listeners.Lock()\n\tdefer conn.listeners.Unlock()\n\tif conn.topics == nil {\n\t\tconn.topics = make(map[string][]string)\n\t}\n\tconn.topics[token] = append(conn.topics[token], topics...)\n\treturn nil\n}", "func (d *distEventBus) setupTopicSub(topicName, subscriberName string) error {\n\tctx := context.TODO()\n\n\t// Create the topic if it doesn't exist yet.\n\td.topic = d.client.Topic(topicName)\n\tif exists, err := d.topic.Exists(ctx); err != nil {\n\t\treturn skerr.Fmt(\"Error checking whether topic exits: %s\", err)\n\t} else if !exists {\n\t\tif d.topic, err = d.client.CreateTopic(ctx, topicName); err != nil {\n\t\t\treturn skerr.Fmt(\"Error creating pubsub topic '%s': %s\", topicName, err)\n\t\t}\n\t}\n\n\t// Create the subscription if it doesn't exist.\n\tsubName := fmt.Sprintf(\"%s+%s\", subscriberName, topicName)\n\td.sub = d.client.Subscription(subName)\n\tif exists, err := d.sub.Exists(ctx); err != nil {\n\t\treturn skerr.Fmt(\"Error checking existence of pubsub subscription '%s': %s\", subName, err)\n\t} else if !exists {\n\t\td.sub, err = d.client.CreateSubscription(ctx, subName, pubsub.SubscriptionConfig{\n\t\t\tTopic: d.topic,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn skerr.Fmt(\"Error creating pubsub subscription '%s': %s\", subName, err)\n\t\t}\n\t}\n\td.sub.ReceiveSettings.MaxOutstandingMessages = MaximumConcurrentPublishesPerTopic\n\t// Make the subscription also the id of this client.\n\td.clientID = subName\n\treturn nil\n}", "func (s *Switch) RegisterDirectProtocol(protocol string) chan service.DirectMessage { // TODO: not used - remove\n\tif s.started == 1 {\n\t\tlog.Panic(\"attempt to register direct protocol after p2p has started\")\n\t}\n\tmchan := make(chan service.DirectMessage, s.config.BufferSize)\n\ts.directProtocolHandlers[protocol] = mchan\n\treturn mchan\n}", "func New(t Topic) (Interface, error) {\n\t// TODO(tmrts): use a LFU and if cache-key space consumption is not a\n\t// problem, migrate to optimized cache replacement algorithms that use\n\t// extra queues\n\t// TODO(tmrts): utilize the eviction callback in LRU\n\tlruCache := lru.NewLRU(t.Capacity, nil)\n\n\tr, err := router.New(RouterPort)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := r.Join(t.Peers); err != nil {\n\t\treturn nil, err\n\t}\n\n\tc := &topic{\n\t\tlru: lruCache,\n\t\tprovider: t.Provider,\n\t\trouter: r,\n\t}\n\n\t// FIXME(tmrts): needs restructuring\n\tsvc, err := service.NewServer(ServicePort, c)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tc.svc = svc\n\n\treturn c, nil\n}", "func RegisterGroupAPIRoute(basePath string, ginEngine *gin.Engine, controllers []IBaseController) {\n\tif !strings.HasPrefix(basePath, \"/\") {\n\t\tbasePath = \"/\" + basePath\n\t}\n\tg := ginEngine.Group(basePath)\n\t{\n\t\troutesControllerMapping(g, controllers)\n\t}\n}", "func (r *HijackRouter) Add(pattern string, resourceType proto.NetworkResourceType, handler func(*Hijack)) error {\n\tr.enable.Patterns = append(r.enable.Patterns, &proto.FetchRequestPattern{\n\t\tURLPattern: pattern,\n\t\tResourceType: resourceType,\n\t})\n\n\treg := regexp.MustCompile(proto.PatternToReg(pattern))\n\n\tr.handlers = append(r.handlers, &hijackHandler{\n\t\tpattern: pattern,\n\t\tregexp: reg,\n\t\thandler: handler,\n\t})\n\n\treturn r.enable.Call(r.caller)\n}", "func Register(ch chan ExecutionEvent, topics ...Topic) {\n\tfor _, t := range topics {\n\t\tsubscriberRegistry[t] = append(subscriberRegistry[t], ch)\n\t}\n}", "func RegisterProvider(name string, provider Provider) {\n\tregistry[name] = provider\n\tpriority = append(priority, name)\n}", "func (cntrolr *Controller) BookmarkRouteRegister(r *mux.Router) {\n\ts := r.PathPrefix(\"/user/{userid}\").Subrouter()\n\ts.Use(cntrolr.auth.AuthUser)\n\tcntrolr.CategoryRouteRegister(s)\n\ts.HandleFunc(\"/bookmark/{pagesize}/{pagenumber}\", cntrolr.GetAllBookmark).Methods(\"GET\")\n\ts.HandleFunc(\"/bookmark/{bookmarkid}\", cntrolr.GetBookmarkByID).Methods(\"GET\")\n\ts.HandleFunc(\"/bookmark/category/{categoryid}/{pagesize}/{pagenumber}\", cntrolr.GetBookmarkByCategory).Methods(\"GET\")\n\ts.HandleFunc(\"/bookmark\", cntrolr.AddBookmark).Methods(\"POST\")\n\ts.HandleFunc(\"/bookmark/{bookmarkid}\", cntrolr.UpdateBookmark).Methods(\"PUT\")\n\ts.HandleFunc(\"/bookmark/{bookmarkid}\", cntrolr.DeleteBookmark).Methods(\"DELETE\")\n}", "func (p *endPoint) RegisterRoute(route *gin.RouterGroup) (err error) {\r\n\troute.GET(p.tasks, list)\r\n\troute.POST(p.tasks, create)\r\n\troute.GET(p.tasks+\"/:\"+p.taskid, get)\r\n\troute.PUT(p.tasks+\"/:\"+p.taskid, update)\r\n\troute.DELETE(p.tasks+\"/:\"+p.taskid, delete)\r\n\troute.PUT(p.taskorders, updateTaskOrders)\r\n\treturn\r\n}", "func (factory *Factory) MeshRouter(provider string) Interface {\n\tswitch {\n\tcase provider == \"none\":\n\t\treturn &NopRouter{}\n\tcase provider == \"kubernetes\":\n\t\treturn &NopRouter{}\n\tcase provider == \"nginx\":\n\t\treturn &IngressRouter{\n\t\t\tlogger: factory.logger,\n\t\t\tkubeClient: factory.kubeClient,\n\t\t\tannotationsPrefix: factory.ingressAnnotationsPrefix,\n\t\t}\n\tcase provider == \"appmesh\":\n\t\treturn &AppMeshRouter{\n\t\t\tlogger: factory.logger,\n\t\t\tflaggerClient: factory.flaggerClient,\n\t\t\tkubeClient: factory.kubeClient,\n\t\t\tappmeshClient: factory.meshClient,\n\t\t}\n\tcase strings.HasPrefix(provider, \"smi:\"):\n\t\tmesh := strings.TrimPrefix(provider, \"smi:\")\n\t\treturn &SmiRouter{\n\t\t\tlogger: factory.logger,\n\t\t\tflaggerClient: factory.flaggerClient,\n\t\t\tkubeClient: factory.kubeClient,\n\t\t\tsmiClient: factory.meshClient,\n\t\t\ttargetMesh: mesh,\n\t\t}\n\tcase provider == \"linkerd\":\n\t\treturn &SmiRouter{\n\t\t\tlogger: factory.logger,\n\t\t\tflaggerClient: factory.flaggerClient,\n\t\t\tkubeClient: factory.kubeClient,\n\t\t\tsmiClient: factory.meshClient,\n\t\t\ttargetMesh: \"linkerd\",\n\t\t}\n\tcase provider == \"contour\":\n\t\treturn &ContourRouter{\n\t\t\tlogger: factory.logger,\n\t\t\tflaggerClient: factory.flaggerClient,\n\t\t\tkubeClient: factory.kubeClient,\n\t\t\tcontourClient: factory.meshClient,\n\t\t}\n\tcase strings.HasPrefix(provider, \"gloo\"):\n\t\tupstreamDiscoveryNs := \"gloo-system\"\n\t\tif strings.HasPrefix(provider, \"gloo:\") {\n\t\t\tupstreamDiscoveryNs = strings.TrimPrefix(provider, \"gloo:\")\n\t\t}\n\t\treturn &GlooRouter{\n\t\t\tlogger: factory.logger,\n\t\t\tflaggerClient: factory.flaggerClient,\n\t\t\tkubeClient: factory.kubeClient,\n\t\t\tglooClient: factory.meshClient,\n\t\t\tupstreamDiscoveryNs: upstreamDiscoveryNs,\n\t\t}\n\tcase strings.HasPrefix(provider, \"supergloo:appmesh\"):\n\t\treturn &AppMeshRouter{\n\t\t\tlogger: factory.logger,\n\t\t\tflaggerClient: factory.flaggerClient,\n\t\t\tkubeClient: factory.kubeClient,\n\t\t\tappmeshClient: factory.meshClient,\n\t\t}\n\tcase strings.HasPrefix(provider, \"supergloo:istio\"):\n\t\treturn &IstioRouter{\n\t\t\tlogger: factory.logger,\n\t\t\tflaggerClient: factory.flaggerClient,\n\t\t\tkubeClient: factory.kubeClient,\n\t\t\tistioClient: factory.meshClient,\n\t\t}\n\tcase strings.HasPrefix(provider, \"supergloo:linkerd\"):\n\t\treturn &SmiRouter{\n\t\t\tlogger: factory.logger,\n\t\t\tflaggerClient: factory.flaggerClient,\n\t\t\tkubeClient: factory.kubeClient,\n\t\t\tsmiClient: factory.meshClient,\n\t\t\ttargetMesh: \"linkerd\",\n\t\t}\n\tdefault:\n\t\treturn &IstioRouter{\n\t\t\tlogger: factory.logger,\n\t\t\tflaggerClient: factory.flaggerClient,\n\t\t\tkubeClient: factory.kubeClient,\n\t\t\tistioClient: factory.meshClient,\n\t\t}\n\t}\n}", "func Register(clusters *clusters.Clusters, plugins *plugin.Plugins, config Config) chi.Router {\n\tplugins.Append(plugin.Plugin{\n\t\tName: \"dashboards\",\n\t\tDisplayName: \"Dashboards\",\n\t\tDescription: \"Create dashboards for your Teams and Applications.\",\n\t\tType: \"dashboards\",\n\t})\n\n\trouter := Router{\n\t\tchi.NewRouter(),\n\t\tclusters,\n\t\tconfig,\n\t}\n\n\trouter.Get(\"/dashboards\", router.getAllDashboards)\n\trouter.Post(\"/dashboards\", router.getDashboards)\n\trouter.Post(\"/dashboard\", router.getDashboard)\n\n\treturn router\n}", "func (follower *Follower) Register(leaderHost string) (err error) {\n\tlog.Printf(\"Registring with Leader...\\n\")\n\tconn, err := net.Dial(\"tcp\", leaderHost)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tmessage := &Message{\n\t\tAction: Message_REGISTER.Enum(),\n\n\t\tId: follower.id[:],\n\t\tHost: proto.String(follower.host),\n\t}\n\n\tdata, err := proto.Marshal(message)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := conn.Write(data); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func RegisterRoutes(r weave.Registry, auth x.Authenticator) {\n\tblogs := NewBlogBucket()\n\tr.Handle(PathCreateBlogMsg, CreateBlogMsgHandler{auth, blogs})\n\tr.Handle(PathCreatePostMsg, CreatePostMsgHandler{auth, NewPostBucket(), blogs})\n\tr.Handle(PathRenameBlogMsg, RenameBlogMsgHandler{auth, blogs})\n\tr.Handle(PathChangeBlogAuthorsMsg, ChangeBlogAuthorsMsgHandler{auth, blogs})\n\tr.Handle(PathSetProfileMsg, SetProfileMsgHandler{auth, NewProfileBucket()})\n}", "func (srv *Server) RegisterPut(ep *Endpoint) (err error) {\n\tif ep == nil || ep.Call == nil {\n\t\treturn nil\n\t}\n\n\tep.Method = RequestMethodPut\n\tep.ResponseType = ResponseTypeNone\n\n\t// Check if the same route already registered.\n\tfor _, rute := range srv.routePuts {\n\t\t_, ok := rute.parse(ep.Path)\n\t\tif ok {\n\t\t\treturn ErrEndpointAmbiguous\n\t\t}\n\t}\n\n\trute, err := newRoute(ep)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsrv.routePuts = append(srv.routePuts, rute)\n\n\treturn nil\n}", "func (mux *URLMux) RegisterTransport(scheme string, opener TransportURLOpener) {\n\tmux.schemes.Register(\"mailer\", \"Transport\", scheme, opener)\n}", "func Register(kind string, props actor.Props) {\n\tnameLookup[kind] = props\n}", "func K8sRegister(auth *restful.WebService) {\n\tPub, Priv = CreateKeyPair()\n\tauth.Path(\"/\").\n\t\tConsumes(restful.MIME_JSON).\n\t\tProduces(restful.MIME_JSON) // you can specify this per route as well\n\t//apiserver http\n\tauth.Route(auth.GET(\"/.well-known/openid-configuration\").To(discoveryHandler))\n\tauth.Route(auth.GET(\"/keys\").To(handlePublicKeys))\n\t//user token http\n\tauth.Route(auth.GET(\"/token\").To(handlerToken))\n\n}", "func (h *HubBub) LookupStrategy(id string) (Strategy, error) {\n\tfor _, s := range h.strategies {\n\t\tif s.ID == id {\n\t\t\treturn s, nil\n\t\t}\n\t}\n\treturn Strategy{}, fmt.Errorf(\"%q not found\", id)\n}", "func RegisterEndpoints(router *fasthttprouter.Router) {\n\thandle(router, \"POST\", \"/submit/question\", SubmitQuestion)\n\thandle(router, \"POST\", \"/submit/answer\", SubmitAnswer)\n\thandleWs(router, wsPath)\n\thandleFS(router, \"/\", \"public\")\n\thandleFS(router, \"/images/*filepath\", \"public/images\")\n\thandleFS(router, \"/js/*filepath\", \"public/js\")\n\thandleFS(router, \"/vendor/*filepath\", \"public/vendor\")\n\thandleFS(router, \"/css/*filepath\", \"public/css\")\n\thandleNotFound(router)\n}", "func RegisterRoutes(r weave.Registry, auth x.Authenticator) {\n\tbucket := NewVoteBucket()\n\ttallies := NewTallyBucket()\n\tr.Handle(pathRecordVote, &recordVoteHandler{auth: auth, bucket: bucket, tallies: tallies})\n}", "func Register(clusters *clusters.Clusters, plugins *plugin.Plugins, config Config) chi.Router {\n\tvar options map[string]interface{}\n\toptions = make(map[string]interface{})\n\toptions[\"webSocketAddress\"] = config.WebSocket.Address\n\toptions[\"ephemeralContainers\"] = config.EphemeralContainers\n\n\tplugins.Append(plugin.Plugin{\n\t\tName: \"resources\",\n\t\tDisplayName: \"Resources\",\n\t\tDescription: \"View and edit Kubernetes resources.\",\n\t\tType: \"resources\",\n\t\tOptions: options,\n\t})\n\n\trouter := Router{\n\t\tchi.NewRouter(),\n\t\tclusters,\n\t\tconfig,\n\t}\n\n\trouter.Get(\"/resources\", router.getResources)\n\trouter.Delete(\"/resources\", router.deleteResource)\n\trouter.Put(\"/resources\", router.patchResource)\n\trouter.Post(\"/resources\", router.createResource)\n\trouter.Get(\"/logs\", router.getLogs)\n\trouter.HandleFunc(\"/terminal\", router.getTerminal)\n\n\treturn router\n}", "func ExposeStrategyFromString(s string) (ExposeStrategy, bool) {\n\tes := ExposeStrategy(s)\n\treturn es, AllExposeStrategies.Has(es)\n}", "func (c *Config) initTopicMap() {\n\ttopicMap[MapperTopicPrefix+c.Device.ID+WatcherTopicSuffix] = c.handleWatchMessage\n\ttopicMap[MapperTopicPrefix+c.Device.ID+SchedulerCreateTopicSuffix] = c.handleScheduleCreateMessage\n\ttopicMap[MapperTopicPrefix+c.Device.ID+SchedulerDeleteTopicSuffix] = c.handleScheduleDeleteMessage\n\ttopicMap[MapperTopicPrefix+c.Device.ID+ActionManagerCreateTopicSuffix] = c.handleActionCreateMessage\n\ttopicMap[MapperTopicPrefix+c.Device.ID+ActionManagerDeleteTopicSuffix] = c.handleActionDeleteMessage\n}", "func (s *Server) Register(service, method string, h http.HandlerFunc) {\n\ts.routes[fmt.Sprintf(\"%s%s.%s\", s.Basepath, service, method)] = h\n}", "func RegisterAPIRoute(ginEngine *gin.Engine, controllers []IBaseController) {\n\troutesControllerMapping(ginEngine, controllers)\n}", "func RegisterProtocol(name string, v interface{}) {\n\troot.Protocols[name] = v\n}", "func (p *JSONProtocol) RegisterWithName(msg interface{}, msgName string) {\n\tt := reflect.TypeOf(msg)\n\tif t.Kind() == reflect.Ptr {\n\t\tt = t.Elem()\n\t}\n\tif msgName == \"\" {\n\t\tmsgName = t.PkgPath() + \"/\" + t.Name()\n\t}\n\tp.types[msgName] = t\n\tp.names[t] = msgName\n}", "func NewStrategy(defaultRegistry DefaultRegistry) imageRepositoryStrategy {\n\treturn imageRepositoryStrategy{kapi.Scheme, kapi.SimpleNameGenerator, defaultRegistry}\n}", "func (r *Router) Register(log LogFunc, routes ...Route) *Router {\n\tfor _, route := range routes {\n\t\tr.inner.Path(route.Path).\n\t\t\tMethods(route.Method).\n\t\t\tHandlerFunc(route.Handler.HTTP(log))\n\t}\n\treturn r\n}", "func (b *Broker) AddTopic(subscriber *Subscriber, keyTopic string, priority uint, action func(msg interface{}) (err error)) {\n\n\tt := &topicSub{priority: priority, action: action, channel: make(chan interface{}, 1)}\n\n\tsubscriber.topic = t\n\n\tsubscriber.debug = keyTopic\n\t//subscriber.byPriority = append(subscriber.byPriority, t)\n\n\t// Sort topics by priority, will be useful to manage priorities after\n\n\tb.addSubscriber(keyTopic, subscriber)\n\n\t// Sort byPriority array\n\t//\tsort.SliceStable(subscriber.byPriority, func(i, j int) bool { return subscriber.byPriority[i].priority < subscriber.byPriority[j].priority })\n}", "func (sr *Sereport) RegisterRoute() *mux.Router {\n\tr := mux.NewRouter()\n\tsr.api.RegisterRoute(r)\n\treturn r\n}", "func (r *Routes) Add(pattern string, f fasthttp.RequestHandler) {\n\tr.Routes[normaliseRoute(pattern)] = f\n}", "func RegisterDriver(drv Driver, scheme string) {\n\tdrivers[scheme] = drv\n}", "func (r *templateRouter) dynamicallyAddRoute(backendKey ServiceAliasConfigKey, route *routev1.Route, backend *ServiceAliasConfig) bool {\n\tif r.dynamicConfigManager == nil {\n\t\treturn false\n\t}\n\n\tlog.V(4).Info(\"dynamically adding route backend\", \"backendKey\", backendKey)\n\tr.dynamicConfigManager.Register(backendKey, route)\n\n\t// If no initial sync was done, don't try to dynamically add the\n\t// route as we will need a reload anyway.\n\tif !r.synced {\n\t\treturn false\n\t}\n\n\terr := r.dynamicConfigManager.AddRoute(backendKey, backend.RoutingKeyName, route)\n\tif err != nil {\n\t\tlog.V(4).Info(\"router will reload as the ConfigManager could not dynamically add route for backend\", \"backendKey\", backendKey, \"error\", err)\n\t\treturn false\n\t}\n\n\t// For each referenced service unit replace the route endpoints.\n\toldEndpoints := []Endpoint{}\n\n\t// As the endpoints have changed, recalculate the weights.\n\tnewWeights := r.calculateServiceWeights(backend.ServiceUnits, backend.PreferPort)\n\tfor key := range backend.ServiceUnits {\n\t\tif service, ok := r.findMatchingServiceUnit(key); ok {\n\t\t\tnewEndpoints := endpointsForAlias(*backend, service)\n\t\t\tlog.V(4).Info(\"for new route backend, replacing endpoints for service\", \"backendKey\", backendKey, \"serviceKey\", key, \"newEndpoints\", newEndpoints)\n\n\t\t\tweight, ok := newWeights[key]\n\t\t\tif !ok {\n\t\t\t\tweight = 0\n\t\t\t}\n\t\t\tif err := r.dynamicConfigManager.ReplaceRouteEndpoints(backendKey, oldEndpoints, newEndpoints, weight); err != nil {\n\t\t\t\tlog.V(4).Info(\"router will reload as the ConfigManager could not dynamically replace endpoints for route backend\",\n\t\t\t\t\t\"backendKey\", backendKey, \"serviceKey\", key, \"error\", err)\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t}\n\n\tlog.V(4).Info(\"dynamically added route backend\", \"backendKey\", backendKey)\n\treturn true\n}", "func (l *Labeltile) SetupRoutes(injectFns ...RouteInjector) {\n\tfor _, fn := range injectFns {\n\t\tfn(l.engine, l.container)\n\t}\n}", "func (client *Client) DescribeRouteDistributionStrategiesWithCallback(request *DescribeRouteDistributionStrategiesRequest, callback func(response *DescribeRouteDistributionStrategiesResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DescribeRouteDistributionStrategiesResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DescribeRouteDistributionStrategies(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func RegisterEndpoints(router *mux.Router) {\n\tfmt.Println(\"Registering wheel endpoints...\")\n\twheelDriver = NewDriver(adaptor.RPi)\n\trouter.HandleFunc(\"/wheels\", setWheelsMovement).Methods(http.MethodPut)\n}", "func Register(plugins *admission.Plugins) {\n\tplugins.Register(\"FooBar\", func(config io.Reader) (admission.Interface, error) {\n\t\treturn New()\n\t})\n}", "func (s *sink) register(sub *sinksub) {\n\ts.trieq0.Add(sub)\n\ts.trieq1.Add(sub)\n}", "func Register(c *client.Client) {\n\tctrlz.RegisterTopic(CreateTopic(c))\n}", "func (c *Control) registerTunnel(rawTunnelReq *msg.ReqTunnel) {\n\tfor _, proto := range strings.Split(rawTunnelReq.Protocol, \"+\") {\n\t\ttunnelReq := *rawTunnelReq\n\t\ttunnelReq.Protocol = proto\n\n\t\tc.conn.Debug(\"Registering new tunnel\")\n\t\tt, err := NewTunnel(&tunnelReq, c)\n\t\tif err != nil {\n\t\t\tack := &msg.NewTunnel{Error: err.Error()}\n\t\t\tif len(c.tunnels) == 0 {\n\t\t\t\t// you can't fail your first tunnel registration\n\t\t\t\t// terminate the control connection\n\t\t\t\tc.stop <- ack\n\t\t\t} else {\n\t\t\t\t// inform client of failure\n\t\t\t\tc.out <- ack\n\t\t\t}\n\n\t\t\t// we're done\n\t\t\treturn\n\t\t}\n\n\t\t// add it to the list of tunnels\n\t\tc.tunnels = append(c.tunnels, t)\n\n\t\t// acknowledge success\n\t\tc.out <- &msg.NewTunnel{\n\t\t\tUrl: t.url,\n\t\t\tProtocol: proto,\n\t\t\tReqId: rawTunnelReq.ReqId,\n\t\t}\n\n\t\trawTunnelReq.Hostname = strings.Replace(t.url, proto+\"://\", \"\", 1)\n\t}\n}", "func RegisterRoutes(cliCtx context.CLIContext, r *mux.Router) {\r\n\t// this line is used by starport scaffolding\r\n\tr.HandleFunc(\"/voter/poll\", ListPollHandler(cliCtx, \"voter\")).Methods(\"GET\")\r\n\tr.HandleFunc(\"/voter/poll\", createPollHandler(cliCtx)).Methods(\"POST\")\r\n}", "func RegisterRoute(w http.ResponseWriter, r *http.Request) {\n\taddr := r.URL.Query().Get(\"address\")\n\tport, err := strconv.Atoi(r.URL.Query().Get(\"port\"))\n\tif (addr == \"\" || port <= 0) || err != nil {\n\t\thttp.Error(w, \"'address' and/or 'port' query params are bad, please fix\", http.StatusBadRequest)\n\t\treturn\n\t}\n\tip := net.ParseIP(addr)\n\tregister(dist.Address{\n\t\tIP: ip,\n\t\tPort: port,\n\t})\n\tw.WriteHeader(204) // 'No Content'\n}", "func AddOutputStrategy(name string, f OutputStrategyCreator) {\n\toutputStrategies[name] = f\n}", "func registerPostRoutes(api iris.Party) {\n\tpostRoutes := api.Party(\"/posts\")\n\tpostRoutes.Use(middlewares.CheckLoginStatus)\n\n\t// subpath of \"/posts\"\n\t// Post Collection and Creation\n\tpostRoutes.Get(\"/\", service.GetRecentPosts).Name = \"GetRecentPosts\"\n\tpostRoutes.Post(\"/\", service.CreatePost).Name = \"CreatePost\"\n\n\t// Get and Delete Post\n\tpostRoutes.Get(\"/{postId:int min(1)}\", middlewares.CheckPostIDExistence,\n\t\tservice.GetPostByID).Name = \"GetPostByID\"\n\tpostRoutes.Delete(\"/{postId:int min(1)}\", middlewares.CheckPostIDExistence,\n\t\tservice.DeletePost).Name = \"DeletePost\"\n\n\t// share a post\n\t// postRoutes.Get(\"/{postId:int min(1)/share\", handler)\n\n\t// liek/un-like a post\n\tpostRoutes.Get(\"/{postId:int min(1)}/like\", middlewares.CheckPostIDExistence,\n\t\tservice.UpvotePost).Name = \"UpvotePost\"\n\n\t// report a post\n\tpostRoutes.Post(\"/{postId:int min(1)}/report\", middlewares.CheckPostIDExistence,\n\t\tservice.CreateReportPost).Name = \"CreateReportPost\"\n\n\t// star a post\n\tpostRoutes.Get(\"/{postId:int min(1)}/star\", middlewares.CheckPostIDExistence,\n\t\tservice.StarPost).Name = \"StarPost\"\n\n\t// share a post\n\tpostRoutes.Get(\"/{postId:int min(1)}/share\", middlewares.CheckPostIDExistence,\n\t\tservice.SharePost).Name = \"SharePost\"\n\n\t// all category names\n\tpostRoutes.Get(\"/categories\", service.GetAllCategoryNames).Name = \"GetAllCategoryNames\"\n}", "func setupRoutes(srv *gin.Engine, s *controller) *gin.Engine {\n\n\tpingroute.NewRouter(s.pingController, srv).RegisterRoutes()\n\tqrcoderoute.NewRouter(s.qrcodeController, srv).RegisterRoutes()\n\tuserroute.NewRouter(s.userController, srv).RegisterRoutes()\n\tsafraroute.NewRouter(s.safraController, srv).RegisterRoutes()\n\n\treturn srv\n}", "func (r *rdsRoute) Add(ns, topic, address string, version uint64) error {\n\tkey := r.getRoutePrefix(ns, topic)\n\trds := r.rdc.Get(util.W, key)\n\tdefer rds.Close()\n\t_, err := rds.Do(\"HSET\", key, address, version)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *Switch) RegisterDirectProtocolWithChannel(protocol string, ingressChannel chan service.DirectMessage) chan service.DirectMessage {\n\tif s.started == 1 {\n\t\tlog.Panic(\"attempting to register direct protocol with channel after p2p has started\")\n\t}\n\ts.directProtocolHandlers[protocol] = ingressChannel\n\treturn ingressChannel\n}", "func (s Service) Register(engine *gin.RouterGroup) {\n\tfor i := 0; i < len(s.Routes); i++ {\n\t\tengine.Handle(s.Routes[i].Method, s.Prefix+s.Routes[i].Url, s.Routes[i].Handler)\n\t}\n}" ]
[ "0.6233871", "0.60150677", "0.5539055", "0.5388272", "0.5353805", "0.5258051", "0.5237158", "0.5210581", "0.52068865", "0.5089592", "0.50649077", "0.50563955", "0.5038952", "0.49194244", "0.48399246", "0.48336083", "0.48249227", "0.48226076", "0.48137185", "0.47765005", "0.47171158", "0.4699946", "0.46942163", "0.4683713", "0.46646646", "0.46525216", "0.46375036", "0.45974588", "0.45926422", "0.4587312", "0.45839214", "0.4583406", "0.45766097", "0.45709026", "0.45647272", "0.45448098", "0.4541369", "0.45290783", "0.45248446", "0.45214218", "0.45185527", "0.45148847", "0.4514022", "0.45135486", "0.45126092", "0.44894752", "0.44888514", "0.44843557", "0.44770533", "0.44698477", "0.44510818", "0.44503424", "0.4443307", "0.44399884", "0.4437437", "0.44363144", "0.44182473", "0.44160986", "0.44102848", "0.43994096", "0.43933317", "0.4386697", "0.43794006", "0.4379104", "0.43753737", "0.43708453", "0.4358872", "0.43536055", "0.43508503", "0.43493518", "0.43478355", "0.43477994", "0.43437767", "0.433444", "0.43284985", "0.4328204", "0.43067694", "0.43033844", "0.42972434", "0.42904246", "0.42841724", "0.42818552", "0.42782813", "0.42462537", "0.42422473", "0.4241987", "0.4241986", "0.42405534", "0.4235132", "0.42311", "0.4230506", "0.42269683", "0.42264813", "0.42255124", "0.4220519", "0.42203653", "0.4217864", "0.42077243", "0.42066112", "0.42003986" ]
0.8255487
0
Write write the payload to a parition in the topic based on the RoutingStrategy specified
func (n *Node) Write(topic, payload string) error { rs := n.getStrategyFor(topic) partID := rs.WhichPartition(payload) parts := n.getPartitionsFor(topic) return parts[partID].Write(payload) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Subscriber) write(mt int, payload []byte) error {\n s.ws.SetWriteDeadline(time.Now().Add(writeWait))\n return s.ws.WriteMessage(mt, payload)\n}", "func (route_p *Router) publish(pkt_p *defs.UDPMsg, addr_p *net.UDPAddr) {\n\tif pkt_p.Flag&defs.UDPMsg_NewNode == defs.UDPMsg_NewNode {\n\t\troute_p.nmap_p.AddNbr(&pkt_p.DstHash, addr_p)\n\t} else {\n\t\troute_p.omap_p.Insert(&(pkt_p.DstHash), &(pkt_p.Payload.Msg), pkt_p.Payload.IsAddr)\n\t}\n\n\tpkt_p.Hops--\n\tif pkt_p.Hops > 0 {\n\t\tnbrs, err := route_p.nmap_p.GetNbr(&pkt_p.DstHash)\n\t\tif err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\tcmds, err := proto.Marshal(pkt_p)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfor _, nbr_p := range nbrs {\n\t\t\t\tif _, err := route_p.ucon_p.WriteToUDP(cmds, nbr_p.Addr_p); err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t} else {\n\t\treturn\n\t}\n}", "func (b *RouteBuilder) Write(sample interface{}) *RouteBuilder {\n\tb.writeSample = sample\n\treturn b\n}", "func (d *RMQ) PublishTo(topic string) error { return nil }", "func SendPayloadToTopic(topic, payload string) error {\n\tif topic == \"\" {\n\t\treturn errors.New(\"topic is missing\")\n\t}\n\n\tl, err := standalone.List()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(l) == 0 {\n\t\treturn errors.New(\"couldn't find a running Dapr instance\")\n\t}\n\n\tapp := l[0]\n\tb := []byte{}\n\n\tif payload != \"\" {\n\t\tb = []byte(payload)\n\t}\n\n\turl := fmt.Sprintf(\"http://localhost:%s/v%s/publish/%s\", fmt.Sprintf(\"%v\", app.HTTPPort), api.RuntimeAPIVersion, topic)\n\t// nolint: gosec\n\tr, err := http.Post(url, \"application/json\", bytes.NewBuffer(b))\n\n\tif r != nil {\n\t\tdefer r.Body.Close()\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (conn *extHost) sendMessageToReplicas(pr *inPutMessage, extSendTimer *common.Timer, watermark *int64) (int64, error) {\n\tvar err error\n\n\twatermarkOnly := pr == nil\n\tif watermarkOnly && conn.destType != shared.DestinationType_LOG {\n\t\tlog.Fatal(\"WatermarkOnly message requested for non LOG destination\")\n\t}\n\tmsg := store.NewAppendMessage()\n\tvar appendMsgAckCh chan *store.AppendMessageAck\n\tif watermarkOnly {\n\t\tif watermark == nil {\n\t\t\tlog.Fatal(\"nil watermark and pr\")\n\t\t}\n\t\tif conn.lastSentWatermark == *watermark {\n\t\t\treturn -1, nil\n\t\t}\n\t}\n\n\tconn.extMetrics.Increment(load.ExtentMetricMsgsIn)\n\tconn.dstMetrics.Increment(load.DstMetricMsgsIn)\n\tconn.hostMetrics.Increment(load.HostMetricMsgsIn)\n\n\t// increment seq-num; do atomically, since this could\n\t// be concurrently queried by the reporter\n\tsequenceNumber := atomic.AddInt64(&conn.seqNo, 1)\n\tmsg.SequenceNumber = common.Int64Ptr(sequenceNumber)\n\tmsg.EnqueueTimeUtc = common.Int64Ptr(conn.getEnqueueTime())\n\tif !watermarkOnly {\n\t\tmsg.Payload = pr.putMsg\n\t\tappendMsgAckCh = make(chan *store.AppendMessageAck, 5)\n\t}\n\tif watermark != nil && conn.lastSentWatermark < *watermark {\n\t\tmsg.FullyReplicatedWatermark = watermark\n\t}\n\n\t// we write the above same message to all the replicas\n\t// even if one of the replicas fail, we consider the message failed\n\t// no need to lock the conn.streams here because the replica set\n\t// for an extent will not change at all\n\terrCh := make(chan error)\n\tfor _, stream := range conn.streams {\n\t\tgo func(replInfo *replicaInfo, aMsg *store.AppendMessage, aMsgAckCh chan *store.AppendMessageAck) {\n\t\t\tpMsg := &replicaPutMsg{\n\t\t\t\tappendMsg: aMsg,\n\t\t\t\tappendMsgAckCh: aMsgAckCh,\n\t\t\t}\n\n\t\t\t// log disabled due to CPU cost\n\t\t\t// conn.logger.WithFields(logger.Fields{`replica`: replica, common.TagSeq: conn.seqNo, `Payload`: msg.Payload,}).Debug(`inputhost: sending data to store: ; seqno: , data`)\n\n\t\t\treplInfo.sendTimer.Reset(replicaSendTimeout)\n\t\t\tselect {\n\t\t\tcase replInfo.conn.putMessagesCh <- pMsg:\n\t\t\tcase <-replInfo.sendTimer.C:\n\t\t\t\terrCh <- ErrTimeout\n\t\t\t\treturn\n\t\t\t}\n\t\t\terrCh <- nil\n\t\t\treturn\n\t\t}(stream, msg, appendMsgAckCh)\n\t}\n\n\t// Wait for all the go routines above; we wait on the errCh to get the response from all replicas\n\tfor replica, stream := range conn.streams {\n\t\terr = <-errCh\n\t\tif err != nil {\n\t\t\tif watermarkOnly {\n\t\t\t\tconn.logger.WithFields(bark.Fields{`replica`: replica, common.TagErr: err, `putMessagesChLength`: len(stream.conn.putMessagesCh)}).Warn(`inputhost: sending fully replicated watermark to replica: , failed with error: ; length of putMsgCh: ;`)\n\t\t\t} else {\n\t\t\t\tconn.logger.WithFields(bark.Fields{`replica`: replica, common.TagErr: err, `putMessagesChLength`: len(stream.conn.putMessagesCh), `replyChLength`: len(stream.conn.replyCh)}).Error(`inputhost: sending msg to replica: , failed with error: ; length of putMsgCh: ; length of replyCh: ;`)\n\t\t\t}\n\t\t\treturn sequenceNumber, err\n\t\t}\n\t}\n\n\tif !watermarkOnly {\n\t\textSendTimer.Reset(replicaSendTimeout)\n\t\t// this is for the extHost's inflight messages for a successful message\n\t\tselect {\n\t\tcase conn.replyClientCh <- writeResponse{pr.putMsg.GetID(), sequenceNumber, appendMsgAckCh, pr.putMsgAckCh, pr.putMsgRecvTime, pr.putMsg.GetUserContext()}:\n\t\tcase <-extSendTimer.C:\n\t\t\tconn.logger.WithField(`lenReplyClientCh`, len(conn.replyClientCh)).Error(`inputhost: exthost: sending msg to the replyClientCh on exthost timed out`)\n\t\t\terr = ErrTimeout\n\t\t}\n\t}\n\tif err == nil && watermark != nil {\n\t\tconn.lastSentWatermark = *watermark\n\t}\n\treturn sequenceNumber, err\n}", "func WritePayload(w io.Writer, session uint16, body []byte, meta uint32) (err error) {\n\thead := BuildHeader(session, body, meta)\n\tif _, err := w.Write(head); err != nil {\n\t\treturn err\n\t}\n\tif _, err := w.Write(body); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *Client) write(mt int, payload []byte) error {\n\tc.ws.SetWriteDeadline(time.Now().Add(writeWait))\n\treturn c.ws.WriteMessage(mt, payload)\n}", "func (self *RouteBuilder) Writes(sample interface{}) *RouteBuilder {\n\tself.writeSample = sample\n\treturn self\n}", "func (ll *LogLink) Write(l CloudwatchLinkAddress, protocol tcpip.NetworkProtocolNumber, header []byte, payload []byte) (int, error) {\n\t// todo: replace with pcap-friendly format\n\tpl := PacketLog{ll.ProtocolToString(protocol), l.Src().String(), l.Dest().String(),\n\t\tbase64.StdEncoding.EncodeToString(header), base64.StdEncoding.EncodeToString(payload)}\n\tplBytes, err := json.Marshal(pl)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tll.writePoller.Cw <- WritePollInput{plBytes, &l}\n\treturn len(plBytes), nil\n}", "func (m *MQTT) Write(msg *message.Batch) error {\n\tm.connMut.RLock()\n\tclient := m.client\n\tm.connMut.RUnlock()\n\n\tif client == nil {\n\t\treturn component.ErrNotConnected\n\t}\n\n\treturn IterateBatchedSend(msg, func(i int, p *message.Part) error {\n\t\tretained := m.conf.Retained\n\t\tif m.retained != nil {\n\t\t\tvar parseErr error\n\t\t\tretained, parseErr = strconv.ParseBool(m.retained.String(i, msg))\n\t\t\tif parseErr != nil {\n\t\t\t\tm.log.Errorf(\"Error parsing boolean value from retained flag: %v \\n\", parseErr)\n\t\t\t}\n\t\t}\n\t\tmtok := client.Publish(m.topic.String(i, msg), m.conf.QoS, retained, p.Get())\n\t\tmtok.Wait()\n\t\tsendErr := mtok.Error()\n\t\tif sendErr == mqtt.ErrNotConnected {\n\t\t\tm.connMut.RLock()\n\t\t\tm.client = nil\n\t\t\tm.connMut.RUnlock()\n\t\t\tsendErr = component.ErrNotConnected\n\t\t}\n\t\treturn sendErr\n\t})\n}", "func (opcuaExport *OpcuaExport) Publish(data interface{}) {\n\tpubTopics := opcuaExport.opcuaBus.pubTopics\n\ttopicConfigs := make([]map[string]string, len(pubTopics))\n\tfor i, pubTopic := range pubTopics {\n\t\ttopicConfigs[i] = map[string]string{\"ns\": \"StreamManager\", \"name\": pubTopic, \"dType\": \"string\"}\n\t}\n\tfor _, topicConfig := range topicConfigs {\n\t\topcuaData := fmt.Sprintf(\"%s %v\", topicConfig[\"name\"], data)\n\t\topcuaExport.opcuaBus.opcuaDatab.Publish(topicConfig, opcuaData)\n\t\tglog.Infof(\"Published data: %v on topic: %s\\n\", opcuaData, topicConfig)\n\t}\n}", "func (c *connection) write(mt int, payload []byte) error {\n\tc.ws.SetWriteDeadline(time.Now().Add(writeWait))\n\treturn c.ws.WriteMessage(mt, payload)\n}", "func (c *connection) write(mt int, payload []byte) error {\n\tc.ws.SetWriteDeadline(time.Now().Add(writeWait))\n\treturn c.ws.WriteMessage(mt, payload)\n}", "func (c *connectSvr) write(mt int, payload []byte) error {\n\tc.ws.SetWriteDeadline(time.Now().Add(writeWait))\n\treturn c.ws.WriteMessage(mt, payload)\n}", "func (c *connection) write(opCode int, payload []byte) error {\n\tc.ws.SetWriteDeadline(time.Now().Add(writeWait))\n\treturn c.ws.WriteMessage(opCode, payload)\n}", "func (n *Node) RegisterStrategy(topic string, strategy RoutingStrategy) {\n\tn.strategyMap[topic] = strategy\n}", "func (w *Writer) Write(key string, req *remote.WriteRequest) error {\n\tdata, err := proto.Marshal(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tm := &sarama.ProducerMessage{\n\t\tTopic: w.topic,\n\t\tKey: sarama.StringEncoder(key),\n\t\tValue: sarama.ByteEncoder(data),\n\t}\n\n\tgo func() {\n\t\tw.producer.Input() <- m\n\n\t\tw.queuedForWrites.Inc()\n\t}()\n\n\treturn nil\n}", "func (t *Transport) Publish(message string) {\n\t// Some voodoo here to actually send the string over the wire\n\n\tlog.Println(\"Got the message: \", message)\n}", "func (c *Connection) write(msg interface{}) {\n c.mutex.Lock()\n c.log(fmt.Sprintf(\"Sending message: %+v\", msg))\n c.socket.WriteJSON(msg)\n c.mutex.Unlock()\n}", "func (c client) Publish(route loadbalancer.Route) (loadbalancer.Result, error) {\n\t_, l4Type := c.name.GetLookupAndType()\n\treq := PublishRequest{Type: l4Type, Route: route}\n\tresp := PublishResponse{}\n\n\tif err := c.client.Call(\"L4.Publish\", req, &resp); err != nil {\n\t\treturn nil, err\n\t}\n\treturn clientResult(resp.Result), nil\n}", "func (h *Handler) Publish(c *session.Client, topic *string, payload *[]byte) {\n\th.logger.Info(fmt.Sprintf(\"Publish() - username: %s, clientID: %s, topic: %s, payload: %s\", c.Username, c.ID, *topic, string(*payload)))\n}", "func (conn *Connection) write(mt int, payload []byte, wsc config.WebSocketSettings) error {\n\tconn.ws.SetWriteDeadline(time.Now().Add(wsc.WriteWait))\n\treturn conn.ws.WriteMessage(mt, payload)\n}", "func (c *TCPClient) WritePassthrough(\n\tmetric aggregated.Metric,\n\tstoragePolicy policy.StoragePolicy,\n) error {\n\tpayload := payloadUnion{\n\t\tpayloadType: passthroughType,\n\t\tpassthrough: passthroughPayload{\n\t\t\tmetric: metric,\n\t\t\tstoragePolicy: storagePolicy,\n\t\t},\n\t}\n\n\tc.metrics.writePassthrough.Inc(1)\n\treturn c.write(metric.ID, metric.TimeNanos, payload)\n}", "func (c *Conn) write(mt int, payload []byte) error {\n\tc.ws.SetWriteDeadline(time.Now().Add(writeWait))\n\treturn c.ws.WriteMessage(mt, payload)\n}", "func SendMessageToRoute(msg interface{}, exchangeName string, routeKey string) {\n sendMessage(msg, DirectExchange, exchangeName, \"\", routeKey, true)\n}", "func Write(s beam.Scope, project, topic string, col beam.PCollection) {\n\ts = s.Scope(\"pubsubio.Write\")\n\n\tpayload := &pipepb.PubSubWritePayload{\n\t\tTopic: pubsubx.MakeQualifiedTopicName(project, topic),\n\t}\n\n\tout := col\n\tif col.Type().Type() == reflectx.ByteSlice {\n\t\tout = beam.ParDo(s, wrapInMessage, col)\n\t}\n\tif out.Type().Type() != pubSubMessageT {\n\t\tpanic(fmt.Sprintf(\"pubsubio.Write only accepts PCollections of %v and %v, received %v\", pubSubMessageT, reflectx.ByteSlice, col.Type().Type()))\n\t}\n\tmarshaled := beam.ParDo(s, marshalMessageFn, out)\n\tbeam.External(s, writeURN, protox.MustEncode(payload), []beam.PCollection{marshaled}, nil, false)\n}", "func (tg *TradesGroup) publish(data interface{}, dataType string, err error) {\n\ttg.bus.outChannel <- schemas.ResultChannel{\n\t\tDataType: dataType,\n\t\tData: data,\n\t\tError: err,\n\t}\n}", "func (m *MemBroker) Publish(topic string, payload interface{}) {\n\tm.ingress <- event{topic, payload}\n}", "func (ne NullEndpoint) Write(p model.AgentPayload) (int, error) {\n\tlog.Debug(\"null endpoint: dropping payload, %d traces, %d stats buckets\", p.Traces, p.Stats)\n\treturn 0, nil\n}", "func (v *ViewView) write(msg string) {\n\tfmt.Fprint(v.Writer, msg)\n}", "func TopicPublish(w http.ResponseWriter, r *http.Request) {\n\t// Init output\n\toutput := []byte(\"\")\n\n\t// Add content type header to the response\n\tcontentType := \"application/json\"\n\tcharset := \"utf-8\"\n\tw.Header().Add(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\t// Get url path variables\n\turlVars := mux.Vars(r)\n\turlTopic := urlVars[\"topic\"]\n\n\t// Grab context references\n\n\trefBrk := gorillaContext.Get(r, \"brk\").(brokers.Broker)\n\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\trefUserUUID := gorillaContext.Get(r, \"auth_user_uuid\").(string)\n\trefRoles := gorillaContext.Get(r, \"auth_roles\").([]string)\n\trefAuthResource := gorillaContext.Get(r, \"auth_resource\").(bool)\n\t// Get project UUID First to use as reference\n\tprojectUUID := gorillaContext.Get(r, \"auth_project_uuid\").(string)\n\n\tresults, err := topics.Find(projectUUID, \"\", urlVars[\"topic\"], \"\", 0, refStr)\n\n\tif err != nil {\n\t\terr := APIErrGenericBackend()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// If not found\n\tif results.Empty() {\n\t\terr := APIErrorNotFound(\"Topic\")\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\tres := results.Topics[0]\n\n\t// Check Authorization per topic\n\t// - if enabled in config\n\t// - if user has only publisher role\n\n\tif refAuthResource && auth.IsPublisher(refRoles) {\n\n\t\tif auth.PerResource(projectUUID, \"topics\", urlTopic, refUserUUID, refStr) == false {\n\t\t\terr := APIErrorForbidden()\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t}\n\n\t// Read POST JSON body\n\tbody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\terr := APIErrorInvalidRequestBody()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Create Message List from Post JSON\n\tmsgList, err := messages.LoadMsgListJSON(body)\n\tif err != nil {\n\t\terr := APIErrorInvalidArgument(\"Message\")\n\t\trespondErr(w, err)\n\t\tlog.Error(string(body[:]))\n\t\treturn\n\t}\n\n\t// check if the topic has a schema associated with it\n\tif res.Schema != \"\" {\n\n\t\t// retrieve the schema\n\t\t_, schemaName, err := schemas.ExtractSchema(res.Schema)\n\t\tif err != nil {\n\t\t\tlog.WithFields(\n\t\t\t\tlog.Fields{\n\t\t\t\t\t\"type\": \"service_log\",\n\t\t\t\t\t\"schema_name\": res.Schema,\n\t\t\t\t\t\"topic_name\": res.Name,\n\t\t\t\t\t\"error\": err.Error(),\n\t\t\t\t},\n\t\t\t).Error(\"Could not extract schema name\")\n\t\t\terr := APIErrGenericInternal(schemas.GenericError)\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\tsl, err := schemas.Find(projectUUID, \"\", schemaName, refStr)\n\n\t\tif err != nil {\n\t\t\tlog.WithFields(\n\t\t\t\tlog.Fields{\n\t\t\t\t\t\"type\": \"service_log\",\n\t\t\t\t\t\"schema_name\": schemaName,\n\t\t\t\t\t\"topic_name\": res.Name,\n\t\t\t\t\t\"error\": err.Error(),\n\t\t\t\t},\n\t\t\t).Error(\"Could not retrieve schema from the store\")\n\t\t\terr := APIErrGenericInternal(schemas.GenericError)\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\tif !sl.Empty() {\n\t\t\terr := schemas.ValidateMessages(sl.Schemas[0], msgList)\n\t\t\tif err != nil {\n\t\t\t\tif err.Error() == \"500\" {\n\t\t\t\t\terr := APIErrGenericInternal(schemas.GenericError)\n\t\t\t\t\trespondErr(w, err)\n\t\t\t\t\treturn\n\t\t\t\t} else {\n\t\t\t\t\terr := APIErrorInvalidData(err.Error())\n\t\t\t\t\trespondErr(w, err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tlog.WithFields(\n\t\t\t\tlog.Fields{\n\t\t\t\t\t\"type\": \"service_log\",\n\t\t\t\t\t\"schema_name\": res.Schema,\n\t\t\t\t\t\"topic_name\": res.Name,\n\t\t\t\t},\n\t\t\t).Error(\"List of schemas was empty\")\n\t\t\terr := APIErrGenericInternal(schemas.GenericError)\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t}\n\n\t// Init message ids list\n\tmsgIDs := messages.MsgIDs{IDs: []string{}}\n\n\t// For each message in message list\n\tfor _, msg := range msgList.Msgs {\n\t\t// Get offset and set it as msg\n\t\tfullTopic := projectUUID + \".\" + urlTopic\n\n\t\tmsgID, rTop, _, _, err := refBrk.Publish(fullTopic, msg)\n\n\t\tif err != nil {\n\t\t\tif err.Error() == \"kafka server: Message was too large, server rejected it to avoid allocation error.\" {\n\t\t\t\terr := APIErrTooLargeMessage(\"Message size too large\")\n\t\t\t\trespondErr(w, err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\terr := APIErrGenericBackend()\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\tmsg.ID = msgID\n\t\t// Assertions for Succesfull Publish\n\t\tif rTop != fullTopic {\n\t\t\terr := APIErrGenericInternal(\"Broker reports wrong topic\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\t// Append the MsgID of the successful published message to the msgIds list\n\t\tmsgIDs.IDs = append(msgIDs.IDs, msg.ID)\n\t}\n\n\t// timestamp of the publish event\n\tpublishTime := time.Now().UTC()\n\n\t// amount of messages published\n\tmsgCount := int64(len(msgList.Msgs))\n\n\t// increment topic number of message metric\n\trefStr.IncrementTopicMsgNum(projectUUID, urlTopic, msgCount)\n\n\t// increment daily count of topic messages\n\tyear, month, day := publishTime.Date()\n\trefStr.IncrementDailyTopicMsgCount(projectUUID, urlTopic, msgCount, time.Date(year, month, day, 0, 0, 0, 0, time.UTC))\n\n\t// increment topic total bytes published\n\trefStr.IncrementTopicBytes(projectUUID, urlTopic, msgList.TotalSize())\n\n\t// update latest publish date for the given topic\n\trefStr.UpdateTopicLatestPublish(projectUUID, urlTopic, publishTime)\n\n\t// count the rate of published messages per sec between the last two publish events\n\tvar dt float64 = 1\n\t// if its the first publish to the topic\n\t// skip the subtraction that computes the DT between the last two publish events\n\tif !res.LatestPublish.IsZero() {\n\t\tdt = publishTime.Sub(res.LatestPublish).Seconds()\n\t}\n\trefStr.UpdateTopicPublishRate(projectUUID, urlTopic, float64(msgCount)/dt)\n\n\t// Export the msgIDs\n\tresJSON, err := msgIDs.ExportJSON()\n\tif err != nil {\n\t\terr := APIErrExportJSON()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Write response\n\toutput = []byte(resJSON)\n\trespondOK(w, output)\n}", "func (c *connection) Write(mt int, payload []byte) error {\r\n\tc.ws.SetWriteDeadline(time.Now().Add(writeWait))\r\n\treturn c.ws.WriteMessage(mt, payload)\r\n}", "func (c *Client) Publish(topic string, payload []byte) error {\n\tif c.connected == false {\n\t\treturn ErrNotConnected\n\t}\n\n\tbuff := bytes.NewBuffer(make([]byte, 0, 1+8+len(topic)+len(payload)))\n\n\terr := buff.WriteByte(PUBLISH)\n\tif err != nil {\n\t\tfmt.Println(\"error writing packet type\")\n\t\treturn err\n\t}\n\n\terr = binary.Write(buff, binary.LittleEndian, uint32(len(topic)))\n\tif err != nil {\n\t\tfmt.Println(\"error writing topic length\")\n\t\treturn err\n\t}\n\n\t_, err = buff.WriteString(topic)\n\tif err != nil {\n\t\tfmt.Println(\"error writing topic data\")\n\t\treturn err\n\t}\n\n\terr = binary.Write(buff, binary.LittleEndian, uint32(len(payload)))\n\tif err != nil {\n\t\tfmt.Println(\"error writing payload length\")\n\t\treturn err\n\t}\n\n\t_, err = buff.Write(payload)\n\tif err != nil {\n\t\tfmt.Println(\"error writing payload data\")\n\t\treturn err\n\t}\n\n\t//c.conn.Write(buff.Bytes())\n\tbuff.WriteTo(c.conn)\n\n\treturn nil\n}", "func (c *Client) write(mt int, payload []byte) error {\n\tc.conn.SetWriteDeadline(time.Now().Add(writeWait))\n\treturn c.conn.WriteMessage(mt, payload)\n}", "func Publish(topic string, data interface{}) (err error) {\n\tvar payload []byte\n\t// topic = prefix + topic\n\tpayload, err = json.Marshal(data)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif defaultProducer == nil {\n\t\treturn\n\t}\n\t// log.Debugf(\"[MQ] Publishing topic=%s payload=%s\", topic, string(payload))\n\treturn defaultProducer.Publish(topic, payload)\n}", "func (e *agentEndpoint) write(msg *submitws.Message) {\n\te.mutex.Lock()\n\tdefer e.mutex.Unlock()\n\tif e.isClosed {\n\t\treturn\n\t}\n\tif err := e.conn.WriteMessage(websocket.BinaryMessage, msg.ToBinary()); err != nil {\n\t\tlogger.WithError(err).Errorf(\"error sending message to agent with id == %s: %v\", e.id, err)\n\t\tif err := e.conn.Close(); err != nil {\n\t\t\tlogger.WithError(err).Errorf(\"error closing connection to agent with id == %s after write error: %v\", e.id, err)\n\t\t}\n\t\te.isClosed = true\n\t}\n}", "func (c *client) Publish(data []byte) error {\n topic := fmt.Sprintf(\n \"iot-2/type/%s/id/%s/evt/usage/fmt/json\",\n c.config.TypeId,\n c.config.DeviceId,\n )\n\n t := c.mqttClient.Publish(topic, 0, false, data)\n t.Wait()\n\n if t.Error() != nil {\n return t.Error()\n }\n\n return nil\n}", "func writer(coord string) {\n\tbroadcast <- coord\n}", "func write(w http.ResponseWriter, status int, payload interface{}) {\n\tw.WriteHeader(status)\n\traw, err := json.Marshal(payload)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\t_, _ = w.Write(raw)\n}", "func Write(tx Transport, host string, data, info []string, options ...TransportOption) error {\n\t// the Kind should configure the transport parameters before\n\n\terr := tx.Connect(host, options...)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"%s: %s\", host, err)\n\t}\n\n\tdefer tx.Close()\n\n\tfor i1, d1 := range data {\n\t\terr := tx.Write(&d1, &info[i1])\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"could not write config %s: %s\", d1, err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (m *MesosMessenger) Route(upid *upid.UPID, msg proto.Message) error {\n\t// if destination is not self, send to outbound.\n\tif !upid.Equal(m.upid) {\n\t\treturn m.Send(upid, msg)\n\t}\n\n\tdata, err := proto.Marshal(msg)\n\tif err != nil {\n\t\treturn err\n\t}\n\tname := getMessageName(msg)\n\treturn m.tr.Inject(&Message{upid, name, msg, data})\n}", "func (msg MsgCreatePDV) Route() string { return RouterKey }", "func (self *Ring) writeToReplicas(sentData *data.DataStore, key int) int {\n i := self.writeToNReplicas(sentData, key, replicaNumber)\n if i == replicaNumber {\n return 1\n }\n return 0\n}", "func (kp *KafkaPacket) Distribute(d interface{}) error {\n\n\t// recover is called here to catch any panic in kafka.NewWriter\n\t// and to release the lock\n\tunlocked := false\n\tdefer func() {\n\t\tif err := recover(); err != nil && !unlocked {\n\t\t\tkrw.writer.Unlock()\n\t\t}\n\t}()\n\n\tkrw.writer.Lock()\n\t// Check for existing Writers. If not existing for this specific Pipe,\n\t// then we would create this Writer object for sending the message.\n\tif _, exist := krw.Writers[kp.pipe]; !exist {\n\t\tif e := kafkaConnect(kp); e != nil {\n\t\t\tkrw.writer.Unlock()\n\t\t\treturn e\n\t\t}\n\n\t\tkrw.Writers[kp.pipe] = kafka.NewWriter(kafka.WriterConfig{\n\t\t\tBrokers: kp.ServersInfo,\n\t\t\tTopic: kp.pipe,\n\t\t\tBalancer: &kafka.RoundRobin{},\n\t\t\tBatchSize: 1,\n\t\t\tQueueCapacity: 1,\n\t\t\tAsync: true,\n\t\t\tDialer: kp.DialerConn,\n\t\t})\n\t}\n\twriter := krw.Writers[kp.pipe]\n\tkrw.writer.Unlock()\n\tunlocked = true\n\n\t// Encode the message before appending into KAFKA Message struct\n\tb, e := Encode(d)\n\tif e != nil {\n\t\treturn fmt.Errorf(\"error: message encoding failed: %s\", e.Error())\n\t}\n\n\t// Place the byte stream into Kafka.Message\n\tkm := kafka.Message{\n\t\tKey: []byte(kp.pipe),\n\t\tValue: b,\n\t}\n\n\t// Write the messgae in the specified Pipe.\n\tif e = writer.WriteMessages(context.Background(), km); e != nil {\n\t\treturn fmt.Errorf(\"error: write message failed: %s\", e.Error())\n\t}\n\n\treturn nil\n}", "func publishToDW(topicID string, object interface{}, info DWDatasetInfo) (string,error) {\n\tlog.Println(\"topic: \", topicID, \" object: \", object, \"datainfo: \", info)\n\tif config.PROFILE == \"dev\" {\n\t\tlog.Println(\"Dev environment: logging message localy\")\n\t} else {\n\t\tt := client.Topic(topicID)\n\t\tdata, _ := json.Marshal(object)\n\t\tpayload := b64.StdEncoding.EncodeToString(data)\n\t\tdwMessage := DWMessage{\n\t\t\tDatasetInfo: info,\n\t\t\tPayload: payload,\n\t\t}\n\t\tdwMessageData, _ := json.Marshal(dwMessage)\n\t\tlog.Println(dwMessageData)\n\t\tresult := t.Publish(ctx, &pubsub.Message{\n\t\t\tData: dwMessageData,\n\t\t})\n\t\t// Block until the result is returned and a server-generated\n\t\t// ID is returned for the published message.\n\t\tid, err := result.Get(ctx)\n\t\tlog.Println(\"message published: \", id)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\treturn id, nil\n\t}\n\treturn \"\", nil\n\n}", "func (_Eth *EthTransactor) Publish(opts *bind.TransactOpts, a common.Address, topic string) (*types.Transaction, error) {\n\treturn _Eth.contract.Transact(opts, \"publish\", a, topic)\n}", "func (producer *Producer) Write(ctx context.Context, msg Msg, topic string) error {\n\tkmsg, result, err := producer.buildMessage(msg, topic)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not build message: %s\", err)\n\t}\n\n\tselect {\n\tcase producer.saramaProducer.Input() <- &kmsg:\n\tcase <-ctx.Done():\n\t\treturn errors.New(\"message write failed: no space in processing channel\")\n\t}\n\n\tselect {\n\tcase err, ok := <-result:\n\t\tif !ok {\n\t\t\treturn errors.New(\"result channel was closed, write result unknown\")\n\t\t}\n\t\treturn err\n\tcase <-ctx.Done():\n\t\treturn errors.New(\"message write failed: context expired, write result unknown, not waiting for it\")\n\t}\n}", "func (t *testObject) WritePacket(_ *stack.Route, pkt stack.PacketBufferPtr) tcpip.Error {\n\tvar prot tcpip.TransportProtocolNumber\n\tvar srcAddr tcpip.Address\n\tvar dstAddr tcpip.Address\n\n\tif t.v4 {\n\t\th := header.IPv4(pkt.NetworkHeader().Slice())\n\t\tprot = tcpip.TransportProtocolNumber(h.Protocol())\n\t\tsrcAddr = h.SourceAddress()\n\t\tdstAddr = h.DestinationAddress()\n\n\t} else {\n\t\th := header.IPv6(pkt.NetworkHeader().Slice())\n\t\tprot = tcpip.TransportProtocolNumber(h.NextHeader())\n\t\tsrcAddr = h.SourceAddress()\n\t\tdstAddr = h.DestinationAddress()\n\t}\n\tt.checkValues(prot, pkt.Data().AsRange().ToSlice(), srcAddr, dstAddr)\n\treturn nil\n}", "func (_Eth *EthTransactorSession) Publish(a common.Address, topic string) (*types.Transaction, error) {\n\treturn _Eth.Contract.Publish(&_Eth.TransactOpts, a, topic)\n}", "func (msg MsgCreateDeployment) Route() string { return RouterKey }", "func (k *Kafka) Publish(ctx context.Context, topic string, value []byte) error {\n\tspan, ctx := tracer.StartSpanFromContext(ctx, \"Publish\")\n\tdefer span.Finish()\n\n\tif err := k.GetWriterByTopic(topic).WriteMessages(ctx, skafka.Message{\n\t\tValue: value,\n\t}); err != nil {\n\t\tspan.SetTag(\"error\", err)\n\t\treturn errors.Wrap(err, \"unable to publish message(s)\")\n\t}\n\n\treturn nil\n}", "func (k *Kafka) Write(key, value []byte) error {\n\tmsg := skafka.Message{\n\t\tKey: key,\n\t\tValue: value,\n\t}\n\n\theaders := make([]skafka.Header, 0)\n\n\tfor headerName, headerValue := range k.Options.Kafka.WriteHeader {\n\t\theaders = append(headers, skafka.Header{\n\t\t\tKey: headerName,\n\t\t\tValue: []byte(headerValue),\n\t\t})\n\t}\n\n\tif len(headers) != 0 {\n\t\tmsg.Headers = headers\n\t}\n\n\tif err := k.Writer.WriteMessages(context.Background(), msg); err != nil {\n\t\treturn errors.Wrap(err, \"unable to publish message(s)\")\n\t}\n\n\tk.log.Infof(\"Successfully wrote message to topic '%s'\", k.Options.Kafka.Topics[0])\n\n\treturn nil\n}", "func publishData(data string, subChannel string) {\n\n\tif key := getMqttKey(); key != \"\" {\n\n\t\tclientWriting.OnError(func(_ *emitter.Client, err emitter.Error) {\n\t\t\tmqttHasWritingPermission = false\n\t\t})\n\n\t\tchannel, data := getMqttChannelPrefix()+subChannel, data\n\t\tclientWriting.Publish(key, channel, data)\n\n\t} else {\n\t\tlog.Println(\"MQTT key not set!!! Not publishing any data...\")\n\t\tmqttKeyStatusCh <- \"Chave do MQTT: Ausente\"\n\t\treturn\n\t}\n}", "func (r *stepRouter) routeOut(ctx context.Context) (int, error) {\n\n\tstepLabel := r.bundle.TestStepLabel\n\tlog := logging.AddField(r.log, \"step\", stepLabel)\n\tlog = logging.AddField(log, \"phase\", \"routeOut\")\n\n\ttargetWriter := newTargetWriter(log, r.timeouts)\n\n\tvar err error\n\n\tlog.Debugf(\"initializing routeOut for %s\", stepLabel)\n\t// `egressTarget` is used to keep track of egress times of a target from a test step\n\tegressTarget := make(map[string]time.Time)\n\n\tfor {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\terr = fmt.Errorf(\"termination requested for routing into %s\", r.bundle.TestStepLabel)\n\t\tcase t, chanIsOpen := <-r.routingChannels.stepOut:\n\t\t\tif !chanIsOpen {\n\t\t\t\tlog.Debugf(\"step output closed\")\n\t\t\t\tr.routingChannels.stepOut = nil\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tif _, targetPresent := egressTarget[t.ID]; targetPresent {\n\t\t\t\terr = fmt.Errorf(\"step %s returned target %+v multiple times\", r.bundle.TestStepLabel, t)\n\t\t\t\tbreak\n\t\t\t}\n\t\t\t// Emit an event signaling that the target has left the TestStep\n\t\t\tif err := r.emitOutEvent(t, nil); err != nil {\n\t\t\t\tlog.Warningf(\"could not emit out event for target %v: %v\", *t, err)\n\t\t\t}\n\t\t\t// Register egress time and forward target to the next routing block\n\t\t\tegressTarget[t.ID] = time.Now()\n\t\t\tif err := targetWriter.writeTimeout(ctx, r.routingChannels.routeOut, t, r.timeouts.MessageTimeout); err != nil {\n\t\t\t\tlog.Panicf(\"could not forward target to the test runner: %+v\", err)\n\t\t\t}\n\t\tcase targetError, chanIsOpen := <-r.routingChannels.stepErr:\n\t\t\tif !chanIsOpen {\n\t\t\t\tlog.Debugf(\"step error closed\")\n\t\t\t\tr.routingChannels.stepErr = nil\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tif _, targetPresent := egressTarget[targetError.Target.ID]; targetPresent {\n\t\t\t\terr = fmt.Errorf(\"step %s returned target %+v multiple times\", r.bundle.TestStepLabel, targetError.Target)\n\t\t\t} else {\n\t\t\t\tif err := r.emitOutEvent(targetError.Target, targetError.Err); err != nil {\n\t\t\t\t\tlog.Warningf(\"could not emit err event for target: %v\", *targetError.Target)\n\t\t\t\t}\n\t\t\t\tegressTarget[targetError.Target.ID] = time.Now()\n\t\t\t\tif err := targetWriter.writeTargetError(ctx, r.routingChannels.targetErr, targetError, r.timeouts.MessageTimeout); err != nil {\n\t\t\t\t\tlog.Panicf(\"could not forward target (%+v) to the test runner: %v\", targetError.Target, err)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t\tif r.routingChannels.stepErr == nil && r.routingChannels.stepOut == nil {\n\t\t\tlog.Debugf(\"output and error channel from step are closed, routeOut should terminate\")\n\t\t\tclose(r.routingChannels.routeOut)\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif err != nil {\n\t\tlog.Debugf(\"routeOut failed: %v\", err)\n\t\treturn 0, err\n\t}\n\treturn len(egressTarget), nil\n\n}", "func (r *Router) Route(m *paho.Publish) {\n\tvar handler HandlerFunc\n\ttopic := string(m.Topic)\n\tres := r.trie.Match(topic)\n\n\tif res.Node == nil {\n\t\thandler = r.DefaultHandler\n\t} else {\n\t\thandler = res.Node.GetHandler()\n\t}\n\n\thandler(m)\n}", "func PublishHandler(topic string, payload string) {\n\tlog.Println(topic, \"-\", payload)\n}", "func (l *Listener) Publish(line []byte, reply *string) error {\n\n\tvar m Message\n\tvar topic Topic\n\n\terr := json.Unmarshal(line, &m)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tm.Visible = true //imposto la visibilità di default a true\n\n\tfmt.Printf(\"Command publish : %+v\\n\", m) //stampa il messaggio che ha ricevuto\n\n\tqueue = insert(*queue, &m) //inserimento\n\n\ttopic.topic = m.Topic\n\n\tflag := 0\n\n\t//Controllo che il topic non sia già presente\n\tfor i := 0; i < len(topics); i++ {\n\t\tif topics[i].topic == topic.topic {\n\t\t\tflag = 1\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif flag == 0 {\n\t\ttopics = append(topics, topic)\n\t}\n\n\t*reply = \"ACK\"\n\n\t//richiama le funzioni che si occupano di inviare i messaggi ai subscriber in base alla semantica\n\tswitch semantic {\n\n\tcase 1:\n\t\tgo send_mess_1()\n\tcase 2:\n\t\tgo send_mess_2()\n\n\t}\n\n\treturn nil\n}", "func (a *agent) Publish(ctx context.Context, job interface{}, b backoff.BackOff) error {\n\tpayload, err := json.Marshal(job)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn backoff.Retry(func() error {\n\t\ta.mu.Lock()\n\t\tdefer a.mu.Unlock()\n\t\treturn a.ch.Publish(\n\t\t\t\"\", // exchange\n\t\t\ta.queueName, // routing key\n\t\t\tfalse, // mandatory\n\t\t\tfalse,\n\t\t\tamqp.Publishing{\n\t\t\t\tMessageId: uuid.New().String(),\n\t\t\t\tDeliveryMode: amqp.Persistent,\n\t\t\t\tContentType: \"text/plain\",\n\t\t\t\tBody: payload,\n\t\t\t})\n\t}, backoff.WithContext(b, a.ctx))\n}", "func (b *BootstrapClient) writeTransitPolicy() error {\n\treturn b.usingVaultRootToken(func() error {\n\t\tfor i := range b.config.transitData {\n\t\t\tif err := b.writeWritePolicy(b.config.transitData[i].EncryptPath); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif err := b.writeWritePolicy(b.config.transitData[i].DecryptPath); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif err := b.writeReadPolicy(b.config.transitData[i].OutputPath); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t})\n}", "func (t *Topic) Publish(msg interface{}) {\n\tfor ch, ctx := range t.subs {\n\t\tt.pubs <- Work{\n\t\t\tctx: ctx,\n\t\t\tch: ch,\n\t\t\tdata: msg,\n\t\t}\n\t}\n}", "func WriteHandler(ctx *routing.Context) error {\n\n\t//Get the Path values\n\tindexPath := strings.Replace(string(ctx.Path()), \"/write/\", \"\", -1)\n\tif indexPath == \"\" {\n\t\tindexPath = \"default\"\n\t}\n\ttmp := strings.SplitAfter(indexPath, \"/\")\n\tindexHead := strings.TrimRight(tmp[0], \"/\")\n\tindexPath = strings.TrimRight(indexPath, \"/\")\n\n\t//Ensure pathing is proper\n\tisValidPathChar := regexp.MustCompile(`^[A-Za-z0-9\\/\\.\\-\\_]+$`).MatchString\n\tfor _, pathchar := range []string{indexPath} {\n\t\tif !isValidPathChar(pathchar) {\n\t\t\tctx.Error(\"invalid path detected\", 400)\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t//Get the Header, validate type, push to Ingestion Enging\n\trawHeaders := string(ctx.Request.Header.Peek(\"Content-Type\"))\n\tif rawHeaders == \"application/x-ndjson\" || rawHeaders == \"application/json\" {\n\t\t//Send the data onward to the Ingestion Engine for indexing\n\t\tdatabase.IngestionEngine(database.LogPacket{\n\t\t\tTimeAtIndex: time.Now().Unix(),\n\t\t\tIndexHead: indexHead,\n\t\t\tIndexPath: indexPath,\n\t\t\tDataBlob: ctx.PostBody(),\n\t\t\tDataType: rawHeaders})\n\t} else {\n\t\tctx.Error(\"invalid headers detected\", 415)\n\t}\n\n\treturn nil\n}", "func (t *Topic) Publish(msg interface{}) {\n\tfor i, f := range t.subs {\n\t\tif *f == nil {\n\t\t\tt.subs = append(t.subs[:i], t.subs[i+1:]...)\n\t\t\tcontinue\n\t\t}\n\t\tcb := *f\n\t\tcb(msg)\n\t}\n}", "func (c *Client) WriteTo(p *Packet, addr net.HardwareAddr) error {\n\tpb, err := p.MarshalBinary()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tf := &ethernet.Frame{\n\t\tDestination: addr,\n\t\tSource: p.SenderHardwareAddr,\n\t\tEtherType: ethernet.EtherTypeARP,\n\t\tPayload: pb,\n\t}\n\n\tfb, err := f.MarshalBinary()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = c.p.WriteTo(fb, &packet.Addr{HardwareAddr: addr})\n\treturn err\n}", "func publish(job IJob, msg message, d TTL) error {\n\t// 创建MQ连接\n\tmq, err := Open(*job.Config())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// 声明队列\n\tif err := declareJob(job, mq); err != nil {\n\t\treturn err\n\t}\n\n\treturn (&postman{\n\t\tmq: mq,\n\t\tjob: job,\n\t\tmsg: msg,\n\t\tttl: d,\n\t}).send()\n}", "func emailWrite(splited []string, length int) {\n if length != 3 { return }\n var new_pack = settings.PackageTCP {\n From: models.From {\n Name: settings.User.Hash,\n },\n To: splited[2],\n Head: models.Head {\n Header: settings.HEAD_EMAIL,\n Mode: settings.MODE_SAVE,\n }, \n Body: \n set_email.title + settings.SEPARATOR +\n set_email.body + settings.SEPARATOR +\n time.Now().Format(time.RFC850),\n }\n connect.CreateRedirectPackage(&new_pack)\n connect.SendInitRedirectPackage(new_pack)\n}", "func (e *EventBus) Publish(topic string, data string) {\n\tt := topic\n\t// Some of our topics for the socket support passing a more specific namespace,\n\t// such as \"backup completed:1234\" to indicate which specific backup was completed.\n\t//\n\t// In these cases, we still need to send the event using the standard listener\n\t// name of \"backup completed\".\n\tif strings.Contains(topic, \":\") {\n\t\tparts := strings.SplitN(topic, \":\", 2)\n\n\t\tif len(parts) == 2 {\n\t\t\tt = parts[0]\n\t\t}\n\t}\n\n\te.mu.RLock()\n\tdefer e.mu.RUnlock()\n\n\t// Acquire a read lock and loop over all the channels registered for the topic. This\n\t// avoids a panic crash if the process tries to unregister the channel while this routine\n\t// is running.\n\tif cp, ok := e.pools[t]; ok {\n\t\tfor _, callback := range cp.callbacks {\n\t\t\tc := *callback\n\t\t\tevt := Event{Data: data, Topic: topic}\n\t\t\t// Using the workerpool with one worker allows us to execute events in a FIFO manner. Running\n\t\t\t// this using goroutines would cause things such as console output to just output in random order\n\t\t\t// if more than one event is fired at the same time.\n\t\t\t//\n\t\t\t// However, the pool submission does not block the execution of this function itself, allowing\n\t\t\t// us to call publish without blocking any of the other pathways.\n\t\t\t//\n\t\t\t// @see https://github.com/pterodactyl/panel/issues/2303\n\t\t\tcp.pool.Submit(func() {\n\t\t\t\tc(evt)\n\t\t\t})\n\t\t}\n\t}\n}", "func (_Eth *EthSession) Publish(a common.Address, topic string) (*types.Transaction, error) {\n\treturn _Eth.Contract.Publish(&_Eth.TransactOpts, a, topic)\n}", "func (zj *ZapJournal) Topic(topic string) Writer {\n\treturn &ZapWriter{\n\t\tlogger: zj.logger.Sugar().Named(topic),\n\t\ttopic: topic,\n\t}\n}", "func (msg MsgSend) Route() string { return RouterKey }", "func writer(g *Gossiper, udpConn net.UDPConn, queue chan *Packet) {\n\t// writing loop\n\t// write every message on queue\n\tfor pkt := range queue {\n\t\tdestination := pkt.Destination\n\t\tgossipPacket := pkt.GossipPacket\n\n\t\tbuf, err := protobuf.Encode(&gossipPacket)\n\t\tcommon.CheckRead(err)\n\t\t_, err = udpConn.WriteToUDP(buf, &destination)\n\t\tcommon.CheckRead(err)\n\t}\n}", "func (e *TarantoolEngine) publish(chID ChannelID, message []byte) error {\n\t// Process service messages\n\tif chID != e.app.config.ControlChannel && chID != e.app.config.AdminChannel {\n\t\tnewMessage, err := e.processMessage(chID, message)\n\t\tif err != nil {\n\t\t\treturn err // if no need further processing\n\t\t}\n\t\tmessage = newMessage\n\t}\n\t// All other messages\n\treturn e.app.handleMsg(chID, message)\n}", "func (p *pahoClient) Publish(c chan error, topic string, qos uint8, payload interface{}) {\n\ttoken := p.client.Publish(topic, qos, true, payload)\n\tc <- p.waitForToken(token)\n}", "func (msg MsgSaveProfile) Route() string { return RouterKey }", "func (msg MsgSaveProfile) Route() string { return RouterKey }", "func (o *GetTransportByIDOK) WriteResponse(rw http.ResponseWriter, producer runtime.Producer) {\n\n\trw.WriteHeader(200)\n\tif o.Payload != nil {\n\t\tpayload := o.Payload\n\t\tif err := producer.Produce(rw, payload); err != nil {\n\t\t\tpanic(err) // let the recovery middleware deal with this\n\t\t}\n\t}\n}", "func (p Producer) PublishToJSON(brokerAddress string, topic string, data interface{}) error {\n\tjsonMsg, err := json.Marshal(data)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\tif err := p.Publish(brokerAddress, topic, string(jsonMsg)); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (fs *FS) writePost(id int64, parent int64, msg Message) error {\n\tvar network bytes.Buffer\n\tenc := gob.NewEncoder(&network)\n\terr := enc.Encode(msg)\n\tif err != nil { return err }\n\n\tpath := fmt.Sprintf(\"%s/posts/%d\", fs.path, id)\n\terr = os.MkdirAll(filepath.Dir(path), 0777)\n\tif err != err { return err }\n\n\tf, err := os.Create(path)\n\tdefer f.Close()\n\tif err != nil { return err }\n\n\t_, err = f.Write(network.Bytes())\n\tif err != nil { return err }\n\n\treturn nil\n}", "func (o *CreateRoutingInstanceUsingPOSTParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// header param Authorization\n\tif err := r.SetHeaderParam(\"Authorization\", o.Authorization); err != nil {\n\t\treturn err\n\t}\n\n\tif o.Request != nil {\n\t\tif err := r.SetBodyParam(o.Request); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func Write(laddr, raddr *net.UDPAddr, message []byte, redundancyFactor uint8) error {\n\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tlog.WithField(\"err\", r).Warn(\"Send failed\")\n\t\t}\n\t}()\n\n\t// Send from same IP that the UDP listener is bound on but choose random port\n\tladdr.Port = 0\n\tconn, err := net.DialUDP(\"udp4\", laddr, raddr)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err = conn.SetWriteBuffer(writeBufferSize); err != nil {\n\t\tlog.WithError(err).Traceln(\"SetWriteBuffer socket problem\")\n\t}\n\n\terr = sendRaptorRFC5053(conn, message, redundancyFactor)\n\t_ = conn.Close()\n\treturn err\n}", "func (n *TaskResponse) Publish(payload []byte) {\n\terr := n.app.ch.Publish(\n\t\tn.app.exchange, // exchange\n\t\tn.replyTo, // routing key\n\t\tfalse, // mandatory\n\t\tfalse, // immediate\n\t\tamqp.Publishing{\n\t\t\tHeaders: amqp.Table{\n\t\t\t\t\"name\": n.app.queueName,\n\t\t\t},\n\t\t\tContentType: \"application/gob\",\n\t\t\tCorrelationId: n.corrID,\n\t\t\tBody: payload,\n\t\t})\n\tif err != nil {\n\t\tlog.Println(\"Failed to send message to RabbitMQ. Error: \", err)\n\t}\n}", "func Publish(t *testing.T, tp *Tapestry) {\n\ttp.Lock()\n\tnode, _, err := tp.RandNode()\n\tif err == nil {\n\t\tkey := RandString(8)\n\t\tdata := make([]byte, DATASIZE)\n\t\ttp.Rand.Read(data) //fills data with random bytes\n\t\terr := node.Store(key, data) //publish data\n\t\tif err != nil {\n\t\t\tt.Errorf(\"Error %v while publishing %v to node %v\", err, key, node.node.Id.String())\n\t\t} else {\n\t\t\tWaitUntilPublished(t, key, node, tp.Nodes)\n\t\t\tpublished := IsPublished(t, key, node, tp.Nodes)\n\t\t\tif published {\n\t\t\t\tOut.Printf(\"Key %v published to node %v\", key, node.node.Id.String())\n\t\t\t\t//if succesful, update tp\n\t\t\t\ttp.Keys = append(tp.Keys, key)\n\t\t\t\ttp.Blobs[key] = data\n\t\t\t} else {\n\t\t\t\tt.Errorf(\"Key %v was not actually published to node %v\", key, node.node.Id)\n\t\t\t}\n\t\t}\n\t}\n\ttp.Unlock()\n}", "func (s *Service) Forward(topicID, msgID, targetAddr, text string) error {\n\tpayload := map[string]interface{}{\n\t\t\"id\": msgID,\n\t\t\"address\": targetAddr,\n\t\t\"text\": text,\n\t}\n\tb, err := json.Marshal(payload)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tpf, err := s.client.Publish(topicID, b, 2, false)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn pf.Wait(5 * time.Second)\n}", "func (h *Hub) Publish(topic string, data interface{}) {\n\th.publish <- Value{topic, data}\n}", "func DestinationWriter(p *pool.Pool, dest pool.Destination) {\n\n\t// Get initial connection.\n\tp.Register(dest)\n\tconn, err := establishConn(p, dest)\n\tif err != nil {\n\t\treturn\n\t}\n\tdefer conn.Close()\n\n\t// Dequeue from destination outbound queue\n\t// and send.\n\tn := 1\n\tfor {\n\t\t// Exponential backoff var\n\t\t// if the channel is empty.\n\t\tif n < 1000 {\n\t\t\tn = n * 2\n\t\t}\n\n\t\t// Need to make sure the connection\n\t\t// exists. It's possible that it becomes\n\t\t// unregistered (therefore doesn't exist) between\n\t\t// checking if it exists and attempting to read from it.\n\t\tp.Lock()\n\t\t_, ok := p.Conns[dest.Name]\n\t\tif !ok {\n\t\t\tp.Unlock()\n\t\t\treturn\n\t\t}\n\n\t\t// Have to do a non-blocking read attempt, otherwise\n\t\t// unlocking the mutex will be blocked.\n\t\tselect {\n\t\tcase m, ok := <-p.Conns[dest.Name]:\n\t\t\tp.Unlock()\n\n\t\t\tif !ok {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t_, err := fmt.Fprintln(conn, *m)\n\t\t\t// If we fail to send, reload the message into the\n\t\t\t// queue and attempt to reconnect.\n\t\t\tif err != nil {\n\t\t\t\tp.Conns[dest.Name] <- m\n\t\t\t\tlog.Printf(\"Destination %s error: %s\\n\", dest.Name, err)\n\n\t\t\t\t// Wait on a connection. If the destination isn't\n\t\t\t\t// registered, err and close this writer.\n\t\t\t\tnewConn, err := establishConn(p, dest)\n\t\t\t\tif err != nil {\n\t\t\t\t\tbreak\n\t\t\t\t} else {\n\t\t\t\t\tconn = newConn\n\t\t\t\t}\n\t\t\t}\n\t\t\t// Reset backoff var.\n\t\t\tn = 1\n\t\tdefault:\n\t\t\tp.Unlock()\n\t\t\ttime.Sleep(time.Duration(n) * time.Millisecond)\n\t\t\tcontinue\n\t\t}\n\n\t}\n}", "func (h *handler) Publish(c *session.Client, topic *string, payload *[]byte) {\n\tif c == nil {\n\t\th.logger.Error(LogErrFailedPublish + ErrClientNotInitialized.Error())\n\t\treturn\n\t}\n\th.logger.Info(fmt.Sprintf(LogInfoPublished, c.ID, *topic))\n\t// Topics are in the format:\n\t// channels/<channel_id>/messages/<subtopic>/.../ct/<content_type>\n\n\tchannelParts := channelRegExp.FindStringSubmatch(*topic)\n\tif len(channelParts) < 2 {\n\t\th.logger.Error(LogErrFailedPublish + (ErrMalformedTopic).Error())\n\t\treturn\n\t}\n\n\tchanID := channelParts[1]\n\tsubtopic := channelParts[2]\n\n\tsubtopic, err := parseSubtopic(subtopic)\n\tif err != nil {\n\t\th.logger.Error(logErrFailedParseSubtopic + err.Error())\n\t\treturn\n\t}\n\n\tmsg := messaging.Message{\n\t\tProtocol: protocol,\n\t\tChannel: chanID,\n\t\tSubtopic: subtopic,\n\t\tPublisher: c.Username,\n\t\tPayload: *payload,\n\t\tCreated: time.Now().UnixNano(),\n\t}\n\n\tfor _, pub := range h.publishers {\n\t\tif err := pub.Publish(msg.Channel, msg); err != nil {\n\t\t\th.logger.Error(LogErrFailedPublishToMsgBroker + err.Error())\n\t\t}\n\t}\n}", "func (ht Transport) Write() base.HeaderValue {\n\tvar vals []string\n\n\tif ht.Protocol == base.StreamProtocolUDP {\n\t\tvals = append(vals, \"RTP/AVP\")\n\t} else {\n\t\tvals = append(vals, \"RTP/AVP/TCP\")\n\t}\n\n\tif ht.Delivery != nil {\n\t\tif *ht.Delivery == base.StreamDeliveryUnicast {\n\t\t\tvals = append(vals, \"unicast\")\n\t\t} else {\n\t\t\tvals = append(vals, \"multicast\")\n\t\t}\n\t}\n\n\tif ht.ClientPorts != nil {\n\t\tports := *ht.ClientPorts\n\t\tvals = append(vals, \"client_port=\"+strconv.FormatInt(int64(ports[0]), 10)+\"-\"+strconv.FormatInt(int64(ports[1]), 10))\n\t}\n\n\tif ht.ServerPorts != nil {\n\t\tports := *ht.ServerPorts\n\t\tvals = append(vals, \"server_port=\"+strconv.FormatInt(int64(ports[0]), 10)+\"-\"+strconv.FormatInt(int64(ports[1]), 10))\n\t}\n\n\tif ht.InterleavedIds != nil {\n\t\tports := *ht.InterleavedIds\n\t\tvals = append(vals, \"interleaved=\"+strconv.FormatInt(int64(ports[0]), 10)+\"-\"+strconv.FormatInt(int64(ports[1]), 10))\n\t}\n\n\tif ht.Mode != nil {\n\t\tif *ht.Mode == TransportModePlay {\n\t\t\tvals = append(vals, \"mode=play\")\n\t\t} else {\n\t\t\tvals = append(vals, \"mode=record\")\n\t\t}\n\t}\n\n\treturn base.HeaderValue{strings.Join(vals, \";\")}\n}", "func (p *Publisher) PublishWithTracing(ctx context.Context, exchange, key, corID string, body []byte) (err error) {\n\tspan, _ := opentracing.StartSpanFromContext(ctx, `publish_key: `+key)\n\tdefer span.Finish()\n\n\text.SpanKind.Set(span, ext.SpanKindProducerEnum)\n\tspan.SetTag(\"key\", key)\n\tspan.SetTag(\"exchange\", exchange)\n\tspan.SetTag(\"corID\", corID)\n\n\tmsg := amqp.Publishing{\n\t\tHeaders: amqp.Table{},\n\t\tContentType: \"application/json\",\n\t\tCorrelationId: corID,\n\t\tBody: body,\n\t\tDeliveryMode: amqp.Persistent,\n\t}\n\n\t// Inject the span context into the AMQP header.\n\tif err := amqptracer.Inject(span, msg.Headers); err != nil {\n\t\tlog.Printf(\"publish: error inject headers: %s\", err)\n\t}\n\n\treturn p.ch.Publish(exchange, key, false, false, msg)\n}", "func (this *BaseRouter) FlashWrite(key string, value string) {\n\tfmt.Println(\"FlashWrite() function\")\n\tflash := beego.NewFlash()\n\tflash.Data[key] = value\n\tflash.Store(&this.Controller)\n}", "func PublishTo(bus *nats.Conn, subject string, msg interface{}) error {\r\n\tdata, err := json.Marshal(&msg)\r\n\tif nil != err {\r\n\t\treturn err\r\n\t}\r\n\r\n\treturn bus.Publish(subject, data)\r\n}", "func (bew *KafkaBEWriter) Write(m Message) error {\n\t// TODO - change the hard-coding of json.Marshal() to a\n\t// strategy pattern.\n\t// This would allow to inject a strategy to write messages.\n\t// For instance, a JSONWriterStrategy or a MapWriterStrategy (which\n\t// would return a map) or an XMLWriterStrategy, etc.\n\tjsonStr, err := json.Marshal(m)\n\tif err != nil {\n\t\treturn ErrUnableToTranscodeMessage\n\t}\n\n\tmsg := &sarama.ProducerMessage{Topic: bew.topic, Value: sarama.StringEncoder(jsonStr)}\n\t_, _, err = bew.producer.SendMessage(msg)\n\tif err != nil {\n\t\treturn ErrUnableToSendMessage\n\t}\n\n\treturn nil\n}", "func (coll *Collector) Write(msg string, extra map[string]interface{}) (err error) {\n\n\tm := gelf.Message{\n\t\tVersion: \"1.1\",\n\t\tHost: coll.host,\n\t\tShort: msg,\n\t\tTimeUnix: float64(time.Now().Unix()),\n\t\tLevel: 6, // info always\n\t\tFacility: \"drone\",\n\t\tExtra: extra,\n\t}\n\n\tif err = coll.writer.WriteMessage(&m); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n\n}", "func (r *RPCTractserverTalker) Write(ctx context.Context, addr string, id core.TractID, version int, b []byte, off int64) core.Error {\n\tpri := priorityFromContext(ctx)\n\trpcid := rpc.GenID()\n\treq := core.WriteReq{ID: id, Version: version, Off: off, Pri: pri, ReqID: rpcid}\n\treq.Set(b, false)\n\tvar reply core.Error\n\tcancel := rpc.CancelAction{Method: core.CancelReqMethod, Req: rpcid}\n\tif err := r.cc.SendWithCancel(ctx, addr, core.WriteMethod, &req, &reply, &cancel); err != nil {\n\t\tlog.Errorf(\"Write RPC error for tract (id: %s, version: %d, offset: %d) on tractserver @%s: %s\", id, version, off, addr, err)\n\t\treturn core.ErrRPC\n\t}\n\tif reply != core.NoError {\n\t\tlog.Errorf(\"Write error for tract (id: %s, version: %d, offset: %d) on tractserver @%s: %s\", id, version, off, addr, reply)\n\t}\n\treturn reply\n}", "func (c *Connection) Publish(payload []byte) error {\n\terr := c.Channel.Publish(\n\t\tc.Config.Exchange,\n\t\tc.Config.RoutingKey,\n\t\tc.Config.Options.Publish.Mandatory,\n\t\tc.Config.Options.Publish.Immediate,\n\t\tamqp.Publishing{\n\t\t\tDeliveryMode: amqp.Persistent,\n\t\t\tContentType: \"text/plain\",\n\t\t\tBody: payload,\n\t\t\tHeaders: c.Headers,\n\t\t})\n\n\treturn err\n}", "func (msg MsgCreateEthBridgeClaim) Route() string { return RouterKey }", "func (msg MsgInsertRow) Route() string { return RouterKey }", "func (c *client) write(mt int, message []byte) error {\n\tc.ws.SetWriteDeadline(time.Now().Add(writeWait))\n\treturn c.ws.WriteMessage(mt, message)\n}", "func (_e *MockTestTransportInstance_Expecter) Write(data interface{}) *MockTestTransportInstance_Write_Call {\n\treturn &MockTestTransportInstance_Write_Call{Call: _e.mock.On(\"Write\", data)}\n}", "func (c *Connection) Write(mt int, payload []byte) error {\n\tc.Conn.SetWriteDeadline(time.Now().Add(WriteWait))\n\treturn c.Conn.WriteMessage(mt, payload)\n}", "func (p *Protocol) WritePacket(t PacketType, body interface{}) error {\n\tpay, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn err\n\t}\n\t//fmt.Println(p.pks(), \"WRITE:\", t, string(pay))\n\traw := make([]byte, 3+len(pay))\n\traw[0] = byte(t)\n\tbinary.BigEndian.PutUint16(raw[1:3], uint16(len(pay)))\n\tcopy(raw[3:], pay)\n\t_, err = p.rwc.Write(raw)\n\treturn err\n}" ]
[ "0.5794445", "0.55988747", "0.5263754", "0.52281415", "0.5177151", "0.50187415", "0.4988992", "0.4955801", "0.49440837", "0.49342078", "0.49267042", "0.49209622", "0.49163938", "0.49163938", "0.4913341", "0.49116528", "0.48997316", "0.4893645", "0.48907122", "0.48559427", "0.48252437", "0.48037475", "0.47929296", "0.47908399", "0.4780583", "0.47755685", "0.47646064", "0.47637722", "0.4756794", "0.472737", "0.46917942", "0.46906662", "0.4683675", "0.46824342", "0.46794882", "0.4672908", "0.46408764", "0.46328294", "0.46237886", "0.4619904", "0.4618582", "0.46046132", "0.46027714", "0.46026966", "0.45996028", "0.45949358", "0.4594401", "0.4594309", "0.4587804", "0.4584031", "0.45800248", "0.45789903", "0.4578499", "0.45505834", "0.4548441", "0.4543476", "0.45425743", "0.45337027", "0.45292628", "0.4527815", "0.45202798", "0.4519993", "0.45146802", "0.45144156", "0.45078236", "0.45066902", "0.45051268", "0.45008335", "0.44976798", "0.44976193", "0.44971937", "0.449164", "0.44866896", "0.44806844", "0.44806844", "0.44719172", "0.44667464", "0.4465588", "0.44651848", "0.44646916", "0.4464163", "0.44592467", "0.44560882", "0.44510555", "0.44483334", "0.44447953", "0.44440287", "0.44294062", "0.44259763", "0.44245625", "0.44234812", "0.44190645", "0.44177786", "0.44087055", "0.44008222", "0.43949917", "0.43888265", "0.4382913", "0.4381804", "0.43794414" ]
0.6585562
0
responsible for (1) discovering if we have seen this topic before, otherwise fetching it
func (n *Node) getPartitionsFor(topic string) []partition { parts, hasParts := n.partitions[topic] if hasParts { return parts } // we haven't seen this before fetch it log.Info("The topic %s is new -- creating it locally", topic) // TODO fetch!!! - for now, we will just make it all local parts = make([]partition, n.config.DefaultPartitionCount) for i := range parts { partName := fmt.Sprintf("%s_%d.jsonl", topic, i) partName = s.ToLower(s.TrimSpace(s.Replace(partName, " ", "_", -1))) lp, err := newLocalPartition(filepath.Join(n.config.DataDirectory, partName), topic) if err != nil { log.Err("Failed to create partition %d for topic %s", i, topic, err) } log.Debug("Created local partition topic %s, %s", topic, partName) parts[i] = lp } n.partitions[topic] = parts return parts }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (ts *TargetSyncer) fetchTopics() ([]string, bool, error) {\n\tnew, err := ts.topicManager.Topics()\n\tif err != nil {\n\t\treturn nil, false, err\n\t}\n\tif len(ts.previousTopics) != len(new) {\n\t\tts.previousTopics = new\n\t\treturn new, true, nil\n\t}\n\tfor i, v := range ts.previousTopics {\n\t\tif v != new[i] {\n\t\t\tts.previousTopics = new\n\t\t\treturn new, true, nil\n\t\t}\n\t}\n\treturn nil, false, nil\n}", "func hasTopic(kz kzoo, topic string) (bool, error) {\n\ttopics, err := kz.Topics()\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tfor _, t := range topics {\n\t\tif t.Name == topic {\n\t\t\treturn true, nil\n\t\t}\n\t}\n\treturn false, nil\n}", "func fetchTopics(topic string) (*github.TopicsSearchResult, error) {\n\tclient := github.NewClient(nil)\n\ttopics, _, err := client.Search.Topics(context.Background(), topic, nil)\n\treturn topics, err\n}", "func (c *Consumer) Topic() string { return c.topic }", "func (m *Medium) Topic(name string) (Topic, bool) {\n\tm.mx.RLock()\n\tdefer m.mx.RUnlock()\n\tt, ok := m.topics[name]\n\treturn t, ok\n}", "func (h *hub) findTopic(topic []string) (found *topicNode, err error) {\n\tfound, rest, _ := h.root.MaybeFindTopic(topic)\n\n\tif len(rest) != 0 {\n\t\treturn nil, errors.New(\n\t\t\tfmt.Sprintf(\"Topic not found: %s\",\n\t\t\t\tstrings.Join(topic, topicDelimeter)))\n\t}\n\n\treturn found, nil\n}", "func topicRespondsToMetadata(t *Topic, client connection.Client) int {\n\tcontroller, err := client.Controller()\n\tif err != nil {\n\t\tlog.Error(\"Failed to get controller from client: %s\", err)\n\t}\n\n\t// Attempt to collect metadata and determine whether it errors out\n\t_, err = controller.GetMetadata(&sarama.MetadataRequest{Version: 0, Topics: []string{t.Name}, AllowAutoTopicCreation: false})\n\tif err != nil {\n\t\treturn 0\n\t}\n\n\treturn 1\n}", "func GetTopic(name string) *Topic {\n\treturn registry.GetTopic(name)\n}", "func (r *viewResolver) Topic(\n\tctx context.Context, view *models.View, topicID string,\n) (*models.TopicValue, error) {\n\tlog.Printf(\"Fetching topic %s\", topicID)\n\tscope := models.Topics(topicQueryMods(view, qm.Where(\"topics.id = ?\", topicID), nil, nil)...)\n\ttopic, err := scope.One(ctx, r.DB)\n\treturn &models.TopicValue{topic, false, view}, err\n}", "func (tm *topicManager) getTopicForName(name string) (log, error) {\n\ttm.topicsMutex.Lock()\n\tdefer tm.topicsMutex.Unlock()\n\n\tlc, ok := tm.topicNamesToLogs[name]\n\tif ok {\n\t\tlc.refCount++\n\t\treturn lc.l, nil\n\t}\n\n\tid, ok := tm.topicNamesToIds[name]\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"No topic with name [%s]\", name)\n\t}\n\n\treturn tm.createNewTopic(name, id)\n}", "func (feed *Feed) Topic() string {\n\tfor _, link := range feed.Link {\n\t\tif link.Rel == \"self\" {\n\t\t\treturn link.Href\n\t\t\tbreak\n\t\t}\n\t}\n\treturn \"\"\n}", "func (cl *Client) waitUnknownTopic(\n\ttopic string,\n\tunknown *unknownTopicProduces,\n) {\n\tcl.cfg.logger.Log(LogLevelInfo, \"waiting for metadata to produce to unknown topic\", \"topic\", topic)\n\tvar after <-chan time.Time\n\tif timeout := cl.cfg.recordTimeout; timeout > 0 {\n\t\ttimer := time.NewTimer(cl.cfg.recordTimeout)\n\t\tdefer timer.Stop()\n\t\tafter = timer.C\n\t}\n\tvar tries int\n\tvar err error\n\tfor err == nil {\n\t\tselect {\n\t\tcase <-cl.ctx.Done():\n\t\t\terr = errClientClosing\n\t\tcase <-after:\n\t\t\terr = errRecordTimeout\n\t\tcase retriableErr, ok := <-unknown.wait:\n\t\t\tif !ok {\n\t\t\t\tcl.cfg.logger.Log(LogLevelInfo, \"done waiting for unknown topic\", \"topic\", topic)\n\t\t\t\treturn // metadata was successful!\n\t\t\t}\n\t\t\tcl.cfg.logger.Log(LogLevelInfo, \"unknown topic wait failed, retrying wait\", \"topic\", topic, \"err\", retriableErr)\n\t\t\ttries++\n\t\t\tif int64(tries) >= cl.cfg.produceRetries {\n\t\t\t\terr = fmt.Errorf(\"no partitions available after attempting to refresh metadata %d times, last err: %w\", tries, retriableErr)\n\t\t\t}\n\t\t}\n\t}\n\n\t// If we errored above, we come down here to potentially clear the\n\t// topic wait and fail all buffered records. However, under some\n\t// extreme conditions, a quickly following metadata update could delete\n\t// our unknown topic, and then a produce could recreate a new unknown\n\t// topic. We only delete and finish promises if the pointer in the\n\t// unknown topic map is still the same.\n\tp := &cl.producer\n\n\tp.unknownTopicsMu.Lock()\n\tdefer p.unknownTopicsMu.Unlock()\n\n\tnowUnknown := p.unknownTopics[topic]\n\tif nowUnknown != unknown {\n\t\treturn\n\t}\n\tcl.cfg.logger.Log(LogLevelInfo, \"unknown topic wait failed, done retrying, failing all records\", \"topic\", topic)\n\n\tdelete(p.unknownTopics, topic)\n\tcl.failUnknownTopicRecords(topic, unknown, err)\n}", "func (a *AllRepositories) Topic() loaders.TopicRepository {\n\treturn a.topic\n}", "func (v *KILLYD) GetTopic(topicName string) *Topic {\n\t// most likely, we already have this topic, so try read lock first.\n\tv.RLock()\n\tt, ok := v.topicMap[topicName]\n\tv.RUnlock()\n\tif ok {\n\t\treturn t\n\t}\n\n\tv.Lock()\n\tt, ok = v.topicMap[topicName]\n\tif ok {\n\t\tv.Unlock()\n\t\treturn t\n\t}\n\tt = NewTopic(topicName, &context{v})\n\tv.topicMap[topicName] = t\n\tv.logf(LOG_INFO, \"TOPIC(%s): created\", t.name)\n\tv.Unlock()\n\treturn t\n}", "func (o *TopicsToResetOffset) GetTopicOk() (*string, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn &o.Topic, true\n}", "func (r *SourceResolver) Topic(ctx context.Context, source *model.Source) (*model.Topic, error) {\n\tresults, err := r.DataLoaders.SourceLoader(ctx).TopicBySource(source.ID)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to get topic from loader\")\n\t}\n\treturn results, nil\n}", "func (log Log) GetTopic(idx uint) (common.Hash, error) {\n\tif len(log.Topics) <= int(idx) {\n\t\treturn common.Hash{}, fmt.Errorf(\"Log: Unable to get topic #%v for %v\", idx, log)\n\t}\n\n\treturn log.Topics[idx], nil\n}", "func (a *HipchatAdapter) Topic(res *Response, strings ...string) error {\n\tfor _, str := range strings {\n\t\t_ = str\n\t}\n\treturn nil\n}", "func (cl *Client) partitionsForTopicProduce(pr promisedRec) (*topicPartitions, *topicPartitionsData) {\n\tp := &cl.producer\n\ttopic := pr.Topic\n\n\ttopics := p.topics.load()\n\tparts, exists := topics[topic]\n\tif exists {\n\t\tif v := parts.load(); len(v.partitions) > 0 {\n\t\t\treturn parts, v\n\t\t}\n\t}\n\n\tif !exists { // topic did not exist: check again under mu and potentially create it\n\t\tp.topicsMu.Lock()\n\t\tdefer p.topicsMu.Unlock()\n\n\t\tif parts, exists = p.topics.load()[topic]; !exists { // update parts for below\n\t\t\t// Before we store the new topic, we lock unknown\n\t\t\t// topics to prevent a concurrent metadata update\n\t\t\t// seeing our new topic before we are waiting from the\n\t\t\t// addUnknownTopicRecord fn. Otherwise, we would wait\n\t\t\t// and never be re-notified.\n\t\t\tp.unknownTopicsMu.Lock()\n\t\t\tdefer p.unknownTopicsMu.Unlock()\n\n\t\t\tp.topics.storeTopics([]string{topic})\n\t\t\tcl.addUnknownTopicRecord(pr)\n\t\t\tcl.triggerUpdateMetadataNow()\n\t\t\treturn nil, nil\n\t\t}\n\t}\n\n\t// Here, the topic existed, but maybe has not loaded partitions yet. We\n\t// have to lock unknown topics first to ensure ordering just in case a\n\t// load has not happened.\n\tp.unknownTopicsMu.Lock()\n\tdefer p.unknownTopicsMu.Unlock()\n\n\tif v := parts.load(); len(v.partitions) > 0 {\n\t\treturn parts, v\n\t}\n\tcl.addUnknownTopicRecord(pr)\n\tcl.triggerUpdateMetadata(false)\n\n\treturn nil, nil // our record is buffered waiting for metadata update; nothing to return\n}", "func (h *hub) findOrCreateTopic(topic []string) (found *topicNode, err error) {\n\tfound, rest, _ := h.root.MaybeFindTopic(topic)\n\n\tif len(rest) != 0 {\n\t\tfound, err = found.CreateChild(rest)\n\t}\n\n\treturn found, err\n}", "func (p *metadataService) onTopicSubscribe(e event.Event) {\n}", "func (c *KafkaClient) Fetch(topic string, partition int32, offset int64) (*FetchResponse, error) {\n\tresponse, err := c.tryFetch(topic, partition, offset)\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\tif response.Error(topic, partition) == ErrNotLeaderForPartition {\n\t\tlog.Infof(\"Sent a fetch reqest to a non-leader broker. Refleshing metadata for topic %s and retrying the request\", topic)\n\t\terr = c.metadata.Refresh([]string{topic})\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresponse, err = c.tryFetch(topic, partition, offset)\n\t}\n\n\treturn response, err\n}", "func GetTopic(tid string) (*Topic, error) {\n\ttidNum, err := strconv.ParseInt(tid, 10, 64)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\to := orm.NewOrm()\n\n\ttopic := new(Topic)\n\n\tqs := o.QueryTable(\"topic\")\n\terr = qs.Filter(\"id\", tidNum).One(topic)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttopic.Views++\n\t_, err = o.Update(topic)\n\n\treturn topic, err\n}", "func (se *SubscriberPanic) Topic() interface{} {\n\treturn se.topic\n}", "func (v *KILLYD) GetExistingTopic(topicName string) (*Topic, error) {\n\tv.RLock()\n\tdefer v.RUnlock()\n\ttopic, ok := v.topicMap[topicName]\n\tif !ok {\n\t\treturn nil, errors.New(\"topic does not exist\")\n\t}\n\treturn topic, nil\n}", "func GetTopicCache() *TopicCache {\n\treturn topicCache\n}", "func (c *MqClient) QueryTopic(topic string) (*proto.TopicInfo, error) {\n\treq := NewMessage()\n\treq.SetCmd(proto.Query)\n\treq.SetTopic(topic)\n\n\tinfo := &proto.TopicInfo{}\n\terr := c.invokeCmd(req, info)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn info, err\n}", "func GetTopic(name string) (*Topic, error) {\n\tentity := Topic{ID: strings.ToLower(name)}\n\terr := DS.GetByKey(&entity)\n\treturn &entity, err\n}", "func (t *topicTrie) matchTopic(topicSlice []string, rs subscription.ClientSubscriptions) {\n\tendFlag := len(topicSlice) == 1\n\tif cnode := t.children[\"#\"]; cnode != nil {\n\t\tsetRs(cnode, rs)\n\t}\n\tif cnode := t.children[\"+\"]; cnode != nil {\n\t\tif endFlag {\n\t\t\tsetRs(cnode, rs)\n\t\t\tif n := cnode.children[\"#\"]; n != nil {\n\t\t\t\tsetRs(n, rs)\n\t\t\t}\n\t\t} else {\n\t\t\tcnode.matchTopic(topicSlice[1:], rs)\n\t\t}\n\t}\n\tif cnode := t.children[topicSlice[0]]; cnode != nil {\n\t\tif endFlag {\n\t\t\tsetRs(cnode, rs)\n\t\t\tif n := cnode.children[\"#\"]; n != nil {\n\t\t\t\tsetRs(n, rs)\n\t\t\t}\n\t\t} else {\n\t\t\tcnode.matchTopic(topicSlice[1:], rs)\n\t\t}\n\t}\n}", "func TopicListOne(w http.ResponseWriter, r *http.Request) {\n\n\t// Init output\n\toutput := []byte(\"\")\n\n\t// Add content type header to the response\n\tcontentType := \"application/json\"\n\tcharset := \"utf-8\"\n\tw.Header().Add(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\t// Grab url path variables\n\turlVars := mux.Vars(r)\n\n\t// Grab context references\n\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\tprojectUUID := gorillaContext.Get(r, \"auth_project_uuid\").(string)\n\n\tresults, err := topics.Find(projectUUID, \"\", urlVars[\"topic\"], \"\", 0, refStr)\n\n\tif err != nil {\n\t\terr := APIErrGenericBackend()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// If not found\n\tif results.Empty() {\n\t\terr := APIErrorNotFound(\"Topic\")\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\tres := results.Topics[0]\n\n\t// Output result to JSON\n\tresJSON, err := res.ExportJSON()\n\tif err != nil {\n\t\terr := APIErrExportJSON()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Write response\n\toutput = []byte(resJSON)\n\trespondOK(w, output)\n\n}", "func (conn *Conn) HasTopic(topic string) bool {\n\tconn.listeners.Lock()\n\tdefer conn.listeners.Unlock()\n\tfor _, g := range conn.topics {\n\t\tfor _, t := range g {\n\t\t\tif topic == t {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\treturn false\n}", "func (c *KafkaCluster) getTopicState(groupID, topicName string) *topicState {\n\tgroup := c.getGroupState(groupID)\n\n\t// Read lock check\n\tc.lock.RLock()\n\tif topic, ok := group[topicName]; ok {\n\t\tc.lock.RUnlock()\n\t\treturn topic\n\t}\n\tc.lock.RUnlock()\n\n\t// Write lock check and possible create\n\tc.lock.Lock()\n\tdefer c.lock.Unlock()\n\n\tif topic, ok := group[topicName]; ok {\n\t\treturn topic\n\t}\n\tgroup[topicName] = &topicState{\n\t\tclaimPartition: c.getClaimPartition(topicName),\n\t\tpartitions: nil,\n\t\tlock: &sync.RWMutex{},\n\t}\n\treturn group[topicName]\n}", "func (p *Partition) Topic() *Topic {\n\treturn p.topic\n}", "func (t *TopicCache) LoadByTopicName(projectName, serviceName, topicName string) (aiven.KafkaTopic, bool) {\n\tt.RLock()\n\tdefer t.RUnlock()\n\n\ttopics, ok := t.internal[projectName+serviceName]\n\tif !ok {\n\t\treturn aiven.KafkaTopic{State: \"CONFIGURING\"}, false\n\t}\n\n\tresult, ok := topics[topicName]\n\tif !ok {\n\t\tresult.State = \"CONFIGURING\"\n\t}\n\n\tlog.Printf(\"[TRACE] retrienve from a topic cache `%+#v` for a topic name `%s`\", result, topicName)\n\n\treturn result, ok\n}", "func (s *Storage) GetChannelByTopic(topic string) (*fetching.RSS, error) {\n\n\tlink := s.getTopicLink(topic)\n\tresp, err := http.Get(link)\n\tif err != nil {\n\t\treturn nil, errors.New(\"getting topic link\")\n\t}\n\trss := &fetching.RSS{}\n\tdefer resp.Body.Close()\n\tdecoder := xml.NewDecoder(resp.Body)\n\terr = decoder.Decode(&rss)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn rss, nil\n}", "func (this *WebController) Topic() {\n\t//选题总览\n\tp := P{}\n\tp[\"news_class\"] = 1\n\tp[\"dh\"] = topic_dh\n\tp[\"old\"] = 0\n\ttotal := D(News).Find(p).Count()\n\tlist := *D(News).Find(p).Sort(\"-date\").Limit(100).All()\n\tif len(list) == 0 {\n\t\tp := P{}\n\t\tp[\"news_class\"] = 1\n\t\tp[\"dh\"] = topic_dh\n\t\tp[\"old\"] = 1\n\t\tlist = *D(News).Find(p).Sort(\"-date\").Limit(100).All()\n\t}\n\tr := P{}\n\tr[\"total\"] = total\n\t//r[\"page\"], _ = this.GetInt(\"page\", 1)\n\tindex := 1\n\tfor _, v := range list {\n\t\t//为每条信息添加序号\n\t\tv[\"index\"] = index\n\t\tindex++\n\t}\n\tr[\"list\"] = list\n\tthis.EchoJsonMsg(r)\n}", "func (c *Client) GetTopic() string {\n\treturn c.topic\n}", "func (p *Publisher) Topic() []string {\n\treturn p.keys\n}", "func (kz *Kazoo) Topic(topic string) *Topic {\n\treturn &Topic{Name: topic, kz: kz}\n}", "func (o *TopicsToResetOffset) GetTopic() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Topic\n}", "func GetOrCreateTopic(name string, ts []string) *Topic {\n\ttopic, err := GetTopic(name)\n\tif err != nil {\n\t\ttopic = NewTopic(name, ts)\n\t\ttopic.Save()\n\t\treturn topic\n\t}\n\treturn topic\n}", "func (ri *retentionInfo) loadRetentionInfo(topic string, wg *sync.WaitGroup) {\n\t// TODO(yukun): If there needs to add lock\n\t// ll, ok := topicMu.Load(topic)\n\t// if !ok {\n\t// \treturn fmt.Errorf(\"topic name = %s not exist\", topic)\n\t// }\n\t// lock, ok := ll.(*sync.Mutex)\n\t// if !ok {\n\t// \treturn fmt.Errorf(\"get mutex failed, topic name = %s\", topic)\n\t// }\n\t// lock.Lock()\n\t// defer lock.Unlock()\n\tdefer wg.Done()\n\tpageEndID := make([]UniqueID, 0)\n\tpageMsgSize := make(map[int64]UniqueID)\n\n\tfixedPageSizeKey, err := constructKey(PageMsgSizeTitle, topic)\n\tif err != nil {\n\t\tlog.Debug(\"ConstructKey failed\", zap.Any(\"error\", err))\n\t\treturn\n\t}\n\tpageMsgSizePrefix := fixedPageSizeKey + \"/\"\n\tpageMsgSizeKeys, pageMsgSizeVals, err := prefixLoad(ri.kv.DB, pageMsgSizePrefix)\n\tif err != nil {\n\t\tlog.Debug(\"PrefixLoad failed\", zap.Any(\"error\", err))\n\t\treturn\n\t}\n\tfor i, key := range pageMsgSizeKeys {\n\t\tendID, err := strconv.ParseInt(key[FixedChannelNameLen+1:], 10, 64)\n\t\tif err != nil {\n\t\t\tlog.Debug(\"ParseInt failed\", zap.Any(\"error\", err))\n\t\t\treturn\n\t\t}\n\t\tpageEndID = append(pageEndID, endID)\n\n\t\tmsgSize, err := strconv.ParseInt(pageMsgSizeVals[i], 10, 64)\n\t\tif err != nil {\n\t\t\tlog.Debug(\"ParseInt failed\", zap.Any(\"error\", err))\n\t\t\treturn\n\t\t}\n\t\tpageMsgSize[endID] = msgSize\n\t}\n\ttopicPageInfo := &topicPageInfo{\n\t\tpageEndID: pageEndID,\n\t\tpageMsgSize: pageMsgSize,\n\t}\n\n\t// Load all acked infos\n\tackedTs := make(map[UniqueID]UniqueID)\n\n\ttopicBeginIDKey := TopicBeginIDTitle + topic\n\ttopicBeginIDVal, err := ri.kv.Load(topicBeginIDKey)\n\tif err != nil {\n\t\treturn\n\t}\n\ttopicBeginID, err := strconv.ParseInt(topicBeginIDVal, 10, 64)\n\tif err != nil {\n\t\tlog.Debug(\"ParseInt failed\", zap.Any(\"error\", err))\n\t\treturn\n\t}\n\n\tackedTsPrefix, err := constructKey(AckedTsTitle, topic)\n\tif err != nil {\n\t\tlog.Debug(\"ConstructKey failed\", zap.Any(\"error\", err))\n\t\treturn\n\t}\n\tkeys, vals, err := prefixLoad(ri.kv.DB, ackedTsPrefix)\n\tif err != nil {\n\t\tlog.Debug(\"PrefixLoad failed\", zap.Any(\"error\", err))\n\t\treturn\n\t}\n\n\tfor i, key := range keys {\n\t\toffset := FixedChannelNameLen + 1\n\t\tackedID, err := strconv.ParseInt((key)[offset:], 10, 64)\n\t\tif err != nil {\n\t\t\tlog.Debug(\"RocksMQ: parse int \" + key[offset:] + \" failed\")\n\t\t\treturn\n\t\t}\n\n\t\tts, err := strconv.ParseInt(vals[i], 10, 64)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tackedTs[ackedID] = ts\n\t}\n\n\tackedSizeKey := AckedSizeTitle + topic\n\tackedSizeVal, err := ri.kv.Load(ackedSizeKey)\n\tif err != nil {\n\t\tlog.Debug(\"Load failed\", zap.Any(\"error\", err))\n\t\treturn\n\t}\n\tvar ackedSize int64\n\tif ackedSizeVal == \"\" {\n\t\tackedSize = 0\n\t} else {\n\t\tackedSize, err = strconv.ParseInt(ackedSizeVal, 10, 64)\n\t\tif err != nil {\n\t\t\tlog.Debug(\"PrefixLoad failed\", zap.Any(\"error\", err))\n\t\t\treturn\n\t\t}\n\t}\n\n\tackedInfo := &topicAckedInfo{\n\t\ttopicBeginID: topicBeginID,\n\t\tackedTs: ackedTs,\n\t\tackedSize: ackedSize,\n\t}\n\n\t//Load last retention timestamp\n\tlastRetentionTsKey := LastRetTsTitle + topic\n\tlastRetentionTsVal, err := ri.kv.Load(lastRetentionTsKey)\n\tif err != nil {\n\t\tlog.Debug(\"Load failed\", zap.Any(\"error\", err))\n\t\treturn\n\t}\n\tvar lastRetentionTs int64\n\tif lastRetentionTsVal == \"\" {\n\t\tlastRetentionTs = math.MaxInt64\n\t} else {\n\t\tlastRetentionTs, err = strconv.ParseInt(lastRetentionTsVal, 10, 64)\n\t\tif err != nil {\n\t\t\tlog.Debug(\"ParseInt failed\", zap.Any(\"error\", err))\n\t\t\treturn\n\t\t}\n\t}\n\n\tri.ackedInfo.Store(topic, ackedInfo)\n\tri.pageInfo.Store(topic, topicPageInfo)\n\tri.lastRetentionTime.Store(topic, lastRetentionTs)\n}", "func popularTopics(p *model.GundemParams) ([]model.Topic, error) {\n\ttopicList := make([]model.Topic, 0)\n\tif len(p.Kategori) > 0 {\n\t\tgundemURL += \"kanal/\" + p.Kategori\n\t} else {\n\t\tgundemURL += \"gundem\"\n\t}\n\n\tfor {\n\t\tresp, err := http.Get(gundemURL + \"?p=\" + strconv.Itoa(p.Page))\n\t\tif err != nil {\n\t\t\treturn nil, errors.New(\"ERROR: internet bağlantınızı kontrol edin\")\n\t\t}\n\n\t\troot, err := html.Parse(resp.Body)\n\t\tif err != nil {\n\t\t\treturn nil, errors.New(\"ERROR: An error occured while parsing body\")\n\t\t}\n\t\tdefer resp.Body.Close()\n\n\t\ttopics := scrape.FindAll(root, topicMatcher)\n\t\tif len(topics) == 0 {\n\t\t\treturn nil, errors.New(\"ERROR: Lütfen parametre değerlerinizi kontrol edin\")\n\t\t}\n\n\t\tfor _, topic := range topics {\n\t\t\tt := model.Topic{}\n\t\t\tt.Title = scrape.Text(topic.FirstChild)\n\n\t\t\tif topic.LastChild.DataAtom == atom.Small {\n\t\t\t\tt.NewEntryCount = scrape.Text(topic.LastChild)\n\t\t\t}\n\n\t\t\ttopicList = append(topicList, t)\n\n\t\t\tif len(topicList) == p.Limit {\n\t\t\t\treturn topicList, nil\n\t\t\t}\n\t\t}\n\t\tp.Page++\n\t}\n}", "func (t *TopicService) Get(id string) (*Topic, error) {\n\tquery := `\n\tquery ($id: ID){\n\t\ttopic(id: $id){\n\t\t\tid,\n\t\t\tname,\n\t\t\tdescription,\n\t\t\tposts{id, title},\n\t\t\thierarchy,\n\t\t\tparent{id},\n\t\t\tancestors{id},\n\t\t\tchildren{id},\n\t\t\tinsertedAt,\n\t\t\tupdatedAt\n\t\t}\n\t}`\n\tvar resp struct {\n\t\tTopic *Topic `json:\"topic\"`\n\t}\n\tvars := map[string]interface{}{\"id\": id}\n\terr := t.client.Do(context.Background(), query, vars, &resp)\n\treturn resp.Topic, err\n}", "func (r *rdsRoute) Exist(ns, topic string) (bool, error) {\n\tkey := r.getRoutePrefix(ns, topic)\n\trds := r.rdc.Get(util.R, key)\n\tdefer rds.Close()\n\treply, err := redis.Int(rds.Do(\"HLEN\", key))\n\tif err != nil || reply == 0 {\n\t\treturn false, err\n\t}\n\treturn true, nil\n}", "func (r *TopicReconciler) Reconcile(req ctrl.Request) (ctrl.Result, error) {\n\tctx := context.Background()\n\t_ = r.Log.WithValues(\"topic\", req.NamespacedName)\n\n\ttopic := &nlptv1.Topic{}\n\tif err := r.Get(ctx, req.NamespacedName, topic); err != nil {\n\t\tklog.Errorf(\"cannot get topic of ctrl req: %+v\", err)\n\t\treturn ctrl.Result{}, nil\n\t}\n\t//klog.Infof(\"get new topic event: %+v\", *topic)\n\t//klog.Infof(\"Status:%s\", topic.Status.Status)\n\n\tif topic.Status.Status == nlptv1.Creating {\n\t\t//klog.Info(\"Current status is Init\")\n\t\tif err := r.Operator.CreateTopic(topic); err != nil {\n\t\t\ttopic.Spec.Url = topic.GetUrl()\n\t\t\ttopic.Status.Status = nlptv1.CreateFailed\n\t\t\ttopic.Status.Message = fmt.Sprintf(\"create topic error:%+v\", err)\n\t\t\tklog.Errorf(\"create topic failed, err: %+v\", err)\n\t\t} else {\n\t\t\ttopic.Spec.Url = topic.GetUrl()\n\t\t\ttopic.Status.Status = nlptv1.Created\n\t\t\ttopic.Status.Message = \"success\"\n\t\t}\n\n\t\t//更新数据库的状态\n\t\t//klog.Infof(\"Final Topic: %+v\", *topic)\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v, err: %+v\", *topic, err)\n\t\t}\n\n\t}\n\n\tif topic.Status.Status == nlptv1.Importing {\n\t\tif ok, _ := r.Operator.isNamespacesExist(topic); ok {\n\t\t\tif err := r.Operator.CreateTopic(topic); err != nil {\n\t\t\t\ttopic.Spec.Url = topic.GetUrl()\n\t\t\t\ttopic.Status.Status = nlptv1.ImportFailed\n\t\t\t\ttopic.Status.Message = fmt.Sprintf(\"create topic error:%+v\", err)\n\t\t\t\tklog.Errorf(\"create topic failed, err: %+v\", err)\n\t\t\t} else {\n\t\t\t\ttopic.Spec.Url = topic.GetUrl()\n\t\t\t\ttopic.Status.Status = nlptv1.ImportSuccess\n\t\t\t\ttopic.Status.Message = \"success\"\n\t\t\t}\n\n\t\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\tif topic.Status.Status == nlptv1.Deleting {\n\t\tif err := r.Operator.DeleteTopic(topic, false); err != nil {\n\t\t\ttopic.Status.Status = nlptv1.DeleteFailed\n\t\t\ttopic.Status.Message = fmt.Sprintf(\"delete topic error: %+v\", err)\n\t\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t\t}\n\t\t} else {\n\t\t\t//删除数据\n\t\t\tif err = r.Delete(ctx, topic); err != nil {\n\t\t\t\tklog.Errorf(\"delete Topic Failed: %+v\", *topic)\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif topic.Status.Status == nlptv1.ForceDeleting {\n\t\tif err := r.Operator.DeleteTopic(topic, true); err != nil {\n\t\t\ttopic.Status.Status = nlptv1.ForceDeleteFailed\n\t\t\ttopic.Status.Message = fmt.Sprintf(\"force delete topic error: %+v\", err)\n\t\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t\t}\n\t\t} else {\n\t\t\t//删除数据\n\t\t\tif err = r.Delete(ctx, topic); err != nil {\n\t\t\t\tklog.Errorf(\"delete Topic Failed: %+v\", *topic)\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif topic.Status.Status == nlptv1.Terminating {\n\t\tif err := r.Operator.TerminateTopic(topic); err != nil {\n\t\t\ttopic.Status.Status = nlptv1.TerminatedFailed\n\t\t\ttopic.Status.Message = fmt.Sprintf(\"terminate topic error: %+v\", err)\n\n\t\t} else {\n\t\t\ttopic.Status.Status = nlptv1.Terminated\n\t\t\ttopic.Status.Message = fmt.Sprintf(\"terminate topic success\")\n\t\t}\n\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t}\n\n\t}\n\n\tif topic.Status.Status == nlptv1.Updating {\n\t\t//增加topic分区\n\t\tif err := r.Operator.AddPartitionsOfTopic(topic); err != nil {\n\t\t\ttopic.Status.Status = nlptv1.UpdateFailed\n\t\t\ttopic.Status.Message = fmt.Sprintf(\"add topic partition error: %+v \", err)\n\t\t\ttopic.Spec.PartitionNum = topic.Spec.OldPartitionNum\n\t\t} else {\n\t\t\ttopic.Status.Status = nlptv1.Updated\n\t\t\ttopic.Status.Message = \"success\"\n\t\t}\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t}\n\t}\n\n\t//删除授权\n\tif topic.Status.AuthorizationStatus == nlptv1.DeletingAuthorization {\n\t\tfor i := 0; i < len(topic.Spec.Permissions); i++ {\n\t\t\tp := topic.Spec.Permissions[i]\n\t\t\tif p.Status.Status == nlptv1.DeletingAuthorization {\n\t\t\t\tif err := r.Operator.DeletePer(topic, &p); err != nil {\n\t\t\t\t\tp.Status.Status = nlptv1.DeleteAuthorizationFailed\n\t\t\t\t\tp.Status.Message = fmt.Sprintf(\"revoke permission error: %+v\", err)\n\t\t\t\t\t//删除失败,将标签重置为true\n\t\t\t\t\ttopic.ObjectMeta.Labels[p.AuthUserID] = \"true\"\n\t\t\t\t\ttopic.Status.AuthorizationStatus = nlptv1.DeleteAuthorizationFailed\n\t\t\t\t} else {\n\t\t\t\t\tpers := topic.Spec.Permissions\n\t\t\t\t\ttopic.Spec.Permissions = append(pers[:i], pers[i+1:]...)\n\t\t\t\t\t//收回权限成功,删除标签\n\t\t\t\t\tdelete(topic.ObjectMeta.Labels, p.AuthUserID)\n\t\t\t\t\ttopic.Status.AuthorizationStatus = nlptv1.DeletedAuthorization\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t}\n\t}\n\n\tif topic.Status.AuthorizationStatus == nlptv1.UpdatingAuthorization {\n\t\t//klog.Infof(\"Start Grant Topic: %+v\", *topic)\n\t\t//授权操作\n\t\tfor i := 0; i < len(topic.Spec.Permissions); i++ {\n\t\t\tp := topic.Spec.Permissions[i]\n\t\t\tif p.Status.Status == nlptv1.UpdatingAuthorization {\n\t\t\t\tif err := r.Operator.GrantPermission(topic, &p); err != nil {\n\t\t\t\t\tp.Status.Status = nlptv1.UpdatingAuthorizationFailed\n\t\t\t\t\tp.Status.Message = fmt.Sprintf(\"modify permission error: %+v\", err)\n\t\t\t\t\ttopic.Status.AuthorizationStatus = nlptv1.UpdatingAuthorizationFailed\n\n\t\t\t\t\t//TODO roll back\n\n\t\t\t\t} else {\n\t\t\t\t\tp.Status.Status = nlptv1.UpdatingAuthorizationSuccess\n\t\t\t\t\tp.Status.Message = \"success\"\n\t\t\t\t\ttopic.Status.AuthorizationStatus = nlptv1.UpdatingAuthorizationSuccess\n\t\t\t\t}\n\t\t\t\ttopic.Spec.Permissions[i] = p\n\t\t\t}\n\t\t}\n\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t}\n\t}\n\tif topic.Status.AuthorizationStatus == nlptv1.Authorizing {\n\t\tklog.Infof(\"Start Grant Topic: %+v\", *topic)\n\t\t//授权操作\n\t\tfor i := 0; i < len(topic.Spec.Permissions); i++ {\n\t\t\tp := topic.Spec.Permissions[i]\n\t\t\tif p.Status.Status == nlptv1.Authorizing {\n\t\t\t\tif err := r.Operator.GrantPermission(topic, &p); err != nil {\n\t\t\t\t\tp.Status.Status = nlptv1.AuthorizeFailed\n\t\t\t\t\tp.Status.Message = fmt.Sprintf(\"grant permission error: %+v\", err)\n\t\t\t\t\ttopic.Status.AuthorizationStatus = nlptv1.AuthorizeFailed\n\t\t\t\t} else {\n\t\t\t\t\tp.Status.Status = nlptv1.Authorized\n\t\t\t\t\tp.Status.Message = \"success\"\n\t\t\t\t\ttopic.Status.AuthorizationStatus = nlptv1.Authorized\n\t\t\t\t}\n\t\t\t\ttopic.Spec.Permissions[i] = p\n\t\t\t}\n\t\t}\n\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t}\n\t}\n\n\tif topic.Status.BindStatus == nlptv1.BindingOrUnBinding {\n\t\tfor appid, application := range topic.Spec.Applications {\n\t\t\tswitch application.Status {\n\t\t\tcase nlptv1.UpdatingAuthorization:\n\t\t\t\tp := nlptv1.Permission{\n\t\t\t\t\tAuthUserID: \"\",\n\t\t\t\t\tAuthUserName: application.ID,\n\t\t\t\t\tActions: application.Actions,\n\t\t\t\t}\n\t\t\t\tif err := r.Operator.GrantPermission(topic, &p); err != nil {\n\t\t\t\t\tapplication.Status = nlptv1.UpdatingAuthorizationFailed\n\t\t\t\t\tapplication.DisplayStatus = nlptv1.ShowStatusMap[application.Status]\n\t\t\t\t\tapplication.Message = fmt.Sprintf(\"update authorization: %+v\", err)\n\t\t\t\t} else {\n\t\t\t\t\tapplication.Status = nlptv1.UpdatingAuthorizationSuccess\n\t\t\t\t\tapplication.DisplayStatus = nlptv1.ShowStatusMap[application.Status]\n\t\t\t\t\tapplication.Message = \"update authorization successfully\"\n\t\t\t\t}\n\t\t\tcase nlptv1.Binding:\n\t\t\t\t//actions := make([]string, 0)\n\t\t\t\t//actions = append(actions, nlptv1.Consume)\n\t\t\t\t//actions = append(actions, nlptv1.Produce)\n\n\t\t\t\tp := nlptv1.Permission{\n\t\t\t\t\tAuthUserID: \"\",\n\t\t\t\t\tAuthUserName: application.ID,\n\t\t\t\t\tActions: application.Actions,\n\t\t\t\t}\n\t\t\t\tif err := r.Operator.GrantPermission(topic, &p); err != nil {\n\t\t\t\t\tapplication.Status = nlptv1.BindFailed\n\t\t\t\t\tapplication.DisplayStatus = nlptv1.ShowStatusMap[application.Status]\n\t\t\t\t\tapplication.Message = fmt.Sprintf(\"bind error: %+v\", err)\n\t\t\t\t} else {\n\t\t\t\t\tapplication.Status = nlptv1.Bound\n\t\t\t\t\tapplication.DisplayStatus = nlptv1.ShowStatusMap[application.Status]\n\t\t\t\t\tapplication.Message = \"bind success\"\n\t\t\t\t}\n\t\t\tcase nlptv1.Unbinding:\n\t\t\t\tp := nlptv1.Permission{\n\t\t\t\t\tAuthUserID: \"\",\n\t\t\t\t\tAuthUserName: application.ID,\n\t\t\t\t}\n\n\t\t\t\tif err := r.Operator.DeletePer(topic, &p); err != nil {\n\t\t\t\t\tapplication.Status = nlptv1.UnbindFailed\n\t\t\t\t\tapplication.DisplayStatus = nlptv1.ShowStatusMap[application.Status]\n\t\t\t\t\tapplication.Message = fmt.Sprintf(\"release error: %+v\", err)\n\n\t\t\t\t} else {\n\t\t\t\t\tapplication.Status = nlptv1.UnbindSuccess\n\t\t\t\t}\n\n\t\t\t}\n\t\t\ttopic.Spec.Applications[appid] = application\n\t\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t\t}\n\t\t}\n\n\t\t//处理解绑定的场景\n\t\tapps := make(map[string]nlptv1.Application)\n\n\t\tfor appid, application := range topic.Spec.Applications {\n\t\t\tif application.Status != nlptv1.UnbindSuccess {\n\t\t\t\tapps[appid] = application\n\t\t\t}\n\t\t}\n\n\t\ttopic.Spec.Applications = apps\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t}\n\n\t}\n\n\t//klog.Infof(\"Final Topic: %+v\", *topic)\n\treturn ctrl.Result{}, nil\n}", "func (_Flytrap *FlytrapSession) Topics(arg0 string) (struct {\n\tIsValue bool\n\tSensitive bool\n\tName string\n\tAddPubCost *big.Int\n\tAddSubCost *big.Int\n\tOwner common.Address\n\tCountry [2]byte\n}, error) {\n\treturn _Flytrap.Contract.Topics(&_Flytrap.CallOpts, arg0)\n}", "func Match(savedTopic, givenTopic string) bool {\n\treturn givenTopic == savedTopic || match(strings.Split(savedTopic, \"/\"), strings.Split(givenTopic, \"/\"))\n}", "func (s *onmemoryStorage) findSubscriberForFetchMessages(ctx context.Context, sl domain.SubscriberLocator) (*onmemorySubscriber, error) {\n\t// This method is called from FetchMessages function.\n\t// It does not want to lock storage during polling.\n\t// So that lock storage in this method instead of the FetchMessages.\n\tunlock, err := s.lock.Lock(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer unlock()\n\n\tch, err := s.getChannel(sl.ChannelID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsbsc := ch.subscribers[sl.SubscriberID]\n\tif sbsc == nil {\n\t\treturn nil, xerrors.Errorf(\"%w\", domain.ErrSubscriptionNotFound)\n\t}\n\treturn sbsc, nil\n}", "func (m *MessageProcessor) shouldBeCached(t topics.Topic) bool {\n\tswitch t {\n\tcase topics.Tx,\n\t\ttopics.Candidate,\n\t\ttopics.NewBlock,\n\t\ttopics.Reduction,\n\t\ttopics.Agreement,\n\t\ttopics.AggrAgreement,\n\t\ttopics.GetCandidate:\n\t\treturn true\n\tdefault:\n\t\treturn false\n\t}\n}", "func (cl *Client) addUnknownTopicRecord(pr promisedRec) {\n\tunknown := cl.producer.unknownTopics[pr.Topic]\n\tif unknown == nil {\n\t\tunknown = &unknownTopicProduces{\n\t\t\tbuffered: make([]promisedRec, 0, 100),\n\t\t\twait: make(chan error, 5),\n\t\t}\n\t\tcl.producer.unknownTopics[pr.Topic] = unknown\n\t}\n\tunknown.buffered = append(unknown.buffered, pr)\n\tif len(unknown.buffered) == 1 {\n\t\tgo cl.waitUnknownTopic(pr.Topic, unknown)\n\t}\n}", "func (r *TopicsService) Get(topic string) *TopicsGetCall {\n\tc := &TopicsGetCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.topic = topic\n\treturn c\n}", "func (t *TopicOperation) CheckTopic(topicName string) bool {\n\tconst resourceName = \"cephbuckettopic\"\n\t_, err := t.k8sh.GetResource(resourceName, topicName)\n\tif err != nil {\n\t\tlogger.Infof(\"%q %q does not exist\", resourceName, topicName)\n\t\treturn false\n\t}\n\n\ttopicARN, _ := t.k8sh.GetResource(resourceName, topicName, \"--output\", \"jsonpath={.status.ARN}\")\n\tif topicARN == \"\" {\n\t\tlogger.Infof(\"%q %q exist, but ARN was not set\", resourceName, topicName)\n\t\treturn false\n\t}\n\n\tlogger.Infof(\"topic ARN is %q\", topicARN)\n\treturn true\n}", "func (n *Notifier) FetchOrCreateConversation(s *session.Session, topic string) (string, error) {\n\tconversationID := n.ConversationIDFromTopic(topic)\n\tif _, f := conversations[conversationID]; f {\n\t\treturn conversationID, nil\n\t}\n\n\tparticipants := []*mixin.Participant{\n\t\t&mixin.Participant{\n\t\t\tUserID: adminID,\n\t\t\tRole: \"ADMIN\",\n\t\t},\n\t}\n\tif _, err := n.CreateConversation(s.Context(), \"GROUP\", conversationID, topic, \"\", \"\", \"\", participants); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tconversations[conversationID] = true\n\tt := Topic{\n\t\tTopic: topic,\n\t\tConversationID: conversationID,\n\t}\n\terr := s.MysqlWrite().Where(\"conversation_id = ?\", conversationID).FirstOrCreate(&t).Error\n\treturn conversationID, err\n}", "func become_discoverable(args []string) {\n\tsongs := get_local_song_info(args[2])\n\tmsg_content := \"\"\n\tfor _, s := range songs {\n\t\tmsg_content += s\n\t}\n\tmsg := prepare_msg(INIT, 0, []byte(msg_content))\n\ttracker := send(*msg, TRACKER_IP+args[1])\n\tdefer tracker.Close()\n}", "func (c *client) createTopic(topicName string) (*pubsub.Topic, error) {\n\tvar topic *pubsub.Topic\n\tctx := context.Background()\n\n\texists, err := c.topicExists(topicName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif !exists {\n\t\ttopic, err = c.client.CreateTopic(ctx, topicName)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\ttopic = c.client.Topic(topicName)\n\t}\n\n\treturn topic, nil\n}", "func (e *Extractor) buildTopic(topic_name string) meta.Topic {\n\treturn meta.Topic{\n\t\tUrn: topic_name,\n\t\tName: topic_name,\n\t\tSource: \"kafka\",\n\t}\n}", "func (suite *BaseSuite) EnsureTopic(ctx context.Context, name string) (*servicebus.TopicEntity, error) {\n\tns := suite.GetNewNamespace()\n\ttm := ns.NewTopicManager()\n\n\tte, err := tm.Get(ctx, name)\n\tif err == nil {\n\t\treturn te, nil\n\t}\n\n\treturn tm.Put(ctx, name)\n}", "func (e *engineImpl) prepareTopic(c context.Context, params topicParams) (topic string, tok string, err error) {\n\t// If given URL, ask the service for name of its default service account.\n\t// FetchServiceInfo implements efficient cache internally, so it's fine to\n\t// call it often.\n\tif strings.HasPrefix(params.publisher, \"https://\") {\n\t\tlogging.Infof(c, \"Fetching info about %q\", params.publisher)\n\t\tserviceInfo, err := authinfo.FetchServiceInfo(c, params.publisher)\n\t\tif err != nil {\n\t\t\tlogging.Errorf(c, \"Failed to fetch info about %q - %s\", params.publisher, err)\n\t\t\treturn \"\", \"\", err\n\t\t}\n\t\tlogging.Infof(c, \"%q is using %q\", params.publisher, serviceInfo.ServiceAccountName)\n\t\tparams.publisher = serviceInfo.ServiceAccountName\n\t}\n\n\ttopic, sub := e.genTopicAndSubNames(c, params.manager.Name(), params.publisher)\n\n\t// Put same parameters in push URL to make them visible in logs. On dev server\n\t// use pull based subscription, since localhost push URL is not valid.\n\tpushURL := \"\"\n\tif inf := info.Get(c); !inf.IsDevAppServer() {\n\t\turlParams := url.Values{}\n\t\turlParams.Add(\"kind\", params.manager.Name())\n\t\turlParams.Add(\"publisher\", params.publisher)\n\t\tpushURL = fmt.Sprintf(\n\t\t\t\"https://%s%s?%s\", inf.DefaultVersionHostname(), e.PubSubPushPath, urlParams.Encode())\n\t}\n\n\t// Create and configure the topic. Do it only once.\n\terr = e.doIfNotDone(c, fmt.Sprintf(\"prepareTopic:v1:%s\", topic), func() error {\n\t\tif e.configureTopic != nil {\n\t\t\treturn e.configureTopic(c, topic, sub, pushURL, params.publisher)\n\t\t}\n\t\treturn configureTopic(c, topic, sub, pushURL, params.publisher, \"\")\n\t})\n\tif err != nil {\n\t\treturn \"\", \"\", err\n\t}\n\n\t// Encode full invocation identifier (job key + invocation ID) into HMAC\n\t// protected token.\n\ttok, err = pubsubAuthToken.Generate(c, nil, map[string]string{\n\t\t\"job\": params.inv.JobKey.StringID(),\n\t\t\"inv\": fmt.Sprintf(\"%d\", params.inv.ID),\n\t}, 0)\n\tif err != nil {\n\t\treturn \"\", \"\", err\n\t}\n\n\treturn topic, tok, nil\n}", "func notSubscribedTopicNotFound(src *v1alpha1.AWSSNSSource) {\n\tsrc.Status.MarkNotSubscribed(v1alpha1.AWSSNSReasonFailedSync,\n\t\tfmt.Sprintf(\"The provided topic %q does not exist\", tTopicARN),\n\t)\n}", "func (c *consumer) fetch() ([]*proto.Message, error) {\n\treq := proto.FetchReq{\n\t\tClientID: c.broker.conf.ClientID,\n\t\tMaxWaitTime: c.conf.RequestTimeout,\n\t\tMinBytes: c.conf.MinFetchSize,\n\t\tTopics: []proto.FetchReqTopic{\n\t\t\t{\n\t\t\t\tName: c.conf.Topic,\n\t\t\t\tPartitions: []proto.FetchReqPartition{\n\t\t\t\t\t{\n\t\t\t\t\t\tID: c.conf.Partition,\n\t\t\t\t\t\tFetchOffset: c.offset,\n\t\t\t\t\t\tMaxBytes: c.conf.MaxFetchSize,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\tvar resErr error\n\tretry := &backoff.Backoff{Min: c.conf.RetryErrWait, Jitter: true}\nconsumeRetryLoop:\n\tfor try := 0; try < c.conf.RetryErrLimit; try++ {\n\t\tif try != 0 {\n\t\t\ttime.Sleep(retry.Duration())\n\t\t}\n\n\t\tconn, err := c.broker.leaderConnection(c.conf.Topic, c.conf.Partition)\n\t\tif err != nil {\n\t\t\tresErr = err\n\t\t\tcontinue\n\t\t}\n\t\tdefer func(lconn *connection) { go c.broker.conns.Idle(lconn) }(conn)\n\n\t\tresp, err := conn.Fetch(&req)\n\t\tresErr = err\n\t\tif _, ok := err.(*net.OpError); ok || err == io.EOF || err == syscall.EPIPE {\n\t\t\tlog.Debugf(\"connection died while fetching messages from %s:%d: %s\",\n\t\t\t\tc.conf.Topic, c.conf.Partition, err)\n\t\t\t_ = conn.Close()\n\t\t\tcontinue\n\t\t}\n\n\t\tif err != nil {\n\t\t\tlog.Debugf(\"cannot fetch messages (try %d): %s\", retry, err)\n\t\t\t_ = conn.Close()\n\t\t\tcontinue\n\t\t}\n\n\t\t// Should only be a single topic/partition in the response, the one we asked about.\n\t\tfor _, t := range resp.Topics {\n\t\t\tfor _, p := range t.Partitions {\n\t\t\t\tif t.Name != c.conf.Topic || p.ID != c.conf.Partition {\n\t\t\t\t\tlog.Warningf(\"fetch response with unexpected data for %s:%d\",\n\t\t\t\t\t\tt.Name, p.ID)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tswitch p.Err {\n\t\t\t\tcase proto.ErrLeaderNotAvailable, proto.ErrNotLeaderForPartition,\n\t\t\t\t\tproto.ErrBrokerNotAvailable, proto.ErrUnknownTopicOrPartition:\n\t\t\t\t\t// Failover happened, so we probably need to talk to a different broker. Let's\n\t\t\t\t\t// kick off a metadata refresh.\n\t\t\t\t\tlog.Warningf(\"cannot fetch messages (try %d): %s\", retry, p.Err)\n\t\t\t\t\tif err := c.broker.cluster.RefreshMetadata(); err != nil {\n\t\t\t\t\t\tlog.Warningf(\"cannot refresh metadata: %s\", err)\n\t\t\t\t\t}\n\t\t\t\t\tcontinue consumeRetryLoop\n\t\t\t\t}\n\t\t\t\treturn p.Messages, p.Err\n\t\t\t}\n\t\t}\n\t\treturn nil, errors.New(\"incomplete fetch response\")\n\t}\n\n\treturn nil, resErr\n}", "func (_Flytrap *FlytrapCallerSession) Topics(arg0 string) (struct {\n\tIsValue bool\n\tSensitive bool\n\tName string\n\tAddPubCost *big.Int\n\tAddSubCost *big.Int\n\tOwner common.Address\n\tCountry [2]byte\n}, error) {\n\treturn _Flytrap.Contract.Topics(&_Flytrap.CallOpts, arg0)\n}", "func (db *trieDB) getMatchedTopicFilter(topicName string) map[string][]packets.Topic {\n\t// system topic\n\tif isSystemTopic(topicName) {\n\t\treturn db.systemTrie.getMatchedTopicFilter(topicName)\n\t}\n\treturn db.userTrie.getMatchedTopicFilter(topicName)\n}", "func (conn *ProtoConnection) ConsumeTopic(msgClb func(messaging.ProtoMessage), topics ...string) error {\n\tconn.multiplexer.rwlock.Lock()\n\tdefer conn.multiplexer.rwlock.Unlock()\n\n\tif conn.multiplexer.started {\n\t\treturn fmt.Errorf(\"ConsumeTopic can be called only if the multiplexer has not been started yet\")\n\t}\n\n\tbyteClb := func(bm *client.ConsumerMessage) {\n\t\tpm := client.NewProtoConsumerMessage(bm, conn.serializer)\n\t\tmsgClb(pm)\n\t}\n\n\tfor _, topic := range topics {\n\t\t// check if we have already consumed the topic\n\t\tvar found bool\n\t\tvar subs *consumerSubscription\n\tLoopSubs:\n\t\tfor _, subscription := range conn.multiplexer.mapping {\n\t\t\tif subscription.manual == true {\n\t\t\t\t// do not mix dynamic and manual mode\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif subscription.topic == topic {\n\t\t\t\tfound = true\n\t\t\t\tsubs = subscription\n\t\t\t\tbreak LoopSubs\n\t\t\t}\n\t\t}\n\n\t\tif !found {\n\t\t\tsubs = &consumerSubscription{\n\t\t\t\tmanual: false, // non-manual example\n\t\t\t\ttopic: topic,\n\t\t\t\tconnectionName: conn.name,\n\t\t\t\tbyteConsMsg: byteClb,\n\t\t\t}\n\t\t\t// subscribe new topic\n\t\t\tconn.multiplexer.mapping = append(conn.multiplexer.mapping, subs)\n\t\t}\n\n\t\t// add subscription to consumerList\n\t\tsubs.byteConsMsg = byteClb\n\t}\n\n\treturn nil\n}", "func (w *Strap) Discover() {\n\trequestEncode(discoveryURL, w.token, &w.resources)\n\n}", "func read_nsq(lookupd_addr string, topic string, logchan string, lreader *logreader) {\n\treader, _ := nsq.NewReader(topic, logchan)\n\treader.AddAsyncHandler(lreader)\n\treader.ConnectToLookupd(lookupd_addr)\n}", "func (b *Builder) Topic() Topic {\n\treturn b.topic\n}", "func (p *postgresqlDatabase) getTopicNIDFromCache(topicName string) (topicNID int64) {\n\tp.topicsMutex.Lock()\n\tdefer p.topicsMutex.Unlock()\n\treturn p.topicNIDs[topicName]\n}", "func (_Contract *ContractCallerSession) GetTopics(account common.Address) ([]string, error) {\n\treturn _Contract.Contract.GetTopics(&_Contract.CallOpts, account)\n}", "func (o LiteSubscriptionOutput) Topic() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *LiteSubscription) pulumi.StringOutput { return v.Topic }).(pulumi.StringOutput)\n}", "func Pull(nc *nats.Conn, topic string, timeout time.Duration, handle nats.MsgHandler) {\n\tgo func() {\n\t\tfor {\n\t\t\tif interrupted == true {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tmsg, err := nc.Request(topic, []byte(\"\"), timeout)\n\t\t\tif err == nats.ErrTimeout {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif err != nil {\n\t\t\t\tfmt.Print(fmt.Sprintf(\"Error: %s\", err))\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\thandle(msg)\n\n\t\t\tmsg.Respond(nil)\n\t\t}\n\t}()\n}", "func (d *Dao) TopicHots(c context.Context) (topics []*clive.TopicHot, err error) {\n\tvar res struct {\n\t\tCode int `json:\"code\"`\n\t\tData struct {\n\t\t\tList []*clive.TopicHot `json:\"list\"`\n\t\t} `json:\"data\"`\n\t}\n\tif err = d.clientAsyn.Get(c, d.topic, \"\", nil, &res); err != nil {\n\t\treturn\n\t}\n\tb, _ := json.Marshal(&res)\n\tlog.Info(\"topichots list url(%v) response(%s)\", d.topic, b)\n\tif res.Code != ecode.OK.Code() {\n\t\terr = errors.Wrap(err, d.topic)\n\t\treturn\n\t}\n\tfor _, t := range res.Data.List {\n\t\ttmp := &clive.TopicHot{}\n\t\t*tmp = *t\n\t\tif err = tmp.TopicJSONChange(); err != nil {\n\t\t\tlog.Error(\"TopicJSONChange error(%v)\", err)\n\t\t\treturn\n\t\t}\n\t\ttopics = append(topics, tmp)\n\t}\n\treturn\n}", "func (p *Partition) GetTopic() string {\n\treturn p.topic\n}", "func (q *Qlient) SubOn(topic string) (chan []byte, error) {\n\tsub, ok := q.subs[topic]\n\tif !ok {\n\t\tvar err error\n\t\tsub, err = q.newConsumer(topic)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn sub, nil\n}", "func (p *DiscoveryProtocol) onDiscoveryRequest(s inet.Stream) {\n\t// get request data\n\tdata := &api.DiscoveryRequest{}\n\tdecodeProtoMessage(data, s)\n\n\t// Log the reception of the message\n\tlog.Printf(\"%s: Received discovery request from %s. Message: %s\", s.Conn().LocalPeer(), s.Conn().RemotePeer(), data.Message)\n\n\t// If the request's TTL expired or\n\t// If I received the same message again, I will skip\n\tif p.requestExpired(data) || p.checkMsgReceived(data) {\n\t\treturn\n\t}\n\t// Storing the received discovery message so as we don't process it again\n\tp.receivedMsgs[data.DiscoveryMsgData.InitHash] = data.DiscoveryMsgData.Expiry\n\n\t// Authenticate integrity and authenticity of the message\n\tif valid := authenticateProtoMsg(data, data.DiscoveryMsgData.MessageData); !valid {\n\t\tlog.Println(\"Failed to authenticate message\")\n\t\treturn\n\t}\n\n\t// Pass this message to my neighbours\n\tp.ForwardMsgToPeers(data, s.Conn().RemotePeer())\n\n\t// Even if there is possibility that we never send a reply to this Node (because of being busy),\n\t// we still store it our our Peerstore, because there is high possibility to\n\t// receive a request again.\n\n\t// If the node who sent this message is different than the initPeerID\n\t// then we need to add the init node to our neighbours before sending the message\n\tinitPeerID, _ := peer.IDB58Decode(data.DiscoveryMsgData.InitNodeID)\n\tif s.Conn().RemotePeer().String() != initPeerID.String() {\n\t\tp.dhtFindAddrAndStore(initPeerID)\n\t}\n\n\tbusy, err := NodeBusy()\n\tcommon.FatalIfErr(err, \"Error on checking if node is busy\")\n\tif busy {\n\t\t// Cache the request for a later time\n\t\tif uint16(len(p.pendingReq)) < p.maxPendingReq {\n\t\t\tp.pendingReq[data] = struct{}{}\n\t\t}\n\t\tlog.Println(\"I am busy at the moment. Returning...\")\n\t\treturn\n\t}\n\n\tp.createSendResponse(data)\n}", "func (ac *Consumer) EnsureTopics(topics []string) error {\n\tnotFound := make([]string, 0)\n\n\tmeta, err := ac.KafkaConsumer.GetMetadata(nil, true, 6000)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, topic := range topics {\n\t\tif _, ok := meta.Topics[topic]; !ok {\n\t\t\tnotFound = append(notFound, topic)\n\t\t}\n\t}\n\n\tif len(notFound) > 0 {\n\t\treturn fmt.Errorf(\"topics not found: %v\", notFound)\n\t}\n\n\treturn nil\n}", "func (s *source) fetch(consumerSession *consumerSession, doneFetch chan<- struct{}) (fetched bool) {\n\treq := s.createReq()\n\n\t// For all returns, if we do not buffer our fetch, then we want to\n\t// ensure our used offsets are usable again.\n\tvar alreadySentToDoneFetch bool\n\tvar buffered bool\n\tdefer func() {\n\t\tif !buffered {\n\t\t\tif req.numOffsets > 0 {\n\t\t\t\treq.usedOffsets.finishUsingAll()\n\t\t\t}\n\t\t\tif !alreadySentToDoneFetch {\n\t\t\t\tdoneFetch <- struct{}{}\n\t\t\t}\n\t\t}\n\t}()\n\n\tif req.numOffsets == 0 { // cursors could have been set unusable\n\t\treturn\n\t}\n\n\t// If our fetch is killed, we want to cancel waiting for the response.\n\tvar (\n\t\tkresp kmsg.Response\n\t\trequested = make(chan struct{})\n\t\tctx, cancel = context.WithCancel(consumerSession.ctx)\n\t)\n\tdefer cancel()\n\n\tbr, err := s.cl.brokerOrErr(ctx, s.nodeID, errUnknownBroker)\n\tif err != nil {\n\t\tclose(requested)\n\t} else {\n\t\tbr.do(ctx, req, func(k kmsg.Response, e error) {\n\t\t\tkresp, err = k, e\n\t\t\tclose(requested)\n\t\t})\n\t}\n\n\tselect {\n\tcase <-requested:\n\t\tfetched = true\n\tcase <-ctx.Done():\n\t\treturn\n\t}\n\n\t// If we had an error, we backoff. Killing a fetch quits the backoff,\n\t// but that is fine; we may just re-request too early and fall into\n\t// another backoff.\n\tif err != nil {\n\t\t// We preemptively allow more fetches (since we are not buffering)\n\t\t// and reset our session because of the error (who knows if kafka\n\t\t// processed the request but the client failed to receive it).\n\t\tdoneFetch <- struct{}{}\n\t\talreadySentToDoneFetch = true\n\t\ts.session.reset()\n\n\t\ts.cl.triggerUpdateMetadata(false) // as good a time as any\n\t\ts.consecutiveFailures++\n\t\tafter := time.NewTimer(s.cl.cfg.retryBackoff(s.consecutiveFailures))\n\t\tdefer after.Stop()\n\t\tselect {\n\t\tcase <-after.C:\n\t\tcase <-ctx.Done():\n\t\t}\n\t\treturn\n\t}\n\ts.consecutiveFailures = 0\n\n\tresp := kresp.(*kmsg.FetchResponse)\n\n\tvar (\n\t\tfetch Fetch\n\t\treloadOffsets listOrEpochLoads\n\t\tpreferreds cursorPreferreds\n\t\tupdateMeta bool\n\t\thandled = make(chan struct{})\n\t)\n\n\t// Theoretically, handleReqResp could take a bit of CPU time due to\n\t// decompressing and processing the response. We do this in a goroutine\n\t// to allow the session to be canceled at any moment.\n\t//\n\t// Processing the response only needs the source's nodeID and client.\n\tgo func() {\n\t\tdefer close(handled)\n\t\tfetch, reloadOffsets, preferreds, updateMeta = s.handleReqResp(req, resp)\n\t}()\n\n\tselect {\n\tcase <-handled:\n\tcase <-ctx.Done():\n\t\treturn\n\t}\n\n\t// The logic below here should be relatively quick.\n\n\tdeleteReqUsedOffset := func(topic string, partition int32) {\n\t\tt := req.usedOffsets[topic]\n\t\tdelete(t, partition)\n\t\tif len(t) == 0 {\n\t\t\tdelete(req.usedOffsets, topic)\n\t\t}\n\t}\n\n\t// Before updating the source, we move all cursors that have new\n\t// preferred replicas and remove them from being tracked in our req\n\t// offsets. We also remove the reload offsets from our req offsets.\n\t//\n\t// These two removals transition responsibility for finishing using the\n\t// cursor from the request's used offsets to the new source or the\n\t// reloading.\n\tpreferreds.eachPreferred(func(c cursorOffsetPreferred) {\n\t\tc.move()\n\t\tdeleteReqUsedOffset(c.from.topic, c.from.partition)\n\t})\n\treloadOffsets.each(deleteReqUsedOffset)\n\n\t// The session on the request was updated; we keep those updates.\n\ts.session = req.session\n\n\t// handleReqResp only parses the body of the response, not the top\n\t// level error code.\n\t//\n\t// The top level error code is related to fetch sessions only, and if\n\t// there was an error, the body was empty (so processing is basically a\n\t// no-op). We process the fetch session error now.\n\tswitch err := kerr.ErrorForCode(resp.ErrorCode); err {\n\tcase kerr.FetchSessionIDNotFound:\n\t\tif s.session.epoch == 0 {\n\t\t\t// If the epoch was zero, the broker did not even\n\t\t\t// establish a session for us (and thus is maxed on\n\t\t\t// sessions). We stop trying.\n\t\t\ts.cl.cfg.logger.Log(LogLevelInfo, \"session failed with SessionIDNotFound while trying to establish a session; broker likely maxed on sessions; continuing on without using sessions\", \"broker\", s.nodeID)\n\t\t\ts.session.kill()\n\t\t} else {\n\t\t\ts.cl.cfg.logger.Log(LogLevelInfo, \"received SessionIDNotFound from our in use session, our session was likely evicted; resetting session\", \"broker\", s.nodeID)\n\t\t\ts.session.reset()\n\t\t}\n\t\treturn\n\tcase kerr.InvalidFetchSessionEpoch:\n\t\ts.cl.cfg.logger.Log(LogLevelInfo, \"resetting fetch session\", \"broker\", s.nodeID, \"err\", err)\n\t\ts.session.reset()\n\t\treturn\n\t}\n\n\tif resp.Version < 7 {\n\t\t// If the version is less than 7, we cannot use fetch sessions,\n\t\t// so we kill them on the first response.\n\t\ts.session.kill()\n\t} else if resp.SessionID > 0 {\n\t\ts.session.bumpEpoch(resp.SessionID)\n\t}\n\n\t// If we moved any partitions to preferred replicas, we reset the\n\t// session. We do this after bumping the epoch just to ensure that we\n\t// have truly reset the session. (TODO switch to usingForgottenTopics)\n\tif len(preferreds) > 0 {\n\t\ts.session.reset()\n\t}\n\n\tif updateMeta && !reloadOffsets.loadWithSessionNow(consumerSession) {\n\t\ts.cl.triggerUpdateMetadataNow()\n\t}\n\n\tif len(fetch.Topics) > 0 {\n\t\tbuffered = true\n\t\ts.buffered = bufferedFetch{\n\t\t\tfetch: fetch,\n\t\t\tdoneFetch: doneFetch,\n\t\t\tusedOffsets: req.usedOffsets,\n\t\t}\n\t\ts.sem = make(chan struct{})\n\t\ts.cl.consumer.addSourceReadyForDraining(s)\n\t}\n\treturn\n}", "func (c *Collection) Topic(p types.Path) types.Path {\n\treturn types.PathFromString(c.Spec.Metadata.Name).Join(p)\n}", "func (c *Client) Topic(target, topic string) error {\n\treturn c.Raw(\"TOPIC %s :%s\", target, topic)\n}", "func (r *Reconciler) deleteTopic(ctx context.Context, topicName string) error {\n\n\t// Setup The Logger\n\tlogger := r.logger.With(zap.String(\"Topic\", topicName))\n\n\t// Attempt To Delete The Topic & Process Results\n\ttopicResults, err := r.adminClient.DeleteTopics(ctx, []string{topicName})\n\tif len(topicResults) > 0 {\n\t\ttopicResultError := topicResults[0].Error\n\t\ttopicResultErrorCode := topicResultError.Code()\n\t\tlogger = logger.With(zap.Any(\"topicResultErrorCode\", topicResultErrorCode))\n\t\tif topicResultErrorCode == kafka.ErrUnknownTopic ||\n\t\t\ttopicResultErrorCode == kafka.ErrUnknownPartition ||\n\t\t\ttopicResultErrorCode == kafka.ErrUnknownTopicOrPart {\n\n\t\t\tlogger.Info(\"Kafka Topic or Partition Not Found - No Deletion Required\")\n\t\t\treturn nil\n\t\t} else if topicResultErrorCode == kafka.ErrInvalidConfig && r.environment.KafkaProvider == env.KafkaProviderValueAzure {\n\t\t\t// While this could be a valid Kafka error, this most likely is coming from our custom EventHub AdminClient\n\t\t\t// implementation and represents the fact that the EventHub Cache does not contain this topic. This can\n\t\t\t// happen when an EventHub could not be created due to exceeding the number of allowable EventHubs. The\n\t\t\t// KafkaChannel is then in an \"UNKNOWN\" state having never been fully reconciled. We want to swallow this\n\t\t\t// error here so that the deletion of the Topic / EventHub doesn't block the deletion of the KafkaChannel.\n\t\t\tlogger.Warn(\"Invalid Kafka Topic Configuration (Likely EventHub Namespace Cache) - Unable To Delete\", zap.Error(err))\n\t\t\treturn nil\n\t\t} else if topicResultErrorCode == kafka.ErrNoError {\n\t\t\tlogger.Info(\"Successfully Deleted Existing Kafka Topic\")\n\t\t\treturn nil\n\t\t} else {\n\t\t\tlogger.Error(\"Failed To Delete Topic (Results)\", zap.Error(err), zap.Any(\"TopicResults\", topicResults))\n\t\t\treturn topicResults[0].Error\n\t\t}\n\t} else if err != nil {\n\t\tlogger.Error(\"Failed To Delete Topic (Error)\", zap.Error(err))\n\t\treturn err\n\t} else {\n\t\tlogger.Warn(\"Received Empty TopicResults From DeleteTopics Request\")\n\t\treturn nil\n\t}\n}", "func (c *ClusterInfo) GetLookupdTopicProducers(topic string, lookupdHTTPAddrs []string) (Producers, error) {\n\tvar producers Producers\n\tvar lock sync.Mutex\n\tvar wg sync.WaitGroup\n\tvar errs []error\n\n\ttype respType struct {\n\t\tProducers Producers `json:\"producers\"`\n\t}\n\n\tfor _, addr := range lookupdHTTPAddrs {\n\t\twg.Add(1)\n\t\tgo func(addr string) {\n\t\t\tdefer wg.Done()\n\n\t\t\tendpoint := fmt.Sprintf(\"http://%s/lookup?topic=%s\", addr, url.QueryEscape(topic))\n\t\t\tc.logf(\"CI: querying nsqlookupd %s\", endpoint)\n\n\t\t\tvar resp respType\n\t\t\terr := c.client.GETV1(endpoint, &resp)\n\t\t\tif err != nil {\n\t\t\t\tlock.Lock()\n\t\t\t\terrs = append(errs, err)\n\t\t\t\tlock.Unlock()\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tlock.Lock()\n\t\t\tdefer lock.Unlock()\n\t\t\tfor _, p := range resp.Producers {\n\t\t\t\tfor _, pp := range producers {\n\t\t\t\t\tif p.HTTPAddress() == pp.HTTPAddress() {\n\t\t\t\t\t\tgoto skip\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tproducers = append(producers, p)\n\t\t\tskip:\n\t\t\t}\n\t\t}(addr)\n\t}\n\twg.Wait()\n\n\tif len(errs) == len(lookupdHTTPAddrs) {\n\t\treturn nil, fmt.Errorf(\"failed to query any nsqlookupd: %s\", ErrList(errs))\n\t}\n\tif len(errs) > 0 {\n\t\treturn producers, ErrList(errs)\n\t}\n\treturn producers, nil\n}", "func Retrieve(keyword string, num int, domain string) []string {\n\toutput := make([]string, num)\n\turl := fmt.Sprintf(\"%s?q=%s&format=json\", domain, keyword)\n\tresponse, err := http.Get(url)\n\tif err != nil {\n\t\tfmt.Printf(\"Error retrieving results %s\\n\\n\", err)\n\t} else {\n\t\tdata, _ := ioutil.ReadAll(response.Body)\n\t\tvar result Result\n\t\tjson.Unmarshal([]byte(string(data)), &result)\n\t\tfor i, s := range result.RelatedTopics {\n\t\t\tif i == num-1 {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\toutput[i] = s.FirstURL\n\t\t}\n\t}\n\treturn output\n\n}", "func (tl TopicList) Find(name string) *Topic {\n\tfor _, topic := range tl {\n\t\tif topic.Name == name {\n\t\t\treturn topic\n\t\t}\n\t}\n\treturn nil\n}", "func (m *Mqtt) subscribe(topic string) error {\n\tm.token = m.client.Subscribe(topic, 0, nil)\n\tm.token.Wait()\n\treturn m.token.Error()\n}", "func (c *Consumer) ReachedEndOfTopic() <-chan struct{} {\n\treturn c.endOfTopicc\n}", "func (d *RetryDownloader) GetTopics() ([]string, error) {\n\tvar (\n\t\ttopics []string\n\t\terr error\n\t)\n\n\terr = d.retry(func() error {\n\t\ttopics, err = d.Downloader.GetTopics()\n\t\treturn err\n\t})\n\n\treturn topics, err\n}", "func (local *Node) Fetch(key string) (isRoot bool, replicas []RemoteNode) {\n\t// TODO: students should implement this\n\treturn\n}", "func (c *Client) Topic(id string) *Topic {\n\treturn &Topic{c.Client.Topic(id), c.projectID, c.sensor}\n}", "func SNSSQSNonexistingTopic(t *testing.T) {\n\tconsumerGroup1 := watcher.NewUnordered()\n\n\t// Set the partition key on all messages so they are written to the same partition. This allows for checking of ordered messages.\n\tmetadata := map[string]string{\n\t\tmessageKey: partition0,\n\t}\n\n\t// subscriber of the given topic\n\tsubscriberApplication := func(appID string, topicName string, messagesWatcher *watcher.Watcher) app.SetupFn {\n\t\treturn func(ctx flow.Context, s common.Service) error {\n\t\t\t// Simulate periodic errors.\n\t\t\tsim := simulate.PeriodicError(ctx, 100)\n\t\t\t// Setup the /orders event handler.\n\t\t\treturn multierr.Combine(\n\t\t\t\ts.AddTopicEventHandler(&common.Subscription{\n\t\t\t\t\tPubsubName: pubsubName,\n\t\t\t\t\tTopic: topicName,\n\t\t\t\t\tRoute: \"/orders\",\n\t\t\t\t}, func(_ context.Context, e *common.TopicEvent) (retry bool, err error) {\n\t\t\t\t\tif err := sim(); err != nil {\n\t\t\t\t\t\treturn true, err\n\t\t\t\t\t}\n\n\t\t\t\t\t// Track/Observe the data of the event.\n\t\t\t\t\tmessagesWatcher.Observe(e.Data)\n\t\t\t\t\tctx.Logf(\"Message Received appID: %s,pubsub: %s, topic: %s, id: %s, data: %s\", appID, e.PubsubName, e.Topic, e.ID, e.Data)\n\t\t\t\t\treturn false, nil\n\t\t\t\t}),\n\t\t\t)\n\t\t}\n\t}\n\n\tpublishMessages := func(metadata map[string]string, sidecarName string, topicName string, messageWatchers ...*watcher.Watcher) flow.Runnable {\n\t\treturn func(ctx flow.Context) error {\n\t\t\t// prepare the messages\n\t\t\tmessages := make([]string, numMessages)\n\t\t\tfor i := range messages {\n\t\t\t\tmessages[i] = fmt.Sprintf(\"partitionKey: %s, message for topic: %s, index: %03d, uniqueId: %s\", metadata[messageKey], topicName, i, uuid.New().String())\n\t\t\t}\n\n\t\t\t// add the messages as expectations to the watchers\n\t\t\tfor _, messageWatcher := range messageWatchers {\n\t\t\t\tmessageWatcher.ExpectStrings(messages...)\n\t\t\t}\n\n\t\t\t// get the sidecar (dapr) client\n\t\t\tclient := sidecar.GetClient(ctx, sidecarName)\n\n\t\t\t// publish messages\n\t\t\tctx.Logf(\"Publishing messages. sidecarName: %s, topicName: %s\", sidecarName, topicName)\n\n\t\t\tvar publishOptions dapr.PublishEventOption\n\n\t\t\tif metadata != nil {\n\t\t\t\tpublishOptions = dapr.PublishEventWithMetadata(metadata)\n\t\t\t}\n\n\t\t\tfor _, message := range messages {\n\t\t\t\tctx.Logf(\"Publishing: %q\", message)\n\t\t\t\tvar err error\n\n\t\t\t\tif publishOptions != nil {\n\t\t\t\t\terr = client.PublishEvent(ctx, pubsubName, topicName, message, publishOptions)\n\t\t\t\t} else {\n\t\t\t\t\terr = client.PublishEvent(ctx, pubsubName, topicName, message)\n\t\t\t\t}\n\t\t\t\trequire.NoError(ctx, err, \"SNSSQSNonexistingTopic - error publishing message\")\n\t\t\t}\n\t\t\treturn nil\n\t\t}\n\t}\n\n\tassertMessages := func(timeout time.Duration, messageWatchers ...*watcher.Watcher) flow.Runnable {\n\t\treturn func(ctx flow.Context) error {\n\t\t\t// assert for messages\n\t\t\tfor _, m := range messageWatchers {\n\t\t\t\tif !m.Assert(ctx, 25*timeout) {\n\t\t\t\t\tctx.Errorf(\"SNSSQSNonexistingTopic - message assertion failed for watcher: %#v\\n\", m)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\treturn nil\n\t\t}\n\t}\n\n\tflow.New(t, \"SNSSQS certification - non-existing topic\").\n\n\t\t// Run subscriberApplication app1\n\t\tStep(app.Run(appID1, fmt.Sprintf(\":%d\", appPort+portOffset*3),\n\t\t\tsubscriberApplication(appID1, topicToBeCreated, consumerGroup1))).\n\n\t\t// Run the Dapr sidecar with ConsumerID \"PUBSUB_AWS_SNSSQS_QUEUE_1\"\n\t\tStep(sidecar.Run(sidecarName1,\n\t\t\tappend(componentRuntimeOptions(),\n\t\t\t\tembedded.WithComponentsPath(\"./components/consumer_one\"),\n\t\t\t\tembedded.WithAppProtocol(protocol.HTTPProtocol, strconv.Itoa(appPort+portOffset*3)),\n\t\t\t\tembedded.WithDaprGRPCPort(strconv.Itoa(runtime.DefaultDaprAPIGRPCPort+portOffset*3)),\n\t\t\t\tembedded.WithDaprHTTPPort(strconv.Itoa(runtime.DefaultDaprHTTPPort+portOffset*3)),\n\t\t\t\tembedded.WithProfilePort(strconv.Itoa(runtime.DefaultProfilePort+portOffset*3)),\n\t\t\t)...,\n\t\t)).\n\t\tStep(fmt.Sprintf(\"publish messages to topicToBeCreated: %s\", topicToBeCreated), publishMessages(metadata, sidecarName1, topicToBeCreated, consumerGroup1)).\n\t\tStep(\"wait\", flow.Sleep(30*time.Second)).\n\t\tStep(\"verify if app1 has recevied messages published to newly created topic\", assertMessages(10*time.Second, consumerGroup1)).\n\t\tRun()\n}", "func recursive(fullTopic string) (topic string, isRecursive bool) {\n\t// Legacy: topics ending in ?recursive are recursive.\n\tif strings.HasSuffix(fullTopic, \"?recursive\") {\n\t\treturn fullTopic[:len(fullTopic)-len(\"?recursive\")], true\n\t}\n\n\t// Topics ending in \"/\" are recursive\n\tif strings.HasSuffix(fullTopic, \"/\") {\n\t\treturn fullTopic, true\n\t}\n\n\treturn fullTopic, false\n}", "func ParseTopicName(topic string) (*TopicObject, error) {\n\tfullName := topic\n\tif !strings.Contains(topic, \"://\") {\n\t\tpathTokens := strings.Split(topic, \"/\")\n\t\tif len(pathTokens) == 3 {\n\t\t\tfullName = \"persistent://\" + pathTokens[0] + \"/\" + pathTokens[1] + \"/\" + pathTokens[2]\n\t\t} else if len(pathTokens) == 1 {\n\t\t\tfullName = \"persistent://public/default/\" + pathTokens[0]\n\t\t} else {\n\t\t\treturn nil, errors.New(\"invalid topic\")\n\t\t}\n\t}\n\n\toriginFullName := fullName\n\tfullName = strings.Replace(fullName, \"://\", \"/\", -1)\n\tpathTokens := strings.Split(fullName, \"/\")\n\tif len(pathTokens) < 4 {\n\t\treturn nil, errors.New(\"invalid topic t\")\n\t}\n\n\tvar domain, property, cluster, namespacePortion, localName string\n\tvar isV2Topic bool\n\tvar numSlashIndexes uint32\n\tdomain = pathTokens[0]\n\tif len(pathTokens) == 4 {\n\t\t// New topic t without cluster t\n\t\tproperty = pathTokens[1]\n\t\tcluster = \"\"\n\t\tnamespacePortion = pathTokens[2]\n\t\tlocalName = pathTokens[3]\n\t\tisV2Topic = true\n\t} else {\n\t\t// Legacy topic t that includes cluster t\n\t\tproperty = pathTokens[1]\n\t\tcluster = pathTokens[2]\n\t\tnamespacePortion = pathTokens[3]\n\t\tlocalName = pathTokens[4]\n\t\tnumSlashIndexes = 4\n\t\tisV2Topic = false\n\t\tslashIndex := strings.IndexFunc(fullName, func(c rune) bool {\n\t\t\tif numSlashIndexes <= 0 {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\tif c == '/' {\n\t\t\t\tnumSlashIndexes--\n\t\t\t}\n\t\t\treturn false\n\t\t})\n\t\tlocalName = fullName[slashIndex:]\n\t}\n\n\treturn &TopicObject{\n\t\tFullName: originFullName,\n\t\tDomain: domain,\n\t\tProperty: property,\n\t\tCluster: cluster,\n\t\tNamespacePortion: namespacePortion,\n\t\tLocalName: localName,\n\t\tIsV2Topic: isV2Topic,\n\t}, nil\n}", "func GetTopicName(completeName string) (*TopicName, error) {\n\tvar topicName TopicName\n\t// The short topic name can be:\n\t// - <topic>\n\t// - <tenant>/<namespace>/<topic>\n\tif !strings.Contains(completeName, \"://\") {\n\t\tparts := strings.Split(completeName, \"/\")\n\t\tswitch len(parts) {\n\t\tcase 3:\n\t\t\tcompleteName = persistent.String() + \"://\" + completeName\n\t\tcase 1:\n\t\t\tcompleteName = persistent.String() + \"://\" + PUBLICTENANT + \"/\" + DEFAULTNAMESPACE + \"/\" + parts[0]\n\t\tdefault:\n\t\t\treturn nil, errors.Errorf(\"Invalid short topic name '%s', it should be \"+\n\t\t\t\t\"in the format of <tenant>/<namespace>/<topic> or <topic>\", completeName)\n\t\t}\n\t}\n\n\t// The fully qualified topic name can be:\n\t// <domain>://<tenant>/<namespace>/<topic>\n\n\tparts := strings.SplitN(completeName, \"://\", 2)\n\n\tdomain, err := ParseTopicDomain(parts[0])\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttopicName.domain = domain\n\n\trest := parts[1]\n\tparts = strings.SplitN(rest, \"/\", 3)\n\tif len(parts) == 3 {\n\t\ttopicName.tenant = parts[0]\n\t\ttopicName.namespace = parts[1]\n\t\ttopicName.topic = parts[2]\n\t\ttopicName.partitionIndex = getPartitionIndex(completeName)\n\t} else {\n\t\treturn nil, errors.Errorf(\"invalid topic name '%s', it should be in the format of \"+\n\t\t\t\"<tenant>/<namespace>/<topic>\", rest)\n\t}\n\n\tif topicName.topic == \"\" {\n\t\treturn nil, errors.New(\"topic name can not be empty\")\n\t}\n\n\tn, err := GetNameSpaceName(topicName.tenant, topicName.namespace)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttopicName.namespaceName = n\n\n\treturn &topicName, nil\n}", "func (this *UnsubscribeMessage) TopicExists(topic []byte) bool {\n\tfor _, t := range this.topics {\n\t\tif bytes.Equal(t, topic) {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (p *metadataService) onTopicUnsubscribe(e event.Event) {\n}", "func (t *topicsPartitions) ensureTopics(topics []string) topicsPartitionsData {\n\tvar cloned bool\n\tcurrent := t.load()\n\tfor _, topic := range topics {\n\t\tif _, exists := current[topic]; !exists {\n\t\t\tif !cloned {\n\t\t\t\tcurrent = t.clone()\n\t\t\t\tcloned = true\n\t\t\t}\n\t\t\tcurrent[topic] = newTopicPartitions()\n\t\t}\n\t}\n\treturn current\n}", "func (d *directConsumer) findNewAssignments() map[string]map[int32]Offset {\n\ttopics := d.tps.load()\n\n\ttoUse := make(map[string]map[int32]Offset, 10)\n\tfor topic, topicPartitions := range topics {\n\t\t// If we are using regex topics, we have to check all\n\t\t// topic regexes to see if any match on this topic.\n\t\tvar useTopic bool\n\t\tif d.cfg.regex {\n\t\t\twant, seen := d.reSeen[topic]\n\t\t\tif !seen {\n\t\t\t\tfor _, re := range d.cfg.topics {\n\t\t\t\t\tif want = re.MatchString(topic); want {\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\td.reSeen[topic] = want\n\t\t\t}\n\t\t\tuseTopic = want\n\t\t} else {\n\t\t\t_, useTopic = d.cfg.topics[topic]\n\t\t}\n\n\t\t// If the above detected that we want to keep this topic, we\n\t\t// set all partitions as usable.\n\t\t//\n\t\t// For internal partitions, we only allow consuming them if\n\t\t// the topic is explicitly specified.\n\t\tif useTopic {\n\t\t\tpartitions := topicPartitions.load()\n\t\t\tif d.cfg.regex && partitions.isInternal {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\ttoUseTopic := make(map[int32]Offset, len(partitions.partitions))\n\t\t\tfor partition := range partitions.partitions {\n\t\t\t\ttoUseTopic[int32(partition)] = d.cfg.resetOffset\n\t\t\t}\n\t\t\ttoUse[topic] = toUseTopic\n\t\t}\n\n\t\t// Lastly, if this topic has some specific partitions pinned,\n\t\t// we set those.\n\t\tfor partition, offset := range d.cfg.partitions[topic] {\n\t\t\ttoUseTopic, exists := toUse[topic]\n\t\t\tif !exists {\n\t\t\t\ttoUseTopic = make(map[int32]Offset, 10)\n\t\t\t\ttoUse[topic] = toUseTopic\n\t\t\t}\n\t\t\ttoUseTopic[partition] = offset\n\t\t}\n\t}\n\n\t// With everything we want to consume, remove what we are already.\n\tfor topic, partitions := range d.using {\n\t\ttoUseTopic, exists := toUse[topic]\n\t\tif !exists {\n\t\t\tcontinue // metadata update did not return this topic (regex or failing load)\n\t\t}\n\t\tfor partition := range partitions {\n\t\t\tdelete(toUseTopic, partition)\n\t\t}\n\t\tif len(toUseTopic) == 0 {\n\t\t\tdelete(toUse, topic)\n\t\t}\n\t}\n\n\tif len(toUse) == 0 {\n\t\treturn nil\n\t}\n\n\t// Finally, toUse contains new partitions that we must consume.\n\t// Add them to our using map and assign them.\n\tfor topic, partitions := range toUse {\n\t\ttopicUsing, exists := d.using[topic]\n\t\tif !exists {\n\t\t\ttopicUsing = make(map[int32]struct{})\n\t\t\td.using[topic] = topicUsing\n\t\t}\n\t\tfor partition := range partitions {\n\t\t\ttopicUsing[partition] = struct{}{}\n\t\t}\n\t}\n\n\treturn toUse\n}", "func (c *ClusterInfo) GetNSQDTopicProducers(topic string, nsqdHTTPAddrs []string) (Producers, error) {\n\tvar producers Producers\n\tvar lock sync.Mutex\n\tvar wg sync.WaitGroup\n\tvar errs []error\n\n\ttype infoRespType struct {\n\t\tVersion string `json:\"version\"`\n\t\tBroadcastAddress string `json:\"broadcast_address\"`\n\t\tHostname string `json:\"hostname\"`\n\t\tHTTPPort int `json:\"http_port\"`\n\t\tTCPPort int `json:\"tcp_port\"`\n\t}\n\n\ttype statsRespType struct {\n\t\tTopics []struct {\n\t\t\tName string `json:\"topic_name\"`\n\t\t} `json:\"topics\"`\n\t}\n\n\tfor _, addr := range nsqdHTTPAddrs {\n\t\twg.Add(1)\n\t\tgo func(addr string) {\n\t\t\tdefer wg.Done()\n\n\t\t\tendpoint := fmt.Sprintf(\"http://%s/stats?format=json&topic=%s&include_clients=false\",\n\t\t\t\taddr, url.QueryEscape(topic))\n\t\t\tc.logf(\"CI: querying nsqd %s\", endpoint)\n\n\t\t\tvar statsResp statsRespType\n\t\t\terr := c.client.GETV1(endpoint, &statsResp)\n\t\t\tif err != nil {\n\t\t\t\tlock.Lock()\n\t\t\t\terrs = append(errs, err)\n\t\t\t\tlock.Unlock()\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tvar producerTopics ProducerTopics\n\t\t\tfor _, t := range statsResp.Topics {\n\t\t\t\tproducerTopics = append(producerTopics, ProducerTopic{Topic: t.Name})\n\t\t\t}\n\n\t\t\tfor _, t := range statsResp.Topics {\n\t\t\t\tif t.Name == topic {\n\t\t\t\t\tendpoint := fmt.Sprintf(\"http://%s/info\", addr)\n\t\t\t\t\tc.logf(\"CI: querying nsqd %s\", endpoint)\n\n\t\t\t\t\tvar infoResp infoRespType\n\t\t\t\t\terr := c.client.GETV1(endpoint, &infoResp)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlock.Lock()\n\t\t\t\t\t\terrs = append(errs, err)\n\t\t\t\t\t\tlock.Unlock()\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\n\t\t\t\t\tversion, err := semver.Parse(infoResp.Version)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tversion, _ = semver.Parse(\"0.0.0\")\n\t\t\t\t\t}\n\n\t\t\t\t\t// if BroadcastAddress/HTTPPort are missing, use the values from `addr` for\n\t\t\t\t\t// backwards compatibility\n\n\t\t\t\t\tif infoResp.BroadcastAddress == \"\" {\n\t\t\t\t\t\tvar p string\n\t\t\t\t\t\tinfoResp.BroadcastAddress, p, _ = net.SplitHostPort(addr)\n\t\t\t\t\t\tinfoResp.HTTPPort, _ = strconv.Atoi(p)\n\t\t\t\t\t}\n\t\t\t\t\tif infoResp.Hostname == \"\" {\n\t\t\t\t\t\tinfoResp.Hostname, _, _ = net.SplitHostPort(addr)\n\t\t\t\t\t}\n\n\t\t\t\t\tlock.Lock()\n\t\t\t\t\tproducers = append(producers, &Producer{\n\t\t\t\t\t\tVersion: infoResp.Version,\n\t\t\t\t\t\tVersionObj: version,\n\t\t\t\t\t\tBroadcastAddress: infoResp.BroadcastAddress,\n\t\t\t\t\t\tHostname: infoResp.Hostname,\n\t\t\t\t\t\tHTTPPort: infoResp.HTTPPort,\n\t\t\t\t\t\tTCPPort: infoResp.TCPPort,\n\t\t\t\t\t\tTopics: producerTopics,\n\t\t\t\t\t})\n\t\t\t\t\tlock.Unlock()\n\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}(addr)\n\t}\n\twg.Wait()\n\n\tif len(errs) == len(nsqdHTTPAddrs) {\n\t\treturn nil, fmt.Errorf(\"failed to query any nsqd: %s\", ErrList(errs))\n\t}\n\tif len(errs) > 0 {\n\t\treturn producers, ErrList(errs)\n\t}\n\treturn producers, nil\n}", "func (ctx *TemplateContext) Topic(index int) (string, error) {\n\tif index >= len(ctx.Parts) {\n\t\treturn \"\", errors.New(\"Topic index out of range\")\n\t}\n\n\treturn ctx.Parts[index], nil\n}", "func (s *Store) lookupPrecedingReplica(key roachpb.RKey) *Replica {\n\ts.mu.RLock()\n\tdefer s.mu.RUnlock()\n\treturn s.mu.replicasByKey.LookupPrecedingReplica(context.Background(), key)\n}", "func (tm *topicManager) createNewTopic(name string, id string) (log, error) {\n\tlc, ok := tm.topicNamesToLogs[name]\n\tif ok {\n\t\tlc.refCount++\n\t\treturn lc.l, nil\n\t}\n\n\t// lazily create the log if we know about the topic/cc ID\n\tl, err := newBadgerLog(tm.dataDir, name, id, tm.cc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttm.topicNamesToLogs[name] = &logContainer{\n\t\tl: l,\n\t\trefCount: 1,\n\t}\n\treturn l, nil\n}", "func (c *KafkaCluster) getTopics() []string {\n\tc.lock.RLock()\n\tdefer c.lock.RUnlock()\n\n\ttopics := make([]string, 0, len(c.topics))\n\tfor topic := range c.topics {\n\t\ttopics = append(topics, topic)\n\t}\n\treturn topics\n}" ]
[ "0.63564193", "0.6155337", "0.6141741", "0.59800786", "0.5890245", "0.5821753", "0.5762925", "0.57450515", "0.57057124", "0.55477715", "0.5542679", "0.5485723", "0.5468279", "0.54534346", "0.54458946", "0.5443599", "0.54278564", "0.5422039", "0.54053885", "0.53741664", "0.5373013", "0.53489786", "0.5345589", "0.5301546", "0.52823603", "0.526613", "0.5253322", "0.52446496", "0.5240215", "0.5228584", "0.51953423", "0.51948506", "0.5181991", "0.5171827", "0.51449525", "0.5131248", "0.51309246", "0.5115074", "0.5091221", "0.508865", "0.5085444", "0.50830066", "0.507734", "0.50761133", "0.50644404", "0.50627023", "0.50547206", "0.5048112", "0.5042066", "0.5025549", "0.5019231", "0.50188285", "0.501676", "0.50008816", "0.4998215", "0.49922428", "0.4984152", "0.49832958", "0.49720785", "0.49627388", "0.49524695", "0.49335414", "0.4928852", "0.49266085", "0.4924453", "0.4915268", "0.49115738", "0.49066764", "0.48977768", "0.4897356", "0.4892903", "0.488258", "0.48755708", "0.48751184", "0.48750865", "0.4874837", "0.4869195", "0.48567012", "0.48542944", "0.4851", "0.48474574", "0.48472714", "0.48405868", "0.48399174", "0.4838859", "0.48309603", "0.48295006", "0.48261833", "0.48178884", "0.48155996", "0.48114833", "0.48078027", "0.4804379", "0.47948158", "0.47938636", "0.47930303", "0.47929746", "0.47892952", "0.47832862", "0.4782743", "0.47815058" ]
0.0
-1
GetFileContentType : Gets file content type
func GetFileContentType(out *os.File) (string, error) { // Only the first 512 bytes are used to sniff the content type. buffer := make([]byte, 512) _, err := out.Read(buffer) if err != nil { return "", err } // Use the net/http package's handy DectectContentType function. Always returns a valid // content-type by returning "application/octet-stream" if no others seemed to match. contentType := http.DetectContentType(buffer) return contentType, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func getFileContentType(out io.Reader) (string, error) {\n\t// Only the first 512 bytes are used to sniff the content type.\n\tbuffer := make([]byte, 512)\n\n\t_, err := out.Read(buffer)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Use the net/http package's handy DectectContentType function. Always returns a valid\n\t// content-type by returning \"application/octet-stream\" if no others seemed to match.\n\tcontentType := http.DetectContentType(buffer)\n\n\treturn contentType, nil\n}", "func (pstFile *File) GetContentType() ([]byte, error) {\n\tcontentType, err := pstFile.Read(2, 8)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif bytes.Equal(contentType, ContentTypePST) {\n\t\treturn ContentTypePST, nil\n\t} else if bytes.Equal(contentType, ContentTypeOST) {\n\t\treturn ContentTypeOST, nil\n\t} else if bytes.Equal(contentType, ContentTypePAB) {\n\t\treturn ContentTypePAB, nil\n\t} else {\n\t\treturn nil, errors.New(\"unsupported content type\")\n\t}\n}", "func (lf *localFile) ContentType() string {\n\tif lf.matcher != nil && lf.matcher.ContentType != \"\" {\n\t\treturn lf.matcher.ContentType\n\t}\n\n\text := filepath.Ext(lf.NativePath)\n\tif mimeType, _, found := lf.mediaTypes.GetFirstBySuffix(strings.TrimPrefix(ext, \".\")); found {\n\t\treturn mimeType.Type\n\t}\n\n\treturn mime.TypeByExtension(ext)\n}", "func ContentType(absoluteFilePath string) (string, error) {\n\tout, err := os.Open(path.Clean(absoluteFilePath))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer out.Close()\n\n\t// Only the first 512 bytes are used to sniff the content type.\n\tbuffer := make([]byte, 512)\n\n\t_, err = out.Read(buffer)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Use the net/http package's handy DectectContentType function. Always returns a valid\n\t// content-type by returning \"application/octet-stream\" if no others seemed to match.\n\tcontentType := http.DetectContentType(buffer)\n\n\treturn contentType, nil\n}", "func GetContentTypeByFileName(name string) string {\n\tswitch strings.ToLower(filepath.Ext(name)) {\n\tcase \".zip\":\n\t\treturn ContentType.ApplicationZip\n\tcase \".csv\":\n\t\treturn ContentType.TextCsv\n\tdefault:\n\t\treturn ContentType.TextPlain // if no extension found\n\t}\n}", "func GetType(filename string) (string, error) {\n\tfile, err := os.Open(filename)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer file.Close()\n\n\tbuf := make([]byte, 512)\n\t_, err = file.Read(buf)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tfiletype := http.DetectContentType(buf)\n\tfor _, ext := range ext {\n\t\tif strings.Contains(ext, filetype[6:]) { // like image/jpeg\n\t\t\treturn filetype, nil\n\t\t}\n\t}\n\n\treturn \"\", errors.New(\"invalid image type\")\n}", "func GetContentType(r *http.Request) string {\n\treturn filteFlag(r.Header.Get(\"Content-Type\"))\n}", "func (rf *HTTPResFile) MediaType() string {\n\t// TODO (b5) - this is super hacky\n\treturn strings.Split(rf.res.Header.Get(\"Content-Type\"), \";\")[0]\n}", "func getMimeType(file string) string {\n\tmm, err := magicmime.New(magicmime.MAGIC_MIME_TYPE | magicmime.MAGIC_SYMLINK | magicmime.MAGIC_ERROR)\n\tif err != nil {\n\t\tnumErrors++\n\t\treturn DefaultMimeType\n\t}\n\n\tmimetype, err := mm.TypeByFile(file)\n\tif err != nil {\n\t\tnumErrors++\n\t\treturn DefaultMimeType\n\t}\n\n\treturn mimetype\n}", "func getFile(fileName string) (bytes []byte, filetype string) {\n\tfile, _ := os.Open(fileName)\n\tdefer file.Close()\n\n\tfileInfo, _ := file.Stat()\n\tsize := fileInfo.Size()\n\tbytes = make([]byte, size)\n\n\tbuffer := bufio.NewReader(file)\n\tbuffer.Read(bytes)\n\n\tfiletype = http.DetectContentType(bytes)\n\treturn\n}", "func (m *AttachmentItem) GetContentType()(*string) {\n return m.contentType\n}", "func (p profiler) ContentType() string {\n\tif p.debug > 0 && p.which != \"profile\" {\n\t\treturn \"text/plain; charset=utf-8\"\n\t}\n\treturn \"application/octet-stream\"\n}", "func (promise AttachmentPromise) GetContentType() string {\n\tif ret, ok := promise.params[\"content-type\"]; ok {\n\t\treturn ret\n\t}\n\treturn \"\"\n}", "func (m *ChatMessageAttachment) GetContentType()(*string) {\n val, err := m.GetBackingStore().Get(\"contentType\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func filterContentType(info tusd.FileInfo) (contentType string, contentDisposition string) {\n filetype := info.MetaData[\"filetype\"]\n\n if reMimeType.MatchString(filetype) {\n // If the filetype from metadata is well formed, we forward use this\n // for the Content-Type header. However, only whitelisted mime types\n // will be allowed to be shown inline in the browser\n contentType = filetype\n if _, isWhitelisted := mimeInlineBrowserWhitelist[filetype]; isWhitelisted {\n contentDisposition = \"inline\"\n } else {\n contentDisposition = \"attachment\"\n }\n } else {\n // If the filetype from the metadata is not well formed, we use a\n // default type and force the browser to download the content.\n contentType = \"application/octet-stream\"\n contentDisposition = \"attachment\"\n }\n\n // Add a filename to Content-Disposition if one is available in the metadata\n if filename, ok := info.MetaData[\"filename\"]; ok {\n contentDisposition += \";filename=\" + strconv.Quote(filename)\n }\n\n return contentType, contentDisposition\n}", "func FileMimeType() string {\n\treturn getRandValue([]string{\"file\", \"mime_type\"})\n}", "func getContentType(r *http.Request) string {\n\theaderValue := r.Header.Get(\"Content-Type\")\n\tif headerValue == \"\" {\n\t\treturn \"\"\n\t}\n\n\theaderValueSplit := strings.Split(headerValue, \";\")\n\treturn headerValueSplit[0]\n}", "func (lfs ListFilesystemSchema) ContentType() string {\n\treturn lfs.rawResponse.Header.Get(\"Content-Type\")\n}", "func getMimeType(filePath string) string {\n\tfileExt := filepath.Ext(filePath)\n\tdefaultMime := \"application/octet-stream\"\n\tif len(fileExt) != 0 {\n\t\tmimeType := mime.TypeByExtension(fileExt)\n\t\tif len(mimeType) != 0 {\n\t\t\treturn mimeType\n\t\t}\n\t}\n\treturn defaultMime\n}", "func (info *Info) GetMimeType() string {\n\tv := info.Attributes[\"content-type\"]\n\tif v != nil {\n\t\treturn v.(string)\n\t}\n\treturn \"application/octet-stream\"\n}", "func (o FileContentBufferResponseOutput) FileType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FileContentBufferResponse) string { return v.FileType }).(pulumi.StringOutput)\n}", "func (c *Context) ContentType() string {\n\treturn filterFlags(c.GetHeader(\"Content-Type\"))\n}", "func (c *Context) ContentType() string {\n\treturn filterFlags(c.requestHeader(\"Content-Type\"))\n}", "func (_BaseAccessControlGroup *BaseAccessControlGroupCaller) GetContentType(opts *bind.CallOpts, position *big.Int) (common.Address, error) {\n\tvar out []interface{}\n\terr := _BaseAccessControlGroup.contract.Call(opts, &out, \"getContentType\", position)\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (m *PrinterDefaults) GetContentType()(*string) {\n val, err := m.GetBackingStore().Get(\"contentType\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (sct *ContentSniffer) ContentType() (string, bool) {\n\tif sct.sniffed {\n\t\treturn sct.ctype, sct.ctype != \"\"\n\t}\n\tsct.sniffed = true\n\t// If ReadAll hits EOF, it returns err==nil.\n\tsct.start, sct.err = ioutil.ReadAll(io.LimitReader(sct.r, sniffBuffSize))\n\n\t// Don't try to detect the content type based on possibly incomplete data.\n\tif sct.err != nil {\n\t\treturn \"\", false\n\t}\n\n\tsct.ctype = http.DetectContentType(sct.start)\n\treturn sct.ctype, true\n}", "func (dr downloadResponse) ContentType() string {\n\treturn dr.rawResponse.Header.Get(\"Content-Type\")\n}", "func GetContentType(str string) (ContentType, error) {\n\tmediaType, _, err := mime.ParseMediaType(str)\n\treturn ContentType(mediaType), err\n}", "func (req *Request) GetContentType() string {\n\treturn req.Req.Header.Get(\"Content-Type\")\n}", "func MimeType(path string) (mime string) {\n\tfile, err := os.Open(path)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn ReaderMimeType(file)\n}", "func (req *DronaRequest) GetContentType() string {\n\treq.Lock()\n\tdefer req.Unlock()\n\treturn req.contentType\n}", "func (b *Buffer) FileType() string {\n\treturn b.Settings[\"filetype\"].(string)\n}", "func (o FileContentBufferResponsePtrOutput) FileType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FileContentBufferResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.FileType\n\t}).(pulumi.StringPtrOutput)\n}", "func (r *ReleaseAsset) GetContentType() string {\n\tif r == nil || r.ContentType == nil {\n\t\treturn \"\"\n\t}\n\treturn *r.ContentType\n}", "func (m *Gzip) GetContentType() []string {\n\tif m != nil {\n\t\treturn m.ContentType\n\t}\n\treturn nil\n}", "func (me TxsdImpactSimpleContentExtensionType) IsFile() bool { return me.String() == \"file\" }", "func (p *Part) ContentType() string {\n\tctype := C.gmime_get_content_type_string(p.gmimePart)\n\tdefer C.g_free(C.gpointer(unsafe.Pointer(ctype)))\n\treturn C.GoString(ctype)\n}", "func (m MediaType) ContentType() string {\n\tif len(m.Type) > 0 && m.Charset != \"\" {\n\t\treturn fmt.Sprintf(\"%s; charset=%s\", m.Type, m.Charset)\n\t}\n\treturn m.Type\n}", "func MimeType() string {\n\treturn getRandValue([]string{\"file\", \"mime_type\"})\n}", "func GetMimeType(imageFormat string) (string, error) {\n\tswitch imageFormat {\n\tcase \"png\":\n\t\treturn \"image/png\", nil\n\tcase \"gif\":\n\t\treturn \"image/gif\", nil\n\tcase \"jpg\":\n\tcase \"jpeg\":\n\t\treturn \"image/jpeg\", nil\n\t}\n\treturn \"\", fmt.Errorf(\"Unknown image format: \", imageFormat)\n}", "func (rpr ReadPathResponse) ContentType() string {\n\treturn rpr.rawResponse.Header.Get(\"Content-Type\")\n}", "func (json Json) getContentType() (contentType string) {\n\treturn json.contentType\n}", "func (o *Object) MimeType(ctx context.Context) string {\n\treturn o.contentType\n}", "func MimeTypeFromFilename(filename string) string {\n\text := path.Ext(filename)\n\tswitch ext {\n\tcase \".js\":\n\t\treturn \"application/javascript\"\n\tcase \".html\":\n\t\treturn \"text/html; charset=utf-8\"\n\tcase \".css\":\n\t\treturn \"text/css; charset=utf-8\"\n\t}\n\treturn \"text/plain; charset=utf-8\"\n}", "func (this *Context) ContentType(val string) string {\n\tvar ctype string\n\tif strings.ContainsRune(val, '/') {\n\t\tctype = val\n\t} else {\n\t\tif !strings.HasPrefix(val, \".\") {\n\t\t\tval = \".\" + val\n\t\t}\n\t\tctype = mime.TypeByExtension(val)\n\t}\n\tif ctype != \"\" {\n\t\tthis.Header().Set(\"Content-Type\", ctype)\n\t}\n\treturn ctype\n}", "func ThreatAssessmentContentTypePFile() *ThreatAssessmentContentType {\n\tv := ThreatAssessmentContentTypeVFile\n\treturn &v\n}", "func getRequestContentType(r *http.Request) string {\n\tct := r.Header.Get(\"Content-Type\")\n\tif ct != \"\" {\n\t\treturn ct\n\t} else {\n\t\treturn DEFAULT_CONTENT_TYPE\n\t}\n}", "func (h *RequestHeader) ContentType() []byte {\n\tif h.disableSpecialHeader {\n\t\treturn peekArgBytes(h.h, []byte(HeaderContentType))\n\t}\n\treturn h.contentType\n}", "func (event FileEvent) GetType() string {\n\treturn \"file\"\n}", "func ContentType(mimeType mime.Type, additionalParams ...params.Header) (string, string) {\n\tif len(additionalParams) > 0 {\n\t\treturn \"Content-Type\", fmt.Sprintf(\"%v%v\", mimeType, params.Aggregate(additionalParams))\n\t} else {\n\t\treturn \"Content-Type\", string(mimeType)\n\t}\n}", "func (dr DownloadResponse) ContentType() string {\n\treturn dr.dr.ContentType()\n}", "func (_AccessIndexor *AccessIndexorCaller) GetContentType(opts *bind.CallOpts, position *big.Int) (common.Address, error) {\n\tvar out []interface{}\n\terr := _AccessIndexor.contract.Call(opts, &out, \"getContentType\", position)\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (o FileContentBufferOutput) FileType() FileContentBufferFileTypePtrOutput {\n\treturn o.ApplyT(func(v FileContentBuffer) *FileContentBufferFileType { return v.FileType }).(FileContentBufferFileTypePtrOutput)\n}", "func Get(p string) (string, error) {\n\tfile, err := os.Open(p)\n\n\tif err != nil {\n\t\tlog.Println(err)\n\t\tos.Exit(1)\n\t}\n\n\tbuff := make([]byte, 512)\n\n\t_, err = file.Read(buff)\n\n\tif err != nil {\n\t\tlog.Println(err)\n\t\tos.Exit(1)\n\t}\n\n\tfiletype := http.DetectContentType(buff)\n\n\text := imgext.Get()\n\n\tfor i := 0; i < len(ext); i++ {\n\t\tif strings.Contains(ext[i], filetype[6:len(filetype)]) {\n\t\t\treturn filetype, nil\n\t\t}\n\t}\n\n\treturn \"\", errors.New(\"Invalid image type\")\n\n}", "func HTTPGetContentType(input interface{}) string {\n\tvar (\n\t\theader http.Header\n\t)\n\n\tswitch input.(type) {\n\tcase *http.Request:\n\t\theader = input.(*http.Request).Header\n\tcase *http.Response:\n\t\theader = input.(*http.Response).Header\n\tcase http.ResponseWriter:\n\t\theader = input.(http.ResponseWriter).Header()\n\tdefault:\n\t\treturn \"\"\n\t}\n\n\tcontentType := header.Get(\"Content-Type\")\n\n\tindex := strings.Index(contentType, \";\")\n\tif index != -1 {\n\t\treturn contentType[:index]\n\t}\n\n\treturn contentType\n}", "func (f *File) Type() TypeFlag { return FileType }", "func (s funcRenderer) ContentType() string {\n\treturn s.contentType\n}", "func TypeByExtension(filePath string) string {\n\ttyp := mime.TypeByExtension(path.Ext(filePath))\n\tif typ == \"\" {\n\t\ttyp = extToMimeType[strings.ToLower(path.Ext(filePath))]\n\t}\n\treturn typ\n}", "func (m *ThreatAssessmentRequest) GetContentType()(*ThreatAssessmentContentType) {\n return m.contentType\n}", "func (o *SoftwarerepositoryCategoryMapper) GetFileType() string {\n\tif o == nil || o.FileType == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.FileType\n}", "func extensionToMimeType(path string) string {\n\tif strings.HasSuffix(path, \".txt\") {\n\t\t// anyone uploading text in anything other than utf-8 needs to be\n\t\t// re-educated.\n\t\treturn \"text/plain; charset=utf-8\"\n\t}\n\n\tif strings.HasSuffix(path, \".png\") {\n\t\treturn \"image/png\"\n\t}\n\n\tif strings.HasSuffix(path, \".jpg\") {\n\t\treturn \"image/jpeg\"\n\t}\n\n\treturn \"application/octet-stream\"\n}", "func (o BlobOutput) ContentType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Blob) pulumi.StringPtrOutput { return v.ContentType }).(pulumi.StringPtrOutput)\n}", "func (o FileContentBufferPtrOutput) FileType() FileContentBufferFileTypePtrOutput {\n\treturn o.ApplyT(func(v *FileContentBuffer) *FileContentBufferFileType {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.FileType\n\t}).(FileContentBufferFileTypePtrOutput)\n}", "func getSpecType(file *anypoint.ExchangeFile, specContent []byte) (string, error) {\n\tif file.Classifier == apic.Wsdl {\n\t\treturn apic.Wsdl, nil\n\t}\n\n\tif specContent != nil {\n\t\tjsonMap := make(map[string]interface{})\n\t\terr := json.Unmarshal(specContent, &jsonMap)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tif _, isSwagger := jsonMap[\"swagger\"]; isSwagger {\n\t\t\treturn apic.Oas2, nil\n\t\t} else if _, isOpenAPI := jsonMap[\"openapi\"]; isOpenAPI {\n\t\t\treturn apic.Oas3, nil\n\t\t}\n\t}\n\treturn \"\", nil\n}", "func (o *S3UploadOpts) GetContentType() *string {\n\treturn getStrPtr(o.ContentType)\n}", "func (_BaseAccessWallet *BaseAccessWalletCaller) GetContentType(opts *bind.CallOpts, position *big.Int) (common.Address, error) {\n\tvar out []interface{}\n\terr := _BaseAccessWallet.contract.Call(opts, &out, \"getContentType\", position)\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (ctx *Context) ContentType(val string, unique bool) {\n\tvar ctype string\n\tif strings.ContainsRune(val, '/') {\n\t\tctype = val\n\t} else {\n\t\tif !strings.HasPrefix(val, \".\") {\n\t\t\tval = \".\" + val\n\t\t}\n\t\tctype = mime.TypeByExtension(val)\n\t}\n\tif ctype != \"\" {\n\t\tctx.SetHeader(\"Content-Type\", ctype, unique)\n\t}\n}", "func (req *Request) ContentType() string {\n\tif req == nil {\n\t\treturn \"\"\n\t}\n\n\tif req._contentType == \"\" {\n\t\tcontentType := req.Request.Header.Get(\"Content-Type\")\n\n\t\tif contentType == \"\" {\n\t\t\treq._contentType = \"text/html\"\n\t\t} else {\n\t\t\treq._contentType = strings.ToLower(strings.TrimSpace(strings.Split(contentType, \";\")[0]))\n\t\t}\n\t}\n\n\treturn req._contentType\n}", "func (res *ResponseRecorder) ContentType() string {\n\treturn res.Header().Get(\"Content-Type\")\n}", "func (bgpr BlobsGetPropertiesResponse) ContentType() string {\n\treturn bgpr.rawResponse.Header.Get(\"Content-Type\")\n}", "func (dgpr DirectoryGetPropertiesResponse) ContentType() string {\n\treturn PathGetPropertiesResponse(dgpr).ContentType()\n}", "func GetContentTypeHeader(header http.Header) string {\n\tcontentType := header.Get(ContentTypeHeader)\n\tif len(contentType) > 0 {\n\t\treturn contentType\n\t}\n\treturn DefaultContentType\n}", "func (m *ThreatAssessmentRequest) GetContentType()(*ThreatAssessmentContentType) {\n val, err := m.GetBackingStore().Get(\"contentType\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*ThreatAssessmentContentType)\n }\n return nil\n}", "func (r *Repo) MimeType(path string, rev int64) (string, error) {\n\tvar (\n\t\tmimetype *C.svn_string_t\n\t\trevisionRoot *C.svn_fs_root_t\n\t)\n\n\tmime := \"\"\n\n\tif e := C.svn_fs_revision_root(&revisionRoot, r.fs, C.svn_revnum_t(rev), r.pool); e != nil {\n\t\treturn mime, makeError(e)\n\t} else {\n\t\tdefer C.svn_fs_close_root(revisionRoot)\n\t}\n\n\tcpath := C.CString(path) // convert to C string\n\tdefer C.free(unsafe.Pointer(cpath))\n\n\tif err := C.FileMimeType(&mimetype, revisionRoot, cpath, r.pool); err != nil {\n\t\treturn mime, makeError(err)\n\t}\n\n\tif mimetype != nil {\n\t\tmime = C.GoString(mimetype.data)\n\t}\n\n\treturn mime, nil\n}", "func GetMimeTypeByFileExt(url string) string {\n\tparts := strings.Split(url, \".\")\n\tvar fileExt = strings.ToLower(strings.TrimSpace(parts[len(parts)-1]))\n\n\tif fileExt == \"jpg\" || fileExt == \"jpeg\" {\n\t\treturn \"image/jpg\"\n\t} else if fileExt == \"png\" {\n\t\treturn \"image/png\"\n\t} else if fileExt == \"webp\" {\n\t\treturn \"image/webp\"\n\t} else if fileExt == \"svg\" {\n\t\treturn \"image/svg+xml\"\n\t} else if fileExt == \"gif\" {\n\t\treturn \"image/gif\"\n\t}\n\n\treturn \"\"\n}", "func TestFileType(t *testing.T) {\n\tres := FileType(\"flag.go\")\n\tif res != 1 {\n\t\tt.Errorf(\"Test FileType(\\\"flag.go\\\") failed, expect 1, but got %d\", res)\n\t}\n}", "func (at AssetType) FileExt() string {\n\tswitch at {\n\tcase AssetTypeAudio:\n\t\treturn \".mp3\"\n\tcase AssetTypeTranscript:\n\t\treturn \".txt\"\n\tcase AssetTypeVideo:\n\t\treturn \".mp4\"\n\tdefault:\n\t\treturn \".json\"\n\t}\n}", "func (gppr GetPathPropertiesResponse) ContentType() string {\n\treturn gppr.rawResponse.Header.Get(\"Content-Type\")\n}", "func (h *ResponseHeader) ContentType() []byte {\n\tcontentType := h.contentType\n\tif !h.noDefaultContentType && len(h.contentType) == 0 {\n\t\tcontentType = defaultContentType\n\t}\n\treturn contentType\n}", "func (m *WorkbookCommentReply) GetContentType()(*string) {\n return m.contentType\n}", "func (bl BlockList) ContentType() string {\n\treturn bl.rawResponse.Header.Get(\"Content-Type\")\n}", "func (o *Object) MimeType(ctx context.Context) (mimeType string) {\n\tif do, ok := o.Object.(fs.MimeTyper); ok {\n\t\tmimeType = do.MimeType(ctx)\n\t}\n\treturn mimeType\n}", "func (lbfr ListBlobsFlatResponse) ContentType() string {\n\treturn lbfr.rawResponse.Header.Get(\"Content-Type\")\n}", "func (x *Message) GetContentType() string {\n\tif v, ok := x.Header[ContentType]; ok && v != \"\" {\n\t\treturn v\n\t}\n\treturn encoding.ContentTypeJSON\n}", "func (o *GetMessagesAllOf) GetContentType() interface{} {\n\tif o == nil {\n\t\tvar ret interface{}\n\t\treturn ret\n\t}\n\treturn o.ContentType\n}", "func getImageFormat(filename string) (string, error) {\n\tmime, _, err := mimetype.DetectFile(filename)\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"can't detect image's type\")\n\t} else {\n\t\treturn mime, nil\n\t}\n}", "func (r reportLog) RequestContentType() string {\n\tif r.RequestHeaders != nil {\n\t\tfor k, v := range r.RequestHeaders {\n\t\t\tif strings.ToLower(k) == \"content-type\" {\n\t\t\t\treturn v\n\t\t\t}\n\t\t}\n\t}\n\treturn \"\"\n}", "func (imr *InvokeMethodResponse) ContentType() string {\n\tm := imr.r.Message\n\tif m == nil {\n\t\treturn \"\"\n\t}\n\n\tcontentType := m.ContentType\n\n\tif m.Data != nil && m.Data.TypeUrl != \"\" {\n\t\tcontentType = ProtobufContentType\n\t}\n\n\treturn contentType\n}", "func (o *Object) MimeType() string {\n\terr := o.loadMetadataIfNotLoaded()\n\tif err != nil {\n\t\treturn \"error/error\"\n\t}\n\treturn o.meta.MimeType\n}", "func (r *Reconciler) getFileType(path string) SourceType {\n\tfragments := strings.Split(path, \".\")\n\tif len(fragments) == 0 {\n\t\treturn SourceTypeUnknown\n\t}\n\n\textension := strings.TrimSpace(fragments[len(fragments)-1])\n\tswitch strings.ToLower(extension) {\n\tcase \"json\":\n\t\treturn SourceTypeJson\n\tcase \"grafonnet\":\n\t\treturn SourceTypeJsonnet\n\tcase \"jsonnet\":\n\t\treturn SourceTypeJsonnet\n\tcase \"yaml\":\n\t\treturn SourceTypeYaml\n\tdefault:\n\t\treturn SourceTypeUnknown\n\t}\n}", "func (st SampleType) Mime() string {\n\tswitch st {\n\tcase S16LE:\n\t\treturn \"l16\"\n\tdefault:\n\t\tpanic(\"Invalid SampleType\")\n\t}\n}", "func DetectMediaType(media io.ReaderAt) string {\n\tif typer, ok := media.(googleapi.ContentTyper); ok {\n\t\treturn typer.ContentType()\n\t}\n\n\ttyp := \"application/octet-stream\"\n\tbuf := make([]byte, 1024)\n\tn, err := media.ReadAt(buf, 0)\n\tbuf = buf[:n]\n\tif err == nil || err == io.EOF {\n\t\ttyp = http.DetectContentType(buf)\n\t}\n\treturn typ\n}", "func (o DeploymentArtifactOutput) MimeType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *DeploymentArtifact) pulumi.StringOutput { return v.MimeType }).(pulumi.StringOutput)\n}", "func (r *Request) ContentType() string {\n\treturn r.contentType\n}", "func (s *Source) mimeType(arg string) (mimeType string, err error) {\n\tif len(arg) > 0 {\n\t\tif strings.HasPrefix(arg, \"//\") {\n\t\t\targ = arg[1:]\n\t\t}\n\t\tif !strings.HasPrefix(arg, \"/\") {\n\t\t\targ = \"/\" + arg\n\t\t}\n\t}\n\targURL, err := url.Parse(arg)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"mimeType: couldn't parse arg %q: %w\", arg, err)\n\t}\n\tmediatype := argURL.Query().Get(\"type\")\n\tif mediatype == \"\" {\n\t\tmediatype = s.URL.Query().Get(\"type\")\n\t}\n\n\tif mediatype == \"\" {\n\t\tmediatype = s.mediaType\n\t}\n\n\t// make it so + doesn't need to be escaped\n\tmediatype = strings.ReplaceAll(mediatype, \" \", \"+\")\n\n\tif mediatype == \"\" {\n\t\text := filepath.Ext(argURL.Path)\n\t\tmediatype = mime.TypeByExtension(ext)\n\t}\n\n\tif mediatype == \"\" {\n\t\text := filepath.Ext(s.URL.Path)\n\t\tmediatype = mime.TypeByExtension(ext)\n\t}\n\n\tif mediatype != \"\" {\n\t\tt, _, err := mime.ParseMediaType(mediatype)\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"MIME type was %q: %w\", mediatype, err)\n\t\t}\n\t\tmediatype = t\n\t\treturn mediatype, nil\n\t}\n\n\treturn textMimetype, nil\n}", "func (s *Structured) GetContentType() string {\n\treturn s.cloudEvent.ContentType\n}", "func (r *Request) contentType() int {\n\tif r.Method == \"HEAD\" || r.Method == \"OPTIONS\" {\n\t\treturn contentNone\n\t}\n\n\tct := r.Header.Get(\"content-type\")\n\tif strings.Contains(ct, \"application/x-www-form-urlencoded\") {\n\t\treturn contentFormData\n\t}\n\n\tif strings.Contains(ct, \"multipart/form-data\") {\n\t\treturn contentMultipart\n\t}\n\n\treturn contentStream\n}", "func MimeFile(filename string) (string, error) {\n\tf, err := os.Open(filename)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer f.Close()\n\n\t// We need up to 512 bytes\n\tb := make([]byte, 512)\n\tn, err := f.Read(b)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Clip buffer to prevent detecting binary files.\n\treturn mime.DetectMimeType(b[:n]), nil\n}", "func (upr UpdatePathResponse) ContentType() string {\n\treturn upr.rawResponse.Header.Get(\"Content-Type\")\n}", "func (o GoogleCloudDatalabelingV1beta1GcsSourceOutput) MimeType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GoogleCloudDatalabelingV1beta1GcsSource) string { return v.MimeType }).(pulumi.StringOutput)\n}" ]
[ "0.81000364", "0.8066238", "0.79145086", "0.7778787", "0.74996734", "0.7295712", "0.7261386", "0.72185165", "0.7110216", "0.6988363", "0.6982667", "0.6970133", "0.6933078", "0.68941605", "0.68878514", "0.6883115", "0.6878635", "0.68280417", "0.6811066", "0.6810509", "0.67394835", "0.6729942", "0.6620254", "0.6589901", "0.65861416", "0.6581592", "0.65774757", "0.6576521", "0.65665454", "0.6561205", "0.65447366", "0.652321", "0.65202326", "0.65115225", "0.64962584", "0.6488923", "0.64872533", "0.64658386", "0.6440806", "0.6410378", "0.6379529", "0.6373523", "0.6371358", "0.635339", "0.6347398", "0.63354945", "0.63091767", "0.62946194", "0.62751377", "0.6274279", "0.62713456", "0.6267831", "0.62607795", "0.62487453", "0.62444484", "0.6242183", "0.62289906", "0.6199824", "0.6199325", "0.6182842", "0.61771435", "0.61622727", "0.615986", "0.6154699", "0.61542493", "0.6122468", "0.6121052", "0.61131626", "0.6099952", "0.60889304", "0.6083196", "0.6077706", "0.6070504", "0.60595477", "0.6056635", "0.6032777", "0.60318565", "0.6028658", "0.60280144", "0.60179234", "0.6001187", "0.59891516", "0.5978959", "0.59687126", "0.5959292", "0.5950229", "0.5949531", "0.5923232", "0.59211314", "0.59122443", "0.589708", "0.58947307", "0.5892826", "0.58870924", "0.5862303", "0.584982", "0.58492017", "0.5843454", "0.58279574", "0.58185065" ]
0.79005975
3
New creates a new topic
func New(name string, baseDir string) (*WALTopic, error) { walDir := filepath.Join(baseDir, name) oneMB := 1 * 1024 * 1024 w, err := walFactory.NewSize(nil, nil, walDir, oneMB, false) if err != nil { return nil, errors.Wrapf(err, "Could not create wal for topic %s", name) } m := make(map[int]uint64) return &WALTopic{name: name, walDir: walDir, wal: w, firstIndexInSegment: m}, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (api *TopicalAPI) TopicNew(w http.ResponseWriter, r *http.Request) {\n\tflashes, _ := api.session.GetFlashes(r, w)\n\tuser, err := api.session.GetUser(r)\n\n\tif err != nil {\n\t\tapi.session.SaveFlash(\"Log in to post a message\", r, w)\n\t\thttp.Redirect(w, r, \"/topics\", 302)\n\t\treturn\n\t}\n\n\tpayload := struct {\n\t\tUser *models.User\n\t\tFlashes []string\n\t}{user, flashes}\n\n\tapi.templates.ExecuteTemplate(w, \"new-topic\", payload)\n}", "func NewTopic(n string, ts []string) *Topic {\n\treturn &Topic{ID: strings.ToLower(n), Name: n, Tags: ts, Created: time.Now()}\n}", "func (v TopicsResource) New(c buffalo.Context) error {\n\treturn c.Render(200, r.Auto(c, &models.Topic{}))\n}", "func NewTopic() *Topic {\n\treturn &Topic{}\n}", "func (c Threads) New(topic string) revel.Result {\n // Validate\n c.Validation.Required(topic)\n c.Validation.MinSize(topic, 3)\n\n if c.Validation.HasErrors() {\n c.Validation.Keep()\n c.FlashParams()\n return c.Redirect(Threads.ShowNew)\n }\n\n // Create thread\n user := c.connected()\n thread := &models.Thread{\n 0,\n 0,\n topic,\n \"\",\n user,\n time.Now(),\n }\n err := c.Txn.Insert(thread)\n if err != nil {\n c.Flash.Error(\"An error occurred, sorry\")\n fmt.Println(err)\n }\n\n return c.Redirect(\"/threads/%d\", thread.ThreadId)\n}", "func (t *TopicService) Create(name, description, parentID string) (*Topic, error) {\n\tquery := `\n\tmutation (\n\t\t$name: String!,\n\t\t$description: String,\n\t\t$parentId: ID\n\t){\n\t\tcreateTopic(\n\t\t\tname: $name, description: $description, parentId: $parentId\n\t\t){ id, name, description }\n\t}`\n\tvar resp struct {\n\t\tTopic *Topic `json:\"createTopic\"`\n\t}\n\tvars := map[string]interface{}{\"name\": name, \"description\": description, \"parentId\": parentID}\n\terr := t.client.Do(context.Background(), query, vars, &resp)\n\treturn resp.Topic, err\n}", "func NewTopic(session *mgo.Session, name string) *Topic {\n\treturn &Topic{\n\t\tsession: session,\n\t\tname: name,\n\t}\n}", "func TopicCreate(w http.ResponseWriter, r *http.Request) {\n\n\t// Init output\n\toutput := []byte(\"\")\n\n\t// Add content type header to the response\n\tcontentType := \"application/json\"\n\tcharset := \"utf-8\"\n\tw.Header().Add(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\t// Grab url path variables\n\turlVars := mux.Vars(r)\n\n\t// Grab context references\n\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\tprojectUUID := gorillaContext.Get(r, \"auth_project_uuid\").(string)\n\n\tpostBody := map[string]string{}\n\tschemaUUID := \"\"\n\n\t// check if there's a request body provided before trying to decode\n\tif r.Body != nil {\n\n\t\tb, err := ioutil.ReadAll(r.Body)\n\n\t\tif err != nil {\n\t\t\terr := APIErrorInvalidRequestBody()\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t\tdefer r.Body.Close()\n\n\t\tif len(b) > 0 {\n\t\t\terr = json.Unmarshal(b, &postBody)\n\t\t\tif err != nil {\n\t\t\t\terr := APIErrorInvalidRequestBody()\n\t\t\t\trespondErr(w, err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tschemaRef := postBody[\"schema\"]\n\n\t\t\t// if there was a schema name provided, check its existence\n\t\t\tif schemaRef != \"\" {\n\t\t\t\t_, schemaName, err := schemas.ExtractSchema(schemaRef)\n\t\t\t\tif err != nil {\n\t\t\t\t\terr := APIErrorInvalidData(err.Error())\n\t\t\t\t\trespondErr(w, err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tsl, err := schemas.Find(projectUUID, \"\", schemaName, refStr)\n\t\t\t\tif err != nil {\n\t\t\t\t\terr := APIErrGenericInternal(err.Error())\n\t\t\t\t\trespondErr(w, err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tif sl.Empty() {\n\t\t\t\t\terr := APIErrorNotFound(\"Schema\")\n\t\t\t\t\trespondErr(w, err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tschemaUUID = sl.Schemas[0].UUID\n\t\t\t}\n\t\t}\n\t}\n\t// Get Result Object\n\tres, err := topics.CreateTopic(projectUUID, urlVars[\"topic\"], schemaUUID, refStr)\n\tif err != nil {\n\t\tif err.Error() == \"exists\" {\n\t\t\terr := APIErrorConflict(\"Topic\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t\terr := APIErrGenericInternal(err.Error())\n\t\trespondErr(w, err)\n\t}\n\n\t// Output result to JSON\n\tresJSON, err := res.ExportJSON()\n\tif err != nil {\n\t\terr := APIErrExportJSON()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Write response\n\toutput = []byte(resJSON)\n\trespondOK(w, output)\n\n}", "func NewTopic() (t *Topic, err error) {\n\tt = &Topic{}\n\n\t// Random values\n\tid := make([]byte, 8)\n\tif _, err = rand.Read(id); err != nil {\n\t\treturn\n\t}\n\tseed1, err := drbg.NewSeed()\n\tif err != nil {\n\t\treturn\n\t}\n\tseed2, err := drbg.NewSeed()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tt.ID, _ = binary.Uvarint(id[0:8])\n\tt.Handle.Seed1 = seed1\n\tt.Handle.Seed2 = seed2\n\tif err = initHandle(&t.Handle); err != nil {\n\t\treturn\n\t}\n\n\t// Create shared secret\n\tpub, priv, err := box.GenerateKey(rand.Reader)\n\tif err != nil {\n\t\treturn\n\t}\n\tvar sharedKey [32]byte\n\tbox.Precompute(&sharedKey, pub, priv)\n\n\tt.Handle.SharedSecret = &sharedKey\n\n\t// Create signing secrets\n\tt.Handle.SigningPublicKey, t.SigningPrivateKey, err = ed25519.GenerateKey(rand.Reader)\n\n\treturn\n}", "func NewTopic(name string) Topic {\n\tct := &channeledTopic{\n\t\tname: name,\n\t\tsubscribers: make(map[uint64]*Subscription, 512), // Sane default space?\n\t}\n\treturn ct\n}", "func NewTopic(node storm.Node) *Topic {\n\treturn &Topic{\n\t\tnode: node,\n\t}\n}", "func (p *nsqPublisher) CreateTopic(topic string) error {\n\t//nolint:noctx\n\treq, err := http.NewRequest(http.MethodPost, fmt.Sprintf(\"http://%s/topic/create?topic=%s\", p.httpEndpoint, topic), nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\treq.Header.Add(\"ContentType\", \"text/plain\")\n\n\tresp, err := p.client.Do(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif resp.StatusCode >= 300 {\n\t\tbodyBytes, err := io.ReadAll(resp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn fmt.Errorf(\"error creating topic: %s\", string(bodyBytes))\n\t}\n\n\t_ = resp.Body.Close()\n\treturn nil\n}", "func (a *DefaultApiService) CreateTopic(ctx _context.Context) ApiCreateTopicRequest {\n\treturn ApiCreateTopicRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func NewTopic(project, name string) Topic {\n\treturn Topic(newResource(project, \"topics\", name))\n}", "func NewTopic(name string) *Topic {\n\treturn &Topic{\n\t\tName: name,\n\t\tChannel: make(chan schema.Message),\n\t\tDepth: 0,\n\t\tbuffer: newQueueBuffer(1000),\n\t\tlock: sync.Mutex{},\n\t}\n}", "func (r *TopicsService) Create(topic *Topic) *TopicsCreateCall {\n\tc := &TopicsCreateCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.topic = topic\n\treturn c\n}", "func (v TopicsResource) Create(c buffalo.Context) error {\n\t// Allocate an empty Topic\n\ttopic := &models.Topic{}\n\n\t// Bind topic to the html form elements\n\tif err := c.Bind(topic); err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\n\t// Get the DB connection from the context\n\ttx, ok := c.Value(\"tx\").(*pop.Connection)\n\tif !ok {\n\t\treturn errors.WithStack(errors.New(\"no transaction found\"))\n\t}\n\n\t// Validate the data from the html form\n\tverrs, err := tx.ValidateAndCreate(topic)\n\tif err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\n\tif verrs.HasAny() {\n\t\t// Make the errors available inside the html template\n\t\tc.Set(\"errors\", verrs)\n\n\t\t// Render again the new.html template that the user can\n\t\t// correct the input.\n\t\treturn c.Render(422, r.Auto(c, topic))\n\t}\n\n\t// If there are no errors set a success message\n\tc.Flash().Add(\"success\", \"Topic was created successfully\")\n\n\t// and redirect to the topics index page\n\treturn c.Render(201, r.Auto(c, topic))\n}", "func NewTopic(ctx *pulumi.Context,\n\tname string, args *TopicArgs, opts ...pulumi.ResourceOption) (*Topic, error) {\n\tif args == nil {\n\t\targs = &TopicArgs{}\n\t}\n\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Topic\n\terr := ctx.RegisterResource(\"google-native:pubsub/v1beta1a:Topic\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (c *Client) CreateTopic(ctx context.Context, id string) (*Topic, error) {\n\ttop, err := c.Client.CreateTopic(ctx, id)\n\treturn &Topic{top, c.projectID, c.sensor}, err\n}", "func CreateTopic(c *client.Client) fw.Topic {\n\treturn &configzTopic{\n\t\tcl: c,\n\t}\n}", "func (tm *topicManager) createNewTopic(name string, id string) (log, error) {\n\tlc, ok := tm.topicNamesToLogs[name]\n\tif ok {\n\t\tlc.refCount++\n\t\treturn lc.l, nil\n\t}\n\n\t// lazily create the log if we know about the topic/cc ID\n\tl, err := newBadgerLog(tm.dataDir, name, id, tm.cc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttm.topicNamesToLogs[name] = &logContainer{\n\t\tl: l,\n\t\trefCount: 1,\n\t}\n\treturn l, nil\n}", "func NewTopic(topic_id string) *Topics {\n\tnew_topic := Topics{\n\t\treg_subscribers: Subscribers{},\n\t\tsub_lock: sync.RWMutex{},\n\t\tsub_topics: TopicsTable{},\n\t\ttopic_lock: sync.RWMutex{},\n\t}\n\tnew_topic.CreateTopic(topic_id)\n\treturn &new_topic\n}", "func New(t Topic) (Interface, error) {\n\t// TODO(tmrts): use a LFU and if cache-key space consumption is not a\n\t// problem, migrate to optimized cache replacement algorithms that use\n\t// extra queues\n\t// TODO(tmrts): utilize the eviction callback in LRU\n\tlruCache := lru.NewLRU(t.Capacity, nil)\n\n\tr, err := router.New(RouterPort)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := r.Join(t.Peers); err != nil {\n\t\treturn nil, err\n\t}\n\n\tc := &topic{\n\t\tlru: lruCache,\n\t\tprovider: t.Provider,\n\t\trouter: r,\n\t}\n\n\t// FIXME(tmrts): needs restructuring\n\tsvc, err := service.NewServer(ServicePort, c)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tc.svc = svc\n\n\treturn c, nil\n}", "func New(top topics.Topic, p interface{}) Message {\n\tsafePayload := convertToSafePayload(p)\n\treturn &simple{category: top, payload: safePayload}\n}", "func (aaa *NotificationService) CreateTopic(input *notification.CreateTopicParams) error {\n\ttoken, err := aaa.TokenRepository.GetToken()\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, badRequest, unauthorized, forbidden, conflict, err := aaa.Client.Notification.CreateTopic(input, client.BearerToken(*token.AccessToken))\n\tif badRequest != nil {\n\t\treturn badRequest\n\t}\n\tif unauthorized != nil {\n\t\treturn unauthorized\n\t}\n\tif forbidden != nil {\n\t\treturn forbidden\n\t}\n\tif conflict != nil {\n\t\treturn conflict\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (client *Client) CreateTopic(request *CreateTopicRequest) (response *CreateTopicResponse, err error) {\n\tresponse = CreateCreateTopicResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func newNode() *topicNode {\n\treturn &topicNode{\n\t\tchildren: children{},\n\t\tclients: make(clientOpts),\n\t\tshared: make(map[string]clientOpts),\n\t}\n}", "func NewTopic(workers int) *Topic {\n\tt := &Topic{\n\t\tsubs: make(map[chan interface{}]context.Context),\n\t\tpubs: make(chan Work),\n\t\tend: make(chan bool),\n\t\tworkerChan: make(chan chan Work),\n\t}\n\tfor i := 0; i < workers; i++ {\n\t\tt.workers = append(t.workers, NewWorker(t.workerChan))\n\t}\n\tt.Start()\n\treturn t\n}", "func (c *client) createTopic(topicName string) (*pubsub.Topic, error) {\n\tvar topic *pubsub.Topic\n\tctx := context.Background()\n\n\texists, err := c.topicExists(topicName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif !exists {\n\t\ttopic, err = c.client.CreateTopic(ctx, topicName)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\ttopic = c.client.Topic(topicName)\n\t}\n\n\treturn topic, nil\n}", "func (tm *topicManager) createTopic(name string) error {\n\ttm.topicsMutex.RLock()\n\t_, ok := tm.topicNamesToIds[name]\n\ttm.topicsMutex.RUnlock()\n\tif ok {\n\t\treturn fmt.Errorf(\"Topic with name [%s] already exists\", name)\n\t}\n\n\t// this allocates a copycat data structure but doesn't create a log structure yet\n\tid, err := tm.cc.AllocateNewDataStructure()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tprotobuf := &pb.TopicMetadataOperation{\n\t\tName: name,\n\t\tDataStructureId: id.String(),\n\t}\n\n\tbites, err := protobuf.Marshal()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttm.proposeCh <- bites\n\treturn nil\n}", "func createTopic(broker *sarama.Broker, topicName string) error {\n\ttopicDetail := &sarama.TopicDetail{}\n\ttopicDetail.NumPartitions = int32(1)\n\ttopicDetail.ReplicationFactor = int16(1)\n\ttopicDetail.ConfigEntries = make(map[string]*string)\n\n\ttopicDetails := make(map[string]*sarama.TopicDetail)\n\ttopicDetails[topicName] = topicDetail\n\trequest := sarama.CreateTopicsRequest{\n\t\tTimeout: time.Second * 15,\n\t\tTopicDetails: topicDetails,\n\t}\n\n\t_, err := broker.CreateTopics(&request)\n\treturn err\n}", "func (r *Reconciler) createTopic(ctx context.Context, topicName string, partitions int32, replicationFactor int16, retentionMillis int64) error {\n\n\t// Get The Logger From The Context\n\tlogger := logging.FromContext(ctx)\n\n\t// Create The TopicDefinition\n\tretentionMillisString := strconv.FormatInt(retentionMillis, 10)\n\ttopicDetail := &sarama.TopicDetail{\n\t\tNumPartitions: partitions,\n\t\tReplicationFactor: replicationFactor,\n\t\tReplicaAssignment: nil, // Currently Not Assigning Partitions To Replicas\n\t\tConfigEntries: map[string]*string{\n\t\t\tcommonconstants.KafkaTopicConfigRetentionMs: &retentionMillisString,\n\t\t},\n\t}\n\n\t// Attempt To Create The Topic & Process TopicError Results (Including Success ;)\n\terr := r.adminClient.CreateTopic(ctx, topicName, topicDetail)\n\tif err != nil {\n\t\tlogger := logger.With(zap.Int16(\"KError\", int16(err.Err)))\n\t\tswitch err.Err {\n\t\tcase sarama.ErrNoError:\n\t\t\tlogger.Info(\"Successfully Created New Kafka Topic (ErrNoError)\")\n\t\t\treturn nil\n\t\tcase sarama.ErrTopicAlreadyExists:\n\t\t\tlogger.Info(\"Kafka Topic Already Exists - No Creation Required\")\n\t\t\treturn nil\n\t\tdefault:\n\t\t\tlogger.Error(\"Failed To Create Topic\")\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tlogger.Info(\"Successfully Created New Kafka Topic (Nil TopicError)\")\n\t\treturn nil\n\t}\n}", "func (g *GooglePubSub) CreateTopic(topicName string) error {\n\t_, err := g.client.CreateTopic(g.ctx, topicName)\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to create topic: %v\", err)\n\t\treturn err\n\t}\n\treturn err\n}", "func newTopicTrie() *topicTrie {\n\treturn newNode()\n}", "func (b *Topics) CreateTopic(topic_id string) {\n\tb.sub_lock.Lock()\n\tif nil == b.sub_topics[topic_id] {\n\t\tb.sub_topics[topic_id] = Subscribers{}\n\t}\n\tb.sub_lock.Unlock()\n}", "func (r *Reconciler) createTopic(ctx context.Context, topicName string, partitions int, replicationFactor int, retentionMillis int64) error {\n\n\t// Setup The Logger\n\tlogger := r.logger.With(zap.String(\"Topic\", topicName))\n\n\t// Create The TopicSpecification\n\ttopicSpecifications := []kafka.TopicSpecification{\n\t\t{\n\t\t\tTopic: topicName,\n\t\t\tNumPartitions: partitions,\n\t\t\tReplicationFactor: replicationFactor,\n\t\t\tConfig: map[string]string{\n\t\t\t\tconstants.KafkaTopicConfigRetentionMs: strconv.FormatInt(retentionMillis, 10),\n\t\t\t},\n\t\t},\n\t}\n\n\t// Attempt To Create The Topic & Process Results\n\ttopicResults, err := r.adminClient.CreateTopics(ctx, topicSpecifications)\n\tif len(topicResults) > 0 {\n\t\ttopicResultError := topicResults[0].Error\n\t\ttopicResultErrorCode := topicResultError.Code()\n\t\tif topicResultErrorCode == kafka.ErrTopicAlreadyExists {\n\t\t\tlogger.Info(\"Kafka Topic Already Exists - No Creation Required\")\n\t\t\treturn nil\n\t\t} else if topicResultErrorCode == kafka.ErrNoError {\n\t\t\tlogger.Info(\"Successfully Created New Kafka Topic\")\n\t\t\treturn nil\n\t\t} else {\n\t\t\tlogger.Error(\"Failed To Create Topic (Results)\", zap.Error(err), zap.Any(\"TopicResults\", topicResults))\n\t\t\treturn topicResults[0].Error\n\t\t}\n\t} else if err != nil {\n\t\tlogger.Error(\"Failed To Create Topic (Error)\", zap.Error(err))\n\t\treturn err\n\t} else {\n\t\tlogger.Warn(\"Received Empty TopicResults From CreateTopics Request\")\n\t\treturn nil\n\t}\n}", "func HandleCreate(w http.ResponseWriter, r *http.Request) {\r\n\tCorsHandler(w)\r\n\tctx := context.Background()\r\n\r\n\tprojectID := GetEnvVar(\"GOOGLE_CLOUD_PROJECT\")\r\n\tclient, err := pubsub.NewClient(ctx, projectID)\r\n\tCheckError(w, \"Error Creating Client\", err)\r\n\tt, err := client.CreateTopic(ctx, \"top3\")\r\n\tCheckError(w, \"Error Creating topic\", err)\r\n\tfmt.Fprintf(w, \"Topic Created Successfully %s\", t)\r\n}", "func (a *DefaultApiService) CreateTopicExecute(r ApiCreateTopicRequest) (Topic, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue Topic\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"DefaultApiService.CreateTopic\")\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, GenericOpenAPIError{error: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/topics\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\tif r.newTopicInput == nil {\n\t\treturn localVarReturnValue, nil, reportError(\"newTopicInput is required and must be specified\")\n\t}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = r.newTopicInput\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = _ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func createTopic(brokerAddress string, topicName string,\n\tnumPartitions int32, replicationFactor int32) {\n\tbrokerAddresses := []string{brokerAddress}\n\n\t// configuration handling\n\t// please note the versioning, it needs to be specified explicitly\n\tconfig := sarama.NewConfig()\n\tconfig.Version = sarama.V2_1_0_0\n\n\t// try to create and initialize cluster admin object\n\t// (it will reach out Kafka broker, so it checks the connection as well)\n\tclusterAdmin, err := sarama.NewClusterAdmin(brokerAddresses, config)\n\n\t// check if cluster admin has been initialized successfully\n\tif err != nil {\n\t\tlog.Fatal(\"Error: create cluster admin: \", err.Error())\n\t}\n\n\t// everything's seems to be ok\n\tlog.Print(\"Cluster admin has been initialized\")\n\n\t// cluster admin needs to be closed properly\n\tdefer func() {\n\t\t// try to close cluster admin\n\t\terr := clusterAdmin.Close()\n\n\t\t// check if cluster admin has been closed successfully\n\t\tif err != nil {\n\t\t\tlog.Fatal(\"Error: close cluster admin: \", err.Error())\n\t\t}\n\t}()\n\n\t// try to create a new topic via cluster admin\n\terr = clusterAdmin.CreateTopic(topicName,\n\t\t&sarama.TopicDetail{\n\t\t\tNumPartitions: numPartitions,\n\t\t\tReplicationFactor: int16(replicationFactor),\n\t\t}, false)\n\n\t// check if topic has been created successfully\n\tif err != nil {\n\t\tlog.Fatalf(\"Error: create topic '%s': %v\", topicName, err.Error())\n\t}\n\n\t// everything's seems to be ok -> topic has been created\n\tlog.Printf(\"Topic '%s' has been created\", topicName)\n}", "func (s Server) AddTopic(w http.ResponseWriter, r *http.Request) {\n\n\ttopicName := strings.TrimSpace(r.FormValue(\"name\"))\n\tif s.MaybeUserError(w, len(topicName) == 0, \"new topic name must not be blank\") {\n\t\treturn\n\t}\n\n\tuser, err := CurrentUser(s.DB, r)\n\tif handleError(w, \"cannot get current user\", err) {\n\t\treturn\n\t}\n\n\ttopic := model.NewTopic(user.ID, topicName)\n\ttopic, err = store.CreateTopic(s.DB, topic)\n\tif handleError(w, \"cannot create topic: %w\", err) {\n\t\treturn\n\t}\n\n\ts.WritePage(w, \"new-topic.html\", map[string]interface{}{\n\t\t\"topic\": topic,\n\t})\n}", "func (kz *Kazoo) Topic(topic string) *Topic {\n\treturn &Topic{Name: topic, kz: kz}\n}", "func New(topics []Topic) Chatroom {\n\troom := Room{\n\t\tIn: make(chan interface{}),\n\t\tOut: make(chan interface{}),\n\t}\n\tchatroom := Chatroom{\n\t\ttopics: topics,\n\t\tRoom: room,\n\t}\n\tgo chatroom.talk(room)\n\treturn chatroom\n}", "func New(address, topic string) (Producer, error) {\n\n\tconfig := nsq.NewConfig()\n\tp, err := nsq.NewProducer(address, config)\n\tif err != nil {\n\t\treturn Producer{}, err\n\t}\n\tif p.Ping() != nil {\n\t\treturn Producer{}, err\n\t}\n\n\treturn Producer{p}, nil\n}", "func createJndiTopicFunc(d *schema.ResourceData, meta interface{}) error {\n\tstate := meta.(*ProviderState)\n\t// These are the only required fields, so init them upfront\n\tjnditopic := semp_client.MsgVpnJndiTopic{\n\t\tTopicName: d.Get(TOPIC_NAME).(string),\n\t\tMsgVpnName: d.Get(MSG_VPN_NAME).(string),\n\t}\n\tpopulateJndiTopicFromResource(&jnditopic, d)\n\n\tresp, _, err := semp_client.JndiApi {\n\t\tConfiguration: state.sempcfg,\n\t}.CreateMsgVpnJndiTopic(jnditopic.MsgVpnName, jnditopic, nil)\n\n\tif err != nil {\n\t\tlog.Println(\"JndiTopicApi.CreateMsgVpnJndiTopic ERROR\")\n\t\treturn err\n\t}\n\tlogSempMeta(\"JndiTopic create\", resp.Meta)\n\t// Must uniquely identify the resource within terraform\n\td.SetId( state.host + \"_\" + jnditopic.MsgVpnName + \"_jndiTopic_\" + jnditopic.TopicName )\n\n\treturn nil\n}", "func (c *PublisherClient) CreateTopic(ctx context.Context, req *pubsubpb.Topic) (*pubsubpb.Topic, error) {\n\tctx = metadata.NewContext(ctx, c.metadata)\n\tvar resp *pubsubpb.Topic\n\terr := gax.Invoke(ctx, func(ctx context.Context) error {\n\t\tvar err error\n\t\tresp, err = c.client.CreateTopic(ctx, req)\n\t\treturn err\n\t}, c.CallOptions.CreateTopic...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func CreateTopic(p *kafka.Producer, topic string) {\n\n\ta, err := kafka.NewAdminClientFromProducer(p)\n\tif err != nil {\n\t\tfmt.Printf(\"Failed to create new admin client from producer: %s\", err)\n\t\tos.Exit(1)\n\t}\n\t// Contexts are used to abort or limit the amount of time\n\t// the Admin call blocks waiting for a result.\n\tctx, cancel := context.WithCancel(context.Background())\n\tdefer cancel()\n\t// Create topics on cluster.\n\t// Set Admin options to wait up to 60s for the operation to finish on the remote cluster\n\tmaxDur, err := time.ParseDuration(\"60s\")\n\tif err != nil {\n\t\tfmt.Printf(\"ParseDuration(60s): %s\", err)\n\t\tos.Exit(1)\n\t}\n\tresults, err := a.CreateTopics(\n\t\tctx,\n\t\t// Multiple topics can be created simultaneously\n\t\t// by providing more TopicSpecification structs here.\n\t\t[]kafka.TopicSpecification{{\n\t\t\tTopic: topic,\n\t\t\tNumPartitions: 1,\n\t\t\tReplicationFactor: 3}},\n\t\t// Admin options\n\t\tkafka.SetAdminOperationTimeout(maxDur))\n\tif err != nil {\n\t\tfmt.Printf(\"Admin Client request error: %v\\n\", err)\n\t\tos.Exit(1)\n\t}\n\tfor _, result := range results {\n\t\tif result.Error.Code() != kafka.ErrNoError && result.Error.Code() != kafka.ErrTopicAlreadyExists {\n\t\t\tfmt.Printf(\"Failed to create topic: %v\\n\", result.Error)\n\t\t\tos.Exit(1)\n\t\t}\n\t\tfmt.Printf(\"%v\\n\", result)\n\t}\n\ta.Close()\n\n}", "func New() Publisher {\n\treturn &publisher{\n\t\ttopics: make(map[string]*topic),\n\t}\n}", "func GetOrCreateTopic(name string, ts []string) *Topic {\n\ttopic, err := GetTopic(name)\n\tif err != nil {\n\t\ttopic = NewTopic(name, ts)\n\t\ttopic.Save()\n\t\treturn topic\n\t}\n\treturn topic\n}", "func (eventNotifications *EventNotificationsV1) CreateTopic(createTopicOptions *CreateTopicOptions) (result *TopicResponse, response *core.DetailedResponse, err error) {\n\treturn eventNotifications.CreateTopicWithContext(context.Background(), createTopicOptions)\n}", "func (svc *SNS) CreateTopic(name string) (*Topic, error) {\n\tarn, err := svc.createTopic(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttopic := NewTopic(svc, arn, name)\n\treturn topic, nil\n}", "func ExampleTopicsClient_CreateOrUpdate() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armservicebus.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := clientFactory.NewTopicsClient().CreateOrUpdate(ctx, \"ArunMonocle\", \"sdk-Namespace-1617\", \"sdk-Topics-5488\", armservicebus.SBTopic{\n\t\tProperties: &armservicebus.SBTopicProperties{\n\t\t\tEnableExpress: to.Ptr(true),\n\t\t},\n\t}, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// You could use response here. We use blank identifier for just demo purposes.\n\t_ = res\n\t// If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t// res.SBTopic = armservicebus.SBTopic{\n\t// \tName: to.Ptr(\"sdk-Topics-5488\"),\n\t// \tType: to.Ptr(\"Microsoft.ServiceBus/Namespaces/Topics\"),\n\t// \tID: to.Ptr(\"/subscriptions/5f750a97-50d9-4e36-8081-c9ee4c0210d4/resourceGroups/ArunMonocle/providers/Microsoft.ServiceBus/namespaces/sdk-Namespace-1617/topics/sdk-Topics-5488\"),\n\t// \tProperties: &armservicebus.SBTopicProperties{\n\t// \t\tAccessedAt: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2017-05-26T20:50:34.32Z\"); return t}()),\n\t// \t\tAutoDeleteOnIdle: to.Ptr(\"P10675199DT2H48M5.4775807S\"),\n\t// \t\tCreatedAt: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2017-05-26T20:50:34.1Z\"); return t}()),\n\t// \t\tDefaultMessageTimeToLive: to.Ptr(\"P10675199DT2H48M5.4775807S\"),\n\t// \t\tDuplicateDetectionHistoryTimeWindow: to.Ptr(\"PT10M\"),\n\t// \t\tEnableBatchedOperations: to.Ptr(true),\n\t// \t\tEnableExpress: to.Ptr(true),\n\t// \t\tEnablePartitioning: to.Ptr(false),\n\t// \t\tMaxMessageSizeInKilobytes: to.Ptr[int64](10240),\n\t// \t\tMaxSizeInMegabytes: to.Ptr[int32](10240),\n\t// \t\tRequiresDuplicateDetection: to.Ptr(false),\n\t// \t\tSizeInBytes: to.Ptr[int64](0),\n\t// \t\tStatus: to.Ptr(armservicebus.EntityStatusActive),\n\t// \t\tSubscriptionCount: to.Ptr[int32](0),\n\t// \t\tSupportOrdering: to.Ptr(true),\n\t// \t\tUpdatedAt: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2017-05-26T20:50:34.32Z\"); return t}()),\n\t// \t},\n\t// }\n}", "func (p *PubsubValueStore) createTopicHandler(topic string) (*topicInfo, error) {\n\tt, err := p.ps.Join(topic)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsub, err := t.Subscribe()\n\tif err != nil {\n\t\t_ = t.Close()\n\t\treturn nil, err\n\t}\n\n\tevts, err := t.EventHandler()\n\tif err != nil {\n\t\tsub.Cancel()\n\t\t_ = t.Close()\n\t}\n\n\tti := &topicInfo{\n\t\ttopic: t,\n\t\tevts: evts,\n\t\tsub: sub,\n\t\tfinished: make(chan struct{}, 1),\n\t}\n\n\treturn ti, nil\n}", "func newTopicManager(dataDir string, cc copycat.CopyCat) (*topicManager, error) {\n\ttm := &topicManager{\n\t\ttopicNamesToIds: make(map[string]string),\n\t\ttopicNamesToLogs: make(map[string]*logContainer),\n\t\ttopicsMutex: &sync.RWMutex{},\n\t\tcc: cc,\n\t\tdataDir: dataDir,\n\t\tlogger: logrus.WithFields(logrus.Fields{}),\n\t}\n\n\tvar err error\n\ttm.proposeCh, tm.commitCh, tm.errorCh, tm.snapshotConsumer, err = cc.SubscribeToDataStructureWithStringID(\"01CFRDSD7PBQZXV8N515RVYTZQ\", tm.snapshotProvider)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tgo tm.serveChannels()\n\treturn tm, nil\n}", "func CreateCreateTopicRequest() (request *CreateTopicRequest) {\n\trequest = &CreateTopicRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"alikafka\", \"2019-09-16\", \"CreateTopic\", \"alikafka\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (ns *Namespace) NewTopic(name string, opts ...TopicOption) (*Topic, error) {\n\ttopic := &Topic{\n\t\tsendingEntity: newSendingEntity(newEntity(name, topicManagementPath(name), ns)),\n\t}\n\n\tfor i := range opts {\n\t\tif err := opts[i](topic); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn topic, nil\n}", "func New() *PubSub {\n\treturn &PubSub{\n\t\tMaxSubs: 20,\n\t\tregistry: make(map[string]*topic),\n\t}\n}", "func CreateTopic(topicName string, partitions int, replicas int) error {\r\n\tconn, err := kafka.Dial(Config.GetString(\"kafka.admin.network\"),\r\n\t\tConfig.GetString(\"kafka.admin.address\"))\r\n\tif err != nil {\r\n\t\treturn err\r\n\t}\r\n\tdefer conn.Close()\r\n\treturn conn.CreateTopics(\r\n\t\tkafka.TopicConfig{\r\n\t\t\tTopic: topicName,\r\n\t\t\tNumPartitions: partitions,\r\n\t\t\tReplicationFactor: replicas,\r\n\t\t},\r\n\t)\r\n}", "func (svc *SNS) createTopic(name string) (topicARN string, err error) {\n\ttopicName := svc.prefix + name\n\tin := &SDK.CreateTopicInput{\n\t\tName: pointers.String(topicName),\n\t}\n\tresp, err := svc.client.CreateTopic(in)\n\tif err != nil {\n\t\tsvc.Errorf(\"error on `CreateTopic` operation; name=%s; error=%s;\", name, err.Error())\n\t\treturn \"\", err\n\t}\n\treturn *resp.TopicArn, nil\n}", "func (c *Client) Topic(id string) *Topic {\n\treturn &Topic{c.Client.Topic(id), c.projectID, c.sensor}\n}", "func (mr *MockC2ClientMockRecorder) NewTopic(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"NewTopic\", reflect.TypeOf((*MockC2Client)(nil).NewTopic), varargs...)\n}", "func New(host string) *PubSub {\n\tps := PubSub{\n\t\thost: host,\n\t}\n\n\t// PRETEND THERE IS A SPECIFIC IMPLEMENTATION.\n\n\treturn &ps\n}", "func NewTopicRouter(br fiber.Router, conn *config.DBConn) {\n\trouter := br.Group(\"/topics\")\n\ttopicRepo := repository.NewTopicRepository(conn)\n\tpostRepo := repository.NewPostRepository(conn)\n\tuserRepo := repository.NewUserRepository(conn)\n\ttopicService := services.NewTopicService(topicRepo, postRepo, userRepo)\n\ttopicController := controllers.NewTopicController(topicService)\n\n\trouter.Get(\"/:id\", topicController.GetOne)\n\trouter.Get(\"/\", topicController.GetAll)\n\trouter.Post(\"/\", topicController.Create)\n}", "func (c *Cluster) AddTopic(toml []byte) error {\n\treturn c.createTopic(toml, false)\n}", "func CreateTopicOperation(k8sh *utils.K8sHelper, manifests installer.CephManifests) *TopicOperation {\n\treturn &TopicOperation{k8sh, manifests}\n}", "func GenerateTopic(name string, s v1alpha1.TopicParameters) *pubsub.Topic {\n\tt := &pubsub.Topic{\n\t\tName: name,\n\t\tLabels: s.Labels,\n\t\tKmsKeyName: gcp.StringValue(s.KmsKeyName),\n\t}\n\tif s.MessageStoragePolicy != nil {\n\t\tt.MessageStoragePolicy = &pubsub.MessageStoragePolicy{\n\t\t\tAllowedPersistenceRegions: s.MessageStoragePolicy.AllowedPersistenceRegions,\n\t\t}\n\t}\n\tif s.MessageRetentionDuration != nil {\n\t\tt.MessageRetentionDuration = gcp.StringValue(s.MessageRetentionDuration)\n\t}\n\treturn t\n}", "func NewWithMetadata(t topics.Topic, payload interface{}, metadata *Metadata) Message {\n\tsafePayload := convertToSafePayload(payload)\n\treturn &simple{category: t, payload: safePayload, metadata: metadata}\n}", "func New() *PubSubChannel {\n return &PubSubChannel{\n subscriptions: make([]*Subscription, 0),\n }\n}", "func newTopicscope(topic, scope string) (ts topicscope) {\n\tif len(topic) == 0 || topic == \"*\" {\n\t\tts.Topic = \"any\"\n\t} else {\n\t\tts.Topic = topic\n\t}\n\tif len(scope) == 0 || scope == \"*\" {\n\t\tts.Scope = \"any\"\n\t} else {\n\t\tts.Scope = scope\n\t}\n\treturn\n}", "func TestSubscriptionTopic(t *testing.T) {\n\tsub := NewFSubscription(\"foo\", nil)\n\tassert.Equal(t, \"foo\", sub.Topic())\n}", "func New(factory func() interface{}) *Server {\n\tt := topic.New()\n\tt.AddSubscriber(1, &subscriber{state: factory()})\n\treturn &Server{topic: t}\n}", "func New(nameClient string) *PubSub {\n\n\tp := &PubSub{}\n\tsubscriptions = make(map[string]chan []byte)\n\n\topts := MQTT.NewClientOptions().AddBroker(\"tcp://127.0.0.1:1883\")\n\topts.SetClientID(nameClient)\n\topts.SetDefaultPublishHandler(f)\n\topts.SetOnConnectHandler(onConnection)\n\topts.SetAutoReconnect(true)\n\tp.Conn = MQTT.NewClient(opts)\n\treturn p\n}", "func (c *Client) Topic(text string) error {\n\tvar args []byte\n\n\targs = combine(args, \"TOPIC \")\n\targs = combine(args, text)\n\n\treturn c.Send(MESSAGE_CHAT_MESSAGE, args)\n}", "func (c *Client) Topic(target, topic string) error {\n\treturn c.Raw(\"TOPIC %s :%s\", target, topic)\n}", "func SubCreate(w http.ResponseWriter, r *http.Request) {\n\n\t// Init output\n\toutput := []byte(\"\")\n\n\t// Add content type header to the response\n\tcontentType := \"application/json\"\n\tcharset := \"utf-8\"\n\tw.Header().Add(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\t// Grab url path variables\n\turlVars := mux.Vars(r)\n\n\t// Grab context references\n\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\trefBrk := gorillaContext.Get(r, \"brk\").(brokers.Broker)\n\tprojectUUID := gorillaContext.Get(r, \"auth_project_uuid\").(string)\n\n\t// Read POST JSON body\n\tbody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\terr := APIErrorInvalidRequestBody()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Parse pull options\n\tpostBody, err := subscriptions.GetFromJSON(body)\n\tif err != nil {\n\t\terr := APIErrorInvalidArgument(\"Subscription\")\n\t\trespondErr(w, err)\n\t\tlog.Error(string(body[:]))\n\t\treturn\n\t}\n\n\ttProject, tName, err := subscriptions.ExtractFullTopicRef(postBody.FullTopic)\n\n\tif err != nil {\n\t\terr := APIErrorInvalidName(\"Topic\")\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\tif topics.HasTopic(projectUUID, tName, refStr) == false {\n\t\terr := APIErrorNotFound(\"Topic\")\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Get current topic offset\n\ttProjectUUID := projects.GetUUIDByName(tProject, refStr)\n\tfullTopic := tProjectUUID + \".\" + tName\n\tcurOff := refBrk.GetMaxOffset(fullTopic)\n\n\tpushEnd := \"\"\n\trPolicy := \"\"\n\trPeriod := 0\n\tmaxMessages := int64(1)\n\n\t//pushWorker := auth.User{}\n\tverifyHash := \"\"\n\n\tif postBody.PushCfg != (subscriptions.PushConfig{}) {\n\n\t\t// check the state of the push functionality\n\t\tpwToken := gorillaContext.Get(r, \"push_worker_token\").(string)\n\t\tpushEnabled := gorillaContext.Get(r, \"push_enabled\").(bool)\n\n\t\tif !pushEnabled {\n\t\t\terr := APIErrorPushConflict()\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\t_, err = auth.GetPushWorker(pwToken, refStr)\n\t\tif err != nil {\n\t\t\terr := APIErrInternalPush()\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\tpushEnd = postBody.PushCfg.Pend\n\t\t// Check if push endpoint is not a valid https:// endpoint\n\t\tif !(isValidHTTPS(pushEnd)) {\n\t\t\terr := APIErrorInvalidData(\"Push endpoint should be addressed by a valid https url\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t\trPolicy = postBody.PushCfg.RetPol.PolicyType\n\t\trPeriod = postBody.PushCfg.RetPol.Period\n\t\tmaxMessages = postBody.PushCfg.MaxMessages\n\n\t\tif rPolicy == \"\" {\n\t\t\trPolicy = subscriptions.LinearRetryPolicyType\n\t\t}\n\n\t\tif maxMessages == 0 {\n\t\t\tmaxMessages = int64(1)\n\t\t}\n\n\t\tif rPeriod <= 0 {\n\t\t\trPeriod = 3000\n\t\t}\n\n\t\tif !subscriptions.IsRetryPolicySupported(rPolicy) {\n\t\t\terr := APIErrorInvalidData(subscriptions.UnSupportedRetryPolicyError)\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\tverifyHash, err = auth.GenToken()\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Could not generate verification hash for subscription %v, %v\", urlVars[\"subscription\"], err.Error())\n\t\t\terr := APIErrGenericInternal(\"Could not generate verification hash\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t}\n\n\t// Get Result Object\n\tres, err := subscriptions.CreateSub(projectUUID, urlVars[\"subscription\"], tName, pushEnd, curOff, maxMessages, postBody.Ack, rPolicy, rPeriod, verifyHash, false, refStr)\n\n\tif err != nil {\n\t\tif err.Error() == \"exists\" {\n\t\t\terr := APIErrorConflict(\"Subscription\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t\terr := APIErrGenericInternal(err.Error())\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Output result to JSON\n\tresJSON, err := res.ExportJSON()\n\tif err != nil {\n\t\terr := APIErrExportJSON()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Write response\n\toutput = []byte(resJSON)\n\trespondOK(w, output)\n\n}", "func TopicPublish(w http.ResponseWriter, r *http.Request) {\n\t// Init output\n\toutput := []byte(\"\")\n\n\t// Add content type header to the response\n\tcontentType := \"application/json\"\n\tcharset := \"utf-8\"\n\tw.Header().Add(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\t// Get url path variables\n\turlVars := mux.Vars(r)\n\turlTopic := urlVars[\"topic\"]\n\n\t// Grab context references\n\n\trefBrk := gorillaContext.Get(r, \"brk\").(brokers.Broker)\n\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\trefUserUUID := gorillaContext.Get(r, \"auth_user_uuid\").(string)\n\trefRoles := gorillaContext.Get(r, \"auth_roles\").([]string)\n\trefAuthResource := gorillaContext.Get(r, \"auth_resource\").(bool)\n\t// Get project UUID First to use as reference\n\tprojectUUID := gorillaContext.Get(r, \"auth_project_uuid\").(string)\n\n\tresults, err := topics.Find(projectUUID, \"\", urlVars[\"topic\"], \"\", 0, refStr)\n\n\tif err != nil {\n\t\terr := APIErrGenericBackend()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// If not found\n\tif results.Empty() {\n\t\terr := APIErrorNotFound(\"Topic\")\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\tres := results.Topics[0]\n\n\t// Check Authorization per topic\n\t// - if enabled in config\n\t// - if user has only publisher role\n\n\tif refAuthResource && auth.IsPublisher(refRoles) {\n\n\t\tif auth.PerResource(projectUUID, \"topics\", urlTopic, refUserUUID, refStr) == false {\n\t\t\terr := APIErrorForbidden()\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t}\n\n\t// Read POST JSON body\n\tbody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\terr := APIErrorInvalidRequestBody()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Create Message List from Post JSON\n\tmsgList, err := messages.LoadMsgListJSON(body)\n\tif err != nil {\n\t\terr := APIErrorInvalidArgument(\"Message\")\n\t\trespondErr(w, err)\n\t\tlog.Error(string(body[:]))\n\t\treturn\n\t}\n\n\t// check if the topic has a schema associated with it\n\tif res.Schema != \"\" {\n\n\t\t// retrieve the schema\n\t\t_, schemaName, err := schemas.ExtractSchema(res.Schema)\n\t\tif err != nil {\n\t\t\tlog.WithFields(\n\t\t\t\tlog.Fields{\n\t\t\t\t\t\"type\": \"service_log\",\n\t\t\t\t\t\"schema_name\": res.Schema,\n\t\t\t\t\t\"topic_name\": res.Name,\n\t\t\t\t\t\"error\": err.Error(),\n\t\t\t\t},\n\t\t\t).Error(\"Could not extract schema name\")\n\t\t\terr := APIErrGenericInternal(schemas.GenericError)\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\tsl, err := schemas.Find(projectUUID, \"\", schemaName, refStr)\n\n\t\tif err != nil {\n\t\t\tlog.WithFields(\n\t\t\t\tlog.Fields{\n\t\t\t\t\t\"type\": \"service_log\",\n\t\t\t\t\t\"schema_name\": schemaName,\n\t\t\t\t\t\"topic_name\": res.Name,\n\t\t\t\t\t\"error\": err.Error(),\n\t\t\t\t},\n\t\t\t).Error(\"Could not retrieve schema from the store\")\n\t\t\terr := APIErrGenericInternal(schemas.GenericError)\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\tif !sl.Empty() {\n\t\t\terr := schemas.ValidateMessages(sl.Schemas[0], msgList)\n\t\t\tif err != nil {\n\t\t\t\tif err.Error() == \"500\" {\n\t\t\t\t\terr := APIErrGenericInternal(schemas.GenericError)\n\t\t\t\t\trespondErr(w, err)\n\t\t\t\t\treturn\n\t\t\t\t} else {\n\t\t\t\t\terr := APIErrorInvalidData(err.Error())\n\t\t\t\t\trespondErr(w, err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tlog.WithFields(\n\t\t\t\tlog.Fields{\n\t\t\t\t\t\"type\": \"service_log\",\n\t\t\t\t\t\"schema_name\": res.Schema,\n\t\t\t\t\t\"topic_name\": res.Name,\n\t\t\t\t},\n\t\t\t).Error(\"List of schemas was empty\")\n\t\t\terr := APIErrGenericInternal(schemas.GenericError)\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t}\n\n\t// Init message ids list\n\tmsgIDs := messages.MsgIDs{IDs: []string{}}\n\n\t// For each message in message list\n\tfor _, msg := range msgList.Msgs {\n\t\t// Get offset and set it as msg\n\t\tfullTopic := projectUUID + \".\" + urlTopic\n\n\t\tmsgID, rTop, _, _, err := refBrk.Publish(fullTopic, msg)\n\n\t\tif err != nil {\n\t\t\tif err.Error() == \"kafka server: Message was too large, server rejected it to avoid allocation error.\" {\n\t\t\t\terr := APIErrTooLargeMessage(\"Message size too large\")\n\t\t\t\trespondErr(w, err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\terr := APIErrGenericBackend()\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\tmsg.ID = msgID\n\t\t// Assertions for Succesfull Publish\n\t\tif rTop != fullTopic {\n\t\t\terr := APIErrGenericInternal(\"Broker reports wrong topic\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\t// Append the MsgID of the successful published message to the msgIds list\n\t\tmsgIDs.IDs = append(msgIDs.IDs, msg.ID)\n\t}\n\n\t// timestamp of the publish event\n\tpublishTime := time.Now().UTC()\n\n\t// amount of messages published\n\tmsgCount := int64(len(msgList.Msgs))\n\n\t// increment topic number of message metric\n\trefStr.IncrementTopicMsgNum(projectUUID, urlTopic, msgCount)\n\n\t// increment daily count of topic messages\n\tyear, month, day := publishTime.Date()\n\trefStr.IncrementDailyTopicMsgCount(projectUUID, urlTopic, msgCount, time.Date(year, month, day, 0, 0, 0, 0, time.UTC))\n\n\t// increment topic total bytes published\n\trefStr.IncrementTopicBytes(projectUUID, urlTopic, msgList.TotalSize())\n\n\t// update latest publish date for the given topic\n\trefStr.UpdateTopicLatestPublish(projectUUID, urlTopic, publishTime)\n\n\t// count the rate of published messages per sec between the last two publish events\n\tvar dt float64 = 1\n\t// if its the first publish to the topic\n\t// skip the subtraction that computes the DT between the last two publish events\n\tif !res.LatestPublish.IsZero() {\n\t\tdt = publishTime.Sub(res.LatestPublish).Seconds()\n\t}\n\trefStr.UpdateTopicPublishRate(projectUUID, urlTopic, float64(msgCount)/dt)\n\n\t// Export the msgIDs\n\tresJSON, err := msgIDs.ExportJSON()\n\tif err != nil {\n\t\terr := APIErrExportJSON()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Write response\n\toutput = []byte(resJSON)\n\trespondOK(w, output)\n}", "func createTopics(kafkaAddress string, topics ...string) error {\n\n\tconn, err := kafka.Dial(\"tcp\", kafkaAddress)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"event-publisher.createTopics\")\n\t}\n\tdefer conn.Close()\n\n\tcontroller, err := conn.Controller()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"event-publisher.createTopics\")\n\t}\n\tvar controllerConn *kafka.Conn\n\tcontrollerConn, err = kafka.Dial(\"tcp\", net.JoinHostPort(controller.Host, strconv.Itoa(controller.Port)))\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"event-publisher.createTopics\")\n\t}\n\tdefer controllerConn.Close()\n\tfor _, topic := range topics {\n\n\t\ttopicConfigs := []kafka.TopicConfig{\n\t\t\tkafka.TopicConfig{\n\t\t\t\tTopic: topic,\n\t\t\t\tNumPartitions: 1,\n\t\t\t\tReplicationFactor: 1,\n\t\t\t},\n\t\t}\n\n\t\terr = controllerConn.CreateTopics(topicConfigs...)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"event-publisher.createTopics\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func (n NotifyForumTopic) construct() NotifyPeerClass { return &n }", "func TestTopicNotExist(t *testing.T) {\n\tdefer log4go.Close()\n\tSetUp(t)\n\ttalosAdminMock.EXPECT().DescribeTopic(&topic.DescribeTopicRequest{topicName}).Return(nil, errors.New(\"topic not exist\")).Times(4)\n\n\ttalosProducer = producer.NewTalosProducerForTest(talosProducerConfig,\n\t\ttalosClientFactoryMock, talosAdminMock,\n\t\t&topic.TopicTalosResourceName{resourceName},\n\t\t&client.SimpleTopicAbnormalCallback{}, &TestCallback{})\n\n\tbigStr := getRandomString(utils.TALOS_SINGLE_MESSAGE_BYTES_MAXIMAL + 1)\n\tlist := make([]*message.Message, 0)\n\tmsg2 := &message.Message{Message: []byte(bigStr)}\n\tlist = append(list, msg2)\n\terr := talosProducer.AddUserMessage(list)\n\tassert.NotNil(t, err)\n}", "func CreateMessage(topic string)(msg Message){\n msg = C.CreateMessage(C.CString(topic))\n return msg;\n}", "func (client *Client) CreateTopicWithCallback(request *CreateTopicRequest, callback func(response *CreateTopicResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *CreateTopicResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.CreateTopic(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (es *EventStore) AddTopic(topic Topic) (string, error) {\n\tstart := time.Now()\n\tdefer func() {\n\t\tmetrics.EventStoreLatency(\"AddTopic\", start)\n\t}()\n\n\tname := strings.ToLower(topic.Name)\n\tschema := topic.Schema\n\n\tif name == \"\" {\n\t\treturn \"\", errors.New(\"Topic name cannot be empty\")\n\t} else if es.getTopicID(name) != \"\" {\n\t\treturn \"\", jh.NewError(errors.New(\"Topic with name already exists\").Error(), http.StatusConflict)\n\t}\n\n\tschemaStr := \"{}\"\n\tif schema != nil {\n\t\tschemaBytes, err := json.Marshal(schema)\n\t\tif err != nil {\n\t\t\treturn \"\", jh.NewError(errors.Wrap(err, \"Error marshalling schema into json\").Error(), http.StatusBadRequest)\n\t\t}\n\t\tschemaStr = string(schemaBytes)\n\t}\n\n\tjsonSchema, ok := es.validateSchema(schemaStr)\n\tif !ok {\n\t\treturn \"\", jh.NewError(errors.New(\"Error adding topic - schema is not in valid JSON format\").Error(), http.StatusBadRequest)\n\t}\n\n\tid := uuid.NewV4().String()\n\tif err := es.ds.AddTopic(RawTopic{\n\t\tID: id,\n\t\tName: name,\n\t\tSchema: schemaStr,\n\t}); err != nil {\n\t\tmetrics.DBError(\"write\")\n\t\treturn \"\", errors.Wrap(err, \"Error adding topic to data source\")\n\t}\n\n\tes.topicMutex.Lock()\n\tes.topicNameToID[name] = id\n\tes.topicIDToName[id] = name\n\tes.topicSchemaPropertiesMap[id] = schema\n\tes.topicSchemaMap[id] = jsonSchema\n\tes.topicMutex.Unlock()\n\n\treturn id, nil\n}", "func (s Site) AddTopic(t string) {\n\ts.Handler.HandleFunc(\"/\"+t+\"/\", s.TopicHandler)\n\ts.Handler.HandleFunc(\"/nominate/\"+t+\"/\", s.NominateHandler)\n}", "func (h *hub) findOrCreateTopic(topic []string) (found *topicNode, err error) {\n\tfound, rest, _ := h.root.MaybeFindTopic(topic)\n\n\tif len(rest) != 0 {\n\t\tfound, err = found.CreateChild(rest)\n\t}\n\n\treturn found, err\n}", "func New(w http.ResponseWriter, r *http.Request) {\r\n\ttmpl.ExecuteTemplate(w, \"New\", nil)\r\n}", "func New(w http.ResponseWriter, r *http.Request) {\r\n\ttmpl.ExecuteTemplate(w, \"New\", nil)\r\n}", "func TopicPost(ctx *context.Context) {\n\tif ctx.User == nil {\n\t\tctx.JSON(403, map[string]interface{}{\n\t\t\t\"message\": \"Only owners could change the topics.\",\n\t\t})\n\t\treturn\n\t}\n\n\tvar topics = make([]string, 0)\n\tvar topicsStr = strings.TrimSpace(ctx.Query(\"topics\"))\n\tif len(topicsStr) > 0 {\n\t\ttopics = strings.Split(topicsStr, \",\")\n\t}\n\n\terr := models.SaveTopics(ctx.Repo.Repository.ID, topics...)\n\tif err != nil {\n\t\tlog.Error(2, \"SaveTopics failed: %v\", err)\n\t\tctx.JSON(500, map[string]interface{}{\n\t\t\t\"message\": \"Save topics failed.\",\n\t\t})\n\t\treturn\n\t}\n\n\tctx.JSON(200, map[string]interface{}{\n\t\t\"status\": \"ok\",\n\t})\n}", "func (zj *ZapJournal) Topic(topic string) Writer {\n\treturn &ZapWriter{\n\t\tlogger: zj.logger.Sugar().Named(topic),\n\t\ttopic: topic,\n\t}\n}", "func (_Flytrap *FlytrapTransactor) AddTopic(opts *bind.TransactOpts, topic string, country [2]byte, addPubCost *big.Int, addSubCost *big.Int, reason string, sensitive bool) (*types.Transaction, error) {\n\treturn _Flytrap.contract.Transact(opts, \"addTopic\", topic, country, addPubCost, addSubCost, reason, sensitive)\n}", "func New(pubsub, blob msg.Transport, change int) *Transport {\n\tif change <= 0 {\n\t\tpanic(\"WTF are you doing?\")\n\t}\n\treturn &Transport{\n\t\tpubsub: pubsub,\n\t\tblob: blob,\n\t\tchange: change,\n\t}\n}", "func New(t T) *Tester {\n\ttt := &Tester{\n\t\tt: t,\n\n\t\tclients: make(map[string]*client),\n\n\t\tcodecs: make(map[string]goka.Codec),\n\t\ttopicQueues: make(map[string]*queue),\n\t\tstorages: make(map[string]storage.Storage),\n\t}\n\ttt.tmgr = NewMockTopicManager(tt, 1, 1)\n\ttt.producer = newProducerMock(tt.handleEmit)\n\n\treturn tt\n}", "func (m *Medium) Add(t Topic) {\n\tm.mx.Lock()\n\tm.topics[t.Name()] = t\n\tm.mx.Unlock()\n}", "func newPubSubMessageBus(topic string) (MessageBus, error) {\n\tvar ctx = context.Background()\n\tservice, err := pubsub.NewService(ctx)\n\tpublisher := pubsub.NewProjectsTopicsService(service)\n\n\tif err != nil {\n\t\tlogrus.Errorf(\"unable to create Pubsub message bus: %v\", err)\n\t\treturn nil, err\n\t}\n\treturn &pubsubMessageBus{ctx, publisher, topic}, nil\n}", "func OpenTopic(client mqtt.Client, topicName string, qos byte) (*pubsub.Topic, error) {\n\tdt, err := openDriverTopic(client, topicName, qos)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn pubsub.NewTopic(dt, nil), nil\n}", "func (*GenericFramework) NewMessage(ctx *MessageContext) {}", "func (t *topicNode) newChild() *topicNode {\n\tn := newNode()\n\tn.parent = t\n\treturn n\n}", "func New(config *MessageProducerConfig) (*messageProducer, error) {\n\tmsgProducer := &messageProducer{\n\t\ttopic: config.Topic,\n\t}\n\n\tproducer, err := nsq.NewProducer(config.Host, nsq.NewConfig())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err := producer.Ping(); err != nil {\n\t\treturn nil, err\n\t}\n\tmsgProducer.producer = producer\n\treturn msgProducer, nil\n}", "func New(w http.ResponseWriter, r *http.Request) {\n\tgetTemplates().ExecuteTemplate(w, \"New\", nil)\n}", "func (conn *Conn) Topic(channel string, topic ...string) {\n\tt := strings.Join(topic, \" \")\n\tif t != \"\" {\n\t\tt = \" :\" + t\n\t}\n\tconn.Raw(TOPIC + \" \" + channel + t)\n}", "func NewHistoriedTopic(name string, length int) HistoriedTopic {\n\treturn TrackHistory(NewTopic(name), length)\n}", "func createThread(writer http.ResponseWriter, request *http.Requet) {\n\tsess, err := session(writer, request)\n\tif err != nil {\n\t\thttp.Redirect(writer, request, \"/login\", 302)\n\t} esle {\n\t\terr = request.ParseForm()\n\t\tif err != nil {\n\t\t\t// danger\n\t\t}\n\t\ttopic := request.PostFromValue(\"topic\")\n\t\tif _, err := user.CreateThread(topic); err != nil {\n\t\t\t// danger\n\t\t}\n\t\thttp.Redirect(writer, request, \"/\", 302)\n\t}\n}" ]
[ "0.8250956", "0.7481734", "0.7237881", "0.7221837", "0.72201616", "0.71867234", "0.7052233", "0.6944838", "0.69335455", "0.6894096", "0.6878326", "0.6877358", "0.68346196", "0.681044", "0.6770804", "0.67220366", "0.6689578", "0.6662674", "0.6654795", "0.66509044", "0.6650095", "0.66346806", "0.6612587", "0.65521246", "0.65015256", "0.6496329", "0.64719015", "0.64554375", "0.64235187", "0.64171046", "0.64093184", "0.63403255", "0.6226633", "0.61929977", "0.61794704", "0.61734176", "0.61636436", "0.60916793", "0.6084251", "0.60056996", "0.59903806", "0.5979106", "0.5963427", "0.59446853", "0.5902215", "0.5899521", "0.58920187", "0.5873588", "0.5858109", "0.58568156", "0.5833557", "0.5829379", "0.5810808", "0.5775788", "0.5758267", "0.575147", "0.5739881", "0.5721438", "0.5720788", "0.5708088", "0.56887734", "0.5675129", "0.56655043", "0.5661537", "0.5638388", "0.56274486", "0.56199694", "0.5619858", "0.5581009", "0.5572244", "0.5555517", "0.5550383", "0.55420125", "0.55304205", "0.5526252", "0.5523592", "0.5520114", "0.5504262", "0.5482557", "0.54463935", "0.5441406", "0.54405046", "0.54209846", "0.54122204", "0.54122204", "0.54099256", "0.5409286", "0.53938496", "0.53908914", "0.53894293", "0.5387008", "0.5385096", "0.5383237", "0.5374357", "0.5368731", "0.53342754", "0.5310458", "0.5307821", "0.52923805", "0.5280344" ]
0.5847545
50
Restore creates a topic with an existing wal
func Restore(name string, walDir string) (*WALTopic, error) { w, err := walFactory.Open(nil, nil, walDir) if err != nil { return nil, errors.Wrapf(err, "Could no open wal for topic %s", name) } m := make(map[int]uint64) return &WALTopic{name: name, walDir: walDir, wal: w, firstIndexInSegment: m}, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *Cluster) RestoreTopic(toml []byte) error {\n\treturn c.createTopic(toml, true)\n}", "func (ctl Controller) Restore(name string) *pitr.Error {\n\terr := ctl.cluster.Stop()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = ctl.cluster.Clear()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tstdout, stderr, runErr := ctl.runner.Run(\"sudo --login --user postgres wal-g backup-fetch %s %s\", ctl.cluster.DataDirectory(), name)\n\n\tif runErr != nil {\n\t\treturn &pitr.Error{\n\t\t\tMessage: runErr.Error(),\n\t\t\tStdout: stdout,\n\t\t\tStderr: stderr,\n\t\t}\n\t}\n\n\tctl.createRecoveryConf(`restore_command = 'bash --login -c \\\"wal-g wal-fetch %f %p\\\"'`)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = ctl.cluster.Start()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (tm *topicManager) createNewTopic(name string, id string) (log, error) {\n\tlc, ok := tm.topicNamesToLogs[name]\n\tif ok {\n\t\tlc.refCount++\n\t\treturn lc.l, nil\n\t}\n\n\t// lazily create the log if we know about the topic/cc ID\n\tl, err := newBadgerLog(tm.dataDir, name, id, tm.cc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttm.topicNamesToLogs[name] = &logContainer{\n\t\tl: l,\n\t\trefCount: 1,\n\t}\n\treturn l, nil\n}", "func (m *MonLeaderDetector) Restore(id int) {\n\t// TODO(student): Implement\n\t_, ok := m.suspected[id]\n\tif ok == true {\n\t\tdelete(m.suspected, id)\n\t\tm.alive[id] = true\n\t}\n\n\t//Publish to subscribers\n\tvar j int\n\tnewLeader := m.Leader()\n\tif m.LeaderChange || m.Allsuspected {\n\t\tfor j < len(m.Channels) {\n\t\t\tm.Channels[j] <- newLeader\n\t\t\tj++\n\t\t}\n\t}\n}", "func (ctl Controller) RestoreToTransactionID(txID int64) *pitr.Error {\n\terr := ctl.cluster.Stop()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = ctl.cluster.Clear()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tstdout, stderr, runErr := ctl.runner.Run(\"sudo --login --user postgres wal-g backup-fetch %s %s\", ctl.cluster.DataDirectory(), \"LATEST\")\n\n\tif runErr != nil {\n\t\treturn &pitr.Error{\n\t\t\tMessage: runErr.Error(),\n\t\t\tStdout: stdout,\n\t\t\tStderr: stderr,\n\t\t}\n\t}\n\n\tctl.createRecoveryConf(\n\t\t`restore_command = 'bash --login -c \\\"wal-g wal-fetch %f %p\\\"'`,\n\t\tfmt.Sprintf(\"recovery_target_xid = %d\", txID),\n\t\t\"recovery_target_action=promote\",\n\t)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif runErr != nil {\n\t\treturn &pitr.Error{\n\t\t\tMessage: runErr.Error(),\n\t\t\tStdout: stdout,\n\t\t\tStderr: stderr,\n\t\t}\n\t}\n\n\terr = ctl.cluster.Start()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func createTopic(broker *sarama.Broker, topicName string) error {\n\ttopicDetail := &sarama.TopicDetail{}\n\ttopicDetail.NumPartitions = int32(1)\n\ttopicDetail.ReplicationFactor = int16(1)\n\ttopicDetail.ConfigEntries = make(map[string]*string)\n\n\ttopicDetails := make(map[string]*sarama.TopicDetail)\n\ttopicDetails[topicName] = topicDetail\n\trequest := sarama.CreateTopicsRequest{\n\t\tTimeout: time.Second * 15,\n\t\tTopicDetails: topicDetails,\n\t}\n\n\t_, err := broker.CreateTopics(&request)\n\treturn err\n}", "func (r *Reconciler) createTopic(ctx context.Context, topicName string, partitions int32, replicationFactor int16, retentionMillis int64) error {\n\n\t// Get The Logger From The Context\n\tlogger := logging.FromContext(ctx)\n\n\t// Create The TopicDefinition\n\tretentionMillisString := strconv.FormatInt(retentionMillis, 10)\n\ttopicDetail := &sarama.TopicDetail{\n\t\tNumPartitions: partitions,\n\t\tReplicationFactor: replicationFactor,\n\t\tReplicaAssignment: nil, // Currently Not Assigning Partitions To Replicas\n\t\tConfigEntries: map[string]*string{\n\t\t\tcommonconstants.KafkaTopicConfigRetentionMs: &retentionMillisString,\n\t\t},\n\t}\n\n\t// Attempt To Create The Topic & Process TopicError Results (Including Success ;)\n\terr := r.adminClient.CreateTopic(ctx, topicName, topicDetail)\n\tif err != nil {\n\t\tlogger := logger.With(zap.Int16(\"KError\", int16(err.Err)))\n\t\tswitch err.Err {\n\t\tcase sarama.ErrNoError:\n\t\t\tlogger.Info(\"Successfully Created New Kafka Topic (ErrNoError)\")\n\t\t\treturn nil\n\t\tcase sarama.ErrTopicAlreadyExists:\n\t\t\tlogger.Info(\"Kafka Topic Already Exists - No Creation Required\")\n\t\t\treturn nil\n\t\tdefault:\n\t\t\tlogger.Error(\"Failed To Create Topic\")\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tlogger.Info(\"Successfully Created New Kafka Topic (Nil TopicError)\")\n\t\treturn nil\n\t}\n}", "func New(name string, baseDir string) (*WALTopic, error) {\n\twalDir := filepath.Join(baseDir, name)\n\toneMB := 1 * 1024 * 1024\n\tw, err := walFactory.NewSize(nil, nil, walDir, oneMB, false)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"Could not create wal for topic %s\", name)\n\t}\n\tm := make(map[int]uint64)\n\treturn &WALTopic{name: name, walDir: walDir, wal: w, firstIndexInSegment: m}, nil\n}", "func createTopic(brokerAddress string, topicName string,\n\tnumPartitions int32, replicationFactor int32) {\n\tbrokerAddresses := []string{brokerAddress}\n\n\t// configuration handling\n\t// please note the versioning, it needs to be specified explicitly\n\tconfig := sarama.NewConfig()\n\tconfig.Version = sarama.V2_1_0_0\n\n\t// try to create and initialize cluster admin object\n\t// (it will reach out Kafka broker, so it checks the connection as well)\n\tclusterAdmin, err := sarama.NewClusterAdmin(brokerAddresses, config)\n\n\t// check if cluster admin has been initialized successfully\n\tif err != nil {\n\t\tlog.Fatal(\"Error: create cluster admin: \", err.Error())\n\t}\n\n\t// everything's seems to be ok\n\tlog.Print(\"Cluster admin has been initialized\")\n\n\t// cluster admin needs to be closed properly\n\tdefer func() {\n\t\t// try to close cluster admin\n\t\terr := clusterAdmin.Close()\n\n\t\t// check if cluster admin has been closed successfully\n\t\tif err != nil {\n\t\t\tlog.Fatal(\"Error: close cluster admin: \", err.Error())\n\t\t}\n\t}()\n\n\t// try to create a new topic via cluster admin\n\terr = clusterAdmin.CreateTopic(topicName,\n\t\t&sarama.TopicDetail{\n\t\t\tNumPartitions: numPartitions,\n\t\t\tReplicationFactor: int16(replicationFactor),\n\t\t}, false)\n\n\t// check if topic has been created successfully\n\tif err != nil {\n\t\tlog.Fatalf(\"Error: create topic '%s': %v\", topicName, err.Error())\n\t}\n\n\t// everything's seems to be ok -> topic has been created\n\tlog.Printf(\"Topic '%s' has been created\", topicName)\n}", "func restore(ccmd *cobra.Command, args []string) error {\n\tkeyBytes, err := getPrivateKey()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar mdk types.MasterDerivationKey\n\tcopy(mdk[:], keyBytes)\n\tcwResponse, err := kmdClient.CreateWallet(WalletName, WalletPassword, kmd.DefaultWalletDriver, mdk)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error creating wallet - %s\", err)\n\t}\n\n\tfmt.Printf(\"Created wallet '%s' with ID: %s\\n\", cwResponse.Wallet.Name, cwResponse.Wallet.ID)\n\tif os.Getenv(\"GOTEST\") == \"true\" {\n\t\tccmd.Print(\"Created wallet successfully.\")\n\t}\n\n\treturn nil\n}", "func TestRestoreInvalidLearner(t *testing.T) {\n\ts := pb.Snapshot{\n\t\tMetadata: pb.SnapshotMetadata{\n\t\t\tIndex: 11, // magic number\n\t\t\tTerm: 11, // magic number\n\t\t\tConfState: pb.ConfState{Nodes: []uint64{1, 2}, Learners: []uint64{3}},\n\t\t},\n\t}\n\n\tstorage := NewMemoryStorage()\n\tsm := newTestRaft(3, []uint64{1, 2, 3}, 10, 1, storage)\n\tdefer closeAndFreeRaft(sm)\n\n\tif sm.isLearner {\n\t\tt.Errorf(\"%x is learner, want not\", sm.id)\n\t}\n\tif ok := sm.restore(s); ok {\n\t\tt.Error(\"restore succeed, want fail\")\n\t}\n}", "func TestRestoreLearnerPromotion(t *testing.T) {\n\ts := pb.Snapshot{\n\t\tMetadata: pb.SnapshotMetadata{\n\t\t\tIndex: 11, // magic number\n\t\t\tTerm: 11, // magic number\n\t\t\tConfState: pb.ConfState{Nodes: []uint64{1, 2, 3}},\n\t\t},\n\t}\n\n\tstorage := NewMemoryStorage()\n\tsm := newTestLearnerRaft(3, []uint64{1, 2}, []uint64{3}, 10, 1, storage)\n\tdefer closeAndFreeRaft(sm)\n\n\tif !sm.isLearner {\n\t\tt.Errorf(\"%x is not learner, want yes\", sm.id)\n\t}\n\n\tif ok := sm.restore(s); !ok {\n\t\tt.Error(\"restore fail, want succeed\")\n\t}\n\n\tif sm.isLearner {\n\t\tt.Errorf(\"%x is learner, want not\", sm.id)\n\t}\n}", "func TestRestoreWithLearner(t *testing.T) {\n\ts := pb.Snapshot{\n\t\tMetadata: pb.SnapshotMetadata{\n\t\t\tIndex: 11, // magic number\n\t\t\tTerm: 11, // magic number\n\t\t\tConfState: pb.ConfState{Nodes: []uint64{1, 2}, Learners: []uint64{3}},\n\t\t},\n\t}\n\n\tstorage := NewMemoryStorage()\n\tsm := newTestLearnerRaft(3, []uint64{1, 2}, []uint64{3}, 10, 1, storage)\n\tdefer closeAndFreeRaft(sm)\n\tif ok := sm.restore(s); !ok {\n\t\tt.Error(\"restore fail, want succeed\")\n\t}\n\n\tif sm.raftLog.lastIndex() != s.Metadata.Index {\n\t\tt.Errorf(\"log.lastIndex = %d, want %d\", sm.raftLog.lastIndex(), s.Metadata.Index)\n\t}\n\tif mustTerm(sm.raftLog.term(s.Metadata.Index)) != s.Metadata.Term {\n\t\tt.Errorf(\"log.lastTerm = %d, want %d\", mustTerm(sm.raftLog.term(s.Metadata.Index)), s.Metadata.Term)\n\t}\n\tif !sm.isLearner {\n\t\tt.Errorf(\"%x is not learner, want yes\", sm.id)\n\t}\n\tsg := sm.nodes()\n\tif len(sg)+len(sm.learnerNodes()) != len(s.Metadata.ConfState.Nodes)+len(s.Metadata.ConfState.Learners) {\n\t\tt.Errorf(\"sm.Nodes = %+v, length not equal with %+v\", sg, s.Metadata.ConfState)\n\t}\n\tfor _, n := range s.Metadata.ConfState.Nodes {\n\t\tif sm.prs[n].IsLearner {\n\t\t\tt.Errorf(\"sm.Node %x isLearner = %s, want %t\", n, sm.prs[n], false)\n\t\t}\n\t}\n\tif len(s.Metadata.ConfState.Nodes) != len(sm.prs) {\n\t\tt.Errorf(\"sm.Nodes = %+v, length not equal with %+v\", sm.prs, s.Metadata.ConfState.Nodes)\n\t}\n\tfor _, n := range s.Metadata.ConfState.Learners {\n\t\tif !sm.learnerPrs[n].IsLearner {\n\t\t\tt.Errorf(\"sm.Node %x isLearner = %s, want %t\", n, sm.prs[n], true)\n\t\t}\n\t}\n\tif len(s.Metadata.ConfState.Learners) != len(sm.learnerPrs) {\n\t\tt.Errorf(\"sm.Nodes = %+v, length not equal with %+v\", sm.learnerPrs, s.Metadata.ConfState.Learners)\n\t}\n\n\tif ok := sm.restore(s); ok {\n\t\tt.Error(\"restore succeed, want fail\")\n\t}\n}", "func (p *PubsubValueStore) createTopicHandler(topic string) (*topicInfo, error) {\n\tt, err := p.ps.Join(topic)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsub, err := t.Subscribe()\n\tif err != nil {\n\t\t_ = t.Close()\n\t\treturn nil, err\n\t}\n\n\tevts, err := t.EventHandler()\n\tif err != nil {\n\t\tsub.Cancel()\n\t\t_ = t.Close()\n\t}\n\n\tti := &topicInfo{\n\t\ttopic: t,\n\t\tevts: evts,\n\t\tsub: sub,\n\t\tfinished: make(chan struct{}, 1),\n\t}\n\n\treturn ti, nil\n}", "func (gWal *GenericWAL) Recover() error {\n entryCh := make(chan *Entry)\n quitCh := make(chan struct{})\n defer close(quitCh)\n go func() {\n if err := gWal.wal.StartRecovery(entryCh, quitCh); err != nil {\n glog.Errorf(\"could not recover state from local wal :: %v\", err)\n entryCh <- nil\n }\n }()\n\n count := 0\n for entry := range entryCh {\n if entry == nil {\n glog.Errorf(\"wal recovery channel is closed unexpectedly\")\n return fmt.Errorf(\"wal error\")\n }\n count++\n\n switch entry.Type {\n case CTypeDone:\n glog.Infof(\"wal recovery is complete because last record is read\")\n close(entryCh)\n\n case CTypeData:\n gWal.updateRecoveryMap(false /* checkpoint */, int64(entry.LSN),\n entry.Data)\n glog.V(1).Infof(\"recovered a delta record with lsn %v\", entry.LSN)\n\n case CTypeCheckpoint:\n gWal.updateRecoveryMap(true /* checkpoint */, -1, entry.Data)\n glog.V(1).Infof(\"recovered a checkpoint record with lsn %v\", entry.LSN)\n\n case CTypeError:\n glog.Errorf(\"wal recovery encountered an unrecoverable error\")\n return fmt.Errorf(\"wal error\")\n\n default:\n glog.Errorf(\"wal recovery received an unknown or invalid record\")\n return fmt.Errorf(\"wal error\")\n }\n }\n\n return nil\n}", "func (s *Store) Restore(rc io.ReadCloser) error {\n\t// TODO\n\treturn nil\n}", "func (tm *topicManager) createTopic(name string) error {\n\ttm.topicsMutex.RLock()\n\t_, ok := tm.topicNamesToIds[name]\n\ttm.topicsMutex.RUnlock()\n\tif ok {\n\t\treturn fmt.Errorf(\"Topic with name [%s] already exists\", name)\n\t}\n\n\t// this allocates a copycat data structure but doesn't create a log structure yet\n\tid, err := tm.cc.AllocateNewDataStructure()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tprotobuf := &pb.TopicMetadataOperation{\n\t\tName: name,\n\t\tDataStructureId: id.String(),\n\t}\n\n\tbites, err := protobuf.Marshal()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttm.proposeCh <- bites\n\treturn nil\n}", "func CreateTopic(c *client.Client) fw.Topic {\n\treturn &configzTopic{\n\t\tcl: c,\n\t}\n}", "func (r *TopicReconciler) Reconcile(req ctrl.Request) (ctrl.Result, error) {\n\tctx := context.Background()\n\t_ = r.Log.WithValues(\"topic\", req.NamespacedName)\n\n\ttopic := &nlptv1.Topic{}\n\tif err := r.Get(ctx, req.NamespacedName, topic); err != nil {\n\t\tklog.Errorf(\"cannot get topic of ctrl req: %+v\", err)\n\t\treturn ctrl.Result{}, nil\n\t}\n\t//klog.Infof(\"get new topic event: %+v\", *topic)\n\t//klog.Infof(\"Status:%s\", topic.Status.Status)\n\n\tif topic.Status.Status == nlptv1.Creating {\n\t\t//klog.Info(\"Current status is Init\")\n\t\tif err := r.Operator.CreateTopic(topic); err != nil {\n\t\t\ttopic.Spec.Url = topic.GetUrl()\n\t\t\ttopic.Status.Status = nlptv1.CreateFailed\n\t\t\ttopic.Status.Message = fmt.Sprintf(\"create topic error:%+v\", err)\n\t\t\tklog.Errorf(\"create topic failed, err: %+v\", err)\n\t\t} else {\n\t\t\ttopic.Spec.Url = topic.GetUrl()\n\t\t\ttopic.Status.Status = nlptv1.Created\n\t\t\ttopic.Status.Message = \"success\"\n\t\t}\n\n\t\t//更新数据库的状态\n\t\t//klog.Infof(\"Final Topic: %+v\", *topic)\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v, err: %+v\", *topic, err)\n\t\t}\n\n\t}\n\n\tif topic.Status.Status == nlptv1.Importing {\n\t\tif ok, _ := r.Operator.isNamespacesExist(topic); ok {\n\t\t\tif err := r.Operator.CreateTopic(topic); err != nil {\n\t\t\t\ttopic.Spec.Url = topic.GetUrl()\n\t\t\t\ttopic.Status.Status = nlptv1.ImportFailed\n\t\t\t\ttopic.Status.Message = fmt.Sprintf(\"create topic error:%+v\", err)\n\t\t\t\tklog.Errorf(\"create topic failed, err: %+v\", err)\n\t\t\t} else {\n\t\t\t\ttopic.Spec.Url = topic.GetUrl()\n\t\t\t\ttopic.Status.Status = nlptv1.ImportSuccess\n\t\t\t\ttopic.Status.Message = \"success\"\n\t\t\t}\n\n\t\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\tif topic.Status.Status == nlptv1.Deleting {\n\t\tif err := r.Operator.DeleteTopic(topic, false); err != nil {\n\t\t\ttopic.Status.Status = nlptv1.DeleteFailed\n\t\t\ttopic.Status.Message = fmt.Sprintf(\"delete topic error: %+v\", err)\n\t\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t\t}\n\t\t} else {\n\t\t\t//删除数据\n\t\t\tif err = r.Delete(ctx, topic); err != nil {\n\t\t\t\tklog.Errorf(\"delete Topic Failed: %+v\", *topic)\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif topic.Status.Status == nlptv1.ForceDeleting {\n\t\tif err := r.Operator.DeleteTopic(topic, true); err != nil {\n\t\t\ttopic.Status.Status = nlptv1.ForceDeleteFailed\n\t\t\ttopic.Status.Message = fmt.Sprintf(\"force delete topic error: %+v\", err)\n\t\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t\t}\n\t\t} else {\n\t\t\t//删除数据\n\t\t\tif err = r.Delete(ctx, topic); err != nil {\n\t\t\t\tklog.Errorf(\"delete Topic Failed: %+v\", *topic)\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif topic.Status.Status == nlptv1.Terminating {\n\t\tif err := r.Operator.TerminateTopic(topic); err != nil {\n\t\t\ttopic.Status.Status = nlptv1.TerminatedFailed\n\t\t\ttopic.Status.Message = fmt.Sprintf(\"terminate topic error: %+v\", err)\n\n\t\t} else {\n\t\t\ttopic.Status.Status = nlptv1.Terminated\n\t\t\ttopic.Status.Message = fmt.Sprintf(\"terminate topic success\")\n\t\t}\n\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t}\n\n\t}\n\n\tif topic.Status.Status == nlptv1.Updating {\n\t\t//增加topic分区\n\t\tif err := r.Operator.AddPartitionsOfTopic(topic); err != nil {\n\t\t\ttopic.Status.Status = nlptv1.UpdateFailed\n\t\t\ttopic.Status.Message = fmt.Sprintf(\"add topic partition error: %+v \", err)\n\t\t\ttopic.Spec.PartitionNum = topic.Spec.OldPartitionNum\n\t\t} else {\n\t\t\ttopic.Status.Status = nlptv1.Updated\n\t\t\ttopic.Status.Message = \"success\"\n\t\t}\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t}\n\t}\n\n\t//删除授权\n\tif topic.Status.AuthorizationStatus == nlptv1.DeletingAuthorization {\n\t\tfor i := 0; i < len(topic.Spec.Permissions); i++ {\n\t\t\tp := topic.Spec.Permissions[i]\n\t\t\tif p.Status.Status == nlptv1.DeletingAuthorization {\n\t\t\t\tif err := r.Operator.DeletePer(topic, &p); err != nil {\n\t\t\t\t\tp.Status.Status = nlptv1.DeleteAuthorizationFailed\n\t\t\t\t\tp.Status.Message = fmt.Sprintf(\"revoke permission error: %+v\", err)\n\t\t\t\t\t//删除失败,将标签重置为true\n\t\t\t\t\ttopic.ObjectMeta.Labels[p.AuthUserID] = \"true\"\n\t\t\t\t\ttopic.Status.AuthorizationStatus = nlptv1.DeleteAuthorizationFailed\n\t\t\t\t} else {\n\t\t\t\t\tpers := topic.Spec.Permissions\n\t\t\t\t\ttopic.Spec.Permissions = append(pers[:i], pers[i+1:]...)\n\t\t\t\t\t//收回权限成功,删除标签\n\t\t\t\t\tdelete(topic.ObjectMeta.Labels, p.AuthUserID)\n\t\t\t\t\ttopic.Status.AuthorizationStatus = nlptv1.DeletedAuthorization\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t}\n\t}\n\n\tif topic.Status.AuthorizationStatus == nlptv1.UpdatingAuthorization {\n\t\t//klog.Infof(\"Start Grant Topic: %+v\", *topic)\n\t\t//授权操作\n\t\tfor i := 0; i < len(topic.Spec.Permissions); i++ {\n\t\t\tp := topic.Spec.Permissions[i]\n\t\t\tif p.Status.Status == nlptv1.UpdatingAuthorization {\n\t\t\t\tif err := r.Operator.GrantPermission(topic, &p); err != nil {\n\t\t\t\t\tp.Status.Status = nlptv1.UpdatingAuthorizationFailed\n\t\t\t\t\tp.Status.Message = fmt.Sprintf(\"modify permission error: %+v\", err)\n\t\t\t\t\ttopic.Status.AuthorizationStatus = nlptv1.UpdatingAuthorizationFailed\n\n\t\t\t\t\t//TODO roll back\n\n\t\t\t\t} else {\n\t\t\t\t\tp.Status.Status = nlptv1.UpdatingAuthorizationSuccess\n\t\t\t\t\tp.Status.Message = \"success\"\n\t\t\t\t\ttopic.Status.AuthorizationStatus = nlptv1.UpdatingAuthorizationSuccess\n\t\t\t\t}\n\t\t\t\ttopic.Spec.Permissions[i] = p\n\t\t\t}\n\t\t}\n\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t}\n\t}\n\tif topic.Status.AuthorizationStatus == nlptv1.Authorizing {\n\t\tklog.Infof(\"Start Grant Topic: %+v\", *topic)\n\t\t//授权操作\n\t\tfor i := 0; i < len(topic.Spec.Permissions); i++ {\n\t\t\tp := topic.Spec.Permissions[i]\n\t\t\tif p.Status.Status == nlptv1.Authorizing {\n\t\t\t\tif err := r.Operator.GrantPermission(topic, &p); err != nil {\n\t\t\t\t\tp.Status.Status = nlptv1.AuthorizeFailed\n\t\t\t\t\tp.Status.Message = fmt.Sprintf(\"grant permission error: %+v\", err)\n\t\t\t\t\ttopic.Status.AuthorizationStatus = nlptv1.AuthorizeFailed\n\t\t\t\t} else {\n\t\t\t\t\tp.Status.Status = nlptv1.Authorized\n\t\t\t\t\tp.Status.Message = \"success\"\n\t\t\t\t\ttopic.Status.AuthorizationStatus = nlptv1.Authorized\n\t\t\t\t}\n\t\t\t\ttopic.Spec.Permissions[i] = p\n\t\t\t}\n\t\t}\n\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t}\n\t}\n\n\tif topic.Status.BindStatus == nlptv1.BindingOrUnBinding {\n\t\tfor appid, application := range topic.Spec.Applications {\n\t\t\tswitch application.Status {\n\t\t\tcase nlptv1.UpdatingAuthorization:\n\t\t\t\tp := nlptv1.Permission{\n\t\t\t\t\tAuthUserID: \"\",\n\t\t\t\t\tAuthUserName: application.ID,\n\t\t\t\t\tActions: application.Actions,\n\t\t\t\t}\n\t\t\t\tif err := r.Operator.GrantPermission(topic, &p); err != nil {\n\t\t\t\t\tapplication.Status = nlptv1.UpdatingAuthorizationFailed\n\t\t\t\t\tapplication.DisplayStatus = nlptv1.ShowStatusMap[application.Status]\n\t\t\t\t\tapplication.Message = fmt.Sprintf(\"update authorization: %+v\", err)\n\t\t\t\t} else {\n\t\t\t\t\tapplication.Status = nlptv1.UpdatingAuthorizationSuccess\n\t\t\t\t\tapplication.DisplayStatus = nlptv1.ShowStatusMap[application.Status]\n\t\t\t\t\tapplication.Message = \"update authorization successfully\"\n\t\t\t\t}\n\t\t\tcase nlptv1.Binding:\n\t\t\t\t//actions := make([]string, 0)\n\t\t\t\t//actions = append(actions, nlptv1.Consume)\n\t\t\t\t//actions = append(actions, nlptv1.Produce)\n\n\t\t\t\tp := nlptv1.Permission{\n\t\t\t\t\tAuthUserID: \"\",\n\t\t\t\t\tAuthUserName: application.ID,\n\t\t\t\t\tActions: application.Actions,\n\t\t\t\t}\n\t\t\t\tif err := r.Operator.GrantPermission(topic, &p); err != nil {\n\t\t\t\t\tapplication.Status = nlptv1.BindFailed\n\t\t\t\t\tapplication.DisplayStatus = nlptv1.ShowStatusMap[application.Status]\n\t\t\t\t\tapplication.Message = fmt.Sprintf(\"bind error: %+v\", err)\n\t\t\t\t} else {\n\t\t\t\t\tapplication.Status = nlptv1.Bound\n\t\t\t\t\tapplication.DisplayStatus = nlptv1.ShowStatusMap[application.Status]\n\t\t\t\t\tapplication.Message = \"bind success\"\n\t\t\t\t}\n\t\t\tcase nlptv1.Unbinding:\n\t\t\t\tp := nlptv1.Permission{\n\t\t\t\t\tAuthUserID: \"\",\n\t\t\t\t\tAuthUserName: application.ID,\n\t\t\t\t}\n\n\t\t\t\tif err := r.Operator.DeletePer(topic, &p); err != nil {\n\t\t\t\t\tapplication.Status = nlptv1.UnbindFailed\n\t\t\t\t\tapplication.DisplayStatus = nlptv1.ShowStatusMap[application.Status]\n\t\t\t\t\tapplication.Message = fmt.Sprintf(\"release error: %+v\", err)\n\n\t\t\t\t} else {\n\t\t\t\t\tapplication.Status = nlptv1.UnbindSuccess\n\t\t\t\t}\n\n\t\t\t}\n\t\t\ttopic.Spec.Applications[appid] = application\n\t\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t\t}\n\t\t}\n\n\t\t//处理解绑定的场景\n\t\tapps := make(map[string]nlptv1.Application)\n\n\t\tfor appid, application := range topic.Spec.Applications {\n\t\t\tif application.Status != nlptv1.UnbindSuccess {\n\t\t\t\tapps[appid] = application\n\t\t\t}\n\t\t}\n\n\t\ttopic.Spec.Applications = apps\n\t\tif err := r.Update(ctx, topic); err != nil {\n\t\t\tklog.Errorf(\"Update Topic Failed: %+v\", *topic)\n\t\t}\n\n\t}\n\n\t//klog.Infof(\"Final Topic: %+v\", *topic)\n\treturn ctrl.Result{}, nil\n}", "func (f *fsm) Restore(rc io.ReadCloser) error {\n\t//o := NewInMemDatastore()\n\t//if err := json.NewDecoder(rc).Decode(&o.m); err != nil {\n\t//\treturn err\n\t//}\n\n\t// Set the state from the snapshot, no lock required according to\n\t// Hashicorp docs.\n\t//f.m = o\n\n\treturn nil\n}", "func (api *TopicalAPI) TopicNew(w http.ResponseWriter, r *http.Request) {\n\tflashes, _ := api.session.GetFlashes(r, w)\n\tuser, err := api.session.GetUser(r)\n\n\tif err != nil {\n\t\tapi.session.SaveFlash(\"Log in to post a message\", r, w)\n\t\thttp.Redirect(w, r, \"/topics\", 302)\n\t\treturn\n\t}\n\n\tpayload := struct {\n\t\tUser *models.User\n\t\tFlashes []string\n\t}{user, flashes}\n\n\tapi.templates.ExecuteTemplate(w, \"new-topic\", payload)\n}", "func TopicCreate(w http.ResponseWriter, r *http.Request) {\n\n\t// Init output\n\toutput := []byte(\"\")\n\n\t// Add content type header to the response\n\tcontentType := \"application/json\"\n\tcharset := \"utf-8\"\n\tw.Header().Add(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\t// Grab url path variables\n\turlVars := mux.Vars(r)\n\n\t// Grab context references\n\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\tprojectUUID := gorillaContext.Get(r, \"auth_project_uuid\").(string)\n\n\tpostBody := map[string]string{}\n\tschemaUUID := \"\"\n\n\t// check if there's a request body provided before trying to decode\n\tif r.Body != nil {\n\n\t\tb, err := ioutil.ReadAll(r.Body)\n\n\t\tif err != nil {\n\t\t\terr := APIErrorInvalidRequestBody()\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t\tdefer r.Body.Close()\n\n\t\tif len(b) > 0 {\n\t\t\terr = json.Unmarshal(b, &postBody)\n\t\t\tif err != nil {\n\t\t\t\terr := APIErrorInvalidRequestBody()\n\t\t\t\trespondErr(w, err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tschemaRef := postBody[\"schema\"]\n\n\t\t\t// if there was a schema name provided, check its existence\n\t\t\tif schemaRef != \"\" {\n\t\t\t\t_, schemaName, err := schemas.ExtractSchema(schemaRef)\n\t\t\t\tif err != nil {\n\t\t\t\t\terr := APIErrorInvalidData(err.Error())\n\t\t\t\t\trespondErr(w, err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tsl, err := schemas.Find(projectUUID, \"\", schemaName, refStr)\n\t\t\t\tif err != nil {\n\t\t\t\t\terr := APIErrGenericInternal(err.Error())\n\t\t\t\t\trespondErr(w, err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tif sl.Empty() {\n\t\t\t\t\terr := APIErrorNotFound(\"Schema\")\n\t\t\t\t\trespondErr(w, err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tschemaUUID = sl.Schemas[0].UUID\n\t\t\t}\n\t\t}\n\t}\n\t// Get Result Object\n\tres, err := topics.CreateTopic(projectUUID, urlVars[\"topic\"], schemaUUID, refStr)\n\tif err != nil {\n\t\tif err.Error() == \"exists\" {\n\t\t\terr := APIErrorConflict(\"Topic\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t\terr := APIErrGenericInternal(err.Error())\n\t\trespondErr(w, err)\n\t}\n\n\t// Output result to JSON\n\tresJSON, err := res.ExportJSON()\n\tif err != nil {\n\t\terr := APIErrExportJSON()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Write response\n\toutput = []byte(resJSON)\n\trespondOK(w, output)\n\n}", "func (t *WALTopic) Close() error {\n\treturn t.wal.Close()\n}", "func (gWal *GenericWAL) updateRecoveryMap(\n checkpoint bool, lsn int64, data []byte) error {\n\n record := &GenericWALRecord{}\n if err := json.Unmarshal(data, record); err != nil {\n glog.Errorf(\"could not parse wal record during recovery :: %v\", err)\n return err\n }\n // We need to explicitly set the lsn field, as \"lsn\" is not present within\n // the \"data\" that is serialized on to the wal. The lsn is\n // generated \"after\" we finish writing \"data\" into the wal.\n record.LSN = lsn\n\n // Recovery happens as follows:\n //\n // Checkpoint records are applied immediately because they always store a\n // consistent snapshot with any live operations.\n //\n // Delta records are stored in a map so that when we see a corresponding\n // revert record, we can ignore it.\n //\n // EXTRA NOTES\n //\n // Ideally, we want to replay all delta records. But sometimes an operation\n // could have failed and returned an error. We don't want to replay such\n // operations after a crash, so we use Revert records to remove such items\n // out of the recovery record map.\n //\n // Similarly, a long running operation wants to record its progress so that\n // it can resume from the middle during recovery. They use Progress records\n // to save their progress and can skip over already completed operations\n // during recovery.\n //\n // Progress records are sometimes necessary for correctness because\n // filesystem state cannot be reconstructed through replay. For example,\n // consider this sequence of operations: etcd.Create, etcd.Remove and\n // etcd.Create. If current state of the system is at the second etcd.Create\n // and a crash here would replay etcd.Remove operation which removes the\n // files owned by second etcd.Create operation. So, progress records will\n // help avoid such cases.\n\n if record.IsRevert {\n delete(gWal.recoveryMap, record.Revert.LSN)\n return nil\n }\n\n if record.IsProgress {\n if _, ok := gWal.recoveryMap[record.Progress.LSN]; ok {\n gWal.recoveryMap[record.Progress.LSN] = record.Progress.NewRecord\n } else {\n glog.Warningf(\"progress record %d is ignored because its initial \"+\n \"record %d is not found in the recovery map\", lsn, record.Progress.LSN)\n }\n return nil\n }\n\n // This includes both delta and checkpoint records.\n gWal.recoveryMap[lsn] = record\n return nil\n}", "func (r *Reconciler) createTopic(ctx context.Context, topicName string, partitions int, replicationFactor int, retentionMillis int64) error {\n\n\t// Setup The Logger\n\tlogger := r.logger.With(zap.String(\"Topic\", topicName))\n\n\t// Create The TopicSpecification\n\ttopicSpecifications := []kafka.TopicSpecification{\n\t\t{\n\t\t\tTopic: topicName,\n\t\t\tNumPartitions: partitions,\n\t\t\tReplicationFactor: replicationFactor,\n\t\t\tConfig: map[string]string{\n\t\t\t\tconstants.KafkaTopicConfigRetentionMs: strconv.FormatInt(retentionMillis, 10),\n\t\t\t},\n\t\t},\n\t}\n\n\t// Attempt To Create The Topic & Process Results\n\ttopicResults, err := r.adminClient.CreateTopics(ctx, topicSpecifications)\n\tif len(topicResults) > 0 {\n\t\ttopicResultError := topicResults[0].Error\n\t\ttopicResultErrorCode := topicResultError.Code()\n\t\tif topicResultErrorCode == kafka.ErrTopicAlreadyExists {\n\t\t\tlogger.Info(\"Kafka Topic Already Exists - No Creation Required\")\n\t\t\treturn nil\n\t\t} else if topicResultErrorCode == kafka.ErrNoError {\n\t\t\tlogger.Info(\"Successfully Created New Kafka Topic\")\n\t\t\treturn nil\n\t\t} else {\n\t\t\tlogger.Error(\"Failed To Create Topic (Results)\", zap.Error(err), zap.Any(\"TopicResults\", topicResults))\n\t\t\treturn topicResults[0].Error\n\t\t}\n\t} else if err != nil {\n\t\tlogger.Error(\"Failed To Create Topic (Error)\", zap.Error(err))\n\t\treturn err\n\t} else {\n\t\tlogger.Warn(\"Received Empty TopicResults From CreateTopics Request\")\n\t\treturn nil\n\t}\n}", "func (r *Reconciler) createRestoreConfig(ctx context.Context, postgresCluster *v1beta1.PostgresCluster,\n\tconfigHash string) error {\n\n\tpostgresClusterWithMockedBackups := postgresCluster.DeepCopy()\n\tpostgresClusterWithMockedBackups.Spec.Backups.PGBackRest.Global = postgresCluster.Spec.\n\t\tDataSource.PGBackRest.Global\n\tpostgresClusterWithMockedBackups.Spec.Backups.PGBackRest.Repos = []v1beta1.PGBackRestRepo{\n\t\tpostgresCluster.Spec.DataSource.PGBackRest.Repo,\n\t}\n\n\treturn r.reconcilePGBackRestConfig(ctx, postgresClusterWithMockedBackups,\n\t\t\"\", configHash, \"\", \"\", []string{})\n}", "func (w *WorkerFSM) Restore(i io.ReadCloser) error {\n return nil\n}", "func (c *client) createTopic(topicName string) (*pubsub.Topic, error) {\n\tvar topic *pubsub.Topic\n\tctx := context.Background()\n\n\texists, err := c.topicExists(topicName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif !exists {\n\t\ttopic, err = c.client.CreateTopic(ctx, topicName)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\ttopic = c.client.Topic(topicName)\n\t}\n\n\treturn topic, nil\n}", "func (f *raftStore) Restore(rc io.ReadCloser) error {\n\tvar o state\n\tif err := json.NewDecoder(rc).Decode(&o); err != nil {\n\t\treturn err\n\t}\n\tf.mu.Lock()\n\tf.inFlight = nil\n\tf.mu.Unlock()\n\n\tf.m.setNewState(o)\n\treturn nil\n}", "func (zj *ZapJournal) Topic(topic string) Writer {\n\treturn &ZapWriter{\n\t\tlogger: zj.logger.Sugar().Named(topic),\n\t\ttopic: topic,\n\t}\n}", "func topic(prefix string, obj metav1.Object) string {\n\treturn fmt.Sprintf(\"%s.%s.%s\", prefix, obj.GetNamespace(), obj.GetName())\n}", "func (d *distEventBus) setupTopicSub(topicName, subscriberName string) error {\n\tctx := context.TODO()\n\n\t// Create the topic if it doesn't exist yet.\n\td.topic = d.client.Topic(topicName)\n\tif exists, err := d.topic.Exists(ctx); err != nil {\n\t\treturn skerr.Fmt(\"Error checking whether topic exits: %s\", err)\n\t} else if !exists {\n\t\tif d.topic, err = d.client.CreateTopic(ctx, topicName); err != nil {\n\t\t\treturn skerr.Fmt(\"Error creating pubsub topic '%s': %s\", topicName, err)\n\t\t}\n\t}\n\n\t// Create the subscription if it doesn't exist.\n\tsubName := fmt.Sprintf(\"%s+%s\", subscriberName, topicName)\n\td.sub = d.client.Subscription(subName)\n\tif exists, err := d.sub.Exists(ctx); err != nil {\n\t\treturn skerr.Fmt(\"Error checking existence of pubsub subscription '%s': %s\", subName, err)\n\t} else if !exists {\n\t\td.sub, err = d.client.CreateSubscription(ctx, subName, pubsub.SubscriptionConfig{\n\t\t\tTopic: d.topic,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn skerr.Fmt(\"Error creating pubsub subscription '%s': %s\", subName, err)\n\t\t}\n\t}\n\td.sub.ReceiveSettings.MaxOutstandingMessages = MaximumConcurrentPublishesPerTopic\n\t// Make the subscription also the id of this client.\n\td.clientID = subName\n\treturn nil\n}", "func TestLessorRecover(t *testing.T) {\n\tlg := zap.NewNop()\n\tdir, be := NewTestBackend(t)\n\tdefer os.RemoveAll(dir)\n\tdefer be.Close()\n\n\tle := newLessor(lg, be, clusterLatest(), LessorConfig{MinLeaseTTL: minLeaseTTL})\n\tdefer le.Stop()\n\tl1, err1 := le.Grant(1, 10)\n\tl2, err2 := le.Grant(2, 20)\n\tif err1 != nil || err2 != nil {\n\t\tt.Fatalf(\"could not grant initial leases (%v, %v)\", err1, err2)\n\t}\n\n\t// Create a new lessor with the same backend\n\tnle := newLessor(lg, be, clusterLatest(), LessorConfig{MinLeaseTTL: minLeaseTTL})\n\tdefer nle.Stop()\n\tnl1 := nle.Lookup(l1.ID)\n\tif nl1 == nil || nl1.ttl != l1.ttl {\n\t\tt.Errorf(\"nl1 = %v, want nl1.ttl= %d\", nl1.ttl, l1.ttl)\n\t}\n\n\tnl2 := nle.Lookup(l2.ID)\n\tif nl2 == nil || nl2.ttl != l2.ttl {\n\t\tt.Errorf(\"nl2 = %v, want nl2.ttl= %d\", nl2.ttl, l2.ttl)\n\t}\n}", "func (kz *Kazoo) Topic(topic string) *Topic {\n\treturn &Topic{Name: topic, kz: kz}\n}", "func (r *Reconciler) reconcileTopic(ctx context.Context, channel *kafkav1alpha1.KafkaChannel) error {\n\n\t// Get The TopicName For Specified Channel\n\ttopicName := util.TopicName(channel)\n\n\t// Get Channel Specific Logger & Add Topic Name\n\tlogger := util.ChannelLogger(r.logger, channel).With(zap.String(\"TopicName\", topicName))\n\n\t// Get The Topic Configuration (First From Channel With Failover To Environment)\n\tnumPartitions := util.NumPartitions(channel, r.environment, r.logger)\n\treplicationFactor := util.ReplicationFactor(channel, r.environment, r.logger)\n\tretentionMillis := util.RetentionMillis(channel, r.environment, r.logger)\n\n\t// Create The Topic (Handles Case Where Already Exists)\n\terr := r.createTopic(ctx, topicName, numPartitions, replicationFactor, retentionMillis)\n\n\t// Log Results & Return Status\n\tif err != nil {\n\t\tcontroller.GetEventRecorder(ctx).Eventf(channel, corev1.EventTypeWarning, event.KafkaTopicReconciliationFailed.String(), \"Failed To Reconcile Kafka Topic For Channel: %v\", err)\n\t\tlogger.Error(\"Failed To Reconcile Topic\", zap.Error(err))\n\t\tchannel.Status.MarkTopicFailed(\"TopicFailed\", fmt.Sprintf(\"Channel Kafka Topic Failed: %s\", err))\n\t} else {\n\t\tlogger.Info(\"Successfully Reconciled Topic\")\n\t\tchannel.Status.MarkTopicTrue()\n\t}\n\treturn err\n}", "func (gWal *GenericWAL) LogCheckpoint(payload interface{},\n version int64) error {\n\n if gWal.lostOwnership {\n return ErrLostOwnership\n }\n\n walRecord := &GenericWALRecord{}\n walRecord.Version = version\n // serialize the payload.\n userData, errMarshal := json.Marshal(payload)\n if errMarshal != nil {\n glog.Errorf(\"could not serialize user payload for WAL :: %v\",\n errMarshal)\n return errMarshal\n }\n walRecord.Payload = userData\n data, errMarshal := json.Marshal(walRecord)\n if errMarshal != nil {\n glog.Errorf(\"could not serialize wal record :: %v\", errMarshal)\n return errMarshal\n }\n\n if err := gWal.wal.StartCheckpoint(); err != nil {\n glog.Errorf(\"could not start new checkpoint :: %v\", err)\n return err\n }\n // TODO: WAL must have the ability to abort a checkpoint on errors,\n // otherwise, there is no way to cancel StartCheckpoint operation, so all\n // future checkpoint attempts will also fail.\n lsn, errWrite := gWal.writeCheckpointRecord(data)\n if errWrite != nil {\n glog.Errorf(\"could not write checkpoint record :: %v\", errWrite)\n gWal.lostOwnership = IsLostOwnership(errWrite)\n return errWrite\n }\n if err := gWal.wal.FinalizeCheckpoint(); err != nil {\n glog.Errorf(\"could not finalize checkpoint :: %v\", err)\n return err\n }\n glog.Infof(\"new checkpoint is recorded with lsn %d at version %d\", lsn,\n version)\n walRecord.LSN = lsn\n return nil\n}", "func (f *FSM) restoreDatabase(tracer *trace.Tracer, reader io.ReadCloser) (bool, error) {\n\tdone := false\n\n\t// The first 8 bytes contain the size of database.\n\tvar dataSize uint64\n\tif err := binary.Read(reader, binary.LittleEndian, &dataSize); err != nil {\n\t\treturn false, errors.Wrap(err, \"failed to read database size\")\n\t}\n\ttracer.Message(\"database size: %d\", dataSize)\n\n\t// Then there's the database data.\n\tdata := make([]byte, dataSize)\n\tif _, err := io.ReadFull(reader, data); err != nil {\n\t\treturn false, errors.Wrap(err, \"failed to read database data\")\n\t}\n\n\t// Next, the size of the WAL.\n\tvar walSize uint64\n\tif err := binary.Read(reader, binary.LittleEndian, &walSize); err != nil {\n\t\treturn false, errors.Wrap(err, \"failed to read wal size\")\n\t}\n\ttracer.Message(\"wal size: %d\", walSize)\n\n\t// Read the WAL data.\n\twal := make([]byte, walSize)\n\tif _, err := io.ReadFull(reader, wal); err != nil {\n\t\treturn false, errors.Wrap(err, \"failed to read wal data\")\n\t}\n\n\t// Read the database path.\n\tbufReader := bufio.NewReader(reader)\n\tfilename, err := bufReader.ReadString(0)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"failed to read database name\")\n\t}\n\tfilename = filename[:len(filename)-1] // Strip the trailing 0\n\ttracer.Message(\"filename: %s\", filename)\n\n\t// XXX TODO: reason about this situation, is it harmful?\n\t// Check that there are no leader connections for this database.\n\t//\n\t// FIXME: we should relax this, as it prevents restoring snapshots \"on\n\t// the fly\".\n\t// conns := f.registry.ConnLeaders(filename)\n\t// if len(conns) > 0 {\n\t// \ttracer.Panic(\"found %d leader connections\", len(conns))\n\t// }\n\n\t// XXX TODO: reason about this situation, is it possible?\n\t//txn := f.transactions.GetByConn(f.connections.Follower(name))\n\t//if txn != nil {\n\t//\tf.logger.Printf(\"[WARN] dqlite: fsm: closing follower in-flight transaction %s\", txn)\n\t//\tf.transactions.Remove(txn.ID())\n\t//}\n\n\t// Close any follower connection, since we're going to overwrite the\n\t// database file.\n\tif f.registry.ConnFollowerExists(filename) {\n\t\ttracer.Message(\"close follower: %s\", filename)\n\t\tfollower := f.registry.ConnFollower(filename)\n\t\tf.registry.ConnFollowerDel(filename)\n\t\tif err := follower.Close(); err != nil {\n\t\t\treturn false, err\n\t\t}\n\t}\n\n\t// At this point there should be not connection open against this\n\t// database, so it's safe to overwrite it.\n\ttxid, err := bufReader.ReadString(0)\n\tif err != nil {\n\t\tif err != io.EOF {\n\t\t\treturn false, errors.Wrap(err, \"failed to read txid\")\n\t\t}\n\t\tdone = true // This is the last database.\n\t}\n\ttracer.Message(\"transaction ID: %s\", txid)\n\n\tvfs := f.registry.Vfs()\n\n\tif err := connection.Restore(vfs, filename, data, wal); err != nil {\n\t\treturn false, err\n\t}\n\n\ttracer.Message(\"open follower: %s\", filename)\n\tif err := f.openFollower(filename); err != nil {\n\t\treturn false, err\n\t}\n\n\tif txid != \"\" {\n\t\t// txid, err := strconv.ParseUint(txid, 10, 64)\n\t\t// if err != nil {\n\t\t// \treturn false, err\n\t\t// }\n\t\t// tracer.Message(\"add transaction: %d\", txid)\n\t\t// conn := f.registry.ConnFollower(filename)\n\t\t// txn := f.registry.TxnFollowerAdd(conn, txid)\n\t\t// if err := txn.Begin(); err != nil {\n\t\t// \treturn false, err\n\t\t// }\n\t}\n\n\treturn done, nil\n}", "func (gWal *GenericWAL) LogRevertRecord(\n record *GenericWALRecord) error {\n\n if gWal.lostOwnership {\n return ErrLostOwnership\n }\n\n revert := &GenericWALRecord{Payload: []byte(`\"\"`)}\n revert.IsRevert = true\n revert.Revert.LSN = record.LSN\n data, errMarshal := json.Marshal(revert)\n if errMarshal != nil {\n glog.Errorf(\"could not serialize undo/rollback record :: %v\", errMarshal)\n return errMarshal\n }\n if _, err := gWal.writeDataRecord(data); err != nil {\n glog.Errorf(\"could not write revert record to wal :: %v\", err)\n gWal.lostOwnership = IsLostOwnership(err)\n return err\n }\n return nil\n}", "func UnmarshalTopic(m map[string]json.RawMessage, result interface{}) (err error) {\n\tobj := new(Topic)\n\terr = core.UnmarshalPrimitive(m, \"id\", &obj.ID)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"description\", &obj.Description)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"name\", &obj.Name)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"updated_at\", &obj.UpdatedAt)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"source_count\", &obj.SourceCount)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalModel(m, \"sources\", &obj.Sources, UnmarshalSourcesListItems)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"subscription_count\", &obj.SubscriptionCount)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalModel(m, \"subscriptions\", &obj.Subscriptions, UnmarshalSubscriptionListItem)\n\tif err != nil {\n\t\treturn\n\t}\n\treflect.ValueOf(result).Elem().Set(reflect.ValueOf(obj))\n\treturn\n}", "func SubCreate(w http.ResponseWriter, r *http.Request) {\n\n\t// Init output\n\toutput := []byte(\"\")\n\n\t// Add content type header to the response\n\tcontentType := \"application/json\"\n\tcharset := \"utf-8\"\n\tw.Header().Add(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\t// Grab url path variables\n\turlVars := mux.Vars(r)\n\n\t// Grab context references\n\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\trefBrk := gorillaContext.Get(r, \"brk\").(brokers.Broker)\n\tprojectUUID := gorillaContext.Get(r, \"auth_project_uuid\").(string)\n\n\t// Read POST JSON body\n\tbody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\terr := APIErrorInvalidRequestBody()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Parse pull options\n\tpostBody, err := subscriptions.GetFromJSON(body)\n\tif err != nil {\n\t\terr := APIErrorInvalidArgument(\"Subscription\")\n\t\trespondErr(w, err)\n\t\tlog.Error(string(body[:]))\n\t\treturn\n\t}\n\n\ttProject, tName, err := subscriptions.ExtractFullTopicRef(postBody.FullTopic)\n\n\tif err != nil {\n\t\terr := APIErrorInvalidName(\"Topic\")\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\tif topics.HasTopic(projectUUID, tName, refStr) == false {\n\t\terr := APIErrorNotFound(\"Topic\")\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Get current topic offset\n\ttProjectUUID := projects.GetUUIDByName(tProject, refStr)\n\tfullTopic := tProjectUUID + \".\" + tName\n\tcurOff := refBrk.GetMaxOffset(fullTopic)\n\n\tpushEnd := \"\"\n\trPolicy := \"\"\n\trPeriod := 0\n\tmaxMessages := int64(1)\n\n\t//pushWorker := auth.User{}\n\tverifyHash := \"\"\n\n\tif postBody.PushCfg != (subscriptions.PushConfig{}) {\n\n\t\t// check the state of the push functionality\n\t\tpwToken := gorillaContext.Get(r, \"push_worker_token\").(string)\n\t\tpushEnabled := gorillaContext.Get(r, \"push_enabled\").(bool)\n\n\t\tif !pushEnabled {\n\t\t\terr := APIErrorPushConflict()\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\t_, err = auth.GetPushWorker(pwToken, refStr)\n\t\tif err != nil {\n\t\t\terr := APIErrInternalPush()\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\tpushEnd = postBody.PushCfg.Pend\n\t\t// Check if push endpoint is not a valid https:// endpoint\n\t\tif !(isValidHTTPS(pushEnd)) {\n\t\t\terr := APIErrorInvalidData(\"Push endpoint should be addressed by a valid https url\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t\trPolicy = postBody.PushCfg.RetPol.PolicyType\n\t\trPeriod = postBody.PushCfg.RetPol.Period\n\t\tmaxMessages = postBody.PushCfg.MaxMessages\n\n\t\tif rPolicy == \"\" {\n\t\t\trPolicy = subscriptions.LinearRetryPolicyType\n\t\t}\n\n\t\tif maxMessages == 0 {\n\t\t\tmaxMessages = int64(1)\n\t\t}\n\n\t\tif rPeriod <= 0 {\n\t\t\trPeriod = 3000\n\t\t}\n\n\t\tif !subscriptions.IsRetryPolicySupported(rPolicy) {\n\t\t\terr := APIErrorInvalidData(subscriptions.UnSupportedRetryPolicyError)\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\tverifyHash, err = auth.GenToken()\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Could not generate verification hash for subscription %v, %v\", urlVars[\"subscription\"], err.Error())\n\t\t\terr := APIErrGenericInternal(\"Could not generate verification hash\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t}\n\n\t// Get Result Object\n\tres, err := subscriptions.CreateSub(projectUUID, urlVars[\"subscription\"], tName, pushEnd, curOff, maxMessages, postBody.Ack, rPolicy, rPeriod, verifyHash, false, refStr)\n\n\tif err != nil {\n\t\tif err.Error() == \"exists\" {\n\t\t\terr := APIErrorConflict(\"Subscription\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t\terr := APIErrGenericInternal(err.Error())\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Output result to JSON\n\tresJSON, err := res.ExportJSON()\n\tif err != nil {\n\t\terr := APIErrExportJSON()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Write response\n\toutput = []byte(resJSON)\n\trespondOK(w, output)\n\n}", "func (s *SchematicServer) generateSchemaTopic(schemaTopic string) error {\n\tconfig := sarama.NewConfig()\n\tconfig.Version = s.val.kafkaVersion\n\tadmin, err := sarama.NewClusterAdmin(s.val.brokerList, config)\n\tif err != nil {\n\t\t// log.Fatal(\"Cannot create cluster admin %s \", err)\n\t\treturn err\n\t}\n\tdefer admin.Close()\n\t// schema topic should be comacted\n\tcleanupPolicy := \"compact\"\n\terr = admin.CreateTopic(schemaTopic, &sarama.TopicDetail{\n\t\tNumPartitions: 1,\n\t\tReplicationFactor: 1,\n\t\tConfigEntries: map[string]*string{\n\t\t\t\"cleanup.policy\": &cleanupPolicy,\n\t\t},\n\t}, false)\n\treturn err\n\n}", "func CreateTopicOperation(k8sh *utils.K8sHelper, manifests installer.CephManifests) *TopicOperation {\n\treturn &TopicOperation{k8sh, manifests}\n}", "func (tm *topicManager) closeTopic(l log) error {\n\tdelete(tm.topicNamesToLogs, l.getTopicName())\n\tdefer l.close()\n\treturn nil\n}", "func (s *SimpleFSM) Restore(kvMap io.ReadCloser) error {\n\n\tkvSnapshot := make(map[string]string)\n\tif err := json.NewDecoder(kvMap).Decode(&kvSnapshot); err != nil {\n\t\treturn err\n\t}\n\n\t// Set the state from the snapshot, no lock required according to\n\t// Hashicorp docs.\n\tfor k, v := range kvSnapshot {\n\t\ts.db.Update(func(tx *bolt.Tx) error {\n\t\t\tb := tx.Bucket(bucket)\n\t\t\terr := b.Put([]byte(k), []byte(v))\n\t\t\treturn err\n\t\t})\n\t}\n\treturn nil\n}", "func (p *nsqPublisher) CreateTopic(topic string) error {\n\t//nolint:noctx\n\treq, err := http.NewRequest(http.MethodPost, fmt.Sprintf(\"http://%s/topic/create?topic=%s\", p.httpEndpoint, topic), nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\treq.Header.Add(\"ContentType\", \"text/plain\")\n\n\tresp, err := p.client.Do(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif resp.StatusCode >= 300 {\n\t\tbodyBytes, err := io.ReadAll(resp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn fmt.Errorf(\"error creating topic: %s\", string(bodyBytes))\n\t}\n\n\t_ = resp.Body.Close()\n\treturn nil\n}", "func (b *raftBadger) Restore(rClose io.ReadCloser) error {\n\tdefer func() {\n\t\tif err := rClose.Close(); err != nil {\n\t\t\t_, _ = fmt.Fprintf(os.Stdout, \"[FINALLY RESTORE] close error %s\\n\", err.Error())\n\t\t}\n\t}()\n\n\t_, _ = fmt.Fprintf(os.Stdout, \"[START RESTORE] read all message from snapshot\\n\")\n\tvar totalRestored int\n\n\tdecoder := json.NewDecoder(rClose)\n\tfor decoder.More() {\n\t\tvar data = &command{}\n\t\terr := decoder.Decode(data)\n\t\tif err != nil {\n\t\t\tif !strings.Contains(err.Error(), \"EOF\") {\n\t\t\t\t_, _ = fmt.Fprintf(os.Stdout, \"[END RESTORE]snap skipped\\n\", totalRestored)\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\t_, _ = fmt.Fprintf(os.Stdout, \"[END RESTORE] error decode data %s\\n\", err.Error())\n\t\t\treturn err\n\t\t}\n\t\tvar key string\n\t\tif key, err = b.gs.Save(data.Key, data.Store, data.Value); err != nil {\n\t\t\t_, _ = fmt.Fprintf(os.Stdout, \"[END RESTORE] error persist data %s\\n\", err.Error())\n\t\t\treturn err\n\t\t}\n\t\tb.logger.Debug(\"restored key\", \"key\", hclog.Fmt(\"%v\", key))\n\n\t\ttotalRestored++\n\t}\n\n\t// read closing bracket\n\t_, err := decoder.Token()\n\tif err != nil && !strings.Contains(err.Error(), \"EOF\") {\n\t\t_, _ = fmt.Fprintf(os.Stdout, \"[END RESTORE] error %s\\n\", err.Error())\n\t\treturn err\n\t}\n\n\t_, _ = fmt.Fprintf(os.Stdout, \"[END RESTORE] success restore %d messages in snapshot\\n\", totalRestored)\n\treturn nil\n}", "func (client *Client) MetastoreCreateKafkaTopic(request *MetastoreCreateKafkaTopicRequest) (response *MetastoreCreateKafkaTopicResponse, err error) {\n\tresponse = CreateMetastoreCreateKafkaTopicResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func cephRBDVolumeRestore(clusterName string, poolName string, volumeName string,\n\tvolumeType string, snapshotName string, userName string) error {\n\t_, err := shared.RunCommand(\n\t\t\"rbd\",\n\t\t\"--id\", userName,\n\t\t\"--cluster\", clusterName,\n\t\t\"--pool\", poolName,\n\t\t\"snap\",\n\t\t\"rollback\",\n\t\t\"--snap\", snapshotName,\n\t\tfmt.Sprintf(\"%s_%s\", volumeType, volumeName))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func NewTopic(session *mgo.Session, name string) *Topic {\n\treturn &Topic{\n\t\tsession: session,\n\t\tname: name,\n\t}\n}", "func (s *KVStore) RestoreSnapshot(snapshot []byte) {\n\tlog.Printf(\"Restore all value from a snapshot\")\n\tread := bytes.NewBuffer(snapshot)\n\tdecoder := gob.NewDecoder(read)\n\tdecoder.Decode(&s.store)\n}", "func (e *Extractor) buildTopic(topic_name string) meta.Topic {\n\treturn meta.Topic{\n\t\tUrn: topic_name,\n\t\tName: topic_name,\n\t\tSource: \"kafka\",\n\t}\n}", "func (bin unixRecycleBin) Restore(trashFilename string) error {\n\ttrashInfoPath := buildTrashInfoPath(bin.Path, trashFilename)\n\ttrashInfo, err := readTrashInfo(trashInfoPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdeletedFilePath := buildTrashFilePath(bin.Path, trashFilename)\n\tif err := fs.Rename(deletedFilePath, trashInfo.Path); err != nil {\n\t\treturn err\n\t}\n\terr = fs.Remove(buildTrashInfoPath(bin.Path, trashFilename))\n\treturn err\n}", "func rollbackSchema(w io.Writer, projectID, schemaID, revisionID string) error {\n\t// projectID := \"my-project-id\"\n\t// schemaID := \"my-schema\"\n\t// revisionID := \"a1b2c3d4\"\n\tctx := context.Background()\n\tclient, err := pubsub.NewSchemaClient(ctx, projectID)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"pubsub.NewSchemaClient: %w\", err)\n\t}\n\tdefer client.Close()\n\n\ts, err := client.RollbackSchema(ctx, schemaID, revisionID)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"RollbackSchema: %w\", err)\n\t}\n\tfmt.Fprintf(w, \"Rolled back a schema: %#v\\n\", s)\n\treturn nil\n}", "func KeyToTopic(key string) string {\n\t// Record-store keys are arbitrary binary. However, pubsub requires UTF-8 string topic IDs.\n\t// Encodes to \"/record/base64url(key)\"\n\treturn \"/record/\" + base64.RawURLEncoding.EncodeToString([]byte(key))\n}", "func (es *EventStore) AddTopic(topic Topic) (string, error) {\n\tstart := time.Now()\n\tdefer func() {\n\t\tmetrics.EventStoreLatency(\"AddTopic\", start)\n\t}()\n\n\tname := strings.ToLower(topic.Name)\n\tschema := topic.Schema\n\n\tif name == \"\" {\n\t\treturn \"\", errors.New(\"Topic name cannot be empty\")\n\t} else if es.getTopicID(name) != \"\" {\n\t\treturn \"\", jh.NewError(errors.New(\"Topic with name already exists\").Error(), http.StatusConflict)\n\t}\n\n\tschemaStr := \"{}\"\n\tif schema != nil {\n\t\tschemaBytes, err := json.Marshal(schema)\n\t\tif err != nil {\n\t\t\treturn \"\", jh.NewError(errors.Wrap(err, \"Error marshalling schema into json\").Error(), http.StatusBadRequest)\n\t\t}\n\t\tschemaStr = string(schemaBytes)\n\t}\n\n\tjsonSchema, ok := es.validateSchema(schemaStr)\n\tif !ok {\n\t\treturn \"\", jh.NewError(errors.New(\"Error adding topic - schema is not in valid JSON format\").Error(), http.StatusBadRequest)\n\t}\n\n\tid := uuid.NewV4().String()\n\tif err := es.ds.AddTopic(RawTopic{\n\t\tID: id,\n\t\tName: name,\n\t\tSchema: schemaStr,\n\t}); err != nil {\n\t\tmetrics.DBError(\"write\")\n\t\treturn \"\", errors.Wrap(err, \"Error adding topic to data source\")\n\t}\n\n\tes.topicMutex.Lock()\n\tes.topicNameToID[name] = id\n\tes.topicIDToName[id] = name\n\tes.topicSchemaPropertiesMap[id] = schema\n\tes.topicSchemaMap[id] = jsonSchema\n\tes.topicMutex.Unlock()\n\n\treturn id, nil\n}", "func Sub(c mqtt.Client, topic string) {\n\tvar choke = make(chan [2]string)\n\n\tvar f mqtt.MessageHandler = func(client mqtt.Client, msg mqtt.Message) {\n\t\tchoke <- [2]string{msg.Topic(), string(msg.Payload())}\n\t}\n\tfor {\n\t\tif token := c.Subscribe(topic, 0, f); token.Wait() && token.Error() != nil {\n\t\t\tmqtt.ERROR.Println(token.Error())\n\t\t\tos.Exit(1)\n\t\t}\n\t\tfor {\n\t\t\tincoming := <-choke\n\t\t\tmqtt.ERROR.Printf(\"Received:TOPIC: %s\\n\", incoming[0])\n\t\t\twriteFile(incoming[1])\n\t\t}\n\t}\n\n}", "func (api *snapshotrestoreAPI) Create(obj *cluster.SnapshotRestore) error {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = apicl.ClusterV1().SnapshotRestore().Create(context.Background(), obj)\n\t\tif err != nil && strings.Contains(err.Error(), \"AlreadyExists\") {\n\t\t\t_, err = apicl.ClusterV1().SnapshotRestore().Update(context.Background(), obj)\n\n\t\t}\n\t\treturn err\n\t}\n\n\tapi.ct.handleSnapshotRestoreEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Created})\n\treturn nil\n}", "func NewTopic(node storm.Node) *Topic {\n\treturn &Topic{\n\t\tnode: node,\n\t}\n}", "func NewTopic(n string, ts []string) *Topic {\n\treturn &Topic{ID: strings.ToLower(n), Name: n, Tags: ts, Created: time.Now()}\n}", "func ExampleTopicsClient_CreateOrUpdate() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armservicebus.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := clientFactory.NewTopicsClient().CreateOrUpdate(ctx, \"ArunMonocle\", \"sdk-Namespace-1617\", \"sdk-Topics-5488\", armservicebus.SBTopic{\n\t\tProperties: &armservicebus.SBTopicProperties{\n\t\t\tEnableExpress: to.Ptr(true),\n\t\t},\n\t}, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// You could use response here. We use blank identifier for just demo purposes.\n\t_ = res\n\t// If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t// res.SBTopic = armservicebus.SBTopic{\n\t// \tName: to.Ptr(\"sdk-Topics-5488\"),\n\t// \tType: to.Ptr(\"Microsoft.ServiceBus/Namespaces/Topics\"),\n\t// \tID: to.Ptr(\"/subscriptions/5f750a97-50d9-4e36-8081-c9ee4c0210d4/resourceGroups/ArunMonocle/providers/Microsoft.ServiceBus/namespaces/sdk-Namespace-1617/topics/sdk-Topics-5488\"),\n\t// \tProperties: &armservicebus.SBTopicProperties{\n\t// \t\tAccessedAt: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2017-05-26T20:50:34.32Z\"); return t}()),\n\t// \t\tAutoDeleteOnIdle: to.Ptr(\"P10675199DT2H48M5.4775807S\"),\n\t// \t\tCreatedAt: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2017-05-26T20:50:34.1Z\"); return t}()),\n\t// \t\tDefaultMessageTimeToLive: to.Ptr(\"P10675199DT2H48M5.4775807S\"),\n\t// \t\tDuplicateDetectionHistoryTimeWindow: to.Ptr(\"PT10M\"),\n\t// \t\tEnableBatchedOperations: to.Ptr(true),\n\t// \t\tEnableExpress: to.Ptr(true),\n\t// \t\tEnablePartitioning: to.Ptr(false),\n\t// \t\tMaxMessageSizeInKilobytes: to.Ptr[int64](10240),\n\t// \t\tMaxSizeInMegabytes: to.Ptr[int32](10240),\n\t// \t\tRequiresDuplicateDetection: to.Ptr(false),\n\t// \t\tSizeInBytes: to.Ptr[int64](0),\n\t// \t\tStatus: to.Ptr(armservicebus.EntityStatusActive),\n\t// \t\tSubscriptionCount: to.Ptr[int32](0),\n\t// \t\tSupportOrdering: to.Ptr(true),\n\t// \t\tUpdatedAt: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2017-05-26T20:50:34.32Z\"); return t}()),\n\t// \t},\n\t// }\n}", "func (t *Task) buildBackup(client k8sclient.Client, backupTypePrefix string) (*velero.Backup, error) {\n\tvar includeClusterResources *bool = nil\n\tannotations, err := t.getAnnotations(client)\n\tif err != nil {\n\t\treturn nil, liberr.Wrap(err)\n\t}\n\tbackupLocation, err := t.getBSL()\n\tif err != nil {\n\t\treturn nil, liberr.Wrap(err)\n\t}\n\tsnapshotLocation, err := t.getVSL()\n\tif err != nil {\n\t\treturn nil, liberr.Wrap(err)\n\t}\n\n\t// Construct a restore name like \"$migrationname-54823-initial\" or \"$migrationname-54823-stage\".\n\t// This will produce a 57 character string max. Note that generateName gracefully handles strings >63 char.\n\tfmtString := fmt.Sprintf(\"%%.%ds\", 55-len(backupTypePrefix))\n\tmigrationNameTruncated := fmt.Sprintf(fmtString, t.Owner.GetName())\n\ttruncatedGenerateName := fmt.Sprintf(\"%s-%s-\", migrationNameTruncated, backupTypePrefix)\n\n\tbackup := &velero.Backup{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tLabels: t.Owner.GetCorrelationLabels(),\n\t\t\tGenerateName: truncatedGenerateName,\n\t\t\tNamespace: migapi.VeleroNamespace,\n\t\t\tAnnotations: annotations,\n\t\t},\n\t\tSpec: velero.BackupSpec{\n\t\t\tIncludeClusterResources: includeClusterResources,\n\t\t\tStorageLocation: backupLocation.Name,\n\t\t\tVolumeSnapshotLocations: []string{snapshotLocation.Name},\n\t\t\tTTL: metav1.Duration{Duration: 720 * time.Hour},\n\t\t\tIncludedNamespaces: t.sourceNamespaces(),\n\t\t\tHooks: velero.BackupHooks{\n\t\t\t\tResources: []velero.BackupResourceHookSpec{},\n\t\t\t},\n\t\t},\n\t}\n\treturn backup, nil\n}", "func (b *BackupEngine) RestoreDBFromLatestBackup(dbDir, walDir string, ro *RestoreOptions) (err error) {\n\tcDbDir := C.CString(dbDir)\n\tcWalDir := C.CString(walDir)\n\n\tvar cErr *C.char\n\tC.rocksdb_backup_engine_restore_db_from_latest_backup(b.c, cDbDir, cWalDir, ro.c, &cErr)\n\terr = fromCError(cErr)\n\n\tC.free(unsafe.Pointer(cDbDir))\n\tC.free(unsafe.Pointer(cWalDir))\n\treturn\n}", "func NewTopicRouter(br fiber.Router, conn *config.DBConn) {\n\trouter := br.Group(\"/topics\")\n\ttopicRepo := repository.NewTopicRepository(conn)\n\tpostRepo := repository.NewPostRepository(conn)\n\tuserRepo := repository.NewUserRepository(conn)\n\ttopicService := services.NewTopicService(topicRepo, postRepo, userRepo)\n\ttopicController := controllers.NewTopicController(topicService)\n\n\trouter.Get(\"/:id\", topicController.GetOne)\n\trouter.Get(\"/\", topicController.GetAll)\n\trouter.Post(\"/\", topicController.Create)\n}", "func (w *WAL) cut() error {\n\t// close old wal file; truncate to avoid wasting space if an early cut\n\toff, serr := w.tail().Seek(0, io.SeekCurrent)\n\tif serr != nil {\n\t\treturn serr\n\t}\n\n\tif err := w.tail().Truncate(off); err != nil {\n\t\treturn err\n\t}\n\n\tif err := w.sync(); err != nil {\n\t\treturn err\n\t}\n\n\tfpath := filepath.Join(w.dir, walName(w.seq()+1, w.enti+1))\n\n\t// create a temp wal file with name sequence + 1, or truncate the existing one\n\tnewTail, err := w.fp.Open()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// update writer and save the previous crc\n\tw.locks = append(w.locks, newTail)\n\tprevCrc := w.encoder.crc.Sum32()\n\tw.encoder, err = newFileEncoder(w.tail().File, prevCrc)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err = w.saveCrc(prevCrc); err != nil {\n\t\treturn err\n\t}\n\n\tif err = w.encoder.encode(&walpb.Record{Type: metadataType, Data: w.metadata}); err != nil {\n\t\treturn err\n\t}\n\n\tif err = w.saveState(&w.state); err != nil {\n\t\treturn err\n\t}\n\n\t// atomically move temp wal file to wal file\n\tif err = w.sync(); err != nil {\n\t\treturn err\n\t}\n\n\toff, err = w.tail().Seek(0, io.SeekCurrent)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err = os.Rename(newTail.Name(), fpath); err != nil {\n\t\treturn err\n\t}\n\tstart := time.Now()\n\tif err = fileutil.Fsync(w.dirFile); err != nil {\n\t\treturn err\n\t}\n\twalFsyncSec.Observe(time.Since(start).Seconds())\n\n\t// reopen newTail with its new path so calls to Name() match the wal filename format\n\tnewTail.Close()\n\n\tif newTail, err = fileutil.LockFile(fpath, os.O_WRONLY, fileutil.PrivateFileMode); err != nil {\n\t\treturn err\n\t}\n\tif _, err = newTail.Seek(off, io.SeekStart); err != nil {\n\t\treturn err\n\t}\n\n\tw.locks[len(w.locks)-1] = newTail\n\n\tprevCrc = w.encoder.crc.Sum32()\n\tw.encoder, err = newFileEncoder(w.tail().File, prevCrc)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tklog.Infof(fmt.Sprintf(\"created a new WAL segment path:%s\", fpath))\n\treturn nil\n}", "func (v *NopObject) Restore(Snapshot string) (err error) {\n\treturn\n}", "func RestoreSnapshot(snap []byte, appConfig string) *StateMachine {\n\tsm := New(appConfig)\n\terr := json.Unmarshal(snap, sm)\n\tif err != nil {\n\t\tglog.Fatal(\"Unable to restore from snapshot: \", err)\n\t}\n\treturn sm\n}", "func (r Restorer) Restore() {\n\tr()\n}", "func removeTopic(subID string, topics []string) (retErr error) {\n\n\t//Creazione nuovo insieme di topic eliminando quelli da rimuovere\n\torigTopics, err := getTopicList(subID)\n\tif err != nil {\n\t\tcommon.Fatal(\"[BROKER] Errore nell'ottenere la topic list\")\n\t}\n\n\tvar newTopics []string\n\n\tfor _, elem := range origTopics {\n\t\tif common.StringListContains(topics, elem) == false {\n\t\t\tnewTopics = append(newTopics, elem)\n\t\t}\n\t}\n\n\t//Se non rimane nessun topic, viene impostato il singolo topic empty\n\t//Nota bene: questo è necessario farlo poichè dynamodb non supporta le liste vuote\n\tif len(newTopics) == 0 {\n\t\tnewTopics = append(newTopics, \"empty\")\n\t}\n\n\tsvc := dynamodb.New(common.Sess)\n\n\tinput := &dynamodb.UpdateItemInput{\n\t\tExpressionAttributeValues: map[string]*dynamodb.AttributeValue{\n\t\t\t\":t\": {\n\t\t\t\tSS: aws.StringSlice(newTopics),\n\t\t\t},\n\t\t},\n\t\tTableName: aws.String(subTableName),\n\t\tKey: map[string]*dynamodb.AttributeValue{\n\t\t\t\"SubID\": {\n\t\t\t\tS: aws.String(subID),\n\t\t\t},\n\t\t},\n\t\tReturnValues: aws.String(\"UPDATED_NEW\"),\n\t\tUpdateExpression: aws.String(\"set Topics = :t\"),\n\t}\n\n\t_, err = svc.UpdateItem(input)\n\tif err != nil {\n\t\tcommon.Fatal(\"[BROKER] Errore nella rimozione di topic. \" + err.Error())\n\t\treturn err\n\t}\n\n\tcommon.Info(\"[BROKER] Topic rimosso con successo\")\n\n\treturn nil\n}", "func (r *Replica) tryRollbackRaftLearner(\n\tctx context.Context,\n\trangeDesc *roachpb.RangeDescriptor,\n\ttarget roachpb.ReplicationTarget,\n\treason kvserverpb.RangeLogEventReason,\n\tdetails string,\n) {\n\trepDesc, ok := rangeDesc.GetReplicaDescriptor(target.StoreID)\n\tisLearnerOrNonVoter := repDesc.GetType() == roachpb.LEARNER || repDesc.GetType() == roachpb.NON_VOTER\n\tif !ok || !isLearnerOrNonVoter {\n\t\t// There's no learner to roll back.\n\t\tlog.Event(ctx, \"learner to roll back not found; skipping\")\n\t\treturn\n\t}\n\n\t// If (for example) the promotion failed because of a context deadline\n\t// exceeded, we do still want to clean up after ourselves, so always use a new\n\t// context (but with the old tags and with some timeout to save this from\n\t// blocking the caller indefinitely).\n\tconst rollbackTimeout = 10 * time.Second\n\n\trollbackFn := func(ctx context.Context) error {\n\t\t_, err := execChangeReplicasTxn(\n\t\t\tctx, rangeDesc, reason, details,\n\t\t\t[]internalReplicationChange{{target: target, typ: internalChangeTypeRemove}},\n\t\t\tchangeReplicasTxnArgs{\n\t\t\t\tdb: r.store.DB(),\n\t\t\t\tliveAndDeadReplicas: r.store.allocator.storePool.liveAndDeadReplicas,\n\t\t\t\tlogChange: r.store.logChange,\n\t\t\t\ttestForceJointConfig: r.store.TestingKnobs().ReplicationAlwaysUseJointConfig,\n\t\t\t\ttestAllowDangerousReplicationChanges: r.store.TestingKnobs().AllowDangerousReplicationChanges,\n\t\t\t})\n\t\treturn err\n\t}\n\trollbackCtx := logtags.WithTags(context.Background(), logtags.FromContext(ctx))\n\tif err := contextutil.RunWithTimeout(\n\t\trollbackCtx, \"learner rollback\", rollbackTimeout, rollbackFn,\n\t); err != nil {\n\t\tlog.Infof(\n\t\t\tctx,\n\t\t\t\"failed to rollback %s %s, abandoning it for the replicate queue: %v\",\n\t\t\trepDesc.GetType(),\n\t\t\ttarget,\n\t\t\terr,\n\t\t)\n\t\tr.store.replicateQueue.MaybeAddAsync(ctx, r, r.store.Clock().NowAsClockTimestamp())\n\t} else {\n\t\tlog.Infof(ctx, \"rolled back %s %s in %s\", repDesc.GetType(), target, rangeDesc)\n\t}\n}", "func TestSubscriptionTopic(t *testing.T) {\n\tsub := NewFSubscription(\"foo\", nil)\n\tassert.Equal(t, \"foo\", sub.Topic())\n}", "func GenerateTopic(name string, s v1alpha1.TopicParameters) *pubsub.Topic {\n\tt := &pubsub.Topic{\n\t\tName: name,\n\t\tLabels: s.Labels,\n\t\tKmsKeyName: gcp.StringValue(s.KmsKeyName),\n\t}\n\tif s.MessageStoragePolicy != nil {\n\t\tt.MessageStoragePolicy = &pubsub.MessageStoragePolicy{\n\t\t\tAllowedPersistenceRegions: s.MessageStoragePolicy.AllowedPersistenceRegions,\n\t\t}\n\t}\n\tif s.MessageRetentionDuration != nil {\n\t\tt.MessageRetentionDuration = gcp.StringValue(s.MessageRetentionDuration)\n\t}\n\treturn t\n}", "func CreateRecovery(cluster, namespace, volumeName, deploymentName, mountPath, pathRestic string) {\n\tvar recovery map[string]interface{}\n\tvar nameRecovery string\n\t// TODO Backend -> local, s3, glusterFS, ...\n\tif cluster == \"ClusterFrom\" {\n\t\trecovery = utils.ReadJson(\"templates/recovery\", \"recovery_s3_template_from\")\n\t\tnameRecovery= \"recoveryFrom\"\n\t} else {\n\t\trecovery = utils.ReadJson(\"templates/recovery\", \"recovery_s3_template_to\")\n\t\tnameRecovery = \"recoveryTo\"\n\t}\n\n\t// Change namespace, name,\n\tauxName := \"recovery-\" + deploymentName\n\tauxName = deploymentName\n\trecovery[\"metadata\"].(map[string]interface{})[\"name\"] = auxName\n\trecovery[\"metadata\"].(map[string]interface{})[\"namespace\"] = namespace\n\trecovery[\"spec\"].(map[string]interface{})[\"workload\"].(map[string]interface{})[\"name\"] = deploymentName\n\trecovery[\"spec\"].(map[string]interface{})[\"paths\"].([]interface{})[0] = mountPath\n\trecovery[\"spec\"].(map[string]interface{})[\"recoveredVolumes\"].([]interface{})[0].(map[string]interface{})[\"mountPath\"] = mountPath\n\n\terr := utils.WriteJson(pathRestic, nameRecovery, recovery)\n\tif err != nil {\n\t\tfmt.Println(\"Error creating \" + auxName)\n\t}\n}", "func (m *subscriptionMigrator) migrateSubscription(sub kymaeventingv1alpha1.Subscription) error {\n\tsubKey := fmt.Sprintf(\"%s/%s\", sub.Namespace, sub.Name)\n\n\tlog.Printf(\"+ Checking Subscription %q\", subKey)\n\tif m.findTriggerForSubscription(sub) != nil {\n\t\tlog.Printf(\"+ Trigger already exists for Subscription %q, skipping\", subKey)\n\t\treturn nil\n\t}\n\n\tlog.Printf(\"+ Trigger not found for Subscription %q\", subKey)\n\ttrigger, err := m.createTriggerForSubscription(sub)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"creating Trigger for Subscription %q\", subKey)\n\t}\n\tlog.Printf(\"+ Trigger \\\"%s/%s\\\" created for Subscription %q\", trigger.Namespace, trigger.Name, subKey)\n\n\treturn nil\n}", "func (b *Topics) CreateTopic(topic_id string) {\n\tb.sub_lock.Lock()\n\tif nil == b.sub_topics[topic_id] {\n\t\tb.sub_topics[topic_id] = Subscribers{}\n\t}\n\tb.sub_lock.Unlock()\n}", "func (a *actuator) Restore(ctx context.Context, ex *extensionsv1alpha1.Extension) error {\n\treturn a.Reconcile(ctx, ex)\n}", "func TestTopicBeDeleteDuringProducerRunning(t *testing.T) {\n\tSetUp(t)\n\n\ttopicInfo2 := &topic.TopicInfo{\n\t\tTopicName: topicName,\n\t\tTopicTalosResourceName: &topic.TopicTalosResourceName{anotherResourceName},\n\t\tOwnerId: ownerId,\n\t}\n\ttopicAttribute2 := &topic.TopicAttribute{\n\t\tPartitionNumber: thrift.Int32Ptr(partitionNumber),\n\t\tMessageRetentionSecs: thrift.Int32Ptr(messageRetentionMs),\n\t}\n\ttopicState2 := &topic.TopicState{\n\t\tTopicStatus: topic.TopicStatus_ACTIVE,\n\t\tCreateTimestamp: utils.CurrentTimeMills(),\n\t}\n\tanother := &topic.Topic{\n\t\tTopicInfo: topicInfo2,\n\t\tTopicAttribute: topicAttribute2,\n\t\tTopicState: topicState2,\n\t}\n\n\tgomock.InOrder(\n\t\ttalosAdminMock.EXPECT().DescribeTopic(&topic.DescribeTopicRequest{topicName}).Return(topic1, nil).Times(1),\n\t\ttalosAdminMock.EXPECT().DescribeTopic(&topic.DescribeTopicRequest{topicName}).Return(another, nil).Times(1),\n\t)\n\n\ttalosProducer = producer.NewTalosProducerForTest(talosProducerConfig,\n\t\ttalosClientFactoryMock, talosAdminMock,\n\t\t&topic.TopicTalosResourceName{resourceName},\n\t\t&client.SimpleTopicAbnormalCallback{}, &TestCallback{})\n\n\t// wait check partition interval\n\ttime.Sleep(time.Duration(checkPartitionInterval*2) * time.Millisecond)\n}", "func (ctl Controller) RestoreLatest() *pitr.Error {\n\treturn ctl.Restore(\"LATEST\")\n}", "func TestStore_Snapshot_And_Restore(t *testing.T) {\n\tt.Parallel()\n\n\ts := MustOpenStore()\n\ts.LeaveFiles = true\n\n\t// Create a bunch of databases in the Store\n\tnDatabases := 5\n\tfor n := 0; n < nDatabases; n++ {\n\t\ts.CreateDatabase(fmt.Sprintf(\"db%d\", n))\n\t}\n\n\t// Test taking a snapshot.\n\tif err := s.Store.Snapshot(); err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\ts.Close()\n\n\t// Test restoring the snapshot taken above.\n\texistingDataPath := s.Path()\n\ts = NewStore(NewConfig(existingDataPath))\n\tif err := s.Open(); err != nil {\n\t\tpanic(err)\n\t}\n\tdefer s.Close()\n\n\t// Wait until the server is ready.\n\tselect {\n\tcase err := <-s.Err():\n\t\tpanic(err)\n\tcase <-s.Ready():\n\t}\n\n\t// Make sure all the data we added to the Store is still there.\n\tfor n := 0; n < nDatabases; n++ {\n\t\tname := fmt.Sprintf(\"db%d\", n)\n\t\tif dbi, err := s.Database(name); err != nil {\n\t\t\tt.Fatal(err)\n\t\t} else if dbi == nil {\n\t\t\tt.Fatalf(\"database not found: %s\", name)\n\t\t} else if dbi.Name != name {\n\t\t\tt.Fatal(name)\n\t\t}\n\t}\n}", "func (_this *RaftNode) replayWAL() *wal.WAL {\n\tw := _this.openWAL()\n\t_, st, ents, err := w.ReadAll()\n\tif err != nil {\n\t\tlog.Fatalf(\"Failed to read WAL. %v\", err)\n\t}\n\n\t// append to storage so raft starts at the right place in log.\n\t_this.raftStorage.Append(ents)\n\t// send nil once lastIndex is published so client knows commit channel is current\n\tif len(ents) > 0 {\n\t\t_this.lastIndex = ents[len(ents)-1].Index\n\t} else {\n\t\t_this.commitC <- nil\n\t}\n\t_this.raftStorage.SetHardState(st)\n\treturn w\n}", "func (tm *TabletManager) restoreToTimeFromBinlog(ctx context.Context, pos replication.Position, restoreTime *vttime.Time) error {\n\t// validate the minimal settings necessary for connecting to binlog server\n\tif binlogHost == \"\" || binlogPort <= 0 || binlogUser == \"\" {\n\t\tlog.Warning(\"invalid binlog server setting, restoring to last available backup.\")\n\t\treturn nil\n\t}\n\n\ttimeoutCtx, cancelFnc := context.WithTimeout(ctx, timeoutForGTIDLookup)\n\tdefer cancelFnc()\n\n\tafterGTIDPos, beforeGTIDPos, err := tm.getGTIDFromTimestamp(timeoutCtx, pos, restoreTime.Seconds)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif afterGTIDPos == \"\" && beforeGTIDPos == \"\" {\n\t\treturn vterrors.New(vtrpcpb.Code_FAILED_PRECONDITION, fmt.Sprintf(\"unable to fetch the GTID for the specified time - %s\", restoreTime.String()))\n\t} else if afterGTIDPos == \"\" && beforeGTIDPos != \"\" {\n\t\tlog.Info(\"no afterGTIDPos found, which implies we reached the end of all GTID events\")\n\t}\n\n\tlog.Infof(\"going to restore upto the GTID - %s\", afterGTIDPos)\n\t// when we don't have before GTID, we will take it as current backup pos's last GTID\n\t// this is case where someone tries to restore just to the 1st event after backup\n\tif beforeGTIDPos == \"\" {\n\t\tbeforeGTIDPos = pos.GTIDSet.Last()\n\t}\n\terr = tm.catchupToGTID(timeoutCtx, afterGTIDPos, beforeGTIDPos)\n\tif err != nil {\n\t\treturn vterrors.Wrapf(err, \"unable to replicate upto desired GTID : %s\", afterGTIDPos)\n\t}\n\n\treturn nil\n}", "func (c *Container) Restore(ctx context.Context, name string) error {\n\treturn c.client.ContainerStart(ctx, c.id, types.ContainerStartOptions{CheckpointID: name})\n}", "func (r *Release) rollback(cs *currentState, p *plan) {\n\trs, ok := cs.releases[r.key()]\n\tif !ok {\n\t\treturn\n\t}\n\n\tif r.Namespace == rs.Namespace {\n\n\t\tcmd := helmCmd(concat([]string{\"rollback\", r.Name, rs.getRevision()}, r.getWait(), r.getTimeout(), r.getNoHooks(), flags.getRunFlags()), \"Rolling back release [ \"+r.Name+\" ] in namespace [ \"+r.Namespace+\" ]\")\n\t\tp.addCommand(cmd, r.Priority, r, []hookCmd{}, []hookCmd{})\n\t\tr.upgrade(p) // this is to reflect any changes in values file(s)\n\t\tp.addDecision(\"Release [ \"+r.Name+\" ] was deleted and is desired to be rolled back to \"+\n\t\t\t\"namespace [ \"+r.Namespace+\" ]\", r.Priority, create)\n\t} else {\n\t\tr.reInstall(p)\n\t\tp.addDecision(\"Release [ \"+r.Name+\" ] is deleted BUT from namespace [ \"+rs.Namespace+\n\t\t\t\" ]. Will purge delete it from there and install it in namespace [ \"+r.Namespace+\" ]\", r.Priority, create)\n\t\tp.addDecision(\"WARNING: rolling back release [ \"+r.Name+\" ] from [ \"+rs.Namespace+\" ] to [ \"+r.Namespace+\n\t\t\t\" ] might not correctly connect to existing volumes. Check https://github.com/Praqma/helmsman/blob/master/docs/how_to/apps/moving_across_namespaces.md\"+\n\t\t\t\" for details if this release uses PV and PVC.\", r.Priority, create)\n\t}\n}", "func CreateTopic(p *kafka.Producer, topic string) {\n\n\ta, err := kafka.NewAdminClientFromProducer(p)\n\tif err != nil {\n\t\tfmt.Printf(\"Failed to create new admin client from producer: %s\", err)\n\t\tos.Exit(1)\n\t}\n\t// Contexts are used to abort or limit the amount of time\n\t// the Admin call blocks waiting for a result.\n\tctx, cancel := context.WithCancel(context.Background())\n\tdefer cancel()\n\t// Create topics on cluster.\n\t// Set Admin options to wait up to 60s for the operation to finish on the remote cluster\n\tmaxDur, err := time.ParseDuration(\"60s\")\n\tif err != nil {\n\t\tfmt.Printf(\"ParseDuration(60s): %s\", err)\n\t\tos.Exit(1)\n\t}\n\tresults, err := a.CreateTopics(\n\t\tctx,\n\t\t// Multiple topics can be created simultaneously\n\t\t// by providing more TopicSpecification structs here.\n\t\t[]kafka.TopicSpecification{{\n\t\t\tTopic: topic,\n\t\t\tNumPartitions: 1,\n\t\t\tReplicationFactor: 3}},\n\t\t// Admin options\n\t\tkafka.SetAdminOperationTimeout(maxDur))\n\tif err != nil {\n\t\tfmt.Printf(\"Admin Client request error: %v\\n\", err)\n\t\tos.Exit(1)\n\t}\n\tfor _, result := range results {\n\t\tif result.Error.Code() != kafka.ErrNoError && result.Error.Code() != kafka.ErrTopicAlreadyExists {\n\t\t\tfmt.Printf(\"Failed to create topic: %v\\n\", result.Error)\n\t\t\tos.Exit(1)\n\t\t}\n\t\tfmt.Printf(\"%v\\n\", result)\n\t}\n\ta.Close()\n\n}", "func ExampleServersClient_BeginCreate_createADatabaseAsAPointInTimeRestore() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armmariadb.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tpoller, err := clientFactory.NewServersClient().BeginCreate(ctx, \"TargetResourceGroup\", \"targetserver\", armmariadb.ServerForCreate{\n\t\tLocation: to.Ptr(\"brazilsouth\"),\n\t\tProperties: &armmariadb.ServerPropertiesForRestore{\n\t\t\tCreateMode: to.Ptr(armmariadb.CreateModePointInTimeRestore),\n\t\t\tRestorePointInTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2017-12-14T00:00:37.467Z\"); return t }()),\n\t\t\tSourceServerID: to.Ptr(\"/subscriptions/ffffffff-ffff-ffff-ffff-ffffffffffff/resourceGroups/SourceResourceGroup/providers/Microsoft.DBforMariaDB/servers/sourceserver\"),\n\t\t},\n\t\tSKU: &armmariadb.SKU{\n\t\t\tName: to.Ptr(\"GP_Gen5_2\"),\n\t\t\tCapacity: to.Ptr[int32](2),\n\t\t\tFamily: to.Ptr(\"Gen5\"),\n\t\t\tTier: to.Ptr(armmariadb.SKUTierGeneralPurpose),\n\t\t},\n\t\tTags: map[string]*string{\n\t\t\t\"ElasticServer\": to.Ptr(\"1\"),\n\t\t},\n\t}, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\tres, err := poller.PollUntilDone(ctx, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to pull the result: %v\", err)\n\t}\n\t// You could use response here. We use blank identifier for just demo purposes.\n\t_ = res\n\t// If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t// res.Server = armmariadb.Server{\n\t// \tName: to.Ptr(\"targetserver\"),\n\t// \tType: to.Ptr(\"Microsoft.DBforMariaDB/servers\"),\n\t// \tID: to.Ptr(\"/subscriptions/ffffffff-ffff-ffff-ffff-ffffffffffff/resourceGroups/testrg/providers/Microsoft.DBforMariaDB/servers/targetserver\"),\n\t// \tLocation: to.Ptr(\"brazilsouth\"),\n\t// \tTags: map[string]*string{\n\t// \t\t\"elasticServer\": to.Ptr(\"1\"),\n\t// \t},\n\t// \tProperties: &armmariadb.ServerProperties{\n\t// \t\tAdministratorLogin: to.Ptr(\"cloudsa\"),\n\t// \t\tEarliestRestoreDate: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2018-03-14T18:02:41.577+00:00\"); return t}()),\n\t// \t\tFullyQualifiedDomainName: to.Ptr(\"targetserver.mariadb.database.azure.com\"),\n\t// \t\tSSLEnforcement: to.Ptr(armmariadb.SSLEnforcementEnumEnabled),\n\t// \t\tStorageProfile: &armmariadb.StorageProfile{\n\t// \t\t\tBackupRetentionDays: to.Ptr[int32](7),\n\t// \t\t\tGeoRedundantBackup: to.Ptr(armmariadb.GeoRedundantBackupEnabled),\n\t// \t\t\tStorageMB: to.Ptr[int32](128000),\n\t// \t\t},\n\t// \t\tUserVisibleState: to.Ptr(armmariadb.ServerStateReady),\n\t// \t\tVersion: to.Ptr(armmariadb.ServerVersionTen3),\n\t// \t},\n\t// \tSKU: &armmariadb.SKU{\n\t// \t\tName: to.Ptr(\"GP_Gen5_2\"),\n\t// \t\tCapacity: to.Ptr[int32](2),\n\t// \t\tFamily: to.Ptr(\"Gen5\"),\n\t// \t\tTier: to.Ptr(armmariadb.SKUTierGeneralPurpose),\n\t// \t},\n\t// }\n}", "func (_Flytrap *FlytrapTransactorSession) AddTopic(topic string, country [2]byte, addPubCost *big.Int, addSubCost *big.Int, reason string, sensitive bool) (*types.Transaction, error) {\n\treturn _Flytrap.Contract.AddTopic(&_Flytrap.TransactOpts, topic, country, addPubCost, addSubCost, reason, sensitive)\n}", "func newSubscriptionMigrator(kymaClient kymaeventingclientset.Interface,\n\tknativeClient kneventingclientset.Interface, namespaces []string) (*subscriptionMigrator, error) {\n\n\tm := &subscriptionMigrator{\n\t\tkymaClient: kymaClient,\n\t\tknativeClient: knativeClient,\n\t}\n\n\tif err := m.populateSubscriptions(namespaces); err != nil {\n\t\treturn nil, err\n\t}\n\tif err := m.populateTriggers(namespaces); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn m, nil\n}", "func (s *StanServer) createSubStore() *subStore {\n\tsubs := &subStore{\n\t\tpsubs: make([]*subState, 0, 4),\n\t\tqsubs: make(map[string]*queueState),\n\t\tdurables: make(map[string]*subState),\n\t\tacks: make(map[string]*subState),\n\t\tinboxes: make(map[string]*subState),\n\t\tstan: s,\n\t}\n\treturn subs\n}", "func RestoreBackup(dto *action_dtos.RestoreBackupDto, service *pgCommon.PostgresServiceInformations) error {\n\t_, err := service.PgoApi.RestoreBackup(&msgs.RestoreRequest{\n\t\tBackrestStorageType: \"s3\",\n\t\tNamespace: service.ClusterInstance.Namespace,\n\t\tFromCluster: dto.OldClusterName,\n\t})\n\n\tif err != nil {\n\t\tlogger.RError(err, \"Unable to progress restore backup action for \"+service.ClusterInstance.Name)\n\t}\n\n\treturn err\n}", "func create(l *lambdasrv.Lambda, version string, note CreateNote) (string, error) {\n\tnoteId := &id{}\n\tstatus, err := lambda.InvokeLambda(l, version, nil, nil, note, noteId)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"saver lambda: %w\", err)\n\t}\n\tif status != http.StatusCreated {\n\t\treturn \"\", fmt.Errorf(\"create note: expected status: %v got %v\", http.StatusCreated, status)\n\t}\n\treturn noteId.ID, nil\n}", "func NewTopic() *Topic {\n\treturn &Topic{}\n}", "func (_Flytrap *FlytrapTransactor) AddTopic(opts *bind.TransactOpts, topic string, country [2]byte, addPubCost *big.Int, addSubCost *big.Int, reason string, sensitive bool) (*types.Transaction, error) {\n\treturn _Flytrap.contract.Transact(opts, \"addTopic\", topic, country, addPubCost, addSubCost, reason, sensitive)\n}", "func (c *consumer) Back() error {\n\tif err := c.store.Back(c.topic, c.ackOffset); err != nil {\n\t\treturn fmt.Errorf(\"backing topic %s with offset %d: %v\", c.topic, c.ackOffset, err)\n\t}\n\n\tc.outstanding = false\n\tc.notifier.NotifyConsumer(c.topic, eventTypeBack)\n\n\treturn nil\n}", "func CreateTopics(ctx context.Context, admin *kafka.AdminClient, details KafkaBroker) {\n\n\t// create topic configs\n\ttopicConfigs := buildTopicConfig(details)\n\tresults, _ := admin.CreateTopics(ctx, topicConfigs)\n\n\tfmt.Println(\"result topic creation\", results)\n\n\tfailedTopics := make(map[string]bool)\n\tfor _, res := range results {\n\t\tif res.Error.Code() == kafka.ErrTopicAlreadyExists {\n\t\t\tfailedTopics[res.Topic] = true\n\t\t}\n\t}\n\n\tif len(failedTopics) > 0 {\n\t\talterTopicConfigs := make([]kafka.TopicSpecification, 0, 0)\n\t\tfor _, topic := range topicConfigs {\n\t\t\tif _, ok := failedTopics[topic.Topic]; ok {\n\t\t\t\talterTopicConfigs = append(alterTopicConfigs, topic)\n\t\t\t}\n\t\t}\n\n\t\talterTopicsCgf := alterConfig(alterTopicConfigs)\n\t\tres, err := admin.AlterConfigs(ctx, alterTopicsCgf)\n\t\tfmt.Println(\"result topic alteration\", res, err)\n\t}\n\n}", "func initAndRestore(dbInfo *DatabaseInfo, targetDir string) error {\n\trestore, err := getMongoRestore(dbInfo, targetDir)\n\tif err != nil {\n\t\tfmt.Printf(\"Mongo restore initialization failed: %s\", err)\n\t\treturn err\n\t}\n\tif result := restore.Restore(); result.Err != nil {\n\t\tfmt.Printf(\"Mongo restore failed: %s\", result.Err)\n\t\treturn result.Err\n\t}\n\treturn nil\n}", "func restoreKV(r *Restore, c *consul.Consul) {\n\trestoredKeyCount := 0\n\terrorCount := 0\n\tfor _, data := range r.JSONData {\n\t\t_, err := c.Client.KV().Put(data, nil)\n\t\tif err != nil {\n\t\t\terrorCount++\n\t\t\tlog.Printf(\"Unable to restore key: %s, %v\", data.Key, err)\n\t\t}\n\t\trestoredKeyCount++\n\t}\n\tlog.Printf(\"[INFO] Restored %v keys with %v errors\", restoredKeyCount, errorCount)\n}", "func NewGenericWAL(wal *WAL) (*GenericWAL, error) {\n gWal := &GenericWAL{\n wal: wal,\n recoveryMap: make(map[int64]*GenericWALRecord),\n }\n return gWal, nil\n}", "func New(path string, logger *persist.Logger, cancel <-chan struct{}, walStopped chan struct{}, settings map[string]bool) (u []Update, w *WAL, err error) {\n\t// Create a new WAL\n\tnewWal := WAL{\n\t\tavailablePages: []uint64{},\n\t\tfilePageCount: 0,\n\t\ttransactionCounter: 0,\n\t\tlogFile: nil,\n\t\tlog: logger,\n\t\tsettings: settings,\n\t}\n\n\t// Try opening the WAL file.\n\tnewWal.logFile, err = os.OpenFile(path, os.O_RDWR, 0600)\n\tif err == nil {\n\t\t// err == nil indicates that there is a WAL file, which means that the\n\t\t// previous shutdown was not clean. Restore the WAL and return the updates\n\t\tnewWal.log.Println(\"WARN: WAL file detected, performing recovery after unclean shutdown.\")\n\n\t\t// Recover WAL and return updates\n\t\tupdates, err := newWal.recover()\n\t\treturn updates, &newWal, err\n\n\t} else if !os.IsNotExist(err) {\n\t\t// the file exists but couldn't be opened\n\t\treturn nil, nil, build.ExtendErr(\"walFile was not opened successfully\", err)\n\t}\n\n\t// Create new empty WAL\n\tnewWal.logFile, err = os.Create(path)\n\tif err != nil {\n\t\treturn nil, nil, build.ExtendErr(\"walFile could not be created\", err)\n\t}\n\n\t// If there were no errors prepare clean shutdown\n\tgo func() {\n\t\tselect {\n\t\tcase <-cancel:\n\t\t}\n\t\tw.logFile.Close()\n\t\tif !w.settings[\"cleanWALFile\"] {\n\t\t\tos.Remove(path)\n\t\t}\n\t\tclose(walStopped)\n\t}()\n\treturn nil, &newWal, nil\n}", "func createTopics(kafkaAddress string, topics ...string) error {\n\n\tconn, err := kafka.Dial(\"tcp\", kafkaAddress)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"event-publisher.createTopics\")\n\t}\n\tdefer conn.Close()\n\n\tcontroller, err := conn.Controller()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"event-publisher.createTopics\")\n\t}\n\tvar controllerConn *kafka.Conn\n\tcontrollerConn, err = kafka.Dial(\"tcp\", net.JoinHostPort(controller.Host, strconv.Itoa(controller.Port)))\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"event-publisher.createTopics\")\n\t}\n\tdefer controllerConn.Close()\n\tfor _, topic := range topics {\n\n\t\ttopicConfigs := []kafka.TopicConfig{\n\t\t\tkafka.TopicConfig{\n\t\t\t\tTopic: topic,\n\t\t\t\tNumPartitions: 1,\n\t\t\t\tReplicationFactor: 1,\n\t\t\t},\n\t\t}\n\n\t\terr = controllerConn.CreateTopics(topicConfigs...)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"event-publisher.createTopics\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func (svc *SNS) createTopic(name string) (topicARN string, err error) {\n\ttopicName := svc.prefix + name\n\tin := &SDK.CreateTopicInput{\n\t\tName: pointers.String(topicName),\n\t}\n\tresp, err := svc.client.CreateTopic(in)\n\tif err != nil {\n\t\tsvc.Errorf(\"error on `CreateTopic` operation; name=%s; error=%s;\", name, err.Error())\n\t\treturn \"\", err\n\t}\n\treturn *resp.TopicArn, nil\n}", "func (t *TopicService) Create(name, description, parentID string) (*Topic, error) {\n\tquery := `\n\tmutation (\n\t\t$name: String!,\n\t\t$description: String,\n\t\t$parentId: ID\n\t){\n\t\tcreateTopic(\n\t\t\tname: $name, description: $description, parentId: $parentId\n\t\t){ id, name, description }\n\t}`\n\tvar resp struct {\n\t\tTopic *Topic `json:\"createTopic\"`\n\t}\n\tvars := map[string]interface{}{\"name\": name, \"description\": description, \"parentId\": parentID}\n\terr := t.client.Do(context.Background(), query, vars, &resp)\n\treturn resp.Topic, err\n}" ]
[ "0.7388679", "0.5776889", "0.54517996", "0.5324671", "0.52316856", "0.52075505", "0.5200593", "0.5074295", "0.50688195", "0.5018047", "0.49563769", "0.49482065", "0.49438968", "0.49104205", "0.49067542", "0.4878728", "0.48616022", "0.48592317", "0.48512042", "0.48401818", "0.48293346", "0.4826911", "0.48152378", "0.48047054", "0.478865", "0.47751322", "0.4694502", "0.46688735", "0.4644747", "0.46117958", "0.46075088", "0.45825455", "0.45693532", "0.45623946", "0.4548324", "0.45469782", "0.4530172", "0.45093954", "0.4502007", "0.44982842", "0.44855177", "0.4485051", "0.44789207", "0.4477266", "0.44754502", "0.44626677", "0.44485766", "0.44431373", "0.44363376", "0.44273862", "0.4421797", "0.44150594", "0.44131383", "0.44121748", "0.44058815", "0.43975303", "0.43871278", "0.43844417", "0.43826082", "0.43683216", "0.43625695", "0.43619534", "0.43587846", "0.43539327", "0.4350482", "0.43497932", "0.43490776", "0.4349014", "0.43482825", "0.43355948", "0.43245918", "0.43240255", "0.43214008", "0.43174917", "0.43117222", "0.43098116", "0.42978573", "0.4289457", "0.42893183", "0.42887723", "0.4284596", "0.4284283", "0.42840865", "0.42825395", "0.42790943", "0.42769918", "0.4274504", "0.42735937", "0.4270964", "0.4269312", "0.42679977", "0.42598197", "0.4256943", "0.424216", "0.424088", "0.4240748", "0.42387018", "0.42381227", "0.4235825", "0.4234888" ]
0.8242817
0
Append appends to the wal of this topic
func (t *WALTopic) Append(data Data) error { var buf bytes.Buffer _ = gob.NewEncoder(&buf).Encode(data) if err := t.wal.Log(buf.Bytes()); err != nil { return errors.Wrapf(err, "Error appending to wal for topic: %s", t.name) } for _, a := range t.readerAlerts { if a.waiting { a.appendC <- true } } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (t *Topic) Append(msg schema.Message) {\n\tt.lock.Lock()\n\tdefer t.lock.Unlock()\n\tt.Depth++\n\tmsg.Offset = t.Depth\n\tt.Channel <- msg\n\tt.buffer.Put(msg)\n}", "func (l *Log) Append(ctx context.Context, msg Message) error {\n\tdata, err := json.Marshal(msg)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar b pgx.Batch\n\tb.Queue(\"begin\")\n\tb.Queue(\"lock table switchover_log in exclusive mode\")\n\tb.Queue(\"insert into switchover_log (id, timestamp, data) values (coalesce((select max(id)+1 from switchover_log), 1), now(), $1)\", data)\n\tb.Queue(\"commit\")\n\tb.Queue(\"rollback\")\n\n\tconn, err := stdlib.AcquireConn(l.db)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer releaseConn(l.db, conn)\n\n\terr = conn.SendBatch(ctx, &b).Close()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (l *InMemoryLog) Append(value []byte) error {\n\tl.Lock()\n\tdefer l.Unlock()\n\tl.create(value)\n\treturn nil\n}", "func (w *tWriter) append(key, value []byte) error {\n\tif w.first == nil {\n\t\tw.first = append([]byte{}, key...)\n\t}\n\tw.last = append(w.last[:0], key...)\n\treturn w.tw.Append(key, value)\n}", "func (w *Writer) Append(msg *sej.Message) error {\n\twriter, err := w.ws[int(w.shard(msg)&w.shardMask)].getOrOpen()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn writer.Append(msg)\n}", "func (r *Raft) AppendToLog_Leader(cmd []byte) {\n\tterm := r.currentTerm\n\tlogVal := LogVal{term, cmd, 0} //make object for log's value field with acks set to 0\n\t//fmt.Println(\"Before putting in log,\", logVal)\n\tr.myLog = append(r.myLog, logVal)\n\t//fmt.Println(\"I am:\", r.Myconfig.Id, \"Added cmd to my log\")\n\n\t//modify metadata after appending\n\t//fmt.Println(\"Metadata before appending,lastLogIndex,prevLogIndex,prevLogTerm\", r.myMetaData.lastLogIndex, r.myMetaData.prevLogIndex, r.myMetaData.prevLogTerm)\n\tlastLogIndex := r.myMetaData.lastLogIndex + 1\n\tr.myMetaData.prevLogIndex = r.myMetaData.lastLogIndex\n\tr.myMetaData.lastLogIndex = lastLogIndex\n\t//fmt.Println(r.myId(), \"Length of my log is\", len(r.myLog))\n\tif len(r.myLog) == 1 {\n\t\tr.myMetaData.prevLogTerm = r.myMetaData.prevLogTerm + 1 //as for empty log prevLogTerm is -2\n\n\t} else if len(r.myLog) > 1 { //explicit check, else would have sufficed too, just to eliminate len=0 possibility\n\t\tr.myMetaData.prevLogTerm = r.myLog[r.myMetaData.prevLogIndex].Term\n\t}\n\t//r.currentTerm = term\n\t//fmt.Println(\"I am leader, Appended to log, last index, its term is\", r.myMetaData.lastLogIndex, r.myLog[lastLogIndex].term)\n\t//fmt.Println(\"Metadata after appending,lastLogIndex,prevLogIndex,prevLogTerm\", r.myMetaData.lastLogIndex, r.myMetaData.prevLogIndex, r.myMetaData.prevLogTerm)\n\tr.setNextIndex_All() //Added-28 march for LogRepair\n\t//Write to disk\n\t//fmt.Println(r.myId(), \"In append_leader, appended to log\", string(cmd))\n\tr.WriteLogToDisk()\n\n}", "func (l *TimestampedLog) Append(ts int64, data []byte) error {\n\tlatest, err := l.latest()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif ts < latest {\n\t\treturn errors.New(\"TimestampedLog.append: wrong timestamp\")\n\t}\n\n\tidx, err := l.addToSize(1)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar buf bytes.Buffer\n\tbuf.Write(util.Uint64To8Bytes(uint64(ts)))\n\tbuf.Write(data)\n\tl.kvw.Set(l.getElemKey(idx), buf.Bytes())\n\treturn nil\n}", "func (s *Storage) Append(smpl *model.Sample) error {\n\ts.mtx.RLock()\n\n\tvar snew model.Sample\n\tsnew = *smpl\n\tsnew.Metric = smpl.Metric.Clone()\n\n\tfor ln, lv := range s.externalLabels {\n\t\tif _, ok := smpl.Metric[ln]; !ok {\n\t\t\tsnew.Metric[ln] = lv\n\t\t}\n\t}\n\ts.mtx.RUnlock()\n\n\tfor _, q := range s.queues {\n\t\tq.Append(&snew)\n\t}\n\treturn nil\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\n\t// Your code here.\n\tif ck.view.Viewnum == 0 {\n\t\tck.UpdateView()\n\t}\n\targs := PutAppendArgs{key, value, ck.me, strconv.FormatInt(nrand(), 10), op}\n\tvar reply PutAppendReply\n\tfor {\n\t\tok := call(ck.view.Primary, \"PBServer.PutAppend\", &args, &reply)\n\t\tif ok {\n\t\t\treturn \n\t\t}\n\t\ttime.Sleep(viewservice.PingInterval)\n\t\tck.UpdateView()\n\t}\n}", "func (m *Medium) Add(t Topic) {\n\tm.mx.Lock()\n\tm.topics[t.Name()] = t\n\tm.mx.Unlock()\n}", "func (r *Raft) AppendToLog_Leader(cmd []byte) {\n\tTerm := r.myCV.CurrentTerm\n\tlogVal := LogVal{Term, cmd, 0} //make object for log's value field with acks set to 0\n\tr.MyLog = append(r.MyLog, logVal)\n\t//modify metaData after appending\n\tLastLogIndex := r.MyMetaData.LastLogIndex + 1\n\tr.MyMetaData.PrevLogIndex = r.MyMetaData.LastLogIndex\n\tr.MyMetaData.LastLogIndex = LastLogIndex\n\tif len(r.MyLog) == 1 {\n\t\tr.MyMetaData.PrevLogTerm = r.MyMetaData.PrevLogTerm + 1 //as for empty log PrevLogTerm is -2\n\n\t} else if len(r.MyLog) > 1 { //explicit check, else would have sufficed too, just to eliminate len=0 possibility\n\t\tr.MyMetaData.PrevLogTerm = r.MyLog[r.MyMetaData.PrevLogIndex].Term\n\t}\n\tr.setNextIndex_All() //Added-28 march for LogRepair\n\tr.WriteLogToDisk()\n\n}", "func (wal *seriesWAL) Append(metricID uint32, tagsHash uint64, seriesID uint32) (err error) {\n\tif err := wal.base.checkPage(seriesEntryLength); err != nil {\n\t\treturn err\n\t}\n\twal.base.putUint32(metricID)\n\twal.base.putUint64(tagsHash)\n\twal.base.putUint32(seriesID)\n\n\treturn nil\n}", "func (s *SharedLog_) Append(data []byte) (LogEntry_, error) {\n\tmutex.Lock()\n\tlog := LogEntry_{r.currentTerm, s.LsnLogToBeAdded, data, false}\n\ts.Entries = append(s.Entries, log)\n\ts.LsnLogToBeAdded++\n\tmutex.Unlock()\n\treturn log, nil\n}", "func (r *Raft) sendAppend(to uint64) bool {\n\tprevIndex := r.Prs[to].Next - 1\n\tprevLogTerm, err := r.RaftLog.Term(prevIndex)\n\tif err != nil {\n\t\tif err == ErrCompacted {\n\t\t\tr.sendSnapshot(to)\n\t\t\treturn false\n\t\t}\n\t\tpanic(err)\n\t}\n\tvar entries []*pb.Entry\n\tn := len(r.RaftLog.entries)\n\tfor i := r.RaftLog.toSliceIndex(prevIndex + 1); i < n; i++ {\n\t\tentries = append(entries, &r.RaftLog.entries[i])\n\t}\n\tmsg := pb.Message{\n\t\tMsgType: pb.MessageType_MsgAppend,\n\t\tFrom: r.id,\n\t\tTo: to,\n\t\tTerm: r.Term,\n\t\tCommit: r.RaftLog.committed,\n\t\tLogTerm: prevLogTerm,\n\t\tIndex: prevIndex,\n\t\tEntries: entries,\n\t}\n\tr.msgs = append(r.msgs, msg)\n\treturn true\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\tDPrintf(\"Clert PutAppend %s, %v\", key, value)\n\targs := PutAppendArgs{RequestId: ck.genMsgId(), PreviousId: ck.previousId, Key: key, Value: value, Op: op}\n\n\tfor ; ; ck.leaderId = (ck.leaderId + 1) % len(ck.servers) {\n\t\tserver := ck.servers[ck.leaderId]\n\t\treply := GetReply{}\n\t\tok := server.Call(\"KVServer.PutAppend\", &args, &reply)\n\t\tif ok && reply.Err == OK {\n\t\t\tDPrintf(\"!!! Clert PutAppend %s, %v DONE\", key, value)\n\t\t\tck.previousId = args.RequestId\n\t\t\treturn\n\t\t}\n\t}\n}", "func (c *Cache) appendEntries(topic, key string, entries Entries, new bool) error {\n\tt, ok := c.topics.Load(topic)\n\tif !ok {\n\t\treturn errors.New(\"Topic does not exist\")\n\t}\n\ttop := t.(*Topic)\n\n\tp, ok := top.partitions.Load(key)\n\tif !ok {\n\t\tnewPart, err := c.newPartition(topic, key)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\ttop.partitions.Store(key, newPart)\n\t\tp = newPart\n\t\tc.router.Update(topic, key, AddPartition)\n\t}\n\tpartition := p.(*Partition)\n\n\tpartition.mu.Lock()\n\tdefer partition.mu.Unlock()\n\tfpos := partition.clog.Tell()\n\n\tvar (\n\t\tfirstAppend *int\n\t\tlastEntry *Entry\n\t\tlastTime time.Time\n\t)\n\tif len(partition.entries) > 0 {\n\t\tlastEntry = partition.entries[len(partition.entries)-1]\n\t\tlastTime = lastEntry.Timestamp\n\t}\n\n\tfor i, entry := range entries {\n\t\tif entry.Timestamp.IsZero() {\n\t\t\t// maybe we want to error out in some cases in the future.\n\t\t\tentry.Timestamp = time.Now()\n\t\t\tif entry.Timestamp.Equal(lastTime) {\n\t\t\t\t// make sure it is unique (in some platform like play.golang.org,\n\t\t\t\t// time.Now() is second-precision)\n\t\t\t\tentry.Timestamp = entry.Timestamp.Add(time.Duration(1))\n\t\t\t}\n\t\t\tlastTime = entry.Timestamp\n\t\t}\n\n\t\tif i > 0 {\n\t\t\tlastEntry = entries[i-1]\n\t\t}\n\t\t// the behavior is to discard the entries that are before the latest\n\t\t// entry in the partition. if other entries in the request are after\n\t\t// though, they are still appended.\n\t\tif lastEntry != nil && entry.Timestamp.Before(lastEntry.Timestamp) {\n\t\t\tcontinue\n\t\t}\n\t\tif firstAppend == nil {\n\t\t\ttmp := i\n\t\t\tfirstAppend = &tmp\n\t\t}\n\t\tif new {\n\t\t\tif err := partition.clog.Append(&commitlog.Entry{\n\t\t\t\tTimestamp: entry.Timestamp,\n\t\t\t\tData: entry.Data}); err != nil {\n\t\t\t\tlog.Error(\"Failed to persist %v: %v\", entry, err)\n\t\t\t\tpartition.clog.Truncate(fpos)\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\tif firstAppend != nil {\n\t\tpartition.entries = append(partition.entries, entries[*firstAppend:]...)\n\t} else {\n\t\treturn errors.New(\"Nothing new to append\")\n\t}\n\n\tc.LastCommit = CacheCommit{\n\t\tKey: fmt.Sprintf(\"%v_%v\", topic, key),\n\t\tTimestamp: entries[entries.Len()-1].Timestamp,\n\t}\n\treturn nil\n}", "func (r *Raft) AppendToLog_Follower(request AppendEntriesReq) {\n\tterm := request.term\n\tcmd := request.entries\n\tindex := request.prevLogIndex + 1\n\tlogVal := LogVal{term, cmd, 0} //make object for log's value field\n\n\tif len(r.myLog) == index {\n\t\tr.myLog = append(r.myLog, logVal) //when trying to add a new entry\n\t} else {\n\t\tr.myLog[index] = logVal //overwriting in case of log repair\n\t\t//fmt.Println(\"Overwiriting!!\")\n\t}\n\t//fmt.Println(r.myId(), \"Append to log\", string(cmd))\n\t//modify metadata after appending\n\t//r.myMetaData.lastLogIndex = r.myMetaData.lastLogIndex + 1\n\t//r.myMetaData.prevLogIndex = r.myMetaData.lastLogIndex\n\t//\tif len(r.myLog) == 1 {\n\t//\t\tr.myMetaData.prevLogTerm = r.myMetaData.prevLogTerm + 1\n\t//\t} else if len(r.myLog) > 1 {\n\t//\t\tr.myMetaData.prevLogTerm = r.myLog[r.myMetaData.prevLogIndex].Term\n\t//\t}\n\n\t//Changed on 4th april, above is wrong in case of overwriting of log\n\tr.myMetaData.lastLogIndex = index\n\tr.myMetaData.prevLogIndex = index - 1\n\tif index == 0 {\n\t\tr.myMetaData.prevLogTerm = r.myMetaData.prevLogTerm + 1 //or simple -1\n\t} else if index >= 1 {\n\t\tr.myMetaData.prevLogTerm = r.myLog[index-1].Term\n\t}\n\n\t//Update commit index\n\tleaderCI := float64(request.leaderCommitIndex)\n\tmyLI := float64(r.myMetaData.lastLogIndex)\n\tif request.leaderCommitIndex > r.myMetaData.commitIndex {\n\t\tif myLI == -1 { //REDUNDANT since Append to log will make sure it is never -1,also must not copy higher CI if self LI is -1\n\t\t\tr.myMetaData.commitIndex = int(leaderCI)\n\t\t} else {\n\t\t\tr.myMetaData.commitIndex = int(math.Min(leaderCI, myLI))\n\t\t}\n\t}\n\t//fmt.Println(r.myId(), \"My CI is:\", r.myMetaData.commitIndex)\n\tr.WriteLogToDisk()\n}", "func (m *MemoryLogger) Append(newEntry LogEntry) {\n\tm.mutex.Lock()\n\tdefer m.mutex.Unlock()\n\tm.Entries[m.index] = newEntry\n\tm.index = (m.index + 1) % maxLogItems\n}", "func (h *HistoricalRecords) Append(tr *TransferRecord) {\n\th.mutex.Lock()\n\th.records = append(h.records, tr)\n\th.mutex.Unlock()\n}", "func (list *WhoWasList) Append(whowas WhoWas) {\n\tlist.accessMutex.Lock()\n\tdefer list.accessMutex.Unlock()\n\n\tif len(list.buffer) == 0 {\n\t\treturn\n\t}\n\n\tvar pos int\n\tif list.start == -1 { // empty\n\t\tpos = 0\n\t\tlist.start = 0\n\t\tlist.end = 1\n\t} else if list.start != list.end { // partially full\n\t\tpos = list.end\n\t\tlist.end = (list.end + 1) % len(list.buffer)\n\t} else if list.start == list.end { // full\n\t\tpos = list.end\n\t\tlist.end = (list.end + 1) % len(list.buffer)\n\t\tlist.start = list.end // advance start as well, overwriting first entry\n\t}\n\n\tlist.buffer[pos] = whowas\n}", "func (it *eventStorage) Append(event *spec.Event) (*spec.Event, error) {\n\ttopic := event.GetTopic()\n\teventToStore := &spec.Event{Id: it.ids.Next(), Topic: topic, Payload: event.GetPayload()}\n\n\tserializedEvent, err := serializeEvent(eventToStore)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbatch := new(leveldb.Batch)\n\tbatch.Put(it.keys.Event(eventToStore), serializedEvent)\n\n\terr = it.db.Write(batch, nil)\n\tif err != nil {\n\t\treturn nil, errors.Internal.New(\"could not write event: [%v, %v]\", topic, eventToStore)\n\t}\n\n\treturn eventToStore, nil\n}", "func (r *Raft) sendAppend(to uint64) bool {\n\t//\tappend entry\n\tlastIndex := r.RaftLog.LastIndex()\n\tprs := r.Prs[to]\n\tmatched := prs.Match\n\t//if matched < lastIndex {\n\tmsg := r.buildMsgWithoutData(pb.MessageType_MsgAppend, to, false)\n\tvar position int\n\t// send empty append,update follower committed index\n\tif matched == r.RaftLog.LastIndex() {\n\t\tposition = len(r.RaftLog.entries)\n\t} else {\n\t\tp, found := r.RaftLog.findByIndex(matched + 1)\n\t\tif !found {\n\t\t\tpanic(\"not found matched index\")\n\t\t}\n\t\tposition = p\n\t}\n\n\tmsg.Entries = entryValuesToPoints(r.RaftLog.entries[position:])\n\tmsg.Index = prs.Match\n\tt, err := r.RaftLog.Term(prs.Match)\n\tif err != nil {\n\t\tpanic(\"error \")\n\t}\n\tmsg.LogTerm = t\n\tmsg.Commit = r.RaftLog.committed\n\tr.appendMsg(msg)\n\t//update prs\n\tr.Prs[to] = &Progress{\n\t\tMatch: prs.Match,\n\t\tNext: lastIndex + 1,\n\t}\n\treturn true\n\t//}\n\t// Your Code Here (2A).\n\t//return false\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\t// You will have to modify this function.\n\n\treqID := atomic.AddInt32(&ck.nextReqID, 1)\n\n\targs := PutAppendArgs{Key: key, Value: value, Op: op, RequestID: reqID, ClientID: ck.id, PrevIndex: -1}\n\treply := PutAppendReply{Index: -1}\n\n\tck.logger.Debugf(\"Put Append calls start, key %s, value %s\", args.Key, args.Value)\n\n\tfor {\n\t\tfor _, server := range ck.servers {\n\n\t\t\t//ck.logger.Debugf(\"Put Append calls start: server %d, key %s, value %s\", index, args.Key, args.Value)\n\t\t\tok := server.Call(\"KVServer.PutAppend\", &args, &reply)\n\t\t\tif ok {\n\t\t\t\tif reply.Err == OK {\n\t\t\t\t\tck.logger.Debugf(\"Put Append End Successfully : %s\", value)\n\t\t\t\t\treturn\n\t\t\t\t} else {\n\t\t\t\t\t//ck.logger.Debugf(\"Put Append Failed : Wrong leader %d\", index)\n\t\t\t\t\targs.PrevIndex = reply.Index\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t//ck.logger.Debugf(\"PutAppend RPC failed: %d\", index)\n\t\t\t}\n\t\t}\n\n\t}\n\n\n\n\n}", "func (r *RollingPolicy) Append(val float64) {\n\tr.add(r.window.Append, val)\n}", "func (r *RollingPolicy) Append(val float64) {\n\tr.add(r.window.Append, val)\n}", "func (tt *Index) Push(recs ...*types.Log) error {\n\tfor _, rec := range recs {\n\t\tif len(rec.Topics) > MaxCount {\n\t\t\treturn ErrTooManyTopics\n\t\t}\n\t\tcount := posToBytes(uint8(1 + len(rec.Topics)))\n\n\t\tid := NewID(rec.BlockNumber, rec.TxHash, rec.Index)\n\n\t\tvar pos int\n\t\tpush := func(topic common.Hash) error {\n\t\t\tkey := topicKey(topic, uint8(pos), id)\n\t\t\terr := tt.table.Topic.Put(key, count)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tkey = otherKey(id, uint8(pos))\n\t\t\terr = tt.table.Other.Put(key, topic.Bytes())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tpos++\n\t\t\treturn nil\n\t\t}\n\n\t\tif err := push(rec.Address.Hash()); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfor _, topic := range rec.Topics {\n\t\t\tif err := push(topic); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t\tbuf := make([]byte, 0, common.HashLength+len(rec.Data))\n\t\tbuf = append(buf, rec.BlockHash.Bytes()...)\n\t\tbuf = append(buf, rec.Data...)\n\n\t\terr := tt.table.Logrec.Put(id.Bytes(), buf)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (w WatWorkspace) Append(name, contents string) error {\n\tpath := filepath.Join(w.root, kWatDirName, name)\n\tfile, err := os.OpenFile(path, os.O_APPEND|os.O_CREATE|os.O_WRONLY, permFile)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"os.OpenFile: %v\", err)\n\t}\n\tdefer file.Close()\n\n\t_, err = file.WriteString(contents)\n\treturn err\n\n}", "func (tr *Repository) Append(ev es.Event) {\n\ttr.DB.Append(ev)\n}", "func (r *Raft) AppendToLog_Follower(request AppendEntriesReq) {\n\tTerm := request.LeaderLastLogTerm\n\tcmd := request.Entries\n\tindex := request.PrevLogIndex + 1\n\tlogVal := LogVal{Term, cmd, 0} //make object for log's value field\n\n\tif len(r.MyLog) == index {\n\t\tr.MyLog = append(r.MyLog, logVal) //when trying to add a new entry\n\t} else {\n\t\tr.MyLog[index] = logVal //overwriting in case of log repair\n\t}\n\n\tr.MyMetaData.LastLogIndex = index\n\tr.MyMetaData.PrevLogIndex = index - 1\n\tif index == 0 {\n\t\tr.MyMetaData.PrevLogTerm = r.MyMetaData.PrevLogTerm + 1 //or simple -1\n\t} else if index >= 1 {\n\t\tr.MyMetaData.PrevLogTerm = r.MyLog[index-1].Term\n\t}\n\tleaderCI := float64(request.LeaderCommitIndex) //Update commit index\n\tmyLI := float64(r.MyMetaData.LastLogIndex)\n\tif request.LeaderCommitIndex > r.MyMetaData.CommitIndex {\n\t\tr.MyMetaData.CommitIndex = int(math.Min(leaderCI, myLI))\n\t}\n\tr.WriteLogToDisk()\n}", "func (t *QueueManager) Append(s *model.Sample) error {\n\tsnew := *s\n\tsnew.Metric = s.Metric.Clone()\n\n\tb := labels.NewBuilder(t.externalLabels)\n\tfor k, v := range s.Metric {\n\t\tif !t.externalLabels.Has(string(k)) {\n\t\t\tb.Set(string(k), string(v))\n\t\t}\n\t}\n\n\tls := relabel.Process(b.Labels(), t.relabelConfigs...)\n\n\t// If there are no labels; don't queue the sample\n\tif len(ls) < 1 {\n\t\treturn nil\n\t}\n\n\tsnew.Metric = make(model.Metric, len(ls))\n\tfor _, label := range ls {\n\t\tsnew.Metric[model.LabelName(label.Name)] = model.LabelValue(label.Value)\n\t}\n\n\tif snew.Metric == nil {\n\t\treturn nil\n\t}\n\n\tt.shardsMtx.RLock()\n\tenqueued := t.shards.enqueue(&snew)\n\tt.shardsMtx.RUnlock()\n\n\tif enqueued {\n\t\tqueueLength.WithLabelValues(t.queueName).Inc()\n\t} else {\n\t\tdroppedSamplesTotal.WithLabelValues(t.queueName).Inc()\n\t\tif t.logLimiter.Allow() {\n\t\t\tlevel.Warn(t.logger).Log(\"msg\", \"Remote storage queue full, discarding sample. Multiple subsequent messages of this kind may be suppressed.\")\n\t\t}\n\t}\n\treturn nil\n}", "func (kv *DisKV) PutAppend(args *PutAppendArgs, reply *PutAppendReply) error {\n\t// Your code here.\n\t// fmt.Printf(\"PUTAPPEND %d %t --- %d\\n\", kv.me, kv.recovery, kv.Seq+1)\n\tif !kv.recovery {\n\t\tkv.mu.Lock()\n\t\tdefer kv.mu.Unlock()\n\n\t\tvar op Op\n\t\tif args.Op == \"Put\" {\n\t\t\top = Op{OpID: args.OpID, Type: \"Put\", Key: args.Key, Value: args.Value,\n\t\t\t\tCk: args.Ck, LastID: args.LastID}\n\t\t} else {\n\t\t\top = Op{OpID: args.OpID, Type: \"Append\", Key: args.Key, Value: args.Value,\n\t\t\t\tCk: args.Ck, LastID: args.LastID}\n\t\t}\n\t\treply.Err = kv.handleReq(args.OpID, key2shard(args.Key), op)\n\t}\n\n\treturn nil\n}", "func (w *BytesWriter) Append(args ...interface{}) {\r\n\tw.args = append(w.args, args...)\r\n}", "func (w *Writer) Append(msg *Message) error {\n\tw.mu.Lock()\n\t// slow but correct: wait for https://github.com/golang/go/issues/14939\n\tdefer w.mu.Unlock()\n\n\tif w.err != nil { // skip if an error already happens\n\t\treturn w.err\n\t}\n\tif len(msg.Key) > math.MaxInt8 {\n\t\treturn errors.New(\"key is too long\")\n\t}\n\tif len(msg.Value) > math.MaxInt32 {\n\t\treturn errors.New(\"value is too long\")\n\t}\n\tmsg.Offset = w.offset\n\tnumWritten, err := WriteMessage(w.w, w.msgBuf, msg)\n\tw.fileLen += int(numWritten)\n\tif err != nil {\n\t\tw.err = err\n\t\treturn err\n\t}\n\tw.offset++\n\tif w.fileLen >= w.SegmentSize {\n\t\tif err := w.closeFile(); err != nil {\n\t\t\tw.err = err\n\t\t\treturn err\n\t\t}\n\t\tvar err error\n\t\tw.file, err = openOrCreate(journalFileName(w.dir, w.offset))\n\t\tif err != nil {\n\t\t\tw.err = err\n\t\t\treturn err\n\t\t}\n\t\tw.fileLen = 0\n\t\tw.w = newBufferWriter(w.file)\n\t}\n\treturn nil\n}", "func sendAppendEntries(s *Sailor, peer string) error {\n\tam := appendMessage{}\n\tam.Term = s.currentTerm\n\tam.LeaderId = s.client.NodeName\n\tam.PrevLogIndex = s.leader.nextIndex[peer] - 1\n\t// This is just some fancy logic to check for the bounds on the log\n\t// e.g. our log has 0 entries, so the prevEntryTerm cannot be pulled from the log\n\tif len(s.log) == 0 {\n\t\tam.PrevLogTerm = 0\n\t\tam.Entries = nil\n\t} else {\n\t\t// If our log is too short to have prevTerm, use 0\n\t\tif int(s.leader.nextIndex[peer])-2 < 0 {\n\t\t\tam.PrevLogTerm = 0\n\t\t} else {\n\t\t\tam.PrevLogTerm = s.log[s.leader.nextIndex[peer]-2].Term\n\t\t}\n\t\t// If our nextIndex is a value we don't have yet, send nothing\n\t\tif s.leader.nextIndex[peer] > uint(len(s.log)) {\n\t\t\tam.Entries = []entry{}\n\t\t} else {\n\t\t\tam.Entries = s.log[s.leader.nextIndex[peer]-1:]\n\t\t}\n\t}\n\n\tam.LeaderCommit = s.volatile.commitIndex\n\tap := messages.Message{}\n\tap.Type = \"appendEntries\"\n\tap.ID = 0\n\tap.Source = s.client.NodeName\n\tap.Value = makePayload(am)\n\treturn s.client.SendToPeer(ap, peer)\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\t// You will have to modify this function.\n\t//lock\n\tck.snum = ck.snum + 1\n\treply := PutAppendReply{}\n\targs := &PutAppendArgs{Key: key, Value: value, Op: op, Snum: ck.snum, Cid: ck.me}\n\n\t//leaderTry := ck.lastLeader\n\tfor {\n\t\t//log.Printf(\"Leaderid: %d\", ck.lastLeader)\n\t\treply = PutAppendReply{}\n\n\t\tok := ck.servers[ck.lastLeader].Call(\"KVServer.PutAppend\", args, &reply)\n\t\t//WrongLeader bool\n\t\t//Err Err\n\t\tDPrintf4(\"Cli %d PutAppend %v to server %d. Get %v, %v\", ck.me, op, ck.lastLeader, ok, reply.WrongLeader)\n\t\tif !ok || reply.WrongLeader {\n\t\t\tck.lastLeader = (ck.lastLeader + 1) % len(ck.servers) //REM: random assign\n\t\t\tcontinue\n\t\t}\n\n\t\tif !reply.WrongLeader {\n\t\t\t//log.Printf(\"Success Leaderid: %d\", ck.lastLeader)\n\t\t\t//REM: possible reply.leader op in paper\n\t\t\t//ck.lastLeader = leaderTry\n\t\t\tbreak\n\t\t}\n\n\t}\n\t//WrongLeader bool\n\t//Err Err\n\treturn\n}", "func (l *LevelDBLog) Append(value []byte) error {\n\tl.Lock()\n\tdefer l.Unlock()\n\n\tif l.db == nil {\n\t\treturn errors.New(\"log database has been closed \")\n\t}\n\n\t// Create and dump entry to bytes, incrementing last index\n\tl.lastIdx++\n\tentry := &Entry{\n\t\tIndex: l.lastIdx,\n\t\tValue: value,\n\t}\n\tdata, err := entry.Dump()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Put the entry into the database\n\tkey := l.makeKey(entry.Index)\n\treturn l.db.Put(key, data, nil)\n}", "func (s *StickyMemberMetadata) AppendTo(dst []byte) []byte {\n\tdst = kbin.AppendArrayLen(dst, len(s.CurrentAssignment))\n\tfor _, assignment := range s.CurrentAssignment {\n\t\tdst = kbin.AppendString(dst, assignment.Topic)\n\t\tdst = kbin.AppendArrayLen(dst, len(assignment.Partitions))\n\t\tfor _, partition := range assignment.Partitions {\n\t\t\tdst = kbin.AppendInt32(dst, partition)\n\t\t}\n\t}\n\tif s.Generation != -1 {\n\t\tdst = kbin.AppendInt32(dst, s.Generation)\n\t}\n\treturn dst\n}", "func (ck *Clerk) Append(key string, value string) {\n\tck.PutAppend(key, value, \"Append\")\n}", "func (ck *Clerk) Append(key string, value string) {\n\tck.PutAppend(key, value, \"Append\")\n}", "func (rbl *RawBytesLog) Append(entry *Entry) error {\n\terr := writeBytesWithLen16(rbl.logFile, entry.Key)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = writeBytesWithLen32(rbl.logFile, entry.Bytes)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\n\t// Your code here.\n\targs := &PutAppendArgs{}\n\targs.Forward = true\n\targs.Key = key\n\targs.Value = value\n\targs.Viewnum = ck.cur_view.Viewnum\n\t// args.Seq = hex.EncodeToString(Hash(key, value, op, time.Now()))\n\targs.Seq = hex.EncodeToString(Hash(key, value, op))\n\tswitch op {\n\tcase \"Put\":\n\t\targs.Op = 0\n\tcase \"Append\":\n\t\targs.Op = 1\n\t}\n\tfor {\n\t\t// log.Println(ck.cur_view.Primary, ck.cur_view.Backup)\n\t\trsp := &PutAppendReply{}\n\t\tok := call(ck.cur_view.Primary, \"PBServer.PutAppend\", args, rsp)\n\t\tif !ok {\n\t\t\t// log.Printf(\"Client PutAppend rpc to %s failed, Key:%s, Value:%s\\n\", ck.cur_view.Primary, key, value)\n\t\t} else if rsp.Err != \"\" {\n\t\t\t// log.Printf(\"Client PutAppend rpc to %s error, Key:%s, Value:%s\\n\", ck.cur_view.Primary, key, value)\n\t\t\t// log.Println(rsp.Err, ck.cur_view.Primary)\n\t\t\tif rsp.Err == ForwardFailed {\n\t\t\t\tck.UpdateView()\n\t\t\t\tif ck.cur_view.Backup == \"\" {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t} else if rsp.Err == DuplicateRequest {\n\t\t\t\tbreak\n\t\t\t}\n\t\t} else {\n\t\t\t// log.Printf(\"Client PutAppend rpc to %s done! Key:%s, Value:%s\\n\", ck.cur_view.Primary, key, value)\n\t\t\tbreak\n\t\t}\n\t\tif ck.UpdateView() != nil { //check if primary has changed\n\t\t\treturn\n\t\t}\n\t\ttime.Sleep(viewservice.PingInterval)\n\t}\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\t// You will have to modify this function.\n\targs := PutAppendArgs{key, value, op}\n\tflag := false\n\tfor {\n\t\tfor i := range ck.servers {\n\t\t\treply := PutAppendReply{}\n\t\t\tok := ck.servers[i].Call(\"KVServer.PutAppend\", &args, &reply)\n\t\t\tif ok == true && reply.Err == OK {\n\t\t\t\tDPrintf(\"PutAppend succ. key:%v, value:%v, op:%v\", key, value, op)\n\t\t\t\tflag = true\n\t\t\t\tbreak\n\t\t\t} else {\n\t\t\t\t//DPrintf(\"PutAppend fail! key:%v value:%v op:%v ret:%v\", key, value, op, reply.Err)\n\t\t\t}\n\t\t}\n\t\tif true == flag {\n\t\t\tbreak\n\t\t}\n\t\ttime.Sleep(1000 * time.Millisecond)\n\t}\n\n}", "func (rf *Raft) StartAppendLog() {\n\tvar count int32 = 1\n\tfor i, _ := range rf.peers {\n\t\tif i == rf.me {\n\t\t\tcontinue\n\t\t}\n\t\tgo func(i int) {\n\t\t\tfor{\n\t\t\t\trf.mu.Lock()\n\t\t\t\t//fmt.Printf(\"follower %d lastlogindex: %v, nextIndex: %v\\n\",i, rf.GetPrevLogIndex(i), rf.nextIndex[i])\n\t\t\t\t//fmt.Print(\"sending log entries from leader %d to peer %d for term %d\\n\", rf.me, i, rf.currentTerm)\n\t\t\t\t//fmt.Print(\"nextIndex:%d\\n\", rf.nextIndex[i])\n\t\t\t\tif rf.state != Leader {\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\targs := AppendEntriesArgs{\n\t\t\t\t\tTerm: rf.currentTerm,\n\t\t\t\t\tLeaderId: rf.me,\n\t\t\t\t\tPrevLogIndex: rf.GetPrevLogIndex(i),\n\t\t\t\t\tPrevLogTerm: rf.GetPrevLogTerm(i),\n\t\t\t\t\tEntries: append(make([]LogEntry, 0), rf.logEntries[rf.nextIndex[i]:]...),\n\t\t\t\t\tLeaderCommit: rf.commitIndex,\n\t\t\t\t}\n\t\t\t\treply := AppendEntriesReply{}\n\t\t\t\trf.mu.Unlock()\n\t\t\t\tok := rf.sendAppendEntries(i, &args, &reply)\n\n\t\t\t\tif !ok {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\trf.mu.Lock()\n\t\t\t\tif rf.state != Leader {\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif reply.Term > rf.currentTerm {\n\t\t\t\t\trf.BeFollower(reply.Term)\n\t\t\t\t\tgo func() {\n\t\t\t\t\t\tsend(rf.appendEntry)\n\t\t\t\t\t}()\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif reply.Success {\n\t\t\t\t\trf.matchIndex[i] = args.PrevLogIndex + len(args.Entries)\n\t\t\t\t\trf.nextIndex[i] = rf.matchIndex[i] + 1\n\t\t\t\t\t//fmt.Print(\"leader: %v, for peer %v, match index: %d, next index: %d, peers: %d\\n\", rf.me, i, rf.matchIndex[i], rf.nextIndex[i], len(rf.peers))\n\t\t\t\t\tatomic.AddInt32(&count, 1)\n\t\t\t\t\tif atomic.LoadInt32(&count) > int32(len(rf.peers)/2) {\n\t\t\t\t\t\t//fmt.Print(\"leader %d reach agreement\\n, args.prevlogindex:%d, len:%d\\n\", rf.me, args.PrevLogIndex, len(args.Entries))\n\t\t\t\t\t\trf.UpdateCommitIndex()\n\t\t\t\t\t}\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\treturn\n\t\t\t\t} else {\n\t\t\t\t\t//fmt.Printf(\"peer %d reset the next index from %d to %d\\n\", i, rf.nextIndex[i], rf.nextIndex[i]-1)\n\t\t\t\t\tif rf.nextIndex[i] > 0 {\n\t\t\t\t\t\trf.nextIndex[i]--\n\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t} else {\n\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\ttime.Sleep(10 * time.Millisecond)\n\t\t\t}\n\t\t}(i)\n\t}\n\n}", "func (t Tee) Append(s *clientmodel.Sample) {\n\tt.Appender1.Append(s)\n\tt.Appender2.Append(s)\n}", "func (thread *Thread) OnAppend(ctx aero.Context, key string, index int, obj interface{}) {\n\tonAppend(thread, ctx, key, index, obj)\n}", "func (s *storage) appendEntry(e *entry) {\n\tassert(e.index == s.lastLogIndex+1)\n\tw := new(bytes.Buffer)\n\tif err := e.encode(w); err != nil {\n\t\tpanic(bug{fmt.Sprintf(\"entry.encode(%d)\", e.index), err})\n\t}\n\tif err := s.log.Append(w.Bytes()); err != nil {\n\t\tpanic(opError(err, \"Log.Append\"))\n\t}\n\ts.lastLogIndex, s.lastLogTerm = e.index, e.term\n}", "func (w *Writer) append(key, value []byte, restart bool) {\n\tnShared := 0\n\tif restart {\n\t\tw.restarts = append(w.restarts, uint32(w.buf.Len()))\n\t} else {\n\t\tnShared = db.SharedPrefixLen(w.prevKey, key)\n\t}\n\tw.prevKey = append(w.prevKey[:0], key...)\n\tw.nEntries++\n\tn := binary.PutUvarint(w.tmp[0:], uint64(nShared))\n\tn += binary.PutUvarint(w.tmp[n:], uint64(len(key)-nShared))\n\tn += binary.PutUvarint(w.tmp[n:], uint64(len(value)))\n\tw.buf.Write(w.tmp[:n])\n\tw.buf.Write(key[nShared:])\n\tw.buf.Write(value)\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\t// You will have to modify this function.\n\targs := &PutAppendArgs{\n\t\tKey: key,\n\t\tValue: value,\n\t\tOp: op,\n\t\tCid: ck.cid,\n\t\tSeqNum: ck.seqNum,\n\t}\n\tidx := ck.lastLeaderId\n\tfor {\n\t\treply := &PutAppendReply{}\n\t\tck.mu.Lock()\n\t\tck.seqNum++\n\t\tck.mu.Unlock()\n\t\t//DPrintf(\"Client request server %v to call PutAppend method, args: %+v\", idx, args)\n\t\tok := ck.servers[idx].Call(\"KVServer.PutAppend\", args, reply)\n\t\t//DPrintf(\"Client request server %v to call PutAppend method finished, reply: %+v\", idx, reply)\n\t\tif ok && !reply.WrongLeader {\n\t\t\t// todo how to deal with ERR\n\t\t\tck.lastLeaderId = idx\n\t\t\t//DPrintf(\"===================Client request server %v to call PutAppend method success!\", idx)\n\t\t\treturn\n\t\t}\n\t\tidx = (idx + 1)%len(ck.servers)\n\t\ttime.Sleep(time.Duration(50)*time.Millisecond)\n\t}\n\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\tck.RunCall(\"RaftKV.PutAppend\",\n\t\tfunc(callerId CallerId) interface{} {\n\t\t\targs := new(PutAppendArgs)\n\t\t\targs.CID = callerId\n\t\t\targs.Op = op\n\t\t\targs.Key = key\n\t\t\targs.Value = value\n\t\t\treturn args\n\t\t},\n\t\tfunc() interface{} {\n\t\t\treturn new(PutAppendReply)\n\t\t},\n\t\tfunc(reply interface{}) (string, Err) {\n\t\t\treturn \"\", reply.(*PutAppendReply).Err\n\t\t})\n}", "func (s *BytesWriter) Append(args ...interface{}) {\n\ts.args = append(s.args, args...)\n}", "func (s *BytesWriter) Append(args ...interface{}) {\n\ts.args = append(s.args, args...)\n}", "func (rf *Raft) AppendEntries(args *AppendEntriesArgs, reply *AppendEntriesReply) {\n\trf.mu.Lock()\n\tdefer rf.mu.Unlock()\n\t// Resetting as we received a heart beat.\n\trf.resetElectionTimer()\n\trf.debug( \"AppendEntries: from LEADER %#v \\n\",args)\n\trf.debug(\"My current state: %#v \\n\", rf)\n\t//1. Reply false if term < currentTerm (§5.1)\n\tif args.Term > rf.currentTerm{\n\t\tif rf.currentState != Follower {\n\t\t\trf.transitionToFollower(args.Term)\n\t\t}\n\t}\n\t//2. Reply false if log doesn’t contain an entry at prevLogIndex\n\t//whose term matches prevLogTerm (§5.3)\n\t//3. If an existing entry conflicts with a new one (same index\n\t//but different terms), delete the existing entry and all that\n\t//follow it (§5.3)\n\t//4. Append any new entries not already in the log\n\t//5. If leaderCommit > commitIndex, set commitIndex =\n\t//\tmin(leaderCommit, index of last new entry)\n\t/////////////Pending implementation point 5 above.\n\tif args.Term < rf.currentTerm{\n\t\treply.Success = false\n\t\treply.Term =rf.currentTerm\n\t\treturn\n\t}\n\n\t// Update my term to that of the leaders\n\trf.currentTerm = args.Term\n\trf.debug(\"Dereferencing %d\",len(rf.log)-1)\n\trf.debug(\"Current log contents %v\", rf.log)\n\n\t// Check first whether it is a heartbeat or an actual append entry.\n\t// If it is heartbeat, then just reset the timer and then go back.\n\t//Otherwise, we need to add the entries into the logs of this peer.\n\t// If this is heart beat, then we know that the command is going to be nil.\n\t// Identify this and return.\n\tlastLogEntryIndex := len(rf.log) - 1\n\tif args.LogEntries == nil {\n\t\t//This is heart beat\n\t\treply.Term = rf.currentTerm\n\t\trf.debug(\"Received a HEART BEAT.\")\n\t}else {\n\t\trf.debug(\"Received an APPEND ENTRY. PROCESSING\")\n\t\tlastLogEntry := rf.log[len(rf.log)-1]\n\t\t//1a\n\t\tif lastLogEntryIndex < args.PreviousLogIndex {\n\t\t\treply.Success = false\n\t\t\treply.NextIndex = lastLogEntryIndex\n\t\t\trf.debug(\"1a \\n\")\n\t\t\treturn\n\t\t}\n\t\t//1b\n\t\tif lastLogEntryIndex > args.PreviousLogIndex {\n\t\t\treply.Success = false\n\t\t\trf.debug(\"Last log entry index --> %d, PreviousLogIndex From LEADER -->%d\", lastLogEntryIndex, args.PreviousLogIndex)\n\t\t\trf.log = rf.log[:len(rf.log)-1]\n\t\t\treturn\n\t\t}\n\t\t//3\n\t\tif lastLogEntry.LastLogTerm != args.PreviousLogTerm {\n\t\t\treply.Success = false\n\t\t\t//Reduce size by 1;\n\t\t\trf.debug(\"3 \\n\")\n\t\t\trf.log = rf.log[:len(rf.log)-1]\n\t\t\treturn\n\t\t}\n\n\t\t// 4 We are good to apply the command.\n\t\trf.printSlice(rf.log, \"Before\")\n\t\trf.debug(\"Printing the entry to be added within the handler %v\", args.LogEntries)\n\t\trf.log = append(rf.log, args.LogEntries...)\n\t\trf.printSlice(rf.log, \"After\")\n\t\trf.debug(\"\\n Applied the command to the log. Log size is -->%d \\n\", len(rf.log))\n\t\t//5\n\t}\n\tif args.LeaderCommit >rf.commitIndex {\n\t\trf.debug(\"5 Update commitIndex. LeaderCommit %v rf.commitIndex %v \\n\",args.LeaderCommit,rf.commitIndex )\n\t\t//Check whether all the entries are committed prior to this.\n\t\toldCommitIndex:=rf.commitIndex\n\t\trf.commitIndex = min(args.LeaderCommit,lastLogEntryIndex+1)\n\t\trf.debug(\"moving ci from %v to %v\", oldCommitIndex, rf.commitIndex)\n\t\t//Send all the received entries into the channel\n\t\tj:=0\n\t\tfor i:=oldCommitIndex ;i<args.LeaderCommit;i++ {\n\t\t\trf.debug(\"Committing %v \",i)\n\t\t\tapplyMsg := ApplyMsg{CommandValid: true, Command: rf.log[i].Command, CommandIndex: i}\n\t\t\tj++\n\t\t\trf.debug(\"Sent a response to the end client \")\n\t\t\trf.debug(\"applyMsg %v\",applyMsg)\n\t\t\trf.applyCh <- applyMsg\n\t\t}\n\t}\n\treply.Success = true\n\t//Check at the last. This is because this way the first HB will be sent immediately.\n\t//timer := time.NewTimer(100 * time.Millisecond)\n}", "func AppendCaller() {\n\t\tfor {\n\t\t\tlog_conn := <-Append_ch\n\t\t\tlogentry :=log_conn.Logentry\n\t\t\tconn:=log_conn.Conn\n\t\t\traft.AppendHeartbeat <- 1 // No need to send heartbeat in this cycle, as sending log entires is also treated as heartbeat\n\t\t\tappendAckcount:=1\n\t\t\tsyncNeeded := false\n\t\t\tvar logentry1 LogEntry\n\t\t\tvar args *AppendRPCArgs // Prepare Arguments, \n\t\t\t/*if logentry.SequenceNumber >= 1 {\t // if Log has more than 2 entries\n\t\t\t\targs = &AppendRPCArgs {\n\t\t\t\t\tr.CurrentTerm,\n\t\t\t\t\tr.LeaderId,\n\t\t\t\t\tlogentry.SequenceNumber-1,\n\t\t\t\t\tr.Log[logentry.SequenceNumber-1].Term,\n\t\t\t\t\tlogentry,\n\t\t\t\t\tr.CommitIndex,\n\t\t\t\t}\n\t\t\t} else { \n\t\t\t\targs = &AppendRPCArgs { // if Log has only one entry or no entry\n\t\t\t\t\tr.CurrentTerm,\n\t\t\t\t\tr.LeaderId,\n\t\t\t\t\t0,\n\t\t\t\t\tr.CurrentTerm,\n\t\t\t\t\tlogentry,\n\t\t\t\t\tr.CommitIndex,\n\t\t\t\t}\n\t\t\t}*/\n\n\t\t\t//fmt.Println(\"Append Recieved \",logentry.SequenceNumber)\n\t\t\t\tvar AppendAck_ch = make (chan int,len(r.ClusterConfigV.Servers)-1)\n\t\t\t\tfor _,server := range r.ClusterConfigV.Servers {\t\t\t\n\t\t\t\t\t\tif server.Id != r.Id {\n\t\t\t\t\t\t\tif(logentry.SequenceNumber>r.NextIndex[server.Id]){\n\t\t\t\t\t\t\t\t\tlogentry1 = r.Log[r.NextIndex[server.Id]]\n\t\t\t\t\t\t\t\t\tsyncNeeded=true\n\t\t\t\t\t\t\t\t}else{\n\t\t\t\t\t\t\t\t\tlogentry1 = logentry\t\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\tif logentry1.SequenceNumber >= 1 {\t\n\t\t\t\t\t\t\t\t\t\targs = &AppendRPCArgs {\n\t\t\t\t\t\t\t\t\t\tr.CurrentTerm,\n\t\t\t\t\t\t\t\t\t\tr.LeaderId,\n\t\t\t\t\t\t\t\t\t\tlogentry1.SequenceNumber-1,\n\t\t\t\t\t\t\t\t\t\tr.Log[logentry1.SequenceNumber-1].Term,\n\t\t\t\t\t\t\t\t\t\tlogentry1,\n\t\t\t\t\t\t\t\t\t\tr.CommitIndex,\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t\targs = &AppendRPCArgs {\n\t\t\t\t\t\t\t\t\tr.CurrentTerm,\n\t\t\t\t\t\t\t\t\tr.LeaderId,\n\t\t\t\t\t\t\t\t\t0,\n\t\t\t\t\t\t\t\t\tr.CurrentTerm,\n\t\t\t\t\t\t\t\t\tlogentry1,\n\t\t\t\t\t\t\t\t\tr.CommitIndex,\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\t\t\t\n\t\t\t\t\t\t\tgo r.sendAppendRpc(server,args,AppendAck_ch,false) // to send Log entry to follower \n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\tfor j:=0;j<len(r.ClusterConfigV.Servers)-1;j++{\n\t\t\t\t\t\t\tid:=<- AppendAck_ch \n\t\t\t\t\t\t\tif(id!=-1 && r.MatchIndex[id]==logentry.SequenceNumber){\n\t\t\t\t\t\t\t\tappendAckcount++\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tif appendAckcount > len(r.ClusterConfigV.Servers)/2 { // If we have majority in log , update commit index\n\t\t\t\t\t\t\t\tr.CommitIndex=logentry.SequenceNumber\n\t\t\t\t\t\t\t\tlogentry.IsCommitted=true\n\t\t\t\t\t\t\tbreak\t\t\t\n\t\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\t/*\tmajorCount:=0\n\t\t\t\t\t\tfor _,serverC:= range r.ClusterConfigV.Servers { // Check if log entry is in majority \n\t\t\t\t\t\t\tif serverC.Id !=r.Id && r.MatchIndex[serverC.Id] == logentry.SequenceNumber {\n\t\t\t\t\t\t\t\tmajorityCount++\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t*/\n\t\t\t\t\tif(logentry.IsCommitted==true){ // If log is committed, write it to log, and send log entry for evaluation on input_ch\n\t\t\t\t\t\t//fmt.Println(\"Commited \",logentry.SequenceNumber)\n\t\t\t\t\t\tr.Log[logentry.SequenceNumber].IsCommitted=true\n\t\t\t\t\t\tlogentry.IsCommitted=true\n\t\t\t\t\t\tr.CommitIndex=logentry.SequenceNumber\n\t\t\t\t\t\tInput_ch <- Log_Conn{logentry, conn}\n\t\t\t\t\t\tr.File.WriteString(strconv.Itoa(logentry.Term)+\" \"+strconv.Itoa(logentry.SequenceNumber)+\" \"+strings.TrimSpace(strings.Replace(string(logentry.Command),\"\\n\",\" \",-1))+\" \"+\n\t\t\" \"+strconv.FormatBool(logentry.IsCommitted))\n\t\t\t\t\t\tr.File.WriteString(\"\\t\\r\\n\");\n\t\t\t\t\t} else { \n\t\t\t\t\t \t\t\t//if syncNeeded==true{ // If Log is not commited, call thsi function to Sync all logs, Logs are sync only till current Logentry, not beyong this even if \n\t\t\t\t\t\t\t\t\t// Leader log has go more entries added while executing this\n\t\t\t\t\t\t\t\t\t//fmt.Println(\"Sync call from append\")\n\t\t\t\t\t\t\t\tsyncNeeded=false\n\t\t\t\t\t\t\t\t//fmt.Println(\"Sync Called from Else\")\n\t\t\t\t\t\t\t\tSyncAllLog(Log_Conn{logentry,conn})\t\t\n\n\t\t\t\t\t\t}\n\t\t\t\t\tif syncNeeded==true{ // If Log is is commited, call thsi function to Sync all logs, Logs are sync only till current Logentry, not beyong this even if \n\t\t\t\t\t\t\t\t\t// Leader log has go more entries added while executing this\n\t\t\t\t\t\t\t\t\t//fmt.Println(\"Sync call from append\")\n\t\t\t\t\t\t//\tfmt.Println(\"Sync Called from syncNeeded == True\")\n\t\t\t\t\t\t\tSyncAllLog(Log_Conn{logentry,conn})\n\t\t\t\t\t\t}\t\t\t\t\t\t\t\t\n\t\t\t\t}\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\t// You will have to modify this function.\n\tck.mu.Lock()\n\tserveri := ck.i\n\tck.sequence++\n\tsequence := ck.sequence\n\tclientName := ck.name\n\tck.mu.Unlock()\n\tfor {\n\t\targs := &PutAppendArgs{\n\t\t\tKey: key,\n\t\t\tValue: value,\n\t\t\tOp: op,\n\t\t\tClientName: clientName,\n\t\t\tSequence: sequence,\n\t\t}\n\t\treply := &PutAppendReply{}\n\t\tDPrintf(\"[CLERK] calling KVServer %d.PutAppend [K=%s] [V=%s] [Op=%s] \", serveri, key, value, op)\n\t\tok := ck.servers[serveri].Call(\"KVServer.PutAppend\", args, reply)\n\t\tif !reply.IsLeader || !ok || reply.Err != \"\" {\n\t\t\tif reply.Err != \"\" {\n\t\t\t\tDPrintf(\"[DEBUG] Clerk.PutAppend err:[%s]\", reply.Err)\n\t\t\t}\n\t\t\tDPrintf(\"[CLERK] KVServer.PutAppend sent to nonleader\")\n\t\t\t// try another server\n\t\t\tck.mu.Lock()\n\t\t\tck.i = int(nrand()) % len(ck.servers)\n\t\t\tserveri = ck.i\n\t\t\tck.mu.Unlock()\n\t\t\ttime.Sleep(30 * time.Millisecond)\n\t\t\tcontinue\n\t\t} else {\n\t\t\tDPrintf(\"[CLERK] KVServer.PutAppend return\")\n\t\t\tbreak\n\t\t}\n\t}\n\n}", "func (c *ConcurrentPreviousSet) Append(message types.Message) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\tc.values[message.Identifier] = message\n}", "func (r *Raft) sendAppend(to uint64) bool {\n\tr.sendAppendEntries(to)\n\treturn true\n}", "func (ck *Clerk) PutAppend(key string, value string, op Type) {\n\targs := PutAppendArgs{key, value, op, nrand()}\n\ti := ck.leader\n\tfor {\n\t\tvar reply PutAppendReply\n\t\tDPrintf(\"CLIENT %d TRYING TO CALL PUT/APPEND on server %d for key %s and value %s\\n\", ck.id, i, key, value)\n\t\tdone := make(chan bool)\n\t\tgo func() {\n\t\t\tok := ck.servers[i].Call(\"RaftKV.PutAppend\", &args, &reply)\n\t\t\tif ok {\n\t\t\t\tdone <- true\n\t\t\t}\n\t\t}()\n\t\tselect {\n\t\tcase <-done:\n\t\t\tDPrintf(\"CLIENT %d GOT PUT/APPEND KVRAFT REPLY from server %d for key %s and value %s: %+v\\n\", ck.id, i, key, value, reply)\n\t\t\tif reply.WrongLeader {\n\t\t\t\ti = (i + 1) % len(ck.servers)\n\t\t\t} else if reply.Err == ErrLostAction {\n\t\t\t\t// retry this server because its the leader\n\t\t\t} else if reply.Err == OK {\n\t\t\t\tck.leader = i\n\t\t\t\tDPrintf(\"CLIENT %d SET LEADER TO %d\\n\", ck.id, i)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase <-time.After(500 * time.Millisecond):\n\t\t\tDPrintf(\"CLIENT %d TIMED OUT ON PUT/APPEND KVRAFT REQUEST for server %d and key %s and value %s\", ck.id, i, key, value)\n\t\t\ti = (i + 1) % len(ck.servers)\n\t\t}\n\t}\n}", "func (s *State) Append(group *mesosfile.Group) {\n\ts.pending <- group\n}", "func (ep *EndPoint) push(msg *Chat_Message) {\n\tif len(ep.Inbox) > ep.retention {\n\t\tep.Inbox = append(ep.Inbox[1:], *msg)\n\t\tep.StartOffset++\n\t} else {\n\t\tep.Inbox = append(ep.Inbox, *msg)\n\t}\n\tep.notifyConsumers()\n}", "func (oplog *OpLog) Append(op *Operation) {\n\toplog.append(op, nil)\n}", "func (s *storage) add(m mqtt.Message, messageid uint16) {\n\ts.mux.Lock()\n\ts.sto[messageid] = m\n\ts.mux.Unlock()\n}", "func (l *LevelDB) Append(entries []pb.Entry) error {\n\tbatch := new(leveldb.Batch)\n\tfor _, e := range entries {\n\t\tk := make([]byte, 8)\n\t\tbinary.LittleEndian.PutUint64(k, uint64(e.Index))\n\t\tb, err := proto.Marshal(&e)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbatch.Put(k, b)\n\t}\n\n\treturn l.db.Write(batch, nil)\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\t// You will have to modify this function.\n\targs := PutAppendArgs{\n\t\tClerkId: ck.clerkId,\n\t\tKey: key,\n\t\tValue: value,\n\t\tOp: op,\n\t}\n\tif !ck.reSendFlag {\n\t\tck.opId++\n\t}\n\targs.OpId = ck.opId\n\treply := PutAppendReply{}\n\tok := ck.servers[ck.leader].Call(\"KVServer.PutAppend\", &args, &reply)\n\tif ok {\n\t\tswitch reply.Err {\n\t\tcase OK:\n\t\t\tck.reSendFlag = false\n\t\tcase ErrWrongLeader:\n\t\t\tck.reSendFlag = true\n\t\t\tck.leader = (ck.leader + 1) % len(ck.servers)\n\t\t\tck.PutAppend(key, value, op)\n\t\t}\n\t} else {\n\t\tck.reSendFlag = true\n\t\tck.leader = (ck.leader + 1) % len(ck.servers)\n\t\tck.PutAppend(key, value, op)\n\t}\n}", "func (wal *BaseWAL) Write(msg WALMessage) error {\n\tif wal == nil {\n\t\treturn nil\n\t}\n\n\tif err := wal.enc.Encode(&TimedWALMessage{tmtime.Now(), msg}); err != nil {\n\t\twal.Logger.Error(\"Error writing msg to consensus wal. WARNING: recover may not be possible for the current height\",\n\t\t\t\"err\", err, \"msg\", msg)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (tm *TableManager) Append(i Index, ts *Table) error {\n\tif ts == nil {\n\t\treturn log.Error(\"csdb.TableManager.Init\", \"err\", errgo.Newf(\"Table pointer cannot be nil for Index %d\", i))\n\t}\n\ttm.mu.Lock()\n\ttm.ts[i] = ts\n\ttm.mu.Unlock() // use defer once there are multiple returns\n\treturn nil\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\t// You will have to modify this function.\n\tpaArgs := &PutAppendArgs{}\n\tpaArgs.Key = key\n\tpaArgs.Value = value\n\tpaArgs.Op = op\n\tpaArgs.ClientID = ck.ID\n\tck.mu.Lock()\n\tpaArgs.CommSeq = ck.commSeq\n\tck.commSeq++\n\tck.mu.Unlock()\n\tDPrintf(\"client: %v reqeust pa ID %d %s key:%s value:%s\", ck.ID, paArgs.CommSeq, op, key, value)\n\tvar leaderNum int\n\tfor true {\n\t\tpaReply := &PutAppendReply{}\n\t\tif ck.LeaderIndex != -1 {\n\t\t\tleaderNum = ck.LeaderIndex\n\t\t} else {\n\t\t\tleaderNum = (int(nrand()) % len(ck.servers))\n\t\t}\n\t\tDPrintf(\"client %v: get leaderNum %d\", ck.ID, leaderNum)\n\t\t//ok := ck.servers[(leaderNum+len(ck.servers)-1)%(len(ck.servers))].Call(\"KVServer.PutAppend\", paArgs, paReply)\n\t\tok := ck.servers[leaderNum].Call(\"KVServer.PutAppend\", paArgs, paReply)\n\t\tif paReply.WrongLeader || !ok {\n\t\t\tDPrintf(\"client: %v ask %d but wrong leader\", ck.ID, leaderNum)\n\t\t\tck.LeaderIndex = -1\n\t\t\ttime.Sleep(10 * time.Millisecond)\n\t\t\t//TODO::optimization needed\n\t\t} else if paReply.Err != Err(\"\") {\n\t\t\tDPrintf(\"kv %s: put meets err\", string(paReply.Err))\n\t\t} else {\n\t\t\tck.LeaderIndex = leaderNum\n\t\t\tbreak\n\t\t}\n\t}\n\tDPrintf(\"client: %v end reqeust pa ID %d %s key:%s \", ck.ID, paArgs.CommSeq, op, key)\n}", "func updateLastAppended(s *followerReplication, req *pb.AppendEntriesRequest) {\n\t// Mark any inflight logs as committed\n\tif logs := req.Entries; len(logs) > 0 {\n\t\tlast := logs[len(logs)-1]\n\t\tatomic.StoreUint64(&s.nextIndex, last.Index+1)\n\t\ts.commitment.match(s.peer.ID, last.Index)\n\t}\n\n\t// Notify still leader\n\ts.notifyAll(true)\n}", "func (tm *topicManager) createNewTopic(name string, id string) (log, error) {\n\tlc, ok := tm.topicNamesToLogs[name]\n\tif ok {\n\t\tlc.refCount++\n\t\treturn lc.l, nil\n\t}\n\n\t// lazily create the log if we know about the topic/cc ID\n\tl, err := newBadgerLog(tm.dataDir, name, id, tm.cc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttm.topicNamesToLogs[name] = &logContainer{\n\t\tl: l,\n\t\trefCount: 1,\n\t}\n\treturn l, nil\n}", "func (e *EventStore) AppendTo(ctx context.Context, streamName goengine.StreamName, streamEvents []goengine.Message) error {\n\treturn e.AppendToWithExecer(ctx, e.db, streamName, streamEvents)\n}", "func (batch *BatchEncoder) Append(key []byte, value []byte) {\n\tvar keyLenByte [8]byte\n\tbinary.BigEndian.PutUint64(keyLenByte[:], uint64(len(key)))\n\tvar valueLenByte [8]byte\n\tbinary.BigEndian.PutUint64(valueLenByte[:], uint64(len(value)))\n\n\tbatch.keyBuf.Write(keyLenByte[:])\n\tbatch.keyBuf.Write(key)\n\n\tbatch.valueBuf.Write(valueLenByte[:])\n\tbatch.valueBuf.Write(value)\n}", "func (c *Chain) Append(mw ...Middleware) *Chain {\n\tnc := make([]Middleware, 0, len(c.mw)+len(mw))\n\tnc = append(nc, c.mw...)\n\tnc = append(nc, mw...)\n\n\treturn &Chain{\n\t\tmw: nc,\n\t}\n}", "func (wal *WAL) Write(bufs ...[]byte) error {\n\tif wal.backlog != nil {\n\t\twal.backlog <- bufs\n\t\treturn nil\n\t} else {\n\t\treturn wal.doWrite(bufs...)\n\t}\n}", "func (kv *ShardKV) PutAppend(args *PutAppendArgs, reply *PutAppendReply) error {\n\t// Your code here.\n\t\n\tkv.mu.Lock()\n\tdefer kv.mu.Unlock()\n\n\tdelete(kv.records, args.Prev)\n\n\tDPrintf(\"%s Received %s, Key=%s,Value=%s\",kv.name,args.Op,args.Key,args.Value)\n\tDPrintf(\"ID=%v\",args.ID)\n\t\n\tif record, ok := kv.records[args.ID]; ok && record.Err==OK {\n\t // we have a record associate with this ID\n\t reply.Err = record.Err\n\n\t DPrintf(\"%s found an ID we have seen before,%s key=%s value=%s\",kv.name,args.Op,args.Key,args.Value)\n\t DPrintf(\"Err=%v\", reply.Err)\n\t return nil\n\t}\n\n//\tdelete(kv.records, args.Prev)\n\n\tvar v = Op{}\n\tv.ID = args.ID\n\tv.Type = args.Op\n\tv.Key = args.Key\n\tv.Value = args.Value\n\t\n\tseq := kv.Log(v)\n\tresult := kv.ReadLog(seq)\n\n\treply.Err = result.Err\n\treply.Done = args.ID\n\n\tkv.px.Done(seq)\n\tkv.seq = seq + 1\n\n\tnewRecord := Record{result.Value,result.Err}\n\tkv.records[args.ID] = &newRecord\n\n\treturn nil\n}", "func (c *Client) PutAppend(key string, value string, op string) {\n\n\treq := PutAppendRequest{\n\t\tKey: key,\n\t\tValue: value,\n\t\tClientInfo: ClientInfo{\n\t\t\tUid: c.Uid,\n\t\t\tSeq: atomic.AddInt64(&c.Seq, 1),\n\t\t},\n\t\tOpType: op,\n\t}\n\ts := key2shard(key)\n\tc.info(\"开始PutAppend %+v\", req)\n\tdefer c.info(\"成功PutAppend %+v\", req)\n\tfor {\n\t\tc.mu.Lock()\n\t\tgid := c.conf.Shards[s]\n\t\tservers, ok := c.conf.Groups[gid]\n\t\tc.mu.Unlock()\n\t\tif ok {\n\t\tround:\n\t\t\tfor _, srvi := range servers {\n\t\t\t\tvar resp PutAppendResponse\n\t\t\t\tsrv := c.make_end(srvi)\n\t\t\t\tsrv.Call(\"ShardKV.PutAppend\", &req, &resp)\n\t\t\t\tswitch resp.RPCInfo {\n\t\t\t\tcase SUCCESS:\n\t\t\t\t\treturn\n\t\t\t\tcase DUPLICATE_REQUEST:\n\t\t\t\t\treturn\n\t\t\t\tcase WRONG_GROUP:\n\t\t\t\t\tbreak round\n\t\t\t\tdefault:\n\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\ttime.Sleep(CLIENT_REQUEST_INTERVAL)\n\t\t// ask controler for the latest configuration.\n\t\tconfig := c.scc.Query(-1)\n\t\tc.mu.Lock()\n\t\tc.conf = config\n\t\tc.mu.Unlock()\n\t}\n}", "func (fsm *fsm) applyAppend(b []byte) interface{} {\n\tvar req api.ProduceRequest\n\terr := proto.Unmarshal(b, &req)\n\tif err != nil {\n\t\treturn err\n\t}\n\toffset, err := fsm.log.Append(req.Record)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn &api.ProduceResponse{Offset: offset}\n}", "func (w *fileWAL) append(entry walEntry) (int, error) {\n\n\t// Make sure we have an open WAL.\n\tif w.f == nil {\n\t\treturn 0, errorInvalidWALFile\n\t}\n\n\t// Buffer writes until the end.\n\tbuf := &bytes.Buffer{}\n\n\tvar err error\n\n\t// Write magic number\n\terr = binary.Write(buf, binary.LittleEndian, walMagic)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\t// Write the operation type\n\terr = binary.Write(buf, binary.LittleEndian, entry.operation)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\t// Write the number of rows\n\terr = binary.Write(buf, binary.LittleEndian, uint16(len(entry.rows)))\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tfor _, row := range entry.rows {\n\t\t// Write source name length\n\t\terr = binary.Write(buf, binary.LittleEndian, uint8(len(row.Source)))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\n\t\t// Write metric name length\n\t\terr = binary.Write(buf, binary.LittleEndian, uint8(len(row.Metric)))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\n\t\t// Write source and metric names\n\t\t_, err = buf.WriteString(row.Source + row.Metric)\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\n\t\t// Write timestamp and value\n\t\terr = binary.Write(buf, binary.LittleEndian, Point{\n\t\t\tTimestamp: row.Timestamp,\n\t\t\tValue: row.Value,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\t// Record the current offset so we can truncate\n\t// later in case something goes wrong.\n\tcurrentOffset, err := w.f.Seek(0, 1)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tw.lastReadOffset = currentOffset\n\n\t// Flush to the file.\n\tn, err := w.f.Write(buf.Bytes())\n\tif err != nil {\n\t\treturn n, err\n\t}\n\n\treturn n, err\n}", "func (rf *Raft) sendAppendEntriesToMultipleFollowers() {\n for !rf.killed() {\n rf.mu.Lock()\n if rf.state != \"Leader\" {\n DLCPrintf(\"Server (%d) is no longer Leader and stop sending Heart Beat\", rf.me)\n rf.mu.Unlock()\n return\n }\n\n for i := 0; i < len(rf.peers) && rf.state == \"Leader\"; i++ {\n if i == rf.me {\n continue\n }else{\n if rf.nextIndex[i] <= rf.snapshottedIndex {\n go rf.sendInstallSnapshotToOneFollower(i, rf.log[0].Term)\n }else{\n go rf.sendAppendEntriesToOneFollower(i)\n }\n }\n }\n if rf.state != \"Leader\" {\n DLCPrintf(\"Server (%d) is no longer Leader and stop sending Heart Beat\", rf.me)\n rf.mu.Unlock()\n return\n }\n rf.commitEntries()\n rf.mu.Unlock()\n\n time.Sleep(100 * time.Millisecond)\n }\n}", "func (m *OutMessage) Append(src []byte) {\n\tp := m.GrowNoZero(len(src))\n\tif p == nil {\n\t\tpanic(fmt.Sprintf(\"Can't grow %d bytes\", len(src)))\n\t}\n\n\tsh := (*reflect.SliceHeader)(unsafe.Pointer(&src))\n\tmemmove(p, unsafe.Pointer(sh.Data), uintptr(sh.Len))\n\n\treturn\n}", "func (kv *ShardKV) PutAppend(args *PutAppendArgs, reply *PutAppendReply) error {\n\tkv.mu.Lock()\n\tdefer kv.mu.Unlock()\n\n\tif args.Seq == kv.seqHistory[args.Me] {\n\t\treply.Err = kv.replyHistory[args.Me].Err\n\t} else if args.Seq > kv.seqHistory[args.Me] {\n\t\tputAppendOp := Op{CID: args.Me, Seq: args.Seq, Op: args.Op, Key: args.Key, Value: args.Value}\n\t\tkv.doOperation(putAppendOp)\n\t}\n\n\treturn nil\n}", "func (h *Handle) NeighAppend(neigh *Neigh) error {\n\treturn h.neighAdd(neigh, syscall.NLM_F_CREATE|syscall.NLM_F_APPEND)\n}", "func (w *Writer) Append(prog []byte, n int64)", "func (this *UnsubscribeMessage) AddTopic(topic []byte) {\n\tif this.TopicExists(topic) {\n\t\treturn\n\t}\n\n\tthis.topics = append(this.topics, topic)\n\tthis.dirty = true\n}", "func (client *MemcachedClient4T) Append(e *common.Element) error {\n\treturn client.store(\"append\", e)\n}", "func (r *Redis) AppendTaskLog(taskID string, now time.Time, system, message, level string) error {\n\tconn := r.pool.Get()\n\tdefer conn.Close()\n\n\ttaskLogsKey := \"sync_tasks#\" + taskID + \":logs\"\n\tlogRecord := TaskLogRecord{\n\t\tTime: now.Format(timestamp.Layout),\n\t\tSystem: system,\n\t\tMessage: message,\n\t\tLevel: level,\n\t}\n\n\t_, err := conn.Do(\"ZADD\", taskLogsKey, now.Unix(), logRecord.Marshal())\n\tnoticeError(err)\n\tif err != nil && err != redis.ErrNil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *Cluster) AppendData(ap *types.AcceptedProposal) error {\n\tlog.Printf(\"[node %d] Received data: %d,%d [%s] - %s\\n\", c.id, ap.Index, ap.Term, ap.Topic, string(ap.Data))\n\n\tt, ok := c.Data.Topics[ap.Topic]\n\tif !ok {\n\t\treturn fmt.Errorf(\"Attempting to append to topic %s which was not found\", c.dataDir)\n\t}\n\n\treturn t.Append(topic.Data{Index: ap.Index, Term: ap.Term, Data: ap.Data})\n}", "func (s *sliding) append(u Update) {\n\tpos := s.start + LogPosition(len(s.log))\n\ts.log = append(s.log, u)\n\ts.addrPos[u.Addr] = pos\n}", "func (r *LoggingRepository) Append(ctx context.Context, teamID, userID string) error {\n\tstart := time.Now()\n\terr := r.upstream.Append(ctx, teamID, userID)\n\n\tlogger := r.logger.With().\n\t\tStr(\"request\", r.requestID(ctx)).\n\t\tStr(\"method\", \"append\").\n\t\tDur(\"duration\", time.Since(start)).\n\t\tStr(\"team\", teamID).\n\t\tStr(\"user\", userID).\n\t\tLogger()\n\n\tif err != nil {\n\t\tlogger.Warn().\n\t\t\tErr(err).\n\t\t\tMsg(\"failed to append member\")\n\t} else {\n\t\tlogger.Debug().\n\t\t\tMsg(\"\")\n\t}\n\n\treturn err\n}", "func (p *Pes) Append(buf []byte) {\n\tp.buf = append(p.buf, buf...)\n}", "func (n *Sub) push(nmsg []byte) {\n\tif n.filter != nil && n.filter(nmsg) {\n\t\treturn\n\t}\n\tfor n.stop {\n\t\ttime.Sleep(time.Minute)\n\t}\n\tmsg := string(nmsg)\n\tfor i := 0; i < len(n.w.Callbacks); i++ {\n\t\tcb := n.w.Callbacks[i]\n\t\tif err := n.retry(cb.URL, string(nmsg), _syncCall); err != nil {\n\t\t\tid, err := n.backupMsg(msg, i)\n\t\t\tif err != nil {\n\t\t\t\tlog.Error(\"group(%s) topic(%s) add msg(%s) backup fail err %v\", n.w.Group, n.w.Topic, string(nmsg), err)\n\t\t\t}\n\t\t\tn.addAsyncRty(id, msg, i)\n\t\t\treturn\n\t\t}\n\t}\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\t// You will have to modify this function.\n\tck.seq++\n\targs := PutAppendArgs{\n\t\tKey: key,\n\t\tValue: value,\n\t\tOp: op,\n\t\tId: ck.id,\n\t\tSeq: ck.seq,\n\t}\n\n\tif op == \"Put\" || op == \"Append\" {\n\t\tfor {\n\t\t\tvar reply PutAppendReply\n\t\t\tif ck.servers[ck.leaderId].Call(\"KVServer.PutAppend\", &args, &reply) && reply.Err==OK {\n\t\t\t\tDPrintf(\"[client %v, commandSeq %v] command(%v) execute, (key:%v, value %v)\",ck.id, ck.seq, op, key, value)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif reply.Err == ErrWrongLeader {\n\t\t\t\tDPrintf(\"[client %v, commandSeq %v] command(%v) send to server(%v) is not leader\", ck.id, ck.seq, op, ck.leaderId)\n\t\t\t} else {\n\t\t\t\tDPrintf(\"cannot connect to server [%v], reply.Err(%v)\", ck.leaderId, reply.Err)\n\t\t\t}\n\t\t\tck.leaderId = ck.randomChooseLeader()\n\t\t}\n\t} else {\n\t\tDPrintf(\"[client %v, commandSeq %v] command (%v) not support\", ck.id, ck.seq, op)\n\t}\n}", "func (k Keeper) AppendMessage(\n\tctx sdk.Context,\n\tmessage types.Message,\n) uint64 {\n\t// Create the message\n\tcount := k.GetMessageCount(ctx)\n\n\t// Set the ID of the appended value\n\tmessage.Id = count\n\n\tstore := prefix.NewStore(ctx.KVStore(k.storeKey), types.KeyPrefix(types.MessageKey))\n\tappendedValue := k.cdc.MustMarshalBinaryBare(&message)\n\tstore.Set(GetMessageIDBytes(message.Id), appendedValue)\n\n\t// Update message count\n\tk.SetMessageCount(ctx, count+1)\n\n\treturn count\n}", "func (rf *Raft) sendEntries() {\n\trf.mu.Lock()\n\tlastLog := rf.getLastLog()\n\trf.mu.Unlock()\n\tfor i := range rf.peers {\n\t\tif i == rf.me {\n\t\t\tcontinue\n\t\t}\n\t\trf.mu.Lock()\n\t\tmatchIndex := rf.LeaderStatus.matchIndex[i]\n\t\tnextIndex := rf.LeaderStatus.nextIndex[i]\n\t\t//DPrintf(\"send entry peer=%v matchIndex=%v lastIndex=%v nextIndex=%v\", i, matchIndex, lastLog.Index, nextIndex)\n\t\tvar req *AppendEntriesArgs\n\t\t// TODO: whether delete ???\n\t\tif matchIndex >= lastLog.Index {\n\t\t\treq = &AppendEntriesArgs{\n\t\t\t\tType: HeartBeat,\n\t\t\t\tTerm: rf.currentTerm,\n\t\t\t\tLeaderId: rf.peerId,\n\t\t\t\tLeaderCommit: rf.commitIndex,\n\t\t\t}\n\t\t\tDPrintf(\"peer=%v send heartbeat to peer=%v\", rf.me, i)\n\t\t} else {\n\t\t\t// TODO: if the logEntries be cutoff after make snapshot, we should shift the start index\n\t\t\tlogEntries := rf.logEntries[matchIndex+1 : min(nextIndex+1, len(rf.logEntries))]\n\t\t\tprevLog := rf.logEntries[matchIndex]\n\t\t\treq = &AppendEntriesArgs{\n\t\t\t\tType: Entries,\n\t\t\t\tTerm: rf.currentTerm,\n\t\t\t\tLeaderId: rf.peerId,\n\t\t\t\tPrevLogIndex: prevLog.Index,\n\t\t\t\tPrevLogTerm: prevLog.Term,\n\t\t\t\tLogEntries: logEntries, // TODO: refine to control each time send message count (case 2B)\n\t\t\t\tLeaderCommit: rf.commitIndex,\n\t\t\t}\n\t\t\t//DPrintf(\"peer=%v send entry=%v to=%v next=%v logEntrySize=%d\", rf.me, rf.logEntries[matchIndex+1 : nextIndex+1], i, nextIndex, len(logEntries))\n\t\t}\n\t\trf.mu.Unlock()\n\t\tgo rf.sendAppendEntries(i, req, &AppendEntriesReply{})\n\t}\n}", "func (ck *Clerk) PutAppend(key string, value string, op string) {\n\targs := PutAppendArgs{}\n\targs.Key = key\n\targs.Value = value\n\targs.Op = op\n\targs.ClientID = ck.clientID\n\targs.RequestID = ck.currentRPCNum\n\n\n\tfor {\n\t\tshard := key2shard(key)\n\t\tgid := ck.config.Shards[shard]\n\t\t// If the gid exists in our current stored configuration. \n\t\tif servers, ok := ck.config.Groups[gid]; ok {\n\t\t\t// try each server for the shard.\n\t\t\t\n\t\t\t\tselectedServer := ck.getRandomServer(gid)\n\t\t\t\tsrv := ck.make_end(servers[selectedServer])\n\t\t\t\t\n\t\t\t\tvar reply PutAppendReply\n\t\t\t\tok := ck.sendRPC(srv, \"ShardKV.PutAppend\", &args, &reply)\n\n\t\t\t\t// Wrong Leader (reset stored leader)\n\t\t\t\tif !ok || (ok && reply.WrongLeader == true) {\n\t\t\t\t\tck.currentLeader[gid] = -1\n\t\t\t\t}\n\n\t\t\t\t// Correct Leader\n\t\t\t\tif ok && reply.WrongLeader == false {\n\t\t\t\t\t//Update stored Leader\n\t\t\t\t\tck.currentLeader[gid] = selectedServer\n\n\t\t\t\t\t// Handle successful reply\n\t\t\t\t\tif (reply.Err == OK) {\n\t\t\t\t\t\tck.DPrintf1(\"Action: PutAppend completed. Sent Args => %+v, Received Reply => %+v \\n\", args, reply)\n\t\t\t\t\t\t// RPC Completed so increment the RPC count by 1.\n\t\t\t\t\t\tck.currentRPCNum = ck.currentRPCNum + 1\n\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tif ok && reply.Err == ErrWrongGroup {\n\t\t\t\t\t// ask master for the latest configuration.\n\t\t\t\t\tck.config = ck.sm.Query(-1)\n\t\t\t\t\ttime.Sleep(100 * time.Millisecond)\n\t\t\t\t}\n\n\t\t} else {\n\t\t\t// ask master for the latest configuration.\n\t\t\tck.config = ck.sm.Query(-1)\n\t\t\ttime.Sleep(100 * time.Millisecond)\n\t\t}\n\t}\n\n\tck.DError(\"Return from PUTAPPEND in ShardKV Client. Should never return from here.\")\n}", "func (r *Raft) handleAppendEntries(m pb.Message) {\n\n\tvar prevPosition = -1\n\tif len(r.RaftLog.entries) == 0 || m.Index < r.RaftLog.entries[0].Index {\n\t\tterm, err := r.RaftLog.storage.Term(m.Index)\n\t\tif err != nil || term != m.LogTerm {\n\t\t\tr.appendMsg(r.buildReject(pb.MessageType_MsgAppendResponse, m.From))\n\t\t\treturn\n\t\t}\n\t} else {\n\t\t//reject if prevPosition entry not findLastMatch\n\t\tvar found bool\n\t\tprevPosition, found = r.RaftLog.findByIndex(m.Index)\n\t\tif !found || r.RaftLog.entries[prevPosition].Term != m.LogTerm {\n\t\t\tr.appendMsg(r.buildReject(pb.MessageType_MsgAppendResponse, m.From))\n\t\t\treturn\n\t\t}\n\t}\n\n\toffset := 0\n\tfor ; offset < len(m.Entries); offset++ {\n\t\tif offset+prevPosition+1 >= len(r.RaftLog.entries) {\n\t\t\tr.RaftLog.append(m.Entries[offset:])\n\t\t\tbreak\n\t\t}\n\t\te1 := r.RaftLog.entries[offset+prevPosition+1]\n\t\te2 := m.Entries[offset]\n\t\tif e1.Index != e2.Index || e1.Term != e2.Term {\n\t\t\tr.RaftLog.entries = r.RaftLog.entries[:offset+prevPosition+1]\n\t\t\tif len(r.RaftLog.entries) > 0 {\n\t\t\t\tlastIndexInLog := r.RaftLog.entries[len(r.RaftLog.entries)-1].Index\n\t\t\t\tif lastIndexInLog < r.RaftLog.stabled {\n\t\t\t\t\tr.RaftLog.stabled = lastIndexInLog\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tr.RaftLog.stabled = 0\n\t\t\t}\n\t\t\tr.RaftLog.append(m.Entries[offset:])\n\t\t\tbreak\n\t\t}\n\t}\n\n\tmsg := r.buildMsgWithoutData(pb.MessageType_MsgAppendResponse, m.From, false)\n\tmsg.Index = r.RaftLog.LastIndex()\n\tr.appendMsg(msg)\n\n\t// update committed\n\tlastIndex := lastIndexInMeg(m)\n\tif m.Commit > r.RaftLog.committed && lastIndex > r.RaftLog.committed {\n\t\tr.RaftLog.committed = min(m.Commit, lastIndex)\n\t}\n\n\t// Your Code Here (2A).\n}", "func (b *BookShelf) Append(book *Book) {\n\tb.books = append(b.books, book)\n\tb.last++\n}", "func (l *Log) writer() {\n\t// Open as O_RDWR (which should get lock) and O_DIRECT.\n\tf, err := os.OpenFile(l.filename, os.O_WRONLY|os.O_APPEND, 0660)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer f.Close()\n\tenc := json.NewEncoder(f)\n\tfor {\n\t\tr, ok := <-l.in\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\t\tif r.m == nil {\n\t\t\tr.err <- fmt.Errorf(\"cannot write nil to wal\")\n\t\t\treturn\n\t\t}\n\t\t// serialize mutation and write to disk\n\t\tif err := enc.Encode(r.m); err != nil {\n\t\t\tr.err <- fmt.Errorf(\"wal encoding: %s\", err.Error())\n\t\t\treturn\n\t\t}\n\t\t// sync\n\t\tif err := f.Sync(); err != nil {\n\t\t\tr.err <- fmt.Errorf(\"wal sync: %s\", err.Error())\n\t\t\treturn\n\t\t}\n\t\tr.err <- nil\n\t\t// send to reader\n\t\tif l.closed {\n\t\t\treturn\n\t\t}\n\t}\n}", "func (r *raft) appendEntriesReceiver(p *AppendEntries) (*AppendEntriesResults, error) {\n\tif p.Term < r.currentTerm {\n\t\treturn &AppendEntriesResults{Term: r.currentTerm, Success: false}, nil\n\t}\n\n\tlastIndex, err := r.log.LastIndex()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif lastIndex < p.PrevLogIndex {\n\t\treturn &AppendEntriesResults{Term: r.currentTerm, Success: false}, nil\n\t}\n\tentries, err := r.log.Read(p.PrevLogIndex, p.PrevLogIndex+1)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif entries[0].Term != p.PrevLogTerm {\n\t\treturn &AppendEntriesResults{Term: r.currentTerm, Success: false}, nil\n\t}\n\t// 3. If an existing entry conflicts with a new one(same index but different terms),\n\t// delete the existing entry and all that follow\n\t// 4. Append any new entries not alredy in the log\n\t// TODO: just overwrite directly, is it most efficient?\n\t//r.log = append(r.log[:p.PrevLogIndex], p.Entries...)\n\tif err := r.log.Write(p.PrevLogIndex+1, p.Entries); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif p.LeaderCommit > r.commitIndex {\n\t\t// attention: the log's last index has been updated\n\t\tlastIndex, err := r.log.LastIndex()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tr.commitIndex = min(p.LeaderCommit, lastIndex)\n\t\t// TODO:apply the new committed log to state machine\n\t\t// and update lastApplied\n\t\tentries, err = r.log.Read(r.lastApplied+1, r.commitIndex+1)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\t// TODO: need save lastApplied to file as currentTerm and votedFor ?\n\t\tr.applier.Apply(entries)\n\t\tr.votedFor = p.LeaderId\n\t\tr.lastApplied = r.commitIndex\n\t}\n\tr.currentTerm = p.Term\n\treturn &AppendEntriesResults{Term: r.currentTerm, Success: true}, nil\n}", "func (l *Lgr) newStreamEntry(streamKey string, msg string, level string, complete bool) {\n // Get new connection from Redis pool.\n conn := l.RedisPool.Get()\n defer conn.Close()\n\n // Add message to log stream.\n if _, err := conn.Do(\"XADD\", streamKey, \"*\", \"msg\", msg, \"level\", level, \"complete\", complete); err != nil {\n l.Errorf(\"error logging to stream: %s\", err.Error())\n }\n}", "func (c *Cluster) AddTopic(toml []byte) error {\n\treturn c.createTopic(toml, false)\n}", "func (l *Lifecycle) Append(hook Hook) {\n\t// Save the caller's stack frame to report file/line number.\n\tif f := fxreflect.CallerStack(2, 0); len(f) > 0 {\n\t\thook.callerFrame = f[0]\n\t}\n\tl.hooks = append(l.hooks, hook)\n}" ]
[ "0.7172584", "0.6377327", "0.57611495", "0.5748846", "0.5728976", "0.5707987", "0.56621695", "0.5645391", "0.55513144", "0.55505306", "0.5547055", "0.55321455", "0.5517304", "0.5488533", "0.54715824", "0.5470633", "0.5447862", "0.5418506", "0.5404934", "0.5399982", "0.53975344", "0.53965235", "0.5375046", "0.53295565", "0.53295565", "0.52958375", "0.5295405", "0.5294363", "0.52781916", "0.52729446", "0.52707654", "0.526941", "0.52657485", "0.5248292", "0.523581", "0.5233371", "0.52330667", "0.5217975", "0.5217975", "0.5209594", "0.5207466", "0.52020955", "0.5201816", "0.5190506", "0.51887834", "0.5187495", "0.5187166", "0.5181439", "0.5175892", "0.5165442", "0.5165442", "0.51644033", "0.5158411", "0.5155751", "0.5152437", "0.5128269", "0.5115329", "0.5102119", "0.5100771", "0.5100118", "0.509994", "0.5086438", "0.50755847", "0.50676787", "0.5064432", "0.50624436", "0.5057373", "0.5055597", "0.5041096", "0.50350314", "0.50230014", "0.50147986", "0.5009445", "0.5008241", "0.49980783", "0.49953473", "0.49923852", "0.4991727", "0.49907213", "0.49903828", "0.49884138", "0.49855018", "0.49834862", "0.49834782", "0.49747807", "0.49672598", "0.49602807", "0.4958491", "0.49415877", "0.49414766", "0.49389717", "0.49249688", "0.4921974", "0.49197665", "0.4917665", "0.49084678", "0.4907669", "0.4894885", "0.48887002", "0.48829356" ]
0.7302169
0
Close closes the underlying wal
func (t *WALTopic) Close() error { return t.wal.Close() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (gWal *GenericWAL) Close() error {\n gWal.wal.Close()\n return nil\n}", "func (wal *seriesWAL) Close() error {\n\treturn wal.base.close()\n}", "func (w *WAL) Close() error {\n\tw.mu.Lock()\n\tdefer w.mu.Unlock()\n\n\tif w.fp != nil {\n\t\tw.fp.Close()\n\t\tw.fp = nil\n\t}\n\n\tif w.tail() != nil {\n\t\tif err := w.sync(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tfor _, l := range w.locks {\n\t\tif l == nil {\n\t\t\tcontinue\n\t\t}\n\t\tif err := l.Close(); err != nil {\n\t\t\tklog.Errorf(fmt.Sprintf(\"failed to unlock during closing wal: %v\", err))\n\t\t}\n\t}\n\n\treturn w.dirFile.Close()\n}", "func (w *fileWAL) close() error {\n\tif w.f != nil {\n\t\terr := w.f.Close()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tw.f = nil\n\tw.lastReadOffset = 0\n\treturn nil\n}", "func (t *LSMTree) Close() error {\n\tif err := t.wal.Close(); err != nil {\n\t\treturn fmt.Errorf(\"failed to close file %s: %w\", t.wal.Name(), err)\n\t}\n\n\treturn nil\n}", "func (wal *WAL) Close() (err error) {\n\twal.closeOnce.Do(func() {\n\t\tselect {\n\t\tcase <-wal.closed:\n\t\t\t// already closed\n\t\t\treturn\n\t\tdefault:\n\t\t\t// continue\n\t\t}\n\n\t\twal.log.Debug(\"Closing\")\n\t\tdefer wal.log.Debug(\"Closed\")\n\n\t\tclose(wal.closed)\n\n\t\tif wal.backlog != nil {\n\t\t\tclose(wal.backlog)\n\t\t\t<-wal.backlogFinished\n\t\t}\n\n\t\twal.mx.Lock()\n\t\tflushErr := wal.writer.Flush()\n\t\tsyncErr := wal.file.Sync()\n\t\twal.mx.Unlock()\n\t\tcloseErr := wal.file.Close()\n\t\tif flushErr != nil {\n\t\t\terr = flushErr\n\t\t}\n\t\tif syncErr != nil {\n\t\t\terr = syncErr\n\t\t}\n\t\terr = closeErr\n\t})\n\n\treturn\n}", "func (self *FileBaseDataStore) Close() {}", "func (mavls *Store) Close() {\r\n\tmavl.ClosePrune()\r\n\tmavls.BaseStore.Close()\r\n\tmlog.Info(\"store mavl closed\")\r\n}", "func (lw *LogWriter) Close() error { return nil }", "func (m *DataStore) Close() {\n}", "func (s *State) Close() {\n\ts.dbFile.Close()\n}", "func (s *State) Close() {\n\ts.dbFile.Close()\n}", "func (store *Store) Close() {\n\tstore.writer.Close()\n}", "func (t *Triplestore) Close() {\n\tt.db.Close()\n}", "func (*Store) Close(context.Context) error { return nil }", "func (s *Store) Close() error {\n\terr2 := s.writer.Close()\n\terr1 := s.db.Close()\n\t//s.reader.Close()\n\tif err1 != nil {\n\t\treturn err1\n\t}\n\treturn err2\n}", "func (t *Tangle) Close() {\n\tt.store.Close()\n\tt.data.Close()\n}", "func (scs *StorageContractSet) Close() (err error) {\n\tscs.db.Close()\n\t_, err = scs.wal.CloseIncomplete()\n\treturn\n}", "func (ms *memoryStorer) Close() error {\n\treturn nil\n}", "func (t *TestStore) Close() error {\n\tos.Remove(t.path)\n\n\treturn t.BoltStore.Close()\n}", "func (s ConsoleIndexStore) Close() error { return nil }", "func (db *DB) Close() {\n\tif db == nil {\n\t\treturn\n\t}\n\n\tif db.roIt != nil {\n\t\tdb.roIt.Close()\n\t}\n\n\tif db.ro != nil {\n\t\tdb.ro.Close()\n\t}\n\n\tif db.wo != nil {\n\t\tdb.wo.Close()\n\t}\n\n\tif db.LevigoDB != nil {\n\t\tdb.LevigoDB.Close()\n\t}\n\t// delete cache AFTER close leveldb or it will hang.\n\t// See cache in http://leveldb.googlecode.com/svn/trunk/doc/index.html\n\tif db.cache != nil {\n\t\tdb.cache.Close()\n\t}\n}", "func (ram *Ram) Close() error {\n\tram.tables = nil\n\n\treturn nil\n}", "func (DummyStore) Close() error {\n\treturn nil\n}", "func (ds *DkStore) Close() {\n\terr := ds.st.Flush()\n\tcheck(err)\n\tds.st.Close()\n\tcheck(err)\n\terr = ds.f.Sync()\n\tcheck(err)\n\terr = ds.f.Close()\n\tcheck(err)\n}", "func (b *Backup) Close() error {\n\tif b == nil {\n\t\treturn errors.New(\"nil sqlite backup\")\n\t}\n\tif b.sb == nil {\n\t\treturn nil\n\t}\n\trv := C.sqlite3_backup_finish(b.sb) // must be called only once\n\tb.sb = nil\n\tif rv != C.SQLITE_OK {\n\t\treturn b.dst.error(rv, \"backup finish failed\")\n\t}\n\treturn nil\n}", "func (s *Store) Close() error {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\n\tif s.opened {\n\t\tclose(s.closing)\n\t}\n\ts.wg.Wait()\n\n\tfor _, sh := range s.shards {\n\t\tif err := sh.Close(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\ts.opened = false\n\ts.shards = nil\n\ts.databaseIndexes = nil\n\n\treturn nil\n}", "func (r *leveldbKV) Close() error {\n\tif r.db != nil {\n\t\tr.db.Close()\n\t}\n\tif r.wo != nil {\n\t\tr.wo.Close()\n\t}\n\tif r.ro != nil {\n\t\tr.ro.Close()\n\t}\n\tif r.opts != nil {\n\t\tr.opts.Close()\n\t}\n\tif r.fp != nil {\n\t\tr.fp.Close()\n\t}\n\tr.db = nil\n\treturn nil\n}", "func (s *Store) Close() error {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\n\ts.recs = []*record{}\n\ts.index = map[uint64]*record{}\n\n\treturn s.backend.Close()\n}", "func (store *sqlStore) Close() error {\n\tif store.db != nil {\n\t\tstore.db.Close()\n\t\tstore.db = nil\n\t}\n\treturn nil\n}", "func (mw *MockWarehouse) Close() {}", "func (l *LogDB) Close() (err error) {\n\tfunc() {\n\t\tl.mu.Lock()\n\t\tdefer l.mu.Unlock()\n\t\terr = firstError(err, l.collection.iterate(func(db *db) error {\n\t\t\treturn db.close()\n\t\t}))\n\t}()\n\terr = firstError(err, l.bsDir.Close())\n\terr = firstError(err, l.dir.Close())\n\treturn firstError(err, l.fileLock.Close())\n}", "func (l *LevelDBLog) Close() (err error) {\n\tl.Lock()\n\tdefer l.Unlock()\n\n\tif err = l.db.Close(); err != nil {\n\t\treturn err\n\t}\n\tl.db = nil\n\treturn nil\n}", "func (d *DB) Close() error { return d.lab.Close() }", "func (s *Store) Close() error {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\ts.m = nil\n\treturn nil\n}", "func (s *FileStore) Close() error {\n\treturn nil\n}", "func (b *blt) Close() error {\n\tb.Lock()\n\tdefer b.Unlock()\n\tif b.db == nil {\n\t\treturn trace.AlreadyExists(\"database %v is already closed\", b.path)\n\t}\n\terr := b.db.Close()\n\tif err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\tb.db = nil\n\treturn nil\n}", "func (l *raftLayer) Close() error { return l.ln.Close() }", "func (s *MemStore) Close() error {\n\treturn nil\n}", "func (s *OnlineDDLStorage) Close() {\n\ts.Lock()\n\tdefer s.Unlock()\n\n\tcloseBaseDB(s.logCtx, s.db)\n}", "func (p *Paket) Close() error {\n\terr := p.file.Close()\n\tp.key = nil\n\tp.table = nil\n\tp.file = nil\n\tp = nil\n\treturn err\n}", "func (p *BinlogFile) Close() {\n\tif p.reader != nil {\n\t\tp.reader.Close()\n\t\tp.reader = nil\n\t}\n}", "func (blt *Bolt) Close() error {\n\tblt.opened = false\n\treturn blt.db.Close()\n}", "func (e *SqliteEndorsementStore) Close() error {\n\treturn e.db.Close()\n}", "func Close() {\n\tif store != nil {\n\t\tstore.Close()\n\t\tstore = nil\n\t}\n}", "func (l *InMemoryLog) Close() error {\n\treturn nil\n}", "func (h *Harness) Close() error {\n\th.t.Helper()\n\tif recErr := recover(); recErr != nil {\n\t\tdefer panic(recErr)\n\t}\n\th.dumpDB() // early as possible\n\n\th.tw.WaitAndAssert(h.t)\n\th.slack.WaitAndAssert()\n\th.email.WaitAndAssert()\n\n\th.mx.Lock()\n\th.closing = true\n\th.mx.Unlock()\n\n\tctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)\n\tdefer cancel()\n\terr := h.backend.Shutdown(ctx)\n\tif err != nil {\n\t\th.t.Error(\"failed to shutdown backend cleanly:\", err)\n\t}\n\th.backendLogs.Close()\n\n\th.slackS.Close()\n\th.twS.Close()\n\n\th.tw.Close()\n\n\th.pgTime.Close()\n\n\tconn, err := pgx.Connect(ctx, DBURL(\"\"))\n\tif err != nil {\n\t\th.t.Error(\"failed to connect to DB:\", err)\n\t}\n\tdefer conn.Close(ctx)\n\t_, err = conn.Exec(ctx, \"drop database \"+sqlutil.QuoteID(h.dbName))\n\tif err != nil {\n\t\th.t.Errorf(\"failed to drop database '%s': %v\", h.dbName, err)\n\t}\n\n\treturn nil\n}", "func (b *BackupEngine) Close() {\n\tC.rocksdb_backup_engine_close(b.c)\n\tb.c = nil\n\tb.db = nil\n}", "func (db *sqlstore) Close() error {\n\treturn db.DB.Close()\n}", "func (l *ChainLedger) Close() {\n\tl.ldb.Close()\n}", "func (l *ChainLedger) Close() {\n\tl.ldb.Close()\n}", "func (mBatch *goLevelDBBatch) Close() {}", "func (rd *RemoteDB) Close() {\n}", "func (mvccs *KVMVCCStore) Close() {\n\tkmlog.Info(\"store kvdb closed\")\n}", "func Close() {\n\tC.Z3_close_log()\n}", "func (s *Store) Close() error {\n\ts.Sql.Close()\n\treturn nil\n}", "func (db *sliceDB) close() {\r\n\tclose(db.quit)\r\n\tdb.lvl.Close()\r\n}", "func (s *Store) Close() {\n\ts.db.Close()\n}", "func (d *Store) Close() error {\n\tif err := d.save(); err != nil {\n\t\treturn err\n\t}\n\treturn d.f.Close()\n}", "func (s *stor) Close() {\n\tif s.d == nil {\n\t\treturn\n\t}\n\ts.d.Close()\n}", "func (t *Tstore) Close() {\n\tlog.Tracef(\"Close\")\n\n\t// Close connections\n\tt.tlog.Close()\n\tt.store.Close()\n}", "func (a *Archiver) Close() error {\n\treturn a.db.Close()\n}", "func (l *Ledger) Close() {\n\tl.baseDB.Close()\n}", "func (db *DB) Close() error {\n\tif db.cancelBgWorker != nil {\n\t\tdb.cancelBgWorker()\n\t}\n\tdb.closeWg.Wait()\n\tdb.mu.Lock()\n\tdefer db.mu.Unlock()\n\tif err := db.writeMeta(); err != nil {\n\t\treturn err\n\t}\n\tif err := db.datalog.close(); err != nil {\n\t\treturn err\n\t}\n\tif err := db.index.close(); err != nil {\n\t\treturn err\n\t}\n\tif err := db.lock.Unlock(); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (w *RWWrapper) Close() {\n\tif w.gz != nil {\n\t\tw.gz.Close()\n\t}\n}", "func (s *LevelDBStore) Close() error {\n\treturn s.db.Close()\n}", "func (db *DB) Close() error {\n\tdb.rwlock.Lock()\n\tdefer db.rwlock.Unlock()\n\n\tdb.metalock.Lock()\n\tdefer db.metalock.Unlock()\n\n\tdb.mmaplock.Lock()\n\tdefer db.mmaplock.Unlock()\n\n\treturn db.close()\n}", "func (*GraphStore) Close(ctx context.Context) error { return nil }", "func (s *Store) Close() error {\n\tif s.Listener != nil {\n\t\ts.Listener.Close()\n\t}\n\tif !s.LeaveFiles {\n\t\tdefer os.RemoveAll(s.Path())\n\t}\n\treturn s.Store.Close()\n}", "func (db *DB) Close() error {\n\n\tdb.rwlock.Lock()\n\tdefer db.rwlock.Unlock()\n\n\tdb.metalock.Lock()\n\tdefer db.metalock.Unlock()\n\n\tdb.mmaplock.RLock()\n\tdefer db.mmaplock.RUnlock()\n\n\treturn db.close()\n}", "func (c *cachestub) Close() {}", "func (s *MemorySink) Close() error { return nil }", "func (dbm *MemDB) Close() error {\n\treturn nil\n}", "func WriterClose(w *zip.Writer,) error", "func (l *networkLayer) Close() error { return l.ln.Close() }", "func (d *Datastore) Close() error {\n return d.DB.Close()\n}", "func (ng *Engine) Close() {\n\t// Clean will un-ref and Close the databases\n\tng.stores.Clean()\n}", "func (store *Storage) Close() {\n\tstore.db.Close()\n}", "func (s *Storage) Close() error {\n\ts.logger.Stop()\n\ts.mtx.Lock()\n\tdefer s.mtx.Unlock()\n\treturn s.rws.Close()\n}", "func (storage *Storage) Close() {\n}", "func (mcs *MemoryCellStore) Close() error {\n\treturn nil\n}", "func (m *DataManager) Close() {\n\tif m.opNeedLock {\n\t\tm.opLock.Lock()\n\t\tdefer m.opLock.Unlock()\n\t}\n\tif m.opdb != nil {\n\t\tm.opdb.Close()\n\t\tm.opdb = nil\n\t}\n\n\tif m.qNeedLock {\n\t\tm.qLock.Lock()\n\t\tdefer m.qLock.Unlock()\n\t}\n\tif m.qdb != nil {\n\t\tm.qdb.Close()\n\t\tm.qdb = nil\n\t}\n}", "func (store *KeyValueStore) Close() {\n\tif store.closeFileOnClose {\n\t\tstore.DB.Close()\n\t}\n\tstore.DB = nil\n}", "func Close() {\n\t_db.Close()\n\t_db = nil\n}", "func (dbWoker *DomainInformationStorage)Close() {\n\tdbWoker.session.Close()\n}", "func (s *Store) Close() {\n\ts.db.Disconnect(dbctx)\n}", "func (mb *metadataBackend) Close() error {\n\treturn mb.db.Close()\n}", "func (db *MemoryCache) Close() error {\n\tdb.lock.Lock()\n\tdefer db.lock.Unlock()\n\n\tdb.db = nil\n\treturn nil\n}", "func (d *BoltDB) Close() {\n\td.Close()\n}", "func (p *Store) Close(context.Context) error {\n\treturn nil\n}", "func (c *cockroachdb) Close() {\n\tlog.Tracef(\"Close\")\n\n\tc.Lock()\n\tdefer c.Unlock()\n\n\tc.shutdown = true\n\tc.recordsdb.Close()\n}", "func (b *BoltDBStorage) Close() error {\n\treturn b.db.Close()\n}", "func Close() {\n\terr := store.Close()\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n}", "func (dal *DAL) Close() error {\n\treturn dal.db.Close()\n}", "func (b *Backend) Close() error { return nil }", "func (c *Cache) Close() {\n\tc.fh.Close()\n}", "func (s *Store) Close() error {\n\treturn s.storeLock.Close()\n}", "func (g *CayleyGraph) Close() {\n\tg.store.Close()\n}", "func (d *Dao) Close() {\n\td.redis.Close()\n\td.mallDB.Close()\n\td.ugcDB.Close()\n\td.ticketDB.Close()\n}", "func (b *BadgerStore) Close() error {\n\tif b.vlogTicker != nil {\n\t\tb.vlogTicker.Stop()\n\t}\n\tif b.mandatoryVlogTicker != nil {\n\t\tb.mandatoryVlogTicker.Stop()\n\t}\n\treturn b.conn.Close()\n}" ]
[ "0.7960339", "0.710267", "0.6919875", "0.67559904", "0.67290163", "0.6673831", "0.6418077", "0.6385415", "0.6304291", "0.629084", "0.6190526", "0.6190526", "0.6162513", "0.6127387", "0.6101584", "0.6093734", "0.60745424", "0.60743195", "0.6074003", "0.60731363", "0.6072704", "0.6070774", "0.60638964", "0.6042525", "0.60317177", "0.6024847", "0.6017662", "0.6009037", "0.59942436", "0.59924597", "0.59862554", "0.5974137", "0.5971498", "0.5971356", "0.59565604", "0.5954908", "0.5936971", "0.5934785", "0.5927571", "0.59187776", "0.5913699", "0.59074193", "0.5901752", "0.5901072", "0.5896395", "0.589464", "0.58862585", "0.5871463", "0.5847102", "0.5844612", "0.5844612", "0.58417964", "0.58405226", "0.5834153", "0.58159107", "0.58098036", "0.57994986", "0.5798836", "0.5789513", "0.5787743", "0.57765836", "0.5769962", "0.5769394", "0.57462466", "0.5746023", "0.5738162", "0.57362694", "0.5733682", "0.5727046", "0.57251376", "0.5724358", "0.5712801", "0.5710099", "0.5704299", "0.5703902", "0.5701958", "0.57011104", "0.56960773", "0.5694718", "0.56945145", "0.5693611", "0.5690692", "0.56900316", "0.5687351", "0.56844884", "0.5679518", "0.567746", "0.5677027", "0.5676001", "0.5670712", "0.5670706", "0.56697947", "0.56658417", "0.56643414", "0.56612295", "0.5655091", "0.56522524", "0.5649732", "0.5642354", "0.56420255" ]
0.6502886
6
NewReader gets a reader of a wal TODO The index might not exist in this wal so we want to find either the index or the next index after the one given This is because during recovery in a multitopic syncable we may not know exactly what we are looking for and we are just checking to make sure we didn't miss anything.
func (t *WALTopic) NewReader(index uint64) (Reader, error) { return t.newWalReader(index) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (wal *WAL) NewReader(name string, offset Offset, bufferSource func() []byte) (*Reader, error) {\n\tr := &Reader{\n\t\tfilebased: filebased{\n\t\t\tdir: wal.dir,\n\t\t\tfileFlags: os.O_RDONLY,\n\t\t\th: newHash(),\n\t\t\tlog: golog.LoggerFor(\"wal.\" + name),\n\t\t},\n\t\twal: wal,\n\t\tbufferSource: bufferSource,\n\t}\n\tif offset != nil {\n\t\toffsetString := sequenceToFilename(offset.FileSequence())\n\t\tif offsetString[0] != '0' {\n\t\t\twal.log.Debugf(\"Converting legacy offset\")\n\t\t\toffset = NewOffset(offset.FileSequence()/1000, offset.Position())\n\t\t}\n\n\t\tfiles, err := ioutil.ReadDir(wal.dir)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Unable to list existing log files: %v\", err)\n\t\t}\n\n\t\tcutoff := sequenceToFilename(offset.FileSequence())\n\t\tfor i, fileInfo := range files {\n\t\t\tisMostRecent := i == len(files)-1\n\t\t\tif fileInfo.Name() >= cutoff {\n\t\t\t\t// Found existing or more recent WAL file\n\t\t\t\tr.fileSequence = filenameToSequence(fileInfo.Name())\n\t\t\t\tif r.fileSequence == offset.FileSequence() {\n\t\t\t\t\t// Exact match, start at right position\n\t\t\t\t\tr.position = offset.Position()\n\t\t\t\t\tif r.position == fileInfo.Size() && !isMostRecent {\n\t\t\t\t\t\t// At end of file and more recent is available, move to next\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\t// Newer WAL file, start at beginning\n\t\t\t\t\tr.position = 0\n\t\t\t\t}\n\t\t\t\topenErr := r.open()\n\t\t\t\tif openErr != nil {\n\t\t\t\t\treturn nil, fmt.Errorf(\"Unable to open existing log file at %v: %v\", fileInfo.Name(), openErr)\n\t\t\t\t}\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\n\tif r.file == nil {\n\t\t// Didn't find WAL file, advance\n\t\terr := r.advance()\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Unable to advance initially: %v\", err)\n\t\t}\n\t\twal.log.Debugf(\"Replaying log starting at %v\", r.file.Name())\n\t}\n\treturn r, nil\n}", "func (s ConsoleIndexStore) GetIndexReader(string) (io.ReadCloser, error) {\n\treturn ioutil.NopCloser(os.Stdin), nil\n}", "func ReadWAL(lg *zap.Logger, waldir string, snap log.Snapshot, unsafeNoFsync bool) (w *log.WAL,\n\twmetadata []byte, st log.HardState, ents []log.LogEntry) {\n\tvar err error\n\n\tst = log.NewEmptyState()\n\trepaired := false\n\tfor {\n\t\tif w, err = log.Open(lg, waldir, snap); err != nil {\n\t\t\tlg.Fatal(\"failed to open WAL\", zap.Error(err))\n\t\t}\n\t\tif unsafeNoFsync {\n\t\t\tw.SetUnsafeNoFsync()\n\t\t}\n\t\tif wmetadata, st, ents, err = w.ReadAll(); err != nil {\n\t\t\tw.Close()\n\t\t\t// we can only repair ErrUnexpectedEOF and we never repair twice.\n\t\t\tif repaired || err != io.ErrUnexpectedEOF {\n\t\t\t\tlg.Fatal(\"failed to read WAL, cannot be repaired\", zap.Error(err))\n\t\t\t}\n\t\t\tif !log.Repair(lg, waldir) {\n\t\t\t\tlg.Fatal(\"failed to repair WAL\", zap.Error(err))\n\t\t\t} else {\n\t\t\t\tlg.Info(\"repaired WAL\", zap.Error(err))\n\t\t\t\trepaired = true\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\treturn w, wmetadata, st, ents\n}", "func (c *ReplicaClient) WALSegmentReader(ctx context.Context, pos litestream.Pos) (_ io.ReadCloser, err error) {\n\tdefer func() { c.resetOnConnError(err) }()\n\n\tsftpClient, err := c.Init(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfilename, err := litestream.WALSegmentPath(c.Path, pos.Generation, pos.Index, pos.Offset)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"cannot determine wal segment path: %w\", err)\n\t}\n\n\tf, err := sftpClient.Open(filename)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tinternal.OperationTotalCounterVec.WithLabelValues(ReplicaClientType, \"GET\").Inc()\n\n\treturn f, nil\n}", "func readWAL(lg *zap.Logger, waldir string, snap walpb.Snapshot, unsafeNoFsync bool) (w *wal.WAL, id, cid types.ID, st raftpb.HardState, ents []raftpb.Entry) {\n\tvar (\n\t\terr error\n\t\twmetadata []byte\n\t)\n\n\trepaired := false\n\tfor {\n\t\tif w, err = wal.Open(lg, waldir, snap); err != nil {\n\t\t\tlg.Fatal(\"failed to open WAL\", zap.Error(err))\n\t\t}\n\t\tif unsafeNoFsync {\n\t\t\tw.SetUnsafeNoFsync()\n\t\t}\n\t\tif wmetadata, st, ents, err = w.ReadAll(); err != nil {\n\t\t\tw.Close()\n\t\t\t// we can only repair ErrUnexpectedEOF and we never repair twice.\n\t\t\tif repaired || !errors.Is(err, io.ErrUnexpectedEOF) {\n\t\t\t\tlg.Fatal(\"failed to read WAL, cannot be repaired\", zap.Error(err))\n\t\t\t}\n\t\t\tif !wal.Repair(lg, waldir) {\n\t\t\t\tlg.Fatal(\"failed to repair WAL\", zap.Error(err))\n\t\t\t} else {\n\t\t\t\tlg.Info(\"repaired WAL\", zap.Error(err))\n\t\t\t\trepaired = true\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\tvar metadata pb.Metadata\n\tpbutil.MustUnmarshal(&metadata, wmetadata)\n\tid = types.ID(metadata.NodeID)\n\tcid = types.ID(metadata.ClusterID)\n\treturn w, id, cid, st, ents\n}", "func (h *HTTPGetter) IndexReader() (io.ReadCloser, error) {\n\tsavePath := path.Join(h.dst, \"index\")\n\tif err := h.underlying.GetFile(savePath, h.idxURL); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn os.Open(savePath)\n}", "func (r *Reader) Read() ([]byte, error) {\n\tfor {\n\t\tlength, err := r.readHeader()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tchecksum, err := r.readHeader()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif length > maxEntrySize {\n\t\t\tfmt.Printf(\"Discarding wal entry of size %v exceeding %v, probably corrupted\\n\", humanize.Bytes(uint64(length)), humanize.Bytes(uint64(maxEntrySize)))\n\t\t\t_, discardErr := io.CopyN(ioutil.Discard, r.reader, int64(length))\n\t\t\tif discardErr == io.EOF {\n\t\t\t\tdiscardErr = nil\n\t\t\t}\n\t\t\treturn nil, discardErr\n\t\t}\n\t\tdata, err := r.readData(length)\n\t\tif data != nil || err != nil {\n\t\t\tif data != nil {\n\t\t\t\tr.h.Reset()\n\t\t\t\tr.h.Write(data)\n\t\t\t\tif checksum != int(r.h.Sum32()) {\n\t\t\t\t\tr.log.Errorf(\"Checksum mismatch, skipping entry\")\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn data, err\n\t\t}\n\t}\n}", "func (x *Index) Read(r io.Reader) error", "func (this *DatastoreOperations) CreateReader(state *DatastoreState, updatedAfter int64) (reader io.Reader, readSize int64, err error) {\n\t// Use the index to find the offset of the first entry matching the condition\n\toffset := state.Index.FindOffsetOfFirstEntryUpdatedAfter(updatedAfter)\n\n\t// If no such entry was found\n\tif offset == -1 {\n\t\t// Return an empty reader with zero length\n\t\treturn EmptyReader{}, 0, nil\n\t}\n\n\t// Create a reader for the range between the offset and the total\n\t// size of the indexed entries (in most cases, this would be the size of the file)\n\treader = NewRangeReader(state.File, offset, int64(state.Size()))\n\n\t// Calculate the read size as the difference between the total size and the read start offset\n\treadSize = state.Size() - offset\n\n\treturn\n}", "func NewWALReader(files []string) *WALReader {\n\tsort.Strings(files)\n\treturn &WALReader{\n\t\tfiles: files,\n\t\tlogger: zap.NewNop(),\n\t\tr: nil,\n\t}\n}", "func IndexRead(x *suffixarray.Index, r io.Reader) error", "func (w *SegmentWAL) Reader() WALReader {\n\treturn newWALReader(w, w.logger)\n}", "func (x *Index) Read(r io.Reader) error {}", "func (r *walReader) nextEntry() (WALEntryType, byte, []byte, error) {\n\tif r.cur >= len(r.wal.files) {\n\t\treturn 0, 0, nil, io.EOF\n\t}\n\tcf := r.wal.files[r.cur]\n\n\tet, flag, b, err := r.entry(cf)\n\t// If we reached the end of the reader, advance to the next one\n\t// and close.\n\t// Do not close on the last one as it will still be appended to.\n\tif err == io.EOF && r.cur < len(r.wal.files)-1 {\n\t\t// Current reader completed, close and move to the next one.\n\t\tif err := cf.Close(); err != nil {\n\t\t\treturn 0, 0, nil, err\n\t\t}\n\t\tr.cur++\n\t\treturn r.nextEntry()\n\t}\n\treturn et, flag, b, err\n}", "func (_this *RaftNode) replayWAL() *wal.WAL {\n\tw := _this.openWAL()\n\t_, st, ents, err := w.ReadAll()\n\tif err != nil {\n\t\tlog.Fatalf(\"Failed to read WAL. %v\", err)\n\t}\n\n\t// append to storage so raft starts at the right place in log.\n\t_this.raftStorage.Append(ents)\n\t// send nil once lastIndex is published so client knows commit channel is current\n\tif len(ents) > 0 {\n\t\t_this.lastIndex = ents[len(ents)-1].Index\n\t} else {\n\t\t_this.commitC <- nil\n\t}\n\t_this.raftStorage.SetHardState(st)\n\treturn w\n}", "func (r *walReader) next() bool {\n\tif r.cur >= len(r.wal.files) {\n\t\treturn false\n\t}\n\tcf := r.wal.files[r.cur]\n\n\t// Save position after last valid entry if we have to truncate the WAL.\n\tlastOffset, err := cf.Seek(0, os.SEEK_CUR)\n\tif err != nil {\n\t\tr.err = err\n\t\treturn false\n\t}\n\n\tet, flag, b, err := r.entry(cf)\n\t// If we reached the end of the reader, advance to the next one\n\t// and close.\n\t// Do not close on the last one as it will still be appended to.\n\tif err == io.EOF {\n\t\tif r.cur == len(r.wal.files)-1 {\n\t\t\treturn false\n\t\t}\n\t\t// Current reader completed, close and move to the next one.\n\t\tif err := cf.Close(); err != nil {\n\t\t\tr.err = err\n\t\t\treturn false\n\t\t}\n\t\tr.cur++\n\t\treturn r.next()\n\t}\n\tif err != nil {\n\t\tr.err = err\n\n\t\tif _, ok := err.(walCorruptionErr); ok {\n\t\t\tr.err = r.truncate(lastOffset)\n\t\t}\n\t\treturn false\n\t}\n\n\tr.curType = et\n\tr.curFlag = flag\n\tr.curBuf = b\n\treturn r.err == nil\n}", "func LoadIndexReader(r io.Reader) (*IndexFile, error) {\n\treturn loadIndex(r)\n}", "func (l *Log) Reader() io.Reader {\n\tl.mu.Lock()\n\tdefer l.mu.Unlock()\n\treaders := make([]io.Reader, len(l.segments))\n\tfor i, segment := range l.segments {\n\t\treaders[i] = &originReader{segment.store, 0}\n\t}\n\treturn io.MultiReader(readers...)\n}", "func (w *fileWAL) readEntry() (walEntry, error) {\n\tentry := walEntry{}\n\tvar err error\n\n\t// Make sure we have an open WAL.\n\tif w.f == nil {\n\t\treturn entry, errorInvalidWALFile\n\t}\n\n\t// Read magic value.\n\tmagic := uint32(0)\n\terr = binary.Read(w.f, binary.LittleEndian, &magic)\n\tif err != nil {\n\t\treturn entry, err\n\t}\n\n\tif magic != walMagic {\n\t\treturn entry, errorInvalidWALMagic\n\t}\n\n\t// Read the operation type.\n\terr = binary.Read(w.f, binary.LittleEndian, &entry.operation)\n\tif err != nil {\n\t\treturn entry, err\n\t}\n\n\t// Read the number of rows.\n\tnumRows := uint16(0)\n\terr = binary.Read(w.f, binary.LittleEndian, &numRows)\n\tif err != nil {\n\t\treturn entry, err\n\t}\n\n\tfor i := uint16(0); i < numRows; i++ {\n\t\trow := Row{}\n\n\t\tsourceNameLength, metricNameLength := uint8(0), uint8(0)\n\n\t\t// Read the source and metric name lengths.\n\t\terr = binary.Read(w.f, binary.LittleEndian, &sourceNameLength)\n\t\tif err != nil {\n\t\t\treturn entry, err\n\t\t}\n\t\terr = binary.Read(w.f, binary.LittleEndian, &metricNameLength)\n\t\tif err != nil {\n\t\t\treturn entry, err\n\t\t}\n\n\t\tsourceAndMetricNames := make([]byte, int(sourceNameLength+metricNameLength))\n\n\t\t_, err = w.f.Read(sourceAndMetricNames)\n\t\tif err != nil {\n\t\t\treturn entry, err\n\t\t}\n\n\t\trow.Source = string(sourceAndMetricNames[:int(sourceNameLength)])\n\t\trow.Metric = string(sourceAndMetricNames[int(sourceNameLength):])\n\n\t\ttmpPoint := Point{}\n\t\terr = binary.Read(w.f, binary.LittleEndian, &tmpPoint)\n\t\tif err != nil {\n\t\t\treturn entry, err\n\t\t}\n\n\t\trow.Timestamp = tmpPoint.Timestamp\n\t\trow.Value = tmpPoint.Value\n\n\t\tentry.rows = append(entry.rows, row)\n\t}\n\n\t// We've decoded everything fine.\n\t// We now update lastReadOffset to the current offset\n\t// in the file.\n\tcurrentOffset, err := w.f.Seek(0, 1)\n\tif err != nil {\n\t\treturn entry, err\n\t}\n\n\tw.lastReadOffset = currentOffset\n\n\treturn entry, err\n}", "func (wal *WalDB) ReadAll(snapshot *raftpb.Snapshot) (id *consensus.RaftIdentity, state *raftpb.HardState, ents []raftpb.Entry, err error) {\n\tif id, err = wal.GetIdentity(); err != nil {\n\t\treturn nil, state, ents, err\n\t}\n\n\tstate, err = wal.GetHardState()\n\tif err != nil {\n\t\treturn id, state, ents, ErrWalGetHardState\n\t}\n\n\tcommitIdx := state.Commit\n\tlastIdx, err := wal.GetRaftEntryLastIdx()\n\tif err != nil {\n\t\treturn id, state, ents, ErrWalGetLastIdx\n\t}\n\n\tvar snapIdx, snapTerm uint64\n\tif snapshot != nil {\n\t\tsnapIdx = snapshot.Metadata.Index\n\t\tsnapTerm = snapshot.Metadata.Term\n\t}\n\n\tlogger.Info().Uint64(\"snapidx\", snapIdx).Uint64(\"snapterm\", snapTerm).Uint64(\"commit\", commitIdx).Uint64(\"last\", lastIdx).Msg(\"read all entries of wal\")\n\n\tstart := snapIdx + 1\n\n\tfor i := start; i <= lastIdx; i++ {\n\t\twalEntry, err := wal.GetRaftEntry(i)\n\t\t// if snapshot is nil, initial confchange entry isn't saved to db\n\t\tif err != nil {\n\t\t\tlogger.Error().Err(err).Uint64(\"idx\", i).Msg(\"failed to get raft entry\")\n\t\t\treturn id, state, nil, err\n\t\t}\n\n\t\tif walEntry.Term < snapTerm {\n\t\t\tlogger.Error().Str(\"wal\", walEntry.ToString()).Err(ErrWalEntryTooLowTerm).Msg(\"invalid wal entry\")\n\t\t\treturn id, state, nil, ErrWalEntryTooLowTerm\n\t\t}\n\n\t\traftEntry, err := wal.convertWalToRaft(walEntry)\n\t\tif err != nil {\n\t\t\treturn id, state, nil, err\n\t\t}\n\n\t\tlogger.Debug().Str(\"walentry\", walEntry.ToString()).Msg(\"read wal entry\")\n\t\tents = append(ents, *raftEntry)\n\t}\n\n\treturn id, state, ents, nil\n}", "func (r *chunkReader) Read(data []byte) (int, error) {\n\tbytesToRead := len(data)\n\tr.l.Debug(\"Start cafs reader Read\", zap.Int(\"length\", bytesToRead))\n\n\tif r.lastChunk && r.rdr == nil {\n\t\treturn 0, io.EOF\n\t}\n\tfor {\n\t\tkey := r.keys[r.idx]\n\t\tif r.rdr == nil {\n\t\t\trdr, err := r.fs.Get(context.Background(), r.pather(key))\n\t\t\tif err != nil {\n\t\t\t\treturn r.readSoFar, err\n\t\t\t}\n\t\t\tr.rdr = rdr\n\t\t}\n\n\t\tn, errRead := r.rdr.Read(data[r.readSoFar:])\n\n\t\tdefer func() {\n\t\t\tif r.MetricsEnabled() && errRead == nil {\n\t\t\t\tr.m.Volume.Blobs.IncBlob(\"read\")\n\t\t\t\tr.m.Volume.Blobs.Size(int64(n), \"read\")\n\t\t\t}\n\t\t\tr.l.Debug(\"End cafs reader Read\", zap.Int(\"length\", bytesToRead))\n\t\t}()\n\n\t\tr.currLeaf = append(r.currLeaf, data[r.readSoFar:r.readSoFar+n]...)\n\t\tif errRead != nil {\n\t\t\tr.rdr.Close() // TODO(fred): nice - why are we ignoring errors here?\n\t\t\tr.readSoFar += n\n\t\t\tif errRead == io.EOF { // we reached the end of the stream for this key\n\t\t\t\tr.idx++\n\t\t\t\tr.rdr = nil\n\t\t\t\tr.lastChunk = r.idx == len(r.keys)\n\t\t\t\tif r.withVerifyHash {\n\t\t\t\t\tnodeOffset := r.idx\n\t\t\t\t\tisLastNode := false\n\n\t\t\t\t\t// NOTE: we follow the checksumming scheme adopted by the writer.\n\t\t\t\t\t// The writer behaves in a way a bit unexpected here: not only offets don't start at zero\n\t\t\t\t\t// as one might expect, but the last node is not flagged as the last one\n\t\t\t\t\t// when the content size is aligned with the leaf size.\n\t\t\t\t\tif r.lastChunk && uint32(len(r.currLeaf)) != r.leafSize {\n\t\t\t\t\t\tnodeOffset--\n\t\t\t\t\t\tisLastNode = true\n\t\t\t\t\t}\n\t\t\t\t\tr.l.Debug(\"cafs reader Read: hash verification\", zap.Stringer(\"key\", key))\n\t\t\t\t\tif err := r.verifyHash(key, r.currLeaf, nodeOffset, isLastNode); err != nil {\n\t\t\t\t\t\treturn 0, err\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tif r.lastChunk { // this was the last chunk, so also EOF for this hash\n\t\t\t\t\tif n == bytesToRead {\n\t\t\t\t\t\treturn n, nil\n\t\t\t\t\t}\n\t\t\t\t\treturn r.readSoFar, io.EOF\n\t\t\t\t}\n\t\t\t\t// move on to the next key\n\t\t\t\tr.currLeaf = make([]byte, 0)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\treturn n, errRead\n\t\t}\n\t\t// we filled up the entire byte slice but still have data remaining in the reader,\n\t\t// we should move on to receive the next buffer\n\t\tr.readSoFar += n\n\t\tif r.readSoFar >= bytesToRead {\n\t\t\tr.readSoFar = 0\n\t\t\t// return without error\n\t\t\treturn bytesToRead, nil\n\t\t}\n\t}\n}", "func ReadRotor(rotor *Rotor, readIndex int) (retIndex int) {\n\tbaseIndex := readIndex + GetRotorOffset(rotor)\n\tbaseIndex = offsetAdjust(baseIndex)\n\n\tretLetter := string(rotor.baseWiring[baseIndex])\n\tretIndex = GetAlphabetIndex(retLetter)\n\treturn retIndex\n}", "func (r *walReader) Err() error {\n\treturn r.err\n}", "func NewReader(k pyxtea.Key, r ReaderAtLen) (*Reader, error) {\n\tn := Reader{k: k, r: r}\n\tbuf := [TrailerLen]byte{}\n\t// Read trailer\n\tif _, err := r.ReadAt(buf[:], int64(r.Len()-TrailerLen)); err != nil {\n\t\treturn nil, fmt.Errorf(\"reading trailer: %w\", err)\n\t}\n\tif err := restruct.Unpack(buf[:], binary.LittleEndian, &n.t); err != nil {\n\t\treturn nil, err\n\t}\n\tif n.t.Signature != 0x12 {\n\t\treturn nil, ErrInvalidSignature\n\t}\n\n\treturn &n, nil\n}", "func NewReader(r io.ReaderAt, size int64) (*zip.Reader, error)", "func (s *Store) ReadFrom(r io.Reader) (int64, error) {\n\tnw := false\n\tif len(s.lines) == 0 {\n\t\tnw = true\n\t}\n\tscanner := bufio.NewScanner(r)\n\ts.lines = []*line{}\n\tfor scanner.Scan() {\n\t\ts.lines = append(s.lines, newLine(scanner.Text()))\n\t}\n\tif nw {\n\t\ts.hash()\n\t} else {\n\t\ts.notifyChange()\n\t}\n\treturn int64(s.Len()), scanner.Err()\n}", "func (hp *hdfsProvider) GetObjReader(ctx context.Context, lom *cluster.LOM) (r io.ReadCloser,\n\texpectedCksm *cos.Cksum, errCode int, err error) {\n\tfilePath := filepath.Join(lom.Bck().Props.Extra.HDFS.RefDirectory, lom.ObjName)\n\tfr, err := hp.c.Open(filePath)\n\tif err != nil {\n\t\terrCode, err = hdfsErrorToAISError(err)\n\t\treturn\n\t}\n\tlom.SetCustomKey(cmn.SourceObjMD, apc.HDFS)\n\tsetSize(ctx, fr.Stat().Size())\n\treturn wrapReader(ctx, fr), nil, 0, nil\n}", "func (r *baseReader) initReader() error {\n\tif len(r.buf) <= indexFooterSize {\n\t\treturn fmt.Errorf(\"block length short:%d than footer size: %d\", len(r.buf), indexFooterSize)\n\t}\n\t// read footer(4+4+4)\n\tfooterPos := len(r.buf) - indexFooterSize\n\tr.tagValueBitmapAt = int(binary.LittleEndian.Uint32(r.buf[footerPos : footerPos+4]))\n\tr.offsetsAt = int(binary.LittleEndian.Uint32(r.buf[footerPos+4 : footerPos+8]))\n\tr.crc32CheckSum = binary.LittleEndian.Uint32(r.buf[footerPos+8 : footerPos+12])\n\t// validate offsets\n\tif !sort.IntsAreSorted([]int{\n\t\t0, r.tagValueBitmapAt, r.offsetsAt, footerPos}) {\n\t\treturn fmt.Errorf(\"invalid footer format\")\n\t}\n\t// read keys\n\tkeys := roaring.New()\n\tif err := encoding.BitmapUnmarshal(keys, r.buf[r.tagValueBitmapAt:]); err != nil {\n\t\treturn err\n\t}\n\tr.keys = keys\n\t// read high keys offsets\n\tr.offsets = encoding.NewFixedOffsetDecoder()\n\t_, err := r.offsets.Unmarshal(r.buf[r.offsetsAt:])\n\treturn err\n}", "func (a *Archive) Reader() RecordReader {\n\treturn &EOFReader{&archiveReader{a: a}}\n}", "func (a *Archive) Reader() RecordReader {\n\treturn &EOFReader{&archiveReader{a: a}}\n}", "func (sg *TiKVSingleGather) GetPhysicalIndexReader(schema *expression.Schema, stats *property.StatsInfo, props ...*property.PhysicalProperty) *PhysicalIndexReader {\n\treader := PhysicalIndexReader{}.Init(sg.SCtx(), sg.SelectBlockOffset())\n\treader.SetStats(stats)\n\treader.SetSchema(schema)\n\treader.childrenReqProps = props\n\treturn reader\n}", "func (lf *localFile) Reader() (io.ReadCloser, error) {\n\tif lf.matcher != nil && lf.matcher.Gzip {\n\t\t// We've got the gzipped contents cached in gzipped.\n\t\t// Note: we can't use lf.gzipped directly as a Reader, since we it discards\n\t\t// data after it is read, and we may read it more than once.\n\t\treturn io.NopCloser(bytes.NewReader(lf.gzipped.Bytes())), nil\n\t}\n\t// Not expected to fail since we did it successfully earlier in newLocalFile,\n\t// but could happen due to changes in the underlying filesystem.\n\treturn lf.fs.Open(lf.NativePath)\n}", "func NewJSONReader(ctx context.Context, logger log.Logger, bkt objstore.InstrumentedBucketReader, dir string, id ulid.ULID) (*JSONReader, error) {\n\tcachefn := filepath.Join(dir, id.String(), block.IndexCacheFilename)\n\tjr, err := newFileJSONReader(logger, cachefn)\n\tif err == nil {\n\t\treturn jr, nil\n\t}\n\n\tif !os.IsNotExist(errors.Cause(err)) && errors.Cause(err) != jsonUnmarshalError {\n\t\treturn nil, errors.Wrap(err, \"read index cache\")\n\t}\n\n\t// Just in case the dir was not created.\n\tif err := os.MkdirAll(filepath.Join(dir, id.String()), os.ModePerm); err != nil {\n\t\treturn nil, errors.Wrap(err, \"create dir\")\n\t}\n\n\t// Try to download index cache file from object store.\n\tif err = objstore.DownloadFile(ctx, logger, bkt.ReaderWithExpectedErrs(bkt.IsObjNotFoundErr), filepath.Join(id.String(), block.IndexCacheFilename), cachefn); err == nil {\n\t\treturn newFileJSONReader(logger, cachefn)\n\t}\n\n\tif !bkt.IsObjNotFoundErr(errors.Cause(err)) && errors.Cause(err) != jsonUnmarshalError {\n\t\treturn nil, errors.Wrap(err, \"download index cache file\")\n\t}\n\n\t// No cache exists on disk yet, build it from the downloaded index and retry.\n\tfn := filepath.Join(dir, id.String(), block.IndexFilename)\n\n\tif err := objstore.DownloadFile(ctx, logger, bkt, filepath.Join(id.String(), block.IndexFilename), fn); err != nil {\n\t\treturn nil, errors.Wrap(err, \"download index file\")\n\t}\n\n\tdefer func() {\n\t\tif rerr := os.Remove(fn); rerr != nil {\n\t\t\tlevel.Error(logger).Log(\"msg\", \"failed to remove temp index file\", \"path\", fn, \"err\", rerr)\n\t\t}\n\t}()\n\n\tif err := WriteJSON(logger, fn, cachefn); err != nil {\n\t\treturn nil, errors.Wrap(err, \"write index cache\")\n\t}\n\n\treturn newFileJSONReader(logger, cachefn)\n}", "func (l *LeechedReadCloser) Read(b []byte) (n int, err error) {\n\tspaceLeft := l.maxBodyLogSize - l.loggedBytesCount\n\tif spaceLeft > 0 {\n\t\t// Let's read the request into our Logger (not all of it maybe), but also let's make sure that\n\t\t// we'll be able to to copy all the content we read in l.data into b\n\t\tn, err := l.originalReadCloser.Read(l.data[l.loggedBytesCount : l.loggedBytesCount+min(int64(len(b)), spaceLeft)])\n\n\t\t// And copy what was read into the original slice\n\t\tcopy(b, l.data[l.loggedBytesCount:l.loggedBytesCount+int64(n)])\n\n\t\t// Let's not forget to increment the pointer on the currently logged amount of bytes\n\t\tl.loggedBytesCount += int64(n)\n\n\t\t// And return what the Read() call we did on the original ReadCloser just returned, shhhhh\n\t\treturn n, err\n\t}\n\n\t// Our leecher is full ? Nevermind, let's just call read on the original Reader. Apart from an\n\t// additional level in the call stack and an if statement, we have no overhead for large bodies :)\n\treturn l.originalReadCloser.Read(b)\n}", "func ReadRotorBackwards(rotor *Rotor, readIndex int) (retIndex int) {\n\t// Alphabet: ABCDEFGHIJKLMNOPQRSTUVWXYZ\n\t// Rotor Base: EKMFLGDQVZNTOWYHXUSPAIBRCJ\n\tbaseIndex := readIndex + (-1)*GetRotorOffset(rotor)\n\tbaseIndex = offsetAdjust(baseIndex)\n\n\tretLetter := GetAlphabetLetter(baseIndex)\n\tretIndex = GetAlphabetIndex(retLetter)\n\treturn retIndex\n}", "func (api *API) GetReader(path string) (r io.Reader, statusCode int, err error) {\n\treq, err := http.NewRequest(\"GET\", api.host+path, nil)\n\tif err != nil {\n\t\treturn nil, 0, err\n\t}\n\n\treturn api.do(req)\n}", "func (gWal *GenericWAL) Recover() error {\n entryCh := make(chan *Entry)\n quitCh := make(chan struct{})\n defer close(quitCh)\n go func() {\n if err := gWal.wal.StartRecovery(entryCh, quitCh); err != nil {\n glog.Errorf(\"could not recover state from local wal :: %v\", err)\n entryCh <- nil\n }\n }()\n\n count := 0\n for entry := range entryCh {\n if entry == nil {\n glog.Errorf(\"wal recovery channel is closed unexpectedly\")\n return fmt.Errorf(\"wal error\")\n }\n count++\n\n switch entry.Type {\n case CTypeDone:\n glog.Infof(\"wal recovery is complete because last record is read\")\n close(entryCh)\n\n case CTypeData:\n gWal.updateRecoveryMap(false /* checkpoint */, int64(entry.LSN),\n entry.Data)\n glog.V(1).Infof(\"recovered a delta record with lsn %v\", entry.LSN)\n\n case CTypeCheckpoint:\n gWal.updateRecoveryMap(true /* checkpoint */, -1, entry.Data)\n glog.V(1).Infof(\"recovered a checkpoint record with lsn %v\", entry.LSN)\n\n case CTypeError:\n glog.Errorf(\"wal recovery encountered an unrecoverable error\")\n return fmt.Errorf(\"wal error\")\n\n default:\n glog.Errorf(\"wal recovery received an unknown or invalid record\")\n return fmt.Errorf(\"wal error\")\n }\n }\n\n return nil\n}", "func (fr *FileReader) readerForOffset(off int64) (io.ReadCloser, error) {\n\tif off < 0 {\n\t\tpanic(\"negative offset\")\n\t}\n\tif off >= fr.size {\n\t\treturn eofReader, nil\n\t}\n\toffRemain := off\n\tparts := fr.ss.Parts\n\tfor len(parts) > 0 && parts[0].Size <= uint64(offRemain) {\n\t\toffRemain -= int64(parts[0].Size)\n\t\tparts = parts[1:]\n\t}\n\tif len(parts) == 0 {\n\t\treturn eofReader, nil\n\t}\n\tp0 := parts[0]\n\tvar rsc blobref.ReadSeekCloser\n\tvar err error\n\tswitch {\n\tcase p0.BlobRef != nil && p0.BytesRef != nil:\n\t\treturn nil, fmt.Errorf(\"part illegally contained both a blobRef and bytesRef\")\n\tcase p0.BlobRef == nil && p0.BytesRef == nil:\n\t\treturn &nZeros{int(p0.Size - uint64(offRemain))}, nil\n\tcase p0.BlobRef != nil:\n\t\trsc, _, err = fr.fetcher.Fetch(p0.BlobRef)\n\tcase p0.BytesRef != nil:\n\t\trsc, err = NewFileReader(fr.fetcher, p0.BytesRef)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\toffRemain += int64(p0.Offset)\n\tif offRemain > 0 {\n\t\tnewPos, err := rsc.Seek(offRemain, os.SEEK_SET)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif newPos != offRemain {\n\t\t\tpanic(\"Seek didn't work\")\n\t\t}\n\t}\n\treturn struct {\n\t\tio.Reader\n\t\tio.Closer\n\t}{\n\t\tio.LimitReader(rsc, int64(p0.Size)),\n\t\trsc,\n\t}, nil\n}", "func getreader(bn int, vname string) (io.Reader, io.Closer) {\n\tfn := config.BucketPath(bn, sourcedir)\n\tfn = path.Join(fn, fmt.Sprintf(\"%s.bin.sz\", vname))\n\tfid, err := os.Open(fn)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\trdr := snappy.NewReader(fid)\n\treturn rdr, fid\n}", "func NewWALDecoder(rd io.Reader) *WALDecoder {\n\treturn &WALDecoder{rd}\n}", "func NewReader(p []byte, _ uint64, readers []io.ReadCloser) (io.ReadCloser, error) {\n\tif len(readers) != 1 {\n\t\treturn nil, errors.New(\"delta: need exactly one reader\")\n\t}\n\n\tif len(p) != 1 {\n\t\treturn nil, errors.New(\"delta: not enough properties\")\n\t}\n\n\treturn &readCloser{\n\t\trc: readers[0],\n\t\tdelta: int(p[0] + 1),\n\t}, nil\n}", "func TestInOutRW(t *testing.T) {\n\tvar els = []string{\"ab\", \"bc\", \"de\", \"fgh\"}\n\tvar outs = []string{\"ab\", \"abbc\", \"abbcde\", \"abbcdefgh\"}\n\n\tl := NewLine()\n\tt.Logf(\"%v %v %v\", els, outs, l)\n\tfor i := range els {\n\t\ts := strings.Join(els[:i+1], \"\")\n\t\tl.Write([]byte(s))\n\t\tb, err := l.ReadAll()\n\t\tif err != nil {\n\t\t\tt.Errorf(\"ReadAll of %s: got %v, want nil\", s, err)\n\t\t}\n\t\tif string(b) != outs[i] {\n\t\t\tt.Errorf(\"Read back %s: got %s, want %s\", s, string(b), s)\n\t\t}\n\t}\n}", "func OpenReader(name string) (*zip.ReadCloser, error)", "func (m *InMemoryRepository) Reader(u fyne.URI) (fyne.URIReadCloser, error) {\n\tpath := u.Path()\n\n\tif path == \"\" {\n\t\treturn nil, fmt.Errorf(\"invalid path '%s'\", path)\n\t}\n\n\t_, ok := m.Data[path]\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"no such path '%s' in InMemoryRepository\", path)\n\t}\n\n\treturn &nodeReaderWriter{path: path, repo: m}, nil\n}", "func TestReadAllNoSupportedDecrypter(t *testing.T) {\n\tmetadata, entries, snapshot := makeWALData(1, 1)\n\tfor i, entry := range entries {\n\t\tr := api.MaybeEncryptedRecord{Data: entry.Data, Algorithm: api.MaybeEncryptedRecord_Algorithm(-3)}\n\t\tdata, err := r.Marshal()\n\t\trequire.NoError(t, err)\n\t\tentries[i].Data = data\n\t}\n\n\ttempdir := createWithWAL(t, OriginalWAL, metadata, snapshot, entries)\n\tdefer os.RemoveAll(tempdir)\n\n\tc := NewWALFactory(encryption.NoopCrypter, encryption.NoopCrypter)\n\twrapped, err := c.Open(tempdir, snapshot)\n\trequire.NoError(t, err)\n\tdefer wrapped.Close()\n\n\t_, _, _, err = wrapped.ReadAll()\n\trequire.Error(t, err)\n\tdefer wrapped.Close()\n}", "func (r *Reference) NewReader(ctx context.Context, path iosrc.URI, readspan nano.Span) (*Reader, error) {\n\tsegspan := r.Span()\n\tspan := segspan.Intersect(readspan)\n\tobjectPath := r.RowObjectPath(path)\n\tif span.Dur == 0 {\n\t\treturn nil, fmt.Errorf(\"segment reader: segment does not intersect provided span: %s chunkspan %v readspan %v\", path, segspan, readspan)\n\t}\n\treader, err := iosrc.NewReader(ctx, objectPath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsr := &Reader{\n\t\tReader: reader,\n\t\tCloser: reader,\n\t\tTotalBytes: r.Size,\n\t\tReadBytes: r.Size,\n\t}\n\tif span == segspan {\n\t\treturn sr, nil\n\t}\n\ts, err := seekindex.Open(ctx, r.SeekObjectPath(path))\n\tif err != nil {\n\t\tif zqe.IsNotFound(err) {\n\t\t\treturn sr, nil\n\t\t}\n\t\treturn nil, err\n\t}\n\tdefer s.Close()\n\trg, err := s.Lookup(ctx, span)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\trg = rg.TrimEnd(sr.TotalBytes)\n\tsr.ReadBytes = rg.Size()\n\tsr.Reader, err = rg.LimitReader(reader)\n\treturn sr, err\n}", "func NewReader(store EntryStore) Reader {\n\treturn defaultReader{\n\t\tstore: store,\n\t}\n}", "func (r *WALReader) Read(cb func(WALEntry) error) error {\n\tfor _, file := range r.files {\n\t\tif err := r.readFile(file, cb); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (p *Project) LogReader(full bool, generation int) (io.ReadCloser, error) {\n\tlogFile := workdir.LogFile(p.Name, generation)\n\tf, err := os.Open(logFile)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn &EmptyReadCloser{}, nil\n\t\t}\n\t\treturn nil, err\n\t}\n\tif full {\n\t\treturn f, nil\n\t}\n\treturn headreader.New(f, 10000), nil // first 10000 bytes\n}", "func NewReader(seed string) (io.Reader, error) {\n\tsum := sha256.Sum256([]byte(seed))\n\tblock, err := aes.NewCipher(sum[:])\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tiv := make([]byte, block.BlockSize())\n\treturn undoAGL{cipher.StreamReader{\n\t\tS: cipher.NewCTR(block, iv),\n\t\tR: devZero{},\n\t}}, nil\n}", "func (b *balanceMonitor) getReader() (solanaClient.Reader, error) {\n\tif b.reader == nil {\n\t\tvar err error\n\t\tb.reader, err = b.newReader()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn b.reader, nil\n}", "func (r *Reader) Next() (*index.Header, error) {\n\thdr, err := r.ir.Next()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// Convert index tar header to corresponding content tar header.\n\tindexToContentHeader(hdr)\n\t// Store header and reset reader for if a read is attempted.\n\tr.hdr = hdr\n\tr.tr = nil\n\treturn hdr, nil\n}", "func NewWithReader(r io.Reader) []*Index {\n\tvar (\n\t\tz = html.NewTokenizer(r)\n\n\t\tcurrentLevel int\n\t\tcurrentText string\n\t\tcurrentLinkText string\n\t\tcurrentAnchor string\n\t\tnodeDeep int\n\n\t\tindexs []*Index\n\t)\n\tfor {\n\t\ttoken := z.Next()\n\t\tif token == html.ErrorToken {\n\t\t\tbreak\n\t\t}\n\t\tif token == html.EndTagToken {\n\t\t\tif nodeDeep == 1 && currentLevel > 0 {\n\t\t\t\tindexs = append(indexs, &Index{\n\t\t\t\t\tLevel: currentLevel,\n\t\t\t\t\tTitle: currentText,\n\t\t\t\t\tLink: currentLinkText,\n\t\t\t\t\tAnchor: currentAnchor,\n\t\t\t\t})\n\t\t\t\tcurrentLevel = 0\n\t\t\t\tcurrentText = \"\"\n\t\t\t\tcurrentLinkText = \"\"\n\t\t\t\tcurrentAnchor = \"\"\n\t\t\t\tnodeDeep--\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tif token == html.StartTagToken {\n\t\t\tname, hasAttr := z.TagName()\n\t\t\tlv := parseIndexLevel(name)\n\n\t\t\tif lv > 0 {\n\t\t\t\tcurrentLevel = lv\n\t\t\t\tif hasAttr {\n\t\t\t\t\tfor {\n\t\t\t\t\t\tk, v, isMore := z.TagAttr()\n\t\t\t\t\t\tif bytes.Equal(k, []byte(\"id\")) {\n\t\t\t\t\t\t\tcurrentAnchor = string(v)\n\t\t\t\t\t\t}\n\t\t\t\t\t\tif !isMore {\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tnodeDeep++\n\t\t\t}\n\n\t\t\tif currentLevel > 0 && string(name) == \"a\" {\n\t\t\t\tif hasAttr {\n\t\t\t\t\tfor {\n\t\t\t\t\t\tk, v, isMore := z.TagAttr()\n\t\t\t\t\t\tif bytes.Equal(k, []byte(\"href\")) {\n\t\t\t\t\t\t\tcurrentLinkText = string(v)\n\t\t\t\t\t\t}\n\t\t\t\t\t\tif !isMore {\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif token == html.TextToken && currentLevel > 0 {\n\t\t\tcurrentText += string(z.Text())\n\t\t}\n\t}\n\tindexs = assembleIndex(indexs)\n\treturn indexs\n}", "func (u *uploader) nextReader() (readerAtSeeker, int, []byte, error) {\n\ttype readerAtSeeker interface {\n\t\tio.ReaderAt\n\t\tio.ReadSeeker\n\t}\n\tswitch r := u.in.Body.(type) {\n\tcase readerAtSeeker:\n\t\tvar err error\n\n\t\tn := u.cfg.PartSize\n\t\tif u.totalSize >= 0 {\n\t\t\tbytesLeft := u.totalSize - u.readerPos\n\n\t\t\tif bytesLeft <= u.cfg.PartSize {\n\t\t\t\terr = io.EOF\n\t\t\t\tn = bytesLeft\n\t\t\t}\n\t\t}\n\t\treader := io.NewSectionReader(r, u.readerPos, n)\n\t\tu.readerPos += n\n\n\t\treturn reader, int(n), nil, err\n\n\tdefault:\n\t\tpart := u.bufferPool.Get().([]byte)\n\t\tn, err := readFillBuf(r, part)\n\t\tu.readerPos += int64(n)\n\n\t\treturn bytes.NewReader(part[0:n]), n, part, err\n\t}\n}", "func NewReader(r io.Reader, key []byte) (wire.Reader, map[string]string, error) {\n\t// Read the metadata with the hash.\n\th := hmac.New(sha256.New, key)\n\tmetadata, err := metadata(r, h)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\t// Determine image compression state. If the metadata doesn't contain\n\t// compression information the default behavior is the \"compressed\" state\n\t// because the default behavior used to be to always compress.\n\tcompression, err := CompressionLevelFromMetadata(metadata)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\t// Pick correct reader\n\tvar cr wire.Reader\n\n\tif compression == CompressionLevelFlateBestSpeed {\n\t\tcr, err = compressio.NewReader(r, key)\n\t} else if compression == CompressionLevelNone {\n\t\tcr, err = compressio.NewSimpleReader(r, key)\n\t} else {\n\t\t// Should never occur, as it has the default path.\n\t\treturn nil, nil, fmt.Errorf(\"metadata contains invalid compression flag value: %v\", compression)\n\t}\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn cr, metadata, nil\n}", "func NewReader(r io.Reader) io.ReadCloser {\n\treturn &reader{\n\t\tlz4Stream: C.LZ4_createStreamDecode(),\n\t\tunderlyingReader: r,\n\t\tisLeft: true,\n\t\t// double buffer needs to use C.malloc to make sure the same memory address\n\t\t// allocate buffers in go memory will fail randomly since GC may move the memory\n\t\tleft: C.malloc(boudedStreamingBlockSize),\n\t\tright: C.malloc(boudedStreamingBlockSize),\n\t}\n}", "func (i *Index) Read(in int64) (out uint32, pos uint64, err error) {\n\tif i.size == 0 {\n\t\treturn 0, 0, lib.Wrap(io.EOF, \"Unable read Index file(it's empty)\")\n\t}\n\n\tif in == -1 {\n\t\tout = uint32((i.size / entWidth) - 1)\n\t} else {\n\t\tout = uint32(in)\n\t}\n\n\tpos = uint64(out) * entWidth\n\n\tif i.size < pos+entWidth {\n\t\treturn 0, 0, lib.Wrap(io.EOF, \"Position is out of file's amount\")\n\t}\n\n\tout = enc.Uint32(i.mmap[pos : pos+offWidth])\n\tpos = enc.Uint64(i.mmap[pos+offWidth : pos+entWidth])\n\n\treturn\n}", "func (base *Base) Reader(ctx context.Context, path string, offset int64) (io.ReadCloser, error) {\n\tctx, done := dcontext.WithTrace(ctx)\n\tdefer done(\"%s.Reader(%q, %d)\", base.Name(), path, offset)\n\n\tif offset < 0 {\n\t\treturn nil, storagedriver.InvalidOffsetError{Path: path, Offset: offset, DriverName: base.StorageDriver.Name()}\n\t}\n\n\tif !storagedriver.PathRegexp.MatchString(path) {\n\t\treturn nil, storagedriver.InvalidPathError{Path: path, DriverName: base.StorageDriver.Name()}\n\t}\n\n\trc, e := base.StorageDriver.Reader(ctx, path, offset)\n\treturn rc, base.setDriverName(e)\n}", "func (r *Reader) R(i int) io.Reader {\n\treturn r.r[len(r.r)-1-i]\n}", "func initZipReader(r io.ReaderAt, size int64) (*zip.Reader, error) {\n\tzr, err := zip.NewReader(r, size)\n\tif err == nil || !errors.Is(err, zip.ErrFormat) {\n\t\treturn zr, err\n\t}\n\tconst BUFSIZE = 4096\n\tvar buf [BUFSIZE + 4]byte\n\tfor i := int64(0); (i-1)*BUFSIZE < size; i++ {\n\t\tlen, err := r.ReadAt(buf[:], i*BUFSIZE)\n\t\tif err != nil && err != io.EOF {\n\t\t\tbreak\n\t\t}\n\t\tn := 0\n\t\tfor {\n\t\t\tif isZoneInfoFile(buf[n:len]) {\n\t\t\t\treturn nil, &ZoneInfoFileError{}\n\t\t\t}\n\t\t\tm := bytes.Index(buf[n:len], []byte(fileHeaderSignatureString))\n\t\t\tif m == -1 {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\toff := i*BUFSIZE + int64(n+m)\n\t\t\tzipSize := size - int64(off)\n\t\t\tsr := io.NewSectionReader(r, int64(off), zipSize)\n\t\t\tif zr, ze := zip.NewReader(sr, zipSize+1); ze == nil {\n\t\t\t\treturn zr, nil\n\t\t\t}\n\t\t\tn += m + 1\n\t\t}\n\t\tif err == io.EOF {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn nil, errors.New(\"No zip file found\")\n}", "func TestReader(r io.Reader, content []byte) error {\n\tif len(content) > 0 {\n\t\tn, err := r.Read(nil)\n\t\tif n != 0 || err != nil {\n\t\t\treturn fmt.Errorf(\"Read(0) = %d, %v, want 0, nil\", n, err)\n\t\t}\n\t}\n\n\tdata, err := io.ReadAll(&smallByteReader{r: r})\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !bytes.Equal(data, content) {\n\t\treturn fmt.Errorf(\"ReadAll(small amounts) = %q\\n\\twant %q\", data, content)\n\t}\n\tn, err := r.Read(make([]byte, 10))\n\tif n != 0 || err != io.EOF {\n\t\treturn fmt.Errorf(\"Read(10) at EOF = %v, %v, want 0, EOF\", n, err)\n\t}\n\n\tif r, ok := r.(io.ReadSeeker); ok {\n\t\t// Seek(0, 1) should report the current file position (EOF).\n\t\tif off, err := r.Seek(0, 1); off != int64(len(content)) || err != nil {\n\t\t\treturn fmt.Errorf(\"Seek(0, 1) from EOF = %d, %v, want %d, nil\", off, err, len(content))\n\t\t}\n\n\t\t// Seek backward partway through file, in two steps.\n\t\t// If middle == 0, len(content) == 0, can't use the -1 and +1 seeks.\n\t\tmiddle := len(content) - len(content)/3\n\t\tif middle > 0 {\n\t\t\tif off, err := r.Seek(-1, 1); off != int64(len(content)-1) || err != nil {\n\t\t\t\treturn fmt.Errorf(\"Seek(-1, 1) from EOF = %d, %v, want %d, nil\", -off, err, len(content)-1)\n\t\t\t}\n\t\t\tif off, err := r.Seek(int64(-len(content)/3), 1); off != int64(middle-1) || err != nil {\n\t\t\t\treturn fmt.Errorf(\"Seek(%d, 1) from %d = %d, %v, want %d, nil\", -len(content)/3, len(content)-1, off, err, middle-1)\n\t\t\t}\n\t\t\tif off, err := r.Seek(+1, 1); off != int64(middle) || err != nil {\n\t\t\t\treturn fmt.Errorf(\"Seek(+1, 1) from %d = %d, %v, want %d, nil\", middle-1, off, err, middle)\n\t\t\t}\n\t\t}\n\n\t\t// Seek(0, 1) should report the current file position (middle).\n\t\tif off, err := r.Seek(0, 1); off != int64(middle) || err != nil {\n\t\t\treturn fmt.Errorf(\"Seek(0, 1) from %d = %d, %v, want %d, nil\", middle, off, err, middle)\n\t\t}\n\n\t\t// Reading forward should return the last part of the file.\n\t\tdata, err := io.ReadAll(&smallByteReader{r: r})\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"ReadAll from offset %d: %v\", middle, err)\n\t\t}\n\t\tif !bytes.Equal(data, content[middle:]) {\n\t\t\treturn fmt.Errorf(\"ReadAll from offset %d = %q\\n\\twant %q\", middle, data, content[middle:])\n\t\t}\n\n\t\t// Seek relative to end of file, but start elsewhere.\n\t\tif off, err := r.Seek(int64(middle/2), 0); off != int64(middle/2) || err != nil {\n\t\t\treturn fmt.Errorf(\"Seek(%d, 0) from EOF = %d, %v, want %d, nil\", middle/2, off, err, middle/2)\n\t\t}\n\t\tif off, err := r.Seek(int64(-len(content)/3), 2); off != int64(middle) || err != nil {\n\t\t\treturn fmt.Errorf(\"Seek(%d, 2) from %d = %d, %v, want %d, nil\", -len(content)/3, middle/2, off, err, middle)\n\t\t}\n\n\t\t// Reading forward should return the last part of the file (again).\n\t\tdata, err = io.ReadAll(&smallByteReader{r: r})\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"ReadAll from offset %d: %v\", middle, err)\n\t\t}\n\t\tif !bytes.Equal(data, content[middle:]) {\n\t\t\treturn fmt.Errorf(\"ReadAll from offset %d = %q\\n\\twant %q\", middle, data, content[middle:])\n\t\t}\n\n\t\t// Absolute seek & read forward.\n\t\tif off, err := r.Seek(int64(middle/2), 0); off != int64(middle/2) || err != nil {\n\t\t\treturn fmt.Errorf(\"Seek(%d, 0) from EOF = %d, %v, want %d, nil\", middle/2, off, err, middle/2)\n\t\t}\n\t\tdata, err = io.ReadAll(r)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"ReadAll from offset %d: %v\", middle/2, err)\n\t\t}\n\t\tif !bytes.Equal(data, content[middle/2:]) {\n\t\t\treturn fmt.Errorf(\"ReadAll from offset %d = %q\\n\\twant %q\", middle/2, data, content[middle/2:])\n\t\t}\n\t}\n\n\tif r, ok := r.(io.ReaderAt); ok {\n\t\tdata := make([]byte, len(content), len(content)+1)\n\t\tfor i := range data {\n\t\t\tdata[i] = 0xfe\n\t\t}\n\t\tn, err := r.ReadAt(data, 0)\n\t\tif n != len(data) || err != nil && err != io.EOF {\n\t\t\treturn fmt.Errorf(\"ReadAt(%d, 0) = %v, %v, want %d, nil or EOF\", len(data), n, err, len(data))\n\t\t}\n\t\tif !bytes.Equal(data, content) {\n\t\t\treturn fmt.Errorf(\"ReadAt(%d, 0) = %q\\n\\twant %q\", len(data), data, content)\n\t\t}\n\n\t\tn, err = r.ReadAt(data[:1], int64(len(data)))\n\t\tif n != 0 || err != io.EOF {\n\t\t\treturn fmt.Errorf(\"ReadAt(1, %d) = %v, %v, want 0, EOF\", len(data), n, err)\n\t\t}\n\n\t\tfor i := range data {\n\t\t\tdata[i] = 0xfe\n\t\t}\n\t\tn, err = r.ReadAt(data[:cap(data)], 0)\n\t\tif n != len(data) || err != io.EOF {\n\t\t\treturn fmt.Errorf(\"ReadAt(%d, 0) = %v, %v, want %d, EOF\", cap(data), n, err, len(data))\n\t\t}\n\t\tif !bytes.Equal(data, content) {\n\t\t\treturn fmt.Errorf(\"ReadAt(%d, 0) = %q\\n\\twant %q\", len(data), data, content)\n\t\t}\n\n\t\tfor i := range data {\n\t\t\tdata[i] = 0xfe\n\t\t}\n\t\tfor i := range data {\n\t\t\tn, err = r.ReadAt(data[i:i+1], int64(i))\n\t\t\tif n != 1 || err != nil && (i != len(data)-1 || err != io.EOF) {\n\t\t\t\twant := \"nil\"\n\t\t\t\tif i == len(data)-1 {\n\t\t\t\t\twant = \"nil or EOF\"\n\t\t\t\t}\n\t\t\t\treturn fmt.Errorf(\"ReadAt(1, %d) = %v, %v, want 1, %s\", i, n, err, want)\n\t\t\t}\n\t\t\tif data[i] != content[i] {\n\t\t\t\treturn fmt.Errorf(\"ReadAt(1, %d) = %q want %q\", i, data[i:i+1], content[i:i+1])\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (r *regulator) Reader(ctx context.Context, path string, offset int64) (io.ReadCloser, error) {\n\tr.enter()\n\tdefer r.exit()\n\n\treturn r.StorageDriver.Reader(ctx, path, offset)\n}", "func (tb *tableManager) read(keyIn uint64) (*Block, error) {\n\tentry, err := tb.getEntry(keyIn)\n\tif err != nil {\n\t\tlog.Println(\"Could not obtain entry.\")\n\t\treturn nil, errors.New(\"Could not obtain entry.\")\n\t}\n\tif entry.flags&flagRemove != 0 {\n\t\t// dataBase should be able to tell if a dirtyKey is marked\n\t\t// for removal so it can write it as removed in log.\n\t\treturn nil, nil\n\t}\n\ttb.updateLRUCacheHead(entry)\n\treturn entry.block, nil\n}", "func (s *Storage) GetLifeline(objRef []byte, fromIndex *string, pulseNumberLt, pulseNumberGt, timestampLte, timestampGte *int64, limit, offset int, sortByIndexAsc bool) ([]models.Record, int, error) {\n\ttimer := prometheus.NewTimer(GetLifelineDuration)\n\tdefer timer.ObserveDuration()\n\n\tquery := s.db.Model(&models.Record{}).Where(\"object_reference = ?\", objRef).Where(\"type = ?\", models.State)\n\n\tquery = filterByPulse(query, pulseNumberLt, pulseNumberGt)\n\n\tquery = filterByTimestamp(query, timestampLte, timestampGte)\n\n\tvar err error\n\tif fromIndex != nil {\n\t\tquery, err = filterRecordsByIndex(query, *fromIndex, sortByIndexAsc)\n\t\tif err != nil {\n\t\t\treturn nil, 0, err\n\t\t}\n\t}\n\n\tquery = sortRecordsByDirection(query, sortByIndexAsc)\n\n\trecords, total, err := getRecords(query, limit, offset)\n\tif err != nil {\n\t\treturn nil, 0, errors.Wrapf(err, \"error while select records for object %v from db\", objRef)\n\t}\n\treturn records, total, nil\n}", "func Open(opts *Opts) (*WAL, error) {\n\twal := &WAL{\n\t\tfilebased: filebased{\n\t\t\tdir: opts.Dir,\n\t\t\tfileFlags: os.O_CREATE | os.O_APPEND | os.O_WRONLY,\n\t\t\th: newHash(),\n\t\t\tlog: golog.LoggerFor(\"wal\"),\n\t\t},\n\t\tclosed: make(chan interface{}),\n\t}\n\n\t// Append a sentinel to the most recent file (just in case it wasn't closed correctly)\n\tsentinelErr := wal.forEachSegmentInReverse(func(fi os.FileInfo, first bool, last bool) (bool, error) {\n\t\tfile, err := os.OpenFile(filepath.Join(wal.dir, fi.Name()), os.O_APPEND|os.O_WRONLY, 0600)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\t\tdefer file.Close()\n\t\t_, err = file.Write(sentinelBytes)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\t\terr = file.Sync()\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\t\treturn false, nil\n\t})\n\tif sentinelErr != nil {\n\t\treturn nil, errors.New(\"Unable to append sentinel to old segment: %v\", sentinelErr)\n\t}\n\n\t// Advance wal to get a new segment file\n\terr := wal.advance()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif opts.SyncInterval <= 0 {\n\t\twal.syncImmediate = true\n\t} else {\n\t\tif opts.MaxMemoryBacklog > 0 {\n\t\t\twal.log.Debugf(\"Enabling in-memory backlog up to %d buffers\", opts.MaxMemoryBacklog)\n\t\t\twal.backlog = make(chan [][]byte, opts.MaxMemoryBacklog)\n\t\t\twal.backlogFinished = make(chan interface{})\n\t\t\tgo wal.writeAsync()\n\t\t}\n\t\tgo wal.sync(opts.SyncInterval)\n\t}\n\n\treturn wal, nil\n}", "func ReadStore(in io.Reader, passphrase []byte) (s *store.Store, err error) {\n\ts = store.NewStore()\n\tbuf := new(bytes.Buffer)\n\n\t// Marshal header and redirect data to buffer\n\tr := io.TeeReader(in, buf)\n\tif err = s.Header.Unmarshal(r); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Initialize crypto stream and HMAC\n\tstream, mac := initStream(passphrase, &s.Header)\n\n\t// Update HMAC by feeding previously read header\n\tio.Copy(mac, buf)\n\n\t// Check header HMAC\n\tif ok, err := checkHMAC(io.TeeReader(in, mac), mac); err != nil {\n\t\treturn nil, err\n\t} else if !ok {\n\t\treturn nil, ErrWrongPass\n\t}\n\n\t// Copy remainder of input stream to buffer\n\tif _, err = io.Copy(buf, in); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Construct ciphertext reader for entry data block\n\tn := int64(buf.Len() - mac.Size())\n\tr = io.LimitReader(io.TeeReader(buf, mac), n)\n\tr = cipher.StreamReader{S: stream, R: r}\n\n\t// Read JSON-encoded entries\n\tdec := json.NewDecoder(r)\n\tif err := dec.Decode(&s.Entries); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Check store HMAC\n\tif ok, err := checkHMAC(buf, mac); err != nil {\n\t\treturn nil, err\n\t} else if !ok {\n\t\treturn nil, ErrWrongPass\n\t}\n\n\treturn\n}", "func New(br *bam.Reader, idxio io.Reader) (*Reader, error) {\n\tidx, err := bam.ReadIndex(idxio)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbx := &Reader{Reader: br, idx: idx}\n\n\tbx.refs = make(map[string]*sam.Reference)\n\tfor _, r := range br.Header().Refs() {\n\t\tbx.refs[r.Name()] = r\n\t}\n\treturn bx, nil\n}", "func (lw *LogWatcher) Read(buf []byte) (int, error) {\n\tvar err error\n\tvar fInfo os.FileInfo\n\n\tif fInfo, err = os.Stat(lw.Filename); err == nil {\n\t\tdoRead := false\n\t\tnewFile := false\n\n\t\tlw.debugf(\"logwatcher.Read: fInfo: %+v\", fInfo)\n\n\t\tif lw.lastFInfo == nil {\n\t\t\tnewFile = true\n\t\t\t// User can pass a checkpointed position.\n\t\t\tlw.lastPos = lw.StartPosition\n\t\t\tlw.debugf(\"logwatcher.Read: newfile, lastpos = %d\", lw.lastPos)\n\t\t} else if !os.SameFile(lw.lastFInfo, fInfo) {\n\t\t\tnewFile = true\n\t\t\tlw.debugf(\"logwatcher.Read: not samefile.\")\n\t\t} else if fInfo.Size() < lw.lastFInfo.Size() {\n\t\t\t// Truncated\n\t\t\tlw.lastPos = 0\n\t\t\tnewFile = true\n\t\t\tlw.debugf(\"logwatcher.Read: truncated.\")\n\t\t} else if fInfo.Size() > lw.lastFInfo.Size() {\n\t\t\t// logfile grew, read it\n\t\t\tdoRead = true\n\t\t\tlw.debugf(\"logwatcher.Read: bigger file reading.\")\n\t\t} else if fInfo.Size() > lw.lastPos {\n\t\t\t// logfile grew, read it\n\t\t\tdoRead = true\n\t\t\tlw.debugf(\"logwatcher.Read: stuff left to read, reading.\")\n\t\t} else {\n\t\t\t// same size, don't read\n\t\t\tlw.debugf(\"logwatcher.Read: no change, ignoring.\")\n\t\t\terr = io.EOF\n\t\t}\n\n\t\tif newFile && fInfo.Size() > 0 {\n\t\t\t// Reset pointers\n\t\t\tlw.lastFInfo = nil\n\t\t\tdoRead = true\n\t\t\tlw.debugf(\"logwatcher.Read: bigger file reading.\")\n\t\t}\n\n\t\tif doRead {\n\t\t\treturn lw.read(fInfo, buf)\n\t\t}\n\t}\n\tlw.debugf(\"logwatcher.Read: Returning 0, %v\", err)\n\treturn 0, err\n}", "func makeBlockchainFromWAL(wal WAL) ([]*types.Block, []*types.Commit, error) {\n\tvar height int64\n\n\t// Search for height marker\n\tgr, found, err := wal.SearchForEndHeight(height, &WALSearchOptions{})\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tif !found {\n\t\treturn nil, nil, fmt.Errorf(\"wal does not contain height %d\", height)\n\t}\n\tdefer gr.Close()\n\n\t// log.Notice(\"Build a blockchain by reading from the WAL\")\n\n\tvar (\n\t\tblocks []*types.Block\n\t\tcommits []*types.Commit\n\t\tthisBlockParts *types.PartSet\n\t\tthisBlockCommit *types.Commit\n\t)\n\n\tdec := NewWALDecoder(gr)\n\tfor {\n\t\tmsg, err := dec.Decode()\n\t\tif err == io.EOF {\n\t\t\tbreak\n\t\t} else if err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\n\t\tpiece := readPieceFromWAL(msg)\n\t\tif piece == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tswitch p := piece.(type) {\n\t\tcase EndHeightMessage:\n\t\t\t// if its not the first one, we have a full block\n\t\t\tif thisBlockParts != nil {\n\t\t\t\tpbb := new(tmproto.Block)\n\t\t\t\tbz, err := io.ReadAll(thisBlockParts.GetReader())\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\t\t\t\terr = proto.Unmarshal(bz, pbb)\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\t\t\t\tblock, err := types.BlockFromProto(pbb)\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\n\t\t\t\tif block.Height != height+1 {\n\t\t\t\t\tpanic(fmt.Sprintf(\"read bad block from wal. got height %d, expected %d\", block.Height, height+1))\n\t\t\t\t}\n\t\t\t\tcommitHeight := thisBlockCommit.Height\n\t\t\t\tif commitHeight != height+1 {\n\t\t\t\t\tpanic(fmt.Sprintf(\"commit doesnt match. got height %d, expected %d\", commitHeight, height+1))\n\t\t\t\t}\n\t\t\t\tblocks = append(blocks, block)\n\t\t\t\tcommits = append(commits, thisBlockCommit)\n\t\t\t\theight++\n\t\t\t}\n\t\tcase *types.PartSetHeader:\n\t\t\tthisBlockParts = types.NewPartSetFromHeader(*p)\n\t\tcase *types.Part:\n\t\t\t_, err := thisBlockParts.AddPart(p)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\tcase *types.Vote:\n\t\t\tif p.Type == tmproto.PrecommitType {\n\t\t\t\tthisBlockCommit = types.NewCommit(p.Height, p.Round,\n\t\t\t\t\tp.BlockID, []types.CommitSig{p.CommitSig()})\n\t\t\t}\n\t\t}\n\t}\n\t// grab the last block too\n\tbz, err := io.ReadAll(thisBlockParts.GetReader())\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tpbb := new(tmproto.Block)\n\terr = proto.Unmarshal(bz, pbb)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tblock, err := types.BlockFromProto(pbb)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tif block.Height != height+1 {\n\t\tpanic(fmt.Sprintf(\"read bad block from wal. got height %d, expected %d\", block.Height, height+1))\n\t}\n\tcommitHeight := thisBlockCommit.Height\n\tif commitHeight != height+1 {\n\t\tpanic(fmt.Sprintf(\"commit doesnt match. got height %d, expected %d\", commitHeight, height+1))\n\t}\n\tblocks = append(blocks, block)\n\tcommits = append(commits, thisBlockCommit)\n\treturn blocks, commits, nil\n}", "func NewlogicalReaderReadInvocation(path string, ident1 *vault.Secret, ident2 error) *logicalReaderReadInvocation {\n\tinvocation := new(logicalReaderReadInvocation)\n\n\tinvocation.Parameters.Path = path\n\n\tinvocation.Results.Ident1 = ident1\n\tinvocation.Results.Ident2 = ident2\n\n\treturn invocation\n}", "func (sr *SequentialReader) Remain() int64 {\n\treturn sr.reader.Size() - sr.Position()\n}", "func (i *index) Read(indexOff int64) (off uint32, indexPos uint64, err error) {\n\tif i.size == 0 {\n\t\treturn 0, 0, io.EOF\n\t}\n\n\trealIndexOff := uint32(indexOff)\n\tif indexOff == -1 {\n\t\trealIndexOff = uint32((i.size / entWidth) - 1)\n\t}\n\n\tindexPos = uint64(realIndexOff) * entWidth\n\tif i.size < indexPos+entWidth {\n\t\treturn 0, 0, io.EOF\n\t}\n\n\tdataOff := enc.Uint32(i.mmap[indexPos : indexPos+offWidth])\n\tdataPos := enc.Uint64(i.mmap[indexPos+offWidth : indexPos+entWidth])\n\treturn dataOff, dataPos, nil\n}", "func (ms *MemStore) ReadFrom(r io.Reader) (int64, error) {\n\tlr := byteio.StickyLittleEndianReader{Reader: r}\n\tms.mu.Lock()\n\tfor {\n\t\tkey := lr.ReadStringX()\n\t\tif lr.Err == io.EOF {\n\t\t\tlr.Err = nil\n\t\t\tbreak\n\t\t}\n\t\tbuf := make(memio.Buffer, lr.ReadUintX())\n\t\tlr.Read(buf)\n\t\tif lr.Err != nil {\n\t\t\tif lr.Err == io.EOF {\n\t\t\t\tlr.Err = io.ErrUnexpectedEOF\n\t\t\t}\n\t\t\tbreak\n\t\t}\n\t\tms.data[key] = buf\n\t}\n\tms.mu.Unlock()\n\treturn lr.Count, lr.Err\n}", "func (master *MasterIndex) Read() error {\n\tf, err := os.OpenFile(master.Filename, os.O_RDWR, 0600)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdec := gob.NewDecoder(f)\n\terr = dec.Decode(master)\n\tf.Close()\n\treturn err\n}", "func checkReader(t *testing.T, r zbuf.Reader, checkReads bool) {\n\tfor expect := 3; expect <= 6; expect++ {\n\t\trec, err := r.Read()\n\t\trequire.NoError(t, err)\n\n\t\tv, err := rec.AccessInt(\"value\")\n\t\trequire.NoError(t, err)\n\n\t\trequire.Equal(t, int64(expect), v, \"Got expected record value\")\n\t}\n\n\trec, err := r.Read()\n\trequire.NoError(t, err)\n\trequire.Nil(t, rec, \"Reached eof after last record in time span\")\n\n\tif checkReads {\n\t\trr, ok := r.(*rangeReader)\n\t\trequire.True(t, ok, \"Can get read stats from index reader\")\n\t\trequire.LessOrEqual(t, rr.reads(), uint64(6), \"Indexed reader did not read the entire file\")\n\t}\n}", "func (s *inMemoryLogStore) Read(id string) (io.ReadCloser, error) {\n\ts.mu.RLock()\n\tdefer s.mu.RUnlock()\n\n\tl, ok := s.logs[id]\n\tif !ok {\n\t\treturn nil, ErrNotFound\n\t}\n\n\tch := make(chan []byte)\n\tl.Mu.Lock()\n\tl.Reader[ch] = struct{}{}\n\tl.Mu.Unlock()\n\treturn ioutil.NopCloser(&logSessionReader{\n\t\tLog: l,\n\t\tR: ch,\n\t}), nil\n}", "func NewReader(cfg Config, plannerCfg PlannerConfig) (*Reader, error) {\n\tplanner, err := NewPlanner(plannerCfg)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tscanner, err := storage.NewChunkScanner(cfg.StorageType, cfg.StorageConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tid := fmt.Sprintf(\"%d_%d\", plannerCfg.FirstShard, plannerCfg.LastShard)\n\n\t// Default to one worker if none is set\n\tif cfg.NumWorkers < 1 {\n\t\tcfg.NumWorkers = 1\n\t}\n\n\treturn &Reader{\n\t\tcfg: cfg,\n\t\tid: id,\n\t\tplanner: planner,\n\t\tscanner: scanner,\n\t\tscanRequestsChan: make(chan chunk.ScanRequest),\n\t\tquit: make(chan struct{}),\n\t}, nil\n}", "func (d *KrakenStorageDriver) Reader(ctx context.Context, path string, offset int64) (io.ReadCloser, error) {\n\tlog.Debugf(\"(*KrakenStorageDriver).Reader %s\", path)\n\tpathType, pathSubType, err := ParsePath(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar reader io.ReadCloser\n\tswitch pathType {\n\tcase _uploads:\n\t\treader, err = d.uploads.reader(path, pathSubType, offset)\n\tcase _blobs:\n\t\treader, err = d.blobs.reader(ctx, path, offset)\n\tdefault:\n\t\treturn nil, InvalidRequestError{path}\n\t}\n\tif err != nil {\n\t\treturn nil, toDriverError(err, path)\n\t}\n\treturn reader, nil\n}", "func (a *Archive) Open(name string) (*Reader, error) {\n\te, err := a.GetFileInfo(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsectionReader := io.NewSectionReader(a.reader, e.Offset, e.CompressedSize)\n\treturn &Reader{\n\t\treader: lz4.NewReader(sectionReader),\n\t}, nil\n}", "func (haa *HistoryArchiveAdapter) GetLedger(sequence uint32) (io.ArchiveLedgerReader, error) {\n\treturn nil, fmt.Errorf(\"not implemented yet\")\n}", "func (w *blobWriter) ReadFrom(r io.Reader) (n int64, err error) {\n\treturn w.BlobWriter.ReadFrom(r)\n}", "func (tfb *TempFileBlock) Reader() (*Reader, error) {\n switch tfb.GetAlgorithm() {\n case \"snappy\":\n\t\tf, err := os.OpenFile(tfb.TempFile, os.O_RDONLY, 0600)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"Error opening file block at \\\"\"+tfb.TempFile+\"\\\" for reading\")\n\t\t}\n\t\treturn &Reader{Reader: snappy.NewReader(bufio.NewReader(f)), File: f}, nil\n case \"gzip\":\n\t\tf, err := os.OpenFile(tfb.TempFile, os.O_RDONLY, 0600)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"Error opening file block at \\\"\"+tfb.TempFile+\"\\\" for reading\")\n\t\t}\n\t\tgr, err := gzip.NewReader(bufio.NewReader(f))\n\t\tif gr != nil {\n\t\t\treturn nil, errors.Wrap(err, \"Error creating gzip reader for temp file block.\")\n\t\t}\n\t\treturn &Reader{ReadCloser: gr, File: f}, nil\n case \"none\":\n\t\tf, err := os.OpenFile(tfb.TempFile, os.O_RDONLY, 0600)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"Error opening file block at \\\"\"+tfb.TempFile+\"\\\" for reading\")\n\t\t}\n return &Reader{Reader: bufio.NewReader(f), File: f}, nil\n }\n return nil, errors.New(\"Unknown compression algorithm\")\n}", "func (this *reader) rollback(n int, err error) ([]byte, []byte, int64, error) {\n\t_, e := this.Seek(int64(-n), 1)\n\tif e != nil {\n\t\terr = fmt.Errorf(\"Reading error(%q); Rollback error(%q)\", err, e)\n\t}\n\treturn nil, nil, 0, err\n}", "func (d *driver) Reader(ctx context.Context, path string, offset int64) (io.ReadCloser, error) {\n\tbaseUrl := d.getBaseUrl(path)\n\n\tinfo, err := d.Bucket.Stat(ctx, path)\n\tif err != nil {\n\t\treturn nil, err;\n\t}\n\n\tif offset > info.Fsize {\n\t\treturn ioutil.NopCloser(bytes.NewReader(nil)), nil\n\t}\n\n\thttpClient := &http.Client{}\n\treq, err := http.NewRequest(\"GET\", baseUrl, nil)\n\treq.Header.Add(\"Range\", \"bytes=\"+strconv.FormatInt(offset, 10)+\"-\")\n\tresp, err := httpClient.Do(req)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tc,_ := ioutil.ReadAll(resp.Body)\n\n\tfmt.Print(\"content\"+string(c)+\"\\n\")\n\n\treturn resp.Body,err\n}", "func FindReaderSize(r io.Reader, needle []byte, bufSize int) (int, error) {\n\tif bufSize < len(needle) {\n\t\tbufSize = 2 * len(needle)\n\t}\n\tneedleLen := len(needle)\n\tbuf := make([]byte, bufSize)\n\tvar off, start int\n\tfor {\n\t\tn, err := io.ReadAtLeast(r, buf[start:], needleLen)\n\t\tif errors.Is(err, io.ErrUnexpectedEOF) {\n\t\t\terr = io.EOF\n\t\t}\n\t\tif n == 0 && errors.Is(err, io.EOF) {\n\t\t\treturn -1, nil\n\t\t}\n\t\t//fmt.Println(off, start, n)\n\t\tif i := bytes.Index(buf[:start+n], needle); i >= 0 {\n\t\t\t//fmt.Printf(\"buf=%q\\n\", buf[i:i+len(needle)])\n\t\t\treturn off + i, nil\n\t\t}\n\t\tif err != nil {\n\t\t\treturn -1, err\n\t\t}\n\t\t// copy the end to the start\n\t\tcopy(buf[0:], buf[start+n-needleLen+1:start+n])\n\t\tif off == 0 {\n\t\t\toff = n - needleLen + 1\n\t\t\tstart = needleLen - 1\n\t\t} else {\n\t\t\toff += n\n\t\t}\n\t\t//fmt.Printf(\"buf=%q n=%d needle=%d off=%d\\n\", buf[:start], n, len(needle), off)\n\t}\n}", "func (tr testReader) ReadAt(b []byte, off int64) (int, error) {\n\tn, err := tr.Reader.ReadAt(b, off)\n\tif err == nil && (int64(n)+off == tr.Size()) {\n\t\treturn n, io.EOF\n\t}\n\treturn n, err\n}", "func (rbl *RawBytesLog) NewReader() (*Reader, error) {\n\tlogFile, err := os.OpenFile(rbl.logFile.Name(), os.O_RDONLY, 0)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treader := &Reader{\n\t\tlogFile: logFile,\n\t}\n\treturn reader, nil\n}", "func (l *Log) Read(off uint64) (*api.Record, error) {\n\tl.mu.Lock()\n\tdefer l.mu.Unlock()\n\tvar s *segment\n\tfor _, segment := range l.segments {\n\t\tif segment.baseOffset <= segment.nextOffset {\n\t\t\ts = segment\n\t\t\tbreak\n\t\t}\n\t}\n\tif s == nil || s.nextOffset <= off {\n\t\treturn nil, api.ErrOffsetOutOfRange{Offset: off}\n\t}\n\treturn s.Read(off)\n}", "func (r *Redis) NewReader(item sdk.CDNItem, format sdk.CDNReaderFormat, from int64, size uint, sort int64) io.ReadCloser {\n\treturn &redis.Reader{\n\t\tStore: r.store,\n\t\tItemID: item.ID,\n\t\tApiRefHash: item.APIRefHash,\n\t\tPrefixKey: redisLruItemCacheKey,\n\t\tSize: size,\n\t\tFrom: from,\n\t\tFormat: format,\n\t\tSort: sort,\n\t}\n}", "func readUntilCRLF(buf []byte, r io.Reader, idx int) ([]byte, error) {\n\t// Read until we find the cRLF or we hit our max possible header length\n\tfor idx < 107 {\n\t\tc, err := r.Read(buf[idx : idx+1])\n\t\tif c != 1 {\n\t\t\treturn nil, errors.New(\"expected to read more bytes, but got none\")\n\t\t}\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif bytes.Equal(buf[idx-1:idx+1], []byte(cRLF)) {\n\t\t\treturn buf[0 : idx-1], nil\n\t\t}\n\t\tidx++\n\t}\n\treturn nil, errors.New(\"gave up after 107 bytes\")\n}", "func (s *Store) ReadFrom(r io.Reader) (int64, error) {\n\t// Read current file version.\n\tuint32Bytes := make([]byte, 4)\n\tn, err := io.ReadFull(r, uint32Bytes)\n\tn64 := int64(n)\n\tif err != nil {\n\t\treturn n64, err\n\t}\n\tvers := binary.LittleEndian.Uint32(uint32Bytes)\n\n\t// Reading files with versions before versCombined is unsupported.\n\tif vers < versCombined {\n\t\treturn n64, ErrUnsupportedVersion\n\t}\n\n\t// Reset store.\n\ts.txs = make(map[blockTx]*btcutil.Tx)\n\ts.sorted = list.New()\n\ts.signed = make(map[blockTx]*signedTx)\n\ts.recv = make(map[blockOutPoint]*recvTxOut)\n\ts.unspent = make(map[btcwire.OutPoint]*recvTxOut)\n\n\t// Read backing transactions and records.\n\tfor {\n\t\t// Read byte header. If this errors with io.EOF, we're done.\n\t\theader := make([]byte, 1)\n\t\tn, err = io.ReadFull(r, header)\n\t\tn64 += int64(n)\n\t\tif err == io.EOF {\n\t\t\treturn n64, nil\n\t\t}\n\n\t\tswitch header[0] {\n\t\tcase backingTxHeader:\n\t\t\t// Read block height.\n\t\t\tn, err = io.ReadFull(r, uint32Bytes)\n\t\t\tn64 += int64(n)\n\t\t\tif err == io.EOF {\n\t\t\t\terr = io.ErrUnexpectedEOF\n\t\t\t}\n\t\t\tif err != nil {\n\t\t\t\treturn n64, err\n\t\t\t}\n\t\t\theight := int32(binary.LittleEndian.Uint32(uint32Bytes))\n\n\t\t\t// Read serialized transaction.\n\t\t\ttx := new(msgTx)\n\t\t\ttxN, err := tx.readFrom(r)\n\t\t\tn64 += txN\n\t\t\tif err == io.EOF {\n\t\t\t\terr = io.ErrUnexpectedEOF\n\t\t\t}\n\t\t\tif err != nil {\n\t\t\t\treturn n64, err\n\t\t\t}\n\n\t\t\t// Add backing tx to store.\n\t\t\tutx := btcutil.NewTx((*btcwire.MsgTx)(tx))\n\t\t\ts.txs[blockTx{*utx.Sha(), height}] = utx\n\n\t\tcase recvTxOutHeader:\n\t\t\t// Read received transaction output record.\n\t\t\trtx := new(recvTxOut)\n\t\t\ttxN, err := rtx.readFrom(r)\n\t\t\tn64 += txN\n\t\t\tif err == io.EOF {\n\t\t\t\terr = io.ErrUnexpectedEOF\n\t\t\t}\n\t\t\tif err != nil {\n\t\t\t\treturn n64, err\n\t\t\t}\n\n\t\t\t// It is an error for the backing transaction to have\n\t\t\t// not already been read.\n\t\t\tif _, ok := s.txs[rtx.blockTx()]; !ok {\n\t\t\t\treturn n64, ErrInconsistantStore\n\t\t\t}\n\n\t\t\t// Add entries to store.\n\t\t\ts.sorted.PushBack(rtx)\n\t\t\tk := blockOutPoint{rtx.outpoint, rtx.Height()}\n\t\t\ts.recv[k] = rtx\n\t\t\tif !rtx.Spent() {\n\t\t\t\ts.unspent[rtx.outpoint] = rtx\n\t\t\t}\n\n\t\tcase signedTxHeader:\n\t\t\t// Read signed (sent) transaction record.\n\t\t\tstx := new(signedTx)\n\t\t\ttxN, err := stx.readFrom(r)\n\t\t\tn64 += txN\n\t\t\tif err == io.EOF {\n\t\t\t\terr = io.ErrUnexpectedEOF\n\t\t\t}\n\t\t\tif err != nil {\n\t\t\t\treturn n64, err\n\t\t\t}\n\n\t\t\t// It is an error for the backing transaction to have\n\t\t\t// not already been read.\n\t\t\tif _, ok := s.txs[stx.blockTx()]; !ok {\n\t\t\t\treturn n64, ErrInconsistantStore\n\t\t\t}\n\n\t\t\t// Add entries to store.\n\t\t\ts.sorted.PushBack(stx)\n\t\t\ts.signed[stx.blockTx()] = stx\n\n\t\tdefault:\n\t\t\treturn n64, errors.New(\"bad magic byte\")\n\t\t}\n\t}\n\n\treturn n64, nil\n}", "func NewReader(r io.Reader) *Reader {\n\treturn &Reader{\n\t\tMaxObjectsLimit: DefaultMaxObjectsLimit,\n\n\t\tr: &trackingReader{r: r},\n\t\toffsets: make(map[int64]core.Hash, 0),\n\t}\n}", "func (cfp *FsPool) GetReader(fileIndex int64) (io.Reader, error) {\n\trs, err := cfp.GetReadSeeker(fileIndex)\n\tif err != nil {\n\t\treturn nil, errors.WithStack(err)\n\t}\n\n\t_, err = rs.Seek(0, io.SeekStart)\n\tif err != nil {\n\t\treturn nil, errors.WithStack(err)\n\t}\n\n\treturn rs, nil\n}", "func (rstra *ReadSeekerToReaderAt) ReadAt(p []byte, offset int64) (n int, err error) {\n\tdefer func() {\n\t\tif state := recover(); state != nil {\n\t\t\terr = log.Wrap(state.(error))\n\t\t}\n\t}()\n\n\toriginalOffset, err := rstra.rs.Seek(0, io.SeekCurrent)\n\tlog.PanicIf(err)\n\n\tdefer func() {\n\t\t_, err := rstra.rs.Seek(originalOffset, io.SeekStart)\n\t\tlog.PanicIf(err)\n\t}()\n\n\t_, err = rstra.rs.Seek(offset, io.SeekStart)\n\tlog.PanicIf(err)\n\n\t// Note that all errors will be wrapped, here. The usage of this method is\n\t// such that typically no specific errors would be expected as part of\n\t// normal operation (in which case we'd check for those first and return\n\t// them directly).\n\tn, err = io.ReadFull(rstra.rs, p)\n\tlog.PanicIf(err)\n\n\treturn n, nil\n}", "func (p *parallelReader) Read() ([][]byte, error) {\n\ttype errIdx struct {\n\t\tidx int\n\t\tbuf []byte\n\t\terr error\n\t}\n\n\terrCh := make(chan errIdx)\n\tcurrReaderIndex := 0\n\tnewBuf := make([][]byte, len(p.readers))\n\n\tif p.offset+p.shardSize > p.shardFileSize {\n\t\tp.shardSize = p.shardFileSize - p.offset\n\t}\n\n\tread := func(currReaderIndex int) {\n\t\tb, err := p.readers[currReaderIndex].ReadChunk(p.offset, p.shardSize)\n\t\terrCh <- errIdx{currReaderIndex, b, err}\n\t}\n\n\treaderCount := 0\n\tfor _, r := range p.readers {\n\t\tif r != nil {\n\t\t\treaderCount++\n\t\t}\n\t}\n\tif readerCount < p.dataBlocks {\n\t\treturn nil, errXLReadQuorum\n\t}\n\n\treaderCount = 0\n\tfor i, r := range p.readers {\n\t\tif r == nil {\n\t\t\tcontinue\n\t\t}\n\t\tgo read(i)\n\t\treaderCount++\n\t\tif readerCount == p.dataBlocks {\n\t\t\tcurrReaderIndex = i + 1\n\t\t\tbreak\n\t\t}\n\t}\n\n\tfor errVal := range errCh {\n\t\tif errVal.err == nil {\n\t\t\tnewBuf[errVal.idx] = errVal.buf\n\t\t\tif p.canDecode(newBuf) {\n\t\t\t\tp.offset += int64(p.shardSize)\n\t\t\t\treturn newBuf, nil\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tp.readers[errVal.idx] = nil\n\t\tfor currReaderIndex < len(p.readers) {\n\t\t\tif p.readers[currReaderIndex] != nil {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tcurrReaderIndex++\n\t\t}\n\n\t\tif currReaderIndex == len(p.readers) {\n\t\t\tbreak\n\t\t}\n\t\tgo read(currReaderIndex)\n\t\tcurrReaderIndex++\n\t}\n\n\treturn nil, errXLReadQuorum\n}", "func NewReader(r io.Reader, rd int) (*Reader, error) {\n\tbg, err := bgzf.NewReader(r, rd)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\th, _ := sam.NewHeader(nil, nil)\n\tbr := &Reader{\n\t\tr: bg,\n\t\th: h,\n\n\t\treferences: int32(len(h.Refs())),\n\t}\n\terr = br.h.DecodeBinary(br.r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbr.lastChunk.End = br.r.LastChunk().End\n\treturn br, nil\n}", "func ReadFromReader(reader io.ReaderAt, offset uint32, blockSize int) ([]byte, uint32, error) {\n\tb, err := ReadFromReader64(reader, uint64(offset*PAD), blockSize)\n\tif err != nil {\n\t\treturn nil, 0, err\n\t}\n\tnextOffset := (offset + ((uint32(16+len(b)))+PAD-1)/PAD)\n\treturn b, uint32(nextOffset), nil\n}", "func ReadFromReader(reader io.ReaderAt, offset uint32, blockSize int) ([]byte, uint32, error) {\n\tb, err := ReadFromReader64(reader, uint64(offset*PAD), blockSize)\n\tif err != nil {\n\t\treturn nil, 0, err\n\t}\n\tnextOffset := (offset + ((uint32(16+len(b)))+PAD-1)/PAD)\n\treturn b, uint32(nextOffset), nil\n}", "func (reader *Reader) Next() (common.Entry, error) {\n\tif reader.index >= len(reader.zipReader.File) {\n\t\treturn nil, io.EOF\n\t}\n\n\tf := reader.zipReader.File[reader.index]\n\n\treader.index++\n\treturn &zipFormatEntry{f}, nil\n}", "func (wpr *Wrapper) ReadAt(p []byte, off int64) (n int, err error) {\n\twrap := NewWrapReader(wpr.rwa, wpr.O+off, wpr.N)\n\tr := io.LimitReader(wrap, wpr.L-off)\n\treturn r.Read(p)\n}" ]
[ "0.6434191", "0.5840932", "0.58075285", "0.567036", "0.55909795", "0.5506269", "0.5485145", "0.5431219", "0.5323112", "0.5318784", "0.52565265", "0.5216578", "0.5212504", "0.5120026", "0.51129574", "0.5033053", "0.5015555", "0.4984612", "0.4940707", "0.49372166", "0.48791194", "0.4876886", "0.48541844", "0.48383093", "0.47947794", "0.47933993", "0.47823817", "0.4764167", "0.4752632", "0.4752632", "0.47421703", "0.47180355", "0.4714134", "0.46576437", "0.46556655", "0.465546", "0.46532312", "0.4651351", "0.4641173", "0.46384114", "0.4634638", "0.46055382", "0.45967302", "0.45948374", "0.4568875", "0.45440307", "0.4539113", "0.45214534", "0.45143178", "0.45000952", "0.44982758", "0.44898564", "0.44837105", "0.44755974", "0.44698718", "0.44666976", "0.44660878", "0.4442243", "0.44308832", "0.44164717", "0.44149733", "0.44108132", "0.44100183", "0.44024602", "0.43884894", "0.4388107", "0.43838218", "0.43836597", "0.43786237", "0.437382", "0.4367027", "0.43659574", "0.4354706", "0.43469247", "0.43453854", "0.4341443", "0.43349457", "0.43320712", "0.43302494", "0.43297872", "0.43276864", "0.4326815", "0.43237177", "0.43199152", "0.43157876", "0.4314704", "0.4310866", "0.43083656", "0.43077463", "0.43032125", "0.43005478", "0.4297051", "0.4296766", "0.4294328", "0.4285045", "0.42842275", "0.42759478", "0.42759478", "0.42758065", "0.42659962" ]
0.5193216
13
TODO This is a testing only function. Can we test a different way?
func (t *WALTopic) addReaderAlert(a *topicReaderAlert) { t.readerAlerts = append(t.readerAlerts, a) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func TestGetNone4A(t *testing.T) {\n}", "func TestGetEmpty4A(t *testing.T) {\n}", "func TestSinglePrewrite4A(t *testing.T) {\n}", "func TestGetValue4A(t *testing.T) {\n}", "func TestGetVersions4A(t *testing.T) {\n}", "func TestAckInstalledApplicationListDuplicateRegression(t *testing.T) {\n\n}", "func TestGetValueTs4A(t *testing.T) {\n}", "func TestPrewriteMultiple4A(t *testing.T) {\n}", "func TestPrewriteWritten4A(t *testing.T) {\n}", "func TestGetDeleted4A(t *testing.T) {\n}", "func TestManifestAPI_Get_UnknownSchema(t *testing.T) {}", "func TestEmptyPrewrite4A(t *testing.T) {\n}", "func TestRestOfInternalCode(t *testing.T) {\n\n\t// In this case unit testing will not help as we need to actually corever\n\t// this package with test. Because real functions hide under internal structures\n\t// which we do not expose, so our previous approach will no longer works.\n\t// Well it works but coverage does not detect that we are testing actual\n\t// implementation\n\n\t// In order to cover this part we will need to either pretend that we are\n\t// testing something or create real integration tests and ensure that mongod\n\t// process is running. In my case I will just fake my testing and do not use\n\t// assert. This way my test will pass either way\n\n\t// Create database context. I use real database, but it is possible to mock\n\t// database and configuration through interfaces.\n\tconf := config.GetConfig()\n\tclient, _ := databases.NewClient(conf)\n\tclient.StartSession()\n\n\tdb := databases.NewDatabase(conf, client)\n\tclient.Connect()\n\tdb.Client()\n\tvar result interface{}\n\t// because we do not care for actual results, we just quickly timeout the\n\t// call and we use incorrect call method\n\ttimeoutCtx, _ := context.WithTimeout(context.Background(), 1*time.Microsecond)\n\tdb.Collection(\"non-fake-existing-collection\").FindOne(timeoutCtx, \"incorrect-value\").Decode(&result)\n\n\t// insert and delete functions seems to panic instead of returning and error.\n\t// I did not investigate anything in this case as this is not our main goal.\n\t// Just define assert panic function and use this panicing function in it.\n\tvar mongoPanics assert.PanicTestFunc\n\n\tmongoPanics = func() {\n\t\tdb.Collection(\"non-fake-existing-collection\").InsertOne(timeoutCtx, result)\n\t}\n\tassert.Panics(t, mongoPanics)\n\n\tmongoPanics = func() {\n\t\tdb.Collection(\"non-fake-existing-collection\").DeleteOne(timeoutCtx, result)\n\t}\n\tassert.Panics(t, mongoPanics)\n\n\t// And it is done. We do not need to have mongo running and our code is\n\t// covered 100%. Well the actual implementation is faked, but it should be\n\t// tested via integration tests, not unit tests.\n\n}", "func TestGetUrl(t *testing.T) {\n\n shards := map[string]string{\n \"us\": \"https://api.us.onelogin.com/\",\n \"eu\": \"https://api.eu.onelogin.com/\",\n }\n\n for shard,url := range shards {\n o := OneLogin{Shard: shard}\n result := o.GetUrl(\"\")\n if result != url {\n t.Errorf(\"GetUrl() for shard %s != %s (actual result was %s)\", shard, url, result)\n }\n }\n}", "func TestCallFunc_arguments(t *testing.T) {\n\n}", "func TestUnmarshalPathologies(t *testing.T) {\n\n}", "func TestPrewriteOverwrite4A(t *testing.T) {\n}", "func TestToOne(t *testing.T) {}", "func TestToOne(t *testing.T) {}", "func TestToOne(t *testing.T) {}", "func TestToOne(t *testing.T) {}", "func TestToOneSet(t *testing.T) {}", "func TestToOneSet(t *testing.T) {}", "func TestToOneSet(t *testing.T) {}", "func TestToOneSet(t *testing.T) {}", "func testFullCycleFilesystem(t *testing.T) {\n\tt.Log(\"TODO\")\n}", "func TestSingleCommit4A(t *testing.T) {\n}", "func TestSqlSMSStorage_GetSMSs(t *testing.T) {\n\n}", "func TestCommitMissingPrewrite4a(t *testing.T) {\n}", "func TestPrewriteWrittenNoConflict4A(t *testing.T) {\n}", "func helper(t *testing.T, expectedErr interface{}, code string) {\n\t_, err := getExportedFunctions(\"\", code)\n\tif err == nil {\n\t\tt.Error(\"Expected error, got nil!\")\n\t}\n\n\texpectedErrType := reflect.TypeOf(expectedErr)\n\tactualErrType := reflect.TypeOf(err)\n\n\tif actualErrType != expectedErrType {\n\t\tt.Error(\"Expected error\", expectedErrType.String(), \"but got\",\n\t\t\tactualErrType.String(), \"with message\", err)\n\t}\n}", "func TestGet_Token(t *testing.T) {\n t.Errorf(\"No tests written yet for Get_Token()\")\n}", "func TestEmptyCommit4A(t *testing.T) {\n}", "func TestGetSlidesDocumentWithFormatInvalidoutPath(t *testing.T) {\n request := createGetSlidesDocumentWithFormatRequest()\n request.outPath = invalidizeTestParamValue(request.outPath, \"outPath\", \"string\").(string)\n e := initializeTest(\"GetSlidesDocumentWithFormat\", \"outPath\", request.outPath)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n _, r, e := getTestApiClient().DocumentApi.GetSlidesDocumentWithFormat(request)\n assertError(t, \"GetSlidesDocumentWithFormat\", \"outPath\", int32(r.StatusCode), e)\n}", "func TestGetSlidesDocumentWithFormatInvalidstorage(t *testing.T) {\n request := createGetSlidesDocumentWithFormatRequest()\n request.storage = invalidizeTestParamValue(request.storage, \"storage\", \"string\").(string)\n e := initializeTest(\"GetSlidesDocumentWithFormat\", \"storage\", request.storage)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n _, r, e := getTestApiClient().DocumentApi.GetSlidesDocumentWithFormat(request)\n assertError(t, \"GetSlidesDocumentWithFormat\", \"storage\", int32(r.StatusCode), e)\n}", "func TestUserstorage(t *testing.T) {\n t.Log(\"*** User data storage and retrieval test ***\")\n\n // initialize user\n u, err := InitUser(\"alice\",\"fubar\")\n if err != nil {\n t.Error(\"Failed to initialize user (\", err, \")\")\n } else {\n t.Log(\"Successfully stored user\", u)\n }\n\n // retrieve user \n v, err := GetUser(\"alice\", \"fubar\")\n if err != nil {\n t.Error(\"Failed to reload user\", err)\n } else {\n t.Log(\"Correctly retrieved user\", v)\n }\n}", "func Test_sampe002(t *testing.T) {\n\n}", "func TestGetSlidesDocumentInvalidstorage(t *testing.T) {\n request := createGetSlidesDocumentRequest()\n request.storage = invalidizeTestParamValue(request.storage, \"storage\", \"string\").(string)\n e := initializeTest(\"GetSlidesDocument\", \"storage\", request.storage)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n r, _, e := getTestApiClient().DocumentApi.GetSlidesDocument(request)\n assertError(t, \"GetSlidesDocument\", \"storage\", r.Code, e)\n}", "func TestGetSlidesDocumentWithFormatInvalidname(t *testing.T) {\n request := createGetSlidesDocumentWithFormatRequest()\n request.name = invalidizeTestParamValue(request.name, \"name\", \"string\").(string)\n e := initializeTest(\"GetSlidesDocumentWithFormat\", \"name\", request.name)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n _, r, e := getTestApiClient().DocumentApi.GetSlidesDocumentWithFormat(request)\n assertError(t, \"GetSlidesDocumentWithFormat\", \"name\", int32(r.StatusCode), e)\n}", "func TestPrewriteLocked4A(t *testing.T) {\n}", "func TestCommitConflictRepeat4A(t *testing.T) {\n}", "func testResource(t *testing.T, s *Service) {\n\tp := &model.ArgRes{\n\t\tResID: 1233,\n\t}\n\tres, err := s.Resource(context.TODO(), p)\n\tif err != nil {\n\t\tt.Logf(\"testResource error(%v) \\n\", err)\n\t\treturn\n\t}\n\tt.Logf(\"testResource res: %+v \\n\", res)\n}", "func TestGetSlidesDocumentInvalidname(t *testing.T) {\n request := createGetSlidesDocumentRequest()\n request.name = invalidizeTestParamValue(request.name, \"name\", \"string\").(string)\n e := initializeTest(\"GetSlidesDocument\", \"name\", request.name)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n r, _, e := getTestApiClient().DocumentApi.GetSlidesDocument(request)\n assertError(t, \"GetSlidesDocument\", \"name\", r.Code, e)\n}", "func TestCallFunc_function(t *testing.T) {\n\n}", "func TestPostSlidesSaveAsInvalidstorage(t *testing.T) {\n request := createPostSlidesSaveAsRequest()\n request.storage = invalidizeTestParamValue(request.storage, \"storage\", \"string\").(string)\n e := initializeTest(\"PostSlidesSaveAs\", \"storage\", request.storage)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n _, r, e := getTestApiClient().DocumentApi.PostSlidesSaveAs(request)\n assertError(t, \"PostSlidesSaveAs\", \"storage\", int32(r.StatusCode), e)\n}", "func TestDelegatorProxyValidatorShares7Steps(t *testing.T) {\n\n}", "func TestPostSlidesSaveAsInvalidname(t *testing.T) {\n request := createPostSlidesSaveAsRequest()\n request.name = invalidizeTestParamValue(request.name, \"name\", \"string\").(string)\n e := initializeTest(\"PostSlidesSaveAs\", \"name\", request.name)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n _, r, e := getTestApiClient().DocumentApi.PostSlidesSaveAs(request)\n assertError(t, \"PostSlidesSaveAs\", \"name\", int32(r.StatusCode), e)\n}", "func TestPostSlidesDocumentInvalidstorage(t *testing.T) {\n request := createPostSlidesDocumentRequest()\n request.storage = invalidizeTestParamValue(request.storage, \"storage\", \"string\").(string)\n e := initializeTest(\"PostSlidesDocument\", \"storage\", request.storage)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n r, _, e := getTestApiClient().DocumentApi.PostSlidesDocument(request)\n assertError(t, \"PostSlidesDocument\", \"storage\", r.Code, e)\n}", "func Test_Client_MapByCallingCode(t *testing.T) {\n\tret := mockClient.MapByCallingCode(\"65\")\n\tassert.Equal(t, ret[0].Name, \"Singapore\")\n}", "func TestPostSlidesSaveAsInvalidoutPath(t *testing.T) {\n request := createPostSlidesSaveAsRequest()\n request.outPath = invalidizeTestParamValue(request.outPath, \"outPath\", \"string\").(string)\n e := initializeTest(\"PostSlidesSaveAs\", \"outPath\", request.outPath)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n _, r, e := getTestApiClient().DocumentApi.PostSlidesSaveAs(request)\n assertError(t, \"PostSlidesSaveAs\", \"outPath\", int32(r.StatusCode), e)\n}", "func Test_Client_MapByAlpha3(t *testing.T) {\n\tret := mockClient.MapByAlpha3(\"SGP\")\n\tassert.Equal(t, ret.Name, \"Singapore\")\n}", "func TestComplexifyRandom(t *testing.T) {\n\n}", "func TestServiceUpdateToJSON_TwoConfig_UpdateActions(t *testing.T) {\n}", "func TestMultiplePrewrites4A(t *testing.T) {\n}", "func sampleFunc() {\n\t// TODO[TK]\n\t// todo\n\t// FIXME\n}", "func TestCommitOverwrite4A(t *testing.T) {\n}", "func TestOneToOneSet(t *testing.T) {}", "func TestOneToOneSet(t *testing.T) {}", "func TestOneToOneSet(t *testing.T) {}", "func TestOneToOneSet(t *testing.T) {}", "func TestOneToOneSet(t *testing.T) {}", "func TestOneToOneSet(t *testing.T) {}", "func TestOneToOneSet(t *testing.T) {}", "func TestOneToOneSet(t *testing.T) {}", "func TestOneToOneSet(t *testing.T) {}", "func TestGetOrder(t *testing.T) {\n\n // ...\n\n}", "func almostOkayFunction() {\n\treturn nil\n}", "func testSingleFileGet(t *testing.T, tg *siatest.TestGroup) {\n\t// Grab the first of the group's renters\n\trenter := tg.Renters()[0]\n\t// Upload file, creating a piece for each host in the group\n\tdataPieces := uint64(1)\n\tparityPieces := uint64(len(tg.Hosts())) - dataPieces\n\tfileSize := 100 + siatest.Fuzz()\n\t_, _, err := renter.UploadNewFileBlocking(fileSize, dataPieces, parityPieces)\n\tif err != nil {\n\t\tt.Fatal(\"Failed to upload a file for testing: \", err)\n\t}\n\n\tfiles, err := renter.Files()\n\tif err != nil {\n\t\tt.Fatal(\"Failed to get renter files: \", err)\n\t}\n\n\tvar file modules.FileInfo\n\tfor _, f := range files {\n\t\tfile, err = renter.File(f.SiaPath)\n\t\tif err != nil {\n\t\t\tt.Fatal(\"Failed to request single file\", err)\n\t\t}\n\t\tif file != f {\n\t\t\tt.Fatal(\"Single file queries does not match file previously requested.\")\n\t\t}\n\t}\n}", "func Test_Client_MapByAlpha2(t *testing.T) {\n\tret := mockClient.MapByAlpha2(\"SG\")\n\tassert.Equal(t, ret.Name, \"Singapore\")\n}", "func TestPostSlidesDocumentInvalidname(t *testing.T) {\n request := createPostSlidesDocumentRequest()\n request.name = invalidizeTestParamValue(request.name, \"name\", \"string\").(string)\n e := initializeTest(\"PostSlidesDocument\", \"name\", request.name)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n r, _, e := getTestApiClient().DocumentApi.PostSlidesDocument(request)\n assertError(t, \"PostSlidesDocument\", \"name\", r.Code, e)\n}", "func TestReadUser(t *testing.T) {\r\n/////////////////////////////////// MOCKING ////////////////////////////////////////////\r\n\tvar batches = []string{\r\n\t\t`CREATE TABLE Users (Id INTEGER PRIMARY KEY AUTOINCREMENT NOT NULL, Name TEXT NOT NULL UNIQUE);`,\r\n\t\t`INSERT INTO Users (Id,Name) VALUES (1,'anonymous');`,\r\n\t}\r\n\t//open pseudo database for function\r\n\tdb, err := sql.Open(\"ramsql\", \"TestReadUser\")\r\n\tif err != nil {\r\n\t\tt.Fatalf(\"Error creating mock sql : %s\\n\", err)\r\n\t}\r\n\tdefer db.Close()\r\n\r\n\t// Exec every line of batch and create database\r\n\tfor _, b := range batches {\r\n\t\t_, err = db.Exec(b)\r\n\t\tif err != nil {\r\n\t\t\tt.Fatalf(\"Error exec query in query: %s\\n Error:%s\", b, err)\r\n\t\t}\r\n\t}\r\n/////////////////////////////////// MOCKING ////////////////////////////////////////////\r\n\r\n\t// Specify test variables and expected results.\r\n\ttests := []struct {\r\n\t\tid int\r\n\t\t// we need to use models.User for passing to object.This is different with \"database.User\".\r\n\t\tresult models.User\r\n\t\terr error\r\n\t}{\r\n\t\t// When give to first parameter(id) 1 , We expect result :1 error nil\r\n\t\t{id: 1, result: models.User{Id: 1, Name: \"anonymous\"}, err: nil},\r\n\t\t// When give to first parameter(id) 1 , We expect result :1 error nil\r\n\t\t//{id: 2, result: models.User{Id: 2, Name: \"test\"}, err: nil},\r\n\t}\r\n\r\n\t// test all of the variables.\r\n\tfor _, test := range tests {\r\n\t\t//get result after test.\r\n\t\ts, err := u.ReadUser(db, test.id)\r\n\t\t// if expected error type nil we need to compare with actual error different way.\r\n\t\tif test.err == nil {\r\n\t\t\t// If test fails give error.It checks expected result and expected error\r\n\t\t\tif err != test.err || s != test.result {\r\n\t\t\t\t// Compare expected error and actual error\r\n\t\t\t\tt.Errorf(\"Error is: %v . Expected: %v\", err, test.err)\r\n\t\t\t\t// Compare expected result and actual result\r\n\t\t\t\tt.Errorf(\"Result is: %v . Expected: %v\", s, test.result)\r\n\t\t\t}\r\n\t\t\t// if expected error type is not nil we need to compare with actual error different way.\r\n\t\t} else {\r\n\t\t\tif err.Error() != test.err.Error() || s != test.result {\r\n\t\t\t\t// Compare expected error and actual error\r\n\t\t\t\tt.Errorf(\"Error is: %v . Expected: %v\", err, test.err)\r\n\t\t\t\t// Compare expected result and actual result\r\n\t\t\t\tt.Errorf(\"Result is: %v . Expected: %v\", s, test.result)\r\n\t\t\t}\r\n\t\t}\r\n\t}\r\n}", "func TestMyInfoProtocolFunctions(t *testing.T) {\n\toptions := &Options{\n\t\tHasHeader: true,\n\t\tRecordDelimiter: \"\\n\",\n\t\tFieldDelimiter: \",\",\n\t\tComments: \"\",\n\t\tName: \"S3Object\", // Default table name for all objects\n\t\tReadFrom: bytes.NewReader([]byte(\"name1,name2,name3,name4\" + \"\\n\" + \"5,is,a,string\" + \"\\n\" + \"random,random,stuff,stuff\")),\n\t\tCompressed: \"\",\n\t\tExpression: \"\",\n\t\tOutputFieldDelimiter: \",\",\n\t\tStreamSize: 20,\n\t}\n\ts3s, err := NewInput(options)\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\tmyVal, _ := s3s.createStatXML()\n\tmyOtherVal, _ := s3s.createProgressXML()\n\n\ttables := []struct {\n\t\tpayloadStatMsg string\n\t\tpayloadProgressMsg string\n\t\texpectedStat int\n\t\texpectedProgress int\n\t}{\n\t\t{myVal, myOtherVal, 233, 243},\n\t}\n\tfor _, table := range tables {\n\t\tvar currBuf = &bytes.Buffer{}\n\t\tif len(s3s.writeStatMessage(table.payloadStatMsg, currBuf).Bytes()) != table.expectedStat {\n\t\t\tt.Error()\n\t\t}\n\t\tcurrBuf.Reset()\n\t\tif len(s3s.writeProgressMessage(table.payloadProgressMsg, currBuf).Bytes()) != table.expectedProgress {\n\t\t\tt.Error()\n\t\t}\n\t}\n}", "func TestParseNode(t *testing.T) {\n}", "func TestCommitMultipleKeys4A(t *testing.T) {\n}", "func TestPostSlidesDocumentInvalidtemplateStorage(t *testing.T) {\n request := createPostSlidesDocumentRequest()\n request.templateStorage = invalidizeTestParamValue(request.templateStorage, \"templateStorage\", \"string\").(string)\n e := initializeTest(\"PostSlidesDocument\", \"templateStorage\", request.templateStorage)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n r, _, e := getTestApiClient().DocumentApi.PostSlidesDocument(request)\n assertError(t, \"PostSlidesDocument\", \"templateStorage\", r.Code, e)\n}", "func TestCommitConflictRace4A(t *testing.T) {\n}", "func TestRecommitKey4A(t *testing.T) {\n}", "func TestData(t *testing.T) { TestingT(t) }", "func TestEmailAlreadyInUse(t *testing.T) {\n handle1 := \"test_handleinuse1\"\n handle2 := \"test_handleinuse2\"\n email := \"[email protected]\"\n\n var account UserAccount\n if err := Create(handle1, email, \"timisadork\", &account); err != nil {\n t.Error(err)\n }\n\n if err := Create(handle2, email, \"timisadork\", &account); err == nil {\n t.Errorf(\"Used the same email twice\")\n }\n\n if err := Delete(account.Key); err != nil {\n t.Error(err)\n }\n}", "func TestPointerTypes(t *testing.T) {\n\n}", "func checkRead(t *testing.T, testname string, b *Builder, s string) {\n\tbytes := b.Bytes()\n\tstr := b.String()\n\tif b.Len() != len(str) {\n\t\tt.Errorf(\"%s: b.Len() == %d, len(b.String()) == %d\", testname, b.Len(), len(str))\n\t}\n\tif string(bytes) != s {\n\t\tt.Errorf(\"%s: string(b.Bytes()) == %q, s == %q\", testname, string(bytes), s)\n\t}\n}", "func TestEmailExists(t *testing.T) {\n handle := \"test_handleexists\"\n email := \"[email protected]\"\n\n var account UserAccount\n if err := Create(handle, email, \"timisadork\", &account); err != nil {\n t.Error(err)\n }\n\n if err := EmailExists(email); err == nil {\n t.Error(\"Handle exists failed\")\n }\n\n if err := Delete(account.Key); err != nil {\n t.Error(err)\n }\n}", "func notTested() string {\n\treturn \"This function isn't tested!\"\n}", "func TestAddress(t *testing.T) {\n addr, err := StringAddress(m_pub2)\n if err != nil {\n t.Errorf(\"%s should have been nil\",err.Error())\n }\n expected_addr := \"1AEg9dFEw29kMgaN4BNHALu7AzX5XUfzSU\"\n if addr != expected_addr {\n t.Errorf(\"\\n%s\\nshould be\\n%s\",addr,expected_addr)\n }\n}", "func testIndexIcon(t *testing.T, s *Service) {\n\tres, err := s.IndexIcon(context.TODO())\n\tif err != nil {\n\t\tt.Logf(\"testIndexIcon error(%v) \\n\", err)\n\t\treturn\n\t}\n\tt.Logf(\"testIndexIcon res: %+v \\n\", res)\n}", "func TestGetSlidesDocumentWithFormatInvalidformat(t *testing.T) {\n request := createGetSlidesDocumentWithFormatRequest()\n request.format = invalidizeTestParamValue(request.format, \"format\", \"string\").(string)\n e := initializeTest(\"GetSlidesDocumentWithFormat\", \"format\", request.format)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n _, r, e := getTestApiClient().DocumentApi.GetSlidesDocumentWithFormat(request)\n assertError(t, \"GetSlidesDocumentWithFormat\", \"format\", int32(r.StatusCode), e)\n}", "func TestMoby21233(t *testing.T) {\n\tgo testTransfer() // G1\n}", "func TestPutSlidesConvertInvalidoutPath(t *testing.T) {\n request := createPutSlidesConvertRequest()\n request.outPath = invalidizeTestParamValue(request.outPath, \"outPath\", \"string\").(string)\n e := initializeTest(\"PutSlidesConvert\", \"outPath\", request.outPath)\n if e != nil {\n t.Errorf(\"Error: %v.\", e)\n return\n }\n _, r, e := getTestApiClient().DocumentApi.PutSlidesConvert(request)\n assertError(t, \"PutSlidesConvert\", \"outPath\", int32(r.StatusCode), e)\n}", "func TestReferences(t *testing.T) {\n}", "func TestGetter(t *testing.T){\n\tvar f Getter =GetterFunc(func(key string) ([]byte,error){\n\t\treturn []byte(key),nil\n\t})\n\texpect := []byte(\"key\")\n\tif v,_ :=f.Get(\"key\");!reflect.DeepEqual(v,expect){\n\t\tt.Errorf(\"callback failed\")\n\t}\n}", "func TestGenString(t *testing.T) {\n\tif genStr := GenString(); genStr != \"hello world\" {\n\t\tt.Error(\"The generated string should be \\\"hello word\\\", got \", genStr)\n\t}\n}", "func testBuildETag(t *testing.T) {\n\tt.Parallel()\n\n\t// base case\n\tpath := \"/\"\n\tformat := modules.SkyfileFormatNotSpecified\n\tvar skylink modules.Skylink\n\terr := skylink.LoadString(\"AACogzrAimYPG42tDOKhS3lXZD8YvlF8Q8R17afe95iV2Q\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\teTag := buildETag(skylink, \"GET\", path, format)\n\tif eTag != \"a58dd03937cacfeaa6974c8d12c1758bc05d8a49777eda3be52a9ba701364804\" {\n\t\tt.Fatal(\"unexpected output\")\n\t}\n\n\t// adjust URL and expect different hash value\n\tpath = \"/foo\"\n\teTag2 := buildETag(skylink, \"GET\", path, format)\n\tif eTag2 == \"\" || eTag2 == eTag {\n\t\tt.Fatal(\"unexpected output\")\n\t}\n\n\t// adjust query and expect different hash value\n\tformat = modules.SkyfileFormatZip\n\teTag3 := buildETag(skylink, \"GET\", path, format)\n\tif eTag3 == \"\" || eTag3 == eTag2 {\n\t\tt.Fatal(\"unexpected output\")\n\t}\n\n\t// adjust skylink and expect different hash value\n\terr = skylink.LoadString(\"BBCogzrAimYPG42tDOKhS3lXZD8YvlF8Q8R17afe95iV2Q\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\teTag4 := buildETag(skylink, \"GET\", path, format)\n\tif eTag4 == \"\" || eTag4 == eTag3 {\n\t\tt.Fatal(\"unexpected output\")\n\t}\n\n\t// adjust method and expect different hash value\n\terr = skylink.LoadString(\"BBCogzrAimYPG42tDOKhS3lXZD8YvlF8Q8R17afe95iV2Q\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\teTag5 := buildETag(skylink, \"HEAD\", path, format)\n\tif eTag5 == \"\" || eTag5 == eTag4 {\n\t\tt.Fatal(\"unexpected output\")\n\t}\n}", "func TestToOneRemove(t *testing.T) {}", "func TestToOneRemove(t *testing.T) {}", "func TestToOneRemove(t *testing.T) {}", "func TestToOneRemove(t *testing.T) {}", "func TestToOneRemove(t *testing.T) {}", "func TestToOneRemove(t *testing.T) {}", "func TestToOneRemove(t *testing.T) {}", "func TestToOneRemove(t *testing.T) {}", "func TestGetAllOrdersForTableID(t *testing.T) {\n\n // ...\n\n}" ]
[ "0.6145971", "0.57826537", "0.5515557", "0.54874736", "0.5458228", "0.5443233", "0.54182136", "0.54031444", "0.538412", "0.5362841", "0.53577185", "0.5339525", "0.5329989", "0.5237156", "0.5230411", "0.5178454", "0.5160178", "0.51488256", "0.51488256", "0.51488256", "0.51488256", "0.51156926", "0.51156926", "0.51156926", "0.51156926", "0.511329", "0.510814", "0.50998676", "0.5079899", "0.50540966", "0.5053124", "0.5050753", "0.50505525", "0.5042255", "0.5039012", "0.5034737", "0.50344735", "0.5029783", "0.5015634", "0.5013481", "0.5010869", "0.50085264", "0.5001484", "0.49938458", "0.49758285", "0.49655616", "0.4959245", "0.4954333", "0.49542427", "0.49494717", "0.49348992", "0.4929759", "0.4927061", "0.49264678", "0.49114478", "0.49066103", "0.4905554", "0.4905554", "0.4905554", "0.4905554", "0.4905554", "0.4905554", "0.4905554", "0.4905554", "0.4905554", "0.49015352", "0.4900474", "0.48972502", "0.4895302", "0.488078", "0.48805153", "0.48707372", "0.48599407", "0.48549044", "0.48397404", "0.4826997", "0.48146012", "0.48028672", "0.48007917", "0.47990593", "0.47910216", "0.47868377", "0.47866374", "0.47860128", "0.47825223", "0.47802413", "0.4777756", "0.47749096", "0.47723523", "0.4770002", "0.47680604", "0.47675666", "0.4764391", "0.4764391", "0.4764391", "0.4764391", "0.4764391", "0.4764391", "0.4764391", "0.4764391", "0.47566327" ]
0.0
-1
Next gets the next AcceptedProposal or blocks
func (r *walTopicReader) Next(ctx context.Context) (*types.AcceptedProposal, error) { return r.next(ctx, false) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p *Pacer) Next() {\n\t// Wait for item to be read from gate.\n\tp.gate <- struct{}{}\n}", "func (b *BlockProcessorQueue) ConfirmedNext() (uint64, bool) {\n\n\tresp := make(chan struct {\n\t\tStatus bool\n\t\tNumber uint64\n\t})\n\treq := Next{ResponseChan: resp}\n\n\tb.ConfirmedNextChan <- req\n\n\tv := <-resp\n\treturn v.Number, v.Status\n\n}", "func (req *Request) Next() ([]blocks.Block, error) {\n\tselect {\n\tcase bs := <-req.bs:\n\t\treturn bs, nil\n\tcase <-req.Done():\n\t\tselect {\n\t\tcase bs := <-req.bs:\n\t\t\treturn bs, nil\n\t\tdefault:\n\t\t\tif req.err != nil {\n\t\t\t\treturn nil, req.err\n\t\t\t}\n\n\t\t\treturn nil, io.EOF\n\t\t}\n\t}\n}", "func (m *ClientMech) Next(_ context.Context, _ []byte) ([]byte, error) {\n\tif m.done {\n\t\treturn nil, fmt.Errorf(\"unexpected challenge\")\n\t}\n\n\tm.done = true\n\treturn nil, nil\n}", "func (pb *PB) Next() bool {}", "func (it *OfferIterator) Next() (*channelpb.Offer, error) {\n\tvar item *channelpb.Offer\n\tif err := it.nextFunc(); err != nil {\n\t\treturn item, err\n\t}\n\titem = it.items[0]\n\tit.items = it.items[1:]\n\treturn item, nil\n}", "func (pb *PB) Next() bool", "func (pn *paxosNode) GetNextProposalNumber(args *paxosrpc.ProposalNumberArgs, reply *paxosrpc.ProposalNumberReply) error {\n\t// Will just give the Max([Nh/k]*k + id , )\n\tkey := args.Key\n\tpxi := pn.getInstance(key)\n\n\tpxi.mu.RLock()\n\tdefer pxi.mu.RUnlock()\n\n\treplyN := (pxi.Nh/pn.numNodes+1)*pn.numNodes + pn.id\n\treply.N = replyN\n\n\treturn nil\n}", "func (p *Player) Next() { p.Player.Call(INTERFACE+\".Player.Next\", 0) }", "func (it *PurchasableOfferIterator) Next() (*channelpb.PurchasableOffer, error) {\n\tvar item *channelpb.PurchasableOffer\n\tif err := it.nextFunc(); err != nil {\n\t\treturn item, err\n\t}\n\titem = it.items[0]\n\tit.items = it.items[1:]\n\treturn item, nil\n}", "func (s *RPC) Next(c context.Context, agentFilter rpc.Filter) (*rpc.Workflow, error) {\n\tmetadata, ok := grpcMetadata.FromIncomingContext(c)\n\tif ok {\n\t\thostname, ok := metadata[\"hostname\"]\n\t\tif ok && len(hostname) != 0 {\n\t\t\tlog.Debug().Msgf(\"agent connected: %s: polling\", hostname[0])\n\t\t}\n\t}\n\n\tfn, err := createFilterFunc(agentFilter)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor {\n\t\tagent, err := s.getAgentFromContext(c)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t} else if agent.NoSchedule {\n\t\t\treturn nil, nil\n\t\t}\n\n\t\ttask, err := s.queue.Poll(c, agent.ID, fn)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t} else if task == nil {\n\t\t\treturn nil, nil\n\t\t}\n\n\t\tif task.ShouldRun() {\n\t\t\tworkflow := new(rpc.Workflow)\n\t\t\terr = json.Unmarshal(task.Data, workflow)\n\t\t\treturn workflow, err\n\t\t}\n\n\t\tif err := s.Done(c, task.ID, rpc.State{}); err != nil {\n\t\t\tlog.Error().Err(err).Msgf(\"mark task '%s' done failed\", task.ID)\n\t\t}\n\t}\n}", "func (stream *VTGateStream) Next() (*sqltypes.Result, error) {\n\tticker := time.Tick(10 * time.Second)\n\tselect {\n\tcase s := <-stream.respChan:\n\t\treturn s, nil\n\tcase <-ticker:\n\t\treturn nil, fmt.Errorf(\"time limit exceeded\")\n\t}\n}", "func (it *TransferableOfferIterator) Next() (*channelpb.TransferableOffer, error) {\n\tvar item *channelpb.TransferableOffer\n\tif err := it.nextFunc(); err != nil {\n\t\treturn item, err\n\t}\n\titem = it.items[0]\n\tit.items = it.items[1:]\n\treturn item, nil\n}", "func (page *PeerAsnListResultPage) Next() error {\n\treturn page.NextWithContext(context.Background())\n}", "func NewAcceptedProposal(data []byte) (*AcceptedProposal, error) {\n\tvar ap AcceptedProposal\n\tdec := gob.NewDecoder(bytes.NewBuffer(data))\n\tif err := dec.Decode(&ap); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &ap, nil\n}", "func (it *GuestPolicyIterator) Next() (*osconfigpb.GuestPolicy, error) {\n\tvar item *osconfigpb.GuestPolicy\n\tif err := it.nextFunc(); err != nil {\n\t\treturn item, err\n\t}\n\titem = it.items[0]\n\tit.items = it.items[1:]\n\treturn item, nil\n}", "func (pn *paxosNode) GetNextProposalNumber(args *paxosrpc.ProposalNumberArgs, reply *paxosrpc.ProposalNumberReply) error {\n\treturn errors.New(\"not implemented\")\n}", "func (px *Paxos) Start(seq int, v interface{}) {\n // Your code here.\n go func(seq int, v interface{}) {\n DPrintf(\"&&&&&&&&&&&&&&& Paxos %v making proposal [%d]..\\n\", px.me, seq)\n n := 0\n ins := px.getProp(seq)\n\n for px.dead==false && ins.status==false { //If the proposal has not been decided yet.\n DPrintf(\"Making for Seq %v..\\n\", seq)\n DPrintf(\"Current Status [np: %v, na: %v, nv: %s, status: %v]...\\n\", ins.n_p, ins.n_a, ins.v_a, ins.status)\n count := 0\n max_n := -1\n max_v := v\n\n //Prepare stage.\n DPrintf(\"$$$$$$$ Prepare Stage $$$$$$$$$\\n\")\n for i := 0; i < len(px.peers); i++ {\n args := &PrepareArgs{seq, n}\n var reply PrepareReply\n if px.sendPrepare(i, args, &reply) && reply.Err == PrepareOK {\n DPrintf(\"Receive Agreement from %v..\\n\", i)\n px.mu.Lock()\n px.decisions[i] = reply.Done\n px.mu.Unlock()\n count = count + 1\n if reply.N_a > max_n {\n max_n = reply.N_a\n max_v = reply.V_a\n }\n } else {\n DPrintf(\"No Agreement from %v..\\n\", i)\n }\n }\n if count < px.majorSize { // do not make majority prepare\n ins = px.getProp(seq)\n n = max(max_n, n) + 1\n DPrintf(\"No Majority Prepare. Change n to [%v]..\\n\", n)\n time.Sleep(20 * time.Millisecond)\n continue\n }\n\n // Accept stage. Will come here if majority are prepared.\n DPrintf(\"$$$$$$$ Accept Stage $$$$$$$$$\\n\")\n count = 0\n for i := 0; i < len(px.peers); i++ {\n args := &AcceptArgs{seq, n, max_v}\n var reply AcceptReply\n if px.sendAccept(i, args, &reply) && reply.Err == AcceptOK {\n px.mu.Lock()\n px.decisions[i] = reply.Done\n px.mu.Unlock()\n count = count + 1\n }\n }\n if count < px.majorSize { // do not make majority Accept\n ins = px.getProp(seq)\n n = max(max_n, n) + 1\n time.Sleep(20 * time.Millisecond)\n continue\n }\n\n // Decide stage\n DPrintf(\"$$$$$$$ Decide Stage $$$$$$$$$\\n\")\n for i := 0; i < len(px.peers); i++ {\n args := &DecideArgs{seq, n, max_v}\n var reply DecideReply\n if px.sendDecide(i, args, &reply) && reply.Err == DecideOK {\n px.mu.Lock()\n px.decisions[i] = reply.Done\n px.mu.Unlock()\n }\n }\n DPrintf(\"After a new Instance %d, history table becomes...\\n\", seq)\n px.PrintHis();\n break\n }\n }(seq, v)\n}", "func (s *Subscriber) Next() (*Message, bool) {\n\tselect {\n\tcase m := <-s.msgs:\n\t\treturn m, false\n\tcase <-s.done:\n\t\treturn nil, true\n\t}\n}", "func (b *BlockProcessorQueue) UnconfirmedNext() (uint64, bool) {\n\n\tresp := make(chan struct {\n\t\tStatus bool\n\t\tNumber uint64\n\t})\n\treq := Next{ResponseChan: resp}\n\n\tb.UnconfirmedNextChan <- req\n\n\tv := <-resp\n\treturn v.Number, v.Status\n\n}", "func (c *Client) next() (rsp *Response, err error) {\n\traw, err := c.r.Next()\n\tif err == nil {\n\t\trsp, err = raw.Parse()\n\t}\n\treturn\n}", "func (message *Message) Next() {\n\tif message == nil {\n\t\treturn\n\t}\n\n\tmessage.resolve(nil)\n}", "func (query *Query) GetNextProposalID() (*model.NextProposalID, error) {\n\tresp, err := query.transport.Query(GetNextProposalIDKey(), ProposalKVStoreKey)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tnextProposalID := new(model.NextProposalID)\n\tif err := query.transport.Cdc.UnmarshalJSON(resp, nextProposalID); err != nil {\n\t\treturn nil, err\n\t}\n\treturn nextProposalID, nil\n}", "func (cm *ConsensusManager) shareProposal(rs *RoundState) {\n\tprivValidator := cm.PrivValidator()\n\tproposal := rs.Proposal\n\tif privValidator == nil || proposal == nil {\n\t\treturn\n\t}\n\tprivValidator.SignProposal(rs.Round, proposal)\n\tblockParts := proposal.BlockParts()\n\tpeers := cm.sw.Peers().List()\n\tif len(peers) == 0 {\n\t\tlog.Warning(\"Could not propose: no peers\")\n\t\treturn\n\t}\n\tnumBlockParts := uint16(len(blockParts))\n\tkbpMsg := cm.makeKnownBlockPartsMessage(rs)\n\tfor i, peer := range peers {\n\t\tpeerState := cm.getPeerState(peer)\n\t\tif !peerState.IsConnected() {\n\t\t\tcontinue // Peer was disconnected.\n\t\t}\n\t\tstartIndex := uint16((i * len(blockParts)) / len(peers))\n\t\t// Create a function that when called,\n\t\t// starts sending block parts to peer.\n\t\tcb := func(peer *p2p.Peer, startIndex uint16) func() {\n\t\t\treturn func() {\n\t\t\t\t// TODO: if the clocks are off a bit,\n\t\t\t\t// peer may receive this before the round flips.\n\t\t\t\tpeer.Send(KnownPartsCh, kbpMsg)\n\t\t\t\tfor i := uint16(0); i < numBlockParts; i++ {\n\t\t\t\t\tpart := blockParts[(startIndex+i)%numBlockParts]\n\t\t\t\t\t// Ensure round hasn't expired on our end.\n\t\t\t\t\tcurrentRS := cm.cs.RoundState()\n\t\t\t\t\tif currentRS != rs {\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t\t// If peer wants the block:\n\t\t\t\t\tif peerState.WantsBlockPart(part) {\n\t\t\t\t\t\tpartMsg := &BlockPartMessage{BlockPart: part}\n\t\t\t\t\t\tpeer.Send(ProposalCh, partMsg)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}(peer, startIndex)\n\t\t// Call immediately or schedule cb for when peer is ready.\n\t\tpeerState.SetRoundCallback(rs.Height, rs.Round, cb)\n\t}\n}", "func (c *Client) Next() goa.Endpoint {\n\tvar (\n\t\tdecodeResponse = DecodeNextResponse(c.decoder, c.RestoreResponseBody)\n\t)\n\treturn func(ctx context.Context, v interface{}) (interface{}, error) {\n\t\treq, err := c.BuildNextRequest(ctx, v)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresp, err := c.NextDoer.Do(req)\n\t\tif err != nil {\n\t\t\treturn nil, goahttp.ErrRequestError(\"spin-broker\", \"next\", err)\n\t\t}\n\t\treturn decodeResponse(resp)\n\t}\n}", "func Next() (int, Response) {\n\treturn 0, Response{}\n}", "func (px *Paxos) Start(seq int, v interface{}) {\n \n // if consensus on this seq number is no longer\n // desired from the client, ignore this request\n if seq < px.Min() {\n return\n }\n\n // access state up front\n px.mu.Lock()\n \n doneMap := make(map[int]int) // local copy of the done map\n for k, v := range px.done { \n doneMap[k] = v\n }\n\n px.freeMemory()\n \n px.mu.Unlock()\n\n // Question could you optimize the initial proposal number?\n\n go func(seq int, v interface{}) { // launch thread\n np := ProposalId { 0, px.me } // initial proposal \n instance := px.getInstance(seq) // retrieve instance from log\n\n // while loop driving consensus:\n for !px.isdead() && instance.Fate == Pending {\n na := NullProposal() // highest na seen\n va := v // va associated with (na, va)\n count := 0 // count prepare_oks\n\n // follows the pseudocode for the Paxos algorithm\n for i := 0; i < len(px.peers); i++ {\n args := PrepareArgs{seq, np}\n var reply PrepareReply\n if px.sendPrepare(i, args, &reply) && reply.Ok {\n doneMap[i] = reply.Done\n count += 1\n if reply.Na.Greater(na) {\n na = reply.Na\n va = reply.Va\n }\n }\n }\n\n // did not gain a majority response during prepare\n // check instance for consensus and try again with\n // the next highest proposal id\n if count < 1 + len(px.peers) / 2 {\n instance = px.getInstance(seq)\n if na.Greater(np) {\n np = np.Next(na)\n } else {\n np = np.Next(np)\n }\n continue // return to top of for loop\n }\n\n count = 0 // count accept_oks\n \n // follows the pseudocode for the Paxos algorithm\n for i := 0; i < len(px.peers); i++ {\n args := AcceptArgs{seq, np, va}\n var reply AcceptReply\n if px.sendAccept(i, args, &reply) && reply.Ok {\n doneMap[i] = reply.Done\n count += 1\n }\n }\n\n // did not gain a majority response during accept\n // check instance for consensus and try again with\n // the next highest proposal id\n if count < 1 + len(px.peers) / 2 {\n instance = px.getInstance(seq)\n if na.Greater(np) {\n np = np.Next(na)\n } else {\n np = np.Next(np)\n }\n continue // return to top of for loop\n }\n\n // consensus achieved, sendDecide()\n for i := 0; i < len(px.peers); i++ {\n args := DecideArgs{seq, np, va}\n var reply DecideReply\n if px.sendDecide(i, args, &reply) && reply.Ok {\n doneMap[i] = reply.Done\n }\n }\n\n // Question: Should we retry if sendDecide() fails?\n\n break // consesus achieved break out of the while loop\n }\n\n // update done pointer to address the local updated copy\n px.mu.Lock()\n \n px.done = doneMap\n px.freeMemory()\n \n px.mu.Unlock()\n\n // Question: Should I check for higher value in px.done?\n // Local copy's done[i] may be less than the global max. \n\n }(seq, v)\n}", "func (query *Query) GetNextProposalID(ctx context.Context) (*model.NextProposalID, error) {\n\tresp, err := query.transport.Query(ctx, getNextProposalIDKey(), ProposalKVStoreKey)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tnextProposalID := new(model.NextProposalID)\n\tif err := query.transport.Cdc.UnmarshalJSON(resp, nextProposalID); err != nil {\n\t\treturn nil, err\n\t}\n\treturn nextProposalID, nil\n}", "func (qi *Items) Next() *pageloader.Request {\n\tqi.Lock()\n\tlog.Println(\"pulling request from the queue\")\n\tif qi.Length == 0 {\n\t\tlog.Println(\"nothing in the queue\")\n\t\treturn nil\n\t}\n\tt := qi.Stack[0]\n\ttCopy := *t\n\tqi.Stack = qi.Stack[1:]\n\tqi.Length--\n\tlog.Printf(\"queue length now: %d\\n\", qi.Length)\n\tqi.Unlock()\n\treturn &tCopy\n}", "func EndBlocker(ctx sdk.Context, keeper Keeper) sdk.Tags {\n\tlogger := ctx.Logger().With(\"module\", \"x/gov\")\n\tresTags := sdk.NewTags()\n\n\tinactiveIterator := keeper.InactiveProposalQueueIterator(ctx, ctx.BlockHeader().Time)\n\tdefer inactiveIterator.Close()\n\tfor ; inactiveIterator.Valid(); inactiveIterator.Next() {\n\t\tvar proposalID uint64\n\n\t\tkeeper.cdc.MustUnmarshalBinaryLengthPrefixed(inactiveIterator.Value(), &proposalID)\n\t\tinactiveProposal, ok := keeper.GetProposal(ctx, proposalID)\n\t\tif !ok {\n\t\t\tpanic(fmt.Sprintf(\"proposal %d does not exist\", proposalID))\n\t\t}\n\n\t\tkeeper.DeleteProposal(ctx, proposalID)\n\t\tkeeper.DeleteDeposits(ctx, proposalID) // delete any associated deposits (burned)\n\n\t\tresTags = resTags.AppendTag(tags.ProposalID, fmt.Sprintf(\"%d\", proposalID))\n\t\tresTags = resTags.AppendTag(tags.ProposalResult, tags.ActionProposalDropped)\n\n\t\tlogger.Info(\n\t\t\tfmt.Sprintf(\"proposal %d (%s) didn't meet minimum deposit of %s (had only %s); deleted\",\n\t\t\t\tinactiveProposal.GetProposalID(),\n\t\t\t\tinactiveProposal.GetTitle(),\n\t\t\t\tkeeper.GetDepositParams(ctx).MinDeposit,\n\t\t\t\tinactiveProposal.GetTotalDeposit(),\n\t\t\t),\n\t\t)\n\t}\n\n\t// fetch active proposals whose voting periods have ended (are passed the block time)\n\tactiveIterator := keeper.ActiveProposalQueueIterator(ctx, ctx.BlockHeader().Time)\n\tdefer activeIterator.Close()\n\tfor ; activeIterator.Valid(); activeIterator.Next() {\n\t\tvar proposalID uint64\n\n\t\tkeeper.cdc.MustUnmarshalBinaryLengthPrefixed(activeIterator.Value(), &proposalID)\n\t\tactiveProposal, ok := keeper.GetProposal(ctx, proposalID)\n\t\tif !ok {\n\t\t\tpanic(fmt.Sprintf(\"proposal %d does not exist\", proposalID))\n\t\t}\n\t\tresult, tallyResults, votingVals := tally(ctx, keeper, activeProposal)\n\n\t\tvar tagValue string\n\t\tif result == PASS {\n\t\t\tkeeper.RefundDeposits(ctx, activeProposal.GetProposalID())\n\t\t\tactiveProposal.SetStatus(StatusPassed)\n\t\t\ttagValue = tags.ActionProposalPassed\n\t\t\tExecute(ctx, keeper, activeProposal)\n\t\t} else if result == REJECT {\n\t\t\tkeeper.RefundDeposits(ctx, activeProposal.GetProposalID())\n\t\t\tactiveProposal.SetStatus(StatusRejected)\n\t\t\ttagValue = tags.ActionProposalRejected\n\t\t} else if result == REJECTVETO {\n\t\t\tkeeper.DeleteDeposits(ctx, activeProposal.GetProposalID())\n\t\t\tactiveProposal.SetStatus(StatusRejected)\n\t\t\ttagValue = tags.ActionProposalRejected\n\t\t}\n\n\t\tactiveProposal.SetTallyResult(tallyResults)\n\t\tkeeper.SetProposal(ctx, activeProposal)\n\t\tkeeper.RemoveFromActiveProposalQueue(ctx, activeProposal.GetVotingEndTime(), activeProposal.GetProposalID())\n\n\t\tlogger.Info(\n\t\t\tfmt.Sprintf(\n\t\t\t\t\"proposal %d (%s) tallied; passed: %v\",\n\t\t\t\tactiveProposal.GetProposalID(), activeProposal.GetTitle(), result,\n\t\t\t),\n\t\t)\n\n\t\tresTags = resTags.AppendTag(tags.ProposalID, fmt.Sprintf(\"%d\", proposalID))\n\t\tresTags = resTags.AppendTag(tags.ProposalResult, tagValue)\n\n\t\tfor _, valAddr := range keeper.GetValidatorSet(ctx, proposalID) {\n\t\t\tif _, ok := votingVals[valAddr.String()]; !ok {\n\t\t\t\tval := keeper.ds.GetValidatorSet().Validator(ctx, valAddr)\n\t\t\t\tif val != nil && val.GetStatus() == sdk.Bonded {\n\t\t\t\t\tkeeper.ds.GetValidatorSet().Slash(ctx,\n\t\t\t\t\t\tval.GetConsAddr(),\n\t\t\t\t\t\tctx.BlockHeight(),\n\t\t\t\t\t\tval.GetBondedTokens().Int64(),\n\t\t\t\t\t\tkeeper.GetTallyParams(ctx).Penalty)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tkeeper.DeleteValidatorSet(ctx, activeProposal.GetProposalID())\n\t}\n\treturn resTags\n}", "func (a *acceptor) receivePropose(propose message) bool {\n\tif a.promised.number() > propose.number() {\n\t\tlog.Printf(\"acceptor: %d [promised: %+v] ignored proposal %+v\", a.id, a.promised, propose)\n\t\treturn false\n\t}\n\tlog.Printf(\"acceptor: %d [promised: %+v, accept: %+v] accepted proposal %+v\", a.id, a.promised, a.accept, propose)\n\ta.accept = propose\n\ta.accept.typ = Accept\n\treturn true\n}", "func (pn *paxosNode) GetNextProposalNumber(args *paxosrpc.ProposalNumberArgs, reply *paxosrpc.ProposalNumberReply) error {\n\tkey := args.Key\n\tpn.nextProposalNumMapLock.Lock()\n\tdefer pn.nextProposalNumMapLock.Unlock()\n\n\tif _, exist := pn.nextProposalNumMap[key]; !exist {\n\t\tpn.nextProposalNumMap[key] = pn.srvId\n\t} else {\n\t\tpn.nextProposalNumMap[key] += pn.numNodes\n\t}\n\n\treply.N = pn.nextProposalNumMap[key]\n\n\tfmt.Printf(\"node %d: key = %s, next prop num = %d\\n\", pn.srvId, key, reply.N)\n\treturn nil\n}", "func (s Service) Next(msg *Message, data interface{}, useMeta map[string]string) error {\n\terr := msg.RawMessage.Ack(true)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif msg.Chain == nil {\n\t\treturn ErrorChainIsEmpty\n\t}\n\n\tchain := SetCurrentItemSuccess(msg.Chain)\n\n\tnextIndex := getCurrentChainIndex(chain)\n\tif nextIndex == -1 {\n\t\treturn ErrorNextIsNotDefined\n\t}\n\n\tnextElement := chain[nextIndex]\n\n\tmeta := Meta{}\n\tmeta.Merge(msg.Meta, useMeta)\n\n\tvar items []interface{}\n\n\tif nextElement.IsMultiple {\n\t\tval := reflect.ValueOf(data)\n\n\t\tif val.Kind() != reflect.Slice {\n\t\t\treturn ErrorDataIsNotArray\n\t\t}\n\n\t\tfor i := 0; i < val.Len(); i++ {\n\t\t\titems = append(items, val.Index(i).Interface())\n\t\t}\n\t} else {\n\t\titems = append(items, data)\n\t}\n\n\tfor _, item := range items {\n\t\tb, err := json.Marshal(item)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tm := InitialMessage{\n\t\t\tCatch: msg.InitialMessage.Catch,\n\t\t\tConfig: msg.Config,\n\t\t\tMeta: meta,\n\t\t\tChain: chain,\n\t\t\tData: b,\n\t\t}\n\n\t\tif err := s.Publish(m); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (q *QuestionnaireT) Next() int {\n\tpg, _ := q.nextInNavi()\n\treturn pg\n}", "func (p *Parser) Next() {\n\tp.curr = p.peek\n\tp.peek = p.l.NextToken()\n}", "func (s *state) next(last bool) (n nonce) {\n\t// tag\n\tt := make([]byte, 1)\n\n\t// prepare block number\n\tb := make([]byte, 4)\n\tbinary.BigEndian.PutUint32(b, s.block)\n\n\t// prepare tag\n\tn = append(n, s.seed...)\n\tn = append(n, b...)\n\tn = append(n, t...)\n\n\tif last {\n\t\tn.last()\n\t}\n\n\t// increment\n\ts.block++\n\n\t// now we have the nonce..\n\t// not sure if we need to apply transformations..\n\treturn\n}", "func EndBlocker(ctx sdk.Context, keeper Keeper) sdk.Tags {\n\tlogger := keeper.Logger(ctx)\n\tresTags := sdk.NewTags()\n\n\tinactiveIterator := keeper.InactiveProposalQueueIterator(ctx, ctx.BlockHeader().Time)\n\tdefer inactiveIterator.Close()\n\tfor ; inactiveIterator.Valid(); inactiveIterator.Next() {\n\t\tvar proposalID uint64\n\n\t\tkeeper.cdc.MustUnmarshalBinaryLengthPrefixed(inactiveIterator.Value(), &proposalID)\n\t\tinactiveProposal, ok := keeper.GetProposal(ctx, proposalID)\n\t\tif !ok {\n\t\t\tpanic(fmt.Sprintf(\"proposal %d does not exist\", proposalID))\n\t\t}\n\n\t\tkeeper.DeleteProposal(ctx, proposalID)\n\t\tkeeper.DeleteDeposits(ctx, proposalID) // delete any associated deposits (burned)\n\n\t\tresTags = resTags.AppendTag(tags.ProposalID, fmt.Sprintf(\"%d\", proposalID))\n\t\tresTags = resTags.AppendTag(tags.ProposalResult, tags.ActionProposalDropped)\n\n\t\tlogger.Info(\n\t\t\tfmt.Sprintf(\"proposal %d (%s) didn't meet minimum deposit of %s (had only %s); deleted\",\n\t\t\t\tinactiveProposal.ProposalID,\n\t\t\t\tinactiveProposal.GetTitle(),\n\t\t\t\tkeeper.GetDepositParams(ctx).MinDeposit,\n\t\t\t\tinactiveProposal.TotalDeposit,\n\t\t\t),\n\t\t)\n\t}\n\n\t// fetch active proposals whose voting periods have ended (are passed the block time)\n\tactiveIterator := keeper.ActiveProposalQueueIterator(ctx, ctx.BlockHeader().Time)\n\tdefer activeIterator.Close()\n\tfor ; activeIterator.Valid(); activeIterator.Next() {\n\t\tvar proposalID uint64\n\n\t\tkeeper.cdc.MustUnmarshalBinaryLengthPrefixed(activeIterator.Value(), &proposalID)\n\t\tactiveProposal, ok := keeper.GetProposal(ctx, proposalID)\n\t\tif !ok {\n\t\t\tpanic(fmt.Sprintf(\"proposal %d does not exist\", proposalID))\n\t\t}\n\t\tpasses, burnDeposits, tallyResults := tally(ctx, keeper, activeProposal)\n\n\t\tvar tagValue, logMsg string\n\n\t\tif burnDeposits {\n\t\t\tkeeper.DeleteDeposits(ctx, activeProposal.ProposalID)\n\t\t} else {\n\t\t\tkeeper.RefundDeposits(ctx, activeProposal.ProposalID)\n\t\t}\n\n\t\tif passes {\n\t\t\thandler := keeper.router.GetRoute(activeProposal.ProposalRoute())\n\t\t\tcacheCtx, writeCache := ctx.CacheContext()\n\n\t\t\t// The proposal handler may execute state mutating logic depending\n\t\t\t// on the proposal content. If the handler fails, no state mutation\n\t\t\t// is written and the error message is logged.\n\t\t\terr := handler(cacheCtx, activeProposal.Content)\n\t\t\tif err == nil {\n\t\t\t\tactiveProposal.Status = StatusPassed\n\t\t\t\ttagValue = tags.ActionProposalPassed\n\t\t\t\tlogMsg = \"passed\"\n\n\t\t\t\t// write state to the underlying multi-store\n\t\t\t\twriteCache()\n\t\t\t} else {\n\t\t\t\tactiveProposal.Status = StatusFailed\n\t\t\t\ttagValue = tags.ActionProposalFailed\n\t\t\t\tlogMsg = fmt.Sprintf(\"passed, but failed on execution: %s\", err.ABCILog())\n\t\t\t}\n\t\t} else {\n\t\t\tactiveProposal.Status = StatusRejected\n\t\t\ttagValue = tags.ActionProposalRejected\n\t\t\tlogMsg = \"rejected\"\n\t\t}\n\n\t\tactiveProposal.FinalTallyResult = tallyResults\n\n\t\tkeeper.SetProposal(ctx, activeProposal)\n\t\tkeeper.RemoveFromActiveProposalQueue(ctx, activeProposal.VotingEndTime, activeProposal.ProposalID)\n\n\t\tlogger.Info(\n\t\t\tfmt.Sprintf(\n\t\t\t\t\"proposal %d (%s) tallied; result: %s\",\n\t\t\t\tactiveProposal.ProposalID, activeProposal.GetTitle(), logMsg,\n\t\t\t),\n\t\t)\n\n\t\tresTags = resTags.AppendTag(tags.ProposalID, fmt.Sprintf(\"%d\", proposalID))\n\t\tresTags = resTags.AppendTag(tags.ProposalResult, tagValue)\n\t}\n\n\treturn resTags\n}", "func (p *DelayProcess) Next() interface{} {\n\treturn p.Successor\n}", "func (iter *PeerAsnListResultIterator) Next() error {\n\treturn iter.NextWithContext(context.Background())\n}", "func Next() {\n\tgo next()\n}", "func (page *VirtualMachineListResultPageClient) Next() error {\n\treturn page.vmlrp.Next()\n}", "func (c *Conn) Next() EventInfo {\n\tvar evt EventInfo\n\n\tif c.nextEvent != nil {\n\t\tevt = *c.nextEvent\n\t\tc.nextEvent = nil\n\t\treturn evt\n\t}\n\tif !c.replied && c.curcmd != noCmd {\n\t\tif c.state == sAuth {\n\t\t\t// send empty challenge instead of auto accept\n\t\t\t// to prevent accidental auth success.\n\t\t\tc.AuthChallenge(nil)\n\t\t} else {\n\t\t\tc.Accept()\n\t\t}\n\t}\n\tif c.state == sStartup {\n\t\tvar announce string\n\t\tc.state = sInitial\n\t\t// log preceeds the banner in case the banner hits an error.\n\t\tc.log(\"#\", \"remote %v at %s\", c.conn.RemoteAddr(),\n\t\t\ttime.Now().Format(TimeFmt))\n\t\tif c.Config.Announce != \"\" {\n\t\t\tannounce = \"\\n\" + c.Config.Announce\n\t\t}\n\t\tif c.Config.SayTime {\n\t\t\tc.replyMulti(220, \"%s %s %s%s\",\n\t\t\t\tc.Config.LocalName, c.Config.SftName,\n\t\t\t\ttime.Now().Format(time.RFC1123Z), announce)\n\t\t} else {\n\t\t\tc.replyMulti(220, \"%s %s%s\", c.Config.LocalName,\n\t\t\t\tc.Config.SftName, announce)\n\t\t}\n\t}\n\n\t// Read and parse client AUTH response. Note that AUTH responses\n\t// are not SMTP commands. During state sAuth, the only events we\n\t// can return are AUTHRESP, AUTHABORT, and ABORT.\n\tif c.state == sAuth {\n\t\tdata := c.readAuthResp()\n\t\tif c.state == sAbort {\n\t\t\tevt.What = ABORT\n\t\t\tc.log(\"#\", \"abort at %v\", time.Now().Format(TimeFmt))\n\t\t\treturn evt\n\t\t}\n\t\tif data == \"*\" {\n\t\t\tc.authDone(false)\n\t\t\tc.reply(ReplyAuthAborted)\n\t\t\tevt.What = AUTHABORT\n\t\t} else {\n\t\t\tc.replied = false\n\t\t\tevt.What = AUTHRESP\n\t\t\tevt.Arg = data\n\t\t}\n\t\treturn evt\n\t}\n\n\t// Read DATA chunk if called for.\n\tif c.state == sData {\n\t\tdata := c.readData()\n\t\tif len(data) > 0 {\n\t\t\tevt.What = GOTDATA\n\t\t\tevt.Arg = data\n\t\t\tc.replied = false\n\t\t\t// This is technically correct; only a *successful*\n\t\t\t// DATA block ends the mail transaction according to\n\t\t\t// the RFCs. An unsuccessful one must be RSET.\n\t\t\tc.state = sPostData\n\t\t\tc.nstate = sHelo\n\t\t\treturn evt\n\t\t}\n\t\t// If the data read failed, c.state will be sAbort and we\n\t\t// will exit in the main loop.\n\t}\n\n\t// Main command loop.\n\tfor {\n\t\tif c.stopme() {\n\t\t\tbreak\n\t\t}\n\n\t\tline := c.readCmd()\n\t\tif line == \"\" {\n\t\t\tbreak\n\t\t}\n\n\t\tres, err := ParseCmd(line)\n\t\tif err != nil {\n\t\t\tc.badcmds++\n\t\t\tc.reply(ReplySyntaxError, err.Error())\n\t\t\tcontinue\n\t\t}\n\t\t// Is this command valid in this state at all?\n\t\t// Since we implicitly support PIPELINING, which can\n\t\t// result in out of sequence commands when earlier ones\n\t\t// fail, we don't count out of sequence commands as bad\n\t\t// commands.\n\t\tt := states[res.Cmd]\n\t\tif t.validin != 0 && (t.validin&c.state) == 0 {\n\t\t\tc.reply(ReplyCmdOutOfSeq)\n\t\t\tcontinue\n\t\t}\n\n\t\t// The command is legitimate. Handle it for real.\n\n\t\t// Handle simple commands that are valid in all states.\n\t\tif t.validin == 0 {\n\t\t\tswitch res.Cmd {\n\t\t\tcase NOOP:\n\t\t\t\tc.reply(ReplyCmdOk)\n\t\t\tcase RSET:\n\t\t\t\t// It's valid to RSET before EHLO and\n\t\t\t\t// doing so can't skip EHLO.\n\t\t\t\tif c.state != sInitial {\n\t\t\t\t\tc.state = sHelo\n\t\t\t\t}\n\t\t\t\tc.reply(ReplyCmdOk)\n\t\t\t\t// RSETs are not delivered to higher levels;\n\t\t\t\t// they are implicit in sudden MAIL FROMs.\n\t\t\tcase QUIT:\n\t\t\t\tc.state = sQuit\n\t\t\t\tc.reply(ReplyGoodbye)\n\t\t\t\t// Will exit at main loop.\n\t\t\tcase HELP:\n\t\t\t\tc.reply(ReplyNoHelp)\n\t\t\tcase STARTTLS:\n\t\t\t\tif c.Config.TLSConfig == nil || c.TLSOn {\n\t\t\t\t\tc.reply(ReplyCmdNotSupported)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tc.reply(ReplyReadyForTLS)\n\t\t\t\tif c.state == sAbort {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\t// Since we're about to start chattering on\n\t\t\t\t// conn outside of our normal framework, we\n\t\t\t\t// must reset both read and write timeouts\n\t\t\t\t// to our TLS setup timeout.\n\t\t\t\tc.conn.SetDeadline(time.Now().Add(c.Config.Limits.TLSSetup))\n\t\t\t\ttlsConn := tls.Server(c.conn, c.Config.TLSConfig)\n\t\t\t\terr := tlsConn.Handshake()\n\t\t\t\tif err != nil {\n\t\t\t\t\tc.log(\"!\", \"TLS setup failed: %v\", err)\n\t\t\t\t\tc.state = sAbort\n\t\t\t\t\tevt.What = TLSERROR\n\t\t\t\t\tevt.Arg = fmt.Sprintf(\"%v\", err)\n\t\t\t\t\treturn evt\n\t\t\t\t}\n\t\t\t\t// With TLS set up, we now want no read and\n\t\t\t\t// write deadlines on the underlying\n\t\t\t\t// connection. So cancel all deadlines by\n\t\t\t\t// providing a zero value.\n\t\t\t\tc.conn.SetReadDeadline(time.Time{})\n\t\t\t\t// switch c.conn to tlsConn.\n\t\t\t\tc.setupConn(tlsConn)\n\t\t\t\tc.TLSOn = true\n\t\t\t\tc.TLSState = tlsConn.ConnectionState()\n\t\t\t\tif c.TLSState.ServerName != \"\" {\n\t\t\t\t\tc.log(\n\t\t\t\t\t\t\"!\",\n\t\t\t\t\t\t\"TLS negociated with cipher 0x%04x protocol 0x%04x server name '%s'\",\n\t\t\t\t\t\tc.TLSState.CipherSuite,\n\t\t\t\t\t\tc.TLSState.Version,\n\t\t\t\t\t\tc.TLSState.ServerName,\n\t\t\t\t\t)\n\t\t\t\t} else {\n\t\t\t\t\tc.log(\"!\",\n\t\t\t\t\t\t\"TLS negociated with cipher 0x%04x protocol 0x%04x\",\n\t\t\t\t\t\tc.TLSState.CipherSuite,\n\t\t\t\t\t\tc.TLSState.Version,\n\t\t\t\t\t)\n\t\t\t\t}\n\t\t\t\t// By the STARTTLS RFC, we return to our state\n\t\t\t\t// immediately after the greeting banner\n\t\t\t\t// and clients must re-EHLO.\n\t\t\t\tc.state = sInitial\n\t\t\tdefault:\n\t\t\t\tc.reply(ReplyCmdNotSupported)\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\t// Full state commands\n\t\tc.nstate = t.next\n\t\tc.replied = false\n\t\tc.curcmd = res.Cmd\n\n\t\tswitch res.Cmd {\n\t\tcase AUTH:\n\t\t\tif c.Config.Auth == nil {\n\t\t\t\tc.reply(ReplyCmdNotSupported)\n\t\t\t\tc.replied = true\n\t\t\t\t// AUTH with no AUTH enabled counts as a\n\t\t\t\t// bad command. This deals with a few people\n\t\t\t\t// who spam AUTH requests at non-supporting\n\t\t\t\t// servers.\n\t\t\t\tc.badcmds++\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif c.authenticated {\n\t\t\t\t// RFC4954, section 4: After an AUTH\n\t\t\t\t// command has been successfully\n\t\t\t\t// completed, no more AUTH commands\n\t\t\t\t// may be issued in the same session.\n\t\t\t\tc.reply(ReplyCmdOutOfSeq)\n\t\t\t\tc.replied = true\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif !c.authMechanismValid(res.Arg) {\n\t\t\t\tc.reply(ReplyCmdParamNotImplemented)\n\t\t\t\tc.replied = true\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// Queue initial auth response for the next\n\t\t\t// round. This way, all auth responses are\n\t\t\t// delivered with event type AUTHRESP.\n\t\t\tc.nextEvent = &EventInfo{What: AUTHRESP, Arg: res.Params}\n\t\t\tres.Params = \"\"\n\t\t\tc.state = sAuth\n\t\tcase MAILFROM, RCPTTO:\n\t\t\t// Verify that the client has authenticated.\n\t\t\t// We do this here because MAIL FROM is the\n\t\t\t// only valid full state command after\n\t\t\t// HELO/EHLO that requires authentication.\n\t\t\tif c.Config.Auth != nil && !c.authenticated {\n\t\t\t\tc.reply(ReplyAuthRequired)\n\t\t\t\tc.replied = true\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// RCPT TO:<> is invalid; reject it. Otherwise\n\t\t\t// defer all address checking to our callers.\n\t\t\tif res.Cmd == RCPTTO && len(res.Arg) == 0 {\n\t\t\t\tc.Reject()\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// reject parameters that we don't accept,\n\t\t\t// which right now is all of them. We reject\n\t\t\t// with the RFC-correct reply instead of a\n\t\t\t// generic one, so we can't use c.Reject().\n\t\t\tif res.Params != \"\" && c.Config.Limits.NoParams && !mimeParam(res) {\n\t\t\t\tc.reply(ReplyCmdParamNotImplemented)\n\t\t\t\tc.replied = true\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\n\t\t// Real, valid, in sequence command. Deliver it to our\n\t\t// caller.\n\t\tevt.What = COMMAND\n\t\tevt.Cmd = res.Cmd\n\t\t// TODO: does this hold down more memory than necessary?\n\t\tevt.Arg = res.Arg\n\t\treturn evt\n\t}\n\n\t// Explicitly mark and notify too many bad commands. This is\n\t// an out of sequence 'reply', but so what, the client will\n\t// see it if they send anything more. It will also go in the\n\t// SMTP command log.\n\tevt.Arg = \"\"\n\tif c.badcmds > c.Config.Limits.BadCmds {\n\t\tc.reply(ReplyTooManyBadCmds)\n\t\tc.state = sAbort\n\t\tevt.Arg = ArgTooManyBadCmds\n\t}\n\tif c.state == sQuit {\n\t\tevt.What = DONE\n\t\tc.log(\"#\", \"finished at %v\", time.Now().Format(TimeFmt))\n\t} else {\n\t\tevt.What = ABORT\n\t\tc.log(\"#\", \"abort at %v\", time.Now().Format(TimeFmt))\n\t}\n\treturn evt\n}", "func (it *insertIterator) next() (*types.Block, error) {\n\t// If we reached the end of the chain, abort\n\tif it.index+1 >= len(it.chain) {\n\t\tit.index = len(it.chain)\n\t\treturn nil, nil\n\t}\n\t// Advance the iterator and wait for verification result if not yet done\n\tit.index++\n\tif len(it.errors) <= it.index {\n\t\tit.errors = append(it.errors, <-it.results)\n\t}\n\tif it.errors[it.index] != nil {\n\t\treturn it.chain[it.index], it.errors[it.index]\n\t}\n\t// Block header valid, run body validation and return\n\treturn it.chain[it.index], it.validator.ValidateBody(it.chain[it.index])\n}", "func (p *offerPool) GetNextOffer(o *Offer) *Offer {\n\tp.RLock()\n\tdefer p.RUnlock()\n\tblog.V(3).Infof(\"GetNextOffer id offer(%d | %s:%s)'s next offer\", o.Id, o.offerId, o.hostname)\n\n\tif p.offerList.Len() == 0 {\n\t\tblog.Infof(\"GetNextOffer offer(%s:%s) pool don't have offers\",\n\t\t\to.offerId, o.hostname)\n\t\treturn nil\n\t}\n\n\tvar offer *Offer\n\n\t_, ok := p.offerIds[o.offerId]\n\tif ok {\n\t\tblog.V(3).Infof(\"getNextOffer offer(%s:%s) exist\", o.offerId, o.hostname)\n\n\t\thead := o.element.Next()\n\t\tfor {\n\t\t\tif head == nil {\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tinnerOffer := head.Value.(*innerOffer)\n\t\t\tblog.V(3).Infof(\"getNextOffer offer(%s:%s) isValid %t\", innerOffer.offerId,\n\t\t\t\tinnerOffer.hostname, innerOffer.isValid)\n\n\t\t\tif innerOffer.isValid {\n\t\t\t\toffer = &Offer{\n\t\t\t\t\telement: head,\n\t\t\t\t\tOffer: innerOffer.offer,\n\t\t\t\t\tId: innerOffer.id,\n\t\t\t\t\tofferId: innerOffer.offerId,\n\t\t\t\t\thostname: innerOffer.hostname,\n\t\t\t\t\tDeltaCPU: innerOffer.deltaCPU,\n\t\t\t\t\tDeltaMem: innerOffer.deltaMem,\n\t\t\t\t\tDeltaDisk: innerOffer.deltaDisk,\n\t\t\t\t}\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tblog.V(3).Infof(\"getNextOffer offer(%s:%s) get next offer\",\n\t\t\t\tinnerOffer.offerId, innerOffer.hostname)\n\t\t\thead = head.Next()\n\t\t}\n\n\t\tif offer != nil {\n\t\t\tblog.V(3).Infof(\"GetNextOffer offer(%d | %s:%s)\", offer.Id, offer.offerId, offer.hostname)\n\t\t} else {\n\t\t\tblog.Infof(\"GetNextOffer offer(%s:%s) don't have next offer\",\n\t\t\t\to.offerId, o.hostname)\n\t\t}\n\n\t\treturn offer\n\t}\n\n\tblog.V(3).Infof(\"getNextOffer offer(%s:%s) don't exist\", o.offerId, o.hostname)\n\thead := p.offerList.Front()\n\tfor {\n\t\tif head == nil {\n\t\t\tbreak\n\t\t}\n\n\t\tinnerOffer := head.Value.(*innerOffer)\n\t\tblog.V(3).Infof(\"getNextOffer id %d offer(%s:%s) isValid %t\",\n\t\t\tinnerOffer.id, innerOffer.offerId, innerOffer.hostname, innerOffer.isValid)\n\n\t\tif innerOffer.isValid {\n\t\t\toffer = &Offer{\n\t\t\t\telement: head,\n\t\t\t\tOffer: innerOffer.offer,\n\t\t\t\tId: innerOffer.id,\n\t\t\t\tofferId: innerOffer.offerId,\n\t\t\t\thostname: innerOffer.hostname,\n\t\t\t\tDeltaCPU: innerOffer.deltaCPU,\n\t\t\t\tDeltaMem: innerOffer.deltaMem,\n\t\t\t\tDeltaDisk: innerOffer.deltaDisk,\n\t\t\t}\n\n\t\t\tblog.V(3).Infof(\"GetNextOffer offer(%d:%s)\", offer.Id, offer.hostname)\n\t\t\treturn offer\n\n\t\t}\n\n\t\thead = head.Next()\n\t}\n\n\tblog.V(3).Infof(\"GetNextOffer offer(%s:%s) don't have next offer\", o.offerId, o.hostname)\n\n\treturn nil\n}", "func (p *Playlist) Next() {\n\tp.ch <- \"next\"\n}", "func (c *CompletionManager) Next() {\n\tif c.verticalScroll+int(c.max)-1 == c.selected {\n\t\tc.verticalScroll++\n\t}\n\tc.selected++\n\tc.update()\n}", "func (prev_block Block) Next(data string) Block {\n\t// TODO\n\tprevHash := prev_block.Hash\n\tgen := prev_block.Generation + 1\n\tdif := prev_block.Difficulty\n\tvar p uint64\n\tvar h []byte\n\n\tb := Block{prevHash, gen, dif, data, p, h}\n\treturn b\n}", "func (px *Paxos) proposer(seq int, val interface{}) {\n\n\tdecided, _ := px.Status(seq)\n\tfor !decided && !px.dead {\n\t\t//get a unique proposal number\n\t\tnum := px.getPropNum(seq)\n\t\tpx.mu.Lock()\n\t\tpx.mu.Unlock()\n\t\t//initialize response array\n\t\tresponses := make([]*PrepareReply, len(px.peers))\n\t\tfor i, _ := range responses {\n\t\t\tresponses[i] = &PrepareReply{false, -1, -1, nil, -1, -1}\n\t\t}\n\n\t\t//piggyback done info\n\t\tdone := -1\n\t\tpx.mu.Lock()\n\t\tif s, ok := px.done[px.me]; ok {\n\t\t\tdone = s\n\t\t}\n\t\tpx.mu.Unlock()\n\n\t\tvar pr PrepareReply\n\t\tprepArgs := &PrepareArgs{seq, num, px.me, done}\n\t\t//send proposals!\n\n\t\tfor i, p := range px.peers {\n\t\t\tif i == px.me { //send prepare to self\n\t\t\t\tif px.RcvPrepare(prepArgs, &pr) == nil {\n\t\t\t\t\tresponses[pr.From] = &pr\n\t\t\t\t\tpx.setDone(pr.From, pr.Done)\n\t\t\t\t}\n\t\t\t} else { //send prepares to peers\n\t\t\t\tif ok, r := px.sendPrepare(seq, p, num); ok {\n\t\t\t\t\tresponses[r.From] = r\n\t\t\t\t\tpx.setDone(r.From, r.Done)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\treject := 0\n\t\t//look at our lovely responses\n\t\tfor _, r := range responses {\n\t\t\tif !r.OK {\n\t\t\t\treject++\n\t\t\t\t//if n_p was too low, try again...\n\t\t\t\tif px.setHighestRumoredProp(seq, r.N_P) {\n\t\t\t\t\ttime.Sleep(time.Duration(rand.Int()%\n\t\t\t\t\t\t30) * time.Millisecond)\n\t\t\t\t\tpx.Start(seq, val)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\t//if not enough peers responded to prepares, start over.\n\t\tif reject >= px.majority {\n\t\t\ttime.Sleep(time.Duration(rand.Int()%30) * time.Millisecond)\n\t\t\tpx.Start(seq, val)\n\t\t\treturn\n\t\t}\n\t\t//val = v_a with highest n_a; choose original val otherwise\n\t\tval = px.chooseHighest(responses, val)\n\n\t\t//piggyback Done info\n\t\tdone = -1\n\t\tpx.mu.Lock()\n\t\tif s, ok := px.done[px.me]; ok {\n\t\t\tdone = s\n\t\t}\n\t\tpx.mu.Unlock()\n\n\t\taccept := 0\n\t\tvar ar AcceptReply\n\t\tacceptArgs := &AcceptArgs{seq, num, val, px.me, done}\n\t\t//send the accept to all peers\n\t\tfor i, p := range px.peers {\n\t\t\tif i == px.me { //send to self\n\t\t\t\tif px.RcvAccept(acceptArgs, &ar) == nil {\n\t\t\t\t\tif ar.OK {\n\t\t\t\t\t\taccept++\n\t\t\t\t\t\tpx.setDone(ar.From, ar.Done)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else { //send to peers\n\t\t\t\tif ok, r := px.sendAccept(seq, p, num, val); ok {\n\t\t\t\t\tpx.setDone(r.From, r.Done)\n\t\t\t\t\tif r.OK {\n\t\t\t\t\t\taccept++\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\t//if not enough accepts succeeded, start over!\n\t\tif accept < px.majority {\n\t\t\ttime.Sleep(time.Duration(rand.Int()%30) * time.Millisecond)\n\t\t\tpx.Start(seq, val)\n\t\t\treturn\n\t\t}\n\n\t\taccept = 0\n\t\t//send decided messages to all\n\t\tfor i, p := range px.peers {\n\t\t\tif i == px.me { //send to self\n\t\t\t\tdr := &DecideReply{}\n\t\t\t\tif px.RcvDecide(&DecideArgs{seq, val}, dr) == nil {\n\t\t\t\t\tif dr.OK {\n\t\t\t\t\t\taccept++\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else { //send to peers\n\t\t\t\tif ok, r := px.sendDecide(seq, p, val); ok {\n\t\t\t\t\tif r.OK {\n\t\t\t\t\t\taccept++\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tif accept < px.majority {\n\t\t\ttime.Sleep(time.Duration(rand.Int()%30) * time.Millisecond)\n\t\t\tpx.Start(seq, val)\n\t\t\treturn\n\t\t}\n\t\tdecided, _ = px.Status(seq)\n\t}\n\treturn\n}", "func (px *Paxos) Start(seq int, v interface{}) {\n\t// Your code here.\n\tpx.mu.Lock()\n\tdefer px.mu.Unlock()\n\tif seq < px.minSeq {\n\t\treturn\n\t}\n\tif _, ok := px.proposerInfo[seq]; ok {\n\t\treturn\n\t}\n\tif seq > px.maxSeq {\n\t\tpx.maxSeq = seq\n\t}\n\tdecided := px.isDecided(seq)\n\tvar maxProposalId ProposalId\n\tif !decided {\n\t\tmaxProposalId = px.maxProposalId[seq]\n\t\tmaxProposalId.S = maxProposalId.S + 1\n\t\tmaxProposalId.Addr = px.peers[px.me]\n\t\tpx.maxProposalId[seq] = maxProposalId\n\t\tvar proposerInfo ProposerInfo\n\t\tproposerInfo.proposalValue = v\n\t\tproposerInfo.prepareResp = make(map[string]PrepareResp)\n\t\tproposerInfo.acceptResp = make(map[string]bool)\n\t\tproposerInfo.decidedResp = make(map[string]bool)\n\t\tpx.proposerInfo[seq] = proposerInfo\n\t}\n\n\tif !decided {\n\t\tgo func() {\n\t\t\tpx.mu.Lock()\n\t\tExit:\n\t\t\tfor {\n\t\t\t\tvar rejected bool\n\t\t\t\tvar ok bool\n\t\t\t\tif seq < px.minSeq {\n\t\t\t\t\tbreak Exit\n\t\t\t\t}\n\t\t\t\tif !px.isPrepareMajority(seq) {\n\t\t\t\t\tfor _, peer := range px.peers {\n\t\t\t\t\t\tif _, ok = px.proposerInfo[seq].prepareResp[peer]; !ok {\n\t\t\t\t\t\t\targs := &PrepareArgs{}\n\t\t\t\t\t\t\targs.ProposalId = maxProposalId\n\t\t\t\t\t\t\targs.Seq = seq\n\t\t\t\t\t\t\tvar reply PrepareResp\n\t\t\t\t\t\t\tif peer != px.peers[px.me] {\n\t\t\t\t\t\t\t\tpx.mu.Unlock()\n\t\t\t\t\t\t\t\tok = call(peer, \"Paxos.Prepare\", args, &reply)\n\t\t\t\t\t\t\t\tpx.mu.Lock()\n\t\t\t\t\t\t\t\tif seq < px.minSeq {\n\t\t\t\t\t\t\t\t\tbreak Exit\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\tok = true\n\t\t\t\t\t\t\t\tpx.prepareUnlock(args, &reply)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tif ok {\n\t\t\t\t\t\t\t\trejected = reply.Rejected\n\t\t\t\t\t\t\t\tif rejected {\n\t\t\t\t\t\t\t\t\tproposerInfo := px.proposerInfo[seq]\n\t\t\t\t\t\t\t\t\tproposerInfo.decided = reply.Decided\n\t\t\t\t\t\t\t\t\t//px.proposerInfo[seq].decided = reply.Decided\n\t\t\t\t\t\t\t\t\tif reply.Decided {\n\t\t\t\t\t\t\t\t\t\tproposerInfo.decidedValue = reply.DecidedValue\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\tpx.proposerInfo[seq] = proposerInfo\n\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t\tpx.proposerInfo[seq].prepareResp[peer] = reply\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tif !rejected && px.isPrepareMajority(seq) && !px.isAcceptMajority(seq) {\n\t\t\t\t\tproposerInfo := px.proposerInfo[seq]\n\t\t\t\t\tproposerInfo.decidedValue = px.decidedValue(seq)\n\t\t\t\t\tpx.proposerInfo[seq] = proposerInfo\n\t\t\t\t\tfor _, peer := range px.peers {\n\t\t\t\t\t\tif _, ok = px.proposerInfo[seq].acceptResp[peer]; !ok {\n\t\t\t\t\t\t\targs := &AcceptArgs{}\n\t\t\t\t\t\t\targs.Seq = seq\n\t\t\t\t\t\t\targs.ProposalId = maxProposalId\n\t\t\t\t\t\t\targs.DecidedValue = px.proposerInfo[seq].decidedValue\n\t\t\t\t\t\t\tvar reply AcceptResp\n\t\t\t\t\t\t\tif peer != px.peers[px.me] {\n\t\t\t\t\t\t\t\tpx.mu.Unlock()\n\t\t\t\t\t\t\t\tok = call(peer, \"Paxos.Accept\", args, &reply)\n\t\t\t\t\t\t\t\tpx.mu.Lock()\n\t\t\t\t\t\t\t\tif seq < px.minSeq {\n\t\t\t\t\t\t\t\t\tbreak Exit\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\tok = true\n\t\t\t\t\t\t\t\tpx.acceptUnlock(args, &reply)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tif ok {\n\t\t\t\t\t\t\t\trejected = reply.Rejected\n\t\t\t\t\t\t\t\tif rejected {\n\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t\tpx.proposerInfo[seq].acceptResp[peer] = true\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tif !rejected && px.isAcceptMajority(seq) && !px.isDecidedAll(seq) {\n\t\t\t\t\t//fmt.Println(\"acceptResp\", px.proposerInfo[seq])\n\t\t\t\t\tproposerInfo := px.proposerInfo[seq]\n\t\t\t\t\tproposerInfo.decided = true\n\t\t\t\t\tpx.proposerInfo[seq] = proposerInfo\n\t\t\t\t\tfor _, peer := range px.peers {\n\t\t\t\t\t\tif !px.isDecidedAll(seq) {\n\t\t\t\t\t\t\tif _, ok = px.proposerInfo[seq].decidedResp[peer]; !ok {\n\t\t\t\t\t\t\t\targs := &DecidedArgs{}\n\t\t\t\t\t\t\t\targs.Seq = seq\n\t\t\t\t\t\t\t\targs.DecidedValue = px.proposerInfo[seq].decidedValue\n\t\t\t\t\t\t\t\tvar reply DecidedResp\n\t\t\t\t\t\t\t\tif peer != px.peers[px.me] {\n\t\t\t\t\t\t\t\t\tpx.mu.Unlock()\n\t\t\t\t\t\t\t\t\tok = call(peer, \"Paxos.Decided\", args, &reply)\n\t\t\t\t\t\t\t\t\tpx.mu.Lock()\n\t\t\t\t\t\t\t\t\tif seq < px.minSeq {\n\t\t\t\t\t\t\t\t\t\tbreak Exit\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t\tok = true\n\t\t\t\t\t\t\t\t\tpx.decidedUnlock(args, &reply)\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\tif ok && seq >= px.minSeq {\n\t\t\t\t\t\t\t\t\tpx.proposerInfo[seq].decidedResp[peer] = true\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tdecidedAll := px.isDecidedAll(seq)\n\t\t\t\tif decidedAll || rejected {\n\t\t\t\t\tbreak Exit\n\t\t\t\t}\n\t\t\t}\n\t\t\tpx.mu.Unlock()\n\t\t}()\n\t}\n}", "func (it *ChannelPartnerRepricingConfigIterator) Next() (*channelpb.ChannelPartnerRepricingConfig, error) {\n\tvar item *channelpb.ChannelPartnerRepricingConfig\n\tif err := it.nextFunc(); err != nil {\n\t\treturn item, err\n\t}\n\titem = it.items[0]\n\tit.items = it.items[1:]\n\treturn item, nil\n}", "func (f *bookFeed) Next() <-chan *BookUpdate {\n\treturn f.c\n}", "func (_CraftingI *CraftingICallerSession) NextItem() (*big.Int, error) {\n\treturn _CraftingI.Contract.NextItem(&_CraftingI.CallOpts)\n}", "func PoolNextBlock(request []string) (uint32, error) {\n\tapikey, userid, err := splitApiKey(request[1])\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tmposClient := mpos.NewMposClient(nil, request[0], apikey, userid, userAgent)\n\tmposClient.SetDebug(debug)\n\tstatus, err := mposClient.GetPoolStatus()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn status.NextNetworkBlock, nil\n}", "func (b *Bootstrapper) sendGetAccepted() {\n\tvdrs := ids.NewShortSet(1)\n\tfor b.pendingSendAccepted.Len() > 0 && b.pendingReceiveAccepted.Len() < MaxOutstandingBootstrapRequests {\n\t\tvdr, _ := b.pendingSendAccepted.Pop()\n\t\t// Add the validator to the set to send the messages to\n\t\tvdrs.Add(vdr)\n\t\t// Add the validator to send pending receipt set\n\t\tb.pendingReceiveAccepted.Add(vdr)\n\t}\n\n\tif vdrs.Len() > 0 {\n\t\tb.Ctx.Log.Debug(\"sent %d more GetAccepted messages with %d more to send\",\n\t\t\tvdrs.Len(),\n\t\t\tb.pendingSendAccepted.Len(),\n\t\t)\n\t\tb.Sender.SendGetAccepted(vdrs, b.RequestID, b.acceptedFrontier)\n\t}\n}", "func (i *Invocation) Next(opts ...InvocationOption) {\n\tvar op InvocationOp\n\tfor _, opt := range opts {\n\t\top = opt(op)\n\t}\n\n\ti.setCallback(op.Func, op.Async)\n\ti.chain.Next(i)\n}", "func (c *Conn) Next(ch chan<- Stanza) {\n\tstanza, err := new(Stanza), (error)(nil)\n\tfor {\n\t\tif stanza.Name, stanza.Value, err = next(c.in); err != nil {\n\t\t\treturn\n\t\t}\n\n\t\tch <- *stanza\n\t}\n}", "func (s *Server) next(ctx context.Context, req *jsonrpc2.Request) (interface{}, error) {\n\tin := Filter{}\n\tif err := json.Unmarshal([]byte(*req.Params), &in); err != nil {\n\t\treturn nil, err\n\t}\n\treturn s.peer.Next(ctx, in)\n}", "func (it *Iterator) Next() (*Message, error) {\n\tm, err := it.puller.Next()\n\n\tif err == nil {\n\t\tm.it = it\n\t\treturn m, nil\n\t}\n\n\tselect {\n\t// If Stop has been called, we return Done regardless the value of err.\n\tcase <-it.closed:\n\t\treturn nil, Done\n\tdefault:\n\t\treturn nil, err\n\t}\n}", "func (i *BlockchainIterator) Next() *Block {\n\tvar block *Block\n\n\terr := i.db.View(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket([]byte(blocksBucket))\n\t\tdata := b.Get(i.nextHash)\n\t\tblock = DeserializeBlock(data)\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\tlog.Panicf(\"迭代时,无法取出 %x : %s\", i.nextHash, err)\n\t}\n\n\ti.nextHash = block.PrevBlockHash\n\n\treturn block\n}", "func (_CraftingI *CraftingISession) NextItem() (*big.Int, error) {\n\treturn _CraftingI.Contract.NextItem(&_CraftingI.CallOpts)\n}", "func (s *BasevhdlListener) EnterNext_statement(ctx *Next_statementContext) {}", "func (i *PolicyIterator) Next() bool {\n\tif !i.started {\n\t\ti.started = true\n\t\treturn true\n\t}\n\n\tpolicy, _, ingress, _, _ := i.items()\n\n\tif i.ruleIdx < len(ingress.Rules)-1 {\n\t\ti.ruleIdx += 1\n\t\treturn true\n\t}\n\n\tif i.peerIdx < len(ingress.Peers)-1 {\n\t\ti.peerIdx += 1\n\t\ti.ruleIdx = 0\n\t\treturn true\n\t}\n\n\tif i.ingressIdx < len(policy.Ingress)-1 {\n\t\ti.ingressIdx += 1\n\t\ti.ruleIdx = 0\n\t\ti.peerIdx = 0\n\t\treturn true\n\t}\n\n\tif i.targetIdx < len(policy.AppliedTo)-1 {\n\t\ti.targetIdx += 1\n\t\ti.ingressIdx = 0\n\t\ti.ruleIdx = 0\n\t\ti.peerIdx = 0\n\t\treturn true\n\t}\n\n\tif i.policyIdx < len(i.policies)-1 {\n\t\ti.policyIdx += 1\n\t\ti.targetIdx = 0\n\t\ti.ingressIdx = 0\n\t\ti.ruleIdx = 0\n\t\ti.peerIdx = 0\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (_PlasmaFramework *PlasmaFrameworkCaller) NextDepositBlock(opts *bind.CallOpts) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _PlasmaFramework.contract.Call(opts, out, \"nextDepositBlock\")\n\treturn *ret0, err\n}", "func (o *CallbackOperator) Next(ctx context.Context) coldata.Batch {\n\treturn o.NextCb(ctx)\n}", "func (c *consumer) Next(ctx context.Context) (val *value, err error) {\n\t// Prevent Next from being called if the consumer already has one outstanding\n\t// unacknowledged message.\n\tif c.outstanding {\n\t\treturn nil, errors.New(\"unacknowledged message outstanding\")\n\t}\n\n\tvar ao int\n\n\t// Repeat trying to get the next value while the topic is either empty or not\n\t// created yet. It may exist sometime in the future.\n\tfor {\n\t\tval, ao, err = c.store.GetNext(c.topic)\n\t\tif !errors.Is(err, errTopicEmpty) && !errors.Is(err, errTopicNotExist) {\n\t\t\tbreak\n\t\t}\n\n\t\tselect {\n\t\tcase <-c.eventChan:\n\t\tcase <-ctx.Done():\n\t\t\treturn nil, errRequestCancelled\n\t\t}\n\t}\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"getting next from store: %v\", err)\n\t}\n\n\tc.ackOffset = ao\n\tc.outstanding = true\n\n\treturn val, err\n}", "func (c *Client) Next() string {\n\treturn c.nodes.Next()\n}", "func (m *Manager) NextPendingSupported() bool {\n\tm.pendingLock.RLock()\n\tdefer m.pendingLock.RUnlock()\n\tif len(m.pending) == 0 {\n\t\treturn true\n\t}\n\n\treturn m.SupportedVersions().Supports(m.pending[0].ProtocolVersion)\n}", "func (_ConsortiumManagement *ConsortiumManagementCaller) Proposal(opts *bind.CallOpts) (struct {\n\tId *big.Int\n\tProposer common.Address\n\tPtype uint8\n\tMember common.Address\n\tCapacity uint8\n\tApproves []common.Address\n\tDisapproves []common.Address\n}, error) {\n\tvar out []interface{}\n\terr := _ConsortiumManagement.contract.Call(opts, &out, \"proposal\")\n\n\toutstruct := new(struct {\n\t\tId *big.Int\n\t\tProposer common.Address\n\t\tPtype uint8\n\t\tMember common.Address\n\t\tCapacity uint8\n\t\tApproves []common.Address\n\t\tDisapproves []common.Address\n\t})\n\n\toutstruct.Id = out[0].(*big.Int)\n\toutstruct.Proposer = out[1].(common.Address)\n\toutstruct.Ptype = out[2].(uint8)\n\toutstruct.Member = out[3].(common.Address)\n\toutstruct.Capacity = out[4].(uint8)\n\toutstruct.Approves = out[5].([]common.Address)\n\toutstruct.Disapproves = out[6].([]common.Address)\n\n\treturn *outstruct, err\n\n}", "func (p PlayerIndex) Next(state State) PlayerIndex {\n\tif p == AdminPlayerIndex || p == ObserverPlayerIndex {\n\t\treturn p\n\t}\n\tp++\n\tif int(p) >= len(state.PlayerStates()) {\n\t\tp = 0\n\t}\n\treturn p\n}", "func (e *ReloadOptRuleBlacklistExec) Next(context.Context, *chunk.Chunk) error {\n\tinternalCtx := kv.WithInternalSourceType(context.Background(), kv.InternalTxnPrivilege)\n\treturn LoadOptRuleBlacklist(internalCtx, e.Ctx())\n}", "func (r *newSelectResult) Next() (NewPartialResult, error) {\n\tre := <-r.results\n\tif re.err != nil {\n\t\treturn nil, errors.Trace(re.err)\n\t}\n\tif re.result == nil {\n\t\treturn nil, nil\n\t}\n\tpr := &newPartialResult{}\n\tpr.rowLen = r.rowLen\n\terr := pr.unmarshal(re.result)\n\treturn pr, errors.Trace(err)\n}", "func (_PlasmaFramework *PlasmaFrameworkSession) NextDepositBlock() (*big.Int, error) {\n\treturn _PlasmaFramework.Contract.NextDepositBlock(&_PlasmaFramework.CallOpts)\n}", "func (b *Balancer) Next() *nsq.Producer { return b.pickNext().client }", "func (c *Conn) Next(ctx context.Context) (i Item, err error) {\n\tvar tx pgx.Tx\n\ttx, err = c.db.Begin(ctx)\n\tif err != nil {\n\t\treturn\n\t}\n\n\trow := tx.QueryRow(ctx, \"SELECT url, priority FROM jobs ORDER BY priority DESC FOR UPDATE skip locked\")\n\n\tif err = row.Scan(&i.URL, &i.Priority); err != nil {\n\t\tif err == pgx.ErrNoRows {\n\t\t\terr = ErrNoItem\n\t\t}\n\n\t\treturn\n\t}\n\n\ti.tx = *&tx\n\n\treturn\n}", "func (r Response) Next() Command {\n\treturn Command{\n\t\tJID: r.IQ.From,\n\t\tSID: r.SID,\n\t\tNode: r.Node,\n\t\tAction: \"next\",\n\t}\n}", "func (consensus *Consensus) requestBlockProposalMessageHandler(remoteMessage *node.RemoteMessage) ([]byte, bool, error) {\n\treplyMsg, err := NewRequestBlockProposalReply(nil)\n\tif err != nil {\n\t\treturn nil, false, err\n\t}\n\n\treplyBuf, err := consensus.localNode.SerializeMessage(replyMsg, false)\n\tif err != nil {\n\t\treturn nil, false, err\n\t}\n\n\tmsgBody := &pb.RequestBlockProposal{}\n\terr = proto.Unmarshal(remoteMessage.Message, msgBody)\n\tif err != nil {\n\t\treturn replyBuf, false, err\n\t}\n\n\tblockHash, err := common.Uint256ParseFromBytes(msgBody.BlockHash)\n\tif err != nil {\n\t\treturn replyBuf, false, err\n\t}\n\n\tblock, err := consensus.getBlockProposal(blockHash)\n\tif err != nil {\n\t\treturn replyBuf, false, err\n\t}\n\n\treplyMsg, err = NewRequestBlockProposalReply(block)\n\tif err != nil {\n\t\treturn replyBuf, false, err\n\t}\n\n\treplyBuf, err = consensus.localNode.SerializeMessage(replyMsg, false)\n\treturn replyBuf, false, err\n}", "func (c *Concurrent) NextNoBlock() ([]interface{}, error) {\n\tselect {\n\tcase value, ok := <-c.values:\n\t\tif ok {\n\t\t\treturn value, nil\n\t\t}\n\t\treturn nil, ErrVChanClosed\n\tdefault:\n\t\treturn nil, ErrNotPrep\n\t}\n}", "func (cs *ConsensusState) addProposalBlock(proposalBlock *tmtypes.QbftBlock) (err error) {\n\tblock := &ttypes.QbftBlock{QbftBlock: proposalBlock}\n\theight, round := block.Header.Height, block.Header.Round\n\tqbftlog.Debug(fmt.Sprintf(\"Consensus receive proposal block. Current: %v/%v/%v\", cs.Height, cs.Round, cs.Step),\n\t\t\"block(H/R/hash)\", fmt.Sprintf(\"%v/%v/%X\", height, round, block.Hash()))\n\n\t// Blocks might be reused, so round mismatch is OK\n\tif cs.Height != height {\n\t\tqbftlog.Debug(\"Received block from wrong height\", \"height\", height, \"round\", round)\n\t\treturn nil\n\t}\n\n\tif cs.begCons.IsZero() {\n\t\tcs.begCons = time.Now()\n\t}\n\n\tif block.Header.Sequence != cs.state.Sequence || block.Header.LastSequence != cs.state.LastSequence {\n\t\tqbftlog.Error(\"invalid proposalBlock sequence\", \"proposalBlock-seq\", block.Header.Sequence,\n\t\t\t\"state-seq\", cs.state.Sequence, \"proposalBlock-lastSeq\", block.Header.LastSequence,\n\t\t\t\"state-lastSeq\", cs.state.LastSequence)\n\t\treturn nil\n\t}\n\tif block.Header.Sequence > 0 && !bytes.Equal(block.Header.ProposerAddr, cs.Validators.GetProposer().Address) {\n\t\tqbftlog.Error(\"invalid proposalBlock proposer\",\n\t\t\t\"proposalBlock-proposer\", fmt.Sprintf(\"%X\", ttypes.Fingerprint(block.Header.ProposerAddr)),\n\t\t\t\"state-proposer\", fmt.Sprintf(\"%X\", ttypes.Fingerprint(cs.Validators.GetProposer().Address)))\n\t\treturn nil\n\t}\n\n\t// We're not expecting a block\n\tif !block.HashesTo(cs.ProposalBlockHash) {\n\t\t// NOTE: this can happen when we've gone to a higher round and\n\t\t// then receive parts from the previous round - not necessarily a bad peer.\n\t\tqbftlog.Debug(\"Received block when we're not expecting any\", \"ProposalBlockHash\", fmt.Sprintf(\"%X\", cs.ProposalBlockHash),\n\t\t\t\"height\", height, \"round\", round, \"hash\", fmt.Sprintf(\"%X\", block.Hash()))\n\t\treturn nil\n\t}\n\n\t// Already have expected proposal block\n\tif block.HashesTo(cs.ProposalBlock.Hash()) {\n\t\tqbftlog.Debug(\"addProposalBlock: already has proposal block\")\n\t\treturn nil\n\t}\n\n\tcs.ProposalBlock = block\n\n\t// NOTE: it's possible to receive proposal block for future rounds without having the proposal\n\tqbftlog.Info(fmt.Sprintf(\"Consensus set proposal block. Current: %v/%v/%v\", cs.Height, cs.Round, cs.Step),\n\t\t\"ProposalBlockHash\", fmt.Sprintf(\"%X\", cs.ProposalBlockHash), \"cost\", types.Since(cs.begCons))\n\n\tif DetachExec() {\n\t\tqbftlog.Info(\"write proposal block in advance\")\n\t\tgo cs.client.CommitBlock(cs.ProposalBlock.Data.Clone())\n\t}\n\n\t// Update Valid* if we can.\n\tprevotes := cs.Votes.Prevotes(cs.Round)\n\tblockID, hasTwoThirds := prevotes.TwoThirdsMajority()\n\tif hasTwoThirds && len(blockID.Hash) == 0 && (cs.ValidRound < cs.Round) {\n\t\tif cs.ProposalBlock.HashesTo(blockID.Hash) {\n\t\t\tqbftlog.Info(\"Updating valid block to new proposal block\",\n\t\t\t\t\"valid-round\", cs.Round, \"valid-block-hash\", cs.ProposalBlock.Hash())\n\t\t\tcs.ValidRound = cs.Round\n\t\t\tcs.ValidBlock = cs.ProposalBlock\n\t\t}\n\t\t// TODO: In case there is +2/3 majority in Prevotes set for some\n\t\t// block and cs.ProposalBlock contains different block, either\n\t\t// proposer is faulty or voting power of faulty processes is more\n\t\t// than 1/3. We should trigger in the future accountability\n\t\t// procedure at this point.\n\t}\n\n\tif cs.Step <= ttypes.RoundStepPropose && cs.isProposalComplete() {\n\t\t// Move onto the next step\n\t\tcs.enterPrevote(cs.Height, cs.Round)\n\t\tif hasTwoThirds { // this is optimisation as this will be triggered when prevote is added\n\t\t\tcs.enterPrecommit(cs.Height, cs.Round)\n\t\t}\n\t} else if cs.Step == ttypes.RoundStepCommit {\n\t\t// If we're waiting on the proposal block...\n\t\tcs.tryFinalizeCommit(cs.Height)\n\t}\n\treturn nil\n}", "func (n *NetIter) Next() bool {\n\treturn n.ips.Next()\n}", "func (i *Iterator) Next() *block.Block {\n\tvar currBlock *block.Block\n\n\terr := i.db.View(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket([]byte(blocksBucket))\n\n\t\tencodeBlock := b.Get(i.currentHash)\n\t\tcurrBlock = block.DeserializeBlock(encodeBlock)\n\n\t\treturn nil\n\t})\n\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\n\ti.currentHash = currBlock.PrevBlockHash\n\n\treturn currBlock\n}", "func (t Session) Next() *remotecommand.TerminalSize {\n\tselect {\n\tcase size := <-t.SizeChan:\n\t\treturn &size\n\tcase <-t.DoneChan:\n\t\treturn nil\n\t}\n}", "func (board *Board)NextProcedure()(nextProcedure Procedure){\n if board.BoardHeadId == 0 {\n Db.First(board)\n }\n currentProcedure := Procedure{Id: board.ProcedureId}\n Db.First(&currentProcedure)\n var existCount int\n if currentProcedure.Board {\n Db.Table(\"board_records\").Where(\"board_id = ? AND procedure_id = ?\", board.Id, currentProcedure.Id).Count(&existCount)\n if existCount == 0 {\n return currentProcedure\n }\n } else {\n // plasmids or prechecks or qualities\n procedureTable := currentProcedure.RecordName\n // samples or reactions\n recordTable := currentProcedure.FlowType + \"s\"\n Db.Table(procedureTable).Joins(\"INNER JOIN \" + currentProcedure.FlowType + \"s ON \" + procedureTable + \".\" + currentProcedure.FlowType + \"_id = \" + recordTable + \".id\").Where(recordTable + \".board_id = ?\", board.Id).Count(&existCount)\n var recordCount int\n Db.Table(recordTable).Where(\"board_id = ?\", board.Id).Count(&recordCount)\n if existCount < recordCount {\n return currentProcedure\n }\n }\n nextFlow := Flow{}\n Db.Where(\"board_head_id = ? AND flows.id > (SELECT id FROM flows WHERE board_head_id = ? AND procedure_id = ?)\", board.BoardHeadId, board.BoardHeadId, board.ProcedureId).Order(\"flows.id\").First(&nextFlow)\n if nextFlow.ProcedureId > 0 {\n nextProcedure.Id = nextFlow.ProcedureId\n Db.Model(board).UpdateColumn(\"procedure_id\", nextFlow.ProcedureId)\n Db.First(&nextProcedure)\n } else {\n Db.Exec(\"UPDATE boards SET status = 'finish' WHERE id = ?\", board.Id)\n //Db.Model(board).UpdateColumns(Board{ProcedureId: 0, Status: \"finish\"})\n }\n return nextProcedure\n}", "func (c *Context) Next() {\n\tc.index++\n\tfor c.index < int8(len(c.handlers)) {\n\t\tif c.IsAborted() {\n\t\t\tbreak\n\t\t}\n\t\t(c.handlers)[c.index](c)\n\t\tc.index++\n\t}\n}", "func (l *List) Choose(ctx context.Context, req *transport.Request) (peer peer.Peer, onFinish func(error), err error) {\n\treturn l.list.Choose(ctx, req)\n}", "func (it *EntitlementIterator) Next() (*channelpb.Entitlement, error) {\n\tvar item *channelpb.Entitlement\n\tif err := it.nextFunc(); err != nil {\n\t\treturn item, err\n\t}\n\titem = it.items[0]\n\tit.items = it.items[1:]\n\treturn item, nil\n}", "func (c *RDContext) Next(data []byte, endStream bool) error {\n\tfilters := c.Stream.dataReceivers\n\n\t// Finished processing the filters - call connection\n\tif c.currentFilter >= len(filters) {\n\t\treturn c.Stream.Connection.SendData(c.Stream, data, endStream)\n\t}\n\n\tnext := filters[c.currentFilter]\n\tc.currentFilter++\n\tc.Stream.setMiddlewareName(next.Name())\n\treturn next.ReceiveData(c, data, endStream)\n}", "func (c *cConn) Next(ctx context.Context, status sbalance.ConnStatus) error {\n\t// Ingoring ctx\n\tif c.conn == nil {\n\t\treturn syscall.EINVAL\n\t}\n\treturn c.get(C.enum_sbalance_conn_status(status))\n}", "func (cs *ConsensusState) enterPropose(height int64, round int) {\n\tif cs.Height != height || round < cs.Round || (cs.Round == round && ttypes.RoundStepPropose <= cs.Step) {\n\t\tqbftlog.Debug(fmt.Sprintf(\"enterPropose(%v/%v): Invalid args. Current step: %v/%v/%v\", height, round, cs.Height, cs.Round, cs.Step))\n\t\treturn\n\t}\n\tqbftlog.Info(fmt.Sprintf(\"enterPropose(%v/%v). Current: %v/%v/%v\", height, round, cs.Height, cs.Round, cs.Step))\n\tif cs.Round == 0 && cs.begCons.IsZero() {\n\t\tcs.begCons = time.Now()\n\t}\n\n\tdefer func() {\n\t\t// Done enterPropose:\n\t\tcs.updateRoundStep(round, ttypes.RoundStepPropose)\n\t\tcs.newStep()\n\n\t\t// If we have the whole proposal + POL, then goto Prevote now.\n\t\t// else, we'll enterPrevote when the rest of the proposal is received (in AddProposalBlockPart),\n\t\t// or else after timeoutPropose\n\t\tif cs.isProposalComplete() {\n\t\t\tcs.enterPrevote(height, cs.Round)\n\t\t}\n\t}()\n\n\t// If we don't get the proposal and all block parts quick enough, enterPrevote\n\tcs.scheduleTimeout(cs.Propose(round), height, round, ttypes.RoundStepPropose)\n\n\t// Nothing more to do if we're not a validator\n\tif cs.privValidator == nil {\n\t\tqbftlog.Debug(\"This node is not a validator\")\n\t\treturn\n\t}\n\n\t// if not a validator, we're done\n\tif !cs.Validators.HasAddress(cs.privValidator.GetAddress()) {\n\t\tqbftlog.Debug(\"This node is not a validator\",\n\t\t\t\"privValidator\", fmt.Sprintf(\"%X\", ttypes.Fingerprint(cs.privValidator.GetAddress())),\n\t\t\t\"Validators\", cs.Validators.String())\n\t\treturn\n\t}\n\tqbftlog.Debug(\"This node is a validator\")\n\n\tif cs.isProposer() {\n\t\tqbftlog.Info(\"enterPropose: Our turn to propose\",\n\t\t\t\"proposer\", fmt.Sprintf(\"%X\", ttypes.Fingerprint(cs.Validators.GetProposer().Address)),\n\t\t\t\"privValidator\", fmt.Sprintf(\"%X\", ttypes.Fingerprint(cs.privValidator.GetAddress())))\n\t\tcs.decideProposal(height, round)\n\t} else {\n\t\tqbftlog.Info(\"enterPropose: Not our turn to propose\",\n\t\t\t\"proposer\", fmt.Sprintf(\"%X\", ttypes.Fingerprint(cs.Validators.GetProposer().Address)),\n\t\t\t\"privValidator\", fmt.Sprintf(\"%X\", ttypes.Fingerprint(cs.privValidator.GetAddress())))\n\t}\n}", "func (e *BsnEndorsementHandler) Handle(requestContext *RequestContext, clientContext *ClientContext) {\n\t//GatewayLog.Logs( \"BSNEndorsementHandler Handle 开始交易提案\",)\n\tif len(requestContext.Opts.Targets) == 0 {\n\t\trequestContext.Error = status.New(status.ClientStatus, status.NoPeersFound.ToInt32(), \"targets were not provided\", nil)\n\t\treturn\n\t}\n\n\t// Endorse Tx\n\tvar TxnHeaderOpts []fab.TxnHeaderOpt\n\tif e.headerOptsProvider != nil {\n\t\tTxnHeaderOpts = e.headerOptsProvider()\n\t}\n\t//GatewayLog.Logs( \"createAndSendTransactionProposal 开始发送交易提案\",)\n\n\ttransactionProposalResponses, proposal, err := createAndSendBsnTransactionProposal(\n\t\tclientContext.Transactor,\n\t\t&requestContext.Request,\n\t\tpeer.PeersToTxnProcessors(requestContext.Opts.Targets),\n\t\tTxnHeaderOpts...,\n\t)\n\t//GatewayLog.Logs( \"Query createAndSendTransactionProposal END\",)\n\trequestContext.Response.Proposal = proposal\n\trequestContext.Response.TransactionID = proposal.TxnID // TODO: still needed?\n\n\tif err != nil {\n\t\trequestContext.Error = err\n\t\treturn\n\t}\n\n\trequestContext.Response.Responses = transactionProposalResponses\n\tif len(transactionProposalResponses) > 0 {\n\t\trequestContext.Response.Payload = transactionProposalResponses[0].ProposalResponse.GetResponse().Payload\n\t\trequestContext.Response.ChaincodeStatus = transactionProposalResponses[0].ChaincodeStatus\n\t}\n\t//GatewayLog.Logs( \"Query EndorsementHandler Handle END\",)\n\t//Delegate to next step if any\n\tif e.next != nil {\n\t\te.next.Handle(requestContext, clientContext)\n\t}\n}", "func (_ConsortiumManagement *ConsortiumManagementCallerSession) Proposal() (struct {\n\tId *big.Int\n\tProposer common.Address\n\tPtype uint8\n\tMember common.Address\n\tCapacity uint8\n\tApproves []common.Address\n\tDisapproves []common.Address\n}, error) {\n\treturn _ConsortiumManagement.Contract.Proposal(&_ConsortiumManagement.CallOpts)\n}", "func (p *parser) next() item {\n\tif p.peekCount > 0 {\n\t\tp.peekCount--\n\t} else {\n\t\tt := p.lex.nextItem()\n\t\t// Skip comments.\n\t\tfor t.typ == itemComment {\n\t\t\tt = p.lex.nextItem()\n\t\t}\n\t\tp.token[0] = t\n\t}\n\tif p.token[p.peekCount].typ == itemError {\n\t\tp.errorf(\"%s\", p.token[p.peekCount].val)\n\t}\n\treturn p.token[p.peekCount]\n}", "func (c *remoteCursor) Next() ([]byte, []byte, error) {\n\treturn c.next()\n}", "func (px *Paxos) Start(seq int, v interface{}) {\n\t// run Paxos algorithm in a new thread(run the Paxos protocol concurrently)\n\t// play the role of proposer\n\n\t// Your code here.\n\n\tpx.clog(DBG_PROPOSER, \"Start\", \"Start seq=%d v=%v\", seq, v)\n\n\t// I'm Proposer\n\tgo func() {\n\t\tn := 0\n\t\tmax_reject_pnum := -1\n\t\tfor {\n\t\t\tif px.dead {\n\t\t\t\t// I'm dead\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tif px.Lslots[seq].Decided {\n\t\t\t\t// locally decided, wouldn't send prepare and accept anymore\n\t\t\t\t// just propagate the decision\n\t\t\t\tpx.send_decided(seq, px.Lslots[seq].V)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tif px.APp[seq]+1 > n {\n\t\t\t\tn = px.APp[seq] + 1\n\t\t\t} else {\n\t\t\t\tn++\n\t\t\t}\n\n\t\t\tif n < max_reject_pnum {\n\t\t\t\tn = max_reject_pnum + 1\n\t\t\t}\n\n\t\t\tpx.clog(DBG_PROPOSER, \"Start\", \"send prepare, seq=%d n=%d\", seq, n)\n\n\t\t\tprepare_ok, p := px.send_prepare(seq, n)\n\t\t\tif !prepare_ok {\n\t\t\t\tmax_reject_pnum = p.PNum\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tnew_p := Proposal{}\n\n\t\t\t// no proposal yet, use v\n\t\t\tif p.PNum == 0 {\n\t\t\t\tnew_p.Value = v\n\t\t\t} else {\n\t\t\t\tnew_p.Value = p.Value\n\t\t\t}\n\n\t\t\tnew_p.PNum = n\n\n\t\t\tpx.clog(DBG_PROPOSER, \"Start\", \"prepare OK, proposal=%v\", new_p)\n\n\t\t\taccept_ok := px.send_accept(seq, new_p)\n\t\t\tif !accept_ok {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tpx.clog(DBG_PROPOSER, \"Start\", \"accept OK\")\n\n\t\t\tpx.send_decided(seq, new_p.Value)\n\n\t\t\tpx.clog(DBG_PROPOSER, \"Start\", \"decided\")\n\t\t\tbreak\n\t\t}\n\t}()\n}", "func (u *Updater) Next() (bool, dht.NodeID, SeekRequest) {\n\tvar ln int\n\tlinks := len(u.network.ID()) * 8\n\n\tln = u.queueLen()\n\n\t// By lazy populating the queue, as responses come back, that can be used in\n\t// later requests.\n\tfor ; ln == 0; ln = u.queueLen() {\n\t\tif u.idx >= links {\n\t\t\treturn false, nil, SeekRequest{}\n\t\t}\n\t\tif u.idx > u.depth {\n\t\t\tu.idx = links - 1\n\t\t}\n\t\tu.queueIdx(u.idx)\n\t\tu.idx++\n\t}\n\tid, sr := u.seekRequest(u.queue[ln-1])\n\tu.Lock()\n\tu.queue = u.queue[:ln-1]\n\tu.Unlock()\n\treturn true, id, sr\n}", "func (_PlasmaFramework *PlasmaFrameworkCallerSession) NextDepositBlock() (*big.Int, error) {\n\treturn _PlasmaFramework.Contract.NextDepositBlock(&_PlasmaFramework.CallOpts)\n}", "func (px *Paxos) Accept(seq int, num int, v interface{}) (prepareNum int, success bool) {\n\tsuccess = true\n\tmajority := len(px.peers)/2 + 1\n\tprepareNum = num\n\tvar acceptedAcceptNum int\n\targs := &AcceptArgs{}\n\targs.Seq = seq\n\targs.Num = num\n\targs.Value = v\n\treplyChan := make(chan *AcceptReply)\n\n\tacceptFunc := func(i int) {\n\t\tvar reply AcceptReply\n\t\tif i == int(px.me) {\n\t\t\tpx.AcceptHandle(args, &reply)\n\t\t\treplyChan <- &reply\n\t\t\treturn\n\t\t}\n\t\tif ok := call(px.peers[i], \"Paxos.AcceptHandle\", args, &reply); ok {\n\t\t\treplyChan <- &reply\n\t\t} else {\n\t\t\treplyChan <- nil\n\t\t}\n\n\t}\n\tfor i := range px.peers {\n\t\tgo acceptFunc(i)\n\t}\n\n\tfor _ = range px.peers {\n\t\treply := <-replyChan\n\t\tif reply != nil {\n\t\t\tacceptedAcceptNum++\n\t\t\tif prepareNum < reply.PrepareNumber {\n\t\t\t\tprepareNum = reply.PrepareNumber\n\t\t\t\tsuccess = false\n\t\t\t}\n\t\t}\n\t}\n\n\t// log.Printf(\"prepare phase finish num: %d/%d, success: %v\\n\", acceptedAcceptNum, len(px.peers), success)\n\tsuccess = success && (acceptedAcceptNum >= majority)\n\treturn\n\n}", "func (t TerminalSession) Next() *remotecommand.TerminalSize {\n\tselect {\n\tcase size := <-t.sizeChan:\n\t\treturn &size\n\t}\n}", "func (b *mpgBuff) next() *concBuff {\n\t// request fill\n\tgo func(buff *concBuff) {\n\t\tb.fill(buff)\n\t}(b.buffer())\n\tb.cur = !b.cur\n\treturn b.buffer()\n}", "func (it *ChannelPartnerLinkIterator) Next() (*channelpb.ChannelPartnerLink, error) {\n\tvar item *channelpb.ChannelPartnerLink\n\tif err := it.nextFunc(); err != nil {\n\t\treturn item, err\n\t}\n\titem = it.items[0]\n\tit.items = it.items[1:]\n\treturn item, nil\n}", "func (kademlia *Kademlia) AskNextNode(target *KademliaID, destination *Contact, findData bool, returnChannel chan interface{}) {\n\tif findData {\n\t\tgo kademlia.network.SendFindDataMessage(target.String(), destination, returnChannel)\n\t} else {\n\t\tgo kademlia.network.SendFindContactMessage(target, destination, returnChannel)\n\t}\n}" ]
[ "0.59082824", "0.5903719", "0.5894827", "0.5835964", "0.56806195", "0.5579939", "0.5564853", "0.55631655", "0.5560681", "0.5522159", "0.5475068", "0.545712", "0.53569907", "0.5316378", "0.530344", "0.52969617", "0.5296468", "0.5285935", "0.5249892", "0.52495056", "0.52451444", "0.5232498", "0.5201004", "0.5198574", "0.51984066", "0.51982933", "0.51872337", "0.5177819", "0.51755774", "0.5164345", "0.5159591", "0.51453507", "0.5135544", "0.5125086", "0.5093949", "0.5082911", "0.5080744", "0.5080038", "0.50744146", "0.5070844", "0.5070695", "0.5064886", "0.5056417", "0.50402284", "0.5023416", "0.5007337", "0.49988985", "0.4984346", "0.4963346", "0.49577466", "0.49555668", "0.4953336", "0.4935227", "0.4934601", "0.49167198", "0.49147248", "0.49145418", "0.49136305", "0.49095836", "0.49051282", "0.49041152", "0.4903162", "0.49026686", "0.48993823", "0.48962715", "0.48929396", "0.4890934", "0.48850447", "0.4883366", "0.48670787", "0.4863908", "0.48609492", "0.48568505", "0.48552275", "0.48548335", "0.485301", "0.4851286", "0.48406774", "0.48388618", "0.48348832", "0.48331967", "0.48319358", "0.4831251", "0.48307896", "0.48242688", "0.48090908", "0.4808299", "0.47981405", "0.47861096", "0.47783118", "0.47756302", "0.47664762", "0.47637257", "0.47633046", "0.4762898", "0.4761374", "0.4742251", "0.47380382", "0.47374782", "0.47367465" ]
0.6375579
0
Create some middleware which swaps out the existing request context with new context.Context value containing the connection pool.
func injectDBGorm(db *gorm.DB) gin.HandlerFunc { return func(c *gin.Context) { c.Set("dbConnection", db) c.Next() } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *AppServer) new(w http.ResponseWriter, r *http.Request, params *AppParams, handlers []Middleware) *Context {\n\t// adjust request id\n\trequestId := r.Header.Get(s.requestId)\n\tif requestId == \"\" {\n\t\trequestId = NewObjectId().Hex()\n\n\t\t// inject request header with new request id\n\t\tr.Header.Set(s.requestId, requestId)\n\t}\n\tw.Header().Set(s.requestId, requestId)\n\n\tctx := s.pool.Get().(*Context)\n\tctx.Request = r\n\tctx.Response = &ctx.writer\n\tctx.Params = params\n\tctx.Logger = s.logger.New(requestId)\n\tctx.settings = nil\n\tctx.frozenSettings = nil\n\tctx.writer.reset(w)\n\tctx.handlers = handlers\n\tctx.index = -1\n\tctx.startedAt = time.Now()\n\tctx.downAfter = ctx.startedAt.Add(s.slowdown)\n\n\treturn ctx\n}", "func Middleware(db Datastore) func(http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tctx := context.WithValue(r.Context(), dbCtxKey, db)\n\n\t\t\t// and call the next with our new context\n\t\t\tr = r.WithContext(ctx)\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func WithMiddleware(c *router.Context, next router.Handler) {\n\tdatabase, err := getDatabaseConnection(c.Context)\n\tif err != nil {\n\t\tlogging.Errorf(c.Context, \"Failed to retrieve a database connection: %s\", err.Error())\n\t\tc.Writer.Header().Set(\"Content-Type\", \"text/plain\")\n\t\tc.Writer.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\tc.Context = With(c.Context, database)\n\tnext(c)\n}", "func NewMiddlewareContext(writer http.ResponseWriter, request *http.Request) *MiddlewareContext {\n\tcontext := &MiddlewareContext{Request: request, ResponceWriter: writer}\n\n\treturn context\n}", "func requestMiddleware(next *handler.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {\n\t\tctx := context.WithValue(req.Context(), \"request\", req)\n\n\t\tnext.ContextHandler(ctx, w, req)\n\t})\n}", "func Bind(req *http.Request) *http.Request {\n // Reuse context if already binded\n if _, ok := req.Context().Value(Key).(Store); ok {\n return req\n }\n // Create new context store\n ctx := context.WithValue(req.Context(), Key, Store{})\n return req.WithContext(ctx)\n}", "func Middleware(bnetAuth *Auth) func(http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\n\t\t\t// return new context with previous context values and added value\n\t\t\tctx := context.WithValue(r.Context(), bnetAuthCtxKey, bnetAuth)\n\n\t\t\t// and call the next with our new context\n\t\t\tr = r.WithContext(ctx)\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func (s *AppServer) reuse(ctx *Context) {\n\ts.pool.Put(ctx)\n}", "func Middleware(fn AppHandler, c interface{}) AppHandler {\n\treturn func(w http.ResponseWriter, r *http.Request) *AppError {\n\t\tr = r.WithContext(context.WithValue(r.Context(), \"env\", c))\n\t\tr = r.WithContext(context.WithValue(r.Context(), \"vars\", mux.Vars(r)))\n\t\treturn fn(w, r)\n\t}\n}", "func RequestContextMiddleware(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tctx, id := BuildContext(r)\n\t\tr = r.WithContext(ctx)\n\t\tw.Header().Set(UUIDHeaderKey, id)\n\n\t\tnext.ServeHTTP(w, r)\n\t})\n}", "func (m *Macross) ReleaseContext(c *Context) {\n\tc.Response.Header.SetServer(\"Macross\")\n\tm.pool.Put(c)\n}", "func Middleware(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tremoteIP := AddressForRequest(r)\n\t\tnext.ServeHTTP(w, r.Clone(context.WithValue(r.Context(), contextKey, remoteIP)))\n\t})\n}", "func adaptContextHandlerFunc(fn HandlerFunc) Handler {\n\treturn func(ctx *Context) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tfn(ctx, w, r)\n\t\t})\n\t}\n}", "func Bootstrap(in *ContextIn) *ContextOut {\n\n\t// transform middleware slice\n\tmiddlewares := make([]mux.MiddlewareFunc, len(in.MiddlewaresToRegister))\n\tfor i, middleware := range in.MiddlewaresToRegister {\n\t\tmiddlewares[i] = middleware\n\t}\n\n\tout := &ContextOut{}\n\tout.Server = &server{\n\t\tport: in.Port,\n\t\troutes: in.RoutesToRegister,\n\t\tmiddlewares: middlewares,\n\t}\n\n\treturn out\n}", "func CtxMiddleware(ctx context.Context, r *http.Request) context.Context {\n\treturn appengine.WithContext(ctx, r)\n}", "func Mongo(dbsession *mgo.Session, contextKey string) func(next http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\tfn := func(w http.ResponseWriter, r *http.Request) {\n\t\t\tsCopy := dbsession.Copy()\n\t\t\tdefer sCopy.Close()\n\n\t\t\tvar ctx context.Context\n\t\t\tvar newReq *http.Request\n\n\t\t\tif contextKey == \"\" {\n\t\t\t\tctx = context.WithValue(r.Context(), \"db\", sCopy)\n\t\t\t\tnewReq = r.WithContext(ctx)\n\t\t\t} else {\n\t\t\t\tctx = context.WithValue(r.Context(), contextKey, sCopy)\n\t\t\t\tnewReq = r.WithContext(ctx)\n\t\t\t}\n\t\t\t// Process request\n\t\t\tnext.ServeHTTP(w, newReq)\n\t\t}\n\t\treturn http.HandlerFunc(fn)\n\t}\n}", "func (cp *ContextPool) Get(rw ResponseWriter, req *Request) (c *Context) {\n\tselect {\n\tcase c = <-cp.c:\n\t// reuse existing context\n\tdefault:\n\t\t// create new context\n\t\tc = NewContext(rw, req)\n\t}\n\treturn\n}", "func UsePool(ctx context.Context, pool *redis.Pool) context.Context {\n\treturn context.WithValue(ctx, &contextKey, pool)\n}", "func Adapt(fn func(http.Handler) http.Handler) Middleware {\n\treturn func(ctx *Context, h http.Handler) http.Handler {\n\t\treturn fn(h)\n\t}\n}", "func (req MinRequest) Context(ctx context.Context) MinRequest {\n\treq.impl = req.impl.Context(ctx)\n\n\treturn req\n}", "func (ds *DatabaseSession) AttachMiddleware() Middleware {\n\treturn func(next http.HandlerFunc) http.HandlerFunc {\n\t\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\t\tconnection := session.session.Clone()\n\n\t\t\tctx := context.WithValue(r.Context(), sessionKey, connection)\n\n\t\t\tgo func() {\n\t\t\t\tselect {\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\tconnection.Close()\n\t\t\t\t}\n\t\t\t}()\n\n\t\t\tnext.ServeHTTP(w, r.WithContext(ctx))\n\t\t}\n\t}\n}", "func (tunnel *TunnelHandler) Use(middleware ...Middleware) {\n\ttunnel.Ctx.Use(middleware...)\n}", "func WrapMiddleware(m func(http.Handler) http.Handler) MiddlewareFunc {\n\treturn func(c Context, next HandlerFunc) (err error) {\n\t\tm(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tc.SetRequest(r)\n\t\t\terr = next(c)\n\t\t})).ServeHTTP(c.Response(), c.Request())\n\t\treturn\n\t}\n}", "func Middleware(next http.Handler) http.Handler {\n\tfn := func(w http.ResponseWriter, r *http.Request) {\n\t\tappengineCtx := appengine.NewContext(r)\n\t\tctx := context.WithValue(r.Context(), contextKeyContext, appengineCtx)\n\t\tnext.ServeHTTP(w, r.WithContext(ctx))\n\t}\n\n\treturn http.HandlerFunc(fn)\n}", "func BindContext(hndl http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {\n\t\tprint(\"Binding context\\n\")\n\t\tctx := OpenCtx(req)\n\t\tprint(\"BindContext: \", ctx, \"\\n\")\n\n\t\tdefer closeCtx(req)\n\t\thndl.ServeHTTP(w, req)\n\t})\n}", "func GraphqlContextMiddleware(next echo.HandlerFunc) echo.HandlerFunc {\n\treturn func(c echo.Context) error {\n\t\tctx := graph.AddHttpResponseToContext(c.Request().Context(), c)\n\t\tc.SetRequest(c.Request().WithContext(ctx))\n\t\treturn next(c)\n\t}\n}", "func connContext(ctx context.Context, c net.Conn) context.Context {\n\treturn ctx\n}", "func newContext(r *http.Request) context.Context {\n\treturn context.Background()\n}", "func DBMiddlware(db *gorm.DB) func(c *gin.Context) {\n\treturn func(c *gin.Context) {\n\t\tc.Set(\"db\", db)\n\t\tc.Next()\n\t}\n\n}", "func (c *fakeRedisConn) SetContext(v interface{}) {}", "func Context(h http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t// user service\n\t\tu := user.New(cognitoID, cognitoSecret)\n\t\tcontext.Set(r, \"userService\", u)\n\n\t\t// session helper\n\t\ts := session.New()\n\t\tcontext.Set(r, \"session\", s)\n\n\t\tvar netTransport = &http.Transport{\n\t\t\tDial: (&net.Dialer{\n\t\t\t\tTimeout: 5 * time.Second,\n\t\t\t}).Dial,\n\t\t\tTLSHandshakeTimeout: 5 * time.Second,\n\t\t}\n\n\t\t// support timeout and net transport.\n\t\tc := &http.Client{\n\t\t\tTimeout: time.Second * 10,\n\t\t\tTransport: netTransport,\n\t\t}\n\n\t\t// http client\n\t\tcontext.Set(r, \"client\", c)\n\n\t\tp := post.New(dynamoTablePosts, dynamoEndpoint, nil)\n\t\tcontext.Set(r, \"postService\", p)\n\n\t\tl := like.New(dynamoTableLikes, dynamoEndpoint, nil)\n\t\tcontext.Set(r, \"likeService\", l)\n\n\t\th.ServeHTTP(w, r)\n\t})\n}", "func GinContextToHttpContext(c *gin.Context) {\n\tctx := context.WithValue(c.Request.Context(), GinContextKey, c)\n\tc.Request = c.Request.WithContext(ctx)\n\tc.Next()\n}", "func Middleware(config Config) func(http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tctx := context.WithValue(r.Context(), KeyEnvironment, config.Environment)\n\t\t\tctx = context.WithValue(ctx, KeyDatabase, config.DB)\n\t\t\tctx = context.WithValue(ctx, KeyGithubAppConfig, config.GithubAppConfig)\n\n\t\t\tnext.ServeHTTP(w, r.WithContext(ctx))\n\t\t})\n\t}\n}", "func Middleware(store *Store) func(h http.Handler) http.Handler {\n\treturn func(h http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(\n\t\t\tw http.ResponseWriter,\n\t\t\tr *http.Request,\n\t\t) {\n\t\t\tctx := r.Context()\n\t\t\tctx = context.WithValue(ctx, contextKey, store)\n\t\t\th.ServeHTTP(w, r.WithContext(ctx))\n\t\t})\n\t}\n}", "func AddToContext(ctx context.Context, reqID string) context.Context {\n\treturn context.WithValue(ctx, reqIDKey, reqID)\n}", "func (aps *ApiServer) setMiddleware() {\n\t/*\n\t\taps.Engine.Use(func(c *gin.Context) {\n\t\t\tstart := time.Now()\n\t\t\tc.Next()\n\t\t\tend := time.Now()\n\t\t\tlatency := end.Sub(start)\n\t\t\tpath := c.Request.URL.Path\n\t\t\tclientIP := c.ClientIP()\n\t\t\tmethod := c.Request.Method\n\t\t\tstatusCode := c.Writer.Status()\n\t\t\tlogger.Info(\"api request\",\n\t\t\t\tzap.Int(\"status_code\", statusCode),\n\t\t\t\tzap.Duration(\"latency\", latency),\n\t\t\t\tzap.String(\"client_ip\", clientIP),\n\t\t\t\tzap.String(\"method\", method),\n\t\t\t\tzap.String(\"path\", path),\n\t\t\t)\n\t\t})\n\t*/\n\taps.Engine.Use(gin.Recovery())\n}", "func WrapMiddleware(m func(http.Handler) http.Handler) MiddlewareFunc {\n\treturn func(next HandlerFunc) HandlerFunc {\n\t\treturn func(c Context) (err error) {\n\t\t\tm(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\t\tc.SetRequest(r)\n\t\t\t\tc.SetResponse(newResponse(w))\n\t\t\t\terr = next(c)\n\t\t\t})).ServeHTTP(c.Response(), c.Request())\n\t\t\treturn\n\t\t}\n\t}\n}", "func AddToRequest(app *App) server.PreHandlerFunc {\n\treturn func(req *http.Request) *http.Request {\n\t\tnewCtx := With(req.Context(), app)\n\t\treturn req.Clone(newCtx)\n\t}\n}", "func newRequestContext(ctx context.Context, req *http.Request) context.Context {\n\treturn context.WithValue(ctx, requestKey, req)\n}", "func prepContext(r *http.Request) *http.Request {\n\tskip := getSkipFile(r)\n\tres := r.WithContext(\n\t\tcontext.WithValue(r.Context(), toSkip, skip+1),\n\t)\n\n\treturn res\n}", "func CtxMiddleware(con *gorm.DB, redis *redis.Client) gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tc.Set(\"db_mysql\", con)\n\t\tc.Set(\"redis\", redis)\t\t\n\t\tc.Next()\n\t}\n}", "func wrapHandler(h func(http.ResponseWriter, *http.Request)) httprouter.Handle {\n\treturn func(w http.ResponseWriter, r *http.Request, ps httprouter.Params) {\n\t\thttpcontext.Set(r, \"params\", ps)\n\t\th(w, r)\n\t}\n}", "func (p *AlertingProxy) createProxyContext(ctx *contextmodel.ReqContext, request *http.Request, response *response.NormalResponse) *contextmodel.ReqContext {\n\tcpy := *ctx\n\tcpyMCtx := *cpy.Context\n\tcpyMCtx.Resp = web.NewResponseWriter(ctx.Req.Method, &safeMacaronWrapper{response})\n\tcpy.Context = &cpyMCtx\n\tcpy.Req = request\n\n\t// If RBAC is enabled, the actions are checked upstream and if the user gets here then it is allowed to do an action against a datasource.\n\t// Some data sources require legacy Editor role in order to perform mutating operations. In this case, we elevate permissions for the context that we\n\t// will provide downstream.\n\t// TODO (yuri) remove this after RBAC for plugins is implemented\n\tif !ctx.SignedInUser.HasRole(org.RoleEditor) {\n\t\tnewUser := *ctx.SignedInUser\n\t\tnewUser.OrgRole = org.RoleEditor\n\t\tcpy.SignedInUser = &newUser\n\t}\n\treturn &cpy\n}", "func Contexter() func(next http.Handler) http.Handler {\n\trnd := templates.HTMLRenderer()\n\tcsrfOpts := CsrfOptions{\n\t\tSecret: setting.SecretKey,\n\t\tCookie: setting.CSRFCookieName,\n\t\tSetCookie: true,\n\t\tSecure: setting.SessionConfig.Secure,\n\t\tCookieHTTPOnly: setting.CSRFCookieHTTPOnly,\n\t\tHeader: \"X-Csrf-Token\",\n\t\tCookieDomain: setting.SessionConfig.Domain,\n\t\tCookiePath: setting.SessionConfig.CookiePath,\n\t\tSameSite: setting.SessionConfig.SameSite,\n\t}\n\tif !setting.IsProd {\n\t\tCsrfTokenRegenerationInterval = 5 * time.Second // in dev, re-generate the tokens more aggressively for debug purpose\n\t}\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(resp http.ResponseWriter, req *http.Request) {\n\t\t\tctx := Context{\n\t\t\t\tResp: NewResponse(resp),\n\t\t\t\tCache: mc.GetCache(),\n\t\t\t\tLocale: middleware.Locale(resp, req),\n\t\t\t\tLink: setting.AppSubURL + strings.TrimSuffix(req.URL.EscapedPath(), \"/\"),\n\t\t\t\tRender: rnd,\n\t\t\t\tSession: session.GetSession(req),\n\t\t\t\tRepo: &Repository{\n\t\t\t\t\tPullRequest: &PullRequest{},\n\t\t\t\t},\n\t\t\t\tOrg: &Organization{},\n\t\t\t\tData: middleware.GetContextData(req.Context()),\n\t\t\t}\n\t\t\tdefer ctx.Close()\n\n\t\t\tctx.Data.MergeFrom(middleware.CommonTemplateContextData())\n\t\t\tctx.Data[\"Context\"] = &ctx\n\t\t\tctx.Data[\"CurrentURL\"] = setting.AppSubURL + req.URL.RequestURI()\n\t\t\tctx.Data[\"Link\"] = ctx.Link\n\t\t\tctx.Data[\"locale\"] = ctx.Locale\n\n\t\t\t// PageData is passed by reference, and it will be rendered to `window.config.pageData` in `head.tmpl` for JavaScript modules\n\t\t\tctx.PageData = map[string]any{}\n\t\t\tctx.Data[\"PageData\"] = ctx.PageData\n\n\t\t\tctx.Req = WithContext(req, &ctx)\n\t\t\tctx.Csrf = PrepareCSRFProtector(csrfOpts, &ctx)\n\n\t\t\t// Get the last flash message from cookie\n\t\t\tlastFlashCookie := middleware.GetSiteCookie(ctx.Req, CookieNameFlash)\n\t\t\tif vals, _ := url.ParseQuery(lastFlashCookie); len(vals) > 0 {\n\t\t\t\t// store last Flash message into the template data, to render it\n\t\t\t\tctx.Data[\"Flash\"] = &middleware.Flash{\n\t\t\t\t\tDataStore: &ctx,\n\t\t\t\t\tValues: vals,\n\t\t\t\t\tErrorMsg: vals.Get(\"error\"),\n\t\t\t\t\tSuccessMsg: vals.Get(\"success\"),\n\t\t\t\t\tInfoMsg: vals.Get(\"info\"),\n\t\t\t\t\tWarningMsg: vals.Get(\"warning\"),\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// prepare an empty Flash message for current request\n\t\t\tctx.Flash = &middleware.Flash{DataStore: &ctx, Values: url.Values{}}\n\t\t\tctx.Resp.Before(func(resp ResponseWriter) {\n\t\t\t\tif val := ctx.Flash.Encode(); val != \"\" {\n\t\t\t\t\tmiddleware.SetSiteCookie(ctx.Resp, CookieNameFlash, val, 0)\n\t\t\t\t} else if lastFlashCookie != \"\" {\n\t\t\t\t\tmiddleware.SetSiteCookie(ctx.Resp, CookieNameFlash, \"\", -1)\n\t\t\t\t}\n\t\t\t})\n\n\t\t\t// If request sends files, parse them here otherwise the Query() can't be parsed and the CsrfToken will be invalid.\n\t\t\tif ctx.Req.Method == \"POST\" && strings.Contains(ctx.Req.Header.Get(\"Content-Type\"), \"multipart/form-data\") {\n\t\t\t\tif err := ctx.Req.ParseMultipartForm(setting.Attachment.MaxSize << 20); err != nil && !strings.Contains(err.Error(), \"EOF\") { // 32MB max size\n\t\t\t\t\tctx.ServerError(\"ParseMultipartForm\", err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\n\t\t\thttpcache.SetCacheControlInHeader(ctx.Resp.Header(), 0, \"no-transform\")\n\t\t\tctx.Resp.Header().Set(`X-Frame-Options`, setting.CORSConfig.XFrameOptions)\n\n\t\t\tctx.Data[\"CsrfToken\"] = ctx.Csrf.GetToken()\n\t\t\tctx.Data[\"CsrfTokenHtml\"] = template.HTML(`<input type=\"hidden\" name=\"_csrf\" value=\"` + ctx.Data[\"CsrfToken\"].(string) + `\">`)\n\n\t\t\t// FIXME: do we really always need these setting? There should be someway to have to avoid having to always set these\n\t\t\tctx.Data[\"DisableMigrations\"] = setting.Repository.DisableMigrations\n\t\t\tctx.Data[\"DisableStars\"] = setting.Repository.DisableStars\n\t\t\tctx.Data[\"EnableActions\"] = setting.Actions.Enabled\n\n\t\t\tctx.Data[\"ManifestData\"] = setting.ManifestData\n\n\t\t\tctx.Data[\"UnitWikiGlobalDisabled\"] = unit.TypeWiki.UnitGlobalDisabled()\n\t\t\tctx.Data[\"UnitIssuesGlobalDisabled\"] = unit.TypeIssues.UnitGlobalDisabled()\n\t\t\tctx.Data[\"UnitPullsGlobalDisabled\"] = unit.TypePullRequests.UnitGlobalDisabled()\n\t\t\tctx.Data[\"UnitProjectsGlobalDisabled\"] = unit.TypeProjects.UnitGlobalDisabled()\n\t\t\tctx.Data[\"UnitActionsGlobalDisabled\"] = unit.TypeActions.UnitGlobalDisabled()\n\n\t\t\tctx.Data[\"AllLangs\"] = translation.AllLangs()\n\n\t\t\tnext.ServeHTTP(ctx.Resp, ctx.Req)\n\t\t})\n\t}\n}", "func AddToContext(name contextKey, value string) func(next http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tr = r.WithContext(context.WithValue(r.Context(), name, value))\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func Wrap(\n\tmiddleware Middleware,\n\thandler Handler,\n) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tmiddleware(handler)(context.Background(), w, r)\n\t}\n}", "func Middleware(next HandlerFunc) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tsession, _ := GetCurrentSession(r)\n\t\tnext(session, w, r)\n\t})\n}", "func Wrap(next http.Handler) httprouter.Handle {\n\treturn func(rw http.ResponseWriter, req *http.Request, params httprouter.Params) {\n\t\treq = req.WithContext(context.WithValue(req.Context(), httprouter.ParamsKey, params))\n\t\tnext.ServeHTTP(rw, req)\n\t}\n}", "func (req *MinRequest) Context(ctx context.Context) *MinRequest {\n\treq.impl = req.impl.Context(ctx)\n\n\treturn req\n}", "func newContext(w http.ResponseWriter, r *http.Request) *Context {\n\treturn &Context{\n\t\tw: w,\n\t\tr: r,\n\t\tdata: nil,\n\t}\n}", "func injectCtx(endpointHandler func(http.ResponseWriter, *http.Request, context.Context)) http.Handler {\n\treturn http.HandlerFunc(func(res http.ResponseWriter, req *http.Request) {\n\t\tendpointHandler(res, req, logs.WithRqId(httpContext, uuid.NewRandom().String()))\n\t})\n}", "func SaveConnInContext(ctx context.Context, c net.Conn) context.Context {\n\treturn context.WithValue(ctx, ConnContextKey, c)\n}", "func statementCtx(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tstID := chi.URLParam(r, \"stID\")\n\t\tvar stmt model.Statement\n\t\terr := driver.DoOne(&stmt, stID, driver.GetOne)\n\t\tif err != nil {\n\t\t\thttp.Error(w, http.StatusText(404), 404)\n\t\t\treturn\n\t\t}\n\t\tctx := context.WithValue(r.Context(), statementContext, &stmt)\n\t\tlog.Printf(\"Data from DB: %+v with ID: %v\", stmt, stID)\n\n\t\tnext.ServeHTTP(w, r.WithContext(ctx))\n\t})\n}", "func ToContext(r *http.Request, w http.ResponseWriter, s *Store) context.Context {\n\treturn context.WithValue(r.Context(), sessionContextKey, s)\n}", "func (r *Request) Context() context.Context", "func wrapHandler(h http.HandlerFunc) httprouter.Handle {\n\treturn func(w http.ResponseWriter, r *http.Request, ps httprouter.Params) {\n\t\tctx := r.Context()\n\t\tr = r.WithContext(ctx)\n\t\th.ServeHTTP(w, r)\n\t}\n}", "func withConnections(db *mgo.Session) adapter {\n\treturn func(h http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tdbsession := db.Copy()\n\t\t\tdefer dbsession.Close() // cleaning up\n\t\t\tdbName := os.Getenv(\"MONGO_DB_NAME\")\n\t\t\tif dbName == \"\" {\n\t\t\t\tdbName = \"sarbacanes_contacts_tests\"\n\t\t\t}\n\t\t\tctx := context.WithValue(r.Context(), lib.MongoDB, dbsession.DB(dbName))\n\t\t\tctx = context.WithValue(ctx, lib.MongoDBSession, dbsession)\n\t\t\th.ServeHTTP(w, r.WithContext(ctx))\n\t\t})\n\t}\n}", "func (g *Glue) newContext(w http.ResponseWriter, r *http.Request) *Context {\n\tctx := &Context{inj.New(), g, newResponseWriter(w)}\n\n\tctx.Register(r)\n\tctx.Register(ctx.rw)\n\t// register our ResponseWriter as an http.ResponseWriter as well for\n\t// net/http HandlerFunc compatibility\n\tctx.RegisterAs(ctx.rw, (*http.ResponseWriter)(nil))\n\t// register this instance with itself\n\tctx.Register(*ctx)\n\treturn ctx\n}", "func (r *Request) Pool(pool Pool) *Request {\n\tpanic(\"TODO\")\n\treturn r\n}", "func AddLoggerToContext() echo.MiddlewareFunc {\n\treturn func(next echo.HandlerFunc) echo.HandlerFunc {\n\t\treturn func(c echo.Context) error {\n\n\t\t\tlogger := log.WithFields(log.Fields{\n\t\t\t\t\"request_id\": c.Response().Header().Get(echo.HeaderXRequestID),\n\t\t\t})\n\t\t\tc.Set(\"logger\", logger)\n\t\t\treturn next(c)\n\t\t}\n\t}\n}", "func (dp *DataPlane) middlewareSetContext(k string, v interface{}) echo.MiddlewareFunc {\n\treturn func(next echo.HandlerFunc) echo.HandlerFunc {\n\t\treturn func(c echo.Context) error {\n\t\t\tc.Set(k, v)\n\t\t\treturn next(c)\n\t\t}\n\t}\n}", "func wrapContext(ctx context.Context, adapter Adapter) contextWrapper {\n\treturn contextWrapper{\n\t\tctx: context.WithValue(ctx, ctxKey, adapter),\n\t\tadapter: adapter,\n\t}\n}", "func genContext(r *http.Request) (context.Context, context.CancelFunc) {\n\twriteTimeout := r.Context().Value(http.ServerContextKey).(*http.Server).WriteTimeout\n\treturn context.WithTimeout(context.Background(), writeTimeout*80/100)\n}", "func NewRequestCtx(w http.ResponseWriter, r *http.Request) context.Context {\n\treturn context.WithValue(r.Context(), ctxName, newLambda(w, r, mux.Vars(r)))\n}", "func (c Conn) WithContext(cx context.Context) *Conn {\n\treturn &Conn{\n\t\tlocal: c.local,\n\t\tremote: c.remote,\n\t\tConn: connCtxOverride{c: cx, Conn: c.Conn},\n\t}\n}", "func (r *oauthProxy) proxyMiddleware(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {\n\t\tnext.ServeHTTP(w, req)\n\n\t\t// @step: retrieve the request scope\n\t\tscope := req.Context().Value(contextScopeName)\n\t\tif scope != nil {\n\t\t\tsc := scope.(*RequestScope)\n\t\t\tif sc.AccessDenied {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\t// @step: add the proxy forwarding headers\n\t\treq.Header.Add(\"X-Forwarded-For\", realIP(req))\n\t\treq.Header.Set(\"X-Forwarded-Host\", req.Host)\n\t\treq.Header.Set(\"X-Forwarded-Proto\", req.Header.Get(\"X-Forwarded-Proto\"))\n\n\t\t// @step: add any custom headers to the request\n\t\tfor k, v := range r.config.Headers {\n\t\t\treq.Header.Set(k, v)\n\t\t}\n\n\t\t// @note: by default goproxy only provides a forwarding proxy, thus all requests have to be absolute and we must update the host headers\n\t\treq.URL.Host = r.endpoint.Host\n\t\treq.URL.Scheme = r.endpoint.Scheme\n\t\tif v := req.Header.Get(\"Host\"); v != \"\" {\n\t\t\treq.Host = v\n\t\t\treq.Header.Del(\"Host\")\n\t\t} else if !r.config.PreserveHost {\n\t\t\treq.Host = r.endpoint.Host\n\t\t}\n\n\t\tif isUpgradedConnection(req) {\n\t\t\tr.log.Debug(\"upgrading the connnection\", zap.String(\"client_ip\", req.RemoteAddr))\n\t\t\tif err := tryUpdateConnection(req, w, r.endpoint); err != nil {\n\t\t\t\tr.log.Error(\"failed to upgrade connection\", zap.Error(err))\n\t\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\t\treturn\n\t\t\t}\n\t\t\treturn\n\t\t}\n\n\t\tr.upstream.ServeHTTP(w, req)\n\t})\n}", "func WithContext(response http.ResponseWriter, request *http.Request, ctx context.Context) (http.ResponseWriter, *http.Request) {\n\tif ca, ok := response.(ContextAware); ok {\n\t\tca.SetContext(ctx)\n\t\treturn response, request.WithContext(ctx)\n\t}\n\n\tif ctx == nil {\n\t\tctx = context.Background()\n\t}\n\n\treturn &contextAwareResponseWriter{response, ctx}, request.WithContext(ctx)\n}", "func adaptHandler(h http.Handler) Handler {\n\treturn func(ctx *Context) http.Handler {\n\t\treturn h\n\t}\n}", "func ContextDB(db *gorm.DB) gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tc.Set(\"db\", db)\n\t\tc.Next()\n\t}\n}", "func (gorilla Gorilla) Middleware(handler http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {\n\t\tdefer gorillaContext.Clear(req)\n\t\tctx := context.WithValue(req.Context(), reader, req)\n\t\thandler.ServeHTTP(w, req.WithContext(ctx))\n\t})\n}", "func WrapperMiddleware(w HandlerFuncWrapper) Middleware { return w }", "func (a *Router) Use(hs ...func(*Context) error) *Router {\n\ta.middlewares = append(a.middlewares, hs...)\n\treturn a\n}", "func (r *Request) WithContext(ctx context.Context) *Request", "func (e *engine) prepareContext(w http.ResponseWriter, r *http.Request) *Context {\n\tctx := acquireContext()\n\tctx.Req = ahttp.AcquireRequest(r)\n\tctx.Res = ahttp.AcquireResponseWriter(w)\n\tctx.reply = acquireReply()\n\tctx.subject = security.AcquireSubject()\n\treturn ctx\n}", "func NewContext(c ctx.Context, req *http.Request, resp http.ResponseWriter) Context {\n\twc := &webCtx{req: req, resp: resp, now: time.Now()}\n\twc.Context, wc.cancel = ctx.WithCancel(c)\n\twc.remoteIP = remoteIP(req)\n\treturn wc\n}", "func WithDB(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tdatabase, err := model.NewDatabase()\n\t\tif err != nil {\n\t\t\thttp.Error(w, http.StatusText(http.StatusInternalServerError),\n\t\t\t\thttp.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tdefer database.Close()\n\n\t\tctx := context.WithValue(r.Context(), BlogContext(\"database\"), database)\n\t\treq := r.WithContext(ctx)\n\n\t\tnext.ServeHTTP(w, req)\n\t})\n}", "func NewContext(w http.ResponseWriter, r *http.Request) (*Context, error) {\n var s *Session\n var u *UserSession\n var err error\n\n s, err = GetSession(w, r)\n if err != nil {\n return nil, err\n }\n\n u, err = GetUserSession(w, r)\n if err != nil {\n return nil, err\n }\n\n data := setup(w, r, u, s)\n\n return &Context{\n W: w,\n R: r,\n Session: s,\n UserSession: u,\n Data: data,\n }, nil\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\treturn handler\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\treturn handler\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\treturn handler\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\treturn handler\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\treturn handler\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\treturn handler\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\treturn handler\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\treturn handler\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\treturn handler\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\treturn handler\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\treturn handler\n}", "func setupGlobalMiddleware(handler http.Handler) http.Handler {\n\treturn handler\n}", "func Middleware(\n\tenv *Env,\n) func(http.Handler) http.Handler {\n\treturn func(h http.Handler) http.Handler {\n\t\treturn middleware{h, env}\n\t}\n}", "func remoteAddrToContext(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {\n\t\tnext.ServeHTTP(w, req.WithContext(context.WithValue(req.Context(), remoteAddrKey, req.RemoteAddr)))\n\t})\n}", "func withDB(s *mgo.Session, h http.Handler) http.Handler {\n\treturn &dbwrapper{dbSession: s, h: h}\n}", "func SetContext(response http.ResponseWriter, ctx context.Context) http.ResponseWriter {\n\tif ca, ok := response.(ContextAware); ok {\n\t\tca.SetContext(ctx)\n\t\treturn response\n\t}\n\n\tif ctx == nil {\n\t\tpanic(\"nil context\")\n\t}\n\n\treturn &contextAwareResponseWriter{response, ctx}\n}", "func SessionMiddleware(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\n\t\t_, err := r.Cookie(\"pinged\")\n\t\tif err != nil {\n\t\t\thttp.Error(w, http.StatusText(401), 401)\n\t\t}\n\n\t\t//TODO: look up session in redis\n\t\tctx1 := context.WithValue(r.Context(), UserID, \"11111111111\")\n\t\tctx2 := context.WithValue(ctx1, Role, \"user\")\n\n\t\tnext.ServeHTTP(w, r.WithContext(ctx2))\n\t})\n}", "func Middleware(conn *storage.Conn, key []byte) func(http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tauth := r.Header.Get(\"Authorization\")\n\n\t\t\t// Allow unauthenticated users in\n\t\t\tif auth == \"\" {\n\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tuserID, err := validateAndGetUserID(key, auth)\n\t\t\tif err != nil {\n\t\t\t\tlog.Error().Err(err).Msg(\"failed validate user for ID\")\n\t\t\t\thttp.Error(w, \"Invalid auth\", http.StatusForbidden)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// Get the user from the database\n\t\t\tadmin, err := getUserByID(r.Context(), conn, userID)\n\t\t\tif err != nil {\n\t\t\t\tlog.Error().Err(err).Msg(\"failed get user for ID\")\n\t\t\t\thttp.Error(w, \"Invalid auth\", http.StatusForbidden)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// Put it in context\n\t\t\tctx := context.WithValue(r.Context(), adminCtxKey, admin)\n\n\t\t\t// And call the next with our new context\n\t\t\tr = r.WithContext(ctx)\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func ToContext(ctx context.Context, reqID string) context.Context {\n\treturn context.WithValue(ctx, ctxKey, reqID)\n}", "func DoCtx(ctx context.Context, req *http.Request, resp interface{}) (*http.Response, error) {\n\tr := req.Clone(ctx)\n\n\treturn Do(r, resp)\n}", "func NewMiddleware(gcpProjectID string, logger *zap.Logger) func(next http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tt := newTraceFromTraceContext(gcpProjectID, r.Header.Get(\"X-Cloud-Trace-Context\"))\n\t\t\tzapFields := append(\n\t\t\t\tzapdriver.TraceContext(t.TraceID, t.SpanID, true, t.ProjectID),\n\t\t\t\tzap.String(\"ip\", r.Header.Get(\"X-Forwarded-For\")),\n\t\t\t)\n\t\t\tctx := ctxzap.ToContext(r.Context(), logger.With(zapFields...))\n\t\t\tnext.ServeHTTP(w, r.WithContext(ctx))\n\t\t})\n\t}\n}", "func RouteContext(next echo.HandlerFunc) echo.HandlerFunc {\n\treturn func(c echo.Context) error {\n\t\tctxParent, cancel := context.WithTimeout(context.Background(), timeoutSeconds*time.Second)\n\t\tdefer cancel()\n\n\t\tsession, err := db.Client().StartSession()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tvar ctx context.Context\n\n\t\tif err = mongo.WithSession(ctxParent, session, func(sctx mongo.SessionContext) error {\n\n\t\t\tusr, err := user.Get(sctx, c.Get(auth.UserIDParam).(string))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tusr, err = user.Inherit(sctx, usr)\n\n\t\t\tctx = context.WithValue(ctxParent, sess.ContextKey(\"user\"), usr)\n\n\t\t\treturn nil\n\t\t}); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = mongo.WithSession(ctx, session, func(sctx mongo.SessionContext) error {\n\t\t\tc.Set(\"ctx\", sctx)\n\n\t\t\treturn next(c)\n\t\t})\n\n\t\tsession.EndSession(ctx)\n\n\t\treturn err\n\t}\n}", "func addDatabaseMiddleware(r *gin.Engine) {\n\tdb := initDB()\n\t// Add database to our context\n\tr.Use(func(c *gin.Context) {\n\t\tc.Set(ContextKeyDB, db)\n\t})\n}", "func (s *server) middleware(n httprouter.Handle) httprouter.Handle {\n\treturn func(w http.ResponseWriter, r *http.Request, ps httprouter.Params) {\n\t\t// Log the basics\n\t\tlog.WithFields(logrus.Fields{\n\t\t\t\"method\": r.Method,\n\t\t\t\"remote-addr\": r.RemoteAddr,\n\t\t\t\"http-protocol\": r.Proto,\n\t\t\t\"headers\": r.Header,\n\t\t\t\"content-length\": r.ContentLength,\n\t\t}).Debugf(\"HTTP Request to %s\", r.URL)\n\n\t\tif r.ContentLength > 0 {\n\t\t\t// Dump payload into logs for visibility\n\t\t\tb, err := ioutil.ReadAll(r.Body)\n\t\t\tif err == nil {\n\t\t\t\tlog.Debugf(\"Dumping Payload for request to %s: %s\", r.URL, b)\n\t\t\t}\n\t\t}\n\n\t\t// Call registered handler\n\t\tn(w, r, ps)\n\t}\n}" ]
[ "0.6319336", "0.6146942", "0.5946056", "0.5754442", "0.575218", "0.5731107", "0.57245153", "0.5641714", "0.55994594", "0.5569156", "0.55382276", "0.55248874", "0.5497842", "0.5477662", "0.5432631", "0.54214084", "0.54184", "0.5410464", "0.53937906", "0.5390533", "0.53903556", "0.5388572", "0.5383331", "0.53752655", "0.53685105", "0.53476995", "0.5347237", "0.5345758", "0.53255105", "0.5313026", "0.5301872", "0.52777064", "0.5277673", "0.52557886", "0.5244464", "0.52314585", "0.5229418", "0.522004", "0.52125067", "0.5210611", "0.5200027", "0.5195665", "0.51681876", "0.5166352", "0.51658", "0.51593196", "0.5150683", "0.5149868", "0.5145594", "0.51448053", "0.51277095", "0.51198816", "0.51197135", "0.51171666", "0.51093906", "0.51078874", "0.51045877", "0.5094818", "0.50874853", "0.5061051", "0.50589865", "0.5050188", "0.50456643", "0.50393695", "0.50388765", "0.5027338", "0.50255376", "0.50195813", "0.5019138", "0.50158525", "0.5013387", "0.5013288", "0.5011728", "0.5008382", "0.4996059", "0.49925974", "0.49906653", "0.49903", "0.49903", "0.49903", "0.49903", "0.49903", "0.49903", "0.49903", "0.49903", "0.49903", "0.49903", "0.49903", "0.49903", "0.49895626", "0.49885246", "0.4982477", "0.49570987", "0.4956122", "0.49467078", "0.4935718", "0.49334678", "0.4930038", "0.49240598", "0.4918755", "0.49112374" ]
0.0
-1
User will inject the databaseTx in the `User` schema
func (gtx *GuardTx) User(user *schema.User) *schema.User { if user == nil { user = &schema.User{ Entity: schema.Entity{DBContract: gtx.dbTx}, } } else { user.DBContract = gtx.dbTx } user.SetValidator(gtx.validator.User) return user }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func User(user tables.Person) error {\n\tfmt.Println(\"preparing\")\n\tstmt, err := mysqlBus.DB.Prepare(\"INSERT INTO Person VALUES(?, ?, ?, ?, ?, ?)\")\n\tfmt.Println(\"prepared\")\n\tfmt.Println(err)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfmt.Println(\"executing\")\n\t_, err = stmt.Exec(user.Username, user.HashedPassword, user.Salt, user.Fname, user.Lname, user.ColorPalette)\n\tfmt.Println(\"executed\")\n\tfmt.Println(err)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfmt.Println(\"no error\")\n\tfmt.Println(err)\n\treturn nil\n}", "func (tx *Tx) User(userName []byte) (*User, error) {\n\tu := &User{\n\t\tTx: tx,\n\t\tName: userName,\n\t}\n\n\treturn u, u.Load()\n}", "func (r *Resolver) User() exec.UserResolver { return &userResolver{r} }", "func (create) User(ctx context.Context, db *sqlx.DB, n NewUser, now time.Time) (*User, error) {\n\tctx, span := global.Tracer(\"service\").Start(ctx, \"internal.data.create.user\")\n\tdefer span.End()\n\n\thash, err := bcrypt.GenerateFromPassword([]byte(n.Password), bcrypt.DefaultCost)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"generating password hash\")\n\t}\n\n\tu := User{\n\t\tID: uuid.New().String(),\n\t\tName: n.Name,\n\t\tEmail: n.Email,\n\t\tPasswordHash: hash,\n\t\tRoles: n.Roles,\n\t\tDateCreated: now.UTC(),\n\t\tDateUpdated: now.UTC(),\n\t}\n\n\tconst q = `INSERT INTO users\n\t\t(user_id, name, email, password_hash, roles, date_created, date_updated)\n\t\tVALUES ($1, $2, $3, $4, $5, $6, $7)`\n\t_, err = db.ExecContext(ctx, q, u.ID, u.Name, u.Email, u.PasswordHash, u.Roles, u.DateCreated, u.DateUpdated)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"inserting user\")\n\t}\n\n\treturn &u, nil\n}", "func init() {\n\tuserFields := schema.User{}.Fields()\n\t_ = userFields\n\t// userDescUsername is the schema descriptor for username field.\n\tuserDescUsername := userFields[1].Descriptor()\n\t// user.UsernameValidator is a validator for the \"username\" field. It is called by the builders before save.\n\tuser.UsernameValidator = userDescUsername.Validators[0].(func(string) error)\n\t// userDescPassword is the schema descriptor for password field.\n\tuserDescPassword := userFields[2].Descriptor()\n\t// user.PasswordValidator is a validator for the \"password\" field. It is called by the builders before save.\n\tuser.PasswordValidator = userDescPassword.Validators[0].(func(string) error)\n\t// userDescEmail is the schema descriptor for email field.\n\tuserDescEmail := userFields[3].Descriptor()\n\t// user.EmailValidator is a validator for the \"email\" field. It is called by the builders before save.\n\tuser.EmailValidator = userDescEmail.Validators[0].(func(string) error)\n\t// userDescCreatedAt is the schema descriptor for created_at field.\n\tuserDescCreatedAt := userFields[4].Descriptor()\n\t// user.DefaultCreatedAt holds the default value on creation for the created_at field.\n\tuser.DefaultCreatedAt = userDescCreatedAt.Default.(func() time.Time)\n\t// userDescUpdatedAt is the schema descriptor for updated_at field.\n\tuserDescUpdatedAt := userFields[5].Descriptor()\n\t// user.DefaultUpdatedAt holds the default value on creation for the updated_at field.\n\tuser.DefaultUpdatedAt = userDescUpdatedAt.Default.(func() time.Time)\n\t// user.UpdateDefaultUpdatedAt holds the default value on update for the updated_at field.\n\tuser.UpdateDefaultUpdatedAt = userDescUpdatedAt.UpdateDefault.(func() time.Time)\n\t// userDescID is the schema descriptor for id field.\n\tuserDescID := userFields[0].Descriptor()\n\t// user.DefaultID holds the default value on creation for the id field.\n\tuser.DefaultID = userDescID.Default.(func() uuid.UUID)\n}", "func (r *Resolver) User() generated.UserResolver { return &userResolver{r} }", "func (r *Resolver) User() generated.UserResolver { return &userResolver{r} }", "func TestCreateTablePutUser(t *testing.T) {\n\n\tdbsql, err := sql.Open(\"postgres\", \"user=postgres dbname=gorm password=simsim sslmode=disable\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tdb, err := InitDB(dbsql)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\terr = db.PutUser(12312)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n}", "func init() {\n\tuserFields := schema.User{}.Fields()\n\t_ = userFields\n\t// userDescLogin is the schema descriptor for login field.\n\tuserDescLogin := userFields[0].Descriptor()\n\t// user.LoginValidator is a validator for the \"login\" field. It is called by the builders before save.\n\tuser.LoginValidator = userDescLogin.Validators[0].(func(string) error)\n\t// userDescName is the schema descriptor for name field.\n\tuserDescName := userFields[1].Descriptor()\n\t// user.NameValidator is a validator for the \"name\" field. It is called by the builders before save.\n\tuser.NameValidator = userDescName.Validators[0].(func(string) error)\n\t// userDescPassword is the schema descriptor for password field.\n\tuserDescPassword := userFields[2].Descriptor()\n\t// user.PasswordValidator is a validator for the \"password\" field. It is called by the builders before save.\n\tuser.PasswordValidator = userDescPassword.Validators[0].(func(string) error)\n\t// userDescIsActive is the schema descriptor for isActive field.\n\tuserDescIsActive := userFields[3].Descriptor()\n\t// user.DefaultIsActive holds the default value on creation for the isActive field.\n\tuser.DefaultIsActive = userDescIsActive.Default.(bool)\n}", "func (r *Resolver) User() UserResolver { return &userResolver{r} }", "func TestTx_SaveUser(t *testing.T) {\n\tdb := NewTestDB()\n\tdefer db.Close()\n\n\tok(t, db.Update(func(tx *gist.Tx) error {\n\t\tok(t, tx.SaveUser(&gist.User{ID: 100, Username: \"john\", AccessToken: \"1234\"}))\n\t\treturn nil\n\t}))\n\n\tok(t, db.View(func(tx *gist.Tx) error {\n\t\tu, _ := tx.User(100)\n\t\tequals(t, &gist.User{ID: 100, Username: \"john\", AccessToken: \"1234\"}, u)\n\t\treturn nil\n\t}))\n}", "func (_UserCrud *UserCrudTransactor) InsertUser(opts *bind.TransactOpts, userAddress common.Address, userEmail string, userAge *big.Int) (*types.Transaction, error) {\n\treturn _UserCrud.contract.Transact(opts, \"insertUser\", userAddress, userEmail, userAge)\n}", "func Create(user User) error {\n\t\n}", "func (f *userFactory) Tx(tx *ent.Tx) *userFactory {\n\treturn f.Client(tx.Client())\n}", "func (m *Manager) Create(ctx context.Context, tx *sql.Tx, user v0.User) error {\n\t_, err := tx.ExecContext(ctx, `\n\t\t\t\tINSERT INTO users (\n\t\t\t\t\tname, \n\t\t\t\t\temail, \n\t\t\t\t\tprimary_public_key, \n\t\t\t\t\trecovery_public_key, \n\t\t\t\t\tsuper_user, \n\t\t\t\t\tauth_level, \n\t\t\t\t\tweight,\n\t\t\t\t\tuser_set\n\t\t\t\t\t) VALUES (?, ?, ?, ?, ?, ?, ?, ?)`,\n\t\tuser.Name,\n\t\tuser.Email,\n\t\tuser.PrimaryPublicKey,\n\t\tuser.RecoveryPublicKey,\n\t\tuser.SuperUser,\n\t\tuser.AuthLevel,\n\t\tuser.Weight,\n\t\tuser.Set,\n\t)\n\treturn err\n}", "func (ec *executionContext) _User(ctx context.Context, sel []query.Selection, obj *model.User) graphql.Marshaler {\n\tfields := graphql.CollectFields(ec.Doc, sel, userImplementors, ec.Variables)\n\n\tout := graphql.NewOrderedMap(len(fields))\n\tfor i, field := range fields {\n\t\tout.Keys[i] = field.Alias\n\n\t\tswitch field.Name {\n\t\tcase \"__typename\":\n\t\t\tout.Values[i] = graphql.MarshalString(\"User\")\n\t\tcase \"id\":\n\t\t\tout.Values[i] = ec._User_id(ctx, field, obj)\n\t\tcase \"username\":\n\t\t\tout.Values[i] = ec._User_username(ctx, field, obj)\n\t\tcase \"email\":\n\t\t\tout.Values[i] = ec._User_email(ctx, field, obj)\n\t\tcase \"displayName\":\n\t\t\tout.Values[i] = ec._User_displayName(ctx, field, obj)\n\t\tcase \"isDefault\":\n\t\t\tout.Values[i] = ec._User_isDefault(ctx, field, obj)\n\t\tcase \"profileImagePath\":\n\t\t\tout.Values[i] = ec._User_profileImagePath(ctx, field, obj)\n\t\tcase \"updatedAt\":\n\t\t\tout.Values[i] = ec._User_updatedAt(ctx, field, obj)\n\t\tcase \"blocked\":\n\t\t\tout.Values[i] = ec._User_blocked(ctx, field, obj)\n\t\tcase \"shadowbanned\":\n\t\t\tout.Values[i] = ec._User_shadowbanned(ctx, field, obj)\n\t\tcase \"possibleUninstall\":\n\t\t\tout.Values[i] = ec._User_possibleUninstall(ctx, field, obj)\n\t\tcase \"lastActiveAt\":\n\t\t\tout.Values[i] = ec._User_lastActiveAt(ctx, field, obj)\n\t\tcase \"joinedFromInvite\":\n\t\t\tout.Values[i] = ec._User_joinedFromInvite(ctx, field, obj)\n\t\tcase \"coinBalance\":\n\t\t\tout.Values[i] = ec._User_coinBalance(ctx, field, obj)\n\t\tcase \"temporaryCoinBalance\":\n\t\t\tout.Values[i] = ec._User_temporaryCoinBalance(ctx, field, obj)\n\t\tcase \"postCount\":\n\t\t\tout.Values[i] = ec._User_postCount(ctx, field, obj)\n\t\tcase \"engagement\":\n\t\t\tout.Values[i] = ec._User_engagement(ctx, field, obj)\n\t\tdefault:\n\t\t\tpanic(\"unknown field \" + strconv.Quote(field.Name))\n\t\t}\n\t}\n\n\treturn out\n}", "func (o *Transaction) User(exec boil.Executor, mods ...qm.QueryMod) userQuery {\n\tqueryMods := []qm.QueryMod{\n\t\tqm.Where(\"user_id=?\", o.UserID),\n\t}\n\n\tqueryMods = append(queryMods, mods...)\n\n\tquery := Users(exec, queryMods...)\n\tqueries.SetFrom(query.Query, \"`users`\")\n\n\treturn query\n}", "func (h *Handler) InsertUser(user models.User) (error){\n\n\thashedPassword, err := bcrypt.GenerateFromPassword([]byte(user.Password), 8)\n\n\tquery := fmt.Sprintf(\"insert into users (first_name, last_name, email, password) values ('%s', '%s', '%s', '%s');\", user.FirstName, user.LastName, user.Email, hashedPassword)\n\n\t_, err = h.DB.Exec(query)\n\tif err != nil {\n\t\tfmt.Printf(\"user_service-InsertUser-Exec: %s\\n\", err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func CreateUser(c *gin.Context) {}", "func init() {\n\tdatabaseModels = append(databaseModels, &models.User{})\n}", "func Example_transactions() {\n\tdb, _ := dbx.Open(\"mysql\", \"user:pass@/example\")\n\n\tdb.Transactional(func(tx *dbx.Tx) error {\n\t\t_, err := tx.Insert(\"user\", dbx.Params{\n\t\t\t\"name\": \"user1\",\n\t\t}).Execute()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t_, err = tx.Insert(\"user\", dbx.Params{\n\t\t\t\"name\": \"user2\",\n\t\t}).Execute()\n\t\treturn err\n\t})\n}", "func (user *User) BeforeCreate(transaction *gorm.DB) error {\n return nil\n}", "func (t *tx) AddUser(user *model.User) error {\n\t// FIXME: handle sql constraint errors\n\terr := t.Create(user).Error\n\n\treturn errors.Wrap(err, \"create user failed\")\n}", "func insertUser(db *sql.DB, u *User) {\n\n\t// name := u.name\n\t// rollno := u.rollno\n\tinsertUserSQL := `INSERT INTO User( name, rollno) VALUES (?, ?)`\n\tstatement, err := db.Prepare(insertUserSQL) \n\n\tif err != nil {\n\t\tlog.Fatalln(err.Error())\n\t}\n\t_, err = statement.Exec(u.name, u.rollno)\n\tif err != nil {\n\t\tlog.Fatalln(err.Error())\n\t}\n}", "func (us UserStorer) Save(schema *Schema, context *Context) (user User, userID aorm.ID, err error) {\n\tvar db = context.DB()\n\n\tif context.Auth.Config.UserModel != nil {\n\t\tcurrentUser := reflect.New(utils.ModelType(context.Auth.Config.UserModel)).Interface()\n\t\tcopier.Copy(currentUser, schema)\n\t\tif us.CreateFunc != nil {\n\t\t\terr = us.CreateFunc(context, currentUser)\n\t\t} else {\n\t\t\terr = db.Create(currentUser).Error\n\t\t}\n\t\treturn currentUser.(User), aorm.IdOf(currentUser), err\n\t}\n\treturn\n}", "func CreateUser(user model.User) {\n\tfmt.Println(user)\n}", "func Insert() error {\n\tuser := &TbUser{\n\t\tName: \"viney\",\n\t\tEmail: \"[email protected]\",\n\t\tCreated: time.Now().Format(\"2006-01-02 15:04:05\"),\n\t}\n\treturn orm.Save(user)\n}", "func registerUser() {\n\tgoes.Register(\n\t\t&User{},\n\t\tFirstNameUpdatedV1{},\n\t\tCreatedV1{},\n\t)\n}", "func init() {\n\torm.RegisterModel(new(User))\n}", "func (db *DataBase) Register(user *models.UserPrivateInfo) (userID int, err error) {\n\n\tvar (\n\t\ttx *sql.Tx\n\t)\n\n\tif tx, err = db.Db.Begin(); err != nil {\n\t\treturn\n\t}\n\tdefer tx.Rollback()\n\n\tif userID, err = db.createPlayer(tx, user); err != nil {\n\t\treturn\n\t}\n\n\tif err = db.createRecords(tx, userID); err != nil {\n\t\treturn\n\t}\n\n\terr = tx.Commit()\n\treturn\n}", "func insertUser(user User) {\n\tcollection := client.Database(\"Go_task\").Collection(\"users\")\n\tinsertResult, err := collection.InsertOne(context.TODO(), user)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tfmt.Println(\"Inserted user with ID:\", insertResult.InsertedID)\n}", "func (db Database) saveUser() error {\n\treturn makeError()\n}", "func SignUpUser(c *gin.Context) {\n\tvar db = models.InitDB()\n\tvar userData models.User\n\terr := c.Bind(&userData)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tfmt.Println(userData)\n\tif err := db.Create(&userData).Error; err != nil {\n\t\tc.JSON(200, gin.H{\n\t\t\t\"creation\": \"false\",\n\t\t})\n\t\treturn\n\t}\n\tc.JSON(200, gin.H{\n\t\t\"creation\": \"true\",\n\t})\n}", "func AddUser(user *e.User) error {\n\treturn db.DB.Transaction(func(tx *gorm.DB) error {\n\t\tif err := tx.Create(&user).Error; err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif user.Type == \"talent\" {\n\t\t\ttx.Create(&e.Talent{\n\t\t\t\tUserID: user.ID,\n\t\t\t\tIsVerified: false,\n\t\t\t})\n\t\t}\n\t\tif user.Type == \"company\" {\n\t\t\ttx.Create(&e.Company{\n\t\t\t\tUserID: user.ID,\n\t\t\t\tName: user.Name,\n\t\t\t\tIsUpdated: false,\n\t\t\t})\n\t\t}\n\t\treturn nil\n\t})\n}", "func (persist *Persister) createUserTable() {\n\n\tvar dbaser = persist.databaser\n\n\tdb, err := sql.Open(dbaser.Driver(), dbaser.Name())\n\tif err != nil {\n\t\tfmt.Println(\"Error on open of database\", err)\n\t\treturn\n\t}\n\tdefer db.Close()\n\n\tvar query = fmt.Sprintf(\n\t\tcreateUserTable,\n\t\tdbaser.IncrementPrimaryKey(),\n\t\tdbaser.DateField())\n\n\t_, err = db.Exec(query)\n\tif err != nil {\n\t\tfmt.Printf(\"Error creating Users table, driver \\\"%s\\\", dbname \\\"%s\\\", query = \\\"%s\\\"\\n\",\n\t\t\tdbaser.Driver(), dbaser.Name(), query)\n\t\tfmt.Println(err)\n\t\treturn\n\t}\n}", "func (u *User) Register(ctx context.Context, user model.User) (*model.User, error) {\n\tspan, _ := jtrace.Tracer.SpanFromContext(ctx, \"register_user\")\n\tdefer span.Finish()\n\tspan.SetTag(\"register\", \"register user model\")\n\n\ttx := mysql.Storage.GetDatabase().Begin()\n\n\tif err := tx.Create(&user).Error; err != nil {\n\t\tlog := logger.GetZapLogger(false)\n\t\tlogger.Prepare(log).\n\t\t\tAppend(zap.Any(\"error\", fmt.Sprintf(\"register user: %s\", err))).\n\t\t\tLevel(zap.ErrorLevel).\n\t\t\tDevelopment().\n\t\t\tCommit(\"env\")\n\t\ttx.Rollback()\n\t\treturn nil, err\n\t}\n\tdefer tx.Commit()\n\n\treturn &user, nil\n}", "func (ua *UserAuth) User(ctx context.Context, db XODB) (*User, error) {\n\treturn UserByUserID(ctx, db, ua.UserID)\n}", "func (s *UserRepository) User(id string) (*akwad.Account, error) {\n\n\treturn nil, nil\n}", "func (userRepo *mockUserRepo) Initialize(ctx context.Context, db *sql.DB) {}", "func setUser(ctx context.Context, data *User) error {\n\t// clear session_token and API_token for user\n\tk := datastore.NameKey(\"Users\", strings.ToLower(data.Username), nil)\n\n\t// New struct, to not add body, author etc\n\n\tif _, err := dbclient.Put(ctx, k, data); err != nil {\n\t\tlog.Println(err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (d *PostgresDatabase) Store(u model.User) error {\n\topMetric.WithLabelValues(\"store-user\").Inc()\n\n\tostart := time.Now()\n\tdefer opDuration.WithLabelValues(\"store-user\").Observe(time.Since(ostart).Seconds())\n\n\tdob, err := time.Parse(\"2006-01-02\", u.DOB)\n\tif err != nil {\n\t\topErrMetric.WithLabelValues(\"store-user\").Inc()\n\t\treturn err\n\t}\n\t_, err = d.Get(u.Username)\n\t// if the user is not present we insert\n\tif err == ErrUserNotFound {\n\t\t_, err := d.db.Query(queryInsertUser, u.Username, dob)\n\t\tif err != nil {\n\t\t\topErrMetric.WithLabelValues(\"store-user\").Inc()\n\t\t\treturn err\n\t\t}\n\t\topMetric.WithLabelValues(\"store-user\").Inc()\n\t\treturn nil\n\t} else if err != nil {\n\t\topErrMetric.WithLabelValues(\"store-user\").Inc()\n\t\treturn err\n\t}\n\n\t// if we got here, the user is already present and we should do an update\n\tstmt, err := d.db.Prepare(queryUpdateUser)\n\t_, err = stmt.Exec(dob, u.Username)\n\tif err != nil {\n\t\topErrMetric.WithLabelValues(\"store-user\").Inc()\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (s *XPackSecurityPutUserService) User(user *XPackSecurityPutUserRequest) *XPackSecurityPutUserService {\n\ts.user = user\n\treturn s\n}", "func (db *UserDatabase) Init() error {\n\tvar err error\n\ttx, err := db.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor _, stmt := range schemaV1 {\n\t\tlog.Println(stmt)\n\t\ttx.MustExec(stmt)\n\t}\n\tdefaultPassword := getDefaultPassword()\n\t_, err = tx.CreateUser(claudia.ApplicationAdminUsername, defaultPassword)\n\tif err != nil {\n\t\treturn err\n\t}\n\tsessionAuthKey := securecookie.GenerateRandomKey(32)\n\tsessionCryptKey := securecookie.GenerateRandomKey(32)\n\tcrt, key := util.GenerateSelfSignedCert()\n\ttx.MustExec(\"INSERT INTO configuration (schema_version, session_auth_key, session_crypt_key, private_key, public_certificate) VALUES ($1, $2, $3, $4, $5)\",\n\t\tSchemaVersion, sessionAuthKey, sessionCryptKey, key, crt)\n\ttx.Commit()\n\ttx, err = db.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\tconf, err := tx.GetConfiguration()\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\ttx.Commit()\n\tlog.Printf(\"Successfully initialized database (schema: %d)\", conf.SchemaVersion)\n\treturn nil\n}", "func init() {\n\tuserFields := schema.User{}.Fields()\n\t_ = userFields\n\t// userDescUserName is the schema descriptor for user_name field.\n\tuserDescUserName := userFields[0].Descriptor()\n\t// user.DefaultUserName holds the default value on creation for the user_name field.\n\tuser.DefaultUserName = userDescUserName.Default.(string)\n\t// userDescEmail is the schema descriptor for email field.\n\tuserDescEmail := userFields[1].Descriptor()\n\t// user.DefaultEmail holds the default value on creation for the email field.\n\tuser.DefaultEmail = userDescEmail.Default.(string)\n\t// userDescTelNum is the schema descriptor for tel_num field.\n\tuserDescTelNum := userFields[2].Descriptor()\n\t// user.DefaultTelNum holds the default value on creation for the tel_num field.\n\tuser.DefaultTelNum = userDescTelNum.Default.(string)\n\t// userDescPassword is the schema descriptor for password field.\n\tuserDescPassword := userFields[3].Descriptor()\n\t// user.DefaultPassword holds the default value on creation for the password field.\n\tuser.DefaultPassword = userDescPassword.Default.(string)\n\t// userDescPasswordStr is the schema descriptor for password_str field.\n\tuserDescPasswordStr := userFields[4].Descriptor()\n\t// user.DefaultPasswordStr holds the default value on creation for the password_str field.\n\tuser.DefaultPasswordStr = userDescPasswordStr.Default.(string)\n\t// userDescRegType is the schema descriptor for reg_type field.\n\tuserDescRegType := userFields[5].Descriptor()\n\t// user.DefaultRegType holds the default value on creation for the reg_type field.\n\tuser.DefaultRegType = userDescRegType.Default.(int8)\n\t// userDescRegisterIP is the schema descriptor for register_ip field.\n\tuserDescRegisterIP := userFields[7].Descriptor()\n\t// user.DefaultRegisterIP holds the default value on creation for the register_ip field.\n\tuser.DefaultRegisterIP = userDescRegisterIP.Default.(int)\n\t// userDescTelStatus is the schema descriptor for tel_status field.\n\tuserDescTelStatus := userFields[8].Descriptor()\n\t// user.DefaultTelStatus holds the default value on creation for the tel_status field.\n\tuser.DefaultTelStatus = userDescTelStatus.Default.(int8)\n\t// userDescStatus is the schema descriptor for status field.\n\tuserDescStatus := userFields[9].Descriptor()\n\t// user.DefaultStatus holds the default value on creation for the status field.\n\tuser.DefaultStatus = userDescStatus.Default.(int8)\n\t// userDescCreatedAt is the schema descriptor for created_at field.\n\tuserDescCreatedAt := userFields[10].Descriptor()\n\t// user.DefaultCreatedAt holds the default value on creation for the created_at field.\n\tuser.DefaultCreatedAt = userDescCreatedAt.Default.(func() time.Time)\n\tuserextendFields := schema.UserExtend{}.Fields()\n\t_ = userextendFields\n\t// userextendDescRealName is the schema descriptor for real_name field.\n\tuserextendDescRealName := userextendFields[1].Descriptor()\n\t// userextend.DefaultRealName holds the default value on creation for the real_name field.\n\tuserextend.DefaultRealName = userextendDescRealName.Default.(string)\n\t// userextendDescIDNumber is the schema descriptor for id_number field.\n\tuserextendDescIDNumber := userextendFields[2].Descriptor()\n\t// userextend.DefaultIDNumber holds the default value on creation for the id_number field.\n\tuserextend.DefaultIDNumber = userextendDescIDNumber.Default.(string)\n\t// userextendDescAge is the schema descriptor for age field.\n\tuserextendDescAge := userextendFields[3].Descriptor()\n\t// userextend.DefaultAge holds the default value on creation for the age field.\n\tuserextend.DefaultAge = userextendDescAge.Default.(int8)\n\t// userextendDescSex is the schema descriptor for sex field.\n\tuserextendDescSex := userextendFields[4].Descriptor()\n\t// userextend.DefaultSex holds the default value on creation for the sex field.\n\tuserextend.DefaultSex = userextendDescSex.Default.(int8)\n\t// userextendDescBirth is the schema descriptor for birth field.\n\tuserextendDescBirth := userextendFields[5].Descriptor()\n\t// userextend.DefaultBirth holds the default value on creation for the birth field.\n\tuserextend.DefaultBirth = userextendDescBirth.Default.(int)\n\t// userextendDescIcon is the schema descriptor for icon field.\n\tuserextendDescIcon := userextendFields[6].Descriptor()\n\t// userextend.DefaultIcon holds the default value on creation for the icon field.\n\tuserextend.DefaultIcon = userextendDescIcon.Default.(string)\n\t// userextendDescNickName is the schema descriptor for nick_name field.\n\tuserextendDescNickName := userextendFields[7].Descriptor()\n\t// userextend.DefaultNickName holds the default value on creation for the nick_name field.\n\tuserextend.DefaultNickName = userextendDescNickName.Default.(string)\n\t// userextendDescImei is the schema descriptor for imei field.\n\tuserextendDescImei := userextendFields[8].Descriptor()\n\t// userextend.DefaultImei holds the default value on creation for the imei field.\n\tuserextend.DefaultImei = userextendDescImei.Default.(string)\n\t// userextendDescOaid is the schema descriptor for oaid field.\n\tuserextendDescOaid := userextendFields[9].Descriptor()\n\t// userextend.DefaultOaid holds the default value on creation for the oaid field.\n\tuserextend.DefaultOaid = userextendDescOaid.Default.(string)\n\t// userextendDescDeviceID is the schema descriptor for device_id field.\n\tuserextendDescDeviceID := userextendFields[10].Descriptor()\n\t// userextend.DefaultDeviceID holds the default value on creation for the device_id field.\n\tuserextend.DefaultDeviceID = userextendDescDeviceID.Default.(string)\n\t// userextendDescSystemName is the schema descriptor for system_name field.\n\tuserextendDescSystemName := userextendFields[11].Descriptor()\n\t// userextend.DefaultSystemName holds the default value on creation for the system_name field.\n\tuserextend.DefaultSystemName = userextendDescSystemName.Default.(string)\n\t// userextendDescSystemVersion is the schema descriptor for system_version field.\n\tuserextendDescSystemVersion := userextendFields[12].Descriptor()\n\t// userextend.DefaultSystemVersion holds the default value on creation for the system_version field.\n\tuserextend.DefaultSystemVersion = userextendDescSystemVersion.Default.(string)\n\t// userextendDescAdid is the schema descriptor for adid field.\n\tuserextendDescAdid := userextendFields[13].Descriptor()\n\t// userextend.DefaultAdid holds the default value on creation for the adid field.\n\tuserextend.DefaultAdid = userextendDescAdid.Default.(string)\n\t// userextendDescGameID is the schema descriptor for game_id field.\n\tuserextendDescGameID := userextendFields[14].Descriptor()\n\t// userextend.DefaultGameID holds the default value on creation for the game_id field.\n\tuserextend.DefaultGameID = userextendDescGameID.Default.(string)\n\t// userextendDescThirdPlatformID is the schema descriptor for third_platform_id field.\n\tuserextendDescThirdPlatformID := userextendFields[15].Descriptor()\n\t// userextend.DefaultThirdPlatformID holds the default value on creation for the third_platform_id field.\n\tuserextend.DefaultThirdPlatformID = userextendDescThirdPlatformID.Default.(int8)\n\t// userextendDescCreatedAt is the schema descriptor for created_at field.\n\tuserextendDescCreatedAt := userextendFields[16].Descriptor()\n\t// userextend.DefaultCreatedAt holds the default value on creation for the created_at field.\n\tuserextend.DefaultCreatedAt = userextendDescCreatedAt.Default.(func() time.Time)\n}", "func (mgr *UserMgr) InsertUser(user *User) {\n\tuser.Password = HashPassword(user.Password)\n\tsql := \"INSERT INTO users (username, email, password, role) VALUES (:username, :email, :password, :role)\"\n\t_, err := mgr.db.NamedExec(sql, user)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n}", "func (s *BasePlSqlParserListener) EnterCreate_user(ctx *Create_userContext) {}", "func (m *mysqlUserRepository) Store(u *domain.User) (err error) {\n\tquery := `INSERT INTO user SET name=?, email=?, password=?, updated_at=?, created_at, deleted_at`\n\n\tstmt, err := m.Conn.Prepare(query)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tres, err := stmt.Exec(u.Name, u.Email, u.Password, time.Now(), nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tlastID, err := res.LastInsertId()\n\tif err != nil {\n\t\treturn\n\t}\n\tu.ID = lastID\n\treturn\n\n}", "func init() {\n\tuserFields := schema.User{}.Fields()\n\t_ = userFields\n\t// userDescApikey is the schema descriptor for apikey field.\n\tuserDescApikey := userFields[5].Descriptor()\n\t// user.DefaultApikey holds the default value on creation for the apikey field.\n\tuser.DefaultApikey = userDescApikey.Default.(func() uuid.UUID)\n\t// userDescActivated is the schema descriptor for activated field.\n\tuserDescActivated := userFields[8].Descriptor()\n\t// user.DefaultActivated holds the default value on creation for the activated field.\n\tuser.DefaultActivated = userDescActivated.Default.(bool)\n\t// userDescLocked is the schema descriptor for locked field.\n\tuserDescLocked := userFields[9].Descriptor()\n\t// user.DefaultLocked holds the default value on creation for the locked field.\n\tuser.DefaultLocked = userDescLocked.Default.(bool)\n\t// userDescUpdatedAt is the schema descriptor for updated_at field.\n\tuserDescUpdatedAt := userFields[10].Descriptor()\n\t// user.DefaultUpdatedAt holds the default value on creation for the updated_at field.\n\tuser.DefaultUpdatedAt = userDescUpdatedAt.Default.(func() time.Time)\n\t// userDescCreatedAt is the schema descriptor for created_at field.\n\tuserDescCreatedAt := userFields[11].Descriptor()\n\t// user.DefaultCreatedAt holds the default value on creation for the created_at field.\n\tuser.DefaultCreatedAt = userDescCreatedAt.Default.(func() time.Time)\n\t// userDescID is the schema descriptor for id field.\n\tuserDescID := userFields[0].Descriptor()\n\t// user.DefaultID holds the default value on creation for the id field.\n\tuser.DefaultID = userDescID.Default.(func() uuid.UUID)\n}", "func (usr *UserAuth) createUserTable(cfg *config.CfgService) (err error) {\n\tuserSchema := []string{\n\t\t\"username TEXT\",\n\t\t\"password TEXT\",\n\t\t\"isAdmin INTEGER\",\n\t\t\"loggedIn INTEGER\",\n\t\t\"ccNumber INTEGER\",\n\t\t\"ccExpiryMonth INTEGER\",\n\t\t\"ccExpiryYear INTEGER\",\n\t\t\"cvv INTEGER\",\n\t\t\"cardName TEXT\"}\n\n\tuserDefaultAdmin := map[string]string{\n\t\t\"username\": \"'admin'\",\n\t\t\"password\": \"'admin'\",\n\t\t\"isAdmin\": \"1\",\n\t\t\"loggedIn\": \"0\",\n\t\t\"ccNumber\": \"\",\n\t\t\"ccExpiryMonth\": \"0\",\n\t\t\"ccExpiryYear\": \"0\",\n\t\t\"cvv\": \"0\",\n\t\t\"cardName\": \"\"}\n\n\tuserDefault := map[string]string{\n\t\t\"username\": \"'user'\",\n\t\t\"password\": \"'user'\",\n\t\t\"isAdmin\": \"0\",\n\t\t\"loggedIn\": \"0\",\n\t\t\"ccNumber\": \"\",\n\t\t\"ccExpiryMonth\": \"0\",\n\t\t\"ccExpiryYear\": \"0\",\n\t\t\"cvv\": \"0\",\n\t\t\"cardName\": \"\"}\n\n\terr = cfg.CreateTable(usr.Name, userSchema)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = cfg.InitUser(usr.Name, userDefaultAdmin, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = cfg.InitUser(usr.Name, userDefault, 1)\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\treturn nil\n}", "func InsertUser(u User, t string) {\n\t// Begin transaction\n\ttx, err := globals.Db.Begin()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer tx.Rollback()\n\n\t// Prepare user insertion and execute\n\tstmt, err := tx.Prepare(\"INSERT INTO tblUsers(fldFirstName, fldLastName, fldEmail, fldPassword) VALUES (?, ?, ?, ?)\")\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer stmt.Close()\n\n\tres, err := stmt.Exec(u.FirstName, u.LastName, u.Email, u.Password)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\t// Get the last inserted row's ID\n\tlastID, err := res.LastInsertId()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\t// Prepare token insertion and execute\n\tstmt, err = tx.Prepare(\"INSERT INTO tblActivationTokens (fldToken,fldFKUserID) VALUES (?, ?)\")\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tres, err = stmt.Exec(t, lastID)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\t// Commit query\n\terr = tx.Commit()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n}", "func GetUser() {\n\tvar user User\n\terr := orm.NewOrm().QueryTable(\"t_user\").Filter(\"Id\", 5).One(&user)\n\tif err == nil {\n\t\tfmt.Println(user.ToString())\n\t}\n}", "func (s *BasePlSqlParserListener) EnterAlter_user(ctx *Alter_userContext) {}", "func createSchema(db *pg.DB) (err error) {\n\t// db.AddQueryHook(dbLogger{})\n\n\trefresh := false\n\n\tmodels := []interface{}{\n\t\t&structs.User{},\n\t\t&structs.Project{},\n\t\t&structs.Webhook{},\n\t\t&structs.IssueEntry{},\n\t\t&structs.Comment{},\n\t\t&structs.InviteCode{},\n\t}\n\n\tfor _, model := range models {\n\t\tif refresh {\n\t\t\tdb.Model(model).DropTable(&orm.DropTableOptions{\n\t\t\t\tIfExists: true,\n\t\t\t\tCascade: true,\n\t\t\t})\n\t\t}\n\n\t\terr = db.Model(model).CreateTable(&orm.CreateTableOptions{\n\t\t\tIfNotExists: true,\n\t\t})\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\n\tif refresh {\n\t\tidGen := idgenerator.NewIDGenerator(1602507674941, 0)\n\n\t\tprintln(\"Create User\")\n\t\tuser := &structs.User{\n\t\t\tID: idGen.GenerateID(),\n\t\t\tName: \"ImRock\",\n\t\t\tAvatar: \"https://cdn.discordapp.com/avatars/143090142360371200/a_70444022ea3e5d73dd00d59c5578b07e.gif?size=1024\",\n\t\t\tUserType: structs.DiscordUser,\n\t\t\tHookID: 143090142360371200,\n\t\t\tProjectIDs: make([]int64, 0),\n\t\t\tIntegration: false,\n\t\t}\n\n\t\t_, err = db.Model(user).Insert()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Create second user\")\n\t\tuser2 := &structs.User{\n\t\t\tID: idGen.GenerateID(),\n\t\t\tName: \"biscuitcord\",\n\t\t\tAvatar: \"https://cdn.discordapp.com/avatars/164297154276360192/4c8f9b0310948cce460613081d074a13.webp?size=1024\",\n\t\t\tUserType: structs.DiscordUser,\n\t\t\tHookID: 164297154276360192,\n\t\t\tProjectIDs: make([]int64, 0),\n\t\t\tIntegration: false,\n\t\t}\n\t\t_, err = db.Model(user2).Insert()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Create Project\")\n\t\tproject := &structs.Project{\n\t\t\tID: idGen.GenerateID(),\n\n\t\t\tCreatedAt: time.Now().UTC(),\n\t\t\tCreatedByID: user.ID,\n\n\t\t\tIntegrations: make([]*structs.User, 0),\n\t\t\tWebhooks: make([]*structs.Webhook, 0),\n\n\t\t\tSettings: structs.ProjectSettings{\n\t\t\t\tDisplayName: \"Welcomer\",\n\t\t\t\tURL: \"https://welcomer.gg\",\n\t\t\t\tArchived: false,\n\t\t\t\tPrivate: false,\n\t\t\t\tLimited: false,\n\t\t\t},\n\n\t\t\tStarredIssues: 0,\n\t\t\tOpenIssues: 0,\n\t\t\tActiveIssues: 0,\n\t\t\tResolvedIssues: 0,\n\t\t}\n\t\t_, err = db.Model(project).Insert()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tprintln(\"Add project to user\")\n\t\tuser.ProjectIDs = append(user.ProjectIDs, project.ID)\n\t\t_, err = db.Model(user).WherePK().Update()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Add project to user 2\")\n\t\tuser2.ProjectIDs = append(user2.ProjectIDs, project.ID)\n\t\t_, err = db.Model(user).WherePK().Update()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Add second user to contributors\")\n\t\tproject.Settings.ContributorIDs = append(project.Settings.ContributorIDs, user2.ID)\n\t\t_, err = db.Model(project).WherePK().Update()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Create Webhooks\")\n\t\twebhook := &structs.Webhook{\n\t\t\tID: idGen.GenerateID(),\n\t\t\tProjectID: project.ID,\n\t\t\tActive: false,\n\t\t\tFailures: 16,\n\t\t\tCreatedAt: time.Now().UTC(),\n\t\t\tCreatedByID: user.ID,\n\t\t\tURL: \"https://welcomer.gg/webhook\",\n\t\t\tType: structs.DiscordWebhook,\n\t\t\tJSONContent: true,\n\t\t\tSecret: \"\",\n\t\t}\n\t\t_, err = db.Model(webhook).Insert()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Create Integration\")\n\t\tintegration := &structs.User{\n\t\t\tID: idGen.GenerateID(),\n\t\t\tName: \"Welcomer\",\n\n\t\t\tUserType: structs.IntegrationUser,\n\n\t\t\tCreatedAt: time.Now().UTC(),\n\n\t\t\tProjectID: project.ID,\n\t\t\tIntegration: true,\n\t\t\tCreatedByID: user.ID,\n\t\t}\n\t\t_, err = db.Model(integration).Insert()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Create user issue\")\n\t\tnow := time.Now().UTC()\n\t\tissue := &structs.IssueEntry{\n\t\t\tID: idGen.GenerateID(),\n\t\t\tProjectID: project.ID,\n\n\t\t\tStarred: false,\n\n\t\t\tType: structs.EntryOpen,\n\t\t\tOccurrences: 1,\n\t\t\tAssigneeID: 0,\n\n\t\t\tError: \"genericError\",\n\t\t\tFunction: \"createSchema(db *pg.DB)\",\n\t\t\tCheckpoint: \"internal/api.go:147\",\n\t\t\tDescription: \"\",\n\t\t\tTraceback: \"\",\n\n\t\t\tLastModified: now,\n\n\t\t\tCreatedAt: now,\n\t\t\tCreatedByID: user.ID,\n\n\t\t\tCommentCount: 0,\n\t\t\tCommentsLocked: false,\n\t\t}\n\t\t_, err = db.Model(issue).Insert()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Increment project issue counter\")\n\t\tproject.OpenIssues++\n\t\t_, err = db.Model(project).WherePK().Update()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Create user issue 2\")\n\t\tnow = time.Now().UTC()\n\t\tissue2 := &structs.IssueEntry{\n\t\t\tID: idGen.GenerateID(),\n\t\t\tProjectID: project.ID,\n\n\t\t\tStarred: false,\n\n\t\t\tType: structs.EntryOpen,\n\t\t\tOccurrences: 5,\n\t\t\tAssigneeID: user.ID,\n\n\t\t\tError: \"panic:\",\n\t\t\tFunction: \"main.main.func1\",\n\t\t\tCheckpoint: \"main.go:11\",\n\t\t\tDescription: \"\",\n\t\t\tTraceback: \"stacktrace from panic: \\ngoroutine 1 [running]:\\nruntime/debug.Stack(0x1042ff18, 0x98b2, 0xf0ba0, 0x17d048)\\n /usr/local/go/src/runtime/debug/stack.go:24 +0xc0\\nmain.main.func1()\\n /tmp/sandbox973508195/main.go:11 +0x60\\npanic(0xf0ba0, 0x17d048)\\n /usr/local/go/src/runtime/panic.go:502 +0x2c0\\nmain.main()\\n /tmp/sandbox973508195/main.go:16 +0x60\",\n\n\t\t\tLastModified: now,\n\n\t\t\tCreatedAt: now,\n\t\t\tCreatedByID: user2.ID,\n\n\t\t\tCommentCount: 0,\n\t\t\tCommentsLocked: false,\n\t\t}\n\t\t_, err = db.Model(issue2).Insert()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Increment project issue counter\")\n\t\tproject.OpenIssues++\n\t\t_, err = db.Model(project).WherePK().Update()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Create integration issue\")\n\t\tnow = time.Now().UTC()\n\t\tissue3 := &structs.IssueEntry{\n\t\t\tID: idGen.GenerateID(),\n\t\t\tProjectID: project.ID,\n\n\t\t\tStarred: false,\n\n\t\t\tType: structs.EntryOpen,\n\t\t\tOccurrences: 1,\n\t\t\tAssigneeID: user2.ID,\n\n\t\t\tError: \"TypeError\",\n\t\t\tFunction: \"\",\n\t\t\tCheckpoint: \"\",\n\t\t\tDescription: \"can only concatenate str (not \\\"int\\\") to str\",\n\t\t\tTraceback: \"Traceback (most recent call last):\\n File \\\"<stdin>\\\", line 1, in <module>\\n File \\\"<stdin>\\\", line 2, in a\\nTypeError: can only concatenate str (not \\\"int\\\") to str\",\n\n\t\t\tLastModified: now,\n\n\t\t\tCreatedAt: now,\n\t\t\tCreatedByID: integration.ID,\n\n\t\t\tCommentCount: 0,\n\t\t\tCommentsLocked: false,\n\t\t}\n\t\t_, err = db.Model(issue3).Insert()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Increment project issue counter\")\n\t\tproject.OpenIssues++\n\t\t_, err = db.Model(project).WherePK().Update()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Create user issue comment\")\n\t\tcontent := \"Test :)\"\n\t\tcomment := &structs.Comment{\n\t\t\tID: idGen.GenerateID(),\n\t\t\tIssueID: issue.ID,\n\n\t\t\tCreatedAt: time.Now().UTC(),\n\t\t\tCreatedByID: user2.ID,\n\n\t\t\tType: structs.Message,\n\t\t\tContent: &content,\n\t\t}\n\t\t_, err = db.Model(comment).Insert()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tissue.CommentCount++\n\n\t\tprintln(\"Create user issue comment2\")\n\t\topen := structs.EntryOpen\n\t\tcomment2 := &structs.Comment{\n\t\t\tID: idGen.GenerateID(),\n\t\t\tIssueID: issue.ID,\n\n\t\t\tCreatedAt: time.Now().UTC(),\n\t\t\tCreatedByID: user2.ID,\n\n\t\t\tType: structs.IssueMarked,\n\t\t\tIssueMarked: &open,\n\t\t}\n\t\t_, err = db.Model(comment2).Insert()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tissue.CommentCount++\n\n\t\tprintln(\"Create user issue comment3\")\n\t\topened := true\n\t\tcomment3 := &structs.Comment{\n\t\t\tID: idGen.GenerateID(),\n\t\t\tIssueID: issue.ID,\n\n\t\t\tCreatedAt: time.Now().UTC(),\n\t\t\tCreatedByID: user.ID,\n\n\t\t\tType: structs.CommentsLocked,\n\t\t\tCommentsOpened: &opened,\n\t\t}\n\t\t_, err = db.Model(comment3).Insert()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tissue.CommentCount++\n\n\t\tprintln(\"Update issue comment count\")\n\t\t_, err = db.Model(issue).WherePK().Update()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Create user 2 issue comment\")\n\t\tcomment4 := &structs.Comment{\n\t\t\tID: idGen.GenerateID(),\n\t\t\tIssueID: issue2.ID,\n\n\t\t\tCreatedAt: time.Now().UTC(),\n\t\t\tCreatedByID: user.ID,\n\n\t\t\tType: structs.CommentsLocked,\n\t\t\tCommentsOpened: &opened,\n\t\t}\n\t\t_, err = db.Model(comment4).Insert()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Update issue2 comment count\")\n\t\tissue2.CommentCount++\n\t\t_, err = db.Model(issue2).WherePK().Update()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Star user issue\")\n\t\tissue2.Starred = true\n\t\t_, err = db.Model(issue2).WherePK().Update()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Update project stars\")\n\t\tproject.StarredIssues++\n\t\t_, err = db.Model(project).WherePK().Update()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Close integration issue\")\n\t\tissue3.Type = structs.EntryResolved\n\t\tissue3.LastModified = time.Now().UTC()\n\t\t_, err = db.Model(issue3).WherePK().Update()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Update project issue counter\")\n\t\tproject.ResolvedIssues++\n\t\tproject.OpenIssues--\n\t\t_, err = db.Model(project).WherePK().Update()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tprintln(\"Create close integration issue comment\")\n\t\tresolved := structs.EntryResolved\n\t\tcomment5 := &structs.Comment{\n\t\t\tID: idGen.GenerateID(),\n\t\t\tIssueID: issue3.ID,\n\n\t\t\tCreatedAt: time.Now().UTC(),\n\t\t\tCreatedByID: user.ID,\n\n\t\t\tType: structs.IssueMarked,\n\t\t\tIssueMarked: &resolved,\n\t\t}\n\t\t_, err = db.Model(comment5).Insert()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (userRepo *PostUserRepository) StoreUser(u entity.User) error {\n\n\t_, err := userRepo.conn.Exec(\"INSERT INTO users (firstname,lastname,email,password)\"+\n\t\t\" values($1, $2, $3, $4)\", u.FirstName, u.LastName, u.Email, u.Password)\n\n\tif err != nil {\n\t\t//panic(err)\n\t\treturn errors.New(\"Insertion has failed\")\n\t}\n\n\treturn nil\n}", "func (ur *UserGormRepo) StoreUser(user *entity.User) (*entity.User, []error) {\n\tusr := user\n\tusr.Password,_ = handler.HashPassword(user.Password)\n\terrs := ur.conn.Create(usr).GetErrors()\n\n\tfor _, err := range errs {\n\t\tpqerr := err.(*pq.Error)\n\t\tfmt.Println(pqerr)\n\t}\n\n\tif len(errs) > 0 {\n\t\treturn nil, errs\n\t}\n\n\treturn usr, nil\n}", "func (user *User) Save() error {\n\tif user.RegistrationDate.IsZero() {\n\t\tuser.RegistrationDate = time.Now()\n\t}\n\tvar q *ara.Query\n\tif user.Key == nil {\n\t\trd, _ := user.RegistrationDate.MarshalJSON()\n\t\tq = ara.NewQuery(`INSERT {\n\t\t\t\tUsername: %q,\n\t\t\t\tEmail: %q,\n\t\t\t\tPassword: %q,\n\t\t\t\tAge: %d,\n\t\t\t\tGender: %q,\n\t\t\t\tLikes: %q,\n\t\t\t\tMeets: %q,\n\t\t\t\tRegistrationDate: %s\n\t\t\t} IN users`,\n\t\t\tuser.Username,\n\t\t\tuser.Email,\n\t\t\tuser.Password,\n\t\t\tuser.Age,\n\t\t\tuser.Gender,\n\t\t\tuser.Likes,\n\t\t\tuser.Meets,\n\t\t\trd,\n\t\t\t)\n\n\t} else {\n\t\tq = ara.NewQuery(`UPDATE %q WITH {\n\t\t\t\tUsername: %q,\n\t\t\t\tEmail: %q,\n\t\t\t\tPassword: %q,\n\t\t\t\tAge: %d,\n\t\t\t\tGender: %q,\n\t\t\t\tLikes: %q,\n\t\t\t\tMeets: %q\n\t\t\t} IN users`,\n\t\t\t*user.Key,\n\t\t\tuser.Username,\n\t\t\tuser.Email,\n\t\t\tuser.Password,\n\t\t\tuser.Age,\n\t\t\tuser.Gender,\n\t\t\tuser.Likes,\n\t\t\tuser.Meets,\n\t\t\t)\n\t}\n\tlog.Println(q)\n\t_, err := db.Run(q)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn err\n\t}\n\tvar users []User\n\tq = ara.NewQuery(`FOR user IN users FILTER user.Username == %q RETURN user`, user.Username).Cache(true).BatchSize(500)\n\tresp, err := db.Run(q)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn err\n\t}\n\tlog.Println(string(resp))\n\terr = json.Unmarshal(resp, &users)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn err\n\t}\n\tlog.Println(users)\n\tif len(users) > 0 {\n\t\t*user = users[0]\n\t\treturn nil\n\t}\n\treturn errors.New(\"prout\")\n}", "func (_UserCrud *UserCrudTransactorSession) InsertUser(userAddress common.Address, userEmail string, userAge *big.Int) (*types.Transaction, error) {\n\treturn _UserCrud.Contract.InsertUser(&_UserCrud.TransactOpts, userAddress, userEmail, userAge)\n}", "func CreateUser(user *User) error {\n\n //Validate fields\n\n //Username must be between 3 and 20 alphanumeric characters\n invalidCharsRegex := regexp.MustCompile(\"[^A-Za-z0-9]+\")\n if len(user.Username) < 3 || len(user.Username) > 20 || invalidCharsRegex.MatchString(user.Username) {\n return errors.New(\"username must be 3 to 20 alphanumeric characters\") \n }\n\n //Passwords must be at least 8 characters\n if len(user.Password) < 8 {\n return errors.New(\"password must be at least 8 characters\")\n }\n\n db, err := bolt.Open(DB_NAME, 0600, nil)\n if err != nil {\n panic(err)\n }\n defer db.Close()\n\n err = db.Update(func(tx *bolt.Tx) error {\n\n b := tx.Bucket([]byte(USER_BUCKET))\n\n //ensure username is not taken\n v := b.Get([]byte(user.Username))\n if v != nil {\n return errors.New(\"username already taken\")\n }\n\n // generate UUID that never changes for this user.\n user.ID = uuid.New().String()\n\n // hash password (NOTE this will modify the user object passed in)\n user.Password = hashPassword(user.Password)\n\n // encode to JSON\n encoded, err := json.Marshal(user)\n if err != nil {\n return err\n }\n\n // put in database\n err = b.Put([]byte(user.Username), encoded)\n\n return err //nil implies commit transaction, otherwise rollback\n })\n\n return err\n}", "func (s *CreateUserEndpoint) saveToDB(user *User) (int, error) {\n\t// implementation removed\n\treturn 0, nil\n}", "func (ur *UserGormRepo) User(user *entity.User) (*entity.User, []error) {\n\tlgusr := user\n\tusr := entity.User{}\n\terrs := ur.conn.Where(\"email = ?\", user.Email).First(&usr).GetErrors()\n\terr := bcrypt.CompareHashAndPassword([]byte(usr.Password), []byte(lgusr.Password))\n\tfmt.Println(err)\n\tif err != nil {\n\t\treturn nil, []error{err}\n\t}\n\n\tif len(errs) > 0 {\n\t\treturn nil, errs\n\t}\n\n\treturn &usr, nil\n}", "func (tx *txDriver) Tx(context.Context) (dialect.Tx, error) { return tx, nil }", "func (tx *txDriver) Tx(context.Context) (dialect.Tx, error) { return tx, nil }", "func init() {\n\tuserHooks := schema.User{}.Hooks()\n\tuser.Hooks[0] = userHooks[0]\n\tuser.Hooks[1] = userHooks[1]\n}", "func Init() {\n\tdb, err = gorm.Open(\"postgres\", \"host=db port=5432 user=LikeTwitterApp-backend dbname=LikeTwitterApp-backend password=LikeTwitterApp-backend sslmode=disable\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tautoMigration()\n\tuser := models.User{\n\t\tID: 1,\n\t\tName: \"aoki\",\n\t\tPosts: []models.Post{{ID: 1, Content: \"tweet1\"}, {ID: 2, Content: \"tweet2\"}},\n\t}\n\tdb.Create(&user)\n}", "func (u *User) StoreUser() error {\n\terr := database.DB.Create(&u).Error\n\treturn err\n}", "func main() {\n\tdb, err := db.OpenConnection()\n\tdb.LogMode(true)\n\n\tif err != nil {\n\t\tlog.WithError(err).Fatal(\"An error occurred while connecting to the database\")\n\t}\n\tdefer db.Close()\n\n\t// Create models\n\tdb.DropTableIfExists(&domain.User{}, &domain.Email{}, &domain.Role{})\n\tdb.CreateTable(&domain.Email{}, &domain.Role{}, &domain.User{})\n\n\t// Create user\n\tid, _ := uuid.Parse(\"8c5df3bc-9fa6-4d73-b79a-9a1cbb35740c\")\n\tuser := domain.User{\n\t\tID: id,\n\t\tEmails: []domain.Email{\n\t\t\t{Email: \"[email protected]\"},\n\t\t},\n\t\tRoles: []domain.Role{\n\t\t\t{Role: \"ROLE_USER\"},\n\t\t},\n\t\tPassword: \"Test1234\",\n\t}\n\n\t// Create developer\n\tid, _ = uuid.Parse(\"b0203081-5dfe-4bb7-87d1-e2c59e2af7b6\")\n\tdeveloper := domain.User{\n\t\tID: id,\n\t\tEmails: []domain.Email{\n\t\t\t{Email: \"[email protected]\"},\n\t\t},\n\t\tRoles: []domain.Role{\n\t\t\t{Role: \"ROLE_USER\"},\n\t\t\t{Role: \"ROLE_DEVELOPER\"},\n\t\t},\n\t\tPassword: \"Test1234\",\n\t}\n\n\t// Create hr\n\tid, _ = uuid.Parse(\"6b59c645-82c3-4e08-b089-f4236a2141b6\")\n\thr := domain.User{\n\t\tID: id,\n\t\tEmails: []domain.Email{\n\t\t\t{Email: \"[email protected]\"},\n\t\t},\n\t\tRoles: []domain.Role{\n\t\t\t{Role: \"ROLE_USER\"},\n\t\t\t{Role: \"ROLE_HR\"},\n\t\t},\n\t\tPassword: \"Test1234\",\n\t}\n\n\t// Create manager\n\tid, _ = uuid.Parse(\"75129bb5-5c12-48a1-8410-bb2630fff9ed\")\n\tmanager := domain.User{\n\t\tID: id,\n\t\tEmails: []domain.Email{\n\t\t\t{Email: \"[email protected]\"},\n\t\t\t{Email: \"[email protected]\"},\n\t\t},\n\t\tRoles: []domain.Role{\n\t\t\t{Role: \"ROLE_USER\"},\n\t\t\t{Role: \"ROLE_MANAGER\"},\n\t\t},\n\t\tPassword: \"Test1234\",\n\t}\n\n\tdb.Create(&user)\n\tdb.Create(&developer)\n\tdb.Create(&hr)\n\tdb.Create(&manager)\n}", "func createUser() User {\n\tuser := User{\n\t\tUsername: \"igor\",\n\t\tPassword: \"please store hashed password\",\n\t\tName: \"Paolo\",\n\t\tSurname: \"Galeone\",\n\t\tEmail: \"please validate the @email . com\",\n\t\tGender: true,\n\t\tBirthDate: time.Now(),\n\t}\n\n\tif e = db.Create(&user); e != nil {\n\t\tpanic(fmt.Sprintf(\"Create(&user) filling fields having no default should work, but got: %s\\n\", e.Error()))\n\t}\n\treturn user\n}", "func (u *User) Create() {\n\tconfig.DB.Create(u)\n}", "func (s *UserStore) Transaction(callback func(*UserStore) error) error {\n\tif callback == nil {\n\t\treturn kallax.ErrInvalidTxCallback\n\t}\n\n\treturn s.Store.Transaction(func(store *kallax.Store) error {\n\t\treturn callback(&UserStore{store})\n\t})\n}", "func (cli *Store) User() *UserRepository {\n\t// TODO: add test store for testing\n\tif cli.userRepository != nil {\n\t\treturn cli.userRepository\n\t}\n\n\tcli.userRepository = &UserRepository{\n\t\tstore: cli,\n\t}\n\n\treturn cli.userRepository\n}", "func SaveUser(user *User) {\n\tDb.Save(&user)\n}", "func CreateUser(\n\tctx context.Context,\n\ttx *sql.Tx,\n\trequest *models.CreateUserRequest) error {\n\tmodel := request.User\n\t// Prepare statement for inserting data\n\tstmt, err := tx.Prepare(insertUserQuery)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"preparing create statement failed\")\n\t}\n\tdefer stmt.Close()\n\tlog.WithFields(log.Fields{\n\t\t\"model\": model,\n\t\t\"query\": insertUserQuery,\n\t}).Debug(\"create query\")\n\t_, err = stmt.ExecContext(ctx, string(model.GetUUID()),\n\t\tcommon.MustJSON(model.GetPerms2().GetShare()),\n\t\tint(model.GetPerms2().GetOwnerAccess()),\n\t\tstring(model.GetPerms2().GetOwner()),\n\t\tint(model.GetPerms2().GetGlobalAccess()),\n\t\tstring(model.GetPassword()),\n\t\tstring(model.GetParentUUID()),\n\t\tstring(model.GetParentType()),\n\t\tbool(model.GetIDPerms().GetUserVisible()),\n\t\tint(model.GetIDPerms().GetPermissions().GetOwnerAccess()),\n\t\tstring(model.GetIDPerms().GetPermissions().GetOwner()),\n\t\tint(model.GetIDPerms().GetPermissions().GetOtherAccess()),\n\t\tint(model.GetIDPerms().GetPermissions().GetGroupAccess()),\n\t\tstring(model.GetIDPerms().GetPermissions().GetGroup()),\n\t\tstring(model.GetIDPerms().GetLastModified()),\n\t\tbool(model.GetIDPerms().GetEnable()),\n\t\tstring(model.GetIDPerms().GetDescription()),\n\t\tstring(model.GetIDPerms().GetCreator()),\n\t\tstring(model.GetIDPerms().GetCreated()),\n\t\tcommon.MustJSON(model.GetFQName()),\n\t\tstring(model.GetDisplayName()),\n\t\tcommon.MustJSON(model.GetAnnotations().GetKeyValuePair()))\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"create failed\")\n\t}\n\n\tmetaData := &common.MetaData{\n\t\tUUID: model.UUID,\n\t\tType: \"user\",\n\t\tFQName: model.FQName,\n\t}\n\terr = common.CreateMetaData(tx, metaData)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = common.CreateSharing(tx, \"user\", model.UUID, model.GetPerms2().GetShare())\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.WithFields(log.Fields{\n\t\t\"model\": model,\n\t}).Debug(\"created\")\n\treturn nil\n}", "func (a UserService) WithTrx(trxHandle *gorm.DB) UserService {\n\ta.userRepository = a.userRepository.WithTrx(trxHandle)\n\ta.userRoleRepository = a.userRoleRepository.WithTrx(trxHandle)\n\n\treturn a\n}", "func (s *Supplier) User() store.UserStore {\n\treturn postgres.NewPgUserStore(s.Pgst)\n}", "func (pg *PG) StoreUser(ctx context.Context, user *models.User, passwordHash []byte) (*models.User, error) {\n\tconst query string = `insert into public.users \n\t\t(name_user, email_addr, pwd_hash,\n\t\tfirst_name, last_name, phone_number, \n\t\tuser_role, description)\n\t\tvalues ($1,$2,$3,$4,$5,$6,$7,$8)\n\t\treturning id_user;`\n\n\tvar uid int64\n\terr := pg.db.QueryRow(ctx, query, user.Username, user.Email, passwordHash,\n\t\tuser.FirstName, user.LastName, user.Phone,\n\t\tuser.UserRole, user.Username+\" TODO: привести БД в соответствии с моделью\").Scan(&uid)\n\tif err != nil {\n\t\treturn nil, errors.WithMessage(err, \"insert user failed\")\n\t}\n\tif uid == 0 {\n\t\treturn nil, errors.Errorf(\"insert user failed, empty id\")\n\t}\n\tuser.ID = uid\n\t// TODO: привести в соответствие с моделью и спекой\n\t// что за статус, что он означает?\n\tuser.UserStatus = 1\n\n\treturn user, nil\n}", "func (usr *User) Insert() error {\n\tif _, err := orm.NewOrm().Insert(usr); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (_UserCrud *UserCrudSession) InsertUser(userAddress common.Address, userEmail string, userAge *big.Int) (*types.Transaction, error) {\n\treturn _UserCrud.Contract.InsertUser(&_UserCrud.TransactOpts, userAddress, userEmail, userAge)\n}", "func (o *Stock) User(exec boil.Executor, mods ...qm.QueryMod) userQuery {\n\tqueryMods := []qm.QueryMod{\n\t\tqm.Where(\"user_id=?\", o.UserID),\n\t}\n\n\tqueryMods = append(queryMods, mods...)\n\n\tquery := Users(exec, queryMods...)\n\tqueries.SetFrom(query.Query, \"`users`\")\n\n\treturn query\n}", "func init() {\n\tuserFields := schema.User{}.Fields()\n\t_ = userFields\n\t// userDescCreatedAt is the schema descriptor for created_at field.\n\tuserDescCreatedAt := userFields[2].Descriptor()\n\t// user.DefaultCreatedAt holds the default value on creation for the created_at field.\n\tuser.DefaultCreatedAt = userDescCreatedAt.Default.(func() time.Time)\n\t// userDescUpdatedAt is the schema descriptor for updated_at field.\n\tuserDescUpdatedAt := userFields[3].Descriptor()\n\t// user.DefaultUpdatedAt holds the default value on creation for the updated_at field.\n\tuser.DefaultUpdatedAt = userDescUpdatedAt.Default.(func() time.Time)\n\t// user.UpdateDefaultUpdatedAt holds the default value on update for the updated_at field.\n\tuser.UpdateDefaultUpdatedAt = userDescUpdatedAt.UpdateDefault.(func() time.Time)\n\t// userDescID is the schema descriptor for id field.\n\tuserDescID := userFields[0].Descriptor()\n\t// user.IDValidator is a validator for the \"id\" field. It is called by the builders before save.\n\tuser.IDValidator = userDescID.Validators[0].(func(int) error)\n}", "func (user *User) Save() {\n\tdb := common.GetDatabase()\n\n\tdb.Save(&user)\n}", "func (ug *userGorm) Create(user *User) error{\n\treturn ug.db.Create(user).Error\n}", "func RegisterUser(db *gorm.DB, w http.ResponseWriter, r *http.Request) {\n\tfmt.Println(\"register\")\n\truser := model.RUser{}\n\tdecoder := json.NewDecoder(r.Body)\n\tif err := decoder.Decode(&ruser); err != nil {\n\t\tRespondError(w, http.StatusBadRequest, \"\")\n\t\tlog.Println(\"decode:\", err.Error())\n\t\treturn\n\t}\n\tdefer r.Body.Close()\n\n\thashedPassword, err := bcrypt.GenerateFromPassword([]byte(ruser.Password), 8)\n\tif err != nil {\n\t\tRespondError(w, http.StatusInternalServerError, \"\")\n\t\tlog.Println(\"hash:\", err.Error())\n\t\treturn\n\t}\n\n\tid, err := uuid.NewUUID()\n\tif err != nil {\n\t\tRespondError(w, http.StatusInternalServerError, \"\")\n\t}\n\n\tuser := model.User{\n\t\tName: ruser.Name,\n\t\tUsername: ruser.Username,\n\t\tPassword: string(hashedPassword),\n\t\tUUID: id.String(),\n\t}\n\n\tif err := db.Save(&user).Error; err != nil {\n\t\tRespondError(w, http.StatusInternalServerError, \"\")\n\t\tlog.Println(\"save:\", err.Error())\n\t\treturn\n\t}\n\tRespondJSON(w, http.StatusCreated, user)\n}", "func (rep *UserRepository) Create(user *models.User) (err error) {\n\tuser.Created = utils.GetTimestampNow()\n\tuser.Updated = utils.GetTimestampNow()\n\terr = databaseConnection.Create(user).Error\n\tif err != nil {\n\t\tlog.Error(0, \"Could not create user: %v\", err)\n\t\treturn\n\t}\n\treturn\n}", "func (ug *userDbHandle) Create(user *User) error {\n\treturn ug.db.Create(user).Error\n}", "func NewUser(db *sql.DB) *User {\n\tvar dbx = sqlx.NewDb(db, \"postgres\")\n\treturn &User{\n\t\tdb: dbx,\n\t}\n}", "func CreateUser(c *gin.Context) {\n\n\tfmt.Println(\"Endpoint Hit: Create A new User\")\n\n\tuser := model.Users{}\n\n\terr := c.Bind(&user)\n\n\tif err != nil {\n\t\tfmt.Print(err)\n\t}\n\n\tfmt.Println(user.Name)\n\n\tfmt.Println(user.Email)\n\tfmt.Println(user.Password)\n\tdb, err := sql.Open(\"mysql\", \"root:password@tcp(127.0.0.1:3306)/twitter\")\n\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\n\t_, errQ := db.Query(\"INSERT INTO users(name, email, password) VALUES (?,?,?)\", user.Name, user.Email, user.Password)\n\n\tif errQ != nil {\n\t\tpanic(err.Error())\n\t}\n\n\tc.JSON(http.StatusOK, gin.H{\n\t\t\"Name\": user.Name,\n\t\t\"Email\": user.Email,\n\t\t\"Password\": user.Password,\n\t})\n\n\tdefer db.Close()\n\n}", "func (ug *userGorm) Create(user *User) error {\n\treturn ug.db.Create(user).Error\n\n}", "func dataSourceUser() *schema.Resource {\n\treturn &schema.Resource{\n\t\tRead: dataSourceUserRead,\n\t\tSchema: userschema.ReadSchema(),\n\t}\n}", "func registerUser(stub shim.ChaincodeStubInterface, args []string) ([]byte, error) {\n\n\tvar user_obj sbiStruct.user\t\n\tvar err error\n\n\tfmt.Println(\"Entering registerUser\")\n\n\tif (len(args) < 1) {\n\t\tfmt.Println(\"Invalid number of args\")\n\t\treturn nil, errors.New(\"Expected atleast one arguments for initiate Transaction\")\n\t}\n\n\tfmt.Println(\"Args [0] is : %v\\n\",args[0])\n\tfmt.Println(\"Args [1] is : %v\\n\",args[1])\n\t\n\t//unmarshal transaction initiation data from UI to \"transactionInitiation\" struct\n\terr = json.Unmarshal([]byte(args[1]), &user_obj)\n\tif err != nil {\n\t\tfmt.Printf(\"Unable to unmarshal createTransaction input transaction initiation : %s\\n\", err)\n\t\treturn nil, nil\n\t}\n\n\tfmt.Println(\"TransactionInitiation object refno variable value is : %s\\n\",trans_obj.TransRefNo);\n\t\n\tGetUserMap(stub)\t\n\n\tuser_map[user_obj.uname] = user_obj\t\n\n\tSetUserMap(stub)\t\n\t\n\tfmt.Printf(\"final user map : %v \\n\", user_map)\t\t\n\t\n\treturn nil, nil\n}", "func CreateUser(c *gin.Context) {\n\tlog.Println(\"CreateUser in db\")\n\tvar user models.User\n\tvar db = db.GetDB()\n\tif err := c.BindJSON(&user); err != nil {\n\t\tc.AbortWithStatusJSON(http.StatusBadRequest, gin.H{\n\t\t\t\"error\": err.Error(),\n\t\t})\n\t\tlog.Println(\"Failed to create user in db\")\n\t\treturn\n\t}\n\t// hash the password\n\tuser.Password = security.HashAndSalt([]byte(user.Password))\n\n\tdb.Create(&user)\n\tc.JSON(http.StatusOK, &user)\n}", "func (t *UserInstanceTable) Setup(ctx context.Context, db *sql.DB) error {\n\ttx, err := db.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = tx.Exec(`\n\tCREATE TABLE IF NOT EXISTS compute_personal (\n\t\trowid INTEGER PRIMARY KEY AUTOINCREMENT,\n owner_uid INT NOT NULL,\n\t instance_uid INT NOT NULL,\n ip VARCHAR(64) NOT NULL,\n\t\tstatus VARCHAR(64) NOT NULL,\n user_sshkey VARCHAR(2048) NOT NULL,\n\n\t\tCONSTRAINT fk_compute_instances\n\t\t\tFOREIGN KEY (instance_uid)\n\t\t\tREFERENCES compute_instances(rowid)\n\t\t\tON DELETE CASCADE\n\t);\n\n CREATE INDEX IF NOT EXISTS compute_personal_owner ON compute_personal(owner_uid);\n\tCREATE INDEX IF NOT EXISTS compute_personal_instance_uid ON compute_personal(instance_uid);\n\t`)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err = tx.Commit(); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func setUsuarioDBLF(newUser user) {\n\t_, err := database.Exec(\"call sp_insert_usuario_carga_masiva(:1,:2,:3,:4,:5,:6,:7)\",\n\t\tnewUser.UserName, newUser.Password, newUser.Nombre,\n\t\tnewUser.Apellido, newUser.FechaNacimiento, newUser.Email,\n\t\tnewUser.Membresia)\n\tif err != nil {\n\t\tfmt.Println(\"Error in Query:\", err)\n\t\treturn\n\t}\n}", "func (ug *userGorm) Create(user *User) error {\n\treturn ug.db.Create(user).Error\n}", "func (ug *userGorm) Create(user *User) error {\n\treturn ug.db.Create(user).Error\n}", "func (c *UserRepoImpl) Create(user *model.User) (*model.User, error) {\n\tif err := c.db.Table(\"user\").Save(&user).Error; err != nil {\n\t\tfmt.Errorf(\"Having error : %w\", err)\n\t\tlogrus.Error(err)\n\t\treturn nil, errors.New(\"add user data : error\")\n\t}\n\treturn user, nil\n}", "func createUser(u *models.User, db *sql.DB) error {\n\tif err := u.CryptPwd(); err != nil {\n\t\treturn fmt.Errorf(\"Cryptage du mot de passe de %s : %v\", u.Name, err)\n\t}\n\tif err := u.Create(db); err != nil {\n\t\treturn fmt.Errorf(\"Création en base de données de %s : %v\", u.Name, err)\n\t}\n\treturn nil\n}", "func (store *dbStore) CreateUser(user *structs.User) error {\r\n\tsqlStatement := fmt.Sprint(\"INSERT INTO user(username) VALUES ('\", user.Username, \"')\")\r\n\r\n\tfmt.Println(sqlStatement)\r\n\r\n\t_, err := store.db.Query(sqlStatement)\r\n\tif err != nil {\r\n\t\tfmt.Printf(\"failed to execute create user query on the database: %v\", err)\r\n\t\treturn err\r\n\t}\r\n\r\n\treturn nil\r\n}", "func TestUser(t *testing.T) {\n\tu := new(entity.User)\n\tu.Userid = \"1111\" //\n\thas, err := Engine.Get(u)\n\tt.Log(has, err)\n}", "func (dbHandler *DbHandler) CreateUser(user user.User) (err error) {\n\tpsw, _ := bcrypt.GenerateFromPassword([]byte(user.Password), bcrypt.DefaultCost)\n\tuser.Password = string(psw)\n\terr = dbHandler.handlers[UserCollection].GetCollection().Insert(user)\n\treturn\n}", "func User(w http.ResponseWriter, r *http.Request, db *mgo.Database) {\n\tuserID := chi.URLParam(r, \"userID\")\n\thelloWorld := \"Hello user \" + userID\n\trender.JSON(w, r, helloWorld)\n}" ]
[ "0.6224481", "0.58505857", "0.58417505", "0.5826436", "0.5734593", "0.5690672", "0.5690672", "0.568731", "0.5683977", "0.56832737", "0.5665492", "0.5661991", "0.56523687", "0.5643033", "0.5639015", "0.56369126", "0.56279415", "0.56095916", "0.56004083", "0.5600074", "0.5595484", "0.5565761", "0.5545583", "0.5542015", "0.5534423", "0.55320984", "0.5531834", "0.55242366", "0.5504431", "0.55030316", "0.5493047", "0.54877263", "0.54838043", "0.54757607", "0.5473926", "0.5470939", "0.5461291", "0.5439798", "0.54238564", "0.5420296", "0.5419646", "0.5419177", "0.5409692", "0.54083896", "0.54037637", "0.5392657", "0.5391155", "0.5381344", "0.53702915", "0.5360898", "0.5357747", "0.5354139", "0.5352433", "0.53491986", "0.53488064", "0.5346243", "0.5344441", "0.53364617", "0.53205574", "0.5318841", "0.53132886", "0.53132886", "0.5313153", "0.53084224", "0.53025484", "0.52941394", "0.52881664", "0.5280728", "0.527682", "0.52731633", "0.527029", "0.5266338", "0.5266307", "0.52503145", "0.5248703", "0.5242318", "0.5240256", "0.5240102", "0.5237818", "0.5235533", "0.5226176", "0.5224732", "0.52229136", "0.5217687", "0.52106", "0.5205586", "0.52050316", "0.5199569", "0.5198464", "0.51941717", "0.5191411", "0.51910585", "0.51837516", "0.51837516", "0.5183274", "0.51754415", "0.5171958", "0.51590437", "0.51574457", "0.5155708" ]
0.71175754
0
Role will inject the databaseTx in the `Role` schema
func (gtx *GuardTx) Role(role *schema.Role) *schema.Role { if role == nil { role = &schema.Role{ Entity: schema.Entity{DBContract: gtx.dbTx}, } } else { role.DBContract = gtx.dbTx } role.SetValidator(gtx.validator.Role) return role }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (j *AuroraJob) Role(role string) Job {\n\tj.jobConfig.Key.Role = role\n\n\t// Will be deprecated\n\tidentity := &aurora.Identity{User: role}\n\tj.jobConfig.Owner = identity\n\tj.jobConfig.TaskConfig.Owner = identity\n\treturn j\n}", "func RoleSeeder(db *gorm.DB) {\n\tdb.Exec(\"TRUNCATE TABLE roles\")\n\tdb.Create(&models.Role{Name: \"admin\"})\n\tdb.Create(&models.Role{Name: \"low\"})\n\tdb.Create(&models.Role{Name: \"middle\"})\n\tdb.Create(&models.Role{Name: \"high\"})\n}", "func (t *AuroraTask) Role(role string) *AuroraTask {\n\tt.task.Job.Role = role\n\treturn t\n}", "func (_Distributor *DistributorTransactor) GrantRole(opts *bind.TransactOpts, role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _Distributor.contract.Transact(opts, \"grantRole\", role, account)\n}", "func (_Distributor *DistributorTransactorSession) GrantRole(role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _Distributor.Contract.GrantRole(&_Distributor.TransactOpts, role, account)\n}", "func (_Distributor *DistributorSession) GrantRole(role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _Distributor.Contract.GrantRole(&_Distributor.TransactOpts, role, account)\n}", "func (_TellorMesosphere *TellorMesosphereTransactor) GrantRole(opts *bind.TransactOpts, role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _TellorMesosphere.contract.Transact(opts, \"grantRole\", role, account)\n}", "func (pge *PgEngine) SetRole(ctx context.Context, tx pgx.Tx, runUID pgtype.Varchar) {\n\tl := log.GetLogger(ctx)\n\tl.Info(\"Setting Role to \", runUID.String)\n\t_, err := tx.Exec(ctx, fmt.Sprintf(\"SET ROLE %v\", runUID.String))\n\tif err != nil {\n\t\tl.WithError(err).Error(\"Error in Setting role\", err)\n\t}\n}", "func DBCreateACLRoleTableTx(ctx context.Context, tx Tx) error {\n\tq := \"CREATE TABLE `acl_role` (`id`VARCHAR(64) NOT NULL PRIMARY KEY,`checksum`CHAR(64),`name` VARCHAR(100) NOT NULL,`description`TEXT,`admin_user_id` VARCHAR(64),`customer_id`VARCHAR(64),`created_at` BIGINT UNSIGNED NOT NULL,`updated_at` BIGINT UNSIGNED,INDEX acl_role_name_index (`name`),INDEX acl_role_customer_id_index (`customer_id`)) ENGINE=InnoDB DEFAULT CHARSET=utf8mb4 COLLATE=utf8mb4_unicode_ci;\"\n\t_, err := tx.ExecContext(ctx, q)\n\treturn err\n}", "func (gb *CurrentGrantBuilder) Role(n string) GrantExecutable {\n\treturn &CurrentGrantExecutable{\n\t\tgrantName: gb.qualifiedName,\n\t\tgrantType: gb.grantType,\n\t\tgranteeName: n,\n\t\tgranteeType: roleType,\n\t}\n}", "func (m RoleMutation) Tx() (*Tx, error) {\n\tif _, ok := m.driver.(*txDriver); !ok {\n\t\treturn nil, fmt.Errorf(\"ent: mutation is not running in a transaction\")\n\t}\n\ttx := &Tx{config: m.config}\n\ttx.init()\n\treturn tx, nil\n}", "func (m RoleMutation) Tx() (*Tx, error) {\n\tif _, ok := m.driver.(*txDriver); !ok {\n\t\treturn nil, fmt.Errorf(\"ent: mutation is not running in a transaction\")\n\t}\n\ttx := &Tx{config: m.config}\n\ttx.init()\n\treturn tx, nil\n}", "func (m RoleMutation) Tx() (*Tx, error) {\n\tif _, ok := m.driver.(*txDriver); !ok {\n\t\treturn nil, fmt.Errorf(\"ent: mutation is not running in a transaction\")\n\t}\n\ttx := &Tx{config: m.config}\n\ttx.init()\n\treturn tx, nil\n}", "func (m RoleMutation) Tx() (*Tx, error) {\n\tif _, ok := m.driver.(*txDriver); !ok {\n\t\treturn nil, fmt.Errorf(\"ent: mutation is not running in a transaction\")\n\t}\n\ttx := &Tx{config: m.config}\n\ttx.init()\n\treturn tx, nil\n}", "func (_TellorMesosphere *TellorMesosphereTransactorSession) GrantRole(role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _TellorMesosphere.Contract.GrantRole(&_TellorMesosphere.TransactOpts, role, account)\n}", "func (t *ACLRole) DBFindTx(ctx context.Context, tx Tx, _params ...interface{}) (bool, error) {\n\tparams := []interface{}{\n\t\torm.Column(\"id\"),\n\t\torm.Column(\"checksum\"),\n\t\torm.Column(\"name\"),\n\t\torm.Column(\"description\"),\n\t\torm.Column(\"admin_user_id\"),\n\t\torm.Column(\"customer_id\"),\n\t\torm.Column(\"created_at\"),\n\t\torm.Column(\"updated_at\"),\n\t\torm.Table(ACLRoleTableName),\n\t}\n\tif len(_params) > 0 {\n\t\tfor _, param := range _params {\n\t\t\tparams = append(params, param)\n\t\t}\n\t}\n\tq, p := orm.BuildQuery(params...)\n\trow := tx.QueryRowContext(ctx, q, p...)\n\tvar _ID sql.NullString\n\tvar _Checksum sql.NullString\n\tvar _Name sql.NullString\n\tvar _Description sql.NullString\n\tvar _AdminUserID sql.NullString\n\tvar _CustomerID sql.NullString\n\tvar _CreatedAt sql.NullInt64\n\tvar _UpdatedAt sql.NullInt64\n\terr := row.Scan(\n\t\t&_ID,\n\t\t&_Checksum,\n\t\t&_Name,\n\t\t&_Description,\n\t\t&_AdminUserID,\n\t\t&_CustomerID,\n\t\t&_CreatedAt,\n\t\t&_UpdatedAt,\n\t)\n\tif err != nil && err != sql.ErrNoRows {\n\t\treturn false, err\n\t}\n\tif _ID.Valid {\n\t\tt.SetID(_ID.String)\n\t}\n\tif _Checksum.Valid {\n\t\tt.SetChecksum(_Checksum.String)\n\t}\n\tif _Name.Valid {\n\t\tt.SetName(_Name.String)\n\t}\n\tif _Description.Valid {\n\t\tt.SetDescription(_Description.String)\n\t}\n\tif _AdminUserID.Valid {\n\t\tt.SetAdminUserID(_AdminUserID.String)\n\t}\n\tif _CustomerID.Valid {\n\t\tt.SetCustomerID(_CustomerID.String)\n\t}\n\tif _CreatedAt.Valid {\n\t\tt.SetCreatedAt(_CreatedAt.Int64)\n\t}\n\tif _UpdatedAt.Valid {\n\t\tt.SetUpdatedAt(_UpdatedAt.Int64)\n\t}\n\treturn true, nil\n}", "func withRole(node *Role) roleOption {\n\treturn func(m *RoleMutation) {\n\t\tm.oldValue = func(context.Context) (*Role, error) {\n\t\t\treturn node, nil\n\t\t}\n\t\tm.id = &node.ID\n\t}\n}", "func withRole(node *Role) roleOption {\n\treturn func(m *RoleMutation) {\n\t\tm.oldValue = func(context.Context) (*Role, error) {\n\t\t\treturn node, nil\n\t\t}\n\t\tm.id = &node.ID\n\t}\n}", "func withRole(node *Role) roleOption {\n\treturn func(m *RoleMutation) {\n\t\tm.oldValue = func(context.Context) (*Role, error) {\n\t\t\treturn node, nil\n\t\t}\n\t\tm.id = &node.ID\n\t}\n}", "func (_TellorMesosphere *TellorMesosphereSession) GrantRole(role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _TellorMesosphere.Contract.GrantRole(&_TellorMesosphere.TransactOpts, role, account)\n}", "func (c *configuration) Role(clientSet ClientSet) *Role {\n\tif clientSet != nil {\n\t\treturn NewRole(clientSet)\n\t}\n\treturn nil\n\n}", "func setRaftRole(t *testing.T, database *db.Node, address string) client.NodeStore {\n\trequire.NoError(t, database.Transaction(context.Background(), func(ctx context.Context, tx *db.NodeTx) error {\n\t\terr := tx.UpdateConfig(map[string]string{\"cluster.https_address\": address})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = tx.CreateRaftNode(address, \"test\")\n\t\treturn err\n\t}))\n\n\tstore := client.NewNodeStore(database.DB(), \"main\", \"raft_nodes\", \"address\")\n\treturn store\n}", "func (_DelegationController *DelegationControllerTransactorSession) GrantRole(role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _DelegationController.Contract.GrantRole(&_DelegationController.TransactOpts, role, account)\n}", "func (_DelegationController *DelegationControllerTransactor) GrantRole(opts *bind.TransactOpts, role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _DelegationController.contract.Transact(opts, \"grantRole\", role, account)\n}", "func (mRoleRepo *MockRoleRepo) StoreRole(role *entity.Role) (*entity.Role, []error) {\n\tRol := role\n\treturn Rol, nil\n}", "func TestRole(t *testing.T) {\n\t// Initializing variables\n\tvar (\n\t\terr error\n\t\ttestDatastore *datastores.ConcreteDatastore\n\t)\n\n\tif testDatastore, err = datastores.NewDatabase(\"myTestDatabase.db\"); err != nil {\n\t\tt.Error(err)\n\t}\n\n\trole1 := model.Role{\n\t\tRoleId: 0,\n\t\tRoleName: \"Role 1\",\n\t\tCanAddAndModifyUsers: true,\n\t\tCanSeeOtherSchedules: true,\n\t\tCanAddProjects: true,\n\t\tCanSeeReports: true,\n\t}\n\n\trole2 := model.Role{\n\t\tRoleId: 0,\n\t\tRoleName: \"Role 2\",\n\t\tCanAddAndModifyUsers: false,\n\t\tCanSeeOtherSchedules: false,\n\t\tCanAddProjects: true,\n\t\tCanSeeReports: true,\n\t}\n\n\trole3 := model.Role{\n\t\tRoleId: 0,\n\t\tRoleName: \"Role 3\",\n\t\tCanAddAndModifyUsers: false,\n\t\tCanSeeOtherSchedules: false,\n\t\tCanAddProjects: false,\n\t\tCanSeeReports: false,\n\t}\n\n\t//\n\t//\tCreateRoles()\n\t//\n\n\tif role1.RoleId, err = testDatastore.CreateRole(role1); err != nil {\n\t\tt.Error(err)\n\t}\n\n\tif role2.RoleId, err = testDatastore.CreateRole(role2); err != nil {\n\t\tt.Error(err)\n\t}\n\n\tif role3.RoleId, err = testDatastore.CreateRole(role3); err != nil {\n\t\tt.Error(err)\n\t}\n\n\tglobals.Log.Debug(\"CreateRole test - PASSED\")\n\n\t//\n\t// Test GetRoles()\n\t//\n\tvar (\n\t\tallRoles model.Roles\n\t\tdefaultRole1 model.Role\n\t\tdefaultRole2 model.Role\n\t\tdefaultRole3 model.Role\n\t)\n\n\t// Fetching the default roles\n\tif defaultRole1, err = testDatastore.GetRole(1); err != nil {\n\t\tt.Error(err)\n\t}\n\n\tif defaultRole2, err = testDatastore.GetRole(2); err != nil {\n\t\tt.Error(err)\n\t}\n\n\tif defaultRole3, err = testDatastore.GetRole(3); err != nil {\n\t\tt.Error(err)\n\t}\n\n\t// Formatting the data\n\troleList := model.Roles{}\n\troleList = append(roleList, defaultRole1)\n\troleList = append(roleList, defaultRole2)\n\troleList = append(roleList, defaultRole3)\n\troleList = append(roleList, role1)\n\troleList = append(roleList, role2)\n\troleList = append(roleList, role3)\n\n\t// Fetching all roles\n\tif allRoles, err = testDatastore.GetRoles(); err != nil {\n\t\tt.Error(err)\n\t}\n\n\t// Verigying the result\n\tif !cmp.Equal(allRoles, roleList) {\n\t\tt.Error(err)\n\t}\n\n\tglobals.Log.Debug(\"GetRoles test - PASSED\")\n\n\t//\n\t// Test GetRole(RoleId)\n\t//\n\tvar role model.Role\n\n\t// Fetching a role\n\tif role, err = testDatastore.GetRole(role1.RoleId); err != nil {\n\t\tt.Error(err)\n\t}\n\n\t// Verifying the data\n\tif !cmp.Equal(role1, role) {\n\t\tt.Error(err)\n\t}\n\n\tglobals.Log.Debug(\"GetRole test - PASSED\")\n\n\t//\n\t// Test GetRoleOfUser\n\t//\n\n\t// Creating some data\n\tcontract := model.Contract{\n\t\tContractId: 0,\n\t\tContractName: \"Contract\",\n\t}\n\n\tif contract.ContractId, err = testDatastore.CreateContract(contract); err != nil {\n\t\tt.Error(err)\n\t}\n\n\tuser := model.User{\n\t\tUserId: 0,\n\t\tContractId: contract.ContractId,\n\t\tRoleId: role1.RoleId,\n\t\tUsername: \"First User\",\n\t\tPassword: \"This is a password\",\n\t\tLastName: \"User\",\n\t\tFirstName: \"First\",\n\t\tMail: \"[email protected]\",\n\t\tTheoricalHoursWorked: 50,\n\t\tVacationHours: 50,\n\t}\n\n\tif user.UserId, err = testDatastore.CreateUser(user); err != nil {\n\t\tt.Error(err)\n\t}\n\n\t// Fetching the role of the user\n\tvar dbRoleOfUser model.Role\n\tif dbRoleOfUser, err = testDatastore.GetRoleOfUser(user.UserId); err != nil {\n\t\tt.Error(err)\n\t}\n\n\t// Verifying the data\n\tif !cmp.Equal(dbRoleOfUser, role1) {\n\t\tt.Error(err)\n\t}\n\n\tglobals.Log.Debug(\"GetRoleOfUser test - PASSED\")\n\n\t//\n\t// Test UpdateRole(Role)\n\t//\n\tvar updatedRole model.Role\n\n\t// Modifying the role\n\trole1.RoleName = \"New role name\"\n\trole1.CanAddAndModifyUsers = false\n\trole1.CanSeeOtherSchedules = false\n\trole1.CanAddProjects = false\n\trole1.CanSeeReports = false\n\n\t// Saving the changes\n\tif _, err = testDatastore.UpdateRole(role1); err != nil {\n\t\tt.Error(err)\n\t}\n\n\t// Getting the role so we can check the changes\n\tif updatedRole, err = testDatastore.GetRole(role1.RoleId); err != nil {\n\t\tt.Error(err)\n\t}\n\n\t// Checking changes\n\tif !cmp.Equal(role1, updatedRole) {\n\t\tt.Error(err)\n\t}\n\n\tglobals.Log.Debug(\"UpdateRole test - PASSED\")\n\n\t//\n\t// Test DeleteRole(RoleId)\n\t//\n\n\t// Creating a role so we can delete it\n\tvar ind int64\n\tif ind, err = testDatastore.CreateRole(model.Role{\n\t\tRoleId: 0,\n\t\tRoleName: \"This is a role\",\n\t}); err != nil {\n\t\tt.Error(err)\n\t}\n\n\t// Deleting it\n\tif err = testDatastore.DeleteRole(ind); err != nil {\n\t\tt.Error(err)\n\t}\n\n\t// trying to get the role we just deleted\n\tif _, err = testDatastore.GetRole(ind); err == nil {\n\t\tt.Error()\n\t}\n\n\tglobals.Log.Debug(\"DeleteRole test - PASSED\")\n\n\ttestDatastore.CloseDatabase()\n}", "func InsertRole(db *sql.DB, name, intro string) error {\n\tresult, err := db.Exec(roleSQLString[mysqlRoleInsert], name, intro, true)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif rows, _ := result.RowsAffected(); rows == 0 {\n\t\treturn errInvalidMysql\n\t}\n\n\treturn nil\n}", "func (t *ACLRole) DBCreateTx(ctx context.Context, tx Tx) (sql.Result, error) {\n\tq := \"INSERT INTO `acl_role` (`acl_role`.`id`,`acl_role`.`checksum`,`acl_role`.`name`,`acl_role`.`description`,`acl_role`.`admin_user_id`,`acl_role`.`customer_id`,`acl_role`.`created_at`,`acl_role`.`updated_at`) VALUES (?,?,?,?,?,?,?,?)\"\n\tchecksum := t.CalculateChecksum()\n\tif t.GetChecksum() == checksum {\n\t\treturn nil, nil\n\t}\n\tt.Checksum = &checksum\n\treturn tx.ExecContext(ctx, q,\n\t\torm.ToSQLString(t.ID),\n\t\torm.ToSQLString(t.Checksum),\n\t\torm.ToSQLString(t.Name),\n\t\torm.ToSQLString(t.Description),\n\t\torm.ToSQLString(t.AdminUserID),\n\t\torm.ToSQLString(t.CustomerID),\n\t\torm.ToSQLInt64(t.CreatedAt),\n\t\torm.ToSQLInt64(t.UpdatedAt),\n\t)\n}", "func (c *Controller) reconcileMySQLRole(dbRClient database.DatabaseRoleInterface, myRole *api.MySQLRole) error {\n\tstatus := myRole.Status\n\t// enable the database secrets engine if it is not already enabled\n\terr := dbRClient.EnableDatabase()\n\tif err != nil {\n\t\tstatus.Conditions = []api.MySQLRoleCondition{\n\t\t\t{\n\t\t\t\tType: \"Available\",\n\t\t\t\tStatus: corev1.ConditionFalse,\n\t\t\t\tReason: \"FailedToEnableDatabase\",\n\t\t\t\tMessage: err.Error(),\n\t\t\t},\n\t\t}\n\n\t\terr2 := c.updatedMySQLRoleStatus(&status, myRole)\n\t\tif err2 != nil {\n\t\t\treturn errors.Wrap(err2, \"failed to update status\")\n\t\t}\n\t\treturn errors.Wrap(err, \"failed to enable database secret engine\")\n\t}\n\n\t// create database config for mysql\n\terr = dbRClient.CreateConfig()\n\tif err != nil {\n\t\tstatus.Conditions = []api.MySQLRoleCondition{\n\t\t\t{\n\t\t\t\tType: \"Available\",\n\t\t\t\tStatus: corev1.ConditionFalse,\n\t\t\t\tReason: \"FailedToCreateDatabaseConfig\",\n\t\t\t\tMessage: err.Error(),\n\t\t\t},\n\t\t}\n\n\t\terr2 := c.updatedMySQLRoleStatus(&status, myRole)\n\t\tif err2 != nil {\n\t\t\treturn errors.Wrap(err2, \"failed to update status\")\n\t\t}\n\t\treturn errors.Wrap(err, \"failed to create database connection config\")\n\t}\n\n\t// create role\n\terr = dbRClient.CreateRole()\n\tif err != nil {\n\t\tstatus.Conditions = []api.MySQLRoleCondition{\n\t\t\t{\n\t\t\t\tType: \"Available\",\n\t\t\t\tStatus: corev1.ConditionFalse,\n\t\t\t\tReason: \"FailedToCreateRole\",\n\t\t\t\tMessage: err.Error(),\n\t\t\t},\n\t\t}\n\n\t\terr2 := c.updatedMySQLRoleStatus(&status, myRole)\n\t\tif err2 != nil {\n\t\t\treturn errors.Wrap(err2, \"failed to update status\")\n\t\t}\n\t\treturn errors.Wrap(err, \"failed to create role\")\n\t}\n\n\tstatus.Conditions = []api.MySQLRoleCondition{}\n\tstatus.Phase = MySQLRolePhaseSuccess\n\tstatus.ObservedGeneration = types.NewIntHash(myRole.Generation, meta_util.GenerationHash(myRole))\n\n\terr = c.updatedMySQLRoleStatus(&status, myRole)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to update MySQLRole status\")\n\t}\n\treturn nil\n}", "func ProvideRoleRepo(engine *core.Engine) RoleRepo {\n\treturn RoleRepo{Engine: engine}\n}", "func createDatabaseRoles(ctx context.Context, client kubernetes.Interface, f flags) error {\n\tif err := createServiceAccount(ctx, client, appdbServiceAccount, f.memberClusterNamespace); err != nil {\n\t\treturn err\n\t}\n\tif err := createServiceAccount(ctx, client, databasePodsServiceAccount, f.memberClusterNamespace); err != nil {\n\t\treturn err\n\t}\n\tif err := createServiceAccount(ctx, client, opsManagerServiceAccount, f.memberClusterNamespace); err != nil {\n\t\treturn err\n\t}\n\tif err := createDatabaseRole(ctx, client, appdbRole, f.memberClusterNamespace); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (_AccessControl *AccessControlTransactor) GrantRole(opts *bind.TransactOpts, role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _AccessControl.contract.Transact(opts, \"grantRole\", role, account)\n}", "func (*TeleportRoleMarshaler) MarshalRole(u Role, opts ...MarshalOption) ([]byte, error) {\n\treturn json.Marshal(u)\n}", "func setupDatabaseRoles(ctx context.Context, clientSet map[string]kubernetes.Interface, f flags) error {\n\tfor _, clusterName := range f.memberClusters {\n\t\tif clusterName != f.sourceCluster {\n\t\t\tif err := copyDatabaseRoles(ctx, clientSet[f.sourceCluster], clientSet[clusterName], f.memberClusterNamespace); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (_AccessControl *AccessControlTransactorSession) GrantRole(role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _AccessControl.Contract.GrantRole(&_AccessControl.TransactOpts, role, account)\n}", "func (_DelegationController *DelegationControllerSession) GrantRole(role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _DelegationController.Contract.GrantRole(&_DelegationController.TransactOpts, role, account)\n}", "func (pge *PgEngine) ResetRole(ctx context.Context, tx pgx.Tx) {\n\tl := log.GetLogger(ctx)\n\tl.Info(\"Resetting Role\")\n\tconst sqlResetRole = `RESET ROLE`\n\t_, err := tx.Exec(ctx, sqlResetRole)\n\tif err != nil {\n\t\tl.WithError(err).Error(\"Error in ReSetting role\", err)\n\t}\n}", "func Role() helmify.Processor {\n\treturn &role{}\n}", "func (b *GroupsEditManagerBuilder) Role(v string) *GroupsEditManagerBuilder {\n\tb.Params[\"role\"] = v\n\treturn b\n}", "func RoleCreate(user_id int64, level_id uint8) (sql.Result, error) {\n\tres, err := database.DB.Exec(\"INSERT INTO role (user_id, level_id) VALUES (?,?)\", user_id, level_id)\n\treturn res, err\n}", "func (d *Dao) Role(c context.Context, user, token string) (nodes *model.CacheData, err error) {\n\tvar (\n\t\treq *http.Request\n\t)\n\tif req, err = http.NewRequest(\"GET\", appsURI, nil); err != nil {\n\t\tlog.Error(\"Status url(%s) error(%v)\", dataURI, err)\n\t\treturn\n\t}\n\treq.Header.Set(\"X-Authorization-Token\", token)\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tvar res struct {\n\t\tCode int `json:\"code\"`\n\t\tData []*model.RoleNode `json:\"data\"`\n\t\tMessage string `json:\"message\"`\n\t\tStatus int `json:\"status\"`\n\t}\n\tif err = d.client.Do(c, req, &res); err != nil {\n\t\tlog.Error(\"d.Status url(%s) res($s) err(%v)\", dataURI, res, err)\n\t\treturn\n\t}\n\tif res.Code != 90000 {\n\t\terr = fmt.Errorf(\"error code :%d\", res.Code)\n\t\tlog.Error(\"Status url(%s) res(%v)\", dataURI, res)\n\t\treturn\n\t}\n\tnodes = &model.CacheData{Data: make(map[int64]*model.RoleNode)}\n\tnodes.CTime = time.Now()\n\tfor _, node := range res.Data {\n\t\tif bytes.Equal(prefix, []byte(node.Path)[0:9]) {\n\t\t\tnode.Path = string([]byte(node.Path)[9:])\n\t\t}\n\t\tnodes.Data[node.ID] = node\n\t}\n\treturn\n}", "func (b *BootstrapClient) mountPostgresRole() error {\n\treturn b.usingVaultRootToken(func() error {\n\t\tmountInfo := &api.MountInput{\n\t\t\tType: SecretBackendPostgres,\n\t\t}\n\t\tif err := b.VaultClient.Sys().Mount(b.config.postgresPolicyMountPath, mountInfo); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// set connection data used by vault to connect to postgres\n\t\tconnectionData := map[string]interface{}{\n\t\t\t\"connection_url\": b.config.postgresRootConnURL,\n\t\t}\n\t\tif _, err := b.VaultClient.Logical().Write(b.config.postgresPolicyRoleConnPath, connectionData); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// create lease configuration\n\t\tleaseData := map[string]interface{}{\n\t\t\t\"lease\": b.config.postgresPolicyRoleLeaseTTL,\n\t\t\t\"lease_max\": b.config.postgresPolicyRoleLeaseMaxTTL,\n\t\t}\n\t\tif _, err := b.VaultClient.Logical().Write(b.config.postgresPolicyRoleLeasePath, leaseData); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// create pg configuration for creating roles\n\t\troleData := map[string]interface{}{\n\t\t\t\"sql\": b.config.postgresPolicyRoleCreateSQL,\n\t\t}\n\t\tif _, err := b.VaultClient.Logical().Write(b.config.postgresPolicyRoleCreatePath, roleData); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t})\n}", "func (store *Store) AddRole(role *graylog.Role) error {\n\tstore.imutex.Lock()\n\tdefer store.imutex.Unlock()\n\tstore.roles[role.Name] = *role\n\treturn nil\n}", "func main() {\n\tdb, err := db.OpenConnection()\n\tdb.LogMode(true)\n\n\tif err != nil {\n\t\tlog.WithError(err).Fatal(\"An error occurred while connecting to the database\")\n\t}\n\tdefer db.Close()\n\n\t// Create models\n\tdb.DropTableIfExists(&domain.User{}, &domain.Email{}, &domain.Role{})\n\tdb.CreateTable(&domain.Email{}, &domain.Role{}, &domain.User{})\n\n\t// Create user\n\tid, _ := uuid.Parse(\"8c5df3bc-9fa6-4d73-b79a-9a1cbb35740c\")\n\tuser := domain.User{\n\t\tID: id,\n\t\tEmails: []domain.Email{\n\t\t\t{Email: \"[email protected]\"},\n\t\t},\n\t\tRoles: []domain.Role{\n\t\t\t{Role: \"ROLE_USER\"},\n\t\t},\n\t\tPassword: \"Test1234\",\n\t}\n\n\t// Create developer\n\tid, _ = uuid.Parse(\"b0203081-5dfe-4bb7-87d1-e2c59e2af7b6\")\n\tdeveloper := domain.User{\n\t\tID: id,\n\t\tEmails: []domain.Email{\n\t\t\t{Email: \"[email protected]\"},\n\t\t},\n\t\tRoles: []domain.Role{\n\t\t\t{Role: \"ROLE_USER\"},\n\t\t\t{Role: \"ROLE_DEVELOPER\"},\n\t\t},\n\t\tPassword: \"Test1234\",\n\t}\n\n\t// Create hr\n\tid, _ = uuid.Parse(\"6b59c645-82c3-4e08-b089-f4236a2141b6\")\n\thr := domain.User{\n\t\tID: id,\n\t\tEmails: []domain.Email{\n\t\t\t{Email: \"[email protected]\"},\n\t\t},\n\t\tRoles: []domain.Role{\n\t\t\t{Role: \"ROLE_USER\"},\n\t\t\t{Role: \"ROLE_HR\"},\n\t\t},\n\t\tPassword: \"Test1234\",\n\t}\n\n\t// Create manager\n\tid, _ = uuid.Parse(\"75129bb5-5c12-48a1-8410-bb2630fff9ed\")\n\tmanager := domain.User{\n\t\tID: id,\n\t\tEmails: []domain.Email{\n\t\t\t{Email: \"[email protected]\"},\n\t\t\t{Email: \"[email protected]\"},\n\t\t},\n\t\tRoles: []domain.Role{\n\t\t\t{Role: \"ROLE_USER\"},\n\t\t\t{Role: \"ROLE_MANAGER\"},\n\t\t},\n\t\tPassword: \"Test1234\",\n\t}\n\n\tdb.Create(&user)\n\tdb.Create(&developer)\n\tdb.Create(&hr)\n\tdb.Create(&manager)\n}", "func (cli *Service) CreateRolePri(req *restful.Request, resp *restful.Response) {\n\n\tlanguage := util.GetActionLanguage(req)\n\townerID := util.GetOwnerID(req.Request.Header)\n\tdefErr := cli.Core.CCErr.CreateDefaultCCErrorIf(language)\n\tctx := util.GetDBContext(context.Background(), req.Request.Header)\n\tdb := cli.Instance.Clone()\n\n\tpathParams := req.PathParameters()\n\tobjID := pathParams[\"bk_obj_id\"]\n\tpropertyID := pathParams[\"bk_property_id\"]\n\tvalue, err := ioutil.ReadAll(req.Request.Body)\n\tif err != nil {\n\t\tblog.Error(\"read json data error :%v\", err)\n\t\tresp.WriteError(http.StatusBadRequest, &meta.RespError{Msg: defErr.New(common.CCErrCommHTTPReadBodyFailed, err.Error())})\n\t\treturn\n\t}\n\tvar roleJSON []string\n\terr = json.Unmarshal([]byte(value), &roleJSON)\n\tif err != nil {\n\t\tblog.Error(\"read json data error :%v\", err)\n\t\tresp.WriteError(http.StatusBadRequest, &meta.RespError{Msg: defErr.New(common.CCErrCommJSONUnmarshalFailed, err.Error())})\n\t\treturn\n\t}\n\tinput := make(map[string]interface{})\n\tinput[common.BKOwnerIDField] = ownerID\n\tinput[common.BKObjIDField] = objID\n\tinput[common.BKPropertyIDField] = propertyID\n\tinput[common.BKPrivilegeField] = roleJSON\n\tinput = util.SetModOwner(input, ownerID)\n\n\terr = db.Table(common.BKTableNamePrivilege).Insert(ctx, input)\n\tif nil != err {\n\t\tblog.Error(\"create role privilege error :%v\", err)\n\t\tresp.WriteError(http.StatusBadRequest, &meta.RespError{Msg: defErr.New(common.CCErrObjectDBOpErrno, err.Error())})\n\t\treturn\n\t}\n\n\tresp.WriteEntity(meta.Response{BaseResp: meta.SuccessBaseResp})\n}", "func (treeNode *TreeNode) insertRole(role Role) {\n\tif treeNode.Role.Id == role.Parent {\n\t\tnewNode := NewTreeNode(role)\n\t\ttreeNode.Subordinates = append(treeNode.Subordinates, newNode)\n\t} else {\n\t\tfor idx := range treeNode.Subordinates {\n\t\t\tsubordinate := treeNode.Subordinates[idx]\n\t\t\tsubordinate.insertRole(role)\n\t\t}\n\t}\n}", "func (conf *ConfigType) Role() Role {\n\treturn conf.role\n}", "func bindMinterRole(address common.Address, caller bind.ContractCaller, transactor bind.ContractTransactor, filterer bind.ContractFilterer) (*bind.BoundContract, error) {\n\tparsed, err := ParsedABI(K_MinterRole)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn bind.NewBoundContract(address, *parsed, caller, transactor, filterer), nil\n}", "func (_AccessControl *AccessControlSession) GrantRole(role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _AccessControl.Contract.GrantRole(&_AccessControl.TransactOpts, role, account)\n}", "func (_PermInterface *PermInterfaceTransactor) AssignAccountRole(opts *bind.TransactOpts, _account common.Address, _orgId string, _roleId string) (*types.Transaction, error) {\n\treturn _PermInterface.contract.Transact(opts, \"assignAccountRole\", _account, _orgId, _roleId)\n}", "func AllRole(cfg *fileConfig) *entity {\n\tfor _, u := range cfg.Role {\n\t\tif u.Get && u.Put && u.Del {\n\t\t\treturn u\n\t\t}\n\t}\n\treturn nil\n}", "func (n *Node) Role() string {\n\t// use the cached version populated by NewNode\n\treturn n.role\n}", "func (m *UserMutation) AddRole(i int) {\n\tif m.addrole != nil {\n\t\t*m.addrole += i\n\t} else {\n\t\tm.addrole = &i\n\t}\n}", "func (epc *EntryPointCreate) SetRole(fpr flowschema.EntryPointRole) *EntryPointCreate {\n\tepc.mutation.SetRole(fpr)\n\treturn epc\n}", "func CreateUserRole(w http.ResponseWriter, r *http.Request) {\n\tfLog := userMgmtLogger.WithField(\"func\", \"CreateUserRole\").WithField(\"RequestID\", r.Context().Value(constants.RequestID)).WithField(\"path\", r.URL.Path).WithField(\"method\", r.Method)\n\n\tiauthctx := r.Context().Value(constants.HansipAuthentication)\n\tif iauthctx == nil {\n\t\thelper.WriteHTTPResponse(r.Context(), w, http.StatusUnauthorized, \"You are not authorized to access this resource\", nil, nil)\n\t\treturn\n\t}\n\n\tparams, err := helper.ParsePathParams(fmt.Sprintf(\"%s/management/user/{userRecId}/role/{roleRecId}\", apiPrefix), r.URL.Path)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\trole, err := RoleRepo.GetRoleByRecID(r.Context(), params[\"roleRecId\"])\n\tif err != nil {\n\t\tfLog.Errorf(\"RoleRepo.GetRoleByRecID got %s\", err.Error())\n\t\thelper.WriteHTTPResponse(r.Context(), w, http.StatusInternalServerError, err.Error(), nil, nil)\n\t\treturn\n\t}\n\tif role == nil {\n\t\thelper.WriteHTTPResponse(r.Context(), w, http.StatusNotFound, fmt.Sprintf(\"Role recid %s not found\", params[\"roleRecId\"]), nil, nil)\n\t\treturn\n\t}\n\n\tauthCtx := iauthctx.(*hansipcontext.AuthenticationContext)\n\tif !authCtx.IsAdminOfDomain(role.RoleDomain) {\n\t\thelper.WriteHTTPResponse(r.Context(), w, http.StatusForbidden, \"You don't have the right to access role with the specified domain\", nil, nil)\n\t\treturn\n\t}\n\n\tuser, err := UserRepo.GetUserByRecID(r.Context(), params[\"userRecId\"])\n\tif err != nil {\n\t\tfLog.Errorf(\"UserRepo.GetUserByRecID got %s\", err.Error())\n\t\thelper.WriteHTTPResponse(r.Context(), w, http.StatusInternalServerError, err.Error(), nil, nil)\n\t\treturn\n\t}\n\tif user == nil {\n\t\thelper.WriteHTTPResponse(r.Context(), w, http.StatusNotFound, fmt.Sprintf(\"User recid %s not found\", params[\"userRecId\"]), nil, nil)\n\t\treturn\n\t}\n\n\t_, err = UserRoleRepo.CreateUserRole(r.Context(), user, role)\n\tif err != nil {\n\t\tfLog.Errorf(\"UserRoleRepo.CreateUserRole got %s\", err.Error())\n\t\thelper.WriteHTTPResponse(r.Context(), w, http.StatusBadRequest, err.Error(), nil, nil)\n\t\treturn\n\t}\n\tRevocationRepo.Revoke(r.Context(), user.Email)\n\thelper.WriteHTTPResponse(r.Context(), w, http.StatusOK, \"User-Role created\", nil, nil)\n}", "func (m *RoleMutation) Role() (r string, exists bool) {\n\tv := m._Role\n\tif v == nil {\n\t\treturn\n\t}\n\treturn *v, true\n}", "func CreateRole(c RoleCreated) (int, error) {\n\tq := `INSERT INTO roles(name, parent_id) \n\t\t VALUES($1, $2) RETURNING id`\n\tdb := GetConnection()\n\n\tdefer db.Close()\n\tvar id int = 0\n\terr := db.QueryRow(q, c.Name, c.ParentId).Scan(&id)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn id, nil\n}", "func (m *UserMutation) Role() (r int, exists bool) {\n\tv := m.role\n\tif v == nil {\n\t\treturn\n\t}\n\treturn *v, true\n}", "func (transactionRepo *mockTransactionRepo) Initialize(ctx context.Context, db *sql.DB) {}", "func (c *VaultController) reconcileMySQLRole(rClient database.DatabaseRoleInterface, role *api.MySQLRole) error {\n\t// create role\n\terr := rClient.CreateRole()\n\tif err != nil {\n\t\t_, err2 := patchutil.UpdateMySQLRoleStatus(\n\t\t\tcontext.TODO(),\n\t\t\tc.extClient.EngineV1alpha1(),\n\t\t\trole.ObjectMeta,\n\t\t\tfunc(status *api.MySQLRoleStatus) *api.MySQLRoleStatus {\n\t\t\t\tstatus.Conditions = kmapi.SetCondition(status.Conditions, kmapi.Condition{\n\t\t\t\t\tType: kmapi.ConditionFailed,\n\t\t\t\t\tStatus: core.ConditionTrue,\n\t\t\t\t\tReason: \"FailedToCreateRole\",\n\t\t\t\t\tMessage: err.Error(),\n\t\t\t\t})\n\t\t\t\treturn status\n\t\t\t}, metav1.UpdateOptions{},\n\t\t)\n\t\treturn utilerrors.NewAggregate([]error{err2, errors.Wrap(err, \"failed to create role\")})\n\t}\n\n\t_, err = patchutil.UpdateMySQLRoleStatus(\n\t\tcontext.TODO(),\n\t\tc.extClient.EngineV1alpha1(),\n\t\trole.ObjectMeta,\n\t\tfunc(status *api.MySQLRoleStatus) *api.MySQLRoleStatus {\n\t\t\tstatus.Phase = MySQLRolePhaseSuccess\n\t\t\tstatus.ObservedGeneration = role.Generation\n\t\t\tstatus.Conditions = kmapi.RemoveCondition(status.Conditions, kmapi.ConditionFailed)\n\t\t\tstatus.Conditions = kmapi.SetCondition(status.Conditions, kmapi.Condition{\n\t\t\t\tType: kmapi.ConditionAvailable,\n\t\t\t\tStatus: core.ConditionTrue,\n\t\t\t\tReason: \"Provisioned\",\n\t\t\t\tMessage: \"role is ready to use\",\n\t\t\t})\n\t\t\treturn status\n\t\t}, metav1.UpdateOptions{},\n\t)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tglog.Infof(\"Successfully processed MySQLRole: %s/%s\", role.Namespace, role.Name)\n\treturn nil\n}", "func (_Roles *RolesRaw) Transfer(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Roles.Contract.RolesTransactor.contract.Transfer(opts)\n}", "func (s *BasePlSqlParserListener) EnterRole_clause(ctx *Role_clauseContext) {}", "func (_PermInterface *PermInterfaceTransactorSession) AssignAdminRole(_orgId string, _account common.Address, _roleId string) (*types.Transaction, error) {\n\treturn _PermInterface.Contract.AssignAdminRole(&_PermInterface.TransactOpts, _orgId, _account, _roleId)\n}", "func (s *StateStore) upsertACLRoleTxn(\n\tindex uint64, txn *txn, role *structs.ACLRole, allowMissingPolicies bool) (bool, error) {\n\n\t// Ensure the role hash is not zero to provide defense in depth. This\n\t// should be done outside the state store, so we do not spend time here\n\t// and thus Raft, when it, can be avoided.\n\tif len(role.Hash) == 0 {\n\t\trole.SetHash()\n\t}\n\n\t// This validation also happens within the RPC handler, but Raft latency\n\t// could mean that by the time the state call is invoked, another Raft\n\t// update has deleted policies detailed in role. Therefore, check again\n\t// while in our write txn.\n\tif !allowMissingPolicies {\n\t\tif err := s.validateACLRolePolicyLinksTxn(txn, role); err != nil {\n\t\t\treturn false, err\n\t\t}\n\t}\n\n\t// This validation also happens within the RPC handler, but Raft latency\n\t// could mean that by the time the state call is invoked, another Raft\n\t// update has already written a role with the same name. We therefore need\n\t// to check we are not trying to create a role with an existing name.\n\texistingRaw, err := txn.First(TableACLRoles, indexName, role.Name)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"ACL role lookup failed: %v\", err)\n\t}\n\n\t// Track our type asserted role, so we only need to do this once.\n\tvar existing *structs.ACLRole\n\n\t// If we did not find an ACL Role within state with the same name, we need\n\t// to check using the ID index as the operator might be performing an\n\t// update on the role name.\n\t//\n\t// If we found an entry using the name index, we need to check that the ID\n\t// matches the object within the request.\n\tif existingRaw == nil {\n\t\texistingRaw, err = txn.First(TableACLRoles, indexID, role.ID)\n\t\tif err != nil {\n\t\t\treturn false, fmt.Errorf(\"ACL role lookup failed: %v\", err)\n\t\t}\n\t\tif existingRaw != nil {\n\t\t\texisting = existingRaw.(*structs.ACLRole)\n\t\t}\n\t} else {\n\t\texisting = existingRaw.(*structs.ACLRole)\n\t\tif existing.ID != role.ID {\n\t\t\treturn false, fmt.Errorf(\"ACL role with name %s already exists\", role.Name)\n\t\t}\n\t}\n\n\t// Depending on whether this is an initial create, or an update, we need to\n\t// check and set certain parameters. The most important is to ensure any\n\t// create index is carried over.\n\tif existing != nil {\n\n\t\t// If the role already exists, check whether the update contains any\n\t\t// difference. If it doesn't, we can avoid a state update as wel as\n\t\t// updates to any blocking queries.\n\t\tif existing.Equals(role) {\n\t\t\treturn false, nil\n\t\t}\n\n\t\trole.CreateIndex = existing.CreateIndex\n\t\trole.ModifyIndex = index\n\t} else {\n\t\trole.CreateIndex = index\n\t\trole.ModifyIndex = index\n\t}\n\n\t// Insert the role into the table.\n\tif err := txn.Insert(TableACLRoles, role); err != nil {\n\t\treturn false, fmt.Errorf(\"ACL role insert failed: %v\", err)\n\t}\n\treturn true, nil\n}", "func (rg *Registry) RegisterRole(role Role, rights ...Right) {\n\trg.roles[role] = rights\n}", "func SetupRole(mgr ctrl.Manager, o controller.Options) error {\n\tname := managed.ControllerName(v1beta1.RoleGroupKind)\n\n\tcps := []managed.ConnectionPublisher{managed.NewAPISecretPublisher(mgr.GetClient(), mgr.GetScheme())}\n\tif o.Features.Enabled(features.EnableAlphaExternalSecretStores) {\n\t\tcps = append(cps, connection.NewDetailsManager(mgr.GetClient(), v1alpha1.StoreConfigGroupVersionKind))\n\t}\n\n\treturn ctrl.NewControllerManagedBy(mgr).\n\t\tNamed(name).\n\t\tWithOptions(o.ForControllerRuntime()).\n\t\tFor(&v1beta1.Role{}).\n\t\tComplete(managed.NewReconciler(mgr,\n\t\t\tresource.ManagedKind(v1beta1.RoleGroupVersionKind),\n\t\t\tmanaged.WithExternalConnecter(&connector{kube: mgr.GetClient(), newClientFn: iam.NewRoleClient}),\n\t\t\tmanaged.WithReferenceResolver(managed.NewAPISimpleReferenceResolver(mgr.GetClient())),\n\t\t\tmanaged.WithConnectionPublishers(),\n\t\t\tmanaged.WithPollInterval(o.PollInterval),\n\t\t\tmanaged.WithLogger(o.Logger.WithValues(\"controller\", name)),\n\t\t\tmanaged.WithInitializers(managed.NewNameAsExternalName(mgr.GetClient()), &tagger{kube: mgr.GetClient()}),\n\t\t\tmanaged.WithRecorder(event.NewAPIRecorder(mgr.GetEventRecorderFor(name))),\n\t\t\tmanaged.WithConnectionPublishers(cps...)))\n}", "func (_Distributor *DistributorTransactorSession) RenounceRole(role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _Distributor.Contract.RenounceRole(&_Distributor.TransactOpts, role, account)\n}", "func (o TriggerOutput) Role() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Trigger) pulumi.StringPtrOutput { return v.Role }).(pulumi.StringPtrOutput)\n}", "func CreateRole(c *gin.Context) {\n\tnewRole := model.Role{}\n\tc.BindJSON(&newRole)\n\n\terr := service.CreateRole(newRole)\n\n\tif err != nil {\n\t\terror := service.GetGormErrorCode(err.Error())\n\t\tc.JSON(500, error)\n\t} else {\n\t\tc.String(200, \"ok\")\n\t}\n}", "func (m *GormIdentityRoleRepository) Create(ctx context.Context, u *IdentityRole) error {\n\tdefer goa.MeasureSince([]string{\"goa\", \"db\", \"identity_role\", \"create\"}, time.Now())\n\tif u.IdentityRoleID == uuid.Nil {\n\t\tu.IdentityRoleID = uuid.NewV4()\n\t}\n\terr := m.db.Create(u).Error\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"identity_role_id\": u.IdentityRoleID,\n\t\t\t\"err\": err,\n\t\t}, \"unable to create the identity role\")\n\t\tif gormsupport.IsUniqueViolation(err, \"uq_identity_role_identity_role_resource\") {\n\t\t\treturn errs.WithStack(errors.NewDataConflictError(err.Error()))\n\t\t}\n\t\tif gormsupport.IsForeignKeyViolation(err, \"identity_role_identity_id_fkey\") {\n\t\t\treturn errs.WithStack(errors.NewNotFoundError(\"identity\", u.IdentityID.String()))\n\t\t}\n\t\tif gormsupport.IsForeignKeyViolation(err, \"identity_role_resource_id_fkey\") {\n\t\t\treturn errs.WithStack(errors.NewNotFoundError(\"resource\", u.ResourceID))\n\t\t}\n\t\tif gormsupport.IsForeignKeyViolation(err, \"identity_role_role_id_fkey\") {\n\t\t\treturn errs.WithStack(errors.NewNotFoundError(\"role\", u.RoleID.String()))\n\t\t}\n\t\treturn errs.WithStack(err)\n\t}\n\tlog.Debug(ctx, map[string]interface{}{\n\t\t\"identity_role_id\": u.IdentityRoleID,\n\t}, \"Identity Role created!\")\n\treturn nil\n}", "func (t *ACLRole) DBCreate(ctx context.Context, db DB) (sql.Result, error) {\n\tq := \"INSERT INTO `acl_role` (`acl_role`.`id`,`acl_role`.`checksum`,`acl_role`.`name`,`acl_role`.`description`,`acl_role`.`admin_user_id`,`acl_role`.`customer_id`,`acl_role`.`created_at`,`acl_role`.`updated_at`) VALUES (?,?,?,?,?,?,?,?)\"\n\tchecksum := t.CalculateChecksum()\n\tif t.GetChecksum() == checksum {\n\t\treturn nil, nil\n\t}\n\tt.Checksum = &checksum\n\treturn db.ExecContext(ctx, q,\n\t\torm.ToSQLString(t.ID),\n\t\torm.ToSQLString(t.Checksum),\n\t\torm.ToSQLString(t.Name),\n\t\torm.ToSQLString(t.Description),\n\t\torm.ToSQLString(t.AdminUserID),\n\t\torm.ToSQLString(t.CustomerID),\n\t\torm.ToSQLInt64(t.CreatedAt),\n\t\torm.ToSQLInt64(t.UpdatedAt),\n\t)\n}", "func (_PermInterface *PermInterfaceTransactor) AssignAdminRole(opts *bind.TransactOpts, _orgId string, _account common.Address, _roleId string) (*types.Transaction, error) {\n\treturn _PermInterface.contract.Transact(opts, \"assignAdminRole\", _orgId, _account, _roleId)\n}", "func (o *StorageVdMemberEpAllOf) SetRole(v string) {\n\to.Role = &v\n}", "func testDBSecrets(t *testing.T, vc *vaultapi.Client, vconf vaultConfig) {\n\tassert := require.New(t)\n\trole := \"myrole\"\n\n\t// Use the database via Vault\n\tvdbc := NewConnector(vconf.dbURI, vc, nil, vconf.path, role,\n\t\tzaptest.NewLogger(t).Sugar())\n\tdb := sql.OpenDB(vdbc)\n\t// This combination is intended to indicate that each statement uses a\n\t// brand new connection, and that connections won't be reused.\n\tdb.SetMaxOpenConns(1)\n\tdb.SetMaxIdleConns(0)\n\t// This requires the role to be configured, so will return an error.\n\terr := vdbc.SetConnMaxLifetime(db)\n\tassert.Error(err)\n\n\t// This will attempt to open a connection, thus read creds from vault,\n\t// thus fail because the role isn't configured.\n\terr = db.Ping()\n\tassert.Error(err)\n\n\tvconf.createRole(t, role, 2, 5)\n\n\t// These should succeed now.\n\terr = vdbc.SetConnMaxLifetime(db)\n\tassert.NoError(err)\n\terr = db.Ping()\n\tassert.NoError(err)\n\n\twatcher, err := vdbc.getWatcher()\n\tassert.NoError(err)\n\tgo watcher.Start()\n\n\t// Make sure we got credentials.\n\tephemeralRoleName := vdbc.username()\n\tassert.NotEmpty(vdbc.username())\n\tassert.NotEmpty(vdbc.password())\n\n\t// We can create an object with the credentials\n\t_, err = db.Exec(\"CREATE TABLE test();\")\n\tassert.NoError(err)\n\n\t// Verify that the user postgres thinks we are is the same as what Vault\n\t// told us.\n\trow := db.QueryRow(`SELECT session_user`)\n\tassert.NoError(err)\n\tvar sessionUser string\n\terr = row.Scan(&sessionUser)\n\tassert.NoError(err)\n\tassert.Equal(ephemeralRoleName, sessionUser)\n\n\t// Wait for a renewal, and drop the table (showing the dropping user is\n\t// the same as the creating one).\n\trenewEvent := <-watcher.RenewCh()\n\tassert.IsType(&vaultapi.RenewOutput{}, renewEvent)\n\t_, err = db.Exec(\"DROP TABLE test;\")\n\tassert.NoError(err)\n\n\t// Re-create the table; then, wait for the old credentials to expire.\n\t_, err = db.Exec(\"CREATE TABLE test();\")\n\tassert.NoError(err)\n\tdoneErr := <-watcher.DoneCh()\n\tassert.NoError(doneErr)\n\n\t// Demonstrate that the new credentials are in use by looking at the\n\t// session user. Because the credential rotation isn't happening in a\n\t// separate goroutine, it will happen in one of the queries in the loop,\n\t// but we don't know which, in advance. This is because the \"done\"\n\t// notification we got above is not synchronized with the one received\n\t// in waitWatcher, so we don't have a guarantee that it will have been\n\t// delivered by the time we next call it.\n\tfor start := time.Now(); err == nil &&\n\t\tsessionUser == ephemeralRoleName &&\n\t\ttime.Now().Before(start.Add(time.Second)); time.Sleep(50 * time.Millisecond) {\n\t\terr = db.QueryRow(`SELECT session_user`).Scan(&sessionUser)\n\t}\n\tassert.NoError(err)\n\tassert.NotEqual(ephemeralRoleName, sessionUser)\n\n\t// Also, we can create new objects, but are unable to modify objects in\n\t// use by the old user.\n\t_, err = db.Exec(\"CREATE TABLE test2();\")\n\tassert.NoError(err)\n\t_, err = db.Exec(\"DROP TABLE test;\")\n\tassert.Error(err)\n\n\t// Run a query that creates objects at the beginning and the end, and is\n\t// long enough that it would have to straddle credential rotation.\n\tephemeralRoleName = vdbc.username()\n\t_, err = db.Exec(\"CREATE TABLE test3(); SELECT pg_sleep(5); CREATE TABLE test4();\")\n\tassert.NoError(err)\n\t_, err = db.Exec(\"SELECT 1\")\n\tassert.NoError(err)\n\tassert.NotEmpty(vdbc.username())\n\tassert.NotEmpty(vdbc.password())\n\tassert.NotEqual(ephemeralRoleName, vdbc.username())\n\n\t// Make sure that table ownership is as expected; both tables created in\n\t// the previous statement, despite crossing a credential rotation, are\n\t// owned by the same user, but they're different from the owner of the\n\t// previous one.\n\trows, err := db.Query(`\n\t\tSELECT tablename, tableowner\n\t\tFROM pg_tables\n\t\tWHERE tablename IN ('test', 'test3', 'test4')`)\n\tassert.NoError(err)\n\towners := make(map[string]string)\n\tfor rows.Next() {\n\t\tvar owner, table string\n\t\terr = rows.Scan(&table, &owner)\n\t\tassert.NoError(err)\n\t\towners[table] = owner\n\t}\n\tassert.NotEqual(owners[\"test2\"], owners[\"test3\"])\n\tassert.Equal(owners[\"test3\"], owners[\"test4\"])\n}", "func (_PermInterface *PermInterfaceTransactorSession) AssignAccountRole(_account common.Address, _orgId string, _roleId string) (*types.Transaction, error) {\n\treturn _PermInterface.Contract.AssignAccountRole(&_PermInterface.TransactOpts, _account, _orgId, _roleId)\n}", "func (m *UserMutation) SetRole(i int) {\n\tm.role = &i\n\tm.addrole = nil\n}", "func (s *GetWorkspaceOutput) SetRole(v string) *GetWorkspaceOutput {\n\ts.Role = &v\n\treturn s\n}", "func (f *awsFetcher) applyAssumeRole(db types.Database) {\n\tdb.SetAWSAssumeRole(f.cfg.AssumeRole.RoleARN)\n\tdb.SetAWSExternalID(f.cfg.AssumeRole.ExternalID)\n}", "func (e *Election) Role() role.Role {\n\treturn e.role\n}", "func (_Distributor *DistributorTransactor) RenounceRole(opts *bind.TransactOpts, role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _Distributor.contract.Transact(opts, \"renounceRole\", role, account)\n}", "func AssignRoleTypeFromDb(dbRole dbmodels.Role) types.Role {\n\n\treturn types.Role{\n\t\tId: dbRole.ID,\n\t\tName: dbRole.Name,\n\t\tDescription: dbRole.Description,\n\t}\n}", "func (_Distributor *DistributorSession) RenounceRole(role [32]byte, account common.Address) (*types.Transaction, error) {\n\treturn _Distributor.Contract.RenounceRole(&_Distributor.TransactOpts, role, account)\n}", "func bindPauserRole(address common.Address, caller bind.ContractCaller, transactor bind.ContractTransactor, filterer bind.ContractFilterer) (*bind.BoundContract, error) {\n\tparsed, err := ParsedABI(K_PauserRole)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn bind.NewBoundContract(address, *parsed, caller, transactor, filterer), nil\n}", "func (db *RoleDB) AddRole(role models.Role) error {\n\t_, err := db.collection.\n\t\tInsertOne(timeoutContext(), role)\n\treturn err\n}", "func (o ClientTlsPolicyIamBindingOutput) Role() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *ClientTlsPolicyIamBinding) pulumi.StringOutput { return v.Role }).(pulumi.StringOutput)\n}", "func (s *Pipeline) SetRole(v string) *Pipeline {\n\ts.Role = &v\n\treturn s\n}", "func (r *mutationResolver) CreateRole(ctx context.Context, input *models.CreateRoleInput) (*user.Role, error) {\n\tpanic(\"not implemented\")\n}", "func (m *VpnConfiguration) SetRole(value *string)() {\n err := m.GetBackingStore().Set(\"role\", value)\n if err != nil {\n panic(err)\n }\n}", "func (ts *tester) createRole() error {\n\tfmt.Print(ts.cfg.EKSConfig.Colorize(\"\\n\\n[yellow]*********************************\\n\"))\n\tfmt.Printf(ts.cfg.EKSConfig.Colorize(\"[light_green]createRole [default](%q)\\n\"), ts.cfg.EKSConfig.ConfigPath)\n\n\tif !ts.cfg.EKSConfig.Role.Create {\n\t\tts.cfg.Logger.Info(\"Role.Create false; skipping creation\")\n\t\treturn aws_iam.ValidateV2(\n\t\t\tts.cfg.Logger,\n\t\t\tts.cfg.IAMAPIV2,\n\t\t\tts.cfg.EKSConfig.Role.Name,\n\t\t\t[]string{\"eks.amazonaws.com\"},\n\t\t\t[]string{\n\t\t\t\t// Prior to April 16, 2020, AmazonEKSServicePolicy was also required and the suggested name was eksServiceRole. With the AWSServiceRoleForAmazonEKS service-linked role, that policy is no longer required for clusters created on or after April 16, 2020.\n\t\t\t\t// ref. https://docs.aws.amazon.com/eks/latest/userguide/service_IAM_role.html\n\t\t\t\t\"arn:aws:iam::aws:policy/AmazonEKSClusterPolicy\",\n\t\t\t},\n\t\t)\n\t}\n\tif ts.cfg.EKSConfig.Role.ARN != \"\" {\n\t\tts.cfg.Logger.Info(\"role already created; no need to create a new one\")\n\t\treturn nil\n\t}\n\tif ts.cfg.EKSConfig.Role.Name == \"\" {\n\t\treturn errors.New(\"cannot create a cluster role with an empty Role.Name\")\n\t}\n\n\tif err := ts._createRole(); err != nil {\n\t\treturn err\n\t}\n\tif err := ts.createPolicy(); err != nil {\n\t\treturn err\n\t}\n\tif err := ts.attachPolicy(); err != nil {\n\t\treturn err\n\t}\n\n\tts.cfg.Logger.Info(\"created a new role and attached policy\",\n\t\tzap.String(\"role-arn\", ts.cfg.EKSConfig.Role.ARN),\n\t\tzap.String(\"role-name\", ts.cfg.EKSConfig.Role.Name),\n\t)\n\treturn nil\n}", "func (mcr *MiddlewareClusterRepo) Transaction() (middleware.Transaction, error) {\n\treturn mcr.Database.Transaction()\n}", "func (e *ClusterElector) Role() Role {\n\tif (e.state & stateLeaderBootStrapping) != 0 {\n\t\treturn RoleUnstable\n\t}\n\treturn e.role\n}", "func (p *ThriftHiveMetastoreClient) GrantRole(ctx context.Context, role_name string, principal_name string, principal_type PrincipalType, grantor string, grantorType PrincipalType, grant_option bool) (r bool, err error) {\n var _args124 ThriftHiveMetastoreGrantRoleArgs\n _args124.RoleName = role_name\n _args124.PrincipalName = principal_name\n _args124.PrincipalType = principal_type\n _args124.Grantor = grantor\n _args124.GrantorType = grantorType\n _args124.GrantOption = grant_option\n var _result125 ThriftHiveMetastoreGrantRoleResult\n if err = p.Client_().Call(ctx, \"grant_role\", &_args124, &_result125); err != nil {\n return\n }\n switch {\n case _result125.O1!= nil:\n return r, _result125.O1\n }\n\n return _result125.GetSuccess(), nil\n}", "func (_PermInterface *PermInterfaceSession) AssignAccountRole(_account common.Address, _orgId string, _roleId string) (*types.Transaction, error) {\n\treturn _PermInterface.Contract.AssignAccountRole(&_PermInterface.TransactOpts, _account, _orgId, _roleId)\n}", "func (o *Content) SetRole(v string) {\n\to.Role = &v\n}", "func (wrc *WorkspaceRoleCreate) Save(ctx context.Context) (*WorkspaceRole, error) {\n\tvar (\n\t\terr error\n\t\tnode *WorkspaceRole\n\t)\n\twrc.defaults()\n\tif len(wrc.hooks) == 0 {\n\t\tif err = wrc.check(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tnode, err = wrc.sqlSave(ctx)\n\t} else {\n\t\tvar mut Mutator = MutateFunc(func(ctx context.Context, m Mutation) (Value, error) {\n\t\t\tmutation, ok := m.(*WorkspaceRoleMutation)\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"unexpected mutation type %T\", m)\n\t\t\t}\n\t\t\tif err = wrc.check(); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\twrc.mutation = mutation\n\t\t\tnode, err = wrc.sqlSave(ctx)\n\t\t\tmutation.done = true\n\t\t\treturn node, err\n\t\t})\n\t\tfor i := len(wrc.hooks) - 1; i >= 0; i-- {\n\t\t\tmut = wrc.hooks[i](mut)\n\t\t}\n\t\tif _, err := mut.Mutate(ctx, wrc.mutation); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn node, err\n}", "func ProvideRoleService(p repo.RoleRepo) RoleServ {\n\treturn RoleServ{Repo: p, Engine: p.Engine}\n}", "func (p *RoleServ) Save(role model.Role) (savedRole model.Role, err error) {\n\n\trole.CompanyID, role.NodeCode, _ = role.ID.Split()\n\n\tif err = role.Validate(action.Save); err != nil {\n\t\tp.Engine.CheckError(err, \"validation failed\")\n\t\treturn\n\t}\n\n\tsavedRole, err = p.Repo.Update(role)\n\tp.Engine.CheckInfo(err, fmt.Sprintf(\"Failed in updating role for %+v\", role))\n\n\treturn\n}", "func (*Role) TableName() string {\n\treturn tablePrefix + \"role\"\n}", "func DeleteRole(dbmap *gorp.DbMap, name string) error {\n\tvar r Role\n\terr := dbmap.SelectOne(&r, \"SELECT * FROM role WHERE name = ?\", name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttx, err := dbmap.Begin()\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\t_, err = tx.Exec(\"DELETE FROM user_role WHERE role_id = ?;\", r.PK)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\t_, err = tx.Exec(\"DELETE FROM role_permission WHERE role_id = ?;\", r.PK)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\t_, err = tx.Delete(&r)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\treturn tx.Commit()\n}", "func (_PermInterface *PermInterfaceSession) AssignAdminRole(_orgId string, _account common.Address, _roleId string) (*types.Transaction, error) {\n\treturn _PermInterface.Contract.AssignAdminRole(&_PermInterface.TransactOpts, _orgId, _account, _roleId)\n}" ]
[ "0.593468", "0.5845933", "0.5706156", "0.56633246", "0.5598586", "0.5546218", "0.5509051", "0.5496969", "0.5492553", "0.54864913", "0.54306674", "0.54306674", "0.54306674", "0.54306674", "0.54260254", "0.54152405", "0.54106504", "0.54106504", "0.54106504", "0.53951716", "0.53845227", "0.5376039", "0.5369332", "0.53581125", "0.5352062", "0.53287375", "0.5319704", "0.5302519", "0.5263734", "0.5246594", "0.5239966", "0.52329415", "0.52317506", "0.52257985", "0.52244264", "0.52031475", "0.51951694", "0.51640564", "0.5128497", "0.5127475", "0.51213294", "0.509723", "0.5096657", "0.5087948", "0.5062552", "0.5044597", "0.5041925", "0.5040424", "0.50048643", "0.5001319", "0.4999967", "0.49917388", "0.498094", "0.4972997", "0.4971274", "0.49353427", "0.492981", "0.4922865", "0.4908459", "0.4906435", "0.48937577", "0.4891197", "0.48812795", "0.48687887", "0.48645532", "0.48633382", "0.4852328", "0.48521584", "0.48499787", "0.48494813", "0.48465854", "0.48436233", "0.4842507", "0.482985", "0.48270565", "0.48253047", "0.48172787", "0.48153087", "0.48151368", "0.4811351", "0.48090073", "0.47983527", "0.47885534", "0.47865462", "0.47833794", "0.47819367", "0.47796476", "0.47777382", "0.47684744", "0.47583568", "0.47552285", "0.47549507", "0.4752199", "0.47505528", "0.47468352", "0.47459227", "0.4745063", "0.47318625", "0.4723495", "0.4717571" ]
0.6932387
0
Permission will inject the databaseTx in the `Permission` schema
func (gtx *GuardTx) Permission(permission *schema.Permission) *schema.Permission { if permission == nil { return &schema.Permission{ Entity: schema.Entity{DBContract: gtx.dbTx}, } } else { permission.DBContract = gtx.dbTx } permission.SetValidator(gtx.validator.Permission) return permission }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (_Storage *StorageTransactor) GrantPermission(opts *bind.TransactOpts, kind uint8, addr common.Address) (*types.Transaction, error) {\n\treturn _Storage.contract.Transact(opts, \"grantPermission\", kind, addr)\n}", "func CreatePermissionSuite(\n\tdb Queryer,\n\tmodel interface{},\n) error {\n\tmType, err := mytype.ParseNodeType(structs.Name(model))\n\tif err != nil {\n\t\tmylog.Log.WithError(err).Debug(util.Trace(\"\"))\n\t\treturn err\n\t}\n\n\tfields := structs.Fields(model)\n\tn := len(fields)*len(accessLevelsWithFields) + len(accessLevelsWithoutFields)\n\tpermissions := make([][]interface{}, 0, n)\n\tfor _, al := range accessLevelsWithFields {\n\t\tfor _, f := range fields {\n\t\t\tid, _ := mytype.NewOID(\"Permission\")\n\t\t\tfield := f.Tag(\"db\")\n\t\t\tpermits := strings.Split(f.Tag(\"permit\"), \"/\")\n\t\t\tfor _, p := range permits {\n\t\t\t\tif strings.ToLower(p) == strings.ToLower(al.String()) {\n\t\t\t\t\tpermissions = append(permissions, []interface{}{\n\t\t\t\t\t\tid,\n\t\t\t\t\t\tal,\n\t\t\t\t\t\tmType,\n\t\t\t\t\t\tmytype.Authenticated,\n\t\t\t\t\t\tfield,\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tfor _, al := range accessLevelsWithoutFields {\n\t\tid, _ := mytype.NewOID(\"Permission\")\n\t\tpermissions = append(permissions, []interface{}{\n\t\t\tid,\n\t\t\tal,\n\t\t\tmType,\n\t\t\tmytype.Authenticated,\n\t\t\tnil,\n\t\t})\n\t}\n\n\ttx, err, newTx := BeginTransaction(db)\n\tif err != nil {\n\t\tmylog.Log.WithError(err).Debug(util.Trace(\"\"))\n\t\treturn err\n\t}\n\tif newTx {\n\t\tdefer RollbackTransaction(tx)\n\t}\n\n\tif err := DeletePermissionSuite(db, model); err != nil {\n\t\tmylog.Log.WithError(err).Debug(util.Trace(\"\"))\n\t\treturn err\n\t}\n\n\tcopyCount, err := tx.CopyFrom(\n\t\tpgx.Identifier{\"permission\"},\n\t\t[]string{\"id\", \"access_level\", \"type\", \"audience\", \"field\"},\n\t\tpgx.CopyFromRows(permissions),\n\t)\n\tif err != nil {\n\t\tif pgErr, ok := err.(pgx.PgError); ok {\n\t\t\tswitch PSQLError(pgErr.Code) {\n\t\t\tdefault:\n\t\t\t\tmylog.Log.WithError(err).Debug(util.Trace(\"\"))\n\t\t\t\treturn err\n\t\t\tcase UniqueViolation:\n\t\t\t\tmylog.Log.Warn(\"permissions already created\")\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t\tmylog.Log.WithError(err).Debug(util.Trace(\"\"))\n\t\treturn err\n\t}\n\n\tif newTx {\n\t\terr = CommitTransaction(tx)\n\t\tif err != nil {\n\t\t\tmylog.Log.WithError(err).Debug(util.Trace(\"\"))\n\t\t\treturn err\n\t\t}\n\t}\n\n\tmylog.Log.WithFields(logrus.Fields{\n\t\t\"n\": copyCount,\n\t\t\"type\": mType,\n\t}).Info(util.Trace(\"created permissions\"))\n\treturn nil\n}", "func initPermission() {\n CorePermissionModuleID = AddModule(\"Core Permission Module\", 1)\n\n model := (*CorePermission)(nil)\n AddModel(model)\n\n // Temp\n InsertInitialModel(\n model,\n &CorePermission{\n ModuleID: CoreGroupModuleID,\n GroupID: 3,\n Read: CorePermissionAll,\n Create: CorePermissionAll,\n Update: CorePermissionNone,\n Delete: CorePermissionNone,\n },\n )\n}", "func (_Storage *StorageTransactorSession) GrantPermission(kind uint8, addr common.Address) (*types.Transaction, error) {\n\treturn _Storage.Contract.GrantPermission(&_Storage.TransactOpts, kind, addr)\n}", "func (_Storage *StorageSession) GrantPermission(kind uint8, addr common.Address) (*types.Transaction, error) {\n\treturn _Storage.Contract.GrantPermission(&_Storage.TransactOpts, kind, addr)\n}", "func (m PermissionMutation) Tx() (*Tx, error) {\n\tif _, ok := m.driver.(*txDriver); !ok {\n\t\treturn nil, fmt.Errorf(\"ent: mutation is not running in a transaction\")\n\t}\n\ttx := &Tx{config: m.config}\n\ttx.init()\n\treturn tx, nil\n}", "func (db *SQLiteDB) SetPermissions(prm Permission) error {\n\n\tquery, err := db.Exec(\"INSERT INTO Permissions (`UserID`, `GroupID` , `Permission`) VALUES (?,?,?) \"+\n\t\t\"ON CONFLICT(`UserID`,`GroupID`,`Permission`) DO UPDATE \"+\n\t\t\"SET `Permission` = Excluded.Permission\", prm.UserID, prm.GroupID, prm.Permission)\n\tif err != nil {\n\t\tdb.AddLogEvent(Log{Event: \"SetPermissions_QueryFailed\", Message: \"Impossible to create the execute the query\", Error: err.Error()})\n\t\treturn err\n\t}\n\trows, err := query.RowsAffected()\n\tif err != nil {\n\t\tdb.AddLogEvent(Log{Event: \"SetPermissions_RowsInfoNotGot\", Message: \"Impossible to get afftected rows\", Error: err.Error()})\n\t\treturn err\n\t}\n\tif rows < 1 {\n\t\tdb.AddLogEvent(Log{Event: \"SetPermissions_NoRowsAffected\", Message: \"No rows affected\"})\n\t\treturn NoRowsAffected{error: errors.New(\"No rows affected from the query\")}\n\t}\n\treturn err\n\n\t/*stmt, err := db.Prepare(\"INSERT INTO Permissions (`User`, `Group` , `Permission`) VALUES (?,?,?) ON CONFLICT(`User`,`Group`,`Permission`) DO UPDATE SET `Permission` = Excluded.Permission\")\n\tif err != nil {\n\t\tdb.AddLogEvent(Log{Event: \"SetPermissions_QueryFailed\", Message: \"The query for the SetPermissions function failed\", Error: err.Error()})\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\n\t//And we execute it passing the parameters\n\trows, err := stmt.Exec(userID, groupID, permissions)\n\n\tif err != nil {\n\t\tdb.AddLogEvent(Log{Event: \"SetPermissions_NotFoundUnknown\", Message: \"The execution of the query for the SetPermissions function failed\", Error: err.Error()})\n\t\treturn err\n\t}\n\n\tres, err := rows.RowsAffected()\n\n\tif err != nil {\n\t\tdb.AddLogEvent(Log{Event: \"SetPermissions_ExecutionQueryError\", Message: \"The fetching of the query results for the SetPermissions function failed\", Error: err.Error()})\n\t\treturn err\n\t}\n\tif res > 0 {\n\t\treturn nil\n\t}\n\tdb.AddLogEvent(Log{Event: \"SetPermissions_NotChangesMade\", Message: \"No changes was made to the database!\", Error: err.Error()})\n\treturn errors.New(\"No changes to the database was made\")*/\n}", "func (o *AuthUserUserPermission) Permission(exec boil.Executor, mods ...qm.QueryMod) authPermissionQuery {\n\tqueryMods := []qm.QueryMod{\n\t\tqm.Where(\"id=?\", o.PermissionID),\n\t}\n\n\tqueryMods = append(queryMods, mods...)\n\n\tquery := AuthPermissions(exec, queryMods...)\n\tqueries.SetFrom(query.Query, \"`auth_permission`\")\n\n\treturn query\n}", "func changePermission(db *sqlite.Driver, acct *Account, newPerm PermLevel) (*Account, error) {\n\tvar err error\n\tvar stmt = fmt.Sprintf(\"update %s set permission_level = ? where id = ?\", tableName)\n\tif _, err = db.Exec(stmt, newPerm, acct.ID); err != nil {\n\t\treturn nil, err\n\t}\n\n\tacct.PermLevel = newPerm\n\treturn acct, nil\n}", "func (o *AuthUserUserPermission) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no auth_user_user_permissions provided for insertion\")\n\t}\n\n\tvar err error\n\n\tnzDefaults := queries.NonZeroDefaultSet(authUserUserPermissionColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tauthUserUserPermissionInsertCacheMut.RLock()\n\tcache, cached := authUserUserPermissionInsertCache[key]\n\tauthUserUserPermissionInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tauthUserUserPermissionColumns,\n\t\t\tauthUserUserPermissionColumnsWithDefault,\n\t\t\tauthUserUserPermissionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authUserUserPermissionType, authUserUserPermissionMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(authUserUserPermissionType, authUserUserPermissionMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.query = fmt.Sprintf(\"INSERT INTO `auth_user_user_permissions` (`%s`) VALUES (%s)\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `auth_user_user_permissions` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, authUserUserPermissionPrimaryKeyColumns))\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into auth_user_user_permissions\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == authUserUserPermissionMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for auth_user_user_permissions\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tauthUserUserPermissionInsertCacheMut.Lock()\n\t\tauthUserUserPermissionInsertCache[key] = cache\n\t\tauthUserUserPermissionInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func CreatePermission(c *gin.Context) {\n\tnewPermission := model.Permission{}\n\tc.BindJSON(&newPermission)\n\n\terr := service.CreatePermission(newPermission)\n\n\tif err != nil {\n\t\terror := service.GetGormErrorCode(err.Error())\n\t\tc.JSON(500, error)\n\t} else {\n\t\tc.String(200, \"ok\")\n\t}\n}", "func (ctx *TestContext) addPermissionGranted(group, item, permission, permissionValue string) {\n\tgroupID := ctx.getReference(group)\n\titemID := ctx.getReference(item)\n\n\tpermissionsGrantedTable := \"permissions_granted\"\n\tkey := strconv.FormatInt(groupID, 10) + \",\" + strconv.FormatInt(itemID, 10)\n\n\tif !ctx.isInDatabase(permissionsGrantedTable, key) {\n\t\tctx.addInDatabase(permissionsGrantedTable, key, map[string]interface{}{\n\t\t\t\"group_id\": groupID,\n\t\t\t\"source_group_id\": groupID,\n\t\t\t\"item_id\": itemID,\n\t\t})\n\t}\n\n\tif permission == \"can_request_help_to\" {\n\t\tpermissionValue = strconv.FormatInt(ctx.getReference(permissionValue), 10)\n\t}\n\n\tif permission == \"is_owner\" {\n\t\tboolValue, err := strconv.ParseBool(permissionValue)\n\t\tif err != nil {\n\t\t\tpanic(fmt.Sprintf(\"%v cannot be parsed as a boolean\", boolValue))\n\t\t}\n\n\t\tctx.dbTables[permissionsGrantedTable][key][permission] = boolValue\n\t} else {\n\t\tctx.dbTables[permissionsGrantedTable][key][permission] = permissionValue\n\t}\n}", "func (s *permStore) Create(ctx context.Context, perm *core.Perm) error {\n\treturn s.db.Lock(func(execer db.Execer, binder db.Binder) error {\n\t\tparams := toParams(perm)\n\t\tstmt, args, err := binder.BindNamed(stmtInsert, params)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t_, err = execer.Exec(stmt, args...)\n\t\treturn err\n\t})\n}", "func (a *Authorizer) AddPermission(name string, description string) error {\n\n\tp := permission{\n\t\tName: name,\n\t\tDescription: description,\n\t}\n\n\terr := a.db.Save(&p)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn a.load()\n\n}", "func (r *mutationResolver) CreatePermission(ctx context.Context, input *models.CreatePermissionInput) (*user.Permission, error) {\n\tpanic(\"not implemented\")\n}", "func (r *Resolver) Permissions() PermissionsResolver { return &permissionsResolver{r} }", "func (auup *AuthUserUserPermission) Insert(ctx context.Context, db DB) error {\n\tswitch {\n\tcase auup._exists: // already exists\n\t\treturn logerror(&ErrInsertFailed{ErrAlreadyExists})\n\tcase auup._deleted: // deleted\n\t\treturn logerror(&ErrInsertFailed{ErrMarkedForDeletion})\n\t}\n\t// insert (primary key generated and returned by database)\n\tconst sqlstr = `INSERT INTO django.auth_user_user_permissions (` +\n\t\t`user_id, permission_id` +\n\t\t`) VALUES (` +\n\t\t`:1, :2` +\n\t\t`) RETURNING id INTO :3`\n\t// run\n\tlogf(sqlstr, auup.UserID, auup.PermissionID)\n\tvar id int64\n\tif _, err := db.ExecContext(ctx, sqlstr, auup.UserID, auup.PermissionID, sql.Out{Dest: &id}); err != nil {\n\t\treturn logerror(err)\n\t} // set primary key\n\tauup.ID = int64(id)\n\t// set exists\n\tauup._exists = true\n\treturn nil\n}", "func (u *user) grant(ctx context.Context, db Database, access string) error {\n\tescapedDbName := pathEscape(db.Name())\n\treq, err := u.conn.NewRequest(\"PUT\", path.Join(u.relPath(), \"database\", escapedDbName))\n\tif err != nil {\n\t\treturn WithStack(err)\n\t}\n\tinput := struct {\n\t\tGrant string `arangodb:\"grant\" json:\"grant\"`\n\t}{\n\t\tGrant: access,\n\t}\n\tif _, err := req.SetBody(input); err != nil {\n\t\treturn WithStack(err)\n\t}\n\tresp, err := u.conn.Do(ctx, req)\n\tif err != nil {\n\t\treturn WithStack(err)\n\t}\n\tif err := resp.CheckStatus(200); err != nil {\n\t\treturn WithStack(err)\n\t}\n\treturn nil\n}", "func (a *Authorizer) AttachPermission(permName string, roleName string) error {\n\n\t// Make sure permission exist\n\n\tp := permission{}\n\n\terr := a.db.One(\"Name\", permName, &p)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tr := role{}\n\terr = a.db.One(\"Name\", roleName, &r)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Look if it's already there\n\tfor _, cp := range r.Permissions {\n\t\tif cp.Name == permName {\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t// Add it\n\tr.Permissions = append(r.Permissions, p)\n\n\t// Save\n\terr = a.db.Save(&r)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn a.load()\n\n}", "func init() {\n\tmigrations.MustRegisterTx(func(db migrations.DB) error {\n\t\t_, err := db.Exec(`\nCREATE TABLE sintls_authorization (\n authorization_id bigserial primary key,\n created_at timestamp not null default now(),\n updated_at timestamp not null default now(),\n name text unique not null,\n secret text not null,\n admin boolean not null default false\n);\nCREATE INDEX ON sintls_authorization(name);\n\nCREATE TABLE sintls_subdomain (\n subdomain_id bigserial primary key,\n created_at timestamp not null default now(),\n updated_at timestamp not null default now(),\n name text unique not null,\n authorization_id bigint not null references sintls_authorization(authorization_id) on delete cascade\n);\nCREATE INDEX ON sintls_subdomain(name);\nCREATE INDEX ON sintls_subdomain(authorization_id);\n\nCREATE TABLE sintls_host (\n host_id bigserial primary key,\n created_at timestamp not null default now(),\n updated_at timestamp not null default now(),\n name text not null,\n subdomain_id bigint not null references sintls_subdomain(subdomain_id) on delete cascade,\n dns_target_a inet,\n dns_target_aaaa inet,\n UNIQUE (name, subdomain_id),\n CHECK (dns_target_a is not null or dns_target_aaaa is not null)\n);\nCREATE INDEX on sintls_host(name);\nCREATE INDEX on sintls_host(subdomain_id);`)\n\t\treturn err\n\t}, func(db migrations.DB) error {\n\t\t_, err := db.Exec(`DROP TABLE sintls_host; DROP TABLE sintls_subdomain; DROP TABLE sintls_authorization;`)\n\t\treturn err\n\t})\n}", "func (db *DB) Security(ctx context.Context) (*Security, error) {\n\ts, err := db.driverDB.Security(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Security{\n\t\tAdmins: Members(s.Admins),\n\t\tMembers: Members(s.Members),\n\t}, err\n}", "func (a *Auth) CheckPermission(ctx context.Context, biz *cmdb.Biz, iamSettings cc.IAM, req *meta.ResourceAttribute) (*pbas.CheckPermissionResp, error) {\n\tkt := kit.FromGrpcContext(ctx)\n\n\tlog := &logrus.Logger{\n\t\tOut: os.Stderr,\n\t\tFormatter: new(logrus.TextFormatter),\n\t\tHooks: make(logrus.LevelHooks),\n\t\tLevel: logrus.DebugLevel,\n\t\tExitFunc: os.Exit,\n\t\tReportCaller: false,\n\t}\n\n\tbkiamlogger.SetLogger(log)\n\n\tactionRequest, err := AdaptIAMResourceOptions(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tactionRequest.Subject = bkiam.NewSubject(\"user\", kt.User)\n\t// i := bkiam.NewIAM(sys.SystemIDBSCP, iamSettings.AppCode, iamSettings.AppSecret, iamSettings.Endpoints[0], \"\")\n\ti := bkiam.NewAPIGatewayIAM(sys.SystemIDBSCP, iamSettings.AppCode, iamSettings.AppSecret, iamSettings.APIURL)\n\tallowed, err := i.IsAllowed(*actionRequest)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp := &pbas.CheckPermissionResp{\n\t\tIsAllowed: false,\n\t\tApplyUrl: \"\",\n\t\tResources: []*pbas.BasicDetail{},\n\t}\n\n\tif allowed {\n\t\tresp.IsAllowed = true\n\t\treturn resp, nil\n\t}\n\n\tif req.GenApplyURL {\n\t\tresp.Resources = append(resp.Resources, &pbas.BasicDetail{\n\t\t\tType: string(req.Type),\n\t\t\tAction: req.Action.String(),\n\t\t\tResourceId: strconv.FormatInt(int64(req.ResourceID), 10),\n\t\t\tTypeName: \"业务\",\n\t\t\tActionName: \"业务访问\",\n\t\t\tResourceName: biz.BizName,\n\t\t})\n\n\t\tapplication, err := AdaptIAMApplicationOptions(req)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\turl, err := i.GetApplyURL(*application, \"\", kt.User)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"gen apply url\")\n\t\t}\n\t\tresp.ApplyUrl = url\n\t}\n\n\treturn resp, nil\n}", "func (UserID UserID) HasPermissionDB(db *sql.DB, permissionid string) bool {\n\treturn false\n}", "func (o *AuthUserUserPermission) SetPermission(exec boil.Executor, insert bool, related *AuthPermission) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(exec); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE `auth_user_user_permissions` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, []string{\"permission_id\"}),\n\t\tstrmangle.WhereClause(\"`\", \"`\", 0, authUserUserPermissionPrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.ID, o.ID}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, updateQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tif _, err = exec.Exec(updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.PermissionID = related.ID\n\n\tif o.R == nil {\n\t\to.R = &authUserUserPermissionR{\n\t\t\tPermission: related,\n\t\t}\n\t} else {\n\t\to.R.Permission = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &authPermissionR{\n\t\t\tPermissionAuthUserUserPermissions: AuthUserUserPermissionSlice{o},\n\t\t}\n\t} else {\n\t\trelated.R.PermissionAuthUserUserPermissions = append(related.R.PermissionAuthUserUserPermissions, o)\n\t}\n\n\treturn nil\n}", "func (p *Permission) CreatePermission() revel.Result {\n\tpermType := models.Permission{}\n\tp.Params.BindJSON(&permType)\n\tif !app.DB.Where(\"name = ?\", permType.Name).First(&permType).RecordNotFound() {\n\t\treturn util.AppResponse{400, \"perm already exists\", nil}\n\t}\n\tapp.DB.Create(&permType)\n\treturn util.AppResponse{200, \"success\", permType}\n}", "func (ppc *PermissionsPolicyCreate) Save(ctx context.Context) (*PermissionsPolicy, error) {\n\tif _, ok := ppc.mutation.CreateTime(); !ok {\n\t\tv := permissionspolicy.DefaultCreateTime()\n\t\tppc.mutation.SetCreateTime(v)\n\t}\n\tif _, ok := ppc.mutation.UpdateTime(); !ok {\n\t\tv := permissionspolicy.DefaultUpdateTime()\n\t\tppc.mutation.SetUpdateTime(v)\n\t}\n\tif _, ok := ppc.mutation.Name(); !ok {\n\t\treturn nil, &ValidationError{Name: \"name\", err: errors.New(\"ent: missing required field \\\"name\\\"\")}\n\t}\n\tif v, ok := ppc.mutation.Name(); ok {\n\t\tif err := permissionspolicy.NameValidator(v); err != nil {\n\t\t\treturn nil, &ValidationError{Name: \"name\", err: fmt.Errorf(\"ent: validator failed for field \\\"name\\\": %w\", err)}\n\t\t}\n\t}\n\tif _, ok := ppc.mutation.IsGlobal(); !ok {\n\t\tv := permissionspolicy.DefaultIsGlobal\n\t\tppc.mutation.SetIsGlobal(v)\n\t}\n\tvar (\n\t\terr error\n\t\tnode *PermissionsPolicy\n\t)\n\tif len(ppc.hooks) == 0 {\n\t\tnode, err = ppc.sqlSave(ctx)\n\t} else {\n\t\tvar mut Mutator = MutateFunc(func(ctx context.Context, m Mutation) (Value, error) {\n\t\t\tmutation, ok := m.(*PermissionsPolicyMutation)\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"unexpected mutation type %T\", m)\n\t\t\t}\n\t\t\tppc.mutation = mutation\n\t\t\tnode, err = ppc.sqlSave(ctx)\n\t\t\tmutation.done = true\n\t\t\treturn node, err\n\t\t})\n\t\tfor i := len(ppc.hooks) - 1; i >= 0; i-- {\n\t\t\tmut = ppc.hooks[i](mut)\n\t\t}\n\t\tif _, err := mut.Mutate(ctx, ppc.mutation); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn node, err\n}", "func (p *AccumuloProxyClient) GrantTablePermission(login []byte, user string, table string, perm TablePermission) (err error) {\n\tif err = p.sendGrantTablePermission(login, user, table, perm); err != nil {\n\t\treturn\n\t}\n\treturn p.recvGrantTablePermission()\n}", "func (m *PatientrightstypeMutation) Permission() (r string, exists bool) {\n\tv := m._Permission\n\tif v == nil {\n\t\treturn\n\t}\n\treturn *v, true\n}", "func (Permission) TableName() string {\n\treturn \"iam_permission\"\n}", "func TransformPermission(permission db.Permission) ResponsePermission {\n\treturn ResponsePermission{\n\t\tKey: permission.Key,\n\t\tDescription: permission.Description,\n\t}\n}", "func (p Permissions) CreateEntry() (Permissions, error) {\n\ts, err := session.NewSession(&aws.Config{\n\t\tRegion: aws.String(os.Getenv(\"DB_REGION\")),\n\t\tEndpoint: aws.String(os.Getenv(\"DB_ENDPOINT\")),\n\t})\n\tif err != nil {\n\t\treturn Permissions{}, fmt.Errorf(\"create entry new session err: %w\", err)\n\t}\n\tperms, err := convertPermissionsToDynamo(p.Permissions, p.Identifier)\n\tif err != nil {\n\t\treturn Permissions{}, fmt.Errorf(\"create convert permissions err: %w\", err)\n\t}\n\tsvc := dynamodb.New(s)\n\titem := map[string]*dynamodb.AttributeValue{\n\t\t\"identifier\": {\n\t\t\tS: aws.String(p.Identifier),\n\t\t},\n\t\t\"permissions\": &perms,\n\t}\n\n\tinput := &dynamodb.PutItemInput{\n\t\tTableName: aws.String(os.Getenv(\"DB_TABLE\")),\n\t\tItem: item,\n\t\tConditionExpression: aws.String(\"attribute_not_exists(#IDENTIFIER)\"),\n\t\tExpressionAttributeNames: map[string]*string{\n\t\t\t\"#IDENTIFIER\": aws.String(\"identifier\"),\n\t\t},\n\t}\n\t_, putErr := svc.PutItem(input)\n\tif putErr != nil {\n\t\tif awsErr, ok := putErr.(awserr.Error); ok {\n\t\t\tswitch awsErr.Code() {\n\t\t\tcase dynamodb.ErrCodeConditionalCheckFailedException:\n\t\t\t\treturn Permissions{}, fmt.Errorf(\"permission identifier already exists: %w\", awsErr.Code())\n\t\t\tcase \"ValidationException\":\n\t\t\t\tfmt.Println(fmt.Sprintf(\"validation err reason: %v\", input))\n\t\t\t\treturn Permissions{}, fmt.Errorf(\"validation error: %w\", awsErr)\n\t\t\tdefault:\n\t\t\t\tfmt.Println(fmt.Sprintf(\"unknown code err reason: %v\", input))\n\t\t\t\treturn Permissions{}, fmt.Errorf(\"unknown code err: %w\", awsErr)\n\t\t\t}\n\t\t} else {\n\t\t\treturn Permissions{}, fmt.Errorf(\"unknown err: %w\", putErr)\n\t\t}\n\t}\n\n\treturn convertDynamoToPermission(item)\n}", "func (sp *SalePermission) Insert(ctx context.Context, key ...interface{}) error {\n\tvar err error\n\tvar dbConn *sql.DB\n\tvar res sql.Result\n\t// if already exist, bail\n\tif sp._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\ttx, err := components.M.GetConnFromCtx(ctx)\n\tif err != nil {\n\t\tdbConn, err = components.M.GetMasterConn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttableName, err := GetSalePermissionTableName(key...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// sql insert query, primary key must be provided\n\tsqlstr := `INSERT INTO ` + tableName +\n\t\t` (` +\n\t\t`spid, desc, action_class` +\n\t\t`) VALUES (` +\n\t\t`?, ?, ?` +\n\t\t`)`\n\n\t// run query\n\tutils.GetTraceLog(ctx).Debug(\"DB\", zap.String(\"SQL\", fmt.Sprint(sqlstr, sp.Spid, sp.Desc, sp.ActionClass)))\n\tif tx != nil {\n\t\tres, err = tx.Exec(sqlstr, sp.Spid, sp.Desc, sp.ActionClass)\n\t} else {\n\t\tres, err = dbConn.Exec(sqlstr, sp.Spid, sp.Desc, sp.ActionClass)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\tsp._exists = true\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// retrieve id\n\tid, err := res.LastInsertId()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set primary key and existence\n\tsp.Spid = uint(id)\n\tsp._exists = true\n\n\treturn nil\n}", "func (data *ThingCreateRequest) Save(ctx context.Context) (*model.Thing, error) {\n // Extract the current user from the request context.\n user := ctx.Value(\"user\").(*model.User)\n\n // Get our database connection.\n dao := database.Instance()\n db := dao.GetORM()\n\n // The model we will be creating.\n var thing model.Thing\n\n // Create our object in the database.\n thing = model.Thing {\n BoxID: data.BoxID,\n Name: data.Name,\n ShortDescription: data.ShortDescription,\n LongDescription: data.LongDescription,\n UnitOfMeasure: data.UnitOfMeasure,\n User: *user,\n Status: data.Status,\n ShareKey: \"\",\n StreetAddress: data.StreetAddress,\n StreetAddressExtra: data.StreetAddressExtra,\n City: data.City,\n Province: data.Province,\n Country: data.Country,\n Postal: data.Postal,\n IsAddressVisible: data.IsAddressVisible,\n }\n db.Create(&thing)\n\n // Create our permission.\n userThingPermission := model.UserThingPermission{\n UserID: user.ID,\n ThingID: thing.ID,\n Type: 1,\n }\n db.Create(&userThingPermission)\n\n // Return our newly created `Thing` object.\n return &thing, nil\n}", "func (bc *BaseController) AddPermission(permission *dream.Permission) (error, *dream.Permission) {\n\n\tresult, err := dream.AddPermission(bc.DB, permission)\n\n\tif err != nil {\n\t\treturn ServiceErr, nil\n\t}\n\n\treturn nil, result\n}", "func (ua *App) AddPermission(permission string, ev *enliven.Enliven, groups ...string) {\n\tdb := database.GetDatabase()\n\n\tperm := Permission{}\n\tdb.Where(&Permission{Name: permission}).First(&perm)\n\n\tif perm.ID == 0 {\n\t\tnewPerm := Permission{\n\t\t\tName: permission,\n\t\t}\n\t\tdb.Create(&newPerm)\n\n\t\t// Adding this new permission to any specified groups.\n\t\tfor _, groupName := range groups {\n\t\t\tgroup := Group{}\n\t\t\tdb.Where(\"Name = ?\", groupName).First(&group)\n\n\t\t\tif group.ID != 0 {\n\t\t\t\tgroup.Permissions = append(group.Permissions, newPerm)\n\t\t\t}\n\n\t\t\tdb.Save(group)\n\t\t}\n\t}\n}", "func (App) Permissions() []evo.Permission { return []evo.Permission{} }", "func (state *Instance) dbUserPermissions(id string) bool {\n\tdb := state.Database\n\tvar val sql.NullString\n\t// Grab ID from table (sure indicator that it is a valid person)\n\terr := db.QueryRow(\"SELECT id FROM permissions WHERE id=(?)\", id).Scan(&val)\n\tif err != nil {\n\t\treturn false\n\t\t// If information is not NULL\n\t} else if val.Valid {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func setupDatabase(database *sql.DB) {\n\t_, err := database.Exec(`\n\t\tCREATE TABLE IF NOT EXISTS resources (\n\t\t\tid INTEGER PRIMARY KEY,\n\t\t\towner_id BLOB,\n\t\t\tstr_id TEXT UNIQUE NOT NULL\n\t\t);\n\t`)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"Could not setup table: 'resources': %s\", err))\n\t}\n\t_, err = database.Exec(`\n\t\tCREATE TABLE IF NOT EXISTS resource_permissions (\n\t\t\tresource_id INTEGER,\n\t\t\tuser_id BLOB,\n\t\t\tpermissions INTEGER NOT NULL,\n\t\t\tPRIMARY KEY(resource_id, user_id)\n\t\t);\n\t`)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"Could not setup table: 'resource_permissions': %s\", err))\n\t}\n\t_, err = database.Exec(`\n\t\tCREATE TABLE IF NOT EXISTS users (\n\t\t\tid BLOB,\n\t\t\tusername TEXT NOT NULL,\n\t\t\tpassword TEXT NOT NULL,\n\t\t\tPRIMARY KEY (id)\n\t\t);\n\t`)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"Could not setup table: 'users': %s\", err))\n\t}\n\n\t_, err = database.Exec(`\n\t\tCREATE INDEX IF NOT EXISTS str_id\n\t\tON resources (str_id);\n\t`)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"Could not create index: 'resources.resource_strid': %s\", err))\n\t}\n\n\t_, err = database.Exec(`\n\t\tCREATE INDEX IF NOT EXISTS username\n\t\tON users (username);\n\t`)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"Could not create index: 'users.username': %s\", err))\n\t}\n}", "func GetAllAdminPermission() (v []AdminPermission, err error) {\n\to := orm.NewOrm()\n\tv = []AdminPermission{}\n\t_, err = o.QueryTable(new(AdminPermission)).RelatedSel().All(&v)\n\tfmt.Println(v)\n\treturn v, err\n}", "func (auup *AuthUserUserPermission) Save(ctx context.Context, db DB) error {\n\tif auup.Exists() {\n\t\treturn auup.Update(ctx, db)\n\t}\n\treturn auup.Insert(ctx, db)\n}", "func (_Bucket *BucketTransactor) SetWritePermission(opts *bind.TransactOpts, _fileId *big.Int, _entity common.Address, _permission bool) (*types.Transaction, error) {\n\treturn _Bucket.contract.Transact(opts, \"setWritePermission\", _fileId, _entity, _permission)\n}", "func withPermission(node *Permission) permissionOption {\n\treturn func(m *PermissionMutation) {\n\t\tm.oldValue = func(context.Context) (*Permission, error) {\n\t\t\treturn node, nil\n\t\t}\n\t\tm.id = &node.ID\n\t}\n}", "func AddAdminPermission(roleID int, ActionID int, Enable int8) (userRole []orm.Params, err error) {\n\to := orm.NewOrm()\n\t_, err = o.Raw(\"INSERT INTO `_admin_permission` (`role_id`,`action_id`,`enable`) VALUES (?, ?, ?)\", &roleID, &ActionID, &Enable).Values(&userRole)\n\tfmt.Println(\"=========\", userRole)\n\treturn userRole, err\n}", "func (_PermInterface *PermInterfaceSession) TransactionAllowed(_sender common.Address, _target common.Address, _value *big.Int, _gasPrice *big.Int, _gasLimit *big.Int, _payload []byte) (bool, error) {\n\treturn _PermInterface.Contract.TransactionAllowed(&_PermInterface.CallOpts, _sender, _target, _value, _gasPrice, _gasLimit, _payload)\n}", "func (_Bucket *BucketTransactorSession) SetWritePermission(_fileId *big.Int, _entity common.Address, _permission bool) (*types.Transaction, error) {\n\treturn _Bucket.Contract.SetWritePermission(&_Bucket.TransactOpts, _fileId, _entity, _permission)\n}", "func (o BucketLoggingV2TargetGrantOutput) Permission() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketLoggingV2TargetGrant) string { return v.Permission }).(pulumi.StringOutput)\n}", "func (_Storage *StorageSession) PermissionAt(kind uint8, index *big.Int) (common.Address, error) {\n\treturn _Storage.Contract.PermissionAt(&_Storage.CallOpts, kind, index)\n}", "func (_PermInterface *PermInterfaceCaller) TransactionAllowed(opts *bind.CallOpts, _sender common.Address, _target common.Address, _value *big.Int, _gasPrice *big.Int, _gasLimit *big.Int, _payload []byte) (bool, error) {\n\tvar out []interface{}\n\terr := _PermInterface.contract.Call(opts, &out, \"transactionAllowed\", _sender, _target, _value, _gasPrice, _gasLimit, _payload)\n\n\tif err != nil {\n\t\treturn *new(bool), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(bool)).(*bool)\n\n\treturn out0, err\n\n}", "func (authService *Controller) populatePermission() error {\n\tpeeringPermission, err := peeringRoles.GetPeeringPermission(authService.clientset, authService)\n\tif err != nil {\n\t\tklog.Error(err)\n\t\treturn err\n\t}\n\n\tif peeringPermission != nil {\n\t\tauthService.peeringPermission = *peeringPermission\n\t}\n\treturn nil\n}", "func (ResourceType) Policy() ent.Policy {\n\t// TODO setup RBAC policies for entities (RBAC based on user's role) such as:\n\t// return authz.NewPolicy(\n\t// \tauthz.WithMutationRules(\n\t// \t\tauthz.ResourceTypeWritePolicyRule(),\n\t// \t),\n\t// )\n\treturn nil\n}", "func (_PermInterface *PermInterfaceCallerSession) TransactionAllowed(_sender common.Address, _target common.Address, _value *big.Int, _gasPrice *big.Int, _gasLimit *big.Int, _payload []byte) (bool, error) {\n\treturn _PermInterface.Contract.TransactionAllowed(&_PermInterface.CallOpts, _sender, _target, _value, _gasPrice, _gasLimit, _payload)\n}", "func GetAllPermissions(c *gin.Context) {\r\n\trole := model.Role{}\r\n\tmenus, err := role.GetMenus()\r\n\tif err != nil {\r\n\t\tutils.InternalError(c, nil, \"数据库操作失败\")\r\n\t\treturn\r\n\t}\r\n\tpermissions, err := role.GetPermissions()\r\n\tif err != nil {\r\n\t\tutils.InternalError(c, nil, \"数据库操作失败\")\r\n\t\treturn\r\n\t}\r\n\tutils.Success(c, gin.H{\"menus\": menus, \"permissions\": permissions}, \"权限查询成功\")\r\n}", "func (m AbilitypatientrightsMutation) Tx() (*Tx, error) {\n\tif _, ok := m.driver.(*txDriver); !ok {\n\t\treturn nil, fmt.Errorf(\"ent: mutation is not running in a transaction\")\n\t}\n\ttx := &Tx{config: m.config}\n\ttx.init()\n\treturn tx, nil\n}", "func (auup *AuthUserUserPermission) AuthPermission(ctx context.Context, db DB) (*AuthPermission, error) {\n\treturn AuthPermissionByID(ctx, db, auup.PermissionID)\n}", "func (q queryManager) checkExecutePermissions(qp dbquery.QueryParsed, pubKey []byte) (bool, error) {\n\treturn true, nil\n}", "func UpdatePermission(c *gin.Context) {\n\tupdatePermission := model.Permission{}\n\tc.BindJSON(&updatePermission)\n\n\terr := service.UpdatePermission(updatePermission)\n\n\tif err != nil {\n\t\tfmt.Println(err.Error())\n\t\terror := service.GetGormErrorCode(err.Error())\n\n\t\tc.JSON(500, error)\n\t} else {\n\t\tc.String(200, \"ok\")\n\t}\n}", "func (s *Service) Permissions(c context.Context, username string) (res *model.Permissions, err error) {\n\tuser := &model.User{}\n\tres = new(model.Permissions)\n\t//根据username 招到userid\n\tres.UID = user.ID\n\tres.Perms = make([]string, 10)\n\tres.Admin = true\n\treturn\n}", "func bindPermInterface(address common.Address, caller bind.ContractCaller, transactor bind.ContractTransactor, filterer bind.ContractFilterer) (*bind.BoundContract, error) {\n\tparsed, err := abi.JSON(strings.NewReader(PermInterfaceABI))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn bind.NewBoundContract(address, parsed, caller, transactor, filterer), nil\n}", "func (o BucketAclV2AccessControlPolicyGrantOutput) Permission() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketAclV2AccessControlPolicyGrant) string { return v.Permission }).(pulumi.StringOutput)\n}", "func (m AuthorizeMutation) Tx() (*Tx, error) {\n\tif _, ok := m.driver.(*txDriver); !ok {\n\t\treturn nil, fmt.Errorf(\"ent: mutation is not running in a transaction\")\n\t}\n\ttx := &Tx{config: m.config}\n\ttx.init()\n\treturn tx, nil\n}", "func (ctx *Context) PrincipalDB(ros ...dbRequestReadOnly) *sql.DB {\n\tvar ro bool\n\tif len(ros) > 0 {\n\t\tfor _, r := range ros {\n\t\t\tif r {\n\t\t\t\tro = true\n\t\t\t}\n\t\t}\n\t}\n\tif !ro {\n\t\treturn ctx.principalDBWrite\n\t}\n\tif ctx.principalDBReadOnly == nil {\n\t\treturn ctx.principalDBWrite\n\t}\n\treturn ctx.principalDBReadOnly\n}", "func (_Bucket *BucketSession) SetWritePermission(_fileId *big.Int, _entity common.Address, _permission bool) (*types.Transaction, error) {\n\treturn _Bucket.Contract.SetWritePermission(&_Bucket.TransactOpts, _fileId, _entity, _permission)\n}", "func (s *API) AddPermission(w http.ResponseWriter, req *http.Request) {\n\tlog.Debug(\"AddPermission\")\n\tw.WriteHeader(http.StatusNotImplemented)\n}", "func (ppc *PermissionsPolicyCreate) Mutation() *PermissionsPolicyMutation {\n\treturn ppc.mutation\n}", "func (authUserUserPermissionL) LoadPermission(e boil.Executor, singular bool, maybeAuthUserUserPermission interface{}) error {\n\tvar slice []*AuthUserUserPermission\n\tvar object *AuthUserUserPermission\n\n\tcount := 1\n\tif singular {\n\t\tobject = maybeAuthUserUserPermission.(*AuthUserUserPermission)\n\t} else {\n\t\tslice = *maybeAuthUserUserPermission.(*AuthUserUserPermissionSlice)\n\t\tcount = len(slice)\n\t}\n\n\targs := make([]interface{}, count)\n\tif singular {\n\t\tif object.R == nil {\n\t\t\tobject.R = &authUserUserPermissionR{}\n\t\t}\n\t\targs[0] = object.PermissionID\n\t} else {\n\t\tfor i, obj := range slice {\n\t\t\tif obj.R == nil {\n\t\t\t\tobj.R = &authUserUserPermissionR{}\n\t\t\t}\n\t\t\targs[i] = obj.PermissionID\n\t\t}\n\t}\n\n\tquery := fmt.Sprintf(\n\t\t\"select * from `auth_permission` where `id` in (%s)\",\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, count, 1, 1),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintf(boil.DebugWriter, \"%s\\n%v\\n\", query, args)\n\t}\n\n\tresults, err := e.Query(query, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to eager load AuthPermission\")\n\t}\n\tdefer results.Close()\n\n\tvar resultSlice []*AuthPermission\n\tif err = queries.Bind(results, &resultSlice); err != nil {\n\t\treturn errors.Wrap(err, \"failed to bind eager loaded slice AuthPermission\")\n\t}\n\n\tif len(resultSlice) == 0 {\n\t\treturn nil\n\t}\n\n\tif singular {\n\t\tobject.R.Permission = resultSlice[0]\n\t\treturn nil\n\t}\n\n\tfor _, local := range slice {\n\t\tfor _, foreign := range resultSlice {\n\t\t\tif local.PermissionID == foreign.ID {\n\t\t\t\tlocal.R.Permission = foreign\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (i *Inventory) Permission() Permission {\n\tif i.devicePermission == nil {\n\t\ti.devicePermission = CreateDevicePermission(i.Raptor)\n\t}\n\treturn i.devicePermission\n}", "func (o KafkaAclOutput) Permission() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *KafkaAcl) pulumi.StringOutput { return v.Permission }).(pulumi.StringOutput)\n}", "func (_Storage *StorageCallerSession) PermissionAt(kind uint8, index *big.Int) (common.Address, error) {\n\treturn _Storage.Contract.PermissionAt(&_Storage.CallOpts, kind, index)\n}", "func (us UserService) CheckPermission(uid string, domain string, policy string) (bool, error) {\n\t//Could it be an alias?\n\tdomainModel := domainDao.GetByCode(domain)\n\trow := menuPermAliasDao.GetByAlias(policy, domainModel.Id)\n\tif row.Id > 0 {\n\t\tpolicy = row.Perms\n\t}\n\tlog.Info(fmt.Sprintf(\"check permission for : %#v,%#v,%#v\", uid, policy, domain))\n\treturn perm.Enforce(uid, policy, \"*\", domain)\n}", "func (m *Manager) GrantPermission(sourceAddr net.Addr, peerIP net.IP) error {\n\t// no permission handler: open\n\tif m.permissionHandler == nil {\n\t\treturn nil\n\t}\n\n\tif m.permissionHandler(sourceAddr, peerIP) {\n\t\treturn nil\n\t}\n\n\treturn errAdminProhibited\n}", "func (bc *BaseController) UpatePermission(permission *dream.Permission) (*dream.Permission, error) {\n\n\tresult, err := dream.UpdatePermission(bc.DB, permission)\n\n\tif err != nil {\n\t\treturn nil, ServiceErr\n\t}\n\n\treturn result, nil\n}", "func hasPermission(e *Engine, tableName string, userId int64, rowUID int64, rowGID int64, f func(Permissions) bool) bool {\n\tperms, ok := e.GetTablePermissions(tableName)\n\t\n\t//Default to disallowing all actions in the absence of permissions\t\n\tif !ok { return false }\n\t\n\tif f(perms.Everyone) {\n\t\treturn true\n\t} else if f(perms.Owner) && userId == rowUID {\n\t\treturn true\n\t} else if f(perms.Group) {\n\t\t//Test if user is in group.\n\t\tb, err := UserInGroup(e, userId, rowGID)\n\t\treturn b && err == nil\n\t}\n\treturn false\n}", "func (_Storage *StorageCaller) Permissions(opts *bind.CallOpts, arg0 uint8) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _Storage.contract.Call(opts, out, \"permissions\", arg0)\n\treturn *ret0, err\n}", "func (ui *UserInfo) AuthorizeDatabase(privilege influxql.Privilege, database string) bool {\n\tif ui.Admin || privilege == influxql.NoPrivileges {\n\t\treturn true\n\t}\n\tp, ok := ui.Privileges[database]\n\treturn ok && (p == privilege || p == influxql.AllPrivileges)\n}", "func IsPermission(err error) bool", "func addPerm(w http.ResponseWriter, r *http.Request) {\r\n\tkey := r.URL.Query().Get(\"key\")\r\n\tp, err := checkKey(key, perm)\r\n\tif err != nil {\r\n\t\thttpError{statusCode: http.StatusForbidden, err: err}.ServeHTTP(w, r)\r\n\t\treturn\r\n\t}\r\n\tif !p.IsWritable {\r\n\t\thttpError{statusCode: http.StatusNotImplemented, err: errors.New(\"Method not implemented\")}.ServeHTTP(w, r)\r\n\t\treturn\r\n\t}\r\n\terr = r.ParseForm()\r\n\tif err != nil {\r\n\t\thttpError{statusCode: http.StatusBadRequest, err: errors.New(\"Bad request\")}.ServeHTTP(w, r)\r\n\t\treturn\r\n\t}\r\n\tnewToken := r.FormValue(\"token\")\r\n\tname := r.FormValue(\"name\")\r\n\tif urls, ok := r.Form[\"url\"]; ok && len(newToken) > 5 && len(name) > 0 {\r\n\t\tvar allPerm []dto.Permission\r\n\t\tfor _, v := range urls {\r\n\t\t\tpair := strings.Split(v, \":\")\r\n\t\t\tif len(pair) == 1 {\r\n\t\t\t\tallPerm = append(allPerm, dto.Permission{URL: pair[0], IsWritable: false})\r\n\t\t\t} else if len(pair) == 2 {\r\n\t\t\t\tflag, _ := strconv.ParseBool(pair[1])\r\n\t\t\t\tallPerm = append(allPerm, dto.Permission{URL: pair[0], IsWritable: flag})\r\n\t\t\t}\r\n\t\t}\r\n\t\tcl := dto.ClientPermission{\r\n\t\t\tName: name,\r\n\t\t\tKey: newToken,\r\n\t\t\tPerm: allPerm,\r\n\t\t}\r\n\t\tlog.Infof(\"Create new tocken %s with permision %v\", cl.Key, cl.Perm)\r\n\t\tstorage, ok := c2cData.GetBoltDbInstance().(c2cData.IPerm)\r\n\t\tif !ok {\r\n\t\t\thttpError{statusCode: http.StatusInternalServerError, err: errors.New(\"Database not supported permissions\")}.ServeHTTP(w, r)\r\n\t\t}\r\n\t\tif err = storage.UpdatePermission(cl); err != nil {\r\n\t\t\thttpError{statusCode: http.StatusInternalServerError, err: errors.New(\"Can not save permission\")}.ServeHTTP(w, r)\r\n\t\t}\r\n\t\tw.Header().Add(\"Content-Type\", \"application/json\")\r\n\t\tw.WriteHeader(http.StatusCreated)\r\n\t\tb, _ := json.Marshal(cl)\r\n\t\tw.Write(b)\r\n\t\treturn\r\n\t}\r\n\thttpError{statusCode: http.StatusBadRequest, err: errors.New(\"Incorrect body in post request\")}.ServeHTTP(w, r)\r\n}", "func requiresPermission(eh endpointHandler, permissionNames []string) endpointHandler {\n\tlog.Trace(\"resource/service:requiresPermission() Entering\")\n\tdefer log.Trace(\"resource/service:requiresPermission() Leaving\")\n\treturn func(w http.ResponseWriter, r *http.Request) error {\n\t\tprivileges, err := commContext.GetUserPermissions(r)\n\t\tif err != nil {\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t\tw.Header().Add(\"Strict-Transport-Security\", \"max-age=63072000; includeSubDomains\")\n\t\t\t_, writeErr := w.Write([]byte(\"Could not get user roles from http context\"))\n\t\t\tif writeErr != nil {\n\t\t\t\tlog.WithError(writeErr).Warn(\"resource/service:requiresPermission() Error while writing response\")\n\t\t\t}\n\t\t\tsecLog.Errorf(\"resource/service:requiresPermission() %s Roles: %v | Context: %v\", message.AuthenticationFailed, permissionNames, r.Context())\n\t\t\treturn errors.Wrap(err, \"resource/service:requiresPermission() Could not get user roles from http context\")\n\t\t}\n\t\treqPermissions := ct.PermissionInfo{Service: constants.TAServiceName, Rules: permissionNames}\n\n\t\t_, foundMatchingPermission := auth.ValidatePermissionAndGetPermissionsContext(privileges, reqPermissions,\n\t\t\ttrue)\n\t\tif !foundMatchingPermission {\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\tsecLog.Errorf(\"resource/service:requiresPermission() %s Insufficient privileges to access %s\", message.UnauthorizedAccess, r.RequestURI)\n\t\t\treturn &privilegeError{Message: \"Insufficient privileges to access \" + r.RequestURI, StatusCode: http.StatusUnauthorized}\n\t\t}\n\t\tw.Header().Add(\"Strict-Transport-Security\", \"max-age=63072000; includeSubDomains\")\n\t\tsecLog.Debugf(\"resource/service:requiresPermission() %s - %s\", message.AuthorizedAccess, r.RequestURI)\n\t\treturn eh(w, r)\n\t}\n}", "func (c Chart) PermissionResource() permissions.Resource {\n\treturn ChartPermissionResource.AppendID(c.ID)\n}", "func (a *Authorizer) DettachPermission(permName string, roleName string) error {\n\n\t// Make sure permission exist\n\tp := permission{}\n\n\terr := a.db.One(\"Name\", permName, &p)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tr := role{}\n\terr = a.db.One(\"Name\", roleName, &r)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Look if it's already there\n\tfor i, cp := range r.Permissions {\n\t\tif cp.Name == permName {\n\t\t\tr.Permissions[i] = r.Permissions[len(r.Permissions)-1]\n\t\t\tr.Permissions[len(r.Permissions)-1] = permission{}\n\t\t\tr.Permissions = r.Permissions[:len(r.Permissions)-1]\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// Save\n\terr = a.db.Save(&r)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn a.load()\n\n}", "func (_Storage *StorageCaller) PermissionAt(opts *bind.CallOpts, kind uint8, index *big.Int) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _Storage.contract.Call(opts, out, \"permissionAt\", kind, index)\n\treturn *ret0, err\n}", "func (o *Constraint) Create(db *gorm.DB) error {\n\treturn db.Create(o).Error\n}", "func (sp *SalePermission) Save(ctx context.Context) error {\n\tif sp.Exists() {\n\t\treturn sp.Update(ctx)\n\t}\n\n\treturn sp.Insert(ctx)\n}", "func (o *AggregatedDomain) CreatePermission(child *Permission) *bambou.Error {\n\n\treturn bambou.CurrentSession().CreateChild(o, child)\n}", "func (db *Permstore) PutPerm(perm *model.Perm) error {\n\treturn meddler.Save(db, permTable, perm)\n}", "func (r *mutationResolver) CreateRolePermissionRelationship(ctx context.Context, roleID string, permissionID string) (*user.Role, error) {\n\tpanic(\"not implemented\")\n}", "func (as *AclStatus) Permission() *hdfs.FsPermissionProto {\n\treturn as.aclStatus.GetPermission()\n}", "func (this *managerStruct) AddPermission(name string, permission string) error {\n\tthis.mutex.Lock()\n\tid := this.getUserId(name)\n\n\t/*\n\t * Check if we have a user with the name provided to us.\n\t */\n\tif id < 0 {\n\t\tthis.mutex.Unlock()\n\t\treturn fmt.Errorf(\"User '%s' does not exist.\", name)\n\t} else {\n\t\tusers := this.users\n\t\tuser := users[id]\n\t\tpermissions := user.permissions\n\t\texists := false\n\n\t\t/*\n\t\t * Check if user already has permission.\n\t\t */\n\t\tfor _, currentPermission := range permissions {\n\n\t\t\t/*\n\t\t\t * Check for permission.\n\t\t\t */\n\t\t\tif currentPermission == permission {\n\t\t\t\texists = true\n\t\t\t}\n\n\t\t}\n\n\t\t/*\n\t\t * Add permission to user if he / she does not already have it.\n\t\t */\n\t\tif !exists {\n\t\t\tuser.permissions = append(permissions, permission)\n\t\t}\n\n\t\tusers[id] = user\n\t\tthis.mutex.Unlock()\n\t\treturn nil\n\t}\n\n}", "func PermissionInit(\n\tcasbinAuth casbin.CasbinAuth,\n\t trans ut.Translator,\n\n\t ) PermissionHandler {\n\treturn &permissionHandler{\n\t\tcasbinAuth,\n\t\ttrans,\n\n\t}\n}", "func PermissionProxy(in interface{}, out interface{}) {\n\tra := reflect.ValueOf(in)\n\touts := proxy.GetInternalStructs(out)\n\tfor _, out := range outs {\n\t\trint := reflect.ValueOf(out).Elem()\n\t\tfor i := 0; i < ra.NumMethod(); i++ {\n\t\t\tmethodName := ra.Type().Method(i).Name\n\t\t\tfield, exists := rint.Type().FieldByName(methodName)\n\t\t\tif !exists {\n\t\t\t\t// log.Printf(\"exclude method %s from fullNode\", methodName)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\trequiredPerm := field.Tag.Get(\"perm\")\n\t\t\tif requiredPerm == \"\" {\n\t\t\t\tpanic(\"missing 'perm' tag on \" + field.Name) // ok\n\t\t\t}\n\t\t\tcurule := defaultRule()\n\t\t\tcurule.Perm = requiredPerm\n\n\t\t\tfn := ra.Method(i)\n\t\t\trint.FieldByName(methodName).Set(reflect.MakeFunc(field.Type, func(args []reflect.Value) (results []reflect.Value) {\n\t\t\t\tctx := args[0].Interface().(context.Context)\n\t\t\t\terrNum := 0\n\t\t\t\tif !auth.HasPerm(ctx, defaultPerms, curule.Perm) {\n\t\t\t\t\terrNum++\n\t\t\t\t\tgoto ABORT\n\t\t\t\t}\n\t\t\t\treturn fn.Call(args)\n\t\t\tABORT:\n\t\t\t\terr := xerrors.Errorf(\"missing permission to invoke '%s'\", methodName)\n\t\t\t\tif errNum&1 == 1 {\n\t\t\t\t\terr = xerrors.Errorf(\"%s (need '%s')\", err, curule.Perm)\n\t\t\t\t}\n\t\t\t\trerr := reflect.ValueOf(&err).Elem()\n\t\t\t\tif fn.Type().NumOut() == 2 {\n\t\t\t\t\treturn []reflect.Value{\n\t\t\t\t\t\treflect.Zero(fn.Type().Out(0)),\n\t\t\t\t\t\trerr,\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\treturn []reflect.Value{rerr}\n\t\t\t}))\n\t\t}\n\t}\n\n}", "func (user *User) BeforeCreate(transaction *gorm.DB) error {\n return nil\n}", "func (pe *Permission) Unwrap() *Permission {\n\ttx, ok := pe.config.driver.(*txDriver)\n\tif !ok {\n\t\tpanic(\"models: Permission is not a transactional entity\")\n\t}\n\tpe.config.driver = tx.drv\n\treturn pe\n}", "func checkPermission(c *gin.Context, enforcer *casbin.Enforcer, d *gorm.DB, user *db.User, roles []string) bool {\n\tr := c.Request\n\tmethod := r.Method\n\tpath := r.URL.Path\n\tusergroup := db.UserGroup{}\n\td.First(&usergroup, user.UserGroupID)\n\tfor _, role := range roles {\n\t\treturn enforcer.Enforce(usergroup.GroupName, role, path, method)\n\t}\n\treturn false\n}", "func (p Permission) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"actions\", p.Actions)\n\tpopulate(objectMap, \"dataActions\", p.DataActions)\n\tpopulate(objectMap, \"notActions\", p.NotActions)\n\tpopulate(objectMap, \"notDataActions\", p.NotDataActions)\n\treturn json.Marshal(objectMap)\n}", "func newPermissionMutation(c config, op Op, opts ...permissionOption) *PermissionMutation {\n\tm := &PermissionMutation{\n\t\tconfig: c,\n\t\top: op,\n\t\ttyp: TypePermission,\n\t\tclearedFields: make(map[string]struct{}),\n\t}\n\tfor _, opt := range opts {\n\t\topt(m)\n\t}\n\treturn m\n}", "func Schema() string {\n\n\t/**\n\tDROP DATABASE users;\n\tDROP DATABASE token;\n\t*/\n\treturn `\n\n\tCREATE TABLE IF NOT EXISTS users (\n\t\tid SERIAL UNIQUE,\n\t\tusername text,\n\t\temail text UNIQUE,\n\t\tpassword text,\n\t\tsecureLevel text DEFAULT 'user',\n\t\tpathfile text,\n\t\tcreated_at TIMESTAMP NOT NULL DEFAULT CURRENT_DATE\n\t\tupdated_at TIMESTAMP NOT NULL DEFAULT CURRENT_DATE\n\t);\n\n\tCREATE TABLE IF NOT EXISTS token (\n\t\ttoken text NOT NULL PRIMARY KEY,\n\t\tis_revoked bool DEFAULT FALSE,\n\t\tuser_id INTEGER REFERENCES users(id) NOT NULL,\n\t\tcreated_at TIMESTAMP NOT NULL DEFAULT CURRENT_DATE\n\t\tupdated_at TIMESTAMP NOT NULL DEFAULT CURRENT_DATE\n\t);\n\n\tCREATE TABLE IF NOT EXISTS notification (\n\t\ttokenNotification text NOT NULL PRIMARY KEY,\n\t\tuser_id INTEGER REFERENCES users(id) NOT NULL,\n\t\tcreated_at TIMESTAMP NOT NULL DEFAULT CURRENT_DATE\n\t\tupdated_at TIMESTAMP NOT NULL DEFAULT CURRENT_DATE\n\t);\n\n\tCREATE TABLE IF NOT EXISTS dataNotification (\n\t\tid SERIAL UNIQUE,\n\t\tuser_id INTEGER REFERENCES users(id) NOT NULL,\n\t\ttitle text,\n\t\tbody text,\n\t\tcreated_at TIMESTAMP NOT NULL DEFAULT CURRENT_DATE\n\t\tupdated_at TIMESTAMP NOT NULL DEFAULT CURRENT_DATE\n\t);\n\n\tCREATE TABLE IF NOT EXISTS email (\n\t\tid SERIAL UNIQUE,\n\t\tuser_id INTEGER REFERENCES users(id) NOT NULL,\n\t\tto text,\n\t\tmsg text,\n\t\tcreated_at TIMESTAMP NOT NULL DEFAULT CURRENT_DATE\n\t\tupdated_at TIMESTAMP NOT NULL DEFAULT CURRENT_DATE\n\t);\n\n\tCREATE TABLE IF NOT EXISTS dataHistory (\n\t\tid SERIAL UNIQUE,\n\t\ttotalUserCount INTEGER DEFAULT 0,\n\t\tuserCount INTEGER DEFAULT 0,\n\t\tnotificationCount INTEGER DEFAULT 0,\n\t\temailCount INTEGER DEFAULT 0,\n\t\tcreated_at TIMESTAMP NOT NULL DEFAULT CURRENT_DATE\n\t\tupdated_at TIMESTAMP NOT NULL DEFAULT CURRENT_DATE\n\t);\n\n\t`\n\n}", "func main() {\n\tdb, err := db.OpenConnection()\n\tdb.LogMode(true)\n\n\tif err != nil {\n\t\tlog.WithError(err).Fatal(\"An error occurred while connecting to the database\")\n\t}\n\tdefer db.Close()\n\n\t// Create models\n\tdb.DropTableIfExists(&domain.User{}, &domain.Email{}, &domain.Role{})\n\tdb.CreateTable(&domain.Email{}, &domain.Role{}, &domain.User{})\n\n\t// Create user\n\tid, _ := uuid.Parse(\"8c5df3bc-9fa6-4d73-b79a-9a1cbb35740c\")\n\tuser := domain.User{\n\t\tID: id,\n\t\tEmails: []domain.Email{\n\t\t\t{Email: \"[email protected]\"},\n\t\t},\n\t\tRoles: []domain.Role{\n\t\t\t{Role: \"ROLE_USER\"},\n\t\t},\n\t\tPassword: \"Test1234\",\n\t}\n\n\t// Create developer\n\tid, _ = uuid.Parse(\"b0203081-5dfe-4bb7-87d1-e2c59e2af7b6\")\n\tdeveloper := domain.User{\n\t\tID: id,\n\t\tEmails: []domain.Email{\n\t\t\t{Email: \"[email protected]\"},\n\t\t},\n\t\tRoles: []domain.Role{\n\t\t\t{Role: \"ROLE_USER\"},\n\t\t\t{Role: \"ROLE_DEVELOPER\"},\n\t\t},\n\t\tPassword: \"Test1234\",\n\t}\n\n\t// Create hr\n\tid, _ = uuid.Parse(\"6b59c645-82c3-4e08-b089-f4236a2141b6\")\n\thr := domain.User{\n\t\tID: id,\n\t\tEmails: []domain.Email{\n\t\t\t{Email: \"[email protected]\"},\n\t\t},\n\t\tRoles: []domain.Role{\n\t\t\t{Role: \"ROLE_USER\"},\n\t\t\t{Role: \"ROLE_HR\"},\n\t\t},\n\t\tPassword: \"Test1234\",\n\t}\n\n\t// Create manager\n\tid, _ = uuid.Parse(\"75129bb5-5c12-48a1-8410-bb2630fff9ed\")\n\tmanager := domain.User{\n\t\tID: id,\n\t\tEmails: []domain.Email{\n\t\t\t{Email: \"[email protected]\"},\n\t\t\t{Email: \"[email protected]\"},\n\t\t},\n\t\tRoles: []domain.Role{\n\t\t\t{Role: \"ROLE_USER\"},\n\t\t\t{Role: \"ROLE_MANAGER\"},\n\t\t},\n\t\tPassword: \"Test1234\",\n\t}\n\n\tdb.Create(&user)\n\tdb.Create(&developer)\n\tdb.Create(&hr)\n\tdb.Create(&manager)\n}", "func (dp *dataProvider) Tx(fn func(*pg.Tx) error) error {\n\treturn wrapError(dp.db.RunInTransaction(func(tx *pg.Tx) error {\n\t\tdefer func(t *pg.Tx) {\n\t\t\tif err := recover(); err != nil {\n\t\t\t\tt.Rollback()\n\t\t\t\t// rethrow the panic once the database is safe\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t}(tx)\n\t\treturn fn(tx)\n\t}))\n}", "func AuthorizeCreate(db *gorm.DB, data *Authorize) (err error) {\n\tdata.Ctime = time.Now().Unix()\n\n\tif err = db.Create(data).Error; err != nil {\n\t\tmus.Logger.Error(\"create authorize error\", zap.Error(err))\n\t\treturn\n\t}\n\treturn\n}", "func (permission *Permission) Allow(mode PermissionMode, roles ...string) *Permission{\n\tif mode == CRUD {\n\t\treturn permission.Allow(Create, roles...).Allow(Update, roles...).Allow(Read, roles...).Allow(Delete, roles...)\n\t}\n\n\tif permission.AllowedRoles[mode] == nil {\n\t\tpermission.AllowedRoles[mode] = []string{}\n\t}\n\tpermission.AllowedRoles[mode] = append(permission.AllowedRoles[mode], roles...)\n\treturn permission\n}", "func (o *Port) CreatePermission(child *Permission) *bambou.Error {\n\n\treturn bambou.CurrentSession().CreateChild(o, child)\n}" ]
[ "0.57659733", "0.562225", "0.5615602", "0.559084", "0.5548221", "0.5522893", "0.548699", "0.54744476", "0.54559094", "0.5383814", "0.5369931", "0.53067356", "0.5277368", "0.5224058", "0.520029", "0.5188981", "0.51543236", "0.50971645", "0.5032364", "0.5014209", "0.49922004", "0.4989081", "0.49621975", "0.49581578", "0.4894114", "0.48494533", "0.48253182", "0.48037213", "0.4800279", "0.47924787", "0.4778681", "0.47739252", "0.4764803", "0.4753442", "0.47534207", "0.47349066", "0.4732796", "0.4727382", "0.4724059", "0.4708516", "0.4693867", "0.46912888", "0.46866518", "0.4680441", "0.4672662", "0.46568817", "0.46502015", "0.4647374", "0.46357933", "0.46335992", "0.46243516", "0.46158707", "0.4600892", "0.4597732", "0.45928022", "0.45907202", "0.4589297", "0.45877054", "0.45823762", "0.4579102", "0.45749307", "0.4573419", "0.4567122", "0.456265", "0.4562314", "0.45599657", "0.4554871", "0.45492345", "0.45469868", "0.45437285", "0.45388335", "0.45300394", "0.45289472", "0.45169416", "0.45133507", "0.4508323", "0.4501731", "0.4499938", "0.44988945", "0.44935247", "0.4479813", "0.44794703", "0.44672576", "0.44578895", "0.44557813", "0.4443585", "0.44432044", "0.4441062", "0.4437571", "0.44372058", "0.4431709", "0.44304237", "0.44209576", "0.44128987", "0.44007447", "0.43992442", "0.43906882", "0.43899179", "0.43886915", "0.43839043" ]
0.6913373
0
Rule will inject the databaseTx in the `Rule` schema
func (gtx *GuardTx) Rule(rule *schema.Rule) *schema.Rule { if rule == nil { return &schema.Rule{ Entity: schema.Entity{DBContract: gtx.dbTx}, } } rule.DBContract = gtx.dbTx rule.SetValidator(gtx.validator.Rule) return rule }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (fw *IPtables) injectRule(rule *IPtablesRule) error {\n\trule2arr := strings.Split(rule.GetBody(), \" \")\n\tif len(rule2arr) < 3 {\n\t\treturn fmt.Errorf(\"In injectRule() not enough elements in rule %s\", rule.GetBody())\n\t}\n\n\truleChain := rule2arr[0]\n\n\tfor i, chain := range fw.chains {\n\t\tif chain.ChainName == ruleChain {\n\t\t\tfw.chains[i].Rules = append(fw.chains[i].Rules, rule)\n\t\t\tlog.Infof(\"In injectRule() adding rule %s into chain %s\", rule.GetBody(), chain.ChainName)\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t// TODO should we create new chain instead of throwing error?\n\treturn fmt.Errorf(\"In injectRule() firewall doesn't manage chain for rule %s\", rule.GetBody())\n}", "func registerRule(app *extkingpin.App) {\n\tcomp := component.Rule\n\tcmd := app.Command(comp.String(), \"Ruler evaluating Prometheus rules against given Query nodes, exposing Store API and storing old blocks in bucket.\")\n\n\tconf := &ruleConfig{}\n\tconf.registerFlag(cmd)\n\n\tlabelStrs := cmd.Flag(\"label\", \"Labels to be applied to all generated metrics (repeated). Similar to external labels for Prometheus, used to identify ruler and its blocks as unique source.\").\n\t\tPlaceHolder(\"<name>=\\\"<value>\\\"\").Strings()\n\ttsdbBlockDuration := extkingpin.ModelDuration(cmd.Flag(\"tsdb.block-duration\", \"Block duration for TSDB block.\").\n\t\tDefault(\"2h\"))\n\ttsdbRetention := extkingpin.ModelDuration(cmd.Flag(\"tsdb.retention\", \"Block retention time on local disk.\").\n\t\tDefault(\"48h\"))\n\tnoLockFile := cmd.Flag(\"tsdb.no-lockfile\", \"Do not create lockfile in TSDB data directory. In any case, the lockfiles will be deleted on next startup.\").Default(\"false\").Bool()\n\twalCompression := cmd.Flag(\"tsdb.wal-compression\", \"Compress the tsdb WAL.\").Default(\"true\").Bool()\n\n\tcmd.Flag(\"data-dir\", \"data directory\").Default(\"data/\").StringVar(&conf.dataDir)\n\tcmd.Flag(\"rule-file\", \"Rule files that should be used by rule manager. Can be in glob format (repeated). Note that rules are not automatically detected, use SIGHUP or do HTTP POST /-/reload to re-read them.\").\n\t\tDefault(\"rules/\").StringsVar(&conf.ruleFiles)\n\tcmd.Flag(\"resend-delay\", \"Minimum amount of time to wait before resending an alert to Alertmanager.\").\n\t\tDefault(\"1m\").DurationVar(&conf.resendDelay)\n\tcmd.Flag(\"eval-interval\", \"The default evaluation interval to use.\").\n\t\tDefault(\"1m\").DurationVar(&conf.evalInterval)\n\tcmd.Flag(\"for-outage-tolerance\", \"Max time to tolerate prometheus outage for restoring \\\"for\\\" state of alert.\").\n\t\tDefault(\"1h\").DurationVar(&conf.outageTolerance)\n\tcmd.Flag(\"for-grace-period\", \"Minimum duration between alert and restored \\\"for\\\" state. This is maintained only for alerts with configured \\\"for\\\" time greater than grace period.\").\n\t\tDefault(\"10m\").DurationVar(&conf.forGracePeriod)\n\tcmd.Flag(\"restore-ignored-label\", \"Label names to be ignored when restoring alerts from the remote storage. This is only used in stateless mode.\").\n\t\tStringsVar(&conf.ignoredLabelNames)\n\n\tconf.rwConfig = extflag.RegisterPathOrContent(cmd, \"remote-write.config\", \"YAML config for the remote-write configurations, that specify servers where samples should be sent to (see https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write). This automatically enables stateless mode for ruler and no series will be stored in the ruler's TSDB. If an empty config (or file) is provided, the flag is ignored and ruler is run with its own TSDB.\", extflag.WithEnvSubstitution())\n\n\treqLogDecision := cmd.Flag(\"log.request.decision\", \"Deprecation Warning - This flag would be soon deprecated, and replaced with `request.logging-config`. Request Logging for logging the start and end of requests. By default this flag is disabled. LogFinishCall: Logs the finish call of the requests. LogStartAndFinishCall: Logs the start and finish call of the requests. NoLogCall: Disable request logging.\").Default(\"\").Enum(\"NoLogCall\", \"LogFinishCall\", \"LogStartAndFinishCall\", \"\")\n\n\tconf.objStoreConfig = extkingpin.RegisterCommonObjStoreFlags(cmd, \"\", false)\n\n\treqLogConfig := extkingpin.RegisterRequestLoggingFlags(cmd)\n\n\tvar err error\n\tcmd.Setup(func(g *run.Group, logger log.Logger, reg *prometheus.Registry, tracer opentracing.Tracer, reload <-chan struct{}, _ bool) error {\n\t\tconf.lset, err = parseFlagLabels(*labelStrs)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"parse labels\")\n\t\t}\n\n\t\tconf.alertQueryURL, err = url.Parse(*conf.alertmgr.alertQueryURL)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"parse alert query url\")\n\t\t}\n\n\t\ttsdbOpts := &tsdb.Options{\n\t\t\tMinBlockDuration: int64(time.Duration(*tsdbBlockDuration) / time.Millisecond),\n\t\t\tMaxBlockDuration: int64(time.Duration(*tsdbBlockDuration) / time.Millisecond),\n\t\t\tRetentionDuration: int64(time.Duration(*tsdbRetention) / time.Millisecond),\n\t\t\tNoLockfile: *noLockFile,\n\t\t\tWALCompression: *walCompression,\n\t\t}\n\n\t\tagentOpts := &agent.Options{\n\t\t\tWALCompression: *walCompression,\n\t\t\tNoLockfile: *noLockFile,\n\t\t}\n\n\t\t// Parse and check query configuration.\n\t\tlookupQueries := map[string]struct{}{}\n\t\tfor _, q := range conf.query.addrs {\n\t\t\tif _, ok := lookupQueries[q]; ok {\n\t\t\t\treturn errors.Errorf(\"Address %s is duplicated for --query flag.\", q)\n\t\t\t}\n\n\t\t\tlookupQueries[q] = struct{}{}\n\t\t}\n\n\t\tconf.queryConfigYAML, err = conf.query.configPath.Content()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(conf.query.sdFiles) == 0 && len(conf.query.addrs) == 0 && len(conf.queryConfigYAML) == 0 {\n\t\t\treturn errors.New(\"no --query parameter was given\")\n\t\t}\n\t\tif (len(conf.query.sdFiles) != 0 || len(conf.query.addrs) != 0) && len(conf.queryConfigYAML) != 0 {\n\t\t\treturn errors.New(\"--query/--query.sd-files and --query.config* parameters cannot be defined at the same time\")\n\t\t}\n\n\t\t// Parse and check alerting configuration.\n\t\tconf.alertmgrsConfigYAML, err = conf.alertmgr.configPath.Content()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(conf.alertmgrsConfigYAML) != 0 && len(conf.alertmgr.alertmgrURLs) != 0 {\n\t\t\treturn errors.New(\"--alertmanagers.url and --alertmanagers.config* parameters cannot be defined at the same time\")\n\t\t}\n\n\t\tconf.alertRelabelConfigYAML, err = conf.alertmgr.alertRelabelConfigPath.Content()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\thttpLogOpts, err := logging.ParseHTTPOptions(*reqLogDecision, reqLogConfig)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"error while parsing config for request logging\")\n\t\t}\n\n\t\ttagOpts, grpcLogOpts, err := logging.ParsegRPCOptions(*reqLogDecision, reqLogConfig)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"error while parsing config for request logging\")\n\t\t}\n\n\t\treturn runRule(g,\n\t\t\tlogger,\n\t\t\treg,\n\t\t\ttracer,\n\t\t\tcomp,\n\t\t\t*conf,\n\t\t\treload,\n\t\t\tgetFlagsMap(cmd.Flags()),\n\t\t\thttpLogOpts,\n\t\t\tgrpcLogOpts,\n\t\t\ttagOpts,\n\t\t\ttsdbOpts,\n\t\t\tagentOpts,\n\t\t)\n\t})\n}", "func rulesetInsert(nkey string, key string, value string) (err error) {\n if ndb.Rdb == nil {\n logs.Error(\"rulesetInsert -- Can't access to database\")\n return errors.New(\"rulesetInsert -- Can't access to database\")\n }\n stmt, err := ndb.Rdb.Prepare(\"insert into ruleset (ruleset_uniqueid, ruleset_param, ruleset_value) values(?,?,?)\")\n if err != nil {\n logs.Error(\"rulesetInsert -- Prepare -> %s\", err.Error())\n return err\n }\n _, err = stmt.Exec(&nkey, &key, &value)\n if err != nil {\n logs.Error(\"rulesetInsert -- Execute -> %s\", err.Error())\n return err\n }\n return nil\n}", "func (m Middleware) Tx(db *sql.DB) TxFunc {\n\treturn func(f func(tx daos.Transaction, w http.ResponseWriter, r *http.Request) error) http.HandlerFunc {\n\t\treturn func(w http.ResponseWriter, r *http.Request) {\n\n\t\t\tt, err := db.Begin()\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\n\t\t\tdefer func() {\n\t\t\t\tl := m.log.WithRequest(r)\n\t\t\t\tif p := recover(); p != nil {\n\t\t\t\t\tt.Rollback()\n\t\t\t\t\tl.Info(\"transaction rollbacked\")\n\t\t\t\t\tpanic(p)\n\t\t\t\t} else if err != nil {\n\t\t\t\t\tt.Rollback()\n\t\t\t\t\tl.Info(\"transaction rollbacked\")\n\t\t\t\t\tpanic(err)\n\t\t\t\t} else {\n\t\t\t\t\terr = t.Commit()\n\t\t\t\t\tl.Info(\"transaction commited\")\n\t\t\t\t}\n\t\t\t}()\n\n\t\t\terr = f(t, w, r)\n\t\t}\n\t}\n}", "func setupDatabase(db *sql.DB) error {\n\ttx, err := db.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer tx.Rollback()\n\n\t// Creating a timescaledb extension for the database\n\tconst ext = `CREATE EXTENSION IF NOT EXISTS timescaledb CASCADE;`\n\tif _, err = tx.Exec(ext); err != nil {\n\t\treturn err\n\t}\n\n\t// creating schema in the database\n\tconst sch = `CREATE SCHEMA IF NOT EXISTS \"audit\"`\n\tif _, err = tx.Exec(sch); err != nil {\n\t\treturn err\n\t}\n\n\t// creating the audit log table\n\tconst tbl = `CREATE TABLE IF NOT EXISTS audit.\"Logs\" (\n\t\t\"Timestamp\" TIMESTAMPTZ NOT NULL,\n\t\t\"UserId\" text NOT NULL,\n\t\t\"Action\" text NOT NULL\n\t );`\n\tif _, err = tx.Exec(tbl); err != nil {\n\t\treturn err\n\t}\n\n\t// creating the hypertable of audit log table for timescaledb\n\tconst hptbl = `SELECT create_hypertable('audit.\"Logs\"', 'Timestamp',if_not_exists => true);`\n\tif _, err = tx.Exec(hptbl); err != nil {\n\t\treturn err\n\t}\n\treturn tx.Commit()\n}", "func (probe *BridgeOfProbe) addRule(rule *Rule) {\n\tlogging.GetLogger().Infof(\"New rule %v added\", rule.UUID)\n\tg := probe.OvsOfProbe.Graph\n\tg.Lock()\n\tdefer g.Unlock()\n\tbridgeNode := probe.BridgeNode\n\tmetadata := graph.Metadata{\n\t\t\"Type\": \"ofrule\",\n\t\t\"cookie\": fmt.Sprintf(\"0x%x\", rule.Cookie),\n\t\t\"table\": rule.Table,\n\t\t\"filters\": rule.Filter,\n\t\t\"actions\": rule.Actions,\n\t\t\"priority\": rule.Priority,\n\t\t\"UUID\": rule.UUID,\n\t}\n\truleNode := g.NewNode(graph.GenID(), metadata)\n\tg.Link(bridgeNode, ruleNode, graph.Metadata{\"RelationType\": \"ownership\"})\n}", "func (tx *txDriver) Tx(context.Context) (dialect.Tx, error) { return tx, nil }", "func (tx *txDriver) Tx(context.Context) (dialect.Tx, error) { return tx, nil }", "func (fw *IPtables) addIPtablesRule(rule *IPtablesRule) error {\n\tif err := fw.Store.addIPtablesRule(rule); err != nil {\n\t\tlog.Error(\"In addIPtablesRule failed to add \", rule.GetBody())\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (dp *dataProvider) Tx(fn func(*pg.Tx) error) error {\n\treturn wrapError(dp.db.RunInTransaction(func(tx *pg.Tx) error {\n\t\tdefer func(t *pg.Tx) {\n\t\t\tif err := recover(); err != nil {\n\t\t\t\tt.Rollback()\n\t\t\t\t// rethrow the panic once the database is safe\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t}(tx)\n\t\treturn fn(tx)\n\t}))\n}", "func (s *BaseDMLListener) EnterEveryRule(ctx antlr.ParserRuleContext) {}", "func (c *Conn) Transaction(fn func(*Conn) error) error {\r\n\tvar (\r\n\t\ttx = c.Begin()\r\n\t\tconn = &Conn{}\r\n\t)\r\n\tcopier.Copy(conn, c)\r\n\tconn.DB = tx\r\n\tif err := fn(conn); err != nil {\r\n\t\ttx.Rollback()\r\n\t\treturn err\r\n\t}\r\n\ttx.Commit()\r\n\treturn nil\r\n}", "func (tx *tx) Run(fn func(*Db) error) (err error) {\n\tdefer func() {\n\t\tif panicErr := recover(); panicErr != nil {\n\t\t\t// Don't try to rollback twice (if we're panicing after trying to rollback already)\n\t\t\t// attempt to rollback transaction\n\t\t\terr = tx.Rollback()\n\t\t\tif err != nil {\n\t\t\t\t// propagate panic\n\t\t\t\tpanic(fmt.Sprintf(\"Rollback error: \"+err.Error()+\"\\nOriginal Panic: %s\", panicErr))\n\t\t\t}\n\n\t\t\t// propagate original error\n\t\t\tpanic(panicErr)\n\t\t}\n\n\t\tif err != nil {\n\t\t\terr2 := tx.Rollback()\n\n\t\t\tif err2 != nil {\n\t\t\t\tpanic(\"Rollback error: \" + err2.Error() + \"\\nOriginal Error: \" + err.Error())\n\t\t\t}\n\t\t\treturn\n\t\t}\n\n\t\terr = tx.Commit()\n\t\treturn\n\t}()\n\n\terr = fn(tx.Db)\n\n\treturn\n}", "func (s *BaseMySqlParserListener) EnterEveryRule(ctx antlr.ParserRuleContext) {}", "func AddRule(r AuditRule) (err error) {\n\tard, _, _, err := r.toKernelAuditRule()\n\tif err != nil {\n\t\treturn\n\t}\n\tclient, err := libaudit.NewAuditClient(nil)\n\tdefer client.Close()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Failed to initialize client\")\n\t}\n\terr = client.AddRule(ard.toWireFormat())\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (db *DB) Transaction(fc func(db *DB) error) (err error) {\n\tpanicked := true\n\ttx := &DB{db.DB.Begin()}\n\n\tdefer func() {\n\t\t// Make sure to rollback when panic, Block error or Commit error\n\t\tif panicked || err != nil {\n\t\t\ttx.Rollback()\n\t\t}\n\t}()\n\n\terr = fc(tx)\n\tif err == nil {\n\t\terr = tx.DB.Commit().Error\n\t}\n\tpanicked = false\n\treturn\n}", "func (rc *RuleCreate) Save(ctx context.Context) (*Rule, error) {\n\tvar (\n\t\terr error\n\t\tnode *Rule\n\t)\n\trc.defaults()\n\tif len(rc.hooks) == 0 {\n\t\tif err = rc.check(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tnode, err = rc.sqlSave(ctx)\n\t} else {\n\t\tvar mut Mutator = MutateFunc(func(ctx context.Context, m Mutation) (Value, error) {\n\t\t\tmutation, ok := m.(*RuleMutation)\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"unexpected mutation type %T\", m)\n\t\t\t}\n\t\t\tif err = rc.check(); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\trc.mutation = mutation\n\t\t\tnode, err = rc.sqlSave(ctx)\n\t\t\tmutation.done = true\n\t\t\treturn node, err\n\t\t})\n\t\tfor i := len(rc.hooks) - 1; i >= 0; i-- {\n\t\t\tmut = rc.hooks[i](mut)\n\t\t}\n\t\tif _, err := mut.Mutate(ctx, rc.mutation); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn node, err\n}", "func (m *SQLIgnoreStore) Create(rule *IgnoreRule) error {\n\tstmt := `INSERT INTO ignorerule (userid, updated_by, expires, query, note)\n\t VALUES(?,?,?,?,?)`\n\n\tret, err := m.vdb.DB.Exec(stmt, rule.Name, rule.Name, rule.Expires.Unix(), rule.Query, rule.Note)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcreatedId, err := ret.LastInsertId()\n\tif err != nil {\n\t\treturn err\n\t}\n\trule.ID = int(createdId)\n\tm.inc()\n\treturn nil\n}", "func (ds *MySQLDatastore) Tx(f func(*sql.Tx) error) error {\n\ttx, err := ds.db.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = f(tx)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\treturn tx.Commit()\n}", "func (t *ACLRole) DBCreateTx(ctx context.Context, tx Tx) (sql.Result, error) {\n\tq := \"INSERT INTO `acl_role` (`acl_role`.`id`,`acl_role`.`checksum`,`acl_role`.`name`,`acl_role`.`description`,`acl_role`.`admin_user_id`,`acl_role`.`customer_id`,`acl_role`.`created_at`,`acl_role`.`updated_at`) VALUES (?,?,?,?,?,?,?,?)\"\n\tchecksum := t.CalculateChecksum()\n\tif t.GetChecksum() == checksum {\n\t\treturn nil, nil\n\t}\n\tt.Checksum = &checksum\n\treturn tx.ExecContext(ctx, q,\n\t\torm.ToSQLString(t.ID),\n\t\torm.ToSQLString(t.Checksum),\n\t\torm.ToSQLString(t.Name),\n\t\torm.ToSQLString(t.Description),\n\t\torm.ToSQLString(t.AdminUserID),\n\t\torm.ToSQLString(t.CustomerID),\n\t\torm.ToSQLInt64(t.CreatedAt),\n\t\torm.ToSQLInt64(t.UpdatedAt),\n\t)\n}", "func (r *BatchV1JobRule) createRule(job *batchV1.Job, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(job)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(job)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(job)\n\t\t},\n\t}\n\treturn rule\n}", "func (d *AppsV1DeploymentRule) createRule(deployment *appsv1.Deployment, ydr *YamlDerivedResource) *rule {\n\tr := &rule{\n\t\tID: d.ID,\n\t\tPrereqs: d.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif d.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn d.Condition(deployment)\n\t\t},\n\t\tMessage: d.Message,\n\t\tLevel: d.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif d.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn d.Fix(deployment)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif d.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn d.FixDescription(deployment)\n\t\t},\n\t}\n\treturn r\n}", "func (c *Schema) WithRule(rule IValidationRule) *Schema {\n\tif c.rules == nil {\n\t\tc.rules = []IValidationRule{}\n\t}\n\tc.rules = append(c.rules, rule)\n\treturn c\n}", "func EnsureRule(position RulePosition, table Table, chain Chain, args ...string) error {\n\tvar exists bool\n\tfullArgs := makeFullArgs(table, opListChain, chain, fullMac)\n\tout, err := run(cmd, fullArgs...)\n\tif err == nil {\n\t\texists = checkIfRuleExists(string(out), args...)\n\t}\n\tif !exists {\n\t\tfullArgs = makeFullArgs(table, operation(position), chain, args...)\n\t\tout, err := run(cmd, fullArgs...)\n\t\tif err != nil {\n\t\t\treturn trace.Wrap(err, \"failed to ensure rule: %s\", out)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *ACLRole) DBFindTx(ctx context.Context, tx Tx, _params ...interface{}) (bool, error) {\n\tparams := []interface{}{\n\t\torm.Column(\"id\"),\n\t\torm.Column(\"checksum\"),\n\t\torm.Column(\"name\"),\n\t\torm.Column(\"description\"),\n\t\torm.Column(\"admin_user_id\"),\n\t\torm.Column(\"customer_id\"),\n\t\torm.Column(\"created_at\"),\n\t\torm.Column(\"updated_at\"),\n\t\torm.Table(ACLRoleTableName),\n\t}\n\tif len(_params) > 0 {\n\t\tfor _, param := range _params {\n\t\t\tparams = append(params, param)\n\t\t}\n\t}\n\tq, p := orm.BuildQuery(params...)\n\trow := tx.QueryRowContext(ctx, q, p...)\n\tvar _ID sql.NullString\n\tvar _Checksum sql.NullString\n\tvar _Name sql.NullString\n\tvar _Description sql.NullString\n\tvar _AdminUserID sql.NullString\n\tvar _CustomerID sql.NullString\n\tvar _CreatedAt sql.NullInt64\n\tvar _UpdatedAt sql.NullInt64\n\terr := row.Scan(\n\t\t&_ID,\n\t\t&_Checksum,\n\t\t&_Name,\n\t\t&_Description,\n\t\t&_AdminUserID,\n\t\t&_CustomerID,\n\t\t&_CreatedAt,\n\t\t&_UpdatedAt,\n\t)\n\tif err != nil && err != sql.ErrNoRows {\n\t\treturn false, err\n\t}\n\tif _ID.Valid {\n\t\tt.SetID(_ID.String)\n\t}\n\tif _Checksum.Valid {\n\t\tt.SetChecksum(_Checksum.String)\n\t}\n\tif _Name.Valid {\n\t\tt.SetName(_Name.String)\n\t}\n\tif _Description.Valid {\n\t\tt.SetDescription(_Description.String)\n\t}\n\tif _AdminUserID.Valid {\n\t\tt.SetAdminUserID(_AdminUserID.String)\n\t}\n\tif _CustomerID.Valid {\n\t\tt.SetCustomerID(_CustomerID.String)\n\t}\n\tif _CreatedAt.Valid {\n\t\tt.SetCreatedAt(_CreatedAt.Int64)\n\t}\n\tif _UpdatedAt.Valid {\n\t\tt.SetUpdatedAt(_UpdatedAt.Int64)\n\t}\n\treturn true, nil\n}", "func Setup(db *gorm.DB) error {\n\tlog.SetOutput(ioutil.Discard)\n\tdefer log.SetOutput(os.Stderr)\n\tadapter := gormadapter.NewAdapterByDB(db)\n\tenf, err := casbin.NewEnforcerSafe(conf, adapter)\n\tif err != nil {\n\t\treturn err\n\t}\n\tenf.EnableLog(false)\n\tenf.LoadPolicy()\n\tenforcer = enf\n\treturn nil\n}", "func InsertIptableRule(version, tableName, chainName, match, target string) error {\n\tif RuleExists(version, tableName, chainName, match, target) {\n\t\tlog.Printf(\"Rule already exists\")\n\t\treturn nil\n\t}\n\n\tcmd := GetInsertIptableRuleCmd(version, tableName, chainName, match, target)\n\treturn RunCmd(version, cmd.Params)\n}", "func (s *BasePlSqlParserListener) EnterAlter_database(ctx *Alter_databaseContext) {}", "func (tbl AssociationTable) Tx() sqlapi.SqlTx {\n\treturn tbl.db.(sqlapi.SqlTx)\n}", "func (service *AlertRuleService) CreateAlertRule(ctx context.Context, rule models.AlertRule, provenance models.Provenance, userID int64) (models.AlertRule, error) {\n\tif rule.UID == \"\" {\n\t\trule.UID = util.GenerateShortUID()\n\t}\n\tinterval, err := service.ruleStore.GetRuleGroupInterval(ctx, rule.OrgID, rule.NamespaceUID, rule.RuleGroup)\n\t// if the alert group does not exists we just use the default interval\n\tif err != nil && errors.Is(err, store.ErrAlertRuleGroupNotFound) {\n\t\tinterval = service.defaultIntervalSeconds\n\t} else if err != nil {\n\t\treturn models.AlertRule{}, err\n\t}\n\trule.IntervalSeconds = interval\n\terr = rule.SetDashboardAndPanelFromAnnotations()\n\tif err != nil {\n\t\treturn models.AlertRule{}, err\n\t}\n\trule.Updated = time.Now()\n\terr = service.xact.InTransaction(ctx, func(ctx context.Context) error {\n\t\tids, err := service.ruleStore.InsertAlertRules(ctx, []models.AlertRule{\n\t\t\trule,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif id, ok := ids[rule.UID]; ok {\n\t\t\trule.ID = id\n\t\t} else {\n\t\t\treturn errors.New(\"couldn't find newly created id\")\n\t\t}\n\n\t\tif err = service.checkLimitsTransactionCtx(ctx, rule.OrgID, userID); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn service.provenanceStore.SetProvenance(ctx, &rule, rule.OrgID, provenance)\n\t})\n\tif err != nil {\n\t\treturn models.AlertRule{}, err\n\t}\n\treturn rule, nil\n}", "func (s *BasePlSqlParserListener) EnterDatabase(ctx *DatabaseContext) {}", "func NewRule(table *schema.Table, chain *schema.Chain, expr []schema.Statement, handle *int, index *int, comment string) *schema.Rule {\n\tc := &schema.Rule{\n\t\tFamily: table.Family,\n\t\tTable: table.Name,\n\t\tChain: chain.Name,\n\t\tExpr: expr,\n\t\tHandle: handle,\n\t\tIndex: index,\n\t\tComment: comment,\n\t}\n\n\treturn c\n}", "func insertRulesetValues(uuid string, param string, value string) (err error) {\n if ndb.RulesetParamExists(uuid, param) {\n updateRulesetValue, _ := ndb.Rdb.Prepare(\"update ruleset set ruleset_value = ? where ruleset_uniqueid = ? and ruleset_param =?;\")\n _, err = updateRulesetValue.Exec(&uuid, &param, &value)\n defer updateRulesetValue.Close()\n } else {\n insertRulesetValues, _ := ndb.Rdb.Prepare(\"insert into ruleset (ruleset_uniqueid, ruleset_param, ruleset_value) values (?,?,?);\")\n _, err = insertRulesetValues.Exec(&uuid, &param, &value)\n defer insertRulesetValues.Close()\n }\n\n if err != nil {\n return err\n }\n return nil\n}", "func (crtEP *MdlPersonCreateExt) BeforeDB(ent interface{}) error {\n\n\t// fmt.Println(\"TypeOf ent:\", reflect.TypeOf(ent))\n\t// fmt.Println(\"ValueOf ent:\", reflect.ValueOf(ent))\n\t// p := ent.(*models.Person)\n\n\t// make changes / validate the content struct pointer (p) here\n\t// p.Name = \"A new field value\"\n\treturn nil\n}", "func setupDB(db *sql.DB) error {\n\tsqlScript, err := ioutil.ReadFile(\"dbSchema.sql\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tstatements := strings.Split(string(sqlScript), \";\")\n\tif len(statements) > 0 {\n\t\tstatements = statements[:len(statements)-1]\n\t}\n\n\tfor _, statement := range statements {\n\t\t_, err = db.Exec(statement)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (crc *CasbinRuleCreate) Save(ctx context.Context) (*CasbinRule, error) {\n\tvar (\n\t\terr error\n\t\tnode *CasbinRule\n\t)\n\tif len(crc.hooks) == 0 {\n\t\tif err = crc.check(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tnode, err = crc.sqlSave(ctx)\n\t} else {\n\t\tvar mut Mutator = MutateFunc(func(ctx context.Context, m Mutation) (Value, error) {\n\t\t\tmutation, ok := m.(*CasbinRuleMutation)\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"unexpected mutation type %T\", m)\n\t\t\t}\n\t\t\tif err = crc.check(); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tcrc.mutation = mutation\n\t\t\tnode, err = crc.sqlSave(ctx)\n\t\t\tmutation.done = true\n\t\t\treturn node, err\n\t\t})\n\t\tfor i := len(crc.hooks) - 1; i >= 0; i-- {\n\t\t\tmut = crc.hooks[i](mut)\n\t\t}\n\t\tif _, err := mut.Mutate(ctx, crc.mutation); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn node, err\n}", "func init() {\n\tmigrations.MustRegisterTx(func(db migrations.DB) error {\n\t\t_, err := db.Exec(`\nCREATE TABLE sintls_authorization (\n authorization_id bigserial primary key,\n created_at timestamp not null default now(),\n updated_at timestamp not null default now(),\n name text unique not null,\n secret text not null,\n admin boolean not null default false\n);\nCREATE INDEX ON sintls_authorization(name);\n\nCREATE TABLE sintls_subdomain (\n subdomain_id bigserial primary key,\n created_at timestamp not null default now(),\n updated_at timestamp not null default now(),\n name text unique not null,\n authorization_id bigint not null references sintls_authorization(authorization_id) on delete cascade\n);\nCREATE INDEX ON sintls_subdomain(name);\nCREATE INDEX ON sintls_subdomain(authorization_id);\n\nCREATE TABLE sintls_host (\n host_id bigserial primary key,\n created_at timestamp not null default now(),\n updated_at timestamp not null default now(),\n name text not null,\n subdomain_id bigint not null references sintls_subdomain(subdomain_id) on delete cascade,\n dns_target_a inet,\n dns_target_aaaa inet,\n UNIQUE (name, subdomain_id),\n CHECK (dns_target_a is not null or dns_target_aaaa is not null)\n);\nCREATE INDEX on sintls_host(name);\nCREATE INDEX on sintls_host(subdomain_id);`)\n\t\treturn err\n\t}, func(db migrations.DB) error {\n\t\t_, err := db.Exec(`DROP TABLE sintls_host; DROP TABLE sintls_subdomain; DROP TABLE sintls_authorization;`)\n\t\treturn err\n\t})\n}", "func (o *AutomodRuleDatum) SetRule(ctx context.Context, exec boil.ContextExecutor, insert bool, related *AutomodRule) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(ctx, exec, boil.Infer()); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE \\\"automod_rule_data\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, []string{\"rule_id\"}),\n\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 2, automodRuleDatumPrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.ID, o.ID}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, updateQuery)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tif _, err = exec.ExecContext(ctx, updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.RuleID = related.ID\n\tif o.R == nil {\n\t\to.R = &automodRuleDatumR{\n\t\t\tRule: related,\n\t\t}\n\t} else {\n\t\to.R.Rule = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &automodRuleR{\n\t\t\tRuleAutomodRuleData: AutomodRuleDatumSlice{o},\n\t\t}\n\t} else {\n\t\trelated.R.RuleAutomodRuleData = append(related.R.RuleAutomodRuleData, o)\n\t}\n\n\treturn nil\n}", "func (r *V1Beta1ExtensionsDeploymentRule) createRule(deployment *v1beta1Extensions.Deployment, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(deployment)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(deployment)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(deployment)\n\t\t},\n\t}\n\treturn rule\n}", "func (s *BasePlSqlParserListener) EnterEveryRule(ctx antlr.ParserRuleContext) {}", "func (c *chain) setup(ipt *iptables.IPTables) error {\n\terr := utils.EnsureChain(ipt, c.table, c.name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Add the rules to the chain\n\tfor _, rule := range c.rules {\n\t\tif err := utils.InsertUnique(ipt, c.table, c.name, false, rule); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Add the entry rules to the entry chains\n\tfor _, entryChain := range c.entryChains {\n\t\tfor _, rule := range c.entryRules {\n\t\t\tr := []string{}\n\t\t\tr = append(r, rule...)\n\t\t\tr = append(r, \"-j\", c.name)\n\t\t\tif err := utils.InsertUnique(ipt, c.table, entryChain, c.prependEntry, r); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func checkCreateRule(t *testing.T, expError bool, tenant, policy, ruleID, dir, fnet, fepg, fip, tnet, tepg, tip, proto, act string, prio, port int) {\n\tpol := client.Rule{\n\t\tTenantName: tenant,\n\t\tPolicyName: policy,\n\t\tRuleID: ruleID,\n\t\tDirection: dir,\n\t\tPriority: prio,\n\t\tFromNetwork: fnet,\n\t\tFromEndpointGroup: fepg,\n\t\tFromIpAddress: fip,\n\t\tToNetwork: tnet,\n\t\tToEndpointGroup: tepg,\n\t\tToIpAddress: tip,\n\t\tProtocol: proto,\n\t\tPort: port,\n\t\tAction: act,\n\t}\n\terr := contivClient.RulePost(&pol)\n\tif err != nil && !expError {\n\t\tt.Fatalf(\"Error creating rule {%+v}. Err: %v\", pol, err)\n\t} else if err == nil && expError {\n\t\tt.Fatalf(\"Create rule {%+v} succeded while expecing error\", pol)\n\t} else if err == nil {\n\t\t// verify rule is created\n\t\t_, err := contivClient.RuleGet(tenant, policy, ruleID)\n\t\tif err != nil {\n\t\t\tt.Fatalf(\"Error getting rule %s/%s/%s. Err: %v\", tenant, policy, ruleID, err)\n\t\t}\n\t}\n}", "func (tbl RecordTable) Using(tx sqlapi.SqlTx) RecordTable {\n\ttbl.db = tx\n\treturn tbl\n}", "func setupTX(t testing.TB) (*reform.DB, *reform.TX) {\n\tt.Helper()\n\n\tdb := setupDB(t)\n\n\ttx, err := db.Begin()\n\trequire.NoError(t, err)\n\treturn db, tx\n}", "func (r *V1ServiceAccountRule) createRule(serviceaccount *v1.ServiceAccount, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(serviceaccount)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(serviceaccount)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(serviceaccount)\n\t\t},\n\t}\n\treturn rule\n}", "func (r *V1NamespaceRule) createRule(namespace *v1.Namespace, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(namespace)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(namespace)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(namespace)\n\t\t},\n\t}\n\treturn rule\n}", "func (r *GenericRule) createRule(resource *Resource, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(resource)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(resource)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(resource)\n\t\t},\n\t}\n\treturn rule\n}", "func (rcb *RuleCreateBulk) Save(ctx context.Context) ([]*Rule, error) {\n\tspecs := make([]*sqlgraph.CreateSpec, len(rcb.builders))\n\tnodes := make([]*Rule, len(rcb.builders))\n\tmutators := make([]Mutator, len(rcb.builders))\n\tfor i := range rcb.builders {\n\t\tfunc(i int, root context.Context) {\n\t\t\tbuilder := rcb.builders[i]\n\t\t\tbuilder.defaults()\n\t\t\tvar mut Mutator = MutateFunc(func(ctx context.Context, m Mutation) (Value, error) {\n\t\t\t\tmutation, ok := m.(*RuleMutation)\n\t\t\t\tif !ok {\n\t\t\t\t\treturn nil, fmt.Errorf(\"unexpected mutation type %T\", m)\n\t\t\t\t}\n\t\t\t\tif err := builder.check(); err != nil {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\t\t\t\tbuilder.mutation = mutation\n\t\t\t\tnodes[i], specs[i] = builder.createSpec()\n\t\t\t\tvar err error\n\t\t\t\tif i < len(mutators)-1 {\n\t\t\t\t\t_, err = mutators[i+1].Mutate(root, rcb.builders[i+1].mutation)\n\t\t\t\t} else {\n\t\t\t\t\t// Invoke the actual operation on the latest mutation in the chain.\n\t\t\t\t\tif err = sqlgraph.BatchCreate(ctx, rcb.driver, &sqlgraph.BatchCreateSpec{Nodes: specs}); err != nil {\n\t\t\t\t\t\tif cerr, ok := isSQLConstraintError(err); ok {\n\t\t\t\t\t\t\terr = cerr\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tmutation.done = true\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\t\t\t\tid := specs[i].ID.Value.(int64)\n\t\t\t\tnodes[i].ID = int(id)\n\t\t\t\treturn nodes[i], nil\n\t\t\t})\n\t\t\tfor i := len(builder.hooks) - 1; i >= 0; i-- {\n\t\t\t\tmut = builder.hooks[i](mut)\n\t\t\t}\n\t\t\tmutators[i] = mut\n\t\t}(i, ctx)\n\t}\n\tif len(mutators) > 0 {\n\t\tif _, err := mutators[0].Mutate(ctx, rcb.builders[0].mutation); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn nodes, nil\n}", "func NewRule(release_id string, rule database.Rule) {\n\tvar release database.Release\n\tdb.DB.Where(\"id = ?\", release_id).First(&release)\n\n\trelease.Rules = append(release.Rules, rule)\n\tdb.DB.Save(&release)\n}", "func (store *SQLStore) execTx(ctx context.Context, fn func(*Queries) error) error {\n\ttx, err := store.db.BeginTx(ctx, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// get back new query object \n\tq := New(tx)\n\terr = fn(q)\n\t// rollback on error - if rollback error return both errors \n\tif err != nil {\n\t\tif rbErr := tx.Rollback(); rbErr != nil {\n\t\t\treturn fmt.Errorf(\"tx err: %v, rb err: %v\", err, rbErr)\n\t\t}\n\t\treturn err\n\t}\n\treturn tx.Commit()\n}", "func validateSchema(db *sql.DB) error {\n\tdriver, err := postgres.WithInstance(db, &postgres.Config{})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// whether we are running a test database or a production (which is in the cloud)\n\t// will use different files to start the db\n\tenvironment, _ := os.LookupEnv(\"ENVIRONMENT\")\n\tvar migrationPath string\n\t// Runs when we use docker (this runs in the cloud either on the dev side or productin side)\n\tmigrationPath = \"file://database/migrations\"\n\tif environment == \"production\" || environment == \"dev\" {\n\t\tfmt.Println(\"Using dev migrations\")\n\t\tmigrationPath = \"file://database/migrations\"\n\t} else {\n\t\t// This runs when we run our tests (this is run in the docker test files)\n\t\tfmt.Println(\"Using test migrations\")\n\t\tmigrationPath = \"file://database/test-migrations\"\n\t}\n\tm, err := migrate.NewWithDatabaseInstance(\n\t\t// for production\n\t\tmigrationPath,\n\n\t\t\"postgres\", driver)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = m.Up()\n\tif err != nil && err != migrate.ErrNoChange {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (store *Store) execTx(ctx context.Context, fn func(*Queries) error) error {\n\ttx, err := store.db.BeginTx(ctx, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tq := New(tx)\n\terr = fn(q)\n\tif err != nil {\n\t\tif rbErr := tx.Rollback(); rbErr != nil {\n\t\t\treturn fmt.Errorf(\"tx err: %v, rb err: %v\", err, rbErr)\n\t\t}\n\t\treturn err\n\t}\n\n\treturn tx.Commit()\n}", "func commitDBTx(t *testing.T, store *Store, db walletdb.DB,\n\tf func(walletdb.ReadWriteBucket)) {\n\n\tt.Helper()\n\n\tdbTx, err := db.BeginReadWriteTx()\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tdefer dbTx.Commit()\n\n\tns := dbTx.ReadWriteBucket(namespaceKey)\n\n\tf(ns)\n}", "func (r *V1ContainerRule) createRule(container *v1.Container, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(container)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(container)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(container)\n\t\t},\n\t}\n\treturn rule\n}", "func (store *Store) execTx(ctx context.Context, fn func(*Queries) error) error {\n\ttx, err := store.db.BeginTx(ctx, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tq := New(tx) // create a new Queries object but instead of passing in sql.DB, we are passing in a sql.Tx Object for transactions\n\terr = fn(q) // call the input function as queries\n\t// If there was an error with the transaction\n\tif err != nil {\n\t\tif rollbackErr := tx.Rollback(); rollbackErr != nil {\n\t\t\treturn fmt.Errorf(\"tx error: %v, rollback error: %v\", err, rollbackErr)\n\t\t}\n\t\treturn err\n\t}\n\treturn tx.Commit()\n}", "func RegisterRule(vr Rule) {\n\tregisterRuleLock.Lock()\n\tdefer registerRuleLock.Unlock()\n\tRegisteredRules = append(RegisteredRules, vr)\n}", "func (fw *IPtables) CreateRules(chain int) error {\n\tlog.Info(\"In CreateRules() for chain\", chain)\n\tfor _, rule := range fw.chains[chain].Rules {\n\t\t// First create rule record in database.\n\t\terr0 := fw.addIPtablesRule(rule)\n\t\tif err0 != nil {\n\t\t\tlog.Error(\"In CreateRules() create db record for iptables rule \", rule.GetBody())\n\t\t\treturn err0\n\t\t}\n\n\t\terr1 := fw.EnsureRule(rule, EnsureFirst)\n\t\tif err1 != nil {\n\t\t\tlog.Error(\"In CreateRules() failed to create install firewall rule \", rule.GetBody())\n\t\t\treturn err1\n\t\t}\n\n\t\t// Finally, set 'active' flag in database record.\n\t\tif err2 := fw.Store.switchIPtablesRule(rule, setRuleActive); err2 != nil {\n\t\t\tlog.Error(\"In CreateRules() iptables rule created, but activation failed \", rule.GetBody())\n\t\t\treturn err2\n\t\t}\n\t}\n\tlog.Info(\"Creating firewall rules success\")\n\treturn nil\n}", "func (db *Database) Setup() error {\n\tdb.Update(func(tx *bbolt.Tx) error {\n\t\t_, err := tx.CreateBucketIfNotExists(leasesBucket)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = tx.CreateBucketIfNotExists(reservationBucket)\n\t\treturn err\n\t})\n\treturn nil\n}", "func Transaction(db *sql.DB, fns ...func(DB) error) error {\n\ttx, err := db.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, fn := range fns {\n\t\terr := fn(tx)\n\t\tif err != nil {\n\t\t\ttx.Rollback()\n\t\t\treturn err\n\t\t}\n\t}\n\n\terr = tx.Commit()\n\terr = interpretScanError(err)\n\treturn err\n}", "func (tbl RecordTable) Tx() sqlapi.SqlTx {\n\treturn tbl.db.(sqlapi.SqlTx)\n}", "func (d *MockDataResyncDSL) StnRule(val *stn.STN_Rule) vppclient.DataResyncDSL {\n\top := dsl.TxnOp{Key: stn.Key(val.RuleName), Value: val}\n\td.Ops = append(d.Ops, op)\n\treturn d\n}", "func (exec *StormPerf) runInTx(fn func(tx storm.Node) error) error {\n\ttx, err := exec.db.Begin(true)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer tx.Rollback()\n\n\tif err := fn(tx); err != nil {\n\t\treturn err\n\t}\n\n\treturn tx.Commit()\n}", "func Transact(db *gorm.DB, tf func(tx *gorm.DB) error) (err error) {\n\tif commonDB, ok := db.CommonDB().(sqlTx); ok && commonDB != nil {\n\t\t// If the db is already in a transaction, just execute tf\n\t\t// and let the outer transaction handle Rollback and Commit.\n\t\treturn tf(db)\n\t}\n\n\ttx := db.Begin()\n\tif tx.Error != nil {\n\t\treturn fmt.Errorf(\"could not start transaction. %s\", err)\n\t}\n\tdefer func() {\n\t\tif p := recover(); p != nil {\n\t\t\ttx.Rollback()\n\t\t\tpanic(p)\n\t\t}\n\t\tif err != nil {\n\t\t\ttx.Rollback()\n\t\t} else {\n\t\t\terr = tx.Commit().Error\n\t\t}\n\t}()\n\treturn tf(tx)\n}", "func (o *Constraint) Create(db *gorm.DB) error {\n\treturn db.Create(o).Error\n}", "func (tbl DbCompoundTable) Using(tx *sql.Tx) DbCompoundTable {\n\ttbl.db = tx\n\treturn tbl\n}", "func (h *IptablesUtilsHandler) IptablesInsertRule(table string, chain string, rulespec ...string) error {\n\tiptablesObject, err := iptables.New()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn iptablesObject.Insert(table, chain, 0, rulespec...)\n}", "func DBCreateACLRoleTableTx(ctx context.Context, tx Tx) error {\n\tq := \"CREATE TABLE `acl_role` (`id`VARCHAR(64) NOT NULL PRIMARY KEY,`checksum`CHAR(64),`name` VARCHAR(100) NOT NULL,`description`TEXT,`admin_user_id` VARCHAR(64),`customer_id`VARCHAR(64),`created_at` BIGINT UNSIGNED NOT NULL,`updated_at` BIGINT UNSIGNED,INDEX acl_role_name_index (`name`),INDEX acl_role_customer_id_index (`customer_id`)) ENGINE=InnoDB DEFAULT CHARSET=utf8mb4 COLLATE=utf8mb4_unicode_ci;\"\n\t_, err := tx.ExecContext(ctx, q)\n\treturn err\n}", "func Tx(ctx context.Context, db *sqlx.DB, opts *sqltx.Options, fn TXFn) (err error) {\n\treturn sqltx.TxHandler(ctx, &sqlxDB{db}, opts, func(tx sqltx.TXer) error {\n\t\treturn fn(tx.(*sqlx.Tx))\n\t})\n}", "func (p *PgDb) CreateEmpTx(r *model.Emp, tx *sqlx.Tx, isValidate bool) (*model.Emp, error) {\n\tcnt := \"postgres (p *PgDb) CreateEmpTx\" // Имя текущего метода для логирования\n\t//mylog.PrintfDebug(\"[DEBUG] %v - START, param: '%+v'\", cnt, r)\n\n\t// Проверяем определен ли контекст транзакции\n\tif tx == nil {\n\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - tx *sqlx.Tx is NULL\", cnt)\n\t\tlog.Printf(errM)\n\t\treturn nil, errors.New(errM)\n\t}\n\n\t//=====================================================================\n\t// Добавить валидацию входной структуры\n\t//=====================================================================\n\n\t// =====================================================================\n\t// Переменная часть кода\n\t// =====================================================================\n\t// Если запускаем с проверками\n\tif isValidate {\n\t\t{ //Если Dept NULL или НЕ существует, то ошибка\n\t\t\tif !r.Deptno.Valid {\n\t\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - r.Deptno is NULL\", cnt)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, errors.New(errM)\n\t\t\t}\n\t\t\tdeptno := int(r.Deptno.Int64)\n\t\t\t// Запрос в транзакции\n\t\t\tdeptExists, err := p.DeptExistsTx(deptno, tx)\n\t\t\tif err != nil {\n\t\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - p.DeptExistsTx(deptno, tx), args = '%v'\", cnt, deptno)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, errors.WithMessage(err, errM)\n\t\t\t}\n\t\t\tif !deptExists {\n\t\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - dept '%v' does not exist\", cnt, deptno)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, errors.New(errM)\n\t\t\t}\n\t\t\t//mylog.PrintfDebug(\"[DEBUG] %v - dept %v exists\", cnt, deptno)\n\t\t}\n\t\t{ //Если Emp существует, то игнорируем\n\t\t\texists, err := p.EmpExistsTx(r.Empno, tx)\n\t\t\tif err != nil {\n\t\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - p.EmpExistsTx(r.Empno, tx), args = '%v'\", cnt, r.Empno)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, errors.WithMessage(err, errM)\n\t\t\t}\n\t\t\t// Если запись существует, то ни чего не делаем, возвращем, что пришло на вход\n\t\t\tif exists {\n\t\t\t\terrM := fmt.Sprintf(\"[WARN] %v - WARN - emp '%v' already exist - nothing to do\", cnt, r.Empno)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, nil\n\t\t\t}\n\t\t\t//mylog.PrintfDebug(\"[DEBUG] %v - emp '%v' does not exist - can be created\", cnt, r.Empno)\n\t\t}\n\t}\n\t// =====================================================================\n\n\t// =====================================================================\n\t// Переменная часть кода\n\t// =====================================================================\n\tstmText := sqlInsertEmpText\n\t// =====================================================================\n\n\t//Выполняем команду\n\tres, err := tx.NamedExec(stmText, r)\n\tif err != nil {\n\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - tx.NamedExec(stmText, r), args = '%+v'\", cnt, r)\n\t\tlog.Printf(errM)\n\t\treturn nil, errors.Wrap(err, errM)\n\t}\n\n\t{ // Необязательная часть - можно удалить в последствии\n\t\t// Проверим количество обработанных строк\n\t\trowCount, err := res.RowsAffected()\n\t\t_ = rowCount\n\t\t//mylog.PrintfDebug(\"[DEBUG] %v -- process %v rows\", cnt, rowCount)\n\t\tif err != nil {\n\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - res.RowsAffected()\", cnt)\n\t\t\tlog.Printf(errM)\n\t\t\treturn nil, errors.Wrap(err, errM)\n\t\t}\n\t}\n\n\t// =====================================================================\n\t// Переменная часть кода\n\t// =====================================================================\n\t// считаем данные обновленные данные - в БД могли быть тригера, которые поменяли данные\n\t// если запустили без проверок, то можно не возвращать результат - он будет запрошен уровнем выше\n\tif isValidate {\n\t\tv, err := p.GetEmpTx(r.Empno, tx)\n\t\tif err != nil {\n\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - p.GetEmpTx(r.Empno, tx), args = '%v'\", cnt, r.Empno)\n\t\t\tlog.Printf(errM)\n\t\t\treturn nil, errors.WithMessage(err, errM)\n\t\t}\n\t\tr = v\n\t}\n\t// =====================================================================\n\n\t//mylog.PrintfDebug(\"[DEBUG] %v - SUCCESS\", cnt)\n\n\treturn r, nil\n}", "func (tbl AssociationTable) Using(tx sqlapi.SqlTx) AssociationTabler {\n\ttbl.db = tx\n\treturn tbl\n}", "func (t *Transaction) Execute(rule *models.Rule) *matches.MatchResult {\n\n\tvar matchResult *matches.MatchResult\n\n\tfor _, variable := range rule.Variables {\n\t\tmapData := TransactionMaps.Get(variable.Name)\n\n\t\tif mapData == nil {\n\t\t\t//TODO log unknown Rule\n\t\t\tfmt.Println(\"Unrecognized variable: \" + variable.Name)\n\t\t\treturn nil\n\t\t}\n\n\t\texecuterModel := &TransactionExecuterModel{t, rule, variable}\n\t\tmatchResult = mapData.executer(executerModel)\n\n\t\tif matchResult.IsMatched {\n\t\t\tif rule.Chain != nil {\n\t\t\t\tmatchResult = t.Execute(rule.Chain)\n\n\t\t\t\tif matchResult == nil {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif !variable.FilterIsNotType && !rule.Operator.OperatorIsNotType {\n\t\t\t\treturn matchResult\n\t\t\t} else if !matchResult.DefaultState {\n\t\t\t\tmatchResult.SetMatch(false)\n\t\t\t}\n\n\t\t} else if !matchResult.IsMatched && !matchResult.DefaultState && (variable.FilterIsNotType || rule.Operator.OperatorIsNotType) {\n\t\t\treturn matchResult.SetMatch(true)\n\t\t}\n\t}\n\n\treturn matchResult\n\n}", "func runSchemaChangesInTxn(\n\tctx context.Context, planner *planner, tableDesc *tabledesc.Mutable, traceKV bool,\n) error {\n\tif len(tableDesc.DrainingNames) > 0 {\n\t\t// Reclaim all the old names. Leave the data and descriptor\n\t\t// cleanup for later.\n\t\tfor _, drain := range tableDesc.DrainingNames {\n\t\t\terr := catalogkv.RemoveObjectNamespaceEntry(ctx, planner.Txn(), planner.ExecCfg().Codec,\n\t\t\t\tdrain.ParentID, drain.ParentSchemaID, drain.Name, false /* KVTrace */)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\ttableDesc.DrainingNames = nil\n\t}\n\n\tif tableDesc.Dropped() {\n\t\treturn nil\n\t}\n\n\t// Only needed because columnBackfillInTxn() backfills\n\t// all column mutations.\n\tdoneColumnBackfill := false\n\n\t// Mutations are processed in multiple steps: First we process all mutations\n\t// for schema changes other than adding check or FK constraints, then we\n\t// validate those constraints, and only after that do we process the\n\t// constraint mutations. We need an in-memory copy of the table descriptor\n\t// that contains newly added columns (since constraints being added can\n\t// reference them), but that doesn't contain constraints (since otherwise we'd\n\t// plan the query assuming the constraint holds). This is a different\n\t// procedure than in the schema changer for existing tables, since all the\n\t// \"steps\" in the schema change occur within the same transaction here.\n\t//\n\t// In the future it would be good to either unify the two implementations more\n\t// or make this in-transaction implementation more principled. We expect\n\t// constraint validation to be refactored and treated as a first-class concept\n\t// in the world of transactional schema changes.\n\n\t// Collect constraint mutations to process later.\n\tvar constraintAdditionMutations []descpb.DescriptorMutation\n\n\t// We use a range loop here as the processing of some mutations\n\t// such as the primary key swap mutations result in queueing more\n\t// mutations that need to be processed.\n\tfor i := 0; i < len(tableDesc.Mutations); i++ {\n\t\tm := tableDesc.Mutations[i]\n\t\t// Skip mutations that get canceled by later operations\n\t\tif discarded, _ := isCurrentMutationDiscarded(tableDesc, m, i+1); discarded {\n\t\t\tcontinue\n\t\t}\n\n\t\timmutDesc := tabledesc.NewBuilder(tableDesc.TableDesc()).BuildImmutableTable()\n\t\tswitch m.Direction {\n\t\tcase descpb.DescriptorMutation_ADD:\n\t\t\tswitch m.Descriptor_.(type) {\n\t\t\tcase *descpb.DescriptorMutation_PrimaryKeySwap:\n\t\t\t\t// Don't need to do anything here, as the call to MakeMutationComplete\n\t\t\t\t// will perform the steps for this operation.\n\t\t\tcase *descpb.DescriptorMutation_ComputedColumnSwap:\n\t\t\t\treturn AlterColTypeInTxnNotSupportedErr\n\t\t\tcase *descpb.DescriptorMutation_Column:\n\t\t\t\tif doneColumnBackfill || !tabledesc.ColumnNeedsBackfill(m.Direction, m.GetColumn()) {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tif err := columnBackfillInTxn(ctx, planner.Txn(), planner.EvalContext(), planner.SemaCtx(), immutDesc, traceKV); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tdoneColumnBackfill = true\n\n\t\t\tcase *descpb.DescriptorMutation_Index:\n\t\t\t\tif err := indexBackfillInTxn(ctx, planner.Txn(), planner.EvalContext(), planner.SemaCtx(), immutDesc, traceKV); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\tcase *descpb.DescriptorMutation_Constraint:\n\t\t\t\t// This is processed later. Do not proceed to MakeMutationComplete.\n\t\t\t\tconstraintAdditionMutations = append(constraintAdditionMutations, m)\n\t\t\t\tcontinue\n\n\t\t\tdefault:\n\t\t\t\treturn errors.AssertionFailedf(\n\t\t\t\t\t\"unsupported mutation: %+v\", m)\n\t\t\t}\n\n\t\tcase descpb.DescriptorMutation_DROP:\n\t\t\t// Drop the name and drop the associated data later.\n\t\t\tswitch t := m.Descriptor_.(type) {\n\t\t\tcase *descpb.DescriptorMutation_Column:\n\t\t\t\tif doneColumnBackfill || !tabledesc.ColumnNeedsBackfill(m.Direction, m.GetColumn()) {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tif err := columnBackfillInTxn(\n\t\t\t\t\tctx, planner.Txn(), planner.EvalContext(), planner.SemaCtx(), immutDesc, traceKV,\n\t\t\t\t); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tdoneColumnBackfill = true\n\n\t\t\tcase *descpb.DescriptorMutation_Index:\n\t\t\t\tif err := indexTruncateInTxn(\n\t\t\t\t\tctx, planner.Txn(), planner.ExecCfg(), planner.EvalContext(), immutDesc, t.Index, traceKV,\n\t\t\t\t); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\tcase *descpb.DescriptorMutation_Constraint:\n\t\t\t\tswitch t.Constraint.ConstraintType {\n\t\t\t\tcase descpb.ConstraintToUpdate_CHECK, descpb.ConstraintToUpdate_NOT_NULL:\n\t\t\t\t\tfor i := range tableDesc.Checks {\n\t\t\t\t\t\tif tableDesc.Checks[i].Name == t.Constraint.Name {\n\t\t\t\t\t\t\ttableDesc.Checks = append(tableDesc.Checks[:i], tableDesc.Checks[i+1:]...)\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\tcase descpb.ConstraintToUpdate_FOREIGN_KEY:\n\t\t\t\t\tfor i := range tableDesc.OutboundFKs {\n\t\t\t\t\t\tfk := &tableDesc.OutboundFKs[i]\n\t\t\t\t\t\tif fk.Name == t.Constraint.Name {\n\t\t\t\t\t\t\tif err := planner.removeFKBackReference(ctx, tableDesc, fk); err != nil {\n\t\t\t\t\t\t\t\treturn err\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\ttableDesc.OutboundFKs = append(tableDesc.OutboundFKs[:i], tableDesc.OutboundFKs[i+1:]...)\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\tcase descpb.ConstraintToUpdate_UNIQUE_WITHOUT_INDEX:\n\t\t\t\t\tfor i := range tableDesc.UniqueWithoutIndexConstraints {\n\t\t\t\t\t\tif tableDesc.UniqueWithoutIndexConstraints[i].Name == t.Constraint.Name {\n\t\t\t\t\t\t\ttableDesc.UniqueWithoutIndexConstraints = append(\n\t\t\t\t\t\t\t\ttableDesc.UniqueWithoutIndexConstraints[:i],\n\t\t\t\t\t\t\t\ttableDesc.UniqueWithoutIndexConstraints[i+1:]...,\n\t\t\t\t\t\t\t)\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\tdefault:\n\t\t\t\t\treturn errors.AssertionFailedf(\n\t\t\t\t\t\t\"unsupported constraint type: %d\", errors.Safe(t.Constraint.ConstraintType))\n\t\t\t\t}\n\n\t\t\tdefault:\n\t\t\t\treturn errors.AssertionFailedf(\"unsupported mutation: %+v\", m)\n\t\t\t}\n\n\t\t}\n\t\tif err := tableDesc.MakeMutationComplete(m); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// If the mutation we processed was a primary key swap, there is some\n\t\t// extra work that needs to be done. Note that we don't need to create\n\t\t// a job to clean up the dropped indexes because those mutations can\n\t\t// get processed in this txn on the new table.\n\t\tif pkSwap := m.GetPrimaryKeySwap(); pkSwap != nil {\n\t\t\t// If any old index had an interleaved parent, remove the\n\t\t\t// backreference from the parent.\n\t\t\t// N.B. This logic needs to be kept up to date with the\n\t\t\t// corresponding piece in (*SchemaChanger).done. It is slightly\n\t\t\t// different because of how it access tables and how it needs to\n\t\t\t// write the modified table descriptors explicitly.\n\t\t\tfor _, idxID := range append(\n\t\t\t\t[]descpb.IndexID{pkSwap.OldPrimaryIndexId}, pkSwap.OldIndexes...) {\n\t\t\t\toldIndex, err := tableDesc.FindIndexWithID(idxID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tif oldIndex.NumInterleaveAncestors() != 0 {\n\t\t\t\t\tancestorInfo := oldIndex.GetInterleaveAncestor(oldIndex.NumInterleaveAncestors() - 1)\n\t\t\t\t\tancestor, err := planner.Descriptors().GetMutableTableVersionByID(ctx, ancestorInfo.TableID, planner.txn)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t\tancestorIdxI, err := ancestor.FindIndexWithID(ancestorInfo.IndexID)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t\tancestorIdx := ancestorIdxI.IndexDesc()\n\t\t\t\t\tfoundAncestor := false\n\t\t\t\t\tfor k, ref := range ancestorIdx.InterleavedBy {\n\t\t\t\t\t\tif ref.Table == tableDesc.ID && ref.Index == oldIndex.GetID() {\n\t\t\t\t\t\t\tif foundAncestor {\n\t\t\t\t\t\t\t\treturn errors.AssertionFailedf(\n\t\t\t\t\t\t\t\t\t\"ancestor entry in %s for %s@%s found more than once\",\n\t\t\t\t\t\t\t\t\tancestor.Name, tableDesc.Name, oldIndex.GetName())\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tancestorIdx.InterleavedBy = append(\n\t\t\t\t\t\t\t\tancestorIdx.InterleavedBy[:k], ancestorIdx.InterleavedBy[k+1:]...)\n\t\t\t\t\t\t\tfoundAncestor = true\n\t\t\t\t\t\t\tif err := planner.writeSchemaChange(ctx, ancestor, descpb.InvalidMutationID,\n\t\t\t\t\t\t\t\tfmt.Sprintf(\"remove interleaved backreference from table %s(%d) \"+\n\t\t\t\t\t\t\t\t\t\"for primary key swap of table %s(%d)\",\n\t\t\t\t\t\t\t\t\tancestor.Name, ancestor.ID, tableDesc.Name, tableDesc.ID,\n\t\t\t\t\t\t\t\t)); err != nil {\n\t\t\t\t\t\t\t\treturn err\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\t// Clear all the mutations except for adding constraints.\n\ttableDesc.Mutations = constraintAdditionMutations\n\n\t// Now that the table descriptor is in a valid state with all column and index\n\t// mutations applied, it can be used for validating check/FK constraints.\n\tfor _, m := range constraintAdditionMutations {\n\t\tconstraint := m.GetConstraint()\n\t\tswitch constraint.ConstraintType {\n\t\tcase descpb.ConstraintToUpdate_CHECK, descpb.ConstraintToUpdate_NOT_NULL:\n\t\t\tif constraint.Check.Validity == descpb.ConstraintValidity_Validating {\n\t\t\t\tif err := validateCheckInTxn(\n\t\t\t\t\tctx, planner.Descriptors().LeaseManager(), &planner.semaCtx, planner.EvalContext(), tableDesc, planner.txn, constraint.Check.Expr,\n\t\t\t\t); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tconstraint.Check.Validity = descpb.ConstraintValidity_Validated\n\t\t\t}\n\t\tcase descpb.ConstraintToUpdate_FOREIGN_KEY:\n\t\t\t// We can't support adding a validated foreign key constraint in the same\n\t\t\t// transaction as the CREATE TABLE statement. This would require adding\n\t\t\t// the backreference to the other table and then validating the constraint\n\t\t\t// for whatever rows were inserted into the referencing table in this\n\t\t\t// transaction, which requires multiple schema changer states across\n\t\t\t// multiple transactions.\n\t\t\t//\n\t\t\t// We could partially fix this by queuing a validation job to run post-\n\t\t\t// transaction. Better yet would be to absorb this into the transactional\n\t\t\t// schema change framework eventually.\n\t\t\t//\n\t\t\t// For now, just always add the FK as unvalidated.\n\t\t\tconstraint.ForeignKey.Validity = descpb.ConstraintValidity_Unvalidated\n\t\tcase descpb.ConstraintToUpdate_UNIQUE_WITHOUT_INDEX:\n\t\t\tif constraint.UniqueWithoutIndexConstraint.Validity == descpb.ConstraintValidity_Validating {\n\t\t\t\tif err := validateUniqueWithoutIndexConstraintInTxn(\n\t\t\t\t\tctx, planner.EvalContext(), tableDesc, planner.txn, constraint.Name,\n\t\t\t\t); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tconstraint.UniqueWithoutIndexConstraint.Validity = descpb.ConstraintValidity_Validated\n\t\t\t}\n\t\tdefault:\n\t\t\treturn errors.AssertionFailedf(\n\t\t\t\t\"unsupported constraint type: %d\", errors.Safe(constraint.ConstraintType))\n\t\t}\n\t}\n\n\t// Finally, add the constraints. We bypass MakeMutationsComplete (which makes\n\t// certain assumptions about the state in the usual schema changer) and just\n\t// update the table descriptor directly.\n\tfor _, m := range constraintAdditionMutations {\n\t\tconstraint := m.GetConstraint()\n\t\tswitch constraint.ConstraintType {\n\t\tcase descpb.ConstraintToUpdate_CHECK, descpb.ConstraintToUpdate_NOT_NULL:\n\t\t\ttableDesc.Checks = append(tableDesc.Checks, &constraint.Check)\n\t\tcase descpb.ConstraintToUpdate_FOREIGN_KEY:\n\t\t\tfk := constraint.ForeignKey\n\t\t\tvar referencedTableDesc *tabledesc.Mutable\n\t\t\t// We don't want to lookup/edit a second copy of the same table.\n\t\t\tselfReference := tableDesc.ID == fk.ReferencedTableID\n\t\t\tif selfReference {\n\t\t\t\treferencedTableDesc = tableDesc\n\t\t\t} else {\n\t\t\t\tlookup, err := planner.Descriptors().GetMutableTableVersionByID(ctx, fk.ReferencedTableID, planner.Txn())\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn errors.Errorf(\"error resolving referenced table ID %d: %v\", fk.ReferencedTableID, err)\n\t\t\t\t}\n\t\t\t\treferencedTableDesc = lookup\n\t\t\t}\n\t\t\treferencedTableDesc.InboundFKs = append(referencedTableDesc.InboundFKs, fk)\n\t\t\ttableDesc.OutboundFKs = append(tableDesc.OutboundFKs, fk)\n\n\t\t\t// Write the other table descriptor here if it's not the current table\n\t\t\t// we're already modifying.\n\t\t\tif !selfReference {\n\t\t\t\tif err := planner.writeSchemaChange(\n\t\t\t\t\tctx, referencedTableDesc, descpb.InvalidMutationID,\n\t\t\t\t\tfmt.Sprintf(\"updating referenced FK table %s(%d) table %s(%d)\",\n\t\t\t\t\t\treferencedTableDesc.Name, referencedTableDesc.ID, tableDesc.Name, tableDesc.ID),\n\t\t\t\t); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\tcase descpb.ConstraintToUpdate_UNIQUE_WITHOUT_INDEX:\n\t\t\ttableDesc.UniqueWithoutIndexConstraints = append(\n\t\t\t\ttableDesc.UniqueWithoutIndexConstraints, constraint.UniqueWithoutIndexConstraint,\n\t\t\t)\n\t\tdefault:\n\t\t\treturn errors.AssertionFailedf(\n\t\t\t\t\"unsupported constraint type: %d\", errors.Safe(constraint.ConstraintType))\n\t\t}\n\t}\n\ttableDesc.Mutations = nil\n\treturn nil\n}", "func (o *AutomodRuleDatum) Rule(mods ...qm.QueryMod) automodRuleQuery {\n\tqueryMods := []qm.QueryMod{\n\t\tqm.Where(\"\\\"id\\\" = ?\", o.RuleID),\n\t}\n\n\tqueryMods = append(queryMods, mods...)\n\n\treturn AutomodRules(queryMods...)\n}", "func (transactionRepo *mockTransactionRepo) Initialize(ctx context.Context, db *sql.DB) {}", "func withTransaction(db *sql.DB, fn func(txn *sql.Tx) error) (err error) {\n\ttxn, err := db.Begin()\n\tif err != nil {\n\t\treturn\n\t}\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\ttxn.Rollback()\n\t\t\tpanic(r)\n\t\t} else if err != nil {\n\t\t\ttxn.Rollback()\n\t\t} else {\n\t\t\terr = txn.Commit()\n\t\t}\n\t}()\n\terr = fn(txn)\n\treturn\n}", "func (t *ACLRole) DBCreateIgnoreDuplicateTx(ctx context.Context, tx Tx) (sql.Result, error) {\n\tq := \"INSERT INTO `acl_role` (`acl_role`.`id`,`acl_role`.`checksum`,`acl_role`.`name`,`acl_role`.`description`,`acl_role`.`admin_user_id`,`acl_role`.`customer_id`,`acl_role`.`created_at`,`acl_role`.`updated_at`) VALUES (?,?,?,?,?,?,?,?) ON DUPLICATE KEY UPDATE `id` = `id`\"\n\tchecksum := t.CalculateChecksum()\n\tif t.GetChecksum() == checksum {\n\t\treturn nil, nil\n\t}\n\tt.Checksum = &checksum\n\treturn tx.ExecContext(ctx, q,\n\t\torm.ToSQLString(t.ID),\n\t\torm.ToSQLString(t.Checksum),\n\t\torm.ToSQLString(t.Name),\n\t\torm.ToSQLString(t.Description),\n\t\torm.ToSQLString(t.AdminUserID),\n\t\torm.ToSQLString(t.CustomerID),\n\t\torm.ToSQLInt64(t.CreatedAt),\n\t\torm.ToSQLInt64(t.UpdatedAt),\n\t)\n}", "func (db *DB) Transaction(ctx context.Context, fn TxHandlerFunc) error {\n\tdb.mu.Lock()\n\tdefer db.mu.Unlock()\n\n\torigin, err := db.master.BeginTx(ctx, nil)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to begin transaction: %v\", err)\n\t}\n\ttx := &Tx{origin}\n\n\tif err := fn(ctx, tx); err != nil {\n\t\tif re := tx.parent.Rollback(); re != nil {\n\t\t\tif re.Error() != sql.ErrTxDone.Error() {\n\t\t\t\treturn fmt.Errorf(\"fialed to rollback: %v\", err)\n\t\t\t}\n\t\t}\n\t\treturn fmt.Errorf(\"failed to execcute transaction: %v\", err)\n\t}\n\treturn tx.parent.Commit()\n}", "func AcidCtx(body AcidTx, db Execer) error {\n\n\t// Make the new Tx.\n\ttx, err := db.Beginx()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer tx.Commit()\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\ttx.Rollback()\n\t\t\tlog.Error(\"Transaction rollback.\")\n\t\t\terr = r.(error)\n\t\t\tlog.Error(err)\n\t\t}\n\t}()\n\n\tbody(tx)\n\treturn err\n}", "func (r *V1Beta1ExtensionsIngressRule) createRule(ingress *v1beta1Extensions.Ingress, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(ingress)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(ingress)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(ingress)\n\t\t},\n\t}\n\treturn rule\n}", "func (tp *TxPool) AddTransactionToDatabaseMempool(txHash *common.Hash, txDesc TxDesc) error {\n\ttx := txDesc.Desc.Tx\n\ttempDesc := TempDesc{\n\t\tStartTime: txDesc.StartTime,\n\t\tIsPushMessage: txDesc.IsFowardMessage,\n\t\tHeight: txDesc.Desc.Height,\n\t\tFee: txDesc.Desc.Fee,\n\t\tFeePerKB: txDesc.Desc.FeePerKB,\n\t}\n\tswitch tx.GetType() {\n\t//==================For PRV Transfer Only\n\tcase common.TxNormalType:\n\t\t{\n\t\t\tnormalTx := tx.(*transaction.Tx)\n\t\t\tvalueTx, err := json.Marshal(normalTx)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tvalueDesc, err := json.Marshal(tempDesc)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\terr = tp.config.DataBaseMempool.AddTransaction(txHash, common.TxNormalType, valueTx, valueDesc)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t//==================For PRV & TxCustomToken Transfer\n\tcase common.TxCustomTokenType:\n\t\t{\n\t\t\tcustomTokenTx := tx.(*transaction.TxCustomToken)\n\t\t\tvalueTx, err := json.Marshal(customTokenTx)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tvalueDesc, err := json.Marshal(tempDesc)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\terr = tp.config.DataBaseMempool.AddTransaction(txHash, common.TxCustomTokenType, valueTx, valueDesc)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\tcase common.TxCustomTokenPrivacyType:\n\t\t{\n\t\t\tcustomTokenPrivacyTx := tx.(*transaction.TxCustomTokenPrivacy)\n\t\t\tvalueTx, err := json.Marshal(customTokenPrivacyTx)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tvalueDesc, err := json.Marshal(tempDesc)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\terr = tp.config.DataBaseMempool.AddTransaction(txHash, common.TxCustomTokenPrivacyType, valueTx, valueDesc)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (s *BaseQueryListener) EnterEveryRule(ctx antlr.ParserRuleContext) {}", "func (t *Table) InsertRulesNow(chain string, rules []Rule) error {\n\tfeatures := t.featureDetector.GetFeatures()\n\n\thashes := CalculateRuleHashes(chain, rules, features)\n\n\tbuf := new(RestoreInputBuilder)\n\tbuf.StartTransaction(t.Name)\n\tfor i, r := range rules {\n\t\tprefixFrag := t.commentFrag(hashes[i])\n\t\tbuf.WriteLine(r.RenderInsertAtRuleNumber(chain, i+1, prefixFrag, features))\n\t}\n\tbuf.EndTransaction()\n\n\tif err := t.execIptablesRestore(buf); err != nil {\n\t\treturn fmt.Errorf(\"writting out buffer: %w\", err)\n\t}\n\n\treturn nil\n}", "func (pg *Postgres) Tx(ctx context.Context, txFunc store.TxFunc) (err error) {\n\ttx := shared.GetTx(ctx)\n\n\tif tx != nil {\n\t\treturn txFunc(ctx)\n\t}\n\n\ttx, err = pg.BeginTx(ctx, nil)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"begin tx failed\")\n\t}\n\n\tctx = shared.WithTx(ctx, tx)\n\n\t//nolint:gocritic\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tif err := tx.Rollback(); err != nil {\n\t\t\t\tlog.Warn(ctx, \"tx rollback failed\", \"err\", err)\n\t\t\t}\n\t\t\tpanic(r)\n\t\t} else if err != nil {\n\t\t\tif err := tx.Rollback(); err != nil {\n\t\t\t\tlog.Warn(ctx, \"tx rollback failed\", \"err\", err)\n\t\t\t}\n\t\t} else {\n\t\t\terr = tx.Commit()\n\t\t}\n\t}()\n\n\terr = txFunc(ctx)\n\n\treturn err\n}", "func (r *BatchV1Beta1CronJobRule) createRule(cronjob *batchV1beta1.CronJob, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(cronjob)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(cronjob)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(cronjob)\n\t\t},\n\t}\n\treturn rule\n}", "func dataSourceDatabase() *schema.Resource {\n\treturn &schema.Resource{\n\t\tReadContext: dataSourcesDatabaseRead,\n\t\tSchema: map[string]*schema.Schema{\n\t\t\t\"instance_name\": &schema.Schema{\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tRequired: true,\n\t\t\t},\n\t\t\t\"username\": &schema.Schema{\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tComputed: true,\n\t\t\t},\n\t\t\t\"password\": &schema.Schema{\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tSensitive: true,\n\t\t\t\tComputed: true,\n\t\t\t},\n\t\t\t\"read_hostgroup\": &schema.Schema{\n\t\t\t\tType: schema.TypeInt,\n\t\t\t\tComputed: true,\n\t\t\t},\n\t\t\t\"write_hostgroup\": &schema.Schema{\n\t\t\t\tType: schema.TypeInt,\n\t\t\t\tComputed: true,\n\t\t\t},\n\t\t\t\"enable_ssl\": &schema.Schema{\n\t\t\t\tType: schema.TypeInt,\n\t\t\t\tComputed: true,\n\t\t\t},\n\t\t\t\"query_rules\": &schema.Schema{\n\t\t\t\tType: schema.TypeList,\n\t\t\t\tComputed: true,\n\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\tSchema: map[string]*schema.Schema{\n\t\t\t\t\t\t\"rule_id\": &schema.Schema{\n\t\t\t\t\t\t\tType: schema.TypeInt,\n\t\t\t\t\t\t\tComputed: true,\n\t\t\t\t\t\t},\n\t\t\t\t\t\t\"username\": &schema.Schema{\n\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\tComputed: true,\n\t\t\t\t\t\t},\n\t\t\t\t\t\t\"active\": &schema.Schema{\n\t\t\t\t\t\t\tType: schema.TypeInt,\n\t\t\t\t\t\t\tComputed: true,\n\t\t\t\t\t\t},\n\t\t\t\t\t\t\"match_digest\": &schema.Schema{\n\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\tComputed: true,\n\t\t\t\t\t\t},\n\t\t\t\t\t\t\"destination_hostgroup\": &schema.Schema{\n\t\t\t\t\t\t\tType: schema.TypeInt,\n\t\t\t\t\t\t\tComputed: true,\n\t\t\t\t\t\t},\n\t\t\t\t\t\t\"apply\": &schema.Schema{\n\t\t\t\t\t\t\tType: schema.TypeInt,\n\t\t\t\t\t\t\tComputed: true,\n\t\t\t\t\t\t},\n\t\t\t\t\t\t\"comment\": &schema.Schema{\n\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\tComputed: true,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\t\"master_instance\": &schema.Schema{\n\t\t\t\tType: schema.TypeMap,\n\t\t\t\tComputed: true,\n\t\t\t},\n\t\t\t\"read_replicas\": &schema.Schema{\n\t\t\t\tType: schema.TypeList,\n\t\t\t\tComputed: true,\n\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\tSchema: map[string]*schema.Schema{\n\t\t\t\t\t\t\"name\": &schema.Schema{\n\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\tComputed: true,\n\t\t\t\t\t\t},\n\t\t\t\t\t\t\"ip_address\": &schema.Schema{\n\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\tComputed: true,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func Guard(description string, predicate func(c *db.Connection, tx *sql.Tx) (bool, error)) GuardFunc {\n\treturn func(ctx context.Context, c *db.Connection, tx *sql.Tx, step Action) error {\n\t\tproceed, err := predicate(c, tx)\n\t\tif err != nil {\n\t\t\tif suite := GetContextSuite(ctx); suite != nil {\n\t\t\t\treturn suite.Error(WithLabel(ctx, description), err)\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\n\t\tif !proceed {\n\t\t\tif suite := GetContextSuite(ctx); suite != nil {\n\t\t\t\tsuite.Skipf(ctx, description)\n\t\t\t}\n\t\t\treturn nil\n\t\t}\n\n\t\terr = step(ctx, c, tx)\n\t\tif err != nil {\n\t\t\tif suite := GetContextSuite(ctx); suite != nil {\n\t\t\t\treturn suite.Error(WithLabel(ctx, description), err)\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\t\tif suite := GetContextSuite(ctx); suite != nil {\n\t\t\tsuite.Applyf(ctx, description)\n\t\t}\n\t\treturn nil\n\t}\n}", "func (m *RuleBasedSubjectSet) SetRule(value *string)() {\n err := m.GetBackingStore().Set(\"rule\", value)\n if err != nil {\n panic(err)\n }\n}", "func (r *V1ServiceRule) createRule(service *v1.Service, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(service)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(service)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(service)\n\t\t},\n\t}\n\treturn rule\n}", "func (p *Postgres) Tx(ctx context.Context, txFunc store.TxFunc) (err error) {\n\ttx := shared.GetTx(ctx)\n\n\tif tx != nil {\n\t\treturn txFunc(ctx)\n\t}\n\n\ttx, err = p.BeginTx(ctx, nil)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"begin tx failed\")\n\t}\n\n\tctx = shared.WithTx(ctx, tx)\n\n\t//nolint:gocritic\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tif err := tx.Rollback(); err != nil {\n\t\t\t\tlog.Warn(ctx, \"tx rollback failed\", \"err\", err)\n\t\t\t}\n\t\t\tpanic(r)\n\t\t} else if err != nil {\n\t\t\tif err := tx.Rollback(); err != nil {\n\t\t\t\tlog.Warn(ctx, \"tx rollback failed\", \"err\", err)\n\t\t\t}\n\t\t} else {\n\t\t\terr = tx.Commit()\n\t\t}\n\t}()\n\n\terr = txFunc(ctx)\n\n\treturn err\n}", "func runInTransaction(fn func(tx *sql.Tx) error) (err error) {\n\tvar tx *sql.Tx\n\tdefer func() {\n\t\tif e := recover(); e != nil && tx != nil {\n\t\t\tif e1 := tx.Rollback(); e1 != nil {\n\t\t\t\terr = e1\n\t\t\t} else {\n\t\t\t\terr = e.(error)\n\t\t\t}\n\t\t}\n\t}()\n\n\tif tx, err = db.Begin(); err == nil {\n\t\terr = fn(tx)\n\t\tif err == nil {\n\t\t\terr = tx.Commit()\n\t\t} else if e := tx.Rollback(); e != nil {\n\t\t\terr = e\n\t\t}\n\t}\n\treturn err\n}", "func RegisterTx(fns ...func(DB) error) error {\n\treturn DefaultCollection.RegisterTx(fns...)\n}", "func (updEP *MdlPersonUpdateExt) BeforeDB(ent interface{}) error {\n\n\t// fmt.Println(\"TypeOf ent:\", reflect.TypeOf(ent))\n\t// fmt.Println(\"ValueOf ent:\", reflect.ValueOf(ent))\n\t// p := ent.(*models.Person)\n\n\t// make changes / validate the content struct pointer (p) here\n\t// p.Name = \"A new field value\"\n\treturn nil\n}", "func (dao *ConfigAuditProcessDao) Transaction(ctx context.Context, f func(ctx context.Context, tx *gdb.TX) error) (err error) {\n\treturn dao.Ctx(ctx).Transaction(ctx, f)\n}", "func (self *PolicyAgent) AddRule(rule *OfnetPolicyRule, ret *bool) error {\n\tvar ipDa *net.IP = nil\n\tvar ipDaMask *net.IP = nil\n\tvar ipSa *net.IP = nil\n\tvar ipSaMask *net.IP = nil\n\tvar md *uint64 = nil\n\tvar mdm *uint64 = nil\n\tvar flag, flagMask uint16\n\tvar flagPtr, flagMaskPtr *uint16\n\tvar err error\n\n\t// make sure switch is connected\n\tif !self.agent.IsSwitchConnected() {\n\t\tself.agent.WaitForSwitchConnection()\n\t}\n\n\t// check if we already have the rule\n\tself.mutex.RLock()\n\tif self.Rules[rule.RuleId] != nil {\n\t\toldRule := self.Rules[rule.RuleId].Rule\n\n\t\tif ruleIsSame(oldRule, rule) {\n\t\t\tself.mutex.RUnlock()\n\t\t\treturn nil\n\t\t} else {\n\t\t\tself.mutex.RUnlock()\n\t\t\tlog.Errorf(\"Rule already exists. new rule: {%+v}, old rule: {%+v}\", rule, oldRule)\n\t\t\treturn errors.New(\"Rule already exists\")\n\t\t}\n\t}\n\tself.mutex.RUnlock()\n\n\tlog.Infof(\"Received AddRule: %+v\", rule)\n\n\t// Parse dst ip\n\tif rule.DstIpAddr != \"\" {\n\t\tipDa, ipDaMask, err = ParseIPAddrMaskString(rule.DstIpAddr)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error parsing dst ip %s. Err: %v\", rule.DstIpAddr, err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// parse src ip\n\tif rule.SrcIpAddr != \"\" {\n\t\tipSa, ipSaMask, err = ParseIPAddrMaskString(rule.SrcIpAddr)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error parsing src ip %s. Err: %v\", rule.SrcIpAddr, err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// parse source/dst endpoint groups\n\tif rule.SrcEndpointGroup != 0 && rule.DstEndpointGroup != 0 {\n\t\tsrcMetadata, srcMetadataMask := SrcGroupMetadata(rule.SrcEndpointGroup)\n\t\tdstMetadata, dstMetadataMask := DstGroupMetadata(rule.DstEndpointGroup)\n\t\tmetadata := srcMetadata | dstMetadata\n\t\tmetadataMask := srcMetadataMask | dstMetadataMask\n\t\tmd = &metadata\n\t\tmdm = &metadataMask\n\t} else if rule.SrcEndpointGroup != 0 {\n\t\tsrcMetadata, srcMetadataMask := SrcGroupMetadata(rule.SrcEndpointGroup)\n\t\tmd = &srcMetadata\n\t\tmdm = &srcMetadataMask\n\t} else if rule.DstEndpointGroup != 0 {\n\t\tdstMetadata, dstMetadataMask := DstGroupMetadata(rule.DstEndpointGroup)\n\t\tmd = &dstMetadata\n\t\tmdm = &dstMetadataMask\n\t}\n\n\t// Setup TCP flags\n\tif rule.IpProtocol == 6 && rule.TcpFlags != \"\" {\n\t\tswitch rule.TcpFlags {\n\t\tcase \"syn\":\n\t\t\tflag = TCP_FLAG_SYN\n\t\t\tflagMask = TCP_FLAG_SYN\n\t\tcase \"syn,ack\":\n\t\t\tflag = TCP_FLAG_ACK | TCP_FLAG_SYN\n\t\t\tflagMask = TCP_FLAG_ACK | TCP_FLAG_SYN\n\t\tcase \"ack\":\n\t\t\tflag = TCP_FLAG_ACK\n\t\t\tflagMask = TCP_FLAG_ACK\n\t\tcase \"syn,!ack\":\n\t\t\tflag = TCP_FLAG_SYN\n\t\t\tflagMask = TCP_FLAG_ACK | TCP_FLAG_SYN\n\t\tcase \"!syn,ack\":\n\t\t\tflag = TCP_FLAG_ACK\n\t\t\tflagMask = TCP_FLAG_ACK | TCP_FLAG_SYN\n\t\tdefault:\n\t\t\tlog.Errorf(\"Unknown TCP flags: %s, in rule: %+v\", rule.TcpFlags, rule)\n\t\t\treturn errors.New(\"Unknown TCP flag\")\n\t\t}\n\n\t\tflagPtr = &flag\n\t\tflagMaskPtr = &flagMask\n\t}\n\t// Install the rule in policy table\n\truleFlow, err := self.policyTable.NewFlow(ofctrl.FlowMatch{\n\t\tPriority: uint16(FLOW_POLICY_PRIORITY_OFFSET + rule.Priority),\n\t\tEthertype: 0x0800,\n\t\tIpDa: ipDa,\n\t\tIpDaMask: ipDaMask,\n\t\tIpSa: ipSa,\n\t\tIpSaMask: ipSaMask,\n\t\tIpProto: rule.IpProtocol,\n\t\tTcpSrcPort: rule.SrcPort,\n\t\tTcpDstPort: rule.DstPort,\n\t\tUdpSrcPort: rule.SrcPort,\n\t\tUdpDstPort: rule.DstPort,\n\t\tMetadata: md,\n\t\tMetadataMask: mdm,\n\t\tTcpFlags: flagPtr,\n\t\tTcpFlagsMask: flagMaskPtr,\n\t})\n\tif err != nil {\n\t\tlog.Errorf(\"Error adding flow for rule {%v}. Err: %v\", rule, err)\n\t\treturn err\n\t}\n\n\t// Point it to next table\n\tif rule.Action == \"allow\" {\n\t\terr = ruleFlow.Next(self.nextTable)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error installing flow {%+v}. Err: %v\", ruleFlow, err)\n\t\t\treturn err\n\t\t}\n\t} else if rule.Action == \"deny\" {\n\t\terr = ruleFlow.Next(self.ofSwitch.DropAction())\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error installing flow {%+v}. Err: %v\", ruleFlow, err)\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tlog.Errorf(\"Unknown action in rule {%+v}\", rule)\n\t\treturn errors.New(\"Unknown action in rule\")\n\t}\n\n\t// save the rule\n\tpRule := PolicyRule{\n\t\tRule: rule,\n\t\tflow: ruleFlow,\n\t}\n\tself.mutex.Lock()\n\tself.Rules[rule.RuleId] = &pRule\n\tself.mutex.Unlock()\n\n\treturn nil\n}", "func (d DB) Transaction(f func(DB) error) error {\n\tif _, ok := d.dbProxy.(*sql.Tx); ok {\n\t\t// Already in a nested transaction\n\t\treturn f(d)\n\t}\n\n\ttx, err := d.dbProxy.(*sql.DB).Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = f(DB{\n\t\tdbProxy: tx,\n\t\tStatementBuilderType: statementBuilder(tx),\n\t})\n\tif err != nil {\n\t\t// Rollback error is ignored as we already have one in progress\n\t\tif err2 := tx.Rollback(); err2 != nil {\n\t\t\tlevel.Warn(util_log.Logger).Log(\"msg\", \"transaction rollback error (ignored)\", \"err\", err2)\n\t\t}\n\t\treturn err\n\t}\n\treturn tx.Commit()\n}", "func Transaction(db *sql.DB, f func()) {\n\n\ttx, err := db.Begin()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\terr := tx.Rollback()\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t\tpanic(r)\n\t\t} else {\n\t\t\terr = tx.Commit()\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t}\n\t}()\n\tf()\n}", "func setupDatabase(database *sql.DB) {\n\t_, err := database.Exec(`\n\t\tCREATE TABLE IF NOT EXISTS resources (\n\t\t\tid INTEGER PRIMARY KEY,\n\t\t\towner_id BLOB,\n\t\t\tstr_id TEXT UNIQUE NOT NULL\n\t\t);\n\t`)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"Could not setup table: 'resources': %s\", err))\n\t}\n\t_, err = database.Exec(`\n\t\tCREATE TABLE IF NOT EXISTS resource_permissions (\n\t\t\tresource_id INTEGER,\n\t\t\tuser_id BLOB,\n\t\t\tpermissions INTEGER NOT NULL,\n\t\t\tPRIMARY KEY(resource_id, user_id)\n\t\t);\n\t`)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"Could not setup table: 'resource_permissions': %s\", err))\n\t}\n\t_, err = database.Exec(`\n\t\tCREATE TABLE IF NOT EXISTS users (\n\t\t\tid BLOB,\n\t\t\tusername TEXT NOT NULL,\n\t\t\tpassword TEXT NOT NULL,\n\t\t\tPRIMARY KEY (id)\n\t\t);\n\t`)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"Could not setup table: 'users': %s\", err))\n\t}\n\n\t_, err = database.Exec(`\n\t\tCREATE INDEX IF NOT EXISTS str_id\n\t\tON resources (str_id);\n\t`)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"Could not create index: 'resources.resource_strid': %s\", err))\n\t}\n\n\t_, err = database.Exec(`\n\t\tCREATE INDEX IF NOT EXISTS username\n\t\tON users (username);\n\t`)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"Could not create index: 'users.username': %s\", err))\n\t}\n}", "func (act *CreateAction) Do() error {\n\t// business sharding db.\n\tsd, err := act.smgr.ShardingDB(act.req.BizId)\n\tif err != nil {\n\t\treturn act.Err(pbcommon.ErrCode_E_DM_ERR_DBSHARDING, err.Error())\n\t}\n\tact.sd = sd\n\tact.tx = act.sd.DB().Begin()\n\n\t// query config template.\n\tif errCode, errMsg := act.queryConfigTemplate(); errCode != pbcommon.ErrCode_E_OK {\n\t\tact.tx.Rollback()\n\t\treturn act.Err(errCode, errMsg)\n\t}\n\n\t// create config template bind relation.\n\tif errCode, errMsg := act.createTemplateBind(); errCode != pbcommon.ErrCode_E_OK {\n\t\tact.tx.Rollback()\n\t\treturn act.Err(errCode, errMsg)\n\t}\n\n\t// create config.\n\tif errCode, errMsg := act.createBindConfig(); errCode != pbcommon.ErrCode_E_OK {\n\t\tact.tx.Rollback()\n\t\treturn act.Err(errCode, errMsg)\n\t}\n\n\t// commit tx.\n\tif err := act.tx.Commit().Error; err != nil {\n\t\tact.tx.Rollback()\n\t\treturn act.Err(pbcommon.ErrCode_E_DM_SYSTEM_UNKNOWN, err.Error())\n\t}\n\n\treturn nil\n}", "func setupDB(DB *gorm.DB) {\n\t// remove original functions to update timestamps, we will maintain that by ourself in models in\n\t// BeforeCreate, BeforeSave methods\n\tDB.Callback().Create().Remove(\"gorm:update_time_stamp\")\n}", "func (user *User) BeforeCreate(transaction *gorm.DB) error {\n return nil\n}" ]
[ "0.5817582", "0.53555834", "0.5226998", "0.5186216", "0.508085", "0.5008556", "0.4957827", "0.4957827", "0.49559155", "0.49460205", "0.48994553", "0.48729715", "0.48546115", "0.48274273", "0.4813973", "0.48122132", "0.47885105", "0.47852197", "0.47779572", "0.47655752", "0.47617877", "0.47612348", "0.4754884", "0.47190854", "0.46930078", "0.46715593", "0.46635315", "0.46629536", "0.46570987", "0.46526444", "0.46497536", "0.4626282", "0.46232662", "0.46075338", "0.46021363", "0.459993", "0.45802516", "0.45746082", "0.4571494", "0.45669496", "0.45653403", "0.4561449", "0.4561246", "0.45583716", "0.45533574", "0.45351747", "0.4528399", "0.4523689", "0.45224053", "0.45153013", "0.45025626", "0.45017695", "0.44906035", "0.44870833", "0.44855362", "0.4478724", "0.44761842", "0.4474118", "0.44670668", "0.4466964", "0.44660226", "0.44636866", "0.4462552", "0.4458273", "0.44581297", "0.44575998", "0.4455761", "0.44501427", "0.4444859", "0.44410142", "0.4434069", "0.44328365", "0.4431348", "0.442381", "0.441844", "0.44180405", "0.44077432", "0.44052348", "0.4402005", "0.44019258", "0.4399329", "0.43985125", "0.43925354", "0.43900397", "0.43868467", "0.43859467", "0.43840742", "0.43790042", "0.4373893", "0.4372848", "0.43675396", "0.43643156", "0.4364043", "0.4362277", "0.43621835", "0.43584147", "0.4355341", "0.43481714", "0.43462926", "0.43458593" ]
0.67752594
0
GetTx function will return specific database transaction
func (gtx *GuardTx) GetTx() *sql.Tx { return gtx.dbTx }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (uts *UnapprovedTransactions) GetTransaction(txID []byte) ([]byte, error) {\n\treturn uts.DB.Get(uts.getTableName(), txID)\n}", "func (db *DB) GetTx() *GetTx {\n\treturn &GetTx{\n\t\tdb: db,\n\t}\n}", "func GetTransaction(_db *gorm.DB, blkHash common.Hash, txHash common.Hash) *Transactions {\n\tvar tx Transactions\n\n\tif err := _db.Where(\"hash = ? and blockhash = ?\", txHash.Hex(), blkHash.Hex()).First(&tx).Error; err != nil {\n\t\treturn nil\n\t}\n\n\treturn &tx\n}", "func GetTx(ctx context.Context) (*firestore.Transaction, bool) {\n\ttx, ok := ctx.Value(txKey).(*firestore.Transaction)\n\treturn tx, ok\n}", "func (c *dummyWavesMDLrpcclient) GetTransaction(txid string) (*model.Transactions, error) {\n\ttransaction, _, err := client.NewTransactionsService(c.MainNET).GetTransactionsInfoID(txid)\n\treturn transaction, err\n}", "func GetTransaction(ctx context.Context) *sql.Tx {\n\tiTx := ctx.Value(Transaction)\n\ttx, _ := iTx.(*sql.Tx) //nolint: errcheck\n\treturn tx\n}", "func GetTx() *TX {\n\ttx := &TX{\n\t\tDB: DB.Begin(),\n\t\tfired: false,\n\t}\n\treturn tx\n}", "func (c *Context) GetTx() interface{} {\n\treturn c.tx\n}", "func (b *Backend) GetTransaction(\n\tctx context.Context,\n\tid sdk.Identifier,\n) (*sdk.Transaction, error) {\n\ttx, err := b.emulator.GetTransaction(id)\n\tif err != nil {\n\t\tswitch err.(type) {\n\t\tcase emulator.NotFoundError:\n\t\t\treturn nil, status.Error(codes.NotFound, err.Error())\n\t\tdefault:\n\t\t\treturn nil, status.Error(codes.Internal, err.Error())\n\t\t}\n\t}\n\n\tb.logger.\n\t\tWithField(\"txID\", id.String()).\n\t\tDebugf(\"💵 GetTransaction called\")\n\n\treturn tx, nil\n}", "func (gw *Gateway) GetTransaction(txid cipher.SHA256) (*visor.Transaction, error) {\n\tvar txn *visor.Transaction\n\tvar err error\n\n\tgw.strand(\"GetTransaction\", func() {\n\t\ttxn, err = gw.v.GetTransaction(txid)\n\t})\n\n\treturn txn, err\n}", "func GetTransferringTx(db *gorp.DbMap) (records []*TxRecord, err error) {\n\t_, err = db.Select(&records, `SELECT * FROM \"record\" WHERE \"state\" = ?`, ExchangeStateTransferring)\n\treturn\n}", "func GetTX(c echo.Context) newrelic.Transaction {\n\ttx := c.Get(\"txn\")\n\tif tx == nil {\n\t\treturn nil\n\t}\n\n\treturn tx.(newrelic.Transaction)\n}", "func GetTX(c echo.Context) newrelic.Transaction {\n\ttx := c.Get(\"txn\")\n\tif tx == nil {\n\t\treturn nil\n\t}\n\n\treturn tx.(newrelic.Transaction)\n}", "func GetTX(c echo.Context) newrelic.Transaction {\n\ttx := c.Get(\"txn\")\n\tif tx == nil {\n\t\treturn nil\n\t}\n\n\treturn tx.(newrelic.Transaction)\n}", "func (s *TXPoolServer) getTransaction(hash common.Uint256) *tx.Transaction {\n\treturn s.txPool.GetTransaction(hash)\n}", "func GetTx(txhash string) (*model.Tx, error) {\n\turl := fmt.Sprintf(bchapi.TxUrl, txhash)\n\tresult, err := bchapi.HttpGet(url, bchapi.ConnTimeoutMS, bchapi.ServeTimeoutMS)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttx, err := model.StringToTx(result)\n\treturn tx, err\n}", "func GetTransaction(transactionID bson.ObjectId) (*models.Transaction, error) {\n\tsession, collection := service.Connect(collectionName)\n\tdefer session.Close()\n\n\ttransaction := models.Transaction{}\n\terr := collection.FindId(transactionID).One(&transaction)\n\n\treturn &transaction, err\n}", "func (data *Data) GetTx(hash chainhash.Hash) (*wire.MsgTx, error) {\n\tdb, err := data.openDb()\n\tdefer data.closeDb(db)\n\tif err != nil {\n\t\tlog.Printf(\"data.openDb Error : %+v\", err)\n\t\treturn nil, err\n\t}\n\tvar bs []byte\n\terr = db.QueryRow(\"SELECT data FROM tx WHERE hash=?\", hash.CloneBytes()).Scan(&bs)\n\tif err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\treturn nil, nil\n\t\t}\n\t\tlog.Printf(\"db.QueryRow Error : %+v\", err)\n\t\treturn nil, err\n\t}\n\ttx, err := data.bsToMsgTx(bs)\n\tif err != nil {\n\t\tlog.Printf(\"data.bsToMsgTx Error : %+v\", err)\n\t\treturn nil, err\n\t}\n\treturn tx, nil\n}", "func (ps *PubsubApi) GetTransaction(hash common.Hash) *rtypes.RPCTx {\n\ttx, txEntry := ps.backend().GetTx(hash)\n\tif tx == nil {\n\t\tlog.Info(\"GetTransaction fail\", \"hash\", hash)\n\t}\n\treturn rtypes.NewRPCTx(tx, txEntry)\n}", "func (r *Repository) TxGet(tx *dbr.Tx, userID int64) (*pb.User, error) {\n\treturn r.get(tx, userID)\n}", "func (c *RPC) GetTransaction(txid string) (*webrpc.TxnResult, error) {\n\ttxn, err := c.rpcClient.GetTransactionByID(txid)\n\tif err != nil {\n\t\treturn nil, RPCError{err}\n\t}\n\n\treturn txn, nil\n}", "func (svc *svc) GetTransaction(ctx context.Context, query model.TransactionQuery) (model.Transaction, error) {\n\ttransaction, err := svc.repo.GetTransaction(ctx, query)\n\tif err != nil {\n\t\treturn transaction, err\n\t}\n\n\treturn transaction, nil\n}", "func (s *Store) GetTx(txid common.Hash) *types.Transaction {\n\ttx, _ := s.rlp.Get(s.table.Txs, txid.Bytes(), &types.Transaction{}).(*types.Transaction)\n\n\treturn tx\n}", "func GetTransaction(id int, db *gorm.DB) Transaction {\n\tvar t Transaction\n\terr := db.Table(\"transactions\").\n\t\tPreload(\"Tags\").\n\t\tFirst(&t, id).\n\t\tError\n\tif err != nil {\n\t\tt.Date = time.Now()\n\t}\n\treturn t\n}", "func (tb *TransactionBuilder) GetTransaction() *types.Transaction {\n\treturn tb.tx\n}", "func (mp *TxPool) GetTransaction(hash Uint256) *Transaction {\n\tmp.RLock()\n\tdefer mp.RUnlock()\n\treturn mp.txnList[hash]\n}", "func (sc *ServerConn) GetTransaction(ctx context.Context, txid string) (*GetTransactionResult, error) {\n\tvar resp GetTransactionResult\n\terr := sc.Request(ctx, \"blockchain.transaction.get\", positional{txid, true}, &resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func getTx(txn *badger.Txn) func([]byte) ([]byte, error) {\n\treturn func(key []byte) ([]byte, error) {\n\t\t// Badger returns an \"item\" upon GETs, we need to copy the actual value\n\t\t// from the item and return it.\n\t\titem, err := txn.Get(key)\n\t\tif err != nil {\n\t\t\tif errors.Is(err, badger.ErrKeyNotFound) {\n\t\t\t\treturn nil, storage.ErrNotFound\n\t\t\t}\n\t\t\treturn nil, err\n\t\t}\n\n\t\tval := make([]byte, item.ValueSize())\n\t\treturn item.ValueCopy(val)\n\t}\n}", "func (tx *txDriver) Tx(context.Context) (dialect.Tx, error) { return tx, nil }", "func (tx *txDriver) Tx(context.Context) (dialect.Tx, error) { return tx, nil }", "func (u *User) GetTransaction(nodeID, transactionID string) (map[string]interface{}, error) {\n\tlog.info(\"========== GET TRANSACTION ==========\")\n\turl := buildURL(path[\"users\"], u.UserID, path[\"nodes\"], nodeID, path[\"transactions\"], transactionID)\n\n\treturn u.do(\"GET\", url, \"\", nil)\n}", "func GetTransaction() gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tsugar, _ := item.New(\"Sugar\", map[string]float64{\"Kabras\": 110, \"Mumias\": 110}, \"kg(s)\")\n\t\tpurchase, message, err := transaction.New(sugar, map[string]float64{\"Nzoia\": 150}, 3)\n\t\tc.JSON(\n\t\t\thttp.StatusOK,\n\t\t\tGetResponse{GetData{purchase}, message, responseerr.GetStrErr(err)},\n\t\t)\n\t}\n}", "func (tx *tX) Transaction() (*tX, error) {\n\treturn tx, nil\n}", "func (tbl DbCompoundTable) Tx() *sql.Tx {\n\treturn tbl.db.(*sql.Tx)\n}", "func (s *Service) GetExplorerTransaction(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tid := r.FormValue(\"id\")\n\n\tdata := &Data{}\n\tdefer func() {\n\t\tif err := json.NewEncoder(w).Encode(data.TX); err != nil {\n\t\t\tutils.Logger().Warn().Err(err).Msg(\"cannot JSON-encode TX\")\n\t\t}\n\t}()\n\tif id == \"\" {\n\t\tutils.Logger().Warn().Msg(\"invalid id parameter\")\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\tdb := s.Storage.GetDB()\n\tbytes, err := db.Get([]byte(GetTXKey(id)))\n\tif err != nil {\n\t\tutils.Logger().Warn().Err(err).Str(\"id\", id).Msg(\"cannot read TX\")\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\ttx := new(Transaction)\n\tif rlp.DecodeBytes(bytes, tx) != nil {\n\t\tutils.Logger().Warn().Str(\"id\", id).Msg(\"cannot convert data from DB\")\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\tdata.TX = *tx\n}", "func (b *DatabaseTestSuiteBase) Tx() sqlx.Ext {\n\treturn b.tx\n}", "func (trs *Transaction) GetTransaction() stored_transactions.Transaction {\n\treturn trs.Trs\n}", "func (tp *TXPool) GetTransaction(hash common.Uint256) *types.Transaction {\n\ttp.RLock()\n\tdefer tp.RUnlock()\n\tif tx := tp.txList[hash]; tx == nil {\n\t\treturn nil\n\t}\n\treturn tp.txList[hash].Tx\n}", "func (r *Redis) GetTx(txID ids.ID) ([]byte, error) {\n\tctx, cancelFn := context.WithTimeout(context.Background(), redisTimeout)\n\tdefer cancelFn()\n\n\tcmd := r.client.Get(ctx, redisIndexKeysTxByID(r.chainID.String(), txID.String()))\n\tif err := cmd.Err(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn cmd.Bytes()\n}", "func (p *PgDb) GetEmpTx(id int, tx *sqlx.Tx) (*model.Emp, error) {\n\tcnt := \"postgres (p *PgDb) GetEmpTx\" // Имя текущего метода для логирования\n\t//mylog.PrintfDebug(\"[DEBUG] %v - START, param: '%+v'\", cnt, id)\n\n\t// =====================================================================\n\t// Переменная часть кода\n\t// =====================================================================\n\tvar stm *sqlx.Stmt\n\tif tx == nil {\n\t\t// Работаем без транзакции\n\t\t//mylog.PrintfDebug(\"[DEBUG] %v - NON TX mode\", cnt)\n\t\tstm = p.sqlSelectEmp\n\t} else {\n\t\t// Помещаем запрос в транзакцию\n\t\tstm = tx.Stmtx(p.sqlSelectEmp)\n\t}\n\n\t// Извлечем из кэша новую структуру\n\tv := model.GetEmp()\n\t// =====================================================================\n\n\t//Выполняем запрос - ожидаем одну строку\n\terr := stm.Get(v, id)\n\tif err != nil {\n\t\t// проверим NO_DATA_FOUND\n\t\tif err == sql.ErrNoRows {\n\t\t\terrM := fmt.Sprintf(\"[DEBUG] %v - stm.Get(&v, id) - NO_DATA_FOUND, param: '%v', SQL err: '%+v'\", cnt, id, err)\n\t\t\tlog.Printf(errM)\n\t\t\treturn nil, nil\n\t\t}\n\t\t// Другие ошибки\n\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - stm.Get(&v, id), args = '%v'\", cnt, id)\n\t\tlog.Printf(errM)\n\t\treturn nil, errors.Wrap(err, errM)\n\t}\n\n\t//mylog.PrintfDebug(\"[DEBUG] %v - SUCCESS\", cnt)\n\n\treturn v, nil\n}", "func (db *DB) GetTx(tx *bolt.Tx, key interface{}, value interface{}) error {\n\tbb := db.bucket(value)\n\tb := tx.Bucket(bb)\n\tif b == nil {\n\t\treturn ErrKeyNotExists\n\t}\n\n\tkb, err := db.encode(key)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata := b.Get(kb)\n\tif data == nil {\n\t\treturn ErrKeyNotExists\n\t}\n\n\treturn db.decode(data, value)\n}", "func (s *server) GetTransaction(ctx context.Context, req *transactionpb.GetTransactionRequest) (*transactionpb.GetTransactionResponse, error) {\n\tlog := logrus.WithFields(logrus.Fields{\n\t\t\"method\": \"GetTransaction\",\n\t\t\"id\": base64.StdEncoding.EncodeToString(req.TransactionId.Value),\n\t})\n\n\tif len(req.TransactionId.Value) != 32 && len(req.TransactionId.Value) != 64 {\n\t\treturn nil, status.Error(codes.Internal, \"invalid transaction signature\")\n\t}\n\n\tresp, err := s.loader.loadTransaction(ctx, req.TransactionId.Value)\n\tif err != nil {\n\t\tlog.WithError(err).Warn(\"failed to load transaction\")\n\t\treturn nil, status.Error(codes.Internal, \"failed to load transaction\")\n\t}\n\treturn resp, nil\n}", "func (api *API) Get(tid string) (*pagarme.Response, *pagarme.Transaction, error) {\n\tresp, err := api.Config.Do(http.MethodGet, \"/transactions/\"+tid, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tif werr := www.ExtractError(resp); werr != nil {\n\t\treturn werr, nil, nil\n\t}\n\tresult := &pagarme.Transaction{}\n\tif err := www.Unmarshal(resp, result); err != nil {\n\t\tapi.Config.Logger.Error(\"could not unmarshal transaction [Get]: \" + err.Error())\n\t\treturn nil, nil, err\n\t}\n\n\treturn www.Ok(), result, nil\n}", "func (c *TransactionClient) Get(ctx context.Context, id int32) (*Transaction, error) {\n\treturn c.Query().Where(transaction.ID(id)).Only(ctx)\n}", "func GetTransaction(hostURL string, hostPort int, hash string) *bytes.Buffer {\n\tparams := make(map[string]interface{})\n\tparams[\"hash\"] = hash\n\treturn makePostRequest(hostURL, hostPort, \"f_transaction_json\", params)\n}", "func (r Virtual_Guest) GetActiveTransaction() (resp datatypes.Provisioning_Version1_Transaction, err error) {\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest\", \"getActiveTransaction\", nil, &r.Options, &resp)\n\treturn\n}", "func GetTransaction(retKey string) (models.TransactionCache, error) {\n\n\tdata, err := redisClient.Get(ctx, retKey).Bytes()\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn models.TransactionCache{}, err\n\t}\n\n\tvar transaction models.TransactionCache\n\n\terr = json.Unmarshal(data, &transaction)\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn models.TransactionCache{}, err\n\t}\n\n\treturn transaction, nil\n}", "func (c *changeTrackerDB) ReadTxn() *txn {\n\treturn &txn{Txn: c.memdb.Txn(false)}\n}", "func (b *Bucket) Tx() *Tx {\n\treturn b.tx\n}", "func (s *TransactionService) Get(walletID, txnID string) (*Transaction, error) {\n\tu := fmt.Sprintf(\"/kms/wallets/%s/transactions/%s\", walletID, txnID)\n\ttxn := &Transaction{}\n\tp := &Params{}\n\tp.SetAuthProvider(s.auth)\n\terr := s.client.Call(http.MethodGet, u, nil, txn, p)\n\treturn txn, err\n}", "func (m *SimulateRequest) GetTx() *Tx {\n\tif m != nil {\n\t\treturn m.Tx\n\t}\n\treturn nil\n}", "func (b *Bitcoind) GetTransaction(txid string) (transaction Transaction, err error) {\n\tr, err := b.client.call(\"gettransaction\", []interface{}{txid})\n\tif err = handleError(err, &r); err != nil {\n\t\treturn\n\t}\n\n\terr = json.Unmarshal(r.Result, &transaction)\n\treturn\n}", "func (s *transactionStore) Get(ctx context.Context, id configapi.TransactionID) (*configapi.Transaction, error) {\n\t// If the transaction is not already in the cache, get it from the underlying primitive.\n\tentry, err := s.transactions.Get(ctx, id)\n\tif err != nil {\n\t\treturn nil, errors.FromAtomix(err)\n\t}\n\ttransaction := entry.Value\n\ttransaction.Index = configapi.Index(entry.Index)\n\ttransaction.Version = uint64(entry.Version)\n\treturn transaction, nil\n}", "func (tbl RecordTable) Tx() sqlapi.SqlTx {\n\treturn tbl.db.(sqlapi.SqlTx)\n}", "func (f *FactoidTransaction) Get(ctx context.Context, c *Client) error {\n\t// TODO: Test this functionality\n\t// If the TransactionID is nil then we have nothing to query for.\n\tif f.TransactionID == nil {\n\t\treturn fmt.Errorf(\"txid is nil\")\n\t}\n\t// If the Transaction is already populated then there is nothing to do. If\n\t// the Hash is nil, we cannot populate it anyway.\n\tif f.IsPopulated() {\n\t\treturn nil\n\t}\n\n\tparams := struct {\n\t\tHash *Bytes32 `json:\"hash\"`\n\t}{Hash: f.TransactionID}\n\tvar result struct {\n\t\tData Bytes `json:\"data\"`\n\t}\n\tif err := c.FactomdRequest(ctx, \"raw-data\", params, &result); err != nil {\n\t\treturn err\n\t}\n\n\tif err := f.UnmarshalBinary(result.Data); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (t *TxAPI) Get(hash string) (*api.ResultTx, error) {\n\tresp, statusCode, err := t.c.call(\"tx_get\", hash)\n\tif err != nil {\n\t\treturn nil, makeReqErrFromCallErr(statusCode, err)\n\t}\n\n\tvar r api.ResultTx\n\tif err = util.DecodeMap(resp, &r); err != nil {\n\t\treturn nil, errors.ReqErr(500, ErrCodeDecodeFailed, \"\", err.Error())\n\t}\n\n\treturn &r, nil\n}", "func (db *goLevelDBTx) Get(key []byte) ([]byte, error) {\n\tres, err := db.tx.Get(key, nil)\n\tif err != nil {\n\t\tif err == errors.ErrNotFound {\n\t\t\treturn nil, ErrNotFoundInDb\n\t\t}\n\t\tllog.Error(\"tx Get\", \"error\", err)\n\t\treturn nil, err\n\t}\n\treturn res, nil\n}", "func (api *API) getTransactionByShortID(shortID string) (types.Transaction, error) {\n\tvar txShortID types.TransactionShortID\n\t_, err := fmt.Sscan(shortID, &txShortID)\n\tif err != nil {\n\t\treturn types.Transaction{}, err\n\t}\n\n\ttxn, found := api.cs.TransactionAtShortID(txShortID)\n\tif !found {\n\t\terr = errNotFound\n\t}\n\treturn txn, err\n}", "func (pool *TxPool) Get(hash common.Hash) *types.Transaction {\n\treturn pool.all.Get(hash)\n}", "func (s *Session) Transaction() *Transaction {\n\t// acquire lock\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\n\treturn s.txn\n}", "func (b Block) GetTransaction(txHash cipher.SHA256) (Transaction, bool) {\n\ttxns := b.Body.Transactions\n\tfor i := range txns {\n\t\tif txns[i].Hash() == txHash {\n\t\t\treturn txns[i], true\n\t\t}\n\t}\n\treturn Transaction{}, false\n}", "func (pgb *ChainDBRPC) GetRawTransaction(txid string) (*dcrjson.TxRawResult, error) {\n\ttxraw, err := rpcutils.GetTransactionVerboseByID(pgb.Client, txid)\n\tif err != nil {\n\t\tlog.Errorf(\"GetRawTransactionVerbose failed for: %s\", txid)\n\t\treturn nil, err\n\t}\n\treturn txraw, nil\n}", "func GetTransactionById(pId int) *DOMAIN.Transaction {\n\t// Project structure\n\ttransaction := DOMAIN.Transaction{}\n\t// Add in Transaction variable, the project where ID is the same that the param\n\tres := getTransactionCollection().Find(db.Cond{\"TransactionID\": pId})\n\n\t//project.ProjectType = GetTypesByProjectId(pId)\n\n\t// Close session when ends the method\n\tdefer session.Close()\n\terr := res.One(&transaction)\n\tif err != nil {\n\t\tlog.Error(err)\n\t\treturn nil\n\t}\n\n\treturn &transaction\n}", "func (client *Client) GetTransaction(txnID string) (*Response, error) {\n\tpath := \"/transaction\"\n\turi := fmt.Sprintf(\"%s%s/%s\", client.apiBaseURL, path, txnID)\n\n\treq, err := http.NewRequest(\"GET\", uri, bytes.NewBuffer([]byte(\"\")))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := client.performRequest(req, \"\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// Handle conversion of Response from an interface{} to Transaction for the user.\n\tvar txn Transaction\n\tif err := json.Unmarshal(resp.Response.([]byte), &txn); err != nil {\n\t\treturn nil, err\n\t}\n\tresp.Response = txn\n\treturn resp, err\n}", "func (db database) Tx(txHash id.Hash) (tx.Tx, error) {\n\tscript := \"SELECT hash, selector, txid, txindex, amount, payload, phash, to_address, nonce, nhash, gpubkey, ghash, version FROM txs WHERE hash = $1\"\n\trow := db.db.QueryRow(script, txHash.String())\n\terr := row.Err()\n\tif err != nil {\n\t\treturn tx.Tx{}, err\n\t}\n\treturn rowToTx(row)\n}", "func (c *Client) GetTransaction(ctx context.Context, txhash string) (GetTransactionResponse, error) {\n\tres, err := c.RpcClient.GetTransactionWithConfig(\n\t\tctx,\n\t\ttxhash,\n\t\trpc.GetTransactionConfig{\n\t\t\tEncoding: rpc.GetTransactionConfigEncodingBase64,\n\t\t},\n\t)\n\terr = checkRpcResult(res.GeneralResponse, err)\n\tif err != nil {\n\t\treturn GetTransactionResponse{}, err\n\t}\n\treturn getTransaction(res)\n}", "func (tbl AssociationTable) Tx() sqlapi.SqlTx {\n\treturn tbl.db.(sqlapi.SqlTx)\n}", "func GetTransaction(txBytes []byte) (*peer.Transaction, error) {\n\ttx := &peer.Transaction{}\n\terr := proto.Unmarshal(txBytes, tx)\n\treturn tx, err\n}", "func (r *OrdersRepository) Tx() (tx *dbr.Tx, err error) {\n\tdb := r.PG.PostgresTrade()\n\n\ttx, err = db.Begin()\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"event\": \"error when begin transaction in postgres\",\n\t\t}).Error(err)\n\t}\n\n\treturn\n}", "func (db *Database) GetAttachedTx() *Transaction {\n\tdb.attachedTxMu.Lock()\n\tdefer db.attachedTxMu.Unlock()\n\n\treturn db.attachedTransaction\n}", "func RetrieveTransaction(ctx context.Context, db *mongo.Collection, _id string) (*Transaction, error) {\n\n\tvar transaction Transaction\n\n\tid, err := primitive.ObjectIDFromHex(_id)\n\tif err != nil {\n\t\treturn nil, apierror.ErrInvalidID\n\t}\n\n\tif err := db.FindOne(ctx, bson.M{\"_id\": id}).Decode(&transaction); err != nil {\n\t\treturn nil, apierror.ErrNotFound\n\t}\n\n\t// fmt.Println(\"&transaction.FinancialAccountID\", &transaction.FinancialAccountID)\n\t// fmt.Printf(\"***************\\n&transaction.FinancialAccountID Type : %T\\n\", &transaction.FinancialAccountID)\n\n\treturn &transaction, nil\n}", "func GetTransaction(txBytes []byte) (*peer.Transaction, error) {\n\ttx := &peer.Transaction{}\n\terr := proto.Unmarshal(txBytes, tx)\n\treturn tx, errors.Wrap(err, \"error unmarshaling Transaction\")\n}", "func (conn Connection) Transaction() (datastore.Transact, error) {\n\tif conn.Config.DbType == config.PGDriver {\n\t\treturn conn.SQLDriver.Transaction()\n\t}\n\treturn conn.Controller.Transaction()\n}", "func (r *BTCRPC) GetTransactionDetail(txhash string) ([]byte, error) {\n\tvar (\n\t\ttx []byte\n\t\terr error\n\t)\n\n\terr = r.Client.Call(\"getrawtransaction\", jsonrpc.Params{txhash, 1}, &tx)\n\treturn tx, err\n}", "func (l *LBucket) Tx() Tx {\n\treturn l.tx\n}", "func (t *Transaction) LastTx() string {\n\treturn t.lastTx\n}", "func (c Client) findTransaction(transactionID string) (*Transaction, error) {\n\tpath := fmt.Sprintf(\"/transactions/%s\", transactionID)\n\treq, err := http.NewRequest(\"GET\", c.getURL(path), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar marshalled map[string]*Transaction\n\treturn marshalled[\"transaction\"], c.executeRequestAndMarshal(req, &marshalled)\n}", "func (k Keeper) GetExtTransaction(ctx sdk.Context, id uint64) types.ExtTransaction {\n\tstore := prefix.NewStore(ctx.KVStore(k.storeKey), types.KeyPrefix(types.ExtTransactionKey))\n\tvar extTransaction types.ExtTransaction\n\tk.cdc.MustUnmarshalBinaryBare(store.Get(GetExtTransactionIDBytes(id)), &extTransaction)\n\treturn extTransaction\n}", "func (t *txLookup) Get(hash common.Hash) *types.Transaction {\n\tt.lock.RLock()\n\tdefer t.lock.RUnlock()\n\n\treturn t.all[hash]\n}", "func (a *transactionUsecase) GetByID(c context.Context, id int64) (*models.Transaction, error) {\n\n\tctx, cancel := context.WithTimeout(c, a.contextTimeout)\n\tdefer cancel()\n\n\tres, err := a.transactionRepo.GetByID(ctx, id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn res, nil\n}", "func GetBytesTransaction(tx *peer.Transaction) ([]byte, error) {\n\tbytes, err := proto.Marshal(tx)\n\treturn bytes, err\n}", "func (b Blockstream) GetTransaction(txHash string) (*wire.MsgTx, error) {\n\turl := fmt.Sprintf(\"%s/tx/%s\", baseURL, txHash)\n\tresp, err := http.Get(url)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\tif resp.StatusCode != http.StatusOK {\n\t\tb, err := ioutil.ReadAll(resp.Body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, fmt.Errorf(\"failed to get a transaction: %s\", b)\n\t}\n\n\tvar tx transaction\n\tif err := json.NewDecoder(resp.Body).Decode(&tx); err != nil {\n\t\treturn nil, err\n\t}\n\n\tmsgTx := wire.NewMsgTx(tx.Version)\n\tmsgTx.LockTime = uint32(tx.Locktime)\n\n\tfor _, vin := range tx.Vin {\n\t\tvoutHash, err := chainhash.NewHashFromStr(vin.Txid)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tsigScript, err := hex.DecodeString(vin.Scriptsig)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tvar witness [][]byte\n\t\tfor _, w := range vin.Witness {\n\t\t\tws, err := hex.DecodeString(w)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\twitness = append(witness, ws)\n\t\t}\n\n\t\tnewInput := wire.NewTxIn(\n\t\t\twire.NewOutPoint(voutHash, vin.Vout),\n\t\t\tsigScript,\n\t\t\twitness,\n\t\t)\n\t\tnewInput.Sequence = uint32(vin.Sequence)\n\n\t\tmsgTx.AddTxIn(newInput)\n\t}\n\n\tfor _, vout := range tx.Vout {\n\t\tpkScript, err := hex.DecodeString(vout.Scriptpubkey)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tmsgTx.AddTxOut(\n\t\t\twire.NewTxOut(\n\t\t\t\tvout.Value,\n\t\t\t\tpkScript,\n\t\t\t),\n\t\t)\n\t}\n\n\tif msgTx.TxHash().String() != tx.Txid {\n\t\treturn nil, fmt.Errorf(\"transaction hash doesn't match\")\n\t}\n\n\treturn msgTx, nil\n}", "func GetTransactionHandler(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\t// retrieve the parameters\n\tparam := make(map[string]uint64)\n\tfor _, key := range []string{\"blockId\", \"txId\"} {\n\t\tparam[key], _ = strconv.ParseUint(vars[\"blockId\"], 10, 64)\n\t}\n\n\ttmp := atomic.LoadUint64(&lastBlock)\n\tif param[\"blockId\"] > tmp {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\terr := fmt.Errorf(\"requested id %d latest %d\", param[\"blockId\"], lastBlock)\n\t\tlog.Println(err.Error())\n\t\t_, _ = w.Write([]byte(err.Error()))\n\t\treturn\n\t}\n\t// retuning anything in the body regardless of any error code\n\t// it may contain\n\t_, _, body, _ := dataCollection.GetTransaction(param[\"blockId\"], param[\"txId\"], config.DefaultRequestsTimeout)\n\twriteResponse(body, &w)\n}", "func GetTransactionById(transaction_id string) ([]byte,error) {\n\tparams := url.Values{}\n\tUrl, err := url.Parse(\"http://\"+bigchaindb_addr+\"/api/v1/transactions/\"+transaction_id)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\t//如果参数中有中文参数,这个方法会进行URLEncode\n\tUrl.RawQuery = params.Encode()\n\turlPath := Url.String()\n\tresp, err := http.Get(urlPath)\n\tdefer resp.Body.Close()\n\ts, err := ioutil.ReadAll(resp.Body)\n\tif err != nil{\n\t\tfmt.Println(\"read resp.Body \",err)\n\t\treturn nil,err\n\t}\n\treturn s, nil\n}", "func (m *MockAtomicLogic) GetDBTx() types.Tx {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"GetDBTx\")\n\tret0, _ := ret[0].(types.Tx)\n\treturn ret0\n}", "func (r Virtual_Guest_Block_Device_Template_Group) GetTransaction() (resp datatypes.Provisioning_Version1_Transaction, err error) {\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest_Block_Device_Template_Group\", \"getTransaction\", nil, &r.Options, &resp)\n\treturn\n}", "func (ch *blockchain) GetTx(h chainhash.Hash) (tx *primitives.Tx, err error) {\n\tloc, err := ch.txidx.GetTx(h)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tblock, err := ch.GetBlock(loc.Block)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn block.Txs[loc.Index], nil\n}", "func (b *BlockQueue) GetTx(address string) []*Tx {\n\t\n\ttransactionQ, ok := b.addrTx[address]\n\tif !ok {\n\t\treturn nil\n\t}\n\t\n\ttransactions := make([]*Tx, 0, transactionQ.Len())\n\titer := transactionQ.Iter()\n\tfor tx, finished := iter.Next(); !finished; tx, finished = iter.Next() {\n\t\ttransactions = append(transactions, tx.(* Tx))\n\t}\n\n\treturn transactions\n}", "func GetToProcessTx(db *gorp.DbMap, confirmedBlkNumber uint64) (records []*TxRecord, err error) {\n\t_, err = db.Select(&records, `SELECT * FROM \"record\" WHERE \"eth_block_number\" <= ? AND \"state\" = ?`,\n\t\tconfirmedBlkNumber, ExchangeStateDetected)\n\treturn\n}", "func (s *Session) TransactionTx(f func(*Session) (interface{}, error), opts *sql.TxOptions) (interface{}, error) {\n\terr := s.BeginTx(opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\td, err := f(s)\n\tif err != nil {\n\t\ts.RollBack()\n\t} else {\n\t\ts.Commit()\n\t}\n\treturn d, err\n}", "func (s *inMemSpannerServer) getTransactionID(session *spannerpb.Session, txSelector *spannerpb.TransactionSelector) []byte {\n\tvar res []byte\n\tif txSelector.GetBegin() != nil {\n\t\t// Start a new transaction.\n\t\tres = s.beginTransaction(session, txSelector.GetBegin()).Id\n\t} else if txSelector.GetId() != nil {\n\t\tres = txSelector.GetId()\n\t}\n\treturn res\n}", "func (db *stateDB) ReadTxn() ReadTransaction {\n\treturn &transaction{db: nil, txn: db.memDB.Txn(false)}\n}", "func (client *clientImpl) GetTransactionByHash(tx string) (val *Transaction, err error) {\n\n\terr = client.Call2(\"eth_getTransactionByHash\", &val, tx)\n\n\treturn\n}", "func (bdm *MySQLDBManager) GetTransactionsObject() (TranactionsInterface, error) {\n\tconn, err := bdm.getConnection()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttxs := Tranactions{}\n\ttxs.DB = &MySQLDB{conn, bdm.Config.TablesPrefix, bdm.Logger}\n\n\treturn &txs, nil\n}", "func (r Virtual_Guest) GetLastTransaction() (resp datatypes.Provisioning_Version1_Transaction, err error) {\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest\", \"getLastTransaction\", nil, &r.Options, &resp)\n\treturn\n}", "func (dao *InfoDao) Transaction(ctx context.Context, f func(ctx context.Context, tx *gdb.TX) error) (err error) {\n\treturn dao.Ctx(ctx).Transaction(ctx, f)\n}", "func (r *RBucket) Tx() Tx {\n\treturn r.tx\n}", "func (d *Driver) Tx(ctx context.Context) (dialect.Tx, error) {\n\treturn d.BeginTx(ctx, nil)\n}", "func (m *TransactionMessage) GetTransaction() (*types.Tx, error) {\n\ttx := &types.Tx{}\n\tif err := tx.UnmarshalText(m.RawTx); err != nil {\n\t\treturn nil, err\n\t}\n\treturn tx, nil\n}", "func GetTxType(tx *types.Transaction, statedb *state.StateDB) types.TxType {\n\tif tx.Type != types.UnknownTx {\n\t\treturn tx.Type\n\t}\n\ttxType, _ := ParseTxType(tx, statedb)\n\treturn txType\n}" ]
[ "0.7570912", "0.7563583", "0.75429225", "0.7470973", "0.74389464", "0.74146205", "0.73161745", "0.7297836", "0.72537345", "0.72484773", "0.7109892", "0.70955324", "0.70955324", "0.70955324", "0.7064121", "0.70506865", "0.6991761", "0.69659305", "0.695677", "0.69502926", "0.69251525", "0.6883312", "0.6882305", "0.68456906", "0.68022215", "0.6793521", "0.6790846", "0.67673075", "0.674065", "0.674065", "0.67366207", "0.6715067", "0.6680047", "0.6667371", "0.6661904", "0.6655149", "0.6592765", "0.6579128", "0.65568125", "0.6541197", "0.6532593", "0.6516974", "0.6516057", "0.65095335", "0.6508026", "0.64929897", "0.6489438", "0.6472942", "0.6471681", "0.64067596", "0.64058346", "0.6382142", "0.6378356", "0.63768625", "0.63734215", "0.63672", "0.63639694", "0.6360252", "0.6356039", "0.6335962", "0.63343793", "0.62951833", "0.6286052", "0.6278786", "0.6276492", "0.62586975", "0.6247625", "0.6242138", "0.62388253", "0.6204472", "0.617306", "0.61718106", "0.6167312", "0.61650646", "0.61647516", "0.6157337", "0.61454105", "0.6144023", "0.6137329", "0.61295724", "0.60981435", "0.6095513", "0.60920304", "0.6079369", "0.6075042", "0.6056288", "0.604868", "0.60460734", "0.6037386", "0.60220885", "0.6016744", "0.6014176", "0.60009176", "0.59806097", "0.59716463", "0.5942189", "0.59413004", "0.591254", "0.590971", "0.59065354" ]
0.7353372
6
mark text as Output type
func (*text) isOutput() { }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (Output) typ() string { return \"output1\" }", "func (self *Output) Type() string {\n\treturn \"output\"\n}", "func (t *typewriter) Typewrite() string {\n\t// Re-use the cached result if already processed.\n\tif t.cur != -1 {\n\t\treturn t.result\n\t}\n\n\tvar buf bytes.Buffer\n\n\tfor {\n\t\tsep, str := t.scanMorpheme()\n\t\tif str == \"\" {\n\t\t\tbreak\n\t\t}\n\n\t\tbuf.WriteString(sep)\n\t\tbuf.WriteString(str)\n\t}\n\n\tt.result = buf.String()\n\treturn t.result\n}", "func (ot OutputType) MarshalText() ([]byte, error) {\n\treturn []byte(ot.String()), nil\n}", "func (cb *printcb) outputText(data string) error {\n\treturn cb.outputBinary([]byte(data))\n}", "func (t *Text) Output() *Text {\n\tfmt.Println(output(t))\n\tt.output = \"\"\n\tt.outputColor = \"\"\n\tt.pipeline = TextPipleline{}\n\treturn t\n}", "func (f *FakeOutput) Type() string { return \"fake_output\" }", "func TestFieldOutputText(t *testing.T) {\n\tfield := NewField()\n\tfield.Name = \"foo\"\n\tfield.Type = \"text\"\n\n\ttag := field.output()\n\n\tassert.Equal(t, \"<input type=\\\"text\\\" name=\\\"foo\\\" id=\\\"foo\\\" value=\\\"\\\" />\", tag)\n}", "func textFmt(w io.Writer, x interface{}, format string) {\n\twriteAny(w, x, false)\n}", "func (t *Text) Text(str string) *Text {\n\tt.output = str\n\tt.textLines++\n\tt.lines[t.textLines] = str\n\treturn t\n}", "func (ot *OutputType) UnmarshalText(text []byte) error {\n\tswitch strings.ToLower(string(text)) {\n\tcase \"file\":\n\t\t*ot = OutputFile\n\tcase \"stderr\":\n\t\t*ot = OutputStdErr\n\tcase \"both\":\n\t\t*ot = OutputBoth\n\t}\n\n\treturn nil\n}", "func (this *Tidy) OutputEncoding(val int) (bool, error) {\n\tswitch val {\n\tcase Raw, Ascii, Latin0, Latin1, Utf8, Iso2022, Mac, Win1252, Ibm858, Utf16le, Utf16be, Utf16, Big5, Shiftjis:\n\t\treturn this.optSetInt(C.TidyOutCharEncoding, (C.ulong)(val))\n\t}\n\treturn false, errors.New(\"Argument val int is out of range (0-13)\")\n}", "func (o RuleTargetOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v RuleTarget) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o GetSmartagFlowLogsLogOutput) OutputType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetSmartagFlowLogsLog) string { return v.OutputType }).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementRegexMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementRegexMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o AnalyzerOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Analyzer) pulumi.StringPtrOutput { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementXssMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementXssMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementNotStatementStatementRegexMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementNotStatementStatementRegexMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (ot OutputType) String() string {\n\tswitch ot {\n\tcase OutputFile:\n\t\treturn \"File\"\n\tcase OutputStdErr:\n\t\treturn \"StdErr\"\n\tcase OutputBoth:\n\t\treturn \"Both\"\n\tdefault:\n\t\treturn \"Unknown\"\n\t}\n}", "func (o WebAclRuleStatementAndStatementStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementByteMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementByteMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o AnalyzerIdentityOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AnalyzerIdentity) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementNotStatementStatementXssMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementNotStatementStatementXssMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func IsText(t Type) bool {\n\treturn int(t)&flagIsText == flagIsText\n}", "func (r renderer) NormalText(out *bytes.Buffer, text []byte) {\n\tout.Write(text)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementRegexMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementRegexMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementSqliMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementSqliMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementOrStatementStatementAndStatementStatementAndStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementOrStatementStatementAndStatementStatementAndStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementOrStatementStatementAndStatementStatementAndStatementStatementByteMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementOrStatementStatementAndStatementStatementAndStatementStatementByteMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementOrStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementOrStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o SourceCodeTokenOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *SourceCodeToken) pulumi.StringOutput { return v.Type }).(pulumi.StringOutput)\n}", "func (*SpeechRecognitionResultText) TypeName() string {\n\treturn \"speechRecognitionResultText\"\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementByteMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementByteMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementOrStatementStatementAndStatementStatementAndStatementStatementRegexMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementOrStatementStatementAndStatementStatementAndStatementStatementRegexMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o DocumentDbOutputDataSourceOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v DocumentDbOutputDataSource) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementNotStatementStatementSqliMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementNotStatementStatementSqliMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementOrStatementStatementRegexMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementOrStatementStatementRegexMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementRegexMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementRegexMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func output(data []byte, encodeType Encode) (string, error) {\n\tswitch encodeType {\n\tcase HEX:\n\t\treturn hex.EncodeToString(data), nil\n\tcase Base64:\n\t\treturn base64.StdEncoding.EncodeToString(data), nil\n\tcase None:\n\t\treturn string(data), nil\n\tdefault:\n\t\treturn \"\", fmt.Errorf(\"secretInfo OutputType unsupport\")\n\t}\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementXssMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementXssMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementOrStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementOrStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementXssMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementXssMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementNotStatementStatementRegexMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementNotStatementStatementRegexMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementOrStatementStatementByteMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementOrStatementStatementByteMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o MfaPingidOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *MfaPingid) pulumi.StringOutput { return v.Type }).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementSqliMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementSqliMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func genericTextOutput(doppelgangerState *userInfo, theMessage messageFromChatChannelToDoppelganger) bool {\n\tvar err error\n\tif theMessage.originator == doppelgangerState.userID {\n\t\t// Our own message -- don't backspace out.\n\t\terr = backspaceOut(doppelgangerState, doppelgangerState.promptLen+doppelgangerState.cursorColumn)\n\t\t_, err = oi.LongWrite(doppelgangerState.writer, []byte(theMessage.parameter+\"\\r\\n\"))\n\t} else {\n\t\t//\n\t\t// Backspace out before outputting message\n\t\t//\n\t\terr = backspaceOut(doppelgangerState, doppelgangerState.promptLen+doppelgangerState.cursorColumn)\n\t\tif err != nil {\n\t\t\t//\n\t\t\t// We are assuming if we got an error, the network connection is\n\t\t\t// closed, and we need to exit the doppelganger because we are\n\t\t\t// done, too.\n\t\t\t//\n\t\t\treturn true\n\t\t}\n\t\t_, err = oi.LongWrite(doppelgangerState.writer, []byte(theMessage.parameter+\"\\r\\n\"))\n\t}\n\tif err != nil {\n\t\t//\n\t\t// We are assuming if we got an error, the network connection is closed,\n\t\t// and we need to exit the doppelganger because we are done, too.\n\t\t//\n\t\treturn true\n\t}\n\treturn false\n}", "func (*HTML) isOutput() {\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementOrStatementStatementRegexMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementOrStatementStatementRegexMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementRegexMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementRegexMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementNotStatementStatementOrStatementStatementXssMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementNotStatementStatementOrStatementStatementXssMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementByteMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementByteMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (p *Parser) SetOutput(out io.Writer) { p.out = out }", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementSqliMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementSqliMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o ReleaseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Release) pulumi.StringOutput { return v.Type }).(pulumi.StringOutput)\n}", "func (c *Cmd) Output() ([]byte, error)", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementOrStatementStatementSqliMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementOrStatementStatementSqliMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementNotStatementStatementSizeConstraintStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementNotStatementStatementSizeConstraintStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementNotStatementStatementXssMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementNotStatementStatementXssMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementOrStatementStatementByteMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementOrStatementStatementByteMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementXssMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementXssMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o EntryOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Entry) pulumi.StringPtrOutput { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementOrStatementStatementXssMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementOrStatementStatementXssMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o BlobOutputDataSourceOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BlobOutputDataSource) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementNotStatementStatementOrStatementStatementSqliMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementNotStatementStatementOrStatementStatementSqliMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o TransformationOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Transformation) pulumi.StringPtrOutput { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (o MetricTargetOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v MetricTarget) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementOrStatementStatementSqliMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementOrStatementStatementSqliMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementNotStatementStatementSqliMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementNotStatementStatementSqliMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementSizeConstraintStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementSizeConstraintStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o TopicRuleElasticsearchOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v TopicRuleElasticsearch) string { return v.Type }).(pulumi.StringOutput)\n}", "func (wm Watermark) isText() bool {\n\treturn wm.Mode == WMText\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementSizeConstraintStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementSizeConstraintStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o FlowOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Flow) pulumi.StringOutput { return v.Type }).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementOrStatementStatementAndStatementStatementAndStatementStatementSizeConstraintStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementOrStatementStatementAndStatementStatementAndStatementStatementSizeConstraintStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementSqliMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementSqliMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (*SpeechRecognitionResultText) TypeID() uint32 {\n\treturn SpeechRecognitionResultTextTypeID\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementOrStatementStatementSizeConstraintStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementOrStatementStatementSizeConstraintStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementSizeConstraintStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementSizeConstraintStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o RawOutputDatasourceOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v RawOutputDatasource) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o TriggerOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Trigger) pulumi.StringOutput { return v.Type }).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementNotStatementStatementSizeConstraintStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementNotStatementStatementSizeConstraintStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (dt *FieldTraits) OutputOnly(p Path) { dt.add(p, FieldTypeOutputOnly) }", "func (o TopicRuleErrorActionElasticsearchOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v TopicRuleErrorActionElasticsearch) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o DocumentDbOutputDataSourceResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v DocumentDbOutputDataSourceResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (*backfiller) OutputTypes() []*types.T {\n\t// No output types.\n\treturn nil\n}", "func (o PowerBIOutputDataSourceOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v PowerBIOutputDataSource) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementOrStatementStatementSizeConstraintStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementOrStatementStatementSizeConstraintStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o RecordSetOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *RecordSet) pulumi.StringOutput { return v.Type }).(pulumi.StringOutput)\n}", "func Output(c Content, o Options) (string, error) {\n\tvar e error\n\tvar formatter format\n\tswitch style := o.Style; style {\n\tcase \"md\":\n\t\tformatter = md{Content: c, Options: o}\n\tcase \"psql\":\n\t\tformatter = psql{Content: c, Options: o}\n\tcase \"jira\":\n\t\tformatter = jira{Content: c, Options: o}\n\tdefault:\n\t\te = errors.New(\"Formatter not implemented error\")\n\t}\n\ttable := formatter.format()\n\treturn table, e\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementSizeConstraintStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementSizeConstraintStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o BlobOutputDataSourceResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BlobOutputDataSourceResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o FieldOutput) Type() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v Field) *string { return v.Type }).(pulumi.StringPtrOutput)\n}", "func (o HostingReleaseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *HostingRelease) pulumi.StringOutput { return v.Type }).(pulumi.StringOutput)\n}", "func (o CsvSerializationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v CsvSerialization) string { return v.Type }).(pulumi.StringOutput)\n}" ]
[ "0.6845131", "0.6313934", "0.62364113", "0.5994303", "0.59707606", "0.5910661", "0.58733034", "0.5823576", "0.5814002", "0.5777523", "0.5733294", "0.5717191", "0.57098264", "0.567048", "0.5668985", "0.56583005", "0.559444", "0.55912685", "0.55904263", "0.5571912", "0.55579597", "0.55523145", "0.55334127", "0.5528834", "0.55210745", "0.55184937", "0.5509682", "0.55082417", "0.55058974", "0.5494632", "0.5485169", "0.54814565", "0.54790896", "0.5476313", "0.54727453", "0.54651254", "0.5457429", "0.5451732", "0.54490453", "0.54475904", "0.54459745", "0.5445719", "0.54431313", "0.54409283", "0.5434003", "0.54327863", "0.54313976", "0.54294896", "0.5423597", "0.5419033", "0.5412907", "0.540858", "0.5398548", "0.53979135", "0.53962094", "0.5394872", "0.5390374", "0.5385729", "0.53824466", "0.5378884", "0.5377182", "0.5374802", "0.5364207", "0.53583956", "0.53573483", "0.53510815", "0.5348263", "0.53443056", "0.5343996", "0.534033", "0.53355414", "0.532987", "0.53250724", "0.5324404", "0.53214234", "0.53132594", "0.5307143", "0.5301415", "0.5294021", "0.5293511", "0.5281606", "0.52709794", "0.52692235", "0.5264753", "0.5261764", "0.52607286", "0.5240339", "0.5229542", "0.5228262", "0.5227985", "0.5227379", "0.5226735", "0.52204907", "0.52160877", "0.5202687", "0.51829904", "0.51700306", "0.5157393", "0.51524174", "0.5146496" ]
0.6452956
1
Class or classes to add
func (h *HTML) Class(c ...string) *HTML { h.classes = append(h.classes, c...) return h }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (ns Nodes) AddClass(a string) {\n}", "func (this *Intro) addClass(c *Class) {\n\tthis.classes = append(this.classes, c)\n}", "func (c FakeClassList) Add(classes ...string) {\n\tfor _, cl := range classes {\n\t\tc[cl] = struct{}{}\n\t}\n}", "func (c *Classifier) AddClasses(classes []string) {\n\terr := c.cFrequency.AddClasses(classes)\n\n\tif nil != err {\n\t\tpanic(err)\n\t}\n}", "func (t *TrainingSet) AddClass(ctx context.Context, class string) (Class, error) {\n\t_, err := tspb.NewTrainingSetClient(t.c.ClientConn).AddClass(t.newContext(ctx), &tspb.AddClassRequest{\n\t\tName: t.name,\n\t\tClass: class,\n\t})\n\tif err != nil {\n\t\treturn Class{}, err\n\t}\n\treturn Class{name: class}, nil\n}", "func addclass(cd *classdto) (int, error) {\n\n\tcdb.mu.Lock()\n\tdefer cdb.mu.Unlock()\n\tID := -1\n\tif class, err := findClassByNameandDate(cd); err != nil {\n\t\tif ID = getMaxID(); ID == -1 {\n\t\t\tcdb.classMap[1] = Class{ClassID: 1, Name: cd.Name, StartDate: cd.StartDate, EndDate: cd.EndDate}\n\t\t\treturn 1, nil\n\t\t}\n\n\t\tcdb.classMap[ID+1] = Class{ClassID: ID + 1, Name: cd.Name, StartDate: cd.StartDate, EndDate: cd.EndDate}\n\t\treturn ID + 1, nil\n\t} else {\n\t\tfmt.Printf(\"class already exists for id %v\\n\", class.ClassID)\n\t\treturn class.ClassID, fmt.Errorf(\"class already exists for id %v\", class.ClassID)\n\t}\n\n}", "func (e *HTMLApplet) Class(v string) *HTMLApplet {\n\te.a[\"class\"] = v\n\treturn e\n}", "func (m *MysqlDriver) AddClass(intactClass *transform.IntactClass) error {\n\tsession := m.conn.NewSession()\n\tdefer session.Close()\n\n\terr := session.Begin()\n\t_, err = session.Insert(intactClass.Class)\n\tif err != nil {\n\t\tsession.Rollback()\n\t\treturn err\n\t}\n\t// set class id\n\tvar announcements []*model.Announcement\n\tfor _, announcement := range intactClass.Announcements {\n\t\tannouncement.ClassId = intactClass.Class.Id\n\t\tannouncements = append(announcements, &announcement.Announcement)\n\t}\n\t_, err = session.Insert(announcements)\n\tif err != nil {\n\t\tsession.Rollback()\n\t\treturn err\n\t}\n\n\treturn session.Commit()\n}", "func (r *EducationSchoolClassesCollectionRequest) Add(ctx context.Context, reqObj *EducationClass) (resObj *EducationClass, err error) {\n\terr = r.JSONRequest(ctx, \"POST\", \"\", reqObj, &resObj)\n\treturn\n}", "func (e *Ex) Class() string {\n\treturn e.class\n}", "func (e *Ex) Class() string {\n\treturn e.class\n}", "func (c FakeClassList) Replace(oldClass, newClass string) {\n\tc.Remove(oldClass)\n\tc.Add(newClass)\n}", "func (a *Attributes) AddClassChanged(v string) bool {\n\tif v == \"\" {\n\t\treturn false // nothing to add\n\t}\n\tif a.Has(\"class\") {\n\t\tnewClass, changed := AddClass(a.Get(\"class\"), v)\n\t\tif changed {\n\t\t\ta.StringSliceMap.Set(\"class\", newClass)\n\t\t}\n\t\treturn changed\n\t} else {\n\t\ta.StringSliceMap.Set(\"class\", v)\n\t\treturn true\n\t}\n}", "func (app *App) createClass(w http.ResponseWriter, r *http.Request) {\n\tclassPayload := decodeClass(r)\n\tmongoResult := app.getById(\"classes\", []byte(classPayload.ClassId))\n\tif mongoResult.Err() != nil {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.WriteHeader(http.StatusOK)\n\t\tresult := app.insertOne(\"classes\", classPayload)\n\t\tjson.NewEncoder(w).Encode(&result)\n\t} else {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tencodeError(w, \"A Class With That ID Already Exists.\")\n\t}\n}", "func (m *MysqlDriver) EnterClass(userID, classID int64) error {\n\tclass, err := m.GetClassByID(classID)\n\tif err != nil {\n\t\tlogger.Sugar.Errorf(\"failed to enter class (get class fail) : %v\", err)\n\t\treturn err\n\t}\n\n\tisChecked := 1\n\tif class.Class.IsCheck == 1 {\n\t\tisChecked = 0\n\t}\n\n\tmember := &model.UserClass{\n\t\tUserId: userID,\n\t\tClassId: classID,\n\t\tIsChecked: isChecked,\n\t}\n\n\ti, err := m.conn.Insert(member)\n\tif err != nil {\n\t\tlogger.Sugar.Errorf(\"failed to insert : %v\", err)\n\t\treturn err\n\t}\n\tif i == 0 {\n\t\tlogger.Sugar.Error(ErrNoRowsAffected.Error())\n\t\treturn ErrNoRowsAffected\n\t}\n\treturn nil\n}", "func (element *Element) Class(values ...string) *Element {\n\treturn element.Attr(\"class\", strings.Join(values, \" \"))\n}", "func (c Class) String() string {\n\tswitch c {\n\tcase Other:\n\t\treturn \"Other\"\n\tcase sot:\n\t\treturn \"sot\"\n\tcase eot:\n\t\treturn \"eot\"\n\tdefault:\n\t\treturn \"Class(\" + strconv.Itoa(int(c)) + \")\"\n\tcase AIClass:\n\t\treturn \"AIClass\"\n\tcase ALClass:\n\t\treturn \"ALClass\"\n\tcase B2Class:\n\t\treturn \"B2Class\"\n\tcase BAClass:\n\t\treturn \"BAClass\"\n\tcase BBClass:\n\t\treturn \"BBClass\"\n\tcase BKClass:\n\t\treturn \"BKClass\"\n\tcase CBClass:\n\t\treturn \"CBClass\"\n\tcase CJClass:\n\t\treturn \"CJClass\"\n\tcase CLClass:\n\t\treturn \"CLClass\"\n\tcase CMClass:\n\t\treturn \"CMClass\"\n\tcase CPClass:\n\t\treturn \"CPClass\"\n\tcase CRClass:\n\t\treturn \"CRClass\"\n\tcase EBClass:\n\t\treturn \"EBClass\"\n\tcase EMClass:\n\t\treturn \"EMClass\"\n\tcase EXClass:\n\t\treturn \"EXClass\"\n\tcase GLClass:\n\t\treturn \"GLClass\"\n\tcase H2Class:\n\t\treturn \"H2Class\"\n\tcase H3Class:\n\t\treturn \"H3Class\"\n\tcase HLClass:\n\t\treturn \"HLClass\"\n\tcase HYClass:\n\t\treturn \"HYClass\"\n\tcase IDClass:\n\t\treturn \"IDClass\"\n\tcase INClass:\n\t\treturn \"INClass\"\n\tcase ISClass:\n\t\treturn \"ISClass\"\n\tcase JLClass:\n\t\treturn \"JLClass\"\n\tcase JTClass:\n\t\treturn \"JTClass\"\n\tcase JVClass:\n\t\treturn \"JVClass\"\n\tcase LFClass:\n\t\treturn \"LFClass\"\n\tcase NLClass:\n\t\treturn \"NLClass\"\n\tcase NSClass:\n\t\treturn \"NSClass\"\n\tcase NUClass:\n\t\treturn \"NUClass\"\n\tcase OPClass:\n\t\treturn \"OPClass\"\n\tcase POClass:\n\t\treturn \"POClass\"\n\tcase PRClass:\n\t\treturn \"PRClass\"\n\tcase QUClass:\n\t\treturn \"QUClass\"\n\tcase RIClass:\n\t\treturn \"RIClass\"\n\tcase SAClass:\n\t\treturn \"SAClass\"\n\tcase SGClass:\n\t\treturn \"SGClass\"\n\tcase SPClass:\n\t\treturn \"SPClass\"\n\tcase SYClass:\n\t\treturn \"SYClass\"\n\tcase WJClass:\n\t\treturn \"WJClass\"\n\tcase XXClass:\n\t\treturn \"XXClass\"\n\tcase ZWClass:\n\t\treturn \"ZWClass\"\n\tcase ZWJClass:\n\t\treturn \"ZWJClass\"\n\t}\n}", "func isClass(ln []byte) bool { return (ln[0] == 'C') }", "func (this *ObjectAdd) Type() value.Type { return value.OBJECT }", "func (p *class) createClass(db *sql.DB) error {\n\treturn errors.New(\"createClass method not implemented\")\n}", "func (cis classInterfaces) add(log logging.Logger, iface *HostFabricInterface) {\n\tnn := int(iface.NumaNode)\n\tif _, exists := cis[iface.NetDevClass]; !exists {\n\t\tcis[iface.NetDevClass] = make(numaNetIfaceMap)\n\t}\n\tif _, exists := cis[iface.NetDevClass][nn]; exists {\n\t\treturn // already have interface for this NUMA\n\t}\n\tlog.Debugf(\"%s class iface %s found for NUMA %d\", iface.NetDevClass,\n\t\tiface.Device, nn)\n\tcis[iface.NetDevClass][nn] = iface\n}", "func (id ID) Class() Class {\n\treturn object_getClass(id)\n}", "func assignClass() string { //To start with, three classes.\n\tlottery := random(1, 3)\n\tvar class string\n\tswitch lottery {\n\tcase 1:\n\t\tclass = \"Commoner\"\n\tcase 2:\n\t\tclass = \"Fighter\"\n\tcase 3:\n\t\tclass = \"Rogue\"\n\t}\n\treturn class\n}", "func parseClass(node *node32) (*ClassType, error) {\n\tclass := &ClassType{}\n\n\tclass.SetToken(&node.token32)\n\n\tfor node := range nodeRange(node) {\n\t\tswitch node.pegRule {\n\t\tcase ruleCLASS:\n\t\tcase ruleIS:\n\t\tcase ruleSPACE:\n\t\tcase ruleEND:\n\t\tcase ruleIDENT:\n\t\t\tclass.name = node.match\n\t\tcase ruleMEMBERDEF:\n\t\t\tm, err := parseClassMembers(node.up)\n\t\t\tclass.members = append(class.members, m)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\tcase ruleFUNC:\n\t\t\tf, err := parseFunction(node.up)\n\t\t\tclass.methods = append(class.methods, f)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\tdefault:\n\t\t\treturn nil, fmt.Errorf(\n\t\t\t\t\"Unexpected %s %s\",\n\t\t\t\tnode.String(),\n\t\t\t\tnode.match,\n\t\t\t)\n\t\t}\n\t}\n\n\tclass, err := autoGenerateGetSet(class)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn class, nil\n}", "func (p *BaseProvider) Add(i interface{}) error {\n\tswitch v := i.(type) {\n\tcase Def:\n\t\treturn p.AddDef(v)\n\tcase *Def:\n\t\treturn p.AddDefPtr(v)\n\tcase []Def:\n\t\treturn p.AddDefSlice(v)\n\tcase []*Def:\n\t\treturn p.AddDefPtrSlice(v)\n\tcase func() Def:\n\t\treturn p.AddDef(v())\n\tcase func() *Def:\n\t\treturn p.AddDefPtr(v())\n\tcase func() []Def:\n\t\treturn p.AddDefSlice(v())\n\tcase func() []*Def:\n\t\treturn p.AddDefPtrSlice(v())\n\tdefault:\n\t\treturn errors.New(\"could not load definition with type \" + reflect.TypeOf(i).String() +\n\t\t\t\" (allowed types: dingo.Def, *dingo.Def, []dingo.Def, []*dingo.Def,\" +\n\t\t\t\" func() dingo.Def, func() *dingo.Def, func() []dingo.Def, func() []*dingo.Def)\")\n\t}\n}", "func newClassResult(typeString, class string, payload map[string]interface{}) map[string]interface{} {\n\tout := map[string]interface{}{\n\t\ttypeKey: typeString,\n\t\tclassKey: class,\n\t}\n\tif payload != nil {\n\t\tout[payloadKey] = payload\n\t}\n\treturn out\n}", "func paramToClassMap(classesMap map[string]phpClass, p Parameter) {\n\tif p.BaseType == phpObject && p.BaseType != p.Type {\n\t\tclassesMap[p.Type] = phpClass{\n\t\t\tName: p.Type,\n\t\t\tFields: p.Properties,\n\t\t}\n\t}\n}", "func findClassByNameandDate(cd *classdto) (Class, error) {\n\n\tflag := false\n\texistingClass := &Class{}\n\tif len(cdb.classMap) != 0 {\n\t\tfor _, class := range cdb.classMap {\n\t\t\tif class.Name == cd.Name && class.StartDate == cd.StartDate && class.EndDate == cd.EndDate {\n\t\t\t\tflag = true // a person can either update an existing class or add a new class with an new date\n\t\t\t\texistingClass = &class\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif flag {\n\t\t\treturn *existingClass, nil\n\t\t}\n\t}\n\treturn *existingClass, fmt.Errorf(\"Class not available\")\n}", "func (ns Nodes) HasClass(cl string) bool {\n\tif len(strings.Split(cl, \" \")) > 1 {\n\t\treturn false\n\t}\n\tclasses := getAttrs(ns, \"class\")\n\tfor _, v := range classes {\n\t\tm := mapFromSplit(v)\n\t\tif _, ok := m[cl]; ok {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func addKnownTypes(scheme *runtime.Scheme) error {\n\t// TODO this gets cleaned up when the types are fixed\n\tscheme.AddKnownTypes(SchemeGroupVersion,\n\t\t&Demo{},\n\t\t&DemoList{},\n\t)\n\treturn nil\n}", "func generateClasses(types map[string]raml.Type, dir string) error {\n\tfor k, t := range types {\n\t\tpc := newClassFromType(t, k)\n\t\tif err := pc.generate(dir); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (c *Client) CreateClass(request *CreateClassRequest) (response *CreateClassResponse, err error) {\n if request == nil {\n request = NewCreateClassRequest()\n }\n response = NewCreateClassResponse()\n err = c.Send(request, response)\n return\n}", "func RegisterClass(object Selector) (Class, error) {\n\tptr := reflect.TypeOf(object)\n\tstrct := ptr.Elem()\n\tif strct.NumField() == 0 || strct.Field(0).Type != reflect.TypeOf(Class(0)) {\n\t\treturn 0, fmt.Errorf(\"objc: need objc.Class as first field: %w\", MismatchError)\n\t}\n\tisa := strct.Field(0)\n\ttag := isa.Tag.Get(\"objc\")\n\tif tag == \"\" {\n\t\treturn 0, fmt.Errorf(\"objc: missing objc tag: %w\", TagFormatError)\n\t}\n\t// split contains the class name and super class name followed by all the Protocols\n\t// start with two for ClassName : SuperClassName\n\tvar split = make([]string, 2)\n\t{\n\t\t// This is a simple parser for the objc tag that looks for the format\n\t\t// \t\"ClassName : SuperClassName <Protocol, ...>\"\n\t\t// It appends to the split variable with the [ClassName, SuperClassName, Protocol, ...]\n\n\t\tvar i int // from tag[0:i] is whatever identifier is next\n\t\tvar r rune // r is the current rune\n\t\tskipSpace := func() {\n\t\t\tfor _, c := range tag {\n\t\t\t\tif c == ' ' {\n\t\t\t\t\ttag = tag[1:]\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tskipSpace()\n\t\t// get ClassName\n\t\tfor i, r = range tag {\n\t\t\tif r == ' ' || r == ':' {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tsplit[0] = tag[0:i] // store ClassName\n\t\ttag = tag[i:]\n\n\t\tskipSpace()\n\n\t\t// check for ':'\n\t\tif len(tag) > 0 && tag[0] != ':' {\n\t\t\treturn 0, fmt.Errorf(\"objc: missing ':': %w\", TagFormatError)\n\t\t}\n\t\ttag = tag[1:] // skip ':'\n\t\tskipSpace()\n\n\t\t// get SuperClassName\n\t\tfor i, r = range tag {\n\t\t\tif r == ' ' {\n\t\t\t\tbreak\n\t\t\t} else if i+1 == len(tag) {\n\t\t\t\t// if this is the last character in the string\n\t\t\t\t// make sure to increment i so that tag[:i]\n\t\t\t\t// includes the last character\n\t\t\t\ti++\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif len(tag) < i {\n\t\t\treturn 0, fmt.Errorf(\"objc: missing SuperClassName: %w\", TagFormatError)\n\t\t}\n\t\tsplit[1] = tag[:i] // store SuperClassName\n\t\ttag = tag[i:] // drop SuperClassName\n\t\tskipSpace()\n\t\tif len(tag) > 0 {\n\t\t\tif tag[0] != '<' {\n\t\t\t\treturn 0, fmt.Errorf(\"objc: expected '<': %w\", TagFormatError)\n\t\t\t}\n\t\t\ttag = tag[1:] // drop '<'\n\t\t\t// get Protocols\n\t\touter:\n\t\t\tfor {\n\t\t\t\tskipSpace()\n\t\t\t\tfor i, r = range tag {\n\t\t\t\t\tswitch r {\n\t\t\t\t\tcase ' ':\n\t\t\t\t\t\tsplit = append(split, tag[:i])\n\t\t\t\t\t\ttag = tag[i:]\n\t\t\t\t\t\tcontinue outer\n\t\t\t\t\tcase ',':\n\t\t\t\t\t\t// If there is actually an identifier - add it.\n\t\t\t\t\t\tif i > 0 {\n\t\t\t\t\t\t\tsplit = append(split, tag[:i])\n\t\t\t\t\t\t\ttag = tag[i:]\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t// Otherwise, drop ','\n\t\t\t\t\t\t\ttag = tag[1:]\n\t\t\t\t\t\t}\n\t\t\t\t\t\tcontinue outer\n\t\t\t\t\tcase '>':\n\t\t\t\t\t\t// If there is actually an identifier - add it.\n\t\t\t\t\t\tif i > 0 {\n\t\t\t\t\t\t\tsplit = append(split, tag[:i])\n\t\t\t\t\t\t\ttag = tag[i:]\n\t\t\t\t\t\t}\n\t\t\t\t\t\tbreak outer\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\treturn 0, fmt.Errorf(\"objc: expected '>': %w\", TagFormatError)\n\t\t\t}\n\t\t}\n\t}\n\tclass := objc_allocateClassPair(GetClass(split[1]), split[0], 0)\n\tif class == 0 {\n\t\treturn 0, fmt.Errorf(\"objc: failed to create class with name '%s'\", split[0])\n\t}\n\tif len(split) > 2 {\n\t\t// Add Protocols\n\t\tfor _, n := range split[2:] {\n\t\t\tif !class.AddProtocol(GetProtocol(n)) {\n\t\t\t\treturn 0, fmt.Errorf(\"objc: couldn't add Protocol %s\", n)\n\t\t\t}\n\t\t}\n\t}\n\t// Add exported methods based on the selectors returned from Selector(string) SEL\n\tfor i := 0; i < ptr.NumMethod(); i++ {\n\t\tmet := ptr.Method(i)\n\t\t// we know this method is the interface one since RegisterClass\n\t\t// requires that the struct implement Selector.\n\t\tif met.Name == \"Selector\" {\n\t\t\tcontinue\n\t\t}\n\t\tsel := object.Selector(met.Name)\n\t\tif sel == 0 {\n\t\t\tcontinue\n\t\t}\n\t\tfn := met.Func.Interface()\n\t\timp, err := func() (imp IMP, err error) {\n\t\t\tdefer func() {\n\t\t\t\tif r := recover(); r != nil {\n\t\t\t\t\terr = fmt.Errorf(\"objc: failed to create IMP: %s\", r)\n\t\t\t\t}\n\t\t\t}()\n\t\t\treturn NewIMP(fn), nil\n\t\t}()\n\t\tif err != nil {\n\t\t\treturn 0, fmt.Errorf(\"objc: couldn't add Method %s: %w\", met.Name, err)\n\t\t}\n\t\tencoding, err := encodeFunc(fn)\n\t\tif err != nil {\n\t\t\treturn 0, fmt.Errorf(\"objc: couldn't add Method %s: %w\", met.Name, err)\n\t\t}\n\t\tif !class.AddMethod(sel, imp, encoding) {\n\t\t\treturn 0, fmt.Errorf(\"objc: couldn't add Method %s\", met.Name)\n\t\t}\n\t}\n\t// Add Ivars\n\t// Start at 1 because we skip the class object which is first\n\tfor i := 1; i < strct.NumField(); i++ {\n\t\tf := strct.Field(i)\n\t\tif f.Name == \"_\" {\n\t\t\tcontinue\n\t\t}\n\t\tsize := f.Type.Size()\n\t\talignment := uint8(math.Log2(float64(f.Type.Align())))\n\t\tenc, err := encodeType(f.Type, false)\n\t\tif err != nil {\n\t\t\treturn 0, fmt.Errorf(\"objc: couldn't add Ivar %s: %w\", f.Name, err)\n\t\t}\n\t\tif !class_addIvar(class, f.Name, size, alignment, enc) {\n\t\t\treturn 0, fmt.Errorf(\"objc: couldn't add Ivar %s\", f.Name)\n\t\t}\n\t\tif offset := class.InstanceVariable(f.Name).Offset(); offset != f.Offset {\n\t\t\treturn 0, fmt.Errorf(\"objc: couldn't add Ivar %s because offset (%d != %d)\", f.Name, offset, f.Offset)\n\t\t}\n\t}\n\tobjc_registerClassPair(class)\n\tif size1, size2 := class.InstanceSize(), strct.Size(); size1 != size2 {\n\t\treturn 0, fmt.Errorf(\"objc: sizes don't match %d != %d: %w\", size1, size2, MismatchError)\n\t}\n\treturn class, nil\n}", "func (s *BaseDMLListener) EnterClassElement(ctx *ClassElementContext) {}", "func addKnownTypes(version string, types ...MilpaObject) {\n\tknownTypes := map[string]reflect.Type{}\n\tfor _, obj := range types {\n\t\tt := reflect.TypeOf(obj)\n\t\tknownTypes[t.Name()] = t\n\t}\n\tversionMap[version] = knownTypes\n}", "func (a *Attributes) Class() string {\n\treturn a.Get(\"class\")\n}", "func updateclass(id int, b *classdto) (int, error) {\n\tcdb.mu.Lock()\n\tdefer cdb.mu.Unlock()\n\t//Checking if the class alredy exists with the same name and type\n\t// var class *Class\n\tnewClass := &Class{}\n\tif oldClass, ok := cdb.classMap[id]; ok {\n\t\tnewClass.ClassID = oldClass.ClassID\n\t\tnewClass.Name = b.Name\n\t\tnewClass.StartDate = b.StartDate\n\t\tnewClass.EndDate = b.EndDate\n\t\tcdb.classMap[id] = *newClass\n\n\t\treturn newClass.ClassID, nil\n\t}\n\treturn -1, fmt.Errorf(\"Class not found in database\")\n\n}", "func (r *InferenceClassificationOverridesCollectionRequest) Add(ctx context.Context, reqObj *InferenceClassificationOverride) (resObj *InferenceClassificationOverride, err error) {\n\terr = r.JSONRequest(ctx, \"POST\", \"\", reqObj, &resObj)\n\treturn\n}", "func createClass(w http.ResponseWriter, r *http.Request) {\n\treqBody, _ := ioutil.ReadAll(r.Body)\n\n\tvar classRequest ClassRequest\n\terr := json.Unmarshal(reqBody, &classRequest)\n\tif err != nil {\n\t\terr = errorResponse(w, InvalidJSON, http.StatusBadRequest)\n\t\tif err != nil {\n\t\t\tfmt.Println(err)\n\t\t}\n\t\treturn\n\t}\n\n\tvar classes []Class\n\tstartDate, err := time.Parse(layoutISO, classRequest.StartDate)\n\tif err != nil {\n\t\terr = errorResponse(w, InvalidDate, http.StatusBadRequest)\n\t\tif err != nil {\n\t\t\tfmt.Println(err)\n\t\t}\n\t\treturn\n\t}\n\tendDate, err := time.Parse(layoutISO, classRequest.EndDate)\n\tif err != nil {\n\t\terr = errorResponse(w, InvalidDate, http.StatusBadRequest)\n\t\tif err != nil {\n\t\t\tfmt.Println(err)\n\t\t}\n\t\treturn\n\t}\n\n\tfor days := 0; days <= int(endDate.Sub(startDate).Hours()/24); days++ {\n\t\tclass := Class{\n\t\t\tId: createID(),\n\t\t\tName: classRequest.Name,\n\t\t\tDate: startDate.Add(time.Hour * 24 * time.Duration(days)),\n\t\t\tCapacity: classRequest.Capacity,\n\t\t}\n\t\tclasses = append(classes, class)\n\t}\n\tDBClasses = append(DBClasses, classes...)\n\n\tw.WriteHeader(http.StatusCreated)\n\terr = json.NewEncoder(w).Encode(classes)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn\n\t}\n}", "func (a *Attributes) HasClass(c string) bool {\n\tvar curClass string\n\tif curClass = a.Get(\"class\"); curClass == \"\" {\n\t\treturn false\n\t}\n\tf := strings.Fields(curClass)\n\tfor _, s := range f {\n\t\tif s == c {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (info *fileInfo) addElaboratedTypes() {\n\tgen := &ast.GenDecl{\n\t\tTokPos: info.importCPos,\n\t\tTok: token.TYPE,\n\t}\n\tnames := make([]string, 0, len(info.elaboratedTypes))\n\tfor name := range info.elaboratedTypes {\n\t\tnames = append(names, name)\n\t}\n\tsort.Strings(names)\n\tfor _, name := range names {\n\t\ttyp := info.elaboratedTypes[name]\n\t\ttypeName := \"C.struct_\" + name\n\t\tobj := &ast.Object{\n\t\t\tKind: ast.Typ,\n\t\t\tName: typeName,\n\t\t}\n\t\ttypeSpec := &ast.TypeSpec{\n\t\t\tName: &ast.Ident{\n\t\t\t\tNamePos: info.importCPos,\n\t\t\t\tName: typeName,\n\t\t\t\tObj: obj,\n\t\t\t},\n\t\t\tType: typ,\n\t\t}\n\t\tobj.Decl = typeSpec\n\t\tgen.Specs = append(gen.Specs, typeSpec)\n\t}\n\tinfo.Decls = append(info.Decls, gen)\n}", "func (*H2F) Class() string {\n\treturn \"TH2F\"\n}", "func (client *Client) Class(name string) *Class {\n\treturn &Class{\n\t\tc: client,\n\t\tName: name,\n\t}\n}", "func UseClassWhitelist(whitelist []string) {\n\tlogrus.WithField(\"whitelist\", whitelist).Warning(\"The default class whitelist has been overridden\")\n\tclassWhitelist = whitelist\n}", "func (*H2D) Class() string {\n\treturn \"TH2D\"\n}", "func newContentTypes(f interface{}, pkg *PackageInfo) *ContentTypes {\n\tcontent := &ContentTypes{\n\t\tpkg: pkg,\n\t}\n\n\tcontent.file = NewPackageFile(pkg, f, &content.ml, nil)\n\tcontent.file.LoadIfRequired(nil)\n\treturn content\n}", "func (m *Manager) AddClassGoSubQuad(className, instance string) {\n\tif m.paramCounter.Top() < len(m.classTable[className].Methods[m.currentFunctionCall.Top()].Params) {\n\t\tlog.Fatalf(\n\t\t\t\"Error: (AddClassGoSubQuad) function %s has too few arguments\",\n\t\t\tm.currentFunctionCall.Top(),\n\t\t)\n\t}\n\n\tif m.currentFunctionCall.Top() != \"init\" {\n\t\tselfDir := m.getCurrentFunctionTable()[m.currentFunction].Vars[instance].Dir\n\t\tm.classTable[className].Methods[m.currentFunctionCall.Top()].Vars[\"self\"].Dir = selfDir\n\t\tinstanceName := fmt.Sprintf(\"self_%d_%s\", selfDir, instance)\n\t\tinstanceElement := NewElement(selfDir, instanceName, constants.ADDR, className)\n\t\tq := Quad{INSTANCE, instanceElement, nil, nil}\n\t\tm.quads = append(m.quads, q)\n\t}\n\n\tdir := m.classTable[className].Methods[m.currentFunctionCall.Top()].Dir\n\tn := NewElement(0, m.currentFunctionCall.Top(), constants.TYPEINT, className)\n\tdirElement := NewElement(dir, \"\", constants.ADDR, \"\")\n\tq := Quad{GOSUB, n, nil, dirElement}\n\tm.quads = append(m.quads, q)\n\n\tif m.classTable[className].Methods[m.currentFunctionCall.Top()].TypeOf != constants.VOID {\n\t\tresultType := m.classTable[className].Methods[m.currentFunctionCall.Top()].TypeOf\n\t\tif resultType == constants.ERR {\n\t\t\tlog.Fatalf(\n\t\t\t\t\"Error: (AddClassGoSubQuad) error in return type %s\",\n\t\t\t\tm.classTable[className].Methods[m.currentFunctionCall.Top()].TypeOf,\n\t\t\t)\n\t\t}\n\n\t\tdir, err := memory.Manager.GetNextAddr(resultType, memory.Temp)\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"Error: (AddClassGoSubQuad) %s\\n\", err)\n\t\t}\n\n\t\tvar resultClass string\n\t\tif m.currentFunctionCall.Top() == \"init\" {\n\t\t\tresultClass = className\n\t\t\tm.getCurrentFunctionTable()[m.currentFunction].Objects = append(m.getCurrentFunctionTable()[m.currentFunction].Objects, className)\n\t\t}\n\t\tresult := NewElement(dir, m.getNextAvail(), resultType, resultClass)\n\t\tm.operands.Push(result)\n\n\t\treturnDir := m.classTable[className].Methods[m.currentFunctionCall.Top()].ReturnDir\n\t\tfuncReturnElement := NewElement(returnDir, \"\", constants.ADDR, \"\")\n\t\tq := Quad{ASSIGN, funcReturnElement, nil, result}\n\t\tm.quads = append(m.quads, q)\n\t}\n\tm.currentFunctionCall.Pop()\n\tm.currentFunctionCallClass.Pop()\n\tm.paramCounter.Pop()\n}", "func (c ClusterNodes) Add() {\n\n}", "func (c *contentTypes) add(partName, contentType string) error {\n\t// Process descrived in ISO/IEC 29500-2 §10.1.2.3\n\tt, params, _ := mime.ParseMediaType(contentType)\n\tcontentType = mime.FormatMediaType(t, params)\n\n\text := strings.ToLower(path.Ext(partName))\n\tif len(ext) == 0 {\n\t\tc.addOverride(partName, contentType)\n\t\treturn nil\n\t}\n\text = ext[1:] // remove dot\n\tc.ensureDefaultsMap()\n\tcurrentType, ok := c.defaults[ext]\n\tif ok {\n\t\tif currentType != contentType {\n\t\t\tc.addOverride(partName, contentType)\n\t\t}\n\t} else {\n\t\tc.addDefault(ext, contentType)\n\t}\n\n\treturn nil\n}", "func newOsType(name string, class OsClass, defDisabled bool, archTypes ...ArchType) OsType {\n\tcheckCalledFromInit()\n\tos := OsType{\n\t\tName: name,\n\t\tField: proptools.FieldNameForProperty(name),\n\t\tClass: class,\n\n\t\tDefaultDisabled: defDisabled,\n\t}\n\tOsTypeList = append(OsTypeList, os)\n\n\tif _, found := commonTargetMap[name]; found {\n\t\tpanic(fmt.Errorf(\"Found Os type duplicate during OsType registration: %q\", name))\n\t} else {\n\t\tcommonTargetMap[name] = Target{Os: os, Arch: CommonArch}\n\t}\n\tosArchTypeMap[os] = archTypes\n\n\treturn os\n}", "func ParseClasses(d *drawing.Drawing, line int, data [][2]string) error {\n\treturn nil\n}", "func addTypeDef(def *TypeDef, env *Env) {\n\tdef.File.TypeDefs = append(def.File.TypeDefs, def)\n\tdef.File.Package.typeDefs = append(def.File.Package.typeDefs, def)\n\tdef.File.Package.typeMap[def.Name] = def\n\tif env != nil {\n\t\t// env should only be nil during initialization of the built-in package;\n\t\t// NewEnv ensures new environments have the built-in types.\n\t\tenv.typeMap[def.Type] = def\n\t}\n}", "func (table *ClassTable) Insert(class Class) (new Class, err error) {\n\t// Query over these fields to ensure an existing class doesn't already exist\n\tclassQuery := ClassQuery{Name: class.Name, Quarter: class.Quarter, Year: class.Year, Creator_id: class.Creator_id}\n\tdata, err := table.connection.Insert(CLASS_TABLE, \"AND\", class, classQuery)\n\tif err != nil {\n\t\treturn\n\t}\n\tutilities.Sugar.Infof(\"Class Data: %v\", data)\n\terr = utilities.FillStruct(data, &new)\n\treturn\n}", "func (u UpdatesCombined) construct() UpdatesClass { return &u }", "func Add(t Type) Type {\n\tTypes.Store(t.Extension, t)\n\treturn t\n}", "func (u Updates) construct() UpdatesClass { return &u }", "func (t Type) Class() Class {\n\ttd, ok := types[t]\n\tif !ok {\n\t\treturn Unclassified\n\t}\n\treturn td.class\n}", "func (hfsc *HfscClass) Type() string {\n\treturn \"hfsc\"\n}", "func (p *Window) Class() string {\n\treturn (`winl.Window`)\n}", "func (c ServiceCollection) Add(s *Service) {\n\tc[s.Type] = s\n}", "func (this *Motto) AddModule(id string, loader ModuleLoader) {\n this.modules[id] = loader\n}", "func init() {\n\tmapType(\"wxUpdateUIEvent\", reflect.TypeOf(updateUIEvent{}))\n}", "func AddType(m *Types) (id int64, err error) {\n\to := orm.NewOrm()\n\tid, err = o.Insert(m)\n\treturn\n}", "func (a *App) AddRunnable(constructor interface{}) {\n\tp := newProvider(constructor)\n\ta.providers = append(a.providers, p)\n\n\tfor _, output := range p.outputTypes {\n\t\tif !_isErrorType(output) {\n\t\t\ta.runnables = append(a.runnables, typeKey{output})\n\t\t}\n\t}\n}", "func WithClass(class string) Option {\n\treturn func(o *nvcdilib) {\n\t\to.class = class\n\t}\n}", "func addKnownTypes(scheme *runtime.Scheme) error {\n\tscheme.AddKnownTypes(SchemeGroupVersion,\n\t\t&Spiderman{},\n\t\t&SpidermanList{},\n\t)\n\tmetav1.AddToGroupVersion(scheme, SchemeGroupVersion)\n\treturn nil\n}", "func Class(name string, properties ...Declaration) (id ClassID) {\n\tid = ClassID(fmt.Sprintf(\"%s_%d\", name, len(styles)))\n\tr := Rule{\n\t\tSelector: \".\" + string(id),\n\t\tDeclarations: properties,\n\t}\n\tstyles = append(styles, r)\n\treturn\n}", "func formToClass(form format, attr Attr, vers int, b *buf) Class {\n\tswitch form {\n\tdefault:\n\t\tb.error(\"cannot determine class of unknown attribute form\")\n\t\treturn 0\n\n\tcase formIndirect:\n\t\treturn ClassUnknown\n\n\tcase formAddr, formAddrx, formAddrx1, formAddrx2, formAddrx3, formAddrx4:\n\t\treturn ClassAddress\n\n\tcase formDwarfBlock1, formDwarfBlock2, formDwarfBlock4, formDwarfBlock:\n\t\t// In DWARF 2 and 3, ClassExprLoc was encoded as a\n\t\t// block. DWARF 4 distinguishes ClassBlock and\n\t\t// ClassExprLoc, but there are no attributes that can\n\t\t// be both, so we also promote ClassBlock values in\n\t\t// DWARF 4 that should be ClassExprLoc in case\n\t\t// producers get this wrong.\n\t\tif attrIsExprloc[attr] {\n\t\t\treturn ClassExprLoc\n\t\t}\n\t\treturn ClassBlock\n\n\tcase formData1, formData2, formData4, formData8, formSdata, formUdata, formData16, formImplicitConst:\n\t\t// In DWARF 2 and 3, ClassPtr was encoded as a\n\t\t// constant. Unlike ClassExprLoc/ClassBlock, some\n\t\t// DWARF 4 attributes need to distinguish Class*Ptr\n\t\t// from ClassConstant, so we only do this promotion\n\t\t// for versions 2 and 3.\n\t\tif class, ok := attrPtrClass[attr]; vers < 4 && ok {\n\t\t\treturn class\n\t\t}\n\t\treturn ClassConstant\n\n\tcase formFlag, formFlagPresent:\n\t\treturn ClassFlag\n\n\tcase formRefAddr, formRef1, formRef2, formRef4, formRef8, formRefUdata, formRefSup4, formRefSup8:\n\t\treturn ClassReference\n\n\tcase formRefSig8:\n\t\treturn ClassReferenceSig\n\n\tcase formString, formStrp, formStrx, formStrpSup, formLineStrp, formStrx1, formStrx2, formStrx3, formStrx4:\n\t\treturn ClassString\n\n\tcase formSecOffset:\n\t\t// DWARF 4 defines four *ptr classes, but doesn't\n\t\t// distinguish them in the encoding. Disambiguate\n\t\t// these classes using the attribute.\n\t\tif class, ok := attrPtrClass[attr]; ok {\n\t\t\treturn class\n\t\t}\n\t\treturn ClassUnknown\n\n\tcase formExprloc:\n\t\treturn ClassExprLoc\n\n\tcase formGnuRefAlt:\n\t\treturn ClassReferenceAlt\n\n\tcase formGnuStrpAlt:\n\t\treturn ClassStringAlt\n\n\tcase formLoclistx:\n\t\treturn ClassLocList\n\n\tcase formRnglistx:\n\t\treturn ClassRngList\n\t}\n}", "func CreateNewClass(c *gin.Context) {\n\n\t// // Initialize database connection\n\tdb := config.DatabaseConn()\n\n\t// Declare parameters\n\tIDClassroom := uuid.Must(uuid.NewV4())\n\tUUIDClassroom := uuid.Must(uuid.NewV4())\n\tClassroomName := c.Param(\"ClassroomName\")\n\tClassroomTime := c.Param(\"ClassroomTime\")\n\tRoom := c.Param(\"Room\")\n\tUUIDParticipants := c.Param(\"UUIDParticipants\")\n\n\t// Bind in one parameters\n\tcreateClassroomPayload := model.Classroom{\n\t\tIDClassroom: IDClassroom,\n\t\tUUIDClassroom: UUIDClassroom,\n\t\tClassroomName: ClassroomName,\n\t\tClassroomTime: ClassroomTime,\n\t\tRoom: Room,\n\t\tUUIDParticipants: UUIDParticipants}\n\n\t// Bind parameter as JSON Parameters\n\tc.BindJSON(&createClassroomPayload)\n\n\t// Save in database\n\tdb.Save(&createClassroomPayload)\n\n\tc.JSON(http.StatusOK, gin.H{\"message\": \"Inserted successfully\"})\n}", "func (u UpdateShort) construct() UpdatesClass { return &u }", "func defToClassMap(classesMap map[string]phpClass, definitions map[string]smd.Definition) {\n\tfor name, def := range definitions {\n\t\tclassesMap[name] = phpClass{Name: name, Fields: propertiesToParams(def.Properties)}\n\t}\n}", "func RegisterTypes(registry interface {\n\t RegisterType(name string, obj any)\n}) {\n\n}", "func newClassy(view Viewer) Classy {\n\treturn classy{\n\t\tchain: alice.New(),\n\t\tpath: \"/\",\n\t\tstructname: getStructName(view),\n\t\tview: view,\n\t}.Name(getViewName(view, true))\n}", "func AddInstrumentClassHook(hookPoint boil.HookPoint, instrumentClassHook InstrumentClassHook) {\n\tswitch hookPoint {\n\tcase boil.BeforeInsertHook:\n\t\tinstrumentClassBeforeInsertHooks = append(instrumentClassBeforeInsertHooks, instrumentClassHook)\n\tcase boil.BeforeUpdateHook:\n\t\tinstrumentClassBeforeUpdateHooks = append(instrumentClassBeforeUpdateHooks, instrumentClassHook)\n\tcase boil.BeforeDeleteHook:\n\t\tinstrumentClassBeforeDeleteHooks = append(instrumentClassBeforeDeleteHooks, instrumentClassHook)\n\tcase boil.BeforeUpsertHook:\n\t\tinstrumentClassBeforeUpsertHooks = append(instrumentClassBeforeUpsertHooks, instrumentClassHook)\n\tcase boil.AfterInsertHook:\n\t\tinstrumentClassAfterInsertHooks = append(instrumentClassAfterInsertHooks, instrumentClassHook)\n\tcase boil.AfterSelectHook:\n\t\tinstrumentClassAfterSelectHooks = append(instrumentClassAfterSelectHooks, instrumentClassHook)\n\tcase boil.AfterUpdateHook:\n\t\tinstrumentClassAfterUpdateHooks = append(instrumentClassAfterUpdateHooks, instrumentClassHook)\n\tcase boil.AfterDeleteHook:\n\t\tinstrumentClassAfterDeleteHooks = append(instrumentClassAfterDeleteHooks, instrumentClassHook)\n\tcase boil.AfterUpsertHook:\n\t\tinstrumentClassAfterUpsertHooks = append(instrumentClassAfterUpsertHooks, instrumentClassHook)\n\t}\n}", "func Add(name string) {\n\n}", "func getClasses(w http.ResponseWriter, r *http.Request) {\n\terr := json.NewEncoder(w).Encode(DBClasses)\n\tif err != nil {\n\t\terr = errorResponse(w, InternalError, http.StatusInternalServerError)\n\t\tif err != nil {\n\t\t\tfmt.Println(err)\n\t\t}\n\t}\n}", "func (a *Actor) Class() *Class {\n\treturn a.class\n}", "func newClass(name, description string, properties map[string]interface{}) class {\n\tpc := class{\n\t\tName: name,\n\t\tDescription: commons.ParseDescription(description),\n\t\tFields: map[string]field{},\n\t}\n\n\t// generate fields\n\tfor k, v := range properties {\n\t\tp := raml.ToProperty(k, v)\n\t\tfield := field{\n\t\t\tName: p.Name,\n\t\t\tRequired: p.Required,\n\t\t}\n\t\tfield.setType(p.Type)\n\n\t\tif field.Type == \"\" { // type is not supported, no need to generate the field\n\t\t\tcontinue\n\t\t}\n\n\t\tfield.buildValidators(p)\n\t\tpc.Fields[p.Name] = field\n\n\t}\n\treturn pc\n}", "func (e *Env) stubClasses(flat *dota.CSVCMsg_FlattenedSerializer) {\n\tserializers := flat.GetSerializers()\n\tfor _, s := range serializers {\n\t\tname := e.symbol(int(s.GetSerializerNameSym()))\n\t\tv := int(s.GetSerializerVersion())\n\t\tc := &class{name: name, version: v}\n\t\tDebug.Printf(\"new class: %s\", c)\n\t\th := e.classes[name]\n\t\tif h == nil {\n\t\t\th = new(classHistory)\n\t\t\te.classes[name] = h\n\t\t}\n\t\th.add(c)\n\t}\n}", "func (t textBody) GetClass() string {\n\treturn \"\"\n}", "func (cs *Commodities) Add(c *Commodity) {\n\tif c.Space == \"template\" && c.ID == \"template\" {\n\t\treturn\n\t}\n\t*cs = append(*cs, c)\n}", "func (e *HTMLTableRow) Class(v string) *HTMLTableRow {\n\te.a[\"class\"] = v\n\treturn e\n}", "func (s *BaseDMLListener) EnterClassOption(ctx *ClassOptionContext) {}", "func (m *Model) Classes() ([]string, error) {\n\tpanic(\"Not implemented\")\n}", "func (s Store) UpdateClass(mac net.HardwareAddr, name string, class string) bool {\n\tl := &Lease{}\n\tl, err := s.leases.Mac(mac)\n\tif err != nil {\n\t\tlogger.Error(\"lease error %s\", err)\n\t\treturn false\n\t}\n\tl.Active = true\n\tl.Distro = name\n\tl.Class = class\n\tlogger.Critical(\"Update %v to class %s of %s\", l.MAC, l.Class, l.Distro)\n\ts.leases.Save(s.DBname)\n\treturn true\n}", "func (m *Manager) updateClass(ctx context.Context, className string,\n\tclass *models.Class) error {\n\tm.Lock()\n\tdefer m.Unlock()\n\n\tvar newName *string\n\n\tif class.Class != className {\n\t\t// the name in the URI and body don't match, so we assume the user wants to rename\n\t\tn := upperCaseClassName(class.Class)\n\t\tnewName = &n\n\t}\n\n\tsemanticSchema := m.state.SchemaFor()\n\n\tvar err error\n\tclass, err = schema.GetClassByName(semanticSchema, className)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tclassNameAfterUpdate := className\n\n\t// First validate the request\n\tif newName != nil {\n\t\terr = m.validateClassNameUniqueness(*newName)\n\t\tclassNameAfterUpdate = *newName\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Validate name / keywords in contextionary\n\tif err = m.validateClassName(ctx, classNameAfterUpdate); err != nil {\n\t\treturn err\n\t}\n\n\t// Validated! Now apply the changes.\n\tclass.Class = classNameAfterUpdate\n\n\terr = m.saveSchema(ctx)\n\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\treturn m.migrator.UpdateClass(ctx, className, newName)\n}", "func (p *Parser) class() (result Value) {\n\tif p.Token == tok.Class {\n\t\tp.Match(tok.Class)\n\t\tif p.Token == tok.Colon {\n\t\t\tp.Match(tok.Colon)\n\t\t}\n\t}\n\tvar base Gnum\n\tbaseName := \"class\"\n\tif p.Token.IsIdent() {\n\t\tbaseName = p.Text\n\t\tbase = p.ckBase(baseName)\n\t\tp.MatchIdent()\n\t}\n\tpos1 := p.EndPos\n\tp.Match(tok.LCurly)\n\tpos2 := p.EndPos\n\tprevClassName := p.className\n\tp.className = p.getClassName()\n\tmems := p.mkClass(baseName)\n\tp.memberList(mems, tok.RCurly, base)\n\tp.setPos(mems, pos1, pos2)\n\tp.className = prevClassName\n\tif cc, ok := mems.(classBuilder); ok {\n\t\treturn &SuClass{Base: base, Lib: p.lib, Name: p.name,\n\t\t\tMemBase: MemBase{Data: cc}}\n\t}\n\treturn mems.(Value)\n}", "func appendHLClass(class *syntaxClass, hl, line []byte) ([]byte, int, *syntaxClass) {\n\tif class.middle != nil {\n\t\tif ix := class.middle.FindIndex(line); len(ix) == 2 && ix[0] == 0 && ix[1] > 0 {\n\t\t\treturn appendHLColor(hl, ix[1], class.hlType), ix[1], class\n\t\t}\n\t}\n\n\tif class.end != nil {\n\t\tif ix := class.end.FindIndex(line); len(ix) == 2 && ix[0] == 0 && ix[1] > 0 {\n\t\t\treturn appendHLColor(hl, ix[1], class.hlType), ix[1], nil\n\t\t}\n\t}\n\tif len(line) > 0 {\n\t\treturn hl, 0, nil\n\t}\n\treturn hl, 0, class\n}", "func (m *Model) InsertClassification(seriesID, classID int) (*model.Class, error) {\n\t_, err := m.db.Exec(qryInsertClassification, seriesID, classID)\n\tif err != nil {\n\t\treturn nil, filterError(err)\n\t}\n\treturn &model.Class{ID: int64(classID)}, nil\n\n}", "func (h Handler) Add(obj runtime.Object) error {\n\tnode := obj.(*corev1.Node)\n\terr := h.nodeReaper.Run(node)\n\tif err != nil {\n\t\th.logger.Errorf(\"Error processing node %q: %s\", node.Name, err)\n\t}\n\treturn nil\n}", "func OpenClass(name, extends string, implements ...string) *Writer {\n\tlog.Println(\"Generate class \", name)\n\tw := NewWriterFile(name + \".java\")\n\n\tmod := PUBLIC\n\tw.Put(mod, \"class \", name)\n\tif extends != \"\" {\n\t\tw.Put(\" extends \", extends)\n\t}\n\tif len(implements) != 0 {\n\t\tw.Put(\" implements\")\n\t\tfor _, x := range implements {\n\t\t\tw.Put(\" \", x)\n\t\t}\n\t}\n\tw.Putln(\"{\\n\")\n\tw.indent++\n\treturn w\n}", "func (r *Registry) Add(modules ...bar.Module) *Registry {\n\tif r.err != nil {\n\t\treturn r\n\t}\n\n\tfor _, module := range modules {\n\t\tif module != nil {\n\t\t\tr.modules = append(r.modules, module)\n\t\t}\n\t}\n\treturn r\n}", "func PrintClass(a interface{}) {\n\tj, _ := json.Marshal(a)\n\tfmt.Println(string(j))\n}", "func (classes ClassMap) Merge(source ClassMap) ClassMap {\n\tfor name, class := range source {\n\t\tif class != nil {\n\t\t\tclasses[name].MethodMap.Merge(class.MethodMap)\n\t\t}\n\t}\n\treturn classes\n}", "func (node *Node) HasClass(class string) bool {\n\tif class == \"\" {\n\t\treturn true\n\t}\n\tcheck, ok := node.Attributes[\"class\"]\n\tif !ok {\n\t\treturn false\n\t}\n\treturn check.Val == class\n}", "func addKnownTypes(scheme *runtime.Scheme) error {\n\tscheme.AddKnownTypes(SchemeGroupVersion,\n\t\t&Project{},\n\t\t&ProjectList{},\n\t\t&ProjectRequest{},\n\t)\n\treturn nil\n}", "func AddTypeName(name string) {\n\tif _, ok := extraTypes[name]; !ok {\n\t\textraTypes[name] = &Type{Kind: TypedefType, Name: name}\n\t}\n}", "func (c FakeClassList) Contains(class string) bool {\n\t_, found := c[class]\n\treturn found\n}", "func add[Type IntOrString](items []Type) (result Type) {\n\tfor _, item := range items {\n\t\tresult = result + item\n\t}\n\treturn\n}", "func (a AttachMenuBots) construct() AttachMenuBotsClass { return &a }" ]
[ "0.7353803", "0.6405253", "0.6077631", "0.57352865", "0.5628054", "0.55100405", "0.5430202", "0.5396594", "0.5327561", "0.52193564", "0.52193564", "0.5143893", "0.5127993", "0.51194954", "0.5070766", "0.5058464", "0.5051829", "0.5024158", "0.49740955", "0.49037462", "0.49021664", "0.48989704", "0.48833072", "0.48573944", "0.4814668", "0.48064616", "0.4800856", "0.47937718", "0.4781198", "0.4761053", "0.47528028", "0.4750695", "0.47439486", "0.4728127", "0.47261265", "0.47191456", "0.46772715", "0.46671656", "0.46620867", "0.46603596", "0.46557122", "0.46552235", "0.46538317", "0.46450213", "0.4637968", "0.46290767", "0.46269593", "0.46184716", "0.4614118", "0.45971847", "0.45943716", "0.4592624", "0.4592207", "0.4589953", "0.45886773", "0.45731196", "0.456791", "0.45585042", "0.45545942", "0.45522073", "0.45513564", "0.45439592", "0.45338526", "0.45257446", "0.45252296", "0.45226684", "0.4522306", "0.45222035", "0.4510286", "0.45095494", "0.45013592", "0.44983256", "0.4494817", "0.44817767", "0.4481198", "0.44765887", "0.44704804", "0.44639057", "0.44636542", "0.44610107", "0.4433141", "0.44213095", "0.44173837", "0.44158483", "0.44147006", "0.4407709", "0.4403189", "0.44014743", "0.4400829", "0.4399321", "0.43992606", "0.439633", "0.43856466", "0.4369503", "0.43694657", "0.43690366", "0.4363966", "0.43628034", "0.43614814", "0.43567178" ]
0.5485202
6
On defines an event handler
func (h *HTML) On(event string, handler EventHandler) *HTML { h.events = append(h.events, eventHandler{event, handler}) return h }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *Client) On(event string, handler Handler) {\n\tc.handlers[event] = handler\n}", "func (m *Main) On(event string, cb EventCallback) {\n\tm.Callback[event] = cb\n}", "func (s *Server) On(event string, f interface{}) {\n\tvalue := reflect.ValueOf(f)\n\n\tif !validHandler(value, reflect.TypeOf(f).Name()) {\n\t\tpanic(\"On: invalid callback for event \" + event)\n\t}\n\n\ts.handlersLock.Lock()\n\ts.handlers[event] = &regMethod{value, value.Type().In(1)}\n\ts.handlersLock.Unlock()\n}", "func (c *Connector) On(event string, callback Callback) {\n\tc.muEvents.Lock()\n\tdefer c.muEvents.Unlock()\n\n\tc.events[event] = callback\n}", "func (player *Player) on(eventName event.Name, handler func(event interface{})) {\n\tplayer.Lock()\n\tplayer.handlers[eventName] = handler\n\tplayer.Unlock()\n\t_ = player.WriteJSON(protocol.NewEventRequest(eventName, protocol.Subscribe))\n}", "func (h *Hub) On(event string, fn EventHandler) *Hub {\n\tif h.listeners[event] == nil {\n\t\th.listeners[event] = []EventHandler{}\n\t}\n\n\th.listeners[event] = append(h.listeners[event], fn)\n\n\treturn h\n}", "func (k *Bot) On(event string, cb Callback) {\n\tk.mutex.Lock()\n\tif _, ok := k.callBacks[event]; !ok {\n\t\tk.callBacks[event] = []Callback{}\n\t}\n\tk.callBacks[event] = append(k.callBacks[event], cb)\n\tk.mutex.Unlock()\n}", "func (o *PluginDnsClient) OnEvent(msg string, a, b interface{}) {}", "func (o *IpfixNsPlugin) OnEvent(msg string, a, b interface{}) {}", "func (c *Client) On(event string, cb ...MessageFunc) {\n\tif cb == nil {\n\t\treturn\n\t}\n\n\tls, ok := c.onEventListeners.Load(event)\n\tif !ok {\n\t\tc.onEventListeners.Store(event, cb)\n\t\treturn\n\t}\n\tcbs, ok := ls.([]MessageFunc)\n\tif !ok {\n\t\tc.onEventListeners.Store(event, cb)\n\t\treturn\n\t}\n\n\tc.onEventListeners.Store(event, append(cbs, cb...))\n}", "func (c *Connection) On(ev string, callback func(interface{})) {\n\tif c.eventMap[ev] == nil {\n\t\tc.eventMap[ev] = new(list.List)\n\t}\n\tc.eventMap[ev].PushBack(callback)\n}", "func (o *PluginIgmpClient) OnEvent(msg string, a, b interface{}) {\n\n}", "func (o *PluginDnsNs) OnEvent(msg string, a, b interface{}) {}", "func (e *EventHandlerFuncs) OnAdd(table string, model Model) {\n\tif e.AddFunc != nil {\n\t\te.AddFunc(table, model)\n\t}\n}", "func (o *IPFixGen) OnEvent(a, b interface{}) {\n\tif o.ipfixPlug.enabled == false {\n\t\treturn\n\t}\n\n\tif o.ipfixPlug.maxTime > 0 {\n\t\tenabledDuration := time.Since(o.ipfixPlug.enabledTime)\n\t\tif enabledDuration >= o.ipfixPlug.maxTime {\n\t\t\to.ipfixPlug.Enable(false)\n\t\t\treturn\n\t\t}\n\t}\n\n\tvar isTemplate bool\n\tswitch v := a.(type) {\n\tcase bool:\n\t\tisTemplate = v\n\tdefault:\n\t\treturn\n\t}\n\tif isTemplate {\n\t\to.sendTemplatePkt()\n\t} else {\n\t\to.sendDataPkt()\n\t}\n\n\tif o.isReachedMaxDataRecordsToSend() && o.isReachedMaxTempRecordsToSend() {\n\t\t// We reached the configured maximum number of records - disable exporting\n\t\to.ipfixPlug.Enable(false)\n\t}\n}", "func (e *RepoEntity) On(event string, listener RepositoryListener) {\n\te.mutex.Lock()\n\tdefer e.mutex.Unlock()\n\t// add listener to the specific event topic\n\te.listeners[event] = append(e.listeners[event], listener)\n}", "func (e *UntypedEmitter) On(handlerFunc Handler) {\n\thv, et, err := reflectHandler(handlerFunc)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\thp := hv.Pointer()\n\n\te.mu.Lock()\n\tdefer e.mu.Unlock()\n\n\tif e.allHandlers == nil {\n\t\te.allHandlers = make(map[eventType]handlers)\n\t\te.allHandlers[et] = handlers{hp: handlerFunc}\n\t\treturn\n\t}\n\tif e.allHandlers[et] == nil {\n\t\te.allHandlers[et] = handlers{hp: handlerFunc}\n\t\treturn\n\t}\n\te.allHandlers[et][hp] = handlerFunc\n}", "func on(t interface{}, f func(e events.Event)) *ListenerFunc {\n\treturn &ListenerFunc{\n\t\tfn: func(e events.Event) {\n\t\t\tif reflect.TypeOf(t) == reflect.TypeOf(e) {\n\t\t\t\tf(e)\n\t\t\t}\n\t\t},\n\t}\n}", "func (emitter *EventEmitter) On(eventName string, listener func(...interface{})) *EventEmitter {\n\treturn emitter.addListener(eventName, listener, false)\n}", "func (h HandlerFunc) Handle(e Event) { h(e) }", "func (a *RedisAction) On() func(event interface{}) {\n\n\tf := func(e interface{}) {\n\t\ta.out <- e.(*docker.APIEvents)\n\t}\n\n\treturn f\n}", "func On(hk ent.Hook, op ent.Op) ent.Hook {\n\treturn If(hk, HasOp(op))\n}", "func (f *FileWatcher) OnEvent(fn OnEventFunc) {\n\tf.onEvent = fn\n}", "func (c *Component) Handler(n string, h EventHandler) {\n\tc.GM.RegisterHandler(n, h)\n}", "func (s *ServicesWidget) OnEvent(event appui.EventCommand) error {\n\tif s.RowCount() > 0 {\n\t\treturn event(s.filteredRows[s.selectedIndex].service.ID)\n\t}\n\treturn nil\n}", "func setEventHandler(fn func())", "func (c *Change) On() *Change {\n\tc.params[\"on\"] = true\n\treturn c\n}", "func (_m *ComponentEventListener) OnEvent(event *common.Event) {\n\t_m.Called(event)\n}", "func (n *EventReceiver) Event(eventName string) {\n\n}", "func (c *Controller) OnAdd(obj interface{}, _ bool) {\n\tif err := c.ensureOpened(obj); err != nil {\n\t\tlog.Errorf(\"OnAdd failed: %v\", err)\n\t}\n}", "func (t *TPLink) On() error {\n\treturn t.send(kOn)\n}", "func (p *Processor) OnEvent() {\n\tp.eventChan <- struct{}{}\n}", "func (e LocalService) OnEvent(ctx context.Context, event models.KeptnContextExtendedCE) error {\n\t// You can grab handle the event and grab a sender to send back started / finished events to keptn\n\t// eventSender := ctx.Value(controlplane.EventSenderKeyType{}).(types.EventSender)\n\treturn nil\n}", "func (e *event) On(name string, fn interface{}) error {\n\te.Lock()\n\tdefer e.Unlock()\n\n\tif fn == nil {\n\t\treturn errors.New(\"fn is nil\")\n\t}\n\tif _, ok := fn.(handle); ok {\n\t\te.events[name] = append(e.events[name], fn)\n\t\treturn nil\n\t}\n\n\tt := reflect.TypeOf(fn)\n\tif t.Kind() != reflect.Func {\n\t\treturn errors.New(\"fn is not a function\")\n\t}\n\tif t.NumOut() != 1 {\n\t\treturn errors.New(\"fn must have one return value\")\n\t}\n\tif t.Out(0) != reflect.TypeOf((*error)(nil)).Elem() {\n\t\treturn errors.New(\"fn must return an error message\")\n\t}\n\tif list, ok := e.events[name]; ok && len(list) > 0 {\n\t\ttt := reflect.TypeOf(list[0])\n\t\tif tt.NumIn() != t.NumIn() {\n\t\t\treturn errors.New(\"fn signature is not equal\")\n\t\t}\n\t\tfor i := 0; i < tt.NumIn(); i++ {\n\t\t\tif tt.In(i) != t.In(i) {\n\t\t\t\treturn errors.New(\"fn signature is not equal\")\n\t\t\t}\n\t\t}\n\t}\n\n\te.events[name] = append(e.events[name], fn)\n\treturn nil\n}", "func (r *jsiiProxy_Repository) OnEvent(id *string, options *awsevents.OnEventOptions) awsevents.Rule {\n\tvar returns awsevents.Rule\n\n\t_jsii_.Invoke(\n\t\tr,\n\t\t\"onEvent\",\n\t\t[]interface{}{id, options},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func eventHandler(evt *elemental.Event) {\n\n\tswitch evt.Identity {\n\n\tcase gaia.ExternalNetworkIdentity.Name:\n\n\t\te := gaia.NewExternalNetwork()\n\t\tif err := evt.Decode(&e); err != nil {\n\t\t\tzap.L().Error(\"Failed to decode event\", zap.Reflect(\"event\", evt))\n\t\t}\n\n\t\tfmt.Printf(\"External network name: %s type %s\\n\", e.Name, evt.Type)\n\n\tcase gaia.NetworkAccessPolicyIdentity.Name:\n\n\t\tp := gaia.NewNetworkAccessPolicy()\n\t\tif err := evt.Decode(&p); err != nil {\n\t\t\tzap.L().Error(\"Failed to decode event\", zap.Reflect(\"event\", evt))\n\t\t}\n\n\t\tfmt.Printf(\"Policy name: %s type %s\\n\", p.Name, evt.Type)\n\n\tdefault:\n\t\tzap.L().Error(\"Received event that was not subscribed\", zap.Reflect(\"event\", evt))\n\t}\n}", "func (l *List) OnEvent(e interface{}) {\n\tswitch e.(type) {\n\tcase sparta.CloseEvent:\n\t\tif l.closeFn != nil {\n\t\t\tl.closeFn(l, e)\n\t\t}\n\t\tl.scroll.OnEvent(e)\n\tcase sparta.ConfigureEvent:\n\t\trect := e.(sparta.ConfigureEvent).Rect\n\t\tl.geometry = rect\n\t\tif l.configFn != nil {\n\t\t\tl.configFn(l, e)\n\t\t}\n\t\tl.scroll.SetProperty(sparta.Geometry, image.Rect(rect.Dx()-10, 0, rect.Dx(), rect.Dy()))\n\t\tl.scroll.SetProperty(ScrollPage, l.geometry.Dy()/sparta.HeightUnit)\n\tcase sparta.CommandEvent:\n\t\tif l.commFn != nil {\n\t\t\tif l.commFn(l, e) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tev := e.(sparta.CommandEvent)\n\t\tif ev.Source == l.scroll {\n\t\t\tl.Update()\n\t\t\treturn\n\t\t}\n\t\tl.parent.OnEvent(e)\n\tcase sparta.ExposeEvent:\n\t\tif l.exposeFn != nil {\n\t\t\tl.exposeFn(l, e)\n\t\t}\n\t\tl.win.SetColor(sparta.Foreground, foreColor)\n\t\tif (l.list != nil) && (l.list.Len() > 0) {\n\t\t\tpos := l.scroll.Property(ScrollPos).(int)\n\t\t\tif pos < 0 {\n\t\t\t\tpos = 0\n\t\t\t}\n\t\t\tpage := l.scroll.Property(ScrollPage).(int)\n\t\t\tfor i := 0; i <= page; i++ {\n\t\t\t\tj := i + pos\n\t\t\t\tif j >= l.list.Len() {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tif l.list.IsSel(j) {\n\t\t\t\t\ty := (i * sparta.HeightUnit) + 2\n\t\t\t\t\tl.win.Text(image.Pt(2, y), \">\")\n\t\t\t\t}\n\t\t\t\tl.win.Text(image.Pt(2+sparta.WidthUnit, (i*sparta.HeightUnit)+2), l.list.Item(j))\n\t\t\t}\n\t\t}\n\t\trect := image.Rect(0, 0, l.geometry.Dx()-1, l.geometry.Dy()-1)\n\t\tl.win.Rectangle(rect, false)\n\tcase sparta.KeyEvent:\n\t\tif s.keyFn != nil {\n\t\t\tif s.keyFn(s, e) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tpos := l.scroll.Property(ScrollPos).(int)\n\t\tpage := l.scroll.Property(ScrollPage).(int)\n\t\tev := e.(sparta.KeyEvent)\n\t\tswitch ev.Key {\n\t\tcase sparta.KeyDown:\n\t\t\tl.scroll.SetProperty(ScrollPos, pos+1)\n\t\tcase sparta.KeyUp:\n\t\t\tl.scroll.SetProperty(ScrollPos, pos-1)\n\t\tcase sparta.KeyPageUp:\n\t\t\tl.scroll.SetProperty(ScrollPos, pos-page)\n\t\tcase sparta.KeyPageDown:\n\t\t\tl.scroll.SetProperty(ScrollPos, pos+page)\n\t\tcase sparta.KeyHome:\n\t\t\tl.scroll.SetProperty(ScrollPos, 0)\n\t\tcase sparta.KeyEnd:\n\t\t\ts.SetProperty(ScrollPos, l.list.Len())\n\t\tdefault:\n\t\t\tl.parent.OnEvent(e)\n\t\t}\n\tcase sparta.MouseEvent:\n\t\tif l.mouseFn != nil {\n\t\t\tif l.mouseFn(l, e) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tpos := l.scroll.Property(ScrollPos).(int)\n\t\tev := e.(sparta.MouseEvent)\n\t\tswitch ev.Button {\n\t\tcase -sparta.MouseWheel:\n\t\t\tl.scroll.SetProperty(ScrollPos, pos+1)\n\t\tcase sparta.MouseWheel:\n\t\t\tl.scroll.SetProperty(ScrollPos, pos-1)\n\t\tcase sparta.MouseLeft:\n\t\t\tp := ((ev.Loc.Y - 2) / sparta.HeightUnit) + pos\n\t\t\tsparta.SendEvent(l.target, sparta.CommandEvent{Source: l, Value: p})\n\t\tcase sparta.MouseRight:\n\t\t\tp := ((ev.Loc.Y - 2) / sparta.HeightUnit) + pos\n\t\t\tsparta.SendEvent(l.target, sparta.CommandEvent{Source: l, Value: -(p + 1)})\n\t\t}\n\t}\n}", "func (o *PluginIPFixClient) OnEvent(msg string, a, b interface{}) {\n\tif o.init == false {\n\t\t/* Client initialization failed */\n\t\treturn\n\t}\n\n\tswitch msg {\n\tcase core.MSG_DG_MAC_RESOLVED:\n\t\tbitMask, ok := a.(uint8)\n\t\tif !ok {\n\t\t\t// failed at type assertion\n\t\t\treturn\n\t\t}\n\t\tif o.dgMacResolved {\n\t\t\t// already resolved, nothing to do\n\t\t\t// shouldn't call OnResolve twice\n\t\t\treturn\n\t\t}\n\t\tresolvedIPv4 := (bitMask & core.RESOLVED_IPV4_DG_MAC) == core.RESOLVED_IPV4_DG_MAC\n\t\tresolvedIPv6 := (bitMask & core.RESOLVED_IPV6_DG_MAC) == core.RESOLVED_IPV6_DG_MAC\n\t\tif (o.isIpv6 && resolvedIPv6) || (!o.isIpv6 && resolvedIPv4) {\n\t\t\to.OnResolve()\n\t\t}\n\t}\n}", "func (c *k8sClient) OnAdd(obj interface{}) {\n\t// if there's an event already in the channel, discard this one,\n\t// this is fine because IngressController always reload everything\n\t// when receiving an event. Same for OnUpdate & OnDelete\n\tselect {\n\tcase c.eventCh <- obj:\n\tdefault:\n\t}\n}", "func (m *MemAdmin) OnEvent(siteID string, ev admin.EventType) error {\n\tresp, ok := m.data[siteID]\n\tif !ok {\n\t\treturn fmt.Errorf(\"site %s not found\", siteID)\n\t}\n\tif ev == admin.EvCreate {\n\t\tresp.CountCreated++ // not a good idea, just for demo\n\t\tm.data[siteID] = resp\n\t}\n\treturn nil\n}", "func (r ResourceEventHandlerFuncs) OnAdd(obj interface{}) {\n\tif r.AddFunc != nil {\n\t\tr.AddFunc(obj)\n\t}\n}", "func (e *EventBus) On(topic string, callback *func(Event)) {\n\te.mu.Lock()\n\tdefer e.mu.Unlock()\n\n\t// Check if this topic has been registered at least once for the event listener, and if\n\t// not create an empty struct for the topic.\n\tif _, exists := e.pools[topic]; !exists {\n\t\te.pools[topic] = &CallbackPool{\n\t\t\tcallbacks: make([]*func(Event), 0),\n\t\t\tpool: workerpool.New(1),\n\t\t}\n\t}\n\n\t// If this callback is not already registered as an event listener, go ahead and append\n\t// it to the array of callbacks for this topic.\n\te.pools[topic].Add(callback)\n}", "func (n *NetImpl) eventHandler(handler interface{}, params ...interface{}) {\n\tcallback := handler.(func(_ *peering.RecvEvent))\n\trecvEvent := params[0].(*peering.RecvEvent)\n\tcallback(recvEvent)\n}", "func (s *LocalSampleTrack) OnBind(f func()) {\n\ts.lock.Lock()\n\ts.onBind = f\n\ts.lock.Unlock()\n}", "func On(hk ent.Hook, op ent.Op) ent.Hook {\n\treturn func(next ent.Mutator) ent.Mutator {\n\t\treturn ent.MutateFunc(func(ctx context.Context, m ent.Mutation) (ent.Value, error) {\n\t\t\tif m.Op().Is(op) {\n\t\t\t\treturn hk(next).Mutate(ctx, m)\n\t\t\t}\n\t\t\treturn next.Mutate(ctx, m)\n\t\t})\n\t}\n}", "func Handler(on func(Event) error) EventHandler {\n\treturn &defaultHandler{\n\t\ton: on,\n\t}\n}", "func (s *WebRTCSender) OnClose(f func()) {\n\ts.onCloseHandler = f\n}", "func (r *jsiiProxy_RepositoryBase) OnEvent(id *string, options *awsevents.OnEventOptions) awsevents.Rule {\n\tvar returns awsevents.Rule\n\n\t_jsii_.Invoke(\n\t\tr,\n\t\t\"onEvent\",\n\t\t[]interface{}{id, options},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (s *Subscription) OnJoin(handler JoinHandler) {\n\ts.events.onJoin = handler\n}", "func (p *pipelineEventCB) onEvents(data []interface{}, acked int) {\n\tp.pushMsg(eventsDataMsg{data: data, total: len(data), acked: acked})\n}", "func (e *EventHandlerFuncs) OnUpdate(table string, old, new Model) {\n\tif e.UpdateFunc != nil {\n\t\te.UpdateFunc(table, old, new)\n\t}\n}", "func (s *Scroll) OnEvent(e interface{}) {\n\tswitch e.(type) {\n\tcase sparta.CloseEvent:\n\t\tif s.closeFn != nil {\n\t\t\ts.closeFn(s, e)\n\t\t}\n\tcase sparta.ConfigureEvent:\n\t\ts.geometry = e.(sparta.ConfigureEvent).Rect\n\t\tif s.configFn != nil {\n\t\t\ts.configFn(s, e)\n\t\t}\n\tcase sparta.CommandEvent:\n\t\tif s.commFn != nil {\n\t\t\tif s.commFn(s, e) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\ts.parent.OnEvent(e)\n\tcase sparta.ExposeEvent:\n\t\tif s.exposeFn != nil {\n\t\t\ts.exposeFn(s, e)\n\t\t}\n\t\ts.win.SetColor(sparta.Foreground, foreColor)\n\t\trect := image.Rect(0, 0, s.geometry.Dx()-1, s.geometry.Dy()-1)\n\t\ts.win.Rectangle(rect, false)\n\t\tif s.size > 0 {\n\t\t\tif s.typ == Vertical {\n\t\t\t\trect.Min.Y = (s.geometry.Dy() * s.pos) / s.size\n\t\t\t\trect.Max.Y = rect.Min.Y + ((s.geometry.Dy() * s.page) / s.size)\n\t\t\t} else {\n\t\t\t\trect.Min.X = (s.geometry.Dx() * s.pos) / s.size\n\t\t\t\trect.Max.X = rect.Min.X + ((s.geometry.Dx() * s.page) / s.size)\n\t\t\t}\n\t\t\ts.win.Rectangle(rect, true)\n\t\t}\n\tcase sparta.KeyEvent:\n\t\tif s.keyFn != nil {\n\t\t\tif s.keyFn(s, e) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tev := e.(sparta.KeyEvent)\n\t\tswitch ev.Key {\n\t\tcase sparta.KeyDown:\n\t\t\tif s.typ == Vertical {\n\t\t\t\ts.SetProperty(ScrollPos, s.pos+1)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase sparta.KeyUp:\n\t\t\tif s.typ == Vertical {\n\t\t\t\ts.SetProperty(ScrollPos, s.pos-1)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase sparta.KeyLeft:\n\t\t\tif s.typ != Vertical {\n\t\t\t\ts.SetProperty(ScrollPos, s.pos-1)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase sparta.KeyRight:\n\t\t\tif s.typ != Vertical {\n\t\t\t\ts.SetProperty(ScrollPos, s.pos+1)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase sparta.KeyPageUp:\n\t\t\tif s.typ == Vertical {\n\t\t\t\ts.SetProperty(ScrollPos, s.pos-s.page)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase sparta.KeyPageDown:\n\t\t\tif s.typ == Vertical {\n\t\t\t\ts.SetProperty(ScrollPos, s.pos+s.page)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase sparta.KeyHome:\n\t\t\ts.SetProperty(ScrollPos, 0)\n\t\t\treturn\n\t\tcase sparta.KeyEnd:\n\t\t\ts.SetProperty(ScrollPos, s.size)\n\t\t\treturn\n\t\t}\n\t\ts.parent.OnEvent(e)\n\tcase sparta.MouseEvent:\n\t\tif s.mouseFn != nil {\n\t\t\tif s.mouseFn(s, e) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tev := e.(sparta.MouseEvent)\n\t\tswitch ev.Button {\n\t\tcase sparta.MouseWheel:\n\t\t\ts.SetProperty(ScrollPos, s.pos-1)\n\t\tcase -sparta.MouseWheel:\n\t\t\ts.SetProperty(ScrollPos, s.pos+1)\n\t\tcase sparta.MouseLeft:\n\t\t\tif s.typ == Vertical {\n\t\t\t\tp := (ev.Loc.Y * s.size) / s.geometry.Dy()\n\t\t\t\ts.SetProperty(ScrollPos, p)\n\t\t\t} else {\n\t\t\t\tp := (ev.Loc.X * s.size) / s.geometry.Dx()\n\t\t\t\ts.SetProperty(ScrollPos, p)\n\t\t\t}\n\t\tcase sparta.MouseRight:\n\t\t\tif s.typ == Vertical {\n\t\t\t\tp := (ev.Loc.Y * s.size) / s.geometry.Dy()\n\t\t\t\ts.SetProperty(ScrollPos, p-s.page)\n\t\t\t} else {\n\t\t\t\tp := (ev.Loc.X * s.size) / s.geometry.Dx()\n\t\t\t\ts.SetProperty(ScrollPos, p-s.page)\n\t\t\t}\n\t\t}\n\t}\n}", "func (a Application) Handler(event json.RawMessage, ctx *apex.Context) (interface{}, error) {\n\tlog.Println(string(event))\n\treq, err := ParseRequest(a.ID, bytes.NewReader(event))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres := NewResponse()\n\n\tswitch req.Type() {\n\tcase \"LaunchRequest\":\n\t\tif a.LaunchHandler != nil {\n\t\t\ta.LaunchHandler(req, res)\n\t\t}\n\tcase \"IntentRequest\":\n\t\tif a.IntentHandler != nil {\n\t\t\ta.IntentHandler(req, res)\n\t\t}\n\tcase \"SessionEndedRequest\":\n\t\tif a.SessionEndedHandler != nil {\n\t\t\ta.SessionEndedHandler(req, res)\n\t\t}\n\tdefault:\n\t\treturn nil, errors.New(\"invalid request type\")\n\t}\n\n\treturn res, nil\n}", "func (b *FontButton) OnChanged(f func(*FontButton)) {\n\tb.onChanged = f\n}", "func (e *EventHandler) OnAdd(obj interface{}) {\n\tc := event.CreateEvent{}\n\n\t// Pull Object out of the object\n\tif o, ok := obj.(client.Object); ok {\n\t\tc.Object = o\n\t} else {\n\t\tlog.Error(nil, \"OnAdd missing Object\",\n\t\t\t\"object\", obj, \"type\", fmt.Sprintf(\"%T\", obj))\n\t\treturn\n\t}\n\n\tfor _, p := range e.predicates {\n\t\tif !p.Create(c) {\n\t\t\treturn\n\t\t}\n\t}\n\n\t// Invoke create handler\n\tctx, cancel := context.WithCancel(e.ctx)\n\tdefer cancel()\n\te.handler.Create(ctx, c, e.queue)\n}", "func (o *IPFixTimerCallback) OnEvent(a, b interface{}) {\n\t// a should be a pointer to the client plugin\n\tipfixPlug := a.(*PluginIPFixClient)\n\t// Get the time now.\n\ttimeNow := currentTime()\n\t// Calculate the uptime\n\tipfixPlug.sysUpTime = uint32(timeNow.Sub(ipfixPlug.sysStartTime).Milliseconds())\n\t// Calculate the unix time\n\tipfixPlug.unixUtcTimeNow = timeNow.UTC().Unix()\n\t// Restart call\n\tipfixPlug.timerw.StartTicks(&ipfixPlug.timer, 1)\n}", "func (s *SimpleClickHandler) OnClick(handler func(bar.Event)) {\n\tif handler == nil {\n\t\thandler = func(e bar.Event) {}\n\t}\n\ts.Store(handler)\n}", "func (g *Gui) onEvent(ev tcell.Event) error {\n\tswitch tev := ev.(type) {\n\tcase *tcell.EventKey:\n\t\tmatched, err := g.execEventBindings(g.currentView, ev)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif matched {\n\t\t\tbreak\n\t\t}\n\t\tif g.currentView != nil && g.currentView.Editable && g.currentView.Editor != nil {\n\t\t\tg.currentView.Editor.Edit(g.currentView, tev.Key(), tev.Rune(), tev.Modifiers())\n\t\t}\n\tcase *tcell.EventMouse:\n\t\tv, _, _, err := g.GetViewRelativeMousePosition(tev)\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t\t// If the key-binding wants to move the cursor, it should call SetCursorFromCurrentMouseEvent()\n\t\t// Not all mouse events will want to do this (eg: scroll wheel)\n\t\tg.CurrentEvent = ev\n\t\tif _, err := g.execEventBindings(v, g.CurrentEvent); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (c *Client) OnPing(cb PingFunc) {\n\tc.onPingListeners = append(c.onPingListeners, cb)\n}", "func (b *Button) OnClick(fn func(Event)) {\n\tb.onClick = fn\n}", "func (*listener) OnConnect() {}", "func (e *EventHandlerFuncs) OnDelete(table string, row Model) {\n\tif e.DeleteFunc != nil {\n\t\te.DeleteFunc(table, row)\n\t}\n}", "func HandleJoin(h func(*Server, string, string)) {\n\teventOnJoin = h\n}", "func (c *k8sClient) OnDelete(obj interface{}) {\n\tselect {\n\tcase c.eventCh <- obj:\n\tdefault:\n\t}\n}", "func (e *Input) OnChanged(fn func(Input *Input)) {\n\te.onTextChange = fn\n}", "func onEventCallback(e event.Event, ctx interface{}) {\n\tservice := ctx.(*qutoService)\n\tservice.eventChan <- e\n}", "func EventHandler(w http.ResponseWriter, r *http.Request) {\n\t// Read the request body\n\tpayload, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer r.Body.Close()\n\n\t// Parse the WebHook payload using go-github\n\tevent, err := github.ParseWebHook(github.WebHookType(r), payload)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tswitch event := event.(type) {\n\tcase *github.CheckSuiteEvent:\n\t\tif action := *event.Action; action != \"requested\" {\n\t\t\treturn\n\t\t}\n\n\t\t// Handle the check suite for each registered job\n\t\tfor _, job := range configuration.Configuration.Jobs {\n\t\t\thandleCheckSuite(*event, job)\n\t\t}\n\t}\n}", "func (tw *TestWatch) OnEvent(operation func()) {\n\tif tw.IsClosed() {\n\t\ttw.t.Fatal(\"The watch has been closed\")\n\t\treturn\n\t}\n\n\t// wait until some goroutine calls Event\n\tif event := <-tw.events; event != nil {\n\t\t// closing the event signals a real Watch sending an event\n\t\tclose(event)\n\n\t\t// invoke the operation after the event is triggered\n\t\toperation()\n\t}\n}", "func (he *HTTPErrors) On(httpStatus int, handler HandlerFunc) {\n\tif httpStatus == http.StatusOK {\n\t\treturn\n\t}\n\n\t/*\thttpHandlerOfficialType := reflect.TypeOf((*http.Handler)(nil)).Elem()\n\t\tif !reflect.TypeOf(handler).Implements(httpHandlerOfficialType) {\n\t\t\t//it is not a http.Handler\n\t\t\t//it is func(res,req) we will convert it to a handler using http.HandlerFunc\n\t\t\thandler = ToHandlerFunc(handler.(func(res http.ResponseWriter, req *http.Request)))\n\t\t}\n\t*/\n\tif errH := he.getByCode(httpStatus); errH != nil {\n\t\terrH.handler = handler\n\t} else {\n\t\the.ErrorHanders = append(he.ErrorHanders, &errorHandler{code: httpStatus, handler: handler})\n\t}\n\n}", "func (bas *BaseService) OnRequest(ctx context.Context, args Args) {}", "func On(lc ListenerConfig) (err error) {\n\tif len(lc.Topic) == 0 {\n\t\terr = ErrTopicRequired\n\t\treturn\n\t}\n\n\tif len(lc.Channel) == 0 {\n\t\terr = ErrChannelRequired\n\t\treturn\n\t}\n\n\tif len(lc.Lookup) == 0 {\n\t\tlc.Lookup = []string{\"localhost:4161\"}\n\t}\n\n\tif lc.HandlerConcurrency == 0 {\n\t\tlc.HandlerConcurrency = 1\n\t}\n\n\tconfig := newListenerConfig(lc)\n\tconsumer, err := nsq.NewConsumer(lc.Topic, lc.Channel, config)\n\tif err != nil {\n\t\treturn\n\t}\n\n\thandler := handleMessage(lc)\n\tconsumer.AddConcurrentHandlers(handler, lc.HandlerConcurrency)\n\terr = consumer.ConnectToNSQLookupds(lc.Lookup)\n\n\treturn\n}", "func (q *queue) OnPush(handler PushHandler) {\n\tq.Lock()\n\tdefer q.Unlock()\n\n\tq.pushHandler = handler\n}", "func (a *AbstractHandler) OnReady() error {\n\treturn nil\n}", "func (outlet *CollectorOutleter) OnEvent(data *util.Data) bool {\n\t// We'll actually receive a few spurious blank events that FileBeat likes to use to keep its registry\n\t// of file offsets up-to-date. We're really only interested in events that have messages, and we're really\n\t// only concerned with the messages themselves. FileBeat creates the events, typically, in the harvester.\n\t// To see the generation of these events look at log.harverster's Run method.\n\tevent := data.GetEvent()\n\tif event.Fields != nil {\n\t\t// We only want to send over events that actually have message fields (which should actually be all\n\t\t// of them, but just in case). So this is just Go's way of saying \"if map event.Fields has a key\n\t\t// 'message' (while also storing the value at that key to 'msg')\"\n\t\tif msg, ok := event.Fields[\"message\"]; ok {\n\t\t\t// \"msg\" at this stage is just a generic interface{}, which is kind of the closest Go has to\n\t\t\t// a void pointer. We want to try to cast it to a string (which it always should be) before sending\n\t\t\t// it down the wire.\n\t\t\tif str, ok := msg.(string); ok {\n\t\t\t\t// Send the line over our channel\n\t\t\t\toutlet.lines <- str\n\t\t\t} else {\n\t\t\t\tlogp.Warn(\"Encountered non string message field: %s\", msg)\n\t\t\t}\n\t\t}\n\t}\n\n\t// The boolean we return indicates whether we were able to enqueue the data or not. For our purposes,\n\t// since we're not actually using a complicated Spool feature like FileBeat, we can just say we were\n\t// able to.\n\treturn true\n}", "func (handler KaffeHandler) OnEvent(event flowdock.Event) flowdock.Message {\n\n\tif (event.Event == flowdock.EventMessage) && (strings.Contains(event.Content, KAFFE_COMMAND)) {\n\n\t\tmessage := &flowdock.MessageFile{Event: flowdock.EventFile}\n\t\tmessage.ThreadId = event.ThreadId\n\t\tmessage.Content.ContentType = \"image/jpeg\"\n\t\tmessage.Content.FileName = \"kahvi-status.jpg\"\n\t\tmessage.Content.Data = readFileToBase64String(KAFFEFILE)\n\n\t\tif message.Content.Data == \"\" {\n\t\t\treturn nil\n\t\t}\n\n\t\treturn message\n\t}\n\n\treturn nil\n}", "func (g *Group) On() error {\n\treturn g.OnContext(context.Background())\n}", "func (e *NotificationEvents) OnPlaying(fn func(n NotificationContainer)) {\n\te.events[\"playing\"] = fn\n}", "func (m Method) On(on string) *jen.Statement {\n\treturn jen.Id(on).Dot(m.function.name)\n}", "func (conn *Conn) OnMessage(f func(string, []byte)) {\n\tconn.onMessage = append(conn.onMessage, f)\n}", "func onEventCallback(e event.Event, ctx interface{}) {\n\tservice := ctx.(*metadataService)\n\tservice.eventChan <- e\n}", "func (c *Client) OnMessage(handler MessageHandler) {\n\tproxy := &eventProxy{client: c, onMessage: handler}\n\tc.client.OnMessage(proxy)\n}", "func (s *EmptyScene) HandleEvents(e sdl.Event) {\n}", "func (s Broker) On(name string) {\n\ts.Gauge(name, 1)\n}", "func (player *Player) OnBookEdited(handler func(event *event.BookEdited)) {\n\tplayer.on(event.NameBookEdited, func(e interface{}) {\n\t\thandler(e.(*event.BookEdited))\n\t})\n}", "func (s *BasecluListener) EnterWhen_handler(ctx *When_handlerContext) {}", "func (n *NullEventReceiver) Event(eventName string) {\n}", "func (eh *EventHandler) OnClick(mouseButton MouseButton, callback func()) *EventHandler {\n\teh.mouseEvents = append(eh.mouseEvents, mouseEvent{mouseButton, callback, IsMouseClicked})\n\treturn eh\n}", "func (self *PhysicsP2) ImpactHandler(event interface{}) {\n self.Object.Call(\"impactHandler\", event)\n}", "func (e *Input) OnSubmit(fn func(Input *Input)) {\n\te.onSubmit = fn\n}", "func (h *Hooks) OnName(handler ...OnNameHandler) {\n\th.app.mutex.Lock()\n\th.onName = append(h.onName, handler...)\n\th.app.mutex.Unlock()\n}", "func (s *WebRTCSender) OnCloseHandler(fn func()) {\n\ts.onCloseHandler = fn\n}", "func (w *BaseWebsocketClient) OnOpen() {}", "func (c *Client) OnEvent(name string) <-chan *Message {\n\treturn c.subs.subscribe(name)\n}", "func HandlePart(h func(*Server, string, string, string)) {\n\teventOnPart = h\n}", "func (p *EventEmitter) On(namaEvents ...string) chan interface{} {\n\t//init new chan using muatan as channel buffer\n\tworkChan := make(chan interface{}, p.muatan)\n\n\tvar updateChanList []chan interface{}\n\tfor _, namaEvent := range namaEvents {\n\t\tupdateChanList, _ = p.mapClients[namaEvent]\n\t\tupdateChanList = append(updateChanList, workChan)\n\t\tp.mapClients[namaEvent] = updateChanList\n\t}\n\tp.mapEvents[workChan] = namaEvents\n\n\treturn workChan\n}", "func (e *entry) OnEnter() {\n\te.button.OnEnter()\n}", "func mentionEvents() {\n\tev.On <- ev.EventHandler{\n\t\tOn: ev.NEW_MENTION,\n\t\tHandler: func(e ev.Event) error {\n\t\t\tuserID := e.Params[\"user_id\"].(bson.ObjectId)\n\t\t\trelatedID := e.Params[\"related_id\"].(bson.ObjectId)\n\t\t\tusers := e.Params[\"users\"].([]bson.ObjectId)\n\t\t\trelated := e.Params[\"related\"].(string)\n\n\t\t\t// Create notification\n\t\t\tif related == \"comment\" {\n\t\t\t\tnotify.Database <- notify.Notification{\n\t\t\t\t\tUserId: userID,\n\t\t\t\t\tType: \"mention\",\n\t\t\t\t\tRelatedId: relatedID,\n\t\t\t\t\tUsers: users,\n\t\t\t\t}\n\t\t\t}\n\t\t\tif related == \"chat\" {\n\t\t\t\tnotify.Database <- notify.Notification{\n\t\t\t\t\tUserId: userID,\n\t\t\t\t\tType: \"chat\",\n\t\t\t\t\tRelatedId: relatedID,\n\t\t\t\t\tUsers: users,\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn nil\n\t\t},\n\t}\n}", "func (c *Controller) OnUpdate(oldObj, newObj interface{}) {\n\tif err := c.closeAbsentPorts(oldObj, newObj); err != nil {\n\t\tlog.Errorf(\"OnUpdate failed: %v\", err)\n\t}\n\tif err := c.ensureOpened(newObj); err != nil {\n\t\tlog.Errorf(\"OnUpdate failed: %v\", err)\n\t}\n}", "func OnStart(ecb EventCallbackFunc, priority ...int) {\n\tAppEventStore().Subscribe(EventOnStart, EventCallback{\n\t\tCallback: ecb,\n\t\tCallOnce: true,\n\t\tpriority: parsePriority(priority...),\n\t})\n}", "func (s *HTTPServer) On(pattern string) *Call {\n\ts.guard.Lock()\n\tdefer s.guard.Unlock()\n\th, ok := s.handlers[pattern]\n\tif !ok {\n\t\th = &handler{\n\t\t\tpattern: pattern,\n\t\t}\n\t\ts.handlers[pattern] = h\n\t\t// register http handler\n\t\ts.mux.Handle(pattern, h)\n\t}\n\tc := &Call{}\n\th.calls = append(h.calls, c)\n\treturn c\n}" ]
[ "0.753045", "0.7305682", "0.6904258", "0.6867544", "0.6787099", "0.6748967", "0.6697568", "0.6642001", "0.6630415", "0.6616909", "0.66079646", "0.6538392", "0.6523674", "0.6486402", "0.64701205", "0.63886946", "0.62199557", "0.62079185", "0.614324", "0.61256766", "0.6095047", "0.60948473", "0.6031322", "0.6023801", "0.6013099", "0.59937984", "0.5962725", "0.592939", "0.59203035", "0.58824927", "0.5866419", "0.5831798", "0.5804825", "0.5802217", "0.57975584", "0.5796096", "0.57906175", "0.5777767", "0.5771977", "0.5740085", "0.57396847", "0.57117367", "0.570067", "0.56961113", "0.56554234", "0.5654442", "0.5645227", "0.56450987", "0.5635348", "0.5632538", "0.56256306", "0.56254673", "0.5625214", "0.5622958", "0.561456", "0.56002235", "0.55976254", "0.5590099", "0.55873805", "0.5584531", "0.55841166", "0.5571994", "0.5559666", "0.55482125", "0.55467564", "0.55429864", "0.5538362", "0.55379945", "0.5526861", "0.551543", "0.5507837", "0.5491647", "0.548866", "0.5477126", "0.54744136", "0.5421211", "0.54199153", "0.54148793", "0.5404722", "0.53789496", "0.5376168", "0.5374101", "0.5362761", "0.5351465", "0.53455496", "0.5338783", "0.53284323", "0.5321369", "0.53130054", "0.5311635", "0.53038996", "0.5302622", "0.53025264", "0.5301956", "0.5294565", "0.52938753", "0.52901006", "0.5288592", "0.5279197", "0.527006" ]
0.72935206
2
mark HTML as Output type
func (*HTML) isOutput() { }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (this *Tidy) OutputHtml(val bool) (bool, error) {\n\treturn this.optSetBool(C.TidyHtmlOut, cBool(val))\n}", "func (this *Tidy) OutputXhtml(val bool) (bool, error) {\n\treturn this.optSetBool(C.TidyXhtmlOut, cBool(val))\n}", "func htmlFmt(w io.Writer, x interface{}, format string) {\n\twriteAny(w, x, true)\n}", "func HTMLContentTypeMiddleware(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tw.Header().Set(\"Content-Type\", \"text/html; charset=UTF-8\")\n\t\tnext.ServeHTTP(w, r)\n\t})\n}", "func Html(resp http.ResponseWriter, content string, code int) error {\n\tresp.Header().Add(\"Content-Type\", \"text/html\")\n\tresp.WriteHeader(code)\n\t_, err := resp.Write([]byte(content))\n\treturn maskAny(err)\n}", "func (Output) typ() string { return \"output1\" }", "func IfReturnHTMLResponse(w http.ResponseWriter, r *http.Request) bool {\n\taccepts := r.Header[\"Accept\"]\n\tfor _, accept := range accepts {\n\t\tfields := strings.Split(accept, \",\")\n\t\tfor _, field := range fields {\n\t\t\tif field == contentTypeHtml {\n\t\t\t\tw.Header().Set(\"Content-Type\", contentTypeHtml)\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\n\treturn false\n}", "func (this *Tidy) OutputEncoding(val int) (bool, error) {\n\tswitch val {\n\tcase Raw, Ascii, Latin0, Latin1, Utf8, Iso2022, Mac, Win1252, Ibm858, Utf16le, Utf16be, Utf16, Big5, Shiftjis:\n\t\treturn this.optSetInt(C.TidyOutCharEncoding, (C.ulong)(val))\n\t}\n\treturn false, errors.New(\"Argument val int is out of range (0-13)\")\n}", "func Output(c Content, o Options) (string, error) {\n\tvar e error\n\tvar formatter format\n\tswitch style := o.Style; style {\n\tcase \"md\":\n\t\tformatter = md{Content: c, Options: o}\n\tcase \"psql\":\n\t\tformatter = psql{Content: c, Options: o}\n\tcase \"jira\":\n\t\tformatter = jira{Content: c, Options: o}\n\tdefault:\n\t\te = errors.New(\"Formatter not implemented error\")\n\t}\n\ttable := formatter.format()\n\treturn table, e\n}", "func (self *Output) Type() string {\n\treturn \"output\"\n}", "func (t FieldType) ToHTML() []byte {\n\treturn nil\n}", "func (self *Encoder) SetEscapeHTML(f bool) {\n if f {\n self.Opts |= EscapeHTML\n } else {\n self.Opts &= ^EscapeHTML\n }\n}", "func (ra *ResponseAsserter) HTML() *HTMLAsserter {\n\t// @TODO do some basic html validation checking\n\treturn newHTMLAsserter(ra, ra.fail)\n}", "func (p Page) IsHTML() bool {\n\treturn p.Type().MediaType() == \"text/html\"\n}", "func Output_(children ...HTML) HTML {\n return Output(nil, children...)\n}", "func writeAny(w io.Writer, x interface{}, html bool) {\n\tswitch v := x.(type) {\n\tcase []byte:\n\t\twriteText(w, v, html)\n\tcase string:\n\t\twriteText(w, strings.Bytes(v), html)\n\tcase ast.Decl:\n\t\twriteNode(w, v, html, &defaultStyler)\n\tcase ast.Expr:\n\t\twriteNode(w, v, html, &defaultStyler)\n\tdefault:\n\t\tif html {\n\t\t\tvar buf bytes.Buffer;\n\t\t\tfmt.Fprint(&buf, x);\n\t\t\twriteText(w, buf.Bytes(), true);\n\t\t} else {\n\t\t\tfmt.Fprint(w, x)\n\t\t}\n\t}\n}", "func (p *Parser) SetOutput(out io.Writer) { p.out = out }", "func outHTML(config *MainConfig, fileFunc FileResultFunc) {\n\n\tindexPath := filepath.Join(config.Outpath, FILE_NAME_HTML_INDEX)\n\terr := SFFileManager.WirteFilepath(indexPath, []byte(assets.HTML_INDEX))\n\n\tif nil != err {\n\t\tfileFunc(indexPath, ResultFileOutFail, err)\n\t} else {\n\t\tfileFunc(indexPath, ResultFileSuccess, nil)\n\t}\n\n\tsrcPath := filepath.Join(config.Outpath, FILE_NAME_HTML_SRC)\n\terr = SFFileManager.WirteFilepath(srcPath, []byte(assets.HTML_SRC))\n\n\tif nil != err {\n\t\tfileFunc(srcPath, ResultFileOutFail, err)\n\t} else {\n\t\tfileFunc(srcPath, ResultFileSuccess, nil)\n\t}\n\n}", "func WriteHTML(w http.ResponseWriter, data []byte, status int) (int, error) {\n\tw.Header().Set(ContentType, ContentHTML)\n\tw.WriteHeader(status)\n\treturn w.Write(data)\n}", "func Output(attrs []htmlgo.Attribute, children ...HTML) HTML {\n\treturn &htmlgo.Tree{Tag: \"output\", Attributes: attrs, Children: children}\n}", "func outputHTML(w http.ResponseWriter, r *http.Request, filepath string) {\n\tfile, err := os.Open(filepath)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tdefer file.Close()\n\n\thttp.ServeContent(w, r, file.Name(), time.Now(), file)\n}", "func Render(typ Type, input any, urlPrefix string, metas map[string]string) []byte {\n\tvar rawBytes []byte\n\tswitch v := input.(type) {\n\tcase []byte:\n\t\trawBytes = v\n\tcase string:\n\t\trawBytes = []byte(v)\n\tdefault:\n\t\tpanic(fmt.Sprintf(\"unrecognized input content type: %T\", input))\n\t}\n\n\turlPrefix = strings.TrimRight(strings.ReplaceAll(urlPrefix, \" \", \"%20\"), \"/\")\n\tvar rawHTML []byte\n\tswitch typ {\n\tcase TypeMarkdown:\n\t\trawHTML = RawMarkdown(rawBytes, urlPrefix)\n\tcase TypeOrgMode:\n\t\trawHTML = RawOrgMode(rawBytes, urlPrefix)\n\tdefault:\n\t\treturn rawBytes // Do nothing if syntax type is not recognized\n\t}\n\n\trawHTML = postProcessHTML(rawHTML, urlPrefix, metas)\n\treturn SanitizeBytes(rawHTML)\n}", "func (ctx *Context) HTML(code int, html string) (err error) {\n\tctx.response.Header().Set(\"Content-Type\", \"text/html; charset=utf-8\")\n\tctx.response.WriteHeader(code)\n\t_, err = ctx.response.Write([]byte(html))\n\treturn\n}", "func (r Redirect) WriteContentType(http.ResponseWriter) {}", "func SetHandler(w io.Writer) {\n\toutput.output = w\n}", "func (t *typewriter) Typewrite() string {\n\t// Re-use the cached result if already processed.\n\tif t.cur != -1 {\n\t\treturn t.result\n\t}\n\n\tvar buf bytes.Buffer\n\n\tfor {\n\t\tsep, str := t.scanMorpheme()\n\t\tif str == \"\" {\n\t\t\tbreak\n\t\t}\n\n\t\tbuf.WriteString(sep)\n\t\tbuf.WriteString(str)\n\t}\n\n\tt.result = buf.String()\n\treturn t.result\n}", "func HTML(w http.ResponseWriter, name string, data interface{}) {\n\tDefault.HTML(w, name, data)\n}", "func renderHTMLPage(w http.ResponseWriter, r *http.Request, content []byte) {\n w.Header().Set(\"Content-Type\", \"text/html\")\n w.Write(content)\n}", "func (r renderer) BlockHtml(out *bytes.Buffer, text []byte) {}", "func (c *Context) HTML(code int, html string) {\n\tc.SetHeader(\"Content-Type\", \"text/html\")\n\tc.Status(code)\n\tc.Writer.Write([]byte(html))\n}", "func (h *Encoder) SetupOutput(output interface{}, ht *rest.HandlerTrait) {\n\th.outputBufferType = reflect.TypeOf(output)\n\th.outputHeadersEncoder = nil\n\th.skipRendering = true\n\n\tif output == nil {\n\t\treturn\n\t}\n\n\t// Enable dynamic headers check in interface mode.\n\tif h.unwrapInterface = reflect.ValueOf(output).Elem().Kind() == reflect.Interface; h.unwrapInterface {\n\t\tenc := form.NewEncoder()\n\t\tenc.SetMode(form.ModeExplicit)\n\t\tenc.SetTagName(string(rest.ParamInHeader))\n\n\t\th.outputHeadersEncoder = enc\n\t}\n\n\trespHeaderMapping := ht.RespHeaderMapping\n\tif len(respHeaderMapping) == 0 && refl.HasTaggedFields(output, string(rest.ParamInHeader)) {\n\t\trespHeaderMapping = make(map[string]string)\n\n\t\trefl.WalkTaggedFields(reflect.ValueOf(output), func(v reflect.Value, sf reflect.StructField, tag string) {\n\t\t\trespHeaderMapping[sf.Name] = tag\n\t\t}, string(rest.ParamInHeader))\n\t}\n\n\tif len(respHeaderMapping) > 0 {\n\t\tenc := form.NewEncoder()\n\t\tenc.SetMode(form.ModeExplicit)\n\t\tenc.RegisterTagNameFunc(func(field reflect.StructField) string {\n\t\t\treturn respHeaderMapping[field.Name]\n\t\t})\n\n\t\th.outputHeadersEncoder = enc\n\t}\n\n\tif h.outputBufferType.Kind() == reflect.Ptr {\n\t\th.outputBufferType = h.outputBufferType.Elem()\n\t}\n\n\tif !rest.OutputHasNoContent(output) {\n\t\th.skipRendering = false\n\t}\n\n\tif _, ok := output.(usecase.OutputWithWriter); ok {\n\t\th.skipRendering = true\n\t\th.outputWithWriter = true\n\t}\n\n\tif ht.SuccessStatus != 0 {\n\t\treturn\n\t}\n\n\tif h.skipRendering && !h.outputWithWriter {\n\t\tht.SuccessStatus = http.StatusNoContent\n\t} else {\n\t\tht.SuccessStatus = http.StatusOK\n\t}\n}", "func (c *C) HTML(name string, data interface{}) {\n\tc.render.Render(c.Writer, name, data)\n}", "func (f *FakeOutput) Type() string { return \"fake_output\" }", "func (e *htmlTag) WriteTo(w io.Writer) (int64, error) {\n\tvar bf bytes.Buffer\n\n\t// Write an open tag.\n\tbf.WriteString(lt)\n\tbf.WriteString(e.tagName)\n\t// Write an id.\n\tif e.id != \"\" {\n\t\tbf.WriteString(space)\n\t\tbf.WriteString(attributeNameID)\n\t\tbf.WriteString(equal)\n\t\tbf.WriteString(doubleQuote)\n\t\tbf.WriteString(e.id)\n\t\tbf.WriteString(doubleQuote)\n\t}\n\t// Write classes.\n\tif len(e.classes) > 0 {\n\t\tbf.WriteString(space)\n\t\tbf.WriteString(attributeNameClass)\n\t\tbf.WriteString(equal)\n\t\tbf.WriteString(doubleQuote)\n\t\tfor i, class := range e.classes {\n\t\t\tif i > 0 {\n\t\t\t\tbf.WriteString(space)\n\t\t\t}\n\t\t\tbf.WriteString(class)\n\t\t}\n\t\tbf.WriteString(doubleQuote)\n\t}\n\t// Write attributes.\n\tif len(e.attributes) > 0 {\n\n\t\tfor _, a := range e.attributes {\n\t\t\tbf.WriteString(space)\n\t\t\tbf.WriteString(a.key)\n\t\t\tif a.value != \"\" {\n\t\t\t\tbf.WriteString(equal)\n\t\t\t\tbf.WriteString(doubleQuote)\n\t\t\t\tbf.WriteString(a.value)\n\t\t\t\tbf.WriteString(doubleQuote)\n\t\t\t}\n\t\t}\n\t}\n\tbf.WriteString(gt)\n\n\t// Write a text value\n\tif e.textValue != \"\" {\n\t\tbf.WriteString(e.textValue)\n\t}\n\n\tif e.containPlainText {\n\t\tbf.WriteString(lf)\n\t}\n\n\t// Write children's HTML.\n\tif i, err := e.writeChildren(&bf); err != nil {\n\t\treturn i, err\n\t}\n\n\t// Write a close tag.\n\tif !e.noCloseTag() {\n\t\tbf.WriteString(lt)\n\t\tbf.WriteString(slash)\n\t\tbf.WriteString(e.tagName)\n\t\tbf.WriteString(gt)\n\t}\n\n\t// Write the buffer.\n\ti, err := w.Write(bf.Bytes())\n\n\treturn int64(i), err\n}", "func renderToHTML(b []byte) ([]byte, error) {\n\tb = convertImports(b)\n\tgmParser := goldmark.New(goldmark.WithRendererOptions(gmhtml.WithUnsafe()), goldmark.WithExtensions(extension.Typographer, extension.Table))\n\tvar out bytes.Buffer\n\tif err := gmParser.Convert(b, &out); err != nil {\n\t\tpanic(err)\n\t}\n\treturn out.Bytes(), nil\n}", "func (f *Template) SetOutput(dest io.Writer) {\n\tf.ac, _ = dest.(*AccessLog)\n\n\tif f.Tmpl == nil {\n\t\ttmpl := template.New(\"\")\n\n\t\ttext := f.Text\n\t\tif text != \"\" {\n\t\t\ttmpl.Funcs(f.Funcs)\n\t\t} else {\n\t\t\ttext = defaultTmplText\n\t\t}\n\n\t\tf.Tmpl = template.Must(tmpl.Parse(text))\n\t}\n}", "func (r renderer) RawHtmlTag(out *bytes.Buffer, tag []byte) {}", "func (t toc) writeToHTML(f string) (err error) {\n\ts := t.toHTMLStr()\n\treturn ioutil.WriteFile(f, []byte(s), 0755)\n}", "func (r *Template) Html() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"html\"])\n}", "func (p *ChangePasswordStep2Page) SetHTML(html template.HTML) {\n\tp.HTML = html\n}", "func httpHTML(w http.ResponseWriter, data interface{}) {\n\tw.WriteHeader(http.StatusOK)\n\tfmt.Fprint(w, data)\n}", "func RenderHTML(w http.ResponseWriter, view Renderable, code int) error {\n\tb, err := view.Render(view)\n\tif err != nil {\n\t\treturn err\n\t}\n\tw.Header().Set(\"Content-Type\", \"text/html; charset=utf-8\")\n\tw.WriteHeader(code)\n\t_, err = w.Write(b)\n\treturn err\n}", "func (logger *Logger) SetOutput(level string, w io.Writer, flag int) {\n\tlevelNum := LevelInt(level)\n\tswitch {\n\tcase INFO == levelNum:\n\t\tlogger.info = nativeLogger(w, level, flag)\n\tcase DEBUG == levelNum:\n\t\tlogger.debug = nativeLogger(w, level, flag)\n\tcase WARN == levelNum:\n\t\tlogger.warn = nativeLogger(w, level, flag)\n\tcase ERROR == levelNum:\n\t\tlogger.err = nativeLogger(w, level, flag)\n\tcase FATAL == levelNum:\n\t\tlogger.fatal = nativeLogger(w, level, flag)\n\tdefault:\n\t}\n}", "func Fmt (output *Data, data webapp.ReqData) {\n\toutput.Type = data[\"Content-Type\"]\n}", "func (rw *RW) WriteHTML(fileName string) (err error) {\n\tfile ,err := os.Open(\"views/\"+fileName)\n\tif err != nil {\n\t\treturn\n\t}\n\tdefer file.Close()\n\tfileBytes ,err := ioutil.ReadAll(file)\n\tif err != nil {\n\t\treturn\n\t}\n\n\trw.W.WriteHeader(200)\n\t_,err = rw.W.Write(fileBytes)\n\treturn\n}", "func HTMLHandler(data string, status int) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tMustWriteHTML(w, []byte(data), status)\n\t})\n}", "func postProcessHTML(rawHTML []byte, urlPrefix string, metas map[string]string) []byte {\n\tstartTags := make([]string, 0, 5)\n\tbuf := bytes.NewBuffer(nil)\n\ttokenizer := html.NewTokenizer(bytes.NewReader(rawHTML))\n\nouterLoop:\n\tfor html.ErrorToken != tokenizer.Next() {\n\t\ttoken := tokenizer.Token()\n\t\tswitch token.Type {\n\t\tcase html.TextToken:\n\t\t\tbuf.Write(RenderSpecialLink([]byte(token.String()), urlPrefix, metas))\n\n\t\tcase html.StartTagToken:\n\t\t\ttagName := token.Data\n\n\t\t\tif tagName == \"img\" {\n\t\t\t\twrapImgWithLink(urlPrefix, buf, token)\n\t\t\t\tcontinue outerLoop\n\t\t\t}\n\n\t\t\tbuf.WriteString(token.String())\n\t\t\t// If this is an excluded tag, we skip processing all output until a close tag is encountered.\n\t\t\tif strings.EqualFold(\"a\", tagName) || strings.EqualFold(\"code\", tagName) || strings.EqualFold(\"pre\", tagName) {\n\t\t\t\tstackNum := 1\n\t\t\t\tfor html.ErrorToken != tokenizer.Next() {\n\t\t\t\t\ttoken = tokenizer.Token()\n\n\t\t\t\t\t// Copy the token to the output verbatim\n\t\t\t\t\tbuf.WriteString(token.String())\n\n\t\t\t\t\t// Stack number doesn't increase for tags without end tags.\n\t\t\t\t\tif token.Type == html.StartTagToken && !com.IsSliceContainsStr(noEndTags, token.Data) {\n\t\t\t\t\t\tstackNum++\n\t\t\t\t\t}\n\n\t\t\t\t\t// If this is the close tag to the outer-most, we are done\n\t\t\t\t\tif token.Type == html.EndTagToken {\n\t\t\t\t\t\tstackNum--\n\t\t\t\t\t\tif stackNum <= 0 && strings.EqualFold(tagName, token.Data) {\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tcontinue outerLoop\n\t\t\t}\n\n\t\t\tif !com.IsSliceContainsStr(noEndTags, tagName) {\n\t\t\t\tstartTags = append(startTags, tagName)\n\t\t\t}\n\n\t\tcase html.EndTagToken:\n\t\t\tif len(startTags) == 0 {\n\t\t\t\tbuf.WriteString(token.String())\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tbuf.Write(leftAngleBracket)\n\t\t\tbuf.WriteString(startTags[len(startTags)-1])\n\t\t\tbuf.Write(rightAngleBracket)\n\t\t\tstartTags = startTags[:len(startTags)-1]\n\t\tdefault:\n\t\t\tbuf.WriteString(token.String())\n\t\t}\n\t}\n\n\tif io.EOF == tokenizer.Err() {\n\t\treturn buf.Bytes()\n\t}\n\n\t// If we are not at the end of the input, then some other parsing error has occurred,\n\t// so return the input verbatim.\n\treturn rawHTML\n}", "func (ctx *Context) HTMLByte(code int, body []byte) {\n\tctx.SetContentType(\"text/html; charset=utf-8\")\n\tctx.SetStatusCode(code)\n\tctx.Response.Body = ioutil.NopCloser(bytes.NewBuffer(body))\n}", "func (c *Context) HTML(code int, name string, data interface{}) {\n\tc.SetHeader(\"Content-Type\", \"text/html\")\n\tc.Status(code)\n\tif err := c.engine.htmlTemplates.ExecuteTemplate(c.Writer, name, data); err != nil {\n\t\tc.Fail(500, err.Error())\n\t}\n}", "func (ctx *Context) HTML(code int, body string) {\n\tctx.SetContentType(\"text/html; charset=utf-8\")\n\tctx.SetStatusCode(code)\n\tctx.WriteString(body)\n}", "func (p *RegistrationPage) SetHTML(html template.HTML) {\n\tp.HTML = html\n}", "func (t tag) Render() string {\n return t.render()\n}", "func _html(ns Nodes, outer bool) string {\n\tif len(ns) == 0 {\n\t\treturn \"\"\n\t}\n\twr := w{}\n\tif outer {\n\t\thtml.Render(&wr, ns[0].Node)\n\t} else {\n\t\tfor _, v := range ns[0].Node.Child {\n\t\t\thtml.Render(&wr, v)\n\t\t}\n\t}\n\treturn wr.s\n}", "func (o AccessCustomPageOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *AccessCustomPage) pulumi.StringOutput { return v.Type }).(pulumi.StringOutput)\n}", "func (*backfiller) OutputTypes() []*types.T {\n\t// No output types.\n\treturn nil\n}", "func (h *Encoder) MakeOutput(w http.ResponseWriter, ht rest.HandlerTrait) interface{} {\n\tif h.outputBufferType == nil {\n\t\treturn nil\n\t}\n\n\toutput := reflect.New(h.outputBufferType).Interface()\n\n\tif h.outputWithWriter {\n\t\tif withWriter, ok := output.(usecase.OutputWithWriter); ok {\n\t\t\tif h.outputHeadersEncoder != nil || ht.SuccessContentType != \"\" {\n\t\t\t\twithWriter.SetWriter(&writerWithHeaders{\n\t\t\t\t\tResponseWriter: w,\n\t\t\t\t\tresponseWriter: h,\n\t\t\t\t\ttrait: ht,\n\t\t\t\t\toutput: output,\n\t\t\t\t})\n\t\t\t} else {\n\t\t\t\twithWriter.SetWriter(w)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn output\n}", "func output(data []byte, encodeType Encode) (string, error) {\n\tswitch encodeType {\n\tcase HEX:\n\t\treturn hex.EncodeToString(data), nil\n\tcase Base64:\n\t\treturn base64.StdEncoding.EncodeToString(data), nil\n\tcase None:\n\t\treturn string(data), nil\n\tdefault:\n\t\treturn \"\", fmt.Errorf(\"secretInfo OutputType unsupport\")\n\t}\n}", "func SetOutput(w io.Writer) { std.SetOutput(w) }", "func TestFieldOutputText(t *testing.T) {\n\tfield := NewField()\n\tfield.Name = \"foo\"\n\tfield.Type = \"text\"\n\n\ttag := field.output()\n\n\tassert.Equal(t, \"<input type=\\\"text\\\" name=\\\"foo\\\" id=\\\"foo\\\" value=\\\"\\\" />\", tag)\n}", "func (o DocumentDbOutputDataSourceResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v DocumentDbOutputDataSourceResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o BlobOutputDataSourceResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BlobOutputDataSourceResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (t *tag) render() string {\n tagType := t.GetType()\n tagFormat := tagsFormat[tagType]\n\n switch t.GetType() {\n case TYPE_OPEN:\n return fmt.Sprintf(tagFormat, t.GetName(), getAttrsToString(t))\n case TYPE_CLOSE:\n return fmt.Sprintf(tagFormat, t.GetName())\n case TYPE_OPEN_CLOSE:\n return fmt.Sprintf(tagFormat, t.GetName(), getAttrsToString(t), t.GetVal(), t.GetName())\n case TYPE_SELF_CLOSED_STRICT:\n if t.GetVal() != \"\" {\n t.SetAttr(\"value\", t.GetVal())\n }\n return fmt.Sprintf(tagFormat, t.GetName(), getAttrsToString(t))\n default:\n return t.GetName()\n }\n}", "func (this *Tidy) ForceOutput(val bool) (bool, error) {\n\treturn this.optSetBool(C.TidyForceOutput, cBool(val))\n}", "func SetContentType(w http.ResponseWriter, asset string) {\n\text := filepath.Ext(asset)\n\tswitch ext {\n\tcase \".png\":\n\t\tfallthrough\n\tcase \".gif\":\n\t\tw.Header().Set(\"Content-Type\", fmt.Sprintf(\"image/%s\", ext))\n\tcase \".woff\":\n\t\tfallthrough\n\tcase \".woff2\":\n\t\tfallthrough\n\tcase \".eot\":\n\t\tfallthrough\n\tcase \".ttf\":\n\t\tw.Header().Set(\"Content-Type\", fmt.Sprintf(\"font/%s\", ext))\n\tcase \".svg\":\n\t\tw.Header().Set(\"Content-Type\", \"image/svg+xml\")\n\tcase \".css\":\n\t\tw.Header().Set(\"Content-Type\", \"text/css\")\n\tcase \".js\":\n\t\tw.Header().Set(\"Content-Type\", \"text/javascript\")\n\tdefault:\n\t\tw.Header().Set(\"Content-Type\", \"text/html; charset=utf-8\")\n\t}\n}", "func (this *Tidy) OutputXml(val bool) (bool, error) {\n\treturn this.optSetBool(C.TidyXmlOut, cBool(val))\n}", "func (p *UpdatePage) SetHTML(html template.HTML) {\n\tp.HTML = html\n}", "func (o DocumentDbOutputDataSourceOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v DocumentDbOutputDataSource) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o GetSmartagFlowLogsLogOutput) OutputType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetSmartagFlowLogsLog) string { return v.OutputType }).(pulumi.StringOutput)\n}", "func (c *Context) HTML(status int, data string) error {\n\tc.SetHeader(\"Content-Type\", \"text/html; charset=utf-8\")\n\treturn c.Bytes(status, []byte(data))\n}", "func (a *Ace) HtmlTemplate(render Renderer) {\n\ta.render = render\n}", "func (r *Reply) HTMLlf(layout, filename string, data Data) *Reply {\n\tr.ContentType(ahttp.ContentTypeHTML.String())\n\tr.Render(&htmlRender{Layout: layout, Filename: filename, ViewArgs: data})\n\treturn r\n}", "func (h HTML) HTML() string {\n\treturn string(h)\n}", "func (o WebAclRuleStatementAndStatementStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementNotStatementStatementRegexMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementNotStatementStatementRegexMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (r *renderer) write(s string, unescaped bool) {\n\tif r.indentNext {\n\t\tr.indentNext = false\n\t\tr.w.WriteString(r.indent)\n\t}\n\tif !unescaped {\n\t\ts = html.EscapeString(s)\n\t}\n\tr.w.WriteString(s)\n}", "func HTML(w http.ResponseWriter, r *http.Request, v string) {\n\trender.HTML(w, r, v)\n}", "func HTML(s string) got.HTML {\n\treturn got.HTML(s)\n}", "func (o WebAclRuleStatementNotStatementStatementRegexMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementRegexMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func ConfigureSetOutput(wr io.Writer) {\n\toutputFunction = func(x string) {\n\t\t_, err := fmt.Fprintln(wr, x)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}\n}", "func (u Unsafe) HTML(ctx context.Context) (HTML, error) {\n\treturn nil, errHTMLOnPrimitive(\"Unsafe\")\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementByteMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementByteMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o BlobOutputDataSourceOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BlobOutputDataSource) string { return v.Type }).(pulumi.StringOutput)\n}", "func (r *Render) HTML(w http.ResponseWriter, name string, data interface{}) {\n\tr.Engine.ExecuteTemplate(w, name, data)\n}", "func (o WebAclRuleStatementOrStatementStatementAndStatementStatementAndStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementOrStatementStatementAndStatementStatementAndStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (c *Collection) Output() bool { return templates.VariableMap(c.Metadata).Bool(\"output\", false) }", "func (o *OutputHandler) createBeautifulHTML() error {\n\terr := o.importFile()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\to.markdownToHTML()\n\n\terr = o.createFile()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (DefaultDispatcher) Write(rw http.ResponseWriter, resp Response) error {\n\tswitch x := resp.(type) {\n\tcase JSONResponse:\n\t\trw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\t\tio.WriteString(rw, \")]}',\\n\") // Break parsing of JavaScript in order to prevent XSSI.\n\t\treturn json.NewEncoder(rw).Encode(x.Data)\n\tcase *TemplateResponse:\n\t\tt, ok := (x.Template).(*template.Template)\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"%T is not a safe template and it cannot be parsed and written\", t)\n\t\t}\n\t\trw.Header().Set(\"Content-Type\", \"text/html; charset=utf-8\")\n\t\tif len(x.FuncMap) == 0 {\n\t\t\tif x.Name == \"\" {\n\t\t\t\treturn t.Execute(rw, x.Data)\n\t\t\t}\n\t\t\treturn t.ExecuteTemplate(rw, x.Name, x.Data)\n\t\t}\n\t\tcloned, err := t.Clone()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcloned = cloned.Funcs(x.FuncMap)\n\t\tif x.Name == \"\" {\n\t\t\treturn cloned.Execute(rw, x.Data)\n\t\t}\n\t\treturn cloned.ExecuteTemplate(rw, x.Name, x.Data)\n\tcase safehtml.HTML:\n\t\trw.Header().Set(\"Content-Type\", \"text/html; charset=utf-8\")\n\t\t_, err := io.WriteString(rw, x.String())\n\t\treturn err\n\tcase FileServerResponse:\n\t\trw.Header().Set(\"Content-Type\", x.ContentType())\n\t\t// The http package will take care of writing the file body.\n\t\treturn nil\n\tcase RedirectResponse:\n\t\thttp.Redirect(rw, x.Request.req, x.Location, int(x.Code))\n\t\treturn nil\n\tcase NoContentResponse:\n\t\trw.WriteHeader(int(StatusNoContent))\n\t\treturn nil\n\tdefault:\n\t\treturn fmt.Errorf(\"%T is not a safe response type and it cannot be written\", resp)\n\t}\n}", "func (o WebAclRuleStatementOrStatementStatementAndStatementStatementAndStatementStatementByteMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementOrStatementStatementAndStatementStatementAndStatementStatementByteMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o EventHubOutputDataSourceResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v EventHubOutputDataSourceResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func writeDocHTML(w http.ResponseWriter, content *contentBuffers, lang string) error {\n\t_, err := w.Write(htmlOpen1)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = w.Write([]byte(lang))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = w.Write(htmlOpen2)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = w.Write(content.head.Bytes())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = w.Write(htmlHeadBody)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = w.Write(content.body.Bytes())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = w.Write(htmlClose)\n\treturn err\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementRegexMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementRegexMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementAndStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementNotStatementStatementOrStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementNotStatementStatementOrStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func Encode(w http.ResponseWriter, accept string, i interface{}, htmlLayout string) {\n\tlog.Println(\"Encoding...\")\n\tlog.Println(accept)\n\tprocessed := false\n\tfor _, a := range strings.Split(accept, \",\") {\n\t\tswitch a {\n\t\tcase \"*/*\", \"application/json\", \"\":\n\t\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\t\tenc := json.NewEncoder(w)\n\t\t\tenc.Encode(i)\n\t\t\tprocessed = true\n\t\tcase \"text/html\":\n\t\t\tlog.Println(\"HTML\")\n\t\t\tw.Header().Set(\"Content-Type\", \"text/html\")\n\t\t\thtml := ToHTML(i, htmlLayout)\n\t\t\tfmt.Fprintln(w, html)\n\t\t\tprocessed = true\n\t\t}\n\t\tif processed {\n\t\t\tbreak\n\t\t}\n\t}\n\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementByteMatchStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementByteMatchStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformationOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WebAclRuleStatementAndStatementStatementOrStatementStatementNotStatementStatementRegexPatternSetReferenceStatementTextTransformation) string {\n\t\treturn v.Type\n\t}).(pulumi.StringOutput)\n}", "func (o RawOutputDatasourceResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v RawOutputDatasourceResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (WriterWrapper) SetOut(filename string) {}", "func normalResponse(w http.ResponseWriter, r *http.Request){\n\trespStr := `<html>\n<head><title> My Custom Response </title> </head>\n<body> <h1> Testing the response headers ...... </h1></body>\n</html>`\nw.Write([]byte(respStr))\n}" ]
[ "0.6400135", "0.6055975", "0.6011148", "0.58667994", "0.5687668", "0.561943", "0.55879533", "0.55276394", "0.5470911", "0.5454469", "0.545375", "0.5367108", "0.53660023", "0.536108", "0.5316962", "0.5314116", "0.5296099", "0.5273366", "0.52713025", "0.5254828", "0.5251903", "0.5235802", "0.5221285", "0.51915514", "0.51729983", "0.51528054", "0.51500666", "0.5148139", "0.5110433", "0.50864744", "0.50842357", "0.50739473", "0.5071545", "0.50691015", "0.50666434", "0.5048121", "0.50396824", "0.50097096", "0.49957404", "0.49876302", "0.49800164", "0.4971295", "0.494561", "0.49405378", "0.4939719", "0.49390963", "0.49385193", "0.49333015", "0.492453", "0.4912866", "0.49069113", "0.48959905", "0.48875076", "0.48748642", "0.48723787", "0.48685327", "0.48630646", "0.48555526", "0.48503974", "0.484757", "0.48447904", "0.48371884", "0.4835455", "0.48275054", "0.4827486", "0.48245165", "0.4817055", "0.48162335", "0.4814617", "0.4814422", "0.48054346", "0.47995633", "0.47960356", "0.4789397", "0.47868866", "0.47863033", "0.47856426", "0.4779764", "0.47794044", "0.47781608", "0.47738522", "0.47724912", "0.47612277", "0.47554404", "0.47525495", "0.47492436", "0.4746553", "0.47426638", "0.47405493", "0.47336054", "0.47322783", "0.47301254", "0.47284845", "0.47267333", "0.47205484", "0.47163665", "0.4715423", "0.471436", "0.47142246", "0.47135586" ]
0.70670927
0
Update is for component to have itself rerendered.
func (s *State) Update() { // fmt.Println("update") s.render() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *renderer) Update(_ tea.Msg, _ *list.Model) tea.Cmd { return nil }", "func (component *Component) Update() {\n\tcomponent.WarnError(component.UpdateWithError())\n}", "func (c *PureComponent) HasUpdated() bool { return false }", "func (v *Component) ComponentDidUpdate(prevProps *Map, prevState *Map) {}", "func (c *Component) OnUpdate() {\n\t// Logger.Trace().Str(\"component\", c.GetName()).Msg(\"OnUpdate\")\n\tif c.customOnUpdate != nil {\n\t\tc.customOnUpdate(c)\n\t}\n}", "func (s *State) RequestUpdate() {\n\ts.update = true\n}", "func (md *MockDisplay) Update(*[64][32]uint8, *[64][32]bool) {}", "func (*ModuleBase) Update(*ggl.Window, float64) {}", "func (v *Component) ComponentWillUpdate(nextProps *Map, nextState *Map) {}", "func (o *CanvasItem) Update() {\n\t//log.Println(\"Calling CanvasItem.Update()\")\n\n\t// Build out the method's arguments\n\tptrArguments := make([]gdnative.Pointer, 0, 0)\n\n\t// Get the method bind\n\tmethodBind := gdnative.NewMethodBind(\"CanvasItem\", \"update\")\n\n\t// Call the parent method.\n\t// void\n\tretPtr := gdnative.NewEmptyVoid()\n\tgdnative.MethodBindPtrCall(methodBind, o.GetBaseObject(), ptrArguments, retPtr)\n\n}", "func (v *Status) Update() error {\n\treturn nil\n}", "func (l *List) Update() {\n\tl.win.Update()\n}", "func (g *Game) Update(display chan<- Frame) {\n\t// apply animations\n\tif g.animator.Step() {\n\t\tg.needsRender = true\n\t}\n\t// render if needed\n\tif g.needsRender {\n\t\tdisplay <- g.Render()\n\t\tg.needsRender = false\n\t}\n}", "func (w *Window) Update() {\n\tw.redraw()\n\tw.refreshWait()\n\tw.resize()\n\tpollEvents()\n}", "func (c *Board) OnUpdate() {\n\tc.Component.OnUpdate()\n}", "func (NilUGauge) Update(v uint64) {}", "func (v *Component) redraw() {\n\tv.Render()\n}", "func (c *SceneController) OnUpdate() {\n\tc.Component.OnUpdate()\n}", "func (b *Bar) Update(progress int64) {\n\tb.control <- progress\n}", "func (s *State) Update(status mesos.TaskStatus) {\n\ts.updates <- status\n}", "func (w* Workermeta) Update(state WorkerState) Workermeta {\n\tw.State = state\n\tw.LastUpdateTime = time.Now().Unix()\n\tnw := *w\n\treturn nw\n}", "func (g *Game) Update() error{\n\treturn nil\n}", "func (l *Label) Update() error {\n\treturn nil\n}", "func (self *Graphics) Update() {\n self.Object.Call(\"update\")\n}", "func (f *falconMeter) Update() {\n\tfalcon.SetMeterCount(f.name, 1)\n}", "func (m *BoolMetric) Update(v bool) {\n\tm.update(v)\n}", "func (v *Component) ShouldComponentUpdate(nextProps *Map, nextState *Map) bool {\n\treturn true\n}", "func (c *PIDController) Update(value float64) float64 {\n\tvar duration time.Duration\n\tif !c.lastUpdate.IsZero() {\n\t\tduration = time.Since(c.lastUpdate)\n\t}\n\tc.lastUpdate = time.Now()\n\treturn c.UpdateDuration(value, duration)\n}", "func (h *consulGRPCSocketHook) Update(req *interfaces.RunnerUpdateRequest) error {\n\th.mu.Lock()\n\tdefer h.mu.Unlock()\n\n\th.alloc = req.Alloc\n\n\tif !h.shouldRun() {\n\t\treturn nil\n\t}\n\n\treturn h.proxy.run(h.alloc)\n}", "func (v *Layer) Update() error {\n\treturn nil\n}", "func (ui *UI) Update(f func()) {\n\tf()\n}", "func (o *sampleUpdateHandler) Update(rw http.ResponseWriter, req *http.Request) {\n\to.UpdateHandler.Update(rw, req)\n}", "func (b *BaseElement) HandleUpdate(update UpdateI) {\n\tdirty := true\n\tswitch u := update.(type) {\n\tcase UpdateValue:\n\t\tb.SetValue(u.Value)\n\tcase UpdateX:\n\t\tb.Style.X = u.Number\n\t\tb.Restyle = true\n\tcase UpdateY:\n\t\tb.Style.Y = u.Number\n\t\tb.Restyle = true\n\tcase UpdateW:\n\t\tb.Style.W = u.Number\n\t\tb.Restyle = true\n\tcase UpdateH:\n\t\tb.Style.H = u.Number\n\t\tb.Restyle = true\n\tcase UpdateDimensions:\n\t\tb.Style.X.Value = u.X.Value\n\t\tb.Style.Y.Value = u.Y.Value\n\t\tb.Style.W.Value = u.W.Value\n\t\tb.Style.H.Value = u.H.Value\n\t\tb.Restyle = true\n\tcase UpdateScroll:\n\t\tb.Style.ScrollLeft = u.Left\n\t\tb.Style.ScrollTop = u.Top\n\t\tb.Restyle = true\n\tcase UpdateScrollLeft:\n\t\tb.Style.ScrollLeft = u.Number\n\t\tb.Restyle = true\n\tcase UpdateScrollTop:\n\t\tb.Style.ScrollTop = u.Number\n\t\tb.Restyle = true\n\tcase UpdateZIndex:\n\t\tb.Style.ZIndex = u.Number\n\tcase UpdateOutlineColor:\n\t\tb.Style.OutlineColor = u\n\tcase UpdateBackgroundColor:\n\t\tb.Style.BackgroundColor = color.NRGBA(u)\n\tcase UpdateForegroundColor:\n\t\tb.Style.ForegroundColor = color.NRGBA(u)\n\tcase UpdateDirt:\n\t\tdirty = u\n\tcase UpdateFocus:\n\t\tb.Focus()\n\tcase UpdateHidden:\n\t\tb.SetHidden(bool(u))\n\t\tif u == false {\n\t\t\tb.Restyle = true\n\t\t}\n\tcase UpdateAlpha:\n\t\tb.Style.Alpha.Set(u)\n\tcase UpdateColorMod:\n\t\tb.Style.ColorMod = color.NRGBA{u.R, u.G, u.B, u.A}\n\tcase UpdateParseStyle:\n\t\tb.Style.Parse(string(u))\n\t}\n\tb.SetDirty(dirty)\n}", "func (g *Game) Update() error {\n\treturn nil\n}", "func (g *Game) Update() error {\n\treturn nil\n}", "func (r *Reporter) Update(message string) error {\n\tselect {\n\tcase r.c <- message:\n\t\treturn nil\n\tdefault:\n\t\treturn fmt.Errorf(\"warning: Reporter buffer is full\")\n\t}\n}", "func (b *Buffer) update() {\n\tb.NumLines = len(b.lines)\n}", "func (cl *RestClient) Update() {\n}", "func UpdateWidget(res http.ResponseWriter, req *http.Request) {\n\tresp := response.New()\n\n\tresp.Render(res, req)\n}", "func (g *Game) Update(){\n\t// update logic here\n\tg.inputs()\n\t// inputs\n\tif !g.gameOver {\n\t}\n}", "func (s *ProtoViewSourceJob) Update(key string, msg proto.Message) error {\n\ts.keysSeen[key] = struct{}{}\n\n\tcurrent, err := s.view.Get(key)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to get object\")\n\t}\n\n\tvar shouldUpdate = true\n\n\tif current != nil {\n\t\tc := current.(proto.Message)\n\t\tshouldUpdate = c == nil || !proto.Equal(c, msg)\n\t}\n\n\tif !shouldUpdate {\n\t\treturn nil\n\t}\n\n\terr = s.emitter.Emit(key, msg)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to emit update\")\n\t}\n\n\treturn nil\n}", "func (self *Graphics) PostUpdate() {\n self.Object.Call(\"postUpdate\")\n}", "func (c *Cherry) Update(dotTime int) error {\n\tif c.eaten == false {\n\t\treturn nil\n\t}\n\treturn nil\n}", "func (i *IpScheduler) OnUpdate(old, new interface{}) {}", "func (s *Stats) Update(c *gin.Context, last time.Time) {\n\tStatsWith(c, s.update(last))\n}", "func (s *State) UpdateChild(g *draw.Buffer, bounds image.Rectangle, c Component) {\n\tg.Push(bounds)\n\tif s.disabled {\n\t\tc.Update(g, s)\n\t\tg.Pop()\n\t\treturn\n\t}\n\tthis := s.current\n\th := s.hovered\n\tv := s.visible\n\tf := s.focusable\n\ts.visible = image.Rectangle{}\n\ts.hovered = h && s.mousePos.Sub(s.bounds.Min).In(bounds)\n\tif s.hovered {\n\t\ts.hoveredC = c\n\t}\n\tb := s.bounds\n\ts.bounds = bounds.Add(s.bounds.Min)\n\ts.current = c\n\ts.focusable = false\n\tif s.focused == s.current {\n\t\ts.focusNext = true\n\t}\n\tc.Update(g, s)\n\tif s.focusable {\n\t\ts.lastFocusable = c\n\t}\n\ts.current = this\n\ts.bounds = b\n\ts.hovered = h\n\tif s.visible.Empty() {\n\t\ts.visible = v\n\t} else {\n\t\ts.visible = s.visible.Add(bounds.Min)\n\t}\n\ts.focusable = f\n\tg.Pop()\n}", "func (g *Game) Update(screen *ebiten.Image) error {\n\tg.view = screen\n\n\t// Perform logical updates\n\tg.camera.Update()\n\n\t// TODO: Add your update logic here\n\tg.handleInput()\n\n\tif ebiten.IsDrawingSkipped() {\n\t\t// When the game is running slowly, the rendering result\n\t\t// will not be adopted.\n\t\treturn nil\n\t}\n\n\t// Render game to screen\n\tg.draw()\n\n\t// TPS counter\n\tfps := fmt.Sprintf(\"TPS: %f/%v\", ebiten.CurrentTPS(), ebiten.MaxTPS())\n\tebitenutil.DebugPrint(g.view, fps)\n\n\treturn nil\n}", "func (t *AudioPlayer) Update(a *app.App, deltaTime time.Duration) {\n\n\tif time.Now().Sub(t.lastUpdate) < 100*time.Millisecond {\n\t\treturn\n\t}\n\tt.pc1.UpdateTime()\n\tt.pc2.UpdateTime()\n\tt.pc3.UpdateTime()\n\tt.pc4.UpdateTime()\n\tt.lastUpdate = time.Now()\n}", "func (t *Points) Update(a *app.App, deltaTime time.Duration) {}", "func (b *Batcher) Update(matched, total int) {\n\tb.ratio = float64(matched) / float64(total)\n}", "func (s *HelloSystem) Update(ctx core.UpdateCtx) {}", "func (job *AnalyzeJob) Update(rowCount int64) {\n\tnow := time.Now()\n\tjob.Mutex.Lock()\n\tjob.RowCount += rowCount\n\tjob.updateTime = now\n\tjob.Mutex.Unlock()\n}", "func (g *Gui) Update(f func(*Gui) error) {\n\tgo func() { g.userEvents <- userEvent{f: f} }()\n}", "func (_m *ITaskActions) SendUpdate() {\n\t_m.Called()\n}", "func (dd *Datadog) Update(component *Component) error {\n\tswitch component.Type {\n\tcase types.ComponentDashboard:\n\t\treturn dd.updateDashboard(component.Dashboard)\n\tcase types.ComponentMonitor:\n\t\treturn dd.updateMonitor(component.Monitor)\n\tcase types.ComponentDowntime:\n\t\treturn dd.updateDowntime(component.Downtime)\n\tcase types.ComponentScreenboard:\n\t\treturn dd.updateScreenBoard(component.ScreenBoard)\n\t}\n\n\treturn ErrInvalidComponentTypeID\n}", "func (r *RadioStation) Update() error {\n\treturn nil\n}", "func (c *Camera) Update() {\n\tif c.renderer == nil {\n\t\treturn\n\t}\n\n\tc.updateZoom()\n\tc.santizeBounds()\n\n\tc.renderer.SetViewport(c.Viewport())\n}", "func (c *common) update() error {\n\tif len(c.flags.update) == 0 {\n\t\treturn nil\n\t}\n\n\targs := append(c.flags.global, c.flags.update...)\n\n\treturn shared.RunCommand(c.ctx, nil, nil, c.commands.update, args...)\n}", "func (view *DetailsView) Update() error {\n\treturn nil\n}", "func (d *Display) Update() error {\n\tif inpututil.IsKeyJustPressed(ebiten.KeyEscape) {\n\t\treturn errors.New(\"user exit\")\n\t}\n\tctx := NewUpdateContext(d.ctx)\n\n\t// update the mouse event registry\n\td.mouseEventRegistry.Update()\n\n\t// call all update handlers\n\tfor i := 0; i < len(d.updateHandlers); i++ {\n\t\tif err := d.updateHandlers[i].Update(ctx); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// draw background\n\tif d.background != nil {\n\t\tif err := d.background.Update(ctx); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// update components\n\tfor _, c := range d.components {\n\t\tif err := c.Update(ctx); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// draw cursor\n\tif d.cursor != nil {\n\t\tif err := d.cursor.Update(ctx); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (info *componentInfo) updateComponent(step float32, runtime *Runtime, context *Context) {\n\tif info.Active == 0 && info.Start != nil {\n\t\truntime.workers.Run(func() {\n\t\t\tinfo.Start.Start(context)\n\t\t\tinfo.Active += 1\n\t\t})\n\t} else if info.Update != nil {\n\t\truntime.workers.Run(func() {\n\t\t\tinfo.Update.Update(context)\n\t\t\tinfo.Active += 1\n\t\t})\n\t}\n}", "func CallUpdated(p *Element) error {\n\tif p == nil {\n\t\treturn nil\n\t}\n\n\tc := p.Component\n\tif c == nil {\n\t\tc = p.ParentComponent().Component\n\t}\n\n\tif c.Hooks.Updated != nil {\n\t\terr := c.Hooks.Updated()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (rs *RenderSystem) Update(frameDelta float32) {\n\t// clear the screen\n\twidth, height := rs.Renderer.GetResolution()\n\trs.gfx.Viewport(0, 0, int32(width), int32(height))\n\trs.gfx.ClearColor(0.25, 0.25, 0.25, 1.0)\n\trs.gfx.Clear(graphics.COLOR_BUFFER_BIT | graphics.DEPTH_BUFFER_BIT)\n\n\t// make the projection and view matrixes\n\tprojection := mgl.Perspective(mgl.DegToRad(60.0), float32(width)/float32(height), 1.0, 100.0)\n\tvar view mgl.Mat4\n\tif rs.Camera != nil {\n\t\tview = rs.Camera.GetViewMatrix()\n\t} else {\n\t\tview = mgl.Ident4()\n\t}\n\n\t// draw stuff the visible entities\n\tfor _, e := range rs.visibleEntities {\n\t\tvisibleEntity, okay := e.(RenderableEntity)\n\t\tif okay {\n\t\t\tr := visibleEntity.GetRenderable()\n\t\t\trs.Renderer.DrawRenderable(r, nil, projection, view, rs.Camera)\n\t\t}\n\t}\n\n\t// draw the screen\n\trs.MainWindow.SwapBuffers()\n}", "func (f *BasicFeature) Update(TimeCurrent uint64, data []*dfedata.InputData, connectionChannel ...chan ConnectionChannelData) {\n}", "func (o *Status) Update() {\n o.Time = time.Now()\n}", "func Update(s *discordgo.Session, m *discordgo.MessageCreate) {\n\tif *R == false {\n\t\ts.ChannelMessageSend(support.Config.FactorioChannelID, \"Server is not running!\")\n\t\treturn\n\t}\n\n\ts.ChannelMessageSend(support.Config.FactorioChannelID, \"Server received factorio client update command.\")\n\t*QuitFlag = 1\n\tio.WriteString(*P, \"/quit\\n\")\n\ttime.Sleep(600 * time.Millisecond)\n\tfor {\n\t\tif *QuitFlag == 2 {\n\t\t\ts.ChannelMessageSend(support.Config.FactorioChannelID, \"server is closed.\")\n\t\t\t*QuitFlag = 0\n\t\t\tbreak\n\t\t}\n\t}\n\n\t*R = false\n\tUpdateCmd = 1\n\n\treturn\n}", "func (pl *PolledList) UpdateCh() <-chan struct{} { return pl.updateCh }", "func (pl *PolledList) UpdateCh() <-chan struct{} { return pl.updateCh }", "func (s *Gauge) Update(time TS.TimeSelect) {\n\tt, _, _ := time.CurTime()\n\ts.SetData(t)\n\ts.SetTitle(t)\n}", "func (c *NullController) Update(tick int) {\n\tvar (\n\t\te = c.entity\n\t\tam = e.AnimationManager()\n\t)\n\tif am.Ready() {\n\t\tam.Select(\"Idle\")\n\t}\n}", "func (t *Pitch) Update(a *app.App, deltaTime time.Duration) {}", "func (t *Pitch) Update(a *app.App, deltaTime time.Duration) {}", "func (m *Float64Metric) Update(v float64) {\n\tm.update(v)\n}", "func (g *GistFile) Update(interface{}) (*http.Response, error) {\n\tpanic(\"implement me\")\n}", "func (g *Game) Update() error {\n\tif gameover {\n\t\treturn nil\n\t}\n\n\treturn nil\n}", "func (c *Candy) Update(g *Game) bool {\n\treturn c.state.Update(g, c)\n}", "func (s *UniformSample) Update(v int64) {\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\ts.count++\n\tif len(s.values) < s.reservoirSize {\n\t\ts.values = append(s.values, v)\n\t} else {\n\t\t// Use circle buffer to eliminate the oldest value\n\t\tidx := s.count % int64(s.reservoirSize)\n\t\ts.values[idx] = v\n\t}\n}", "func (g userGists) ShouldComponentUpdate(this *gr.This, next gr.Cops) bool {\n\treturn g.State().HasChanged(next.State, \"gists\")\n}", "func (h *UpdateHandler) Update(rw http.ResponseWriter, req *http.Request) {\n\tstartTime := time.Now()\n\n\tdefer func() {\n\t\th.metrics.HTTPCreateUpdateTime(time.Since(startTime))\n\t}()\n\n\trequest, err := io.ReadAll(req.Body)\n\tif err != nil {\n\t\tcommon.WriteError(rw, http.StatusBadRequest, err)\n\n\t\treturn\n\t}\n\n\tlogger.Debug(\"Processing update request\", logfields.WithRequestBody(request))\n\n\tresponse, err := h.doUpdate(request)\n\tif err != nil {\n\t\tcommon.WriteError(rw, err.(*common.HTTPError).Status(), err)\n\n\t\treturn\n\t}\n\tcommon.WriteResponse(rw, http.StatusOK, response)\n}", "func (w *Watcher) Update() {\n\tw.Action = true\n\tfits := w.SessionKey[:2]\n\tfmt.Println(\"[!] Attempting to update watcher: %s\", fits)\n\twriten, err := w.Connection.Write([]byte(\"Y\"))\n\tif writen != len([]byte(\"Y\")) {\n\t\tfmt.Println(\"[!]Error writting: unable to write\")\n\t}\n\tif err != nil {\n\t\tfmt.Printf(\"%s\", err)\n\t}\n\n}", "func (d *delegate) NotifyUpdate(n *memberlist.Node) {\n\tlevel.Debug(d.logger).Log(\"received\", \"NotifyUpdate\", \"node\", n.Name, \"addr\", n.Address())\n}", "func (_m *Repository) Update(p *entity.Person, commitChan <-chan bool, doneChan chan<- bool) {\n\t_m.Called(p, commitChan, doneChan)\n}", "func (s *Scroll) Update() {\n\ts.win.Update()\n}", "func (v *Filter) Update() error {\n\treturn nil\n}", "func (g *Game) Update(screen *ebiten.Image) error {\n\treturn nil\n}", "func (c *Client) Update() goa.Endpoint {\n\treturn func(ctx context.Context, v interface{}) (interface{}, error) {\n\t\tinv := goagrpc.NewInvoker(\n\t\t\tBuildUpdateFunc(c.grpccli, c.opts...),\n\t\t\tEncodeUpdateRequest,\n\t\t\tnil)\n\t\tres, err := inv.Invoke(ctx, v)\n\t\tif err != nil {\n\t\t\treturn nil, goa.Fault(err.Error())\n\t\t}\n\t\treturn res, nil\n\t}\n}", "func (self *Graphics) PreUpdate() {\n self.Object.Call(\"preUpdate\")\n}", "func Update() {\n\n}", "func (w *Watcher) DoUpdate() {\n\tfor _, i := range w.Inputs {\n\t\tif w.closed {\n\t\t\treturn\n\t\t}\n\t\tnewStats, err := i.GetStats()\n\t\tif err == nil {\n\t\t\tif i.LastErr != nil || w.isSameAs(i.KnownStats, newStats) == false {\n\t\t\t\ti.Update(newStats)\n\t\t\t}\n\t\t} else {\n\t\t\tif i.OnErrored != nil {\n\t\t\t\ti.OnErrored(i, err)\n\t\t\t}\n\t\t\tw.LogFunc(\"Failed to resolve: %v\", err)\n\t\t}\n\t}\n}", "func (pm *basePackageManager) Update() error {\n\t_, _, err := RunCommandWithRetry(pm.cmder.UpdateCmd(), nil)\n\treturn err\n}", "func (t *Timer) Update() {\n\tt.currentTime = t.getCurrentTimeMs()\n\tdelta := t.currentTime - t.lastTime\n\n\tt.tick = float64(delta) * 0.001\n\tt.lastTime = t.currentTime\n}", "func (p *PodsWidget) Update() error {\n\tif !p.pause {\n\t\tpods, err := p.apiClient.GetPodsMetrics(p.filter, p.sortorder)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\trows := make([][]string, len(pods))\n\t\tfor i, pod := range pods {\n\t\t\trows[i] = make([]string, 11)\n\t\t\trows[i][0] = pod.Namespace\n\t\t\trows[i][1] = pod.Name\n\t\t\trows[i][2] = fmt.Sprintf(\"%d/%d\", pod.ContainersReady, pod.ContainersCount)\n\t\t\trows[i][3] = pod.Status\n\t\t\trows[i][4] = fmt.Sprintf(\"%d\", pod.Restarts)\n\t\t\trows[i][5] = fmt.Sprintf(\"%dm\", pod.CPU)\n\t\t\trows[i][6] = helpers.RenderCPUMax(pod.CPUMax, pod.CPUMaxContainerCount, int64(pod.ContainersCount))\n\t\t\trows[i][7] = helpers.FormatBytes(pod.Memory)\n\t\t\trows[i][8] = helpers.RenderMemoryMax(pod.MemoryMax, pod.MemoryMaxContainerCount, int64(pod.ContainersCount))\n\t\t\trows[i][9] = pod.IP\n\t\t\trows[i][10] = helpers.FormatDuration(time.Now().Sub(pod.CreationDate))\n\t\t}\n\n\t\tp.Rows = rows\n\t}\n\n\treturn nil\n}", "func (NilTimer) Update(time.Duration) {}", "func (p *PageView) Rerender() {\n\ttimeKey := time.Now().UnixNano()\n\tp.lastTimeKey = timeKey\n\tgo func() {\n\t\ttime.Sleep(800 * time.Millisecond)\n\t\tif timeKey == p.lastTimeKey {\n\t\t\tvecty.Rerender(p)\n\t\t}\n\t}()\n}", "func (s *EmptyScene) Update() {\n}", "func (sp *ServiceProcessor) Update(event controller.Event) error {\n\tif ksChange, isKSChange := event.(*controller.KubeStateChange); isKSChange {\n\t\treturn sp.propagateDataChangeEv(ksChange)\n\t}\n\n\tif addPod, isAddPod := event.(*podmanager.AddPod); isAddPod {\n\t\treturn sp.ProcessNewPod(addPod.Pod.Namespace, addPod.Pod.Name)\n\t}\n\tif deletePod, isDeletePod := event.(*podmanager.DeletePod); isDeletePod {\n\t\treturn sp.ProcessDeletingPod(deletePod.Pod.Namespace, deletePod.Pod.Name)\n\t}\n\n\tif _, isNodeUpdate := event.(*nodesync.NodeUpdate); isNodeUpdate {\n\t\treturn sp.renderNodePorts()\n\t}\n\n\treturn nil\n}", "func (b Plugin) Update() error {\n\treturn nil\n}", "func Update(req handler.Request, prevModel *Model, currentModel *Model) (handler.ProgressEvent, error) {\n\t// Add your code here:\n\t// * Make API calls (use req.Session)\n\t// * Mutate the model\n\t// * Check/set any callback context (req.CallbackContext / response.CallbackContext)\n\n\t// Construct a new handler.ProgressEvent and return it\n\tresponse := handler.ProgressEvent{\n\t\tOperationStatus: handler.Success,\n\t\tMessage: \"Update complete\",\n\t\tResourceModel: currentModel,\n\t}\n\n\treturn response, nil\n\n\t// Not implemented, return an empty handler.ProgressEvent\n\t// and an error\n\treturn handler.ProgressEvent{}, errors.New(\"Not implemented: Update\")\n}", "func (nv *NetView) Update() {\n\tif !nv.IsVisible() || !nv.HasLayers() {\n\t\treturn\n\t}\n\tvs := nv.Scene()\n\tupdt := vs.UpdateStart()\n\tnv.UpdateImpl()\n\tvs.UpdateEnd(updt)\n}", "func (t *StandardTimer) Update(d time.Duration) {\n\tt.mutex.Lock()\n\tdefer t.mutex.Unlock()\n\tt.histogram.Update(int64(d))\n\tt.meter.Mark(1)\n}" ]
[ "0.739471", "0.7180391", "0.7151857", "0.6852027", "0.67297417", "0.6516826", "0.6448432", "0.6317419", "0.6290429", "0.6266985", "0.6240016", "0.6225204", "0.6130606", "0.6125892", "0.6121056", "0.61098975", "0.6102079", "0.6094033", "0.6089738", "0.60849196", "0.60818887", "0.60601234", "0.60566497", "0.604679", "0.6039725", "0.60317504", "0.602329", "0.60176927", "0.601041", "0.60071033", "0.5988684", "0.5980281", "0.5973557", "0.59378386", "0.59378386", "0.59161747", "0.59160525", "0.5915103", "0.590533", "0.5882886", "0.5880514", "0.5879305", "0.5863307", "0.58526844", "0.5831093", "0.58232796", "0.5817", "0.5815649", "0.580157", "0.5792371", "0.5787555", "0.57770735", "0.57684493", "0.576593", "0.57433736", "0.57301456", "0.5722555", "0.57047933", "0.57047117", "0.5704467", "0.57043856", "0.570402", "0.5672171", "0.56652105", "0.5664052", "0.56625104", "0.5660584", "0.5660584", "0.56567425", "0.5651829", "0.5646247", "0.5646247", "0.56429505", "0.56424284", "0.5632425", "0.5632244", "0.5631095", "0.56281126", "0.5609607", "0.5609054", "0.5605489", "0.56005055", "0.55998594", "0.55966854", "0.55955863", "0.5589095", "0.558877", "0.55862176", "0.5585593", "0.5583543", "0.55741507", "0.5554694", "0.5549732", "0.55487806", "0.55412555", "0.5541001", "0.5517275", "0.5511247", "0.55048704", "0.5497264" ]
0.73830247
1
/ NewComboBoxSelector :initializer of combo box selector
func NewComboBoxSelector(label string, list []string) *ComboBoxSelector { obj := new(ComboBoxSelector) obj.SelectedItem = list[0] // initialize widgets obj.Cell = widgets.NewQWidget(nil, 0) obj.box = widgets.NewQComboBox(obj.Cell) obj.box.AddItems(list) obj.textLabel = widgets.NewQLabel2(label, obj.Cell, 0) // layout layout := widgets.NewQHBoxLayout() layout.AddWidget(obj.textLabel, 0, 0) layout.AddWidget(obj.box, 0, 0) // apply layout obj.Cell.SetLayout(layout) // action connection obj.box.ConnectCurrentIndexChanged(func(index int) { obj.SelectedItem = list[index] //fmt.Println(obj.SelectedItem) }) return obj }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewCombobox() *Combobox {\n\tc := new(Combobox)\n\n\tc.c = C.uiNewCombobox()\n\n\tC.pkguiComboboxOnSelected(c.c)\n\n\tc.ControlBase = NewControlBase(c, uintptr(unsafe.Pointer(c.c)))\n\treturn c\n}", "func (app *controlsTestApplication) ForComboBox() *controls.ComboBoxBuilder {\n\treturn controls.NewComboBoxBuilder(app.ForLabel(), app.rectRenderer)\n}", "func NewSelector(ctx Context) Selector {\n\tstate := getState(ctx)\n\tstate.dispatcher.selectorSequence++\n\treturn NewNamedSelector(ctx, fmt.Sprintf(\"selector-%v\", state.dispatcher.selectorSequence))\n}", "func NewSelector() Selector {\n\treturn internalSelector(nil)\n}", "func newComboUtility(combinator Combinator) *comboUtility {\n\treturn &comboUtility{\n\t\tcombinator: combinator,\n\t\tchildren: bag.New(),\n\t\treset: true,\n\t}\n}", "func createObjectPicker(n int64, distrib string) (ObjectSelector, error) {\n\n\tswitch distrib {\n\tcase ZIPFIAN_OBJECT_PICK:\n\t\tvar x = new(Zipf)\n\t\tx.SetParams(n, 0.8, 99)\n\t\treturn x, nil\n\tcase UNIFORM_OBJECT_PICK:\n\t\tvar y = new(Uniform)\n\t\ty.SetParams(n, 99)\n\t\treturn y, nil\n\t}\n\treturn nil, errors.New(\"Not a valid distribution for object selection\")\n}", "func createObjectPicker(n int64, distrib string) (ObjectSelector, error) {\n\n\tswitch distrib {\n\tcase ZIPFIAN_OBJECT_PICK:\n\t\tvar x = new(Zipf)\n\t\tx.SetParams(n, 0.8, 99)\n\t\treturn x, nil\n\tcase UNIFORM_OBJECT_PICK:\n\t\tvar y = new(Uniform)\n\t\ty.SetParams(n, 99)\n\t\treturn y, nil\n\t}\n\n\tfmt.Println(\"what is the distribution :\", distrib)\n\treturn new(Uniform), errors.New(\"Not a valid distribution for object selection\")\n}", "func NewCFNSelector(prompt Prompter) *CFNSelector {\n\treturn &CFNSelector{\n\t\tprompt: prompt,\n\t}\n}", "func New(opts ...Option) selector.Selector {\n\treturn NewBuilder(opts...).Build()\n}", "func GuiComboBox(bounds Rectangle, text string, active int) int {\n\tctext := C.CString(text)\n\tdefer C.free(unsafe.Pointer(ctext))\n\tcbounds := *bounds.cptr()\n\tres := C.GuiComboBox(cbounds, ctext, C.int(int32(active)))\n\treturn int(int32(res))\n}", "func comboBoxFinalizer(cb *ComboBox) {\n\truntime.SetFinalizer(cb, func(cb *ComboBox) { gobject.Unref(cb) })\n}", "func (t *OpenconfigSystem_System_Logging_Console_Selectors) NewSelector(Facility E_OpenconfigSystemLogging_SYSLOG_FACILITY, Severity E_OpenconfigSystemLogging_SyslogSeverity) (*OpenconfigSystem_System_Logging_Console_Selectors_Selector, error){\n\n\t// Initialise the list within the receiver struct if it has not already been\n\t// created.\n\tif t.Selector == nil {\n\t\tt.Selector = make(map[OpenconfigSystem_System_Logging_Console_Selectors_Selector_Key]*OpenconfigSystem_System_Logging_Console_Selectors_Selector)\n\t}\n\n\tkey := OpenconfigSystem_System_Logging_Console_Selectors_Selector_Key{\n\t\tFacility: Facility,\n\t\tSeverity: Severity,\n\t}\n\n\t// Ensure that this key has not already been used in the\n\t// list. Keyed YANG lists do not allow duplicate keys to\n\t// be created.\n\tif _, ok := t.Selector[key]; ok {\n\t\treturn nil, fmt.Errorf(\"duplicate key %v for list Selector\", key)\n\t}\n\n\tt.Selector[key] = &OpenconfigSystem_System_Logging_Console_Selectors_Selector{\n\t\tFacility: Facility,\n\t\tSeverity: Severity,\n\t}\n\n\treturn t.Selector[key], nil\n}", "func GuiComboBox(bounds Rectangle, text string, active int32) int32 {\n\tcbounds, _ := *(*C.Rectangle)(unsafe.Pointer(&bounds)), cgoAllocsUnknown\n\ttext = safeString(text)\n\tctext, _ := unpackPCharString(text)\n\tcactive, _ := (C.int)(active), cgoAllocsUnknown\n\t__ret := C.GuiComboBox(cbounds, ctext, cactive)\n\truntime.KeepAlive(text)\n\t__v := (int32)(__ret)\n\treturn __v\n}", "func (t *OpenconfigOfficeAp_System_Logging_Console_Selectors) NewSelector(Facility E_OpenconfigSystemLogging_SYSLOG_FACILITY, Severity E_OpenconfigSystemLogging_SyslogSeverity) (*OpenconfigOfficeAp_System_Logging_Console_Selectors_Selector, error) {\n\n\t// Initialise the list within the receiver struct if it has not already been\n\t// created.\n\tif t.Selector == nil {\n\t\tt.Selector = make(map[OpenconfigOfficeAp_System_Logging_Console_Selectors_Selector_Key]*OpenconfigOfficeAp_System_Logging_Console_Selectors_Selector)\n\t}\n\n\tkey := OpenconfigOfficeAp_System_Logging_Console_Selectors_Selector_Key{\n\t\tFacility: Facility,\n\t\tSeverity: Severity,\n\t}\n\n\t// Ensure that this key has not already been used in the\n\t// list. Keyed YANG lists do not allow duplicate keys to\n\t// be created.\n\tif _, ok := t.Selector[key]; ok {\n\t\treturn nil, fmt.Errorf(\"duplicate key %v for list Selector\", key)\n\t}\n\n\tt.Selector[key] = &OpenconfigOfficeAp_System_Logging_Console_Selectors_Selector{\n\t\tFacility: Facility,\n\t\tSeverity: Severity,\n\t}\n\n\treturn t.Selector[key], nil\n}", "func newChoiceBuilder(choiceDef *ChoiceDef) ChoiceBuilder {\n\treturn &chosenBuilder{\n\t\tchoiceDef: choiceDef,\n\t}\n}", "func SelectorCli(label string, options ...string) (string, error) {\n\ts := promptui.Select{\n\t\tLabel: label,\n\t\tItems: options,\n\t}\n\n\t_, result, err := s.Run()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treturn result, nil\n}", "func NewNamedSelector(ctx Context, name string) Selector {\n\treturn &selectorImpl{name: name}\n}", "func NewSelectorFilter(filter string) LinkFilter {\n\treturn NewLinkFilter(selectorFilter, selectorFilterCode, filter)\n}", "func NewCmdSelector(f kcmdutil.Factory, streams genericclioptions.IOStreams) *cobra.Command {\n\tcmd := set.NewCmdSelector(f, streams)\n\tcmd.Long = setSelectorLong\n\tcmd.Example = setSelectorExample\n\n\treturn cmd\n}", "func NewFromRoundRobinSelector(ss *RoundRobinServerList) *Client {\n\treturn &Client{\n\t\tselector: ss,\n\t\tDisableCAS: false,\n\t}\n}", "func (l *labelInfo) genSelector() clusterservice.Selector {\n\treturn clusterservice.NewSelector().SelectByLabel(\n\t\tl.allLabels(), clusterservice.EQ,\n\t)\n}", "func newSelection(pools *csp.CSPList, opts ...buildOption) *selection {\n\ts := &selection{pools: pools, policies: &policyList{map[priority][]policy{}}}\n\tfor _, o := range opts {\n\t\tif o != nil {\n\t\t\to(s)\n\t\t}\n\t}\n\twithDefaultSelection(s)\n\treturn s\n}", "func NewSelect(table string, columns []string) Select {\n\treturn Select{\n\t\tcolumns: columns,\n\t\tBaseQuery: BaseQuery{table: table},\n\t}\n}", "func (s *Selector) Create(from time.Time) {\n\tdays := make([]int, 31)\n\tfor d := 0; d < 31; d++ { // TODO: Actual number of days / month (change dynamically on selection?).\n\t\tdays[d] = d + 1\n\t}\n\tnumYears := 5\n\tyears := make([]int, numYears)\n\tfor i := 0; i < numYears; i++ {\n\t\tyears[i] = from.Year() - i\n\t}\n\t*s = Selector{\n\t\tSelectedYear: from.Year(),\n\t\tSelectedMonth: from.Month() - 1, // -1 to give [0, 11]\n\t\tSelectedDay: from.Day(),\n\t\tSelectedTime: from.Format(\"15:04\"),\n\t\tDaysInMonth: days,\n\t\tMonths: make([]time.Month, 12),\n\t\tYears: years,\n\t}\n\tfor i := 1; i <= 12; i++ {\n\t\ts.Months[i-1] = time.Month(i)\n\t}\n}", "func GetSelector() *Selector {}", "func (t *OpenconfigSystem_System_Logging_RemoteServers_RemoteServer_Selectors) NewSelector(Facility E_OpenconfigSystemLogging_SYSLOG_FACILITY, Severity E_OpenconfigSystemLogging_SyslogSeverity) (*OpenconfigSystem_System_Logging_RemoteServers_RemoteServer_Selectors_Selector, error){\n\n\t// Initialise the list within the receiver struct if it has not already been\n\t// created.\n\tif t.Selector == nil {\n\t\tt.Selector = make(map[OpenconfigSystem_System_Logging_RemoteServers_RemoteServer_Selectors_Selector_Key]*OpenconfigSystem_System_Logging_RemoteServers_RemoteServer_Selectors_Selector)\n\t}\n\n\tkey := OpenconfigSystem_System_Logging_RemoteServers_RemoteServer_Selectors_Selector_Key{\n\t\tFacility: Facility,\n\t\tSeverity: Severity,\n\t}\n\n\t// Ensure that this key has not already been used in the\n\t// list. Keyed YANG lists do not allow duplicate keys to\n\t// be created.\n\tif _, ok := t.Selector[key]; ok {\n\t\treturn nil, fmt.Errorf(\"duplicate key %v for list Selector\", key)\n\t}\n\n\tt.Selector[key] = &OpenconfigSystem_System_Logging_RemoteServers_RemoteServer_Selectors_Selector{\n\t\tFacility: Facility,\n\t\tSeverity: Severity,\n\t}\n\n\treturn t.Selector[key], nil\n}", "func (c ComboBox) Items() []Selection {\n\tresult := make([]Selection, len(c))\n\tfor i := range c {\n\t\tresult[i] = Selection{Name: c[i].Name, Value: strconv.Itoa(i + 1)}\n\t}\n\treturn result\n}", "func (b *Builder) Build(info base.PickerBuildInfo) gBalancer.Picker {\n\tif len(info.ReadySCs) == 0 {\n\t\t// Block the RPC until a new picker is available via UpdateState().\n\t\treturn base.NewErrPicker(gBalancer.ErrNoSubConnAvailable)\n\t}\n\n\tnodes := make([]selector.Node, 0)\n\tfor conn, info := range info.ReadySCs {\n\t\tins, _ := info.Address.Attributes.Value(\"rawServiceInstance\").(*registry.ServiceInstance)\n\t\tnodes = append(nodes, &grpcNode{\n\t\t\tNode: selector.NewNode(info.Address.Addr, ins),\n\t\t\tsubConn: conn,\n\t\t})\n\t}\n\tp := &Picker{\n\t\tselector: b.builder.Build(),\n\t}\n\tp.selector.Apply(nodes)\n\treturn p\n}", "func newSelector() map[string]string {\n\treturn map[string]string{selectorKey: string(uuid.NewUUID())}\n}", "func CargaComboMagnitudes(Magnitudes CatalogoModel.CatalogoMgo, ID string) string {\n\ttempl := ``\n\n\tif ID != \"\" {\n\t\ttempl = `<option value=\"\">--SELECCIONE--</option>`\n\t} else {\n\t\ttempl = `<option value=\"\" selected>--SELECCIONE--</option>`\n\t}\n\n\tfor _, v := range Magnitudes.Valores {\n\t\tif ID == v.ID.Hex() {\n\t\t\ttempl += `<option value=\"` + v.ID.Hex() + `\" selected>` + v.Valor + `</option>`\n\t\t} else {\n\t\t\ttempl += `<option value=\"` + v.ID.Hex() + `\">` + v.Valor + `</option>`\n\t\t}\n\t}\n\treturn templ\n}", "func (*nodePickerBuilder) Build(info base.PickerBuildInfo) balancer.V2Picker {\n\tif len(info.ReadySCs) == 0 {\n\t\treturn base.NewErrPickerV2(balancer.ErrNoSubConnAvailable)\n\t}\n\n\tvar scs []balancer.SubConn\n\tfor sc := range info.ReadySCs {\n\t\tscs = append(scs, sc)\n\t}\n\n\treturn &nodePicker{\n\t\tsubConns: scs,\n\t}\n}", "func CargaComboListaCostos(ID string) string {\n\t\t\tListaCostos := GetAll()\n\n\t\t\ttempl := ``\n\n\t\t\tif ID != \"\" {\n\t\t\t\ttempl = `<option value=\"\">--SELECCIONE--</option> `\n\t\t\t} else {\n\t\t\t\ttempl = `<option value=\"\" selected>--SELECCIONE--</option> `\n\t\t\t}\n\n\t\t\tfor _, v := range ListaCostos {\n\t\t\t\tif ID == v.ID.Hex() {\n\t\t\t\t\ttempl += `<option value=\" ` + v.ID.Hex() + `\" selected> ` + v.Nombre + ` </option> `\n\t\t\t\t} else {\n\t\t\t\t\ttempl += `<option value=\" ` + v.ID.Hex() + `\"> ` + v.Nombre + ` </option> `\n\t\t\t\t}\n\n\t\t\t}\n\t\t\treturn templ\n\t\t}", "func GenerateSelector(object CommonInterface) (labels.Selector, error) {\n\tls := &metav1.LabelSelector{\n\t\tMatchLabels: GenerateLabels(object),\n\t}\n\treturn metav1.LabelSelectorAsSelector(ls)\n}", "func GenerateSelector(object CommonInterface) (labels.Selector, error) {\n\tls := &metav1.LabelSelector{\n\t\tMatchLabels: GenerateLabels(object),\n\t}\n\treturn metav1.LabelSelectorAsSelector(ls)\n}", "func CargaComboUnidades(ID string) string {\n\tUnidades := GetAll()\n\n\ttempl := ``\n\n\tif ID != \"\" {\n\t\ttempl = `<option value=\"\">--SELECCIONE--</option>`\n\t} else {\n\t\ttempl = `<option value=\"\" selected>--SELECCIONE--</option>`\n\t}\n\n\tfor _, v := range Unidades {\n\t\tif ID == v.ID.Hex() {\n\t\t\ttempl += `<option value=\"` + v.ID.Hex() + `\" selected>` + v.Magnitud + `</option>`\n\t\t} else {\n\t\t\ttempl += `<option value=\"` + v.ID.Hex() + `\">` + v.Magnitud + `</option>`\n\t\t}\n\n\t}\n\treturn templ\n}", "func New() *Bricker {\n\treturn &Bricker{\n\t\tconnection: make(map[string]connector.Connector),\n\t\tfirst: \"\",\n\t\tuids: make(map[uint32]string),\n\t\tsubscriber: make(map[hash.Hash]map[string]Subscriber),\n\t\tchoosers: make([]uint8, 0)}\n}", "func comboBoxTextFinalizer(ct *ComboBoxText) {\n\truntime.SetFinalizer(ct, func(ct *ComboBoxText) { gobject.Unref(ct) })\n}", "func createSelector(values map[string]string) labels.Selector {\n\tselector := labels.NewSelector()\n\tfor k, v := range values {\n\t\treq, err := labels.NewRequirement(k, \"=\", []string{v})\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tselector = selector.Add(*req)\n\t}\n\n\treturn selector\n}", "func NewSelectorRegex(s *Selector) (*SelectorRegex, error) {\n\tsr := new(SelectorRegex)\n\tvar err error\n\tsr.selector = s\n\tsr.groupRegex, err = regexp.Compile(anchorRegex(s.Gvk.Group))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsr.versionRegex, err = regexp.Compile(anchorRegex(s.Gvk.Version))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsr.kindRegex, err = regexp.Compile(anchorRegex(s.Gvk.Kind))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsr.nameRegex, err = regexp.Compile(anchorRegex(s.Name))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsr.namespaceRegex, err = regexp.Compile(anchorRegex(s.Namespace))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn sr, nil\n}", "func NewSelectorCollector(config *Config) *SelectorCollector {\n\tif config == nil {\n\t\treturn nil\n\t}\n\n\tif !config.Verify() || config.Type != COLLECTBYSELECTOR || len(config.ValueRuleMap.Items) < 3 {\n\t\tseelog.Errorf(\"config name:%s is unavailable, please check your collectorConfig.xml\", config.Name)\n\t\treturn nil\n\t}\n\n\tselectorMap := make(map[string][]string)\n\n\tfor _, value := range config.ValueRuleMap.Items {\n\t\tif value.Name == \"\" || value.Rule == \"\" {\n\t\t\tseelog.Errorf(\"config name:%s contains valueRuleMap item with empty name or rule, this item will be ignored.\", config.Name)\n\t\t\tcontinue\n\t\t}\n\n\t\tif value.Name == \"table\" {\n\t\t\tselectorMap[value.Name] = []string{value.Rule}\n\t\t} else if value.Attr != \"\" {\n\t\t\tselectorMap[value.Name] = []string{value.Rule, value.Attr}\n\t\t} else {\n\t\t\tselectorMap[value.Name] = []string{value.Rule}\n\t\t}\n\t}\n\n\t// Most website appear their ip list as table, So table item is required.\n\t// For other situation, you can implement your own method.\n\tif v, ok := selectorMap[\"table\"]; !ok || v[0] == \"\" {\n\t\tseelog.Errorf(\"config name:%s table selector's path should not be empty\", config.Name)\n\t\treturn nil\n\t}\n\n\tparameters := strings.Split(config.UrlParameters, \",\")\n\turls := util.MakeUrls(config.UrlFormat, parameters)\n\treturn &SelectorCollector{\n\t\tconfiguration: config,\n\t\turls: urls,\n\t\tselectorMap: selectorMap,\n\t}\n}", "func NewDefaultSelector(ds datastore.Datastorer) DefaultSelector {\n\treturn DefaultSelector{ds}\n}", "func (t *OpenconfigOfficeAp_System_Logging_RemoteServers_RemoteServer_Selectors) NewSelector(Facility E_OpenconfigSystemLogging_SYSLOG_FACILITY, Severity E_OpenconfigSystemLogging_SyslogSeverity) (*OpenconfigOfficeAp_System_Logging_RemoteServers_RemoteServer_Selectors_Selector, error) {\n\n\t// Initialise the list within the receiver struct if it has not already been\n\t// created.\n\tif t.Selector == nil {\n\t\tt.Selector = make(map[OpenconfigOfficeAp_System_Logging_RemoteServers_RemoteServer_Selectors_Selector_Key]*OpenconfigOfficeAp_System_Logging_RemoteServers_RemoteServer_Selectors_Selector)\n\t}\n\n\tkey := OpenconfigOfficeAp_System_Logging_RemoteServers_RemoteServer_Selectors_Selector_Key{\n\t\tFacility: Facility,\n\t\tSeverity: Severity,\n\t}\n\n\t// Ensure that this key has not already been used in the\n\t// list. Keyed YANG lists do not allow duplicate keys to\n\t// be created.\n\tif _, ok := t.Selector[key]; ok {\n\t\treturn nil, fmt.Errorf(\"duplicate key %v for list Selector\", key)\n\t}\n\n\tt.Selector[key] = &OpenconfigOfficeAp_System_Logging_RemoteServers_RemoteServer_Selectors_Selector{\n\t\tFacility: Facility,\n\t\tSeverity: Severity,\n\t}\n\n\treturn t.Selector[key], nil\n}", "func NewArraySelector(array, index Expression) Expression {\n\treturn &arraySelector{\n\t\tarray: array,\n\t\tindex: index,\n\t}\n}", "func NewDynamicSelect(onKillAction func(), channels []ChannelEntry) *DynamicSelect {\n\t// both aggregators, on close notifier, and internal kill chan.\n\ta := make(chan dsWrapper)\n\tp := make(chan dsWrapper)\n\to := make(chan closeWrapper)\n\td := make(chan interface{})\n\n\t// guarded channels\n\tk := make(chan interface{}, 1)\n\tkg := make(chan interface{}, 1)\n\tl := make(chan []ChannelEntry)\n\tlg := make(chan interface{}, 1)\n\n\t// prime the guards.\n\tkg <- unit\n\tlg <- unit\n\n\treturn &DynamicSelect{\n\t\tonKillAction: onKillAction,\n\t\tload: l,\n\t\tloadGuard: lg,\n\t\tchannels: channels,\n\t\taggregator: a,\n\t\talive: true,\n\t\tdone: d,\n\t\tkill: k,\n\t\tkillGuard: kg,\n\t\tkillHeard: false,\n\t\tpriorityAggregator: p,\n\t\tonClose: o,\n\t}\n}", "func Selector(s selector.Selector) Option {\n\treturn func(o *Options) {\n\t\to.Selector = s\n\t}\n}", "func Selector(s selector.Selector) Option {\n\treturn func(o *Options) {\n\t\to.Selector = s\n\t}\n}", "func (c *Combobox) Append(text string) {\n\tctext := C.CString(text)\n\tC.uiComboboxAppend(c.c, ctext)\n\tfreestr(ctext)\n}", "func (p *parser) nextCombo() token.Token {\n\tt := p.next()\n\tt2 := p.next()\n\tif t.Type == token.RIGHT_CARET && t2.Type == token.RIGHT_CARET {\n\t\treturn token.Token{Type: token.RSHIFT, Pos: t.Pos}\n\t}\n\tp.backup(1)\n\treturn t\n}", "func (me TxsdSelectionAnswerTypeSequenceStyleSuggestion) IsCombobox() bool {\n\treturn me.String() == \"combobox\"\n}", "func (c *Combobox) OnSelected(f func(*Combobox)) {\n\tc.onSelected = f\n}", "func NewCBusOptions(connect bool, smart bool, idmon bool, exstat bool, monitor bool, monall bool, pun bool, pcn bool, srchk bool) *_CBusOptions {\n\treturn &_CBusOptions{Connect: connect, Smart: smart, Idmon: idmon, Exstat: exstat, Monitor: monitor, Monall: monall, Pun: pun, Pcn: pcn, Srchk: srchk}\n}", "func NewPicker(store MediumSourceStorer) *Picker {\n\treturn &Picker{\n\t\tstore: store,\n\t}\n}", "func cellRendererComboFinalizer(cl *CellRendererCombo) {\n\truntime.SetFinalizer(cl, func(cl *CellRendererCombo) { gobject.Unref(cl) })\n}", "func NewColorPicker() ColorPicker {\n\timageColors := make(map[string]Color)\n\n\treturn &colorPicker{\n\t\timageColors: imageColors,\n\t}\n}", "func NewDefaultSelector() (*DefaultSelector, error) {\n\n\ts := &DefaultSelector{\n\t\tsources: make(map[string]*source),\n\t}\n\n\treturn s, nil\n}", "func NewAutocompleterFromPool(pool *redis.Pool, name string) *Autocompleter {\n\treturn &Autocompleter{name: name, pool: pool}\n}", "func DefaultCodesSelector() CodesSelector {\n\tvar sel CodesSelector\n\tsel.FieldSelector.Code = \"code\"\n\tsel.FieldSelector.Perms = \"perms\"\n\treturn sel\n}", "func GuiColorPicker(bounds Rectangle, color Color) Color {\n\tccolor := *color.cptr()\n\tcbounds := *bounds.cptr()\n\tres := C.GuiColorPicker(cbounds, ccolor)\n\treturn newColorFromPointer(unsafe.Pointer(&res))\n}", "func createNodeSelector(nodeName string) *v1.NodeSelector {\n\treturn &v1.NodeSelector{\n\t\tNodeSelectorTerms: []v1.NodeSelectorTerm{\n\t\t\tv1.NodeSelectorTerm{\n\t\t\t\tMatchExpressions: []v1.NodeSelectorRequirement{\n\t\t\t\t\tv1.NodeSelectorRequirement{\n\t\t\t\t\t\tKey: \"kubernetes.io/hostname\",\n\t\t\t\t\t\tOperator: v1.NodeSelectorOpIn,\n\t\t\t\t\t\tValues: []string{nodeName},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func New(prompt string, choices []*Choice) *Selection {\n\treturn &Selection{\n\t\tChoices: choices,\n\t\tPrompt: prompt,\n\t\tFilterPrompt: DefaultFilterPrompt,\n\t\tTemplate: DefaultTemplate,\n\t\tConfirmationTemplate: DefaultConfirmationTemplate,\n\t\tFilter: FilterContainsCaseInsensitive,\n\t\tFilterInputPlaceholderStyle: lipgloss.NewStyle().Foreground(lipgloss.Color(\"240\")),\n\t\tKeyMap: NewDefaultKeyMap(),\n\t\tFilterPlaceholder: DefaultFilterPlaceholder,\n\t\tExtendedTemplateScope: template.FuncMap{},\n\t\tOutput: os.Stdout,\n\t\tInput: os.Stdin,\n\t}\n}", "func (e *exprHelper) NewSelect(operand ast.Expr, field string) ast.Expr {\n\treturn e.exprFactory.NewSelect(e.nextMacroID(), operand, field)\n}", "func (p *parser) peekCombo() token.Token {\n\tt := p.next()\n\tif p.tokensAvailable() > 0 {\n\t\tt2 := p.peek()\n\t\tif t.Type == token.RIGHT_CARET && t2.Type == token.RIGHT_CARET {\n\t\t\tp.backup(1)\n\t\t\treturn token.Token{Type: token.RSHIFT, Pos: t.Pos}\n\t\t}\n\t}\n\tp.backup(1)\n\treturn t\n}", "func NewFieldSelector(base Expression, index int) Expression {\n\treturn &fieldSelector{\n\t\tbase: base,\n\t\tindex: index,\n\t}\n}", "func NewSelectBuilder() *SelectBuilder {\n\treturn &SelectBuilder{}\n}", "func GuiColorPicker(bounds Rectangle, color Color) Color {\n\tcbounds, _ := *(*C.Rectangle)(unsafe.Pointer(&bounds)), cgoAllocsUnknown\n\tccolor, _ := *(*C.Color)(unsafe.Pointer(&color)), cgoAllocsUnknown\n\t__ret := C.GuiColorPicker(cbounds, ccolor)\n\t__v := *newColorRef(unsafe.Pointer(&__ret)).convert()\n\treturn __v\n}", "func init() {\n\tSchemeBuilder.Register(&BucketClass{}, &BucketClassList{})\n}", "func init() {\n\tSchemeBuilder.Register(&BucketClass{}, &BucketClassList{})\n}", "func SelectorLabelsWithComponent(name, instance, component string) map[string]string {\n\tlabels := SelectorLabels(name, instance)\n\tlabels[ApplicationComponentLabelKey] = component\n\n\treturn labels\n}", "func NewCaption(ctx context.Context, uri string) (Caption, error) {\n\n\tu, err := url.Parse(uri)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to parse URI for NewCaption, %w\", err)\n\t}\n\n\tscheme := u.Scheme\n\n\ti, err := captions.Driver(ctx, scheme)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to derive driver for '%s' caption scheme, %w\", scheme, err)\n\t}\n\n\tfn := i.(CaptionInitializeFunc)\n\n\tcaption, err := fn(ctx, uri)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"CaptionInitializeFunc failed, %w\", err)\n\t}\n\n\treturn caption, nil\n}", "func NewPickUp() Pickup {\n p := Pickup{name: \"Pickup\", vehicle: \"Pickup\", speed: 60, capacity: 2, isPrivate: true}\n return p\n}", "func NewSelect(comments Comments, exprs SelectExprs, selectOptions []string, from TableExprs, where *Where, groupBy GroupBy, having *Where) *Select {\n\t////var cache *bool\n\t////var distinct, straightJoinHint, sqlFoundRows bool\n\t//\n\t//for _, option := range selectOptions {\n\t//\tswitch strings.ToLower(option) {\n\t//\tcase DistinctStr:\n\t//\t\tdistinct = true\n\t//\tcase SQLCacheStr:\n\t//\t\ttruth := true\n\t//\t\tcache = &truth\n\t//\tcase SQLNoCacheStr:\n\t//\t\ttruth := false\n\t//\t\tcache = &truth\n\t//\tcase StraightJoinHint:\n\t//\t\tstraightJoinHint = true\n\t//\t\t//case SQLCalcFoundRowsStr:\n\t//\t\t//\tsqlFoundRows = true\n\t//\t}\n\t//}\n\treturn &Select{\n\t\t//Cache: cache,\n\t\tComments: comments,\n\t\t//Distinct: distinct,\n\t\t//StraightJoinHint: straightJoinHint,\n\t\t//SQLCalcFoundRows: sqlFoundRows,\n\t\tSelectExprs: exprs,\n\t\tFrom: from,\n\t\tWhere: where,\n\t\tGroupBy: groupBy,\n\t\tHaving: having,\n\t}\n}", "func newChainClient[C chains.Config, R presenters.ChainResource[C], P TableRenderer, P2 ~[]P](c *Client, name string) ChainClient[C, R, P, P2] {\n\treturn &chainClient[C, R, P, P2]{\n\t\tClient: c,\n\t\tpath: \"/v2/chains/\" + name,\n\t}\n}", "func NewBucket(boptions BucketOptions) (b *Bucket) {\n\tb = new(Bucket)\n\tb.boptions = boptions\n\tb.chs = make(map[string]*Channel, boptions.ChannelSize)\n\treturn\n}", "func MakeSelector(in map[string]string) labels.Selector {\n\tset := make(labels.Set)\n\tfor key, val := range in {\n\t\tset[key] = val\n\t}\n\treturn set.AsSelector()\n}", "func Choice(s *string, choices []string, title, id, class string, valid Validator) (jquery.JQuery, error) {\n\tj := jq(\"<select>\").AddClass(ClassPrefix + \"-choice\").AddClass(class)\n\tj.SetAttr(\"title\", title).SetAttr(\"id\", id)\n\tif *s == \"\" {\n\t\t*s = choices[0]\n\t}\n\tindex := -1\n\tfor i, c := range choices {\n\t\tif c == *s {\n\t\t\tindex = i\n\t\t}\n\t\tj.Append(jq(\"<option>\").SetAttr(\"value\", c).SetText(c))\n\t}\n\tif index == -1 {\n\t\treturn jq(), fmt.Errorf(\"Default of '%s' is not among valid choices\", *s)\n\t}\n\tj.SetData(\"prev\", index)\n\tj.SetProp(\"selectedIndex\", index)\n\tj.Call(jquery.CHANGE, func(event jquery.Event) {\n\t\tnewS := event.Target.Get(\"value\").String()\n\t\tnewIndex := event.Target.Get(\"selectedIndex\").Int()\n\t\tif valid != nil && !valid.Validate(newS) {\n\t\t\tnewIndex = int(j.Data(\"prev\").(float64))\n\t\t\tj.SetProp(\"selectedIndex\", newIndex)\n\t\t}\n\t\t*s = choices[int(newIndex)]\n\t\tj.SetData(\"prev\", newIndex)\n\t})\n\treturn j, nil\n}", "func NewMockisAclSelector_PktSelector(ctrl *gomock.Controller) *MockisAclSelector_PktSelector {\n\tmock := &MockisAclSelector_PktSelector{ctrl: ctrl}\n\tmock.recorder = &MockisAclSelector_PktSelectorMockRecorder{mock}\n\treturn mock\n}", "func (gb *gcpBalancer) regeneratePicker() {\n\tgb.mu.RLock()\n\tdefer gb.mu.RUnlock()\n\n\tif gb.state == connectivity.TransientFailure {\n\t\tgb.picker = newErrPicker(balancer.ErrTransientFailure)\n\t\treturn\n\t}\n\treadyRefs := []*subConnRef{}\n\n\t// Select ready subConns from subConn map.\n\tfor sc, scState := range gb.scStates {\n\t\tif scState == connectivity.Ready {\n\t\t\treadyRefs = append(readyRefs, gb.scRefs[sc])\n\t\t}\n\t}\n\tgb.picker = newGCPPicker(readyRefs, gb)\n}", "func ConstructConsumer(id int, proxy Proxy) Consumer {\n\treturn Consumer{\n\t\tid: id,\n\t\tch: make(chan Widget, MAX_NUM_WIDGETS),\n\t\tproxy: proxy,\n\t}\n}", "func parseSelector(str string) (*types.Selector, error) {\n\tparts := strings.SplitAfterN(str, \":\", 2)\n\tif len(parts) < 2 {\n\t\treturn nil, fmt.Errorf(\"selector \\\"%s\\\" must be formatted as type:value\", str)\n\t}\n\n\ts := &types.Selector{\n\t\t// Strip the trailing delimiter\n\t\tType: strings.TrimSuffix(parts[0], \":\"),\n\t\tValue: parts[1],\n\t}\n\treturn s, nil\n}", "func NewChoice(allowedValues ...string) Choice {\n\treturn Choice{AllowedValues: allowedValues}\n}", "func NewCmdToolboxInstanceSelector(f commandutils.Factory, out io.Writer) *cobra.Command {\n\toptions := &InstanceSelectorOptions{}\n\n\tcommandline := cli.New(\n\t\t\"instance-selector INSTANCE_GROUP\",\n\t\ttoolboxInstanceSelectorShort,\n\t\ttoolboxInstanceSelectorLong,\n\t\ttoolboxInstanceSelectorExample,\n\t\tnil,\n\t)\n\tcommandline.Command.Args = func(cmd *cobra.Command, args []string) error {\n\t\toptions.ClusterName = rootCommand.ClusterName(true)\n\n\t\tif options.ClusterName == \"\" {\n\t\t\treturn fmt.Errorf(\"--name is required\")\n\t\t}\n\n\t\tif len(args) == 0 {\n\t\t\treturn fmt.Errorf(\"must specify name of instance group to create\")\n\t\t}\n\n\t\toptions.InstanceGroupName = args[0]\n\n\t\tif len(args) != 1 {\n\t\t\treturn fmt.Errorf(\"can only create one instance group at a time\")\n\t\t}\n\n\t\tif err := processAndValidateFlags(&commandline); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tsetInstanceSelectorOpts(options, &commandline)\n\n\t\treturn nil\n\t}\n\tcommandline.Command.ValidArgsFunction = func(cmd *cobra.Command, args []string, toComplete string) ([]string, cobra.ShellCompDirective) {\n\t\tcommandutils.ConfigureKlogForCompletion()\n\t\tif len(args) == 1 && rootCommand.ClusterName(false) == \"\" {\n\t\t\treturn []string{\"--name\"}, cobra.ShellCompDirectiveNoFileComp\n\t\t}\n\t\treturn nil, cobra.ShellCompDirectiveNoFileComp\n\t}\n\tcommandline.Command.RunE = func(cmd *cobra.Command, args []string) error {\n\t\treturn RunToolboxInstanceSelector(cmd.Context(), f, out, &commandline, options)\n\t}\n\n\tcpuArchs := []string{cpuArchitectureAMD64, cpuArchitectureARM64}\n\tcpuArchDefault := cpuArchitectureAMD64\n\tplacementGroupStrategies := []string{\"cluster\", \"partition\", \"spread\"}\n\tusageClasses := []string{usageClassSpot, usageClassOndemand}\n\tusageClassDefault := usageClassOndemand\n\toutputDefault := \"yaml\"\n\tdryRunDefault := false\n\tclusterAutoscalerDefault := true\n\tnodeCountMinDefault := 1\n\tnodeCountMaxDefault := 10\n\tmaxResultsDefault := 20\n\n\t// Instance Group Node Configurations\n\n\tcommandline.IntFlag(nodeCountMin, nil, &nodeCountMinDefault, \"Minimum number of nodes\")\n\tcommandline.IntFlag(nodeCountMax, nil, &nodeCountMaxDefault, \"Maximum number of nodes\")\n\tcommandline.IntFlag(nodeVolumeSize, nil, nil, \"Instance volume size (in GiB) for nodes\")\n\tcommandline.StringSliceFlag(nodeSecurityGroups, nil, nil, \"Pre-created additional security groups for nodes\")\n\tcommandline.Command.RegisterFlagCompletionFunc(nodeSecurityGroups, completeSecurityGroup)\n\tcommandline.BoolFlag(clusterAutoscaler, nil, &clusterAutoscalerDefault, \"Add auto-discovery tags for cluster-autoscaler to manage the instance-group\")\n\n\t// Aggregate Filters\n\n\tcommandline.StringFlag(instanceTypeBase, nil, nil, \"Base instance type to retrieve similarly specified instance types\", nil)\n\tcommandline.Command.RegisterFlagCompletionFunc(instanceTypeBase, completeMachineType)\n\tcommandline.BoolFlag(flexible, nil, nil, \"Retrieve a group of instance types spanning multiple generations based on opinionated defaults and user overridden resource filters\")\n\tcommandline.IntFlag(instanceGroupCount, nil, nil, \"Number of instance groups to create with different vcpus-to-memory ratios, starting at 1:2 and doubling\")\n\n\t// Raw Filters\n\n\tcommandline.IntMinMaxRangeFlags(vcpus, nil, nil, \"Number of vcpus available to the instance type\")\n\tcommandline.ByteQuantityMinMaxRangeFlags(memory, nil, nil, \"Amount of memory available (Example: 4gb)\")\n\tcommandline.Command.RegisterFlagCompletionFunc(memory, cobra.NoFileCompletions)\n\tcommandline.Command.RegisterFlagCompletionFunc(memory+\"-min\", cobra.NoFileCompletions)\n\tcommandline.Command.RegisterFlagCompletionFunc(memory+\"-max\", cobra.NoFileCompletions)\n\tcommandline.RatioFlag(vcpusToMemoryRatio, nil, nil, \"Ratio of vcpus to memory in MiB. (Example: 1:2)\")\n\tcommandline.Command.RegisterFlagCompletionFunc(vcpusToMemoryRatio, cobra.NoFileCompletions)\n\tcommandline.StringOptionsFlag(cpuArchitecture, nil, &cpuArchDefault, fmt.Sprintf(\"CPU architecture [%s]\", strings.Join(cpuArchs, \", \")), append(cpuArchs, cpuArchitectureX8664))\n\tcommandline.Command.RegisterFlagCompletionFunc(cpuArchitecture, func(cmd *cobra.Command, args []string, toComplete string) ([]string, cobra.ShellCompDirective) {\n\t\treturn cpuArchs, cobra.ShellCompDirectiveNoFileComp\n\t})\n\tcommandline.IntMinMaxRangeFlags(gpus, nil, nil, \"Number of GPUs (Example: 4)\")\n\tcommandline.ByteQuantityMinMaxRangeFlags(gpuMemory, nil, nil, \"GPUs' total memory (Example: 4gb)\")\n\tcommandline.Command.RegisterFlagCompletionFunc(gpuMemory, cobra.NoFileCompletions)\n\tcommandline.Command.RegisterFlagCompletionFunc(gpuMemory+\"-min\", cobra.NoFileCompletions)\n\tcommandline.Command.RegisterFlagCompletionFunc(gpuMemory+\"-max\", cobra.NoFileCompletions)\n\tcommandline.StringOptionsFlag(placementGroupStrategy, nil, nil, fmt.Sprintf(\"Placement group strategy: [%s]\", strings.Join(placementGroupStrategies, \", \")), placementGroupStrategies)\n\tcommandline.Command.RegisterFlagCompletionFunc(placementGroupStrategy, func(cmd *cobra.Command, args []string, toComplete string) ([]string, cobra.ShellCompDirective) {\n\t\treturn placementGroupStrategies, cobra.ShellCompDirectiveNoFileComp\n\t})\n\tcommandline.StringOptionsFlag(usageClass, nil, &usageClassDefault, fmt.Sprintf(\"Usage class: [%s]\", strings.Join(usageClasses, \", \")), usageClasses)\n\tcommandline.Command.RegisterFlagCompletionFunc(usageClass, func(cmd *cobra.Command, args []string, toComplete string) ([]string, cobra.ShellCompDirective) {\n\t\treturn usageClasses, cobra.ShellCompDirectiveNoFileComp\n\t})\n\tcommandline.BoolFlag(enaSupport, nil, nil, \"Instance types where ENA is supported or required\")\n\tcommandline.BoolFlag(burstSupport, nil, nil, \"Burstable instance types\")\n\tcommandline.StringSliceFlag(subnets, nil, nil, \"Subnet(s) in which to create the instance group. One of Availability Zone like eu-west-1a or utility-eu-west-1a,\")\n\tcommandline.Command.RegisterFlagCompletionFunc(subnets, completeClusterSubnet(f, commandline.Flags[subnets].(*[]string)))\n\tcommandline.IntMinMaxRangeFlags(networkInterfaces, nil, nil, \"Number of network interfaces (ENIs) that can be attached to the instance\")\n\tcommandline.RegexFlag(allowList, nil, nil, \"List of allowed instance types to select from w/ regex syntax (Example: m[3-5]\\\\.*)\")\n\tcommandline.Command.RegisterFlagCompletionFunc(allowList, cobra.NoFileCompletions)\n\tcommandline.RegexFlag(denyList, nil, nil, \"List of instance types which should be excluded w/ regex syntax (Example: m[1-2]\\\\.*)\")\n\tcommandline.Command.RegisterFlagCompletionFunc(denyList, cobra.NoFileCompletions)\n\n\t// Output Flags\n\n\tcommandline.IntFlag(maxResults, nil, &maxResultsDefault, \"Maximum number of instance types to return back\")\n\tcommandline.BoolFlag(dryRun, nil, &dryRunDefault, \"Only print the object that would be created, without creating it. This flag can be used to create a cluster YAML or JSON manifest.\")\n\tcommandline.StringFlag(output, commandline.StringMe(\"o\"), &outputDefault, \"Output format. One of json or yaml. Used with the --dry-run flag.\", nil)\n\tcommandline.Command.RegisterFlagCompletionFunc(output, func(cmd *cobra.Command, args []string, toComplete string) ([]string, cobra.ShellCompDirective) {\n\t\treturn []string{\"json\", \"yaml\"}, cobra.ShellCompDirectiveNoFileComp\n\t})\n\n\treturn commandline.Command\n}", "func s(name, key string) *corev1.SecretKeySelector {\n\treturn &corev1.SecretKeySelector{\n\t\tKey: key,\n\t\tLocalObjectReference: corev1.LocalObjectReference{\n\t\t\tName: name,\n\t\t},\n\t}\n}", "func NewBuilder(opts ...Option) selector.Builder {\n\tvar option options\n\tfor _, opt := range opts {\n\t\topt(&option)\n\t}\n\treturn &selector.DefaultBuilder{\n\t\tBalancer: &Builder{},\n\t\tNode: &direct.Builder{},\n\t}\n}", "func New(o *Opt) *Client {\n\treturn &Client{\n\t\to: o,\n\t}\n}", "func New(opts ...OptionFunc) Currency {\n\tc := Currency{}\n\tc.applyDefaults()\n\tc.Option(opts...)\n\treturn c\n}", "func (c *Combobox) SetSelected(index int) {\n\tC.uiComboboxSetSelected(c.c, C.int(index))\n}", "func New(opts ...OptionFunc) corev1.Container {\n\tc := &container{}\n\tfor _, o := range opts {\n\t\to(c)\n\t}\n\treturn c.asContainer()\n}", "func getSelectorFromString(str string) (labels.Selector, error) {\n\tlabelSelector, err := v1.ParseToLabelSelector(str)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tselector, err := v1.LabelSelectorAsSelector(labelSelector)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn selector, nil\n}", "func InitBrokers(opts ...OptionFunc) interfaces.Broker {\n\tbrokerInst := new(brokerInstance)\n\tfor _, opt := range opts {\n\t\topt(brokerInst)\n\t}\n\n\treturn brokerInst\n}", "func NewSubscriptionsCompanySubscriptionItemRequestBuilderInternal(pathParameters map[string]string, requestAdapter i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.RequestAdapter)(*SubscriptionsCompanySubscriptionItemRequestBuilder) {\n m := &SubscriptionsCompanySubscriptionItemRequestBuilder{\n BaseRequestBuilder: *i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.NewBaseRequestBuilder(requestAdapter, \"{+baseurl}/directory/subscriptions/{companySubscription%2Did}{?%24select,%24expand}\", pathParameters),\n }\n return m\n}", "func SelectorBlueprintsBySelectValue(optionTemplates []OptionTemplate, selector string) []PropertyBlueprint {\n\treturn selectorMetadataByFunc(\n\t\toptionTemplates,\n\t\tselector,\n\t\tfunc(optionTemplate OptionTemplate) string {\n\t\t\treturn optionTemplate.SelectValue\n\t\t})\n}", "func NewSelectorWithMinRequired(ec2api ec2iface.EC2API) *AvailabilityZoneSelector {\n\tavoidZones := map[string]bool{}\n\n\treturn &AvailabilityZoneSelector{\n\t\tec2api: ec2api,\n\t\tstrategy: NewMinRequiredNumberRandomStrategy(),\n\t\trules: []ZoneUsageRule{NewZonesToAvoidRule(avoidZones)},\n\t}\n}", "func NewCompaniesCompanyItemRequestBuilderInternal(pathParameters map[string]string, requestAdapter i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.RequestAdapter)(*CompaniesCompanyItemRequestBuilder) {\n m := &CompaniesCompanyItemRequestBuilder{\n BaseRequestBuilder: *i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.NewBaseRequestBuilder(requestAdapter, \"{+baseurl}/financials/companies/{company%2Did}{?%24select,%24expand}\", pathParameters),\n }\n return m\n}", "func NewConCurrency(limit int) *C {\n\treturn &C{\n\t\twg: &sync.WaitGroup{},\n\t\tch: make(chan struct{}, limit),\n\t\tnumber: limit,\n\t}\n}", "func NewPeerPicker(tracker *Tracker, list *memberlist.Memberlist) PeerPicker {\n\treturn PeerPicker{\n\t\ttracker: tracker,\n\t\tourName: list.LocalNode().Name,\n\t}\n}", "func initCbClient(platformBroker cbPlatformBroker) error {\n\tlog.Println(\"[DEBUG] initCbClient - Initializing the ClearBlade client\")\n\n\tcbBroker.client = cb.NewDeviceClientWithAddrs(*(platformBroker.platformURL), *(platformBroker.messagingURL), *(platformBroker.systemKey), *(platformBroker.systemSecret), *(platformBroker.username), *(platformBroker.password))\n\n\tfor err := cbBroker.client.Authenticate(); err != nil; {\n\t\tlog.Printf(\"[ERROR] initCbClient - Error authenticating %s: %s\\n\", platformBroker.name, err.Error())\n\t\tlog.Println(\"[ERROR] initCbClient - Will retry in 1 minute...\")\n\n\t\t// sleep 1 minute\n\t\ttime.Sleep(time.Duration(time.Minute * 1))\n\t\terr = cbBroker.client.Authenticate()\n\t}\n\n\t//Retrieve adapter configuration data\n\tlog.Println(\"[INFO] main - Retrieving adapter configuration...\")\n\tgetAdapterConfig()\n\n\tlog.Println(\"[DEBUG] initCbClient - Initializing MQTT\")\n\tcallbacks := cb.Callbacks{OnConnectionLostCallback: OnConnectLost, OnConnectCallback: OnConnect}\n\tif err := cbBroker.client.InitializeMQTTWithCallback(platformBroker.clientID, \"\", 30, nil, nil, &callbacks); err != nil {\n\t\tlog.Fatalf(\"[FATAL] initCbClient - Unable to initialize MQTT connection with %s: %s\", platformBroker.name, err.Error())\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *Client) Select(dbname string) (client *Client, err error) {\n\tc.mux.RLock()\n\tif c.config.IsEqualDB(dbname) {\n\t\tc.mux.RUnlock()\n\n\t\treturn c, nil\n\t}\n\n\tconfig, err := c.config.NewWithDB(dbname)\n\tif err != nil {\n\t\tc.mux.RUnlock()\n\t\treturn\n\t}\n\n\tname := config.Name()\n\n\t// first, try loading a client from default manager\n\tclient, err = DefaultMgr.NewClientWithLogger(name)\n\tif err == nil {\n\t\tc.mux.RUnlock()\n\n\t\treturn client, nil\n\t}\n\n\tc.mux.RUnlock()\n\n\t// second, register new client for default manager\n\tc.mux.Lock()\n\tdefer c.mux.Unlock()\n\n\tDefaultMgr.Add(name, config)\n\n\treturn DefaultMgr.NewClientWithLogger(name)\n}", "func New(lib io.CUILib) *CUI {\n\tcatList := component.NewListComponent(\"Categories\")\n\ttaskList := component.NewListComponent(\"Tasks\")\n\tstateBar := component.NewTextComponent(\"State\")\n\tdescBox := component.NewTextComponent(\"Description\")\n\tc := &CUI{\n\t\tCUILib: lib,\n\t\tgrid: component.NewGridComponent(\n\t\t\tmap[string]component.InteractiveComponent{\n\t\t\t\tcomponent.DefaultActivated: catList,\n\t\t\t\t\"h\": catList,\n\t\t\t\t\"<Left>\": catList,\n\t\t\t\t\"l\": taskList,\n\t\t\t\t\"<Right>\": taskList,\n\t\t\t},\n\t\t\tui.NewRow(9.0/10,\n\t\t\t\tui.NewCol(2.0/10, catList),\n\t\t\t\tui.NewCol(8.0/10,\n\t\t\t\t\tui.NewRow(5.0/10, taskList),\n\t\t\t\t\tui.NewRow(5.0/10, descBox),\n\t\t\t\t),\n\t\t\t),\n\t\t\tui.NewRow(1.0/10,\n\t\t\t\tui.NewCol(1.0/1, stateBar),\n\t\t\t),\n\t\t),\n\t\ttaskList: taskList,\n\t\tcatList: catList,\n\t\tstateBar: stateBar,\n\t\tdescBox: descBox,\n\t\tcomponents: []component.Component{\n\t\t\ttaskList,\n\t\t\tcatList,\n\t\t},\n\t}\n\treturn c\n}", "func SelectorFromSet(ls Set) Selector {\n\treturn SelectorFromValidatedSet(ls)\n}", "func NewSelectorWithDefaults(ec2api ec2iface.EC2API) *AvailabilityZoneSelector {\n\tavoidZones := map[string]bool{}\n\n\treturn &AvailabilityZoneSelector{\n\t\tec2api: ec2api,\n\t\tstrategy: NewRecommendedNumberRandomStrategy(),\n\t\trules: []ZoneUsageRule{NewZonesToAvoidRule(avoidZones)},\n\t}\n}" ]
[ "0.6398603", "0.60383093", "0.54640085", "0.5442935", "0.5414595", "0.52587616", "0.5221845", "0.519968", "0.5109608", "0.50104415", "0.49778694", "0.49130982", "0.49013048", "0.47839087", "0.47829998", "0.47143096", "0.47012314", "0.4693571", "0.4685004", "0.46644437", "0.46361268", "0.46007046", "0.45777676", "0.4559086", "0.45439368", "0.45267153", "0.4523236", "0.45232326", "0.45106873", "0.45029345", "0.45025072", "0.44884554", "0.44765624", "0.44765624", "0.4468099", "0.4454933", "0.445182", "0.44454074", "0.44342172", "0.4427904", "0.43926176", "0.43811917", "0.4356124", "0.4339909", "0.43117428", "0.43117428", "0.42903942", "0.42817962", "0.42759323", "0.42707092", "0.42583355", "0.42518604", "0.42463866", "0.4245053", "0.42434102", "0.42254323", "0.41862598", "0.41434732", "0.4134076", "0.41203836", "0.4091885", "0.40773317", "0.40670443", "0.4066715", "0.4064195", "0.40531164", "0.40531164", "0.4047968", "0.40330186", "0.40287432", "0.40186417", "0.3997915", "0.3996922", "0.39865598", "0.3971232", "0.39640373", "0.39636365", "0.3961921", "0.39563188", "0.3945611", "0.39426902", "0.3938693", "0.39359984", "0.39253768", "0.39229873", "0.39219952", "0.3917672", "0.39174914", "0.39151517", "0.39111117", "0.3909353", "0.38984233", "0.38924092", "0.388894", "0.38824466", "0.386233", "0.38613844", "0.3858908", "0.38565004", "0.3847688" ]
0.8261668
0
TODO solve this problem by DFS & WFS
func sumOfLeftLeavesDFS(root *TreeNode) int { sum := 0 var dfs func(r *TreeNode) dfs = func(r *TreeNode) { if r == nil { return } if r.Left != nil && r.Left.Left == nil && r.Left.Right == nil { sum += r.Left.Val } dfs(r.Left) dfs(r.Right) } dfs(root) return sum }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func search(start *node, end *node , sp [676]wordlists, ep [676]wordlists) (*node){\n if start.name == end.name {\n return start\n }\n\n var st_index, en_index uint16\n var _arr_st_index, _arr_en_index []uint16\n var wlen uint8\n\n var cur [110000]node\n var child node\n var ndC uint64\n ndC = 0\n ftr := []node{}\n exp := []node{}\n depth := 0\n\n ftr = append(ftr, *start)\n\n for len(ftr) != 0 {\n cur[ndC], ftr = ftr[0], ftr[1:len(ftr)]\n\n if in_queue(cur[ndC], exp) {\n continue\n }\n\n child.parent = &cur[ndC]\n child.cost = cur[ndC].cost+1\n\n st_index = 26*(uint16(cur[ndC].name[0])-'a') + (uint16(cur[ndC].name[1])-'a')\n if !in_index_arr(st_index, _arr_st_index) {\n _arr_st_index = append(_arr_st_index, st_index)\n\n child.end = false\n for i:=0; i<len(ep[st_index].word); i++ {\n child.name = ep[st_index].word[i]\n if child.name == end.name {\n return &child\n }\n ftr = append(ftr, child)\n } \n } \n\n wlen = uint8(len(cur[ndC].name))\n en_index = 26*(uint16(cur[ndC].name[wlen-2])-'a') + (uint16(cur[ndC].name[wlen-1])-'a')\n if !in_index_arr(en_index, _arr_en_index) {\n _arr_en_index = append(_arr_en_index, en_index)\n\n child.end = true\n for i:=0; i<len(sp[en_index].word); i++ {\n child.name = sp[en_index].word[i]\n if child.name == end.name {\n return &child\n }\n ftr = append(ftr, child)\n } \n }\n if cur[ndC].cost != depth {\n depth = cur[ndC].cost\n }\n exp = append(exp, cur[ndC])\n ndC++\n }\n\n child.name = \"\"\n return &child\n}", "func dfs(curr string, prev string, depth int, orbits map[string][]string) int {\n\tsum := 0\n\tfor _, neighbor := range orbits[curr] {\n\t\tif neighbor == prev {\n\t\t\tcontinue\n\t\t}\n\t\tif neighbor == \"SAN\" {\n\t\t\treturn depth\n\t\t}\n\t\tsum += dfs(neighbor, curr, depth+1, orbits)\n\t}\n\treturn sum\n}", "func wBfs(startNode *node) *node {\n\tvar bfsQueue []*node\n\tbfsQueue = append(bfsQueue, startNode)\n\tcurrNode := startNode\n\tvar endNode *node\n\tfor currNode.nodeType != \"end\" {\n\t\tcurrNode.visited = true\n\t\t//Sorting the weight array of the current node se the cheapest nodes are scanned first\n\t\tstart := 0\n\t\tfor i := 0; i < len(currNode.weights); i++ {\n\t\t\tmin := currNode.weights[start]\n\t\t\tminPos := start\n\t\t\tfor i := start; i < len(currNode.weights); i++ {\n\t\t\t\tnewMin := currNode.weights[i]\n\t\t\t\tif newMin < min {\n\t\t\t\t\tmin = newMin\n\t\t\t\t\tminPos = i\n\t\t\t\t}\n\t\t\t}\n\t\t\t//Swap the min and start in weight and adjacentNodes arrays\n\t\t\tcurrNode.weights[minPos] = currNode.weights[start]\n\t\t\tcurrNode.weights[start] = min\n\n\t\t\thelper := currNode.adjacentNodes[minPos]\n\t\t\tcurrNode.adjacentNodes[minPos] = currNode.adjacentNodes[start]\n\t\t\tcurrNode.adjacentNodes[start] = helper\n\n\t\t\tstart++\n\t\t}\n\t\tfor i := 0; i < len(currNode.adjacentNodes); i++ {\n\t\t\tif !currNode.adjacentNodes[i].visited {\n\t\t\t\tbfsQueue = append(bfsQueue, currNode.adjacentNodes[i])\n\t\t\t\tcurrNode.adjacentNodes[i].parrent = currNode\n\t\t\t}\n\t\t}\n\t\tcurrNode = bfsQueue[1]\n\t\tbfsQueue = bfsQueue[1:]\n\n\t\tif currNode.nodeType == \"end\" {\n\t\t\tendNode = currNode\n\t\t\tbreak\n\t\t}\n\t}\n\treturn endNode\n}", "func BFS(mat [][]int, startPos Position, state State, searchVal, maxDepth int) BFSResult {\n\tres := BFSResult{foodDepth: maxDepth, nearestSnakeDepth: maxDepth}\n\tqueue := list.New()\n\tqueue.PushBack(BFSNode{startPos, 1})\n\tmatrix := duplicateMatrix(mat)\n\t//fmt.Println(matrix)\n\tfor queue.Len() > 0 {\n\t\t// Get next element\n\t\tcurr := queue.Front()\n\t\tnode := curr.Value.(BFSNode)\n\t\tpos := node.pos\n\t\tqueue.Remove(curr)\n\n\t\t// Max depth reached\n\t\tif node.depth > maxDepth { break }\n\n\t\t// Test for in bounds\n\t\tif !isPositionInBounds(pos, state) { continue }\n\n\t\t// Don't look at the same value\n\t\tif matrix[pos.y][pos.x] == 1 { continue }\n\n\t\t// Test if we found closest enemy position\n\t\tif res.nearestSnakeDepth == maxDepth {\n\t\t\tif matrix[pos.y][pos.x] == 2 {\n\t\t\t\tres.nearestSnakeDepth = int(math.Abs(float64(pos.x) - float64(startPos.x)))\n\t\t\t\tres.nearestSnakeDepth = int(math.Abs(float64(pos.y) - float64(startPos.y)))\n\t\t\t}\n\t\t}\n\n\t\tif !isValidPosition(pos, state) { continue }\n\n\t\t// Test if we found the closes food position\n\t\tif res.foodDepth == maxDepth {\n\t\t\tif matrix[pos.y][pos.x] == 3 {\n\t\t\t\tres.foodDepth = int(math.Abs(float64(pos.x) - float64(startPos.x)))\n\t\t\t\tres.foodDepth += int(math.Abs(float64(pos.y) - float64(startPos.y)))\n\t\t\t}\n\t\t}\n\n\t\t// Mark as visited\n\t\tmatrix[pos.y][pos.x] = 1\n\n\t\t// Add next elements to queue\n\t\taddPositionsToQueue(queue, matrix, pos, state, node.depth + 1)\n\n\t\t// Update max depth\n\t\tif node.depth + 1 > res.emptySpaceDepth {\n\t\t\tres.emptySpaceDepth = node.depth + 1\n\t\t}\n\t}\n\n\treturn res\n}", "func solve(board [][]byte) {\n\t// 并查集\n\t//if len(board) == 0 {\n\t//\treturn\n\t//}\n\t//m, n := len(board), len(board[0])\n\t//parent, di, dj := make([]int, m*n+1), [4]int{-1, 0, 0, 1}, [4]int{0, -1, 1, 0}\n\t//for i, _ := range parent {\n\t//\tparent[i] = i\n\t//}\n\t//find := func(parent []int, p int) int {\n\t//\tfor p != parent[p] {\n\t//\t\tparent[p], p = parent[parent[p]], parent[parent[p]]\n\t//\t}\n\t//\treturn p\n\t//}\n\t//union := func(parent []int, p, q int) {\n\t//\trootP, rootQ := find(parent, p), find(parent, q)\n\t//\tif rootP != rootQ {\n\t//\t\tparent[rootP] = rootQ\n\t//\t}\n\t//}\n\t//var dfs func(i, j int)\n\t//dfs = func(i, j int) {\n\t//\tfmt.Println(i, j)\n\t//\tif i < 0 || i >= m || j < 0 || j >= n || board[i][j] == 'X' || find(parent, i*n+j) == len(parent)-1 {\n\t//\t\treturn\n\t//\t}\n\t//\tunion(parent, i*n+j, len(parent)-1)\n\t//\tfor k := 0; k < 4; k++ {\n\t//\t\tdfs(i+di[k], j+dj[k])\n\t//\t}\n\t//}\n\t//for j := 0; j < n; j++ {\n\t//\tif board[0][j] == 'O' {\n\t//\t\tdfs(0, j)\n\t//\t}\n\t//}\n\t//for i := 1; i < m; i++ {\n\t//\tif board[i][0] == 'O' {\n\t//\t\tdfs(i, 0)\n\t//\t}\n\t//}\n\t//for i := 1; i < m; i++ {\n\t//\tif board[i][n-1] == 'O' {\n\t//\t\tdfs(i, n-1)\n\t//\t}\n\t//}\n\t//for j := 1; j < n-1; j++ {\n\t//\tif board[m-1][j] == 'O' {\n\t//\t\tdfs(m-1, j)\n\t//\t}\n\t//}\n\t//for i := 0; i < m; i++ {\n\t//\tfor j := 0; j < n; j++ {\n\t//\t\tif board[i][j] == 'O' && find(parent, i*n+j) != len(parent)-1 {\n\t//\t\t\tboard[i][j] = 'X'\n\t//\t\t}\n\t//\t}\n\t//}\n\n\t// BFS:不会OOM,重点:四连通 添加了很多重复的元素到 队列 中\n\tif len(board) == 0 {\n\t\treturn\n\t}\n\tm, n := len(board), len(board[0])\n\tqueue, di, dj := list.New(), [4]int{-1, 0, 0, 1}, [4]int{0, -1, 1, 0}\n\tvisited := make([][]bool, m) // BFS oom 的关键不是队列中的元素多了,而是 四连通 添加了很多重复的元素到 队列 中\n\tfor i := 0; i < m; i++ {\n\t\tvisited[i] = make([]bool, n)\n\t}\n\tbfs := func(i, j int) {\n\t\tqueue.PushBack([2]int{i, j}) // 采用 list 作为队列\n\t\tvisited[i][j] = true\n\t\tfor queue.Len() > 0 {\n\t\t\tcurr := queue.Remove(queue.Front()).([2]int) // 新写法\n\t\t\tfor k := 0; k < 4; k++ {\n\t\t\t\tnI, nJ := curr[0]+di[k], curr[1]+dj[k]\n\t\t\t\tif nI < 0 || nI >= m || nJ < 0 || nJ >= n || board[nI][nJ] == 'X' || visited[nI][nJ] {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tqueue.PushBack([2]int{nI, nJ})\n\t\t\t\tvisited[nI][nJ] = true\n\t\t\t}\n\t\t}\n\t}\n\tfor i := 0; i < m; i++ {\n\t\tfor j := 0; j < n; j++ {\n\t\t\tif board[i][j] == 'O' && (i == 0 || i == m-1 || j == 0 || j == n-1) {\n\t\t\t\tbfs(i, j)\n\t\t\t}\n\t\t}\n\t}\n\tfor i := 0; i < m; i++ {\n\t\tfor j := 0; j < n; j++ {\n\t\t\tif board[i][j] == 'O' && !visited[i][j] {\n\t\t\t\tboard[i][j] = 'X'\n\t\t\t}\n\t\t}\n\t}\n\n\t// BFS:难点,out of memory(太多O),重点:见上一种 BFS\n\t//if len(board) == 0 {\n\t//\treturn\n\t//}\n\t//m, n := len(board), len(board[0])\n\t//queue, di, dj := make([]Point, 0), [4]int{-1, 0, 0, 1}, [4]int{0, -1, 1, 0}\n\t//var bfs func()\n\t//bfs = func() {\n\t//\t//for idx := 0; idx < len(queue); idx++ {\t// out of memory(太多O),\n\t//\tfor len(queue) > 0 { // out of memory(太多O),那么就要换种写法\n\t//\t\tlength := len(queue)\n\t//\t\tfor idx := 0; idx < length; idx++ {\n\t//\t\t\ti, j := queue[idx].i, queue[idx].j\n\t//\t\t\tboard[i][j] = '!'\n\t//\t\t\tfor k := 0; k < 4; k++ {\n\t//\t\t\t\tnI, nJ := i+di[k], j+dj[k]\n\t//\t\t\t\tif nI < 0 || nI >= m || nJ < 0 || nJ >= n || board[nI][nJ] == 'X' || board[nI][nJ] == '!' {\n\t//\t\t\t\t\tcontinue\n\t//\t\t\t\t}\n\t//\t\t\t\tqueue = append(queue, Point{nI, nJ})\t// 作为 四连通 添加重复元素的案例\n\t//\t\t\t}\n\t//\t\t}\n\t//\t\tqueue = queue[length:]\n\t//\t}\n\t//}\n\t//for i := 0; i < m; i++ {\n\t//\tfor j := 0; j < n; j++ {\n\t//\t\tif board[i][j] == 'O' && (i == 0 || i == m-1 || j == 0 || j == n-1) {\n\t//\t\t\tqueue = append(queue, Point{i, j})\n\t//\t\t\tbfs()\n\t//\t\t}\n\t//\t}\n\t//}\n\t//for i := 0; i < m; i++ {\n\t//\tfor j := 0; j < n; j++ {\n\t//\t\tswitch board[i][j] {\n\t//\t\tcase '!':\n\t//\t\t\tboard[i][j] = 'O'\n\t//\t\tcase 'O':\n\t//\t\t\tboard[i][j] = 'X'\n\t//\t\t}\n\t//\t}\n\t//}\n\n\t// DFS\n\t//m := len(board) - 1\n\t//if m == -1 {\n\t//\treturn\n\t//}\n\t//n := len(board[0]) - 1\n\t//dx, dy := [4]int{-1, 0, 0, 1}, [4]int{0, -1, 1, 0}\n\t//var dfs func(i, j int)\n\t//dfs = func(i, j int) {\n\t//\tif i < 0 || i > m || j < 0 || j > n || board[i][j] == 'X' || board[i][j] == '!' {\n\t//\t\treturn\n\t//\t}\n\t//\tboard[i][j] = '!'\n\t//\tfor k := 0; k < 4; k++ {\n\t//\t\tdfs(i+dx[k], j+dy[k])\n\t//\t}\n\t//}\n\t//for i := 0; i <= m; i++ {\n\t//\tfor j := 0; j <= n; j++ {\n\t//\t\tif i == 0 || i == m || j == 0 || j == n && board[i][j] == 'O' { // 处理边界\n\t//\t\t\tdfs(i, j)\n\t//\t\t}\n\t//\t}\n\t//}\n\t//for i := 0; i <= m; i++ {\n\t//\tfor j := 0; j <= n; j++ {\n\t//\t\tswitch board[i][j] { // 处理中央和边界\n\t//\t\tcase '!':\n\t//\t\t\tboard[i][j] = 'O'\n\t//\t\tcase 'O':\n\t//\t\t\tboard[i][j] = 'X'\n\t//\t\t}\n\t//\t}\n\t//}\n\n\t// DFS:失败\n\t//if len(board) == 0 {\n\t//\treturn\n\t//}\n\t//r, c := len(board), len(board[0])\n\t//dx, dy := [4]int{-1, 0, 0, 1}, [4]int{0, -1, 1, 0}\n\t//visited := make([]bool, r*c) // 加 visited 也解决不了“封闭”的bug\n\t//var dfs func(i, j int) // 有返回值的情况,行不通,因为当 [i,j] 3面都是 X 时,进来的地方,进来的地方和 [i,j] 互相依赖,无法判断\n\t//dfs = func(i, j int) {\n\t//\tfmt.Println(i, j)\n\t//\tif board[i][j] == 'X' { // 这里也要判断 visited[i*c+j],因为前面可能提前触到边界返回了\n\t//\t\treturn\n\t//\t}\n\t//\tvisited[i*c+j] = true\n\t//\t//ans := false\n\t//\tfor k := 0; k < 4; k++ {\n\t//\t\tnI, nJ := i+dx[k], j+dy[k]\n\t//\t\tif nI < 0 || nI == r || nJ < 0 || nJ == c {\n\t//\t\t\treturn\n\t//\t\t}\n\t//\t\tif visited[i*c+j] {\n\t//\t\t\tcontinue\n\t//\t\t}\n\t//\t\tdfs(nI, nJ)\n\t//\t}\n\t//\tboard[i][j] = 'X'\n\t//}\n\t//for i := 0; i < r; i++ {\n\t//\tfor j := 0; j < c; j++ {\n\t//\t\tif board[i][j] == 'O' && !visited[i*c+j] {\n\t//\t\t\tdfs(i, j)\n\t//\t\t}\n\t//\t}\n\t//}\n}", "func dfs_eventualSafeNodes(end_start [][]int,cur int,is_safe []bool,visited []bool){\n\tif len(end_start[cur]) == 0{\n\t\treturn\n\t}\n\tif visited[cur]{\n\t\treturn\n\t}\n\tvisited[cur] = true\n\tfor _,next := range end_start[cur]{\n\t\tis_safe[next] = false\n\t\tdfs_eventualSafeNodes(end_start,next,is_safe,visited)\n\t}\n}", "func (f *Func) dfsOrig(entry *Block, succFn linkedBlocks, semi, vertex, label, parent []ID) ID {\n\tn := ID(0)\n\ts := make([]*Block, 0, 256)\n\ts = append(s, entry)\n\n\tfor len(s) > 0 {\n\t\tv := s[len(s)-1]\n\t\ts = s[:len(s)-1]\n\t\t// recursing on v\n\n\t\tif semi[v.ID] != 0 {\n\t\t\tcontinue // already visited\n\t\t}\n\t\tn++\n\t\tsemi[v.ID] = n\n\t\tvertex[n] = v.ID\n\t\tlabel[v.ID] = v.ID\n\t\t// ancestor[v] already zero\n\t\tfor _, e := range succFn(v) {\n\t\t\tw := e.b\n\t\t\t// if it has a dfnum, we've already visited it\n\t\t\tif semi[w.ID] == 0 {\n\t\t\t\t// yes, w can be pushed multiple times.\n\t\t\t\ts = append(s, w)\n\t\t\t\tparent[w.ID] = v.ID // keep overwriting this till it is visited.\n\t\t\t}\n\t\t}\n\t}\n\treturn n\n}", "func dfs(cur, parent int, adjs [][]int) (int, int) {\n\tnodes, cuts := 1, 0\n\tfor _, adj := range adjs[cur] {\n\t\tif adj == parent {\n\t\t\tcontinue\n\t\t}\n\t\tadjNodes, adjCuts := dfs(adj, cur, adjs)\n\t\tnodes += adjNodes\n\t\tcuts += adjCuts\n\t\tif adjNodes%2 == 0 {\n\t\t\tcuts++\n\t\t}\n\t}\n\treturn nodes, cuts\n}", "func initAdjMap(pxMap [][]bool) {\n\tfor x := 0; x < sizeX; x++ {\n\t\tfor y := 0; y < sizeY; y++ {\n\t\t\tif pxMap[x][y] && nodeMap[x][y].nodeType != \"dEnd\" && nodeMap[x][y].nodeType != \"path\" {\n\t\t\t\t//scanning x++\n\t\t\t\tix := x\n\t\t\t\tiy := y\n\t\t\t\tfor pxMap[ix][y] {\n\t\t\t\t\tif ix+1 < sizeX {\n\t\t\t\t\t\tix++\n\t\t\t\t\t} else {\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\n\t\t\t\t\tif nodeMap[ix][iy].nodeType != \"dEnd\" && nodeMap[ix][iy].nodeType != \"wall\" && nodeMap[ix][iy].nodeType != \"path\" {\n\t\t\t\t\t\tnodeMap[x][y].weights = append(nodeMap[x][y].weights, ix-x)\n\t\t\t\t\t\tnodeMap[x][y].adjacentNodes = append(nodeMap[x][y].adjacentNodes, &nodeMap[ix][iy])\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\t//scanning y++\n\t\t\t\tix = x\n\t\t\t\tiy = y\n\t\t\t\tfor pxMap[x][iy] {\n\t\t\t\t\tif iy+1 < sizeY {\n\t\t\t\t\t\tiy++\n\t\t\t\t\t} else {\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t\tif nodeMap[ix][iy].nodeType != \"dEnd\" && nodeMap[ix][iy].nodeType != \"wall\" && nodeMap[ix][iy].nodeType != \"path\" {\n\t\t\t\t\t\tnodeMap[x][y].weights = append(nodeMap[x][y].weights, iy-y)\n\t\t\t\t\t\tnodeMap[x][y].adjacentNodes = append(nodeMap[x][y].adjacentNodes, &nodeMap[ix][iy])\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\t//scanning x--\n\t\t\t\tix = x\n\t\t\t\tiy = y\n\t\t\t\tfor pxMap[ix][y] {\n\t\t\t\t\tif ix-1 >= 0 {\n\t\t\t\t\t\tix--\n\t\t\t\t\t} else {\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\n\t\t\t\t\tif nodeMap[ix][iy].nodeType != \"dEnd\" && nodeMap[ix][iy].nodeType != \"wall\" && nodeMap[ix][iy].nodeType != \"path\" {\n\t\t\t\t\t\tnodeMap[x][y].weights = append(nodeMap[x][y].weights, x-ix)\n\t\t\t\t\t\tnodeMap[x][y].adjacentNodes = append(nodeMap[x][y].adjacentNodes, &nodeMap[ix][iy])\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\t//scanning y--\n\t\t\t\tix = x\n\t\t\t\tiy = y\n\t\t\t\tfor pxMap[x][iy] {\n\t\t\t\t\tif iy-1 >= 0 {\n\t\t\t\t\t\tiy--\n\t\t\t\t\t} else {\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t\tif nodeMap[ix][iy].nodeType != \"dEnd\" && nodeMap[ix][iy].nodeType != \"wall\" && nodeMap[ix][iy].nodeType != \"path\" {\n\t\t\t\t\t\tnodeMap[x][y].weights = append(nodeMap[x][y].weights, y-iy)\n\t\t\t\t\t\tnodeMap[x][y].adjacentNodes = append(nodeMap[x][y].adjacentNodes, &nodeMap[ix][iy])\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func dfs(grid [][]byte, M int, N int, x int, y int) {\n\tif x < 0 || x >= M || y < 0 || y >= N {\n\t\t// Out of bound, do nothing.\n\t\treturn\n\t}\n\n\tif grid[y][x] == '0' {\n\t\t// Visited already, do nothing.\n\t\treturn\n\t}\n\tgrid[y][x] = '0'\n\n\tdfs(grid, M, N, x-1, y) // Go left\n\tdfs(grid, M, N, x, y-1) // Go up\n\tdfs(grid, M, N, x+1, y) // Go right\n\tdfs(grid, M, N, x, y+1) // Go down\n}", "func searchDiagnosisPaths(graph *simple.DirectedGraph, nodeCount int) ([][]int64, error) {\n\tvar queue NodeQueue\n\tvisited := make([]bool, nodeCount)\n\tnodePathCache := make([][][]int64, nodeCount)\n\tsinkNodes := make([]int64, 0)\n\n\t// Validate the graph contains start node with id of 0.\n\tstart := graph.Node(0)\n\tif start == nil {\n\t\treturn nil, fmt.Errorf(\"start node not found in graph\")\n\t}\n\n\t// Set start node as visited and enqueue all nodes that can reach directly from it.\n\tvisited[start.ID()] = true\n\tfromNodes := graph.From(start.ID())\n\tfor fromNodes.Next() {\n\t\tfromNode := fromNodes.Node()\n\t\tqueue.Enqueue(fromNode)\n\t}\n\n\t// Initialize node path cache with start node.\n\tnodePaths := make([][]int64, 0)\n\tnodePaths = append(nodePaths, []int64{start.ID()})\n\tnodePathCache[start.ID()] = nodePaths\n\n\tfor queue.Len() != 0 {\n\t\t// Dequeue a node from queue and retrieve all nodes that can reach directly to or from current node.\n\t\tcurrent := queue.Dequeue()\n\t\ttoNodes := graph.To(current.ID())\n\t\tfromNodes := graph.From(current.ID())\n\n\t\t// Skip current node if it has already been visited.\n\t\tif visited[current.ID()] {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Set current node as visited if all nodes that can reach directly to current node are visited.\n\t\t// Otherwise, enqueue current node.\n\t\tvisited[current.ID()] = true\n\t\tfor toNodes.Next() {\n\t\t\ttoNode := toNodes.Node()\n\t\t\tif !visited[toNode.ID()] {\n\t\t\t\tvisited[current.ID()] = false\n\t\t\t\tqueue.Enqueue(current)\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif visited[current.ID()] {\n\t\t\t// Update node path of current node with visited node that can reach directly to current node.\n\t\t\ttoNodes.Reset()\n\t\t\tfor toNodes.Next() {\n\t\t\t\ttoNode := toNodes.Node()\n\t\t\t\tnodePaths := nodePathCache[current.ID()]\n\t\t\t\tif nodePaths == nil {\n\t\t\t\t\tnodePaths = make([][]int64, 0)\n\t\t\t\t}\n\t\t\t\ttoNodePaths := nodePathCache[toNode.ID()]\n\t\t\t\tfor _, toNodePath := range toNodePaths {\n\t\t\t\t\tnodePath := make([]int64, len(toNodePath))\n\t\t\t\t\tcopy(nodePath, toNodePath)\n\t\t\t\t\tnodePath = append(nodePath, current.ID())\n\t\t\t\t\tnodePaths = append(nodePaths, nodePath)\n\t\t\t\t}\n\t\t\t\t// Node path appended by current node is updated as node path of current node.\n\t\t\t\tnodePathCache[current.ID()] = nodePaths\n\t\t\t}\n\n\t\t\t// Enqueue all nodes that can reach directly from current node if current node is visited.\n\t\t\tsink := true\n\t\t\tfor fromNodes.Next() {\n\t\t\t\tsink = false\n\t\t\t\tfromNode := fromNodes.Node()\n\t\t\t\tqueue.Enqueue(fromNode)\n\t\t\t}\n\t\t\t// Set current node as sink if its outdegree is 0.\n\t\t\tif sink {\n\t\t\t\tsinkNodes = append(sinkNodes, current.ID())\n\t\t\t}\n\t\t}\n\t}\n\n\t// Set diagnosis paths with all node paths of nodes which has outdegree of 0.\n\tdiagnosisPaths := make([][]int64, 0)\n\tfor _, id := range sinkNodes {\n\t\tpaths := nodePathCache[id]\n\t\tdiagnosisPaths = append(diagnosisPaths, paths...)\n\t}\n\n\treturn diagnosisPaths, nil\n}", "func DFS(graph map[int]*Vertex, i *Vertex, pass int) {\n\n\ti.Explored = true\n\n\tif pass == 2 {\n\t\ts.AddMember(i.ID)\n\t}\n\n\tfor _, v := range i.Edges {\n\n\t\tvertex := graph[v]\n\n\t\tif !vertex.Explored {\n\t\t\tDFS(graph, vertex, pass)\n\t\t}\n\t}\n\n\tif pass == 1 {\n\t\tt++\n\t\tmagicalOrderMap[t] = i.ID\n\t}\n\n}", "func (m *Map) Bfs(loc Location, obj Item) LocDir {\n\tisGoal := func(curLoc Location) bool {\n\t\tif m.Item(curLoc) == obj {\n\t\t\treturn true\n\t\t}\n\t\treturn false\n\t}\n\t//ret acts as the default return when either nothing is found\n\t//or the goal state is the cell.\n\tret := LocDir{loc, NoMovement}\n\tret.Loc = -1\n\tret.Dir = NoMovement\n\n\tvar depth int\n\tfrontier := new(Queue) // make this Queue a type of LocDir\n\tvar inFrontier = make(map[Location]bool)\n\tvar explored = make(map[Location]bool) // the keys are only locations\n\n\tif isGoal(loc) {\n\t\treturn ret\n\t}\n\n\tfrontier.Push(ret) // frontier is queue of LocDir\n\tinFrontier[loc] = true // keys to inFrontier are simply locations\n\t// I'm not sure whether I should set the keys to frontier to be a LocDir\n\t// as well.\n\n\tfor {\n\t\tif frontier.Len() == 0 || depth > 10 {\n\t\t\treturn ret\n\t\t}\n\n\t\tcurLoc := frontier.Pop().(LocDir)\n\t\tinFrontier[curLoc.Loc] = false\n\t\texplored[curLoc.Loc] = true\n\n\t\t// Loop over adjacent Locations, action is a LocDir structure\n\t\tfor _, action := range m.Adjacent(curLoc.Loc) {\n\t\t\t//if child not in explored or frontier\n\t\t\tif !explored[action.Loc] || !inFrontier[action.Loc] {\n\t\t\t\tif isGoal(action.Loc) {\n\t\t\t\t\treturn action\n\t\t\t\t}\n\t\t\t\tfrontier.Push(action)\n\t\t\t\tinFrontier[action.Loc] = true\n\t\t\t}\n\t\t}\n\t\tdepth++\n\t}\n\treturn ret\n}", "func bild() {\n\n root.word = words[0]\n d := 0 \n var( \n tmp ,tmp2 *Node\n )\n\n for i := range words{\n if i == 0 {continue}\n if words[i] == \"\" {continue}\n \n tmp = new(Node)\n (*tmp).word = words[i]\n\n d = dist(root.word, (*tmp).word)\n if d == 0{\n //fmt.Printf(\"problem %s %d\",words[i],i)\n }\n tmp2 = &root\n\n for ;;{\n \n d = dist(((*tmp).word),((*tmp2).word))\n\n if ((*tmp2).child)[d] == nil {\n ((*tmp2).child)[d] = tmp\n break\n } else{\n tmp2 = ((*tmp2).child)[d]\n } \n } \n }\n}", "func dfs(start, cur string, stack []string, adj map[string]map[string][]*gographviz.Edge) map[string]bool {\n\t// Check for cycle or cross.\n\tif in(stack, cur) {\n\t\tif cur == start {\n\t\t\treturn set(stack) // Found a cycle, return the current stack as a set.\n\t\t}\n\t\treturn nil // Found a cross, just return.\n\t}\n\n\tr := map[string]bool{}\n\tchildStack := append(stack, cur)\n\t// Loop over all possible destinations of cur.\n\tfor dst := range adj[cur] {\n\t\t// Add all nodes that are in a cycle as found by the recursive call\n\t\t// to dfs.\n\t\tunion(r, dfs(start, dst, childStack, adj))\n\t}\n\treturn r\n}", "func bfs(adj [][]bool, begin, end point) int {\n\tfr := new(frontier)\n\tvisited := make(map[point]bool)\n\theap.Push(fr, step{begin, 0})\n\tfor fr.Len() != 0 {\n\t\tnext := heap.Pop(fr).(step)\n\t\tif next.point == end {\n\t\t\treturn next.steps\n\t\t}\n\t\tfor _, p := range neighbors(adj, next.point) {\n\t\t\tif !visited[p] {\n\t\t\t\theap.Push(fr, step{p, next.steps + 1})\n\t\t\t\tvisited[p] = true\n\t\t\t}\n\t\t}\n\t}\n\treturn -1\n}", "func dfSUtil(v int, visited []bool, adj_list [][]int, conn map[int]int) {\n\n\tvisited[v] = true\n\n\tconn[v] = v\n\n\tadj := adj_list[v-1]\n\n\tfor i := 0; i < len(adj); i++ {\n\t\tif !visited[adj[i]] {\n\t\t\tdfSUtil(adj[i], visited, adj_list, conn)\n\t\t}\n\t}\n}", "func DFS(currentNode *BasicBlock, nodes []*UnionFindNode, number map[*BasicBlock]int, last []int, current int) int {\n\tnodes[current].Init(currentNode, current)\n\tnumber[currentNode] = current\n\n\tlastid := current\n\tfor ll := currentNode.OutEdges().Front(); ll != nil; ll = ll.Next() {\n\t\tif target := ll.Value.(*BasicBlock); number[target] == unvisited {\n\t\t\tlastid = DFS(target, nodes, number, last, lastid+1)\n\t\t}\n\t}\n\tlast[number[currentNode]] = lastid\n\treturn lastid\n}", "func (graph *Graph) search(multi bool) bool {\n\n\tdefer func() {\n\t\tgraph.openHeap.Clear()\n\t}()\n\n\tif graph.start == nil {\n\t\tpanic(\"not set start point\")\n\t}\n\n\tif graph.end == nil {\n\t\tpanic(\"not set end point\")\n\t}\n\n\tstartTile := graph.Tiles[graph.start.Y][graph.start.X]\n\tgraph.start.Attr = startTile.Attr\n\tstartTile.IsCount = true\n\tstartTile.Attr = START\n\n\tendTile := graph.Tiles[graph.end.Y][graph.end.X]\n\tgraph.end.Attr = endTile.Attr\n\tendTile.IsCount = false\n\tendTile.Attr = END\n\n\tgraph.openHeap.Put(startTile)\n\n\tfor {\n\t\tif itile, ok := graph.openHeap.Pop(); ok {\n\t\t\ttile := itile.(*Tile)\n\n\t\t\tif tile == endTile {\n\n\t\t\t\tgraph.pathlist = make([]Path, 0)\n\n\t\t\t\tvar path Path\n\t\t\t\tpath = append(path, tile)\n\t\t\t\tif multi {\n\t\t\t\t\tgraph.multiPath(tile, startTile, path)\n\t\t\t\t\tsort.Sort(graph.pathlist)\n\t\t\t\t} else {\n\t\t\t\t\tgraph.singlePath(tile, startTile, path)\n\t\t\t\t}\n\n\t\t\t\treturn true\n\t\t\t}\n\n\t\t\tfor _, ntile := range graph.neighbor.GetNeighbor(graph, tile) {\n\t\t\t\tif ntile.IsCount == false && ntile.Attr != BLOCK {\n\t\t\t\t\tgraph.countCost.Cost(graph, ntile, tile)\n\t\t\t\t\tgraph.countWeight.Weight(graph, ntile, tile)\n\t\t\t\t\tntile.IsCount = true\n\t\t\t\t\t// 处理ntile权值\n\t\t\t\t\tgraph.openHeap.Put(ntile)\n\t\t\t\t}\n\t\t\t}\n\n\t\t} else {\n\t\t\t// log.Println(\"path can not found\")\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn false\n}", "func (f *FibonacciHeap) consolidate() {\n\tvar used = make([]*Node, f.size) //We use a slice to track collisions in node.degree\n\tused[f.min.degree] = f.min\n\tcur := f.min.right\n\n\tfor used[cur.degree] != cur { //We always go right, so if we placed something to slice, made a lap, and nothing changed, consolidation is finished\n\t\tf.updateMin(cur)\n\t\tif used[cur.degree] == nil { //If yet no other node with same degree recorder - record current\n\t\t\tused[cur.degree] = cur\n\t\t\tcur = cur.right\n\t\t} else {\n\n\t\t\tbusy := used[cur.degree]\n\t\t\tfather, son := cur, busy\n\t\t\tif busy.key.LessThen(cur.key) { //make father point to lighter node, son to heavier one\n\t\t\t\tfather, son = son, father\n\t\t\t} else if father.key.EqualsTo(son.key) { //make sure f.min is always father\n\t\t\t\tif son == f.min {\n\t\t\t\t\tfather, son = son, father\n\t\t\t\t}\n\t\t\t}\n\t\t\tson.left.right = son.right\n\t\t\tson.right.left = son.left //cut the son from his local linked-list\n\n\t\t\tnext := cur.right //remember next to be right from current cur, it can change later\n\n\t\t\tif father.child == nil { //If father has no children - son is the first\n\t\t\t\tfather.child = son\n\t\t\t\tson.left, son.right = son, son\n\t\t\t} else { //else integrate son into children linked-list\n\t\t\t\tson.left, son.right = father.child, father.child.right\n\t\t\t\tfather.child.right.left = son\n\t\t\t\tfather.child.right = son\n\t\t\t}\n\n\t\t\tused[cur.degree] = nil\n\t\t\tson.parent = father\n\t\t\tfather.degree++\n\t\t\tcur = next\n\t\t}\n\t\tf.updateMin(cur)\n\t}\n\n}", "func walker(state []int, results chan<- []int, wg *sync.WaitGroup) {\n\tdefer wg.Done()\n\tif len(state) > 3 {\n\t\tresults <- state\n\t}\n\n\t/*\n\t\tlastDot is the index of the last dot in the current walker function\n\t\tstate array.\n\t*/\n\tlastDot := state[len(state)-1]\n\n\tswitch getDotType(lastDot) {\n\tcase CORNER:\n\t\tfor i := 0; i < 9; i++ {\n\t\t\tif !contains(state, i) {\n\t\t\t\tns := append(append(make([]int, 0), state...), i)\n\t\t\t\tif getDotType(i) != CORNER {\n\t\t\t\t\twg.Add(1)\n\t\t\t\t\tgo walker(ns, results, wg)\n\t\t\t\t} else {\n\t\t\t\t\tmiddleDot := getMiddle(lastDot, i)\n\t\t\t\t\tif contains(state, middleDot) {\n\t\t\t\t\t\twg.Add(1)\n\t\t\t\t\t\tgo walker(ns, results, wg)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\tcase SIDE:\n\t\tfor i := 0; i < 9; i++ {\n\t\t\tif !contains(state, i) {\n\t\t\t\tns := append(append(make([]int, 0), state...), i)\n\t\t\t\tif getDotType(i) != SIDE {\n\t\t\t\t\twg.Add(1)\n\t\t\t\t\tgo walker(ns, results, wg)\n\t\t\t\t} else if lastDot+i == 8 {\n\t\t\t\t\tif contains(state, 4) {\n\t\t\t\t\t\twg.Add(1)\n\t\t\t\t\t\tgo walker(ns, results, wg)\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\twg.Add(1)\n\t\t\t\t\tgo walker(ns, results, wg)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\tcase CENTER:\n\t\tfor i := 0; i < 9; i++ {\n\t\t\tif !contains(state, i) {\n\t\t\t\tns := append(append(make([]int, 0), state...), i)\n\t\t\t\twg.Add(1)\n\t\t\t\tgo walker(ns, results, wg)\n\t\t\t}\n\t\t}\n\t}\n}", "func computeLocationPath(fvdl FVDL, input int) []int {\n\tlog.Entry().Debug(\"Computing for ID \", input)\n\t// Find the successors of input\n\tvar subnodes []int\n\tvar result []int\n\tfor j := 0; j < len(fvdl.UnifiedNodePool.Node[input-1].Reason.Trace.Primary.Entry); j++ {\n\t\tif fvdl.UnifiedNodePool.Node[input-1].Reason.Trace.Primary.Entry[j].NodeRef.RefId != 0 && fvdl.UnifiedNodePool.Node[input-1].Reason.Trace.Primary.Entry[j].NodeRef.RefId != (input-1) {\n\t\t\tsubnodes = append(subnodes, fvdl.UnifiedNodePool.Node[input-1].Reason.Trace.Primary.Entry[j].NodeRef.RefId+1)\n\t\t}\n\t}\n\tresult = append(result, input)\n\tlog.Entry().Debug(\"Successors: \", subnodes)\n\tfor j := 0; j < len(subnodes); j++ {\n\t\tresult = append(result, computeLocationPath(fvdl, subnodes[j])...)\n\t}\n\tlog.Entry().Debug(\"Finishing computing for ID \", input)\n\treturn result\n}", "func solve(board [][]byte) {\n\t// 2020-08-20 15:12 @bingohuang\n\t// 算法:1、深度优先搜索\n\t// 复杂度:O(MxN)/O(MxN)\n\t// 效率:执行耗时:36 ms,击败了9.94% 的Go用户\n\t//\t\t\t内存消耗:6 MB,击败了38.39% 的Go用户\n\tif len(board) == 0 || len(board[0]) == 0 {\n\t\treturn\n\t}\n\tr, c := len(board), len(board[0])\n\tvar dfs func([][]byte, int, int)\n\tdfs = func(board [][]byte, x, y int) {\n\t\tif x < 0 || x >= r || y < 0 || y >= c || board[x][y] != 'O' {\n\t\t\treturn\n\t\t}\n\t\tboard[x][y] = 'A'\n\t\tdfs(board, x+1, y)\n\t\tdfs(board, x-1, y)\n\t\tdfs(board, x, y+1)\n\t\tdfs(board, x, y-1)\n\t}\n\t// 左右两列搜索\n\tfor i := 0; i < r; i++ {\n\t\tdfs(board, i, 0)\n\t\tdfs(board, i, c-1)\n\t}\n\n\t// 上下两行搜索\n\tfor i := 0; i < c; i++ {\n\t\tdfs(board, 0, i)\n\t\tdfs(board, r-1, i)\n\t}\n\n\t// 对 A 标记,重新标记为O,对O标记为X\n\tfor i := 0; i < r; i++ {\n\t\tfor j := 0; j < c; j++ {\n\t\t\tif board[i][j] == 'A' {\n\t\t\t\tboard[i][j] = 'O'\n\t\t\t} else if board[i][j] == 'O' {\n\t\t\t\tboard[i][j] = 'X'\n\t\t\t}\n\t\t}\n\t}\n\n}", "func dijkstra(origin *Vertex, destination *Vertex, v []*Vertex, e []*Edge) []string {\r\n\t// Forcing our vertices to start at distance 0 (in case of the algorithm running multiple times on the same data)\r\n\tfor _, vertex := range v {\r\n\t\tvertex.Distance = 0\r\n\t}\r\n\r\n\t// Initialisation of various Queues\r\n\tqueue := VertexQueue{Elements: []*Vertex{origin}}\r\n\tvisited := VertexQueue{Elements: []*Vertex{}}\r\n\tnode, err := &Vertex{}, errors.New(\"\")\r\n\r\n\t// Iterate over all of the elements of the queue until there are no more vertices (max O(len(v)))\r\n\tfor queue.Size() > 0 {\r\n\t\t// Always work on what is closest to the current vertex in the queue.\r\n\t\tnode, err = queue.Pop()\r\n\t\tif err != nil {\r\n\t\t\tfmt.Println(errors.New(\"no nodes in queue\"))\r\n\t\t\tos.Exit(3)\r\n\t\t}\r\n\r\n\t\t// Avoid repetitions\r\n\t\tvisited.Append(node)\r\n\r\n\t\t// Filtering out the edges that are linked to the current node\r\n\t\tfilteredEdges := make([]*Edge, 0, len(e))\r\n\t\tfor _, edge := range e {\r\n\t\t\tif edge.Start.Label == node.Label || edge.End.Label == node.Label {\r\n\t\t\t\tfilteredEdges = append(filteredEdges, edge)\r\n\t\t\t}\r\n\t\t}\r\n\r\n\t\t// Sorting the edges by distance\r\n\t\tsort.Slice(filteredEdges, func(i, j int) bool {\r\n\t\t\treturn filteredEdges[i].Value > filteredEdges[j].Value\r\n\t\t})\r\n\r\n\t\t// If the vertex is a well\r\n\t\tif len(filteredEdges) == 1 &&\r\n\t\t\t(filteredEdges[0].Start != origin && filteredEdges[0].End != destination &&\r\n\t\t\t\tfilteredEdges[0].Start != destination && filteredEdges[0].End != origin) {\r\n\t\t\t// Set the node to be unreachable by the backtrace\r\n\t\t\tfor _, edge := range e {\r\n\t\t\t\tif edge.Start == filteredEdges[0].Start || edge.End == filteredEdges[0].Start {\r\n\t\t\t\t\tfilteredEdges[0].Start.Distance = math.MaxInt32\r\n\t\t\t\t\tbreak\r\n\t\t\t\t} else if edge.End == filteredEdges[0].End || edge.Start == filteredEdges[0].End {\r\n\t\t\t\t\tfilteredEdges[0].End.Distance = math.MaxInt32\r\n\t\t\t\t\tbreak\r\n\t\t\t\t}\r\n\t\t\t}\r\n\t\t} else {\r\n\t\t\t// Iterate on all available edges.\r\n\t\t\tfor _, edge := range filteredEdges {\r\n\t\t\t\t// Determine the actual direction of the edge.\r\n\t\t\t\tstart, end := &Vertex{}, &Vertex{}\r\n\t\t\t\tif edge.Start.Label == node.Label {\r\n\t\t\t\t\tstart = edge.Start\r\n\t\t\t\t\tend = edge.End\r\n\t\t\t\t} else {\r\n\t\t\t\t\tstart = edge.End\r\n\t\t\t\t\tend = edge.Start\r\n\t\t\t\t}\r\n\r\n\t\t\t\t// If end distance not yet set.\r\n\t\t\t\tif end.Label != origin.Label && ((end.Distance == 0 && end.Distance < start.Distance+edge.Value) || (end.Distance > start.Distance+edge.Value)) {\r\n\t\t\t\t\tend.Distance = start.Distance + edge.Value\r\n\t\t\t\t}\r\n\r\n\t\t\t\t// If arrival at destination, empty queue.\r\n\t\t\t\tif node.Label == destination.Label {\r\n\t\t\t\t\tqueue.DequeueWhere(func(v Vertex) bool { return true })\r\n\t\t\t\t\t// Otherwise append all further nodes to the queue.\r\n\t\t\t\t} else if !visited.Contains(*end) && ((queue.Contains(*end) && queue.FilterWhere(func(v Vertex) bool { return v.Label == end.Label })[0].Distance > end.Distance) || !queue.Contains(*end)) {\r\n\t\t\t\t\tqueue.Append(end)\r\n\t\t\t\t}\r\n\t\t\t}\r\n\t\t}\r\n\r\n\t\tsort.Slice(queue.Elements, func(i, j int) bool {\r\n\t\t\treturn queue.Elements[i].Distance > queue.Elements[j].Distance\r\n\t\t})\r\n\t}\r\n\r\n\t// Path of labels for the trace.\r\n\tpath := []*Vertex{destination}\r\n\tqueue = VertexQueue{Elements: []*Vertex{destination}}\r\n\tvisited = VertexQueue{Elements: []*Vertex{}}\r\n\tnode = &Vertex{}\r\n\r\n\t// Backtrace\r\n\tfor queue.Size() > 0 {\r\n\t\tnode, err = queue.Pop()\r\n\t\tif err != nil {\r\n\t\t\tfmt.Println(errors.New(\"no nodes in queue\"))\r\n\t\t\tos.Exit(3)\r\n\t\t}\r\n\t\tvisited.Append(node)\r\n\t\tfilteredEdges := make([]*Edge, 0, len(e))\r\n\t\tfor _, edge := range e {\r\n\t\t\tif (edge.Start.Label == node.Label || edge.End.Label == node.Label) && edge.Value != math.MaxInt32 {\r\n\t\t\t\tfilteredEdges = append(filteredEdges, edge)\r\n\t\t\t}\r\n\t\t}\r\n\r\n\t\tsort.Slice(filteredEdges, func(i, j int) bool {\r\n\t\t\treturn filteredEdges[i].Value > filteredEdges[j].Value\r\n\t\t})\r\n\r\n\t\tfor _, edge := range filteredEdges {\r\n\t\t\tstart, end := &Vertex{}, &Vertex{}\r\n\t\t\tif edge.Start.Label == node.Label {\r\n\t\t\t\tstart = edge.Start\r\n\t\t\t\tend = edge.End\r\n\t\t\t} else {\r\n\t\t\t\tstart = edge.End\r\n\t\t\t\tend = edge.Start\r\n\t\t\t}\r\n\r\n\t\t\tif node.Label == origin.Label {\r\n\t\t\t\tqueue.DequeueWhere(func(_ Vertex) bool { return true })\r\n\t\t\t} else if !visited.Contains(*end) && !queue.Contains(*end) && start.Distance-end.Distance == edge.Value {\r\n\t\t\t\t// If the node hasn't been visited and isn't planned for visit yet, and that the edge value corresponds to the delta of the distance from the start.\r\n\t\t\t\tqueue.Append(end)\r\n\t\t\t\tpath = append(path, end)\r\n\t\t\t}\r\n\t\t}\r\n\r\n\t\tsort.Slice(queue.Elements, func(i, j int) bool {\r\n\t\t\treturn queue.Elements[i].Distance > queue.Elements[j].Distance\r\n\t\t})\r\n\t}\r\n\r\n\t// Reverse the path to obtain the path in the right direction.\r\n\tfor i, j := 0, len(path)-1; i < j; i, j = i+1, j-1 {\r\n\t\tpath[i], path[j] = path[j], path[i]\r\n\t}\r\n\r\n\t// Format to string\r\n\tresult := []string{}\r\n\tfor _, vertex := range path {\r\n\t\tresult = append(result, vertex.Label)\r\n\t}\r\n\treturn result\r\n}", "func dfs(nodes *[]int, u *TreeNode) {\r\n\t// base\r\n\tif u == nil {\r\n\t\treturn\r\n\t}\r\n\t// recursion\r\n\tif u.Left != nil && u.Right != nil {\r\n\t\tdfs(nodes, u.Left)\r\n\t\tdfs(nodes, u.Right)\r\n\t\treturn\r\n\t}\r\n\tif u.Left != nil {\r\n\t\t*nodes = append(*nodes, u.Left.Val)\r\n\t\tdfs(nodes, u.Left)\r\n\t}\r\n\tif u.Right != nil {\r\n\t\t*nodes = append(*nodes, u.Right.Val)\r\n\t\tdfs(nodes, u.Right)\r\n\t}\r\n}", "func main() {\n graph := createGraph()\n graph.addEdge(1, 2)\n graph.addEdge(2, 3)\n graph.addEdge(2, 4)\n graph.addEdge(3, 4)\n graph.addEdge(1, 5)\n graph.addEdge(5, 6)\n graph.addEdge(5, 7)\n\n visited := make(set)\n\n dfs(graph, 1, visited, func(node int) {\n fmt.Print(node, \" \")\n })\n}", "func dfs(grid [][]int, row, col int) {\n\t// check if out of bound and cell is an unvisited islan\n\tif row < 0 || col < 0 || row >= len(grid) || col >= len(grid[0]) || grid[row][col] != 1 {\n\t\treturn\n\t}\n\n\t// mark cell as visited\n\tgrid[row][col] = 2\n\n\t// check neighbors\n\tdfs(grid, row+1, col)\n\tdfs(grid, row-1, col)\n\tdfs(grid, row, col+1)\n\tdfs(grid, row, col-1)\n}", "func totalBFS(data set, root int) int64 {\n\tn := int64(1) // node count\n\tr := int64(0) // repeated connections\n\n\tvar Q []int\n\tvar u, v int\n\n\tQ = append(Q, root)\n\tfor len(Q) > 0 {\n\t\tQ, u = Q[1:], Q[0] // dequeue\n\n\t\tfor i := 0; i < len(data[u].adj); i++ {\n\t\t\tv = data[u].adj[i]\n\n\t\t\tif data[v].color == 0 {\n\t\t\t\tdata[v].color = 1\n\n\t\t\t\tQ = append(Q, v) // enqueue\n\n\t\t\t\tn++ // new node\n\t\t\t} else if data[v].color == 1 {\n\t\t\t\tr++ // count repeated connection\n\t\t\t}\n\t\t}\n\n\t\tdata[u].color = 2\n\t}\n\n\t//fmt.Printf(\"%d, %d\\n\", n, r)\n\n\t// calculate the total\n\ttotal := ((n - 1) * n * (n + 1)) / 3\n\ttotal += n * (n - 1) * r\n\n\treturn total\n}", "func dfs(node string, visited map[string]int, symphony *Symphony, path []string) (bool, []string) {\n\tif visited[node] == 1 {\n\t\treturn true, path // cyclic dependent\n\t}\n\tif visited[node] == 2 {\n\t\treturn false, path\n\t}\n\t// 1 = temporarily visited\n\tvisited[node] = 1\n\tpath = append(path, node)\n\tdeps := symphony.tasks[node].Deps\n\tfor _, dep := range deps {\n\t\tif cyclic, path := dfs(dep, visited, symphony, path); cyclic {\n\t\t\treturn true, path\n\t\t}\n\t}\n\t// 2 = permanently visited\n\tvisited[node] = 2\n\n\treturn false, path\n}", "func dfs(nodes []node, target string, currNodeIndex, depth int) int {\n\tcurrNode := nodes[currNodeIndex]\n\tif currNode.name == target {\n\t\treturn depth\n\t}\n\tfor _, child := range nodes[currNodeIndex].children {\n\t\tresult := dfs(nodes, target, getNodeIndex(nodes, child), depth+1)\n\t\tif result != -1 {\n\t\t\treturn result\n\t\t}\n\t}\n\treturn -1\n}", "func DFS(g Graph, id ID) []ID {\n\tif _, err := g.GetNode(id); err != nil {\n\t\treturn nil\n\t}\n\n\ts := []ID{id}\n\tvisited := make(map[ID]bool)\n\trs := []ID{}\n\n\t// while S is not empty:\n\tfor len(s) != 0 {\n\n\t\tu := s[len(s)-1]\n\t\ts = s[:len(s)-1 : len(s)-1]\n\n\t\t// if u is not visited yet:\n\t\tif _, ok := visited[u]; !ok {\n\t\t\t// label u as visited\n\t\t\tvisited[u] = true\n\n\t\t\trs = append(rs, u)\n\n\t\t\t// for each vertex w adjacent to u:\n\t\t\tcmap, _ := g.GetTargets(u)\n\t\t\tfor _, w := range cmap {\n\t\t\t\t// if w is not visited yet:\n\t\t\t\tif _, ok := visited[w.ID()]; !ok {\n\t\t\t\t\ts = append(s, w.ID()) // S.push(w)\n\t\t\t\t}\n\t\t\t}\n\t\t\tpmap, _ := g.GetSources(u)\n\t\t\tfor _, w := range pmap {\n\t\t\t\t// if w is not visited yet:\n\t\t\t\tif _, ok := visited[w.ID()]; !ok {\n\t\t\t\t\ts = append(s, w.ID()) // S.push(w)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rs\n}", "func dfs(grid [][]byte, rowIdx int, colIdx int, maxRows int, maxCols int) {\n\tif (rowIdx < 0) || (colIdx < 0) || (rowIdx >= maxRows) || (colIdx >= maxCols) || string(grid[rowIdx][colIdx]) == \"0\" {\n\t\treturn\n\t}\n\n\tgrid[rowIdx][colIdx] = []byte(\"0\")[0]\n\n\tdfs(grid, rowIdx - 1, colIdx, maxRows, maxCols)\n\tdfs(grid, rowIdx + 1, colIdx, maxRows, maxCols)\n\tdfs(grid, rowIdx, colIdx - 1, maxRows, maxCols)\n\tdfs(grid, rowIdx, colIdx + 1, maxRows, maxCols)\n}", "func spfa(s int, t int)(bool){\n\tgPre = make([]int, nodeNum) // initialize gPre\n\tfor i:=0; i<len(gPre); i ++ {\n\t\tgPre[i] = -1 // default -1\n\t}\n\n\tgPath = make([]int, nodeNum) // initialize gPath\n\tfor i:=0; i<len(gPath); i ++ {\n\t\tgPath[i] = -1 // default -1\n\t}\n\n\tgDist = make([]int, nodeNum) // initialize gDist\n\tfor i:=0; i<len(gDist); i ++ {\n\t\tgDist[i] = int(^uint(0) >> 1) // default INT_MAX\n\t}\n\n\tgDist[s] = 0\n\tQ := list.New() // Q is the loose queue, it record all node , from which the min cost to another node may change\n\tQ.PushBack(s)\n\tfor Q.Len() > 0 {\n\t\tu_element := Q.Front()\n\t\tQ.Remove(u_element)\n\t\tu, err := u_element.Value.(int)\n\t\tif(!err){\n\t\t\tpanic(err)\n\t\t}\n\t\tfor e := gHead[u]; e != -1; e = gEdges[e].next { // visit all edges has node u as their src node\n\t\t\tv := gEdges[e].to\n\t\t\tif gEdges[e].vol > 0 && gDist[u]+gEdges[e].cost < gDist[v] { // if edge e has availiable capacity and node v's current min cost is more than that from node u to v\n\t\t\t\tgDist[v] = gDist[u] + gEdges[e].cost // update node v's min cost\n\t\t\t\tgPre[v] = u\n\t\t\t\tgPath[v] = e\n\t\t\t\tQ.PushBack(v) // because v's min cost has changed, so we need to check if the nodes that node v can reach can change its min cost\n\t\t\t}\n\t\t}\n\t}\n\n\tif gPre[t] == -1 {\n\t\treturn false\n\t}\n\treturn true\n\n}", "func (s *defaultSearcher) dfs(args searchArgs) {\n\toutEdges := args.nodeToOutEdges[args.root]\n\tif args.statusMap[args.root] == onstack {\n\t\tlog.Warn(\"The input call graph contains a cycle. This can't be represented in a \" +\n\t\t\t\"flame graph, so this path will be ignored. For your record, the ignored path \" +\n\t\t\t\"is:\\n\" + strings.TrimSpace(s.pathStringer.pathAsString(args.path, args.nameToNodes)))\n\t\treturn\n\t}\n\tif len(outEdges) == 0 {\n\t\targs.buffer.WriteString(s.pathStringer.pathAsString(args.path, args.nameToNodes))\n\t\targs.statusMap[args.root] = discovered\n\t\treturn\n\t}\n\targs.statusMap[args.root] = onstack\n\tfor _, edge := range outEdges {\n\t\ts.dfs(searchArgs{\n\t\t\troot: edge.Dst,\n\t\t\tpath: append(args.path, *edge),\n\t\t\tnodeToOutEdges: args.nodeToOutEdges,\n\t\t\tnameToNodes: args.nameToNodes,\n\t\t\tbuffer: args.buffer,\n\t\t\tstatusMap: args.statusMap,\n\t\t})\n\t}\n\targs.statusMap[args.root] = discovered\n}", "func bfs(root, target *node) *vector.Vector {\n\tqueue := &nodeQueue{vector.Vector{}}\n\troot.num = 0\n\theap.Push(queue, root)\n\n\tfor queue.Len() > 0 {\n\t\tn := heap.Pop(queue).(*node)\n\t\tn.visited = true\n\n\t\tfmt.Printf(\"Visited %v\\n\", n.name)\n\n\t\tif n == target {\n\t\t\tpath := vector.Vector{}\n\t\t\tpath.Push(n)\n\t\t\tfor n.prev != nil {\n\t\t\t\tn = n.prev\n\t\t\t\tpath.Push(n)\n\t\t\t}\n\t\t\treturn &path\n\t\t}\n\t\tfor ei := range n.edges.Iter() {\n\t\t\te := ei.(*edge)\n\t\t\tif !e.to.visited {\n\t\t\t\te.to.prev = n\n\t\t\t\te.to.num = n.num + 1\n\t\t\t\theap.Push(queue, e.to)\n\t\t\t}\n\t\t}\n\t}\n\treturn nil // never reached\n}", "func dfs(g *Graph, current int, visited set, visitFunction func(int)) {\n if _, seen := visited[current]; seen {\n return\n }\n\n visited[current] = true\n visitFunction(current)\n\n for neighbour := range g.adjList[current] {\n dfs(g, neighbour, visited, visitFunction)\n }\n}", "func (this *FindElements) Dfs(root *TreeNode, val int) {\n\t// base case\n\tif root == nil {\n\t\treturn\n\t}\n\n\t// add to seen\n\tthis.seen[val] = true\n\n\t// go left, right\n\tthis.Dfs(root.Left, 2*val+1)\n\tthis.Dfs(root.Right, 2*val+2)\n}", "func (cs *ColStore) makeWay(node1, node2 *ColStoreNode) {\n\tswitch {\n\tcase node1.Col.Max < node2.Col.Min:\n\t\t// The node2 starts after node1 ends, there's no overlap\n\t\t//\n\t\t// Node1 |----|\n\t\t// Node2 |----|\n\t\tif node1.Next != nil {\n\t\t\tif node1.Next.Col.Min <= node2.Col.Max {\n\t\t\t\tcs.makeWay(node1.Next, node2)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tcs.addNode(node1, node2, node1.Next)\n\t\t\treturn\n\t\t}\n\t\tcs.addNode(node1, node2, nil)\n\t\treturn\n\n\tcase node1.Col.Min > node2.Col.Max:\n\t\t// Node2 ends before node1 begins, there's no overlap\n\t\t//\n\t\t// Node1 |-----|\n\t\t// Node2 |----|\n\t\tif node1.Prev != nil {\n\t\t\tif node1.Prev.Col.Max >= node2.Col.Min {\n\t\t\t\tcs.makeWay(node1.Prev, node2)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tcs.addNode(node1.Prev, node2, node1)\n\t\t\treturn\n\t\t}\n\t\tcs.addNode(nil, node2, node1)\n\t\treturn\n\n\tcase node1.Col.Min == node2.Col.Min && node1.Col.Max == node2.Col.Max:\n\t\t// Exact match\n\t\t//\n\t\t// Node1 |xxx|\n\t\t// Node2 |---|\n\n\t\tprev := node1.Prev\n\t\tnext := node1.Next\n\t\tcs.removeNode(node1)\n\t\tcs.addNode(prev, node2, next)\n\t\t// Remove node may have set the root to nil\n\t\tif cs.Root == nil {\n\t\t\tcs.Root = node2\n\t\t}\n\t\treturn\n\n\tcase node1.Col.Min > node2.Col.Min && node1.Col.Max < node2.Col.Max:\n\t\t// Node2 envelopes node1\n\t\t//\n\t\t// Node1 |xx|\n\t\t// Node2 |----|\n\n\t\tprev := node1.Prev\n\t\tnext := node1.Next\n\t\tcs.removeNode(node1)\n\t\tswitch {\n\t\tcase prev == node2:\n\t\t\tnode2.Next = next\n\t\tcase next == node2:\n\t\t\tnode2.Prev = prev\n\t\tdefault:\n\t\t\tcs.addNode(prev, node2, next)\n\t\t}\n\n\t\tif node2.Prev != nil && node2.Prev.Col.Max >= node2.Col.Min {\n\t\t\tcs.makeWay(prev, node2)\n\t\t}\n\t\tif node2.Next != nil && node2.Next.Col.Min <= node2.Col.Max {\n\t\t\tcs.makeWay(next, node2)\n\t\t}\n\n\t\tif cs.Root == nil {\n\t\t\tcs.Root = node2\n\t\t}\n\n\tcase node1.Col.Min < node2.Col.Min && node1.Col.Max > node2.Col.Max:\n\t\t// Node2 bisects node1:\n\t\t//\n\t\t// Node1 |---xx---|\n\t\t// Node2 |--|\n\t\tnewCol := node1.Col.copyToRange(node2.Col.Max+1, node1.Col.Max)\n\t\tnewNode := &ColStoreNode{Col: newCol}\n\t\tcs.addNode(node1, newNode, node1.Next)\n\t\tnode1.Col.Max = node2.Col.Min - 1\n\t\tcs.addNode(node1, node2, newNode)\n\t\treturn\n\n\tcase node1.Col.Max >= node2.Col.Min && node1.Col.Min < node2.Col.Min:\n\t\t// Node2 overlaps node1 at some point above it's minimum:\n\t\t//\n\t\t// Node1 |----xx|\n\t\t// Node2 |-------|\n\t\tnext := node1.Next\n\t\tnode1.Col.Max = node2.Col.Min - 1\n\t\tif next == node2 {\n\t\t\treturn\n\t\t}\n\t\tcs.addNode(node1, node2, next)\n\t\tif next != nil && next.Col.Min <= node2.Col.Max {\n\t\t\tcs.makeWay(next, node2)\n\t\t}\n\t\treturn\n\n\tcase node1.Col.Min <= node2.Col.Max && node1.Col.Min > node2.Col.Min:\n\t\t// Node2 overlaps node1 at some point below it's maximum:\n\t\t//\n\t\t// Node1: |------|\n\t\t// Node2: |----xx|\n\t\tprev := node1.Prev\n\t\tnode1.Col.Min = node2.Col.Max + 1\n\t\tif prev == node2 {\n\t\t\treturn\n\t\t}\n\t\tcs.addNode(prev, node2, node1)\n\t\tif prev != nil && prev.Col.Max >= node2.Col.Min {\n\t\t\tcs.makeWay(node1.Prev, node2)\n\t\t}\n\t\treturn\n\t}\n\treturn\n}", "func BFS(g Graph, id ID) []ID {\n\tif _, err := g.GetNode(id); err != nil {\n\t\treturn nil\n\t}\n\n\tq := []ID{id}\n\tvisited := make(map[ID]bool)\n\tvisited[id] = true\n\trs := []ID{id}\n\n\t// while Q is not empty:\n\tfor len(q) != 0 {\n\n\t\tu := q[0]\n\t\tq = q[1:len(q):len(q)]\n\n\t\t// for each vertex w adjacent to u:\n\t\tcmap, _ := g.GetTargets(u)\n\t\tfor _, w := range cmap {\n\t\t\t// if w is not visited yet:\n\t\t\tif _, ok := visited[w.ID()]; !ok {\n\t\t\t\tq = append(q, w.ID()) // Q.push(w)\n\t\t\t\tvisited[w.ID()] = true // label w as visited\n\n\t\t\t\trs = append(rs, w)\n\t\t\t}\n\t\t}\n\t\tpmap, _ := g.GetSources(u)\n\t\tfor _, w := range pmap {\n\t\t\t// if w is not visited yet:\n\t\t\tif _, ok := visited[w.ID()]; !ok {\n\t\t\t\tq = append(q, w.ID()) // Q.push(w)\n\t\t\t\tvisited[w.ID()] = true // label w as visited\n\n\t\t\t\trs = append(rs, w.ID())\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rs\n}", "func floodfillByDir(t Tile) {\n\tvar visited = make(map[Tile]bool)\n\tvisited[t] = true\n\n\tvar queue = list.New()\n\tqueue.PushBack(t)\n\n\tfor queue.Len() > 0 {\n\t\ts := queue.Front()\n\t\t//type switch\n\t\tfor _, neigh := range s.Value.(Tile).neighbours {\n\t\t\tif !visited[neigh] {\n\t\t\t\tvisited[neigh] = true\n\t\t\t\tqueue.PushBack(neigh)\n\t\t\t}\n\t\t}\n\t\tqueue.Remove(s)\n\t}\n}", "func dfs(G [][]int, y, x int) int {\r\n\t// base\r\n\tif y < 0 || y >= len(G) || x < 0 || x >= len(G[0]) || G[y][x] == 1 {\r\n\t\treturn 0\r\n\t}\r\n\tG[y][x] = 1\r\n\t// recursion\r\n\tcnt := 1\r\n\tcnt += dfs(G, y-1, x) + dfs(G, y+1, x)\r\n\tcnt += dfs(G, y, x-1) + dfs(G, y, x+1)\r\n\treturn cnt\r\n}", "func pacificAtlanticBFS(mat [][]int) [][]int {\n\tres := [][]int{}\n\tif len(mat) == 0 || len(mat[0]) == 0 {\n\t\treturn res\n\t}\n\n\tm, n := len(mat), len(mat[0])\n\n\t// p[i][j] 表示,[i][j] 可以让水流到 Pacific 的点\n\t// a[i][j] 表示,[i][j] 可以让水流到 Atlantic 的点\n\tp, a := make([][]bool, m), make([][]bool, m)\n\tfor i := 0; i < m; i++ {\n\t\tp[i] = make([]bool, n)\n\t\ta[i] = make([]bool, n)\n\t}\n\t// pQueue 是所有能够让水流到 Pacific 的点的队列\n\t// aQueue 是所有能够让水流到 Atlantic 的点的队列\n\t// 初始化 pQueue 和 aQueue\n\tpQueue := [][]int{}\n\taQueue := [][]int{}\n\t// 左边可进pQueue,右边进aQueue\n\tfor i := 0; i < m; i++ {\n\t\tp[i][0] = true\n\t\tpQueue = append(pQueue, []int{i, 0})\n\t\ta[i][n-1] = true\n\t\taQueue = append(aQueue, []int{i, n - 1})\n\t}\n\t// 上边进pQueue,下边进aQueue\n\tfor j := 0; j < n; j++ {\n\t\tp[0][j] = true\n\t\tpQueue = append(pQueue, []int{0, j})\n\t\ta[m-1][j] = true\n\t\taQueue = append(aQueue, []int{m - 1, j})\n\t}\n\n\tds := [][]int{{-1, 0}, {1, 0}, {0, -1}, {0, 1}}\n\tbfs := func(queue [][]int, rec [][]bool) {\n\t\tfor len(queue) > 0 {\n\t\t\tc := queue[0]\n\t\t\tqueue = queue[1:]\n\t\t\tfor _, d := range ds {\n\t\t\t\ti, j := c[0]+d[0], c[1]+d[1]\n\t\t\t\tif 0 <= i && i < m &&\n\t\t\t\t\t0 <= j && j < n &&\n\t\t\t\t\t!rec[i][j] &&\n\t\t\t\t\tmat[c[0]][c[1]] <= mat[i][j] { // i,j可达是流动方向则进队列且\n\t\t\t\t\trec[i][j] = true\n\t\t\t\t\tqueue = append(queue, []int{i, j})\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\tbfs(pQueue, p)\n\tbfs(aQueue, a)\n\tfor i := 0; i < m; i++ {\n\t\tfor j := 0; j < n; j++ {\n\t\t\tif p[i][j] && a[i][j] {\n\t\t\t\tres = append(res, []int{i, j})\n\t\t\t}\n\t\t}\n\t}\n\n\treturn res\n}", "func getShortestPath(sourceX, sourceY, targetX, targetY int, cave [][]Square) (int, int, int) {\n\tdistances := make([][]int, len(cave))\n\tfor y := range distances {\n\t\tdistances[y] = make([]int, len(cave[y]))\n\t\tfor x := range distances[y] {\n\t\t\tdistances[y][x] = math.MaxInt64\n\t\t}\n\t}\n\n\tvisited := make([][]bool, len(cave))\n\tfor y := range visited {\n\t\tvisited[y] = make([]bool, len(cave[y]))\n\t}\n\n\t// Start with the target, and compute the distances back to the source.\n\ttentative := make([]int, 1)\n\ttentative[0] = targetX<<16 + targetY\n\tdistances[targetY][targetX] = 0\n\n\tfmt.Printf(\"Looking for distances from (%d,%d) to (%d,%d)\\n\", sourceX, sourceY, targetX, targetY)\n\n\tfor len(tentative) > 0 {\n\t\t// Find the tentative cell with the shortest distance\n\t\tcur := 0\n\t\tminDistance := math.MaxInt64\n\t\tfor i, pos := range tentative {\n\t\t\tx, y := (pos >> 16), pos&0xffff\n\t\t\tif distances[y][x] < minDistance {\n\t\t\t\tminDistance = distances[y][x]\n\t\t\t\tcur = i\n\t\t\t}\n\t\t}\n\n\t\t// Remove that cell from the tentative list\n\t\tcurX, curY := tentative[cur]>>16, tentative[cur]&0xffff\n\t\ttentative = append(tentative[:cur], tentative[cur+1:]...)\n\t\tvisited[curY][curX] = true\n\n\t\t//fmt.Printf(\"Visited (%d,%d), distance is %d (queue is %d long)\\n\", curX, curY, distances[curY][curX], len(tentative))\n\n\t\t// Consider the neighbors.\n\t\tfor _, move := range directions {\n\t\t\tnewX := curX + move[0]\n\t\t\tnewY := curY + move[1]\n\n\t\t\t// Skip if we're past the edge of the cave\n\t\t\tif newX < 0 || newY < 0 || newX >= len(cave[0]) || newY >= len(cave) {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Skip if it's not empty.\n\t\t\tif cave[newY][newX] != Empty {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// If we haven't visited that cell, set a tentative distance\n\t\t\tif !visited[newY][newX] {\n\t\t\t\tvisited[newY][newX] = true\n\t\t\t\ttentative = append(tentative, newX<<16+newY)\n\t\t\t\tdistances[newY][newX] = distances[curY][curX] + 1\n\t\t\t\t//fmt.Printf(\"- First visit to (%d,%d), set distance to %d\\n\", newX, newY, distances[newY][newX])\n\t\t\t} else {\n\t\t\t\tif distances[newY][newX] > distances[curY][curX] {\n\t\t\t\t\tdistances[newY][newX] = distances[curY][curX] + 1\n\t\t\t\t\t//fmt.Printf(\"- Revisit to (%d,%d), set distance to %d\\n\", newX, newY, distances[newY][newX])\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\t//fmt.Printf(\"-> Distances from (%d,%d) to (%d,%d):\\n\", sourceX, sourceY, targetX, targetY)\n\t//dumpDistances(sourceX, sourceY, targetX, targetY, distances)\n\n\t// We've got the distance, now pick a direction.\n\tbestDistance := math.MaxInt64\n\tbestDX, bestDY := 0, 0\n\tfor _, move := range directions {\n\t\tnewX := sourceX + move[0]\n\t\tnewY := sourceY + move[1]\n\n\t\t// Skip if we're past the edge of the cave\n\t\tif newX < 0 || newY < 0 || newX >= len(cave[0]) || newY >= len(cave) {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Skip if it's not empty.\n\t\tif cave[newY][newX] != Empty {\n\t\t\tcontinue\n\t\t}\n\n\t\tif distances[newY][newX] != math.MaxInt64 && distances[newY][newX]+1 < bestDistance {\n\t\t\tbestDistance = distances[newY][newX] + 1\n\t\t\tbestDX, bestDY = move[0], move[1]\n\t\t}\n\t}\n\n\tif bestDistance != math.MaxInt64 {\n\t\tbestDistance++\n\t}\n\n\tfmt.Printf(\"-> Shortest path from (%d, %d) to (%d, %d) is %d long going (%d,%d)\\n\", sourceX, sourceY, targetX, targetY, bestDistance, bestDX, bestDY)\n\treturn bestDX, bestDY, bestDistance\n}", "func searchPath(nodes vector.Vector, from, to int) {\n\tpath := bfs(nodes[from].(*node), nodes[to].(*node))\n\n\tfor path.Len() > 0 {\n\t\tfmt.Printf(\"-> %v \", path.Pop().(*node).name)\n\t}\n\tfmt.Printf(\"\\n -- Done --\\nCleaning up..\\n\")\n\tclearNodes(nodes)\n}", "func IterativeStore(key ID, value []byte) int {\n\n\tprevDistance := key.Xor(ThisNode.ThisContact.NodeID)\n\t\n\t//var closestNode kademlia.FoundNode\n\tclosestNode := ThisNode.ThisContact\n\tkClosestNodes := make(Bucket,20)\n \n\thostPortStr := get_host_port(ThisNode.ThisContact)\n\t\n\t//closestnode may want to be its own function that we call from FindNode, or at least\n\t//that code should be in FindNode, since we need to populate res.Nodes with more than one bucket\n\tfor true {\n\t\t//fmt.Printf(\"%s\\n\",hostPortStr)\n\t\tclient, err := rpc.DialHTTP(\"tcp\", hostPortStr)\n\t\tif err != nil {\n\t\t\t//fmt.Printf(\"1\\n\")\n\t\t\tlog.Fatal(\"DialHTTP: \", err)\n\t\t}\n\t\treq := new(FindNodeRequest)\n\t\treq.MsgID = NewRandomID()\n\t\treq.NodeID = key\n\t\n\t\tvar res FindNodeResult\n\t\t//if FindNode works, all of the closest nodes should be in res.\n\t\terr = client.Call(\"Kademlia.FindNode\", req, &res)\n \t\tif err != nil {\n\t\t\tlog.Fatal(\"Call: \", err)\n \t\t}\n client.Close()\n \t\t// obviously we need to do something with the array here, not just take the first element\n \t\t//fmt.Printf(\"Node 0: %v\\n\",res.Nodes[0])\n \t\tnextClosestNode, dist := res.Nodes[0], key.Xor(res.Nodes[0].NodeID)\n \t\tfor i:= 0; i < len(res.Nodes); i++ {\n \t\t\tif res.Nodes[i].Port != 0 {\n if res.Nodes[i].NodeID.Xor(key).Less(dist) {\n dist = res.Nodes[i].NodeID.Xor(key)\n nextClosestNode = res.Nodes[i]\n }\n // update kClosestNodes\n replace_idx := -1\n for j := 0; j < len(kClosestNodes); j++ {\n if kClosestNodes[j] == nil {\n kClosestNodes[j] = res.Nodes[i].ToContactPtr()\n replace_idx = -1\n break\n } else if res.Nodes[i].ToContactPtr().NodeID.Xor(key).Less(kClosestNodes[j].NodeID.Xor(key)) {\n if replace_idx != -1 {\n if kClosestNodes[replace_idx].NodeID.Xor(key).Less(kClosestNodes[j].NodeID.Xor(key)) {\n replace_idx = j\n }\n }\n }\n }\n if replace_idx != -1 {\n kClosestNodes[replace_idx] = res.Nodes[i].ToContactPtr()\n }\n }\n \t\t}\n \t\tcurDistance := key.Xor(nextClosestNode.NodeID)\n \t\tif !curDistance.Less(prevDistance) {\n \t\t\tbreak\n \t\t} else {\n \t\t\tclosestNode = nextClosestNode.ToContactPtr()\n prevDistance = curDistance\n \t\t}\n \t\thostPortStr = get_host_port(closestNode)\n\t\t}\n\thostPortStr = get_host_port(closestNode)\n\tstore(hostPortStr, key, value)\n // replicate data across k closest nodes\n for i:=0; i < len(kClosestNodes); i++ {\n if kClosestNodes[i] != nil {\n if !kClosestNodes[i].NodeID.Equals(closestNode.NodeID) {\n hostPortStr = get_host_port(kClosestNodes[i])\n store(hostPortStr, key, value)\n }\n }\n }\n\t//fmt.Printf(\"%v\\n\",closestNode.NodeID)\n\treturn 1\n}", "func allPathsSourceTarget(graph [][]int) [][]int {\n ans := make([][]int, 0)\n bfs := make([][]int, 0)\n bfs = append(bfs, []int{0})\n\n for ; len(bfs) > 0 ; {\n cur := bfs[0]\n if cur[len(cur) - 1] == len(graph) - 1 {\n ans = append(ans, cur)\n }\n for _, val := range graph[cur[len(cur) - 1] ] {\n x := make([]int,len(cur))\n copy(x,cur)\n x = append(x, val)\n bfs = append(bfs, x)\n }\n bfs = bfs[1:]\n }\n return ans\n}", "func Bfs(g *Graph, starter int) {\n\tif g.eNum == 0 {\n\t\treturn\n\t}\n\tfmt.Printf(\"BFS begin with node %d\\n\", starter)\n\tisVisited := make(map[int]bool, g.vNum)\n\tdistance := make(map[int]int, g.vNum)\n\tisVisited[starter] = true\n\tdistance[starter] = 0\n\tvar queue []Vertex\n\tqueue = append(queue, g.adjacencyList[starter])\n\tfor len(queue) > 0 {\n\t\tcurrentNode := queue[0]\n\t\tqueue = queue[1:len(queue)]\n\t\tcurrentDis := distance[currentNode.value]\n\t\tfmt.Printf(\"node %d from %d (starter)'s distance is: %d\\n\", currentNode.value, starter, currentDis)\n\t\tadjNode := currentNode.edgeList.Front()\n\t\tfor adjNode != nil {\n\t\t\tadjNodeVal := adjNode.Value.(int)\n\t\t\tif is, ok := isVisited[adjNodeVal]; !is || !ok {\n\t\t\t\tqueue = append(queue, g.adjacencyList[adjNodeVal])\n\t\t\t\tisVisited[adjNodeVal] = true\n\t\t\t\tdistance[adjNodeVal] = currentDis + 1\n\t\t\t}\n\t\t\tadjNode = adjNode.Next()\n\t\t}\n\t}\n}", "func siftDown(kvs []_MapPair, lo, hi, first int) {\n root := lo\n for {\n child := 2*root + 1\n if child >= hi {\n break\n }\n if child+1 < hi && kvs[first+child].k < kvs[first+child+1].k {\n child++\n }\n if kvs[first+root].k >= kvs[first+child].k {\n return\n }\n swap(kvs, first+root, first+child)\n root = child\n }\n}", "func main() {\n\tg, nodesMap := graphLineByLine()\n\n\tstart := nodesMap[\"shiny gold\"]\n\tcount := 0\n\tgraph.BFS(g, start, func(v, w int, _ int64) {\n\t\t//fmt.Println(v, \"to\", w)\n\t\tcount++\n\n\t})\n\n\tfmt.Println(count)\n}", "func SPath(par *tp.Par, task *tp.PathTask, nodes map[int]*tp.Node, cLog io.Writer) map[*tp.Node][]*tp.PQItem {\r\n\t//count paths build\r\n\tif task.BanLink == nil {\r\n\t\tpar.PathBuildCount[0]++\r\n\t} else {\r\n\t\tpar.PathBuildCount[1]++\r\n\t}\r\n\tnumJ := len(task.TRIP)\r\n\tCurrNd := task.I\r\n\tpqItem := tp.PQItem{Nd: CurrNd, ParentItem: nil, TP: task.TP, IMP: 0, Index: 0, Visited: false}\r\n\tpq := p.MakeHeap(&pqItem)\r\n\tpqItemAll := make(map[*tp.Node]*tp.PQItem)\r\n\tpqItemAll[CurrNd] = &pqItem\r\n\t//loop until all zone nodes have been visited\r\n\tCurrpqItem, _ := pq.Pop()\r\n\tfor {\r\n\t\tCurrpqItem.Visited = true\r\n\t\tCurrNd = CurrpqItem.Nd\r\n\t\tif _, ok := task.TRIP[CurrNd]; ok { //check if all destination nodes have been visited\r\n\t\t\tnumJ--\r\n\t\t\tif numJ == 0 {\r\n\t\t\t\tbreak\r\n\t\t\t}\r\n\t\t}\r\n\t\tif task.DisttoDest > 0 && CurrpqItem.Dist > task.DisttoDest { //current path has exceed the max length\r\n\t\t\treturn nil\r\n\t\t}\r\n\t\tif CurrNd.TYPE == par.NodeTypeZone && CurrNd != task.I { //skip zone node\r\n\t\t\tgoto Check\r\n\t\t}\r\n\t\tfor _, link := range CurrNd.DNLINKS {\r\n\t\t\tif link == task.BanLink { //for decision node path\r\n\t\t\t\tcontinue\r\n\t\t\t}\r\n\t\t\tif element, ok := pqItemAll[link.NODEB]; ok {\r\n\t\t\t\tif element.Visited {\r\n\t\t\t\t\tcontinue\r\n\t\t\t\t} else {\r\n\t\t\t\t\t//htime := math.Sqrt(math.Pow(CurrNd.X-task.DestNode.X, 2)+math.Pow(CurrNd.Y-task.DestNode.Y, 2)) / 5280\r\n\t\t\t\t\timpedent := element.IMP + link.CgTime[element.TP-1]*link.TimeWeight[element.TP-1] +\r\n\t\t\t\t\t\tlink.TollRate[element.TP-1]*par.VOTFactor[task.VCLS]// + htime\r\n\t\t\t\t\tif impedent < element.IMP {\r\n\t\t\t\t\t\telement.Time = element.Time + link.CgTime[element.TP-1]\r\n\t\t\t\t\t\ttimepd := int(element.Time/15) + task.TP\r\n\t\t\t\t\t\tif timepd > 96 {\r\n\t\t\t\t\t\t\ttimepd = timepd - 96\r\n\t\t\t\t\t\t}\r\n\t\t\t\t\t\telement.TP = timepd\r\n\t\t\t\t\t\telement.IMP = impedent\r\n\t\t\t\t\t\telement.Dist = element.Dist + link.DIST\r\n\t\t\t\t\t\telement.Toll = element.Toll + link.TollRate[element.TP-1]\r\n\t\t\t\t\t\telement.FFTime = element.FFTime + link.FFTime\r\n\t\t\t\t\t\telement.ParentItem = CurrpqItem\r\n\t\t\t\t\t\telement.Link = link\r\n\t\t\t\t\t}\r\n\t\t\t\t}\r\n\t\t\t} else { //node was not checked\r\n\t\t\t\t//htime := math.Sqrt(math.Pow(CurrNd.X-task.DestNode.X, 2)+math.Pow(CurrNd.Y-task.DestNode.Y, 2)) / 5280\r\n\t\t\t\timpedent := CurrpqItem.IMP + link.CgTime[CurrpqItem.TP-1]*link.TimeWeight[CurrpqItem.TP-1] +\r\n\t\t\t\t\tlink.TollRate[CurrpqItem.TP-1]*par.VOTFactor[task.VCLS]// + htime\r\n\t\t\t\ttime := CurrpqItem.Time + link.CgTime[CurrpqItem.TP-1]\r\n\t\t\t\tperceivedtime := CurrpqItem.PcvTime + link.CgTime[CurrpqItem.TP-1]*link.TimeWeight[CurrpqItem.TP-1]\r\n\t\t\t\tdist := CurrpqItem.Dist + link.DIST\r\n\t\t\t\ttoll := CurrpqItem.Toll + link.TollRate[CurrpqItem.TP-1]\r\n\t\t\t\tfftime := CurrpqItem.FFTime + link.FFTime\r\n\t\t\t\ttimepd := int(time/15) + task.TP\r\n\t\t\t\tif timepd > 96 {\r\n\t\t\t\t\ttimepd = timepd - 96\r\n\t\t\t\t}\r\n\t\t\t\tnewpqItem := tp.PQItem{Nd: link.NODEB, Link: link, ParentItem: CurrpqItem, IMP: impedent, PcvTime: perceivedtime,\r\n\t\t\t\t\tTime: time, Dist: dist, Toll: toll, FFTime: fftime, TP: timepd, Index: 0, Visited: false}\r\n\t\t\t\tpqItemAll[link.NODEB] = &newpqItem\r\n\t\t\t\tpq.Insert(&newpqItem)\r\n\t\t\t}\r\n\t\t}\r\n\tCheck:\r\n\t\tv, ok := pq.Pop()\r\n\t\tif ok {\r\n\t\t\tCurrpqItem = v\r\n\t\t} else {\r\n\t\t\treturn nil\r\n\t\t}\r\n\t}\r\n\r\n\t//Get paths\r\n\t//var path map[*tp.Node][]*tp.PQItem\r\n\tpath := make(map[*tp.Node][]*tp.PQItem)\r\n\tfor nd := range task.TRIP {\r\n\t\tpqI := pqItemAll[nd]\r\n\t\tvar nodelist []*tp.PQItem\r\n\t\tnodelist = append(nodelist, pqI)\r\n\t\t//var pathWrite []int\r\n\t\t//pathWrite = append(pathWrite, pqI.Nd.N)\r\n\t\tfor pqI.ParentItem != nil {\r\n\t\t\tpqI = pqI.ParentItem\r\n\t\t\tnodelist = append(nodelist, pqI)\r\n\t\t\t//\tpathWrite = append(pathWrite, pqI.Nd.N)\r\n\t\t}\r\n\t\t//if task.Path != nil && found {\r\n\t\t//\tfmt.Printf(\"path %v\\n\", pathWrite)\r\n\t\t//}\r\n\t\tpath[nd] = nodelist\r\n\t}\r\n\treturn path\r\n}", "func dls(zacStav Stav, hlbka int, finAuto string, poc *int)(*Stav) {\n\tzacStav.Hlbka = 0\n\tzacStav.Mapa = vytvorMapuZAut(zacStav.Auta)\n\t//hash mapa, kde sa uchovavaju navstivene stavy\n\thshMapa := make(map[uint64]int)\n\t//stavy, ktore su na okraji (este neprehladane)\n\tokraj := make([]Stav, 0)\n\tokraj = append(okraj, zacStav)\n\n\t//prehladavany stav\n\tvar prehlStav Stav\n\n\tfor len(okraj) != 0 {\n\t\t//pop stavu, ktory sa ide navstivit\n\t\tprehlStav = okraj[len(okraj)-1]\n\t\tokraj = okraj[:len(okraj)-1]\n\t\th := hash(prehlStav)\n\n\t\t//ak sa nasiel finalny stav, vrat ho, inak pokracuj v prehladavani\n\t\tif prehlStav.Hlbka == hlbka && is_state_final(prehlStav, finAuto) {\n\t\t\treturn &prehlStav\n\t\t} else if hshMapa[h] == 0 || hshMapa[h] > prehlStav.Hlbka{\n\t\t\t//ak stav este nebol navstiveny, alebo bol navstiveny, ale hlbsie -> navstiv ho\n\t\t\thshMapa[hash(prehlStav)] = prehlStav.Hlbka\n\t\t\t//ak je hlbka vramci ohranicenia, najdi deti a pushni do stacku\n\t\t\tif prehlStav.Hlbka < hlbka{\n\t\t\t\t*poc += 1\n\t\t\t\tchildren := najdiDeti(prehlStav)\n\t\t\t\tokraj = append(okraj, children...)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func BFS(g *Graph, start *Vertex) map[int]*Vertex { //\n\n\tfor _, v := range g.Vertices {\n\t\tv.Parent, v.Distance, v.InQueue = nil, 0, false\n\t}\n\n\tq := &queue{}\n\tvisited := map[int]*Vertex{}\n\tcurrent := start\n\n\tfor {\n\n\t\tvisited[current.Key] = current\n\n\t\tfor _, v := range current.Vertices {\n\t\t\tif _, ok := visited[v.Key]; ok || v.InQueue {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tq.enqueue(v)\n\t\t\tv.InQueue = true\n\t\t\tv.Parent = current\n\t\t\tv.Distance = current.Distance + 1\n\t\t}\n\n\t\tcurrent = q.dequeue()\n\t\tif current == nil {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn visited\n}", "func dijstra(g adjList, n int, s int, e int) ([]int, []int) {\n\t// g - adjacency list of a weighted graph\n\t// n - the number of nodes in the graph\n\t// s - the index of the starting node ( 0 <= s < n )\n\t// e - the index of the end node ( 0 <= e < n )\n\tvisited := make([]bool, n)\n\tdistance := make([]int, n)\n\n\t// keep track of the previous node we took\n\t// to get to the current node\n\tprevious := make([]int, n)\n\n\tfor i := range visited {\n\t\tvisited[i] = false\n\t}\n\n\tfor i := range distance {\n\t\tdistance[i] = math.MaxInt64\n\t}\n\n\tdistance[s] = 0\n\t// Set Min option to true for minheap\n\tminheap := pqueue.NewHeap(pqueue.Options{\n\t\tMin: true,\n\t})\n\n\tminheap.InsertPriority(string(s), 0)\n\n\tfor minheap.Length() != 0 {\n\n\t\tstringAtIndex, min := minheap.Poll()\n\t\tintegerAtIndex, _ := strconv.Atoi(stringAtIndex)\n\n\t\t// current node is integerAtIndex\n\t\tvisited[integerAtIndex] = true\n\n\t\t// optimization to ignore stale index\n\t\t// (index, min_dis) pair\n\t\tif distance[integerAtIndex] < min {\n\t\t\tcontinue\n\t\t}\n\n\t\t// loop through all the neighbours of\n\t\t// the current node\n\t\tcn := g[integerAtIndex].head\n\t\tfor cn != nil {\n\n\t\t\tif visited[cn.vertex] {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tnewdist := distance[integerAtIndex] + cn.weight\n\t\t\tif newdist < distance[cn.vertex] {\n\t\t\t\tprevious[cn.vertex] = integerAtIndex\n\t\t\t\tdistance[cn.vertex] = newdist\n\t\t\t\tminheap.InsertPriority(strconv.Itoa(cn.vertex), newdist)\n\t\t\t}\n\n\t\t\tif cn.next == nil {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tcn = cn.next\n\t\t}\n\n\t\t// Optimise here to stop early.\n\t\tif integerAtIndex == e {\n\t\t\treturn distance, previous\n\t\t}\n\n\t}\n\treturn distance, previous\n}", "func (g *Graph) Bfs(root string, goal string, maxDepth int) (bool, *Vertex) {\n\n\t// Preconditions\n\tif len(root) == 0 {\n\t\tlog.Fatal(\"Root vertex is empty\")\n\t}\n\n\tif len(goal) == 0 {\n\t\tlog.Fatal(\"Goal vertex is empty\")\n\t}\n\n\tif maxDepth < 0 {\n\t\tlog.Fatalf(\"Maximum depth is invalid: %v\\n\", maxDepth)\n\t}\n\n\t// Set of the identifiers of discovered vertices\n\tdiscovered := set.New()\n\tdiscovered.Insert(root)\n\n\t// Queue to hold the vertices to visit\n\tq := queue.New()\n\tq.Enqueue(NewVertex(root, 0))\n\n\t// While there are vertices in the queue to check\n\tfor q.Len() > 0 {\n\n\t\t// Take a vertex from the queue\n\t\tv := q.Dequeue().(Vertex)\n\n\t\t// If the vertex is the goal, then return\n\t\tif v.Identifier == goal {\n\t\t\treturn true, &v\n\t\t}\n\n\t\t// Depth of any vertices adjacent to v\n\t\tnewDepth := v.Depth + 1\n\n\t\t// If the adjacent vertices are within the range\n\t\tif newDepth <= maxDepth {\n\n\t\t\t// Get a list of the adjacent vertices\n\t\t\tw := g.AdjacentTo(v.Identifier)\n\n\t\t\t// Walk through each of the adjacent vertices\n\t\t\tfor _, adjIdentifier := range w {\n\n\t\t\t\t// If the vertex hasn't been seen before\n\t\t\t\tif !discovered.Has(adjIdentifier) {\n\n\t\t\t\t\t// Add the identifier to the set of discovered identifiers\n\t\t\t\t\tdiscovered.Insert(adjIdentifier)\n\n\t\t\t\t\t// Put the vertex on the queue\n\t\t\t\t\tnewVertex := NewVertex(adjIdentifier, newDepth)\n\t\t\t\t\tnewVertex.Parent = &v\n\t\t\t\t\tq.Enqueue(newVertex)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\t// The goal was not found\n\treturn false, nil\n}", "func zopfliComputeShortestPath(num_bytes uint, position uint, ringbuffer []byte, ringbuffer_mask uint, params *encoderParams, dist_cache []int, hasher *h10, nodes []zopfliNode) uint {\n\tvar max_backward_limit uint = maxBackwardLimit(params.lgwin)\n\tvar max_zopfli_len uint = maxZopfliLen(params)\n\tvar model zopfliCostModel\n\tvar queue startPosQueue\n\tvar matches [2 * (maxNumMatchesH10 + 64)]backwardMatch\n\tvar store_end uint\n\tif num_bytes >= hasher.StoreLookahead() {\n\t\tstore_end = position + num_bytes - hasher.StoreLookahead() + 1\n\t} else {\n\t\tstore_end = position\n\t}\n\tvar i uint\n\tvar gap uint = 0\n\tvar lz_matches_offset uint = 0\n\tnodes[0].length = 0\n\tnodes[0].u.cost = 0\n\tinitZopfliCostModel(&model, &params.dist, num_bytes)\n\tzopfliCostModelSetFromLiteralCosts(&model, position, ringbuffer, ringbuffer_mask)\n\tinitStartPosQueue(&queue)\n\tfor i = 0; i+hasher.HashTypeLength()-1 < num_bytes; i++ {\n\t\tvar pos uint = position + i\n\t\tvar max_distance uint = brotli_min_size_t(pos, max_backward_limit)\n\t\tvar skip uint\n\t\tvar num_matches uint\n\t\tnum_matches = findAllMatchesH10(hasher, &params.dictionary, ringbuffer, ringbuffer_mask, pos, num_bytes-i, max_distance, gap, params, matches[lz_matches_offset:])\n\t\tif num_matches > 0 && backwardMatchLength(&matches[num_matches-1]) > max_zopfli_len {\n\t\t\tmatches[0] = matches[num_matches-1]\n\t\t\tnum_matches = 1\n\t\t}\n\n\t\tskip = updateNodes(num_bytes, position, i, ringbuffer, ringbuffer_mask, params, max_backward_limit, dist_cache, num_matches, matches[:], &model, &queue, nodes)\n\t\tif skip < longCopyQuickStep {\n\t\t\tskip = 0\n\t\t}\n\t\tif num_matches == 1 && backwardMatchLength(&matches[0]) > max_zopfli_len {\n\t\t\tskip = brotli_max_size_t(backwardMatchLength(&matches[0]), skip)\n\t\t}\n\n\t\tif skip > 1 {\n\t\t\t/* Add the tail of the copy to the hasher. */\n\t\t\thasher.StoreRange(ringbuffer, ringbuffer_mask, pos+1, brotli_min_size_t(pos+skip, store_end))\n\n\t\t\tskip--\n\t\t\tfor skip != 0 {\n\t\t\t\ti++\n\t\t\t\tif i+hasher.HashTypeLength()-1 >= num_bytes {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tevaluateNode(position, i, max_backward_limit, gap, dist_cache, &model, &queue, nodes)\n\t\t\t\tskip--\n\t\t\t}\n\t\t}\n\t}\n\n\tcleanupZopfliCostModel(&model)\n\treturn computeShortestPathFromNodes(num_bytes, nodes)\n}", "func traverseBoard() {\n\tvar nextPos *BoardLocation\n\tvar loc *BoardLocation\n\tvar edgePairID edgePairID\n\tvar progress int\n\tvar edgePairList *tileEdgePairList\n\tvar ok bool\n\tvar numberInterations uint64\n\n\tvar highestProgress int\n\n\t// need to have current location set to 1st tile to start TODO\n\tloc = &board.loc[0][0]\n\tloc.edgePairList = loc.edgePairMap[calcEdgePairID(0, 0)]\n\tloc.index = 0\n\tloc.listSize = 1\n\n\tprogress = 0\n\thighestProgress = 0\n\n\tfor {\n\t\tnumberInterations++\n\t\t// are there still tiles left to try in the edge pair list on the current location\n\t\tif loc.index < loc.edgePairList.availableNoTiles {\n\t\t\tloc.tile = loc.edgePairList.tiles[loc.index].tile\n\t\t\tloc.tile.rotation = loc.edgePairList.tiles[loc.index].rotationForEdgePair\n\t\t\tloc.noTimesVisited++\n\t\t\tif progress >= highestProgress {\n\t\t\t\tfmt.Println(board)\n\t\t\t\thighestProgress = progress\n\t\t\t\tfmt.Println(\"Placed:\", progress, time.Now().Format(time.RFC850))\n\t\t\t\tfmt.Println(\"Number of iterations:\", numberInterations)\n\t\t\t\tif progress == (board.width*board.height)-1 {\n\t\t\t\t\tfmt.Println(board)\n\t\t\t\t\tfmt.Println(\"finished solution \") // TODO Print out proper solution\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Now see if there is a valid EP in next location\n\t\t\tnextPos = loc.traverseNext\n\t\t\tedgePairID = nextPos.getEdgePairIDForLocation()\n\n\t\t\tedgePairList, ok = nextPos.edgePairMap[edgePairID]\n\t\t\tif ok && edgePairList.availableNoTiles > 0 { // valid edgepair list and it has tiles on it.\n\t\t\t\tprogress++\n\n\t\t\t\t// remove the tile from the EP lists\n\t\t\t\tloc.tile.edgePairLists[0].removeTile(loc.tile.positionInEdgePairList[0])\n\t\t\t\tloc.tile.edgePairLists[1].removeTile(loc.tile.positionInEdgePairList[1])\n\t\t\t\tloc.tile.edgePairLists[2].removeTile(loc.tile.positionInEdgePairList[2])\n\t\t\t\tloc.tile.edgePairLists[3].removeTile(loc.tile.positionInEdgePairList[3])\n\t\t\t\t// move to next position on board\n\t\t\t\tnextPos.edgePairList = edgePairList\n\t\t\t\tnextPos.index = 0\n\t\t\t\tnextPos.listSize = edgePairList.availableNoTiles // just for debug\n\t\t\t\tloc = nextPos\n\t\t\t\tcontinue\n\t\t\t} else { // move onto next tile in list\n\t\t\t\tloc.index++\n\t\t\t\tcontinue\n\t\t\t}\n\t\t} else {\n\t\t\t// backtrack\n\t\t\t//fmt.Println(\"Backtracking\")\n\t\t\tprogress--\n\t\t\t// traverse to previous tile\n\t\t\tloc = loc.traversePrev\n\t\t\t// remove the tile in the previous location ...\n\t\t\tloc.tile.edgePairLists[3].restoreTile()\n\t\t\tloc.tile.edgePairLists[2].restoreTile()\n\t\t\tloc.tile.edgePairLists[1].restoreTile()\n\t\t\tloc.tile.edgePairLists[0].restoreTile()\n\t\t\t//loc.tile = nil // probably not required but handy, used when we print out progress ...\n\t\t\tloc.index++\n\n\t\t}\n\t}\n\n\tfmt.Println(\"Number of iterations to solution:\", numberInterations)\n}", "func findDos() ([]dos, error) {\r\n\r\n\tvar d []dos\r\n\tfor true {\r\n\r\n\t\t//fmt.Printf(\"Visited Person: %v, %d\\n\\n\", buff.visitedPerson, len(buff.visitedPerson))\r\n\r\n\t\tfor _, person := range buff.visit {\r\n\r\n\t\t\t/*fmt.Printf(\"%s\\n\\n\", person)\r\n\t\t\tif buff.visitedPerson[person] {\r\n\t\t\t\tcontinue\r\n\t\t\t}\r\n\t\t\tbuff.visitedPerson[person] = true\r\n\t\t\t*/\r\n\r\n\t\t\tperson1, err := fetchData(person)\r\n\t\t\tif err != nil {\r\n\t\t\t\tif strings.Contains(err.Error(), \"looking for beginning of value\") {\r\n\t\t\t\t\tcontinue\r\n\t\t\t\t}\r\n\t\t\t\treturn nil, err\r\n\t\t\t}\r\n\r\n\t\t\tfor _, p1movie := range person1.Movies {\r\n\t\t\t\tif buff.p2Movies[p1movie.Url].Url == p1movie.Url {\r\n\t\t\t\t\tif _, found := buff.link[person1.Url]; found {\r\n\t\t\t\t\t\td = append(d, buff.link[person1.Url], dos{p1movie.Name, person1.Name, p1movie.Role, buff.person2.Name, buff.p2Movies[p1movie.Url].Role})\r\n\t\t\t\t\t} else {\r\n\t\t\t\t\t\td = append(d, dos{p1movie.Name, person1.Name, p1movie.Role, buff.person2.Name, buff.p2Movies[p1movie.Url].Role})\r\n\t\t\t\t\t}\r\n\t\t\t\t\treturn d, nil\r\n\t\t\t\t}\r\n\t\t\t}\r\n\r\n\t\t\t// Find new nodes to continue searching\r\n\t\t\tfor _, p1movie := range person1.Movies {\r\n\r\n\t\t\t\tif buff.visited[p1movie.Url] {\r\n\t\t\t\t\tcontinue\r\n\t\t\t\t}\r\n\r\n\t\t\t\tbuff.visited[p1movie.Url] = true\r\n\r\n\t\t\t\tp1moviedetail, err := fetchData(p1movie.Url)\r\n\t\t\t\tif err != nil {\r\n\t\t\t\t\tif strings.Contains(err.Error(), \"looking for beginning of value\") {\r\n\t\t\t\t\t\tcontinue\r\n\t\t\t\t\t}\r\n\t\t\t\t\treturn nil, err\r\n\t\t\t\t}\r\n\r\n\t\t\t\tfor _, p1moviecast := range p1moviedetail.Cast {\r\n\r\n\t\t\t\t\tif buff.visited[p1moviecast.Url] {\r\n\t\t\t\t\t\tcontinue\r\n\t\t\t\t\t}\r\n\r\n\t\t\t\t\tbuff.visited[p1moviecast.Url] = true\r\n\t\t\t\t\tbuff.visit = append(buff.visit, p1moviecast.Url)\r\n\t\t\t\t\tbuff.link[p1moviecast.Url] = dos{p1movie.Name, person1.Name, p1movie.Role, p1moviecast.Name, p1moviecast.Role}\r\n\t\t\t\t}\r\n\r\n\t\t\t\tfor _, p1moviecrew := range p1moviedetail.Crew {\r\n\r\n\t\t\t\t\tif buff.visited[p1moviecrew.Url] {\r\n\t\t\t\t\t\tcontinue\r\n\t\t\t\t\t}\r\n\r\n\t\t\t\t\tbuff.visited[p1moviecrew.Url] = true\r\n\t\t\t\t\tbuff.visit = append(buff.visit, p1moviecrew.Url)\r\n\t\t\t\t\tbuff.link[p1moviecrew.Url] = dos{p1movie.Name, person1.Name, p1movie.Role, p1moviecrew.Name, p1moviecrew.Role}\r\n\t\t\t\t}\r\n\r\n\t\t\t}\r\n\t\t}\r\n\r\n\t\t//fmt.Printf(\"Visit: %v, %d\\n\\n\", buff.visit, len(buff.visit))\r\n\t\t//fmt.Printf(\"Visited: %v, %d\\n\\n\", buff.visited, len(buff.visited))\r\n\r\n\t}\r\n\r\n\treturn d, nil\r\n}", "func (m *Map) BfsUpdate(loc Location, fn uFunc) LocDir {\n\n\t//ret acts as the default return when either nothing is found\n\t//or the goal state is the cell.\n\tret := LocDir{loc, NoMovement}\n\tret.Loc = -1\n\tret.Dir = NoMovement\n\n\tvar depth int\n\tfrontier := new(Queue) // make this Queue a type of LocDir\n\tvar inFrontier = make(map[Location]bool)\n\tvar explored = make(map[Location]bool) // the keys are only locations\n\n\tfrontier.Push(ret) // frontier is queue of LocDir\n\tinFrontier[loc] = true // keys to inFrontier are simply locations\n\t// I'm not sure whether I should set the keys to frontier to be a LocDir\n\t// as well.\n\n\tfor {\n\t\t// Depth should be the viewRadius, but I'm not sure how to get this at\n\t\t// runtime yet.\n\t\tif frontier.Len() == 0 || depth > 7 {\n\t\t\treturn ret\n\t\t}\n\n\t\tcurLoc := frontier.Pop().(LocDir)\n\t\tinFrontier[curLoc.Loc] = false\n\t\texplored[curLoc.Loc] = true\n\n\t\tfn(curLoc.Loc, depth, m) //update function call\n\n\t\t// Loop over adjacent Locations, action is a LocDir structure\n\t\tfor _, action := range m.Adjacent(curLoc.Loc) {\n\t\t\t//if child not in explored or frontier\n\t\t\tif !explored[action.Loc] || !inFrontier[action.Loc] {\n\t\t\t\tfrontier.Push(action)\n\t\t\t\tinFrontier[action.Loc] = true\n\t\t\t}\n\t\t}\n\t\tdepth++\n\t}\n\treturn ret\n}", "func Dijkstra(g Graph, src uint32, weightFn func(uint32, uint32) float32, withPreds bool) DijkstraState {\n\tnv := g.NumVertices()\n\tvertLevel := make([]uint32, nv)\n\tfor i := u0; i < nv; i++ {\n\t\tvertLevel[i] = unvisited\n\t}\n\tcurLevel := make([]uint32, 0, nv)\n\tnextLevel := make([]uint32, 0, nv)\n\tnLevel := uint32(2)\n\tparents := make([]uint32, nv)\n\tpathcounts := make([]uint32, nv)\n\tdists := make([]float32, nv)\n\n\tpreds := make([][]uint32, 0)\n\tif withPreds {\n\t\tpreds = make([][]uint32, nv)\n\t}\n\n\tfor i := range dists {\n\t\tdists[i] = maxDist\n\t}\n\n\tvertLevel[src] = 0\n\tdists[src] = 0\n\tparents[src] = src\n\tpathcounts[src] = 1\n\tcurLevel = append(curLevel, src)\n\tfor len(curLevel) > 0 {\n\t\tfor _, u := range curLevel {\n\t\t\tfor _, v := range g.OutNeighbors(u) {\n\t\t\t\talt := min(maxDist, dists[u]+weightFn(u, v))\n\t\t\t\tif vertLevel[v] == unvisited { // if not visited\n\t\t\t\t\tdists[v] = alt\n\t\t\t\t\tparents[v] = u\n\t\t\t\t\tpathcounts[v] += pathcounts[u]\n\t\t\t\t\tif withPreds {\n\t\t\t\t\t\tpreds[v] = append(preds[v], u)\n\t\t\t\t\t}\n\t\t\t\t\tnextLevel = append(nextLevel, v)\n\t\t\t\t\tvertLevel[v] = nLevel\n\t\t\t\t} else {\n\t\t\t\t\tif alt < dists[v] {\n\t\t\t\t\t\tdists[v] = alt\n\t\t\t\t\t\tparents[v] = u\n\t\t\t\t\t\tpathcounts[v] = 0\n\t\t\t\t\t\tif withPreds {\n\t\t\t\t\t\t\tpreds[v] = preds[v][:0]\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\tif alt == dists[v] {\n\t\t\t\t\t\tpathcounts[v] += pathcounts[u]\n\t\t\t\t\t\tif withPreds {\n\t\t\t\t\t\t\tpreds[v] = append(preds[v], u)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tfmt.Printf(\"completed level %d, size = %d\\n\", nLevel-1, len(nextLevel))\n\t\tnLevel++\n\t\tcurLevel = curLevel[:0]\n\t\tcurLevel, nextLevel = nextLevel, curLevel\n\t\tzuint32.SortBYOB(curLevel, nextLevel[:nv])\n\t}\n\tpathcounts[src] = 1\n\tparents[src] = 0\n\tif withPreds {\n\t\tpreds[src] = preds[src][:0]\n\t}\n\tds := DijkstraState{\n\t\tParents: parents,\n\t\tDists: dists,\n\t\tPathcounts: pathcounts,\n\t\tPredecessors: preds,\n\t}\n\treturn ds\n}", "func (m *MarkovChain) longPaths(source string, n int) []string {\n\t// set up min weights\n\tdistances := make(map[string]float64)\n\tdistances[source] = 0\n\n\t// store removed nodes so we can fetch the closest values and check if something has been removed\n\tremoved_nodes := make(map[string]bool)\n\tclosest_files := make([]string, 0)\n\tnRemoved := 0\n\n\t// store current guesses\n\tqueue := heap.MakeMinHeapFloat64()\n\n\t// relax all edges from source\n\tsrc_node, ok := m.nodes[source]\n\tif !ok {\n\t\tlog.Fatalf(\"THIS SHOULD NEVER HAPPEN [source node not found] %v -> %v\", source, m.nodes)\n\t}\n\n\t// initialize with all of the adjacencies of the source node\n\tfor _, neighbor := range src_node.adjacencies {\n\t\t// weights are the negated log of the edge ratio -> min path weight becomes max product (max probability)\n\t\tweight := -math.Log((float64(neighbor.count) / float64(src_node.count)))\n\t\tdistances[neighbor.name] = weight\n\t\tqueue.Insert(neighbor.name, weight)\n\t}\n\n\t// now run Dijkstra's\n\tfor queue.Size > 0 && nRemoved < n {\n\t\tname := queue.ExtractMin()\n\t\tnode := m.nodes[name]\n\t\testimate := distances[name]\n\t\t// this file is close in probability, so remove it from valid candidates\n\t\tremoved_nodes[name] = true\n\t\tclosest_files = append(closest_files, name)\n\n\t\t// iterate through all neighbors of this file\n\t\tfor _, transition := range node.adjacencies {\n\t\t\t// check if neighbor file has been seen before\n\t\t\tif _, ok := distances[transition.name]; !ok {\n\t\t\t\t// not seen before, set probability estimate and insert into heap\n\t\t\t\tdistances[transition.name] = math.Inf(1)\n\t\t\t\tqueue.Insert(transition.name, math.Inf(1))\n\t\t\t}\n\n\t\t\tif _, ok := removed_nodes[name]; (!ok && transition.name != source) {\n\t\t\t\t// this neighbor has not been removed already and is not the source node\n\t\t\t\t// then try to relax weight estimate\n\t\t\t\tweight := -math.Log((float64(transition.count) / float64(node.count)))\n\t\t\t\tif (weight + estimate) < distances[transition.name] {\n\t\t\t\t\t// then relax this edge\n\t\t\t\t\tdistances[transition.name] = (weight + estimate)\n\t\t\t\t\t// this will insert if not already found\n\t\t\t\t\tqueue.ChangeKey(transition.name, (weight + estimate))\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn closest_files\n}", "func bfsBlizzardWalk(bz blizzards, start, end int) (blizzards, int) {\n\tsteps := []int{0, -bz.w, bz.w, -1, 1} // wait, up, down, left, right\n\tstack := map[int]struct{}{start: {}}\n\tfor t := 0; ; t++ {\n\t\tbz = bz.next()\n\t\tnewStack := make(map[int]struct{})\n\t\tfor p := range stack {\n\t\t\tfor _, step := range steps {\n\t\t\t\tnp := p + step\n\t\t\t\tif np >= 0 && np < len(bz.valley) && bz.valley[np] == clear {\n\t\t\t\t\tif np == end {\n\t\t\t\t\t\treturn bz, t + 1\n\t\t\t\t\t}\n\t\t\t\t\tnewStack[np] = struct{}{}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tstack = newStack\n\t}\n}", "func printMST(parent []int, vertices int, graph [][]float64, k int) {\n\n\tfmt.Println(\"Edge Weight\")\n\n\tfor i := 1; i < vertices; i++ {\n\t\tif comp.numberOfComponents != 1 {\n\t\t\tfmt.Printf(\"%d - %d %f \\n\", comp.compMaps[k][parent[i]], comp.compMaps[k][i], graph[i][parent[i]])\n\t\t\tadj_forest[comp.compMaps[k][parent[i]]-1][comp.compMaps[k][i]-1] = graph[i][parent[i]]\n\t\t\tadj_forest[comp.compMaps[k][i]-1][comp.compMaps[k][parent[i]]-1] = graph[i][parent[i]]\n\t\t} else {\n\t\t\tfmt.Printf(\"%d - %d %f \\n\", comp.compMaps[k][parent[i]+1], comp.compMaps[k][i+1], graph[i][parent[i]])\n\t\t\tadj_forest[comp.compMaps[k][parent[i]]][comp.compMaps[k][i]] = graph[i][parent[i]]\n\t\t\tadj_forest[comp.compMaps[k][i]][comp.compMaps[k][parent[i]]] = graph[i][parent[i]]\n\t\t}\n\n\t}\n\n}", "func dfs(cur, n int, res *[]int) {\r\n\tif cur > n {\r\n\t\treturn\r\n\t}\r\n\t*res = append(*res, cur)\r\n\tfor i := 0; i < 10; i++ {\r\n\t\tif 10 * cur + i > n {\r\n\t\t\treturn\r\n\t\t}\r\n\t\tdfs(10 * cur + i, n, res)\r\n\t}\r\n}", "func SolutionDFS(root *BinaryTreeNode) []*list.List {\n\tresult := []*list.List{}\n\n\tvar dfs func(*BinaryTreeNode, int)\n\n\tdfs = func(root *BinaryTreeNode, level int) {\n\t\tif root == nil {\n\t\t\treturn\n\t\t}\n\n\t\t// found new level\n\t\tif level == len(result) {\n\t\t\tresult = append(result, list.New())\n\t\t}\n\t\tresult[level].PushBack(root.Value)\n\t\tdfs(root.Left, level+1)\n\t\tdfs(root.Right, level+1)\n\n\t}\n\n\tdfs(root, 0)\n\n\treturn result\n}", "func FindPath(mp Graph, start, end Node) ([]Node, error) {\n\tmapCapacity := int(end - start)\n\tif mapCapacity < 0 {\n\t\tmapCapacity = -mapCapacity\n\t}\n\tif mapCapacity > maxDefaultMapCapacity {\n\t\tmapCapacity = maxDefaultMapCapacity\n\t}\n\t// The open list is ordered by the sum of current cost + heuristic cost\n\tstate := newState(mapCapacity)\n\t// Add the start node to the openlist\n\tpCost, err := mp.HeuristicCost(start, end)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tstate.addNodeInfo(&nodeInfo{\n\t\tnode: start,\n\t\tparent: -1,\n\t\tcost: 0.0,\n\t\tpredictedCost: float32(pCost),\n\t})\n\n\tedgeSlice := make([]Edge, 0, 8)\n\tfor {\n\t\tcurrent := state.popBest()\n\t\tif current == nil {\n\t\t\treturn nil, ErrImpossible\n\t\t}\n\t\tif current.node == end {\n\t\t\t// If we reached the end node then we know the optimal path. Traverse\n\t\t\t// it (backwards) and return an array of node IDs.\n\t\t\treturn state.pathToNode(current), nil\n\t\t}\n\t\tif current.cost >= state.maxCost {\n\t\t\tcontinue\n\t\t}\n\t\tif dbg, ok := mp.(Debug); ok {\n\t\t\tdbg.VisitedNode(current.node, current.parent, float64(current.cost), float64(current.predictedCost))\n\t\t}\n\t\tneighbors, err := mp.Neighbors(current.node, edgeSlice[:0])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tfor _, edge := range neighbors {\n\t\t\t// Don't try go backwards\n\t\t\tif edge.Node == current.parent {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Cost for the neighbor node is the current cost plus the\n\t\t\t// cost to get to that node.\n\t\t\tcost := current.cost + float32(edge.Cost)\n\n\t\t\tni := state.info[edge.Node]\n\t\t\tif ni == nil {\n\t\t\t\t// We haven't seen this node so add it to the open list.\n\t\t\t\tpCost, err := mp.HeuristicCost(edge.Node, end)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\t\t\t\tni = &nodeInfo{\n\t\t\t\t\tnode: edge.Node,\n\t\t\t\t\tparent: current.node,\n\t\t\t\t\tcost: cost,\n\t\t\t\t\tpredictedCost: float32(pCost),\n\t\t\t\t}\n\t\t\t\tstate.addNodeInfo(ni)\n\t\t\t} else if cost < ni.cost {\n\t\t\t\t// We've seen this node and the current path is cheaper\n\t\t\t\t// so update the changed info and add it to the open list\n\t\t\t\t// (replacing if necessary).\n\t\t\t\tni.parent = current.node\n\t\t\t\tni.cost = cost\n\t\t\t\tif ni.index >= 0 {\n\t\t\t\t\tstate.updateNodeInfo(ni)\n\t\t\t\t} else {\n\t\t\t\t\tstate.addNodeInfo(ni)\n\t\t\t\t}\n\t\t\t} else if edge.Node == end {\n\t\t\t\tif cost < state.maxCost {\n\t\t\t\t\tstate.maxCost = cost\n\t\t\t\t}\n\t\t\t\tif pp, ok := mp.(PossiblePath); ok {\n\t\t\t\t\tpath := append(state.pathToNode(current), end)\n\t\t\t\t\tpp.PossiblePath(path, float64(cost))\n\t\t\t\t}\n\t\t\t\tni = nil\n\t\t\t}\n\t\t\tif ni != nil && edge.Node == end {\n\t\t\t\tif cost < state.maxCost {\n\t\t\t\t\tstate.maxCost = cost\n\t\t\t\t}\n\t\t\t\tif pp, ok := mp.(PossiblePath); ok {\n\t\t\t\t\tpp.PossiblePath(state.pathToNode(ni), float64(ni.cost))\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func dfs(candidates []int, target int, s int, solution []int, sum int, ans *[][]int) {\n\tif sum == target {\n\t\ts := make([]int, len(solution))\n\t\tcopy(s, solution)\n\t\t*ans = append(*ans, s)\n\t\treturn\n\t}\n\n\tfor i := s; i < len(candidates); i++ {\n\t\tif sum+candidates[i] > target {\n\t\t\t// Stop further iterations if sum+candidates[i]\n\t\t\t// has already exceeded the value of target.\n\t\t\t// (We've sorted candidates already.)\n\t\t\treturn\n\t\t}\n\n\t\tsolution = append(solution, candidates[i]) // Push\n\t\t// Different from pure combination problem,\n\t\t// we can use current candidate repeatedly,\n\t\t// thus pass 'i' instead of 'i+1' to next DFS.\n\t\tdfs(candidates, target, i, solution, sum+candidates[i], ans)\n\t\tsolution = solution[:len(solution)-1] // Pop\n\t}\n}", "func (g Graph) ShortestPath(orig Place, by Accessor) PathMap {\n\n\tinf := math.Inf(1)\n\tnone := Place{} // zero val\n\tvar d pdata // temp var for data\n\n\t// 1. mark all nodes unvisitied. create a set of all unvisited nodes\n\t// call the unvisited set\n\t// 2. assign to every node a tentative distance value: zero for initial node\n\t// and infinity (\"unvisited\") for all others. Set initial node as current.\n\tnodes := make(PathMap, len(g))\n\tfor k := range g {\n\t\tnodes[k] = pdata{Dist: inf}\n\t}\n\n\tcurrent := orig\n\td = nodes[current]\n\td.Dist = 0\n\tnodes[current] = d\n\n\tfound := false // aka done\n\n\tfor !found {\n\t\t// fmt.Println(\"current\", current, nodes[current])\n\t\tif current == none {\n\t\t\treturn nil\n\t\t}\n\n\t\t// 3. for the current node, consider all its unvisited neighbors and\n\t\t// calculate their tentative distances through the current node. Compare\n\t\t// the newly calculated tentative distance to the currently assigned value\n\t\t// and assign the smaller value.\n\t\tfor n, w := range g[current] {\n\t\t\tif !nodes[n].visited { // n in unvisited set\n\t\t\t\ttentative := nodes[current].Dist + by(w)\n\t\t\t\td = nodes[n]\n\t\t\t\tif d.Dist > tentative {\n\t\t\t\t\td.Dist = tentative\n\t\t\t\t\td.parent = current\n\t\t\t\t\td.Hops = nodes[d.parent].Hops + 1\n\t\t\t\t\tnodes[n] = d\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\t// 4. when we are done considering all the unvisited neighbors of the\n\t\t// current node, mark the current node as visited and remove it from the\n\t\t// unvisited set. A visited node will never be checked again.\n\t\td = nodes[current]\n\t\td.visited = true\n\t\tnodes[current] = d\n\n\t\t// 5. A) if all nodes are marked visited (unvisited set is empty)\n\t\t// OR B) if the smallest tentative distance among nodes in the unvisited set\n\t\t// is infinity (no path possible)\n\t\t// The algorithm is finished.\n\t\t// TODO: termination case B\n\t\tunvisitedcount := 0\n\t\tfor _, d := range nodes {\n\t\t\tif !d.visited {\n\t\t\t\tunvisitedcount++\n\t\t\t}\n\t\t}\n\n\t\tfound = unvisitedcount == 0\n\t\tif found {\n\t\t\tcontinue\n\t\t}\n\n\t\t// 6. Otherwise, select the unvisited node that is marked with the smallest\n\t\t// tentative value, set it as the \"current\" and go back to step 3.\n\t\tminDist := inf // pos infinity\n\t\tminPlace := Place{}\n\t\tfor node, d := range nodes {\n\t\t\tif !d.visited && d.Dist < minDist {\n\t\t\t\tminDist = d.Dist\n\t\t\t\tminPlace = node\n\t\t\t}\n\t\t}\n\t\tcurrent = minPlace\n\t\tfound = minDist == inf // termination case 5B above\n\t}\n\n\treturn nodes\n}", "func (level *Level) bfsearch(start Pos) {\n\tedge := make([]Pos, 0, 8)\n\tedge = append(edge, start)\n\tvisited := make(map[Pos]bool)\n\tvisited[start] = true\n\tlevel.Debug = visited\n\n\tfor len(edge) > 0 {\n\t\tcurrent := edge[0]\n\t\tedge = edge[1:]\n\t\tfor _, next := range getNeighbours(level, current) {\n\t\t\tif !visited[next] {\n\t\t\t\tedge = append(edge, next)\n\t\t\t\tvisited[next] = true\n\t\t\t}\n\t\t}\n\t}\n}", "func dfs(board [][]byte, rowIndex int, columnIndex int, currentLetterIndex int, word string) bool {\n\t/*\n\t \tASFBA\n\t board =\n\t [[\"A\",\"B\",\"C\",\"E\"],\n\t [\"S\",\"F\",\"C\",\"S\"],\n\t [\"A\",\"D\",\"E\",\"E\"]]\n\n\t word = \"ABCCED\"\n\t*/\n\tif currentLetterIndex == len(word) {\n\t\treturn true\n\t}\n\n\trows := len(board)\n\tcolumns := len(board[0])\n\n\tif rowIndex < 0 || rowIndex >= rows || columnIndex < 0 || columnIndex >= columns {\n\t\treturn false\n\t}\n\n\tletterThatIWant := word[currentLetterIndex]\n\tcurrentBoardLetter := board[rowIndex][columnIndex]\n\n\tif currentBoardLetter != letterThatIWant {\n\t\treturn false\n\t}\n\n\tboard[rowIndex][columnIndex] = ' '\n\n\twordHasBeenFound := dfs(board, rowIndex-1, columnIndex, currentLetterIndex+1, word) ||\n\t\tdfs(board, rowIndex+1, columnIndex, currentLetterIndex+1, word) ||\n\t\tdfs(board, rowIndex, columnIndex-1, currentLetterIndex+1, word) ||\n\t\tdfs(board, rowIndex, columnIndex+1, currentLetterIndex+1, word)\n\n\tboard[rowIndex][columnIndex] = currentBoardLetter\n\n\treturn wordHasBeenFound\n}", "func findDirections(x, y int, length *int, value int, directions []dirTable) int {\n num := 0\n numChecks := 0\n if value != wall || (getMaze(x, y) == path && setCell(x, y, check, noUpdate, *length, numChecks)) {\n minLength := [4]int {*length, *length, *length, *length}\n len := *length\n for {\n setInt(&dspLength, len)\n dirLength := [4]int {len, len, len, len}\n offset := rand.Intn(4)\n for i := 0; i < 4; i++ {\n dir := &stdDirection[(i + offset) % 4]\n num += look(dir.heading, x, y, dir.x, dir.y, num, value, directions, &dirLength[i] , &minLength[i], &numChecks)\n }\n if num > 0 || len < 0 {\n break\n }\n minLength := min( min(minLength[0], minLength[1]), min(minLength[2], minLength[3]) )\n if minLength <= 0 {\n minLength = 1\n }\n len -= minLength\n }\n if len == *length && len < getInt(&depth) {\n len++\n }\n *length = len\n if getMaze(x, y) == check {\n setMaze(x, y, path)\n }\n }\n if getInt(&maxChecks) < numChecks {\n setInt(&maxChecks , numChecks)\n }\n return (num);\n}", "func (g AdjacencyList) DepthFirst(start NI, options ...TraverseOption) {\n\tcf := &config{start: start}\n\tfor _, o := range options {\n\t\to(cf)\n\t}\n\tb := cf.visBits\n\tif b == nil {\n\t\tn := bits.New(len(g))\n\t\tb = &n\n\t} else if b.Bit(int(cf.start)) != 0 {\n\t\treturn\n\t}\n\tif cf.pathBits != nil {\n\t\tcf.pathBits.ClearAll()\n\t}\n\tvar df func(NI) bool\n\tdf = func(n NI) bool {\n\t\tb.SetBit(int(n), 1)\n\t\tif cf.pathBits != nil {\n\t\t\tcf.pathBits.SetBit(int(n), 1)\n\t\t}\n\n\t\tif cf.nodeVisitor != nil {\n\t\t\tcf.nodeVisitor(n)\n\t\t}\n\t\tif cf.okNodeVisitor != nil {\n\t\t\tif !cf.okNodeVisitor(n) {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\n\t\tif cf.rand == nil {\n\t\t\tfor x, to := range g[n] {\n\t\t\t\tif cf.arcVisitor != nil {\n\t\t\t\t\tcf.arcVisitor(n, x)\n\t\t\t\t}\n\t\t\t\tif cf.okArcVisitor != nil {\n\t\t\t\t\tif !cf.okArcVisitor(n, x) {\n\t\t\t\t\t\treturn false\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tif b.Bit(int(to)) != 0 {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif !df(to) {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tto := g[n]\n\t\t\tfor _, x := range cf.rand.Perm(len(to)) {\n\t\t\t\tif cf.arcVisitor != nil {\n\t\t\t\t\tcf.arcVisitor(n, x)\n\t\t\t\t}\n\t\t\t\tif cf.okArcVisitor != nil {\n\t\t\t\t\tif !cf.okArcVisitor(n, x) {\n\t\t\t\t\t\treturn false\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tif b.Bit(int(to[x])) != 0 {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif !df(to[x]) {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif cf.pathBits != nil {\n\t\t\tcf.pathBits.SetBit(int(n), 0)\n\t\t}\n\t\treturn true\n\t}\n\tdf(cf.start)\n}", "func (g *graph) find_str_conn_component() {\n\tif len(g.nodes) <= 0 { return }\n\n\tchk := &checklist{make([]int,len(g.nodes)),0}\n\tfor i,_ := range g.nodes {\n\t\tif chk.nodes_count[i] == 0 {\n\t\t\tg.dfs(chk,i)\n\t\t}\n\t}\n\n\tg_rev := g.reverse()\n\tchk_rev := &checklist{make([]int,len(g.nodes)),0}\n\tcomponent := list.New()\n\tnum_checked := 0\n\n\tfor chk_rev.count <= len(g.nodes) {\n\t\tidx := g_rev.find_max_value(chk)\n\t\t\n\t\tif idx < 0 { break }\n\n\t\tif chk_rev.nodes_count[idx] == 0 {\n\t\t\tg_rev.dfs(chk_rev,idx)\n\t\t} else { continue }\n\n\t\tlist := list.New()\n\t\tfor i,c := range chk_rev.nodes_count {\n\t\t\tif c > num_checked && c <= chk_rev.count {\n\t\t\t\tlist.PushBack(i)\n\t\t\t}\n\t\t}\n\t\tnum_checked = chk_rev.count\n\t\tcomponent.PushBack(list)\n\t}\n\n\t//Print out strongly connected components\n\ti := 0\n\tfor l := range component.Iter() {\n\t\ti++\n\t\tfmt.Printf(\"strongly connected component %d:\\n\", i)\n\t\tfor sub := range l.(*list.List).Iter() {\n\t\t\tfmt.Printf(\"%d\\t\", sub.(int))\n\t\t}\n\t\tfmt.Println(\"\\n\")\n\t}\n\n}", "func search(graph []*node, from int) []int {\n\t// fmt.Printf(\"starting at node %d in a graph of size %d\\n\", from, len(graph))\n\tvar queue []*node\n\tqueue = append(queue, graph[from])\n\tfor len(queue) > 0 {\n\t\tcurrent := queue[0]\n\t\tqueue = queue[1:]\n\t\tfor _, neigh := range current.neighbors {\n\t\t\tif neigh.dist == 0 {\n\t\t\t\tneigh.dist = 6 + current.dist\n\t\t\t\tqueue = append(queue, neigh)\n\t\t\t}\n\n\t\t}\n\t}\n\tvar result []int\n\tfor _, n := range graph {\n\t\tif n.id == from {\n\t\t\tcontinue\n\t\t}\n\t\tresult = append(result, n.dist)\n\t}\n\treturn result\n}", "func makePath(m GridMap, nodes []node, goali int) (path []Loc) {\n\tif nodes[goali].parent < 0 {\n\t\treturn\n\t}\n\n\tvar rev []Loc\n\tstride := m.Height()\n\tfor i := goali; nodes[i].parent >= 0; i = nodes[i].parent {\n\t\tx, y := i/stride, i%stride\n\t\trev = append(rev, Loc{x, y})\n\t}\n\tfor _, loc := range rev {\n\t\tpath = append(path, loc)\n\t}\n\treturn\n}", "func BFS(s string) {\n\tvar ll linklist.LL\n\n\tv := graphs.Gmap[s]\n\n\tv.C = 1\n\n\tll = ll.LlStart()\n\tll = ll.LlEnqueue(v)\n\n\tfor !ll.LlEmpty() {\n\t\tu := ll.LlPop()\n\t\tw := u.(*graphs.Vertex)\n\t\tt := w.Adj\n\n\t\tif t != nil {\n\t\t\tfor i := 0; i < len(t); i++ {\n\t\t\t\tif t[i].C == 0 {\n\t\t\t\t\tt[i].C = 1\n\t\t\t\t\tt[i].TI = w.TI + 1\n\t\t\t\t\tt[i].P = w\n\t\t\t\t\tll = ll.LlEnqueue(t[i])\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n}", "func minKnightMoves(ex int, ey int) int {\r\n\tif ex == 0 && ey == 0 {\r\n\t\treturn 0\r\n\t}\r\n\tq := [][]int{}\r\n\tq = append(q, []int{0, 0})\r\n\tdir := [][]int{{1, 2}, {2, 1}, {2, -1}, {1, -2}, {-1, 2}, {-2, 1}, {-1, -2}, {-2, -1}}\r\n\tsteps := 0\r\n\tseen := map[int]map[int]bool{0: map[int]bool{0: true}}\r\n\t// BFS loop\r\n\tfor len(q) > 0 {\r\n\t\tsteps++\r\n\t\tn := len(q)\r\n\t\tfor i := 0; i < n; i++ {\r\n\t\t\tt := q[0]\r\n\t\t\tq = q[1:]\r\n\t\t\tfor _, d := range dir {\r\n\t\t\t\tnx := t[0] + d[0]\r\n\t\t\t\tny := t[1] + d[1]\r\n\t\t\t\tif nx == ex && ny == ey {\r\n\t\t\t\t\treturn steps\r\n\t\t\t\t}\r\n\t\t\t\tif _, exists := seen[nx]; !exists {\r\n\t\t\t\t\tseen[nx] = map[int]bool{}\r\n\t\t\t\t}\r\n\t\t\t\tif _, exists := seen[nx][ny]; !exists &&\r\n\t\t\t\t\t(nx*ex >= 0 && ny*ey >= 0 || steps < 3) {\r\n\t\t\t\t\tseen[nx][ny] = true\r\n\t\t\t\t\tq = append(q, []int{nx, ny})\r\n\t\t\t\t}\r\n\t\t\t}\r\n\t\t}\r\n\t}\r\n\treturn -1\r\n}", "func improve(graph ls.IGraph) (bool, ls.IGraph, []int) {\n\t//subset of dependent with biggest param\n\tmainSet := make([]int, 0)\n\t//subset of independent with biggest param\n\tmainIndependetSet := make([]int, 0)\n\t//biggest diff between subsets\n\tbestParam := 0\n\n\t//check all subsets\n\tfor i := graph.GetAmountOfIndependent(); i < graph.AmountOfVertex(); i++ {\n\t\t//intermediate subset of dependent\n\t\tsubSet := make([]int, 1)\n\t\t//first vertex must be added to sub set any way\n\t\tsubSet[0] = i\n\n\t\t//set of banned vertex - vertex that cannot create independent subset\n\t\tedgeSet := make([]int, 0)\n\t\tedgeSet = betterslice.AppendWithOutRepeatInt(edgeSet, graph.GetEdges(i)...)\n\n\t\tfor j := i + 1; j < graph.AmountOfVertex(); j++ {\n\t\t\t//check is vertex banned\n\t\t\tjflag := !isInSubSet(j, edgeSet)\n\n\t\t\t//if vertex not banned\n\t\t\tif jflag {\n\t\t\t\t//add vertex to subset\n\t\t\t\tsubSet = append(subSet, j)\n\t\t\t\t//make subset of banned vertex bigger\n\t\t\t\t//todo: to think avout better perforamnce, @AppendWithOutRepeatInt can be replaced with @append\n\t\t\t\tedgeSet = betterslice.AppendWithOutRepeatInt(edgeSet, graph.GetEdges(j)...)\n\t\t\t}\n\t\t}\n\n\t\t//subset of independent vertex that connected to vertex in dependent subset\n\t\tindependetVertexSet := make([]int, 0)\n\n\t\t//fill independentVertexSet\n\t\tfor _, vertex := range subSet {\n\t\t\tfor _, i := range graph.GetEdges(vertex) {\n\t\t\t\tif i < graph.GetAmountOfIndependent() && !isInSubSet(i, independetVertexSet) {\n\t\t\t\t\tindependetVertexSet = append(independetVertexSet, i)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\t//count value of transition\n\t\tparam := len(subSet) - len(independetVertexSet)\n\t\t//update best value\n\t\tif param > bestParam {\n\t\t\tbestParam = param\n\t\t\tmainSet = subSet\n\t\t\tmainIndependetSet = independetVertexSet\n\t\t}\n\t}\n\n\t//if no positiov transions then stop\n\tif bestParam <= 0 {\n\t\tnewOrd := make([]int, graph.AmountOfVertex())\n\t\tnewOrd = initOrd(newOrd)\n\t\treturn false, graph, newOrd\n\t}\n\n\t//constract new order of vertex\n\tnewOrd := constractNewOrder(mainSet, mainIndependetSet, graph.GetAmountOfIndependent(), graph.AmountOfVertex())\n\n\t//renumber vertex with new ordet\n\tgraph.RenumVertex(newOrd)\n\t//set new amount of independent subset\n\tgraph.SetAmountOfIndependent(graph.GetAmountOfIndependent() + bestParam)\n\treturn true, graph, newOrd\n}", "func printGraphDFS(adjMatrix [7][7]int, StartVertex int) {\n\tv := len(adjMatrix)\n\tvisited := make([]bool, v)\n\t//This loop is to check edge like 5-6\n\tfor i := 0; i < v; i++ {\n\t\tif visited[i] == false {\n\t\t\tprintDfsHelper(adjMatrix, visited, StartVertex)\n\t\t}\n\t}\n\n}", "func traversalBFS(current []Node, filterFn func(node Node) (gotoNextLayer bool), processFn func(node Node) (gotoNextLayer bool), isRoot bool) (gotoNextLayer bool) {\n\tif len(current) == 0 {\n\t\treturn false\n\t}\n\t// Step 1: brothers layer\n\tnextBrothers := []Node{}\n\tfor _, node := range current {\n\t\t// filter root\n\t\tif isRoot {\n\t\t\tif !filterFn(node) {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t\tif !processFn(node) {\n\t\t\treturn false\n\t\t}\n\t\t// filter brothers\n\t\tnextBrothers = append(nextBrothers, node)\n\t}\n\n\t// Step 2: children layer\n\tnextChildren := []Node{}\n\t// filter children\n\tfor _, node := range nextBrothers {\n\t\t// Scan node for nodes to include.\n\t\tnextChildren = append(nextChildren, filterChildren(node, node.LeftNodes(), filterFn)...)\n\t\tnextChildren = append(nextChildren, filterChildren(node, node.MiddleNodes(), filterFn)...)\n\t\tnextChildren = append(nextChildren, filterChildren(node, node.RightNodes(), filterFn)...)\n\t}\n\ttraversalBFS(nextChildren, filterFn, processFn, false)\n\treturn true\n}", "func sumOfDistancesInTree(N int, edges [][]int) []int {\n \n}", "func searchBestOpenings(x, y *int) {\n bestPathLen := 0\n bestTurnCnt := 0\n bestStart := 2\n bestFinish := 2\n saveDelay := getInt(&delay) // don't print updates while solving for best openings\n setInt(&delay, 0)\n\n for i := 0; i < width; i++ {\n for j := 0; j < width; j++ {\n start := 2*(i + 1)\n finish := 2*(j + 1)\n *x = start\n *y = finish\n if getMaze(getInt(&begX), start - 1) != wall && getMaze(getInt(&begX), start + 1) != wall {; continue; }\n if getMaze(getInt(&endX), finish - 1) != wall && getMaze(getInt(&endX), finish + 1) != wall {; continue; }\n createOpenings(x, y)\n solveMaze(x, y)\n if getInt(&pathLen) > bestPathLen ||\n (getInt(&pathLen) == bestPathLen &&\n getInt(&turnCnt) > bestTurnCnt) {\n bestStart = start\n bestFinish = finish\n bestTurnCnt = getInt(&turnCnt)\n bestPathLen = getInt(&pathLen)\n setInt(&solveLength, getInt(&pathLen))\n }\n restoreMaze()\n deleteOpenings()\n incInt(&numSolves)\n }\n }\n addInt(&sumsolveLength, getInt(&solveLength))\n if viewFlag {\n setInt(&delay, saveDelay) // only restore delay value if view solve flag is set\n }\n *x = bestStart\n *y = bestFinish\n createOpenings(x, y)\n}", "func (st *state) visit(x *declNode) {\n\tcheckCanonical(x)\n\tx.index = st.index\n\tx.lowlink = st.index\n\tst.index++\n\n\tst.stack = append(st.stack, x) // push\n\tassert(x.scc == 0, \"node revisited\")\n\tx.scc = -1\n\n\tfor y := range x.intRefs {\n\t\t// Loop invariant: x is canonical.\n\n\t\ty := y.find()\n\n\t\tif x == y {\n\t\t\tcontinue // nodes already coalesced\n\t\t}\n\n\t\tswitch {\n\t\tcase y.scc > 0:\n\t\t\t// y is already a collapsed SCC\n\n\t\tcase y.scc < 0:\n\t\t\t// y is on the stack, and thus in the current SCC.\n\t\t\tif y.index < x.lowlink {\n\t\t\t\tx.lowlink = y.index\n\t\t\t}\n\n\t\tdefault:\n\t\t\t// y is unvisited; visit it now.\n\t\t\tst.visit(y)\n\t\t\t// Note: x and y are now non-canonical.\n\n\t\t\tx = x.find()\n\n\t\t\tif y.lowlink < x.lowlink {\n\t\t\t\tx.lowlink = y.lowlink\n\t\t\t}\n\t\t}\n\t}\n\tcheckCanonical(x)\n\n\t// Is x the root of an SCC?\n\tif x.lowlink == x.index {\n\t\t// Coalesce all nodes in the SCC.\n\t\tfor {\n\t\t\t// Pop y from stack.\n\t\t\ti := len(st.stack) - 1\n\t\t\ty := st.stack[i]\n\t\t\tst.stack = st.stack[:i]\n\n\t\t\tcheckCanonical(x)\n\t\t\tcheckCanonical(y)\n\n\t\t\tif x == y {\n\t\t\t\tbreak // SCC is complete.\n\t\t\t}\n\t\t\tcoalesce(x, y)\n\t\t}\n\n\t\t// Accumulate union of extRefs over\n\t\t// internal edges (to other SCCs).\n\t\tfor y := range x.intRefs {\n\t\t\ty := y.find()\n\t\t\tif y == x {\n\t\t\t\tcontinue // already coalesced\n\t\t\t}\n\t\t\tassert(y.scc == 1, \"edge to non-scc node\")\n\t\t\tfor z := range y.extRefs {\n\t\t\t\tif x.extRefs == nil {\n\t\t\t\t\tx.extRefs = make(symbolSet)\n\t\t\t\t}\n\t\t\t\tx.extRefs[z] = true // extRefs: x U= y\n\t\t\t}\n\t\t}\n\n\t\tx.scc = 1\n\t}\n}", "func (solver *MapSolver)GetMCMF(s int, t int)(map[string]int, int, []int){\n\tresult := make(map[string]int) // there is how many srcNode-dstNode? the distance between srcNode and dstNode is 1\n\tcost := 0\n\tflow := make([]int, data.RESOURCEDIMENSION)\n\n\tcounter := 0\n\tfor true {\n\t\tcounter ++\n\t\tfmt.Println(counter)\n\t\tif isExsit, newCost, paths, arcs :=solver.GetSPFA(s, t); isExsit {\n\t\t\t// paths[0] must start from startNode and after 0, all path start from machine Node.\n\t\t\tvar newFlow []int\n\n\t\t\t// all paths ends at machineNode\n\t\t\t// first path start from applicationNode\n\t\t\tfor i, path := range paths{\n\t\t\t\tif(i == 0){\n\t\t\t\t\t// get the task of this path\n\t\t\t\t\tnode := data.NodeList[path[1]] // taskNode\n\t\t\t\t\ttaskNode, ok := node.(*data.TaskNode)\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\tpanic(\"NodeList[paths[0][2]] must be *TaskNode type\")\n\t\t\t\t\t}\n\t\t\t\t\ttask := taskNode.Task\n\t\t\t\t\t// get flow of this path\n\n\t\t\t\t\tpathFlow := data.NodeList[path[1]].GetLeftInArcs()[0].CopyCapacity() // the flow of this task.\n\n\t\t\t\t\tnewFlow = pathFlow\n\n\t\t\t\t\t// sub capacity of the arc on this path\n\t\t\t\t\tsolver.UpdateCapacityOfPath(arcs, i, pathFlow)\n\n\t\t\t\t\t// we need to update the arc's capacity. from this machine to end node, we need to sub it with pathFlow\n\t\t\t\t\tmachineNode, ok := data.NodeList[path[len(path)-1]].(*data.MachineNode)\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\tpanic(\"data.NodeList[path[len(path)-1]] must be *MachineNode type\")\n\t\t\t\t\t}\n\t\t\t\t\t// update capacity of arc from this machine node to end node\n\t\t\t\t\tsolver.UpdateCapacityOfMachine(machineNode, pathFlow, true)\n\n\t\t\t\t\t// update machine's ScheduledTasks map\n\t\t\t\t\tsolver.UpdateScheduledMachineTaskMap(task, machineNode, arcs, i, true)\n\t\t\t\t}else {\n\t\t\t\t\t// path[i] (i > 0) all start from machine node, end at another machine node. and the second node is the template node,\n\t\t\t\t\tstartMachineNode, ok := data.NodeList[path[0]].(*data.MachineNode)\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\tpanic(\"data.NodeList[paths[i][0]] must be *MachineNode type if i > 0\")\n\t\t\t\t\t}\n\t\t\t\t\tendMachineNode, ok := data.NodeList[path[len(path)-1]].(*data.MachineNode)\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\tpanic(\"data.NodeList[path[len(path)-1]] must be *MachineNode type\")\n\t\t\t\t\t}\n\n\t\t\t\t\ttemplateNode, ok := data.NodeList[path[1]].(*data.TemplateNode) // get template node\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\tpanic(\"data.NodeList[paths[i][1]] must be *TemplateNode type if i > 0\")\n\t\t\t\t\t}\n\n\t\t\t\t\t// get task\n\t\t\t\t\ttask := templateNode.SourceTasks[0]\n\n\t\t\t\t\t// get flow of this path\n\t\t\t\t\tpathFlow := task.GetCapacity()\n\n\t\t\t\t\t// update the start machine's capacity after regret\n\t\t\t\t\tsolver.UpdateCapacityOfMachine(startMachineNode, pathFlow, false)\n\n\t\t\t\t\t// update start machine's ScheduledTasks map\n\t\t\t\t\tsolver.UpdateScheduledMachineTaskMap(task, startMachineNode, arcs, i, false)\n\n\t\t\t\t\tsolver.UpdateCapacityOfPath(arcs, i, pathFlow)\n\n\t\t\t\t\t// we need to update the arc's capacity. from this machine to end node, we need to sub it with pathFlow\n\t\t\t\t\tsolver.UpdateCapacityOfMachine(endMachineNode, pathFlow, true)\n\n\t\t\t\t\t// update start machine's ScheduledTasks map\n\t\t\t\t\tsolver.UpdateScheduledMachineTaskMap(task, endMachineNode, arcs, i, true)\n\t\t\t\t}\n\t\t\t}\n\t\t\tflow = solver.AddFlow(flow, newFlow)\n\t\t\tcost += solver.GetCost(newCost, newFlow)\n\t\t\t// take apart of paths, add all of them to result\n\t\t\tfmt.Println(paths)\n\t\t\tfor _, path := range paths {\n\t\t\t\tvar preNodeIndexStr string\n\t\t\t\tfor i, nodeIndex := range path {\n\t\t\t\t\tif i == 0 {\n\t\t\t\t\t\tpreNodeIndexStr = strconv.Itoa(nodeIndex)\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}else {\n\t\t\t\t\t\tpostNodeIndexStr := strconv.Itoa(nodeIndex)\n\t\t\t\t\t\tkey := preNodeIndexStr+\"-\"+postNodeIndexStr\n\t\t\t\t\t\tif _, ok :=result[key]; ok {\n\t\t\t\t\t\t\tresult[key] ++\n\t\t\t\t\t\t}else{\n\t\t\t\t\t\t\tresult[key] = 1\n\t\t\t\t\t\t}\n\t\t\t\t\t\tpreNodeIndexStr = postNodeIndexStr\n\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\n\t\t}else {\n\t\t\tbreak\n\t\t}\n\t}\n\n\n\treturn result, cost, flow\n}", "func findPaths2(m int, n int, maxMove int, startRow int, startColumn int) int {\n\t// 값이 클 수 있어 modulo 모듈 값을 리턴하라고 문제에 써있음\n\tmodulo := 1000000000 + 7\n\tif startRow < 0 || startRow == m ||\n\t\tstartColumn < 0 || startColumn == n {\n\t\treturn 1\n\t}\n\tif maxMove == 0 {\n\t\treturn 0\n\t}\n\t// 4방향 각각에 대해서 boundary 를 벗어나는지 경우들을 모두 더한다.\n\treturn findPaths(m, n, maxMove-1, startRow-1, startColumn)%modulo +\n\t\tfindPaths(m, n, maxMove-1, startRow, startColumn-1)%modulo +\n\t\tfindPaths(m, n, maxMove-1, startRow+1, startColumn)%modulo +\n\t\tfindPaths(m, n, maxMove-1, startRow, startColumn+1)%modulo\n}", "func DepthFirstSearchUtil(forest []int, visited map[int]bool, count int) int {\n\tfor key, val := range forest {\n\t\tif !visited[key] {\n\t\t\tvisited[key] = true\n\t\t\treturn count + DepthFirstSearchUtil(forest, visited, val)\n\t\t}\n\t}\n}", "func dfs_minSideJumps(obstacles []int, l int, pos int, line int, memo [][3]int) int {\n\tif pos == l-1 {\n\t\treturn 0\n\t}\n\tif memo[pos][line-1] != 0 {\n\t\treturn memo[pos][line-1]\n\t}\n\tif obstacles[pos+1] == line {\n\t\tvar res int = 2147483647\n\t\tfor i := 1; i <= 3; i++ {\n\t\t\tif i != line && obstacles[pos] != i {\n\t\t\t\tsteps := dfs_minSideJumps(obstacles, l, pos+1, i, memo)\n\t\t\t\tif steps != 2147483647 {\n\t\t\t\t\tres = min_int(res, 1+steps)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tmemo[pos][line-1] = res\n\t} else {\n\t\tmemo[pos][line-1] = dfs_minSideJumps(obstacles, l, pos+1, line, memo)\n\t}\n\treturn memo[pos][line-1]\n}", "func treeSQR( sqrArray[] string, arrayPos int) int{\n\n//Function check to see if the there is any shapes in the in the listArray\n//if not it changes the structures as it moves to the end\nif listArray[0] != \"SQR\" { \n\t\t\tlistArray[1]= \"<inst>\"\n\t\t\tlistArray[0] = \"SQR\"\n\t\tarrayPos++\n\t\t\n\t\t// Called the function so it can be processed with the valid format \n\t\ttreeSQR(sqrArray[0:],arrayPos)\t\n\t}else{ if listArray[1] == \"\" || listArray[1] == \"<inst>\"{ // after transforming it is place in a format that can be parsed \n\t\n\t\t\tif sqrArray[arrayPos] == \"SQR\"{ // Ensure we are not Validating a Shape\n\t\t\t\tarrayPos++\n\t\t\t}\n\t\t\t\n\t\t\t// Retrieve the Coordinated from the array\n\t\t\t// Proceeding to the next value\n\t\t\tvar curCoord string=sqrArray[arrayPos]\n\t\t\tarrayPos++\n\t\t\tvar secCoord string=sqrArray[arrayPos]\n\t\t\t//arrayPos++\t\t\n\t\t\t\n\t\t\t// Using Slices we get each Values \n\t\t\tvar test string \n\t\t\tx:=curCoord[0:1]\n\t\t\ty:=curCoord[1:2]\n\t\t\tyy:=secCoord[0:1]\n\t\t\txx:=secCoord[1:2]\n\t\t\tarrayPos=arrayPos-2\n\t\t\t\n\t\t\t//The Printing format for the lower part of the tree\n\t\t\tfmt.Printf(\"\\n |\\n\")\n\t\t\tfmt.Printf(\"SQR\\n/ \\\\\\n\")\n\t\t\tfmt.Printf(\"<coord>,<coord>\\n\")\n\t\t\tfmt.Printf(\" / \\\\ / \\\\\\n\")\n\t\t\tfmt.Printf(\"<x><y>,<x><y>\\n\"+x+\" \"+y+\" \"+yy+\" \"+xx)\n\t\t\n\t\t\tlistArray[0] = \"<inst>\"\n\t\t\tlistArray[1] = \"<inst_list>\"\n\t\t\t\t\t\t\n\t\t\ttempCount=tempCount-1\n\t\t\t\tif(tempCount >= 0){\t\t\t\t\n\t\t\t\t\tlistArray[tempCount]=\"\"\n\t\t\t\t\t\n\t\t\t\t}\n\t\t\treturn arrayPos\t\n\t\t}\n\t\treturn arrayPos\n\t\t }\n\t\n\treturn arrayPos\n}", "func bfsPush(root *node, indexes [][]int32) {\n\tif len(indexes) == 0 || root == nil {\n\t\treturn\n\t}\n\tqueue := []*node{root}\n\tfor _, curNodeValues := range indexes {\n\t\tleftValue, rightValue := curNodeValues[0], curNodeValues[1]\n\t\tif leftValue != -1 {\n\t\t\troot.left = &node{leftValue, nil, nil}\n\t\t\tqueue = append(queue, root.left)\n\t\t}\n\t\tif rightValue != -1 {\n\t\t\troot.right = &node{rightValue, nil, nil}\n\t\t\tqueue = append(queue, root.right)\n\t\t}\n\t\tqueue = queue[1:]\n\t\tif len(queue) == 0 {\n\t\t\treturn\n\t\t}\n\t\troot = queue[0]\n\t}\n}", "func find(gs []*Geodesic, v Vector, minDistSq float64, start int) int {\n\tgs0 := gs[0]\n\tnextStart := start\n\tneighbors := gs0.Faces[start].Neighbors\n\tfor _, n := range neighbors {\n\t\tiDistSq := DistSq(gs0.Centers[n], v)\n\t\tif iDistSq < minDistSq {\n\t\t\tminDistSq = iDistSq\n\t\t\tnextStart = n\n\t\t}\n\t}\n\n\tif len(gs) == 1 {\n\t\t// There's a bug related to correctly classifying the neighbor of a\n\t\t// pentagon face, so this corrects for that.\n\t\tif nextStart == start {\n\t\t\treturn nextStart\n\t\t}\n\t\treturn find(gs, v, minDistSq, nextStart)\n\t}\n\n\treturn find(gs[1:], v, minDistSq, nextStart)\n}", "func connectedComponents(adj_list [][]int) {\n\n\tcomponents := make(map[int]int)\n\tvisited := make([]bool, V+1)\n\n\tcountConnSize := 0\n\tnumComponents := 0\n\n\tcomp.components = make([][]int, V, V) // I have hard coded 3 as the number of components. As I am not able to determine it dynamically\n\n\tfor i := range comp.components {\n\t\tcomp.components[i] = make([]int, 0, V)\n\t}\n\n\tfor v := 1; v <= V; v++ {\n\n\t\tif visited[v] == false {\n\t\t\tdfSUtil(v, visited, adj_list, components)\n\n\t\t\tif len(components) == V {\n\n\t\t\t\tfor x := range components {\n\t\t\t\t\tcomp.components[0] = append(comp.components[0], x)\n\t\t\t\t}\n\t\t\t\tcomp.connected = true\n\t\t\t\tnumComponents = 1\n\n\t\t\t} else {\n\n\t\t\t\tnumComponents++\n\n\t\t\t\tfor x := range components {\n\t\t\t\t\t//fmt.Printf(\"%d \",x)\n\t\t\t\t\tcomp.components[numComponents-1] = append(comp.components[numComponents-1], x)\n\t\t\t\t}\n\t\t\t\tcountConnSize++\n\t\t\t\tcomp.connected = false\n\t\t\t}\n\n\t\t\tfor x := range components {\n\t\t\t\tdelete(components, x)\n\t\t\t}\n\t\t}\n\t\tcomp.numberOfComponents = numComponents\n\t}\n}", "func Day8Part1(filepath string) any {\n\tvar res int\n\n\t// open file\n\treadFile, _ := os.Open(filepath)\n\n\t// read line\n\tfileScanner := bufio.NewScanner(readFile)\n\tfileScanner.Split(bufio.ScanLines)\n\n\t// parse map in to [][]int\n\ttreeMap := parseMap(fileScanner)\n\t// fmt.Println(treeMap)\n\n\t// init visible trees 2D array\n\tvisible := make([][]bool, len(treeMap))\n\tfor i := range visible {\n\t\tvisible[i] = make([]bool, len(treeMap[0]))\n\t}\n\n\t// look from left to r\n\tfor i := range treeMap {\n\t\tmax := 0\n\t\tfor j := range treeMap[0] {\n\t\t\tif i == 0 || i == len(treeMap)-1 || j == 0 || j == len(treeMap[0])-1 || treeMap[i][j] > max {\n\t\t\t\tvisible[i][j] = true\n\t\t\t\tmax = treeMap[i][j]\n\t\t\t}\n\t\t}\n\t}\n\n\t// look from right to l\n\tfor i := range treeMap {\n\t\tmax := 0\n\t\tfor j := len(treeMap[0]) - 1; j >= 0; j-- {\n\t\t\tif i == 0 || i == len(treeMap)-1 || j == 0 || j == len(treeMap[0])-1 || treeMap[i][j] > max {\n\t\t\t\tvisible[i][j] = true\n\t\t\t\tmax = treeMap[i][j]\n\t\t\t}\n\t\t}\n\t}\n\n\t// look from up to down\n\tfor j := 0; j <= len(treeMap[0])-1; j++ {\n\t\tmax := 0\n\t\tfor i := 0; i <= len(treeMap)-1; i++ {\n\t\t\tif i == 0 || i == len(treeMap)-1 || j == 0 || j == len(treeMap[0])-1 || treeMap[i][j] > max {\n\t\t\t\tvisible[i][j] = true\n\t\t\t\tmax = treeMap[i][j]\n\t\t\t}\n\t\t}\n\t}\n\n\t// look from down to up\n\tfor j := 0; j <= len(treeMap[0])-1; j++ {\n\t\tmax := 0\n\t\tfor i := len(treeMap) - 1; i >= 0; i-- {\n\t\t\tif i == 0 || i == len(treeMap)-1 || j == 0 || j == len(treeMap[0])-1 || treeMap[i][j] > max {\n\t\t\t\tvisible[i][j] = true\n\t\t\t\tmax = treeMap[i][j]\n\t\t\t}\n\t\t}\n\t}\n\n\t// traverse visible trees 2D array and count visibles\n\tfor i := range visible {\n\t\tfor j := range visible[i] {\n\t\t\tif visible[i][j] {\n\t\t\t\tres++\n\t\t\t}\n\t\t}\n\t}\n\n\treturn res\n}", "func nearestNeighbor(r, g, b float64, key int, node *Tree) (smallestDistance float64, URLIndex int) {\n\t//base case\n\tif node == nil {\n\t\tsmallestDistance, URLIndex = -1, -1\n\t\treturn //smallestDistance and URLIndex cannot be negative, check for -1 in recursive step\n\t} else {\n\t\t//smallest distance is set to value from recursive call, this value is the squared euclidean distance\n\t\t//take the square root if you want the actual distance\n\n\t\tvar isLeftOfPlane bool //used to tell if point r, g ,b is left of the plane created by the current node\n\t\tnextKey := (key + 1) % 3\n\t\tswitch key {\n\t\tcase 0:\n\t\t\tif r < (*node).R {\n\t\t\t\tsmallestDistance, URLIndex = nearestNeighbor(r, g, b, nextKey, (*node).Left)\n\t\t\t\tisLeftOfPlane = true\n\t\t\t} else {\n\t\t\t\tsmallestDistance, URLIndex = nearestNeighbor(r, g, b, nextKey, (*node).Right)\n\t\t\t\tisLeftOfPlane = false\n\t\t\t}\n\t\tcase 1:\n\t\t\tif g < (*node).G {\n\t\t\t\tsmallestDistance, URLIndex = nearestNeighbor(r, g, b, nextKey, (*node).Left)\n\t\t\t\tisLeftOfPlane = true\n\t\t\t} else {\n\t\t\t\tsmallestDistance, URLIndex = nearestNeighbor(r, g, b, nextKey, (*node).Right)\n\t\t\t\tisLeftOfPlane = false\n\t\t\t}\n\t\tcase 2:\n\t\t\tif b < (*node).B {\n\t\t\t\tsmallestDistance, URLIndex = nearestNeighbor(r, g, b, nextKey, (*node).Left)\n\t\t\t\tisLeftOfPlane = true\n\t\t\t} else {\n\t\t\t\tsmallestDistance, URLIndex = nearestNeighbor(r, g, b, nextKey, (*node).Right)\n\t\t\t\tisLeftOfPlane = false\n\t\t\t}\n\t\t}\n\n\t\tdistanceToThisNode := math.Pow(r-(*node).R, 2) + math.Pow(g-(*node).G, 2) + math.Pow(b-(*node).B, 2)\n\n\t\t//check if value in smallest distance is -1, if it is then we\n\t\t//set smallest distance to distance from the current node\n\t\tif smallestDistance == -1 {\n\t\t\tsmallestDistance, URLIndex = distanceToThisNode, (*node).URLIndex\n\t\t} else if smallestDistance > distanceToThisNode {\n\t\t\t//else if smallest distance is bigger than distance from current node set to distance from this node\n\t\t\tsmallestDistance, URLIndex = distanceToThisNode, (*node).URLIndex\n\t\t}\n\n\t\t//calculate if the sphere with center coordinates r, g, b and radius equal to smallestDistance intersects with\n\t\t//the xy-plane, xz-plane, or yz-plane(depending on the value of key) that contains the point R, G, B\n\t\t//from the current node\n\t\t//if it does calculate the nearest neighbor on the other side of the plane and save it to potential smallest\n\t\tvar potentialSmallest float64\n\t\tvar potentialURL int\n\t\tswitch key {\n\t\tcase 0:\n\t\t\t//yz plane\n\t\t\tif (isLeftOfPlane) && (r+math.Sqrt(smallestDistance) > node.R) {\n\t\t\t\t//sphere intersects\n\t\t\t\tpotentialSmallest, potentialURL = nearestNeighbor(r, g, b, nextKey, (*node).Right)\n\t\t\t} else if (!isLeftOfPlane) && (r-math.Sqrt(smallestDistance) < node.R) {\n\t\t\t\t//sphere intersects\n\t\t\t\tpotentialSmallest, potentialURL = nearestNeighbor(r, g, b, nextKey, (*node).Left)\n\t\t\t} else {\n\t\t\t\tpotentialSmallest = -1\n\t\t\t}\n\t\tcase 1:\n\t\t\t//xz plane\n\t\t\tif (isLeftOfPlane) && (g+math.Sqrt(smallestDistance) > node.G) {\n\t\t\t\t//sphere intersects\n\t\t\t\tpotentialSmallest, potentialURL = nearestNeighbor(r, g, b, nextKey, (*node).Right)\n\t\t\t} else if (!isLeftOfPlane) && (g-math.Sqrt(smallestDistance) < node.G) {\n\t\t\t\t//sphere intersects\n\t\t\t\tpotentialSmallest, potentialURL = nearestNeighbor(r, g, b, nextKey, (*node).Left)\n\t\t\t} else {\n\t\t\t\t//sphere doesnt intersect\n\t\t\t\tpotentialSmallest = -1\n\t\t\t}\n\t\tcase 2:\n\t\t\t//xy plane\n\t\t\tif (isLeftOfPlane) && (b+math.Sqrt(smallestDistance) > node.B) {\n\t\t\t\t//sphere intersects\n\t\t\t\tpotentialSmallest, potentialURL = nearestNeighbor(r, g, b, nextKey, (*node).Right)\n\t\t\t} else if (!isLeftOfPlane) && (b-math.Sqrt(smallestDistance) < node.B) {\n\t\t\t\t//sphere intersects\n\t\t\t\tpotentialSmallest, potentialURL = nearestNeighbor(r, g, b, nextKey, (*node).Left)\n\t\t\t} else {\n\t\t\t\t//sphere doesnt intersect\n\t\t\t\tpotentialSmallest = -1\n\t\t\t}\n\t\t}\n\n\t\t//check if potential smallest is valid, if it is then check it is smaller than smallestDistance\n\t\tif potentialSmallest != -1 && potentialSmallest < smallestDistance {\n\t\t\tsmallestDistance, URLIndex = potentialSmallest, potentialURL\n\t\t}\n\t\treturn\n\t}\n}", "func followPath(x, y *int) bool {\n directions := make([]dirTable, 4, 4)\n lastDir := 0\n length := -1\n setCell(*x, *y, solved, noUpdate, 0, 0)\n for getInt(&begX) <= *x && *x <= getInt(&endX) {\n num := findDirections(*x, *y, &length, path, directions)\n if num == 0 {\n break\n }\n followDir(x, y, directions[0], lastDir)\n if threads > 1 && num > 1 && getBool(&solvedFlag) == false {\n for i := 1; i < num; i++ {\n incInt(&numThreads)\n followDir(x, y, directions[i], lastDir)\n go solve(*x + directions[i].x, *y + directions[i].y)\n }\n }\n *x += directions[0].x\n *y += directions[0].y\n }\n if *x > getInt(&endX) {\n setBool(&solvedFlag, true)\n return true\n } else {\n return false\n }\n}", "func finder(result, n []Node, p Node, id int64) []Node {\n\tif len(result) > 0 {\n\t\treturn result\n\t}\n\tfor _, node := range n {\n\t\tif node.ID == id {\n\t\t\tresult = append(result, node)\n\t\t\treturn result\n\t\t}\n\t\tif len(node.Nodes) > 0 {\n\t\t\tfinder(result, node.Nodes, node, id)\n\t\t}\n\t}\n\treturn n\n}", "func makeTree(todo, done []Node, out chan bool) {\r\n\tfor i, p := range todo {\r\n\t\tvar minDist float64 = 1000\r\n\t\tvar minNode int\r\n\t\tfor j := 0; j < len(done)-len(todo)+i; j++ {\r\n\t\t\ttmp := dist(p.data, done[j].data)\r\n\t\t\tif tmp < minDist {\r\n\t\t\t\tminDist = tmp\r\n\t\t\t\tminNode = j\r\n\t\t\t}\r\n\t\t}\r\n\t\tdone[minNode].children = append(done[minNode].children, &todo[i])\r\n\r\n\t}\r\n\tout <- true\r\n}", "func testDFS() {\n\tg := NewGraph(4)\n\tg.AddEdge(0, 1)\n\tg.AddEdge(0, 2)\n\tg.AddEdge(1, 2)\n\tg.AddEdge(2, 0)\n\tg.AddEdge(2, 3)\n\tg.AddEdge(3, 3)\n\tlog.Printf(\"Graph %v\", g.String())\n\tg.RecursiveDFS(0)\n}", "func (gph *Graph) DFSStack(source int, target int) bool {\n\tcount := gph.count\n\tvisited := make([]bool, count)\n\tvar curr int\n\tstk := stack.New()\n\tpath := []int{}\n\tvisited[source] = true\n\tstk.Push(source)\n\n\tfor stk.Len() != 0 {\n\t\tcurr = stk.Pop().(int)\n\t\tpath = append(path,curr)\n\t\thead := gph.Edges[curr]\n\t\tfor head != nil {\n\t\t\tif visited[head.destination] == false {\n\t\t\t\tvisited[head.destination] = true\n\t\t\t\tstk.Push(head.destination)\n\t\t\t}\n\t\t\thead = head.next\n\t\t}\n\t}\n\tfmt.Println(\"DFS Path is : \", path)\n return visited[target]\n}", "func canFinish(numCourses int, prerequisites [][]int) bool {\n\t// 20200424\n\t// BFS\n\t// 执行耗时:20 ms,击败了45.76% 的Go用户\n\t// 内存消耗:5.7 MB,击败了100.00% 的Go用户\n\t//indegrees := make([]int, numCourses) // 记录每门课程前置应修的课程数\n\t//for _, req := range prerequisites {\n\t//\tindegrees[req[0]]++\n\t//}\n\t//queue := list.New() // 装入度为0的课程,即没有依赖可直接修的课程\n\t//for i := 0; i < numCourses; i++ {\n\t//\tif indegrees[i] == 0 {\n\t//\t\tqueue.PushBack(i)\n\t//\t}\n\t//}\n\t//for queue.Len() > 0 {\n\t//\tcourse := queue.Remove(queue.Front()).(int)\n\t//\tnumCourses-- // 修course这门课\n\t//\tfor _, req := range prerequisites {\n\t//\t\tif req[1] != course {\n\t//\t\t\tcontinue\n\t//\t\t}\n\t//\t\t// course修过了,依赖course的课程也可以修了\n\t//\t\tindegrees[req[0]]--\n\t//\t\t// req[0]这门课的前置课程都修完了\n\t//\t\tif indegrees[req[0]] == 0 {\n\t//\t\t\tqueue.PushBack(req[0])\n\t//\t\t}\n\t//\t}\n\t//}\n\t//return numCourses == 0\n\n\t// 20200424\n\t// BFS-优化\n\t// 执行耗时:12 ms,击败了93.22% 的Go用户\n\t// 内存消耗:5.8 MB,击败了100.00% 的Go用户\n\t//// 记录每门课程前置应修的课程数\n\t//indegrees := make([]int, numCourses)\n\t//// 记录每门课的后修课程\n\t//nexts := make([][]int, numCourses)\n\t//for _, req := range prerequisites {\n\t//\tindegrees[req[0]] += 1\n\t//\tnexts[req[1]] = append(nexts[req[1]], req[0])\n\t//}\n\t//queue := list.New()\n\t//// 装入度为0的课程,即没有依赖可直接修的课程\n\t//for i := 0; i < numCourses; i++ {\n\t//\tif indegrees[i] == 0 {\n\t//\t\tqueue.PushBack(i)\n\t//\t}\n\t//}\n\t//for queue.Len() > 0 {\n\t//\tcourse := queue.Remove(queue.Front()).(int)\n\t//\tnumCourses-- // 修course这门课\n\t//\tfor _, next := range nexts[course] {\n\t//\t\t// course修过了,依赖course的课程也可以修了\n\t//\t\tindegrees[next]--\n\t//\t\tif indegrees[next] == 0 { // 前置课程都修完了\n\t//\t\t\tqueue.PushBack(next)\n\t//\t\t}\n\t//\t}\n\t//}\n\t//return numCourses == 0\n\n\t// 20200424\n\t// DFS\n\t// 执行耗时:20 ms,击败了45.76% 的Go用户\n\t// 内存消耗:5.7 MB,击败了100.00% 的Go用户\n\t//flags := make([]int, numCourses)\n\t//var dfs func(course int) bool\n\t//dfs = func(course int) bool {\n\t//\tif flags[course] == 1 {\n\t//\t\treturn false\n\t//\t}\n\t//\tif flags[course] == -1 {\n\t//\t\treturn true\n\t//\t}\n\t//\tflags[course] = 1\n\t//\tfor _, req := range prerequisites {\n\t//\t\tif req[0] == course && !dfs(req[1]) {\n\t//\t\t\treturn false\n\t//\t\t}\n\t//\t\t/* 以下写法也是对的;两种写法都是对course邻居做判断\n\t//\t\tif req[1] == course && !dfs(req[0], prerequisites, flags) {\n\t//\t\t\treturn false\n\t//\t\t}*/\n\t//\t}\n\t//\tflags[course] = -1\n\t//\treturn true\n\t//}\n\t//for i := 0; i < numCourses; i++ {\n\t//\tif !dfs(i) {\n\t//\t\treturn false\n\t//\t}\n\t//}\n\t//return true\n\n\t// 20200424\n\t// DFS-优化\n\t// 执行耗时:12 ms,击败了93.22% 的Go用户\n\t// 内存消耗:6.1 MB,击败了100.00% 的Go用户\n\tneighbors := make([][]int, numCourses)\n\tfor _, req := range prerequisites {\n\t\t// 写成 neighbors[req[1]] = append(neighbors[req[1]], req[0]) 也对,都是统计邻居\n\t\tneighbors[req[0]] = append(neighbors[req[0]], req[1])\n\t}\n\tflags := make([]int, numCourses)\n\tvar dfs func(course int) bool\n\tdfs = func(course int) bool {\n\t\tif flags[course] == 1 {\n\t\t\treturn false\n\t\t}\n\t\tif flags[course] == -1 {\n\t\t\treturn true\n\t\t}\n\t\tflags[course] = 1\n\t\tfor _, neighbor := range neighbors[course] {\n\t\t\tif !dfs(neighbor) {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t\tflags[course] = -1\n\t\treturn true\n\t}\n\tfor i := 0; i < numCourses; i++ {\n\t\tif !dfs(i) {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func findPaths2(m int, n int, maxMove int, startRow int, startColumn int) int {\n\tprevGrid := make([][]int, m)\n\tnextGrid := make([][]int, m)\n\tfor i := 0; i < m; i++ {\n\t\tprevGrid[i] = make([]int, n)\n\t\tnextGrid[i] = make([]int, n)\n\t}\n\n\tres := 0\n\tprevGrid[startRow][startColumn] = 1\n\n\tfor move := 0; move < maxMove; move++ {\n\t\tfor x := 0; x < m; x++ {\n\t\t\tfor y := 0; y < n; y++ {\n\t\t\t\tv := prevGrid[x][y] % modulo\n\t\t\t\tif v == 0 {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tprevGrid[x][y] = 0\n\n\t\t\t\tif x == 0 {\n\t\t\t\t\tres += v\n\t\t\t\t} else {\n\t\t\t\t\tnextGrid[x-1][y] += v\n\t\t\t\t}\n\n\t\t\t\tif x == m-1 {\n\t\t\t\t\tres += v\n\t\t\t\t} else {\n\t\t\t\t\tnextGrid[x+1][y] += v\n\t\t\t\t}\n\n\t\t\t\tif y == 0 {\n\t\t\t\t\tres += v\n\t\t\t\t} else {\n\t\t\t\t\tnextGrid[x][y-1] += v\n\t\t\t\t}\n\n\t\t\t\tif y == n-1 {\n\t\t\t\t\tres += v\n\t\t\t\t} else {\n\t\t\t\t\tnextGrid[x][y+1] += v\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tprevGrid, nextGrid = nextGrid, prevGrid\n\t}\n\n\treturn res % modulo\n}", "func SPBF(g *gs.Graph, src, dst *gs.Vertex) (string, bool) {\n\tsrc.StampD = 0\n\n\t// for each vertex u ∈ g.V\n\tvertices := g.GetVertices()\n\tfor _, vtx := range *vertices {\n\t\tif vtx == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\t// relax\n\t\tedges := g.GetEdges()\n\t\tfor _, edge := range *edges {\n\t\t\tif edge == nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\ta := edge.(*gs.Edge).Dst.StampD\n\t\t\tb := edge.(*gs.Edge).Src.StampD\n\t\t\tc := edge.(*gs.Edge).Weight\n\n\t\t\tif a > b+int64(c) {\n\t\t\t\tedge.(*gs.Edge).Dst.StampD = b + int64(c)\n\t\t\t}\n\n\t\t\t// Update Prev\n\t\t\tif edge.(*gs.Edge).Dst.Prev.Len() == 0 {\n\t\t\t\tedge.(*gs.Edge).Dst.Prev.PushBack(edge.(*gs.Edge).Src)\n\t\t\t} else {\n\t\t\t\tex := false\n\t\t\t\tpvs := edge.(*gs.Edge).Dst.Prev\n\t\t\t\tfor _, v := range *pvs {\n\t\t\t\t\tif v == nil {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\t// if fmt.Sprintf(\"%v\", v.(*gs.Vertex).ID) == fmt.Sprintf(\"%v\", edge.(*gs.Edge).Src.ID) {\n\t\t\t\t\tif v.(*gs.Vertex) == edge.(*gs.Edge).Src {\n\t\t\t\t\t\tex = true\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tif ex == false {\n\t\t\t\t\tedge.(*gs.Edge).Dst.Prev.PushBack(edge.(*gs.Edge).Src)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\tedges := g.GetEdges()\n\tfor _, edge := range *edges {\n\t\tif edge == nil {\n\t\t\tcontinue\n\t\t}\n\t\ta := edge.(*gs.Edge).Dst.StampD\n\t\tb := edge.(*gs.Edge).Src.StampD\n\t\tc := edge.(*gs.Edge).Weight\n\n\t\tif a > b+int64(c) {\n\t\t\treturn \"There is negative weighted cycle (No Shortest Path)\", false\n\t\t}\n\t}\n\tresult := slice.NewSequence()\n\tTrackSPBF(g, src, dst, dst, result)\n\n\tvar rs string\n\tfor _, v := range *result {\n\t\tif v == nil {\n\t\t\tcontinue\n\t\t}\n\t\trs += fmt.Sprintf(\"%v(=%v) → \", v.(*gs.Vertex).ID, v.(*gs.Vertex).StampD)\n\t}\n\n\treturn rs[:len(rs)-5], true\n}", "func backwardDiffAnalysis(wrt, sortedNodes Nodes) (retVal NodeSet, err error) {\n\tsymdiffLogf(\"Backwards analysis\")\n\tenterLogScope()\n\tdefer leaveLogScope()\n\n\tif !wrt.AllSameGraph() {\n\t\treturn nil, errors.New(\"The supplied output Nodes are not the same graph\")\n\t}\n\n\tdiffSet := wrt.mapSet()\n\tsymdiffLogf(\"wrt:%d diffset: %d\", len(wrt), len(diffSet))\n\tsymdiffLogf(\"%v\", diffSet)\n\tsymdiffLogf(\"sorted: %d\", sortedNodes)\n\n\tenterLogScope()\n\tfor i := len(sortedNodes) - 1; i >= 0; i-- {\n\t\tn := sortedNodes[i]\n\t\tsymdiffLogf(\"working on %v. Has %d children\", n, len(n.children))\n\n\t\tvar op SDOp\n\t\tvar ok bool\n\t\tvar diffs []bool\n\t\tif op, ok = n.op.(SDOp); ok {\n\t\t\tdiffs = op.DiffWRT(len(n.children))\n\t\t}\n\n\t\tsymdiffLogf(\"differentiable WRT: %v\", diffs)\n\t\tenterLogScope()\n\t\tsymdiffLogf(\"Children: %v\", n.children)\n\t\tif len(diffs) == 0 {\n\t\t\t// check if this makes nodes unreachable. If it does, then error out\n\t\t\tif n.isStmt {\n\t\t\t\tsymdiffLogf(\"Statement nodes are Non differentiable!\")\n\t\t\t\tleaveLogScope()\n\t\t\t\tcontinue\n\t\t\t} else if n.isInput() {\n\t\t\t\tsymdiffLogf(\"Input nodes are Non differentiable\")\n\t\t\t\tleaveLogScope()\n\t\t\t\tcontinue\n\t\t\t} else if len(n.children) == 0 {\n\t\t\t\tsymdiffLogf(\"Leaf nodes have no children\")\n\t\t\t\tleaveLogScope()\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tg := n.g\n\t\t\tfor _, child := range n.children {\n\t\t\t\tparents := graph.NodesOf(g.To(child.ID()))\n\t\t\t\tif len(parents) == 1 && len(child.children) > 0 {\n\t\t\t\t\tleaveLogScope()\n\t\t\t\t\treturn nil, errors.Errorf(\"Being unable to differentiate %v would leave a portion of the graph unreachable. Unable to continue\", n)\n\t\t\t\t}\n\t\t\t}\n\t\t\tsymdiffLogf(\"SKIPPING... Non differentiable!\")\n\t\t\tleaveLogScope()\n\t\t\tcontinue\n\t\t}\n\n\tinner:\n\t\tfor j, child := range n.children {\n\t\t\td := diffs[j]\n\t\t\tif diffSet.Contains(child) && d {\n\t\t\t\tsymdiffLogf(\"Adding %x to differentiable set\", child.ID())\n\t\t\t\tdiffSet.Add(n)\n\t\t\t\tbreak inner\n\t\t\t}\n\t\t}\n\t\tleaveLogScope()\n\t}\n\tleaveLogScope()\n\treturn diffSet, nil\n}" ]
[ "0.66430527", "0.65681833", "0.64927596", "0.6254573", "0.6130751", "0.6112519", "0.60927993", "0.598578", "0.59538746", "0.592007", "0.5894179", "0.58892846", "0.5865991", "0.584865", "0.582699", "0.58195686", "0.58023083", "0.5763093", "0.5759495", "0.5752892", "0.57520384", "0.5705026", "0.5689605", "0.5687628", "0.56869984", "0.56851596", "0.565474", "0.5650283", "0.56456876", "0.5630327", "0.5601719", "0.55998045", "0.55846655", "0.55817854", "0.55731016", "0.5565276", "0.556314", "0.55442077", "0.5535402", "0.55311877", "0.5527629", "0.5515576", "0.5506539", "0.5505252", "0.5497821", "0.54968685", "0.54943514", "0.54915434", "0.548952", "0.54884803", "0.5483641", "0.54736936", "0.5466382", "0.5461871", "0.5456825", "0.5445304", "0.544192", "0.54402506", "0.54350334", "0.5434543", "0.54166174", "0.5409092", "0.54076326", "0.53883654", "0.5387515", "0.5385612", "0.5364268", "0.5359906", "0.53596735", "0.5357349", "0.5350829", "0.53493863", "0.53472763", "0.5337414", "0.53303427", "0.53268945", "0.5320586", "0.531624", "0.530922", "0.5307778", "0.5305624", "0.5287264", "0.5284919", "0.52791923", "0.5272863", "0.52717996", "0.5268847", "0.5268512", "0.5265996", "0.5264726", "0.5264714", "0.52609503", "0.5255909", "0.5253737", "0.5253065", "0.52496487", "0.5240801", "0.52401567", "0.5236123", "0.5228175", "0.5224424" ]
0.0
-1
NewFileInfo returns a new FileInfo for testing.
func NewFileInfo( t *testing.T, path string, externalPath string, isImport bool, moduleIdentity bufmoduleref.ModuleIdentity, commit string, ) bufmoduleref.FileInfo { fileInfo, err := bufmoduleref.NewFileInfo( path, filepath.Clean(filepath.FromSlash(externalPath)), isImport, moduleIdentity, commit, ) require.NoError(t, err) return fileInfo }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewFileInfo(fPath string, info fs.FileInfo) FileInfo {\n\treturn &fileInfo{\n\t\tfullPath: fPath,\n\t\tFileInfo: info,\n\t}\n}", "func NewFileInfo(name string, size int64, mode os.FileMode, modTime, accessTime time.Time) *FileInfo {\n\treturn &FileInfo{\n\t\tname: filepath.Base(name),\n\t\tsize: size,\n\t\tmode: mode & (os.ModePerm | os.ModeDir),\n\t\tmodTime: modTime,\n\t\taccessTime: accessTime,\n\t}\n}", "func NewMockFileInfo(name string, isDir bool) MockFileInfo {\n\treturn MockFileInfo{\n\t\tname: name,\n\t\tdir: isDir,\n\t}\n}", "func NewFmFileInfo(provePrivKey []byte) *fmFileInfo {\n\tblks := make([]*fmBlockInfo, 0)\n\treturn &fmFileInfo{\n\t\tBlocks: blks,\n\t\tProvePrivKey: provePrivKey,\n\t}\n}", "func newDirFileInfo(name string) os.FileInfo {\n\treturn &bindataFileInfo{\n\t\tname: name,\n\t\tsize: 0,\n\t\tmode: os.FileMode(2147484068), // equal os.FileMode(0644)|os.ModeDir\n\t\tmodTime: time.Time{}}\n}", "func NewMockFileInfo(ctrl *gomock.Controller) *MockFileInfo {\n\tmock := &MockFileInfo{ctrl: ctrl}\n\tmock.recorder = &MockFileInfoMockRecorder{mock}\n\treturn mock\n}", "func FromFileInfo(dir string, i os.FileInfo) filesystem.File {\n\n\tpath := utils.LinuxPath(fmt.Sprintf(\"%s/%s\", dir, i.Name()))\n\tfile := filesystem.File{\n\t\tFileName: i.Name(),\n\t\tFilePath: path,\n\t\tFileMode: i.Mode(),\n\t\tFileSize: i.Size(),\n\t\tFileModTime: i.ModTime(),\n\t}\n\treturn file\n}", "func wrapFileInfo(t testing.TB, fi os.FileInfo) os.FileInfo {\n\t// get the underlying stat_t and modify the values\n\tstat := fi.Sys().(*syscall.Stat_t)\n\tstat.Mode = mockFileInfoMode\n\tstat.Uid = mockFileInfoUID\n\tstat.Gid = mockFileInfoGID\n\n\t// wrap the os.FileInfo so we can return a modified stat_t\n\tres := wrappedFileInfo{\n\t\tFileInfo: fi,\n\t\tsys: stat,\n\t\tmode: mockFileInfoMode,\n\t}\n\n\treturn res\n}", "func (i Info) FileInfo() os.FileInfo {\n\treturn &FileInfo{i}\n}", "func NewInfo(createResponse string, modified *time.Time) (os.FileInfo, error) {\n\n\telements := strings.SplitN(createResponse, \" \", 3)\n\tif len(elements) != 3 {\n\t\treturn nil, fmt.Errorf(\"invalid download createResponse: %v\", createResponse)\n\t}\n\tisDir := strings.HasPrefix(elements[0], \"D\")\n\tmodeLiteral := string(elements[0][1:])\n\tmode, err := strconv.ParseInt(modeLiteral, 8, 64)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"invalid mode: %v\", modeLiteral)\n\t}\n\tsizeLiteral := elements[1]\n\tsize, err := strconv.ParseInt(sizeLiteral, 10, 64)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"invalid size: %v\", modeLiteral)\n\t}\n\tname := strings.Trim(elements[2], \"\\r\\n\")\n\tif modified == nil {\n\t\tnow := time.Now()\n\t\tmodified = &now\n\t}\n\treturn file.NewInfo(name, size, os.FileMode(mode), *modified, isDir), nil\n}", "func parseFileInfo(info *FileInfo) tree.FileInfo {\n\tif info == nil {\n\t\treturn nil\n\t}\n\tfileType := tree.TYPE_UNKNOWN\n\tif info.Type != nil {\n\t\tfileType = tree.Type(*info.Type)\n\t}\n\tmode := tree.Mode(info.GetMode())\n\thasMode := tree.Mode(info.GetHasMode())\n\tmodTime := time.Time{}\n\tif info.ModTime != nil {\n\t\tmodTime = time.Unix(0, *info.ModTime)\n\t}\n\tsize := info.GetSize()\n\tid := info.GetInode()\n\treturn tree.NewFileInfo(info.Path, fileType, mode, hasMode, modTime, size, id, tree.Hash{tree.HashType(info.GetHashType()), info.HashData})\n}", "func NewInfo(version string, t time.Time) File {\n\tvar info = struct {\n\t\tVersion string\n\t\tTime time.Time\n\t}{version, t}\n\tjs, err := json.Marshal(info)\n\tif err != nil {\n\t\t// json.Marshal only fails for bad types; there are no bad types in info.\n\t\tpanic(\"unexpected json.Marshal failure\")\n\t}\n\treturn MemFile(js, t)\n}", "func newFile(name, contents string) (*file, error) {\n\tif name == \"\" {\n\t\treturn nil, fmt.Errorf(\"files cannot have empty names\")\n\t}\n\n\treturn &file{\n\t\tname: name,\n\t\tcontents: contents,\n\t}, nil\n}", "func newFileInfoGroup(dir string) (*fileInfoGroup, error) {\n\tfsi, err := ioutil.ReadDir(dir)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor k := len(fsi) - 1; k >= 0; k-- {\n\t\t// remove directories, and files without the correct extension\n\t\tif fsi[k].IsDir() || !strings.HasSuffix(fsi[k].Name(), \".gz\") {\n\t\t\tfsi = append(fsi[:k], fsi[k+1:]...)\n\t\t}\n\t}\n\n\treturn &fileInfoGroup{dir: dir, fsi: fsi}, nil\n}", "func (fs FS) NewStat() (Stat, error) {\n\treturn fs.Stat()\n}", "func NewFile(path string, info os.FileInfo) *File {\n\tfile := &File{Path: path}\n\tfile.Load(info)\n\treturn file\n}", "func NewFile(path string, info os.FileInfo) *File {\n\tfile := new(File)\n\tfile.Path = filepath.ToSlash(path)\n\tfile.Load(info)\n\treturn file\n}", "func NewFile(f os.FileInfo, dir string) File {\n\tvar t string\n\tif f.IsDir() {\n\t\tt = DIR\n\t} else {\n\t\tt = FILE\n\t}\n\treturn File{\n\t\tName: f.Name(),\n\t\tType: t,\n\t\tPath: dir,\n\t\tModified: f.ModTime(),\n\t\tSize: f.Size(),\n\t}\n}", "func newFile(t *testing.T, text string) (*os.File, error) {\n\treturn newFileWithPrefix(t, \"woke-\", text)\n}", "func (h *Header) FileInfo() os.FileInfo {\n\treturn headerFileInfo{h}\n}", "func newDataFile(pathFrom, pathTo string, info os.FileInfo) *DataFile {\n\treturn &DataFile{\n\t\tname: info.Name(),\n\t\textension: filepath.Ext(pathFrom),\n\n\t\tpathFrom: pathFrom,\n\t\tpathTo: pathTo,\n\n\t\tmode: info.Mode(),\n\t\tsize: info.Size(),\n\t}\n}", "func New() *file {\n\treturn &file{}\n}", "func NewFile(fs *Filesystem, parent *File, name string, mode os.FileMode, modTime time.Time, isDir bool) *File {\n\tff := &File{\n\t\tfs: fs,\n\t\tparent: parent,\n\t\tname: name,\n\t\tmode: mode,\n\t\tmodTime: modTime,\n\t\tisDir: isDir,\n\t\tcontent: []byte(\"\"),\n\t\tentries: make(map[string]*File),\n\t}\n\tif parent != nil {\n\t\tparent.entries[name] = ff\n\t}\n\treturn ff\n}", "func NewFileInfoCreateParams() *FileInfoCreateParams {\n\treturn &FileInfoCreateParams{\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (e *Entry) makeInfo(hash tree.Hash) tree.FileInfo {\n\tif e.notFound {\n\t\treturn tree.NewFileInfo(e.RelativePath(), e.Type(), 0, 0, time.Time{}, 0, 0, tree.Hash{})\n\t}\n\tinode, _ := e.Id()\n\treturn tree.NewFileInfo(e.RelativePath(), e.Type(), e.Mode(), e.HasMode(), e.ModTime(), e.Size(), inode, hash)\n}", "func (f *Fs) newObjectInfo(src fs.ObjectInfo) *ObjectInfo {\n\treturn &ObjectInfo{\n\t\tObjectInfo: src,\n\t\tf: f,\n\t\tmeta: nil,\n\t}\n}", "func (cs *Record) SetFileInfo(filename string, version string) {\n\tcs.Filename = filename\n\tcs.Version = version\n}", "func serializeFileInfo(info tree.FileInfo) *FileInfo {\n\tpath := info.RelativePath()\n\n\t// This is the root, which doesn't have metadata.\n\t// Serialize a custom object, so we don't panic.\n\tif _, ok := info.(*dtdiff.Entry); ok && len(path) == 0 {\n\t\tfileType := FileType(tree.TYPE_DIRECTORY)\n\t\treturn &FileInfo{\n\t\t\tPath: path,\n\t\t\tType: &fileType,\n\t\t}\n\t}\n\n\tsize := info.Size()\n\tfileType := FileType(info.Type())\n\tmode := uint32(info.Mode())\n\thasMode := uint32(info.HasMode())\n\thashType := uint32(info.Hash().Type)\n\tfileInfo := &FileInfo{\n\t\tPath: path,\n\t\tType: &fileType,\n\t\tMode: &mode,\n\t\tHasMode: &hasMode,\n\t\tSize: &size,\n\t\tHashType: &hashType,\n\t\tHashData: info.Hash().Data,\n\t}\n\tif !info.ModTime().IsZero() {\n\t\tmodTime := info.ModTime().UnixNano()\n\t\tfileInfo.ModTime = &modTime\n\t}\n\tif info.Inode() != 0 {\n\t\tfileInfo.Inode = proto.Uint64(info.Inode())\n\t}\n\treturn fileInfo\n}", "func InfoNew(opts ...func(*Info)) *Info {\n\tinf := &Info{\n\t\tVersion: version,\n\t\tUUID: createV4UUID(),\n\t}\n\tfor _, f := range opts {\n\t\tf(inf)\n\t}\n\treturn inf\n}", "func New(filePath string) File {\n\treturn File{\n\t\tFilepath: filePath,\n\t}\n}", "func New(name string) *File {\n\treturn &File{Name: name}\n}", "func (w *walker) updateFileInfo(file, curFile protocol.FileInfo) protocol.FileInfo {\n\tif file.Type == protocol.FileInfoTypeFile && runtime.GOOS == \"windows\" {\n\t\t// If we have an existing index entry, copy the executable bits\n\t\t// from there.\n\t\tfile.Permissions |= (curFile.Permissions & 0111)\n\t}\n\tfile.Version = curFile.Version.Update(w.ShortID)\n\tfile.ModifiedBy = w.ShortID\n\tfile.LocalFlags = w.LocalFlags\n\treturn file\n}", "func NewFileObject(obj *swift.ObjectOpenFile) (f *FileObject, err error) {\n\tf = &FileObject{obj: obj, loc: 0}\n\tf.length, err = obj.Length()\n\treturn f, err\n}", "func New(name string, current, target float64) *VersionInfo {\n\treturn &VersionInfo{name, current, target}\n}", "func NewFile(s *Super, i *proto.InodeInfo) fs.Node {\n\treturn &File{super: s, info: i}\n}", "func NewFileInfoThumbnail(mime string) *FileInfoThumbnail {\n\tthis := FileInfoThumbnail{}\n\tthis.Mime = mime\n\treturn &this\n}", "func newCacheItem(isFile bool) *cacheItem {\n\treturn &cacheItem{atime: time.Now(), isFile: isFile}\n}", "func (o *FileInfoCreateParams) WithInfo(info *models.FileInfo) *FileInfoCreateParams {\n\to.SetInfo(info)\n\treturn o\n}", "func NewInfo(path string) *Info {\n\tc := &Info{\n\t\tpath: path,\n\t}\n\n\tc.Load()\n\n\treturn c\n}", "func (r ReadBase) FileInfo(path, module string) (FileInfo, error) {\n\tbase, path, err := r.getPath(path, module)\n\tif err != nil {\n\t\treturn FileInfo{}, err\n\t}\n\treturn fileInfo(base, path)\n}", "func NewFile(patterns []Pattern, depth int, index int, entry fs.DirEntry, oldParent, newParent string) File {\n\tname := entry.Name()\n\tresult, newName := ApplyPattern(patterns, depth, name)\n\treturn File{\n\t\tDepth: depth,\n\t\tIndex: index,\n\t\tName: name,\n\t\tNewName: newName,\n\t\tResult: result,\n\t\tParent: oldParent,\n\t\tNewParent: newParent,\n\t\tIsDir: entry.IsDir(),\n\t}\n}", "func newFile(datadir, what, uuid string) (*File, error) {\n\ttimestamp := time.Now().UTC()\n\tdir := path.Join(datadir, what, timestamp.Format(\"2006/01/02\"))\n\terr := os.MkdirAll(dir, 0755)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tname := dir + \"/ndt7-\" + what + \"-\" + timestamp.Format(\"20060102T150405.000000000Z\") + \".\" + uuid + \".jsonl.gz\"\n\t// My assumption here is that we have nanosecond precision and hence it's\n\t// unlikely to have conflicts. If I'm wrong, O_EXCL will let us know.\n\tfp, err := os.OpenFile(name, os.O_WRONLY|os.O_CREATE|os.O_EXCL, 0644)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\twriter, err := gzip.NewWriterLevel(fp, gzip.BestSpeed)\n\tif err != nil {\n\t\tfp.Close()\n\t\treturn nil, err\n\t}\n\treturn &File{\n\t\tWriter: writer,\n\t\tFp: fp,\n\t}, nil\n}", "func newFileCache(opts fileCacheOptions) *fileCache {\n\treturn &fileCache{\n\t\ttoks: make(map[string]cacheToken),\n\t\tch: newCache(opts.Handles),\n\n\t\topts: opts,\n\t}\n}", "func newFileStatFromGetFileInformationByHandle(path string, h syscall.Handle) (fs *fileStat, err error) {\n\tvar d syscall.ByHandleFileInformation\n\terr = syscall.GetFileInformationByHandle(h, &d)\n\tif err != nil {\n\t\treturn nil, &PathError{Op: \"GetFileInformationByHandle\", Path: path, Err: err}\n\t}\n\n\tvar ti windows.FILE_ATTRIBUTE_TAG_INFO\n\terr = windows.GetFileInformationByHandleEx(h, windows.FileAttributeTagInfo, (*byte)(unsafe.Pointer(&ti)), uint32(unsafe.Sizeof(ti)))\n\tif err != nil {\n\t\tif errno, ok := err.(syscall.Errno); ok && errno == windows.ERROR_INVALID_PARAMETER {\n\t\t\t// It appears calling GetFileInformationByHandleEx with\n\t\t\t// FILE_ATTRIBUTE_TAG_INFO fails on FAT file system with\n\t\t\t// ERROR_INVALID_PARAMETER. Clear ti.ReparseTag in that\n\t\t\t// instance to indicate no symlinks are possible.\n\t\t\tti.ReparseTag = 0\n\t\t} else {\n\t\t\treturn nil, &PathError{Op: \"GetFileInformationByHandleEx\", Path: path, Err: err}\n\t\t}\n\t}\n\n\treturn &fileStat{\n\t\tname: basename(path),\n\t\tFileAttributes: d.FileAttributes,\n\t\tCreationTime: d.CreationTime,\n\t\tLastAccessTime: d.LastAccessTime,\n\t\tLastWriteTime: d.LastWriteTime,\n\t\tFileSizeHigh: d.FileSizeHigh,\n\t\tFileSizeLow: d.FileSizeLow,\n\t\tvol: d.VolumeSerialNumber,\n\t\tidxhi: d.FileIndexHigh,\n\t\tidxlo: d.FileIndexLow,\n\t\tReserved0: ti.ReparseTag,\n\t\t// fileStat.path is used by os.SameFile to decide if it needs\n\t\t// to fetch vol, idxhi and idxlo. But these are already set,\n\t\t// so set fileStat.path to \"\" to prevent os.SameFile doing it again.\n\t}, nil\n}", "func New(f *os.File) *Metadata {\n\treturn &Metadata{\n\t\tfile: f,\n\t\ttags: make(map[string]string),\n\t}\n}", "func (o *OpenapiProcessFileAllOf) SetFileInfo(v OpenapiOpenApiSpecificationRelationship) {\n\to.FileInfo = &v\n}", "func NewFile(fd uintptr, name string) *os.File", "func convertFileInfo(\n\tin os.FileInfo,\n\tsymlinkTarget string,\n\tuserRegistry sys.UserRegistry,\n\tgroupRegistry sys.GroupRegistry) (out *FileInfo, err error) {\n\t// Grab system-specific info.\n\tstatT, ok := in.Sys().(*syscall.Stat_t)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"Unexpected sys value: %v\", in.Sys())\n\t}\n\n\tif statT.Size < 0 {\n\t\tpanic(fmt.Sprintf(\"Unexpected size: %d\", statT.Size))\n\t}\n\n\t// Create the basic struct.\n\tout = &FileInfo{\n\t\tName: in.Name(),\n\t\tPermissions: in.Mode() & permissionBits,\n\t\tUid: sys.UserId(statT.Uid),\n\t\tGid: sys.GroupId(statT.Gid),\n\t\tMTime: in.ModTime(),\n\t\tSize: uint64(statT.Size),\n\t\tContainingDevice: statT.Dev,\n\t\tInode: statT.Ino,\n\t\tTarget: symlinkTarget,\n\t}\n\n\t// Attempt to look up user info.\n\tusername, err := userRegistry.FindById(out.Uid)\n\n\tif _, ok := err.(sys.NotFoundError); ok {\n\t\terr = nil\n\t} else if err != nil {\n\t\treturn nil, fmt.Errorf(\"Looking up user: %v\", err)\n\t} else {\n\t\tout.Username = &username\n\t}\n\n\t// Attempt to look up group info.\n\tgroupname, err := groupRegistry.FindById(out.Gid)\n\n\tif _, ok := err.(sys.NotFoundError); ok {\n\t\terr = nil\n\t} else if err != nil {\n\t\treturn nil, fmt.Errorf(\"Looking up group: %v\", err)\n\t} else {\n\t\tout.Groupname = &groupname\n\t}\n\n\t// Convert the type.\n\ttypeBits := in.Mode() & (os.ModeType | os.ModeCharDevice)\n\tswitch typeBits {\n\tcase 0:\n\t\tout.Type = TypeFile\n\tcase os.ModeDir:\n\t\tout.Type = TypeDirectory\n\tcase os.ModeSymlink:\n\t\tout.Type = TypeSymlink\n\tcase os.ModeDevice:\n\t\tout.Type = TypeBlockDevice\n\t\tout.DeviceNumber = statT.Rdev\n\tcase os.ModeDevice | os.ModeCharDevice:\n\t\tout.Type = TypeCharDevice\n\t\tout.DeviceNumber = statT.Rdev\n\tcase os.ModeNamedPipe:\n\t\tout.Type = TypeNamedPipe\n\tcase os.ModeSocket:\n\t\tout.Type = TypeSocket\n\tdefault:\n\t\treturn out, fmt.Errorf(\"Unhandled mode: %v\", in.Mode())\n\t}\n\n\treturn out, nil\n}", "func Generate(filePath string, info os.FileInfo) (*FileStat, error) {\n\tfs := &FileStat{\n\t\tpath: filePath,\n\t\tname: info.Name(),\n\t\tsize: info.Size(),\n\t\tmode: info.Mode(),\n\t\tmodTime: info.ModTime(),\n\t\tisDir: info.IsDir(),\n\t}\n\n\tif !fs.isDir {\n\t\tfile, err := os.Open(filePath)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tbuf, err := ioutil.ReadAll(file)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tfs.Reader = *bytes.NewReader(buf)\n\t}\n\n\treturn fs, nil\n}", "func NewFile(filename string, offset int64, length int) (*File, error) {\n\tlogger.Debugf(\"Will try to map new %s, %d, %d\", filename, offset, length)\n\n\tf, err := os.Create(filename)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer f.Close()\n\n\t_, err = f.Seek(int64(length-1), 0)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\t_, err = f.Write([]byte(\"\\000\"))\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tmmap, err := doMap(f, offset, length)\n\tlogger.Debugf(\"Mapped a new file @%v for %d\", mmap.mmap, mmap.size)\n\n\treturn mmap, err\n}", "func FileHeaderFileInfo(h *zip.FileHeader,) os.FileInfo", "func New() (GitInfo, error) {\n\treturn NewWithPath(\"\")\n}", "func NewMockFile(name string, content []byte, files ...MockFile) MockFile {\n\t// If any files have been passed in, this is a directory.\n\tvar isDir bool\n\tif len(files) != 0 {\n\t\tisDir = true\n\t}\n\n\tinfo := NewMockFileInfo(name, isDir)\n\n\tm := MockFile{\n\t\tFiles: files,\n\t\tInfo: info,\n\t\tcontent: content,\n\t}\n\n\treturn m\n}", "func (s service) FileInfo(\n\tctx context.Context,\n\tfileID string,\n) (data FileInfo, err error) {\n\terr = s.svc.ResourceOp(\n\t\tctx,\n\t\thttp.MethodGet,\n\t\tfileInfoFormat,\n\t\t&fileInfoParams{FileID: fileID},\n\t\t&data,\n\t)\n\treturn\n}", "func NewFs() *Fs { return &Fs{make(map[string]*file)} }", "func New(r ReadAtSeeker) (file *File, err error) {\n\t// TODO(u): Figure out which headers that should always be parsed.\n\t// * DOS header\n\t// - Contains no relevant information, but is required to locate the\n\t// file header offset.\n\t// * File header\n\t// - Specifies the intended architecture of the binary, and the number\n\t// of sections contained within the file.\n\t// * Optional header\n\t// - Specifies the code, data and image base addresses, the entry point\n\t// point address, and the data directories.\n\treturn &File{r: r}, nil\n}", "func NewFile(filename string, base, size int) *File {\n\tif base < 0 {\n\t\tbase = 1\n\t}\n\treturn &File{sync.RWMutex{}, filename, index(base), index(size), []index{0}, nil}\n}", "func (n *Node) NewFile(name string) (*File, error) {\n\tfullpath := filepath.Join(n.basepath, name)\n\t_, err := os.Stat(fullpath)\n\n\tif !os.IsNotExist(err) {\n\t\treturn nil, ErrFileAlreadyExists\n\t}\n\n\tuid := NewFileUID(name)\n\n\treturn &File{\n\t\tName: name,\n\t\tID: uid,\n\t}, nil\n}", "func NewFileSystem(t mockConstructorTestingTNewFileSystem) *FileSystem {\n\tmock := &FileSystem{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func NewFile(base string) (*FileStore, error) {\n\tf, err := os.Stat(base)\n\n\t// If the directory does not exist...\n\tif os.IsNotExist(err) {\n\t\t// ...we try to create it.\n\t\tif err := os.Mkdir(base, 0750); err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Error creating '%s': %s\", base, err)\n\t\t}\n\t} else if err != nil {\n\t\t// Could be a permission problem, let's check.\n\t\treturn nil, handleErr(base, err)\n\t} else if !f.IsDir() {\n\t\t// It exists and we can access it, but it is simply not a directory.\n\t\treturn nil, fmt.Errorf(\"'%s' is not a directory\", base)\n\t}\n\n\t// All good.\n\tstore := &FileStore{basepath: base}\n\n\treturn store, nil\n}", "func NewInfo() *Info {\n\treturn &Info{\n\t\tTypes: map[string]bool{},\n\t}\n}", "func New(m string) error {\n\treturn errors{\n\t\tmsg: m,\n\t\tfileInfo: getFileInfo(),\n\t}\n}", "func NewUserInfo() *UserInfo {\n\treturn &UserInfo{}\n}", "func NewFile(name string) (Capabilities, error) {\n\treturn newFile(name)\n}", "func (f *FileName) GetFileInfo() (*FileInfo, error) {\n\treturn f.conn.GetFileInfo(f.ID)\n}", "func NewFileManager(log backend.LogFunc) *FileManager {\n\treturn &FileManager{\n\t\tpatch: make(map[string][]*plugin.Generated),\n\t\tindex: make(map[string]int),\n\t\tcount: make(map[string]int),\n\t\tlog: log,\n\t}\n}", "func NewFileFilter(t mockConstructorTestingTNewFileFilter) *FileFilter {\n\tmock := &FileFilter{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func makeNewFileVersion(writer http.ResponseWriter,\n\tfileMainPath string,\n\tfilename string,\n\tversion int) error {\n\n\tdirExists, err := exists(fileMainPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !dirExists {\n\t\terr := os.MkdirAll(fileMainPath, 0777)\n\t\tif err != nil {\n\t\t\tlog.Println(err)\n\t\t\thttp.Error(writer, \"Could not write new directory\", 500)\n\t\t\treturn err\n\t\t}\n\t}\n\n\tfileExists, _ := exists(fmt.Sprintf(\"%s/%s\", fileMainPath, filename))\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !fileExists {\n\t\tlog.Println(\"about to make new file:\")\n\t\tlog.Println(fmt.Sprintf(\"%s/%s\", fileMainPath, filename))\n\t\tf, err := os.Create(fmt.Sprintf(\"%s/%s\", fileMainPath, filename))\n\t\tif err != nil {\n\t\t\tlog.Print(err)\n\t\t\thttp.Error(writer, \"Could not create new file\", 500)\n\t\t\treturn err\n\t\t}\n\t\t_, err = f.WriteString(fmt.Sprintf(\"%d\\n\", version))\n\t\tif err != nil {\n\t\t\tlog.Print(err)\n\t\t\thttp.Error(writer, \"Could not write to file\", 500)\n\t\t\treturn err\n\t\t}\n\t\tf.Close()\n\n\t}\n\treturn nil\n}", "func NewFile(ctx context.Context, dirent *fs.Dirent, name string, flags fs.FileFlags, i *inodeOperations, handles *handles) *fs.File {\n\t// Remote file systems enforce readability/writability at an offset,\n\t// see fs/9p/vfs_inode.c:v9fs_vfs_atomic_open -> fs/open.c:finish_open.\n\tflags.Pread = true\n\tflags.Pwrite = true\n\n\tif fs.IsFile(dirent.Inode.StableAttr) {\n\t\t// If cache policy is \"remote revalidating\", then we must\n\t\t// ensure that we have a host FD. Otherwise, the\n\t\t// sentry-internal page cache will be used, and we can end up\n\t\t// in an inconsistent state if the remote file changes.\n\t\tcp := dirent.Inode.InodeOperations.(*inodeOperations).session().cachePolicy\n\t\tif cp == cacheRemoteRevalidating && handles.Host == nil {\n\t\t\tpanic(fmt.Sprintf(\"remote-revalidating cache policy requires gofer to donate host FD, but file %q did not have host FD\", name))\n\t\t}\n\t}\n\n\tf := &fileOperations{\n\t\tinodeOperations: i,\n\t\thandles: handles,\n\t\tflags: flags,\n\t}\n\tif flags.Write {\n\t\tif err := dirent.Inode.CheckPermission(ctx, fs.PermMask{Execute: true}); err == nil {\n\t\t\topensWX.Increment()\n\t\t\tlog.Warningf(\"Opened a writable executable: %q\", name)\n\t\t}\n\t}\n\tif handles.Host != nil {\n\t\topensHost.Increment()\n\t} else {\n\t\topens9P.Increment()\n\t}\n\treturn fs.NewFile(ctx, dirent, flags, f)\n}", "func newFileStatFromWin32finddata(d *syscall.Win32finddata) *fileStat {\n\treturn &fileStat{\n\t\tFileAttributes: d.FileAttributes,\n\t\tCreationTime: d.CreationTime,\n\t\tLastAccessTime: d.LastAccessTime,\n\t\tLastWriteTime: d.LastWriteTime,\n\t\tFileSizeHigh: d.FileSizeHigh,\n\t\tFileSizeLow: d.FileSizeLow,\n\t\tReserved0: d.Reserved0,\n\t}\n}", "func NewFile() *File {\n\treturn (*File)(allocFileMemory(1))\n}", "func (e *Entry) Info() tree.FileInfo {\n\treturn e.makeInfo(tree.Hash{})\n}", "func (m MockFile) Stat() (os.FileInfo, error) {\n\treturn m.Info, nil\n}", "func newFile(fs *FileSystem, bucket, key string) (*File, error) {\n\tif fs == nil {\n\t\treturn nil, errors.New(\"non-nil gs.FileSystem pointer is required\")\n\t}\n\tif bucket == \"\" || key == \"\" {\n\t\treturn nil, errors.New(\"non-empty strings for Bucket and Key are required\")\n\t}\n\tkey = utils.CleanPrefix(key)\n\treturn &File{\n\t\tfileSystem: fs,\n\t\tbucket: bucket,\n\t\tkey: key,\n\t}, nil\n}", "func (graphPtr *CrlGraph) newFile(path string, filename string) (*os.File, error) {\n\tfullPath := path + \"/\" + filename + \".dot\"\n\tfile, err := os.OpenFile(fullPath, os.O_RDWR|os.O_CREATE, 0644)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"CrlGraph.newFile failed\")\n\t}\n\treturn file, nil\n}", "func NewFile(name string, modTime time.Time, size int64, compressed bool, data []byte) *File {\n\treturn &File{\n\t\tname: path.Base(ToEFSPath(name)),\n\t\tsize: size,\n\t\tmodTime: modTime,\n\t\tcompressed: compressed,\n\t\tdata: data,\n\t}\n}", "func SameFile(fi1, fi2 os.FileInfo,) bool", "func newLogFileCompactInfo() *logFileCompactInfo {\n\treturn &logFileCompactInfo{\n\t\tmms: make(map[string]*logFileMeasurementCompactInfo),\n\t}\n}", "func (sp SourceSpec) NewFilesystem(base string) *Filesystem {\n\treturn &Filesystem{SourceSpec: sp, Base: base}\n}", "func newZipFile(fName string) (*zipFile, error) {\n\tf, err := os.Create(fName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &zipFile{\n\t\tf,\n\t\tzip.NewWriter(f),\n\t}, nil\n}", "func New(base fsio.ReadSeekerAt) (*FS, error) {\n\tsize, err := fsio.GetSize(base)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tzr, err := zip.NewReader(base, size)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &FS{zr}, nil\n}", "func New(f *os.File) iterator.Iterator {\n\tscanner := bufio.NewScanner(f)\n\t// read in a line to initialize the state of the\n\t// file iterator.\n\tvar line *string\n\tif scanner.Scan() {\n\t\tline = utils.StrPtr(scanner.Text())\n\t}\n\n\treturn &FileIterator{\n\t\tf: f,\n\t\tscanner: scanner,\n\t\tline: line,\n\t}\n}", "func New(path string) *FileData {\n\treturn &FileData{\n\t\tpath: path,\n\t}\n}", "func NewFile(filename string) (*File, error) {\n\n\tlf := &File{filename: filename}\n\treturn lf, lf.ReOpen()\n}", "func NewFileByInfoHash(baseDir string) ClientImpl {\n\treturn NewFileWithCustomPathMaker(baseDir, infoHashPathMaker)\n}", "func New() *FileData {\n\treturn &FileData{\n\t\tData:\t\t\"aGVsbG8=\",\n\t\tChecksum:\t\"2cf24dba5fb0a30e26e83b2ac5b9e29e1b161e5c1fa7425e73043362938b9824\",\n\t\tName:\t\t\"hello.txt\",\n\t}\n}", "func NewVersionInfo() VersionInfo {\n\tver := VersionInfo{\n\t\tVersion: \"dev\",\n\t\tBuildDate: BuildDate,\n\t\tGitCommit: GitCommit,\n\t\tGitState: \"\",\n\t}\n\tif Version != \"\" {\n\t\tver.Version = Version\n\t\tver.GitState = \"clean\"\n\t}\n\tif GitState != \"\" {\n\t\tver.GitState = GitState\n\t}\n\treturn ver\n}", "func ToFileAttribute(info os.FileInfo) *FileAttribute {\n\tf := FileAttribute{}\n\n\tm := info.Mode()\n\tf.FileMode = uint32(m)\n\tif info.IsDir() {\n\t\tf.Type = FileTypeDirectory\n\t} else if m&os.ModeSymlink != 0 {\n\t\tf.Type = FileTypeLink\n\t} else if m&os.ModeCharDevice != 0 {\n\t\tf.Type = FileTypeCharacter\n\t\t// TODO: set major/minor dev number\n\t\t//f.SpecData = 0,0\n\t} else if m&os.ModeDevice != 0 {\n\t\tf.Type = FileTypeBlock\n\t\t// TODO: set major/minor dev number\n\t\t//f.SpecData = 0,0\n\t} else if m&os.ModeSocket != 0 {\n\t\tf.Type = FileTypeSocket\n\t} else if m&os.ModeNamedPipe != 0 {\n\t\tf.Type = FileTypeFIFO\n\t} else {\n\t\tf.Type = FileTypeRegular\n\t}\n\t// The number of hard links to the file.\n\tf.Nlink = 1\n\n\tif a := file.GetInfo(info); a != nil {\n\t\tf.Nlink = a.Nlink\n\t\tf.UID = a.UID\n\t\tf.GID = a.GID\n\t}\n\n\tf.Filesize = uint64(info.Size())\n\tf.Used = uint64(info.Size())\n\tf.Atime = ToNFSTime(info.ModTime())\n\tf.Mtime = f.Atime\n\tf.Ctime = f.Atime\n\treturn &f\n}", "func NewFile(domain, instance, class, id string) (File, error) {\n\tvar f File\n\tvar err error\n\n\t// Get EXE directory\n\texecutable, err := os.Executable()\n\tif err != nil {\n\t\treturn f, errors.Wrap(err, \"os.executable\")\n\t}\n\texeDir := filepath.Dir(executable)\n\n\tstateDir := filepath.Join(exeDir, \"xestate\")\n\tif _, err = os.Stat(stateDir); os.IsNotExist(err) {\n\t\terr = os.Mkdir(stateDir, 0644)\n\t}\n\tif err != nil {\n\t\treturn f, errors.Wrap(err, \"os.mkdir\")\n\t}\n\n\tf.Name = filepath.Join(stateDir, fileName(domain, instance, class, id))\n\treturn f, nil\n}", "func newFileResultLogger(base string) *FileResultLogger {\n\tthis := new(FileResultLogger)\n\tif err := this.createOrIgnore(base); err != nil {\n\t\tthis.lastError = err\n\t\treturn this\n\t}\n\tthis.baseDir = base\n\treturn this\n}", "func NewFile(name string, size int) *File {\n\treturn &File{\n\t\tname: name,\n\t\tsize: size,\n\t}\n}", "func NewFileEntry(name string, size uint64) FileEntry {\n\treturn FileEntry{name, size}\n}", "func newFileFromParams(upload *Upload, params *common.File) *File {\n\tfile := &File{}\n\tfile.upload = upload\n\tfile.metadata = params\n\tfile.Name = params.Name\n\tfile.Size = params.Size\n\treturn file\n}", "func DebugNewFile(s string) io.WriteCloser {\n\treturn currentDebugProvider.NewFile(s)\n}", "func (fs *FileSystem) NewFile(name string, data []byte) (f File, err error) {\n\tf = File{\n\t\tName: name,\n\t\tPermissions: os.FileMode(0644),\n\t\tSize: len(data),\n\t\tCreated: time.Now(),\n\t\tModified: time.Now(),\n\t\tData: data,\n\t}\n\n\tif fs.doCompression {\n\t\tf.IsCompressed = true\n\t\tf.Data = compressByte(data)\n\t\tf.Size = len(f.Data)\n\t}\n\n\tif fs.encryptPassphrase != \"\" {\n\t\tf.IsEncrypted = true\n\t\t// TODO: do encryption\n\t}\n\treturn\n}", "func (l *Location) NewFile(filePath string) (vfs.File, error) {\n\tif l == nil {\n\t\treturn nil, errors.New(\"non-nil sftp.Location pointer receiver is required\")\n\t}\n\tif filePath == \"\" {\n\t\treturn nil, errors.New(\"non-empty string filePath is required\")\n\t}\n\terr := utils.ValidateRelativeFilePath(filePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tnewFile := &File{\n\t\tfileSystem: l.fileSystem,\n\t\tAuthority: l.Authority,\n\t\tpath: utils.EnsureLeadingSlash(path.Join(l.path, filePath)),\n\t}\n\treturn newFile, nil\n}", "func newLocalFile(fs afero.Fs, nativePath, slashpath string, m *Matcher, mt media.Types) (*localFile, error) {\n\tf, err := fs.Open(nativePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer f.Close()\n\tlf := &localFile{\n\t\tNativePath: nativePath,\n\t\tSlashPath: slashpath,\n\t\tfs: fs,\n\t\tmatcher: m,\n\t\tmediaTypes: mt,\n\t}\n\tif m != nil && m.Gzip {\n\t\t// We're going to gzip the content. Do it once now, and cache the result\n\t\t// in gzipped. The UploadSize is the size of the gzipped content.\n\t\tgz := gzip.NewWriter(&lf.gzipped)\n\t\tif _, err := io.Copy(gz, f); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif err := gz.Close(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tlf.UploadSize = int64(lf.gzipped.Len())\n\t} else {\n\t\t// Raw content. Just get the UploadSize.\n\t\tinfo, err := f.Stat()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tlf.UploadSize = info.Size()\n\t}\n\treturn lf, nil\n}", "func New(source string, body io.ReadSeeker, lastModified time.Time, hash HashFn) (*File, error) {\n\tdigest, size, hashErr := hash(body)\n\tif hashErr != nil {\n\t\treturn nil, hashErr\n\t}\n\t// Prevent creating a file from a source containing metadata.\n\tif size < MetaFileMaxSize {\n\t\tbody.Seek(0, io.SeekStart)\n\t\tif meta, err := ioutil.ReadAll(body); err != nil {\n\t\t\treturn nil, err\n\t\t} else if ValidateMeta(meta) == nil {\n\t\t\treturn nil, fmt.Errorf(\"%w: use sync to interact with metafiles directly\", os.ErrInvalid)\n\t\t}\n\t}\n\tbody.Seek(0, io.SeekStart)\n\tfile := &File{\n\t\tName: digest,\n\t\tSource: source,\n\t\tSize: size,\n\t\tLastModified: lastModified,\n\t\tBody: body,\n\t}\n\tfile.Meta = NewMetaFromFile(file)\n\treturn file, nil\n}", "func NewFile(name string, size uint64, metafilehash []byte) File {\n\treturn File{\n\t\tName: name,\n\t\tSize: size,\n\t\tMetafileHash: metafilehash,\n\t}\n}", "func New(filePath string) (logSeeker *LogSeeker, err error) {\n\tlogSeeker = &LogSeeker{\n\t\tFilePath: filePath,\n\t}\n\n\tlogSeeker.file, err = os.Open(filePath)\n\treturn\n}" ]
[ "0.7941719", "0.79101765", "0.761692", "0.7125709", "0.70744866", "0.6888411", "0.6767215", "0.6567179", "0.65616375", "0.64753735", "0.62839127", "0.61712617", "0.60467106", "0.5909766", "0.5876665", "0.58575416", "0.58478796", "0.58459", "0.58442307", "0.5804826", "0.5786716", "0.5779668", "0.5768729", "0.57548225", "0.5754736", "0.57183844", "0.56810427", "0.56283146", "0.56274", "0.5618619", "0.55886585", "0.55196214", "0.5511704", "0.5509233", "0.54844767", "0.5478997", "0.54660153", "0.5456838", "0.5456257", "0.5433286", "0.5431722", "0.54104185", "0.5385776", "0.5345911", "0.5343281", "0.533728", "0.52668625", "0.52661175", "0.52486455", "0.5240612", "0.5240223", "0.5237605", "0.52137643", "0.52069867", "0.51952684", "0.5170103", "0.5169479", "0.51519406", "0.514723", "0.5142929", "0.51407033", "0.51394606", "0.51350945", "0.51328754", "0.51251316", "0.51212937", "0.51015776", "0.5093336", "0.50920135", "0.5088409", "0.50850403", "0.5082375", "0.50818217", "0.50804245", "0.5079766", "0.5078995", "0.5074674", "0.507363", "0.5060707", "0.50562686", "0.50555223", "0.50537026", "0.50517035", "0.5041869", "0.50374174", "0.50362927", "0.50321764", "0.5030514", "0.50214267", "0.5018518", "0.50154483", "0.50106007", "0.500872", "0.5000846", "0.49983937", "0.4993606", "0.49935403", "0.49908593", "0.4987715", "0.49823317" ]
0.7401893
3
branchName takes the root directory and relative path to the directory and returns the branch name
func branchName(root, dirRelPath, openAPIFileName string) string { name := filepath.Base(dirRelPath) _, err := os.Stat(filepath.Join(root, dirRelPath, openAPIFileName)) if !os.IsNotExist(err) { // add Pkg: prefix indicating that it is a separate package as it has // openAPIFile return fmt.Sprintf("Pkg: %s", name) } return name }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (g *github) GetBranchName() string { return g.branchName }", "func (o FunctionBuildConfigSourceRepoSourceOutput) BranchName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v FunctionBuildConfigSourceRepoSource) *string { return v.BranchName }).(pulumi.StringPtrOutput)\n}", "func (ref RefName) BranchName() string {\n\treturn ref.nameWithoutPrefix(BranchPrefix)\n}", "func (o TriggerBuildSourceRepoSourceOutput) BranchName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TriggerBuildSourceRepoSource) *string { return v.BranchName }).(pulumi.StringPtrOutput)\n}", "func (o FunctionBuildConfigSourceRepoSourcePtrOutput) BranchName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FunctionBuildConfigSourceRepoSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.BranchName\n\t}).(pulumi.StringPtrOutput)\n}", "func (o TriggerBuildSourceRepoSourcePtrOutput) BranchName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TriggerBuildSourceRepoSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.BranchName\n\t}).(pulumi.StringPtrOutput)\n}", "func gitBranchName() string {\n\t// branch name variable set by Github Actions\n\tif branch, isset := os.LookupEnv(\"GITHUB_HEAD_REF\"); isset && branch != \"\" {\n\t\treturn \"origin/\" + branch\n\t}\n\tif branch, isset := os.LookupEnv(\"GITHUB_REF\"); isset && branch != \"\" {\n\t\treturn \"origin/\" + strings.TrimPrefix(branch, \"refs/heads/\")\n\t}\n\tbranch := getCmdOutput(\"git rev-parse --abbrev-ref HEAD\")\n\treturn branch\n}", "func (p *PipelineActivity) BranchName() string {\n\tpipelineName := p.Spec.Pipeline\n\tif pipelineName == \"\" {\n\t\treturn \"\"\n\t}\n\tpaths := strings.Split(pipelineName, \"/\")\n\tbranch := paths[len(paths)-1]\n\tp.Spec.GitBranch = branch\n\treturn branch\n}", "func branchName() (string, string) {\n\tbranch := gitBranchName()\n\treleaseName := strings.TrimPrefix(branch, \"origin/\")\n\n\treturn releaseName, branch\n}", "func Branch() string {\n\treturn run.Capture(\"git rev-parse --abbrev-ref HEAD\")\n}", "func (g *GitLocal) Branch(dir string) (string, error) {\n\treturn g.GitCLI.Branch(dir)\n}", "func GitBranch(dir string) (branch string, err error) {\n\tcmd := Cmd(dir, \"git rev-parse --abbrev-ref HEAD\")\n\tout, err := cmd.Output()\n\tif err != nil {\n\t\treturn\n\t}\n\tbranch = strings.TrimSpace(string(out))\n\tif branch == \"HEAD\" {\n\t\terr = ErrGitDetached\n\t}\n\treturn\n}", "func (c *config) branch(name string) (output string, err error) {\n\tlog.Printf(\"creating branch: %v\", name)\n\n\tdefaultCommand := []string{\"branch\", name}\n\n\treturn c.command(defaultCommand...)\n\n}", "func (s *splicer) branch(name string) error {\n\treturn s.gitCall(\"checkout\", \"-B\", name, \"master\")\n}", "func (o TriggerTriggerTemplateOutput) BranchName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TriggerTriggerTemplate) *string { return v.BranchName }).(pulumi.StringPtrOutput)\n}", "func (self *Repository) Branch(path string) error { return nil }", "func (o TriggerTriggerTemplatePtrOutput) BranchName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TriggerTriggerTemplate) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.BranchName\n\t}).(pulumi.StringPtrOutput)\n}", "func GetBranchFromRef(ref string) string {\n\tparts := strings.Split(ref, \"/\")\n\treturn strings.Join(parts[2:], \"/\")\n}", "func (st *buildStatus) branch() string {\n\tif st.SubRev != \"\" {\n\t\treturn st.SubRevBranch\n\t}\n\treturn st.RevBranch\n}", "func GetCurrentBranch() string {\n\tcmd := exec.Command(\"git\", \"branch\")\n\tvar out bytes.Buffer\n\tcmd.Stdout = &out\n\tcmd.Run()\n\tlines := strings.Split(out.String(), \"\\n\")\n\tfor _, line := range lines {\n\t\tif strings.Contains(line, \"*\") {\n\t\t\tbranch := strings.Replace(line, \"*\", \"\", -1)\n\t\t\tbranch = strings.TrimSpace(branch)\n\t\t\treturn branch\n\t\t}\n\t}\n\treturn \"\"\n}", "func (ge *GollumEvent) BranchName() string {\n\treturn ge.raw.Payload[\"ref\"].(string)\n}", "func GitBranch(e *editarea.EditArea) string {\n\tcmdOut, err := exec.Command(\"git\", \"rev-parse\", \"--abbrev-ref\", \"HEAD\").Output()\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\treturn string(cmdOut)\n}", "func (self *StatusCommands) BranchBeingRebased() string {\n\tfor _, dir := range []string{\"rebase-merge\", \"rebase-apply\"} {\n\t\tif bytesContent, err := os.ReadFile(filepath.Join(self.repoPaths.WorktreeGitDirPath(), dir, \"head-name\")); err == nil {\n\t\t\treturn strings.TrimSpace(string(bytesContent))\n\t\t}\n\t}\n\treturn \"\"\n}", "func BranchNameFromArchiveRefDir(dir string) (data.BranchName, bool) {\n\tif !strings.HasPrefix(dir, ArchivedRevDirPrefix) {\n\t\treturn \"\", false\n\t}\n\n\trev, err := strconv.ParseInt(dir[len(ArchivedRevDirPrefix):], 10, 64)\n\tif err != nil {\n\t\treturn \"\", false\n\t}\n\n\treturn data.MakeRevBranchName(kbfsmd.Revision(rev)), true\n}", "func (r *Repo) BranchRef() string {\n\treturn fmt.Sprintf(\"refs/heads/%s\", r.Branch)\n}", "func binName(rel, prefix, repoRoot string) string {\n\treturn pathtools.RelBaseName(rel, prefix, repoRoot)\n}", "func GetDefaultBranchName(dbOwner, dbFolder, dbName string) (branchName string, err error) {\n\tdbQuery := `\n\t\tSELECT db.default_branch\n\t\tFROM sqlite_databases AS db\n\t\tWHERE db.user_id = (\n\t\t\t\tSELECT user_id\n\t\t\t\tFROM users\n\t\t\t\tWHERE lower(user_name) = lower($1)\n\t\t\t)\n\t\t\tAND db.folder = $2\n\t\t\tAND db.db_name = $3\n\t\t\tAND db.is_deleted = false`\n\tvar b pgx.NullString\n\terr = pdb.QueryRow(dbQuery, dbOwner, dbFolder, dbName).Scan(&b)\n\tif err != nil {\n\t\tif err != pgx.ErrNoRows {\n\t\t\tlog.Printf(\"Error when retrieving default branch name for database '%s%s%s': %v\\n\", dbOwner,\n\t\t\t\tdbFolder, dbName, err)\n\t\t} else {\n\t\t\tlog.Printf(\"No default branch name exists for database '%s%s%s'. This shouldn't happen\\n\", dbOwner,\n\t\t\t\tdbFolder, dbName)\n\t\t}\n\t\treturn\n\t}\n\tif b.Valid {\n\t\tbranchName = b.String\n\t}\n\treturn\n}", "func (o TriggerGithubPullRequestOutput) Branch() pulumi.StringOutput {\n\treturn o.ApplyT(func(v TriggerGithubPullRequest) string { return v.Branch }).(pulumi.StringOutput)\n}", "func GetTrackingBranchName(branchName string) string {\n\treturn \"origin/\" + branchName\n}", "func (prrce *PullRequestReviewCommentEvent) BranchName() string {\n\treturn \"\"\n}", "func checkBranchName(repo *models.Repository, name string) error {\n\tgitRepo, err := git.OpenRepository(repo.RepoPath())\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer gitRepo.Close()\n\n\tbranches, _, err := GetBranches(repo, 0, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, branch := range branches {\n\t\tif branch.Name == name {\n\t\t\treturn models.ErrBranchAlreadyExists{\n\t\t\t\tBranchName: branch.Name,\n\t\t\t}\n\t\t} else if (len(branch.Name) < len(name) && branch.Name+\"/\" == name[0:len(branch.Name)+1]) ||\n\t\t\t(len(branch.Name) > len(name) && name+\"/\" == branch.Name[0:len(name)+1]) {\n\t\t\treturn models.ErrBranchNameConflict{\n\t\t\t\tBranchName: branch.Name,\n\t\t\t}\n\t\t}\n\t}\n\n\tif _, err := gitRepo.GetTag(name); err == nil {\n\t\treturn models.ErrTagAlreadyExists{\n\t\t\tTagName: name,\n\t\t}\n\t}\n\n\treturn nil\n}", "func youtrackBranchNameFilter(branchName string) string {\n\tmatches := youtrackBranchNameRegEx.FindStringSubmatch(branchName)\n\tif matches != nil {\n\t\treturn fmt.Sprintf(\"%s-%s\", strings.ToUpper(matches[1]), matches[2])\n\t}\n\treturn \"\"\n}", "func (machine *Dishwasher) Branch(branch string) {\r\n machine.Append(func() (string, error) {\r\n // IDEA take care of special cases, such as the master branch\r\n // checking if branch exists\r\n branchExists := false\r\n output, oops := RunCommand(\"git branch\")\r\n if oops != nil {\r\n return string(output), oops\r\n }\r\n branches := strings.Split(output, \"\\n\")\r\n for _, rawMaybe := range branches {\r\n maybe := strings.Replace(rawMaybe, \"*\", \" \", -1)\r\n maybe = strings.TrimSpace(maybe)\r\n if maybe == branch {\r\n branchExists = true\r\n }\r\n }\r\n\r\n // creating branch if necessary\r\n if !branchExists {\r\n branchCmd := fmt.Sprintf(\"git branch %s\", branch)\r\n output, oops = RunCommand(branchCmd)\r\n if oops != nil {\r\n return string(output), oops\r\n }\r\n }\r\n\r\n // switching to target branch\r\n checkoutCmd := fmt.Sprintf(\"git checkout %s\", branch)\r\n output, oops = RunCommand(checkoutCmd)\r\n return string(output), oops\r\n })\r\n}", "func (o TriggerGithubPullRequestPtrOutput) Branch() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TriggerGithubPullRequest) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Branch\n\t}).(pulumi.StringPtrOutput)\n}", "func (n Node) CrumbName(url string) string {\n\treturn filepath.Base(url)\n}", "func checkoutLocalBranch(config *GitXargsConfig, ref *plumbing.Reference, worktree *git.Worktree, remoteRepository *github.Repository, localRepository *git.Repository) (plumbing.ReferenceName, error) {\n\tlogger := logging.GetLogger(\"git-xargs\")\n\n\t// BranchName is a global variable that is set in cmd/root.go. It is override-able by the operator via the --branch-name or -b flag. It defaults to \"git-xargs\"\n\n\tbranchName := plumbing.NewBranchReferenceName(config.BranchName)\n\tlogger.WithFields(logrus.Fields{\n\t\t\"Branch Name\": branchName,\n\t\t\"Repo\": remoteRepository.GetName(),\n\t}).Debug(\"Created branch\")\n\n\t// Create a branch specific to the multi repo script runner\n\tco := &git.CheckoutOptions{\n\t\tHash: ref.Hash(),\n\t\tBranch: branchName,\n\t\tCreate: true,\n\t}\n\n\t// Attempt to checkout the new tool-specific branch on which the supplied command will be executed\n\tcheckoutErr := worktree.Checkout(co)\n\n\tif checkoutErr != nil {\n\t\tlogger.WithFields(logrus.Fields{\n\t\t\t\"Error\": checkoutErr,\n\t\t\t\"Repo\": remoteRepository.GetName(),\n\t\t}).Debug(\"Error creating new branch\")\n\n\t\t// Track the error checking out the branch\n\t\tconfig.Stats.TrackSingle(BranchCheckoutFailed, remoteRepository)\n\n\t\treturn branchName, errors.WithStackTrace(checkoutErr)\n\t}\n\n\t// Pull latest code from remote branch if it exists to avoid fast-forwarding errors\n\tpo := &git.PullOptions{\n\t\tRemoteName: \"origin\",\n\t\tReferenceName: branchName,\n\t\tAuth: &http.BasicAuth{\n\t\t\tUsername: remoteRepository.GetOwner().GetLogin(),\n\t\t\tPassword: os.Getenv(\"GITHUB_OAUTH_TOKEN\"),\n\t\t},\n\t\tProgress: os.Stdout,\n\t}\n\n\tpullErr := worktree.Pull(po)\n\n\tif pullErr != nil {\n\n\t\tif pullErr == plumbing.ErrReferenceNotFound {\n\t\t\t// The suppled branch just doesn't exist yet on the remote - this is not a fatal error and will\n\t\t\t// allow the new branch to be pushed in pushLocalBranch\n\t\t\tconfig.Stats.TrackSingle(BranchRemoteDidntExistYet, remoteRepository)\n\t\t\treturn branchName, nil\n\t\t}\n\n\t\t// Track the error pulling the latest from the remote branch\n\t\tconfig.Stats.TrackSingle(BranchRemotePullFailed, remoteRepository)\n\n\t\treturn branchName, errors.WithStackTrace(pullErr)\n\t}\n\n\treturn branchName, nil\n}", "func (ref RefName) ForBranchName() string {\n\treturn ref.nameWithoutPrefix(ForPrefix)\n}", "func RenameBranch(repo *models.Repository, doer *user_model.User, gitRepo *git.Repository, from, to string) (string, error) {\n\tif from == to {\n\t\treturn \"target_exist\", nil\n\t}\n\n\tif gitRepo.IsBranchExist(to) {\n\t\treturn \"target_exist\", nil\n\t}\n\n\tif !gitRepo.IsBranchExist(from) {\n\t\treturn \"from_not_exist\", nil\n\t}\n\n\tif err := repo.RenameBranch(from, to, func(isDefault bool) error {\n\t\terr2 := gitRepo.RenameBranch(from, to)\n\t\tif err2 != nil {\n\t\t\treturn err2\n\t\t}\n\n\t\tif isDefault {\n\t\t\terr2 = gitRepo.SetDefaultBranch(to)\n\t\t\tif err2 != nil {\n\t\t\t\treturn err2\n\t\t\t}\n\t\t}\n\n\t\treturn nil\n\t}); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tnotification.NotifyDeleteRef(doer, repo, \"branch\", \"refs/heads/\"+from)\n\tnotification.NotifyCreateRef(doer, repo, \"branch\", \"refs/heads/\"+to)\n\n\treturn \"\", nil\n}", "func validateBranchName(version string) error {\n\tcmd := exec.Command(\"git\", \"branch\", \"--show-current\")\n\tstdout, err := cmd.Output()\n\tif err != nil {\n\t\treturn err\n\t}\n\tbranch := strings.TrimSpace(string(stdout))\n\texpectedBranch := fmt.Sprintf(\"gopls-release-branch.%s\", strings.TrimPrefix(semver.MajorMinor(version), \"v\"))\n\tif branch != expectedBranch {\n\t\treturn fmt.Errorf(\"expected release branch %s, got %s\", expectedBranch, branch)\n\t}\n\treturn nil\n}", "func GenBranchName(text string) string {\n\ttoRemove := regexp.MustCompile(\"(^[[:^alnum:]]+|[[:^alnum:]]$)\")\n\ttoUnderscore := regexp.MustCompile(\"[[:^alnum:]]+\")\n\tbranchName := strings.ToLower(text)\n\tbranchName = toRemove.ReplaceAllString(branchName, \"\")\n\tbranchName = toUnderscore.ReplaceAllString(branchName, \"_\")\n\treturn branchName\n}", "func (s *Service) SetBranchName(v string) *Service {\n\ts.BranchName = &v\n\treturn s\n}", "func GetGitBranch(dir string) (string, error) {\n\treturn runGit(dir, \"rev-parse\", \"--abbrev-ref\", \"HEAD\")\n}", "func getModuleBranch(moduleName string, proj *model.Project) (string, error) {\n\t// find the module of the patch\n\tfor _, module := range proj.Modules {\n\t\tif module.Name == moduleName {\n\t\t\treturn module.Branch, nil\n\t\t}\n\t}\n\treturn \"\", errors.Errorf(\"module '%s' unknown or not found\", moduleName)\n}", "func GetAppBranch() (string, error) {\n\tcmdArgs := []string{\"rev-parse\", \"--abbrev-ref\", \"HEAD\"}\n\tbranchName, err := exec.Command(\"/usr/bin/git\", cmdArgs...).Output()\n\treturn string(branchName), err\n}", "func (t *Tree) Branch(b string) error {\n\tfor name, leaf := range t.packageMap {\n\t\tif leaf.IsRoot() {\n\t\t\tif err := t.branchBetween(name, b); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func getBranch(urlstr string, branch string, dst string) {\n\tlog.Debugf(\"Getting branch %s\", branch)\n\tif sh.DirExists(dst) {\n\t\tlog.Infof(\"Folder exists, skipping cloning %s\", dst)\n\t\tlog.Infof(\"Checking out %s\", branch)\n\t\tif oldPwd := sh.Pwd(); !(oldPwd == dst) {\n\t\t\tsh.Cd(dst)\n\t\t\tsh.SetE(exec.Command(\"git\", \"checkout\", branch))\n\t\t\tsh.Cd(oldPwd)\n\t\t} else {\n\t\t\tsh.SetE(exec.Command(\"git\", \"checkout\", branch))\n\t\t}\n\t} else {\n\t\tlog.Infof(\"Cloning into %s\", dst)\n\t\tcloneCmd := []string{\n\t\t\t// don't verify the ssl certificate (I've run into trouble with it)\n\t\t\t\"-c\", \"http.sslVerify=false\",\n\t\t\t\"clone\", urlstr, dst,\n\t\t\t// only clone this branch, with two commits of history\n\t\t\t\"--branch=\" + branch, \"--single-branch\",\n\t\t\t\"--depth\", \"2\",\n\t\t}\n\t\tsh.SetE(exec.Command(\"git\", cloneCmd...))\n\t}\n\tlog.Debugf(\"Done getting branch %s\", branch)\n}", "func GetBranch() string {\n\tv := Map[\"branch\"]\n\treturn v\n}", "func GenerateBranch() string {\n\treturn \"z9hG4bK-\" + GenerateTag()\n}", "func (c *CreateEvent) GetMasterBranch() string {\n\tif c == nil || c.MasterBranch == nil {\n\t\treturn \"\"\n\t}\n\treturn *c.MasterBranch\n}", "func CurrentBranch() (string, error) {\n\tresult, err := exec.Command(\"git\", \"symbolic-ref\", \"--short\", \"HEAD\").Output()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn strings.TrimSpace(string(result)), nil\n}", "func (o TriggerGithubPushOutput) Branch() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TriggerGithubPush) *string { return v.Branch }).(pulumi.StringPtrOutput)\n}", "func (o TriggerGithubPushPtrOutput) Branch() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TriggerGithubPush) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Branch\n\t}).(pulumi.StringPtrOutput)\n}", "func (v Repository) TrackBranch(branch string) string {\n\tif branch == \"\" {\n\t\tbranch = v.GetHead()\n\t}\n\tif branch == \"\" {\n\t\treturn \"\"\n\t}\n\tbranch = strings.TrimPrefix(branch, config.RefsHeads)\n\n\tcfg := v.Config()\n\treturn strings.TrimPrefix(cfg.Get(\"branch.\"+branch+\".merge\"), config.RefsHeads)\n}", "func (s *CreateServiceInput) SetBranchName(v string) *CreateServiceInput {\n\ts.BranchName = &v\n\treturn s\n}", "func (r ReferenceName) IsBranch() bool {\n\treturn strings.HasPrefix(string(r), refHeadPrefix)\n}", "func GetMatchingBranch(input string) (string, error) {\n\tvar foundBranches []string\n\n\tloweredInput := strings.ToLower(input)\n\tfor _, branch := range GetBranches() {\n\t\tloweredBranch := strings.ToLower(branch)\n\t\tif loweredBranch == loweredInput {\n\t\t\treturn input, nil\n\t\t} else if strings.Contains(loweredBranch, loweredInput) {\n\t\t\tfoundBranches = append(foundBranches, branch)\n\t\t}\n\t}\n\n\tif len(foundBranches) > 1 {\n\t\treturn \"\", fmt.Errorf(\"multiple branches found: %s\", strings.Join(foundBranches, \", \"))\n\t} else if len(foundBranches) == 1 {\n\t\treturn foundBranches[0], nil\n\t}\n\n\tlog.Errorf(\"Branch not found: %s. We have %d known branches\", input, len(branches))\n\n\t// branch not found in local list, but maybe it was created recently -> let's try it if jenkins accept it\n\treturn input, nil\n}", "func (r *Repository) GetMasterBranch() string {\n\tif r == nil || r.MasterBranch == nil {\n\t\treturn \"\"\n\t}\n\treturn *r.MasterBranch\n}", "func (b Branch) Name() string {\n\treturn polName\n}", "func (p *PushEventRepository) GetMasterBranch() string {\n\tif p == nil || p.MasterBranch == nil {\n\t\treturn \"\"\n\t}\n\treturn *p.MasterBranch\n}", "func refFromCtx(ctx context.Context) string {\n\treturn DefaultBranchName\n}", "func Basename(path string) string {\n\treturn filepath.Base(path)\n}", "func createBranch(r *git.Repository, branchName string) error {\n\twt, err := r.Worktree()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = wt.Checkout(&git.CheckoutOptions{\n\t\tBranch: plumbing.NewBranchReferenceName(branchName),\n\t\tCreate: true,\n\t})\n\tif err != nil {\n\t\tif strings.Contains(err.Error(), \"already exists\") {\n\t\t\tlogDebug(\"Reusing branch \" + branchName)\n\t\t\treturn wt.Checkout(&git.CheckoutOptions{\n\t\t\t\tBranch: plumbing.NewBranchReferenceName(branchName),\n\t\t\t\tCreate: false,\n\t\t\t})\n\t\t}\n\t\treturn err\n\t}\n\tlogDebug(\"Creating branch with name \" + branchName)\n\treturn err\n}", "func (novis *Novis) GetBranch(name string) *Branch {\n\treturn novis.Get(name)\n}", "func BranchHeadRoot(ctx context.Context, db *doltdb.DoltDB, brName string) (*doltdb.RootValue, error) {\n\tcs, err := doltdb.NewCommitSpec(brName)\n\tif err != nil {\n\t\treturn nil, doltdb.RootValueUnreadable{RootType: doltdb.HeadRoot, Cause: err}\n\t}\n\n\tcm, err := db.Resolve(ctx, cs, nil)\n\tif err != nil {\n\t\treturn nil, doltdb.RootValueUnreadable{RootType: doltdb.HeadRoot, Cause: err}\n\t}\n\n\tbranchRoot, err := cm.GetRootValue(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn branchRoot, nil\n}", "func (r *RepositoryContentFileOptions) GetBranch() string {\n\tif r == nil || r.Branch == nil {\n\t\treturn \"\"\n\t}\n\treturn *r.Branch\n}", "func CurrentBranchByRoot(ctx context.Context, dEnv *env.DoltEnv, replay ReplayRootFn, nerf NeedsRebaseFn) error {\n\treplayCommit := wrapReplayRootFn(replay)\n\theadRef, err := dEnv.RepoStateReader().CWBHeadRef()\n\tif err != nil {\n\t\treturn nil\n\t}\n\treturn rebaseRefs(ctx, dEnv.DbData(), replayCommit, nerf, headRef)\n}", "func (u GitlabUpstream) BranchURL(repoURL *url.URL, branch string) (string, error) {\n\tif branch != \"master\" {\n\t\trepoURL.Path = repoURL.Path + \"/-/tree/\" + branch\n\t}\n\treturn repoURL.String(), nil\n}", "func Basename(file string) string {\n\treturn path.Base(file)\n}", "func branchPropertiesFromName(name string) (string, string, color.Attribute) {\n\tif strings.Contains(name, \"feature/\") {\n\t\treturn \"feature\", \"develop\", color.FgGreen\n\t} else if strings.Contains(name, \"bugfix/\") {\n\t\treturn \"bugfix\", \"develop\", color.FgYellow\n\t} else if strings.Contains(name, \"hotfix/\") {\n\t\treturn \"hotfix\", \"master\", color.FgRed\n\t}\n\treturn \"other\", name, color.FgWhite\n}", "func (h *branchesService) determineCommitBranch(\n\trepo *Repo, c *Commit, branchesChildren map[string][]string,\n) *Branch {\n\t// At this point, if a commit c has possible branches in c.Branches[], they will all be\n\t// live git branches. However, on return the c.Branches[] may contain deleted or ambiguous\n\t// branches as well\n\n\tif branch := h.hasOnlyOneBranch(c); branch != nil {\n\t\t// Commit only has one branch, it must have been an actual branch tip originally, use that\n\t\treturn branch\n\t} else if branch := h.isLocalRemoteBranch(c); branch != nil {\n\t\t// Commit has only local and its remote branch, prefer remote remote branch\n\t\treturn branch\n\t} else if branch := h.hasParentChildSetBranch(c, branchesChildren); branch != nil {\n\t\t// The commit has several possible branches, and one is set as parent of the others by the user\n\t\treturn branch\n\t} else if branch := h.hasChildrenPriorityBranch(c, branchesChildren); branch != nil {\n\t\t// The commit has several possible branches, and one of the children's branches is set as the\n\t\t// the parent branch of the other children's branches\n\t\treturn branch\n\t} else if branch := h.isSameChildrenBranches(c); branch != nil {\n\t\t// Commit has no branch but has 2 children with same branch\n\t\treturn branch\n\t} else if branch := h.isMergedDeletedRemoteBranchTip(repo, c); branch != nil {\n\t\t// Commit has no branch and no children, but has a merge child, the commit is a tip\n\t\t// of a deleted branch. It might be a deleted remote branch. Lets try determine branch name\n\t\t// based on merge child's subject or use a generic branch name based on commit id\n\t\treturn branch\n\t} else if branch := h.isMergedDeletedBranchTip(repo, c); branch != nil {\n\t\t// Commit has no branch and no children, but has a merge child, the commit is a tip\n\t\t// of a deleted remote branch, lets try determine branch name based on merge child's\n\t\t// subject or use a generic branch name based on commit id\n\t\treturn branch\n\t} else if branch := h.hasOneChildInDeletedBranch(c); branch != nil {\n\t\t// Commit is middle commit in a deleted branch with only one child above, use same branch\n\t\treturn branch\n\t} else if branch := h.hasOneChildWithLikelyBranch(c); branch != nil {\n\t\t// Commit multiple possible git branches but has one child, which has a likely known branch, use same branch\n\t\treturn branch\n\t} else if branch := h.hasMainBranch(c); branch != nil {\n\t\t// Commit, has several possible branches, and one is in the priority list, e.g. main, master, ...\n\t\treturn branch\n\t} else if branch := h.hasBranchNameInSubject(repo, c); branch != nil {\n\t\t// A branch name could be parsed form the commit subject or a child subject.\n\t\t// The commit will be set to that branch and also if above (first child) commits have\n\t\t// ambiguous branches, the will be reset to same branch as well. This will 'repair' branch\n\t\t// when a parsable commit subjects are encountered.\n\t\treturn branch\n\t} else if branch := h.hasOnlyOneChild(c); branch != nil {\n\t\t// Commit has one child commit and not merge commits, reuse that child commit branch\n\t\treturn branch\n\t} else if branch := h.isChildAmbiguousBranch(c); branch != nil {\n\t\t// one of the commit children is a ambiguous branch, reuse same ambiguous branch\n\t\treturn branch\n\t}\n\n\t// Commit, has several possible branches, and we could not determine which branch is best,\n\t// create a new ambiguous branch. Later commits may fix this by parsing subjects of later\n\t// commits, or the user has to manually set the branch.\n\treturn repo.addAmbiguousBranch(c)\n}", "func (r *Repository) GetDefaultBranch() string {\n\tif r == nil || r.DefaultBranch == nil {\n\t\treturn \"\"\n\t}\n\treturn *r.DefaultBranch\n}", "func refFromCtx(ctx context.Context) string {\n\treturn depot.DefaultBranchName\n}", "func CurrentBranch(ctx context.Context, dEnv *env.DoltEnv, replay ReplayCommitFn, nerf NeedsRebaseFn) error {\n\theadRef, err := dEnv.RepoStateReader().CWBHeadRef()\n\tif err != nil {\n\t\treturn nil\n\t}\n\treturn rebaseRefs(ctx, dEnv.DbData(), replay, nerf, headRef)\n}", "func (v Repository) LocalTrackBranch(branch string) string {\n\tif branch == \"\" {\n\t\tbranch = v.GetHead()\n\t}\n\tif branch == \"\" {\n\t\treturn \"\"\n\t}\n\tbranch = strings.TrimPrefix(branch, config.RefsHeads)\n\n\tcfg := v.Config()\n\ttrack := strings.TrimPrefix(cfg.Get(\"branch.\"+branch+\".merge\"), config.RefsHeads)\n\tremote := cfg.Get(\"branch.\" + branch + \".remote\")\n\tif remote == \"\" || track == \"\" {\n\t\treturn \"\"\n\t}\n\treturn v.RemoteMatchingBranch(remote, track)\n}", "func (a *Action) GetBranch() string {\n\treturn strings.TrimPrefix(a.RefName, git.BranchPrefix)\n}", "func branchChangeID(br string) string {\n\tout, err := exec.Command(\"git\", \"show\", br, \"--\").CombinedOutput()\n\tif err != nil {\n\t\tlog.Printf(\"Error running git show %v: %v: %s\", br, err, out)\n\t}\n\tif m := changeRx.FindSubmatch(out); m != nil {\n\t\treturn string(m[1])\n\t}\n\treturn \"\"\n}", "func (g *GitUtil) GetBranch() (string, error) {\n\tref, err := g.Repository.Head()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif !ref.Name().IsBranch() {\n\t\tbranches, err := g.Repository.Branches()\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\tvar currentBranch string\n\t\tfound := branches.ForEach(func(p *plumbing.Reference) error {\n\n\t\t\tif p.Name().IsBranch() && p.Name().Short() != \"origin\" {\n\t\t\t\tcurrentBranch = p.Name().Short()\n\t\t\t\treturn fmt.Errorf(\"break\")\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\n\t\tif found != nil {\n\t\t\tlog.Debugf(\"Found branch from HEAD %s\", currentBranch)\n\t\t\treturn currentBranch, nil\n\t\t}\n\n\t\treturn \"\", fmt.Errorf(\"no branch found, found %s, please checkout a branch (git checkout -b <BRANCH>)\", ref.Name().String())\n\t}\n\tlog.Debugf(\"Found branch %s\", ref.Name().Short())\n\treturn ref.Name().Short(), nil\n}", "func branchGit() {\n\n}", "func GitBranch(tb testing.TB, branch string) {\n\ttb.Helper()\n\tout, err := fakeGit(\"branch\", branch)\n\trequire.NoError(tb, err)\n\trequire.Empty(tb, out)\n}", "func (o RegistryTaskSourceTriggerOutput) Branch() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v RegistryTaskSourceTrigger) *string { return v.Branch }).(pulumi.StringPtrOutput)\n}", "func IsBranch(name string) bool {\n\treturn plumbing.ReferenceName(name).IsBranch()\n}", "func GetBranchSha(branchName string) string {\n\treturn command.New(\"git\", \"rev-parse\", branchName).Output()\n}", "func commitLogFileName(rootPath, name string) string {\n\treturn fmt.Sprintf(\"%s/%s.hnsw.commitlog\", rootPath, name)\n}", "func GetBranch(name string) *Branch {\n\treturn novis.Get(name)\n}", "func (s RepositoryBranch) String() string {\n\treturn awsutil.Prettify(s)\n}", "func BuildBranch(branch string) error {\n\tcurrDir, err := os.Getwd()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err := buildSetup(cloneDir, branch); err != nil {\n\t\treturn err\n\t}\n\treturn buildAndClean(currDir, cloneDir)\n}", "func (r *Repository) Name() string {\n\tre := regexp.MustCompile(\"/([^/]*)\\\\.git$\")\n\tmatch := re.FindStringSubmatch(r.Git)\n\tif len(match) > 0 {\n\t\treturn match[1]\n\t}\n\n\treturn \"\"\n}", "func mergeBranch(c *cli.Context) error {\n\tif !dit.CheckDitFolderPresent() {\n\t\treturn ErrNotInitialized\n\t}\n\tif c.NArg() == 0 {\n\t\treturn ErrIncorrectOperands\n\t}\n\theadHash := dit.ReadHeadBranch(c.Args().First())\n\tmergedCommit, _ := dit.GetCommit(headHash)\n\tcurrentCommit := dit.GetHeadCommit()\n\tsplitCommit := dit.GetLCA(currentCommit.CommitID, mergedCommit.CommitID)\n\tif splitCommit.CommitID == mergedCommit.CommitID {\n\t\tfmt.Println(\"Given branch is an ancestor of the current branch.\")\n\t\treturn nil\n\t}\n\tif splitCommit.CommitID == currentCommit.CommitID {\n\t\tif err := dit.ApplyCommitToWD(currentCommit, mergedCommit); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdit.ResetIndex()\n\t\trefsHead := dit.GetRefsHead()\n\t\tif err := ioutil.WriteFile(path.Join(config.RootPath, refsHead), []byte(mergedCommit.CommitID), config.DefaultFilePerm); err != nil {\n\t\t\treturn errors.Wrap(err, \"Failed to Set HEAD to the new commit\")\n\t\t}\n\t\tfmt.Println(\"Current branch fast-forwared to \", mergedCommit.CommitID)\n\t\treturn nil\n\t}\n\n\tcurrentIndex, err := dit.ReadIndex()\n\tif err != nil {\n\t\tcurrentIndex = &dit.Index{\n\t\t\tStagedFiles: make(map[string]string),\n\t\t\tRemovedFiles: make(map[string]bool),\n\t\t}\n\t}\n\tfileMap := dit.JoinMapString(currentCommit.Blob, mergedCommit.Blob)\n\tisConflict := false\n\tconflictedFiles := []string{}\n\t// first rule\n\tfor fileName := range fileMap {\n\t\tcurrentFileHash, isCurrentTracked := currentCommit.Blob[fileName]\n\t\tsplitFileHash, isSplitTracked := splitCommit.Blob[fileName]\n\t\tmergeFileHash, isMergedTracked := mergedCommit.Blob[fileName]\n\t\tswitch {\n\t\t// File tracked at split point and in current head\n\t\tcase isSplitTracked && isCurrentTracked && isMergedTracked:\n\t\t\tif (splitFileHash != mergeFileHash) && (splitFileHash == currentFileHash) {\n\t\t\t\t// Any files that have been modified in the given branch since the split point,\n\t\t\t\t// but not modified in the current branch since the split point should be changed\n\t\t\t\t// to their versions in the given branch (checked out from the commit at the front of the given branch).\n\t\t\t\t// These files should then all be automatically staged.\n\t\t\t\tblob, _ := dit.GetBlob(mergeFileHash)\n\t\t\t\tif err := dit.OverwriteFileWithBlob(blob, fileName); err != nil {\n\t\t\t\t\treturn errors.Wrapf(err, \"Failed to overwrite %s \\n\", fileName)\n\t\t\t\t}\n\t\t\t\tcurrentIndex.StagedFiles[fileName] = mergeFileHash\n\t\t\t} else if (splitFileHash != mergeFileHash) && (splitFileHash != currentFileHash) && (currentFileHash != mergeFileHash) {\n\t\t\t\t// \"Modified in different ways\" can mean that the contents of both are changed and different from other\n\t\t\t\tblobHead, _ := dit.GetBlob(currentFileHash)\n\t\t\t\tblobMerge, _ := dit.GetBlob(mergeFileHash)\n\t\t\t\tif err := dit.OverwriteFileWithConflictedBlob(blobHead, blobMerge, fileName); err != nil {\n\t\t\t\t\treturn errors.Wrapf(err, \"Failed to overwrite %s \\n\", fileName)\n\t\t\t\t}\n\t\t\t}\n\t\tcase isSplitTracked && isCurrentTracked && !isMergedTracked:\n\t\t\t// Any files present at the split point, unmodified in the current branch branch, and\n\t\t\t// absent in the given branch should be removed (and untracked).\n\t\t\tif splitFileHash == currentFileHash {\n\t\t\t\tcurrentIndex.RemovedFiles[fileName] = true\n\t\t\t\tos.Remove(fileName)\n\t\t\t} else if splitFileHash != currentFileHash {\n\t\t\t\tfmt.Println(fileName, \" Conflict\")\n\t\t\t\tblobHead, _ := dit.GetBlob(currentFileHash)\n\t\t\t\tblobMerge, _ := dit.GetBlob(mergeFileHash)\n\t\t\t\tif err := dit.OverwriteFileWithConflictedBlob(blobHead, blobMerge, fileName); err != nil {\n\t\t\t\t\treturn errors.Wrapf(err, \"Failed to overwrite %s \\n\", fileName)\n\t\t\t\t}\n\t\t\t}\n\t\tcase isSplitTracked && !isCurrentTracked && isMergedTracked:\n\t\t\tif splitFileHash != mergeFileHash {\n\t\t\t\tfmt.Println(fileName, \" Conflict\")\n\t\t\t\tblobHead, _ := dit.GetBlob(currentFileHash)\n\t\t\t\tblobMerge, _ := dit.GetBlob(mergeFileHash)\n\t\t\t\tif err := dit.OverwriteFileWithConflictedBlob(blobHead, blobMerge, fileName); err != nil {\n\t\t\t\t\treturn errors.Wrapf(err, \"Failed to overwrite %s \\n\", fileName)\n\t\t\t\t}\n\t\t\t}\n\t\tcase isSplitTracked && !isCurrentTracked && !isMergedTracked:\n\t\tcase !isSplitTracked && isCurrentTracked && isMergedTracked:\n\t\t\tif currentFileHash != mergeFileHash {\n\t\t\t\tfmt.Println(fileName, \" Conflict\")\n\t\t\t\tblobHead, _ := dit.GetBlob(currentFileHash)\n\t\t\t\tblobMerge, _ := dit.GetBlob(mergeFileHash)\n\t\t\t\tif err := dit.OverwriteFileWithConflictedBlob(blobHead, blobMerge, fileName); err != nil {\n\t\t\t\t\treturn errors.Wrapf(err, \"Failed to overwrite %s \\n\", fileName)\n\t\t\t\t}\n\t\t\t}\n\t\tcase !isSplitTracked && isCurrentTracked && !isMergedTracked:\n\t\tcase !isSplitTracked && !isCurrentTracked && isMergedTracked:\n\t\t\t// Any files that were not present at the split point and are present only\n\t\t\t// in the given branch should be checked out and staged.\n\t\t\tblob, _ := dit.GetBlob(mergeFileHash)\n\t\t\tif err := dit.OverwriteFileWithBlob(blob, fileName); err != nil {\n\t\t\t\treturn errors.Wrapf(err, \"Failed to overwrite %s \\n\", fileName)\n\t\t\t}\n\t\t\tcurrentIndex.StagedFiles[fileName] = mergeFileHash\n\t\tcase !isSplitTracked && !isCurrentTracked && !isMergedTracked:\n\t\t}\n\t}\n\tif isConflict {\n\t\tfmt.Println(\"Encountered a merge conflict.\")\n\t\tfmt.Println(\"Conflicted Files : \", conflictedFiles)\n\t\treturn nil\n\t}\n\t// Create a commit\n\treturn nil\n}", "func (o BranchProtectionOutput) Branch() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *BranchProtection) pulumi.StringOutput { return v.Branch }).(pulumi.StringOutput)\n}", "func TestGitCommandCurrentBranchName(t *testing.T) {\n\ttype scenario struct {\n\t\ttestName string\n\t\tcommand func(string, ...string) *exec.Cmd\n\t\ttest func(string, string, error)\n\t}\n\n\tscenarios := []scenario{\n\t\t{\n\t\t\t\"says we are on the master branch if we are\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.Equal(t, \"git\", cmd)\n\t\t\t\treturn secureexec.Command(\"echo\", \"master\")\n\t\t\t},\n\t\t\tfunc(name string, displayname string, err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t\tassert.EqualValues(t, \"master\", name)\n\t\t\t\tassert.EqualValues(t, \"master\", displayname)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"falls back to git `git branch --contains` if symbolic-ref fails\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\n\t\t\t\tswitch args[0] {\n\t\t\t\tcase \"symbolic-ref\":\n\t\t\t\t\tassert.EqualValues(t, []string{\"symbolic-ref\", \"--short\", \"HEAD\"}, args)\n\t\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t\tcase \"branch\":\n\t\t\t\t\tassert.EqualValues(t, []string{\"branch\", \"--contains\"}, args)\n\t\t\t\t\treturn secureexec.Command(\"echo\", \"* master\")\n\t\t\t\t}\n\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\tfunc(name string, displayname string, err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t\tassert.EqualValues(t, \"master\", name)\n\t\t\t\tassert.EqualValues(t, \"master\", displayname)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"handles a detached head\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.EqualValues(t, \"git\", cmd)\n\n\t\t\t\tswitch args[0] {\n\t\t\t\tcase \"symbolic-ref\":\n\t\t\t\t\tassert.EqualValues(t, []string{\"symbolic-ref\", \"--short\", \"HEAD\"}, args)\n\t\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t\tcase \"branch\":\n\t\t\t\t\tassert.EqualValues(t, []string{\"branch\", \"--contains\"}, args)\n\t\t\t\t\treturn secureexec.Command(\"echo\", \"* (HEAD detached at 123abcd)\")\n\t\t\t\t}\n\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\tfunc(name string, displayname string, err error) {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t\tassert.EqualValues(t, \"123abcd\", name)\n\t\t\t\tassert.EqualValues(t, \"(HEAD detached at 123abcd)\", displayname)\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\t\"bubbles up error if there is one\",\n\t\t\tfunc(cmd string, args ...string) *exec.Cmd {\n\t\t\t\tassert.Equal(t, \"git\", cmd)\n\t\t\t\treturn secureexec.Command(\"test\")\n\t\t\t},\n\t\t\tfunc(name string, displayname string, err error) {\n\t\t\t\tassert.Error(t, err)\n\t\t\t\tassert.EqualValues(t, \"\", name)\n\t\t\t\tassert.EqualValues(t, \"\", displayname)\n\t\t\t},\n\t\t},\n\t}\n\n\tfor _, s := range scenarios {\n\t\tt.Run(s.testName, func(t *testing.T) {\n\t\t\tgitCmd := NewDummyGitCommand()\n\t\t\tgitCmd.OSCommand.Command = s.command\n\t\t\ts.test(gitCmd.CurrentBranchName())\n\t\t})\n\t}\n}", "func packageFilename(pwd, relativePath string) string {\n\tfullPath := filepath.Join(pwd, relativePath)\n\treturn strings.TrimPrefix(strings.TrimPrefix(fullPath, filepath.Join(gopath(), \"src\")), \"/\")\n}", "func NewBranchReferenceName(name string) ReferenceName {\n\treturn ReferenceName(refHeadPrefix + name)\n}", "func dirName(filename string) string {\n\tif !strings.HasSuffix(filename, \"/\") {\n\t\tfilename += \"/\"\n\t}\n\treturn filename\n}", "func (g *GitLab) BranchHead(ctx context.Context, u *model.User, r *model.Repo, branch string) (string, error) {\n\ttoken := common.UserToken(ctx, r, u)\n\tclient, err := newClient(g.url, token, g.SkipVerify)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t_repo, err := g.getProject(ctx, client, r.Owner, r.Name)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tb, _, err := client.Branches.GetBranch(_repo.ID, branch, gitlab.WithContext(ctx))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn b.Commit.ID, nil\n}", "func (v Branch) ShortName() string {\n\treturn strings.TrimPrefix(v.Name, config.RefsHeads)\n}", "func (c *CheckSuite) GetHeadBranch() string {\n\tif c == nil || c.HeadBranch == nil {\n\t\treturn \"\"\n\t}\n\treturn *c.HeadBranch\n}", "func getRepoName(dir string) (string, error) {\n\tr, err := git.PlainOpen(dir)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"error opening git dir %s: %w\", dir, err)\n\t}\n\trm, err := r.Remote(defaultRemote)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"error finding remote %s in git dir %s: %w\", defaultRemote, dir, err)\n\t}\n\n\t// validate remote URL\n\tremoteURL, err := url.Parse(rm.Config().URLs[0])\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"error parsing remote URL: %w\", err)\n\t}\n\ttrimmedRemotePath := strings.TrimSuffix(remoteURL.Path, \"/\")\n\tsplitRemotePath := strings.Split(trimmedRemotePath, \"/\")\n\t// expect path to be /owner/repo\n\tif len(splitRemotePath) != 3 {\n\t\treturn \"\", fmt.Errorf(\"expected owner/repo, got %s\", trimmedRemotePath)\n\t}\n\treturn splitRemotePath[len(splitRemotePath)-1], nil\n}", "func Branch(branch string) GitOptions {\n\treturn func(o *options) error {\n\t\to.branch = branch\n\t\treturn nil\n\t}\n}", "func isMainBranch(br string) bool {\n\treturn br == \"master\" || strings.HasPrefix(br, \"dev.\")\n}", "func (p *Project) GetDefaultBranch() (string, error) {\n\tcmd := exec.Command(\"git\", \"remote\", \"show\", \"origin\")\n\tcmd.Dir = workdir.ProjectDir(p.Name)\n\treader, err := stdoutStderrReader(cmd, nil)\n\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tbuf := new(bytes.Buffer)\n\tbuf.ReadFrom(reader)\n\tstdout := buf.String()\n\tre := regexp.MustCompile(\"HEAD branch: (\\\\S+)\")\n\tgroup := re.FindStringSubmatch(stdout)\n\n\treturn group[1], nil\n}" ]
[ "0.68889207", "0.6779807", "0.67637867", "0.6708616", "0.66863376", "0.6600948", "0.6599172", "0.656859", "0.65478766", "0.6514856", "0.6429341", "0.63615274", "0.6339297", "0.628713", "0.6250935", "0.6203753", "0.61460423", "0.61223817", "0.60922194", "0.60919046", "0.6079676", "0.60549384", "0.5982161", "0.5962964", "0.5946766", "0.5924643", "0.59219044", "0.5902689", "0.58879274", "0.58367586", "0.57895786", "0.5785761", "0.5775609", "0.5719136", "0.56891394", "0.5687745", "0.564978", "0.563795", "0.5632848", "0.5629636", "0.56295073", "0.56246454", "0.5581682", "0.5549915", "0.55417436", "0.5540002", "0.5501593", "0.54947114", "0.54943687", "0.5492059", "0.54674023", "0.5466856", "0.54464006", "0.5420409", "0.54165983", "0.5406866", "0.5405447", "0.53832155", "0.5378979", "0.5363275", "0.5349607", "0.53382355", "0.53282845", "0.53241587", "0.53059524", "0.530512", "0.5291353", "0.52899945", "0.5289596", "0.528023", "0.5252154", "0.5247498", "0.52463794", "0.52351296", "0.5226568", "0.5224793", "0.52168113", "0.5215356", "0.5194293", "0.51683617", "0.5168172", "0.51665545", "0.5155848", "0.51553106", "0.51019365", "0.50959575", "0.50954014", "0.5094241", "0.5090418", "0.5090399", "0.5085443", "0.50778556", "0.5073105", "0.50722057", "0.5045385", "0.50427115", "0.5040441", "0.50375175", "0.502713", "0.5023958" ]
0.75781703
0
Write writes the ascii tree to p.Writer
func (p TreeWriter) Write(nodes []*yaml.RNode) error { switch p.Structure { case TreeStructurePackage: return p.packageStructure(nodes) case TreeStructureGraph: return p.graphStructure(nodes) } // If any resource has an owner reference, default to the graph structure. Otherwise, use package structure. for _, node := range nodes { if owners, _ := node.Pipe(yaml.Lookup("metadata", "ownerReferences")); owners != nil { return p.graphStructure(nodes) } } return p.packageStructure(nodes) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p *GameTree) writeTree(w *bufio.Writer, n TreeNodeIdx, needs bool, nMov int, nMovPerLine int) (err error) {\n\tdefer u(tr(\"writeTree\"))\n\tif needs == true {\n\t\tif nMov > 0 {\n\t\t\terr = w.WriteByte('\\n')\n\t\t\tnMov = 0\n\t\t}\n\t\terr = w.WriteByte('(')\n\t}\n\tif err == nil {\n\t\tif nMov == nMovPerLine {\n\t\t\terr = w.WriteByte('\\n')\n\t\t\tnMov = 0\n\t\t}\n\t\terr = w.WriteByte(';')\n\t\t// write the node\n\t\ttyp := p.treeNodes[n].TNodType\n\t\tswitch typ {\n\t\tcase GameInfoNode:\n\t\t\t// fmt.Println(\"writing GameInfoNode\\n\")\n\t\t\terr = p.writeProperties(w, n, true)\n\t\tcase InteriorNode:\n\t\t\t// fmt.Println(\"writing InteriorNode\\n\")\n\t\t\terr = p.writeProperties(w, n, false)\n\t\tcase BlackMoveNode:\n\t\t\t_, err = w.WriteString(\"B[\")\n\t\t\t_, err = w.Write(SGFCoords(ah.NodeLoc(p.treeNodes[n].propListOrNodeLoc), p.IsFF4()))\n\t\t\terr = w.WriteByte(']')\n\t\t\tnMov += 1\n\t\tcase WhiteMoveNode:\n\t\t\t_, err = w.WriteString(\"W[\")\n\t\t\t_, err = w.Write(SGFCoords(ah.NodeLoc(p.treeNodes[n].propListOrNodeLoc), p.IsFF4()))\n\t\t\terr = w.WriteByte(']')\n\t\t\tnMov += 1\n\t\tdefault:\n\t\t\tfmt.Println(\"*** unsupported TreeNodeType in writeTree\")\n\t\t\terr = errors.New(\"writeTree: unsupported TreeNodeType\" + strconv.FormatInt(int64(typ), 10))\n\t\t\treturn err\n\t\t}\n\t\tif err == nil {\n\t\t\t// write the children\n\t\t\tlastCh := p.treeNodes[n].Children\n\t\t\tif lastCh != nilTreeNodeIdx && err == nil {\n\t\t\t\tch := p.treeNodes[lastCh].NextSib\n\t\t\t\tchNeeds := (lastCh != ch)\n\t\t\t\terr = p.writeTree(w, ch, chNeeds, nMov, nMovPerLine)\n\t\t\t\tfor ch != lastCh && err == nil {\n\t\t\t\t\tch = p.treeNodes[ch].NextSib\n\t\t\t\t\t//\t\t\t\t\tnMov += 1\n\t\t\t\t\terr = p.writeTree(w, ch, chNeeds, nMov, nMovPerLine)\n\t\t\t\t}\n\t\t\t}\n\t\t\tif (err == nil) && (needs == true) {\n\t\t\t\terr = w.WriteByte(')')\n\t\t\t}\n\t\t}\n\t}\n\treturn err\n}", "func (decTree *Tree) WriteTree(filename string) {\n\tfile, err := os.Create(filename)\n\tif err != nil {\n\t\tfmt.Println(\"Error opening output file: \", filename)\n\t\treturn\n\t}\n\n\tcurrNode := decTree\n\tvar treeStack []*Tree\n\n\ttreeLen := 1\n\tfor treeLen != 0 {\n\t\tfile.WriteString(nodeToStr(currNode.Details))\n\n\t\tif currNode.Details.Leaf == false {\n\t\t\ttreeStack = append(treeStack, currNode.Right)\n\t\t\tcurrNode = currNode.Left\n\t\t\ttreeLen++\n\t\t} else {\n\t\t\t//get the length of the tree and set curr to the last element in the list\n\t\t\ttreeLen--\n\n\t\t\tif treeLen > 0 {\n\t\t\t\tcurrNode, treeStack = treeStack[treeLen-1], treeStack[:treeLen-1]\n\t\t\t}\n\t\t}\n\t}\n\n\tfile.Close()\n}", "func WriteTree(writer io.Writer, hierarchy *Hierarchy, includeEmpty bool) {\n\ttree := assembleTree(hierarchy)\n\tkeys := make([]string, len(tree))\n\ti := 0\n\tfor k := range tree {\n\t\tkeys[i] = k\n\t\ti++\n\t}\n\tsort.Strings(keys)\n\tfor _, key := range keys {\n\t\tb := tree[key]\n\t\twriteBranch(writer, b, \"\", hierarchy, includeEmpty)\n\t}\n}", "func (n *Node) WriteTree(padding ...int) string {\n\tvar indent int\n\tif len(padding) == 1 {\n\t\tindent = padding[0]\n\t}\n\n\tvar s string\n\tif n.val != \"\" {\n\t\tfor i := 0; i < indent; i++ {\n\t\t\ts += \" \"\n\t\t}\n\t\ts += n.key + \" : \" + n.val + \"\\n\"\n\t} else {\n\t\tfor i := 0; i < indent; i++ {\n\t\t\ts += \" \"\n\t\t}\n\t\ts += n.key + \" :\" + \"\\n\"\n\t\tfor _, nn := range n.nodes {\n\t\t\ts += nn.WriteTree(indent + 1)\n\t\t}\n\t}\n\treturn s\n}", "func (node *URLNode) WriteTree(writer io.Writer) {\n\tif _, err := writer.Write([]byte(node.GenerateTree())); err != nil {\n\t\tlog.Error(err)\n\t}\n}", "func (t *ASCIITree) PrintTree(w io.Writer) {\n\tancestorPrefix := \"\"\n\tfor _, parent := range t.Ancestors() {\n\t\tif parent.Level() <= 0 {\n\t\t\tcontinue\n\t\t}\n\t\tif parent.Last() {\n\t\t\tancestorPrefix += \" \"\n\t\t} else {\n\t\t\tancestorPrefix += \" │\"\n\t\t}\n\t}\n\n\tmyPrefix := \"\"\n\tmultilinePrefix := \"\"\n\tif t.Level() > 0 {\n\t\tif t.Last() {\n\t\t\tif t.Empty() {\n\t\t\t\tmyPrefix += \" └── \"\n\t\t\t\tmultilinePrefix += \" \"\n\t\t\t} else {\n\t\t\t\tmyPrefix += \" └─┬ \"\n\t\t\t\tmultilinePrefix += \" └─┬ \"\n\t\t\t}\n\t\t} else {\n\t\t\tif t.Empty() {\n\t\t\t\tmyPrefix += \" ├── \"\n\t\t\t\tmultilinePrefix += \" │ \"\n\t\t\t} else {\n\t\t\t\tmyPrefix += \" ├─┬ \"\n\t\t\t\tmultilinePrefix += \" │ │ \"\n\t\t\t}\n\t\t}\n\t}\n\n\tif t.Text != \"\" {\n\t\tlines := strings.Split(t.Text, \"\\n\")\n\t\tfmt.Fprintf(w, \"%s%s%s\\n\", ancestorPrefix, myPrefix, lines[0])\n\t\tfor _, line := range lines[1:] {\n\t\t\tfmt.Fprintf(w, \"%s%s%s\\n\", ancestorPrefix, multilinePrefix, line)\n\t\t}\n\t}\n\n\tfor _, child := range t.children {\n\t\tchild.PrintTree(w)\n\t}\n}", "func (t *Tree) Print(w io.Writer, f IterateFunc, itemSiz int) {\n\n\tfmt.Fprintf(w, \"treeNode-+-Left \\t / Left High\\n\")\n\tfmt.Fprintf(w, \" | \\t = Equal\\n\")\n\tfmt.Fprintf(w, \" +-Right\\t \\\\ Right High\\n\\n\")\n\n\tmaxHeight := t.Height()\n\n\tif f != nil && t.root != nil {\n\t\td := &printData{0, itemSiz, make([]byte, maxHeight), 0, f, w}\n\t\td.printer(t.root)\n\t}\n}", "func writeNode(w io.Writer, fset *token.FileSet, x interface{}) {\n\t// convert trailing tabs into spaces using a tconv filter\n\t// to ensure a good outcome in most browsers (there may still\n\t// be tabs in comments and strings, but converting those into\n\t// the right number of spaces is much harder)\n\t//\n\t// TODO(gri) rethink printer flags - perhaps tconv can be eliminated\n\t// with an another printer mode (which is more efficiently\n\t// implemented in the printer than here with another layer)\n\tmode := printer.TabIndent | printer.UseSpaces\n\terr := (&printer.Config{Mode: mode, Tabwidth: tabwidth}).Fprint(&tconv{output: w}, fset, x)\n\tif err != nil {\n\t\tlog.Print(err)\n\t}\n}", "func Write(root fidlgen.Root, out io.Writer) error {\n\tfor _, e := range Elements(root) {\n\t\tfmt.Fprintf(out, \"%v\\n\", e)\n\t}\n\treturn nil\n}", "func (n *Node) Write(ctx context.Context, w io.Writer) (int, error) {\n\twritten := 0\n\ti := 0\n\tvar err error\n\n\ti, err = fmt.Fprint(w, \"<\", n.Tag)\n\twritten += i\n\tif err != nil {\n\t\treturn written, err\n\t}\n\n\ti, err = n.Attributes.Write(w, \"\")\n\twritten += i\n\tif err != nil {\n\t\treturn written, err\n\t}\n\n\ti, err = fmt.Fprint(w, \">\")\n\twritten += i\n\tif err != nil {\n\t\treturn written, err\n\t}\n\n\ti, err = Write(ctx, w, n.Inner)\n\twritten += i\n\tif err != nil {\n\t\treturn written, err\n\t}\n\n\tif !n.SelfClosing {\n\t\ti, err = fmt.Fprint(w, \"</\", n.Tag, \">\")\n\t\twritten += i\n\t\tif err != nil {\n\t\t\treturn written, err\n\t\t}\n\t}\n\n\treturn written, nil\n}", "func encodeTree(hmt *Tree, finalTree *string) {\n\tif hmt == nil {\n\t\treturn\n\t}\n\t\n\tif hmt.LeftNode == nil && hmt.RightNode == nil{\n\t\t*finalTree += \"1\" + string(hmt.Char)\n\t} else {\n\t\t*finalTree += \"0\"\n\t}\n\tencodeTree(hmt.LeftNode, finalTree)\n\tencodeTree(hmt.RightNode, finalTree) \n}", "func WriteTree(c *git.Client) string {\n\tidx, err := c.GitDir.ReadIndex()\n\tif err != nil {\n\t\treturn err.Error()\n\t}\n\tsha1, err := idx.WriteTree(c)\n\tif err != nil {\n\t\treturn err.Error()\n\t}\n\treturn sha1.String()\n}", "func (bpt *BplusTree) writeTree(printLayout bool) {\n\tdefer glog.Flush()\n\tnode, _ := bpt.fetch(bpt.rootKey)\n\tif node == nil {\n\t\tglog.Errorf(\"failed to fetch root key: %v\", bpt.rootKey)\n\t\treturn\n\t}\n\t// Print tree layout.\n\tif printLayout == true {\n\t\tbpt.writeLayout()\n\t}\n\n\t// Go to the left most leaf node and start printing in order.\n\tfor node != nil {\n\t\tif node.IsLeaf {\n\t\t\tbreak\n\t\t}\n\t\tnode, _ = bpt.fetch(node.Children[0].NodeKey)\n\t\tif node == nil {\n\t\t\tglog.Errorf(\"failed to fetch key: %v\", node.Children[0].NodeKey)\n\t\t\treturn\n\t\t}\n\t}\n\n\tif node == nil {\n\t\tglog.Infof(\"tree is empty\")\n\t\treturn\n\t}\n\n\tindex := 0\n\tfor {\n\t\tglog.Infof(\"leaf node: %d (DK: %v, NK: %v, XK: %v, PK: %v)\\n\",\n\t\t\tindex, node.DataKey, node.NodeKey, node.NextKey, node.PrevKey)\n\t\tfor _, child := range node.Children {\n\t\t\tglog.Infof(\"\\t%v\\n\", child)\n\t\t}\n\n\t\tif node.NextKey.IsNil() {\n\t\t\tbreak\n\t\t}\n\n\t\tif !node.NextKey.IsNil() {\n\t\t\tnextKey := node.NextKey\n\t\t\tnode, _ = bpt.fetch(nextKey)\n\t\t\tif node == nil {\n\t\t\t\tglog.Errorf(\"failed to fetch key: %v\", nextKey)\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tindex++\n\t}\n}", "func (n *Int64Node) Write(depthFirst bool, reverse bool, w io.Writer, indentStr string) error {\n\tstack := make([]*Int64Node, 0, 4)\n\tindent := 0\n\tfn := func(n *Int64Node, evt Event) (err error) {\n\t\tswitch evt {\n\t\tcase EnterEvent:\n\t\t\tstack = append(stack, n)\n\t\t\tindent++\n\t\tcase VisitEvent:\n\t\t\tfor i := 0; i < indent; i++ {\n\t\t\t\tio.WriteString(w, indentStr)\n\t\t\t}\n\t\t\tio.WriteString(w, fmt.Sprintf(\"%v\\n\", n.Value))\n\t\t\t//fallthrough\n\t\tcase BackUpEvent:\n\t\t\tstack = stack[:len(stack)-1]\n\t\t\tindent--\n\t\tdefault:\n\t\t\terr = errors.New(fmt.Sprintf(\"Unknown event: %v\", evt))\n\t\t}\n\t\treturn\n\t}\n\t_ = stack\n\treturn n.Walk(depthFirst, reverse, fn)\n}", "func (t *BPTree) WriteNodes(rwMode RWMode, syncEnable bool, flag int) error {\n\tvar (\n\t\tn *Node\n\t\ti int\n\t\terr error\n\t)\n\n\tfd, err := os.OpenFile(t.Filepath, os.O_CREATE|os.O_RDWR, 0644)\n\tdefer fd.Close()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tqueue = nil\n\n\tenqueue(t.root)\n\n\tfor queue != nil {\n\t\tn = dequeue()\n\n\t\t_, err := t.WriteNode(n, -1, syncEnable, fd)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif n != nil {\n\t\t\tif !n.isLeaf {\n\t\t\t\tfor i = 0; i <= n.KeysNum; i++ {\n\t\t\t\t\tc, _ := n.pointers[i].(*Node)\n\t\t\t\t\tenqueue(c)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func writeNode(w io.Writer, node interface{}, html bool, styler printer.Styler) {\n\tmode := printer.UseSpaces;\n\tif html {\n\t\tmode |= printer.GenHTML\n\t}\n\t(&printer.Config{mode, *tabwidth, styler}).Fprint(w, node);\n}", "func (this *Codec) serialize(root *TreeNode) string {\n if root == nil {\n return \"x\"\n }\n return strconv.Itoa(root.Val) + \",\" + this.serialize(root.Left)+ \",\" + this.serialize(root.Right)\n}", "func (t *DiskTree) Print(w io.Writer, n *Node, prefix string) {\n\tdn := t.dnodeFromNode(n)\n\tif len(dn.Children) == 0 {\n\t\tfmt.Fprintf(w, \"%s %s\\n\", Decode(prefix), n.Value)\n\t} else {\n\t\tfor _, c := range dn.Children {\n\t\t\tcnode := t.dnodeFromHash(c)\n\t\t\tt.Print(w, cnode.toMem(), prefix+cnode.Edgename)\n\t\t}\n\t\tif len(n.Value) != 0 {\n\t\t\tfmt.Fprintf(w, \"%s %s\\n\", Decode(prefix), n.Value)\n\t\t}\n\t}\n}", "func WriteASCII(w io.Writer, t []Triangle) error {\n\tvar err error\n\n\tprintf := func(format string, a ...interface{}) {\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\t_, err = fmt.Fprintf(w, format, a...)\n\t}\n\tprintf(\"solid object\\n\")\n\tfor _, tt := range t {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tprintf(\"facet normal %f %f %f\\n\", tt.N[0], tt.N[1], tt.N[2])\n\t\tprintf(\" outer loop\\n\")\n\t\tfor _, v := range tt.V {\n\t\t\tprintf(\" vertex %f %f %f\\n\", v[0], v[1], v[2])\n\t\t}\n\t\tprintf(\" endloop\\n\")\n\t\tprintf(\"endfacet\\n\")\n\t}\n\tprintf(\"endsolid object\\n\")\n\treturn nil\n}", "func (g Index) WriteTree(c *Client) (TreeID, error) {\n\tsha1, err := writeIndexEntries(c, \"\", g.Objects)\n\tif err != nil && err != ObjectExists {\n\t\treturn TreeID{}, err\n\t}\n\treturn sha1, nil\n}", "func (fmt *FixedMerkleTree) Write(b []byte) (int, error) {\n\n\tfmt.writeLock.Lock()\n\tdefer fmt.writeLock.Unlock()\n\tif fmt.isFinal {\n\t\treturn 0, goError.New(\"cannot write. Tree is already finalized\")\n\t}\n\n\tfor i, j := 0, MaxMerkleLeavesSize-fmt.writeCount; i < len(b); i, j = j, j+MaxMerkleLeavesSize {\n\t\tif j > len(b) {\n\t\t\tj = len(b)\n\t\t}\n\t\tprevWriteCount := fmt.writeCount\n\t\tfmt.writeCount += int(j - i)\n\t\tcopy(fmt.writeBytes[prevWriteCount:fmt.writeCount], b[i:j])\n\n\t\tif fmt.writeCount == MaxMerkleLeavesSize {\n\t\t\t// data fragment reached 64KB, so send this slice to write to leaf hashes\n\t\t\terr := fmt.writeToLeaves(fmt.writeBytes)\n\t\t\tif err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t\tfmt.writeCount = 0 // reset writeCount\n\t\t}\n\t}\n\treturn len(b), nil\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tif root == nil {\n\t\treturn \"\"\n\t}\n\tvar res string\n\tp := root\n\tq := new(Queue)\n\tq.EnQueue(p)\n\n\tfor !q.IsEmpty() {\n\t\tnode := q.DeQueue()\n\t\tswitch t := node.(type) {\n\t\tcase nil:\n\t\t\tres += \" \"\n\t\tcase *TreeNode:\n\t\t\tres += strconv.Itoa(t.Val)\n\t\t\tq.EnQueue(t.Left)\n\t\t\tq.EnQueue(t.Right)\n\t\tdefault:\n\t\t\tbreak\n\t\t}\n\t}\n\tfmt.Print(res)\n\treturn res\n}", "func (n *nodeReaderWriter) Write(p []byte) (int, error) {\n\n\t// guarantee that the path exists\n\t_, ok := n.repo.Data[n.path]\n\tif !ok {\n\t\tn.repo.Data[n.path] = []byte{}\n\t}\n\n\t// overwrite the file if we haven't already started writing to it\n\tif !n.writing {\n\t\tn.repo.Data[n.path] = make([]byte, 0)\n\t\tn.writing = true\n\t}\n\n\t// copy the data into the node buffer\n\tcount := 0\n\tstart := n.writeCursor\n\tfor ; n.writeCursor < start+len(p); n.writeCursor++ {\n\t\t// extend the file if needed\n\t\tif len(n.repo.Data) < n.writeCursor+len(p) {\n\t\t\tn.repo.Data[n.path] = append(n.repo.Data[n.path], 0)\n\t\t}\n\t\tn.repo.Data[n.path][n.writeCursor] = p[n.writeCursor-start]\n\t\tcount++\n\t}\n\n\treturn count, nil\n}", "func Encode(node ipld.Node, w io.Writer) error {\n\t// 1KiB can be allocated on the stack, and covers most small nodes\n\t// without having to grow the buffer and cause allocations.\n\tenc := make([]byte, 0, 1024)\n\n\tenc, err := AppendEncode(enc, node)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = w.Write(enc)\n\treturn err\n}", "func (t *TreeStorage) Write(ctx context.Context, nodes []storage.Node) error {\n\tms := make([]*spanner.Mutation, 0, len(nodes))\n\tfor _, node := range nodes {\n\t\t// TODO(pavelkalinnikov): Consider doing just Insert when it is clear what\n\t\t// semantic the callers need.\n\t\tms = append(ms, spanner.InsertOrUpdate(\"TreeNodes\",\n\t\t\t[]string{\"TreeID\", \"ShardID\", \"NodeID\", \"NodeHash\"},\n\t\t\t[]interface{}{t.id, t.opts.shardID(node.ID), packNodeID(node.ID), node.Hash}))\n\t}\n\t_, err := t.c.Apply(ctx, ms)\n\treturn err\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tstrs := make([]string, 0)\n\tvar preorder func(*TreeNode)\n\tpreorder = func(node *TreeNode) {\n\t\tif node == nil {\n\t\t\tstrs = append(strs, this.null)\n\t\t\treturn\n\t\t}\n\n\t\tstrs = append(strs, strconv.Itoa(node.Val))\n\t\tpreorder(node.Left)\n\t\tpreorder(node.Right)\n\t}\n\tpreorder(root)\n\treturn strings.Join(strs, this.sep)\n}", "func WriteASCII(w io.Writer, t []Triangle) error {\n\tbw := bufio.NewWriterSize(w, writeBufSize)\n\tvar err error\n\n\tprintf := func(format string, a ...interface{}) {\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\t_, err = fmt.Fprintf(bw, format, a...)\n\t}\n\tprintf(\"solid object\\n\")\n\tfor _, tt := range t {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tprintf(\"facet normal %f %f %f\\n\", tt.N[0], tt.N[1], tt.N[2])\n\t\tprintf(\" outer loop\\n\")\n\t\tfor _, v := range tt.V {\n\t\t\tprintf(\" vertex %f %f %f\\n\", v[0], v[1], v[2])\n\t\t}\n\t\tprintf(\" endloop\\n\")\n\t\tprintf(\"endfacet\\n\")\n\t}\n\tprintf(\"endsolid object\\n\")\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn bw.Flush()\n}", "func stringify(n *Node, level int) {\n\tif n != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"---[ \"\n\t\tlevel++\n\t\tstringify(n.left, level)\n\t\tfmt.Printf(format+\"%d\\n\", n.key)\n\t\tstringify(n.right, level)\n\t}\n}", "func Write(node *parser.Node) []byte {\n\tif node == nil {\n\t\treturn nil\n\t}\n\tbuf := &bytes.Buffer{}\n\tif len(node.Value) > 0 {\n\t\tbuf.Write([]byte(strings.ToUpper(node.Value)))\n\t\tfor _, flag := range node.Flags {\n\t\t\tbuf.Write([]byte(\" \"))\n\t\t\tbuf.Write([]byte(flag))\n\t\t}\n\t\tswitch node.Value {\n\t\tcase command.Onbuild:\n\t\t\tif node.Next != nil && len(node.Next.Children) > 0 {\n\t\t\t\tbuf.Write([]byte(\" \"))\n\t\t\t\tbuf.Write(Write(node.Next.Children[0]))\n\t\t\t}\n\t\t\treturn buf.Bytes()\n\t\tcase command.Env, command.Label:\n\t\t\tfor n := node.Next; n != nil; n = n.Next {\n\t\t\t\tif buf.Len() > 0 {\n\t\t\t\t\tbuf.Write([]byte(\" \"))\n\t\t\t\t}\n\t\t\t\tbuf.Write([]byte(n.Value))\n\t\t\t\tbuf.Write([]byte(\"=\"))\n\t\t\t\tif n.Next != nil {\n\t\t\t\t\tbuf.Write([]byte(n.Next.Value))\n\t\t\t\t}\n\t\t\t\tn = n.Next\n\t\t\t}\n\t\t\tbuf.Write([]byte(\"\\n\"))\n\t\t\treturn buf.Bytes()\n\t\tdefault:\n\t\t\tif node.Attributes[\"json\"] {\n\t\t\t\tvar values []string\n\t\t\t\tfor n := node.Next; n != nil; n = n.Next {\n\t\t\t\t\tvalues = append(values, n.Value)\n\t\t\t\t}\n\t\t\t\tout, _ := json.Marshal(values)\n\t\t\t\tbuf.Write([]byte(\" \"))\n\t\t\t\tbuf.Write(out)\n\t\t\t\tbuf.Write([]byte(\"\\n\"))\n\t\t\t\treturn buf.Bytes()\n\t\t\t}\n\t\t\tfor n := node.Next; n != nil; n = n.Next {\n\t\t\t\tif buf.Len() > 0 {\n\t\t\t\t\tbuf.Write([]byte(\" \"))\n\t\t\t\t}\n\t\t\t\tbuf.Write([]byte(n.Value))\n\t\t\t}\n\t\t\tbuf.Write([]byte(\"\\n\"))\n\t\t\treturn buf.Bytes()\n\t\t}\n\t}\n\tfor _, child := range node.Children {\n\t\tbuf.Write(Write(child))\n\t}\n\treturn buf.Bytes()\n}", "func (e EmptyNode) EncodeBinary(*io.BinWriter) {\n}", "func (tree *GameTree) WriteFile(fileName string, nMovPerLine int) (err error) {\n\tdefer u(tr(\"WriteFile\"))\n\t// old parms to Open(fileName, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, filePERM)\n\tf, err := os.Create(fileName)\n\tif err != nil {\n\t\treturn errors.New(\"OpenFile:\" + fileName + \" \" + err.Error())\n\t}\n\tdefer f.Close() // TODO: should this be conditional on not being closed?\n\tw := bufio.NewWriter(f)\n\tif w == nil {\n\t\treturn errors.New(\"nil from NewWriter:\" + fileName + \" \" + err.Error())\n\t}\n\terr = tree.writeParseTree(w, nMovPerLine)\n\tif err != nil {\n\t\treturn errors.New(\"Error:\" + fileName + \" \" + err.Error())\n\t}\n\terr = w.Flush()\n\terr = f.Close()\n\treturn err\n}", "func (t *BinaryNode) PrintAscii() {\n\tfmt.Println(t.AsciiBuilder())\n}", "func (e *Escpos) WriteNode(name string, params map[string]string, data string) {\n\tcstr := \"\"\n\tif data != \"\" {\n\t\tstr := data[:]\n\t\tif len(data) > 40 {\n\t\t\tstr = fmt.Sprintf(\"%s ...\", data[0:40])\n\t\t}\n\t\tcstr = fmt.Sprintf(\" => '%s'\", str)\n\t}\n\n\tif e.Verbose {\n\t\tlog.Println(\"Write: %s => %+v%s\\n\", name, params, cstr)\n\t}\n\n\tswitch name {\n\tcase \"text\":\n\t\te.Text(params, data)\n\tcase \"feed\":\n\t\te.Feed(params)\n\tcase \"cut\":\n\t\te.FeedAndCut(params)\n\tcase \"pulse\":\n\t\te.Pulse()\n\tcase \"image\":\n\t\te.Image(params, data)\n\t}\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\treturn this.rserialize(root, \"\")\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tif root == nil {\n\t\treturn \"nil\"\n\t}\n\treturn strconv.Itoa(root.Val) + \",\" + this.serialize(root.Left) + \",\" + this.serialize(root.Right)\n}", "func (t *ASCIITree) String() string {\n\tvar buffer bytes.Buffer\n\tt.PrintTree(&buffer)\n\treturn string(buffer.Bytes())\n}", "func (m *Manager) Write(p []byte) (n int, err error) {\n\tm.lock.Lock()\n\tdefer m.lock.Unlock()\n\n\tif len(p) > 0 {\n\t\terr = m.clear()\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t\tm.dangle = p[len(p)-1] != '\\n'\n\t\tn, err = os.Stdout.Write(p)\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t\terr = m.redraw()\n\t}\n\n\treturn n, err\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\ts := \"\"\n\n\tif root == nil {\n\t\treturn s\n\t}\n\n\ts = s + strconv.Itoa(root.Val)\n\n\tif root.Left != nil {\n\t\ts = s + \",\" + this.serialize(root.Left)\n\t} else {\n\t\ts = s + \",\" + \"null\"\n\t}\n\n\tif root.Right != nil {\n\t\ts = s + \",\" + this.serialize(root.Right)\n\t} else {\n\t\ts = s + \",\" + \"null\"\n\t}\n\n\treturn s\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tvar res string\n\n\tvar dfs func(node *TreeNode)\n\tdfs = func(node *TreeNode) {\n\t\tif node == nil {\n\t\t\tres += \"null,\"\n\t\t\treturn\n\t\t}\n\n\t\tres += fmt.Sprintf(\"%+v\", node.Val) + \",\"\n\t\tdfs(node.Left)\n\t\tdfs(node.Right)\n\t}\n\n\tdfs(root)\n\n\treturn res\n}", "func (n *ReflectNode) Write(ctx context.Context, w io.Writer) (int, error) {\n\twritten := 0\n\ti := 0\n\tvar err error\n\n\ti, err = fmt.Fprint(w, \"<\", n.Tag)\n\twritten += i\n\tif err != nil {\n\t\treturn written, err\n\t}\n\n\ti, err = n.writeAttributes(w)\n\twritten += i\n\tif err != nil {\n\t\treturn written, err\n\t}\n\n\ti, err = fmt.Fprint(w, \">\")\n\twritten += i\n\tif err != nil {\n\t\treturn written, err\n\t}\n\n\ti, err = n.writeInner(ctx, w)\n\twritten += i\n\tif err != nil {\n\t\treturn written, err\n\t}\n\n\tif !n.SelfClosing {\n\t\ti, err = fmt.Fprint(w, \"</\", n.Tag, \">\")\n\t\twritten += i\n\t\tif err != nil {\n\t\t\treturn written, err\n\t\t}\n\t}\n\n\treturn written, nil\n}", "func (tree *BinarySearchTree) String() {\n\ttree.lock.Lock()\n\tdefer tree.lock.Unlock()\n\tfmt.Println(\"************************************************\")\n\tstringify(tree.rootNode, 0)\n\tfmt.Println(\"************************************************\")\n}", "func (t *Trie) WriteToBytes(writer io.Writer) error {\n\treturn gob.NewEncoder(writer).Encode(t)\n}", "func writeTeam(codec Codec, t Team, dir string) error {\n\tt.Children = nil\n\n\tbuf, err := codec.Encode(&t)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := ioutil.WriteFile(filepath.Join(dir, teamFile), buf, 0644); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (bpt *BplusTree) writeLayout() {\n\tleafIdx := 0\n\tnodeIdx := 0\n\tlevelIdx := 0\n\n\tif !bpt.initialized || bpt.rootKey.IsNil() {\n\t\treturn\n\t}\n\n\trootNode, _ := bpt.fetch(bpt.rootKey)\n\tif rootNode == nil {\n\t\tglog.Errorf(\"failed to fetch root key: %v. can not print the tree.\",\n\t\t\tbpt.rootKey)\n\t\treturn\n\t}\n\tglog.Infof(\"dumping the tree layout.. numChildren: %d\\n\",\n\t\tlen(rootNode.Children))\n\tnodeList := rootNode.Children\n\tnodeLensList := make([]int, 1)\n\tnodeLensList[0] = len(rootNode.Children)\n\tnumElems := nodeLensList[0]\n\tnumNodesAtLevel := 0\n\tprintLevel := true\n\tglog.Infof(\"level -- 0 <root: %v>\\n\", rootNode)\n\tif rootNode.IsLeaf {\n\t\treturn\n\t}\n\tfor i := 0; i < numElems; i++ {\n\t\tif printLevel {\n\t\t\tglog.Infof(\"level -- %d \", levelIdx+1)\n\t\t\tprintLevel = false\n\t\t}\n\t\tnode, _ := bpt.fetch(nodeList[i].NodeKey)\n\t\tif node == nil {\n\t\t\tglog.Errorf(\"failed to fetch root key: %v\", nodeList[i].NodeKey)\n\t\t\treturn\n\t\t}\n\n\t\tif node.IsLeaf {\n\t\t\tglog.Infof(\"level:%d <tree-L-node :%d, node: %v> \", levelIdx+1, leafIdx, node)\n\t\t\tleafIdx++\n\t\t} else {\n\t\t\tglog.Infof(\"level:%d <tree-I-node :%d, node: %v> \", levelIdx+1, nodeIdx, node)\n\t\t\tnodeList = append(nodeList, node.Children...)\n\t\t\tnumElems += len(node.Children)\n\t\t\tnumNodesAtLevel += len(node.Children)\n\t\t}\n\t\tnodeIdx++\n\t\tif nodeIdx >= nodeLensList[levelIdx] {\n\t\t\tnodeLensList = append(nodeLensList, numNodesAtLevel)\n\t\t\tlevelIdx++\n\t\t\tnodeIdx = 0\n\t\t\tnumNodesAtLevel = 0\n\t\t\tglog.Infof(\"\\n\")\n\t\t\tprintLevel = true\n\t\t}\n\t}\n\tglog.Infof(\"done.. dumping the layout\\n\")\n\tglog.Infof(\"----------------------------\\n\")\n}", "func (this *Codec) serialize(root *TreeNode) string {\n var s string \n res:=helpSerialize(root,s)\n fmt.Println(res)\n return res\n}", "func (c *Codec) serialize(root *TreeNode) string {\n\treturn rserialize(root, \"\")\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tthis.s(root)\n\treturn \"[\" + strings.Join(this.data, \",\") + \"]\"\n}", "func (node *Node) printTree1(out *bytes.Buffer, isRight bool, indent string) {\n\n\tif (node.Left != nil) {\n\t\tstr := \" \"\n\t\tif isRight {\n\t\t\tstr = \" | \"\n\t\t}\n\t\tstr = indent + str\n\t\tnode.Left.printTree1(out, false, str)\n\t}\n\n\tout.Write([]byte(indent))\n\tif (isRight) {\n\t\tout.Write([]byte(\"\\\\\"))\n\t} else {\n\t\tout.Write([]byte (\"/\"))\n\t}\n\tout.Write([]byte(\"--\"))\n\n\tnode.printNodeValue(out)\n\n\tif (node.Right != nil) {\n\t\tstr := \" | \"\n\t\tif isRight {\n\t\t\tstr = \" \"\n\t\t}\n\t\tstr = indent + str\n\t\tnode.Right.printTree1(out, true, str)\n\t}\n\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\n\tif root == nil {\n\t\treturn \"[null]\"\n\t}\n\n\tnodes := []*TreeNode{root}\n\tcode := \"[\" + strconv.Itoa(root.Val)\n\n\tfor len(nodes) > 0 {\n\t\tnewNodes := []*TreeNode{}\n\n\t\tfor i := range nodes {\n\n\t\t\tif nodes[i].Left != nil {\n\t\t\t\tnewNodes = append(newNodes, nodes[i].Left)\n\t\t\t\tcode = code + \",\" + strconv.Itoa(nodes[i].Left.Val)\n\t\t\t} else {\n\t\t\t\tcode = code + \",null\"\n\t\t\t}\n\n\t\t\tif nodes[i].Right != nil {\n\t\t\t\tnewNodes = append(newNodes, nodes[i].Right)\n\t\t\t\tcode = code + \",\" + strconv.Itoa(nodes[i].Right.Val)\n\t\t\t} else {\n\t\t\t\tcode = code + \",null\"\n\t\t\t}\n\n\t\t}\n\n\t\tnodes = newNodes\n\n\t}\n\tcode = code + \"]\"\n\treturn code\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\treturn dfsSerial(root, \"\")\n}", "func (eln *EmptyLeafNode) Serialize(w io.Writer) {\n\tw.Write([]byte{byte(NodeTypeEmptyLeaf)})\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\ttmp := []string{}\n\ts(root, &tmp)\n\tthis.SerializeStr = strings.Join(tmp, \",\")\n\treturn this.SerializeStr\n}", "func (n *Node) print() string {\n\n\th := n.height()\n\tprintMap = make(map[int][]int, h)\n\tfor i := 1; i <= h; i++ {\n\t\tn.printByLevel(i)\n\t}\n\tfor key := h; key > 0; key-- {\n\t\tfor j := h; j > key; j-- {\n\t\t\tfor _, k := range printMap[j] {\n\t\t\t\tif arrayutils.InInt(printMap[key], k) {\n\t\t\t\t\tprintMap[key] = arrayutils.RemoveByValue[int](printMap[key], k)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\ts := fmt.Sprintf(\"Tree: %+v\", printMap)\n\tprintMap = nil\n\n\treturn s\n}", "func WriteNode(root interface{}, path util.Path, value interface{}) error {\n\tpc, _, err := getPathContext(&PathContext{Node: root}, path, path, true)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn WritePathContext(pc, value, false)\n}", "func (t *treeNode) print(height int) {\n\tif t == nil {\n\t\treturn\n\t}\n\tformat := \"--[\"\n\tt.Right.print(height + 1)\n\tfmt.Printf(\"%*s%d\\n\", 7*(height+1), format, t.Value)\n\tt.Left.print(height + 1)\n}", "func stringify(treeNode *TreeNode, level int) {\n\tif treeNode != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"***> \"\n\t\tlevel++\n\t\tstringify(treeNode.leftNode, level)\n\t\tfmt.Printf(format+\"%d\\n\", treeNode.key)\n\t\tstringify(treeNode.rightNode, level)\n\t}\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tif root == nil {\n\t\treturn \"\"\n\t}\n\n\tres := make([]string, 0)\n\tqueue := make([]*TreeNode, 0)\n\tqueue = append(queue, root)\n\tfor len(queue) > 0 {\n\t\tl := len(queue)\n\t\tfor i := 0; i < l; i++ {\n\t\t\tnode := queue[0]\n\t\t\tqueue = queue[1:]\n\t\t\tif node != nil {\n\t\t\t\tres = append(res, strconv.Itoa(node.Val))\n\t\t\t\tqueue = append(queue, node.Left, node.Right)\n\t\t\t} else {\n\t\t\t\tres = append(res, \"#\")\n\t\t\t}\n\n\t\t}\n\t}\n\treturn strings.Join(res, \",\")\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tif root == nil {\n\t\treturn \"\"\n\t}\n\tans := make([]string, 0, 10)\n\tserialize(root, &ans)\n\n\treturn strings.Join(ans, \",\")\n}", "func (t *BinaryNode) AsciiBuilder() string {\n\tvar proot *AsciiNode\n\tvar xmin int\n\tif t == nil {\n\t\treturn AsciiBuilder.String()\n\t}\n\n\tproot = BuildAsciiTree(t)\n\tFillEdgeLen(proot)\n\n\tfor i := 0; i < proot.Height; i++ {\n\t\tLprofile[i] = math.MaxInt32\n\t}\n\tComuputeLprofile(proot, 0, 0)\n\txmin = 0\n\tfor i := 0; i < proot.Height; i++ {\n\t\txmin = Min(xmin, Lprofile[i])\n\t}\n\n\tif proot.Height > MaxHeight {\n\t\tfmt.Printf(\"The tree is too fucking high than %d! how high do you want to be?\\n\",\n\t\t\tMaxHeight)\n\t}\n\tfor i := 0; i < proot.Height; i++ {\n\t\tPrintNext = 0\n\t\tPrintLevel(proot, -xmin, i)\n\t\tfmt.Fprintf(&AsciiBuilder, \"%s\", \"\\n\")\n\t}\n\ts := AsciiBuilder.String()\n\tAsciiBuilder.Reset()\n\treturn s\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tnodes := []*TreeNode{ root }\n\tstrList := []string{}\n\tfor i := 0; i < len(nodes); {\n\t\tcnt := len(nodes)\n\t\tfor ; i < cnt; i++ {\n\t\t\tnode := nodes[i]\n\t\t\tif node == nil {\n\t\t\t\tstrList = append(strList, \"null\")\n\t\t\t} else {\n\t\t\t\tstrList = append(strList, strconv.Itoa(node.Val))\n\t\t\t\tnodes = append(nodes, node.Left)\n\t\t\t\tnodes = append(nodes, node.Right)\n\t\t\t}\n\t\t}\n\t}\n\treturn \"[\" + strings.Join(strList, \"\") + \"]\"\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tnodeValues := []int{}\n\t// preoder 노드 탐색\n\t_to_string(root, &nodeValues)\n\n\t// 노드 값을 공백으로 구분한 스트링으로 리턴\n\tr := \"\"\n\tfor i := 0; i < len(nodeValues); i++ {\n\t\tr += fmt.Sprintf(\"%d \", nodeValues[i])\n\t}\n\treturn strings.TrimSpace(r)\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tstrList := make([]string, 0)\n\tvar solve func(root *TreeNode)\n\tsolve = func(root *TreeNode) {\n\t\tif root == nil {\n\t\t\tstrList = append(strList, \"$\")\n\t\t\treturn\n\t\t}\n\t\tstrList = append(strList, strconv.Itoa(root.Val))\n\t\tsolve(root.Left)\n\t\tsolve(root.Right)\n\t}\n\tsolve(root)\n\tfmt.Println(strings.Join(strList, \",\"))\n\treturn strings.Join(strList, \",\")\n}", "func (f *Forest) dump(t *testing.T) {\n\tt.Logf(\"---------------- TRIE BEGIN ------------------\")\n\tchildNodes := make(map[string]*Node, len(f.nodes))\n\tfor _, n := range f.nodes {\n\t\tfor _, childHash := range n.branches {\n\t\t\tif len(childHash) != 0 {\n\t\t\t\tchildNodes[childHash.KeyForMap()] = f.nodes[childHash.KeyForMap()]\n\t\t\t}\n\t\t}\n\t}\n\n\tfor nodeHash := range f.nodes {\n\t\tif _, isChild := childNodes[nodeHash]; !isChild {\n\t\t\tf.nodes[nodeHash].printNode(\" Ω\", 0, f, t)\n\t\t}\n\t}\n\tt.Logf(\"---------------- TRIE END --------------------\")\n}", "func (enc *HuffmanEncoder) WriteHeader() error {\n\t// for iterative tree walking use savedict\n\t// for recursive, use rsavedict\n\n\t// if err := savedict(enc.bw, enc.root); err != nil {\n\tif err := rsavedict(enc.bw, enc.root); err != nil {\t\t// recursive version\n\t\treturn err\n\t}\n\treturn enc.bw.WriteBit(bs.Zero) // end of dictionary indicator\n}", "func (n *Node) Print(out *os.File) {\n\tn.printLevel(out, 0)\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\n\tif root == nil {\n\t\treturn \"\"\n\t}\n\n\tqueue := []*TreeNode{root}\n\tc := []string{strconv.Itoa(root.Val)}\n\n\tfor len(queue) > 0 {\n\t\tl := len(queue)\n\t\tfor i := 0; i < l; i++ {\n\t\t\tif queue[i].Left != nil {\n\t\t\t\tqueue = append(queue, queue[i].Left)\n\t\t\t}\n\t\t\tif queue[i].Right != nil {\n\t\t\t\tqueue = append(queue, queue[i].Right)\n\t\t\t}\n\t\t\tadd(&c, queue[i].Left)\n\t\t\tadd(&c, queue[i].Right)\n\t\t}\n\t\tqueue = queue[l:]\n\t}\n\n\tres := strings.Join(c, \",\")\n\treturn res\n}", "func Write(obj any, writer io.Writer) error {\n\tenc := toml.NewEncoder(writer)\n\treturn enc.Encode(obj)\n}", "func (head *Node) WriteHeader(w *bitio.Writer, freq map[uint8]uint) (err error) {\n\tvar nEncoded uint32\n\tfor _, v := range freq {\n\t\tnEncoded += uint32(v)\n\t}\n\n\t// Write total number of encoded symbols\n\tw.TryWriteBitsUnsafe(uint64(nEncoded), 32)\n\n\t// Write total number of symbols in graph\n\tw.TryWriteBitsUnsafe(uint64(len(freq)), 8)\n\n\t// Write encoding tree information\n\tif err = head.writeHeader(w); err != nil {\n\t\treturn err\n\t}\n\tw.TryWriteBitsUnsafe(0, 1)\n\treturn w.TryError\n}", "func Encode(input string) (string, *tree.Node[string]) {\n\t// Create huffman tree and map\n\thTree := getHuffmanTree(input)\n\tprintTree(hTree)\n\thMap := getHuffmanEncodingMap(hTree)\n\tbuilder := strings.Builder{}\n\tfor i := 0; i < len(input); i++ {\n\t\tbuilder.WriteString(hMap[string(input[i])])\n\t}\n\treturn builder.String(), hTree\n}", "func (t *Tree) Print() {\n\tvar findMaxLevel func(*Node) int\n\tfindMaxLevel = func(r *Node) int {\n\t\tif r == nil {\n\t\t\treturn 0\n\t\t}\n\n\t\tleftLevel := findMaxLevel(r.left)\n\t\trightLevel := findMaxLevel(r.right)\n\n\t\tif leftLevel > rightLevel {\n\t\t\treturn leftLevel + 1\n\t\t} else {\n\t\t\treturn rightLevel + 1\n\t\t}\n\t}\n\n\tprintSpace := func(n int) {\n\t\tfor i := 0; i < n; i++ {\n\t\t\tfmt.Print(\" \")\n\t\t}\n\t}\n\n\tvar printLevel func([]*Node, int, int)\n\tprintLevel = func(nList []*Node, l int, maxLevel int) {\n\t\tinitalSpaces := int(math.Pow(2, float64(maxLevel-l))) - 1\n\t\tseparaterSpaces := int(math.Pow(2, float64(maxLevel-l+1))) - 1\n\n\t\tisAllElementsNil := true\n\n\t\tprintSpace(initalSpaces)\n\t\tnewList := []*Node{}\n\t\tfor _, n := range nList {\n\t\t\tif n != nil {\n\t\t\t\tisAllElementsNil = false\n\t\t\t\tfmt.Print(n.value)\n\t\t\t\tnewList = append(newList, n.left)\n\t\t\t\tnewList = append(newList, n.right)\n\t\t\t} else {\n\t\t\t\tfmt.Print(\" \")\n\t\t\t\tnewList = append(newList, nil)\n\t\t\t\tnewList = append(newList, nil)\n\t\t\t}\n\t\t\tprintSpace(separaterSpaces)\n\t\t}\n\n\t\tfmt.Println(\"\")\n\n\t\tif !isAllElementsNil {\n\t\t\tprintLevel(newList, l+1, maxLevel)\n\t\t}\n\t}\n\n\tmaxLevel := findMaxLevel(t.r)\n\tnList := []*Node{t.r}\n\tprintLevel(nList, 1, maxLevel)\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tif root == nil {\n\t\treturn \"\"\n\t}\n\tmaps := make(map[int64]int)\n\n\tmaps[1] = root.Val\n\tinnerSerialize(root.Right, &maps, 3)\n\tinnerSerialize(root.Left, &maps, 2)\n\tfmt.Println(maps)\n\n\treturn stringfy(&maps)\n}", "func (n *fullNode) EncodeSER(w io.Writer) error {\n\tvar nodes [17]node\n\n\tfor i, child := range n.Children {\n\t\tif child != nil {\n\t\t\tnodes[i] = child\n\t\t} else {\n\t\t\tnodes[i] = nilValueNode\n\t\t}\n\t}\n\treturn ser.Encode(w, nodes)\n}", "func (n *Node) String() string { return fmt.Sprint(n.Data) }", "func writeNodeText(n *Node, b *bytes.Buffer) {\n\tswitch n.Type() {\n\tcase TextNode:\n\t\tb.WriteString(n.Name())\n\tcase ElementNode:\n\t\tfor c, err := n.FirstChild(); err == nil; c, err = c.NextSibling() {\n\t\t\twriteNodeText(c, b)\n\t\t}\n\t}\n}", "func (z *Corrupter) Write(p []byte) (n int, err error) {\n\tx := xml.NewDecoder(bytes.NewReader(p))\n\t_, err = x.Token()\n\tif err == nil {\n\t\terr = x.Skip()\n\t}\n\tif err == nil {\n\t\treturn 0, errors.New(\"zalgo: cannot consume XML\")\n\t}\n\n\th := html.NewTokenizer(bytes.NewReader(p))\n\tvar f bool\nL:\n\tfor {\n\t\tt := h.Next()\n\t\tswitch t {\n\t\tcase html.ErrorToken:\n\t\t\tif h.Err() == io.EOF {\n\t\t\t\tbreak L\n\t\t\t}\n\t\tcase html.StartTagToken, html.EndTagToken, html.SelfClosingTagToken, html.DoctypeToken:\n\t\t\tf = true\n\t\t}\n\t}\n\tif f {\n\t\treturn 0, errors.New(\"zalgo: cannot consume HTML\")\n\t}\n\n\tvar _n int\n\tn = z.n\n\tdefer func() {\n\t\tn = z.n - n\n\t}()\n\tfor _, r := range string(p) {\n\t\tif _, ok := zalgoChars[r]; ok {\n\t\t\tcontinue\n\t\t}\n\t\tz.b = z.b[:utf8.RuneLen(r)]\n\t\tutf8.EncodeRune(z.b, r)\n\t\t_n, err = z.w.Write(z.b)\n\t\tz.n += _n\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tif z.Zalgo != nil && z.Zalgo(z.n, r, z) {\n\t\t\tcontinue\n\t\t}\n\t\tfor i := real(z.Up); i > 0; i-- {\n\t\t\tif rnd.Float64() < imag(z.Up) {\n\t\t\t\t_n, err = fmt.Fprintf(z.w, \"%c\", up[rnd.Intn(len(up))])\n\t\t\t\tz.n += _n\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tfor i := real(z.Middle); i > 0; i-- {\n\t\t\tif rnd.Float64() < imag(z.Middle) {\n\t\t\t\t_n, err = fmt.Fprintf(z.w, \"%c\", middle[rnd.Intn(len(middle))])\n\t\t\t\tz.n += _n\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tfor i := real(z.Down); i > 0; i-- {\n\t\t\tif rnd.Float64() < imag(z.Down) {\n\t\t\t\t_n, err = fmt.Fprintf(z.w, \"%c\", down[rnd.Intn(len(down))])\n\t\t\t\tz.n += _n\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn\n}", "func (d *DriverFile) Write(folder string) {\n\ttag := \"NikuDataBus\"\n\tif folder == constant.FolderDebug {\n\t\ttag = \"XOGOutput\"\n\t}\n\tr, _ := regexp.Compile(\"(?s)<\" + tag + \"(.*)</\" + tag + \">\")\n\tstr := r.FindString(d.xogXML)\n\tif str == constant.Undefined {\n\t\tstr = d.xogXML\n\t}\n\tioutil.WriteFile(folder+d.Type+\"/\"+d.Path, []byte(str), os.ModePerm)\n}", "func (x *Index) Write(w io.Writer) error", "func (f *Fragment) Encode(w io.Writer) error {\n\ttraf := f.Moof.Traf\n\terr := traf.OptimizeTfhdTrun()\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor _, b := range f.Children {\n\t\terr := b.Encode(w)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (idx *Tree) WriteState(out io.Writer) (n int, err error) {\n\treturn idx.writeState(out)\n}", "func Encode(node ipld.Node, w io.Writer) error {\n\treturn dageth_trie.Encode(node, w)\n}", "func writeTree(in *inode, disk []byte, root *disklayout.ExtentNode, mockExtentBlkSize uint64) []byte {\n\trootData := binary.Marshal(nil, binary.LittleEndian, root.Header)\n\tfor _, ep := range root.Entries {\n\t\trootData = binary.Marshal(rootData, binary.LittleEndian, ep.Entry)\n\t}\n\n\tcopy(in.diskInode.Data(), rootData)\n\n\tvar fileData []byte\n\tfor _, ep := range root.Entries {\n\t\tif root.Header.Height == 0 {\n\t\t\tfileData = append(fileData, writeFileDataToExtent(disk, ep.Entry.(*disklayout.Extent))...)\n\t\t} else {\n\t\t\tfileData = append(fileData, writeTreeToDisk(disk, ep)...)\n\t\t}\n\t}\n\treturn fileData\n}", "func (w *FilteredLevelWriter) Write(p []byte) (int, error) {\n\treturn w.Writer.Write(p)\n}", "func stringify(n *BinarySearchNode, level int, builder *strings.Builder) {\n\tif n != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"---[ \"\n\t\tlevel++\n\t\tstringify(n.left, level, builder)\n\t\tbuilder.WriteString(fmt.Sprintf(format+\"%d\\n\", n.value))\n\t\tstringify(n.right, level, builder)\n\t}\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tif root == nil {\n\t\treturn \"[]\"\n\t}\n\n\telements := []string{}\n\tqueue := []*TreeNode{}\n\tqueue = append(queue, root)\n\n\tfor len(queue) > 0 {\n\t\tnode := queue[0]\n\t\tqueue = queue[1:]\n\t\tif node.Val == NULL_VAL {\n\t\t\telements = append(elements, NULL_STR)\n\t\t\tcontinue\n\t\t}\n\n\t\tif node.Left != nil {\n\t\t\tqueue = append(queue, node.Left)\n\t\t} else {\n\t\t\tqueue = append(queue, &TreeNode{\n\t\t\t\tVal: NULL_VAL,\n\t\t\t})\n\t\t}\n\n\t\tif node.Right != nil {\n\t\t\tqueue = append(queue, node.Right)\n\t\t} else {\n\t\t\tqueue = append(queue, &TreeNode{\n\t\t\t\tVal: NULL_VAL,\n\t\t\t})\n\t\t}\n\n\t\telements = append(elements, strconv.Itoa(node.Val))\n\t}\n\n\treturn \"[\" + strings.Join(elements, \",\") + \"]\"\n}", "func (e ChainEntry) Write() {\n\tif e.disabled {\n\t\treturn\n\t}\n\t// first find writer for level\n\t// if none, stop\n\te.Entry.l.closeEntry(e.Entry)\n\te.Entry.l.finalizeIfContext(e.Entry)\n\te.Entry.enc.Release()\n\n\tif e.exit {\n\t\te.Entry.l.exit(1)\n\t}\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tif root == nil {\n\t\treturn \"\"\n\t}\n\tvar result []string\n\tqueue := []*TreeNode{root}\n\tfor len(queue) > 0 {\n\t\troot = queue[0]\n\t\tqueue = queue[1:]\n\t\tif root == nil {\n\t\t\tresult = append(result, \"\")\n\t\t\tcontinue\n\t\t}\n\t\tresult = append(result, strconv.Itoa(root.Val))\n\t\tqueue = append(queue, root.Left, root.Right)\n\t}\n\treturn strings.Join(result, \",\")\n}", "func (r *Node) Out(formatString string, args ...interface{}) {\n\tutil.Out.Output(2, fmt.Sprintf(\"(%v/%v) %v\\n\", r.Self, r.State, fmt.Sprintf(formatString, args...)))\n}", "func (d *Driver) Write(path string, data []byte) error {\n\t_, stat, err := d.conn.Get(path)\n\tif err != nil && err == zk.ErrNoNode {\n\t\terr := d.makePath(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tif err != nil && err != zk.ErrNoNode {\n\t\treturn err\n\t}\n\n\t_, er := d.conn.Set(path, data, stat.Version)\n\tif er != nil {\n\t\treturn errors.Wrap(err, \"Error writing data to node. Path: \"+path)\n\t}\n\treturn nil\n}", "func writeNodeText(n *html.Node, b *bytes.Buffer) {\n\tswitch n.Type {\n\tcase html.TextNode:\n\t\tb.WriteString(n.Data)\n\tcase html.ElementNode:\n\t\tfor c := n.FirstChild; c != nil; c = c.NextSibling {\n\t\t\twriteNodeText(c, b)\n\t\t}\n\t}\n}", "func main() {\n\troot := TreeNode{\n\t\tVal: 1,\n\t\tLeft: &TreeNode{\n\t\t\tVal: 2,\n\t\t\tLeft: nil,\n\t\t\tRight: nil,\n\t\t},\n\t\tRight: &TreeNode{\n\t\t\tVal: 3,\n\t\t\tLeft: &TreeNode{\n\t\t\t\tVal: 4,\n\t\t\t\tRight: nil,\n\t\t\t\tLeft: nil,\n\t\t\t},\n\t\t\tRight: &TreeNode{\n\t\t\t\tVal: 5,\n\t\t\t\tRight: nil,\n\t\t\t\tLeft: nil,\n\t\t\t},\n\t\t},\n\t}\n\tobj := Constructor()\n\tdata := obj.serialize(&root)\n\tfmt.Println(data)\n}", "func (nc *Ctd) WriteASCII(mapFormat map[string]string, hdr []string) {\n\t// define 2 files, profiles header and data\n\tvar filename string\n\n\t//\tif _, err := os.Stat(\"/path/to/whatever\"); err == nil {\n\t//\t\t// path/to/whatever exists\n\t//\t}\n\n\t// build filenames\n\tstr := nc.Attributes[\"cycle_mesure\"]\n\tstr = strings.Replace(str, \"\\r\", \"\", -1)\n\theaderFilename := fmt.Sprintf(\"%s/ascii/%s.ctd\", outputDir, strings.ToLower(str))\n\tfilename = fmt.Sprintf(\"%s/ascii/%s%s_ctd\", outputDir, strings.ToLower(str), prefixAll)\n\t//fmt.Println(headerFilename)\n\t//fmt.Println(filename)\n\n\t// open header file for writing result\n\tfidHdr, err := os.Create(headerFilename)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer fidHdr.Close()\n\n\t// use buffered mode for writing\n\tfbufHdr := bufio.NewWriter(fidHdr)\n\n\t// open ASCII file for writing result\n\tfid, err := os.Create(filename)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer fid.Close()\n\n\t// use buffered mode for writing\n\tfbuf := bufio.NewWriter(fid)\n\n\t// write header to string\n\tstr = fmt.Sprintf(\"%s %s %s %s %s %s\\n\",\n\t\tnc.Attributes[\"cycle_mesure\"],\n\t\tnc.Attributes[\"plateforme\"],\n\t\tnc.Attributes[\"institute\"],\n\t\tnc.Attributes[\"type_instrument\"],\n\t\tnc.Attributes[\"instrument_number\"],\n\t\tnc.Attributes[\"pi\"])\n\n\t// write first line header on header file and ascii file\n\tfmt.Fprintf(fbufHdr, str)\n\tfmt.Fprintf(fbuf, str)\n\n\t// display on screen\n\tfmt.Printf(\"%s\", str)\n\n\t// write physical parameters in second line\n\tstr = \"\"\n\tfor _, key := range hdr {\n\t\tfmt.Fprintf(fbuf, \"%s \", key)\n\t\tfmt.Fprintf(debug, \"%s \", key)\n\t}\n\t// append new line\n\t//fmt.Fprintln(fbuf_ascii, \"\\n\")\n\n\t// write second line header on ascii file\n\tfmt.Fprintln(fbuf, str)\n\n\t// display on screen\n\tfmt.Printf(\"%s\", str)\n\n\t// get data (slices) from nc struct\n\tlen1D := nc.Dimensions[\"TIME\"]\n\tlen2D := nc.Dimensions[\"DEPTH\"]\n\ttime := nc.Variables.flatten(\"TIME\")\n\tlat := nc.Variables.flatten(\"LATITUDE\")\n\tlon := nc.Variables.flatten(\"LONGITUDE\")\n\tprofile := nc.Variables.flatten(\"PROFILE\")\n\tbath := nc.Variables.flatten(\"BATH\")\n\n\t// loop over each profile\n\tfor x := 0; x < len1D; x++ {\n\t\tstr = \"\"\n\t\t// write profile informations to ASCII data file with DEPTH = -1\n\t\tt1 := NewTimeFromJulian(time[x])\n\t\tt2 := NewTimeFromJulianDay(nc.ExtraFloat[fmt.Sprintf(\"ETDD:%d\", int(profile[x]))], t1)\n\t\t// TODOS: adapt profile format to stationPrefixLength\n\t\tfmt.Fprintf(fbuf, \"%05.0f %4d %f %f %f %s\",\n\t\t\tprofile[x],\n\t\t\tcodeForProfile,\n\t\t\tt1.JulianDayOfYear(),\n\t\t\tlat[x],\n\t\t\tlon[x],\n\t\t\tt1.Format(\"20060102150405\"))\n\n\t\t// write profile informations to header file, max depth CTD and\n\t\t// bathymetrie are in meters\n\t\tstr = fmt.Sprintf(\"%05.0f %s %s %s %s %4.4g %4.4g %s %s\\n\",\n\t\t\tprofile[x],\n\t\t\tt1.Format(\"02/01/2006 15:04:05\"),\n\t\t\tt2.Format(\"02/01/2006 15:04:05\"),\n\t\t\tDecimalPosition2String(lat[x], \"NS\"),\n\t\t\tDecimalPosition2String(lon[x], \"EW\"),\n\t\t\tnc.ExtraFloat[fmt.Sprintf(\"DEPTH:%d\", int(profile[x]))],\n\t\t\tbath[x],\n\t\t\tnc.ExtraString[fmt.Sprintf(\"TYPE:%d\", int(profile[x]))],\n\t\t\tcfg.Ctd.CruisePrefix+nc.ExtraString[fmt.Sprintf(\"PRFL_NAME:%d\", int(profile[x]))])\n\n\t\t// write profile information to header file\n\t\tfmt.Fprintf(fbufHdr, str)\n\n\t\t// display on screen\n\t\tfmt.Printf(\"%s\", str)\n\n\t\t// fill last header columns with 1e36\n\t\tfor i := 0; i < len(hdr)-6; i++ {\n\t\t\tfmt.Fprintf(fbuf, \" %g\", 1e36)\n\t\t}\n\t\tfmt.Fprintln(fbuf) // add newline\n\n\t\t// loop over each level\n\t\tfor y := 0; y < len2D; y++ {\n\t\t\t// goto next profile when max depth reach\n\t\t\tif nc.Variables.get(\"PRES\", x, y).(float64) >=\n\t\t\t\tnc.ExtraFloat[fmt.Sprintf(\"PRES:%d\", int(profile[x]))] {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tfmt.Fprintf(fbuf, \"%05.0f \", profile[x])\n\t\t\t// loop over each physical parameter (key) in the rigth order\n\t\t\tfor _, key := range hdr {\n\t\t\t\t// if key not in map, goto next key\n\t\t\t\tif _, ok := nc.Variables[key]; ok {\n\t\t\t\t\t// fill 2D slice\n\t\t\t\t\tdata := nc.Variables.get(key, x, y)\n\t\t\t\t\t// print data with it's format, change format for FillValue\n\t\t\t\t\tif data == 1e36 {\n\t\t\t\t\t\tfmt.Fprintf(fbuf, \"%g \", data)\n\t\t\t\t\t} else {\n\t\t\t\t\t\tfmt.Fprintf(fbuf, mapFormat[key]+\" \", data)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\tfmt.Fprintf(fbuf, \"\\n\")\n\n\t\t}\n\t\tfbuf.Flush()\n\t\tfbufHdr.Flush()\n\t}\n}", "func (t *BPTree) WriteNode(n *Node, off int64, syncEnable bool, fd *os.File) (number int, err error) {\n\tbn, err := t.ToBinary(n)\n\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tif off == -1 {\n\t\toff = n.Address\n\t}\n\n\tnumber, err = fd.WriteAt(bn, off)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tif syncEnable {\n\t\terr = fd.Sync()\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\treturn\n}", "func PrintLevel(node *AsciiNode, x, level int) {\n\tvar i, isLeft int\n\tif node == nil {\n\t\treturn\n\t}\n\n\tisLeft = bool2int(node.ParentDir == -1)\n\tif level == 0 {\n\t\tfor i = 0; i < x-PrintNext-((node.LabelLen-isLeft)/2); i++ {\n\t\t\tfmt.Fprintf(&AsciiBuilder, \"%s\", \" \")\n\t\t}\n\t\tPrintNext += i\n\t\tfmt.Fprintf(&AsciiBuilder, \"%s\", node.Label)\n\t\tPrintNext += node.LabelLen\n\t} else if node.EdgeLen >= level {\n\t\tif node.Left != nil {\n\t\t\tfor i = 0; i < x-PrintNext-level; i++ {\n\t\t\t\tfmt.Fprintf(&AsciiBuilder, \"%s\", \" \")\n\t\t\t}\n\t\t\tPrintNext += i\n\t\t\tfmt.Fprintf(&AsciiBuilder, \"%s\", \"/\")\n\t\t\tPrintNext++\n\t\t}\n\t\tif node.Right != nil {\n\t\t\tfor i = 0; i < x-PrintNext+level; i++ {\n\t\t\t\tfmt.Fprintf(&AsciiBuilder, \"%s\", \" \")\n\t\t\t}\n\t\t\tPrintNext += i\n\t\t\tfmt.Fprintf(&AsciiBuilder, \"%s\", \"\\\\\")\n\t\t\tPrintNext++\n\t\t}\n\t} else {\n\t\tPrintLevel(node.Left, x-node.EdgeLen-1, level-node.EdgeLen-1)\n\t\tPrintLevel(node.Right, x+node.EdgeLen+1, level-node.EdgeLen-1)\n\t}\n}", "func (r *renderer) write(s string, unescaped bool) {\n\tif r.indentNext {\n\t\tr.indentNext = false\n\t\tr.w.WriteString(r.indent)\n\t}\n\tif !unescaped {\n\t\ts = html.EscapeString(s)\n\t}\n\tr.w.WriteString(s)\n}", "func (c Node) Save() error {\n\t// Create node directories if they don't exist yet\n\t_, err := fileutil.MakeDirectory(c.NodeDirectory())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn ioutil.WriteFile(\n\t\tc.NodeFile(),\n\t\tdata,\n\t\tos.ModePerm,\n\t)\n}", "func Encode(in, out *os.File) {\n\tcounts := count(in)\n\tp := makePQ(counts)\n\th := makeHuffman(p)\n\tm := make(map[byte]string)\n\tfillMap(h, m, \"\")\n\tfor k, v := range m {\n\t\tfmt.Printf(\"k: %c, v: %s\\n\", k, v)\n\t}\n}", "func (tree *UTree) String() string {\r\n\tstr := \"RedBlackTree\\n\"\r\n\tif !tree.Empty() {\r\n\t\toutput(tree.root, \"\", true, &str)\r\n\t}\r\n\treturn str\r\n}", "func PrintTree(pager *Pager, pageNum uint32, indentLevel uint32) {\n\tvar page *Page = GetPage(pager, pageNum)\n\tvar numKeys, child uint32\n\tswitch GetNodeType(page.Mem[:]) {\n\tcase TypeLeafNode:\n\t\tnumKeys = *LeafNodeNumCells(page.Mem[:])\n\t\tindent(indentLevel)\n\t\tfmt.Printf(\"- Leaf num of cells: %v\\n\", numKeys)\n\t\tfor i := uint32(0); i < numKeys; i++ {\n\t\t\tindent(indentLevel + 1)\n\t\t\tfmt.Printf(\"- (Leaf cell num: %v, key: %v)\\n\", i, *LeafNodeKey(page.Mem[:], i))\n\t\t}\n\tcase TypeInternalNode:\n\t\tnumKeys = *InternalNodeNumKeys(page.Mem[:])\n\t\tindent(indentLevel)\n\t\tfmt.Printf(\"- Internal num of cells: %v\\n\", numKeys)\n\t\tfor i := uint32(0); i < numKeys; i++ {\n\t\t\tchild = *InternalNodeChild(page.Mem[:], i)\n\t\t\tPrintTree(pager, child, indentLevel+1)\n\n\t\t\tindent(indentLevel + 1)\n\t\t\tfmt.Printf(\"- (Internal cell num: %v, key: %v)\\n\", i, *InternalNodeKey(page.Mem[:], i))\n\t\t}\n\t\tchild = *internalNodeRightChildPtr(page.Mem[:])\n\t\tPrintTree(pager, child, indentLevel+1)\n\t}\n}", "func (n *TreeNode) ToString() string {\n\tif n == nil {\n\t\treturn \"nil\"\n\t}\n\tif n.Left == nil && n.Right == nil {\n\t\treturn strconv.Itoa(n.Data)\n\t}\n\treturn fmt.Sprintf(\"%d => (%s, %s)\", n.Data, n.Left.ToString(), n.Right.ToString())\n}", "func (g *Graph) Write(w io.Writer, format string) error {\n\treturn templates.ExecuteTemplate(w, format+\".tmpl\", g)\n}" ]
[ "0.73633575", "0.69991916", "0.6758368", "0.66080135", "0.6491745", "0.6239188", "0.6143039", "0.6132252", "0.60777396", "0.5980392", "0.5865827", "0.58224255", "0.58132666", "0.57983017", "0.57665694", "0.5751684", "0.5702447", "0.5665232", "0.56497645", "0.56465477", "0.5635306", "0.5626989", "0.56175405", "0.5539121", "0.5529355", "0.55181575", "0.5496982", "0.5470584", "0.5438289", "0.5438159", "0.54065466", "0.5393971", "0.53929263", "0.53758365", "0.536827", "0.53597134", "0.534817", "0.53393036", "0.5331544", "0.5319176", "0.53162616", "0.53114504", "0.52929866", "0.5284792", "0.52763766", "0.5270401", "0.5262186", "0.525746", "0.5257262", "0.52537733", "0.52531695", "0.5239344", "0.52389604", "0.52371526", "0.5227423", "0.52272224", "0.5225275", "0.5209748", "0.52066624", "0.519816", "0.51919025", "0.5190787", "0.5186109", "0.5178241", "0.51664203", "0.51588315", "0.51554704", "0.5154763", "0.5151439", "0.5140004", "0.51207316", "0.5083377", "0.5083022", "0.5072452", "0.5072117", "0.5060994", "0.5054673", "0.5048995", "0.50440866", "0.50432837", "0.50417066", "0.5038207", "0.50262153", "0.5021967", "0.5013516", "0.5008753", "0.50017565", "0.49890363", "0.49514407", "0.49482992", "0.49426305", "0.4941783", "0.4932414", "0.49318883", "0.49315017", "0.49296844", "0.4928016", "0.49109674", "0.4898421", "0.48854244" ]
0.53567725
36
Tree adds this node to the root
func (a node) Tree(root treeprint.Tree) error { sort.Sort(a) branch := root var err error // generate a node for the Resource if a.RNode != nil { branch, err = a.p.doResource(a.RNode, "Resource", root) if err != nil { return err } } // attach children to the branch for _, n := range a.children { if err := n.Tree(branch); err != nil { return err } } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (tree *BinaryTree) Add(value int) {\n\tif tree.root == nil {\n\t\ttree.root = &(node{Value: value})\n\t} else {\n\t\ttree.root.add(value)\n\t}\n\n}", "func (t *Tree) Add(val float64) {\n\tt.root = t.root.add(val)\n}", "func AddRootNode(db *sql.DB, id int64, name string) error {\n\t// move all other nodes to right, if exits\n\tvar sql bytes.Buffer\n\tsql.WriteString(moveOnAddSQL)\n\t_, err := db.Exec(sql.String(), 0, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\tsql.Reset()\n\n\t// insert root\n\tsql.WriteString(insertSQL)\n\targs := []interface{}{id, name, 0, 1, 1, 2}\n\n\tresult, err := db.Exec(sql.String(), args...)\n\tif err != nil {\n\t\treturn nil\n\t}\n\taffected, _ := result.RowsAffected()\n\tif affected < 1 {\n\t\treturn errors.New(\"nested: inserting root affected none\")\n\t}\n\n\treturn nil\n}", "func (t *Tree) Add(content IContent) error {\n\tid := content.GetID()\n\tvar err error\n\n\tcurrent := t.rootNode\n\tfor depth := 0; depth < t.depth; depth++ {\n\t\tvar b byte\n\t\tif b, err = util.GetBitFromByteArray(depth, id); err != nil {\n\t\t\treturn errors.New(\"Error inserting\")\n\t\t}\n\n\t\tif b == 1 {\n\t\t\tif current.left == nil {\n\t\t\t\tcurrent.left = &node{}\n\t\t\t\tcurrent.left.parent = current\n\t\t\t}\n\t\t\tcurrent = current.left\n\t\t} else {\n\t\t\tif current.right == nil {\n\t\t\t\tcurrent.right = &node{}\n\t\t\t\tcurrent.right.parent = current\n\t\t\t}\n\t\t\tcurrent = current.right\n\t\t}\n\t}\n\n\tif current.leaf == nil {\n\t\tcurrent.leaf = NewLeaf()\n\t}\n\n\tcurrent.leaf.Add(content)\n\treturn nil\n}", "func (t *Tree)AddNode(val int){\n\tt.root = AddNodeUtil(t.root, val)\n}", "func (b *VersionTree) Add(value *Value) {\n\tif b.Root == nil {\n\t\tb.Root = &Tree{nil, value, nil}\n\t} else {\n\t\tinsert(b.Root, value)\n\t}\n}", "func (n *TreeNode) rootify(p *Pos, model ModelInterface) {\n\tn.step = 0\n\tn.side = 1\n\tn.first = true\n\tn.parent = nil\n\tn.children = n.children[:0]\n\tn.Expand(p, model)\n}", "func (bt *BinarySearchTree) Add(val int) {\n\tif bt.root == nil {\n\t\tbt.root = &TreeNode{Val: val}\n\t} else {\n\t\tbt.addTreeNode(bt.root, &TreeNode{Val: val})\n\t}\n}", "func (tb *TreeBuilder) Add(node bufr.Node) {\n if debug.DEBUG {\n fmt.Println(\"++++++++++++++ START ADD NODE ++++++++++++++\")\n defer fmt.Println(\"++++++++++++++ DONE ADD NODE ++++++++++++++\")\n }\n if debug.DEBUG {\n fmt.Println(node)\n }\n tb.node.AddMember(node)\n}", "func (t *Tree) Add(value interface{}) {\n\tnode := makeNode(value)\n\tif t.root == nil {\n\t\tt.root = node\n\t} else {\n\t\tadded := false\n\t\tt.BFS(func(n *Node) bool {\n\t\t\tif n.Left == nil {\n\t\t\t\tnode.Parent = n\n\t\t\t\tn.Left = node\n\t\t\t\tadded = true\n\t\t\t} else if n.Right == nil {\n\t\t\t\tnode.Parent = n\n\t\t\t\tn.Right = node\n\t\t\t\tadded = true\n\t\t\t}\n\t\t\treturn added\n\t\t})\n\t}\n}", "func CalTree(data [][]byte) {\r\n\tvar Root Node\r\n\tRoot.GenerateRoot(data, true)\r\n}", "func (tree *Tree) insert( m int) {\nif tree != nil {\nif tree.LeftNode == nil {\ntree.LeftNode = &Tree{nil,m,nil}\n} else {\nif tree.RightNode == nil {\ntree.RightNode = &Tree{nil,m,nil}\n} else {\nif tree.LeftNode != nil {\ntree.LeftNode.insert(m)\n} else {\ntree.RightNode.insert(m)\n}\n}\n}\n} else {\ntree = &Tree{nil,m,nil}\n}\n}", "func (self *StateObject) updateRoot(db trie.Database) {\n\tself.updateTrie(db)\n\tself.data.Root = self.trie.Hash()\n}", "func (r *RBTree) Add(key, value interface{}) {\n\tr.root = r.add(r.root, key, value)\n\t//root always Black\n\tr.root.color = BLACK\n}", "func (delegateObject *delegateObject) updateRoot(db Database) {\n\tdelegateObject.updateTrie(db)\n\tdelegateObject.data.Root = delegateObject.trie.Hash()\n}", "func addTreeNode(stack *stack, node *Node, item compare.Lesser) *Node {\n\tstack.init(node)\n\tdefer stack.reset()\n\n\tif node == nil {\n\t\t// case 1: new root\n\t\treturn &Node{\n\t\t\tItem: item,\n\t\t\tColor: Black,\n\t\t}\n\t}\n\n\tfor node != nil {\n\t\tswitch {\n\t\tcase item.Less(node.Item):\n\t\t\tstack.push(node, Left)\n\t\t\tnode = node.Left\n\t\tcase node.Item.Less(item):\n\t\t\tstack.push(node, Right)\n\t\t\tnode = node.Right\n\t\tdefault:\n\t\t\tnode.Item = item\n\t\t\treturn stack.root()\n\t\t}\n\t}\n\n\tstack.bindChild(&Node{\n\t\tItem: item,\n\t\tColor: Red,\n\t})\n\n\taddTreeNodeBalance(stack)\n\n\troot := stack.root()\n\troot.Color = Black\n\n\treturn root\n}", "func (t *ASCIITree) Add(children ...*ASCIITree) {\n\tfor _, child := range children {\n\t\tchild.Parent = t\n\t}\n\tt.children = append(t.children, children...)\n}", "func (self *StateStore) AddMerkleTreeRoot(txRoot common.Hash) error {\n\tkey := self.getMerkleTreeKey()\n\tself.merkleTree.AppendHash(txRoot)\n\terr := self.merkleHashStore.Flush()\n\tif err != nil {\n\t\treturn err\n\t}\n\ttreeSize := self.merkleTree.TreeSize()\n\thashes := self.merkleTree.Hashes()\n\tvalue := bytes.NewBuffer(make([]byte, 0, 8+len(hashes)*common.HashLength))\n\terr = serialization.WriteUint64(value, treeSize)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor _, hash := range hashes {\n\t\terr = hash.Serialize(value)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tself.store.BatchPut(key, value.Bytes())\n\treturn nil\n}", "func (t *tree) Root() *node {\n\treturn t.root\n}", "func (tb *TreeBuilder) Push(node bufr.Node) {\n if debug.DEBUG {\n fmt.Println(\">>>>>>>>>>>>>> START PUSH NODE >>>>>>>>>>>>>>\")\n defer fmt.Println(\">>>>>>>>>>>>>> DONE PUSH NODE >>>>>>>>>>>>>>\")\n }\n if debug.DEBUG {\n fmt.Println(len(tb.stack), node)\n }\n tb.stack = append(tb.stack, tb.node)\n tb.node = node\n if debug.DEBUG {\n fmt.Println(len(tb.stack))\n }\n}", "func (t *Tree) Root() *TreeNode {\n\treturn t.root\n}", "func (t *Tree) UpdateRoot(p *Pos, model ModelInterface) {\n\tif t.p == nil || t.p.Hash() != p.Hash() {\n\t\tt.p = p\n\t\tt.root = t.NewTreeNode(nil, 0, false, 1, true)\n\t\tt.root.rootify(p, model)\n\t}\n}", "func (n *ForNode) Tree() *Tree { return n.tree }", "func (t *Tree) AddNode(inputHash hash.Imprint, options ...InputHashOption) error {\n\tif t == nil || len(inputHash) == 0 {\n\t\treturn errors.New(errors.KsiInvalidArgumentError)\n\t}\n\n\tif t.aggrRoot != nil {\n\t\tlog.Error(\"Trying to add new leaf to a closed tree.\")\n\t\treturn errors.New(errors.KsiInvalidStateError).AppendMessage(\"Tree is closed.\")\n\t}\n\n\t// Handle input options.\n\topts := inputHashOptions{tree: t}\n\tfor _, setter := range options {\n\t\tif setter == nil {\n\t\t\treturn errors.New(errors.KsiInvalidArgumentError).AppendMessage(\"Provided option is nil.\")\n\t\t}\n\t\tif err := setter(&opts); err != nil {\n\t\t\treturn errors.KsiErr(err).AppendMessage(\"Unable to apply input hash option.\")\n\t\t}\n\t}\n\n\t// Verify that there is spare space for adding extra data to the tree.\n\theight, err := t.expectedHeight(opts.level, opts.meta.val != nil)\n\tif err != nil {\n\t\treturn errors.KsiErr(err).AppendMessage(\"Unable to add node to the tree.\")\n\t}\n\tif height > t.maxLevel {\n\t\treturn errors.New(errors.KsiBufferOverflow).AppendMessage(\"Tree max level overflow.\")\n\t}\n\n\txi := &TreeNode{\n\t\ttree: t,\n\t\thshValue: append([]byte(nil), inputHash...),\n\t\tlevel: opts.level,\n\t\tleafValue: inputHash,\n\t\tuserCtx: opts.userCtx,\n\t}\n\tt.leafs = append(t.leafs, xi)\n\tif err := t.notifyRecordListener(xi); err != nil {\n\t\treturn err\n\t}\n\n\t// In case additional metadata is provided, make a sub-tree.\n\tif opts.meta.val != nil {\n\t\tmdi := TreeNode{\n\t\t\ttree: t,\n\t\t\thshValue: opts.meta.bin,\n\t\t\tleafValue: opts.meta.val,\n\t\t}\n\t\tif err := t.notifyMetadataListener(&mdi); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif xi, err = t.joinNodes(xi, &mdi); err != nil {\n\t\t\treturn errors.KsiErr(err).AppendMessage(\"Failed to aggregate nodes.\")\n\t\t}\n\t}\n\n\tif t.useBlindingMask() {\n\t\tmask, err := t.calculateBlindingMask()\n\t\tif err != nil {\n\t\t\treturn errors.KsiErr(err).AppendMessage(\"Failed to calculate blinding mask.\")\n\t\t}\n\t\tif mask == nil {\n\t\t\treturn errors.New(errors.KsiInvalidFormatError).AppendMessage(\"Inconsistent blinding mask.\")\n\t\t}\n\t\tlog.Debug(fmt.Sprintf(\"Blinding mask for node[%d]: %s\", t.leafCount, mask))\n\n\t\tmi := &TreeNode{\n\t\t\ttree: t,\n\t\t\thshValue: append([]byte(nil), mask...),\n\t\t\tleafValue: mask,\n\t\t}\n\t\tif xi, err = t.joinNodes(mi, xi); err != nil {\n\t\t\treturn errors.KsiErr(err).AppendMessage(\"Failed to aggregate nodes.\")\n\t\t}\n\t}\n\n\tif err = t.insertNode(xi, 0); err != nil {\n\t\treturn err\n\t}\n\tt.lastLeaf = hash.Imprint(xi.hshValue)\n\tlog.Debug(fmt.Sprintf(\"Lash leaf at %d: %s\", t.leafCount, t.lastLeaf))\n\tt.leafCount++\n\n\treturn nil\n}", "func newDemoTree() TreeRoot {\n\tn1 := newNode(1)\n\n\tn2 := newNode(2)\n\tn1.Left = n2\n\n\tn3 := newNode(3)\n\tn1.Right = n3\n\n\tn4 := newNode(4)\n\tn2.Left = n4\n\n\tn5 := newNode(5)\n\tn2.Right = n5\n\n\tn6 := newNode(6)\n\tn3.Left = n6\n\n\tn7 := newNode(7)\n\tn3.Right = n7\n\n\tn8 := newNode(8)\n\tn4.Left = n8\n\n\tn9 := newNode(9)\n\tn4.Right = n9\n\n\tn10 := newNode(10)\n\tn5.Left = n10\n\n\tn11 := newNode(11)\n\tn5.Right = n11\n\n\tn12 := newNode(12)\n\tn6.Left = n12\n\n\tn13 := newNode(13)\n\tn6.Right = n13\n\n\tn14 := newNode(14)\n\tn7.Left = n14\n\n\tn15 := newNode(15)\n\tn7.Right = n15\n\n\tn16 := newNode(16)\n\tn14.Left = n16\n\n\tn17 := newNode(17)\n\tn14.Right = n17\n\n\treturn n1\n}", "func (self *TTree) addnodes(aMethod string, aNodes []*TNode, aIsHook bool) {\n\t//fmt.Println(\"self.Root\", self.Root)\n\t// 获得对应方法[POST,GET...]\n\tcn := self.Root[aMethod]\n\tif cn == nil {\n\n\t\t// 初始化Root node\n\t\tcn = &TNode{\n\t\t\tChildren: TSubNodes{},\n\t\t}\n\t\tself.Root[aMethod] = cn\n\t}\n\n\tvar p *TNode = cn // 复制方法对应的Root\n\n\t// 层级插入Nodes的Node到Root\n\tfor idx, _ := range aNodes {\n\t\tp = cn.addnode(p, aNodes, idx, aIsHook)\n\t}\n}", "func (self *TTree) addNodes(method string, nodes []*treeNode, isHook bool) {\n\t//fmt.Println(\"self.Root\", self.Root)\n\t// 获得对应方法[POST,GET...]\n\tcn := self.root[method]\n\tif cn == nil {\n\n\t\t// 初始化Root node\n\t\tcn = &treeNode{\n\t\t\tChildren: subNodes{},\n\t\t}\n\t\tself.root[method] = cn\n\t}\n\n\tvar p *treeNode = cn // 复制方法对应的Root\n\n\t// 层级插入Nodes的Node到Root\n\tfor idx := range nodes {\n\t\tp = cn.addNode(self, p, nodes, idx, isHook)\n\t}\n}", "func (self *treeNode) addNode(tree *TTree, parent *treeNode, nodes []*treeNode, i int, isHook bool) *treeNode {\n\tif len(parent.Children) == 0 {\n\t\tparent.Children = make([]*treeNode, 0)\n\t}\n\n\t// 如果:找到[已经注册]的分支节点则从该节继续[查找/添加]下一个节点\n\tfor _, n := range parent.Children {\n\t\tif n.Equal(nodes[i]) {\n\t\t\t// 如果:插入的节点层级已经到末尾,则为该节点注册路由\n\t\t\tif i == len(nodes)-1 {\n\t\t\t\t// 原始路由会被替换\n\t\t\t\tif isHook {\n\t\t\t\t\tn.Route.CombineHandler(nodes[i].Route)\n\t\t\t\t} else {\n\t\t\t\t\tn.Route = nodes[i].Route\n\t\t\t\t\ttree.Count.Inc()\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn n\n\t\t}\n\t}\n\n\t// 如果:该节点没有对应分支则插入同级的nodes为新的分支\n\tparent.Children = append(parent.Children, nodes[i])\n\tsort.Sort(parent.Children)\n\treturn nodes[i]\n}", "func (tb *TreeBuilder) Add(bytes []byte, level uint) error {\n\tif len(tb.levels) == 0 {\n\t\ttb.levels = []*TreeBuilderNode{new(TreeBuilderNode)}\n\t}\n\ttb.levels[0].Chunks = append(tb.levels[0].Chunks, bytes)\n\tfor _, n := range tb.levels {\n\t\tn.size += uint64(len(bytes))\n\t}\n\tfor i := uint(0); i < level; i++ {\n\t\tif i == uint(len(tb.levels))-1 {\n\t\t\ttb.levels = append(tb.levels, &TreeBuilderNode{\n\t\t\t\tsize: tb.levels[i].size,\n\t\t\t})\n\t\t}\n\t\tvar n Node = tb.levels[i]\n\t\tif tb.F != nil {\n\t\t\tvar err error\n\t\t\tn, err = tb.F(tb.levels[i])\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\ttb.levels[i+1].Nodes = append(tb.levels[i+1].Nodes, n)\n\t\ttb.levels[i] = &TreeBuilderNode{\n\t\t\toffset: tb.levels[i+1].offset + tb.levels[i+1].size,\n\t\t}\n\t}\n\treturn nil\n}", "func Add(name string, root *Node) {\n\tmsgDir[name] = root\n}", "func (t *Tree) Root() *node {\n\treturn t.root\n}", "func (t *Tree) Insert(rr dns.RR) {\n\tvar d int\n\tt.Root, d = t.Root.insert(rr)\n\tt.Count += d\n\tt.Root.Color = Black\n}", "func insert(root *Node, data int) *Node{\n\t//if the tree has not been created yet, insert in to root.\n\tif root == nil {\n\t\troot = &Node{data, nil, nil}\n\t\treturn root\n\t} \n\n\t//if the data is less than the data contained in the root\n\t//recursivly insert into the left node.\n\tif data < root.Data {\n\t\troot.Left = insert(root.Left, data)\n\t} \n\n\t//if the data is greater than the data contained in the root,\n\t//recursivly insert into the right node\n\tif data > root.Data {\n\t\troot.Right = insert(root.Right, data)\n\t}\n\treturn root\n}", "func (tn *TreeNode) Add(child *TreeNode) *TreeNode {\n\tif _, ok := tn.Children[child.Name]; !ok {\n\t\ttn.Children[child.Name] = child\n\t} else {\n\t\ttn.Children[child.Name].Total += child.Total\n\t}\n\treturn tn.Children[child.Name]\n}", "func (r *RPCClientRoot) Tree(\n\tfeed cipher.PubKey,\n\tnonce uint64,\n\tseq uint64,\n) (\n\ttree string,\n\terr error,\n) {\n\terr = r.r.c.Call(\"root.Tree\", RootSelector{feed, nonce, seq}, &tree)\n\treturn\n}", "func (t *Tree) Add(v int) {\n\tn := Node{v, nil, nil}\n\tt.r = add(t.r, n)\n}", "func (obj *StateObject) updateRoot(db Database) {\n\tobj.updateTrie(db)\n\tobj.data.StorageRoot = obj.trie.Hash()\n}", "func AddNewNode(root *Node, node *Node) *Node {\n\t// set the new node to red\n\tnode.Color = Red\n\n\troot = addOneNode(root, Left, node)\n\n\t// reset root color\n\troot.Color = Black\n\n\treturn root\n}", "func (tree *Tree) Insert(value int) {\n\ttree.root, _ = tree.root.Insert(value)\n}", "func (t *RBTree) Add(item compare.Lesser) {\n\tt.lock.Lock()\n\tdefer t.lock.Unlock()\n\n\tt.Node = addTreeNode(t.stack, t.Node, item)\n}", "func (t *Trie) Add(k string, v int64) {\n\n\tt.lock.Lock()\n\tdefer t.lock.Unlock()\n\n\tif len(k) == 0 || v < 0 {\n\t\treturn\n\t}\n\tbytes := []byte(k)\n\n\tcurr := t.root\n\tfor _, b := range bytes {\n\n\t\tif curr.children[b] == nil {\n\t\t\tcurr.children[b] = newNode()\n\t\t}\n\n\t\tcurr.children[b].parent = curr\n\t\tcurr = curr.children[b]\n\t\tcurr.key = b\n\t}\n\n\tcurr.val.add(v)\n}", "func (f *FS) Root() (fs.Node, error) {\n\treturn &Node{fs: f}, nil\n}", "func (root *Node) add(p string, blob *Blob) {\n\tfullPath := strings.Split(p, \"/\")\n\tcurr := root\n\tvar nodePath string\n\t// A blob may be used by many packages, so the size of each blob is the total space allocated to\n\t// that blob in blobfs.\n\t// We divide the size by the total number of packages that depend on it to get a rough estimate of\n\t// the size of the individual blob.\n\tsize := blob.size / int64(len(blob.dep))\n\tcurr.size += size\n\tfor _, name := range fullPath {\n\t\tname = strings.TrimSuffix(name, \".meta\")\n\t\tnodePath = filepath.Join(nodePath, name)\n\t\tif _, ok := curr.children[name]; !ok {\n\t\t\ttarget := newNode(nodePath)\n\t\t\ttarget.parent = curr\n\t\t\tcurr.children[name] = target\n\t\t}\n\t\tcurr = curr.children[name]\n\t\tcurr.size += size\n\t}\n}", "func (tree *Tree) Insert(key int, value string) {\n\ttree.Root = tree.Root.insert(key, value)\n}", "func (menu *Menu) AddToTree(menuNode *Menu) {\n\n\tif menuNode.ParentId == 0 {\n\n\t\tmenu.Children = append(menu.Children, menuNode)\n\t} else if menuNode.ParentId == menu.ID {\n\n\t\tmenu.Children = append(menu.Children, menuNode)\n\n\t} else {\n\n\t\tfor _, u := range menu.Children {\n\n\t\t\tu.AddToTree(menuNode)\n\t\t}\n\n\t}\n\n}", "func (r *Root) Root() (fs.Node, error) {\n\tdebug.Log(\"Root()\")\n\treturn r, nil\n}", "func newRoot(view *View, leafAllocation int64) *root {\n\tif leafAllocation < 10 {\n\t\tleafAllocation = 10\n\t}\n\tleafNum := 3 - ((leafAllocation - 1) % 3) + leafAllocation\n\tnodeNum := (leafNum - 1) / 3\n\tr := new(root)\n\tr.leaves = make([]leaf, leafNum, leafNum)\n\tfor i := 0; i < len(r.leaves)-2; i++ {\n\t\tr.leaves[i].nextFree = &r.leaves[i+1]\n\t}\n\tr.nodes = make([]node, nodeNum, nodeNum)\n\tfor i := 0; i < len(r.nodes)-2; i++ {\n\t\tr.nodes[i].nextFree = &r.nodes[i+1]\n\t}\n\tr.freeNode = &r.nodes[0]\n\tr.freeLeaf = &r.leaves[0]\n\trootNode := r.newNode(view)\n\tr.rootNode = rootNode\n\treturn r\n}", "func (n Node) GenerateRoot(data [][]byte, printFlag bool) (RootNode *Node) {\r\n\r\n\t// Prepare leaf nodes\r\n\tvar nodes []*Node\r\n\r\n\tfor i := 0; i < len(data); i = i + 1 {\r\n\t\tnodes = append(nodes, &Node{\r\n\t\t\tNodeData: data[i],\r\n\t\t\tNodeHash: n.CalSHA256Hash(data[i]),\r\n\t\t})\r\n\t}\r\n\r\n\t// Print the leaf node if printFlag is ON\r\n\tif printFlag == true {\r\n\t\tfmt.Printf(\"\\nLeaf Nodes: \\n\")\r\n\t\tfor i := 0; i < len(nodes); i = i + 1 {\r\n\t\t\tfmt.Printf(\"%x, [%s]\\n\", nodes[i].NodeHash, nodes[i].NodeData)\r\n\t\t}\r\n\t}\r\n\r\n\t// Build tree from bottom layer\r\n\tfor {\r\n\t\tif len(nodes) != 1 {\r\n\r\n\t\t\t// If the number of node in bottom layer is not even, replicate and push the last node\r\n\t\t\tif len(nodes)%2 == 1 {\r\n\t\t\t\tnodes = append(nodes, &Node{\r\n\t\t\t\t\tNodeData: nodes[len(nodes)-1].NodeData,\r\n\t\t\t\t\tNodeHash: nodes[len(nodes)-1].NodeHash,\r\n\t\t\t\t})\r\n\t\t\t}\r\n\r\n\t\t\t// Build the upper level nodes\r\n\t\t\tvar upperNodes []*Node\r\n\t\t\tfor i := 0; i < len(nodes); i = i + 2 {\r\n\t\t\t\tupperNodes = append(upperNodes, &Node{\r\n\t\t\t\t\tNodeData: bytes.Join([][]byte{nodes[i].NodeHash, nodes[i+1].NodeHash}, []byte{}),\r\n\t\t\t\t\tNodeHash: n.CalSHA256Hash(bytes.Join([][]byte{nodes[i].NodeHash, nodes[i+1].NodeHash}, []byte{})),\r\n\t\t\t\t})\r\n\t\t\t}\r\n\r\n\t\t\t// Replace leaf node with upper level nodes.\r\n\t\t\t// Continue tree building until Root node is reached, i.e. len(nodes) == 1\r\n\t\t\tnodes = upperNodes\r\n\r\n\t\t\t// Print the upper node if printFlag is ON\r\n\t\t\tif printFlag == true {\r\n\t\t\t\tfmt.Printf(\"\\nUpper Nodes: \\n\")\r\n\t\t\t\tfor i := 0; i < len(nodes); i = i + 1 {\r\n\t\t\t\t\tfmt.Printf(\" Hash[%d]\t%x\\n\", i, nodes[i].NodeHash)\r\n\t\t\t\t\tfmt.Printf(\" Data[%d]\t%x\\n\", i, nodes[i].NodeData)\r\n\t\t\t\t}\r\n\t\t\t}\r\n\r\n\t\t} else {\r\n\t\t\tif printFlag == true {\r\n\t\t\t\tfmt.Printf(\"\\n\")\r\n\t\t\t}\r\n\t\t\tbreak\r\n\t\t}\r\n\t}\r\n\tRootNode = nodes[0]\r\n\treturn RootNode\r\n}", "func (p *ancestorTree) Add(blkID ids.ID, parentID ids.ID) {\n\tp.childToParent[blkID] = parentID\n\n\tchildren := p.parentToChildren[parentID]\n\tchildren.Add(blkID)\n\tp.parentToChildren[parentID] = children\n}", "func (r *Root) Root() (fs.Node, error) {\n\treturn newDir(nil, r.registry), nil\n}", "func (r *root) newNode(view *View) (n *node) {\n\tif r.freeNode == nil {\n\t\tn = &node{view: *view, disposable: true}\n\t} else {\n\t\tn = r.freeNode\n\t\tr.freeNode = n.nextFree\n\t\tn.view = *view\n\t}\n\tr.newLeaves(view, &n.children)\n\treturn\n}", "func (treeNode *TreeNode) Insert(son1 *TreeNode, son2 *TreeNode) *TreeNode {\n\tvar root *TreeNode\n\tvar item Item\n\n\titem.Symbol = 0\n\titem.Weight = son1.Value.Weight + son2.Value.Weight\n\troot, _ = root.New(item)\n\n\t//This condition may be avoided if the heap is well managed.\n\tif son1.Value.Weight < son2.Value.Weight {\n\t\troot.Left = son1\n\t\troot.Right = son2\n\t} else {\n\t\troot.Left = son2\n\t\troot.Right = son1\n\t}\n\n\treturn root\n}", "func (tr *Tree) Insert(cell store_pb.RecordID, data unsafe.Pointer, extra uint64) {\n\tif tr.root == nil {\n\t\ttr.root = new(nodeT)\n\t}\n\ttr.insert(tr.root, cell, data, extra, 128-nBits)\n\ttr.len++\n}", "func (o *Octree) Insert(p *Point) {\n\tif o.root == nil {\n\t\to.root = newLeaf(p)\n\t\treturn\n\t}\n\n\to.root = o.root.Insert(p)\n}", "func addNode(curNode **node, downUpOrNext int) {\n\t\tvar newNode *node\n\n\t\tnewNode = &node{}\n\t\tnewNode.Name = scan.name\n\t\tnewNode.Parent = *curNode // nest it, depth+=1\n\n\t\t// add created newNode to Children of Parent\n\t(*curNode).Children = append((*curNode).Children, newNode)\n\n\t\tif downUpOrNext == 1 {\n\t\t\t*curNode = newNode\n\t\t} else if downUpOrNext == 2 {\n\t\t\t*curNode = (*curNode).Parent\n\t\t}\n\n\t\tscan.name = \"\"\n}", "func (n *NodeBuilder) Root() *TagNode {\n\treturn n.root\n}", "func NewWithRoot(root Value) Tree {\n\treturn &Node{Value: root}\n}", "func (g *Graph) UpdateRoot(root Transaction) {\n\tptr := &root\n\n\tg.Lock()\n\n\tg.depthIndex[root.Depth] = append(g.depthIndex[root.Depth], ptr)\n\tg.eligibleIndex.ReplaceOrInsert((*sortByDepthTX)(ptr))\n\n\tg.transactions[root.ID] = ptr\n\n\tif g.indexer != nil {\n\t\tg.indexer.Index(hex.EncodeToString(root.ID[:]))\n\t}\n\n\tg.height = root.Depth + 1\n\n\tg.Unlock()\n\n\tg.UpdateRootDepth(root.Depth)\n}", "func (e *Entry) Tree() tree.Tree {\n\treturn e.root\n}", "func (fs *fsMutable) initRoot() (err error) {\n\t_, found := fs.lookupTree.Get(formKey(fuseops.RootInodeID))\n\tif found {\n\t\treturn\n\t}\n\terr = fs.createNode(\n\t\tformLookupKey(fuseops.RootInodeID, rootPath),\n\t\tfuseops.RootInodeID,\n\t\trootPath,\n\t\tnil,\n\t\tfuseutil.DT_Directory,\n\t\ttrue)\n\treturn\n}", "func (t *Tree) Add(pattern string, handlers []baa.HandlerFunc, nameNode *Node) *Tree {\n\t// find the common prefix\n\tmatched := 0\n\tfor ; matched < len(pattern) && matched < len(t.pattern) && pattern[matched] == t.pattern[matched]; matched++ {\n\t}\n\n\t// no prefix\n\tif matched == 0 {\n\t\treturn nil\n\t}\n\n\tif matched == len(t.pattern) {\n\t\t// the node pattern is the same as the pattern: make the current node as data node\n\t\tif matched == len(pattern) {\n\t\t\tif handlers != nil {\n\t\t\t\tif t.handlers != nil {\n\t\t\t\t\tpanic(\"the route is be exists: \" + t.String())\n\t\t\t\t}\n\t\t\t\tt.handlers = handlers\n\t\t\t\tt.nameNode = nameNode\n\t\t\t}\n\t\t\treturn t\n\t\t}\n\n\t\t// the node pattern is a prefix of the pattern: create a child node\n\t\tpattern = pattern[matched:]\n\t\tfor _, child := range t.schildren {\n\t\t\tif node := child.Add(pattern, handlers, nameNode); node != nil {\n\t\t\t\treturn node\n\t\t\t}\n\t\t}\n\n\t\t// no child match, to be a new child\n\t\treturn t.addChild(pattern, handlers, nameNode)\n\t}\n\n\t// the pattern is a prefix of node pattern: create a new node instead of child\n\tif matched == len(pattern) {\n\t\tnode := NewTree(t.pattern[matched:], t.handlers)\n\t\tnode.nameNode = t.nameNode\n\t\tnode.schildren = t.schildren\n\t\tnode.rchildren = t.rchildren\n\t\tnode.parent = t\n\t\tt.pattern = pattern\n\t\tt.format = []byte(t.pattern)\n\t\tt.handlers = handlers\n\t\tt.nameNode = nameNode\n\t\tt.schildren = []*Tree{node}\n\t\tt.rchildren = nil\n\t\treturn t\n\t}\n\n\t// the node pattern shares a partial prefix with the key: split the node pattern\n\tnode := NewTree(t.pattern[matched:], t.handlers)\n\tnode.nameNode = t.nameNode\n\tnode.schildren = t.schildren\n\tnode.rchildren = t.rchildren\n\tnode.parent = t\n\tt.pattern = pattern[:matched]\n\tt.format = []byte(t.pattern)\n\tt.handlers = nil\n\tt.nameNode = nil\n\tt.schildren = nil\n\tt.rchildren = nil\n\tt.schildren = append(t.schildren, node)\n\treturn t.addChild(pattern[matched:], handlers, nameNode)\n}", "func (t *Tree) AddIteratively(data int) (err error) {\n\tt.Total += data\n\tt.NodeCount++\n\tif data < 0 {\n\t\treturn ErrPositiveIntegers\n\t}\n\tNodeToAdd := Node{Data: data}\n\tif t.Root == nil {\n\t\tt.Root = &NodeToAdd\n\t\treturn\n\t}\n\tt.addIteratively(t.Root, NodeToAdd)\n\tif t.Root != nil {\n\t\tnewTree := rebalance(t)\n\t\tif newTree.Root != nil {\n\t\t\t*t = newTree\n\t\t}\n\n\t}\n\treturn\n}", "func (t *Tree) AddRecusively(data int) (err error) {\n\tt.Total += data\n\tt.NodeCount++\n\tif data < 0 {\n\t\treturn ErrPositiveIntegers\n\t}\n\tNodeToAdd := Node{Data: data}\n\tif t.Root == nil {\n\t\tt.Root = &NodeToAdd\n\t\treturn\n\t}\n\tt.addRecusively(t.Root, NodeToAdd)\n\tif t.Root != nil {\n\t\tnewTree := rebalance(t)\n\t\tif newTree.Root != nil {\n\t\t\t*t = newTree\n\t\t}\n\n\t}\n\treturn\n}", "func insertTreeNode(rootNode *TreeNode, newTreeNode *TreeNode) {\n\tif newTreeNode.key < rootNode.key {\n\t\tif rootNode.leftNode == nil {\n\t\t\trootNode.leftNode = newTreeNode\n\t\t} else {\n\t\t\tinsertTreeNode(rootNode.leftNode, newTreeNode)\n\t\t}\n\t} else {\n\t\tif rootNode.rightNode == nil {\n\t\t\trootNode.rightNode = newTreeNode\n\t\t} else {\n\t\t\tinsertTreeNode(rootNode.rightNode, newTreeNode)\n\t\t}\n\t}\n}", "func (tree *Tree) Insert(value interface{}) {\n\tcurrent := tree.naiveInsert(value)\n\n\t// Bubble up while the current node's priority is lower than its parent's\n\tfor current.Parent != nil && compare(current.Priority, current.Parent.Priority) < 0 {\n\t\tif current == current.Parent.Left {\n\t\t\tcurrent.Parent.rightRotate()\n\t\t} else {\n\t\t\tcurrent.Parent.leftRotate()\n\t\t}\n\t}\n\tif current.Parent == nil {\n\t\ttree.Root = current\n\t}\n}", "func (log Logger) Root(root Data) Logger {\n\tnewRoot := Data{}\n\tfor k, v := range log.root {\n\t\tnewRoot[k] = v\n\t}\n\tfor k, v := range root {\n\t\tnewRoot[k] = v\n\t}\n\tlog.root = newRoot\n\treturn log\n}", "func (n *Node) add(key interface{}, value interface{}) *Node {\n\tx := newNode()\n\tx.parent = n\n\tx.key = key\n\tx.value = value\n\treturn x\n}", "func (tree *Tree) Put(key interface{}, value interface{}) {\n\tvar insertedNode *Node\n\tif tree.Root == nil {\n\t\t// Assert key is of comparator's type for initial tree\n\t\ttree.Comparator(key, key)\n\t\ttree.Root = &Node{Key: key, Value: value, color: red}\n\t\tinsertedNode = tree.Root\n\t} else {\n\t\tnode := tree.Root\n\t\tloop := true\n\t\tfor loop {\n\t\t\tcompare := tree.Comparator(key, node.Key)\n\t\t\tswitch {\n\t\t\tcase compare == 0:\n\t\t\t\tnode.Key = key\n\t\t\t\tnode.Value = value\n\t\t\t\treturn\n\t\t\tcase compare < 0:\n\t\t\t\tif node.Left == nil {\n\t\t\t\t\tnode.Left = &Node{Key: key, Value: value, color: red}\n\t\t\t\t\tinsertedNode = node.Left\n\t\t\t\t\tloop = false\n\t\t\t\t} else {\n\t\t\t\t\tnode = node.Left\n\t\t\t\t}\n\t\t\tcase compare > 0:\n\t\t\t\tif node.Right == nil {\n\t\t\t\t\tnode.Right = &Node{Key: key, Value: value, color: red}\n\t\t\t\t\tinsertedNode = node.Right\n\t\t\t\t\tloop = false\n\t\t\t\t} else {\n\t\t\t\t\tnode = node.Right\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tinsertedNode.Parent = node\n\t}\n\ttree.insertCase1(insertedNode)\n\ttree.size++\n}", "func (t *Tree) Insert(val int) {\n\tt.Root.insert(val)\n}", "func insert(root *Node) *Node {\n\tfor i := 0; i < 10; i++ {\n\t\tnode := &Node{val: i}\n\t\tif root == nil {\n\t\t\troot = node\n\t\t} else if root.left == nil {\n\t\t\tnode.left, root = root, node\n\t\t} else {\n\t\t\tnode.left, node.right = root.left, root\n\t\t\troot.left = nil\n\t\t\troot = node\n\t\t}\n\t}\n\treturn root\n}", "func (bst *Bst) Root() *Node {\n\treturn bst.root\n}", "func (m *trie) Add(pip *PrefixedIP) {\n\tip := pip.IP()\n\tpl := pip.PrefixLength()\n\tcurrent := m.root\n\tfor i := byte(0); i < pl; i++ {\n\t\tchild := (ip[i/bitsPerByte] >> (bitsPerByte - 1 - i%bitsPerByte)) & 1\n\t\tif current.Children[child] == nil {\n\t\t\tcurrent.Children[child] = &node{}\n\t\t}\n\t\tcurrent = current.Children[child]\n\t\tif current.IsEnd {\n\t\t\treturn\n\t\t}\n\t}\n\tcurrent.IsEnd = true\n}", "func (tree *DNFTree) CreateRoot(phi br.ClauseSet, isFinal bool) int {\n\treturn tree.CreateNodeEntry(phi, 0, isFinal)\n}", "func (tree *LinkedTree) Add(key string, data interface{}) (bool, error) {\n\ttree.mux.Lock()\n\tdefer tree.mux.Unlock()\n\tif tree.root == nil {\n\t\ttree.root = &LinkedTreeNode{\n\t\t\tKey: key,\n\t\t\tData: data,\n\t\t}\n\t\ttree.count = 1\n\t\treturn true, nil\n\t}\n\tok, err := tree.root.Add(key, data)\n\tif ok == true {\n\t\ttree.count++\n\t}\n\treturn ok, err\n}", "func (r *Raft) addNode(id uint64) {\n\t// Your Code Here (3A).\n}", "func (r *Raft) addNode(id uint64) {\n\t// Your Code Here (3A).\n}", "func (tree *Trie) Insert(key string, value interface{}) {\n\tnode := tree.root\n\trunes := []rune(key)\n\tlevel := uint(1)\n\n\tfor _, r := range runes {\n\t\tif child, ok := node.children[r]; ok {\n\t\t\tnode = child\n\t\t} else {\n\t\t\tnew := NewNode(r, node)\n\t\t\tnode.children[r] = new\n\t\t\tnode = new\n\t\t}\n\t\tlevel++\n\t}\n\tnode.data = value\n\tnode.terminal = true\n}", "func (d *Document) Root() Node {\n\treturn Node{0, d.rev, d}\n}", "func insertNode(root *Node, target *Node, key string, n *Node) (*Node, error) {\n\tif target.whatAmI == \"rootleaf\" {\n\t\tnewRoot := new(Node)\n\t\ttarget.whatAmI = \"leaf\"\n\t\tnewRoot.whatAmI = \"root\"\n\t\tnewRoot.records = append(newRoot.records, Record{key: key, value: \"\"})\n\t\tnewRoot.childPtrs = append(newRoot.childPtrs, root, n)\n\t\treturn newRoot, nil\n\t}\n\n\tinsertKeyAndPtr(target, key, n)\n\n\t// Nodes need to split once they get to 6 entries.\n\tif len(target.records) == 6 { //split needed\n\n\t\tsplitKey := target.records[2].key\n\t\tdestinationNode, err := findParent(root, target)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Can't split because...\", err)\n\t\t}\n\n\t\tright := splitNode(root, target, splitKey)\n\n\t\t// Need to make a new root node.\n\t\tif target.whatAmI == \"root\" {\n\t\t\treturn splitRoot(target, right, splitKey), nil\n\t\t}\n\n\t\troot, err = insertNode(root, destinationNode, splitKey, right)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Recursive call to insertNode failed because of \", err)\n\t\t}\n\t}\n\n\treturn root, nil\n}", "func (tree *AcceptTree) Add(next *AcceptEntry) {\n\tif tree.Value == nil {\n\t\ttree.Value = next\n\t} else if next.CompareTo(tree.Value) > 0 {\n\t\tif tree.Left == nil {\n\t\t\ttree.Left = &AcceptTree{Value: next, Size: 1}\n\t\t} else {\n\t\t\ttree.Left.Add(next)\n\t\t}\n\t} else {\n\t\tif tree.Right == nil {\n\t\t\ttree.Right = &AcceptTree{Value: next, Size: 1}\n\t\t} else {\n\t\t\ttree.Right.Add(next)\n\t\t}\n\t}\n\ttree.Size++\n}", "func (node *Node) Root() *Node {\n\tfor node.Parent != nil {\n\t\tnode = node.Parent\n\t}\n\treturn node\n}", "func insertRootRecur(data Comparable, parent *Node, dir int) (ok bool, newNode *Node) {\r\n\tswitch parent.data.Compare(data) {\r\n\tcase equalTo:\r\n\t\treturn false, parent\r\n\tcase greaterThan:\r\n\t\tdir = right\r\n\tcase lessThan:\r\n\t\tdir = left\r\n\t}\r\n\tvar save *Node\r\n\r\n\tif parent.links[dir] == nil { //Found insertion point\r\n\t\tsave = &Node{data: data}\r\n\t\tok = true\r\n\t} else {\r\n\t\tok, save = insertRootRecur(data, parent.links[dir], dir)\r\n\t}\r\n\t//This is simply moving the parent to be the node child\r\n\t// and relink the \"replaced\" child to the parent (where the node is supposed to be link)\r\n\tparent.links[dir] = save.links[flip(dir)] //relink node's child under parent\r\n\tsave.links[flip(dir)] = parent // link parent as node's child\r\n\r\n\treturn ok, save\r\n}", "func (s *DatatypeGeoShape) Tree(tree string) *DatatypeGeoShape {\n\ts.tree = tree\n\treturn s\n}", "func (stateObj *stateObject) updateRoot(db StateDatabase) {\n\t// If nothing changed, don't bother with hashing anything\n\tif stateObj.updateTrie(db) == nil {\n\t\treturn\n\t}\n\t// Track the amount of time wasted on hashing the storage trie\n\tif metrics.EnabledExpensive {\n\t\tdefer func(start time.Time) { stateObj.db.StorageHashes += time.Since(start) }(time.Now())\n\t}\n\tstateObj.data.Root = stateObj.trie.Hash()\n}", "func (CodeNode) AddChild(n Node) {\n}", "func (t *DiskTree) addChild(root *disknode, edgename, key string, value []string) {\n\tnewnode := new(disknode)\n\tnewnode.Value = value\n\tnewnode.Edgename = edgename\n\tnewnode.Key = key\n\tnewnode.Hash = smash(key)\n\tnewnode.Parent = root.Hash\n\tif root.Children == nil {\n\t\troot.Children = make(map[string]string)\n\t}\n\troot.Children[string(edgename[0])] = newnode.Hash\n\tdebugf(\"add disk child %+v\\n\", newnode)\n\tt.write(root)\n\tt.write(newnode)\n}", "func (ft *functionTree) setRoot(f *function) error {\n\tif ft.root != nil {\n\t\treturn errors.New(\"setRoot: root already exists\")\n\t}\n\tft.root = f\n\treturn nil\n}", "func (n *node) Insert(p *Point) octant {\n\tidx := findOctantIdx(n, p)\n\tn.leafs[idx] = n.leafs[idx].Insert(p)\n\tn.updateMassCenter()\n\treturn n\n}", "func (n *ForNode) SetTree(a *Tree) {\n\tn.tree = a\n}", "func treeAdd(repo *git.Repository, tree *git.Tree, key string, valueId *git.Oid) (t *git.Tree, err error) {\n\t/*\n\t** // Primitive but convenient tracing for debugging recursive calls to treeAdd.\n\t** // Uncomment this block for debug output.\n\t**\n\t** var callString string\n\t** if tree != nil {\n\t**\t\tcallString = fmt.Sprintf(\" treeAdd %v:\\t\\t%s\\t\\t\\t= %v\", tree.Id(), key, valueId)\n\t**\t} else {\n\t**\t\tcallString = fmt.Sprintf(\" treeAdd %v:\\t\\t%s\\t\\t\\t= %v\", tree, key, valueId)\n\t**\t}\n\t**\tfmt.Printf(\" %s\\n\", callString)\n\t**\tdefer func() {\n\t**\t\tif t != nil {\n\t**\t\t\tfmt.Printf(\"-> %s => %v\\n\", callString, t.Id())\n\t**\t\t} else {\n\t**\t\t\tfmt.Printf(\"-> %s => %v\\n\", callString, err)\n\t**\t\t}\n\t**\t}()\n\t */\n\tif valueId == nil {\n\t\treturn tree, nil\n\t}\n\tkey = treePath(key)\n\tbase, leaf := path.Split(key)\n\to, err := repo.Lookup(valueId)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar builder *git.TreeBuilder\n\tif tree == nil {\n\t\tbuilder, err = repo.TreeBuilder()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\tbuilder, err = repo.TreeBuilderFromTree(tree)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tdefer builder.Free()\n\t// The specified path has only 1 component (the \"leaf\")\n\tif base == \"\" || base == \"/\" {\n\t\t// If val is a string, set it and we're done.\n\t\t// Any old value is overwritten.\n\t\tif _, isBlob := o.(*git.Blob); isBlob {\n\t\t\tif err := builder.Insert(leaf, valueId, 0100644); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tnewTreeId, err := builder.Write()\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tnewTree, err := lookupTree(repo, newTreeId)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\treturn newTree, nil\n\t\t}\n\t\t// If val is not a string, it must be a subtree.\n\t\t// Return an error if it's any other type than Tree.\n\t\toTree, ok := o.(*git.Tree)\n\t\tif !ok {\n\t\t\treturn nil, fmt.Errorf(\"value must be a blob or subtree\")\n\t\t}\n\t\tvar subTree *git.Tree\n\t\tvar oldSubTree *git.Tree\n\t\tif tree != nil {\n\t\t\toldSubTree, err = treeScope(repo, tree, leaf)\n\t\t\t// FIXME: distinguish \"no such key\" error (which\n\t\t\t// FIXME: distinguish a non-existing previous tree (continue with oldTree==nil)\n\t\t\t// from other errors (abort and return an error)\n\t\t\tif err == nil {\n\t\t\t\tdefer oldSubTree.Free()\n\t\t\t}\n\t\t}\n\t\t// If that subtree already exists, merge the new one in.\n\t\tif oldSubTree != nil {\n\t\t\tsubTree = oldSubTree\n\t\t\tfor i := uint64(0); i < oTree.EntryCount(); i++ {\n\t\t\t\tvar err error\n\t\t\t\te := oTree.EntryByIndex(i)\n\t\t\t\tsubTree, err = treeAdd(repo, subTree, e.Name, e.Id)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\t// If the key is /, we're replacing the current tree\n\t\tif key == \"/\" {\n\t\t\treturn subTree, nil\n\t\t}\n\n\t\t// if the subtree is nill at this point, then we've\n\t\t// hit an error, and should error now -- tycho\n\t\tif subTree == nil {\n\t\t\treturn nil, fmt.Errorf(\"subTree for %s nil, aborting\", key)\n\t\t}\n\n\t\t// Otherwise we're inserting into the current tree\n\t\terr := builder.Insert(leaf, subTree.Id(), 040000)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tnewTreeId, err := builder.Write()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tnewTree, err := lookupTree(repo, newTreeId)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn newTree, nil\n\t}\n\tsubtree, err := treeAdd(repo, nil, leaf, valueId)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn treeAdd(repo, tree, base, subtree.Id())\n}", "func (t *Tree) Insert (value, data string) error {\n\tif t.Root == nil {\n\t\tt.Root = &Node{Value: value, Data: data}\n\t\treturn nil\n\t}\n\treturn t.Root.Insert(value, data)\t// this may throw error\n}", "func (tp *Template) Root(name string) *Template {\n\ttp.root = name\n\treturn tp\n}", "func (self *TNode) addnode(aParent *TNode, aNodes []*TNode, i int, aIsHook bool) *TNode {\n\tif len(aParent.Children) == 0 {\n\t\taParent.Children = make([]*TNode, 0)\n\t}\n\n\t// 如果:找到[已经注册]的分支节点则从该节继续[查找/添加]下一个节点\n\tfor _, n := range aParent.Children {\n\t\tif n.Equal(aNodes[i]) {\n\t\t\t// 如果:插入的节点层级已经到末尾,则为该节点注册路由\n\t\t\tif i == len(aNodes)-1 {\n\t\t\t\t// 原始路由会被替换\n\t\t\t\tif aIsHook {\n\t\t\t\t\tn.Route.CombineController(aNodes[i].Route)\n\t\t\t\t} else {\n\t\t\t\t\tn.Route = aNodes[i].Route\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn n\n\t\t}\n\t}\n\n\t// 如果:该节点没有对应分支则插入同级的aNodes为新的分支\n\taParent.Children = append(aParent.Children, aNodes[i])\n\tsort.Sort(aParent.Children)\n\treturn aNodes[i]\n}", "func (e *Tree) Add(err error) *Tree {\n\tif err == (error)(nil) {\n\t\treturn nil\n\t}\n\tif e == (*Tree)(nil) {\n\t\treturn nil\n\t}\n\tif et, ok := err.(Tree); ok {\n\t\terr = &et\n\t}\n\te.errs = append(e.errs, err)\n\treturn e\n}", "func (rt *RTree) insert(height int, newEntry entry, first bool) error { //first is needed in case of overflowTreatment, it should normaly be true\n\t//[I1]\tChooseSubtree with height as a parameter to find the node N\n\tn := rt.chooseSubtree(newEntry.mbr, height)\n\t//If an internal entry is re-inserted, the node's parent pointer must be updated\n\tif height >= 1 {\n\t\tnewEntry.child.parent = n\n\t}\n\t//[I2]\tAppend newEntry to n if room, else call OverflowTreatment [for reinsertion or split]\n\tn.entries = append(n.entries, newEntry)\n\tif len(n.entries) >= RTree_M+1 { // n is full -> call overflowTreatment\n\t\tdidSplit, nn := rt.overflowTreatment(n, first) //OT finds the appropriate height from n.height\n\t\tif didSplit {\n\t\t\t//[I3]\tif OverflowTreatment was called and a split was performed: propagate OT upwards\n\t\t\tif nn.height == rt.root.height { // if root was split: create a new root\n\t\t\t\tnewRoot := node{\n\t\t\t\t\tparent: nil,\n\t\t\t\t\tentries: make([]entry, 0, RTree_M+1),\n\t\t\t\t\theight: rt.root.height + 1,\n\t\t\t\t}\n\t\t\t\tnEntry := entry{mbr: n.recalculateMBR(), child: n}\n\t\t\t\tnnEntry := entry{mbr: nn.recalculateMBR(), child: nn}\n\t\t\t\tnewRoot.entries = append(newRoot.entries, nEntry)\n\t\t\t\tnewRoot.entries = append(newRoot.entries, nnEntry)\n\t\t\t\tn.parent = &newRoot\n\t\t\t\tnn.parent = &newRoot\n\t\t\t\trt.root = &newRoot\n\t\t\t\t//fmt.Printf(\"Root was split...^ new height is %d\\n\", newRoot.height)\n\t\t\t\treturn nil //The root has no MBR, so there is no need to adjust any MBRs\n\t\t\t}\n\t\t\t// n was split into n & nn -> insert nn into the tree at the same height\n\t\t\terr := rt.insert(nn.height+1, entry{mbr: nn.recalculateMBR(), child: nn}, true)\n\t\t\tCheckErr(err, \"failed to insert nn\")\n\t\t}\n\t}\n\t//[I4]\tAdjust all MBR in the insertion path\n\tfor n.height < rt.root.height {\n\t\tpIdx, err := n.parentEntriesIdx()\n\t\tCheckErr(err, \"insert had some trouble adjusting the MBR...\")\n\t\tn.parent.entries[pIdx].mbr = n.recalculateMBR()\n\t\tn = n.parent\n\t}\n\treturn nil\n}", "func (t *Tree) AddNode(node *Tree) {\n\tnode.parent = t\n\tt.nodes = append(t.nodes, node)\n\tsort.Sort(byOrder(t.nodes))\n}", "func (t *Trie) Add(keyword string) {\n\tchars := []rune(keyword)\n\tif len(chars) == 0 {\n\t\treturn\n\t}\n\n\tt.mu.Lock()\n\tnode := t.root\n\tfor _, char := range chars {\n\t\tif _, ok := node.children[char]; !ok {\n\t\t\tnode.children[char] = newTrieNode()\n\t\t}\n\t\tnode = node.children[char]\n\t}\n\tnode.end = true\n\tt.mu.Unlock()\n}", "func (novis *Novis) Add(name, path string, params ...string) *Branch {\n\treturn novis.Root.Add(name, path, params...)\n}", "func (rbTree *RBTree)insert(curr *treeNode, prev *treeNode, key MapKey, value *MapData) {\n\t// replace value\n\tif curr != nil {\n\t\tcurr.Value = value\n\t\treturn\n\t}\n\t// new node\n\tvar node = &treeNode{\n\t\tleft: nil,\n\t\tright: nil,\n\t\tKey: key,\n\t\tValue: value,\n\t}\n\tif curr == prev {\n\t\t// insert root node\n\t\tnode.father = nil\n\t\trbTree.root = node\n\t\trbTree.mLeft = node\n\t\trbTree.mRight = node\n\t} else if key < prev.Key {\n\t\t// insert prev's left\n\t\tprev.left = node\n\t\tnode.father = prev\n\t\tif prev == rbTree.mLeft {\n\t\t\trbTree.mLeft = node\n\t\t}\n\t} else {\n\t\t// insert prev's right\n\t\tprev.right = node\n\t\tnode.father = prev\n\t\tif prev == rbTree.mRight {\n\t\t\trbTree.mRight = node\n\t\t}\n\t}\n\trbTree.Size++\n\trbTree.insertReshape(node)\n}", "func (bst BST) _add(value Producto, tmp *Node) *Node {\n\tif tmp == nil {\n\t\treturn &Node{value: value}\n\t} else if value.Codigo == tmp.value.Codigo{\n\t\ttmp.value.Cantidad = tmp.value.Cantidad + value.Cantidad\n\t} else if value.Codigo > tmp.value.Codigo {\n\t\ttmp.right = bst._add(value, tmp.right)\n\t} else {\n\t\ttmp.left = bst._add(value, tmp.left)\n\t}\n\treturn tmp\n}" ]
[ "0.6451382", "0.64198977", "0.6349391", "0.62001956", "0.61917245", "0.6170597", "0.61382675", "0.61298287", "0.61081576", "0.6092689", "0.596722", "0.5927524", "0.5906705", "0.5861521", "0.5852074", "0.5836093", "0.5803622", "0.5765342", "0.573831", "0.57127416", "0.564222", "0.56267595", "0.5611519", "0.5601563", "0.5578302", "0.5577062", "0.5565886", "0.55653137", "0.5564152", "0.5561961", "0.5555352", "0.5547453", "0.55437964", "0.5542539", "0.5540922", "0.5535473", "0.5502173", "0.54997814", "0.54945487", "0.548141", "0.54795355", "0.5475852", "0.5468327", "0.5467736", "0.54583246", "0.545151", "0.5445599", "0.5445445", "0.5429577", "0.54269874", "0.5422205", "0.54113746", "0.54007304", "0.5395951", "0.53893524", "0.53851974", "0.53836703", "0.5381693", "0.53697455", "0.5353764", "0.5346721", "0.5337983", "0.5336984", "0.5332794", "0.53281057", "0.532743", "0.5309337", "0.5306478", "0.53056717", "0.5303607", "0.52969545", "0.52939254", "0.52928627", "0.5288234", "0.5287362", "0.5287362", "0.5274416", "0.5265579", "0.5260424", "0.5257141", "0.5255917", "0.52556", "0.52470624", "0.52445835", "0.5244312", "0.52413636", "0.52313685", "0.5229657", "0.52237344", "0.52192837", "0.5217219", "0.5216779", "0.52134037", "0.521009", "0.5208987", "0.5202762", "0.5201356", "0.519634", "0.5191525", "0.51854557" ]
0.6134371
7
graphStructure writes the tree using owners for structure
func (p TreeWriter) graphStructure(nodes []*yaml.RNode) error { resourceToOwner := map[string]*node{} root := &node{} // index each of the nodes by their owner for _, n := range nodes { ownerVal, err := ownerToString(n) if err != nil { return err } var owner *node if ownerVal == "" { // no owner -- attach to the root owner = root } else { // owner found -- attach to the owner var found bool owner, found = resourceToOwner[ownerVal] if !found { // initialize the owner if not found resourceToOwner[ownerVal] = &node{p: p} owner = resourceToOwner[ownerVal] } } nodeVal, err := nodeToString(n) if err != nil { return err } val, found := resourceToOwner[nodeVal] if !found { // initialize the node if not found -- may have already been initialized if it // is the owner of another node resourceToOwner[nodeVal] = &node{p: p} val = resourceToOwner[nodeVal] } val.RNode = n owner.children = append(owner.children, val) } for k, v := range resourceToOwner { if v.RNode == nil { return fmt.Errorf( "owner '%s' not found in input, but found as an owner of input objects", k) } } // print the tree tree := treeprint.New() if err := root.Tree(tree); err != nil { return err } _, err := io.WriteString(p.Writer, tree.String()) return err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (q *Qualifier) Graph(w io.Writer) {\n\tfmt.Fprintf(w, \"digraph {\\n\")\n\tq.root.Graph(w, 0, \"[root]\")\n\tfmt.Fprintf(w, \"}\\n\")\n}", "func WriteTree(writer io.Writer, hierarchy *Hierarchy, includeEmpty bool) {\n\ttree := assembleTree(hierarchy)\n\tkeys := make([]string, len(tree))\n\ti := 0\n\tfor k := range tree {\n\t\tkeys[i] = k\n\t\ti++\n\t}\n\tsort.Strings(keys)\n\tfor _, key := range keys {\n\t\tb := tree[key]\n\t\twriteBranch(writer, b, \"\", hierarchy, includeEmpty)\n\t}\n}", "func (p *GameTree) writeTree(w *bufio.Writer, n TreeNodeIdx, needs bool, nMov int, nMovPerLine int) (err error) {\n\tdefer u(tr(\"writeTree\"))\n\tif needs == true {\n\t\tif nMov > 0 {\n\t\t\terr = w.WriteByte('\\n')\n\t\t\tnMov = 0\n\t\t}\n\t\terr = w.WriteByte('(')\n\t}\n\tif err == nil {\n\t\tif nMov == nMovPerLine {\n\t\t\terr = w.WriteByte('\\n')\n\t\t\tnMov = 0\n\t\t}\n\t\terr = w.WriteByte(';')\n\t\t// write the node\n\t\ttyp := p.treeNodes[n].TNodType\n\t\tswitch typ {\n\t\tcase GameInfoNode:\n\t\t\t// fmt.Println(\"writing GameInfoNode\\n\")\n\t\t\terr = p.writeProperties(w, n, true)\n\t\tcase InteriorNode:\n\t\t\t// fmt.Println(\"writing InteriorNode\\n\")\n\t\t\terr = p.writeProperties(w, n, false)\n\t\tcase BlackMoveNode:\n\t\t\t_, err = w.WriteString(\"B[\")\n\t\t\t_, err = w.Write(SGFCoords(ah.NodeLoc(p.treeNodes[n].propListOrNodeLoc), p.IsFF4()))\n\t\t\terr = w.WriteByte(']')\n\t\t\tnMov += 1\n\t\tcase WhiteMoveNode:\n\t\t\t_, err = w.WriteString(\"W[\")\n\t\t\t_, err = w.Write(SGFCoords(ah.NodeLoc(p.treeNodes[n].propListOrNodeLoc), p.IsFF4()))\n\t\t\terr = w.WriteByte(']')\n\t\t\tnMov += 1\n\t\tdefault:\n\t\t\tfmt.Println(\"*** unsupported TreeNodeType in writeTree\")\n\t\t\terr = errors.New(\"writeTree: unsupported TreeNodeType\" + strconv.FormatInt(int64(typ), 10))\n\t\t\treturn err\n\t\t}\n\t\tif err == nil {\n\t\t\t// write the children\n\t\t\tlastCh := p.treeNodes[n].Children\n\t\t\tif lastCh != nilTreeNodeIdx && err == nil {\n\t\t\t\tch := p.treeNodes[lastCh].NextSib\n\t\t\t\tchNeeds := (lastCh != ch)\n\t\t\t\terr = p.writeTree(w, ch, chNeeds, nMov, nMovPerLine)\n\t\t\t\tfor ch != lastCh && err == nil {\n\t\t\t\t\tch = p.treeNodes[ch].NextSib\n\t\t\t\t\t//\t\t\t\t\tnMov += 1\n\t\t\t\t\terr = p.writeTree(w, ch, chNeeds, nMov, nMovPerLine)\n\t\t\t\t}\n\t\t\t}\n\t\t\tif (err == nil) && (needs == true) {\n\t\t\t\terr = w.WriteByte(')')\n\t\t\t}\n\t\t}\n\t}\n\treturn err\n}", "func Marshal(g *graph.DirectedGraph) ([]byte, error) {\n\tvar b bytes.Buffer\n\n\t// Static graph configuration attributes\n\n\tb.WriteString(\"strict digraph bridge {\\n\")\n\tb.WriteByte('\\n')\n\tb.WriteString(\"graph [\\n\")\n\tb.WriteString(\" rankdir=LR\\n\")\n\tb.WriteString(\"]\\n\")\n\tb.WriteByte('\\n')\n\tb.WriteString(\"node [\\n\")\n\tb.WriteString(\" fontname=\\\"Helvetica\\\"\\n\")\n\tb.WriteString(\" shape=plain\\n\")\n\tb.WriteString(\"]\\n\")\n\tb.WriteByte('\\n')\n\n\t// Vertices\n\n\t// Index of vertices already converted to node, for faster access\n\t// during sorting of edges.\n\t// The keys used in the map are also the ones used in the graph.DirectedGraph\n\tvertIndex := make(map[interface{}]*node)\n\n\tsortedNodes := make(nodeList, 0, len(g.Vertices()))\n\tfor k, v := range g.Vertices() {\n\t\tn := graphVertexToNode(v)\n\t\tvertIndex[k] = n\n\n\t\tsortedNodes = append(sortedNodes, n)\n\t\tsort.Sort(sortedNodes)\n\t}\n\n\tfor _, n := range sortedNodes {\n\t\tdotN, err := n.marshalDOT()\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"marshaling node to DOT: %w\", err)\n\t\t}\n\t\tb.Write(dotN)\n\t}\n\n\tb.WriteByte('\\n')\n\n\t// Edges\n\n\tsortedDownEdges := make(downEdgesList, 0, len(g.DownEdges()))\n\tfor tailVertKey, headVerts := range g.DownEdges() {\n\t\ttailNode := vertIndex[tailVertKey]\n\n\t\tsortedHeadNodes := make(nodeList, 0, len(headVerts))\n\t\tfor headVertKey := range headVerts {\n\t\t\tsortedHeadNodes = append(sortedHeadNodes, vertIndex[headVertKey])\n\t\t}\n\t\tsort.Sort(sortedHeadNodes)\n\n\t\tsortedDownEdges = append(sortedDownEdges, downEdges{\n\t\t\ttail: tailNode,\n\t\t\theads: sortedHeadNodes,\n\t\t})\n\t}\n\tsort.Sort(sortedDownEdges)\n\n\tfor _, e := range sortedDownEdges {\n\t\tb.WriteString(e.tail.id() + \" -> {\")\n\t\tfor _, h := range e.heads {\n\t\t\tb.WriteByte(' ')\n\t\t\tb.WriteString(h.id())\n\t\t}\n\t\tb.WriteString(\" }\\n\")\n\t}\n\n\tb.WriteByte('\\n')\n\n\tb.WriteString(\"}\\n\")\n\n\treturn b.Bytes(), nil\n}", "func (p TreeWriter) Write(nodes []*yaml.RNode) error {\n\tswitch p.Structure {\n\tcase TreeStructurePackage:\n\t\treturn p.packageStructure(nodes)\n\tcase TreeStructureGraph:\n\t\treturn p.graphStructure(nodes)\n\t}\n\n\t// If any resource has an owner reference, default to the graph structure. Otherwise, use package structure.\n\tfor _, node := range nodes {\n\t\tif owners, _ := node.Pipe(yaml.Lookup(\"metadata\", \"ownerReferences\")); owners != nil {\n\t\t\treturn p.graphStructure(nodes)\n\t\t}\n\t}\n\treturn p.packageStructure(nodes)\n}", "func writeJSON(search string, fileName string, nodeArr []nodeJSONStruct, edgeArr []edgeStruct, searchArr []*admin.Group) {\n var (\n jsonFile *os.File\n encoder *json.Encoder\n )\n\n if nodeArr != nil && edges.arr != nil {\n jsonFile = createJSONFile(outputJSONParentsDir, fileName)\n defer jsonFile.Close() \n\n // https://medium.com/eaciit-engineering/better-way-to-read-and-write-json-file-in-golang-9d575b7254f2 \n // also https://www.golangprograms.com/golang-writing-struct-to-json-file.html\n encoder = json.NewEncoder(jsonFile)\n encoder.Encode(parentsJSONStruct{SearchIdentity: search, NodeArr: nodeArr, EdgeArr: edgeArr})\n\n } else if searchArr != nil {\n jsonFile = createJSONFile(outputJSONSearchDir, fileName)\n defer jsonFile.Close() \n\n // https://medium.com/eaciit-engineering/better-way-to-read-and-write-json-file-in-golang-9d575b7254f2 \n // also https://www.golangprograms.com/golang-writing-struct-to-json-file.html\n encoder = json.NewEncoder(jsonFile)\n encoder.Encode(searchJSONStruct{SearchPrefix: search, GroupArr: searchArr})\n }\n}", "func (n *NetworkBuilder) NetworkGraph(w io.Writer) error {\n\tnodes := make(map[string]dot.Node)\n\tnodesByID := make(map[string]dot.Node)\n\tgraph := dot.NewGraph(dot.Directed)\n\n\tfor _, nr := range n.NetResources {\n\t\tnode := graph.Node(strings.Join([]string{nr.NodeId, nr.Iprange.String()}, \"\\n\")).Box()\n\t\t// set special style for \"hidden\" nodes\n\t\tif len(nr.PubEndpoints) == 0 {\n\t\t\tnode.Attr(\"style\", \"dashed\")\n\t\t\tnode.Attr(\"color\", \"blue\")\n\t\t\tgraph.AddToSameRank(\"hidden nodes\", node)\n\t\t}\n\t\tnodes[nr.WireguardPublicKey] = node\n\t\tnodesByID[nr.NodeId] = node\n\t}\n\n\t// add external access\n\tfor _, ea := range n.AccessPoints {\n\t\tnode := graph.Node(strings.Join([]string{\"External network\", ea.Subnet.String()}, \"\\n\")).Box()\n\t\t// set style for hidden nodes\n\t\tnode.Attr(\"style\", \"dashed\")\n\t\tnode.Attr(\"color\", \"green\")\n\t\tgraph.AddToSameRank(\"external access\", node)\n\t\t// add link to access point\n\t\tedge := graph.Edge(node, nodesByID[ea.NodeID], n.Iprange.String())\n\t\tif ea.IP4 {\n\t\t\tedge.Attr(\"color\", \"blue\")\n\t\t}\n\t\tnodes[ea.WGPublicKey] = node\n\t}\n\n\tfor _, nr := range n.NetResources {\n\t\tfor _, peer := range nr.Peers {\n\t\t\tallowedIPs := make([]string, 0, len(peer.AllowedIprange)/2)\n\t\t\tfor _, aip := range peer.AllowedIprange {\n\t\t\t\tif !isCGN(aip) {\n\t\t\t\t\tallowedIPs = append(allowedIPs, aip.String())\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tedge := graph.Edge(nodes[nr.WireguardPublicKey], nodes[peer.PublicKey], strings.Join(allowedIPs, \"\\n\"))\n\t\t\tif peer.Endpoint == \"\" {\n\t\t\t\t// connections to this peer are IPv4 -> blue, and can not be initiated by this node -> dashed\n\t\t\t\tedge.Attr(\"color\", \"blue\").Attr(\"style\", \"dashed\")\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif net.ParseIP(peer.Endpoint[:strings.LastIndex(peer.Endpoint, \":\")]).To4() != nil {\n\t\t\t\t// IPv4 connection -> blue\n\t\t\t\tedge.Attr(\"color\", \"blue\")\n\t\t\t}\n\t\t}\n\t}\n\n\tgraph.Write(w)\n\treturn nil\n}", "func (t *BPTree) WriteNodes(rwMode RWMode, syncEnable bool, flag int) error {\n\tvar (\n\t\tn *Node\n\t\ti int\n\t\terr error\n\t)\n\n\tfd, err := os.OpenFile(t.Filepath, os.O_CREATE|os.O_RDWR, 0644)\n\tdefer fd.Close()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tqueue = nil\n\n\tenqueue(t.root)\n\n\tfor queue != nil {\n\t\tn = dequeue()\n\n\t\t_, err := t.WriteNode(n, -1, syncEnable, fd)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif n != nil {\n\t\t\tif !n.isLeaf {\n\t\t\t\tfor i = 0; i <= n.KeysNum; i++ {\n\t\t\t\t\tc, _ := n.pointers[i].(*Node)\n\t\t\t\t\tenqueue(c)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (decTree *Tree) WriteTree(filename string) {\n\tfile, err := os.Create(filename)\n\tif err != nil {\n\t\tfmt.Println(\"Error opening output file: \", filename)\n\t\treturn\n\t}\n\n\tcurrNode := decTree\n\tvar treeStack []*Tree\n\n\ttreeLen := 1\n\tfor treeLen != 0 {\n\t\tfile.WriteString(nodeToStr(currNode.Details))\n\n\t\tif currNode.Details.Leaf == false {\n\t\t\ttreeStack = append(treeStack, currNode.Right)\n\t\t\tcurrNode = currNode.Left\n\t\t\ttreeLen++\n\t\t} else {\n\t\t\t//get the length of the tree and set curr to the last element in the list\n\t\t\ttreeLen--\n\n\t\t\tif treeLen > 0 {\n\t\t\t\tcurrNode, treeStack = treeStack[treeLen-1], treeStack[:treeLen-1]\n\t\t\t}\n\t\t}\n\t}\n\n\tfile.Close()\n}", "func DrawGraph(filename string, s spn.SPN) {\n\tfile, err := os.Create(filename)\n\n\tif err != nil {\n\t\tfmt.Printf(\"Error. Could not create file [%s].\\n\", filename)\n\t\tpanic(err)\n\t}\n\tdefer file.Close()\n\n\tfmt.Fprintf(file, \"graph {\\n\")\n\n\t// If the SPN is itself an univariate distribution, create a graph with a single node.\n\tif s.Type() == \"leaf\" {\n\t\tfmt.Fprintf(file, \"X1 [label=<X<sub>1</sub>>,shape=circle];\\n\")\n\t\tfmt.Fprintf(file, \"}\")\n\t\tfile.Close()\n\t\treturn\n\t}\n\n\t// Else, BFS the SPN and write nodes to filename.\n\tnvars, nsums, nprods := 0, 0, 0\n\tqueue := common.Queue{}\n\tqueue.Enqueue(&BFSPair{Spn: s, Pname: \"\", Weight: -1.0})\n\tfor !queue.Empty() {\n\t\tcurrpair := queue.Dequeue().(*BFSPair)\n\t\tcurr, pname, pw := currpair.Spn, currpair.Pname, currpair.Weight\n\t\tch := curr.Ch()\n\t\tnch := len(ch)\n\n\t\tname := \"N\"\n\t\tcurrt := curr.Type()\n\n\t\t// In case it is a sum node. Else product node.\n\t\tif currt == \"sum\" {\n\t\t\tname = fmt.Sprintf(\"S%d\", nsums)\n\t\t\tfmt.Fprintf(file, \"%s [label=\\\"+\\\",shape=circle];\\n\", name)\n\t\t\tnsums++\n\t\t} else if currt == \"product\" {\n\t\t\tname = fmt.Sprintf(\"P%d\", nprods)\n\t\t\tfmt.Fprintf(file, \"%s [label=<&times;>,shape=circle];\\n\", name)\n\t\t\tnprods++\n\t\t}\n\n\t\t// If pname is empty, then it is the root node. Else, link parent node to current node.\n\t\tif pname != \"\" {\n\t\t\tif pw >= 0 {\n\t\t\t\tfmt.Fprintf(file, \"%s -- %s [label=\\\"%.3f\\\"];\\n\", pname, name, pw)\n\t\t\t} else {\n\t\t\t\tfmt.Fprintf(file, \"%s -- %s\\n\", pname, name)\n\t\t\t}\n\t\t}\n\n\t\tvar w []float64\n\t\tif curr.Type() == \"sum\" {\n\t\t\tw = (curr.(*spn.Sum).Weights())\n\t\t}\n\t\t// For each children, run the BFS.\n\t\tfor i := 0; i < nch; i++ {\n\t\t\tc := ch[i]\n\n\t\t\t// If leaf, then simply write to the graphviz dot file. Else, recurse the BFS.\n\t\t\tif c.Type() == \"leaf\" {\n\t\t\t\tcname := fmt.Sprintf(\"X%d\", nvars)\n\t\t\t\tfmt.Fprintf(file, \"%s [label=<X<sub>%d</sub>>,shape=circle];\\n\", cname, c.Sc()[0])\n\t\t\t\tnvars++\n\t\t\t\tif currt == \"sum\" {\n\t\t\t\t\tfmt.Fprintf(file, \"%s -- %s [label=\\\"%.3f\\\"]\\n\", name, cname, w[i])\n\t\t\t\t} else {\n\t\t\t\t\tfmt.Fprintf(file, \"%s -- %s\\n\", name, cname)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\ttw := -1.0\n\t\t\t\tif w != nil {\n\t\t\t\t\ttw = w[i]\n\t\t\t\t}\n\t\t\t\tqueue.Enqueue(&BFSPair{Spn: c, Pname: name, Weight: tw})\n\t\t\t}\n\t\t}\n\t}\n\n\tfmt.Fprintf(file, \"}\")\n}", "func (g *Graph) writeGraph(output io.Writer) error {\n\theader := \"p tw \" + strconv.Itoa(g.numvert) + \" \" + strconv.Itoa(g.numedge) + \"\\n\"\n\tio.WriteString(output, header)\n\n\tline := \"\"\n\tfor edge, value := range g.edges {\n\t\tif value {\n\t\t\tline = strconv.Itoa(edge.from.vert) + \" \" + strconv.Itoa(edge.to.vert) + \"\\n\"\n\t\t\tio.WriteString(output, line)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (node *URLNode) WriteTree(writer io.Writer) {\n\tif _, err := writer.Write([]byte(node.GenerateTree())); err != nil {\n\t\tlog.Error(err)\n\t}\n}", "func Map(w io.Writer, is interface{}, spec *Spec) error {\n\n\tvar vertexAppendix string\n\tvar edgeAppendix string\n\n\tfmt.Fprintln(w, \"strict graph {\")\n\tfmt.Fprintln(w, \" node [shape=\\\"circle\\\"];\")\n\n\tswitch is.(type) {\n\n\tcase *graph.Graph:\n\n\t\t// Initialize nodes\n\t\tfor _, v := range is.(*graph.Graph).V {\n\n\t\t\t// Visualize mst\n\t\t\tif spec.MST == true {\n\t\t\t\tvertexAppendix = \" color=\\\"red\\\" style=\\\"bold,filled\\\"\"\n\t\t\t}\n\n\t\t\tfmt.Fprintf(w, \" %d [label=\\\"%+v\\\"%s]\\n\", v.Serial, v.ID, vertexAppendix)\n\n\t\t}\n\n\t\t// Initialize edges\n\t\tfor _, e := range is.(*graph.Graph).E {\n\n\t\t\t// Visualize mst\n\t\t\tif (is.(*graph.Graph).V[e.EndpointA].GetEdgeByEndpoint(e.EndpointB).InMST == true ||\n\t\t\t\tis.(*graph.Graph).V[e.EndpointB].GetEdgeByEndpoint(e.EndpointA).InMST == true) && spec.MST == true {\n\t\t\t\tedgeAppendix = \" color=\\\"red\\\" style=\\\"bold\\\"\"\n\t\t\t}\n\n\t\t\tif is.(*graph.Graph).Type == graph.UNDIRECTED {\n\t\t\t\tfmt.Fprintf(w, \" %d -- %d [label=\\\"%d\\\"%s]\\n\", e.EndpointA, e.EndpointB, e.Weight, edgeAppendix)\n\t\t\t} else {\n\t\t\t\tfmt.Fprintf(w, \" %d -> %d [label=\\\"%d\\\"%s]\\n\", e.EndpointA, e.EndpointB, e.Weight, edgeAppendix)\n\t\t\t}\n\n\t\t\tedgeAppendix = \"\"\n\n\t\t}\n\t// case []graph.SEdge:\n\t// case []*graph.Node:\n\tdefault:\n\t\treturn fmt.Errorf(\"Unsupported Type\")\n\t}\n\n\tfmt.Fprintln(w, \"}\")\n\n\treturn nil\n\n}", "func Marshal(g *ag.Graph) ([]byte, error) {\n\tgv, err := graphviz.BuildGraph(g, graphviz.Options{ColoredTimeSteps: true})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn []byte(gv.String()), nil\n}", "func (bpt *BplusTree) writeLayout() {\n\tleafIdx := 0\n\tnodeIdx := 0\n\tlevelIdx := 0\n\n\tif !bpt.initialized || bpt.rootKey.IsNil() {\n\t\treturn\n\t}\n\n\trootNode, _ := bpt.fetch(bpt.rootKey)\n\tif rootNode == nil {\n\t\tglog.Errorf(\"failed to fetch root key: %v. can not print the tree.\",\n\t\t\tbpt.rootKey)\n\t\treturn\n\t}\n\tglog.Infof(\"dumping the tree layout.. numChildren: %d\\n\",\n\t\tlen(rootNode.Children))\n\tnodeList := rootNode.Children\n\tnodeLensList := make([]int, 1)\n\tnodeLensList[0] = len(rootNode.Children)\n\tnumElems := nodeLensList[0]\n\tnumNodesAtLevel := 0\n\tprintLevel := true\n\tglog.Infof(\"level -- 0 <root: %v>\\n\", rootNode)\n\tif rootNode.IsLeaf {\n\t\treturn\n\t}\n\tfor i := 0; i < numElems; i++ {\n\t\tif printLevel {\n\t\t\tglog.Infof(\"level -- %d \", levelIdx+1)\n\t\t\tprintLevel = false\n\t\t}\n\t\tnode, _ := bpt.fetch(nodeList[i].NodeKey)\n\t\tif node == nil {\n\t\t\tglog.Errorf(\"failed to fetch root key: %v\", nodeList[i].NodeKey)\n\t\t\treturn\n\t\t}\n\n\t\tif node.IsLeaf {\n\t\t\tglog.Infof(\"level:%d <tree-L-node :%d, node: %v> \", levelIdx+1, leafIdx, node)\n\t\t\tleafIdx++\n\t\t} else {\n\t\t\tglog.Infof(\"level:%d <tree-I-node :%d, node: %v> \", levelIdx+1, nodeIdx, node)\n\t\t\tnodeList = append(nodeList, node.Children...)\n\t\t\tnumElems += len(node.Children)\n\t\t\tnumNodesAtLevel += len(node.Children)\n\t\t}\n\t\tnodeIdx++\n\t\tif nodeIdx >= nodeLensList[levelIdx] {\n\t\t\tnodeLensList = append(nodeLensList, numNodesAtLevel)\n\t\t\tlevelIdx++\n\t\t\tnodeIdx = 0\n\t\t\tnumNodesAtLevel = 0\n\t\t\tglog.Infof(\"\\n\")\n\t\t\tprintLevel = true\n\t\t}\n\t}\n\tglog.Infof(\"done.. dumping the layout\\n\")\n\tglog.Infof(\"----------------------------\\n\")\n}", "func insertAncestryNodes(conn types.TGConnection, gof types.TGGraphObjectFactory) map[string]types.TGNode {\n\tfmt.Println(\">>>>>>> Entering InsertAncestryNodes: Insert Few Family Nodes with individual properties <<<<<<<\")\n\n\tvar houseMemberTable = make(map[string]types.TGNode, 0)\n\n\tgmd, err := conn.GetGraphMetadata(true)\n\tif err != nil {\n\t\tfmt.Println(\">>>>>>> Returning from InsertAncestryNodes - error during conn.GetGraphMetadata <<<<<<<\")\n\t\treturn nil\n\t}\n\n\ttestNodeType, err := gmd.GetNodeType(\"houseMemberType\")\n\tif err != nil {\n\t\tfmt.Println(\">>>>>>> Returning from InsertAncestryNodes - error during conn.GetNodeType('houseMemberType') <<<<<<<\")\n\t\treturn nil\n\t}\n\tif testNodeType != nil {\n\t\tfmt.Printf(\">>>>>>> 'houseMemberType' is found with %d attributes <<<<<<<\\n\", len(testNodeType.GetAttributeDescriptors()))\n\t} else {\n\t\tfmt.Println(\">>>>>>> 'houseMemberType' is not found from meta data fetch <<<<<<<\")\n\t\treturn nil\n\t}\n\n\tfor _, houseMember := range HouseMemberData {\n\t\tnode1, err := gof.CreateNodeInGraph(testNodeType)\n\t\tif err != nil {\n\t\t\tfmt.Println(\">>>>>>> Returning from InsertAncestryNodes - error during gof.CreateNode(node1) <<<<<<<\")\n\t\t\treturn nil\n\t\t}\n\t\t_ = node1.SetOrCreateAttribute(\"memberName\", houseMember.MemberName)\n\t\t_ = node1.SetOrCreateAttribute(\"crownName\", houseMember.CrownName)\n\t\t_ = node1.SetOrCreateAttribute(\"houseHead\", houseMember.HouseHead)\n\t\t_ = node1.SetOrCreateAttribute(\"yearBorn\", houseMember.YearBorn)\n\t\t_ = node1.SetOrCreateAttribute(\"yearDied\", houseMember.YearDied)\n\t\t_ = node1.SetOrCreateAttribute(\"crownTitle\", houseMember.CrownTitle)\n\n\t\tif houseMember.ReignStart != \"\" {\n\t\t\treignStart, _ := time.Parse(\"02 Jan 2006\", houseMember.ReignStart)\n\t\t\t_ = node1.SetOrCreateAttribute(\"reignStart\", reignStart)\n\t\t} else {\n\t\t\t_ = node1.SetOrCreateAttribute(\"reignStart\", nil)\n\t\t}\n\n\t\tif houseMember.ReignStart != \"\" {\n\t\t\treignEnd, _ := time.Parse(\"02 Jan 2006\", houseMember.ReignEnd)\n\t\t\t_ = node1.SetOrCreateAttribute(\"reignEnd\", reignEnd)\n\t\t} else {\n\t\t\t_ = node1.SetOrCreateAttribute(\"reignEnd\", nil)\n\t\t}\n\n\t\terr = conn.InsertEntity(node1)\n\t\tif err != nil {\n\t\t\tfmt.Println(\">>>>>>> Returning from InsertAncestryNodes w/ error during conn.InsertEntity(node1) <<<<<<<\")\n\t\t\treturn nil\n\t\t}\n\n\t\t_, err = conn.Commit()\n\t\tif err != nil {\n\t\t\tfmt.Println(\">>>>>>> Returning from InsertAncestryNodes w/ error during conn.Commit() <<<<<<<\")\n\t\t\treturn nil\n\t\t}\n\t\tfmt.Printf(\">>>>>>> Inside InsertAncestryNodes: Successfully added node '%+v'<<<<<<<\\n\", houseMember.MemberName)\n\t\thouseMemberTable[houseMember.MemberName] = node1\n\t} // End of for loop\n\tfmt.Println(\">>>>>>> Successfully added nodes w/ NO ERRORS !!! <<<<<<<\")\n\n\tfmt.Println(\">>>>>>> Returning from InsertAncestryNodes w/ NO ERRORS !!! <<<<<<<\")\n\treturn houseMemberTable\n}", "func main() {\n\troot := TreeNode{\n\t\tVal: 1,\n\t\tLeft: &TreeNode{\n\t\t\tVal: 2,\n\t\t\tLeft: nil,\n\t\t\tRight: nil,\n\t\t},\n\t\tRight: &TreeNode{\n\t\t\tVal: 3,\n\t\t\tLeft: &TreeNode{\n\t\t\t\tVal: 4,\n\t\t\t\tRight: nil,\n\t\t\t\tLeft: nil,\n\t\t\t},\n\t\t\tRight: &TreeNode{\n\t\t\t\tVal: 5,\n\t\t\t\tRight: nil,\n\t\t\t\tLeft: nil,\n\t\t\t},\n\t\t},\n\t}\n\tobj := Constructor()\n\tdata := obj.serialize(&root)\n\tfmt.Println(data)\n}", "func (graph *graphRW) Save() {\n\tgraph.parent.rwLock.Lock()\n\tdefer graph.parent.rwLock.Unlock()\n\n\tdestGraph := graph.parent.graph\n\n\t// propagate newly registered mappings\n\tfor mapName, mapping := range graph.mappings {\n\t\tif _, alreadyReg := destGraph.mappings[mapName]; !alreadyReg {\n\t\t\tdestGraph.mappings[mapName] = mapping\n\t\t}\n\t}\n\n\t// apply deleted nodes\n\tfor _, key := range graph.deleted {\n\t\tif node, has := destGraph.nodes[key]; has {\n\t\t\t// remove metadata\n\t\t\tif node.metadataAdded {\n\t\t\t\tif mapping, hasMapping := destGraph.mappings[node.metadataMap]; hasMapping {\n\t\t\t\t\tmapping.Delete(node.label)\n\t\t\t\t}\n\t\t\t}\n\t\t\t// remove node from graph\n\t\t\tdelete(destGraph.nodes, key)\n\t\t}\n\t\tgraph.newRevs[key] = true\n\t}\n\tgraph.deleted = []string{}\n\n\t// apply new/changes nodes\n\tfor key, node := range graph.nodes {\n\t\tif !node.dataUpdated && !node.targetsUpdated && !node.sourcesUpdated {\n\t\t\tcontinue\n\t\t}\n\n\t\t// update metadata\n\t\tif !node.metaInSync {\n\t\t\t// update metadata map\n\t\t\tif mapping, hasMapping := destGraph.mappings[node.metadataMap]; hasMapping {\n\t\t\t\tif node.metadataAdded {\n\t\t\t\t\tif node.metadata == nil {\n\t\t\t\t\t\tmapping.Delete(node.label)\n\t\t\t\t\t\tnode.metadataAdded = false\n\t\t\t\t\t} else {\n\t\t\t\t\t\tprevMeta, _ := mapping.GetValue(node.label)\n\t\t\t\t\t\tif !reflect.DeepEqual(prevMeta, node.metadata) {\n\t\t\t\t\t\t\tmapping.Update(node.label, node.metadata)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t} else if node.metadata != nil {\n\t\t\t\t\tmapping.Put(node.label, node.metadata)\n\t\t\t\t\tnode.metadataAdded = true\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\t// mark node for recording during RW-handle release\n\t\t// (ignore if only sources have been updated)\n\t\tif node.dataUpdated || node.targetsUpdated {\n\t\t\tif _, newRev := graph.newRevs[key]; !newRev {\n\t\t\t\tgraph.newRevs[key] = false\n\t\t\t}\n\t\t\tgraph.newRevs[key] = graph.newRevs[key] || node.dataUpdated\n\t\t}\n\n\t\t// copy changed node to the actual graph\n\t\tnodeCopy := node.copy()\n\t\tnodeCopy.graph = destGraph\n\t\tdestGraph.nodes[key] = newNode(nodeCopy)\n\n\t\t// use copy-on-write targets+sources for the write-handle\n\t\tcowTargets := nodeCopy.targets\n\t\tnodeCopy.targets = node.targets\n\t\tnode.targets = cowTargets\n\t\tcowSources := nodeCopy.sources\n\t\tnodeCopy.sources = node.sources\n\t\tnode.sources = cowSources\n\n\t\t// working copy is now in-sync\n\t\tnode.dataUpdated = false\n\t\tnode.targetsUpdated = false\n\t\tnode.sourcesUpdated = false\n\t\tnode.metaInSync = true\n\t}\n}", "func BuildAncestryGraph() {\n\tfmt.Println(\">>>>>>> Entering BuildAncestryGraph <<<<<<<\")\n\tconnFactory := connection.NewTGConnectionFactory()\n\tconn, err := connFactory.CreateConnection(ancestryUrl, ancestryUser, ancestryPassword, nil)\n\tif err != nil {\n\t\tfmt.Println(\">>>>>>> Returning from BuildAncestryGraph - error during CreateConnection <<<<<<<\")\n\t\treturn\n\t}\n\n\terr = conn.Connect()\n\tif err != nil {\n\t\tfmt.Println(\">>>>>>> Returning from BuildAncestryGraph - error during conn.Connect <<<<<<<\")\n\t\treturn\n\t}\n\n\tgof, err := conn.GetGraphObjectFactory()\n\tif err != nil {\n\t\tfmt.Println(\">>>>>>> Returning from BuildAncestryGraph - error during conn.GetGraphObjectFactory <<<<<<<\")\n\t\treturn\n\t}\n\tif gof == nil {\n\t\tfmt.Println(\">>>>>>> Returning from BuildAncestryGraph - Graph Object Factory is null <<<<<<<\")\n\t\treturn\n\t}\n\n\tfmt.Println(\">>>>>>> Inside BuildAncestryGraph: About to InsertAncestryNodes <<<<<<<\")\n\thouseMemberTable := insertAncestryNodes(conn, gof)\n\tfmt.Println(\">>>>>>> Inside BuildAncestryGraph: About to InsertRelationEdges <<<<<<<\")\n\tinsertRelationEdges(conn, gof, houseMemberTable)\n\tfmt.Println(\">>>>>>> Inside BuildAncestryGraph: Napoleon Bonaparte Ancestry Graph Created Successfully <<<<<<<\")\n\n\terr = conn.Disconnect()\n\tif err != nil {\n\t\tfmt.Println(\">>>>>>> Returning from BuildAncestryGraph - error during conn.Disconnect <<<<<<<\")\n\t\treturn\n\t}\n\tfmt.Println(\">>>>>>> Returning from BuildAncestryGraph - successfully disconnected. <<<<<<<\")\n}", "func (d *Graph) buildTree() {\n\tconfig := build.Default\n\n\t// For each package, look for the dependencies and build out a tree\n\tfor p := range d.Pkgs {\n\t\tpkg, _ := config.Import(d.Pkgs[p], d.SrcDir, build.AllowBinary)\n\t\timports := pkg.Imports\n\n\t\t// Iterate through the imports and build our tree\n\t\tfor i := range imports {\n\t\t\t// The full path of our current import\n\t\t\tpath := imports[i]\n\n\t\t\t// When dealing with multiple packages, we can't assume that imports\n\t\t\t// are unique. Thus the nodes may already exist and we shouldn't do any work\n\t\t\tif d.Nodes[path] != nil {\n\t\t\t\td.Nodes[path].IsDuplicate = true\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Ignore the GO standard library imports\n\t\t\tif _, ok := stdlib[strings.Split(path, \"/\")[0]]; ok && !d.StdLib {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Keep track when traversing the path\n\t\t\tvar currentNode = &Node{\n\t\t\t\tPath: path,\n\t\t\t\tIsDep: true,\n\t\t\t\tIsCoreDep: strings.HasPrefix(path, strings.Split(d.Pkgs[p], \"/\")[0]),\n\t\t\t}\n\n\t\t\t// Keep track of the number of dependencies\n\t\t\td.TotalDeps++\n\n\t\t\t// Link our dependency node to its ancestors\n\t\t\tfor path != \"\" {\n\t\t\t\t// Constant time lookup to all of our nodes\n\t\t\t\t// based on their full path string\n\t\t\t\td.Nodes[path] = currentNode\n\n\t\t\t\t// Keep popping off the tip of the path\n\t\t\t\tpath, _ = filepath.Split(path)\n\n\t\t\t\tif len(path) > 0 {\n\t\t\t\t\t// Trailing slash in file path causes issues, remove it\n\t\t\t\t\tif strings.HasSuffix(path, \"/\") {\n\t\t\t\t\t\tpath = path[:len(path)-1]\n\t\t\t\t\t}\n\n\t\t\t\t\t// Create nodes for all directory paths if they don't exist\n\t\t\t\t\tif d.Nodes[path] == nil {\n\t\t\t\t\t\tcurrentNode.addParent(&Node{\n\t\t\t\t\t\t\tPath: path,\n\t\t\t\t\t\t})\n\n\t\t\t\t\t\t// Change the current node to the newly created item\n\t\t\t\t\t\tcurrentNode = currentNode.Parent\n\t\t\t\t\t} else {\n\t\t\t\t\t\t// Otherwise, assume the common ancestor already has it's tree built\n\t\t\t\t\t\tcurrentNode.addParent(d.Nodes[path])\n\t\t\t\t\t\tcurrentNode = nil\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// currentNode will be nil if there was already a common ancestor --\n\t\t\t// which means the root node already exists for that import path\n\t\t\tif currentNode != nil {\n\t\t\t\td.RootNode.addChild(currentNode)\n\t\t\t}\n\t\t}\n\t}\n}", "func writegraphs() {\n\n\tfor i := 2; i <= 10; i++ {\n\t\tg, err := buildGraph(false)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\terr = makegrid(i, &g)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\t//g, err = cfi(&g)\n\t\t//if err != nil {\n\t\t//\tlog.Fatal(err)\n\t\t//}\n\n\t\texpandgraph(&g)\n\n\t\tfilename := \"expgrid\" + strconv.Itoa(i) + \".gr\"\n\t\tfmt.Println(filename, \" created\")\n\t\tfile, err := os.Create(filename)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\terr = g.writeGraph(file)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t}\n}", "func WriteGraph(g *Graph, w io.Writer) error {\n\tenc := gob.NewEncoder(w)\n\treturn enc.Encode(g)\n}", "func BuildGraph(rawGraph [][]string, timeStamps []time.Time) *CommitGraph {\n\tnodeMap := make(map[string]*Node, len(rawGraph))\n\tfor idx, rawNode := range rawGraph {\n\t\thash := rawNode[0]\n\t\tnodeMap[hash] = &Node{\n\t\t\tHash: hash,\n\t\t\tParents: make([]*Node, len(rawNode)-1),\n\t\t\tTimestamp: timeStamps[idx],\n\t\t}\n\t}\n\n\tfor _, rawNode := range rawGraph {\n\t\tfor idx, p := range rawNode[1:] {\n\t\t\tnodeMap[rawNode[0]].Parents[idx] = nodeMap[p]\n\t\t}\n\t}\n\n\treturn &CommitGraph{\n\t\tNodes: nodeMap,\n\t}\n}", "func (n *Node) WriteTree(padding ...int) string {\n\tvar indent int\n\tif len(padding) == 1 {\n\t\tindent = padding[0]\n\t}\n\n\tvar s string\n\tif n.val != \"\" {\n\t\tfor i := 0; i < indent; i++ {\n\t\t\ts += \" \"\n\t\t}\n\t\ts += n.key + \" : \" + n.val + \"\\n\"\n\t} else {\n\t\tfor i := 0; i < indent; i++ {\n\t\t\ts += \" \"\n\t\t}\n\t\ts += n.key + \" :\" + \"\\n\"\n\t\tfor _, nn := range n.nodes {\n\t\t\ts += nn.WriteTree(indent + 1)\n\t\t}\n\t}\n\treturn s\n}", "func createNewNodeNetworkObject(writer *bufio.Writer, sourceOsmNode *osm.Node) {\n\ttags := sourceOsmNode.TagMap()\n\n\t// Punktnetzwerk 'Fahrrad'\n\tnewOsmNode := *sourceOsmNode // copy content (don't modify origin/source node)\n\tnewOsmNode.ID = 0\n\tnewOsmNode.Tags = []osm.Tag{} // remove all source tags\n\trefValue, found := tags[\"icn_ref\"]\n\tif found {\n\t\ttag := osm.Tag{Key: \"node_network\", Value: \"node_bicycle\"}\n\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\ttag = osm.Tag{Key: \"name\", Value: refValue}\n\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\twriteNewNodeObject(writer, &newOsmNode)\n\t} else {\n\t\trefValue, found = tags[\"ncn_ref\"]\n\t\tif found {\n\t\t\ttag := osm.Tag{Key: \"node_network\", Value: \"node_bicycle\"}\n\t\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\t\ttag = osm.Tag{Key: \"name\", Value: refValue}\n\t\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\t\twriteNewNodeObject(writer, &newOsmNode)\n\t\t} else {\n\t\t\trefValue, found = tags[\"rcn_ref\"]\n\t\t\tif found {\n\t\t\t\ttag := osm.Tag{Key: \"node_network\", Value: \"node_bicycle\"}\n\t\t\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\t\t\ttag = osm.Tag{Key: \"name\", Value: refValue}\n\t\t\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\t\t\twriteNewNodeObject(writer, &newOsmNode)\n\t\t\t} else {\n\t\t\t\trefValue, found = tags[\"lcn_ref\"]\n\t\t\t\tif found {\n\t\t\t\t\ttag := osm.Tag{Key: \"node_network\", Value: \"node_bicycle\"}\n\t\t\t\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\t\t\t\ttag = osm.Tag{Key: \"name\", Value: refValue}\n\t\t\t\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\t\t\t\twriteNewNodeObject(writer, &newOsmNode)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\t// Punktnetzwerk 'Wandern'\n\tnewOsmNode = *sourceOsmNode // copy content (don't modify origin/source node)\n\tnewOsmNode.ID = 0\n\tnewOsmNode.Tags = []osm.Tag{} // remove all source tags\n\trefValue, found = tags[\"iwn_ref\"]\n\tif found {\n\t\ttag := osm.Tag{Key: \"node_network\", Value: \"node_hiking\"}\n\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\ttag = osm.Tag{Key: \"name\", Value: refValue}\n\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\twriteNewNodeObject(writer, &newOsmNode)\n\t} else {\n\t\trefValue, found = tags[\"nwn_ref\"]\n\t\tif found {\n\t\t\ttag := osm.Tag{Key: \"node_network\", Value: \"node_hiking\"}\n\t\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\t\ttag = osm.Tag{Key: \"name\", Value: refValue}\n\t\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\t\twriteNewNodeObject(writer, &newOsmNode)\n\t\t} else {\n\t\t\trefValue, found = tags[\"rwn_ref\"]\n\t\t\tif found {\n\t\t\t\ttag := osm.Tag{Key: \"node_network\", Value: \"node_hiking\"}\n\t\t\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\t\t\ttag = osm.Tag{Key: \"name\", Value: refValue}\n\t\t\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\t\t\twriteNewNodeObject(writer, &newOsmNode)\n\t\t\t} else {\n\t\t\t\trefValue, found = tags[\"lwn_ref\"]\n\t\t\t\tif found {\n\t\t\t\t\ttag := osm.Tag{Key: \"node_network\", Value: \"node_hiking\"}\n\t\t\t\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\t\t\t\ttag = osm.Tag{Key: \"name\", Value: refValue}\n\t\t\t\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\t\t\t\twriteNewNodeObject(writer, &newOsmNode)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\t// Punktnetzwerk 'Inline-Skaten'\n\tnewOsmNode = *sourceOsmNode // copy content (don't modify origin/source node)\n\tnewOsmNode.ID = 0\n\tnewOsmNode.Tags = []osm.Tag{} // remove all source tags\n\trefValue, found = tags[\"rin_ref\"]\n\tif found {\n\t\ttag := osm.Tag{Key: \"node_network\", Value: \"node_inline_skates\"}\n\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\ttag = osm.Tag{Key: \"name\", Value: refValue}\n\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\twriteNewNodeObject(writer, &newOsmNode)\n\t}\n\n\t// Punktnetzwerk 'Reiten'\n\tnewOsmNode = *sourceOsmNode // copy content (don't modify origin/source node)\n\tnewOsmNode.ID = 0\n\tnewOsmNode.Tags = []osm.Tag{} // remove all source tags\n\trefValue, found = tags[\"rhn_ref\"]\n\tif found {\n\t\ttag := osm.Tag{Key: \"node_network\", Value: \"node_horse\"}\n\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\ttag = osm.Tag{Key: \"name\", Value: refValue}\n\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\twriteNewNodeObject(writer, &newOsmNode)\n\t}\n\n\t// Punktnetzwerk 'Kanu'\n\tnewOsmNode = *sourceOsmNode // copy content (don't modify origin/source node)\n\tnewOsmNode.ID = 0\n\tnewOsmNode.Tags = []osm.Tag{} // remove all source tags\n\trefValue, found = tags[\"rpn_ref\"]\n\tif found {\n\t\ttag := osm.Tag{Key: \"node_network\", Value: \"node_canoe\"}\n\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\ttag = osm.Tag{Key: \"name\", Value: refValue}\n\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\twriteNewNodeObject(writer, &newOsmNode)\n\t}\n\n\t// Punktnetzwerk 'Motorboot'\n\tnewOsmNode = *sourceOsmNode // copy content (don't modify origin/source node)\n\tnewOsmNode.ID = 0\n\tnewOsmNode.Tags = []osm.Tag{} // remove all source tags\n\trefValue, found = tags[\"rmn_ref\"]\n\tif found {\n\t\ttag := osm.Tag{Key: \"node_network\", Value: \"node_motorboat\"}\n\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\ttag = osm.Tag{Key: \"name\", Value: refValue}\n\t\tnewOsmNode.Tags = append(newOsmNode.Tags, tag)\n\t\twriteNewNodeObject(writer, &newOsmNode)\n\t}\n}", "func (tg *TrieGrapher) Graph(path string) error {\n\tf, err := os.Create(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\n\ttg.w = f\n\n\tfmt.Fprintln(f, \"digraph T {\")\n\tfmt.Fprintln(f, \"\\tnodesep=0.2; ranksep=0.4; splines=false; outputorder=edgesfirst;\")\n\tfmt.Fprintln(f, \"\\tnode [shape=circle, style=filled, fillcolor=white, fixedsize=true];\")\n\tfmt.Fprintln(f, \"\\tedge [arrowsize=0.5];\")\n\n\t// Will recursivelly call graphState on every state (which is in use).\n\ttg.graphState(RootState, EmptyCell)\n\n\tfmt.Fprintln(f, \"}\")\n\n\treturn nil\n}", "func (bpt *BplusTree) writeTree(printLayout bool) {\n\tdefer glog.Flush()\n\tnode, _ := bpt.fetch(bpt.rootKey)\n\tif node == nil {\n\t\tglog.Errorf(\"failed to fetch root key: %v\", bpt.rootKey)\n\t\treturn\n\t}\n\t// Print tree layout.\n\tif printLayout == true {\n\t\tbpt.writeLayout()\n\t}\n\n\t// Go to the left most leaf node and start printing in order.\n\tfor node != nil {\n\t\tif node.IsLeaf {\n\t\t\tbreak\n\t\t}\n\t\tnode, _ = bpt.fetch(node.Children[0].NodeKey)\n\t\tif node == nil {\n\t\t\tglog.Errorf(\"failed to fetch key: %v\", node.Children[0].NodeKey)\n\t\t\treturn\n\t\t}\n\t}\n\n\tif node == nil {\n\t\tglog.Infof(\"tree is empty\")\n\t\treturn\n\t}\n\n\tindex := 0\n\tfor {\n\t\tglog.Infof(\"leaf node: %d (DK: %v, NK: %v, XK: %v, PK: %v)\\n\",\n\t\t\tindex, node.DataKey, node.NodeKey, node.NextKey, node.PrevKey)\n\t\tfor _, child := range node.Children {\n\t\t\tglog.Infof(\"\\t%v\\n\", child)\n\t\t}\n\n\t\tif node.NextKey.IsNil() {\n\t\t\tbreak\n\t\t}\n\n\t\tif !node.NextKey.IsNil() {\n\t\t\tnextKey := node.NextKey\n\t\t\tnode, _ = bpt.fetch(nextKey)\n\t\t\tif node == nil {\n\t\t\t\tglog.Errorf(\"failed to fetch key: %v\", nextKey)\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tindex++\n\t}\n}", "func WriteTree(c *git.Client) string {\n\tidx, err := c.GitDir.ReadIndex()\n\tif err != nil {\n\t\treturn err.Error()\n\t}\n\tsha1, err := idx.WriteTree(c)\n\tif err != nil {\n\t\treturn err.Error()\n\t}\n\treturn sha1.String()\n}", "func (graphS Graph) traverse(f *os.File, vertex *dag.Vertex, done *[]*dag.Vertex, fname string) error {\n graph := graphS.g\n\n var err error\n // Check if we are in done[]; if we are, we don't need to do anything\n if sliceContains(*done, vertex) {\n return nil\n }\n\n // We set this here to avoid loops\n *done = append(*done, vertex)\n\n // Loop over children\n children, err := graph.Successors(vertex)\n if err != nil {\n return fmt.Errorf(\"Unable to get children of %s with %w\", vertex.ID, err)\n }\n\n for _, child := range children {\n // Add the line to the DOT\n _, err = f.WriteString(fmt.Sprintf(\"\\\"%s\\\" -> \\\"%s\\\"\\n\", vertex.ID, child.ID))\n if err != nil {\n return fmt.Errorf(\"Unable to write to %s with %w\", fname, err)\n }\n // Recurse to children\n err = graphS.traverse(f, child, done, fname)\n if err != nil {\n return err\n }\n }\n\n return nil\n}", "func treeGene( geneArray[] string, arrayPos int, shapeType string) int{\n\n//Function check to see if the there is any shapes in the in the listArray\n//if not it changes the structures as it moves to the end\n\tif listArray[0] != shapeType { \n\t\t\tlistArray[1]= \"<inst>\"\n\t\t\tlistArray[0] = shapeType\n\t\tarrayPos++\n\t\ttreeGene(geneArray[0:],arrayPos,shapeType)\t// Called the function so it can be processed with the valid format \t\n\t}else{ if listArray[1] == \"\" || listArray[1] == \"<inst>\"{ // after transforming it is place in a format that can be parsed \n\t\t\tif geneArray[arrayPos] == shapeType{ \t// Ensure we are not Validating a Shape\n\t\t\t\tarrayPos++\n\t\t\t}\n\t\t\t// Retrieve the Coordinated from the array\n\t\t\tvar curCoord string=geneArray[arrayPos] \n\t\t\tarrayPos++\n\t\t\t\n\t\t\t// Using Slices we get each Values \n\t\t\tx:=curCoord[0:1] \n\t\t\ty:=curCoord[1:2]\n\t\t\tfmt.Printf(\"\\n |\\n\")\n\t\t\t\n\t\t\t// The Only Difference with Circle is that it Has a Coordinated\n\t\t\tfmt.Printf(shapeType+\"\\n / \\\\\\n\")\n\t\t\tif shapeType == \"CIR\"{\n\t\t\t\tfmt.Printf(\"<coord>\\n\")\n\t\t\t}\t\t\t\n\t\t\tfmt.Printf(\"<x><y>\\n\"+x+\" \"+y)\n\t\t\t\n\t\t}\n\t\t }\n\n\treturn arrayPos\n}", "func Graph(w http.ResponseWriter, g *model.Graph, params *Params) {\n\tgj, err := json.Marshal(g)\n\tif err != nil {\n\t\tlog.Printf(\"Could not execute graph editor template: %v\", err)\n\t\thttp.Error(w, \"Could not execute graph editor template\", http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\td := &editorInput{\n\t\tParams: params,\n\t\tGraph: g,\n\t\tGraphJSON: string(gj),\n\t\tPartTypes: model.PartTypes,\n\t\tPartTypesByCategory: model.PartTypesByCategory,\n\t}\n\td.Licenses = []license{\n\t\t{\"Shenzhen Go\", \"/.static/misc/LICENSE\"},\n\t\t{\"Ace (code editor)\", \"/.static/js/ace/LICENSE\"},\n\t\t{\"Chromium Hterm\", \"/.static/js/hterm/LICENSE\"},\n\t}\n\tif err := graphEditorTemplate.Execute(w, d); err != nil {\n\t\tlog.Printf(\"Could not execute graph editor template: %v\", err)\n\t\thttp.Error(w, \"Could not execute graph editor template\", http.StatusInternalServerError)\n\t}\n}", "func (l *Lista) GraphNodes(i int) string{\n\tAux := l.Inicio\n\tnodos := \"\"\n\tj := 0\n\tfor Aux != nil{\n\t\tnodos = nodos + \"a\" + strconv.Itoa(i) + \"Node\" + strconv.Itoa(j) + \" [label=\\\"\"+ Aux.Dato.Nombre +\"\\\"]\\n\"\n\t\tj++\n\t\tAux = Aux.Siguiente\n\t}\n\tk := 0\n\tnodos = nodos + \"struct:f\" + strconv.Itoa(i)\n\tAux = l.Inicio\n\tfor Aux != nil{\n\t\tnodos = nodos + \" -> a\" + strconv.Itoa(i) + \"Node\" + strconv.Itoa(k)\n\t\tk++\n\t\tAux = Aux.Siguiente\n\t}\n\tnodos = nodos + \";\\n\"\n\treturn nodos\n}", "func writeDotGraph(outf *os.File, start *node, id string) {\n\tdone := make(map[*node]bool)\n\tvar show func(*node)\n\tshow = func(u *node) {\n\t\tif u.accept {\n\t\t\tfmt.Fprintf(outf, \" %v[style=filled,color=green];\\n\", u.n)\n\t\t}\n\t\tdone[u] = true\n\t\tfor _, e := range u.e {\n\t\t\t// We use -1 to denote the dead end node in DFAs.\n\t\t\tif e.dst.n == -1 {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tlabel := \"\"\n\t\t\truneToDot := func(r rune) string {\n\t\t\t\tif strconv.IsPrint(r) {\n\t\t\t\t\treturn fmt.Sprintf(\"%v\", string(r))\n\t\t\t\t}\n\t\t\t\treturn fmt.Sprintf(\"U+%X\", int(r))\n\t\t\t}\n\t\t\tswitch e.kind {\n\t\t\tcase kRune:\n\t\t\t\tlabel = fmt.Sprintf(\"[label=%q]\", runeToDot(e.r))\n\t\t\tcase kWild:\n\t\t\t\tlabel = \"[color=blue]\"\n\t\t\tcase kClass:\n\t\t\t\tlabel = \"[label=\\\"[\"\n\t\t\t\tif e.negate {\n\t\t\t\t\tlabel += \"^\"\n\t\t\t\t}\n\t\t\t\tfor i := 0; i < len(e.lim); i += 2 {\n\t\t\t\t\tlabel += runeToDot(e.lim[i])\n\t\t\t\t\tif e.lim[i] != e.lim[i+1] {\n\t\t\t\t\t\tlabel += \"-\" + runeToDot(e.lim[i+1])\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tlabel += \"]\\\"]\"\n\t\t\t}\n\t\t\tfmt.Fprintf(outf, \" %v -> %v%v;\\n\", u.n, e.dst.n, label)\n\t\t}\n\t\tfor _, e := range u.e {\n\t\t\tif !done[e.dst] {\n\t\t\t\tshow(e.dst)\n\t\t\t}\n\t\t}\n\t}\n\tfmt.Fprintf(outf, \"digraph %v {\\n 0[shape=box];\\n\", id)\n\tshow(start)\n\tfmt.Fprintln(outf, \"}\")\n}", "func writeNewNodeObject(writer *bufio.Writer, newOsmNode *osm.Node) {\n\tnewOsmNode.ID = newNodeID\n\tnewNodeID++\n\n\tdata, err := xml.MarshalIndent(newOsmNode, \" \", \" \")\n\tif err != nil {\n\t\tlog.Fatalf(\"error <%v> at xml.MarshalIndent()\", err)\n\t}\n\t_, err = fmt.Fprintf(writer, \"%s\\n\", string(data))\n\tif err != nil {\n\t\tlog.Fatalf(\"error writing output file: %v\", err)\n\t}\n}", "func (node *Node) PrintStructure(indent int, character string) {\n\tfor i := 0; i < indent; i++ {\n\t\tfmt.Print(character)\n\t}\n\tfmt.Println(node.Data)\n\tfor _, child := range node.Children {\n\t\tchild.PrintStructure(indent+1, character)\n\t}\n\tif len(node.Children) == 0 {\n\t\treturn\n\t}\n\tfor i := 0; i < indent; i++ {\n\t\tfmt.Print(character)\n\t}\n\tfmt.Println(node.Data)\n}", "func mermaidGraph(c *flow.Controller) string {\n\tw := &strings.Builder{}\n\tfmt.Fprintln(w, \"graph TD\")\n\tfor i, t := range c.Tasks() {\n\t\tfmt.Fprintf(w, \" t%d(\\\"%s [%s]\\\")\\n\", i, t.Path(), t.State())\n\t\tfor _, t := range t.Dependencies() {\n\t\t\tfmt.Fprintf(w, \" t%d-->t%d\\n\", i, t.Index())\n\t\t}\n\t}\n\treturn w.String()\n}", "func (g Index) WriteTree(c *Client) (TreeID, error) {\n\tsha1, err := writeIndexEntries(c, \"\", g.Objects)\n\tif err != nil && err != ObjectExists {\n\t\treturn TreeID{}, err\n\t}\n\treturn sha1, nil\n}", "func toJsGraph(g *gen.Graph) jsGraph {\n\tgraph := jsGraph{}\n\tfor _, n := range g.Nodes {\n\t\tnode := jsNode{ID: n.Name}\n\t\tfor _, f := range n.Fields {\n\t\t\tnode.Fields = append(node.Fields, jsField{\n\t\t\t\tName: f.Name,\n\t\t\t\tType: f.Type.String(),\n\t\t\t})\n\t\t}\n\t\tgraph.Nodes = append(graph.Nodes, node)\n\t\tfor _, e := range n.Edges {\n\t\t\tif e.IsInverse() {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tgraph.Edges = append(graph.Edges, jsEdge{\n\t\t\t\tFrom: n.Name,\n\t\t\t\tTo: e.Type.Name,\n\t\t\t\tLabel: e.Name,\n\t\t\t})\n\t\t}\n\n\t}\n\treturn graph\n}", "func (n *Node) WriteMetadata(owner *userpb.UserId) (err error) {\n\tnodePath := n.InternalPath()\n\tif err = xattr.Set(nodePath, xattrs.ParentidAttr, []byte(n.ParentID)); err != nil {\n\t\treturn errors.Wrap(err, \"decomposedfs: could not set parentid attribute\")\n\t}\n\tif err = xattr.Set(nodePath, xattrs.NameAttr, []byte(n.Name)); err != nil {\n\t\treturn errors.Wrap(err, \"decomposedfs: could not set name attribute\")\n\t}\n\tif err = xattr.Set(nodePath, xattrs.BlobIDAttr, []byte(n.BlobID)); err != nil {\n\t\treturn errors.Wrap(err, \"decomposedfs: could not set blobid attribute\")\n\t}\n\tif err = xattr.Set(nodePath, xattrs.BlobsizeAttr, []byte(fmt.Sprintf(\"%d\", n.Blobsize))); err != nil {\n\t\treturn errors.Wrap(err, \"decomposedfs: could not set blobsize attribute\")\n\t}\n\tif owner == nil {\n\t\tif err = xattr.Set(nodePath, xattrs.OwnerIDAttr, []byte(\"\")); err != nil {\n\t\t\treturn errors.Wrap(err, \"decomposedfs: could not set empty owner id attribute\")\n\t\t}\n\t\tif err = xattr.Set(nodePath, xattrs.OwnerIDPAttr, []byte(\"\")); err != nil {\n\t\t\treturn errors.Wrap(err, \"decomposedfs: could not set empty owner idp attribute\")\n\t\t}\n\t\tif err = xattr.Set(nodePath, xattrs.OwnerTypeAttr, []byte(\"\")); err != nil {\n\t\t\treturn errors.Wrap(err, \"decomposedfs: could not set empty owner type attribute\")\n\t\t}\n\t} else {\n\t\tif err = xattr.Set(nodePath, xattrs.OwnerIDAttr, []byte(owner.OpaqueId)); err != nil {\n\t\t\treturn errors.Wrap(err, \"decomposedfs: could not set owner id attribute\")\n\t\t}\n\t\tif err = xattr.Set(nodePath, xattrs.OwnerIDPAttr, []byte(owner.Idp)); err != nil {\n\t\t\treturn errors.Wrap(err, \"decomposedfs: could not set owner idp attribute\")\n\t\t}\n\t\tif err = xattr.Set(nodePath, xattrs.OwnerTypeAttr, []byte(utils.UserTypeToString(owner.Type))); err != nil {\n\t\t\treturn errors.Wrap(err, \"decomposedfs: could not set owner idp attribute\")\n\t\t}\n\t}\n\treturn\n}", "func (this *Codec) serialize(root *TreeNode) string {\n if root == nil {\n return \"x\"\n }\n return strconv.Itoa(root.Val) + \",\" + this.serialize(root.Left)+ \",\" + this.serialize(root.Right)\n}", "func (project *Project) createGraph() error {\n\tg.AddNode(project.Name)\n\tfor _, ig := range project.Csproj.ItemGroups {\n\t\tfor _, pr := range ig.ProjectReferences {\n\t\t\tcsprojFilePath := strings.Replace(path.Join(path.Dir(project.Path), pr.Include), \"\\\\\", \"/\", -1)\n\t\t\tcsproj, err := parseCsproj(csprojFilePath)\n\t\t\tif err != nil {\n\t\t\t\terr = fmt.Errorf(\"cannot parse csproj file referenced in %s: %v\", project.Path, err)\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tfileName := path.Base(strings.Replace(pr.Include, \"\\\\\", \"/\", -1))\n\t\t\tfileNameNoExt := strings.Replace(fileName, path.Ext(fileName), \"\", -1)\n\t\t\tdep := &Project{\n\t\t\t\tName: fileNameNoExt,\n\t\t\t\tPath: csprojFilePath,\n\t\t\t\tCsproj: csproj,\n\t\t\t}\n\n\t\t\tif project.Name == \"\" || dep.Name == \"\" {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tg.AddEdge(project.Name, dep.Name)\n\n\t\t\tdep.createGraph()\n\t\t}\n\t}\n\treturn nil\n}", "func buildTree(path string, vfileStack []*vfs.VFile) error {\n\n\terrFmt := \"build tree: %v\\n\"\n\n\tstats, err := ioutil.ReadDir(path)\n\tif err != nil {\n\t\treturn fmt.Errorf(errFmt, err)\n\t}\n\n\tfor _, node := range stats {\n\n\t\tnodePath := fmt.Sprintf(\"%s/%s\", path, node.Name())\n\n\t\tif node.IsDir() {\n\t\t\tvDir := vfs.NewFile(node)\n\t\t\t// append virtual directory to last element in vfs\n\t\t\tvfileStack[len(vfileStack)-1].Append(vDir)\n\t\t\t// virtual directory becomes new last\n\t\t\t// element of vfs within recursive call\n\t\t\terr := buildTree(nodePath, append(vfileStack, vDir))\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(errFmt, err)\n\t\t\t}\n\n\t\t\tcontinue\n\t\t}\n\n\t\tfileBytes, err := ioutil.ReadFile(nodePath)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(errFmt, err)\n\t\t}\n\n\t\tbuf := bytes.NewBuffer(fileBytes)\n\n\t\t// strip json of white space to save space in generated binary\n\t\tif strings.HasSuffix(node.Name(), \".json\") {\n\t\t\tbuf, err = reEncodeJson(buf)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(errFmt, err)\n\t\t\t}\n\t\t}\n\n\t\tvfile := vfs.NewFile(node, buf.Bytes()...)\n\t\tvfileStack[len(vfileStack)-1].Append(vfile)\n\n\t\tfmt.Printf(\"ESSENCE: embedded file: %s/%s\\n\", path, node.Name())\n\t}\n\n\treturn nil\n}", "func saveNodes() {\n\tf, err := os.Create(\"nodes.dat\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer f.Close()\n\tb, err := json.Marshal(nodes)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tif _, err := f.Write(b); err != nil {\n\t\tpanic(err)\n\t}\n}", "func (p *printer) printStructure(gn *GenericNode) {\n\t// traverse siblings\n\tfor s := gn; s != nil; s = s.next {\n\t\tswitch s.NodeType {\n\t\tcase Declaration:\n\t\t\t// print attributes\n\t\tcase Element:\n\t\t\tif p.pretty {\n\t\t\t\tfmt.Println(\"\")\n\t\t\t}\n\t\t\t// can have children and siblings which must be handled\n\t\t\tfmt.Print(\"<\" + string(s.Name) + \">\")\n\t\t\tp.traverseDepth(s)\n\t\t\tfmt.Print(\"</\" + string(s.Name) + \">\")\n\t\tcase Data:\n\t\t\t// just print and return\n\t\t\tfmt.Print(string(s.Value))\n\t\tcase Cdata:\n\t\t\t// cdata needs to be embedded in a CDATA structure\n\t\t\tfmt.Print(`<![CDATA[` + string(s.Value) + `]]`)\n\t\tcase Comment:\n\t\t\tfmt.Print(\"<!--\" + string(s.Value) + \"-->\")\n\t\tcase Doctype:\n\t\t\tfmt.Print(\"<!DOCTYPE \" + string(s.Value) + \">\")\n\t\t\tp.traverseDepth(s)\n\t\tcase Pi:\n\t\t\tfmt.Print(\"<?\" + string(s.Name) + \" \" + string(s.Value))\n\t\tcase Document:\n\t\t\tp.traverseDepth(s)\n\t\tdefault:\n\t\t\tpanic(\"unknown node type\")\n\t\t}\n\n\t}\n\n}", "func main() {\n\torbits := tree{}\n\n\tif len(os.Args) < 2 {\n\t\tlog.Fatalf(\"Missing parameter, provide file name!\")\n\t\treturn\n\t}\n\n\t// raw reading of the file\n\tdata, err := ioutil.ReadFile(os.Args[1])\n\tif err != nil {\n\t\tlog.Fatalf(\"Can't read file: %v\\n\", os.Args[1])\n\t\tpanic(err)\n\t}\n\n\t// take the read file and convert it from strings to ints\n\tfor _, line := range bytes.Split(data, []byte(\"\\n\")) {\n\t\tplanets := strings.Split(strings.TrimSpace(string(line)), \")\")\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"Could not extract planets from line. %v\\n\", err)\n\t\t}\n\n\t\torbits[planets[0]] = append(orbits[planets[0]], planets[1])\n\t}\n\n\tspew.Dump(\"Result:\")\n\t//\tspew.Dump(orbits)\n\n\t//\tfmt.Printf(\"nodes:%d\\n\", orbits.height(\"COM\"))\n\t//\tfmt.Printf(\"count:%d\\n\", orbits.count(\"COM\", 0))\n\tfmt.Printf(\"p:%v\\n\", orbits.parent(\"B\"))\n\tfmt.Printf(\"p:%v\\n\", orbits.parent(\"COM\"))\n\tfmt.Printf(\"q:%v\\n\", orbits.ancestry(\"YOU\"))\n\tfmt.Printf(\"q:%v\\n\", orbits.ancestry(\"SAN\"))\n\tfmt.Printf(\"q:%v\\n\", orbits.transfers(\"YOU\", \"SAN\"))\n\n}", "func (this *Graph) Cluster() []*Graph {\n /*\n\n Algorithm synopsis:\n\n Loop over the Starters, for each unvisited Starter,\n define an empty sub-graph and, put it into the toVisit set\n\n Loop over the toVisit node set, for each node in it, \n skip if already visited\n add the node to the sub-graph\n remove the nodes into the hasVisited node set\n put all its incoming and outgoing edge into the the toWalk set while\n stop at the hub nodes (edges from the hub nodes are not put in the toWalk set)\n then iterate through the toWalk edge set \n skip if already walked\n add the edge to the sub-graph\n put its connected nodes into the toVisit node set\n remove the edge from the toWalk edge set into the hasWalked edge set\n\n */\n \n // sub-graph index\n sgNdx := -1\n sgRet := make([]*Graph,0)\n\n toVisit := make(nodeSet); hasVisited := make(nodeSet)\n toWalk := make(edgeSet); hasWalked := make(edgeSet)\n\n for starter := range *this.Starters() {\n // define an empty sub-graph and, put it into the toVisit set\n sgRet = append(sgRet, NewGraph(gographviz.NewGraph())); sgNdx++; \n sgRet[sgNdx].Attrs = this.Attrs\n sgRet[sgNdx].SetDir(this.Directed)\n graphName := fmt.Sprintf(\"%s_%03d\\n\", this.Name, sgNdx);\n sgRet[sgNdx].SetName(graphName)\n toVisit.Add(starter)\n hubVisited := make(nodeSet)\n for len(toVisit) > 0 { for nodep := range toVisit {\n toVisit.Del(nodep); //print(\"O \")\n if this.IsHub(nodep) && hasVisited.Has(nodep) && !hubVisited.Has(nodep) { \n // add the already-visited but not-in-this-graph hub node to the sub-graph\n sgRet[sgNdx].AddNode(nodep)\n hubVisited.Add(nodep)\n continue \n }\n if hasVisited.Has(nodep) { continue }\n //spew.Dump(\"toVisit\", nodep)\n // add the node to the sub-graph\n sgRet[sgNdx].AddNode(nodep)\n // remove the nodes into the hasVisited node set\n hasVisited.Add(nodep)\n // stop at the hub nodes\n if this.IsHub(nodep) { continue }\n // put all its incoming and outgoing edge into the the toWalk set\n noden := nodep.Name\n for _, ep := range this.EdgesToParents(noden) {\n toWalk.Add(ep)\n }\n for _, ep := range this.EdgesToChildren(noden) {\n toWalk.Add(ep)\n }\n for edgep := range toWalk {\n toWalk.Del(edgep); //print(\"- \")\n if hasWalked.Has(edgep) { continue }\n //spew.Dump(\"toWalk\", edgep)\n sgRet[sgNdx].Edges.Add(edgep)\n // put its connected nodes into the toVisit node set\n toVisit.Add(this.Lookup(edgep.Src))\n toVisit.Add(this.Lookup(edgep.Dst))\n // remove the edge into the hasWalked edge set\n hasWalked.Add(edgep)\n }\n }}\n //spew.Dump(sgNdx)\n }\n return sgRet\n}", "func DrawGraphTools(filename string, s spn.SPN) {\n\tfile, err := os.Create(filename)\n\n\tif err != nil {\n\t\tfmt.Printf(\"Error. Could not create file [%s].\\n\", filename)\n\t\tpanic(err)\n\t}\n\tdefer file.Close()\n\n\toutname := utils.StringConcat(filename[0:len(filename)-len(filepath.Ext(filename))], \".png\")\n\n\tfmt.Fprintf(file, \"from graph_tool.all import *\\n\\n\")\n\tfmt.Fprintf(file, \"g = Graph(directed=True)\\n\")\n\tfmt.Fprintf(file, \"vcolors = g.new_vertex_property(\\\"string\\\")\\n\")\n\tfmt.Fprintf(file, \"vnames = g.new_vertex_property(\\\"string\\\")\\n\")\n\tfmt.Fprintf(file, \"enames = g.new_edge_property(\\\"string\\\")\\n\\n\")\n\tfmt.Fprintf(file, \"def add_node(name, type):\\n\\tv=g.add_vertex()\\n\\tvnames[v]=name\\n\\t\"+\n\t\t\"vcolors[v]=type\\n\\treturn v\\n\\n\")\n\tfmt.Fprintf(file, \"def add_edge(o, t, name):\\n\\te=g.add_edge(o, t)\\n\\tenames[e]=name\\n\\treturn e\\n\\n\")\n\tfmt.Fprintf(file, \"def add_edge_nameless(o, t):\\n\\te=g.add_edge(o, t)\\n\\treturn e\\n\\n\\n\")\n\n\t// If the SPN is itself an univariate distribution, create a graph with a single node.\n\tif s.Type() == \"leaf\" {\n\t\tfmt.Fprintf(file, \"add_node(\\\"X\\\")\\n\\n\")\n\t\tfmt.Fprintf(file, \"g.vertex_properties[\\\"name\\\"]=vnames\\n\")\n\t\tfmt.Fprintf(file, \"g.vertex_properties[\\\"color\\\"]=vcolors\\n\")\n\t\tfmt.Fprintf(file, \"\\ngraph_draw(g, vertex_text=g.vertex_properties[\\\"name\\\"], \"+\n\t\t\t\"edge_text=enames, vertex_fill_color=g.vertex_properties[\\\"color\\\"], output=\\\"%s\\\")\\n\",\n\t\t\toutname)\n\t\treturn\n\t}\n\n\t// Else, BFS the SPN and write nodes to filename.\n\tnvars, nsums, nprods := 0, 0, 0\n\tqueue := common.Queue{}\n\tqueue.Enqueue(&BFSPair{Spn: s, Pname: \"\", Weight: -1.0})\n\tfor !queue.Empty() {\n\t\tcurrpair := queue.Dequeue().(*BFSPair)\n\t\tcurr, pname, pw := currpair.Spn, currpair.Pname, currpair.Weight\n\t\tch := curr.Ch()\n\t\tnch := len(ch)\n\n\t\tname := \"N\"\n\t\tcurrt := curr.Type()\n\n\t\t// In case it is a sum node. Else product node.\n\t\tif currt == \"sum\" {\n\t\t\tname = fmt.Sprintf(\"S%d\", nsums)\n\t\t\tfmt.Fprintf(file, \"%s = add_node(\\\"+\\\", \\\"#ff3300\\\")\\n\", name)\n\t\t\tnsums++\n\t\t} else if currt == \"product\" {\n\t\t\tname = fmt.Sprintf(\"P%d\", nprods)\n\t\t\tfmt.Fprintf(file, \"%s = add_node(\\\"*\\\", \\\"#669900\\\")\\n\", name)\n\t\t\tnprods++\n\t\t}\n\n\t\t// If pname is empty, then it is the root node. Else, link parent node to current node.\n\t\tif pname != \"\" {\n\t\t\tif pw >= 0 {\n\t\t\t\tfmt.Fprintf(file, \"add_edge(%s, %s, \\\"%.3f\\\")\\n\", pname, name, pw)\n\t\t\t} else {\n\t\t\t\tfmt.Fprintf(file, \"add_edge_nameless(%s, %s)\\n\", pname, name)\n\t\t\t}\n\t\t}\n\n\t\tvar w []float64\n\t\tif curr.Type() == \"sum\" {\n\t\t\tw = (curr.(*spn.Sum).Weights())\n\t\t}\n\t\t// For each children, run the BFS.\n\t\tfor i := 0; i < nch; i++ {\n\t\t\tc := ch[i]\n\n\t\t\t// If leaf, then simply write to the graphviz dot file. Else, recurse the BFS.\n\t\t\tif c.Type() == \"leaf\" {\n\t\t\t\tcname := fmt.Sprintf(\"X%d\", nvars)\n\t\t\t\tfmt.Fprintf(file, \"%s = add_node(\\\"X_%d\\\", \\\"#0066ff\\\")\\n\", cname, c.Sc()[0])\n\t\t\t\tnvars++\n\t\t\t\tif currt == \"sum\" {\n\t\t\t\t\tfmt.Fprintf(file, \"add_edge(%s, %s, \\\"%.3f\\\")\\n\", name, cname, w[i])\n\t\t\t\t} else {\n\t\t\t\t\tfmt.Fprintf(file, \"add_edge_nameless(%s, %s)\\n\", name, cname)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\ttw := -1.0\n\t\t\t\tif w != nil {\n\t\t\t\t\ttw = w[i]\n\t\t\t\t}\n\t\t\t\tqueue.Enqueue(&BFSPair{Spn: c, Pname: name, Weight: tw})\n\t\t\t}\n\t\t}\n\t}\n\n\tfmt.Fprintf(file, \"g.vertex_properties[\\\"name\\\"]=vnames\\n\")\n\tfmt.Fprintf(file, \"g.vertex_properties[\\\"color\\\"]=vcolors\\n\")\n\t//fmt.Fprintf(file, \"\\ngraph_draw(g, vertex_text=g.vertex_properties[\\\"name\\\"], \"+\n\t//\"edge_text=enames, vertex_fill_color=g.vertex_properties[\\\"color\\\"], \"+\n\t//\"output_size=[16384, 16384], output=\\\"%s\\\", bg_color=[1, 1, 1, 1])\\n\", outname)\n\tfmt.Fprintf(file, \"\\ngraph_draw(g, \"+\n\t\t\"edge_text=enames, vertex_fill_color=g.vertex_properties[\\\"color\\\"], \"+\n\t\t\"output_size=[16384, 16384], output=\\\"%s\\\", bg_color=[1, 1, 1, 1])\\n\", outname)\n}", "func encodeTree(hmt *Tree, finalTree *string) {\n\tif hmt == nil {\n\t\treturn\n\t}\n\t\n\tif hmt.LeftNode == nil && hmt.RightNode == nil{\n\t\t*finalTree += \"1\" + string(hmt.Char)\n\t} else {\n\t\t*finalTree += \"0\"\n\t}\n\tencodeTree(hmt.LeftNode, finalTree)\n\tencodeTree(hmt.RightNode, finalTree) \n}", "func (p InodePlugin) GraphDefinition() map[string]mp.Graphs {\n\treturn map[string]mp.Graphs{\n\t\t\"inode.count.#\": {\n\t\t\tLabel: \"Inode\",\n\t\t\tUnit: \"integer\",\n\t\t\tMetrics: []mp.Metrics{\n\t\t\t\t{Name: \"used\", Label: \"used\"},\n\t\t\t\t{Name: \"free\", Label: \"free\"},\n\t\t\t\t{Name: \"total\", Label: \"total\"},\n\t\t\t},\n\t\t},\n\t\t\"inode.percentage.#\": {\n\t\t\tLabel: \"Inode Percentage\",\n\t\t\tUnit: \"percentage\",\n\t\t\tMetrics: []mp.Metrics{\n\t\t\t\t{Name: \"used\", Label: \"used %\"},\n\t\t\t},\n\t\t},\n\t}\n}", "func PrintTree(g *adjacencylist.AdjacencyList, s, v *adjacencylist.AdjListVertex) {\n\tif v == s {\n\t\tfmt.Println(s)\n\t} else if v.P != nil {\n\t\tPrintTree(g, s, v.P)\n\t\tfmt.Println(v)\n\t} else {\n\t\tfmt.Println(\"A path between these two vertices does not exist :(\")\n\t}\n}", "func (this *Codec) serialize(root *TreeNode) string {\n var s string \n res:=helpSerialize(root,s)\n fmt.Println(res)\n return res\n}", "func (graph *DirGraph) createStrongConnections() {\n\tvar l int\n\tl = 0\n\tvar nodes []*GraphNode\n\tnodes = graph.setFinishingOrder()\n\tgraph.reset()\n\tvar i int\n\tfor i = len(nodes) - 1; i >= 0; i-- {\n\t\tvar node *GraphNode\n\t\tnode = nodes[i]\n\t\tif node.visited == false {\n\t\t\tvar m int\n\t\t\tm = l\n\t\t\tl++\n\t\t\tMarkStrongConnections(node, graph, m)\n\t\t}\n\t}\n}", "func (node *URLNode) generateTree(tabSize int) string {\n\tsubTree := \"\"\n\tfor _, child := range node.children {\n\t\tline := strings.Repeat(\"\\t\", tabSize)\n\t\tline += \"└── \"\n\t\tsubTree += line + child.generateTree(tabSize+1)\n\t}\n\treturn node.url + \"\\n\" + subTree\n}", "func (g DotGraph) String() string {\n\tcontent := []string{\"digraph \" + g.name + \" {\"}\n\tcontent = append(content, g.createGlobalAttributes()...)\n\n\tfor from, deps := range g.edges {\n\t\tnodeStylePattern := g.createNodeOptionsPatternString(from)\n\t\tcontent = append(content, from+nodeStylePattern)\n\n\t\tfor _, to := range deps {\n\t\t\tif from == `\"\"` && to.nodeID == `\"\"` {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tedgeStyle := g.createEdgeOptionsPatternString(to.nodeID, to.description)\n\t\t\tcontent = append(content, from+\"->\"+to.nodeID+edgeStyle)\n\t\t}\n\t}\n\n\tcontent = append(content, \"}\")\n\n\treturn strings.Join(content, \"\\n\")\n}", "func (f *Fragment) Encode(w io.Writer) error {\n\ttraf := f.Moof.Traf\n\terr := traf.OptimizeTfhdTrun()\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor _, b := range f.Children {\n\t\terr := b.Encode(w)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func PrettyPrint(graph *Graph) {\n\tfmt.Println(graph.Name)\n\tfmt.Println(\"\")\n\tstr := recursiveString(graph.Root, 0, true)\n\tfmt.Printf(\"%s\", str)\n}", "func printGroupStructure(g *clcv2.GroupInfo, indent string) {\n\tvar groupLine string\n\n\tif g.Type != \"default\" { // 'Archive' or similar: make it stand out\n\t\tgroupLine = fmt.Sprintf(\"%s[%s]/\", indent, g.Name)\n\t} else {\n\t\tgroupLine = fmt.Sprintf(\"%s%s/\", indent, g.Name)\n\t}\n\n\tif showFlags.GroupID {\n\t\tfmt.Printf(\"%-70s %s\\n\", groupLine, g.ID)\n\t} else {\n\t\tfmt.Printf(\"%s\\n\", groupLine)\n\t}\n\n\tfor _, s := range g.Servers {\n\t\tfmt.Printf(\"%s%s\\n\", indent+\" \", s)\n\t}\n\n\tfor _, g := range g.Groups {\n\t\tprintGroupStructure(g, indent+\" \")\n\t}\n}", "func (g *Graph) Write(w io.Writer, format string) error {\n\treturn templates.ExecuteTemplate(w, format+\".tmpl\", g)\n}", "func ownerToString(node *yaml.RNode) (string, error) {\n\tmeta, err := node.GetMeta()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tnamespace := meta.Namespace\n\n\towners, err := node.Pipe(yaml.Lookup(\"metadata\", \"ownerReferences\"))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif owners == nil {\n\t\treturn \"\", nil\n\t}\n\n\telements, err := owners.Elements()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif len(elements) == 0 {\n\t\treturn \"\", err\n\t}\n\towner := elements[0]\n\tvar kind, name string\n\n\tif value := owner.Field(\"kind\"); !value.IsNilOrEmpty() {\n\t\tkind = value.Value.YNode().Value\n\t}\n\tif value := owner.Field(\"name\"); !value.IsNilOrEmpty() {\n\t\tname = value.Value.YNode().Value\n\t}\n\n\treturn fmt.Sprintf(\"%s %s/%s\", kind, namespace, name), nil\n}", "func createGraph() *Graph {\n var g = Graph{}\n g.adjList = make(map[int]set)\n return &g\n}", "func saveAs(name string, node encoding.BinaryMarshaler) error {\n\tb, err := node.MarshalBinary()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err := ioutil.WriteFile(name, b, 0660); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func writeInode(file *os.File, index int64, inode *iNode) {\n\tfile.Seek(index, 0)\n\t//Empezamos el proceso de guardar en binario la data en memoria del struct\n\tvar binaryDisc bytes.Buffer\n\tbinary.Write(&binaryDisc, binary.BigEndian, inode)\n\twriteNextBytes(file, binaryDisc.Bytes())\n}", "func printMST(parent []int, vertices int, graph [][]float64, k int) {\n\n\tfmt.Println(\"Edge Weight\")\n\n\tfor i := 1; i < vertices; i++ {\n\t\tif comp.numberOfComponents != 1 {\n\t\t\tfmt.Printf(\"%d - %d %f \\n\", comp.compMaps[k][parent[i]], comp.compMaps[k][i], graph[i][parent[i]])\n\t\t\tadj_forest[comp.compMaps[k][parent[i]]-1][comp.compMaps[k][i]-1] = graph[i][parent[i]]\n\t\t\tadj_forest[comp.compMaps[k][i]-1][comp.compMaps[k][parent[i]]-1] = graph[i][parent[i]]\n\t\t} else {\n\t\t\tfmt.Printf(\"%d - %d %f \\n\", comp.compMaps[k][parent[i]+1], comp.compMaps[k][i+1], graph[i][parent[i]])\n\t\t\tadj_forest[comp.compMaps[k][parent[i]]][comp.compMaps[k][i]] = graph[i][parent[i]]\n\t\t\tadj_forest[comp.compMaps[k][i]][comp.compMaps[k][parent[i]]] = graph[i][parent[i]]\n\t\t}\n\n\t}\n\n}", "func (t *TreeStorage) Write(ctx context.Context, nodes []storage.Node) error {\n\tms := make([]*spanner.Mutation, 0, len(nodes))\n\tfor _, node := range nodes {\n\t\t// TODO(pavelkalinnikov): Consider doing just Insert when it is clear what\n\t\t// semantic the callers need.\n\t\tms = append(ms, spanner.InsertOrUpdate(\"TreeNodes\",\n\t\t\t[]string{\"TreeID\", \"ShardID\", \"NodeID\", \"NodeHash\"},\n\t\t\t[]interface{}{t.id, t.opts.shardID(node.ID), packNodeID(node.ID), node.Hash}))\n\t}\n\t_, err := t.c.Apply(ctx, ms)\n\treturn err\n}", "func (g *Group) SetOwner(svg *Svg) {\n\tg.Owner = svg\n\tfor _, gn := range g.Elements {\n\t\tswitch gn.(type) {\n\t\tcase *Group:\n\t\t\tgn.(*Group).Owner = g.Owner\n\t\t\tgn.(*Group).SetOwner(svg)\n\t\tcase *Path:\n\t\t\tgn.(*Path).group = g\n\t\t}\n\t}\n}", "func (graph *Graph) BuildGraph(count int, templateHTMLPath string, dumpFolder string) error {\n\tgraph.Nodes = []Node{}\n\tpageDocs, error := graph.ReadPages(count, templateHTMLPath)\n\tgraph.CreateAllPages(pageDocs)\n\tgraph.createOutboundLinks()\n\tgraph.dumpSite(dumpFolder)\n\treturn error\n}", "func (g *unionGraph) Union(a, b unionNode) {\n\tinA := g.find(a)\n\tinB := g.find(b)\n\tif inA != inB {\n\t\tif inA.rank > inB.rank {\n\t\t\tinB.parent = inA\n\t\t} else if inA.rank < inB.rank {\n\t\t\tinA.parent = inB\n\t\t} else { // inA.rank == inB.rank\n\t\t\tinA.parent = inB\n\t\t\tinB.rank++\n\t\t}\n\t}\n}", "func printGraph(format string, dependencies map[string]map[string]bool) {\n\tswitch format {\n\tcase formatDigraph:\n\t\tprintDigraph(os.Stdout, dependencies)\n\tcase formatGraphviz:\n\t\tprintGraphviz(os.Stdout, dependencies)\n\t}\n}", "func Generate(w io.Writer, g goflow.GraphRenderer) error {\n\ttmpl, err := template.\n\t\tNew(\"template\").\n\t\tFuncs(template.FuncMap{\n\t\t\t\"DepsNameAndTypes\": func(fields []goflow.Field) string {\n\t\t\t\ts := make([]string, len(fields))\n\t\t\t\tfor i, f := range fields {\n\t\t\t\t\ts[i] = fmt.Sprintf(\"%s %s\", f.Name, f.Type)\n\t\t\t\t}\n\t\t\t\treturn strings.Join(s, \"\\n\")\n\t\t\t},\n\t\t\t\"NameAndTypes\": func(fields []goflow.Field) string {\n\t\t\t\ts := make([]string, len(fields))\n\t\t\t\tfor i, f := range fields {\n\t\t\t\t\ts[i] = fmt.Sprintf(\"%s %s\", f.Name, f.Type)\n\t\t\t\t}\n\t\t\t\treturn strings.Join(s, \", \")\n\t\t\t},\n\t\t\t\"Names\": func(fields []goflow.Field) string {\n\t\t\t\ts := make([]string, len(fields))\n\t\t\t\tfor i, f := range fields {\n\t\t\t\t\ts[i] = f.Name\n\t\t\t\t}\n\t\t\t\treturn strings.Join(s, \", \")\n\t\t\t},\n\t\t}).\n\t\tParse(tmplStr)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn tmpl.Execute(w, nodeSorter{g})\n}", "func createGraph3() *NodeG {\n n1 := NodeG{1, nil}\n n2 := NodeG{2, nil}\n n3 := NodeG{3, nil}\n n1.neighbors = append(n1.neighbors, &n2)\n n2.neighbors = append(n2.neighbors, &n3)\n\n return &n1\n}", "func (a *RelationAssembler) GenerateJSON() bytes.Buffer {\n\n\t// buffer to store the json data\n\tvar buffer bytes.Buffer\n\n\t// keep track of sub-relations\n\tvar current int\n\tvar relations = make([]*gosmparse.Relation, 1)\n\trelations[0] = a.Relation\n\n\t// synchronize goroutines\n\tvar wg = &sync.WaitGroup{}\n\twg.Add(1)\n\n\t// write all members and sub members to buffer\n\tgo func() {\n\t\tdefer wg.Done()\n\t\tfor current < len(relations) {\n\n\t\t\t// contains too many child relations\n\t\t\tif len(relations) >= MAX_MEMBER_RELATIONS {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\twriteRelation(&buffer, a, &relations, relations[current])\n\t\t\tcurrent++\n\t\t}\n\t}()\n\n\t// done\n\twg.Wait()\n\n\t// debug max sub relations for this entity\n\t// fmt.Printf(\"%d: %d\\n\", a.Relation.ID, len(relations))\n\n\treturn buffer\n}", "func DumpToGraphviz(tables map[string]Table) {\n\tfmt.Printf(\"graph schema {\\n\")\n\tfmt.Printf(\" layout=fdp;\\n\")\n\tfmt.Printf(\" K=0.15;\\n\")\n\tfmt.Printf(\" maxiter=1000;\\n\")\n\tfmt.Printf(\" start=0;\\n\\n\")\n\n\tfor _, table := range tables {\n\t\tfmt.Printf(\"\\\"%s\\\" [shape=box];\\n\", table.Name)\n\n\t\tfor _, column := range table.Columns {\n\t\t\t_, primary := table.PKColumns[column]\n\t\t\tcolor := \"transparent\"\n\t\t\tif primary {\n\t\t\t\tcolor = \"gainsboro\"\n\t\t\t}\n\t\t\tfmt.Printf(\"\\\"%s-%s\\\" [label=\\\"\\\" xlabel=\\\"%s\\\" style=filled fillcolor=\\\"%s\\\"];\\n\", table.Name, column, column, color)\n\t\t\tfmt.Printf(\"\\\"%s\\\" -- \\\"%s-%s\\\";\\n\", table.Name, table.Name, column)\n\t\t}\n\n\t\tif len(table.PKSequence) > 0 {\n\t\t\tfmt.Printf(\"\\\"%s-id-%s\\\" [label=\\\"%s\\\" shape=note];\\n\", table.Name, table.PKSequence, table.PKSequence)\n\t\t\tfmt.Printf(\"\\\"%s-id\\\" -- \\\"%s-id-%s\\\" [style=dashed];\\n\", table.Name, table.Name, table.PKSequence)\n\t\t}\n\n\t\tfor _, index := range table.UniqueIndexes {\n\t\t\tlabel := \"unique\"\n\t\t\tif len(table.MainUniqueIndexName) > 0 {\n\t\t\t\tif strings.Compare(index.Name, table.MainUniqueIndexName) == 0 {\n\t\t\t\t\tlabel = \"unique main\"\n\t\t\t\t}\n\t\t\t}\n\t\t\tfmt.Printf(\"\\\"%s\\\" [label=\\\"%s\\\" shape=tab];\\n\", index.Name, label)\n\n\t\t\tfor _, indexColumn := range index.Columns {\n\t\t\t\tfmt.Printf(\"\\\"%s-%s\\\" -- \\\"%s\\\" [style=dashed];\\n\", table.Name, indexColumn, index.Name)\n\t\t\t}\n\t\t}\n\n\t\tfor i, reference := range table.References {\n\t\t\tfmt.Printf(\"\\\"%s-%s-%d\\\" [label=\\\"\\\" shape=diamond];\\n\", table.Name, reference.TableName, i)\n\n\t\t\tfor column, foreignColumn := range reference.ColumnMapping {\n\t\t\t\tfmt.Printf(\"\\\"%s-%s-%d\\\" -- \\\"%s-%s\\\";\\n\", table.Name, reference.TableName, i, table.Name, column)\n\t\t\t\tfmt.Printf(\"\\\"%s-%s-%d\\\" -- \\\"%s-%s\\\";\\n\", table.Name, reference.TableName, i, reference.TableName, foreignColumn)\n\t\t\t}\n\t\t}\n\t}\n\n\tfmt.Printf(\"}\")\n}", "func TransactionGraph(sourceOutput SiacoinOutputID, edges []TransactionGraphEdge) ([]Transaction, error) {\n\t// Generating the transaction graph based on a set of edges is non-trivial.\n\t//\n\t// Step 1: Generate a map of nodes. Each node records which nodes use it for\n\t// input, and which nodes use it for output. The map goes from node index to\n\t// node data.\n\t//\n\t// Step 2: Create a list of outputs that need to be added to a transaction.\n\t// The first element of this list will be node 0, which uses the source\n\t// output as its input.\n\t//\n\t// Step 3: For each node in the list, check whether that node has already\n\t// been added to a transaction for its outputs. If so, skip that node.\n\t//\n\t// Step 4: For the nodes whose outputs do not yet appear in a transaction,\n\t// create a transaction to house that node. Then follow each output of the\n\t// node to the inputs of the destination nodes.\n\t//\n\t// Step 5: For each input in a destination node, follow that input back to\n\t// the node that created the output. If that output already appears in a\n\t// transaction, the graph is invalid and an error must be returned. If that\n\t// node's outputs do not appear in a transaction yet, that node's inputs\n\t// need to be checked. If that node's inputs do not appear in a transaction\n\t// yet, the current transaction has to be put on hold and the transaction\n\t// for those inputs needs to be created by following the inputs back to\n\t// their corresponding outputs and starting back at step 2.\n\t//\n\t// Step 6: As the transactions are searched, any outputs created by the\n\t// transaction will need to be added to the list of outputs that haven't had\n\t// their transactions finished yet to be searched later.\n\t//\n\t// Step 7: Once all transaction diagrams are complete, translate into\n\t// transactions.\n\t//\n\t// In short, the algorithm we use is essentially a recursive\n\t// depth-first-search that builds the correct transaction graph, and then\n\t// the transactions are processed in an order that allows us to create all\n\t// of their IDs.\n\n\t// Basic input validation.\n\tif len(edges) < 1 {\n\t\treturn nil, errors.New(\"no graph specificed\")\n\t}\n\n\t// Check that the first value of 'sources' is zero, and that the rest of the\n\t// array is sorted.\n\tif edges[0].Source != 0 {\n\t\treturn nil, errors.New(\"first edge must speficy node 0 as the parent\")\n\t}\n\tif edges[0].Dest != 1 {\n\t\treturn nil, errors.New(\"first edge must speficy node 1 as the child\")\n\t}\n\tlatest := edges[0].Source\n\tfor _, edge := range edges {\n\t\tif edge.Source < latest {\n\t\t\treturn nil, errors.New(\"'sources' input is not sorted\")\n\t\t}\n\t\tlatest = edge.Source\n\t}\n\n\t// Create the set of output ids, and fill out the input ids for the source\n\t// transaction.\n\tbiggest := 0\n\tfor _, edge := range edges {\n\t\tif edge.Dest > biggest {\n\t\t\tbiggest = edge.Dest\n\t\t}\n\t}\n\ttxnInputs := make([][]SiacoinOutputID, biggest+1)\n\ttxnInputs[0] = []SiacoinOutputID{sourceOutput}\n\n\t// Go through the nodes bit by bit and create outputs.\n\t// Fill out the outputs for the source.\n\ti, j := 0, 0\n\tts := make([]Transaction, edges[len(edges)-1].Source+1)\n\tfor i < len(edges) {\n\t\tvar t Transaction\n\n\t\t// Grab the inputs for this transaction.\n\t\tfor _, outputID := range txnInputs[j] {\n\t\t\tt.SiacoinInputs = append(t.SiacoinInputs, SiacoinInput{\n\t\t\t\tParentID: outputID,\n\t\t\t})\n\t\t}\n\n\t\t// Grab the outputs for this transaction.\n\t\tstartingPoint := i\n\t\tcurrent := edges[i].Source\n\t\tfor i < len(edges) && edges[i].Source == current {\n\t\t\tt.SiacoinOutputs = append(t.SiacoinOutputs, SiacoinOutput{\n\t\t\t\tValue: edges[i].Value,\n\t\t\t\tUnlockHash: UnlockConditions{}.UnlockHash(),\n\t\t\t})\n\t\t\tif !edges[i].Fee.IsZero() {\n\t\t\t\tt.MinerFees = append(t.MinerFees, edges[i].Fee)\n\t\t\t}\n\t\t\ti++\n\t\t}\n\n\t\t// Record the inputs for the next transactions.\n\t\tfor k := startingPoint; k < i; k++ {\n\t\t\ttxnInputs[edges[k].Dest] = append(txnInputs[edges[k].Dest], t.SiacoinOutputID(uint64(k-startingPoint)))\n\t\t}\n\t\tts[j] = t\n\t\tj++\n\t}\n\n\treturn ts, nil\n}", "func (g *Graph) GobEncode() ([]byte, error) {\n\tgGob := graphGob{[]string{}, map[string]map[string]int{}}\n\n\t// add vertexes and edges to gGob\n\tfor key, v := range g.vertexes {\n\t\tgGob.Vertexes = append(gGob.Vertexes, key)\n\n\t\tgGob.Edges[key] = map[string]int{}\n\n\t\t// for each neighbor...\n\t\tfor neighbor, weight := range v.neighbors {\n\t\t\t// save the edge connection to the neighbor into the edges map\n\t\t\tgGob.Edges[key][neighbor.key] = weight\n\t\t}\n\t}\n\n\t// encode gGob\n\tbuf := &bytes.Buffer{}\n\tenc := gob.NewEncoder(buf)\n\terr := enc.Encode(gGob)\n\n\treturn buf.Bytes(), err\n}", "func (node *Node) printTree1(out *bytes.Buffer, isRight bool, indent string) {\n\n\tif (node.Left != nil) {\n\t\tstr := \" \"\n\t\tif isRight {\n\t\t\tstr = \" | \"\n\t\t}\n\t\tstr = indent + str\n\t\tnode.Left.printTree1(out, false, str)\n\t}\n\n\tout.Write([]byte(indent))\n\tif (isRight) {\n\t\tout.Write([]byte(\"\\\\\"))\n\t} else {\n\t\tout.Write([]byte (\"/\"))\n\t}\n\tout.Write([]byte(\"--\"))\n\n\tnode.printNodeValue(out)\n\n\tif (node.Right != nil) {\n\t\tstr := \" | \"\n\t\tif isRight {\n\t\t\tstr = \" \"\n\t\t}\n\t\tstr = indent + str\n\t\tnode.Right.printTree1(out, true, str)\n\t}\n\n}", "func (wr *plainGenomeWriter) writeNetworkNode(n *network.NNode) error {\n\ttrait_id := 0\n\tif n.Trait != nil {\n\t\ttrait_id = n.Trait.Id\n\t}\n\tact_str, err := utils.NodeActivators.ActivationNameFromType(n.ActivationType)\n\tif err == nil {\n\t\t_, err = fmt.Fprintf(wr.w, \"%d %d %d %d %s\", n.Id, trait_id, n.NodeType(),\n\t\t\tn.NeuronType, act_str)\n\t}\n\treturn err\n}", "func EncodeDAG(ctx context.Context, dag format.NodeGetter, e Encoder, nd format.Node, r Recoverability) (format.Node, error) {\n\tif len(nd.Links()) == 0 {\n\t\treturn nd, nil\n\t}\n\n\tfor _, l := range nd.Links() {\n\t\tnd, err := l.GetNode(ctx, dag)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tend, err := EncodeDAG(ctx, dag, e, nd, r)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif !nd.Cid().Equals(end.Cid()) {\n\t\t\tl.Size, err = end.Size()\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\tl.Cid = end.Cid()\n\t\t}\n\t}\n\n\treturn e.Encode(ctx, nd, r)\n}", "func createGraph1() *NodeG {\n n1 := NodeG{2, nil}\n n2 := NodeG{4, nil}\n n1.neighbors = append(n1.neighbors, &n2)\n n1.neighbors = append(n1.neighbors, &n1)\n n2.neighbors = append(n2.neighbors, &n1)\n\n // fmt.Println(\">>>>>>1\")\n // fmt.Println(n1)\n // fmt.Println(n2)\n // fmt.Println(\">>>>>>2\")\n\n return &n1\n}", "func main() {\n\troot := &TreeNode{\n\t\tVal: 2,\n\t\tLeft: &TreeNode{\n\t\t\tVal: 1,\n\t\t},\n\t\tRight: &TreeNode{\n\t\t\tVal: 3,\n\t\t},\n\t}\n\tprintTreeNodeByDFS(root)\n\tfmt.Println()\n\n\tser := Constructor()\n\ttreeString := ser.serialize(root)\n\tfmt.Println(treeString)\n\tans := ser.deserialize(treeString)\n\t// printTreeNodeByBFS(ans)\n\tprintTreeNodeByDFS(ans)\n\tfmt.Println()\n}", "func (g *Graph) Node(gvk schema.GroupVersionKind, obj metav1.Object) *Node {\n\tapiVersion, kind := gvk.ToAPIVersionAndKind()\n\tnode := &Node{\n\t\tTypeMeta: metav1.TypeMeta{\n\t\t\tAPIVersion: apiVersion,\n\t\t\tKind: kind,\n\t\t},\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tUID: obj.GetUID(),\n\t\t\tNamespace: obj.GetNamespace(),\n\t\t\tName: obj.GetName(),\n\t\t\tAnnotations: FilterByValue(obj.GetAnnotations(), func(v string) bool {\n\t\t\t\treturn !strings.HasPrefix(v, \"{\") && !strings.HasPrefix(v, \"[\")\n\t\t\t}),\n\t\t\tLabels: obj.GetLabels(),\n\t\t},\n\t}\n\n\tif n, ok := g.Nodes[obj.GetUID()]; ok {\n\t\tif len(n.GetAnnotations()) != 0 {\n\t\t\tnode.SetAnnotations(n.GetAnnotations())\n\t\t}\n\t\tif len(n.GetLabels()) != 0 {\n\t\t\tnode.SetLabels(n.GetLabels())\n\t\t}\n\t}\n\n\tg.Nodes[obj.GetUID()] = node\n\n\tfor _, ownerRef := range obj.GetOwnerReferences() {\n\t\towner := g.Node(\n\t\t\tschema.FromAPIVersionAndKind(ownerRef.APIVersion, ownerRef.Kind),\n\t\t\t&metav1.ObjectMeta{\n\t\t\t\tUID: ownerRef.UID,\n\t\t\t\tName: ownerRef.Name,\n\t\t\t\tNamespace: obj.GetNamespace(),\n\t\t\t},\n\t\t)\n\t\tg.Relationship(owner, kind, node)\n\t}\n\n\treturn node\n}", "func (g *CallGraph) WriteGraphviz(w io.Writer) error {\n\tif g.edges == nil {\n\t\tif err := callgraph.GraphVisitEdges(g.cg, g.populateEdges); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tbufw := bufio.NewWriter(w)\n\tbufw.WriteString(\"digraph callgraph {\\n\")\n\t// Instead of using template..\n\tfor _, edge := range g.edges {\n\t\tbufw.WriteString(fmt.Sprintf(\" %q -> %q\\n\", edge.Caller, edge.Callee))\n\t}\n\tbufw.WriteString(\"}\\n\")\n\tbufw.Flush()\n\treturn nil\n}", "func (m *ConnectedOrganizationMembers) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n err := m.SubjectSet.Serialize(writer)\n if err != nil {\n return err\n }\n {\n err = writer.WriteStringValue(\"connectedOrganizationId\", m.GetConnectedOrganizationId())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"description\", m.GetDescription())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\treturn dfsSerial(root, \"\")\n}", "func create(\n\ttrx storage.Transaction,\n\ttxId merkle.Digest,\n\townerData OwnerData,\n\towner *account.Account,\n) {\n\t// increment the count for owner\n\tnKey := owner.Bytes()\n\tcount := trx.Get(storage.Pool.OwnerNextCount, nKey)\n\tif nil == count {\n\t\tcount = []byte{0, 0, 0, 0, 0, 0, 0, 0}\n\t} else if uint64ByteSize != len(count) {\n\t\tlogger.Panic(\"OwnerNextCount database corrupt\")\n\t}\n\tnewCount := make([]byte, uint64ByteSize)\n\tbinary.BigEndian.PutUint64(newCount, binary.BigEndian.Uint64(count)+1)\n\ttrx.Put(storage.Pool.OwnerNextCount, nKey, newCount, []byte{})\n\n\t// write to the owner list\n\toKey := append(owner.Bytes(), count...)\n\ttrx.Put(storage.Pool.OwnerList, oKey, txId[:], []byte{})\n\n\t// write new index record\n\tdKey := append(owner.Bytes(), txId[:]...)\n\ttrx.Put(storage.Pool.OwnerTxIndex, dKey, count, []byte{})\n\n\t// save owner data record\n\ttrx.Put(storage.Pool.OwnerData, txId[:], ownerData.Pack(), []byte{})\n}", "func writeNode(w io.Writer, fset *token.FileSet, x interface{}) {\n\t// convert trailing tabs into spaces using a tconv filter\n\t// to ensure a good outcome in most browsers (there may still\n\t// be tabs in comments and strings, but converting those into\n\t// the right number of spaces is much harder)\n\t//\n\t// TODO(gri) rethink printer flags - perhaps tconv can be eliminated\n\t// with an another printer mode (which is more efficiently\n\t// implemented in the printer than here with another layer)\n\tmode := printer.TabIndent | printer.UseSpaces\n\terr := (&printer.Config{Mode: mode, Tabwidth: tabwidth}).Fprint(&tconv{output: w}, fset, x)\n\tif err != nil {\n\t\tlog.Print(err)\n\t}\n}", "func (g *graph) String() string{\n\ts := \"Graph: \\n\"\n\tfor i,list := range g.nodes {\n\t\ts += \"Node \" + strconv.Itoa(i) + \":\\t\"\n\t\tfor n := range list.Iter() {\n\t\t\ts += strconv.Itoa(n.(int)) + \"\\t\"\n\t\t}\n\t\ts += \"\\n\"\n\t}\n\treturn s\n}", "func (s *ShortenBlock) nodeWrite(node *Node, data []byte) error {\n\tvar newID string\n\tvar err error\n\tlog.Debugf(\"writing %d bytes to node\", len(data))\n\tif newID, err = s.shortener.Write(data); err != nil {\n\t\treturn err\n\t}\n\tlog.Tracef(\"node id changed from %s to %s\", node.id, newID)\n\n\tnode.id = newID\n\tfor {\n\t\tnode = node.parent\n\t\tlog.Tracef(\"updating parent node %s\", node.id)\n\n\t\tvar childIDs []string\n\t\tfor _, child := range node.children {\n\t\t\tchildIDs = append(childIDs, child.id)\n\t\t}\n\t\tnewData := strings.Join(childIDs, \",\")\n\t\tlog.Tracef(\"new child nodes are %s\", newData)\n\n\t\tvar newID string\n\t\tlog.Debugf(\"writing %d bytes to node parent\", len(newData))\n\t\tif newID, err = s.shortener.Write([]byte(newData)); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tnode.id = newID\n\t\tfor _, child := range node.children {\n\t\t\tchild.parent = node\n\t\t}\n\t\tif node.parent == nil {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn nil\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tnodes := []*TreeNode{ root }\n\tstrList := []string{}\n\tfor i := 0; i < len(nodes); {\n\t\tcnt := len(nodes)\n\t\tfor ; i < cnt; i++ {\n\t\t\tnode := nodes[i]\n\t\t\tif node == nil {\n\t\t\t\tstrList = append(strList, \"null\")\n\t\t\t} else {\n\t\t\t\tstrList = append(strList, strconv.Itoa(node.Val))\n\t\t\t\tnodes = append(nodes, node.Left)\n\t\t\t\tnodes = append(nodes, node.Right)\n\t\t\t}\n\t\t}\n\t}\n\treturn \"[\" + strings.Join(strList, \"\") + \"]\"\n}", "func (g *NewGraph) PrintGraph() {\n\tfor i := range g.NodeSets {\n\t\tfmt.Println(g.NodeSets[i])\n\t}\n}", "func main() {\n graph := createGraph()\n graph.addEdge(1, 2)\n graph.addEdge(2, 3)\n graph.addEdge(2, 4)\n graph.addEdge(3, 4)\n graph.addEdge(1, 5)\n graph.addEdge(5, 6)\n graph.addEdge(5, 7)\n\n visited := make(set)\n\n dfs(graph, 1, visited, func(node int) {\n fmt.Print(node, \" \")\n })\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tif root == nil {\n\t\treturn \"\"\n\t}\n\n\tres := make([]string, 0)\n\tqueue := make([]*TreeNode, 0)\n\tqueue = append(queue, root)\n\tfor len(queue) > 0 {\n\t\tl := len(queue)\n\t\tfor i := 0; i < l; i++ {\n\t\t\tnode := queue[0]\n\t\t\tqueue = queue[1:]\n\t\t\tif node != nil {\n\t\t\t\tres = append(res, strconv.Itoa(node.Val))\n\t\t\t\tqueue = append(queue, node.Left, node.Right)\n\t\t\t} else {\n\t\t\t\tres = append(res, \"#\")\n\t\t\t}\n\n\t\t}\n\t}\n\treturn strings.Join(res, \",\")\n}", "func (c cluster) WriteOrder(hash string) []node {\n\treturn hashOrder(hash, len(c.GetNeighbors())+1, c.WriteRing())\n}", "func main() {\n\n nodes := []graph.Node{}\n router := Router{ \"routerA\",1 }\n router2 := Router{ \"routerB\",2 }\n subnet := Subnet{ \"subnet1\", 10}\n nodes = append(nodes, router)\n// nodes = append(nodes, subnet)\n g := graph.NewGraph(nodes)\n g.AddNode(subnet)\n g.AddNode(router2)\n\n\n g.SetEdge(router, subnet)\n g.SetEdge(router2, subnet)\n\n g.Dump()\n\n// weight := float64(40)\n// edge := g.NewWeightedEdge(router, subnet, weight)\n// g.SetWeightedEdge(edge)\n\n// fmt.Printf(\"%v\\n\", g)\n// g.Dump()\n\n/*\n self := 0.0 // the cost of self connection\n absent := 10.0 // the wieght returned for absent edges\n\n graph := simple.NewWeightedUndirectedGraph(self, absent)\n fmt.Printf(\"%v\\n\", graph)\n\n var id int64\n //var node simple.Node\n\n id = 0\n from := simple.Node(id)\n graph.AddNode(from)\n\n id = 1\n to := simple.Node(id)\n graph.AddNode(to)\n\n id = 2\n from2 := simple.Node(id)\n graph.AddNode(from2)\n\n id = 3\n to2 := simple.Node(id)\n graph.AddNode(to2)\n\n\n nodeA := graph.Node(int64(2))\n\n\n\n fmt.Printf(\"%v\\n\", graph)\n\n nodes := graph.Nodes()\n fmt.Printf(\"%v\\n\", nodes)\n fmt.Printf(\"%v\\n\", nodeA)\n\n weight := float64(40)\n edge := graph.NewWeightedEdge(from, to, weight)\n graph.SetWeightedEdge(edge)\n\n edge2 := graph.NewWeightedEdge(from2, to2, weight)\n graph.SetWeightedEdge(edge2)\n\n fmt.Printf(\"%v\\n\", graph)\n edges := graph.Edges()\n fmt.Printf(\"%v\\n\", edges)\n\n edge_ := graph.Edge(int64(0) ,int64(1))\n fmt.Printf(\"%v\\n\", edge_)\n*/\n}", "func (i *IndexBuilder) walk() error {\n\treturn i.g.CommitIter(func(branch string, c *git.Commit, tree *git.Tree, head bool) error {\n\t\tlog.Infof(\"Inspecting commit-id '%s/%s'\", branch, c.Id().String())\n\t\tvar chartDirs []string\n\t\tvar err error\n\t\tif head && branch == \"master\" {\n\t\t\tlog.Infof(\"HEAD: Retrieving all chart directories...\")\n\t\t\tchartDirs, err = i.listAllChartDirs()\n\t\t} else {\n\t\t\tchartDirs, err = i.listModifiedChartDirs(c, tree)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tlog.Debugf(\"Chart directories: '%v'\", chartDirs)\n\t\treturn i.inspectDirs(chartDirs, branch, c, head)\n\t})\n}", "func (agent *MerkleAgent) Serialize() []byte{\n\tsize := config.Size\n\tret := make([]byte, 4 + 4 + size + int(agent.H) * size)\n\tbinary.LittleEndian.PutUint32(ret[0:4], agent.H)\n\tbinary.LittleEndian.PutUint32(ret[4:8], uint32(size))\n\tcopy(ret[8:8+size], agent.root[:])\n\toffset := 8 + size\n\tfor i := 0; i < int(agent.H); i++{\n\t\tcopy(ret[offset:offset+size], agent.auth[i][:])\n\t\toffset += size\n\t}\n\tfor i := 0; i < int(agent.H); i++{\n\t\ttreeHashBytes := agent.treeHashStacks[i].Serialize()\n\t\tret = append(ret, treeHashBytes...)\n\t}\n\tfor _, node := range agent.nodeHouse{\n\t\tret = append(ret, node...)\n\t}\n\treturn ret\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tif root == nil {\n\t\treturn \"nil\"\n\t}\n\treturn strconv.Itoa(root.Val) + \",\" + this.serialize(root.Left) + \",\" + this.serialize(root.Right)\n}", "func stringify(n *Node, level int) {\n\tif n != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"---[ \"\n\t\tlevel++\n\t\tstringify(n.left, level)\n\t\tfmt.Printf(format+\"%d\\n\", n.key)\n\t\tstringify(n.right, level)\n\t}\n}", "func (e EmptyNode) EncodeBinary(*io.BinWriter) {\n}", "func encodeNodeWithType(n Node, w *io.BinWriter) {\n\tw.WriteB(byte(n.Type()))\n\tn.EncodeBinary(w)\n}", "func Generate(data io.ReadSeeker, dataSize int64, treeReader io.ReadSeeker, treeWriter io.WriteSeeker, dataAndTreeInSameFile bool) ([]byte, error) {\n\tlayout := InitLayout(dataSize, dataAndTreeInSameFile)\n\n\tnumBlocks := (dataSize + layout.blockSize - 1) / layout.blockSize\n\n\t// If the data is in the same file as the tree, zero pad the last data\n\t// block.\n\tbytesInLastBlock := dataSize % layout.blockSize\n\tif dataAndTreeInSameFile && bytesInLastBlock != 0 {\n\t\tzeroBuf := make([]byte, layout.blockSize-bytesInLastBlock)\n\t\tif _, err := treeWriter.Seek(0, io.SeekEnd); err != nil && err != io.EOF {\n\t\t\treturn nil, err\n\t\t}\n\t\tif _, err := treeWriter.Write(zeroBuf); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Store the current offset, so we can set it back once verification\n\t// finishes.\n\torigOffset, err := data.Seek(0, io.SeekCurrent)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer data.Seek(origOffset, io.SeekStart)\n\n\t// Read from the beginning of both data and treeReader.\n\tif _, err := data.Seek(0, io.SeekStart); err != nil && err != io.EOF {\n\t\treturn nil, err\n\t}\n\n\tif _, err := treeReader.Seek(0, io.SeekStart); err != nil && err != io.EOF {\n\t\treturn nil, err\n\t}\n\n\tvar root []byte\n\tfor level := 0; level < layout.numLevels(); level++ {\n\t\tfor i := int64(0); i < numBlocks; i++ {\n\t\t\tbuf := make([]byte, layout.blockSize)\n\t\t\tvar (\n\t\t\t\tn int\n\t\t\t\terr error\n\t\t\t)\n\t\t\tif level == 0 {\n\t\t\t\t// Read data block from the target file since level 0 includes hashes\n\t\t\t\t// of blocks in the input data.\n\t\t\t\tn, err = data.Read(buf)\n\t\t\t} else {\n\t\t\t\t// Read data block from the tree file since levels higher than 0 are\n\t\t\t\t// hashing the lower level hashes.\n\t\t\t\tn, err = treeReader.Read(buf)\n\t\t\t}\n\n\t\t\t// err is populated as long as the bytes read is smaller than the buffer\n\t\t\t// size. This could be the case if we are reading the last block, and\n\t\t\t// break in that case. If this is the last block, the end of the block\n\t\t\t// will be zero-padded.\n\t\t\tif n == 0 && err == io.EOF {\n\t\t\t\tbreak\n\t\t\t} else if err != nil && err != io.EOF {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\t// Hash the bytes in buf.\n\t\t\tdigest := sha256.Sum256(buf)\n\n\t\t\tif level == layout.rootLevel() {\n\t\t\t\troot = digest[:]\n\t\t\t}\n\n\t\t\t// Write the generated hash to the end of the tree file.\n\t\t\tif _, err = treeWriter.Write(digest[:]); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t\t// If the generated digests do not round up to a block, zero-padding the\n\t\t// remaining of the last block. But no need to do so for root.\n\t\tif level != layout.rootLevel() && numBlocks%layout.hashesPerBlock() != 0 {\n\t\t\tzeroBuf := make([]byte, layout.blockSize-(numBlocks%layout.hashesPerBlock())*layout.digestSize)\n\t\t\tif _, err := treeWriter.Write(zeroBuf[:]); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t\tnumBlocks = (numBlocks + layout.hashesPerBlock() - 1) / layout.hashesPerBlock()\n\t}\n\treturn root, nil\n}" ]
[ "0.6092704", "0.586756", "0.58613664", "0.57382333", "0.5712016", "0.56290585", "0.5610613", "0.5602439", "0.55215484", "0.54591393", "0.5436809", "0.5413327", "0.53986627", "0.5312661", "0.5216152", "0.51965225", "0.514187", "0.5128386", "0.5080789", "0.5075414", "0.5053383", "0.5034032", "0.5033987", "0.499279", "0.49819502", "0.49766588", "0.4963234", "0.49425358", "0.4932012", "0.4922566", "0.4915826", "0.49048176", "0.490085", "0.48961908", "0.48816642", "0.48592773", "0.48552522", "0.48508236", "0.48384094", "0.48306322", "0.4815688", "0.47982213", "0.47787225", "0.47553867", "0.47498515", "0.47466788", "0.4731142", "0.47304294", "0.47137746", "0.47136247", "0.4710455", "0.47006166", "0.46816114", "0.4680425", "0.4679492", "0.46757782", "0.46745726", "0.46611688", "0.46491674", "0.46379918", "0.4636279", "0.46334285", "0.4624189", "0.4620127", "0.46134523", "0.46006948", "0.4594411", "0.45867515", "0.45862913", "0.45790902", "0.45780545", "0.45650476", "0.4564029", "0.45560074", "0.45549473", "0.45525247", "0.4544659", "0.45418578", "0.45385662", "0.45352012", "0.4534263", "0.45309746", "0.4526098", "0.45239624", "0.45168886", "0.45143354", "0.45099995", "0.4508674", "0.45071885", "0.44904825", "0.44784206", "0.4470868", "0.44708264", "0.44685042", "0.44628704", "0.44603235", "0.44561446", "0.44504684", "0.44480306", "0.44463006" ]
0.7600591
0
nodeToString generates a string to identify the node matches ownerToString format
func nodeToString(node *yaml.RNode) (string, error) { meta, err := node.GetMeta() if err != nil { return "", err } return fmt.Sprintf("%s %s/%s", meta.Kind, meta.Namespace, meta.Name), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ownerToString(node *yaml.RNode) (string, error) {\n\tmeta, err := node.GetMeta()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tnamespace := meta.Namespace\n\n\towners, err := node.Pipe(yaml.Lookup(\"metadata\", \"ownerReferences\"))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif owners == nil {\n\t\treturn \"\", nil\n\t}\n\n\telements, err := owners.Elements()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif len(elements) == 0 {\n\t\treturn \"\", err\n\t}\n\towner := elements[0]\n\tvar kind, name string\n\n\tif value := owner.Field(\"kind\"); !value.IsNilOrEmpty() {\n\t\tkind = value.Value.YNode().Value\n\t}\n\tif value := owner.Field(\"name\"); !value.IsNilOrEmpty() {\n\t\tname = value.Value.YNode().Value\n\t}\n\n\treturn fmt.Sprintf(\"%s %s/%s\", kind, namespace, name), nil\n}", "func NodeToString(node core.Node) string {\n\tvar b strings.Builder\n\tfmt.Println()\n\tnodeRecurse(node, 0, &b)\n\treturn b.String()\n}", "func nodeToString(fs *token.FileSet, n ast.Node) string {\n\tb := bytes.NewBuffer([]byte{})\n\tprinter.Fprint(b, fs, n)\n\treturn b.String()\n}", "func StringFromNode(n Node) string {\n\treturn strings.Join([]string{n.Address(), n.PublicKey().String()}, \"/\")\n}", "func (n *Node) String() string {\n\treturn fmt.Sprintf(\"%s [%d]\", n.ID_, n.Group_)\n}", "func (n *node) String() string {\n\treturn \"<\" + n.view.String() + \"-\\n\" + n.children[0].String() + \", \\n\" + n.children[1].String() + \", \\n\" + n.children[2].String() + \", \\n\" + n.children[3].String() + \">\"\n}", "func (n *TreeNode) ToString() string {\n\tif n == nil {\n\t\treturn \"nil\"\n\t}\n\tif n.Left == nil && n.Right == nil {\n\t\treturn strconv.Itoa(n.Data)\n\t}\n\treturn fmt.Sprintf(\"%d => (%s, %s)\", n.Data, n.Left.ToString(), n.Right.ToString())\n}", "func (n *Node) String() string { return fmt.Sprint(n.Data) }", "func (n Node) String() string {\n\treturn fmt.Sprintf(\"<Node id=%s, data=%s>\", n.ID, n.Data)\n}", "func (n Node) String() string {\n\treturn fmt.Sprintf(\"ID: %s\\nName: %s\\nPort: %s\\nDesc: %s\\nLoc: %s\", n.ID, n.Name, n.Port, n.Desc, n.Loc)\n}", "func (n *Node) String() string {\n\treturn fmt.Sprintf(\"%t %t %v %s\", n.leaf, n.Dup, n.Hash, n.C)\n}", "func (n NodeID) String() string {\n\treturn fmt.Sprintf(\"%x\", n[:])\n}", "func (n *Nodes) String() string {\n\treturn fmt.Sprintf(\"%s\", *n)\n}", "func (n NodeID) String() string {\n\treturn hexutil.Encode(n[:])\n}", "func (n Node) String() string {\n\tstr := \"\"\n\tn.FuncDownMeFirst(0, nil, func(k Ki, level int, d interface{}) bool {\n\t\tfor i := 0; i < level; i++ {\n\t\t\tstr += \"\\t\"\n\t\t}\n\t\tstr += k.Name() + \"\\n\"\n\t\treturn true\n\t})\n\treturn str\n}", "func NodesToString(collection *list.List) string {\n\tif !log.IsEnabledFor(logging.DEBUG) {\n\t\treturn \"\"\n\t}\n\n\tresult := \"\"\n\tfor el := collection.Front(); el != nil; el = el.Next() {\n\t\tresult = result + \"\\n\" + NodeToString(el.Value.(*CandidateNode))\n\t}\n\treturn result\n}", "func (n Node) String() string {\n\t// using json's marshal indent to format and return node information\n\ts, _ := json.MarshalIndent(n, \"\", \"\\t\")\n\treturn string(s)\n}", "func (id NodeID) String() string {\n\tunversionedID := id.unversioned()\n\treturn base58.CheckEncode(unversionedID[:], byte(id.Version().Number))\n}", "func (n Node) String() string {\n\tc := n.GetConfig()\n\n\ts := n.Info.Name() + \"\\n\"\n\tif n.getDepth() == 0 || c.FullPaths {\n\t\ts = n.Path + \"\\n\"\n\t}\n\n\tfor _, v := range n.Children {\n\t\tif !c.DisableIndentation {\n\t\t\ts += v.generatePrefix()\n\t\t}\n\t\ts += v.String()\n\t}\n\n\treturn s\n}", "func (nd *Node) String() string {\n\treturn nd.name + \" \" + nd.host.String()\n}", "func (n *Node) String() string {\n\ts := fmt.Sprintf(\"Level: %d\\nValues: %d\\nChildren: %t\\nBox: %s\\n\",\n\t\tn.level, len(n.values), n.children[0] != nil, n.boundingBox.String())\n\n\tif n.children[0] != nil {\n\t\tfor i, _ := range n.children {\n\t\t\ts = s + n.children[i].String()\n\t\t}\n\t}\n\n\treturn s\n}", "func (n *Node) String() (s string) {\n\treturn fmt.Sprintf(\"\\\"%s\\\" ==> %s/%d %s %s/%d\", n.Original, n.Lref, n.Lval, n.Op, n.Rref, n.Rval)\n}", "func (n *Node) String() string {\n\tvar links string\n\tfor k, n := range n.Nodes {\n\t\tlinks += fmt.Sprintf(\"%s:%s \", k, n.Name)\n\t}\n\tif len(n.Nodes) > 0 {\n\t\tlinks = links[:len(links)-1]\n\t}\n\treturn fmt.Sprintf(\"name=%s links=map[%s]\", n.Name, links)\n}", "func str(n *html.Node) string {\n if n == nil {\n return \"nil\"\n }\n switch n.Type {\n case html.ErrorNode:\n return fmt.Sprintf(`ErrNode(\"%s\")`, n.Data)\n case html.TextNode:\n return fmt.Sprintf(\"%dB of text\", len([]byte(n.Data)))\n case html.DocumentNode:\n return fmt.Sprintf(\"Document\")\n case html.ElementNode:\n strs := make([]string, 0, len(n.Attr) + 1)\n strs = append(strs, n.Data)\n for _, attr := range n.Attr {\n strs = append(strs, fmt.Sprintf(`%s=\"%s\"`, attr.Key, attr.Val))\n }\n return \"<\" + strings.Join(strs, \" \") + \">\"\n case html.CommentNode:\n return fmt.Sprintf(\"%dB of comments\", len([]byte(n.Data)))\n case html.DoctypeNode:\n return fmt.Sprintf(\"doctype:%s\", n.Data)\n }\n panic(`Invalid node type.`)\n}", "func (n *Node) String() string {\n\treturn fmt.Sprintf(\"%v|%d\", n.addr, n.currentNumConnections)\n}", "func (n *Node) String() string {\n\tif n.ServerStartTime.IsZero() {\n\t\treturn fmt.Sprintf(\"{Redis ID: %s, role: %s, master: %s, link: %s, status: %s, addr: %s, slots: %s, len(migratingSlots): %d, len(importingSlots): %d}\",\n\t\t\tn.ID, n.GetRole(), n.MasterReferent, n.LinkState, n.FailStatus, n.IPPort(), SlotSlice(n.Slots), len(n.MigratingSlots), len(n.ImportingSlots))\n\t}\n\treturn fmt.Sprintf(\"{Redis ID: %s, role: %s, master: %s, link: %s, status: %s, addr: %s, slots: %s, len(migratingSlots): %d, len(importingSlots): %d, ServerStartTime: %s}\",\n\t\tn.ID, n.GetRole(), n.MasterReferent, n.LinkState, n.FailStatus, n.IPPort(), SlotSlice(n.Slots), len(n.MigratingSlots), len(n.ImportingSlots), n.ServerStartTime.Format(\"2006-01-02 15:04:05\"))\n}", "func (o *V0037Node) GetOwner() string {\n\tif o == nil || o.Owner == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Owner\n}", "func (s Node) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (n *Node) String() string {\n\tname := n.Field\n\tif n.IsArray {\n\t\tname = \"array[\" + n.TypeName + \"]\"\n\t}\n\n\tif n.ChildNode != nil {\n\t\treturn name + \".\" + n.ChildNode.String()\n\t}\n\tif len(n.TypeName) > 0 {\n\t\treturn name + \" : \" + n.TypeName\n\t}\n\treturn name\n}", "func (n *Node) String() string {\n\treturn fmt.Sprintf(\"%v\", n.data)\n}", "func (n *Node) String() string {\n\tif n.Children.Len() == 0 {\n\t\tfmt.Sprintf(\"(%.2f)\", n.Value)\n\t\treturn fmt.Sprintf(\"(%.2f)\", n.Value)\n\t}\n\tfmt.Sprintf(\"(%.2f: %s)\", n.Value, n.Children)\n\treturn fmt.Sprintf(\"(%.2f: %s)\", n.Value, n.Children)\n}", "func (p NodeSet) ToString(str func(node Node) string) (items []string) {\n\tif p.Err != nil {\n\t\treturn nil\n\t}\n\tp.Data.ForEach(func(node Node) error {\n\t\titems = append(items, str(node))\n\t\treturn nil\n\t})\n\treturn\n}", "func (n *Node[T]) String() string {\n\tif n == nil {\n\t\treturn \"nil\"\n\t}\n\treturn fmt.Sprintf(\"%v\\n%s\\n%s\", n.Value,\n\t\tindentString(n.Left.String()),\n\t\tindentString(n.Right.String()))\n}", "func (n *Node) String() string {\n\treturn n.recString(0)\n}", "func (n *Node) toxstring() string {\n\treturn strings.Replace(n.Nodestr, \"/\", \"+\", -1)\n}", "func (nodeID *NodeID) String() string {\n\treturn hex.EncodeToString(nodeID[0:IDLength])\n}", "func (n NamespaceNode) String() string {\n\treturn string(n)\n}", "func (n Node) String() string {\n\treturn n.pubKey.String()\n}", "func (n *Node) String() string {\n\treturn n.TcpAddr.String()\n}", "func (n NodesID) String() string {\n\treturn fmt.Sprintf(\"%x\", n[:])\n}", "func (i *IpldRawNode) String() string {\n\treturn fmt.Sprintf(\"<IpldRawNode %s>\", i.cid)\n}", "func String(node SQLNode) string {\n\tif node == nil {\n\t\treturn \"<nil>\"\n\t}\n\n\tbuf := NewTrackedBuffer(nil)\n\tbuf.Myprintf(\"%v\", node)\n\treturn buf.String()\n}", "func (n *Node) String() string {\n\treturn n.Name\n}", "func (jfn jobFileNode) String() string {\n\treturn fmt.Sprintf(\"JobFileNode<PARENT=[%s] NAME=[%s]>\", jfn.jobNode.parentNodePath, jfn.info.Name())\n}", "func (s GetNodeOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s CreateNodeOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (n *Node) String() string {\n\tvar b strings.Builder\n\tNewPrinting(WithWriter(&b), WithColSep(\" \")).RunNode(n)\n\treturn b.String()\n}", "func (t AXNodeID) String() string {\n\treturn string(t)\n}", "func (node NodeInfo) String() string {\n\treturn \"NodeInfo: {\\n\\tnodeID:\" + strconv.Itoa(node.NodeID) + \", \\n\\tnodeIPAddr:\" + node.NodeIPAddr + \", \\n\\tport:\" + node.Port + \" \\n}\"\n}", "func (me TAttlistGeneralNoteOwner) String() string { return xsdt.Token(me).String() }", "func (node *ListNode) String() string {\n\tif node == nil {\n\t\treturn \"<nil>\"\n\t}\n\n\tvar vals []string\n\tfor node != nil {\n\t\tvals = append(vals, strconv.Itoa(node.Val))\n\t\tnode = node.Next\n\t}\n\treturn strings.Join(vals, \"->\")\n}", "func (n *Node) String() string {\r\n\treturn fmt.Sprintf(\"%v\", n.val)\r\n}", "func (r NodeRole) ToString() string {\n\tswitch r {\n\tcase NodeRoleLeader:\n\t\treturn \"leader\"\n\tcase NodeRoleCandidate:\n\t\treturn \"candidate\"\n\tcase NodeRoleFollower:\n\t\treturn \"follower\"\n\tdefault:\n\t\treturn \"UNKNOWN\"\n\t}\n}", "func (n *Node) String() string {\n\tbuf := bytes.NewBuffer(nil)\n\tcn := n\n\n\tfor {\n\t\tif cn.next == nil {\n\t\t\tbreak\n\t\t}\n\t\tif buf.Len() != 0 {\n\t\t\tfmt.Fprint(buf, \",\")\n\t\t} else {\n\t\t\tfmt.Fprint(buf, \"[\")\n\n\t\t}\n\t\tfmt.Fprintf(buf, \"%v\", cn.next.data)\n\t\tcn = cn.next\n\t}\n\tfmt.Fprint(buf, \"]\")\n\n\treturn buf.String()\n}", "func stringify(treeNode *TreeNode, level int) {\n\tif treeNode != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"***> \"\n\t\tlevel++\n\t\tstringify(treeNode.leftNode, level)\n\t\tfmt.Printf(format+\"%d\\n\", treeNode.key)\n\t\tstringify(treeNode.rightNode, level)\n\t}\n}", "func ElemToString(elm *pb.PathElem) string {\n\tb := &strings.Builder{}\n\twriteElem(b, elm)\n\treturn b.String()\n}", "func (s OwnerIdentifier) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (v ResourceNode) String() string {\n\tpubKey, err := stratos.Bech32ifyPubKey(stratos.Bech32PubKeyTypeAccPub, v.PubKey)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn fmt.Sprintf(`ResourceNode:{\n\t\tNetwork Id:\t %s\n \t\tPubkey:\t\t\t\t%s\n \t\tSuspend:\t\t\t%v\n \t\tStatus:\t\t\t\t%s\n \t\tTokens:\t\t\t\t%s\n\t\tOwner Address: \t\t%s\n \t\tDescription:\t\t%s\n \t\tCreationTime:\t\t%s\n\t}`, v.NetworkID, pubKey, v.Suspend, v.Status, v.Tokens, v.OwnerAddress, v.Description, v.CreationTime)\n}", "func stringify(n *Node, level int) {\n\tif n != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"---[ \"\n\t\tlevel++\n\t\tstringify(n.left, level)\n\t\tfmt.Printf(format+\"%d\\n\", n.key)\n\t\tstringify(n.right, level)\n\t}\n}", "func (url NodeURL) String() string {\n\tif url.ID.IsZero() {\n\t\treturn url.Address\n\t}\n\treturn url.ID.String() + \"@\" + url.Address\n}", "func (a LinkOwnershipClass) String() string {\n\tswitch a {\n\tcase OwnerToOwner:\n\t\treturn \"owner-to-owner\"\n\tcase OwnerToAny:\n\t\treturn \"owner-to-any\"\n\tcase AnyToOwner:\n\t\treturn \"any-to-owner\"\n\tdefault:\n\t\treturn \"<unknown link ownership class>\"\n\t}\n}", "func (s NodeSummary) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (n *Node) String() string {\n\tvar str string\n\tvalue := n.Value\n\tfor {\n\t\tstr += fmt.Sprintf(\"%d \", n.Value)\n\t\tn = n.Next\n\t\tif n.Value == value {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn str\n}", "func (id NodeConfigurationId) String() string {\n\tcomponents := []string{\n\t\tfmt.Sprintf(\"Subscription: %q\", id.SubscriptionId),\n\t\tfmt.Sprintf(\"Resource Group Name: %q\", id.ResourceGroupName),\n\t\tfmt.Sprintf(\"Automation Account Name: %q\", id.AutomationAccountName),\n\t\tfmt.Sprintf(\"Node Configuration Name: %q\", id.NodeConfigurationName),\n\t}\n\treturn fmt.Sprintf(\"Node Configuration (%s)\", strings.Join(components, \"\\n\"))\n}", "func (ref ostreeReference) StringWithinTransport() string {\n\treturn fmt.Sprintf(\"%s@%s\", ref.image, ref.repo)\n}", "func (n *Node) String() string {\n\tu := url.URL{Scheme: \"enode\"}\n\tif n.Incomplete() {\n\t\tu.Host = fmt.Sprintf(\"%x\", n.ID[:])\n\t} else {\n\t\taddr := net.TCPAddr{IP: n.IP, Port: int(n.TCP)}\n\t\tu.User = url.User(fmt.Sprintf(\"%x\", n.ID[:]))\n\t\tu.Host = addr.String()\n\t\tif n.UDP != n.TCP {\n\t\t\tu.RawQuery = \"discport=\" + strconv.Itoa(int(n.UDP))\n\t\t}\n\t}\n\treturn u.String()\n}", "func (nt NodeValue) String() string {\n\treturn string(nt)\n}", "func (n ExpandedNodeID) String() string {\n\tb := new(strings.Builder)\n\tif n.ServerIndex > 0 {\n\t\tfmt.Fprintf(b, \"svr=%d;\", n.ServerIndex)\n\t}\n\tif len(n.NamespaceURI) > 0 {\n\t\tfmt.Fprintf(b, \"nsu=%s;\", n.NamespaceURI)\n\t}\n\tswitch n2 := n.NodeID.(type) {\n\tcase NodeIDNumeric:\n\t\tb.WriteString(n2.String())\n\tcase NodeIDString:\n\t\tb.WriteString(n2.String())\n\tcase NodeIDGUID:\n\t\tb.WriteString(n2.String())\n\tcase NodeIDOpaque:\n\t\tb.WriteString(n2.String())\n\tdefault:\n\t\tb.WriteString(\"i=0\")\n\t}\n\treturn b.String()\n}", "func formatNode(fset *token.FileSet, node interface{}) string {\n\tvar buf bytes.Buffer\n\terr := format.Node(&buf, fset, node)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"never here: %v\", err))\n\t}\n\treturn buf.String()\n}", "func (n dapNode) String() string {\n\treturn n.name\n}", "func (o RepositoryAssociationRepositoryBitbucketOutput) Owner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v RepositoryAssociationRepositoryBitbucket) string { return v.Owner }).(pulumi.StringOutput)\n}", "func (binaryTree *BinaryTree) ToString() string {\n\tvar buffer bytes.Buffer\n\n\tif !binaryTree.Empty() {\n\t\tvar queue []*BinaryTreeNode = []*BinaryTreeNode{binaryTree.root}\n\n\t\tfor len(queue) > 0 {\n\t\t\tvar node *BinaryTreeNode = queue[0]\n\t\t\tbuffer.WriteString(strconv.Itoa(node.value) + \" \")\n\n\t\t\tif node.left != nil {\n\t\t\t\tqueue = append(queue, node.left)\n\t\t\t}\n\n\t\t\tif node.right != nil {\n\t\t\t\tqueue = append(queue, node.right)\n\t\t\t}\n\n\t\t\tqueue = queue[1:]\n\t\t}\n\t}\n\n\treturn buffer.String()\n}", "func (o SubnetOutput) OwnerId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Subnet) pulumi.StringOutput { return v.OwnerId }).(pulumi.StringOutput)\n}", "func (n *Node) Owner() *userpb.UserId {\n\treturn n.SpaceRoot.owner\n}", "func (jdn jobDirectoryNode) String() string {\n\treturn fmt.Sprintf(\"JobDirectoryNode<PARENT=[%s] NAME=[%s]>\", jdn.jobNode.parentNodePath, jdn.info.Name())\n}", "func (s IpOwner) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (j JSONMetaContainerNode) String() string {\n\treturn fmt.Sprintf(\"(%q -> %s)\", j.Path(), j.ContentString(0))\n}", "func (t *Tree) formatStr() []byte {\n\tvar s []byte\n\ts = append(s, t.format...)\n\tif t.parent != nil {\n\t\tt := t.parent.formatStr()\n\t\tt = append(t, s...)\n\t\ts = t\n\t}\n\treturn s\n}", "func (head *Node) String() string {\n\tvar sb strings.Builder\n\tseen := make(map[*Node]int)\n\tcurrent := head\n\tdepth := 0\n\tfor current.Next != nil {\n\t\tsb.WriteString(fmt.Sprintf(\"(%v) -> \", current.Data))\n\t\tif d, ok := seen[current.Next]; ok {\n\t\t\tif d == 0 {\n\t\t\t\tsb.WriteString(\"HEAD\")\n\t\t\t\treturn sb.String()\n\t\t\t}\n\t\t\tsb.WriteString(fmt.Sprintf(\"HEAD~%d\", d))\n\t\t\treturn sb.String()\n\t\t}\n\t\tseen[current] = depth\n\t\tcurrent = current.Next\n\t\tdepth++\n\t}\n\tsb.WriteString(fmt.Sprintf(\"(%v)\", current.Data))\n\treturn sb.String()\n}", "func (r *RouteTable) OwnerId() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"ownerId\"])\n}", "func stringify(n *BinarySearchNode, level int, builder *strings.Builder) {\n\tif n != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"---[ \"\n\t\tlevel++\n\t\tstringify(n.left, level, builder)\n\t\tbuilder.WriteString(fmt.Sprintf(format+\"%d\\n\", n.value))\n\t\tstringify(n.right, level, builder)\n\t}\n}", "func (tree *BinarySearchTree) String() {\n\ttree.lock.Lock()\n\tdefer tree.lock.Unlock()\n\tfmt.Println(\"************************************************\")\n\tstringify(tree.rootNode, 0)\n\tfmt.Println(\"************************************************\")\n}", "func (this *TreeNode) String() string {\n\treturn this.Name.String() + this.Colon.String() +\n\t\tthis.Pattern.String()\n}", "func (conn *Conn) ToString() string {\n\tnode := conn.node()\n\tif strings.Count(node.host, \":\") > 0 {\n\t\treturn fmt.Sprintf(\"[%s]:%d\", node.host, node.port)\n\t}\n\treturn fmt.Sprintf(\"%s:%d\", node.host, node.port)\n}", "func (o BucketReplicationConfigurationRuleDestinationAccessControlTranslationOutput) Owner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketReplicationConfigurationRuleDestinationAccessControlTranslation) string { return v.Owner }).(pulumi.StringOutput)\n}", "func (t *Transaction) Owner() string {\n\treturn utils.EncodeToBase64(t.owner.Bytes())\n}", "func astToString(tree []interface{}) string {\n\tif len(tree) == 1 {\n\t\treturn toString(tree[0])\n\t}\n\tif len(tree) == 2 {\n\t\top, val := toString(tree[0]), toInterfaces(tree[1])\n\t\ts := astToString(val)\n\t\treturn \"( \" + op + \" \" + s + \" )\"\n\t}\n\top, l, r := toString(tree[0]), toInterfaces(tree[1]), toInterfaces(tree[2])\n\ts1 := astToString(l)\n\ts2 := astToString(r)\n\treturn \"( \" + s1 + \" \" + op + \" \" + s2 + \" )\"\n}", "func (o BucketReplicationConfigRuleDestinationAccessControlTranslationOutput) Owner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketReplicationConfigRuleDestinationAccessControlTranslation) string { return v.Owner }).(pulumi.StringOutput)\n}", "func FormatNode(fset *token.FileSet, n ast.Node) string {\n\tvar buf strings.Builder\n\tif err := printer.Fprint(&buf, fset, n); err != nil {\n\t\treturn \"\"\n\t}\n\treturn buf.String()\n}", "func (r *RawNode) String() string {\n\treturn r.raw\n}", "func (n *Node) String() string {\n\tvar process func(n *Node) string\n\tprocess = func(n *Node) string {\n\t\tif n == nil {\n\t\t\treturn \"\"\n\t\t}\n\t\tswitch n.Operation {\n\t\tcase OperationNoop:\n\t\t\treturn \"(\" + process(n.Left) + \"???\" + process(n.Right) + \")\"\n\t\tcase OperationAdd:\n\t\t\treturn \"(\" + process(n.Left) + \" + \" + process(n.Right) + \")\"\n\t\tcase OperationSubtract:\n\t\t\treturn \"(\" + process(n.Left) + \" - \" + process(n.Right) + \")\"\n\t\tcase OperationMultiply:\n\t\t\treturn \"(\" + process(n.Left) + \" * \" + process(n.Right) + \")\"\n\t\tcase OperationDivide:\n\t\t\treturn \"(\" + process(n.Left) + \" / \" + process(n.Right) + \")\"\n\t\tcase OperationModulus:\n\t\t\treturn \"(\" + process(n.Left) + \" % \" + process(n.Right) + \")\"\n\t\tcase OperationExponentiation:\n\t\t\treturn \"(\" + process(n.Left) + \"^\" + process(n.Right) + \")\"\n\t\tcase OperationNegate:\n\t\t\treturn \"-(\" + process(n.Left) + \")\"\n\t\tcase OperationVariable:\n\t\t\treturn n.Value\n\t\tcase OperationImaginary:\n\t\t\treturn n.Value + \"i\"\n\t\tcase OperationNumber:\n\t\t\treturn n.Value\n\t\tcase OperationNotation:\n\t\t\tif n.Left.Operation == OperationImaginary {\n\t\t\t\treturn n.Left.Value + \"e\" + process(n.Right) + \"i\"\n\t\t\t}\n\t\t\treturn process(n.Left) + \"e\" + process(n.Right)\n\t\tcase OperationNaturalExponentiation:\n\t\t\treturn \"(e^\" + process(n.Left) + \")\"\n\t\tcase OperationNatural:\n\t\t\treturn \"e\"\n\t\tcase OperationPI:\n\t\t\treturn \"pi\"\n\t\tcase OperationNaturalLogarithm:\n\t\t\treturn \"log(\" + process(n.Left) + \")\"\n\t\tcase OperationSquareRoot:\n\t\t\treturn \"sqrt(\" + process(n.Left) + \")\"\n\t\tcase OperationCosine:\n\t\t\treturn \"cos(\" + process(n.Left) + \")\"\n\t\tcase OperationSine:\n\t\t\treturn \"sin(\" + process(n.Left) + \")\"\n\t\tcase OperationTangent:\n\t\t\treturn \"tan(\" + process(n.Left) + \")\"\n\t\t}\n\t\treturn \"\"\n\t}\n\treturn process(n)\n}", "func ToString(ctx context.Context) (ret string) {\n\tfather, ok := ctx.Value(key).(*constNode)\n\tif !ok {\n\t\treturn \"\"\n\t}\n\tDone(ctx)\n\tconst fmtStr = \"%s%s (%dms %d%%)\\n\"\n\tvar levelPrint func(level int, node *constNode, prefix string)\n\tlevelPrint = func(level int, node *constNode, prefix string) {\n\t\tvar (\n\t\t\tlastTabs string\n\t\t\tnoLastTabs string\n\t\t)\n\t\tnoLastTabs = prefix + \"├─\"\n\t\tlastTabs = prefix + \"└─\"\n\t\tfor i, child := range node.child {\n\t\t\ttabs := noLastTabs\n\t\t\tif i == len(node.child)-1 {\n\t\t\t\ttabs = lastTabs\n\t\t\t}\n\t\t\tchildCostMs := child.cost().Milliseconds()\n\t\t\tfatherCostMs := node.cost().Milliseconds()\n\t\t\tradio := int64(0)\n\t\t\tif fatherCostMs > 0 {\n\t\t\t\tradio = childCostMs * 100 / fatherCostMs\n\t\t\t}\n\t\t\tret += fmt.Sprintf(fmtStr, tabs, child.title, childCostMs, radio)\n\t\t\tif len(child.child) > 0 {\n\t\t\t\tif i == len(node.child)-1 {\n\t\t\t\t\tlevelPrint(level+1, child, prefix+\" \")\n\t\t\t\t} else {\n\t\t\t\t\tlevelPrint(level+1, child, prefix+\"│ \")\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tret += fmt.Sprintf(fmtStr, \"\", father.title, father.cost().Milliseconds(), 100)\n\tlevelPrint(0, father, \"\")\n\treturn\n}", "func (r *Document) Owner() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"owner\"])\n}", "func (n *Node) String() string {\n\treturn stringify.Struct(\"GetElement\",\n\t\tstringify.StructField(\"key\", n.key),\n\t\tstringify.StructField(\"value\", n.value),\n\t\tstringify.StructField(\"left\", n.left),\n\t\tstringify.StructField(\"right\", n.right),\n\t)\n}", "func (o GetRulesRuleOutput) SourceOwner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetRulesRule) string { return v.SourceOwner }).(pulumi.StringOutput)\n}", "func printNode(n *ind.IndicatorNode, ident string) string {\n\tvar l strings.Builder\n\tif n.Operator != \"\" {\n\t\tl.WriteString(ident + \"Operator: \" + n.Operator + \"\\n\")\n\t}\n\tif n.Pattern != nil {\n\t\tl.WriteString(ident + \"Pattern:\\n\")\n\t\tl.WriteString(ident + \" type: \" + n.Pattern.Type + \"\\n\")\n\t\tl.WriteString(ident + \" value: \" + n.Pattern.Value + \"\\n\")\n\t}\n\tif len(n.Children) > 0 {\n\t\tl.WriteString(ident + \"Children: [\\n\")\n\t\tfor i, child := range n.Children {\n\t\t\tl.WriteString(ident + strconv.Itoa(i) + \"\\n\" + printNode(child, ident+\" \"))\n\t\t}\n\t\tl.WriteString(ident + \"]\\n\")\n\t}\n\n\treturn l.String()\n}", "func (neuron *Neuron) String() string {\n\tnodeType := \" Node\"\n\tif neuron.IsInput() {\n\t\tnodeType = \" Input node\"\n\t} else if neuron.IsOutput() {\n\t\tnodeType = \"Output node\"\n\t}\n\treturn fmt.Sprintf(\"%s ID %d has these input connections: %v\", nodeType, neuron.neuronIndex, neuron.InputNodes)\n}", "func (l FuncNode) String() string {\n\treturn fmt.Sprintf(\"[FuncNode %s %s ]\", l.typ, l.label)\n}", "func (s GetNodeInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (rr RepositoryReference) String() string {\n\treturn fmt.Sprintf(\"%s/%s\", rr.OwnerName, rr.Name)\n}" ]
[ "0.7972814", "0.77023315", "0.7609072", "0.6673394", "0.6602986", "0.65908784", "0.64567", "0.6454524", "0.6390594", "0.63837796", "0.6312542", "0.6307958", "0.6304605", "0.6244567", "0.6232909", "0.6209375", "0.61758256", "0.6165706", "0.61587185", "0.6129427", "0.612889", "0.6069242", "0.6062437", "0.6042759", "0.6027129", "0.6007346", "0.5997849", "0.5970087", "0.59699047", "0.59655", "0.5955803", "0.59478956", "0.59398264", "0.5934361", "0.5928963", "0.5922466", "0.59139585", "0.59130514", "0.58887696", "0.5874164", "0.58720803", "0.5861633", "0.5845774", "0.5836033", "0.5832042", "0.58206093", "0.58134687", "0.5792452", "0.5784707", "0.57694083", "0.5760431", "0.5755976", "0.57488275", "0.57385486", "0.573701", "0.5732068", "0.5731266", "0.57228655", "0.5705399", "0.5704774", "0.5694856", "0.5658866", "0.5645017", "0.56154925", "0.5612733", "0.56016517", "0.5600037", "0.5598942", "0.55969125", "0.5593106", "0.5592897", "0.55921817", "0.557543", "0.55694515", "0.556173", "0.5555675", "0.5535931", "0.5532957", "0.5526893", "0.5517229", "0.5516016", "0.55016553", "0.5476155", "0.54641104", "0.5461537", "0.54558915", "0.5454945", "0.54520667", "0.5449767", "0.54429", "0.54186577", "0.5417826", "0.5416622", "0.5408771", "0.5398434", "0.53963447", "0.5375486", "0.5374176", "0.53735906", "0.5373433" ]
0.74435776
3
ownerToString generate a string to identify the owner matches nodeToString format
func ownerToString(node *yaml.RNode) (string, error) { meta, err := node.GetMeta() if err != nil { return "", err } namespace := meta.Namespace owners, err := node.Pipe(yaml.Lookup("metadata", "ownerReferences")) if err != nil { return "", err } if owners == nil { return "", nil } elements, err := owners.Elements() if err != nil { return "", err } if len(elements) == 0 { return "", err } owner := elements[0] var kind, name string if value := owner.Field("kind"); !value.IsNilOrEmpty() { kind = value.Value.YNode().Value } if value := owner.Field("name"); !value.IsNilOrEmpty() { name = value.Value.YNode().Value } return fmt.Sprintf("%s %s/%s", kind, namespace, name), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o RepositoryAssociationRepositoryBitbucketOutput) Owner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v RepositoryAssociationRepositoryBitbucket) string { return v.Owner }).(pulumi.StringOutput)\n}", "func (o BucketReplicationConfigurationRuleDestinationAccessControlTranslationOutput) Owner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketReplicationConfigurationRuleDestinationAccessControlTranslation) string { return v.Owner }).(pulumi.StringOutput)\n}", "func (o BucketReplicationConfigRuleDestinationAccessControlTranslationOutput) Owner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketReplicationConfigRuleDestinationAccessControlTranslation) string { return v.Owner }).(pulumi.StringOutput)\n}", "func (o *V0037Node) GetOwner() string {\n\tif o == nil || o.Owner == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Owner\n}", "func (t *Transaction) Owner() string {\n\treturn utils.EncodeToBase64(t.owner.Bytes())\n}", "func (r *Document) Owner() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"owner\"])\n}", "func (o SubnetOutput) OwnerId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Subnet) pulumi.StringOutput { return v.OwnerId }).(pulumi.StringOutput)\n}", "func (o TopicPolicyOutput) Owner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *TopicPolicy) pulumi.StringOutput { return v.Owner }).(pulumi.StringOutput)\n}", "func (o TriggerGithubOutput) Owner() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TriggerGithub) *string { return v.Owner }).(pulumi.StringPtrOutput)\n}", "func (o RepositoryAssociationRepositoryBitbucketPtrOutput) Owner() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *RepositoryAssociationRepositoryBitbucket) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Owner\n\t}).(pulumi.StringPtrOutput)\n}", "func (o RepositoryAssociationRepositoryGithubEnterpriseServerOutput) Owner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v RepositoryAssociationRepositoryGithubEnterpriseServer) string { return v.Owner }).(pulumi.StringOutput)\n}", "func (o TriggerGithubPtrOutput) Owner() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TriggerGithub) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Owner\n\t}).(pulumi.StringPtrOutput)\n}", "func (o LookupManagedPrefixListResultOutput) OwnerId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupManagedPrefixListResult) string { return v.OwnerId }).(pulumi.StringOutput)\n}", "func (s *Stream) Owner() string {\n\tif s.OwnerRaw == nil {\n\t\treturn \"\"\n\t}\n\n\treturn *s.OwnerRaw\n}", "func (n *Node) Owner() *userpb.UserId {\n\treturn n.SpaceRoot.owner\n}", "func (o GetAggregateConfigRulesRuleOutput) SourceOwner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetAggregateConfigRulesRule) string { return v.SourceOwner }).(pulumi.StringOutput)\n}", "func (o RepositoryAssociationRepositoryGithubEnterpriseServerPtrOutput) Owner() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *RepositoryAssociationRepositoryGithubEnterpriseServer) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Owner\n\t}).(pulumi.StringPtrOutput)\n}", "func (s OwnerIdentifier) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (o BucketReplicationConfigRuleDestinationAccessControlTranslationPtrOutput) Owner() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketReplicationConfigRuleDestinationAccessControlTranslation) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Owner\n\t}).(pulumi.StringPtrOutput)\n}", "func (me TAttlistGeneralNoteOwner) String() string { return xsdt.Token(me).String() }", "func (o GetRulesRuleOutput) SourceOwner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetRulesRule) string { return v.SourceOwner }).(pulumi.StringOutput)\n}", "func (o BucketReplicationConfigurationRuleDestinationAccessControlTranslationPtrOutput) Owner() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketReplicationConfigurationRuleDestinationAccessControlTranslation) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Owner\n\t}).(pulumi.StringPtrOutput)\n}", "func TopLevelOwnerKey(ctx context.Context, obj metav1.Object, kubeContext string, kind string) string {\n\tfor {\n\t\tor := obj.GetOwnerReferences()\n\t\tif or == nil {\n\t\t\treturn fmt.Sprintf(\"%s-%s\", kind, obj.GetName())\n\t\t}\n\t\tvar err error\n\t\tkind = or[0].Kind\n\t\tobj, err = ownerMetaObject(ctx, obj.GetNamespace(), kubeContext, or[0])\n\t\tif err != nil {\n\t\t\tlog.Entry(ctx).Warnf(\"unable to get owner from reference: %v\", or[0])\n\t\t\treturn \"\"\n\t\t}\n\t}\n}", "func (r *RouteTable) OwnerId() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"ownerId\"])\n}", "func (o SnapshotOutput) OwnerAccount() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Snapshot) pulumi.StringOutput { return v.OwnerAccount }).(pulumi.StringOutput)\n}", "func (o LookupMulticastDomainResultOutput) OwnerId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupMulticastDomainResult) string { return v.OwnerId }).(pulumi.StringOutput)\n}", "func (a Account) ShowOwner() string {\n\treturn a.owner\n}", "func (_NodeSpace *NodeSpaceCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _NodeSpace.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (s IpOwner) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (o *StorageNetAppCloudTargetAllOf) GetOwner() string {\n\tif o == nil || o.Owner == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Owner\n}", "func (_Node *NodeCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _Node.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (l *loadBalancer) Owner() string {\n\tif l.shared {\n\t\treturn \"\"\n\t}\n\n\tfor _, ingresses := range l.ingresses {\n\t\tfor _, ingress := range ingresses {\n\t\t\treturn fmt.Sprintf(\"%s/%s\", ingress.Namespace, ingress.Name)\n\t\t}\n\t}\n\n\treturn \"\"\n}", "func (o VirtualGatewayOutput) ResourceOwner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *VirtualGateway) pulumi.StringOutput { return v.ResourceOwner }).(pulumi.StringOutput)\n}", "func (o LookupStreamingImageResultOutput) Owner() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v LookupStreamingImageResult) *string { return v.Owner }).(pulumi.StringPtrOutput)\n}", "func (o VirtualGatewayOutput) MeshOwner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *VirtualGateway) pulumi.StringOutput { return v.MeshOwner }).(pulumi.StringOutput)\n}", "func (n *Node) Owner() (*userpb.UserId, error) {\n\tif n.owner != nil {\n\t\treturn n.owner, nil\n\t}\n\n\towner := &userpb.UserId{}\n\n\t// FIXME ... do we return the owner of the reference or the owner of the target?\n\t// we don't really know the owner of the target ... and as the reference may point anywhere we cannot really find out\n\t// but what are the permissions? all? none? the gateway has to fill in?\n\t// TODO what if this is a reference?\n\tnodePath := n.InternalPath()\n\t// lookup parent id in extended attributes\n\tvar attrBytes []byte\n\tvar err error\n\t// lookup ID in extended attributes\n\tattrBytes, err = xattr.Get(nodePath, xattrs.OwnerIDAttr)\n\tswitch {\n\tcase err == nil:\n\t\towner.OpaqueId = string(attrBytes)\n\tcase isAttrUnset(err), isNotFound(err):\n\t\tfallthrough\n\tdefault:\n\t\treturn nil, err\n\t}\n\n\t// lookup IDP in extended attributes\n\tattrBytes, err = xattr.Get(nodePath, xattrs.OwnerIDPAttr)\n\tswitch {\n\tcase err == nil:\n\t\towner.Idp = string(attrBytes)\n\tcase isAttrUnset(err), isNotFound(err):\n\t\tfallthrough\n\tdefault:\n\t\treturn nil, err\n\t}\n\n\t// lookup type in extended attributes\n\tattrBytes, err = xattr.Get(nodePath, xattrs.OwnerTypeAttr)\n\tswitch {\n\tcase err == nil:\n\t\towner.Type = utils.UserTypeMap(string(attrBytes))\n\tcase isAttrUnset(err), isNotFound(err):\n\t\tfallthrough\n\tdefault:\n\t\t// TODO the user type defaults to invalid, which is the case\n\t\terr = nil\n\t}\n\n\tn.owner = owner\n\treturn n.owner, err\n}", "func (a LinkOwnershipClass) String() string {\n\tswitch a {\n\tcase OwnerToOwner:\n\t\treturn \"owner-to-owner\"\n\tcase OwnerToAny:\n\t\treturn \"owner-to-any\"\n\tcase AnyToOwner:\n\t\treturn \"any-to-owner\"\n\tdefault:\n\t\treturn \"<unknown link ownership class>\"\n\t}\n}", "func nodeToString(fs *token.FileSet, n ast.Node) string {\n\tb := bytes.NewBuffer([]byte{})\n\tprinter.Fprint(b, fs, n)\n\treturn b.String()\n}", "func (img Image) GetOwner() string {\n\treturn \"\"\n}", "func (o LookupVirtualRouterResultOutput) ResourceOwner() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupVirtualRouterResult) string { return v.ResourceOwner }).(pulumi.StringOutput)\n}", "func NodeToString(node core.Node) string {\n\tvar b strings.Builder\n\tfmt.Println()\n\tnodeRecurse(node, 0, &b)\n\treturn b.String()\n}", "func (me TAttlistMedlineCitationOwner) String() string { return xsdt.Token(me).String() }", "func (b *FollowUpBuilder) Owner(value string) *FollowUpBuilder {\n\tb.owner = value\n\tb.bitmap_ |= 512\n\treturn b\n}", "func (theAccount Account) Owner() string {\n\treturn theAccount.owner\n}", "func (s *SOC) OwnerAddress() []byte {\n\treturn s.owner\n}", "func (_BaseContentSpace *BaseContentSpaceCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _BaseContentSpace.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (o *V0037Node) SetOwner(v string) {\n\to.Owner = &v\n}", "func (_Onesplitaudit *OnesplitauditCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _Onesplitaudit.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (_BaseGroupFactory *BaseGroupFactoryCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _BaseGroupFactory.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func nodeToString(node *yaml.RNode) (string, error) {\n\tmeta, err := node.GetMeta()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn fmt.Sprintf(\"%s %s/%s\", meta.Kind, meta.Namespace, meta.Name), nil\n}", "func (_LvStreamRightsHolder *LvStreamRightsHolderCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _LvStreamRightsHolder.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func getOwnerSidString(sd []byte) (string, error) {\n\t// Make sure we have enough bytes to safely read the required fields.\n\tif len(sd) < int(unsafe.Sizeof(SECURITY_DESCRIPTOR_RELATIVE{})) {\n\t\treturn \"\", fmt.Errorf(\"Short Security Descriptor: %d bytes!\", len(sd))\n\t}\n\n\t// Only valid revision is 1, verify that.\n\trevision := getRevision(sd)\n\tif revision != SID_REVISION {\n\t\treturn \"\", fmt.Errorf(\"Invalid SID revision (%d), expected %d!\", revision, SID_REVISION)\n\t}\n\n\t// SECURITY_DESCRIPTOR_RELATIVE.OffsetOwner.\n\toffsetOwner := binary.LittleEndian.Uint32(sd[4:8])\n\tif offsetOwner >= uint32(len(sd)) {\n\t\treturn \"\", fmt.Errorf(\"offsetOwner (%d) points outside Security Descriptor of size %d bytes!\",\n\t\t\toffsetOwner, len(sd))\n\t}\n\n\tsidStr, err := sidToString(sd[offsetOwner:])\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn \"O:\" + sidStr, nil\n}", "func (_BaseContentFactory *BaseContentFactoryCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _BaseContentFactory.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (network *VirtualNetwork) Owner() *genruntime.ResourceReference {\n\tgroup, kind := genruntime.LookupOwnerGroupKind(network.Spec)\n\treturn &genruntime.ResourceReference{\n\t\tGroup: group,\n\t\tKind: kind,\n\t\tName: network.Spec.Owner.Name,\n\t}\n}", "func (_LvRecordableStream *LvRecordableStreamCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _LvRecordableStream.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (as *AclStatus) Owner() string {\n\treturn as.aclStatus.GetOwner()\n}", "func (o *V0037Node) GetOwnerOk() (*string, bool) {\n\tif o == nil || o.Owner == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Owner, true\n}", "func (_Trebuchet *TrebuchetCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _Trebuchet.contract.Call(opts, out, \"owner\")\n\treturn *ret0, err\n}", "func (_Smartchef *SmartchefCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _Smartchef.contract.Call(opts, out, \"owner\")\n\treturn *ret0, err\n}", "func (_ChpRegistry *ChpRegistryCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _ChpRegistry.contract.Call(opts, out, \"owner\")\n\treturn *ret0, err\n}", "func (o LookupGatewayResultOutput) OwnerAccountId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupGatewayResult) string { return v.OwnerAccountId }).(pulumi.StringOutput)\n}", "func (f *HubCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := f.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func OwnerAddr(wk *key.Key) NodeOpt {\n\treturn func(opts *nodeOpts) error {\n\t\topts.ownerKey = wk\n\t\treturn nil\n\t}\n}", "func (_BaseAccessControlGroup *BaseAccessControlGroupCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _BaseAccessControlGroup.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (_FCToken *FCTokenCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _FCToken.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (_BurnableToken *BurnableTokenCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _BurnableToken.contract.Call(opts, out, \"owner\")\n\treturn *ret0, err\n}", "func (_DetailedTestToken *DetailedTestTokenCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _DetailedTestToken.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (peering *VirtualNetworksVirtualNetworkPeering) Owner() *genruntime.ResourceReference {\n\tgroup, kind := genruntime.LookupOwnerGroupKind(peering.Spec)\n\treturn &genruntime.ResourceReference{\n\t\tGroup: group,\n\t\tKind: kind,\n\t\tName: peering.Spec.Owner.Name,\n\t}\n}", "func (topic *Topic) Owner() *genruntime.ResourceReference {\n\tgroup, kind := genruntime.LookupOwnerGroupKind(topic.Spec)\n\treturn &genruntime.ResourceReference{\n\t\tGroup: group,\n\t\tKind: kind,\n\t\tName: topic.Spec.Owner.Name,\n\t}\n}", "func (_OracleMgr *OracleMgrCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _OracleMgr.contract.Call(opts, out, \"owner\")\n\treturn *ret0, err\n}", "func ownerOrCurrentUser(gr *GroupResource, owner *string) error {\n\tif owner == nil || (*owner) == \"\" {\n\t\tcurrent, err := gr.client.Users.Current()\n\t\tif err == nil {\n\t\t\treturn err\n\t\t}\n\t\t*owner = current.User.Username\n\t}\n\treturn nil\n}", "func (_TokenVesting *TokenVestingCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _TokenVesting.contract.Call(opts, out, \"owner\")\n\treturn *ret0, err\n}", "func (o RouteOutput) InstanceOwnerId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Route) pulumi.StringOutput { return v.InstanceOwnerId }).(pulumi.StringOutput)\n}", "func (_Gatekeeper *GatekeeperCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _Gatekeeper.contract.Call(opts, out, \"owner\")\n\treturn *ret0, err\n}", "func (_BaseContentFactoryExt *BaseContentFactoryExtCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _BaseContentFactoryExt.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (_Caller *CallerCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _Caller.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (c *Client)GetOwner(name string)string{\n\treturn strings.Split(name,\"-\")[0]\n}", "func CodeOwner() string {\n\towner, _ := parseCodeRepoRaw()\n\treturn owner\n\n}", "func (_ElvTokenHelper *ElvTokenHelperCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _ElvTokenHelper.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (_BaseFactory *BaseFactoryCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _BaseFactory.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (k Keeper) GetMembersOwner(ctx sdk.Context, id uint64) string {\n\treturn k.GetMembers(ctx, id).Creator\n}", "func (_Container *ContainerCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _Container.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (_BaseContent *BaseContentCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _BaseContent.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (_ZKOnacci *ZKOnacciSession) OwnerOf(tokenId *big.Int) (common.Address, error) {\n\treturn _ZKOnacci.Contract.OwnerOf(&_ZKOnacci.CallOpts, tokenId)\n}", "func (rule *NamespacesTopicsSubscriptionsRule) Owner() *genruntime.ResourceReference {\n\tgroup, kind := genruntime.LookupOwnerGroupKind(rule.Spec)\n\treturn &genruntime.ResourceReference{\n\t\tGroup: group,\n\t\tKind: kind,\n\t\tName: rule.Spec.Owner.Name,\n\t}\n}", "func LabelOwnerCyclone() string {\n\treturn LabelOwner + \"=\" + OwnerCyclone\n}", "func (p Order) GetOwnerID() string {\n\treturn p.OwnerID.Hex()\n}", "func (_MetaObject *MetaObjectCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _MetaObject.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (_HasNoEther *HasNoEtherCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _HasNoEther.contract.Call(opts, out, \"owner\")\n\treturn *ret0, err\n}", "func (_PBridge *PBridgeCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _PBridge.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (_Editable *EditableCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _Editable.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (o OwnershipLevel) String() string {\n\treturn string(o)\n}", "func (_SingleAuto *SingleAutoCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _SingleAuto.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (_Cakevault *CakevaultCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _Cakevault.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (_LvRecording *LvRecordingCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _LvRecording.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (rr RepositoryReference) String() string {\n\treturn fmt.Sprintf(\"%s/%s\", rr.OwnerName, rr.Name)\n}", "func (_Votes *VotesCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _Votes.contract.Call(opts, out, \"owner\")\n\treturn *ret0, err\n}", "func (_ZKOnacci *ZKOnacciCallerSession) OwnerOf(tokenId *big.Int) (common.Address, error) {\n\treturn _ZKOnacci.Contract.OwnerOf(&_ZKOnacci.CallOpts, tokenId)\n}", "func (_BaseAccessWalletFactory *BaseAccessWalletFactoryCaller) Owner(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _BaseAccessWalletFactory.contract.Call(opts, &out, \"owner\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (s S3BucketOwner) String() string {\n\treturn awsutil.Prettify(s)\n}" ]
[ "0.70522296", "0.6943506", "0.69375956", "0.688022", "0.6856853", "0.6838874", "0.68302846", "0.67942965", "0.6789248", "0.67787945", "0.6762392", "0.6750272", "0.6650149", "0.6641561", "0.65653723", "0.65565586", "0.6555649", "0.6503313", "0.6494172", "0.6487006", "0.6482137", "0.6471201", "0.645364", "0.6440409", "0.6432808", "0.642556", "0.6406507", "0.6387009", "0.63457984", "0.63235337", "0.63074094", "0.6306351", "0.6280077", "0.62060845", "0.61961925", "0.61756915", "0.612887", "0.6067165", "0.6057717", "0.6056137", "0.6043704", "0.60313666", "0.6030846", "0.60295045", "0.6012226", "0.5989447", "0.59846216", "0.5968691", "0.59572625", "0.59387803", "0.5932242", "0.5915297", "0.5914082", "0.5913642", "0.5901566", "0.5900776", "0.59001124", "0.5883652", "0.5883495", "0.5882085", "0.588059", "0.5879416", "0.5879324", "0.58781403", "0.58723134", "0.587159", "0.5869158", "0.58592904", "0.5850257", "0.5847005", "0.5845539", "0.58360106", "0.5832202", "0.58314687", "0.5831405", "0.5829872", "0.5823703", "0.58122474", "0.580278", "0.57990545", "0.5797809", "0.57946867", "0.5776687", "0.57766616", "0.5776448", "0.57763547", "0.5774533", "0.5768984", "0.5768075", "0.57646984", "0.57616866", "0.5760717", "0.5756278", "0.57524693", "0.5746969", "0.5740973", "0.5740736", "0.5739714", "0.5732278", "0.5731846" ]
0.8615006
0
index indexes the Resources by their package
func (p TreeWriter) index(nodes []*yaml.RNode) map[string][]*yaml.RNode { // index the ResourceNodes by package indexByPackage := map[string][]*yaml.RNode{} for i := range nodes { meta, err := nodes[i].GetMeta() if err != nil || meta.Kind == "" { // not a resource continue } pkg := filepath.Dir(meta.Annotations[kioutil.PathAnnotation]) indexByPackage[pkg] = append(indexByPackage[pkg], nodes[i]) } return indexByPackage }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (pi *PackageIndexer) Index(pack *Package) string {\n\tpi.mutex.Lock() \n\tdefer pi.mutex.Unlock()\n\t// foreach loop over the package's dependencies \n\tfor _, dep := range pack.deps {\n\t\t// query for each dependency\n\t\tif dep.name != \"\" && pi.Query(dep.name) == FAIL {\n\t\t\t// dependency not installed, cannot be indexed \n\t\t\treturn FAIL\n\t\t}\n\t}\n\t// package dependencies exist: update/add package \n\tpi.packs[pack.name] = pack\n\n\treturn OK \n}", "func (w *exportWriter) writeIndex(index map[types.Object]uint64) {\n\ttype pkgObj struct {\n\t\tobj types.Object\n\t\tname string // qualified name; differs from obj.Name for type params\n\t}\n\t// Build a map from packages to objects from that package.\n\tpkgObjs := map[*types.Package][]pkgObj{}\n\n\t// For the main index, make sure to include every package that\n\t// we reference, even if we're not exporting (or reexporting)\n\t// any symbols from it.\n\tif w.p.localpkg != nil {\n\t\tpkgObjs[w.p.localpkg] = nil\n\t}\n\tfor pkg := range w.p.allPkgs {\n\t\tpkgObjs[pkg] = nil\n\t}\n\n\tfor obj := range index {\n\t\tname := w.p.exportName(obj)\n\t\tpkgObjs[obj.Pkg()] = append(pkgObjs[obj.Pkg()], pkgObj{obj, name})\n\t}\n\n\tvar pkgs []*types.Package\n\tfor pkg, objs := range pkgObjs {\n\t\tpkgs = append(pkgs, pkg)\n\n\t\tsort.Slice(objs, func(i, j int) bool {\n\t\t\treturn objs[i].name < objs[j].name\n\t\t})\n\t}\n\n\tsort.Slice(pkgs, func(i, j int) bool {\n\t\treturn w.exportPath(pkgs[i]) < w.exportPath(pkgs[j])\n\t})\n\n\tw.uint64(uint64(len(pkgs)))\n\tfor _, pkg := range pkgs {\n\t\tw.string(w.exportPath(pkg))\n\t\tw.string(pkg.Name())\n\t\tw.uint64(uint64(0)) // package height is not needed for go/types\n\n\t\tobjs := pkgObjs[pkg]\n\t\tw.uint64(uint64(len(objs)))\n\t\tfor _, obj := range objs {\n\t\t\tw.string(obj.name)\n\t\t\tw.uint64(index[obj.obj])\n\t\t}\n\t}\n}", "func IndexResourcesByName(items []types.Resource) map[string]types.Resource {\n\tindexed := make(map[string]types.Resource, len(items))\n\tfor _, item := range items {\n\t\tindexed[GetResourceName(item)] = item\n\t}\n\treturn indexed\n}", "func indexPackages(pkgs map[string]*ast.Package, prefix string) error {\n\tfor name, pkg := range pkgs {\n\t\tpath := prefix\n pack := name\n\t\t//fmt.Println(\"Inspecting \", path)\n\n\t\tast.Inspect(pkg, func(n ast.Node) bool {\n\n\t\t\tswitch x := n.(type) {\n\t\t\t//Packages\n\t\t\tcase *ast.Package:\n\t\t\t\tif x.Name != \"\" {\n\t\t\t\t\t//update index and docMap if necessary\n\t\t\t\t\tdocTerm := updateIndex(x.Name, pack, path)\n\t\t\t\t\t//update docTerm\n\t\t\t\t\tdocTerm.Packages += 1\n\t\t\t\t}\n\t\t\t\tbreak\n\n\t\t\t//Imports\n\t\t\tcase *ast.ImportSpec:\n\t\t\t\tif x.Path.Value != \"\" {\n\t\t\t\t\t//update index and docMap if necessary\n\t\t\t\t\tdocTerm := updateIndex(strings.Replace(x.Path.Value, \"\\\"\", \"\", -1), pack, path)\n\t\t\t\t\t//update docTerm\n\t\t\t\t\tdocTerm.Imports += 1\n\t\t\t\t}\n\t\t\t\tbreak\n\n\t\t\t//Functions\n\t\t\tcase *ast.FuncDecl:\n\t\t\t\tif x.Name.Name != \"\" {\n\t\t\t\t\t//Name tokenize function\n\t\t\t\t\tfor _, n := range tokenizeCamelCase(x.Name.Name) {\n\t\t\t\t\t\t//update index and docMap if necessary\n\t\t\t\t\t\tdocTerm := updateIndex(n, pack, path)\n\t\t\t\t\t\t//update docTerm\n\t\t\t\t\t\tdocTerm.Functions += 1\n\t\t\t\t\t}\n\n\t\t\t\t\t//Add comments to index\n\t\t\t\t\tif x.Doc != nil && *commentParse {\n\t\t\t\t\t\tcomment := \"\"\n\t\t\t\t\t\tfor _, c := range x.Doc.List {\n\t\t\t\t\t\t\tcomment += c.Text\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\tcomment = strings.Replace(comment, \"//\", \"\", -1)\n\t\t\t\t\t\tcomment = strings.ToLower(comment)\n\n\t\t\t\t\t\twords := strings.Fields(comment)\n\n\t\t\t\t\t\tfor _, word := range words {\n\t\t\t\t\t\t\tdocTerm := updateIndex(word, pack, path)\n\t\t\t\t\t\t\tdocTerm.Functions += 1\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tbreak\n\n\t\t\tcase *ast.TypeSpec:\n\t\t\t\tif x.Name.Name != \"\" {\n\t\t\t\t\t//Name tokenize function\n\t\t\t\t\tfor _, n := range tokenizeCamelCase(x.Name.Name) {\n\t\t\t\t\t\t//update index and docMap if necessary\n\t\t\t\t\t\tdocTerm := updateIndex(n, pack, path)\n\t\t\t\t\t\t//update docTerm\n\t\t\t\t\t\tdocTerm.Types += 1\n\t\t\t\t\t}\n\n\t\t\t\t\t//Add comments to index\n\t\t\t\t\tif x.Doc != nil && *commentParse {\n\t\t\t\t\t\tcomment := \"\"\n\t\t\t\t\t\tfor _, c := range x.Doc.List {\n\t\t\t\t\t\t\tcomment += c.Text\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\tcomment = strings.Replace(comment, \"//\", \"\", -1)\n\t\t\t\t\t\tcomment = strings.ToLower(comment)\n\n\t\t\t\t\t\twords := strings.Fields(comment)\n\n\t\t\t\t\t\tfor _, word := range words {\n\t\t\t\t\t\t\tdocTerm := updateIndex(word, pack, path)\n\t\t\t\t\t\t\tdocTerm.Types += 1\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tbreak\n\t\t\t}\n\t\t\treturn true\n\t\t})\n\t}\n\n\treturn nil\n}", "func IndexResourcesByName(items []envoy_types.ResourceWithTTL) map[string]envoy_types.ResourceWithTTL {\n\tindexed := make(map[string]envoy_types.ResourceWithTTL, len(items))\n\tfor _, item := range items {\n\t\tkey := GetResourceName(item.Resource)\n\t\tindexed[key] = item\n\t}\n\treturn indexed\n}", "func (api *MediaApi) index(c *routing.Context) error {\n\t// --- fetch search data\n\tsearchFields := []string{\"title\", \"type\", \"path\", \"created\", \"modified\"}\n\tsearchData := utils.GetSearchConditions(c, searchFields)\n\t// ---\n\n\t// --- fetch sort data\n\tsortFields := []string{\"title\", \"type\", \"path\", \"created\", \"modified\"}\n\tsortData := utils.GetSortFields(c, sortFields)\n\t// ---\n\n\ttotal, _ := api.dao.Count(searchData)\n\n\tlimit, page := utils.GetPaginationSettings(c, total)\n\n\tutils.SetPaginationHeaders(c, limit, total, page)\n\n\titems := []models.Media{}\n\n\tif total > 0 {\n\t\titems, _ = api.dao.GetList(limit, limit*(page-1), searchData, sortData)\n\n\t\titems = daos.ToAbsMediaPaths(items)\n\t}\n\n\treturn c.Write(items)\n}", "func index(pkg *pkg) error {\n\n\t// ensure dependencies are indexed\n\tfor _, dependency := range pkg.Dependencies {\n\t\tif _, ok := indexRead(dependency); !ok {\n\t\t\treturn missingDependencies\n\t\t}\n\t}\n\n\t// if this index already exists we need to just update dependencies\n\texistingPkg, ok := indexRead(pkg.Name)\n\tif ok {\n\t\treturn updateDependents(existingPkg, pkg)\n\t}\n\n\t// update any dependants of this package\n\tupdateDependents(nil, pkg)\n\n\t// add the new index (possibly replacing the old)\n\tindexWrite(pkg.Name, pkg)\n\n\treturn nil\n}", "func Index(realms map[string]*cloudformation.Realm, name, repo, dir, description string) j.ObjectType {\n\tfields := []j.Type{\n\t\td.Import(),\n\t\td.Pkg(name, path.Join(repo, dir, \"main.libsonnet\"), description),\n\t}\n\n\tfor _, realm := range realms {\n\t\timp := filepath.Join(GenPrefix, realm.N(\"realm\"), MainFile)\n\t\tfields = append(fields, j.Hidden(j.Import(realm.Name, imp)))\n\t}\n\n\tSortFields(fields)\n\n\treturn j.Object(\"\", fields...)\n}", "func (router *Router) getResources(w http.ResponseWriter, r *http.Request) {\n\tclusterNames := r.URL.Query()[\"cluster\"]\n\tnamespaces := r.URL.Query()[\"namespace\"]\n\tname := r.URL.Query().Get(\"name\")\n\tresource := r.URL.Query().Get(\"resource\")\n\tpath := r.URL.Query().Get(\"path\")\n\tparamName := r.URL.Query().Get(\"paramName\")\n\tparam := r.URL.Query().Get(\"param\")\n\n\tlog.WithFields(logrus.Fields{\"clusters\": clusterNames, \"namespaces\": namespaces, \"name\": name, \"resource\": resource, \"path\": path, \"paramName\": paramName, \"param\": param}).Tracef(\"getResources\")\n\n\tvar resources []Resources\n\n\t// Loop through all the given cluster names and get for each provided name the cluster interface. After that we\n\t// check if the resource was provided via the forbidden resources list.\n\tfor _, clusterName := range clusterNames {\n\t\tcluster := router.clusters.GetCluster(clusterName)\n\t\tif cluster == nil {\n\t\t\terrresponse.Render(w, r, nil, http.StatusBadRequest, \"Invalid cluster name\")\n\t\t\treturn\n\t\t}\n\n\t\tif router.isForbidden(resource) {\n\t\t\terrresponse.Render(w, r, nil, http.StatusForbidden, fmt.Sprintf(\"Access for resource %s is forbidding\", resource))\n\t\t\treturn\n\t\t}\n\n\t\t// If the namespaces slice is nil, we retrieve the resource for all namespaces. If a list of namespaces was\n\t\t// provided we loop through all the namespaces and return the resources for these namespaces. All results are\n\t\t// added to the resources slice, which is then returned by the api.\n\t\tif namespaces == nil {\n\t\t\tlist, err := cluster.GetResources(r.Context(), \"\", name, path, resource, paramName, param)\n\t\t\tif err != nil {\n\t\t\t\terrresponse.Render(w, r, err, http.StatusBadRequest, \"Could not get resources\")\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tvar tmpResources map[string]interface{}\n\t\t\terr = json.Unmarshal(list, &tmpResources)\n\t\t\tif err != nil {\n\t\t\t\terrresponse.Render(w, r, err, http.StatusInternalServerError, \"Could not unmarshal resources\")\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tresources = append(resources, Resources{\n\t\t\t\tCluster: clusterName,\n\t\t\t\tNamespace: \"\",\n\t\t\t\tResources: tmpResources,\n\t\t\t})\n\t\t} else {\n\t\t\tfor _, namespace := range namespaces {\n\t\t\t\tlist, err := cluster.GetResources(r.Context(), namespace, name, path, resource, paramName, param)\n\t\t\t\tif err != nil {\n\t\t\t\t\terrresponse.Render(w, r, err, http.StatusBadRequest, \"Could not get resources\")\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tvar tmpResources map[string]interface{}\n\t\t\t\terr = json.Unmarshal(list, &tmpResources)\n\t\t\t\tif err != nil {\n\t\t\t\t\terrresponse.Render(w, r, err, http.StatusInternalServerError, \"Could not unmarshal resources\")\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tresources = append(resources, Resources{\n\t\t\t\t\tCluster: clusterName,\n\t\t\t\t\tNamespace: namespace,\n\t\t\t\t\tResources: tmpResources,\n\t\t\t\t})\n\t\t\t}\n\t\t}\n\t}\n\n\tlog.WithFields(logrus.Fields{\"count\": len(resources)}).Tracef(\"getResources\")\n\trender.JSON(w, r, resources)\n}", "func (s *Server) getIndexes(w http.ResponseWriter, r *http.Request) {\n\tfs, err := s.db.List(\"file\")\n\tif err != nil {\n\t\ts.logf(\"error listing files from mpd for building indexes: %v\", err)\n\t\twriteXML(w, errGeneric)\n\t\treturn\n\t}\n\tfiles := indexFiles(fs)\n\n\twriteXML(w, func(c *container) {\n\t\tc.Indexes = &indexesContainer{\n\t\t\tLastModified: time.Now().Unix(),\n\t\t}\n\n\t\t// Incremented whenever it's time to create a new index for a new\n\t\t// initial letter\n\t\tidx := -1\n\n\t\tvar indexes []index\n\n\t\t// A set of initial characters, used to deduplicate the addition of\n\t\t// nwe indexes\n\t\tseenChars := make(map[rune]struct{}, 0)\n\n\t\tfor _, f := range files {\n\t\t\t// Filter any non-top level items\n\t\t\tif strings.Contains(f.Name, string(os.PathSeparator)) {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Initial rune is used to create an index name\n\t\t\tc, _ := utf8.DecodeRuneInString(f.Name)\n\t\t\tname := string(c)\n\n\t\t\t// If initial rune is a digit, put index under a numeric section\n\t\t\tif unicode.IsDigit(c) {\n\t\t\t\tc = '#'\n\t\t\t\tname = \"#\"\n\t\t\t}\n\n\t\t\t// If a new rune appears, create a new index for it\n\t\t\tif _, ok := seenChars[c]; !ok {\n\t\t\t\tseenChars[c] = struct{}{}\n\t\t\t\tindexes = append(indexes, index{Name: name})\n\t\t\t\tidx++\n\t\t\t}\n\n\t\t\tindexes[idx].Artists = append(indexes[idx].Artists, artist{\n\t\t\t\tName: f.Name,\n\t\t\t\tID: strconv.Itoa(f.ID),\n\t\t\t})\n\t\t}\n\n\t\tc.Indexes.Indexes = indexes\n\t})\n}", "func (d *docsIndexer) indexPackage(p *packages.Package) (docsPackage, error) {\n\tvar (\n\t\tpkgDocsMarkdown string\n\t\tconsts []constVarDocs\n\t\tvars []constVarDocs\n\t\ttypes []typeDocs\n\t\tfuncs []funcDocs\n\t\temitted = make(emittedDocumentationResults, 64)\n\t)\n\tfor _, file := range p.Syntax {\n\t\tfilename := p.Fset.Position(file.Pos()).Filename\n\t\tif !strings.HasPrefix(filename, d.i.projectRoot) {\n\t\t\t// Omit files (such as those generated by `go test`) that aren't in the project root\n\t\t\t// because those are not externally accessible under any circumstance.\n\t\t\tcontinue\n\t\t}\n\t\tfileDocs, err := d.indexFile(p, file, filepath.Base(filename), strings.HasSuffix(filename, \"_test.go\"))\n\t\tif err != nil {\n\t\t\treturn docsPackage{}, errors.Wrap(err, \"file \"+filename)\n\t\t}\n\t\tpkgDocsMarkdown += fileDocs.pkgDocsMarkdown\n\t\tfor _, c := range fileDocs.consts {\n\t\t\tconsts = append(consts, c)\n\t\t\temitted[c.def] = c.ID\n\t\t}\n\t\tfor _, v := range fileDocs.vars {\n\t\t\tvars = append(vars, v)\n\t\t\temitted[v.def] = v.ID\n\t\t}\n\t\tfor _, t := range fileDocs.types {\n\t\t\ttypes = append(types, t)\n\t\t\temitted[t.def] = t.ID\n\t\t}\n\t\tfor _, f := range fileDocs.funcs {\n\t\t\tfuncs = append(funcs, f)\n\t\t\temitted[f.def] = f.ID\n\t\t}\n\t}\n\n\trootPkgPath := d.rootPkgPath()\n\tshortestUniquePkgPath := strings.TrimPrefix(strings.TrimPrefix(pkgPathStdStrip(p.PkgPath), rootPkgPath), \"/\")\n\n\tvisibilityTags := []protocol.Tag{}\n\tif strings.Contains(p.PkgPath, \"/internal/\") || strings.HasSuffix(p.Name, \"_test\") {\n\t\tvisibilityTags = append(visibilityTags, protocol.TagPrivate)\n\t}\n\tif isDeprecated(pkgDocsMarkdown) {\n\t\tvisibilityTags = append(visibilityTags, protocol.TagDeprecated)\n\t}\n\tpkgTags := make([]protocol.Tag, len(visibilityTags))\n\tcopy(pkgTags, visibilityTags)\n\tpkgTags = append(pkgTags, protocol.TagPackage)\n\n\tpkgPathElements := strings.Split(pkgPathStdStrip(p.PkgPath), \"/\")\n\tpackageDocsID := (&documentationResult{\n\t\tDocumentation: protocol.Documentation{\n\t\t\tIdentifier: pkgPathElements[len(pkgPathElements)-1],\n\t\t\tSearchKey: shortestUniquePkgPath,\n\t\t\tNewPage: true,\n\t\t\tTags: pkgTags,\n\t\t},\n\t\tLabel: protocol.NewMarkupContent(\"Package \"+p.Name, protocol.PlainText),\n\t\tDetail: protocol.NewMarkupContent(pkgDocsMarkdown, protocol.Markdown),\n\t}).emit(d.i.emitter)\n\n\tnewSection := func(label, identifier string, children []uint64) uint64 {\n\t\tsectionID := (&documentationResult{\n\t\t\tDocumentation: protocol.Documentation{\n\t\t\t\tIdentifier: identifier,\n\t\t\t\tSearchKey: \"\", // don't index sections of documentation for search\n\t\t\t\tNewPage: false,\n\t\t\t\tTags: visibilityTags,\n\t\t\t},\n\t\t\tLabel: protocol.NewMarkupContent(label, protocol.PlainText),\n\t\t\tDetail: protocol.NewMarkupContent(\"\", protocol.PlainText),\n\t\t}).emit(d.i.emitter)\n\t\t_ = d.i.emitter.EmitDocumentationChildrenEdge(children, sectionID)\n\t\treturn sectionID\n\t}\n\n\tvar sections []uint64\n\t// Emit a \"Constants\" section\n\tsort.Slice(consts, func(i, j int) bool {\n\t\treturn sortName(\"\", \"\", consts[i].name, consts[j].name)\n\t})\n\tif len(consts) > 0 {\n\t\tvar children []uint64\n\t\tfor _, constDocs := range consts {\n\t\t\tchildren = append(children, constDocs.ID)\n\t\t}\n\t\tsections = append(sections, newSection(\"Constants\", \"const\", children))\n\t}\n\n\t// Emit a \"Variables\" section\n\tsort.Slice(vars, func(i, j int) bool {\n\t\treturn sortName(\"\", \"\", vars[i].name, vars[j].name)\n\t})\n\tif len(vars) > 0 {\n\t\tvar children []uint64\n\t\tfor _, varDocs := range vars {\n\t\t\tchildren = append(children, varDocs.ID)\n\t\t}\n\t\tsections = append(sections, newSection(\"Variables\", \"var\", children))\n\t}\n\n\t// Emit methods as children of their receiver types, functions as children of the type they\n\t// produce.\n\tsort.Slice(types, func(i, j int) bool {\n\t\treturn sortName(\"\", \"\", types[i].name, types[j].name)\n\t})\n\tsort.Slice(funcs, func(i, j int) bool {\n\t\treturn sortName(funcs[i].recvTypeName, funcs[j].recvTypeName, funcs[i].name, funcs[j].name)\n\t})\n\temittedMethods := map[uint64]struct{}{}\n\tfor _, typeDocs := range types {\n\t\tvar children []uint64\n\t\tfor _, funcDocs := range funcs {\n\t\t\tif _, emitted := emittedMethods[funcDocs.ID]; emitted {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif funcDocs.recvType == nil {\n\t\t\t\tvar matches int\n\t\t\t\tfor _, resultTypeExpr := range funcDocs.resultTypes {\n\t\t\t\t\tresultType := p.TypesInfo.TypeOf(resultTypeExpr)\n\t\t\t\t\tif dereference(resultType) == dereference(typeDocs.typ) {\n\t\t\t\t\t\tmatches++\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tif matches == 1 {\n\t\t\t\t\t// The function is only a child of the type it produces if there was one match.\n\t\t\t\t\t// If it returned multiple types, better off keeping it separate from both.\n\t\t\t\t\temittedMethods[funcDocs.ID] = struct{}{}\n\t\t\t\t\tchildren = append(children, funcDocs.ID)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tfor _, funcDocs := range funcs {\n\t\t\tif _, emitted := emittedMethods[funcDocs.ID]; emitted {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif funcDocs.recvType != nil {\n\t\t\t\trecvType := p.TypesInfo.TypeOf(funcDocs.recvType)\n\t\t\t\tif dereference(recvType) == dereference(typeDocs.typ) {\n\t\t\t\t\temittedMethods[funcDocs.ID] = struct{}{}\n\t\t\t\t\tchildren = append(children, funcDocs.ID)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif len(children) > 0 {\n\t\t\t_ = d.i.emitter.EmitDocumentationChildrenEdge(children, typeDocs.ID)\n\t\t}\n\t}\n\n\t// Emit a \"Types\" section\n\tif len(types) > 0 {\n\t\tvar children []uint64\n\t\tfor _, typeDocs := range types {\n\t\t\tchildren = append(children, typeDocs.ID)\n\t\t}\n\t\tsections = append(sections, newSection(\"Types\", \"type\", children))\n\t}\n\n\t// Emit a \"Functions\" section\n\tif len(funcs) > 0 {\n\t\tvar children []uint64\n\t\tfor _, funcDocs := range funcs {\n\t\t\tif _, emitted := emittedMethods[funcDocs.ID]; emitted {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tchildren = append(children, funcDocs.ID)\n\t\t}\n\t\tif len(children) > 0 {\n\t\t\tsections = append(sections, newSection(\"Functions\", \"func\", children))\n\t\t}\n\t}\n\n\treturn docsPackage{\n\t\tID: packageDocsID,\n\t\tPath: pkgPathStdStrip(p.PkgPath),\n\t\temitted: emitted,\n\t\tchildren: sections,\n\t}, nil\n}", "func (i *Indexer) indexDocumentation() error {\n\tvar (\n\t\td = &docsIndexer{i: i}\n\t\tmu sync.Mutex\n\t\tdocsPackages []docsPackage\n\t\temitted = make(emittedDocumentationResults, 4096)\n\t\temittedPackagesByPath = make(map[string]uint64, 32)\n\t\terrs error\n\t)\n\ti.visitEachPackage(\"Indexing documentation\", func(p *packages.Package) {\n\t\t// Index the package without the lock, for parallelism.\n\t\tdocsPkg, err := d.indexPackage(p)\n\n\t\t// Acquire the lock; note that multierror.Append could also be racy and hence we hold the\n\t\t// lock even for the error check. In practice, this is not where most of the work is done\n\t\t// (indexPackage is) so this is fine.\n\t\tmu.Lock()\n\t\tdefer mu.Unlock()\n\t\tif err != nil {\n\t\t\terrs = multierror.Append(errs, errors.Wrap(err, \"package \"+p.Name))\n\t\t\treturn\n\t\t}\n\t\temitted.addAll(docsPkg.emitted)\n\t\tdocsPackages = append(docsPackages, docsPkg)\n\t\temittedPackagesByPath[docsPkg.Path] = docsPkg.ID\n\t})\n\n\t// Find the root package path (e.g. \"github.com/sourcegraph/sourcegraph\").\n\trootPkgPath := d.rootPkgPath()\n\n\t// Build an understanding of all pages in the workspace.\n\ttype page struct {\n\t\tid uint64 // the page itself\n\t\tchildren []uint64 // the children pages of this one\n\t}\n\tpagesByPath := map[string]*page{}\n\tfor _, docsPkg := range docsPackages {\n\t\trelPackagePath := d.relPackagePath(docsPkg.Path, rootPkgPath)\n\t\tif _, exists := pagesByPath[relPackagePath]; exists {\n\t\t\tpanic(\"invariant: no duplicate paths\")\n\t\t}\n\t\tpagesByPath[relPackagePath] = &page{id: docsPkg.ID}\n\t}\n\n\t// Emit the root documentationResult which will link all packages in this project to the\n\t// project itself. If the root of the workspace is a Go package, this may already exist\n\t// and would be that Go package's documentation.\n\tif rootPage, ok := pagesByPath[\"\"]; ok {\n\t\t_ = i.emitter.EmitDocumentationResultEdge(rootPage.id, i.projectID)\n\t} else {\n\t\t// Emit a blank index page.\n\t\trootDocumentationID := (&documentationResult{\n\t\t\tDocumentation: protocol.Documentation{\n\t\t\t\tIdentifier: \"\",\n\t\t\t\tSearchKey: \"\",\n\t\t\t\tNewPage: true,\n\t\t\t\tTags: []protocol.Tag{},\n\t\t\t},\n\t\t\tLabel: protocol.NewMarkupContent(\"\", protocol.PlainText),\n\t\t\tDetail: protocol.NewMarkupContent(\"\", protocol.PlainText),\n\t\t}).emit(i.emitter)\n\t\t_ = i.emitter.EmitDocumentationResultEdge(rootDocumentationID, i.projectID)\n\t\tpagesByPath[\"\"] = &page{id: rootDocumentationID}\n\t}\n\n\t// What we have now is pages for each package in the workspace, e.g.:\n\t//\n\t// \t/ (root index page)\n\t// \t/internal/lib/protocol (package page)\n\t// \t/internal/lib/util (package page)\n\t// \t/router/mux (package page)\n\t//\n\t// What we want ot add in is index pages (blank pages) for each parent path so we end up with:\n\t//\n\t// \t/ (root index page)\n\t// \t/internal (index page)\n\t// \t/internal/lib (index page)\n\t// \t/internal/lib/protocol (package page)\n\t// \t/internal/lib/util (package page)\n\t// \t/router (index page)\n\t// \t/router/mux (package page)\n\t//\n\t// Note: the actual paths do not have a leading slash.\n\tsort.Slice(docsPackages, func(i, j int) bool {\n\t\treturn docsPackages[i].Path < docsPackages[j].Path\n\t})\n\tfor _, docsPkg := range docsPackages {\n\t\trelPackagePath := d.relPackagePath(docsPkg.Path, rootPkgPath)\n\t\tpkgPathElements := strings.Split(relPackagePath, \"/\") // [\"internal\", \"lib\", \"protocol\"]\n\n\t\t// Walk over each path: \"internal\", \"internal/lib\", \"internal/lib/protocol\" and emit an\n\t\t// index page for each that does not have it.\n\t\tcurrentPath := \"\"\n\t\tfor _, element := range pkgPathElements {\n\t\t\tcurrentPath = path.Join(currentPath, element)\n\t\t\t_, ok := pagesByPath[currentPath]\n\t\t\tif ok {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tcurrentPathElements := strings.Split(currentPath, \"/\")\n\t\t\tparentPath := path.Join(currentPathElements[:len(currentPathElements)-1]...)\n\n\t\t\t// Emit an index page at this path since one does not exist.\n\t\t\tpageID := (&documentationResult{\n\t\t\t\tDocumentation: protocol.Documentation{\n\t\t\t\t\tIdentifier: element,\n\t\t\t\t\tSearchKey: \"\", // don't index for search\n\t\t\t\t\tNewPage: true,\n\t\t\t\t\tTags: []protocol.Tag{},\n\t\t\t\t},\n\t\t\t\tLabel: protocol.NewMarkupContent(\"\", protocol.PlainText),\n\t\t\t\tDetail: protocol.NewMarkupContent(\"\", protocol.PlainText),\n\t\t\t}).emit(i.emitter)\n\t\t\tparentPage, ok := pagesByPath[parentPath]\n\t\t\tif !ok {\n\t\t\t\tpanic(\"invariant: parentPage should always exist(1)\")\n\t\t\t}\n\t\t\tparentPage.children = append(parentPage.children, pageID)\n\t\t\tpagesByPath[currentPath] = &page{id: pageID}\n\t\t}\n\t}\n\n\t// Finalize children of pages.\n\tfor _, docsPkg := range docsPackages {\n\t\trelPackagePath := d.relPackagePath(docsPkg.Path, rootPkgPath)\n\n\t\t// Attach the children sections of the page (consts/vars/etc) as children of the page itself.\n\t\tpage, ok := pagesByPath[relPackagePath]\n\t\tif !ok {\n\t\t\tpanic(\"invariant: page should always exist\")\n\t\t}\n\t\tpage.children = append(page.children, docsPkg.children...)\n\n\t\t// Attach package documentation pages as children of their parent (either another package\n\t\t// documentation page, or a blank index page.)\n\t\tif relPackagePath == \"\" {\n\t\t\t// root is not a child of anything.\n\t\t\tcontinue\n\t\t}\n\t\tpkgPathElements := strings.Split(relPackagePath, \"/\") // [\"internal\", \"lib\", \"protocol\"]\n\t\tparentPath := path.Join(pkgPathElements[:len(pkgPathElements)-1]...)\n\t\tparentPage, ok := pagesByPath[parentPath]\n\t\tif !ok {\n\t\t\tpanic(\"invariant: parentPage should always exist(2)\")\n\t\t}\n\t\tparentPage.children = append(parentPage.children, docsPkg.ID)\n\t}\n\n\t// Emit children edges of all pages.\n\tfor _, page := range pagesByPath {\n\t\t_ = i.emitter.EmitDocumentationChildrenEdge(page.children, page.id)\n\t}\n\n\ti.emittedDocumentationResults = emitted\n\ti.emittedDocumentationResultsByPackagePath = emittedPackagesByPath\n\treturn errs\n}", "func (i *indexer) Index() (*Stats, error) {\n\tpkgs, err := i.packages()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn i.index(pkgs)\n}", "func (h *HTTPApi) listIndex(w http.ResponseWriter, r *http.Request, ps httprouter.Params) {\n\tcollections := h.storageNode.Datasources[ps.ByName(\"datasource\")].GetMeta().Databases[ps.ByName(\"dbname\")].ShardInstances[ps.ByName(\"shardinstance\")].Collections[ps.ByName(\"collectionname\")]\n\n\t// Now we need to return the results\n\tif bytes, err := json.Marshal(collections.Indexes); err != nil {\n\t\t// TODO: log this better?\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t} else {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(bytes)\n\t}\n}", "func (i indexer) Index(ctx context.Context, req IndexQuery) (\n\tresp *IndexResult, err error) {\n\n\tlog.Info(\"index [%v] root [%v] len_dirs=%v len_files=%v\",\n\t\treq.Key, req.Root, len(req.Dirs), len(req.Files))\n\tstart := time.Now()\n\t// Setup the response\n\tresp = NewIndexResult()\n\tif err = req.Normalize(); err != nil {\n\t\tlog.Info(\"index [%v] error: %v\", req.Key, err)\n\t\tresp.Error = errs.NewStructError(err)\n\t\treturn\n\t}\n\n\t// create index shards\n\tvar nshards int\n\tif nshards = i.cfg.NumShards; nshards == 0 {\n\t\tnshards = 1\n\t}\n\tnshards = utils.MinInt(nshards, maxShards)\n\ti.shards = make([]index.IndexWriter, nshards)\n\ti.root = getRoot(i.cfg, &req)\n\n\tfor n := range i.shards {\n\t\tname := path.Join(i.root, shardName(req.Key, n))\n\t\tixw, err := getIndexWriter(ctx, name)\n\t\tif err != nil {\n\t\t\tresp.Error = errs.NewStructError(err)\n\t\t\treturn resp, nil\n\t\t}\n\t\ti.shards[n] = ixw\n\t}\n\n\tfs := getFileSystem(ctx, i.root)\n\trepo := newRepoFromQuery(&req, i.root)\n\trepo.SetMeta(i.cfg.RepoMeta, req.Meta)\n\tresp.Repo = repo\n\n\t// Add query Files and scan Dirs for files to index\n\tnames, err := i.scanner(fs, &req)\n\tch := make(chan int, nshards)\n\tchnames := make(chan string, 100)\n\tgo func() {\n\t\tfor _, name := range names {\n\t\t\tchnames <- name\n\t\t}\n\t\tclose(chnames)\n\t}()\n\treqch := make(chan par.RequestFunc, nshards)\n\tfor _, shard := range i.shards {\n\t\treqch <- indexShard(&i, &req, shard, fs, chnames, ch)\n\t}\n\tclose(reqch)\n\terr = par.Requests(reqch).WithConcurrency(nshards).DoWithContext(ctx)\n\tclose(ch)\n\n\t// Await results, each indicating the number of files scanned\n\tfor num := range ch {\n\t\trepo.NumFiles += num\n\t}\n\n\trepo.NumShards = len(i.shards)\n\t// Flush our index shard files\n\tfor _, shard := range i.shards {\n\t\tshard.Flush()\n\t\trepo.SizeIndex += ByteSize(shard.IndexBytes())\n\t\trepo.SizeData += ByteSize(shard.DataBytes())\n\t\tlog.Debug(\"index flush %v (data) %v (index)\",\n\t\t\trepo.SizeData, repo.SizeIndex)\n\t}\n\trepo.ElapsedIndexing = time.Since(start)\n\trepo.TimeUpdated = time.Now().UTC()\n\n\tvar msg string\n\tif err != nil {\n\t\trepo.State = ERROR\n\t\tresp.SetError(err)\n\t\tmsg = \"error: \" + resp.Error.Error()\n\t} else {\n\t\trepo.State = OK\n\t\tmsg = \"ok \" + fmt.Sprintf(\n\t\t\t\"(%v files, %v data, %v index)\",\n\t\t\trepo.NumFiles, repo.SizeData, repo.SizeIndex)\n\t}\n\tlog.Info(\"index [%v] %v [%v]\", req.Key, msg, repo.ElapsedIndexing)\n\treturn\n}", "func indexesReport(c *clients.Client, response handle.ResponseHandle) error {\n\treq, err := http.NewRequest(\"GET\", c.Base()+\"/config/indexes\", nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn util.Execute(c, req, response)\n}", "func indexRead(key string) (*pkg, bool) {\n\tlocker.RLock()\n\tdefer locker.RUnlock()\n\n\tp, ok := indexedPkgs[key]\n\treturn p, ok\n}", "func index() string {\n\tvar buffer bytes.Buffer\n\tvar id = 0\n\tvar class = 0\n\tbuffer.WriteString(indexTemplate)\n\tlock.Lock()\n\tfor folderName, folder := range folders {\n\t\tbuffer.WriteString(fmt.Sprintf(\"<h2>%s</h2>\", folderName))\n\t\tfor _, source := range folder {\n\t\t\tif !anyNonRead(source) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tsort.Sort(source)\n\t\t\tbuffer.WriteString(fmt.Sprintf(\"<h3>%s</h3>\", source.Title))\n\t\t\tbuffer.WriteString(fmt.Sprintf(`<button onClick=\"hideAll('source_%d'); return false\">Mark all as read</button>`, class))\n\t\t\tbuffer.WriteString(\"<ul>\")\n\n\t\t\tfor _, entry := range source.Entries {\n\t\t\t\tif entry.Read {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tbuffer.WriteString(fmt.Sprintf(`<li id=\"entry_%d\">`, id))\n\t\t\t\tbuffer.WriteString(fmt.Sprintf(`<button class=\"source_%d\" onClick=\"hide('entry_%d', '%s'); return false\">Mark Read</button> `, class, id, entry.Url))\n\t\t\t\tbuffer.WriteString(fmt.Sprintf(`<a href=\"%s\">%s</a>`, entry.Url, entry.Title))\n\t\t\t\tbuffer.WriteString(\"</li>\")\n\t\t\t\tid += 1\n\t\t\t}\n\t\t\tbuffer.WriteString(\"</ul>\")\n\t\t\tclass += 1\n\t\t}\n\t}\n\tlock.Unlock()\n\tbuffer.WriteString(\"</body></html>\")\n\treturn buffer.String()\n}", "func getIndexLayout(clusterUrl string) ([]*IndexerNode, error) {\n\n\tcinfo, err := clusterInfoCache(clusterUrl)\n\tif err != nil {\n\t\tlogging.Errorf(\"Planner::getIndexLayout: Error from connecting to cluster at %v. Error = %v\", clusterUrl, err)\n\t\treturn nil, err\n\t}\n\n\t// find all nodes that has a index http service\n\t// If there is any indexer node that is not in active state (e.g. failover), then planner will skip those indexers.\n\t// Note that if the planner is invoked by the rebalancer, the rebalancer will receive callback ns_server if there is\n\t// an indexer node fails over while planning is happening.\n\tnids := cinfo.GetNodesByServiceType(common.INDEX_HTTP_SERVICE)\n\n\tlist := make([]*IndexerNode, 0)\n\tnumIndexes := 0\n\n\tfor _, nid := range nids {\n\n\t\t// create an empty indexer object using the indexer host name\n\t\tnode, err := createIndexerNode(cinfo, nid)\n\t\tif err != nil {\n\t\t\tlogging.Errorf(\"Planner::getIndexLayout: Error from initializing indexer node. Error = %v\", err)\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// assign server group\n\t\tnode.ServerGroup = cinfo.GetServerGroup(nid)\n\n\t\t// obtain the admin port for the indexer node\n\t\taddr, err := cinfo.GetServiceAddress(nid, common.INDEX_HTTP_SERVICE)\n\t\tif err != nil {\n\t\t\tlogging.Errorf(\"Planner::getIndexLayout: Error from getting service address for node %v. Error = %v\", node.NodeId, err)\n\t\t\treturn nil, err\n\t\t}\n\t\tnode.RestUrl = addr\n\n\t\t// Read the index metadata from the indexer node.\n\t\tlocalMeta, err := getLocalMetadata(addr)\n\t\tif err != nil {\n\t\t\tlogging.Errorf(\"Planner::getIndexLayout: Error from reading index metadata for node %v. Error = %v\", node.NodeId, err)\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// get the node UUID\n\t\tnode.NodeUUID = localMeta.NodeUUID\n\t\tnode.IndexerId = localMeta.IndexerId\n\t\tnode.StorageMode = localMeta.StorageMode\n\n\t\t// convert from LocalIndexMetadata to IndexUsage\n\t\tindexes, err := ConvertToIndexUsages(localMeta, node)\n\t\tif err != nil {\n\t\t\tlogging.Errorf(\"Planner::getIndexLayout: Error for converting index metadata to index usage for node %v. Error = %v\", node.NodeId, err)\n\t\t\treturn nil, err\n\t\t}\n\n\t\tnode.Indexes = indexes\n\t\tnumIndexes += len(indexes)\n\t\tlist = append(list, node)\n\t}\n\n\tif numIndexes != 0 {\n\t\tfor _, node := range list {\n\t\t\tif !common.IsValidIndexType(node.StorageMode) {\n\t\t\t\terr := errors.New(fmt.Sprintf(\"Fail to get storage mode\tfrom %v. Storage mode = %v\", node.RestUrl, node.StorageMode))\n\t\t\t\tlogging.Errorf(\"Planner::getIndexLayout: Error = %v\", err)\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn list, nil\n}", "func BuildPackageIndex(pkgs []*Package) PackageIndex {\n\tidx := PackageIndex{\n\t\tPackages: pkgs,\n\t\tnameMap: make(map[string]*Package, len(pkgs)),\n\t\tcharMap: make(map[string][]*CompletionItem),\n\t}\n\n\tfor _, pkg := range pkgs {\n\t\taddPackageToIndex(pkg, &idx)\n\t}\n\n\treturn idx\n}", "func loadIndexs() {\n\tdb := open()\n\tindexs = make(map[string][]*Index)\n\tdb.View(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket([]byte(INDEX_BUCKET))\n\t\tif b == nil {\n\t\t\tlogger.Infof(\"bucket[%s] not exist\", INDEX_BUCKET)\n\t\t\treturn nil\n\t\t}\n\t\tc := b.Cursor()\n\t\tfor k, v := c.First(); k != nil; k, v = c.Next() {\n\t\t\tkey := string(k)\n\t\t\tvar _indexs []string\n\t\t\terr := json.Unmarshal(v, &_indexs)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorf(\"parse index[%s] error -> %v\", k, err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t__indexs := make([]*Index, len(_indexs)) \n\t\t\t// parse index\n\t\t\tfor i, _index := range _indexs {\n\t\t\t\tsps :=strings.Split(_index, INDEX_SPLIT) \n\t\t\t\tindex := &Index {\n\t\t\t\t\tbucket: key,\n\t\t\t\t\tindexs: sps,\n\t\t\t\t}\n\t\t\t\t__indexs[i] = index\n\t\t\t}\n\t\t\tindexs[key] = __indexs\n\t\t}\n\t\treturn nil\n\t})\n}", "func (resource *Resource) GetResourceIndex(component Component) int {\n\tfor i, currentResource := range component.GetResources() {\n\t\tif currentResource.Group == resource.Group && currentResource.Version == resource.Version && currentResource.Kind == resource.Kind {\n\t\t\tif currentResource.Name == resource.Name && currentResource.Namespace == resource.Namespace {\n\t\t\t\treturn i\n\t\t\t}\n\t\t}\n\t}\n\n\treturn -1\n}", "func getFilesFromIndex(p string, r io.Reader) ([]*FileInfo, Paragraph, error) {\n\treturn getFilesFromRelease(p, r)\n}", "func Indexes() map[string]string {\n\treturn instance.getIndexes()\n}", "func indexWrite(key string, p *pkg) {\n\tlocker.Lock()\n\tdefer locker.Unlock()\n\n\tindexedPkgs[key] = p\n}", "func getResources(client *ApiClient) (map[schema.GroupVersion][]metav1.APIResource, error) {\n\n\tresourceLists, err := client.ClientSet.Discovery().ServerPreferredResources()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tversionResource := map[schema.GroupVersion][]metav1.APIResource{}\n\n\tfor _, apiResourceList := range resourceLists {\n\t\tversion, err := schema.ParseGroupVersion(apiResourceList.GroupVersion)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"unable to parse GroupVersion %v\",err)\n\t\t}\n\n\t\tversionResource[version] = uniqResources(apiResourceList.APIResources)\n\t}\n\n\treturn versionResource, nil\n}", "func IndexDirectory(fs afero.Fs, path string, url string, now *time.Time) (*IndexFile, error) {\n\tarchives, err := afero.Glob(fs, filepath.Join(path, \"*.tgz\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif len(archives) == 0 {\n\t\treturn nil, errors.New(\"no packages discovered\")\n\t}\n\tindex := newIndexFile(now)\n\tops := filesDigest(fs, archives)\n\tpvs := Map(ops, url)\n\tfor _, pv := range pvs {\n\t\terr = index.AddPackageVersion(pv)\n\t\t// on error we report and continue\n\t\tif err != nil {\n\t\t\tfmt.Print(err.Error())\n\t\t}\n\t}\n\tindex.sortPackages()\n\treturn index, nil\n}", "func (i ImageIndexer) ExportFromIndex(request ExportFromIndexRequest) error {\n\t// set a temp directory\n\tworkingDir, err := ioutil.TempDir(\"./\", tmpDirPrefix)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer os.RemoveAll(workingDir)\n\n\t// extract the index database to the file\n\tdatabaseFile, err := i.getDatabaseFile(workingDir, request.Index, request.CaFile, request.SkipTLSVerify, request.PlainHTTP)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdb, err := sqlite.Open(databaseFile)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer db.Close()\n\n\tdbQuerier := sqlite.NewSQLLiteQuerierFromDb(db)\n\n\t// fetch all packages from the index image if packages is empty\n\tif len(request.Packages) == 0 {\n\t\trequest.Packages, err = dbQuerier.ListPackages(context.TODO())\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tbundles, err := getBundlesToExport(dbQuerier, request.Packages)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ti.Logger.Infof(\"Preparing to pull bundles %+q\", bundles)\n\n\t// Creating downloadPath dir\n\tif err := os.MkdirAll(request.DownloadPath, 0777); err != nil {\n\t\treturn err\n\t}\n\n\tvar errs []error\n\tvar wg sync.WaitGroup\n\twg.Add(len(bundles))\n\tvar mu = &sync.Mutex{}\n\n\tsem := make(chan struct{}, concurrencyLimitForExport)\n\n\tfor bundleImage, bundleDir := range bundles {\n\t\tgo func(bundleImage string, bundleDir bundleDirPrefix) {\n\t\t\tdefer wg.Done()\n\n\t\t\tsem <- struct{}{}\n\t\t\tdefer func() {\n\t\t\t\t<-sem\n\t\t\t}()\n\n\t\t\t// generate a random folder name if bundle version is empty\n\t\t\tif bundleDir.bundleVersion == \"\" {\n\t\t\t\tbundleDir.bundleVersion = strconv.Itoa(rand.Intn(10000))\n\t\t\t}\n\t\t\texporter := bundle.NewExporterForBundle(bundleImage, filepath.Join(request.DownloadPath, bundleDir.pkgName, bundleDir.bundleVersion), request.ContainerTool)\n\t\t\tif err := exporter.Export(request.SkipTLSVerify, request.PlainHTTP); err != nil {\n\t\t\t\terr = fmt.Errorf(\"exporting bundle image:%s failed with %s\", bundleImage, err)\n\t\t\t\tmu.Lock()\n\t\t\t\terrs = append(errs, err)\n\t\t\t\tmu.Unlock()\n\t\t\t}\n\t\t}(bundleImage, bundleDir)\n\t}\n\t// Wait for all the go routines to finish export\n\twg.Wait()\n\n\tif errs != nil {\n\t\treturn utilerrors.NewAggregate(errs)\n\t}\n\n\tfor _, packageName := range request.Packages {\n\t\terr := generatePackageYaml(dbQuerier, packageName, filepath.Join(request.DownloadPath, packageName))\n\t\tif err != nil {\n\t\t\terrs = append(errs, err)\n\t\t}\n\t}\n\treturn utilerrors.NewAggregate(errs)\n}", "func (m *ProjectIndexer) Index(resource *models.Project, doc solr.Document) solr.Document {\n\tdoc.Set(\"type_ssi\", \"Project\")\n\tdoc.Set(\"title_tesi\", resource.Title)\n\tdoc.Set(\"alternative_title_tesim\", resource.AlternativeTitle)\n\tdoc.Set(\"start_date_ssi\", resource.StartDate)\n\tdoc.Set(\"end_date_ssi\", resource.EndDate)\n\n\treturn doc\n}", "func (d *Descriptor) ImageIndex() (v1.ImageIndex, error) {\n\tswitch d.MediaType {\n\tcase types.DockerManifestSchema1, types.DockerManifestSchema1Signed:\n\t\t// We don't care to support schema 1 images:\n\t\t// https://github.com/google/go-containerregistry/issues/377\n\t\treturn nil, newErrSchema1(d.MediaType)\n\tcase types.OCIManifestSchema1, types.DockerManifestSchema2:\n\t\t// We want an index but the registry has an image, nothing we can do.\n\t\treturn nil, fmt.Errorf(\"unexpected media type for ImageIndex(): %s; call Image() instead\", d.MediaType)\n\tcase types.OCIImageIndex, types.DockerManifestList:\n\t\t// These are expected.\n\tdefault:\n\t\t// We could just return an error here, but some registries (e.g. static\n\t\t// registries) don't set the Content-Type headers correctly, so instead...\n\t\tlogs.Warn.Printf(\"Unexpected media type for ImageIndex(): %s\", d.MediaType)\n\t}\n\treturn d.remoteIndex(), nil\n}", "func (idx *ManualIndex) Index() error {\n\tvar buf bytes.Buffer\n\n\tfor pkg := range idx.packages {\n\t\t_, err := fmt.Fprintf(&buf, \"\\x00%s\", pkg)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tidx.index = suffixarray.New(buf.Bytes())\n\treturn nil\n}", "func index(w http.ResponseWriter, r *http.Request){\n\terr := templ.ExecuteTemplate(w, \"index\", nil)\n\tif err != nil {\n\t\tfmt.Print(err.Error())\n\t}\n}", "func (c *clusterCache) listResources(ctx context.Context, resClient dynamic.ResourceInterface, callback func(*pager.ListPager) error) (string, error) {\n\tif err := c.listSemaphore.Acquire(ctx, 1); err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer c.listSemaphore.Release(1)\n\tvar retryCount int64 = 0\n\tresourceVersion := \"\"\n\tlistPager := pager.New(func(ctx context.Context, opts metav1.ListOptions) (runtime.Object, error) {\n\t\tvar res *unstructured.UnstructuredList\n\t\tvar listRetry wait.Backoff\n\n\t\tif c.listRetryUseBackoff {\n\t\t\tlistRetry = retry.DefaultBackoff\n\t\t} else {\n\t\t\tlistRetry = retry.DefaultRetry\n\t\t}\n\n\t\tlistRetry.Steps = int(c.listRetryLimit)\n\t\terr := retry.OnError(listRetry, c.listRetryFunc, func() error {\n\t\t\tvar ierr error\n\t\t\tres, ierr = resClient.List(ctx, opts)\n\t\t\tif ierr != nil {\n\t\t\t\t// Log out a retry\n\t\t\t\tif c.listRetryLimit > 1 && c.listRetryFunc(ierr) {\n\t\t\t\t\tretryCount += 1\n\t\t\t\t\tc.log.Info(fmt.Sprintf(\"Error while listing resources: %v (try %d/%d)\", ierr, retryCount, c.listRetryLimit))\n\t\t\t\t}\n\t\t\t\treturn ierr\n\t\t\t}\n\t\t\tresourceVersion = res.GetResourceVersion()\n\t\t\treturn nil\n\t\t})\n\t\treturn res, err\n\t})\n\tlistPager.PageBufferSize = c.listPageBufferSize\n\tlistPager.PageSize = c.listPageSize\n\n\treturn resourceVersion, callback(listPager)\n}", "func TagIndex(c *gin.Context) {\n\tdb := models.GetDB()\n\tvar tags []models.Tag\n\tdb.Preload(\"Posts\").Order(\"title asc\").Find(&tags)\n\th := DefaultH(c)\n\th[\"Title\"] = \"タグ一覧\"\n\th[\"Tags\"] = tags\n\tc.HTML(http.StatusOK, \"tags/index\", h)\n}", "func (l *PackageList) PrepareIndex() {\n\tif l.indexed {\n\t\treturn\n\t}\n\n\tl.packagesIndex = make([]*Package, l.Len())\n\tl.providesIndex = make(map[string][]*Package, 128)\n\n\ti := 0\n\tfor _, p := range l.packages {\n\t\tl.packagesIndex[i] = p\n\t\ti++\n\n\t\tfor _, provides := range p.Provides {\n\t\t\tl.providesIndex[provides] = append(l.providesIndex[provides], p)\n\t\t}\n\t}\n\n\tsort.Sort(l)\n\n\tl.indexed = true\n}", "func (*SearchAllResourcesResponse) Descriptor() ([]byte, []int) {\n\treturn file_google_cloud_asset_v1_asset_service_proto_rawDescGZIP(), []int{23}\n}", "func getItems(repoIndex *index.Index, url string) []FileInfo {\n\tvar items []FileInfo\n\n\tfor _, os := range repoIndex.Data.Keys() {\n\t\tfor _, arch := range repoIndex.Data[os].Keys() {\n\t\t\tfor _, category := range repoIndex.Data[os][arch].Keys() {\n\t\t\t\tfor _, version := range repoIndex.Data[os][arch][category] {\n\t\t\t\t\titems = append(items, FileInfo{\n\t\t\t\t\t\tFile: version.File,\n\t\t\t\t\t\tURL: url + \"/\" + version.Path + \"/\" + version.File,\n\t\t\t\t\t\tOS: os,\n\t\t\t\t\t\tArch: arch,\n\t\t\t\t\t\tSize: version.Size,\n\t\t\t\t\t})\n\n\t\t\t\t\tif len(version.Variations) != 0 {\n\t\t\t\t\t\tfor _, subVersion := range version.Variations {\n\t\t\t\t\t\t\titems = append(items, FileInfo{\n\t\t\t\t\t\t\t\tFile: subVersion.File,\n\t\t\t\t\t\t\t\tURL: url + \"/\" + subVersion.Path + \"/\" + subVersion.File,\n\t\t\t\t\t\t\t\tOS: os,\n\t\t\t\t\t\t\t\tArch: arch,\n\t\t\t\t\t\t\t\tSize: subVersion.Size,\n\t\t\t\t\t\t\t})\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn items\n}", "func generateSignedIndex(scenario string, signer digestedRefSigner, keys ...*cosign.KeysBytes) (layers []v1.Layer, manifests map[string]v1.Manifest, hash v1.Hash, err error) {\n\tmanifests = map[string]v1.Manifest{}\n\n\t// Generating two manifests A and B\n\tl, m, h, err := generateSignedManifest(scenario+\" manifest A\", signDigestedRef)\n\tif err != nil {\n\t\treturn nil, nil, v1.Hash{}, trace.Wrap(err)\n\t}\n\tlayers = append(layers, l...)\n\tmaps.Copy(manifests, m)\n\tmanifestA := m[h.String()]\n\tmanifestADesc, err := descriptorFromManifest(manifestA, &v1.Platform{\n\t\tArchitecture: \"arm64\",\n\t\tOS: \"linux\",\n\t})\n\tif err != nil {\n\t\treturn nil, nil, v1.Hash{}, trace.Wrap(err)\n\t}\n\n\tl, m, h, err = generateSignedManifest(scenario+\" manifest B\", signDigestedRef)\n\tif err != nil {\n\t\treturn nil, nil, v1.Hash{}, trace.Wrap(err)\n\t}\n\tlayers = append(layers, l...)\n\tmaps.Copy(manifests, m)\n\tmanifestB := m[h.String()]\n\tmanifestBDesc, err := descriptorFromManifest(manifestB, &v1.Platform{\n\t\tArchitecture: \"amd64\",\n\t\tOS: \"linux\",\n\t})\n\tif err != nil {\n\t\treturn nil, nil, v1.Hash{}, trace.Wrap(err)\n\t}\n\n\t// Referencing both manifests in an index\n\tindex := v1.IndexManifest{\n\t\tSchemaVersion: 2,\n\t\tMediaType: types.DockerManifestList,\n\t\tManifests: []v1.Descriptor{manifestADesc, manifestBDesc},\n\t}\n\n\t_, _, indexDigest, err := contentSizeAndHash(index)\n\tif err != nil {\n\t\treturn nil, nil, v1.Hash{}, trace.Wrap(err)\n\t}\n\thash = indexDigest\n\tindexRef, err := name.NewDigest(imagePath + \"@\" + indexDigest.String())\n\n\t// Don't sign when no keys are provided\n\tif len(keys) == 0 {\n\t\treturn\n\t}\n\n\t// Signing the index\n\tsigLayers, sigManifest, err := makeSignature(indexRef, signer, keys...)\n\tif err != nil {\n\t\treturn nil, nil, v1.Hash{}, trace.Wrap(err)\n\t}\n\tmanifests[Triangulate(indexDigest)] = sigManifest\n\tlayers = append(layers, sigLayers...)\n\treturn\n}", "func index(w http.ResponseWriter, req *http.Request, ctx httputil.Context) (e *httputil.Error) {\n\tif req.URL.Path != \"/\" {\n\t\tnotFound(w, req)\n\t\treturn\n\t}\n\tm := newManager(ctx)\n\n\tres, err := m.Index()\n\tif err != nil {\n\t\te = httputil.Errorf(err, \"couldn't query for test results\")\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"text/html\")\n\tif err := T(\"index/index.html\").Execute(w, res); err != nil {\n\t\te = httputil.Errorf(err, \"error executing index template\")\n\t}\n\treturn\n}", "func getTagDocuments(p config.Config, swagger *openapi3.Swagger, allDocuments docs.Index) docs.Index {\n\tio.WriteString(os.Stdout, fmt.Sprintf(\"\\033[1m %s\\033[0m (%v tags)\\n\", \"Tags\", len(swagger.Tags)))\n\tfor _, tag := range swagger.Tags {\n\n\t\tvar document docs.Document\n\n\t\t// Basics.\n\t\tdocument.Site = p.Name\n\t\tdocument.Title = tag.Name\n\t\tdocument.Section = \"\"\n\t\tdocument.Subsection = \"\"\n\n\t\t// URLs.\n\t\trel_url := fmt.Sprintf(\"#tag/%s\", strings.Replace(tag.Name, \" \", \"-\", -1))\n\t\tfull_url := fmt.Sprintf(\"%s%s\", p.URL, rel_url)\n\t\tdocument.URL = full_url\n\t\tdocument.RelativeURL = fmt.Sprintf(\"/%s\", rel_url)\n\n\t\t// DocumentID hash.\n\t\th := sha1.New()\n\t\th.Write([]byte(full_url))\n\t\tdocument.DocumentID = fmt.Sprintf(\"%x\", h.Sum(nil))\n\n\t\t// Match `config.yaml` rank, and use React primary/secondary designation.\n\t\tdocument.Rank = p.Rank\n\t\tif p.Rank == 1 {\n\t\t\tdocument.Source = \"primary\"\n\t\t} else {\n\t\t\tdocument.Source = \"secondary\"\n\t\t}\n\n\t\t// Document body text.\n\t\tdocument.Text = strings.Replace(tag.Description, \"\\n\", \" \", -1)\n\n\t\t// Document description.\n\t\tdocument.Description = strings.Replace(tag.Description, \"\\n\", \" \", -1)\n\n\t\t// Append the document.\n\t\tallDocuments.Documents = append(allDocuments.Documents, document)\n\t}\n\n\treturn allDocuments\n}", "func List(indexName string, options types.ListingOptions, indexes map[string]string) (*bleve.SearchResult, error) {\n\tcwd, _ := os.Getwd()\n\tp := cwd + viper.GetString(\"storage.basedir\") + \"/indexes/\" + indexName\n\n\tindex, err := openIndex(p)\n\n\titems_by_page := options.PageSize\n\tif items_by_page == 0 {\n\t\titems_by_page = viper.GetUint32(\"modules.all.items_per_page\")\n\t}\n\n\t// bleve start with page 1\n\tpage := options.Page\n\tif page <= 0 {\n\t\tpage = 1\n\t} else {\n\t\tpage++\n\t}\n\n\t// field scoping\n\tfields := []string{}\n\tfor k, v := range indexes {\n\t\tfields = append(fields, k+\":\"+v)\n\t}\n\tfieldscope := strings.Join(fields, \" \")\n\n\tvar searchRequest *bleve.SearchRequest\n\n\tif options.Q == \"\" {\n\n\t\tbq := bleve.NewBooleanQuery()\n\t\tbq.Must = bleve.NewMatchQuery(fieldscope)\n\t\tbq.Should = bleve.NewMatchAllQuery()\n\t\tquery := bq\n\t\tsearchRequest = bleve.NewSearchRequestOptions(query, int(items_by_page), int((page-1)*items_by_page), false)\n\t} else {\n\t\tquery := bleve.NewFuzzyQuery(fieldscope + options.Q)\n\t\tsearchRequest = bleve.NewSearchRequestOptions(query, int(items_by_page), int((page-1)*items_by_page), false)\n\t}\n\t// default sort order is id desc\n\tsortOrder := []string{\"-_id\"}\n\n\tif options.OrderBy != \"\" {\n\t\tsortOrder = strings.Split(strings.ReplaceAll(options.OrderBy, \" \", \"\"), \",\")\n\t}\n\n\t//searchRequest.Fields = strings.Split(strings.ReplaceAll(options.Fields, \" \", \"\"), \",\")\n\t// todo implement options.Filter\n\n\tsearchRequest.SortBy(sortOrder)\n\n\tres, err := index.Search(searchRequest)\n\treturn res, err\n}", "func registerSymbolIndex(sdk sdkProvider, packages []string, verbose bool) {\n\tfor _, pkg := range packages {\n\t\t// pkg should end with \".far\", otherwise the publish function should fail.\n\t\tsymbolIndexJsonFile := pkg[:len(pkg)-4] + \".symbol-index.json\"\n\t\tif _, err := os.Stat(symbolIndexJsonFile); err != nil {\n\t\t\t// File doesn't exist or is not readable.\n\t\t\tcontinue\n\t\t}\n\n\t\targs := []string{\"debug\", \"symbol-index\", \"add\", symbolIndexJsonFile}\n\t\tif verbose {\n\t\t\tfmt.Printf(\"Running command: ffx %v\\n\", args)\n\t\t}\n\t\t// The command outputs nothing if succeeds, and outputs error messages if fails,\n\t\t// which is sufficient for our users. Use interactive=true here allows the\n\t\t// command to output.\n\t\tsdk.RunFFX(args, true)\n\t}\n}", "func indexHandler(w http.ResponseWriter, req *http.Request) {\n\tlayout, err := template.ParseFile(PATH_PUBLIC + TEMPLATE_LAYOUT)\n\tif err != nil {\n\t\thttp.Error(w, ERROR_TEMPLATE_NOT_FOUND, http.StatusNotFound)\n\t\treturn\n\t}\n\tindex, err := template.ParseFile(PATH_PUBLIC + TEMPLATE_INDEX)\n\t//artical, err := template.ParseFile(PATH_PUBLIC + TEMPLATE_ARTICAL)\n\tif err != nil {\n\t\thttp.Error(w, ERROR_TEMPLATE_NOT_FOUND, http.StatusNotFound)\n\t\treturn\n\t}\n\tmapOutput := map[string]interface{}{\"Title\": \"炫酷的网站技术\" + TITLE, \"Keyword\": KEYWORD, \"Description\": DESCRIPTION, \"Base\": BASE_URL, \"Url\": BASE_URL, \"Carousel\": getAddition(PREFIX_INDEX), \"Script\": getAddition(PREFIX_SCRIPT), \"Items\": leveldb.GetRandomContents(20, &Filter{})}\n\tcontent := []byte(index.RenderInLayout(layout, mapOutput))\n\tw.Write(content)\n\tgo cacheFile(\"index\", content)\n}", "func Index(w http.ResponseWriter, r *http.Request) {\n\t//index_routes := []string{\"providers\"}\n\tvar index_routes []string\n\tindex_routes = GetJobs()\n\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\tw.WriteHeader(http.StatusOK)\n\tif err := json.NewEncoder(w).Encode(index_routes); err != nil {\n\t\tpanic(err)\n\t}\n}", "func (o MongoDBCollectionResourceOutput) Indexes() MongoIndexArrayOutput {\n\treturn o.ApplyT(func(v MongoDBCollectionResource) []MongoIndex { return v.Indexes }).(MongoIndexArrayOutput)\n}", "func getIndexStats(clusterUrl string, plan *Plan) error {\n\n\tcinfo, err := clusterInfoCache(clusterUrl)\n\tif err != nil {\n\t\tlogging.Errorf(\"Planner::getIndexStats: Error from connecting to cluster at %v. Error = %v\", clusterUrl, err)\n\t\treturn err\n\t}\n\n\t// find all nodes that has a index http service\n\tnids := cinfo.GetNodesByServiceType(common.INDEX_HTTP_SERVICE)\n\n\tfor _, nid := range nids {\n\n\t\t// Find the indexer host name\n\t\tnodeId, err := getIndexerHost(cinfo, nid)\n\t\tif err != nil {\n\t\t\tlogging.Errorf(\"Planner::getIndexStats: Error from initializing indexer node. Error = %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t// obtain the admin port for the indexer node\n\t\taddr, err := cinfo.GetServiceAddress(nid, common.INDEX_HTTP_SERVICE)\n\t\tif err != nil {\n\t\t\tlogging.Errorf(\"Planner::getIndexStats: Error from getting service address for node %v. Error = %v\", nodeId, err)\n\t\t\treturn err\n\t\t}\n\n\t\t// Read the index stats from the indexer node.\n\t\tstats, err := getLocalStats(addr)\n\t\tif err != nil {\n\t\t\tlogging.Errorf(\"Planner::getIndexStats: Error from reading index stats for node %v. Error = %v\", nodeId, err)\n\t\t\treturn err\n\t\t}\n\n\t\t// look up the corresponding indexer object based on the nodeId\n\t\tindexer := findIndexerByNodeId(plan.Placement, nodeId)\n\t\tstatsMap := stats.ToMap()\n\n\t\t/*\n\t\t\tCpuUsage uint64 `json:\"cpuUsage,omitempty\"`\n\t\t\tDiskUsage uint64 `json:\"diskUsage,omitempty\"`\n\t\t*/\n\n\t\tvar actualStorageMem uint64\n\t\t// memory_used_storage constains the total storage consumption,\n\t\t// including fdb overhead, main index and back index. This also\n\t\t// includes overhead (skip list / back index).\n\t\tif memUsedStorage, ok := statsMap[\"memory_used_storage\"]; ok {\n\t\t\tactualStorageMem = uint64(memUsedStorage.(float64))\n\t\t}\n\n\t\t// memory_used is the memory used by indexer. This includes\n\t\t// golang in-use heap space, golang idle heap space, and\n\t\t// storage memory manager space (e.g. jemalloc heap space).\n\t\tvar actualTotalMem uint64\n\t\tif memUsed, ok := statsMap[\"memory_used\"]; ok {\n\t\t\tactualTotalMem = uint64(memUsed.(float64))\n\t\t}\n\n\t\t// memory_quota is user specified memory quota.\n\t\tif memQuota, ok := statsMap[\"memory_quota\"]; ok {\n\t\t\tplan.MemQuota = uint64(memQuota.(float64))\n\t\t}\n\n\t\t// uptime\n\t\tvar elapsed uint64\n\t\tif uptimeStat, ok := statsMap[\"uptime\"]; ok {\n\t\t\tuptime := uptimeStat.(string)\n\t\t\tif duration, err := time.ParseDuration(uptime); err == nil {\n\t\t\t\telapsed = uint64(duration.Seconds())\n\t\t\t}\n\t\t}\n\n\t\t// cpu core in host. This is the actual num of cpu core, not cpu quota.\n\t\t/*\n\t\t\tvar actualCpuCore uint64\n\t\t\tif cpuCore, ok := statsMap[\"num_cpu_core\"]; ok {\n\t\t\t\tactualCpuCore = uint64(cpuCore.(float64))\n\t\t\t}\n\t\t*/\n\n\t\t// cpu utilization for the indexer process\n\t\tvar actualCpuUtil float64\n\t\tif cpuUtil, ok := statsMap[\"cpu_utilization\"]; ok {\n\t\t\tactualCpuUtil = cpuUtil.(float64) / 100\n\t\t}\n\n\t\tvar totalDataSize uint64\n\t\tvar totalMutation uint64\n\t\tvar totalScan uint64\n\t\tfor _, index := range indexer.Indexes {\n\n\t\t\t/*\n\t\t\t\tCpuUsage uint64 `json:\"cpuUsage,omitempty\"`\n\t\t\t\tDiskUsage uint64 `json:\"diskUsage,omitempty\"`\n\t\t\t*/\n\n\t\t\tvar key string\n\n\t\t\tindexName := index.GetDisplayName()\n\n\t\t\t// items_count captures number of key per index\n\t\t\tkey = fmt.Sprintf(\"%v:%v:items_count\", index.Bucket, indexName)\n\t\t\tif itemsCount, ok := statsMap[key]; ok {\n\t\t\t\tindex.NumOfDocs = uint64(itemsCount.(float64))\n\t\t\t}\n\n\t\t\t// data_size is the total key size of index, excluding back index overhead.\n\t\t\t// Therefore data_size is typically smaller than index sizing equation which\n\t\t\t// includes overhead for back-index.\n\t\t\tkey = fmt.Sprintf(\"%v:%v:data_size\", index.Bucket, indexName)\n\t\t\tif dataSize, ok := statsMap[key]; ok {\n\t\t\t\tindex.ActualMemUsage = uint64(dataSize.(float64))\n\t\t\t\ttotalDataSize += index.ActualMemUsage\n\t\t\t}\n\n\t\t\t// avg_sec_key_size is currently unavailable in 4.5. To estimate,\n\t\t\t// the key size, it divides index data_size by items_count. This\n\t\t\t// contains sec key size + doc key size + main index overhead (74 bytes).\n\t\t\t// Subtract 74 bytes to get sec key size.\n\t\t\tkey = fmt.Sprintf(\"%v:%v:avg_sec_key_size\", index.Bucket, indexName)\n\t\t\tif avgSecKeySize, ok := statsMap[key]; ok {\n\t\t\t\tindex.AvgSecKeySize = uint64(avgSecKeySize.(float64))\n\t\t\t} else if !index.IsPrimary {\n\t\t\t\t// Aproximate AvgSecKeySize. AvgSecKeySize includes both\n\t\t\t\t// sec key len + doc key len\n\t\t\t\tif index.NumOfDocs != 0 && index.ActualMemUsage != 0 {\n\t\t\t\t\tindex.ActualKeySize = index.ActualMemUsage / index.NumOfDocs\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// These stats are currently unavailable in 4.5.\n\t\t\tkey = fmt.Sprintf(\"%v:%v:avg_doc_key_size\", index.Bucket, indexName)\n\t\t\tif avgDocKeySize, ok := statsMap[key]; ok {\n\t\t\t\tindex.AvgDocKeySize = uint64(avgDocKeySize.(float64))\n\t\t\t} else if index.IsPrimary {\n\t\t\t\t// Aproximate AvgDocKeySize. Subtract 74 bytes for main\n\t\t\t\t// index overhead\n\t\t\t\tif index.NumOfDocs != 0 && index.ActualMemUsage != 0 {\n\t\t\t\t\tindex.ActualKeySize = index.ActualMemUsage / index.NumOfDocs\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// These stats are currently unavailable in 4.5.\n\t\t\tkey = fmt.Sprintf(\"%v:%v:avg_arr_size\", index.Bucket, indexName)\n\t\t\tif avgArrSize, ok := statsMap[key]; ok {\n\t\t\t\tindex.AvgArrSize = uint64(avgArrSize.(float64))\n\t\t\t}\n\n\t\t\t// These stats are currently unavailable in 4.5.\n\t\t\tkey = fmt.Sprintf(\"%v:%v:avg_arr_key_size\", index.Bucket, indexName)\n\t\t\tif avgArrKeySize, ok := statsMap[key]; ok {\n\t\t\t\tindex.AvgArrKeySize = uint64(avgArrKeySize.(float64))\n\t\t\t}\n\n\t\t\t// These stats are currently unavailable in 4.5.\n\t\t\tkey = fmt.Sprintf(\"%v:%v:avg_drain_rate\", index.Bucket, indexName)\n\t\t\tif avgMutationRate, ok := statsMap[key]; ok {\n\t\t\t\tindex.MutationRate = uint64(avgMutationRate.(float64))\n\t\t\t\ttotalMutation += index.MutationRate\n\t\t\t} else {\n\t\t\t\tkey = fmt.Sprintf(\"%v:%v:num_flush_queued\", index.Bucket, indexName)\n\t\t\t\tif flushQueuedStat, ok := statsMap[key]; ok {\n\t\t\t\t\tflushQueued := uint64(flushQueuedStat.(float64))\n\n\t\t\t\t\tif flushQueued != 0 {\n\t\t\t\t\t\tindex.MutationRate = flushQueued / elapsed\n\t\t\t\t\t\ttotalMutation += index.MutationRate\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// These stats are currently unavailable in 4.5.\n\t\t\tkey = fmt.Sprintf(\"%v:%v:avg_scan_rate\", index.Bucket, indexName)\n\t\t\tif avgScanRate, ok := statsMap[key]; ok {\n\t\t\t\tindex.ScanRate = uint64(avgScanRate.(float64))\n\t\t\t\ttotalScan += index.ScanRate\n\t\t\t} else {\n\t\t\t\tkey = fmt.Sprintf(\"%v:%v:num_rows_returned\", index.Bucket, indexName)\n\t\t\t\tif rowReturnedStat, ok := statsMap[key]; ok {\n\t\t\t\t\trowReturned := uint64(rowReturnedStat.(float64))\n\n\t\t\t\t\tif rowReturned != 0 {\n\t\t\t\t\t\tindex.ScanRate = rowReturned / elapsed\n\t\t\t\t\t\ttotalScan += index.ScanRate\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\t// compute the estimated memory usage for each index. This also computes\n\t\t// the aggregated indexer mem usage. Mem usage can be 0 if\n\t\t// 1) there is no index stats\n\t\t// 2) index has no data (datasize = 0) (e.g. deferred index)\n\t\tfor _, index := range indexer.Indexes {\n\t\t\tratio := float64(0)\n\t\t\tif totalDataSize != 0 {\n\t\t\t\tratio = float64(index.ActualMemUsage) / float64(totalDataSize)\n\t\t\t}\n\n\t\t\tindex.ActualMemUsage = uint64(float64(actualStorageMem) * ratio)\n\n\t\t\tif actualTotalMem > actualStorageMem {\n\t\t\t\tindex.ActualMemOverhead = uint64(float64(actualTotalMem-actualStorageMem) * ratio)\n\t\t\t} else {\n\t\t\t\tindex.ActualMemOverhead = 0\n\t\t\t}\n\n\t\t\tif index.ActualMemUsage != 0 {\n\t\t\t\tindex.NoUsage = false\n\t\t\t}\n\n\t\t\tindexer.ActualMemUsage += index.ActualMemUsage\n\t\t\tindexer.ActualMemOverhead += index.ActualMemOverhead\n\t\t}\n\n\t\t// compute the estimated cpu usage for each index. This also computes the\n\t\t// aggregated indexer cpu usage. CPU usge can be 0 if\n\t\t// 1) there is no index stats\n\t\t// 2) index has no scan or mutation (e.g. deferred index)\n\t\tfor _, index := range indexer.Indexes {\n\n\t\t\tmutationRatio := float64(0)\n\t\t\tif totalMutation != 0 {\n\t\t\t\tmutationRatio = float64(index.MutationRate) / float64(totalMutation)\n\t\t\t}\n\n\t\t\tscanRatio := float64(0)\n\t\t\tif totalScan != 0 {\n\t\t\t\tscanRatio = float64(index.ScanRate) / float64(totalScan)\n\t\t\t}\n\n\t\t\tratio := mutationRatio\n\t\t\tif scanRatio != 0 {\n\t\t\t\tif mutationRatio != 0 {\n\t\t\t\t\t// mutation uses 5 times less cpu than scan\n\t\t\t\t\tratio = ((mutationRatio / 5) + scanRatio) / 2\n\t\t\t\t} else {\n\t\t\t\t\tratio = scanRatio\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tusage := float64(actualCpuUtil) * ratio\n\n\t\t\tif usage > 0 {\n\t\t\t\tindex.ActualCpuUsage = usage\n\t\t\t\tindex.NoUsage = false\n\t\t\t}\n\n\t\t\tindexer.ActualCpuUsage += index.ActualCpuUsage\n\t\t}\n\t}\n\n\treturn nil\n}", "func indexHandler(w http.ResponseWriter, r *http.Request) {\n\tdata := &Index{\n\t\tTitle: \"Image gallery\",\n\t\tBody: \"Welcome to the image gallery.\",\n\t}\n\tfor name, img := range images {\n\t\tdata.Links = append(data.Links, Link{\n\t\t\tURL: \"/image/\" + name,\n\t\t\tTitle: img.Title,\n\t\t})\n\t}\n\tif err := indexTemplate.Execute(w, data); err != nil {\n\t\tlog.Println(err)\n\t}\n}", "func bookIndex(w http.ResponseWriter, r *http.Request, _ httprouter.Params) {\n\tmtx.Lock()\n\tdefer mtx.Unlock()\n\tbooks := make([]*model.Book, len(bookstore))\n\ti := 0\n\tfor _, v := range bookstore {\n\t\tbooks[i] = v\n\t\ti++\n\t}\n\tres := &common.ResBody{\n\t\tErr: common.OK,\n\t\tData: books,\n\t}\n\tcommon.WriteJson(w, res, http.StatusOK)\n}", "func (d *Dataset) Indexes() []string {\n\t//grab indexes\n\tvar indexes []string\n\n\tindexFiles, err := ioutil.ReadDir(filepath.Join(path.Dir(d.path), \".gitdb/index/\", d.Name()))\n\tif err != nil {\n\t\treturn indexes\n\t}\n\n\tfor _, indexFile := range indexFiles {\n\t\tindexes = append(indexes, strings.TrimSuffix(indexFile.Name(), \".json\"))\n\t}\n\n\treturn indexes\n}", "func LoadImageManifestFromIndex(indexPath string) ([]ImageManifest, error) {\n\tdata, err := ioutil.ReadFile(indexPath)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"read index: %w\", err)\n\t}\n\n\tvar list ManifestList\n\tif err := json.Unmarshal(data, &list); err != nil {\n\t\treturn nil, fmt.Errorf(\"decode manifest: %w\", err)\n\t}\n\n\treturn list.Manifests, nil\n}", "func ForAllIndexes(ctx context.Context, repo restic.Repository,\n\tfn func(id restic.ID, index *Index, oldFormat bool, err error) error) error {\n\n\tdebug.Log(\"Start\")\n\n\ttype FileInfo struct {\n\t\trestic.ID\n\t\tSize int64\n\t}\n\n\tvar m sync.Mutex\n\n\t// track spawned goroutines using wg, create a new context which is\n\t// cancelled as soon as an error occurs.\n\twg, ctx := errgroup.WithContext(ctx)\n\n\tch := make(chan FileInfo)\n\t// send list of index files through ch, which is closed afterwards\n\twg.Go(func() error {\n\t\tdefer close(ch)\n\t\treturn repo.List(ctx, restic.IndexFile, func(id restic.ID, size int64) error {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn nil\n\t\t\tcase ch <- FileInfo{id, size}:\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t})\n\n\t// a worker receives an index ID from ch, loads the index, and sends it to indexCh\n\tworker := func() error {\n\t\tvar buf []byte\n\t\tfor fi := range ch {\n\t\t\tdebug.Log(\"worker got file %v\", fi.ID.Str())\n\t\t\tvar err error\n\t\t\tvar idx *Index\n\t\t\toldFormat := false\n\n\t\t\tbuf, err = repo.LoadAndDecrypt(ctx, buf[:0], restic.IndexFile, fi.ID)\n\t\t\tif err == nil {\n\t\t\t\tidx, oldFormat, err = DecodeIndex(buf, fi.ID)\n\t\t\t}\n\n\t\t\tm.Lock()\n\t\t\terr = fn(fi.ID, idx, oldFormat, err)\n\t\t\tm.Unlock()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t}\n\n\t// run workers on ch\n\twg.Go(func() error {\n\t\treturn RunWorkers(loadIndexParallelism, worker)\n\t})\n\n\treturn wg.Wait()\n}", "func (i IndexFile) sortPackages() {\n\tfor _, versions := range i.Entries {\n\t\tsort.Sort(sort.Reverse(versions))\n\t}\n}", "func (*SearchAllResourcesRequest) Descriptor() ([]byte, []int) {\n\treturn file_google_cloud_asset_v1_asset_service_proto_rawDescGZIP(), []int{22}\n}", "func list_arches(w rest.ResponseWriter, r *rest.Request) {\n\t// Use caching to reduce calls to the Dropbox API\n\tcache_path := \"arches\"\n\tdata, found := cache_instance.Get(cache_path)\n\tif found {\n\t\tif cached, ok := data.([]string); ok {\n\t\t\tw.WriteJson(cached)\n\t\t\treturn\n\t\t} else {\n\t\t\tlog.Println(\"Error: Unable to retrieve from cache\")\n\t\t}\n\t}\n\n\tarches := []string{}\n\tdirectories := get_directories(cache_instance, db, \"/\")\n\tfor _, arch := range directories {\n\t\tarches = append(arches, strings.Replace(arch.Path, \"/\", \"\", -1))\n\t}\n\tcache_instance.Set(cache_path, arches, 0)\n\tw.WriteJson(arches)\n}", "func indexFile(name string) string {\r\n\treturn name + \".files\"\r\n}", "func GetIndexByMediaType(index *imagespec.Index, mt string) (*imagespec.Descriptor, error) {\n\tfor _, d := range index.Manifests {\n\t\tif d.MediaType == mt {\n\t\t\treturn &d, nil\n\t\t}\n\t}\n\treturn nil, ErrMediaTypeNotFound\n}", "func CreateIndexByScanDir(targetDir string, indexFileName string, public_url string) (index Index) {\n\n\tindex = Index{Sitemaps: []Sitemap{}}\n\n\tfs, err := ioutil.ReadDir(targetDir)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tfor _, f := range fs {\n\t\tif strings.HasSuffix(f.Name(), \".xml.gz\") && !strings.HasSuffix(indexFileName, f.Name()) {\n\t\t\tlastModified := f.ModTime()\n\t\t\tindex.Sitemaps = append(index.Sitemaps, Sitemap{Loc: public_url + f.Name(), LastMod: &lastModified})\n\t\t}\n\t}\n\treturn\n}", "func (k8sc *Client) _search(ctx context.Context, groups, categories, kinds, namespaces, versions, names, labels, containers string) ([]SearchResult, error) {\n\n\t// normalize params\n\tgroups = strings.ToLower(groups)\n\tcategories = strings.ToLower(categories)\n\tkinds = strings.ToLower(kinds) // means resources in K8s API term (i.e. pods, services, etc)\n\tnamespaces = strings.ToLower(namespaces)\n\tversions = strings.ToLower(versions)\n\tlabels = strings.ToLower(labels)\n\tcontainers = strings.ToLower(containers)\n\n\tlogrus.Debugf(\n\t\t\"Search filters groups:[%v]; categories:[%v]; kinds:[%v]; namespaces:[%v]; versions:[%v]; names:[%v]; labels:[%v] containers:[%s]\",\n\t\tgroups, categories, kinds, namespaces, versions, names, labels, containers,\n\t)\n\n\t// Build a groups-resource Map that maps each\n\t// selected group to its associated resources.\n\tgroupResMap := make(map[schema.GroupVersion]*metav1.APIResourceList)\n\tswitch {\n\tcase groups == \"\" && kinds == \"\" && versions == \"\" && categories == \"\":\n\t\t// no groups, no kinds (resources), no versions, no categories provided\n\t\treturn nil, fmt.Errorf(\"search: at least one of {groups, kinds, versions, or categories} is required\")\n\tcase groups == \"\" && kinds == \"\" && versions != \"\" && categories == \"\":\n\t\t// only versions provided\n\t\treturn nil, fmt.Errorf(\"search: versions must be provided with at least one of {groups, kinds, or categories}\")\n\tdefault:\n\t\t// build a group-to-resources map, based on the passed parameters.\n\t\t// first, extract groups needed to build the map\n\t\tvar groupList *metav1.APIGroupList\n\t\tif groups != \"\" {\n\t\t\tgroupList = &metav1.APIGroupList{}\n\t\t\tgroupSlice := splitParamList(groups)\n\n\t\t\t// adjust for legacy group name \"core\" -> \"\" empty\n\t\t\tfor i := 0; i < len(groupSlice); i++ {\n\t\t\t\tgroupSlice[i] = toLegacyGrpName(groupSlice[i])\n\t\t\t}\n\n\t\t\tserverGroups, err := k8sc.Disco.ServerGroups()\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"search: failed to get server groups: %w\", err)\n\t\t\t}\n\t\t\t// for each server group, match specified group name from param\n\t\t\tfor _, grp := range serverGroups.Groups {\n\t\t\t\tif sliceContains(groupSlice, grp.Name) {\n\t\t\t\t\tgroupList.Groups = append(groupList.Groups, grp)\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tserverGroups, err := k8sc.Disco.ServerGroups()\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"search: failed to get server groups: %w\", err)\n\t\t\t}\n\t\t\tgroupList = serverGroups\n\t\t}\n\n\t\t// extract resources names (kinds param) and versions params\n\t\tverSlice := splitParamList(versions)\n\t\tresSlice := splitParamList(kinds)\n\t\tcatSlice := splitParamList(categories)\n\n\t\t// next, for each groupVersion pair\n\t\t// retrieve a set of resources associated with it\n\t\tfor _, grp := range groupList.Groups {\n\t\t\tfor _, ver := range grp.Versions {\n\t\t\t\t// only select ver if it can be matched, otherwise continue to next ver\n\t\t\t\tif versions != \"\" && !sliceContains(verSlice, ver.Version) {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\t// grab all available resources for group/ver\n\t\t\t\tgroupVersion := schema.GroupVersion{Group: grp.Name, Version: ver.Version}\n\t\t\t\tresList, err := k8sc.Disco.ServerResourcesForGroupVersion(groupVersion.String())\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, fmt.Errorf(\"search: failed to get resources for groups: %w\", err)\n\t\t\t\t}\n\n\t\t\t\t// for each resource in group/ver\n\t\t\t\t// attempt to match it with provided resources name (kinds)\n\t\t\t\tresultList := &metav1.APIResourceList{GroupVersion: groupVersion.String()}\n\t\t\t\tfor _, resource := range resList.APIResources {\n\t\t\t\t\t// filter resources on names if provided (kinds param)\n\t\t\t\t\tif kinds != \"\" && !sliceContains(resSlice, resource.Kind) && !sliceContains(resSlice, resource.Name) {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\t// filter resources on categories if specified\n\t\t\t\t\tif categories != \"\" && !sliceContains(catSlice, resource.Categories...) {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tresultList.APIResources = append(resultList.APIResources, resource)\n\t\t\t\t}\n\t\t\t\tgroupResMap[groupVersion] = resultList\n\t\t\t}\n\t\t}\n\t}\n\n\t// prepare namespaces\n\tvar nsList []string\n\tif namespaces != \"\" {\n\t\tnsList = splitParamList(namespaces)\n\t} else {\n\t\tnsNames, err := getNamespaces(ctx, k8sc)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tnsList = nsNames\n\t}\n\n\t// Collect resource objects using the grou-to-resources map\n\tvar finalResults []SearchResult\n\tlogrus.Debugf(\"searching through %d groups\", len(groupResMap))\n\tfor groupVer, resourceList := range groupResMap {\n\t\tfor _, resource := range resourceList.APIResources {\n\t\t\tlistOptions := metav1.ListOptions{\n\t\t\t\tLabelSelector: labels,\n\t\t\t}\n\t\t\tgvr := schema.GroupVersionResource{Group: groupVer.Group, Version: groupVer.Version, Resource: resource.Name}\n\t\t\t// gather found resources\n\t\t\tvar results []SearchResult\n\t\t\tif resource.Namespaced {\n\t\t\t\tfor _, ns := range nsList {\n\t\t\t\t\tlogrus.Debugf(\"searching for %s objects in [group=%s; namespace=%s; labels=%v]\",\n\t\t\t\t\t\tresource.Name, groupVer, ns, listOptions.LabelSelector,\n\t\t\t\t\t)\n\t\t\t\t\tlist, err := k8sc.Client.Resource(gvr).Namespace(ns).List(ctx, listOptions)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlogrus.Debugf(\n\t\t\t\t\t\t\t\"WARN: failed to get %s objects in [group=%s; namespace=%s; labels=%v]: %s\",\n\t\t\t\t\t\t\tresource.Name, groupVer, ns, listOptions.LabelSelector, err,\n\t\t\t\t\t\t)\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tif len(list.Items) == 0 {\n\t\t\t\t\t\tlogrus.Debugf(\n\t\t\t\t\t\t\t\"WARN: found 0 %s in [group=%s; namespace=%s; labels=%v]\",\n\t\t\t\t\t\t\tresource.Name, groupVer, ns, listOptions.LabelSelector,\n\t\t\t\t\t\t)\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\n\t\t\t\t\tlogrus.Debugf(\"found %d %s in [group=%s; namespace=%s; labels=%v]\",\n\t\t\t\t\t\tlen(list.Items), resource.Name, groupVer, ns, listOptions.LabelSelector,\n\t\t\t\t\t)\n\t\t\t\t\tresult := SearchResult{\n\t\t\t\t\t\tListKind: list.GetKind(),\n\t\t\t\t\t\tResourceName: resource.Name,\n\t\t\t\t\t\tResourceKind: resource.Kind,\n\t\t\t\t\t\tNamespaced: resource.Namespaced,\n\t\t\t\t\t\tNamespace: ns,\n\t\t\t\t\t\tGroupVersionResource: gvr,\n\t\t\t\t\t\tList: list,\n\t\t\t\t\t}\n\t\t\t\t\tresults = append(results, result)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tlogrus.Debugf(\"searching for %s objects in [group=%s; non-namespced; labels=%v]\",\n\t\t\t\t\tresource.Name, groupVer, listOptions.LabelSelector,\n\t\t\t\t)\n\n\t\t\t\tlist, err := k8sc.Client.Resource(gvr).List(ctx, listOptions)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogrus.Debugf(\n\t\t\t\t\t\t\"WARN: failed to get %s objects in [group=%s; non-namespaced; labels=%v]: %s\",\n\t\t\t\t\t\tresource.Name, groupVer, listOptions.LabelSelector, err,\n\t\t\t\t\t)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif len(list.Items) == 0 {\n\t\t\t\t\tlogrus.Debugf(\n\t\t\t\t\t\t\"WARN: found 0 %s in [group=%s; non-namespaced; labels=%v]\",\n\t\t\t\t\t\tresource.Name, groupVer, listOptions.LabelSelector,\n\t\t\t\t\t)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tlogrus.Debugf(\"found %d %s in [group=%s; non-namespaced; labels=%v]\",\n\t\t\t\t\tlen(list.Items), resource.Name, groupVer, listOptions.LabelSelector,\n\t\t\t\t)\n\n\t\t\t\tresult := SearchResult{\n\t\t\t\t\tListKind: list.GetKind(),\n\t\t\t\t\tResourceKind: resource.Kind,\n\t\t\t\t\tResourceName: resource.Name,\n\t\t\t\t\tNamespaced: resource.Namespaced,\n\t\t\t\t\tGroupVersionResource: gvr,\n\t\t\t\t\tList: list,\n\t\t\t\t}\n\t\t\t\tresults = append(results, result)\n\t\t\t}\n\n\t\t\t// apply name filters\n\t\t\tlogrus.Debugf(\"applying filters on %d results\", len(results))\n\t\t\tfor _, result := range results {\n\t\t\t\tfilteredResult := result\n\t\t\t\tif len(containers) > 0 && result.ListKind == \"PodList\" {\n\t\t\t\t\tfilteredResult = filterPodsByContainers(result, containers)\n\t\t\t\t\tlogrus.Debugf(\"found %d %s with container filter [%s]\", len(filteredResult.List.Items), filteredResult.ResourceName, containers)\n\t\t\t\t}\n\t\t\t\tif len(names) > 0 {\n\t\t\t\t\tfilteredResult = filterByNames(result, names)\n\t\t\t\t\tlogrus.Debugf(\"found %d %s with name filter [%s]\", len(filteredResult.List.Items), filteredResult.ResourceName, names)\n\t\t\t\t}\n\t\t\t\tfinalResults = append(finalResults, filteredResult)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn finalResults, nil\n}", "func getIndexContent(name string) string {\n\treturn fmt.Sprintf(`export { default } from \"./%s\";`, name)\n}", "func (epi *epInfo) resources(portFilter *filter.RegexFilter, l *logger.Logger) (resources []*pb.Resource) {\n\tfor _, eps := range epi.Subsets {\n\t\t// There is usually one port, but there can be multiple ports, e.g. 9313\n\t\t// and 9314.\n\t\tfor _, port := range eps.Ports {\n\t\t\t// For unnamed ports, use port number.\n\t\t\tportName := port.Name\n\t\t\tif portName == \"\" {\n\t\t\t\tportName = strconv.FormatInt(int64(port.Port), 10)\n\t\t\t}\n\n\t\t\tif portFilter != nil && !portFilter.Match(portName, l) {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tfor _, addr := range eps.Addresses {\n\t\t\t\t// We name the resource as <endpoints_name>_<IP>_<port>\n\t\t\t\tresName := fmt.Sprintf(\"%s_%s_%s\", epi.Metadata.Name, addr.IP, portName)\n\t\t\t\tresources = append(resources, &pb.Resource{\n\t\t\t\t\tName: proto.String(resName),\n\t\t\t\t\tIp: proto.String(addr.IP),\n\t\t\t\t\tPort: proto.Int(port.Port),\n\t\t\t\t\tLabels: epi.Metadata.Labels,\n\t\t\t\t})\n\t\t\t}\n\t\t}\n\t}\n\treturn\n}", "func templatesIndexTmpl() (*asset, error) {\n\tpath := \"/Volumes/Code/go/src/github.com/schollz/cowyo/templates/index.tmpl\"\n\tname := \"templates/index.tmpl\"\n\tbytes, err := bindataRead(path, name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfi, err := os.Stat(path)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"Error reading asset info %s at %s: %v\", name, path, err)\n\t}\n\n\ta := &asset{bytes: bytes, info: fi}\n\treturn a, err\n}", "func (m MatchedResources) Query(client *ApiClient, namespace string) (map[string][]byte, error) {\n\tvar err error\n\tresult := make(map[string][]byte)\n\n\tfor res, object := range m.WantRes{\n\t\tfor namespaced, g := range m.Gvr {\n\t\t\tfor _, gvr := range g {\n\t\t\t\tif res != gvr.Resource {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tj := 0\n\t\t\t\tfor _, k := range object{\n\t\t\t\t\tif namespaced {\n\t\t\t\t\t\tf, err = client.DynClient.Resource(gvr).Namespace(namespace).Get(context.TODO(), k, metav1.GetOptions{})\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\treturn nil, err\n\t\t\t\t\t\t}\n\t\t\t\t\t} else {\n\t\t\t\t\t\tf, err = client.DynClient.Resource(gvr).Get(context.TODO(), k, metav1.GetOptions{})\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\treturn nil, err\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\terr = deepCleaning(f)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn nil, err\n\t\t\t\t\t}\n\n\t\t\t\t\toutput, err := utils.GetPrettyYaml(f)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn nil, err\n\t\t\t\t\t}\n\t\t\t\t\tresult[fmt.Sprintf(\"%s-%v\", res, j)] = output\n\n\t\t\t\t\tj++\n\t\t\t\t}\n\t\t\t}\n\n\t\t}\n\t}\n\treturn result, nil\n\n}", "func (d *docsIndexer) indexFile(p *packages.Package, f *ast.File, fileName string, isTestFile bool) (fileDocs, error) {\n\tvar result fileDocs\n\tresult.pkgDocsMarkdown = godocToMarkdown(f.Doc.Text())\n\n\t// Collect each top-level declaration.\n\tvar initIndex int = 1\n\tfor _, decl := range f.Decls {\n\t\tswitch node := decl.(type) {\n\t\tcase *ast.GenDecl:\n\t\t\tgenDeclDocs := d.indexGenDecl(p, f, node, isTestFile)\n\t\t\tresult.consts = append(result.consts, genDeclDocs.consts...)\n\t\t\tresult.vars = append(result.vars, genDeclDocs.vars...)\n\t\t\tresult.types = append(result.types, genDeclDocs.types...)\n\t\tcase *ast.FuncDecl:\n\t\t\t// Functions, methods\n\t\t\tif node.Name.Name == \"_\" {\n\t\t\t\t// Not only is it not exported, it cannot be referenced outside this package at all.\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tresult.funcs = append(result.funcs, d.indexFuncDecl(p.Fset, p, node, fileName, &initIndex, isTestFile))\n\t\t}\n\t}\n\n\t// Emit documentation for all constants.\n\tfor i, constDocs := range result.consts {\n\t\temittedID := constDocs.result().emit(d.i.emitter)\n\t\tconstDocs.ID = emittedID\n\t\tresult.consts[i] = constDocs\n\t}\n\n\t// Emit documentation for all variables.\n\tfor i, varDocs := range result.vars {\n\t\temittedID := varDocs.result().emit(d.i.emitter)\n\t\tvarDocs.ID = emittedID\n\t\tresult.vars[i] = varDocs\n\t}\n\n\t// Emit documentation for all types (struct/interface/other type definitions)\n\tfor i, typeDocs := range result.types {\n\t\temittedID := typeDocs.result().emit(d.i.emitter)\n\t\ttypeDocs.ID = emittedID\n\t\tresult.types[i] = typeDocs\n\t}\n\n\t// Emit documentation for all funcs/methods.\n\tfor i, funcDocs := range result.funcs {\n\t\temittedID := funcDocs.result().emit(d.i.emitter)\n\t\tfuncDocs.ID = emittedID\n\t\tresult.funcs[i] = funcDocs\n\t}\n\treturn result, nil\n}", "func writeIndexEntries(p []*post, o string, t postListing) error {\n\tvar m string\n\tswitch t {\n\tcase index:\n\t\tm = \"index_template.html\"\n\tcase rss:\n\t\tm = \"rss_template.rss\"\n\tcase archive:\n\t\tm = \"archive_template.html\"\n\t}\n\te, err := template.ParseFiles(filepath.Join(templatesrc, m))\n\tif checkError(err) {\n\t\treturn err\n\t}\n\n\tf, err := os.Create(o)\n\tif checkError(err) {\n\t\treturn err\n\t}\n\terr = e.Execute(f, p)\n\tif checkError(err) {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (m *CompositeIndexer) Map(resources []models.Resource) []solr.Document {\n\tdocs := make([]solr.Document, len(resources))\n\tfor i, v := range resources {\n\t\tif v == nil {\n\t\t\tcontinue\n\t\t}\n\t\tdocs[i] = m.mapOne(v)\n\t}\n\treturn docs\n}", "func (s pipelineResourceNamespaceLister) List(selector labels.Selector) (ret []*v1alpha1.PipelineResource, err error) {\n\terr = cache.ListAllByNamespace(s.indexer, s.namespace, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1alpha1.PipelineResource))\n\t})\n\treturn ret, err\n}", "func IndexHandler(plugins map[string]*tools.Plugin, db *services.Database) httprouter.Handle {\n\treturn func(w http.ResponseWriter, r *http.Request, _ httprouter.Params) {\n\t\tqueryValues := r.URL.Query()\n\t\tquery := make(map[string]interface{})\n\t\tfor k, v := range queryValues {\n\t\t\tquery[k] = strings.Join(v, \"\")\n\t\t}\n\n\t\tdata := make(map[string]interface{})\n\t\tif query[\"provider\"] == nil {\n\t\t\tdata[\"provider\"] = \"\"\n\t\t} else {\n\t\t\tdata[\"provider\"] = query[\"provider\"]\n\t\t}\n\n\t\tif query[\"provider\"] == \"\" {\n\t\t\tdelete(query, \"provider\")\n\t\t}\n\n\t\tif query[\"table\"] == nil {\n\t\t\tquery[\"table\"] = \"images\"\n\t\t}\n\t\tdata[\"table\"] = query[\"table\"]\n\n\t\tcount, err := db.CountEntries(query)\n\t\tif err != nil {\n\t\t\tlog.Println(err)\n\t\t}\n\t\tlog.Printf(\"query %s has %d elements\\n\", query, count)\n\t\tdata[\"count\"] = count\n\n\t\tshown := services.MaxValues\n\t\tif count < shown {\n\t\t\tshown = count\n\t\t}\n\t\tdata[\"shown\"] = shown\n\n\t\tlog.Printf(\"render results for %s\\n\", query)\n\t\tresults, err := db.ReadEntries(query)\n\t\tif err != nil {\n\t\t\tlog.Println(err)\n\t\t}\n\n\t\telements := make([]string, 0)\n\t\tvar element map[string]interface{}\n\t\tfor id := range results {\n\t\t\telement = results[id].(map[string]interface{})\n\t\t\tprovider := element[\"provider\"].(string)\n\t\t\tplugin := plugins[provider]\n\t\t\trender, err := plugin.Present(element, \"\")\n\t\t\tif err == nil {\n\t\t\t\telements = append(elements, render)\n\t\t\t} else {\n\t\t\t\tlog.Println(err)\n\t\t\t}\n\t\t}\n\t\tdata[\"elements\"] = elements\n\t\trenderTemplate(w, \"index.html\", data)\n\t}\n}", "func sortResources(resources []astmodel.InternalTypeName) []astmodel.InternalTypeName {\n\tsort.Slice(resources, func(i, j int) bool {\n\t\tiVal := resources[i]\n\t\tjVal := resources[j]\n\n\t\treturn iVal.PackageReference().PackageName() < jVal.PackageReference().PackageName() ||\n\t\t\tiVal.PackageReference().PackageName() < jVal.PackageReference().PackageName() && iVal.Name() < jVal.Name()\n\t})\n\n\treturn resources\n}", "func InitSearchIndex() {\n\tfor t := range item.Types {\n\t\terr := search.MapIndex(t)\n\t\tif err != nil {\n\t\t\tlog.Fatalln(err)\n\t\t\treturn\n\t\t}\n\t\tSortContent(t)\n\t}\n}", "func (o Model) RebuildIndexes(pattern string) error {\n\t// Quick exit in case no index exists\n\tif o.IndexSet == nil || len(o.IndexSet.Indexes) == 0 {\n\t\treturn nil\n\t}\n\n\tp := res.Pattern(pattern)\n\tif !p.IsValid() {\n\t\treturn errors.New(\"invalid pattern\")\n\t}\n\n\t// Drop existing index entries\n\tfor _, idx := range o.IndexSet.Indexes {\n\t\terr := o.BadgerDB.DB.DropPrefix([]byte(idx.Name))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tt := reflect.TypeOf(o.Type)\n\n\t// Create a prefix to seek from\n\tridPrefix := pattern\n\ti := p.IndexWildcard()\n\tif i >= 0 {\n\t\tridPrefix = pattern[:i]\n\t}\n\n\t// Create new index entries in a single transaction\n\treturn o.BadgerDB.DB.Update(func(txn *badger.Txn) error {\n\t\tit := txn.NewIterator(badger.DefaultIteratorOptions)\n\t\tdefer it.Close()\n\t\tprefix := []byte(ridPrefix)\n\t\tfor it.Seek(prefix); it.ValidForPrefix(prefix); it.Next() {\n\t\t\t// Ensure the key matches the pattern\n\t\t\tif !p.Matches(string(it.Item().Key())) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// Load item and unmarshal it\n\t\t\titem := it.Item()\n\t\t\tv := reflect.New(t)\n\t\t\terr := item.Value(func(dta []byte) error {\n\t\t\t\treturn json.Unmarshal(dta, v.Interface())\n\t\t\t})\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\t// Loop through indexes and generate a new entry per index\n\t\t\tfor _, idx := range o.IndexSet.Indexes {\n\t\t\t\trname := item.KeyCopy(nil)\n\t\t\t\tidxKey := idx.getKey(rname, idx.Key(v.Elem().Interface()))\n\t\t\t\terr = txn.SetEntry(&badger.Entry{Key: idxKey, Value: nil, UserMeta: typeIndex})\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t})\n}", "func Main(adds []string) j.Type {\n\tindex := j.Import(\"\", IndexFile)\n\tif len(adds) == 0 {\n\t\treturn index\n\t}\n\telems := []j.Type{index}\n\tfor _, a := range adds {\n\t\ta = filepath.Join(CustomPrefix, filepath.Base(a))\n\t\tadd := j.Import(\"\", a)\n\t\telems = append(elems, add)\n\t}\n\n\treturn j.Add(\"\", elems...)\n}", "func FindPhysicalIndexes(url string, alias string) []string {\n\tindexResponse := infoResponse{}\n\t_, err := MakeJSONRequest(http.MethodGet, fmt.Sprintf(\"%s/%s\", url, alias), \"\", &indexResponse)\n\tindexes := make([]string, 0)\n\n\t// error could mean a variety of things, but we'll figure that out later\n\tif err != nil {\n\t\treturn indexes\n\t}\n\n\t// our top level key is our physical index name\n\tfor key := range indexResponse {\n\t\tindexes = append(indexes, key)\n\t}\n\n\t// reverse sort order should put our newest index first\n\tsort.Sort(sort.Reverse(sort.StringSlice(indexes)))\n\treturn indexes\n}", "func index(w http.ResponseWriter, r *http.Request) {\n\tfmt.Fprint(w, \"index de uma função\")\n}", "func (app *application) makeSearchIndexMap() *search.IndexSlice {\r\n\tindSlice := search.IndexSlice{map[string][]int{}, map[string][]int{}, map[string][]int{}}\r\n\tproductForSearch, err := app.products.GetSearchProducts()\r\n\tif err != nil {\r\n\t\tapp.errorLog.Println(err)\r\n\t\treturn nil\r\n\t}\r\n\tindSlice.Add(productForSearch)\r\n\treturn &indSlice\r\n}", "func (as *API) Index(ctx context.Context, req *pbreq.Index) (*pbresp.Index, error) {\n\tswitch req.GetType() {\n\tcase \"ipld\":\n\t\tbreak\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"invalid data type '%s'\", req.GetType())\n\t}\n\n\tvar name = req.GetIdentifier()\n\tvar reindex = req.GetReindex()\n\tmetaData, err := as.lens.Magnify(name, reindex)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to perform indexing for '%s': %s\",\n\t\t\tname, err.Error())\n\t}\n\n\tvar resp *lens.Object\n\tif !reindex {\n\t\tif resp, err = as.lens.Store(name, metaData); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\tb, err := as.lens.Get(name)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to find ID for object '%s'\", name)\n\t\t}\n\t\tid, err := uuid.FromBytes(b)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"invalid uuid found for '%s' ('%s'): %s\",\n\t\t\t\tname, string(b), err.Error())\n\t\t}\n\t\tif resp, err = as.lens.Update(id, name, metaData); err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to update object: %s\", err.Error())\n\t\t}\n\t}\n\n\treturn &pbresp.Index{\n\t\tId: resp.LensID.String(),\n\t\tKeywords: metaData.Summary,\n\t}, nil\n}", "func updateIndex(indexName string, objects []algoliasearch.Object) error {\n\n\tindex := algoliaClient.InitIndex(indexName)\n\terr := populateIndex(index, objects)\n\tif err != nil {\n\t\treturn errors.New(\"Error updating index -\" + err.Error())\n\t}\n\n\treturn nil\n}", "func (db *Database) Indexes(label string) ([]*Index, error) {\n\turi := join(db.Url, \"schema/index\", label)\n\tresult := []*Index{}\n\tne := NeoError{}\n\tresp, err := db.Session.Get(uri, nil, &result, &ne)\n\tif err != nil {\n\t\treturn result, err\n\t}\n\tif resp.Status() == 404 {\n\t\treturn result, NotFound\n\t}\n\tif resp.Status() != 200 {\n\t\treturn result, ne\n\t}\n\tfor _, idx := range result {\n\t\tidx.db = db\n\t}\n\treturn result, nil\n}", "func (s *server) handleIndex(FSS fs.FS) http.HandlerFunc {\n\ttype AppConfig struct {\n\t\tAvatarService string\n\t\tToastTimeout int\n\t\tAllowGuests bool\n\t\tAllowRegistration bool\n\t\tDefaultLocale string\n\t\tAuthMethod string\n\t\tAppVersion string\n\t\tCookieName string\n\t\tPathPrefix string\n\t\tAPIEnabled bool\n\t\tCleanupGuestsDaysOld int\n\t\tCleanupStoryboardsDaysOld int\n\t\tShowActiveCountries bool\n\t}\n\ttype UIConfig struct {\n\t\tAnalyticsEnabled bool\n\t\tAnalyticsID string\n\t\tAppConfig AppConfig\n\t\tActiveAlerts []interface{}\n\t}\n\n\ttmpl := s.getIndexTemplate(FSS)\n\n\tappConfig := AppConfig{\n\t\tAvatarService: viper.GetString(\"config.avatar_service\"),\n\t\tToastTimeout: viper.GetInt(\"config.toast_timeout\"),\n\t\tAllowGuests: viper.GetBool(\"config.allow_guests\"),\n\t\tAllowRegistration: viper.GetBool(\"config.allow_registration\") && viper.GetString(\"auth.method\") == \"normal\",\n\t\tDefaultLocale: viper.GetString(\"config.default_locale\"),\n\t\tAuthMethod: viper.GetString(\"auth.method\"),\n\t\tAPIEnabled: viper.GetBool(\"config.allow_external_api\"),\n\t\tAppVersion: s.config.Version,\n\t\tCookieName: s.config.FrontendCookieName,\n\t\tPathPrefix: s.config.PathPrefix,\n\t\tCleanupGuestsDaysOld: viper.GetInt(\"config.cleanup_guests_days_old\"),\n\t\tCleanupStoryboardsDaysOld: viper.GetInt(\"config.cleanup_storyboards_days_old\"),\n\t\tShowActiveCountries: viper.GetBool(\"config.show_active_countries\"),\n\t}\n\n\tActiveAlerts = s.database.GetActiveAlerts()\n\n\tdata := UIConfig{\n\t\tAnalyticsEnabled: s.config.AnalyticsEnabled,\n\t\tAnalyticsID: s.config.AnalyticsID,\n\t\tAppConfig: appConfig,\n\t}\n\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tdata.ActiveAlerts = ActiveAlerts // get latest alerts from memory\n\n\t\tif embedUseOS {\n\t\t\ttmpl = s.getIndexTemplate(FSS)\n\t\t}\n\n\t\ttmpl.Execute(w, data)\n\t}\n}", "func getPackageRefs(resources []astmodel.InternalTypeName) []astmodel.PackageReference {\n\tpackageRefs := make([]astmodel.PackageReference, 0, len(resources)+1)\n\t// Package reference for return type\n\tpackageRefs = append(packageRefs, astmodel.KubernetesResourceType.PackageReference())\n\n\tfor _, typeDef := range resources {\n\t\tpackageRefs = append(packageRefs, typeDef.PackageReference())\n\t}\n\n\treturn packageRefs\n}", "func serveIndex(w http.ResponseWriter, r *http.Request, bs buildSpec, br *buildResult) {\n\txreq := request{bs, \"\", pageIndex}\n\txlink := xreq.link()\n\n\ttype versionLink struct {\n\t\tVersion string\n\t\tURLPath string\n\t\tSuccess bool\n\t\tActive bool\n\t}\n\ttype response struct {\n\t\tErr error\n\t\tLatestVersion string\n\t\tVersionLinks []versionLink\n\t}\n\n\t// Do a lookup to the goproxy in the background, to list the module versions.\n\tc := make(chan response, 1)\n\tgo func() {\n\t\tt0 := time.Now()\n\t\tdefer func() {\n\t\t\tmetricGoproxyListDuration.Observe(time.Since(t0).Seconds())\n\t\t}()\n\n\t\tmodPath, err := module.EscapePath(bs.Mod)\n\t\tif err != nil {\n\t\t\tc <- response{fmt.Errorf(\"bad module path: %v\", err), \"\", nil}\n\t\t\treturn\n\t\t}\n\t\tu := fmt.Sprintf(\"%s%s/@v/list\", config.GoProxy, modPath)\n\t\tmreq, err := http.NewRequestWithContext(r.Context(), \"GET\", u, nil)\n\t\tif err != nil {\n\t\t\tc <- response{fmt.Errorf(\"%w: preparing new http request: %v\", errServer, err), \"\", nil}\n\t\t\treturn\n\t\t}\n\t\tmreq.Header.Set(\"User-Agent\", userAgent)\n\t\tresp, err := http.DefaultClient.Do(mreq)\n\t\tif err != nil {\n\t\t\tc <- response{fmt.Errorf(\"%w: http request: %v\", errServer, err), \"\", nil}\n\t\t\treturn\n\t\t}\n\t\tdefer resp.Body.Close()\n\t\tif resp.StatusCode != 200 {\n\t\t\tmetricGoproxyListErrors.WithLabelValues(fmt.Sprintf(\"%d\", resp.StatusCode)).Inc()\n\t\t\tc <- response{fmt.Errorf(\"%w: http response from goproxy: %v\", errRemote, resp.Status), \"\", nil}\n\t\t\treturn\n\t\t}\n\t\tbuf, err := io.ReadAll(resp.Body)\n\t\tif err != nil {\n\t\t\tc <- response{fmt.Errorf(\"%w: reading versions from goproxy: %v\", errRemote, err), \"\", nil}\n\t\t\treturn\n\t\t}\n\t\tl := []versionLink{}\n\t\tfor _, s := range strings.Split(string(buf), \"\\n\") {\n\t\t\tif s != \"\" {\n\t\t\t\tvbs := bs\n\t\t\t\tvbs.Version = s\n\t\t\t\tsuccess := fileExists(filepath.Join(vbs.storeDir(), \"recordnumber\"))\n\t\t\t\tp := request{vbs, \"\", pageIndex}.link()\n\t\t\t\tlink := versionLink{s, p, success, p == xlink}\n\t\t\t\tl = append(l, link)\n\t\t\t}\n\t\t}\n\t\tsort.Slice(l, func(i, j int) bool {\n\t\t\treturn semver.Compare(l[i].Version, l[j].Version) > 0\n\t\t})\n\t\tvar latestVersion string\n\t\tif len(l) > 0 {\n\t\t\tlatestVersion = l[0].Version\n\t\t}\n\t\tc <- response{nil, latestVersion, l}\n\t}()\n\n\t// Non-emptiness means we'll serve the error page instead of doing a SSE request for events.\n\tvar output string\n\tif br == nil {\n\t\tif buf, err := readGzipFile(filepath.Join(bs.storeDir(), \"log.gz\")); err != nil {\n\t\t\tif !os.IsNotExist(err) {\n\t\t\t\tfailf(w, \"%w: reading log.gz: %v\", errServer, err)\n\t\t\t\treturn\n\t\t\t}\n\t\t\t// For not-exist, we'll continue below to build.\n\t\t} else {\n\t\t\toutput = string(buf)\n\t\t}\n\t}\n\n\t// Construct links to other goversions, targets.\n\ttype goversionLink struct {\n\t\tGoversion string\n\t\tURLPath string\n\t\tSuccess bool\n\t\tSupported bool\n\t\tActive bool\n\t}\n\tgoversionLinks := []goversionLink{}\n\tnewestAllowed, supported, remaining := installedSDK()\n\tfor _, goversion := range supported {\n\t\tgvbs := bs\n\t\tgvbs.Goversion = goversion\n\t\tsuccess := fileExists(filepath.Join(gvbs.storeDir(), \"recordnumber\"))\n\t\tp := request{gvbs, \"\", pageIndex}.link()\n\t\tgoversionLinks = append(goversionLinks, goversionLink{goversion, p, success, true, p == xlink})\n\t}\n\tfor _, goversion := range remaining {\n\t\tgvbs := bs\n\t\tgvbs.Goversion = goversion\n\t\tsuccess := fileExists(filepath.Join(gvbs.storeDir(), \"recordnumber\"))\n\t\tp := request{gvbs, \"\", pageIndex}.link()\n\t\tgoversionLinks = append(goversionLinks, goversionLink{goversion, p, success, false, p == xlink})\n\t}\n\n\ttype targetLink struct {\n\t\tGoos string\n\t\tGoarch string\n\t\tURLPath string\n\t\tSuccess bool\n\t\tActive bool\n\t}\n\ttargetLinks := []targetLink{}\n\tfor _, target := range targets.get() {\n\t\ttbs := bs\n\t\ttbs.Goos = target.Goos\n\t\ttbs.Goarch = target.Goarch\n\t\tsuccess := fileExists(filepath.Join(tbs.storeDir(), \"recordnumber\"))\n\t\tp := request{tbs, \"\", pageIndex}.link()\n\t\ttargetLinks = append(targetLinks, targetLink{target.Goos, target.Goarch, p, success, p == xlink})\n\t}\n\n\ttype variantLink struct {\n\t\tVariant string // \"default\" or \"stripped\"\n\t\tTitle string // Displayed on hover in UI.\n\t\tURLPath string\n\t\tSuccess bool\n\t\tActive bool\n\t}\n\tvar variantLinks []variantLink\n\taddVariant := func(v, title string, stripped bool) {\n\t\tvbs := bs\n\t\tvbs.Stripped = stripped\n\t\tsuccess := fileExists(filepath.Join(vbs.storeDir(), \"recordnumber\"))\n\t\tp := request{vbs, \"\", pageIndex}.link()\n\t\tvariantLinks = append(variantLinks, variantLink{v, title, p, success, p == xlink})\n\t}\n\taddVariant(\"default\", \"\", false)\n\taddVariant(\"stripped\", \"Symbol table and debug information stripped, reducing binary size.\", true)\n\n\tpkgGoDevURL := \"https://pkg.go.dev/\" + path.Join(bs.Mod+\"@\"+bs.Version, bs.Dir[1:]) + \"?tab=doc\"\n\n\tresp := <-c\n\n\tvar filesizeGz string\n\tif br == nil {\n\t\tbr = &buildResult{buildSpec: bs}\n\t} else {\n\t\tif info, err := os.Stat(filepath.Join(bs.storeDir(), \"binary.gz\")); err == nil {\n\t\t\tfilesizeGz = fmt.Sprintf(\"%.1f MB\", float64(info.Size())/(1024*1024))\n\t\t}\n\t}\n\n\tprependDir := xreq.Dir\n\tif prependDir == \"/\" {\n\t\tprependDir = \"\"\n\t}\n\n\tvar newerText, newerURL string\n\tif xreq.Goversion != newestAllowed && newestAllowed != \"\" && xreq.Version != resp.LatestVersion && resp.LatestVersion != \"\" {\n\t\tnewerText = \"A newer version of both this module and the Go toolchain is available\"\n\t} else if xreq.Version != resp.LatestVersion && resp.LatestVersion != \"\" {\n\t\tnewerText = \"A newer version of this module is available\"\n\t} else if xreq.Goversion != newestAllowed && newestAllowed != \"\" {\n\t\tnewerText = \"A newer Go toolchain version is available\"\n\t}\n\tif newerText != \"\" {\n\t\tnbs := bs\n\t\tnbs.Version = resp.LatestVersion\n\t\tnbs.Goversion = newestAllowed\n\t\tnewerURL = request{nbs, \"\", pageIndex}.link()\n\t}\n\n\tfavicon := \"/favicon.ico\"\n\tif output != \"\" {\n\t\tfavicon = \"/favicon-error.png\"\n\t} else if br.Sum == \"\" {\n\t\tfavicon = \"/favicon-building.png\"\n\t}\n\targs := map[string]interface{}{\n\t\t\"Favicon\": favicon,\n\t\t\"Success\": br.Sum != \"\",\n\t\t\"Sum\": br.Sum,\n\t\t\"Req\": xreq, // eg \"/\" or \"/cmd/x\"\n\t\t\"DirAppend\": xreq.appendDir(), // eg \"\" or \"cmd/x/\"\n\t\t\"DirPrepend\": prependDir, // eg \"\" or /cmd/x\"\n\t\t\"GoversionLinks\": goversionLinks,\n\t\t\"TargetLinks\": targetLinks,\n\t\t\"VariantLinks\": variantLinks,\n\t\t\"Mod\": resp,\n\t\t\"GoProxy\": config.GoProxy,\n\t\t\"DownloadFilename\": xreq.downloadFilename(),\n\t\t\"PkgGoDevURL\": pkgGoDevURL,\n\t\t\"GobuildVersion\": gobuildVersion,\n\t\t\"GobuildPlatform\": gobuildPlatform,\n\t\t\"VerifierKey\": config.VerifierKey,\n\t\t\"GobuildsOrgVerifierKey\": gobuildsOrgVerifierKey,\n\t\t\"NewerText\": newerText,\n\t\t\"NewerURL\": newerURL,\n\n\t\t// Whether we will do SSE request for updates.\n\t\t\"InProgress\": br.Sum == \"\" && output == \"\",\n\n\t\t// Non-empty on failure.\n\t\t\"Output\": output,\n\n\t\t// Below only meaningful when \"success\".\n\t\t\"Filesize\": fmt.Sprintf(\"%.1f MB\", float64(br.Filesize)/(1024*1024)),\n\t\t\"FilesizeGz\": filesizeGz,\n\t}\n\n\tif br.Sum == \"\" {\n\t\tw.Header().Set(\"Cache-Control\", \"no-store\")\n\t}\n\n\tif err := buildTemplate.Execute(w, args); err != nil {\n\t\tfailf(w, \"%w: executing template: %v\", errServer, err)\n\t}\n}", "func (a *APIGen) Indices(ctx context.Context, filter string) ([]Index, error) {\n\tpanic(\"Should Not Be Called from Gen Pattern.\")\n}", "func loadIndex(ctx context.Context, repo restic.Repository, id restic.ID) (*index.Index, error) {\n\tbuf, err := repo.LoadUnpacked(ctx, restic.IndexFile, id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tidx, oldFormat, err := index.DecodeIndex(buf, id)\n\tif oldFormat {\n\t\tfmt.Fprintf(os.Stderr, \"index %v has old format\\n\", id.Str())\n\t}\n\treturn idx, err\n}", "func Index() (int, error) {\n\t\tfmt.Println(\"Loading resources\")\n\t\terr := InitDB()\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t\tcoursesHTML, _ := ioutil.ReadFile(config.Local.DefaultFN)\n\t\tdoc := soup.HTMLParse(string(coursesHTML))\n\t\ttables := doc.FindAll(\"table\", \"class\", \"datadisplaytable\")\n\t\tregistrar := start(tables)\n\t\tif config.CatSecret != nil {\n\t\t\t\tcat := handleCatalog()\n\t\t\t\tindexCatalog(cat, registrar)\n\t\t}\n\t\tCommit(registrar)\n\t\treturn 0, nil\n}", "func (sqliteCtx *SqliteCtx) BookIndex(w http.ResponseWriter, r *http.Request) {\n\tquery := fmt.Sprintf(\"SELECT * FROM %s\", tableName)\n\trows, err := sqliteCtx.db.Query(query)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tdefer rows.Close()\n\tbook := new(model.Book)\n\tscannedBooks := []*model.Book{}\n\tfor rows.Next() {\n\t\terr = rows.Scan(&book.ID, &book.Title, &book.Genres,\n\t\t\t&book.Pages, &book.Price)\n\t\tif err != nil {\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tscannedBooks = append(scannedBooks, book)\n\t}\n\tjsonScannedBooks, err := json.MarshalIndent(scannedBooks, \"\", \" \")\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\tw.Write(jsonScannedBooks)\n}", "func (g genericPlugin) List(gvk schema.GroupVersionKind, namespace string,\n\tclient plugin.KubernetesConnector) ([]helm.KubernetesResource, error) {\n\n\tvar returnData []helm.KubernetesResource\n\treturn returnData, nil\n}", "func ImageIndexSchema() *gojsonschema.Schema {\n\treturn loadSchema(\"image-index.schema.json\")\n}", "func (m *Entry) SearchData() *EntryIndex {\n\ttags := []string{}\n\tif m.Tags != nil {\n\t\tfor _, tag := range m.Tags {\n\t\t\ttags = append(tags, tag.Name)\n\t\t}\n\t}\n\n\timages := []string{}\n\tif m.Images != nil && !funcs.IsImgFallback(m.Blog.Url) {\n\t\tfor _, image := range m.Images {\n\t\t\timages = append(images, image.Src)\n\t\t}\n\t}\n\tif len(images) <= 0 {\n\t\timages = append(images, image.CachedRandomSrc(\"large\"))\n\t}\n\n\tidx := &EntryIndex{\n\t\tEntry: *m,\n\t\tTags: tags,\n\t\tImages: images,\n\t}\n\n\tif m.Blog != nil {\n\t\tidx.BlogName = m.Blog.Name\n\t\tidx.BlogMediatype = m.Blog.Mediatype\n\t\tidx.BlogAdsensetype = m.Blog.Adsensetype\n\t}\n\n\tif m.Video != nil {\n\t\tm.Video.LoadRelated()\n\n\t\tif m.Video.Divas != nil {\n\t\t\tfor _, diva := range m.Video.Divas {\n\t\t\t\tidx.VideoDivas = append(idx.VideoDivas, diva.Name)\n\t\t\t\tidx.VideoBracups = append(idx.VideoBracups, diva.Bracup)\n\t\t\t}\n\t\t}\n\n\t\tif m.Video.Site != nil {\n\t\t\tidx.VideoDomain = m.Video.Site.Domain\n\t\t}\n\n\t\tidx.VideoDuration = m.Video.Duration\n\t}\n\n\tif m.Picture != nil {\n\t\tm.Picture.LoadRelated()\n\n\t\tfor _, c := range m.Picture.Characters {\n\t\t\tidx.PictureCharacters = append(idx.PictureCharacters, c.Name)\n\t\t\tidx.PictureBracups = append(idx.PictureBracups, c.Bracup)\n\t\t}\n\n\t\tif m.Picture.Anime != nil {\n\t\t\tidx.PictureAnime = m.Picture.Anime.Name\n\t\t\tidx.PictureAlias = m.Picture.Anime.Alias\n\t\t\tidx.PictureAuthor = m.Picture.Anime.Author\n\t\t\tidx.PictureWorks = m.Picture.Anime.Works\n\t\t}\n\t}\n\n\tvar (\n\t\tn string\n\t\ts *Score\n\t)\n\n\tif m.Scores != nil {\n\t\tfor _, s = range m.Scores {\n\t\t\tn = fmt.Sprintf(\"%sScore\", strings.Title(s.Name))\n\t\t\tattr.SetField(idx, n, s.Count)\n\t\t}\n\t}\n\n\tif m.Blog != nil {\n\t\tif m.Blog.Scores != nil {\n\t\t\tfor _, s = range m.Blog.Scores {\n\t\t\t\tn = fmt.Sprintf(\"%sScore\", strings.Title(s.Name))\n\t\t\t\tattr.SetField(idx, n, s.Count)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn idx\n}", "func addAllFieldIndexes(ctx context.Context, indexer client.FieldIndexer) error {\n\tif err := indexer.IndexField(ctx, &gardencorev1beta1.Project{}, gardencore.ProjectNamespace, func(obj client.Object) []string {\n\t\tproject, ok := obj.(*gardencorev1beta1.Project)\n\t\tif !ok {\n\t\t\treturn []string{\"\"}\n\t\t}\n\t\tif project.Spec.Namespace == nil {\n\t\t\treturn []string{\"\"}\n\t\t}\n\t\treturn []string{*project.Spec.Namespace}\n\t}); err != nil {\n\t\treturn fmt.Errorf(\"failed to add indexer to Project Informer: %w\", err)\n\t}\n\n\tif err := indexer.IndexField(ctx, &gardencorev1beta1.Shoot{}, gardencore.ShootSeedName, func(obj client.Object) []string {\n\t\tshoot, ok := obj.(*gardencorev1beta1.Shoot)\n\t\tif !ok {\n\t\t\treturn []string{\"\"}\n\t\t}\n\t\tif shoot.Spec.SeedName == nil {\n\t\t\treturn []string{\"\"}\n\t\t}\n\t\treturn []string{*shoot.Spec.SeedName}\n\t}); err != nil {\n\t\treturn fmt.Errorf(\"failed to add indexer to Shoot Informer: %w\", err)\n\t}\n\n\tif err := indexer.IndexField(ctx, &seedmanagementv1alpha1.ManagedSeed{}, seedmanagement.ManagedSeedShootName, func(obj client.Object) []string {\n\t\tms, ok := obj.(*seedmanagementv1alpha1.ManagedSeed)\n\t\tif !ok {\n\t\t\treturn []string{\"\"}\n\t\t}\n\t\tif ms.Spec.Shoot == nil {\n\t\t\treturn []string{\"\"}\n\t\t}\n\t\treturn []string{ms.Spec.Shoot.Name}\n\t}); err != nil {\n\t\treturn fmt.Errorf(\"failed to add indexer to ManagedSeed Informer: %w\", err)\n\t}\n\n\treturn nil\n}", "func GetIndexPkgs(page int) (pkgs []hv.PkgInfo) {\n\terr := x.Limit(100, (page-1)*100).Asc(\"rank\").Find(&pkgs)\n\tif err != nil {\n\t\tbeego.Error(\"models.GetIndexPkgs ->\", err)\n\t}\n\treturn pkgs\n}", "func (c *Checker) registerIdxPer(pkg string) {\n\tc.Packages[pkg] = append(c.Packages[pkg], len(c.Violations)-1)\n}", "func (c *Client) ItemTypeIndex() (*response.ItemTypeIndex, error) {\n\tvar data *d3.ItemTypeIndex\n\n\tep := endpointItemTypeIndex(c.region)\n\n\tq, err := c.get(ep, &data)\n\n\tif nil != err {\n\t\treturn nil, err\n\t}\n\n\treturn &response.ItemTypeIndex{\n\t\tData: data,\n\t\tEndpoint: ep,\n\t\tQuota: q,\n\t\tRegion: c.region,\n\t}, nil\n}", "func (d *Deployment) Scan(ctx context.Context, gvr, fqn string, wait bool) (Refs, error) {\n\tns, n := client.Namespaced(fqn)\n\too, err := d.GetFactory().List(d.GVR(), ns, wait, labels.Everything())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trefs := make(Refs, 0, len(oo))\n\tfor _, o := range oo {\n\t\tvar dp appsv1.Deployment\n\t\terr = runtime.DefaultUnstructuredConverter.FromUnstructured(o.(*unstructured.Unstructured).Object, &dp)\n\t\tif err != nil {\n\t\t\treturn nil, errors.New(\"expecting Deployment resource\")\n\t\t}\n\t\tswitch gvr {\n\t\tcase \"v1/configmaps\":\n\t\t\tif !hasConfigMap(&dp.Spec.Template.Spec, n) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\trefs = append(refs, Ref{\n\t\t\t\tGVR: d.GVR(),\n\t\t\t\tFQN: client.FQN(dp.Namespace, dp.Name),\n\t\t\t})\n\t\tcase \"v1/secrets\":\n\t\t\tfound, err := hasSecret(d.Factory, &dp.Spec.Template.Spec, dp.Namespace, n, wait)\n\t\t\tif err != nil {\n\t\t\t\tlog.Warn().Err(err).Msgf(\"scanning secret %q\", fqn)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif !found {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\trefs = append(refs, Ref{\n\t\t\t\tGVR: d.GVR(),\n\t\t\t\tFQN: client.FQN(dp.Namespace, dp.Name),\n\t\t\t})\n\t\tcase \"v1/persistentvolumeclaims\":\n\t\t\tif !hasPVC(&dp.Spec.Template.Spec, n) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\trefs = append(refs, Ref{\n\t\t\t\tGVR: d.GVR(),\n\t\t\t\tFQN: client.FQN(dp.Namespace, dp.Name),\n\t\t\t})\n\t\tcase \"scheduling.k8s.io/v1/priorityclasses\":\n\t\t\tif !hasPC(&dp.Spec.Template.Spec, n) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\trefs = append(refs, Ref{\n\t\t\t\tGVR: d.GVR(),\n\t\t\t\tFQN: client.FQN(dp.Namespace, dp.Name),\n\t\t\t})\n\t\t}\n\n\t}\n\n\treturn refs, nil\n}", "func ApplicationIndex(w http.ResponseWriter, r *http.Request) {\n\tdb, err := database.Connect()\n\tdefer db.Close()\n\tif err != nil {\n\t\tlog.Printf(\"Database error: '%s'\\n\", err)\n\t\tjre := jsonutil.NewJSONResponseError(\n\t\t\thttp.StatusInternalServerError,\n\t\t\t\"there was an error when attempting to connect to the database\")\n\t\tjsonutil.RespondJSONError(w, jre)\n\t\treturn\n\t}\n\tapps, err := resources.AllApplications(db)\n\tif err != nil {\n\t\tlog.Printf(\"Error when retrieving applications: '%s'\\n\", err)\n\t\tjre := jsonutil.NewJSONResponseError(\n\t\t\thttp.StatusInternalServerError,\n\t\t\t\"there was an error when attempting to connect to the database\")\n\t\tjsonutil.RespondJSONError(w, jre)\n\t\treturn\n\t}\n\tfor i := range apps {\n\t\tif len(apps[i].Pods) == 0 {\n\t\t\tapps[i].Pods = make([]resources.Pod, 0)\n\t\t}\n\t\tif len(apps[i].Environments) == 0 {\n\t\t\tapps[i].Environments = make([]resources.Environment, 0)\n\t\t}\n\t}\n\tjson.NewEncoder(w).Encode(apps)\n}", "func loadIndex(data []byte) (*repo.IndexFile, error) {\n\ti := &repo.IndexFile{}\n\tif err := yaml.Unmarshal(data, i); err != nil {\n\t\treturn i, err\n\t}\n\n\ti.SortEntries()\n\tif i.APIVersion == \"\" {\n\t\treturn i, repo.ErrNoAPIVersion\n\t}\n\treturn i, nil\n}", "func loadIndex(r io.Reader) (*IndexFile, error) {\n\ti := &IndexFile{}\n\tif err := json.NewDecoder(r).Decode(i); err != nil {\n\t\treturn i, err\n\t}\n\ti.SortEntries()\n\treturn i, nil\n}", "func loadIndex(data []byte) (*repo.IndexFile, error) {\n\ti := &repo.IndexFile{}\n\tif err := yaml.Unmarshal(data, i); err != nil {\n\t\treturn i, err\n\t}\n\n\ti.SortEntries()\n\n\tif i.APIVersion == \"\" {\n\t\treturn i, repo.ErrNoAPIVersion\n\t}\n\n\treturn i, nil\n}", "func (c *Controller) ListResources(w http.ResponseWriter, r *http.Request) {\n\tresources := c.Dao.GetResources()\n\tlog.Println(resources)\n\n\tdata, _ := json.Marshal(resources)\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tw.WriteHeader(http.StatusOK)\n\tw.Write(data)\n\treturn\n}", "func loadIndex(data []byte) (*repo.IndexFile, error) {\n\ti := &repo.IndexFile{}\n\tif err := yaml.Unmarshal(data, i); err != nil {\n\t\tklog.Error(err, \"Unmarshal failed. Data: \", data)\n\t\treturn i, err\n\t}\n\n\ti.SortEntries()\n\n\tif i.APIVersion == \"\" {\n\t\treturn i, repo.ErrNoAPIVersion\n\t}\n\n\treturn i, nil\n}", "func ListEndpointGroupResults(r *http.Request, cfg config.Config) (int, http.Header, []byte, error) {\n\n\t//STANDARD DECLARATIONS START\n\tcode := http.StatusOK\n\th := http.Header{}\n\toutput := []byte(\"\")\n\terr := error(nil)\n\tcontentType := \"application/xml\"\n\tcharset := \"utf-8\"\n\t//STANDARD DECLARATIONS END\n\n\tcontentType, err = respond.ParseAcceptHeader(r)\n\th.Set(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\tif err != nil {\n\t\tcode = http.StatusNotAcceptable\n\t\toutput, _ = respond.MarshalContent(respond.NotAcceptableContentType, contentType, \"\", \" \")\n\t\treturn code, h, output, err\n\t}\n\n\t// Parse the request into the input\n\turlValues := r.URL.Query()\n\tvars := mux.Vars(r)\n\n\ttenantDbConfig, err := authentication.AuthenticateTenant(r.Header, cfg)\n\tif err != nil {\n\t\tif err.Error() == \"Unauthorized\" {\n\t\t\tcode = http.StatusUnauthorized\n\t\t\tout := respond.UnauthorizedMessage\n\t\t\toutput = out.MarshalTo(contentType)\n\t\t\treturn code, h, output, err\n\t\t}\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\tsession, err := mongo.OpenSession(tenantDbConfig)\n\tdefer mongo.CloseSession(session)\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\treport := reports.MongoInterface{}\n\terr = mongo.FindOne(session, tenantDbConfig.Db, \"reports\", bson.M{\"info.name\": vars[\"report_name\"]}, &report)\n\n\tif err != nil {\n\t\tcode = http.StatusBadRequest\n\t\tmessage := \"The report with the name \" + vars[\"report_name\"] + \" does not exist\"\n\t\toutput, err := createErrorMessage(message, contentType) //Render the response into XML or JSON\n\t\th.Set(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\t\treturn code, h, output, err\n\t}\n\n\tinput := endpointGroupResultQuery{\n\t\tbasicQuery{\n\t\t\tName: vars[\"lgroup_name\"],\n\t\t\tGranularity: urlValues.Get(\"granularity\"),\n\t\t\tFormat: contentType,\n\t\t\tStartTime: urlValues.Get(\"start_time\"),\n\t\t\tEndTime: urlValues.Get(\"end_time\"),\n\t\t\tReport: report,\n\t\t\tVars: vars,\n\t\t}, \"\",\n\t}\n\n\ttenantDB := session.DB(tenantDbConfig.Db)\n\terrs := input.Validate(tenantDB)\n\tif len(errs) > 0 {\n\t\tout := respond.BadRequestSimple\n\t\tout.Errors = errs\n\t\toutput = out.MarshalTo(contentType)\n\t\tcode = 400\n\t\treturn code, h, output, err\n\t}\n\n\tif vars[\"lgroup_type\"] != report.GetEndpointGroupType() {\n\t\tcode = http.StatusBadRequest\n\t\tmessage := \"The report \" + vars[\"report_name\"] + \" does not define endpoint group type: \" + vars[\"lgroup_type\"] + \". Try using \" + report.GetEndpointGroupType() + \" instead.\"\n\t\toutput, err := createErrorMessage(message, contentType) //Render the response into XML or JSON\n\t\th.Set(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\t\treturn code, h, output, err\n\t}\n\n\tresults := []EndpointGroupInterface{}\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\t// Construct the query to mongodb based on the input\n\tfilter := bson.M{\n\t\t\"date\": bson.M{\"$gte\": input.StartTimeInt, \"$lte\": input.EndTimeInt},\n\t\t\"report\": report.ID,\n\t}\n\n\tif input.Name != \"\" {\n\t\tfilter[\"name\"] = input.Name\n\t}\n\n\t// Select the granularity of the search daily/monthly\n\tif input.Granularity == \"daily\" {\n\t\tcustomForm[0] = \"20060102\"\n\t\tcustomForm[1] = \"2006-01-02\"\n\t\tquery := DailyEndpointGroup(filter)\n\t\terr = mongo.Pipe(session, tenantDbConfig.Db, \"endpoint_group_ar\", query, &results)\n\t} else if input.Granularity == \"monthly\" {\n\t\tcustomForm[0] = \"200601\"\n\t\tcustomForm[1] = \"2006-01\"\n\t\tquery := MonthlyEndpointGroup(filter)\n\t\terr = mongo.Pipe(session, tenantDbConfig.Db, \"endpoint_group_ar\", query, &results)\n\t}\n\n\t// mongo.Find(session, tenantDbConfig.Db, \"endpoint_group_ar\", bson.M{}, \"_id\", &results)\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\toutput, err = createEndpointGroupResultView(results, report, input.Format)\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\treturn code, h, output, err\n}", "func (*BulkIndexResponse) Descriptor() ([]byte, []int) {\n\treturn file_protobuf_index_proto_rawDescGZIP(), []int{15}\n}" ]
[ "0.59473515", "0.5910294", "0.5895813", "0.5877585", "0.58192575", "0.5749591", "0.57174927", "0.5698926", "0.56244725", "0.5540453", "0.5508983", "0.54928124", "0.54442203", "0.5425963", "0.54177755", "0.53597957", "0.5350733", "0.53459483", "0.5278293", "0.5273926", "0.5261139", "0.52554137", "0.51712054", "0.51565874", "0.51518506", "0.5120092", "0.51103544", "0.5074563", "0.50634915", "0.5046172", "0.5038586", "0.5031231", "0.5029448", "0.5024264", "0.50147945", "0.5007662", "0.5005348", "0.50038105", "0.49910134", "0.4987291", "0.49823233", "0.49649528", "0.49636778", "0.49608055", "0.49551806", "0.4951384", "0.49468595", "0.4936837", "0.49364686", "0.4936058", "0.4932655", "0.49323234", "0.49250054", "0.4924357", "0.49242046", "0.4916827", "0.49165785", "0.49115825", "0.48935425", "0.48804677", "0.48755366", "0.48735952", "0.48502085", "0.48481622", "0.4845581", "0.48446596", "0.48368245", "0.48326224", "0.4831592", "0.4826705", "0.4821561", "0.48182786", "0.48156506", "0.48050603", "0.48029026", "0.47956264", "0.47951144", "0.47830245", "0.47827467", "0.47818053", "0.47815794", "0.47794363", "0.4779048", "0.47739217", "0.4773371", "0.47678214", "0.47620618", "0.47597635", "0.47574255", "0.4751009", "0.47492483", "0.47482526", "0.47458857", "0.4741442", "0.47408527", "0.47390735", "0.47371244", "0.4734788", "0.47268516", "0.47263882" ]
0.72815484
0
sort sorts the Resources in the index in display order and returns the ordered keys for the index Packages are sorted by package name Resources within a package are sorted by: [filename, namespace, name, kind, apiVersion]
func (p TreeWriter) sort(indexByPackage map[string][]*yaml.RNode) []string { var keys []string for k := range indexByPackage { pkgNodes := indexByPackage[k] sort.Slice(pkgNodes, func(i, j int) bool { return compareNodes(pkgNodes[i], pkgNodes[j]) }) keys = append(keys, k) } // return the package names sorted lexicographically sort.Strings(keys) return keys }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func sortResources(resources []astmodel.InternalTypeName) []astmodel.InternalTypeName {\n\tsort.Slice(resources, func(i, j int) bool {\n\t\tiVal := resources[i]\n\t\tjVal := resources[j]\n\n\t\treturn iVal.PackageReference().PackageName() < jVal.PackageReference().PackageName() ||\n\t\t\tiVal.PackageReference().PackageName() < jVal.PackageReference().PackageName() && iVal.Name() < jVal.Name()\n\t})\n\n\treturn resources\n}", "func (i IndexFile) sortPackages() {\n\tfor _, versions := range i.Entries {\n\t\tsort.Sort(sort.Reverse(versions))\n\t}\n}", "func SortResources(resources []*metav1.APIResourceList) {\n\tsort.SliceStable(resources, func(i, j int) bool {\n\t\tleft := resources[i]\n\t\tleftGV, _ := schema.ParseGroupVersion(left.GroupVersion)\n\t\t// not checking error because it should be impossible to fail to parse data coming from the\n\t\t// apiserver\n\t\tif leftGV.Group == \"extensions\" {\n\t\t\t// always sort extensions at the bottom by saying left is \"greater\"\n\t\t\treturn false\n\t\t}\n\n\t\tright := resources[j]\n\t\trightGV, _ := schema.ParseGroupVersion(right.GroupVersion)\n\t\t// not checking error because it should be impossible to fail to parse data coming from the\n\t\t// apiserver\n\t\tif rightGV.Group == \"extensions\" {\n\t\t\t// always sort extensions at the bottom by saying left is \"less\"\n\t\t\treturn true\n\t\t}\n\n\t\treturn i < j\n\t})\n}", "func orderStackResourceKeys(m map[string]StackResource) []string {\n\tret := make([]string, len(m))\n\ti := 0\n\n\tfor k := range m {\n\t\tret[i] = k\n\t\ti++\n\t}\n\tsort.Sort(sort.Reverse(sort.StringSlice(ret)))\n\treturn ret\n}", "func Sort(sortMetricName string, sortType string, rawMetrics *FormatedLevelMetric) (*FormatedLevelMetric, int) {\n\tdefer func() {\n\t\tif err := recover(); err != nil {\n\t\t\tglog.Errorln(err)\n\t\t\tdebug.PrintStack()\n\t\t}\n\t}()\n\n\tif sortMetricName == \"\" {\n\t\treturn rawMetrics, -1\n\t}\n\n\t// default sort type is descending order\n\tif sortType == \"\" {\n\t\tsortType = ResultSortTypeDesc\n\t}\n\n\tvar currentResourceMap = make(map[string]int)\n\n\t// {<Resource Name>: <Ordering>}\n\tvar indexMap = make(map[string]int)\n\ti := 0\n\n\t// each metricItem is the result for a specific metric name\n\t// so we find the metricItem with sortMetricName, and sort it\n\tfor _, metricItem := range rawMetrics.Results {\n\t\t// only vector type result can be sorted\n\t\tif metricItem.Data.ResultType == ResultTypeVector && metricItem.Status == MetricStatusSuccess {\n\t\t\tif metricItem.MetricName == sortMetricName {\n\t\t\t\tif sortType == ResultSortTypeAsc {\n\t\t\t\t\t// asc\n\t\t\t\t\tsort.Sort(FormatedMetricDataWrapper{metricItem.Data, func(p, q *map[string]interface{}) bool {\n\t\t\t\t\t\tvalue1 := (*p)[ResultItemValue].([]interface{})\n\t\t\t\t\t\tvalue2 := (*q)[ResultItemValue].([]interface{})\n\t\t\t\t\t\tv1, _ := strconv.ParseFloat(value1[len(value1)-1].(string), 64)\n\t\t\t\t\t\tv2, _ := strconv.ParseFloat(value2[len(value2)-1].(string), 64)\n\t\t\t\t\t\tif v1 == v2 {\n\t\t\t\t\t\t\tresourceName1 := (*p)[ResultItemMetric].(map[string]interface{})[ResultItemMetricResourceName]\n\t\t\t\t\t\t\tresourceName2 := (*q)[ResultItemMetric].(map[string]interface{})[ResultItemMetricResourceName]\n\t\t\t\t\t\t\treturn resourceName1.(string) < resourceName2.(string)\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\treturn v1 < v2\n\t\t\t\t\t}})\n\t\t\t\t} else {\n\t\t\t\t\t// desc\n\t\t\t\t\tsort.Sort(FormatedMetricDataWrapper{metricItem.Data, func(p, q *map[string]interface{}) bool {\n\t\t\t\t\t\tvalue1 := (*p)[ResultItemValue].([]interface{})\n\t\t\t\t\t\tvalue2 := (*q)[ResultItemValue].([]interface{})\n\t\t\t\t\t\tv1, _ := strconv.ParseFloat(value1[len(value1)-1].(string), 64)\n\t\t\t\t\t\tv2, _ := strconv.ParseFloat(value2[len(value2)-1].(string), 64)\n\n\t\t\t\t\t\tif v1 == v2 {\n\t\t\t\t\t\t\tresourceName1 := (*p)[ResultItemMetric].(map[string]interface{})[ResultItemMetricResourceName]\n\t\t\t\t\t\t\tresourceName2 := (*q)[ResultItemMetric].(map[string]interface{})[ResultItemMetricResourceName]\n\t\t\t\t\t\t\treturn resourceName1.(string) > resourceName2.(string)\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\treturn v1 > v2\n\t\t\t\t\t}})\n\t\t\t\t}\n\n\t\t\t\tfor _, r := range metricItem.Data.Result {\n\t\t\t\t\t// record the ordering of resource_name to indexMap\n\t\t\t\t\t// example: {\"metric\":{ResultItemMetricResourceName: \"Deployment:xxx\"},\"value\":[1541142931.731,\"3\"]}\n\t\t\t\t\tresourceName, exist := r[ResultItemMetric].(map[string]interface{})[ResultItemMetricResourceName]\n\t\t\t\t\tif exist {\n\t\t\t\t\t\tif _, exist := indexMap[resourceName.(string)]; !exist {\n\t\t\t\t\t\t\tindexMap[resourceName.(string)] = i\n\t\t\t\t\t\t\ti = i + 1\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// iterator all metric to find max metricItems length\n\t\t\tfor _, r := range metricItem.Data.Result {\n\t\t\t\tk, ok := r[ResultItemMetric].(map[string]interface{})[ResultItemMetricResourceName]\n\t\t\t\tif ok {\n\t\t\t\t\tcurrentResourceMap[k.(string)] = 1\n\t\t\t\t}\n\t\t\t}\n\n\t\t}\n\t}\n\n\tvar keys []string\n\tfor k := range currentResourceMap {\n\t\tkeys = append(keys, k)\n\t}\n\tsort.Strings(keys)\n\n\tfor _, resource := range keys {\n\t\tif _, exist := indexMap[resource]; !exist {\n\t\t\tindexMap[resource] = i\n\t\t\ti = i + 1\n\t\t}\n\t}\n\n\t// sort other metric\n\tfor i := 0; i < len(rawMetrics.Results); i++ {\n\t\tre := rawMetrics.Results[i]\n\t\tif re.Data.ResultType == ResultTypeVector && re.Status == MetricStatusSuccess {\n\t\t\tsortedMetric := make([]map[string]interface{}, len(indexMap))\n\t\t\tfor j := 0; j < len(re.Data.Result); j++ {\n\t\t\t\tr := re.Data.Result[j]\n\t\t\t\tk, exist := r[ResultItemMetric].(map[string]interface{})[ResultItemMetricResourceName]\n\t\t\t\tif exist {\n\t\t\t\t\tindex, exist := indexMap[k.(string)]\n\t\t\t\t\tif exist {\n\t\t\t\t\t\tsortedMetric[index] = r\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\trawMetrics.Results[i].Data.Result = sortedMetric\n\t\t}\n\t}\n\n\treturn rawMetrics, len(indexMap)\n}", "func (d *Default) OrderResources(ctx context.Context, sp *spec.Spec, opts *DefaultOptions) error {\n\tsort.Slice(sp.Paths, func(i, j int) bool {\n\t\tp1, p2 := sp.Paths[i], sp.Paths[j]\n\n\t\treturn p1.Name < p2.Name\n\t})\n\n\tfor _, p := range sp.Paths {\n\t\tsort.Slice(p.Operations, func(i, j int) bool {\n\t\t\to1, o2 := p.Operations[i], p.Operations[j]\n\n\t\t\treturn o1.Name < o2.Name\n\t\t})\n\n\t\tfor _, o := range p.Operations {\n\t\t\tsort.Slice(o.Parameters, func(i, j int) bool {\n\t\t\t\tp1, p2 := o.Parameters[i], o.Parameters[j]\n\n\t\t\t\treturn p1.Name < p2.Name\n\t\t\t})\n\n\t\t\tsort.Slice(o.Responses, func(i, j int) bool {\n\t\t\t\tr1, r2 := o.Responses[i], o.Responses[j]\n\n\t\t\t\treturn r1.Name < r2.Name\n\t\t\t})\n\n\t\t\tfor _, cb := range o.Callbacks {\n\t\t\t\tfor _, cbPath := range cb {\n\t\t\t\t\tsort.Slice(cb, func(i, j int) bool {\n\t\t\t\t\t\tp1, p2 := cb[i], cb[j]\n\n\t\t\t\t\t\treturn p1.Name < p2.Name\n\t\t\t\t\t})\n\n\t\t\t\t\tfor _, cbOp := range cbPath.Operations {\n\t\t\t\t\t\tsort.Slice(cbOp.Parameters, func(i, j int) bool {\n\t\t\t\t\t\t\tp1, p2 := cbOp.Parameters[i], cbOp.Parameters[j]\n\n\t\t\t\t\t\t\treturn p1.Name < p2.Name\n\t\t\t\t\t\t})\n\n\t\t\t\t\t\tsort.Slice(cbOp.Responses, func(i, j int) bool {\n\t\t\t\t\t\t\tr1, r2 := cbOp.Responses[i], cbOp.Responses[j]\n\n\t\t\t\t\t\t\treturn r1.Name < r2.Name\n\t\t\t\t\t\t})\n\t\t\t\t\t}\n\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\treturn nil\n}", "func orderStackCRDKeys(m map[string]apiextensions.CustomResourceDefinition) []string {\n\tret := make([]string, len(m))\n\ti := 0\n\n\tfor k := range m {\n\t\tret[i] = k\n\t\ti++\n\t}\n\tsort.Sort(sort.Reverse(sort.StringSlice(ret)))\n\treturn ret\n}", "func sortByKind(manifests []releaseutil.Manifest) []releaseutil.Manifest {\n\tordering := kuberesource.InstallOrder\n\tks := newKindSorter(manifests, ordering)\n\tsort.Sort(ks)\n\treturn ks.manifests\n}", "func SortByKind(manifests []*resource.Resource, ordering SortOrder) []*resource.Resource {\n\tks := newKindSorter(manifests, ordering)\n\tsort.Sort(ks)\n\treturn ks.resources\n}", "func SortResourcesByType(resources []Resource) {\n\tsort.Sort(resourcesSortedByType(resources))\n}", "func sortedMapKeysbyName(m *jsonschema.Index) []string {\n\tvar schemas []*jsonschema.Schema\n\tfor _, v := range *m {\n\t\tschemas = append(schemas, v)\n\t}\n\tsort.Sort(byName(schemas))\n\n\tvar keys []string\n\tfor _, v := range schemas {\n\t\tkeys = append(keys, v.Pointer)\n\t}\n\treturn keys\n}", "func orderStackIconKeys(m map[string]*v1alpha1.IconSpec) []string {\n\tret := make([]string, len(m))\n\ti := 0\n\n\tfor k := range m {\n\t\tret[i] = k\n\t\ti++\n\t}\n\tsort.Sort(sort.Reverse(sort.StringSlice(ret)))\n\treturn ret\n}", "func normalizeAPIGroupResources(apiGroupResource *restmapper.APIGroupResources) []metav1.APIResource {\n\tvar versionedResources []metav1.APIResource\n\tfor version, vr := range apiGroupResource.VersionedResources {\n\t\tfor _, resource := range vr {\n\t\t\tresource.Group = apiGroupResource.Group.Name\n\t\t\tresource.Version = version\n\t\t\tversionedResources = append(versionedResources, resource)\n\t\t}\n\t}\n\n\t// Ensure deterministic output.\n\tpreferredVersion := apiGroupResource.Group.PreferredVersion.Version\n\tsort.SliceStable(versionedResources, func(i, j int) bool {\n\t\tif versionedResources[i].Version == versionedResources[j].Version {\n\t\t\treturn versionedResources[i].Name < versionedResources[j].Name\n\t\t}\n\n\t\t// preferred version\n\t\tif versionedResources[i].Version == preferredVersion {\n\t\t\treturn true\n\t\t}\n\t\tif versionedResources[j].Version == preferredVersion {\n\t\t\treturn false\n\t\t}\n\n\t\t// compare kube-like version\n\t\t// Versions will be sorted based on GA/alpha/beta first and then major and minor versions.\n\t\t// e.g. v2, v1, v1beta2, v1beta1, v1alpha1.\n\t\treturn version.CompareKubeAwareVersionStrings(versionedResources[i].Version, versionedResources[j].Version) > 0\n\t})\n\n\t// pick out preferred version or highest semantic version\n\tregistered := make(map[string]bool)\n\tvar normalizedVersionResources []metav1.APIResource\n\tfor _, vr := range versionedResources {\n\t\tif registered[vr.Name] {\n\t\t\tcontinue\n\t\t}\n\t\tnormalizedVersionResources = append(normalizedVersionResources, vr)\n\t\tregistered[vr.Name] = true\n\t}\n\treturn normalizedVersionResources\n}", "func (v ResourceNodes) Sort() {\n\tsort.Sort(v)\n}", "func (tf tFiles) sortByKey(icmp *iComparer) {\n\tsort.Sort(&tFilesSortByKey{tFiles: tf, icmp: icmp})\n}", "func (p TreeWriter) index(nodes []*yaml.RNode) map[string][]*yaml.RNode {\n\t// index the ResourceNodes by package\n\tindexByPackage := map[string][]*yaml.RNode{}\n\tfor i := range nodes {\n\t\tmeta, err := nodes[i].GetMeta()\n\t\tif err != nil || meta.Kind == \"\" {\n\t\t\t// not a resource\n\t\t\tcontinue\n\t\t}\n\t\tpkg := filepath.Dir(meta.Annotations[kioutil.PathAnnotation])\n\t\tindexByPackage[pkg] = append(indexByPackage[pkg], nodes[i])\n\t}\n\treturn indexByPackage\n}", "func (ri RecipeIndex) sortedSlice() []Recipe {\n\tvar rs []Recipe\n\tfor _, v := range ri {\n\t\trs = append(rs, v)\n\t}\n\tsort.Slice(rs, func(i, j int) bool {\n\t\treturn len(rs[i].Installers) < len(rs[j].Installers)\n\t})\n\treturn rs\n}", "func (i IndexFile) SortEntries() {\n\tfor _, versions := range i.Entries {\n\t\tsort.Sort(sort.Reverse(versions))\n\t}\n}", "func sortedMapKeys(m *jsonschema.Index) []string {\n\tvar keys []string\n\tfor k, _ := range *m {\n\t\tkeys = append(keys, k)\n\t}\n\tsort.Strings(keys)\n\treturn keys\n}", "func sortLayer(layer []task.Task, idToDisplayName map[string]string) []task.Task {\n\tsortKeys := make([]string, 0, len(layer))\n\tsortKeyToTask := make(map[string]task.Task)\n\tfor _, t := range layer {\n\t\t// Construct a key to sort by, consisting of all dependency names, sorted alphabetically,\n\t\t// followed by the task name\n\t\tsortKeyWords := make([]string, 0, len(t.DependsOn)+1)\n\t\tfor _, dep := range t.DependsOn {\n\t\t\tdepName, ok := idToDisplayName[dep.TaskId]\n\t\t\t// Cross-variant dependencies will not be included in idToDisplayName\n\t\t\tif !ok {\n\t\t\t\tdepName = dep.TaskId\n\t\t\t}\n\t\t\tsortKeyWords = append(sortKeyWords, depName)\n\t\t}\n\t\tsort.Strings(sortKeyWords)\n\t\tsortKeyWords = append(sortKeyWords, t.DisplayName)\n\t\tsortKey := strings.Join(sortKeyWords, \" \")\n\t\tsortKeys = append(sortKeys, sortKey)\n\t\tsortKeyToTask[sortKey] = t\n\t}\n\tsort.Strings(sortKeys)\n\tsortedLayer := make([]task.Task, 0, len(layer))\n\tfor _, sortKey := range sortKeys {\n\t\tsortedLayer = append(sortedLayer, sortKeyToTask[sortKey])\n\t}\n\treturn sortedLayer\n}", "func getSortedKeys(modules map[string]*TerraformModule) []string {\n\tkeys := []string{}\n\tfor key := range modules {\n\t\tkeys = append(keys, key)\n\t}\n\n\tsort.Strings(keys)\n\n\treturn keys\n}", "func TestResourceListSorting(t *testing.T) {\n\tsortedResourceList := make([]string, len(resourceList))\n\tcopy(sortedResourceList, resourceList)\n\tsort.Strings(sortedResourceList)\n\tfor i := 0; i < len(resourceList); i++ {\n\t\tif resourceList[i] != sortedResourceList[i] {\n\t\t\tt.Errorf(\"Expected resourceList[%d] = \\\"%s\\\", resourceList is not correctly sorted.\", i, sortedResourceList[i])\n\t\t\tbreak\n\t\t}\n\t}\n}", "func (s SecretVersions) Sort() {\n\tsort.Sort(s)\n}", "func getResources(client *ApiClient) (map[schema.GroupVersion][]metav1.APIResource, error) {\n\n\tresourceLists, err := client.ClientSet.Discovery().ServerPreferredResources()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tversionResource := map[schema.GroupVersion][]metav1.APIResource{}\n\n\tfor _, apiResourceList := range resourceLists {\n\t\tversion, err := schema.ParseGroupVersion(apiResourceList.GroupVersion)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"unable to parse GroupVersion %v\",err)\n\t\t}\n\n\t\tversionResource[version] = uniqResources(apiResourceList.APIResources)\n\t}\n\n\treturn versionResource, nil\n}", "func (p FileInfos) Sort() { sort.Sort(p) }", "func (p *PackageList) Sort(criteria string) {\n\tswitch criteria {\n\tcase \"alpha\":\n\t\talphaSort(p)\n\tcase \"score\":\n\t\tscoreSort(p)\n\tcase \"stars\":\n\t\tstarSort(p)\n\tcase \"imports\":\n\t\timportSort(p)\n\t}\n}", "func sortIngressByCreationTime(ingr []*knetworking.Ingress) []*knetworking.Ingress {\n\tsort.Slice(ingr, func(i, j int) bool {\n\t\t// If creation time is the same, then behavior is nondeterministic. In this case, we can\n\t\t// pick an arbitrary but consistent ordering based on name and namespace, which is unique.\n\t\t// CreationTimestamp is stored in seconds, so this is not uncommon.\n\t\tif ingr[i].CreationTimestamp == ingr[j].CreationTimestamp {\n\t\t\tin := ingr[i].Name + \".\" + ingr[i].Namespace\n\t\t\tjn := ingr[j].Name + \".\" + ingr[j].Namespace\n\t\t\treturn in < jn\n\t\t}\n\t\treturn ingr[i].CreationTimestamp.Before(&ingr[j].CreationTimestamp)\n\t})\n\treturn ingr\n}", "func getPackageRefs(resources []astmodel.InternalTypeName) []astmodel.PackageReference {\n\tpackageRefs := make([]astmodel.PackageReference, 0, len(resources)+1)\n\t// Package reference for return type\n\tpackageRefs = append(packageRefs, astmodel.KubernetesResourceType.PackageReference())\n\n\tfor _, typeDef := range resources {\n\t\tpackageRefs = append(packageRefs, typeDef.PackageReference())\n\t}\n\n\treturn packageRefs\n}", "func sortDescription(v interface{}) {\n\tslice := reflect.ValueOf(v)\n\tvalues := toValueSlice(slice)\n\tsort.Slice(values, func(i, j int) bool {\n\t\tnameI := values[i].FieldByName(\"Name\").String()\n\t\tnameJ := values[j].FieldByName(\"Name\").String()\n\t\tif nameI == nameJ {\n\t\t\tkindI := values[i].FieldByName(\"Kind\").String()\n\t\t\tkindJ := values[j].FieldByName(\"Kind\").String()\n\t\t\tif kindI == kindJ {\n\t\t\t\tversionI := values[i].FieldByName(\"Version\").String()\n\t\t\t\tversionJ := values[j].FieldByName(\"Version\").String()\n\t\t\t\treturn version.CompareKubeAwareVersionStrings(versionI, versionJ) > 0\n\t\t\t}\n\t\t\treturn kindI < kindJ\n\t\t}\n\t\treturn nameI < nameJ\n\t})\n\tfor i := 0; i < slice.Len(); i++ {\n\t\tslice.Index(i).Set(values[i])\n\t}\n}", "func sortNodesByUsage(nodes []NodeUsage) {\n\tsort.Slice(nodes, func(i, j int) bool {\n\t\tti := nodes[i].usage[v1.ResourceMemory].Value() + nodes[i].usage[v1.ResourceCPU].MilliValue() + nodes[i].usage[v1.ResourcePods].Value()\n\t\ttj := nodes[j].usage[v1.ResourceMemory].Value() + nodes[j].usage[v1.ResourceCPU].MilliValue() + nodes[j].usage[v1.ResourcePods].Value()\n\n\t\t// extended resources\n\t\tfor name := range nodes[i].usage {\n\t\t\tif !isBasicResource(name) {\n\t\t\t\tti = ti + nodes[i].usage[name].Value()\n\t\t\t\ttj = tj + nodes[j].usage[name].Value()\n\t\t\t}\n\t\t}\n\n\t\t// To return sorted in descending order\n\t\treturn ti > tj\n\t})\n}", "func sortConfig(provider objsort.Namespaced) objsort.Config {\n\treturn objsort.Config{\n\t\tNamespacedIndicator: func(gvk schema.GroupVersionKind) (bool, error) {\n\t\t\tret, err := provider(gvk)\n\t\t\tif err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\t\t\treturn ret, nil\n\t\t},\n\t\tOrderingProvider: ordering,\n\t}\n}", "func fetchGroupVersionResources(d discovery.DiscoveryInterface, apiGroups []*metav1.APIGroup) (map[schema.GroupVersion]*metav1.APIResourceList, map[schema.GroupVersion]error) {\n\tgroupVersionResources := make(map[schema.GroupVersion]*metav1.APIResourceList)\n\tfailedGroups := make(map[schema.GroupVersion]error)\n\n\twg := &sync.WaitGroup{}\n\tresultLock := &sync.Mutex{}\n\tfor _, apiGroup := range apiGroups {\n\t\tfor _, version := range apiGroup.Versions {\n\t\t\tgroupVersion := schema.GroupVersion{Group: apiGroup.Name, Version: version.Version}\n\t\t\twg.Add(1)\n\t\t\tgo func() {\n\t\t\t\tdefer wg.Done()\n\t\t\t\tdefer utilruntime.HandleCrash()\n\n\t\t\t\tapiResourceList, err := d.ServerResourcesForGroupVersion(groupVersion.String())\n\n\t\t\t\t// lock to record results\n\t\t\t\tresultLock.Lock()\n\t\t\t\tdefer resultLock.Unlock()\n\n\t\t\t\tif err != nil {\n\t\t\t\t\t// TODO: maybe restrict this to NotFound errors\n\t\t\t\t\tfailedGroups[groupVersion] = err\n\t\t\t\t}\n\t\t\t\tif apiResourceList != nil {\n\t\t\t\t\t// even in case of error, some fallback might have been returned\n\t\t\t\t\tgroupVersionResources[groupVersion] = apiResourceList\n\t\t\t\t}\n\t\t\t}()\n\t\t}\n\t}\n\twg.Wait()\n\n\treturn groupVersionResources, failedGroups\n}", "func getAllAPIVersions(apiResourceLists []*metav1.APIResourceList) map[string]apiVersions {\n\ttempKindToAPIVersions := make(map[string]apiVersions)\n\n\tfor _, apiResourceList := range apiResourceLists {\n\t\tlastKind := \"\"\n\t\tfor _, apiResource := range apiResourceList.APIResources {\n\t\t\tif apiResource.Kind == lastKind {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tversion, ok := tempKindToAPIVersions[apiResource.Kind]\n\t\t\tif !ok {\n\t\t\t\ttempKindToAPIVersions[apiResource.Kind] = apiVersions{}\n\t\t\t}\n\n\t\t\tgvk := strings.Join([]string{apiResourceList.GroupVersion, apiResource.Kind}, \"/\")\n\t\t\tversion.gvks = append(version.gvks, gvk)\n\t\t\ttempKindToAPIVersions[apiResource.Kind] = version\n\t\t\tlastKind = apiResource.Kind\n\t\t}\n\t}\n\n\treturn tempKindToAPIVersions\n}", "func Sort(sortKeys []string, reverseList []bool, data []Record) {\n\tsplitSortKeys := make([][]string, len(sortKeys))\n\tfor i, sortKey := range sortKeys {\n\t\tsplitSortKeys[i] = strings.Split(sortKey, \".\")\n\t}\n\n\tless := func(i, j int) (l bool) {\n\t\tvar reverse bool\n\t\tdefer func() {\n\t\t\tif reverse {\n\t\t\t\tl = !l\n\t\t\t}\n\t\t}()\n\tSORTKEYLOOP:\n\t\tfor sortKeyIdx, keyParts := range splitSortKeys {\n\t\t\treverse = reverseList[sortKeyIdx]\n\t\t\t// TODO: record could (and should) point at the CollectionFields which will tell us types\n\t\t\tiVal, iok := data[i].Get(keyParts)\n\t\t\tjVal, jok := data[j].Get(keyParts)\n\t\t\tif !iok || !jok {\n\t\t\t\t// if both don't have this field, continue on\n\t\t\t\tif iok == jok {\n\t\t\t\t\tcontinue SORTKEYLOOP\n\t\t\t\t}\n\t\t\t\tl = iok\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif iVal == nil || jVal == nil {\n\t\t\t\t// if both are nil, continue on\n\t\t\t\tif iVal == jVal {\n\t\t\t\t\tcontinue SORTKEYLOOP\n\t\t\t\t}\n\t\t\t\tl = iVal != nil\n\t\t\t\treturn\n\t\t\t}\n\t\t\tswitch iValTyped := iVal.(type) {\n\t\t\tcase string:\n\t\t\t\tjValTyped := jVal.(string)\n\t\t\t\tif iValTyped != jValTyped {\n\t\t\t\t\tl = iValTyped < jValTyped\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\tcase int:\n\t\t\t\tjValTyped := jVal.(int)\n\t\t\t\tif iValTyped != jValTyped {\n\t\t\t\t\tl = iValTyped < jValTyped\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\tcase int64:\n\t\t\t\tjValTyped := jVal.(int64)\n\t\t\t\tif iValTyped != jValTyped {\n\t\t\t\t\tl = iValTyped < jValTyped\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\tcase float64:\n\t\t\t\tjValTyped := jVal.(float64)\n\t\t\t\tif iValTyped != jValTyped {\n\t\t\t\t\tl = iValTyped < jValTyped\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\tcase bool:\n\t\t\t\tjValTyped := jVal.(bool)\n\t\t\t\tif iValTyped != jValTyped {\n\t\t\t\t\tl = !iValTyped && jValTyped\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t// TODO: return error? At this point if all return false, I'm not sure what happens\n\t\t\tdefault:\n\t\t\t\tpanic(\"Unknown type\")\n\t\t\t\tl = false\n\t\t\t\treturn\n\n\t\t\t}\n\t\t}\n\t\tl = false\n\t\treturn\n\t}\n\tsort.Slice(data, less)\n}", "func (p *provider) GetResources(_ context.Context) (map[string]tfsdk.ResourceType, diag.Diagnostics) {\n\treturn map[string]tfsdk.ResourceType{\n\t\t\"hashicups_order\": resourceOrderType{},\n\t}, nil\n}", "func TopologicalSort(bundle *smith.Bundle) (*SortedData, error) {\n\tgraph := newGraph(len(bundle.Spec.Resources))\n\n\tfor _, res := range bundle.Spec.Resources {\n\t\tgraph.addVertex(res.Name)\n\t}\n\n\tfor _, res := range bundle.Spec.Resources {\n\t\tfor _, d := range res.DependsOn {\n\t\t\tgraph.addEdge(res.Name, d)\n\t\t}\n\t}\n\n\tsorted, err := graph.topologicalSort()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tgraphData := SortedData{\n\t\tGraph: graph,\n\t\tSortedVertices: sorted,\n\t}\n\n\tlog.Printf(\"Sorted graph: %v\", sorted)\n\n\treturn &graphData, nil\n}", "func SortTags(tags []v1alpha1.Tag, ecrTags []ecr.Tag) {\n\tsort.Slice(tags, func(i, j int) bool {\n\t\treturn tags[i].Key < tags[j].Key\n\t})\n\n\tsort.Slice(ecrTags, func(i, j int) bool {\n\t\treturn *ecrTags[i].Key < *ecrTags[j].Key\n\t})\n}", "func sortKeys(v []reflect.Value) []reflect.Value {\n\tif len(v) <= 1 {\n\t\treturn v\n\t}\n\tswitch v[0].Kind() {\n\tcase reflect.Float32, reflect.Float64:\n\t\tsort.Sort(rvFloats{v})\n\tcase reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:\n\t\tsort.Sort(rvInts{v})\n\tcase reflect.String:\n\t\tsort.Sort(rvStrings{v})\n\tcase reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr:\n\t\tsort.Sort(rvUints{v})\n\t}\n\treturn v\n}", "func revisionListSortFunc(revisionList *servingv1.RevisionList) func(i int, j int) bool {\n\treturn func(i, j int) bool {\n\t\ta := revisionList.Items[i]\n\t\tb := revisionList.Items[j]\n\n\t\t// By Namespace\n\t\taNamespace := a.Namespace\n\t\tbNamespace := b.Namespace\n\t\tif aNamespace != bNamespace {\n\t\t\treturn aNamespace < bNamespace\n\t\t}\n\n\t\t// By Service\n\t\taService := a.Labels[serving.ServiceLabelKey]\n\t\tbService := b.Labels[serving.ServiceLabelKey]\n\n\t\tif aService != bService {\n\t\t\treturn aService < bService\n\t\t}\n\n\t\t// By Generation\n\t\t// Convert configuration generation key from string to int for avoiding string comparison.\n\t\tagen, err := strconv.Atoi(a.Labels[serving.ConfigurationGenerationLabelKey])\n\t\tif err != nil {\n\t\t\treturn a.Name < b.Name\n\t\t}\n\t\tbgen, err := strconv.Atoi(b.Labels[serving.ConfigurationGenerationLabelKey])\n\t\tif err != nil {\n\t\t\treturn a.Name < b.Name\n\t\t}\n\n\t\tif agen != bgen {\n\t\t\treturn agen > bgen\n\t\t}\n\t\treturn a.Name < b.Name\n\t}\n}", "func (uq *UploadQueue) sort() {\n\tsort.Slice(uq.Files, func(i, j int) bool {\n\t\treturn len(uq.Files[i].Data[0]) > len(uq.Files[j].Data[0])\n\t})\n}", "func listResources(clt resourcesAPIGetter, r *http.Request, resourceKind string) (*types.ListResourcesResponse, error) {\n\tvalues := r.URL.Query()\n\n\tlimit, err := queryLimitAsInt32(values, \"limit\", defaults.MaxIterationLimit)\n\tif err != nil {\n\t\treturn nil, trace.Wrap(err)\n\t}\n\n\t// Sort is expected in format `<fieldName>:<asc|desc>` where\n\t// index 0 is fieldName and index 1 is direction.\n\t// If a direction is not set, or is not recognized, it defaults to ASC.\n\tvar sortBy types.SortBy\n\tsortParam := values.Get(\"sort\")\n\tif sortParam != \"\" {\n\t\tvals := strings.Split(sortParam, \":\")\n\t\tif vals[0] != \"\" {\n\t\t\tsortBy.Field = vals[0]\n\t\t\tif len(vals) > 1 && vals[1] == \"desc\" {\n\t\t\t\tsortBy.IsDesc = true\n\t\t\t}\n\t\t}\n\t}\n\n\tstartKey := values.Get(\"startKey\")\n\treq := proto.ListResourcesRequest{\n\t\tResourceType: resourceKind,\n\t\tLimit: limit,\n\t\tStartKey: startKey,\n\t\tSortBy: sortBy,\n\t\tPredicateExpression: values.Get(\"query\"),\n\t\tSearchKeywords: client.ParseSearchKeywords(values.Get(\"search\"), ' '),\n\t\tUseSearchAsRoles: values.Get(\"searchAsRoles\") == \"yes\",\n\t}\n\n\treturn clt.ListResources(r.Context(), req)\n}", "func (w *exportWriter) writeIndex(index map[types.Object]uint64) {\n\ttype pkgObj struct {\n\t\tobj types.Object\n\t\tname string // qualified name; differs from obj.Name for type params\n\t}\n\t// Build a map from packages to objects from that package.\n\tpkgObjs := map[*types.Package][]pkgObj{}\n\n\t// For the main index, make sure to include every package that\n\t// we reference, even if we're not exporting (or reexporting)\n\t// any symbols from it.\n\tif w.p.localpkg != nil {\n\t\tpkgObjs[w.p.localpkg] = nil\n\t}\n\tfor pkg := range w.p.allPkgs {\n\t\tpkgObjs[pkg] = nil\n\t}\n\n\tfor obj := range index {\n\t\tname := w.p.exportName(obj)\n\t\tpkgObjs[obj.Pkg()] = append(pkgObjs[obj.Pkg()], pkgObj{obj, name})\n\t}\n\n\tvar pkgs []*types.Package\n\tfor pkg, objs := range pkgObjs {\n\t\tpkgs = append(pkgs, pkg)\n\n\t\tsort.Slice(objs, func(i, j int) bool {\n\t\t\treturn objs[i].name < objs[j].name\n\t\t})\n\t}\n\n\tsort.Slice(pkgs, func(i, j int) bool {\n\t\treturn w.exportPath(pkgs[i]) < w.exportPath(pkgs[j])\n\t})\n\n\tw.uint64(uint64(len(pkgs)))\n\tfor _, pkg := range pkgs {\n\t\tw.string(w.exportPath(pkg))\n\t\tw.string(pkg.Name())\n\t\tw.uint64(uint64(0)) // package height is not needed for go/types\n\n\t\tobjs := pkgObjs[pkg]\n\t\tw.uint64(uint64(len(objs)))\n\t\tfor _, obj := range objs {\n\t\t\tw.string(obj.name)\n\t\t\tw.uint64(index[obj.obj])\n\t\t}\n\t}\n}", "func (responses Responses) SortByReleaseName() {\n\tsort.Slice(responses, func(i, j int) bool {\n\t\treturn responses[i].ReleaseName < responses[j].ReleaseName\n\t})\n}", "func sortGroups(ctx context.Context, log logrus.FieldLogger, client gcs.Stater, configPath gcs.Path, gridPrefix string, groups []*configpb.TestGroup) (map[string]int64, error) {\n\tgroupedPaths := make(map[gcs.Path]*configpb.TestGroup, len(groups))\n\tpaths := make([]gcs.Path, 0, len(groups))\n\tfor _, tg := range groups {\n\t\ttgp, err := testGroupPath(configPath, gridPrefix, tg.Name)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"%s bad group path: %w\", tg.Name, err)\n\t\t}\n\t\tgroupedPaths[*tgp] = tg\n\t\tpaths = append(paths, *tgp)\n\t}\n\n\tgenerationPaths := gcs.LeastRecentlyUpdated(ctx, log, client, paths)\n\tgenerations := make(map[string]int64, len(generationPaths))\n\tfor i, p := range paths {\n\t\ttg := groupedPaths[p]\n\t\tgroups[i] = tg\n\t\tgenerations[tg.Name] = generationPaths[p]\n\t}\n\n\treturn generations, nil\n}", "func sortAlert() {\n\n\tswitch strings.ToLower(sortLabel) {\n\tcase \"severity\":\n\t\tsort.Sort(severitySorter(allAlertData))\n\tcase \"starts\":\n\t\tsort.Sort(startAtSorter(allAlertData))\n\tcase \"ends\":\n\t\tsort.Sort(endsAtSorter(allAlertData))\n\tcase \"duration\":\n\t\tsort.Sort(durationSorter(allAlertData))\n\tdefault:\n\t\treturn\n\t}\n}", "func sortDescriptors(v interface{}) {\n\tslice := reflect.ValueOf(v)\n\tvalues := toValueSlice(slice)\n\tsort.Slice(values, func(i, j int) bool {\n\t\treturn values[i].FieldByName(\"Path\").String() < values[j].FieldByName(\"Path\").String()\n\t})\n\tfor i := 0; i < slice.Len(); i++ {\n\t\tslice.Index(i).Set(values[i])\n\t}\n}", "func sortVersions(slice []string) []string {\n\tsort.Sort(VersionSlice(slice))\n\treturn slice\n}", "func (a KubectlLayerApplier) GetResources(ctx context.Context, layer layers.Layer) (resources []kraanv1alpha1.Resource, err error) {\n\tlogging.TraceCall(a.getLog(layer))\n\tdefer logging.TraceExit(a.getLog(layer))\n\n\tsourceHrs, clusterHrs, err := a.GetSourceAndClusterHelmReleases(ctx, layer)\n\tif err != nil {\n\t\treturn nil, errors.WithMessagef(err, \"%s - failed to get helm releases\", logging.CallerStr(logging.Me))\n\t}\n\n\tfor key, source := range sourceHrs {\n\t\tresource := kraanv1alpha1.Resource{\n\t\t\tNamespace: source.GetNamespace(),\n\t\t\tName: source.GetName(),\n\t\t\tKind: \"helmreleases.helm.toolkit.fluxcd.io\",\n\t\t\tLastTransitionTime: metav1.Now(),\n\t\t\tStatus: \"Unknown\",\n\t\t}\n\t\thr, ok := clusterHrs[key]\n\t\tif ok {\n\t\t\ta.logDebug(\"HelmRelease in AddonsLayer source directory and on cluster\", layer, logging.GetObjKindNamespaceName(source)...)\n\t\t\tresources = append(resources, a.getResourceInfo(layer, resource, hr.Status.Conditions))\n\t\t} else {\n\t\t\t// this resource exists in the source directory but not on the cluster\n\t\t\ta.logDebug(\"HelmRelease in AddonsLayer source directory but not on cluster\", layer, logging.GetObjKindNamespaceName(source)...)\n\t\t\tresource.Status = kraanv1alpha1.NotDeployed\n\t\t\tresources = append(resources, resource)\n\t\t}\n\t}\n\n\tfor key, hr := range clusterHrs {\n\t\tresource := kraanv1alpha1.Resource{\n\t\t\tNamespace: hr.GetNamespace(),\n\t\t\tName: hr.GetName(),\n\t\t\tKind: \"helmreleases.helm.toolkit.fluxcd.io\",\n\t\t\tLastTransitionTime: metav1.Now(),\n\t\t\tStatus: \"Unknown\",\n\t\t}\n\t\t_, ok := sourceHrs[key]\n\t\tif !ok {\n\t\t\ta.logDebug(\"HelmRelease not in AddonsLayer source directory but on cluster\", layer, \"name\", clusterHrs[key])\n\t\t\tresources = append(resources, a.getResourceInfo(layer, resource, hr.Status.Conditions))\n\t\t}\n\t}\n\treturn resources, err\n}", "func ActionApiResources() carapace.Action {\n\treturn carapace.ActionCallback(func(c carapace.Context) carapace.Action {\n\t\treturn carapace.ActionExecCommand(\"kubectl\", \"api-resources\", \"--output=name\", \"--cached\")(func(output []byte) carapace.Action {\n\t\t\tlines := strings.Split(string(output), \"\\n\")\n\t\t\tfor index, line := range lines {\n\t\t\t\tlines[index] = strings.SplitN(line, \".\", 2)[0]\n\t\t\t}\n\t\t\treturn carapace.ActionValues(lines[:len(lines)-1]...)\n\t\t})\n\t})\n}", "func sortProjects(projects []*projectInfo) {\n\tsort.Sort(ByName{projects})\n}", "func (m *mapper) fetchGroupVersionResources(groupName string, versions ...string) (map[schema.GroupVersion]*metav1.APIResourceList, error) {\n\tgroupVersionResources := make(map[schema.GroupVersion]*metav1.APIResourceList)\n\tfailedGroups := make(map[schema.GroupVersion]error)\n\n\tfor _, version := range versions {\n\t\tgroupVersion := schema.GroupVersion{Group: groupName, Version: version}\n\n\t\tapiResourceList, err := m.client.ServerResourcesForGroupVersion(groupVersion.String())\n\t\tif err != nil {\n\t\t\tfailedGroups[groupVersion] = err\n\t\t}\n\t\tif apiResourceList != nil {\n\t\t\t// even in case of error, some fallback might have been returned.\n\t\t\tgroupVersionResources[groupVersion] = apiResourceList\n\t\t}\n\t}\n\n\tif len(failedGroups) > 0 {\n\t\treturn nil, &discovery.ErrGroupDiscoveryFailed{Groups: failedGroups}\n\t}\n\n\treturn groupVersionResources, nil\n}", "func sortModules(modules []models.Module) []models.Module {\n\tfor i, m := range modules {\n\t\tif m.Root {\n\t\t\tmodules = append(modules[:i], modules[i+1:]...)\n\t\t\treturn append([]models.Module{m}, modules...)\n\t\t}\n\t}\n\n\treturn modules\n}", "func sortChannelsByName(pkg *apimanifests.PackageManifest) {\n\tsort.Slice(pkg.Channels, func(i int, j int) bool {\n\t\treturn pkg.Channels[i].Name < pkg.Channels[j].Name\n\t})\n}", "func sortRevisions(revisionList *servingv1.RevisionList) {\n\t// sort revisionList by configuration generation key\n\tsort.SliceStable(revisionList.Items, revisionListSortFunc(revisionList))\n}", "func (cm *CategoryMap) Sort() {\n\tsort.Ints(cm.index)\n}", "func XlaKeyValueSort(scope *Scope, keys tf.Output, values tf.Output) (sorted_keys tf.Output, sorted_values tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"XlaKeyValueSort\",\n\t\tInput: []tf.Input{\n\t\t\tkeys, values,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0), op.Output(1)\n}", "func (s *ServicesWidget) Sort() {\n\ts.Lock()\n\tdefer s.Unlock()\n\tswitch s.sortMode {\n\tcase docker.SortByServiceName:\n\t\ts.sortMode = docker.SortByServiceImage\n\tcase docker.SortByServiceImage:\n\t\ts.sortMode = docker.SortByServiceName\n\t}\n}", "func sortDependencies(ctx context.Context, ranker DepsRanker, missingRuleDeps map[*bazel.Rule]map[ClassName][]bazel.Label) {\n\tstopwatch := time.Now()\n\tfor _, classToLabels := range missingRuleDeps {\n\t\tfor _, labels := range classToLabels {\n\t\t\tsort.Slice(labels, func(i, j int) bool { return ranker.Less(ctx, labels[i], labels[j]) })\n\t\t}\n\t}\n\tlog.Printf(\"Ranking dependencies (%dms)\", int64(time.Now().Sub(stopwatch)/time.Millisecond))\n}", "func (o Actions) Sort() {\n\tfor k := range o {\n\t\tsort.Sort(o[k])\n\t}\n}", "func (cache *Cache) Keys () []string {\n keys := make([]string, len(cache.FilePrints))\n n := 0;\n for k := range cache.FilePrints { keys[n] = k; n++ }\n if len(keys) > 1 { sort.Strings(keys) }\n return keys\n}", "func orderStackGroupKeys(m map[string]StackGroup) []string {\n\tret := make([]string, len(m))\n\ti := 0\n\n\tfor k := range m {\n\t\tret[i] = k\n\t\ti++\n\t}\n\tsort.Sort(sort.Reverse(sort.StringSlice(ret)))\n\treturn ret\n}", "func (l *PackageList) PrepareIndex() {\n\tif l.indexed {\n\t\treturn\n\t}\n\n\tl.packagesIndex = make([]*Package, l.Len())\n\tl.providesIndex = make(map[string][]*Package, 128)\n\n\ti := 0\n\tfor _, p := range l.packages {\n\t\tl.packagesIndex[i] = p\n\t\ti++\n\n\t\tfor _, provides := range p.Provides {\n\t\t\tl.providesIndex[provides] = append(l.providesIndex[provides], p)\n\t\t}\n\t}\n\n\tsort.Sort(l)\n\n\tl.indexed = true\n}", "func (d *Dry) SortImages() {\n\td.state.mutex.Lock()\n\tdefer d.state.mutex.Unlock()\n\tswitch d.state.SortImagesMode {\n\tcase drydocker.SortImagesByRepo:\n\t\td.state.SortImagesMode = drydocker.SortImagesByID\n\tcase drydocker.SortImagesByID:\n\t\td.state.SortImagesMode = drydocker.SortImagesByCreationDate\n\tcase drydocker.SortImagesByCreationDate:\n\t\td.state.SortImagesMode = drydocker.SortImagesBySize\n\tcase drydocker.SortImagesBySize:\n\t\td.state.SortImagesMode = drydocker.SortImagesByRepo\n\n\tdefault:\n\t}\n\td.dockerDaemon.SortImages(d.state.SortImagesMode)\n\td.state.changed = true\n\n}", "func (ns *Namespace) Sort(ctx context.Context, l any, args ...any) (any, error) {\n\tif l == nil {\n\t\treturn nil, errors.New(\"sequence must be provided\")\n\t}\n\n\tseqv, isNil := indirect(reflect.ValueOf(l))\n\tif isNil {\n\t\treturn nil, errors.New(\"can't iterate over a nil value\")\n\t}\n\n\tctxv := reflect.ValueOf(ctx)\n\n\tvar sliceType reflect.Type\n\tswitch seqv.Kind() {\n\tcase reflect.Array, reflect.Slice:\n\t\tsliceType = seqv.Type()\n\tcase reflect.Map:\n\t\tsliceType = reflect.SliceOf(seqv.Type().Elem())\n\tdefault:\n\t\treturn nil, errors.New(\"can't sort \" + reflect.ValueOf(l).Type().String())\n\t}\n\n\tcollator := langs.GetCollator1(ns.deps.Conf.Language())\n\n\t// Create a list of pairs that will be used to do the sort\n\tp := pairList{Collator: collator, sortComp: ns.sortComp, SortAsc: true, SliceType: sliceType}\n\tp.Pairs = make([]pair, seqv.Len())\n\n\tvar sortByField string\n\tfor i, l := range args {\n\t\tdStr, err := cast.ToStringE(l)\n\t\tswitch {\n\t\tcase i == 0 && err != nil:\n\t\t\tsortByField = \"\"\n\t\tcase i == 0 && err == nil:\n\t\t\tsortByField = dStr\n\t\tcase i == 1 && err == nil && dStr == \"desc\":\n\t\t\tp.SortAsc = false\n\t\tcase i == 1:\n\t\t\tp.SortAsc = true\n\t\t}\n\t}\n\tpath := strings.Split(strings.Trim(sortByField, \".\"), \".\")\n\n\tswitch seqv.Kind() {\n\tcase reflect.Array, reflect.Slice:\n\t\tfor i := 0; i < seqv.Len(); i++ {\n\t\t\tp.Pairs[i].Value = seqv.Index(i)\n\t\t\tif sortByField == \"\" || sortByField == \"value\" {\n\t\t\t\tp.Pairs[i].Key = p.Pairs[i].Value\n\t\t\t} else {\n\t\t\t\tv := p.Pairs[i].Value\n\t\t\t\tvar err error\n\t\t\t\tfor i, elemName := range path {\n\t\t\t\t\tv, err = evaluateSubElem(ctxv, v, elemName)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn nil, err\n\t\t\t\t\t}\n\t\t\t\t\tif !v.IsValid() {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\t// Special handling of lower cased maps.\n\t\t\t\t\tif params, ok := v.Interface().(maps.Params); ok {\n\t\t\t\t\t\tv = reflect.ValueOf(params.GetNested(path[i+1:]...))\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tp.Pairs[i].Key = v\n\t\t\t}\n\t\t}\n\n\tcase reflect.Map:\n\t\tkeys := seqv.MapKeys()\n\t\tfor i := 0; i < seqv.Len(); i++ {\n\t\t\tp.Pairs[i].Value = seqv.MapIndex(keys[i])\n\n\t\t\tif sortByField == \"\" {\n\t\t\t\tp.Pairs[i].Key = keys[i]\n\t\t\t} else if sortByField == \"value\" {\n\t\t\t\tp.Pairs[i].Key = p.Pairs[i].Value\n\t\t\t} else {\n\t\t\t\tv := p.Pairs[i].Value\n\t\t\t\tvar err error\n\t\t\t\tfor i, elemName := range path {\n\t\t\t\t\tv, err = evaluateSubElem(ctxv, v, elemName)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn nil, err\n\t\t\t\t\t}\n\t\t\t\t\tif !v.IsValid() {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\t// Special handling of lower cased maps.\n\t\t\t\t\tif params, ok := v.Interface().(maps.Params); ok {\n\t\t\t\t\t\tv = reflect.ValueOf(params.GetNested(path[i+1:]...))\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tp.Pairs[i].Key = v\n\t\t\t}\n\t\t}\n\t}\n\n\tcollator.Lock()\n\tdefer collator.Unlock()\n\n\treturn p.sort(), nil\n}", "func sortObjectsByKindAndName(objs []*Object) []*Object {\n\tsort.SliceStable(objs, func(i, j int) bool {\n\t\ta := objs[i]\n\t\tb := objs[j]\n\n\t\taKind := ObjectKind(a)\n\t\tbKind := ObjectKind(b)\n\t\taName, err := ObjectName(a)\n\t\tif err != nil {\n\t\t\treturn false // Move a to end of slice\n\t\t}\n\t\tbName, err := ObjectName(b)\n\t\tif err != nil {\n\t\t\treturn true // Move b to end of slice\n\t\t}\n\n\t\tif aKind == bKind {\n\t\t\treturn aName < bName\n\t\t}\n\t\treturn aKind < bKind\n\t})\n\treturn objs\n}", "func indexPackages(pkgs map[string]*ast.Package, prefix string) error {\n\tfor name, pkg := range pkgs {\n\t\tpath := prefix\n pack := name\n\t\t//fmt.Println(\"Inspecting \", path)\n\n\t\tast.Inspect(pkg, func(n ast.Node) bool {\n\n\t\t\tswitch x := n.(type) {\n\t\t\t//Packages\n\t\t\tcase *ast.Package:\n\t\t\t\tif x.Name != \"\" {\n\t\t\t\t\t//update index and docMap if necessary\n\t\t\t\t\tdocTerm := updateIndex(x.Name, pack, path)\n\t\t\t\t\t//update docTerm\n\t\t\t\t\tdocTerm.Packages += 1\n\t\t\t\t}\n\t\t\t\tbreak\n\n\t\t\t//Imports\n\t\t\tcase *ast.ImportSpec:\n\t\t\t\tif x.Path.Value != \"\" {\n\t\t\t\t\t//update index and docMap if necessary\n\t\t\t\t\tdocTerm := updateIndex(strings.Replace(x.Path.Value, \"\\\"\", \"\", -1), pack, path)\n\t\t\t\t\t//update docTerm\n\t\t\t\t\tdocTerm.Imports += 1\n\t\t\t\t}\n\t\t\t\tbreak\n\n\t\t\t//Functions\n\t\t\tcase *ast.FuncDecl:\n\t\t\t\tif x.Name.Name != \"\" {\n\t\t\t\t\t//Name tokenize function\n\t\t\t\t\tfor _, n := range tokenizeCamelCase(x.Name.Name) {\n\t\t\t\t\t\t//update index and docMap if necessary\n\t\t\t\t\t\tdocTerm := updateIndex(n, pack, path)\n\t\t\t\t\t\t//update docTerm\n\t\t\t\t\t\tdocTerm.Functions += 1\n\t\t\t\t\t}\n\n\t\t\t\t\t//Add comments to index\n\t\t\t\t\tif x.Doc != nil && *commentParse {\n\t\t\t\t\t\tcomment := \"\"\n\t\t\t\t\t\tfor _, c := range x.Doc.List {\n\t\t\t\t\t\t\tcomment += c.Text\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\tcomment = strings.Replace(comment, \"//\", \"\", -1)\n\t\t\t\t\t\tcomment = strings.ToLower(comment)\n\n\t\t\t\t\t\twords := strings.Fields(comment)\n\n\t\t\t\t\t\tfor _, word := range words {\n\t\t\t\t\t\t\tdocTerm := updateIndex(word, pack, path)\n\t\t\t\t\t\t\tdocTerm.Functions += 1\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tbreak\n\n\t\t\tcase *ast.TypeSpec:\n\t\t\t\tif x.Name.Name != \"\" {\n\t\t\t\t\t//Name tokenize function\n\t\t\t\t\tfor _, n := range tokenizeCamelCase(x.Name.Name) {\n\t\t\t\t\t\t//update index and docMap if necessary\n\t\t\t\t\t\tdocTerm := updateIndex(n, pack, path)\n\t\t\t\t\t\t//update docTerm\n\t\t\t\t\t\tdocTerm.Types += 1\n\t\t\t\t\t}\n\n\t\t\t\t\t//Add comments to index\n\t\t\t\t\tif x.Doc != nil && *commentParse {\n\t\t\t\t\t\tcomment := \"\"\n\t\t\t\t\t\tfor _, c := range x.Doc.List {\n\t\t\t\t\t\t\tcomment += c.Text\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\tcomment = strings.Replace(comment, \"//\", \"\", -1)\n\t\t\t\t\t\tcomment = strings.ToLower(comment)\n\n\t\t\t\t\t\twords := strings.Fields(comment)\n\n\t\t\t\t\t\tfor _, word := range words {\n\t\t\t\t\t\t\tdocTerm := updateIndex(word, pack, path)\n\t\t\t\t\t\t\tdocTerm.Types += 1\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tbreak\n\t\t\t}\n\t\t\treturn true\n\t\t})\n\t}\n\n\treturn nil\n}", "func sortKeys(m map[int]int) []int {\r\n\t// TODO: Implement sortKeys function.\r\n\tkeys:= make([]int,0,len(m))\r\n\tfor k := range m {\r\n\t\tkeys = append(keys,k)\r\n\t}\r\n su:=SortUser{m,keys}\r\n sort.Sort(&su)\r\n\treturn su.Keys\r\n}", "func (r *Compare) Compare() (map[string][]schema.GroupVersionResource, error) {\n\tpreferredSrcResourceList, err := collectPreferredResources(r.SrcDiscovery)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsrcCRDResource, err := collectPreferredCRDResource(r.SrcDiscovery)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdstResourceList, err := collectNamespacedResources(r.DstDiscovery)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tpreferredSrcResourceList, err = r.excludeCRDs(preferredSrcResourceList, srcCRDResource, r.SrcClient)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresourcesDiff := r.compareResources(preferredSrcResourceList, dstResourceList)\n\tincompatibleGVKs, err := convertToGVRList(resourcesDiff)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Don't report an incompatibleGVK if user settings will skip resource anyways\n\texcludedResources := toStringSlice(settings.ExcludedInitialResources.Union(toSet(r.Plan.Status.ExcludedResources)))\n\tfilteredGVKs := []schema.GroupVersionResource{}\n\tfor _, gvr := range incompatibleGVKs {\n\t\tskip := false\n\t\tfor _, resource := range excludedResources {\n\t\t\tif strings.EqualFold(gvr.Resource, resource) {\n\t\t\t\tskip = true\n\t\t\t}\n\t\t}\n\t\tif !skip {\n\t\t\tfilteredGVKs = append(filteredGVKs, gvr)\n\t\t}\n\t}\n\n\treturn r.collectIncompatibleMapping(filteredGVKs)\n}", "func sortExtras(extras map[interface{}][]interface{}) (ret []extra) {\n\tfor k, v := range extras {\n\t\tret = append(ret, extra{k, v})\n\t}\n\tsort.Slice(ret, func(i, j int) bool {\n\t\treturn fmt.Sprint(ret[i].Key) < fmt.Sprint(ret[j].Key)\n\t})\n\treturn\n}", "func orderedPropertyKeys(required []string, m map[string]spec.Schema, isResource bool) []string {\n\tsort.Strings(required)\n\n\tif isResource {\n\t\tmkeys := make(map[string]struct{})\n\t\tfor k := range m {\n\t\t\tmkeys[k] = struct{}{}\n\t\t}\n\t\tfor _, special := range []string{\"metadata\", \"kind\", \"apiVersion\"} {\n\t\t\tif !isRequired(special, required) {\n\t\t\t\tif _, ok := mkeys[special]; ok {\n\t\t\t\t\trequired = append([]string{special}, required...)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\tkeys := make([]string, len(m)-len(required))\n\ti := 0\n\tfor k := range m {\n\t\tif !isRequired(k, required) {\n\t\t\tkeys[i] = k\n\t\t\ti++\n\t\t}\n\t}\n\tsort.Strings(keys)\n\treturn append(required, keys...)\n}", "func (c *Clients) GetKubeResources(r *ReleaseData) (map[string]interface{}, error) {\n\tlog.Printf(\"Getting resources for %s\", r.Name)\n\tif r.Manifest == \"\" {\n\t\treturn nil, errors.New(\"manifest not provided in the request\")\n\t}\n\tresources := map[string]interface{}{}\n\tinfos, err := c.getManifestDetails(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tnamespace := \"default\"\n\tfor _, info := range infos {\n\t\tvar spec interface{}\n\t\tkind := info.Object.GetObjectKind().GroupVersionKind().GroupKind().Kind\n\t\tv := kube.AsVersioned(info)\n\t\tif checkSize(resources, ResourcesOutputSize) {\n\t\t\tbreak\n\t\t}\n\n\t\tif stringInSlice(reflect.TypeOf(v).String(), ResourcesOutputIgnoredTypes) {\n\t\t\tcontinue\n\t\t}\n\t\tinner := make(map[string]interface{})\n\t\tname, ok := ScanFromStruct(v, \"ObjectMeta.Name\")\n\t\tif !ok {\n\t\t\tcontinue\n\t\t}\n\t\tns, ok := ScanFromStruct(v, \"ObjectMeta.Namespace\")\n\t\tif ok {\n\t\t\tnamespace = fmt.Sprint(ns)\n\t\t}\n\t\tif stringInSlice(reflect.TypeOf(v).String(), ResourcesOutputIncludedSpec) {\n\t\t\tspec, ok = ScanFromStruct(v, \"Spec\")\n\t\t\tif ok {\n\t\t\t\tspec = structToMap(spec)\n\t\t\t}\n\t\t}\n\t\tstatus, ok := ScanFromStruct(v, \"Status\")\n\t\tif ok {\n\t\t\tstatus = structToMap(status)\n\t\t}\n\t\tinner = map[string]interface{}{\n\t\t\tfmt.Sprint(name): map[string]interface{}{\n\t\t\t\t\"Namespace\": namespace,\n\t\t\t\t\"Spec\": spec,\n\t\t\t\t\"Status\": status,\n\t\t\t},\n\t\t}\n\t\tif IsZero(resources[kind]) {\n\t\t\tresources[kind] = map[string]interface{}{}\n\t\t}\n\t\ttemp := resources[kind].(map[string]interface{})\n\t\tresources[kind] = mergeMaps(temp, inner)\n\t}\n\treturn resources, nil\n}", "func Sort(releasePackages []Compilable) ([]Compilable, error) {\n\tsortedPackages := []Compilable{}\n\n\tincomingEdges, outgoingEdges := getEdgeMaps(releasePackages)\n\tnoIncomingEdgesSet := []Compilable{}\n\n\tfor pkg, edgeList := range incomingEdges {\n\t\tif len(edgeList) == 0 {\n\t\t\tnoIncomingEdgesSet = append(noIncomingEdgesSet, pkg)\n\t\t}\n\t}\n\tfor len(noIncomingEdgesSet) > 0 {\n\t\telem := noIncomingEdgesSet[0]\n\t\tnoIncomingEdgesSet = noIncomingEdgesSet[1:]\n\n\t\tsortedPackages = append([]Compilable{elem}, sortedPackages...)\n\n\t\tfor _, pkg := range outgoingEdges[elem] {\n\t\t\tincomingEdges[pkg] = removeFromList(incomingEdges[pkg], elem)\n\t\t\tif len(incomingEdges[pkg]) == 0 {\n\t\t\t\tnoIncomingEdgesSet = append(noIncomingEdgesSet, pkg)\n\t\t\t}\n\t\t}\n\t}\n\tfor _, edges := range incomingEdges {\n\t\tif len(edges) > 0 {\n\t\t\treturn nil, errors.New(\"Circular dependency detected while sorting packages\")\n\t\t}\n\t}\n\treturn sortedPackages, nil\n}", "func IndexResourcesByName(items []envoy_types.ResourceWithTTL) map[string]envoy_types.ResourceWithTTL {\n\tindexed := make(map[string]envoy_types.ResourceWithTTL, len(items))\n\tfor _, item := range items {\n\t\tkey := GetResourceName(item.Resource)\n\t\tindexed[key] = item\n\t}\n\treturn indexed\n}", "func IndexResourcesByName(items []types.Resource) map[string]types.Resource {\n\tindexed := make(map[string]types.Resource, len(items))\n\tfor _, item := range items {\n\t\tindexed[GetResourceName(item)] = item\n\t}\n\treturn indexed\n}", "func (c *clusterCache) listResources(ctx context.Context, resClient dynamic.ResourceInterface, callback func(*pager.ListPager) error) (string, error) {\n\tif err := c.listSemaphore.Acquire(ctx, 1); err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer c.listSemaphore.Release(1)\n\tvar retryCount int64 = 0\n\tresourceVersion := \"\"\n\tlistPager := pager.New(func(ctx context.Context, opts metav1.ListOptions) (runtime.Object, error) {\n\t\tvar res *unstructured.UnstructuredList\n\t\tvar listRetry wait.Backoff\n\n\t\tif c.listRetryUseBackoff {\n\t\t\tlistRetry = retry.DefaultBackoff\n\t\t} else {\n\t\t\tlistRetry = retry.DefaultRetry\n\t\t}\n\n\t\tlistRetry.Steps = int(c.listRetryLimit)\n\t\terr := retry.OnError(listRetry, c.listRetryFunc, func() error {\n\t\t\tvar ierr error\n\t\t\tres, ierr = resClient.List(ctx, opts)\n\t\t\tif ierr != nil {\n\t\t\t\t// Log out a retry\n\t\t\t\tif c.listRetryLimit > 1 && c.listRetryFunc(ierr) {\n\t\t\t\t\tretryCount += 1\n\t\t\t\t\tc.log.Info(fmt.Sprintf(\"Error while listing resources: %v (try %d/%d)\", ierr, retryCount, c.listRetryLimit))\n\t\t\t\t}\n\t\t\t\treturn ierr\n\t\t\t}\n\t\t\tresourceVersion = res.GetResourceVersion()\n\t\t\treturn nil\n\t\t})\n\t\treturn res, err\n\t})\n\tlistPager.PageBufferSize = c.listPageBufferSize\n\tlistPager.PageSize = c.listPageSize\n\n\treturn resourceVersion, callback(listPager)\n}", "func SortImages(images []image.Info, pattern policy.Pattern) SortedImageInfos {\n\treturn sortImages(images, pattern)\n}", "func (tf tFiles) sortByNum() {\n\tsort.Sort(&tFilesSortByNum{tFiles: tf})\n}", "func (*ListRequest_SortingParams) Descriptor() ([]byte, []int) {\n\treturn file_api_proto_rawDescGZIP(), []int{2, 1}\n}", "func (r *EventList) sort() {\n\tsort.Slice(\n\t\tr.Items,\n\t\tfunc(i, j int) bool {\n\t\t\treturn r.Items[i].id() < r.Items[j].id()\n\t\t})\n}", "func getDeviceResources(deviceGroupStats []*api.DeviceGroupStats) []string {\n\tstatsSummaryMap := buildDeviceStatsSummaryMap(deviceGroupStats)\n\n\tresult := make([]string, 0, len(statsSummaryMap))\n\tfor id, stats := range statsSummaryMap {\n\t\tresult = append(result, id+\"|\"+stats.String())\n\t}\n\n\tsort.Strings(result)\n\n\treturn result\n}", "func SortFilesByDate(files []File) []File {\n\tsort.Slice(files, func(i, j int) bool {\n\t\tif files[i].State == Uploaded {\n\t\t\treturn files[i].UploadedTimestamp > files[j].UploadedTimestamp\n\t\t}\n\t\treturn files[i].PublishedTimestamp > files[j].PublishedTimestamp\n\t})\n\treturn files\n}", "func (*ListPostsRequest_Sort) Descriptor() ([]byte, []int) {\n\treturn file_view_grpc_blog_api_proto_rawDescGZIP(), []int{19, 1}\n}", "func (g GroupedResponses) SortedFileNames() []string {\n\tvar keys []string\n\tfor k := range g {\n\t\tkeys = append(keys, k)\n\t}\n\tsort.Strings(keys)\n\treturn keys\n}", "func (self *Encoder) SortKeys() *Encoder {\n self.Opts |= SortMapKeys\n return self\n}", "func (regionEnv *RegionEnv) sortUnstructured() {\n\tvar newList []*unstructured.Unstructured\n\n\tfor _, obj := range regionEnv.Unstructured {\n\t\tapiVersion := obj.GetAPIVersion()\n\t\t// Don't add to the list if it is the wrong networking object or if it has pods.\n\t\tif (apiVersion == \"networking.istio.io/v1beta1\" && regionEnv.ClusterSettings[\"ISTIO\"] != \"true\") ||\n\t\t\t(apiVersion == \"contour.heptio.com/v1beta1\" && regionEnv.ClusterSettings[\"ISTIO\"] == \"true\") {\n\t\t\tcontinue\n\t\t}\n\t\tnewList = append(newList, obj)\n\t}\n\tregionEnv.Unstructured = newList\n}", "func DependencySort(ks []HelmRelease) ([]HelmRelease, error) {\n\tn := make(graph)\n\tlookup := map[string]*HelmRelease{}\n\tfor i := 0; i < len(ks); i++ {\n\t\tn[ks[i].Name] = after(ks[i].Spec.DependsOn)\n\t\tlookup[ks[i].Name] = &ks[i]\n\t}\n\tsccs := tarjanSCC(n)\n\tvar sorted []HelmRelease\n\tvar unsortable CircularDependencyError\n\tfor i := 0; i < len(sccs); i++ {\n\t\ts := sccs[i]\n\t\tif len(s) != 1 {\n\t\t\tunsortable = append(unsortable, s)\n\t\t\tcontinue\n\t\t}\n\t\tif k, ok := lookup[s[0]]; ok {\n\t\t\tsorted = append(sorted, *k.DeepCopy())\n\t\t}\n\t}\n\tif unsortable != nil {\n\t\tfor i, j := 0, len(unsortable)-1; i < j; i, j = i+1, j-1 {\n\t\t\tunsortable[i], unsortable[j] = unsortable[j], unsortable[i]\n\t\t}\n\t\treturn nil, unsortable\n\t}\n\treturn sorted, nil\n}", "func (a *Account) Resources() ([]string, error) {\n\tt := a.SkyTable()\n\tresults, err := t.Query(\"SELECT count() GROUP BY resource\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresources := make([]string, 0, len(results))\n\tif results, ok := results[\"resource\"].(map[string]interface{}); ok {\n\t\tfor resource, _ := range results {\n\t\t\tresources = append(resources, resource)\n\t\t}\n\t}\n\tsort.Sort(sort.StringSlice(resources))\n\treturn resources, nil\n}", "func SortRecords(r []Record) {\n\tsort.Slice(r, func(i, j int) bool {\n\t\treturn r[i].Version < r[j].Version\n\t})\n}", "func (p *Proxy) SortKey() string {\n\treturn fmt.Sprintf(\"%s %s:%d/%s\", p.Name, p.ListenIP, p.ListenPort, p.ListenProto)\n}", "func (res *Resource) SortableAttrs(columns ...string) []string {\n\tif len(columns) != 0 || res.sortableAttrs == nil {\n\t\tif len(columns) == 0 {\n\t\t\tcolumns = res.ConvertSectionToStrings(res.indexSections)\n\t\t}\n\t\tres.sortableAttrs = &[]string{}\n\t\tscope := res.GetAdmin().Config.DB.NewScope(res.Value)\n\t\tfor _, column := range columns {\n\t\t\tif field, ok := scope.FieldByName(column); ok && field.DBName != \"\" {\n\t\t\t\tattrs := append(*res.sortableAttrs, column)\n\t\t\t\tres.sortableAttrs = &attrs\n\t\t\t}\n\t\t}\n\t}\n\treturn *res.sortableAttrs\n}", "func (AppImageConfigSortKey) Values() []AppImageConfigSortKey {\n\treturn []AppImageConfigSortKey{\n\t\t\"CreationTime\",\n\t\t\"LastModifiedTime\",\n\t\t\"Name\",\n\t}\n}", "func sortTags(tags []string) {\n\tsort.Slice(tags, func(i, j int) bool {\n\t\timatch := semverRegex.FindStringSubmatch(tags[i])\n\t\tjmatch := semverRegex.FindStringSubmatch(tags[j])\n\t\tif len(imatch) < 5 {\n\t\t\treturn false\n\t\t}\n\t\tif len(jmatch) < 5 {\n\t\t\treturn true\n\t\t}\n\n\t\t// Matches must be numbers due to regex they are parsed from.\n\t\tiM, _ := strconv.Atoi(imatch[1])\n\t\tjM, _ := strconv.Atoi(jmatch[1])\n\t\tim, _ := strconv.Atoi(imatch[2])\n\t\tjm, _ := strconv.Atoi(jmatch[2])\n\t\tip, _ := strconv.Atoi(imatch[3])\n\t\tjp, _ := strconv.Atoi(jmatch[3])\n\n\t\t// weight each level of semver for comparison\n\t\tiTotal := iM*marjorWeight + im*minorWeight + ip*patchWeight\n\t\tjTotal := jM*marjorWeight + jm*minorWeight + jp*patchWeight\n\n\t\t// de-rank all prereleases by a major version\n\t\tif imatch[4] != \"\" {\n\t\t\tiTotal -= marjorWeight\n\t\t}\n\t\tif jmatch[4] != \"\" {\n\t\t\tjTotal -= marjorWeight\n\t\t}\n\n\t\treturn iTotal > jTotal\n\t})\n}", "func getFilteredAPIGroupResources(cl discovery.DiscoveryInterface, groupFilterPredicate GroupFilterPredicate) ([]*restmapper.APIGroupResources, error) {\n\tgs, rs, err := filteredServerGroupsAndResources(cl, groupFilterPredicate)\n\tif rs == nil || gs == nil {\n\t\treturn nil, err\n\t\t// TODO track the errors and update callers to handle partial errors.\n\t}\n\trsm := map[string]*metav1.APIResourceList{}\n\tfor _, r := range rs {\n\t\trsm[r.GroupVersion] = r\n\t}\n\n\tvar result []*restmapper.APIGroupResources\n\tfor _, group := range gs {\n\t\tgroupResources := &restmapper.APIGroupResources{\n\t\t\tGroup: *group,\n\t\t\tVersionedResources: make(map[string][]metav1.APIResource),\n\t\t}\n\t\tfor _, version := range group.Versions {\n\t\t\tresources, ok := rsm[version.GroupVersion]\n\t\t\tif !ok {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tgroupResources.VersionedResources[version.Version] = resources.APIResources\n\t\t}\n\t\tresult = append(result, groupResources)\n\t}\n\treturn result, nil\n}", "func ResourceCatalogSortBy_Values() []string {\n\treturn []string{\n\t\tResourceCatalogSortByCreationTime,\n\t}\n}", "func (info ResourceInfo) Resources(opts options.Get) [][]string {\n\tincludedHeaders := info.Headers(opts)\n\tresult := make([][]string, len(info.data))\n\n\t// for each resource\n\tfor i, fieldMap := range info.data {\n\n\t\t// for each of the columns that we want to display\n\t\tline := make([]string, len(includedHeaders))\n\t\tfor j, h := range includedHeaders {\n\t\t\tval, ok := fieldMap[h]\n\t\t\tif !ok {\n\t\t\t\tval = \"\"\n\t\t\t}\n\t\t\tline[j] = val\n\t\t}\n\t\tresult[i] = line\n\t}\n\treturn result\n}", "func getResourcesForCrd(ctx context.Context, dynClient dynamic.Interface, crd *unstructured.Unstructured, log logr.Logger) ([]byte, error) {\n\tversions := getList(crd.Object, \"spec\", \"versions\")\n\tif len(versions) == 0 {\n\t\tlog.V(INFO).Info(\"crd has no version, skipping\", \"crd\", crd.GetName())\n\n\t\treturn nil, nil\n\t}\n\n\tgroup := getString(crd.Object, \"spec\", \"group\")\n\tresource := getString(crd.Object, \"spec\", \"names\", \"plural\")\n\tversion := getString(versions[0], \"name\")\n\n\tres := schema.GroupVersionResource{Group: group, Resource: resource, Version: version}\n\n\tresources, err := dynClient.Resource(res).List(ctx, metav1.ListOptions{})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to load resources for CRD %s: %w\", res, err)\n\t}\n\n\tresourceBuffer := bytes.Buffer{}\n\n\tfor j := range resources.Items {\n\t\tresourceYaml, err := ToCleanedK8sResourceYAML(&resources.Items[j])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tresourceBuffer.WriteString(\"---\\n\")\n\t\tresourceBuffer.Write(resourceYaml)\n\t}\n\n\treturn resourceBuffer.Bytes(), nil\n}", "func (um *Manager) upgradeGroupVersion(ctx context.Context, groupVersion string) error {\n\t// new client to get updated restmapper\n\tc, err := client.New(um.mgr.GetConfig(), client.Options{Scheme: um.mgr.GetScheme(), Mapper: nil})\n\tif err != nil {\n\t\treturn err\n\t}\n\tum.client = c\n\tif err := um.ensureCRDExists(ctx); err != nil {\n\t\tlog.Info(\"required crd has not been deployed \", \"CRD\", crdName)\n\t\treturn err\n\t}\n\t// get all resource kinds\n\tresourceList, err := um.getAllKinds(groupVersion)\n\tif err != nil {\n\t\t// If the resource doesn't exist, it doesn't need upgrading\n\t\tif errors.IsNotFound(err) {\n\t\t\treturn nil\n\t\t}\n\t\treturn err\n\t}\n\n\tresourceGV := strings.Split(resourceList.GroupVersion, \"/\")\n\tgroup := resourceGV[0]\n\tversion := resourceGV[1]\n\n\t// For some reason we have seen duplicate kinds, suppress that\n\tuniqueKinds := make(map[string]bool)\n\tfor i := range resourceList.APIResources {\n\t\tuniqueKinds[resourceList.APIResources[i].Kind] = true\n\t}\n\n\t// get resource for each Kind\n\tfor kind := range uniqueKinds {\n\t\tlog.Info(\"resource\", \"kind\", kind, \"group\", group, \"version\", version)\n\t\tresourceGvk := schema.GroupVersionKind{\n\t\t\tGroup: group,\n\t\t\tVersion: version,\n\t\t\tKind: kind + \"List\",\n\t\t}\n\t\tinstanceList := &unstructured.UnstructuredList{}\n\t\tinstanceList.SetGroupVersionKind(resourceGvk)\n\t\terr := um.client.List(ctx, instanceList)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tlog.Info(\"resource count\", \"count\", len(instanceList.Items))\n\t\tupdateResources := make(map[util.KindVersionName]unstructured.Unstructured, len(instanceList.Items))\n\t\t// get each resource\n\t\tfor _, item := range instanceList.Items {\n\t\t\tkey := util.GetUniqueKey(item)\n\t\t\tupdateResources[key] = item\n\t\t}\n\n\t\tif len(updateResources) > 0 {\n\t\t\turloop := &updateResourceLoop{\n\t\t\t\tur: updateResources,\n\t\t\t\tclient: um.client,\n\t\t\t\tstop: make(chan struct{}),\n\t\t\t\tstopped: make(chan struct{}),\n\t\t\t}\n\t\t\tlog.Info(\"starting update resources loop\", \"group\", group, \"version\", version, \"kind\", kind)\n\t\t\tgo urloop.update(ctx)\n\t\t}\n\t}\n\treturn nil\n}", "func ResourceKey(group, version, kind string) string {\n\tif group == \"\" {\n\t\tgroup = \"core\"\n\t}\n\treturn \"k8s_\" + ToSnake(group) + \"_\" + version + \"_\" + ToSnake(kind)\n}", "func (p *Provider) Resources() []terraform.ResourceType {\n\tkeys := make([]string, 0, len(p.ResourcesMap))\n\tfor k, _ := range p.ResourcesMap {\n\t\tkeys = append(keys, k)\n\t}\n\tsort.Strings(keys)\n\n\tresult := make([]terraform.ResourceType, 0, len(keys))\n\tfor _, k := range keys {\n\t\tresult = append(result, terraform.ResourceType{\n\t\t\tName: k,\n\t\t})\n\t}\n\n\treturn result\n}", "func uniqResources(resources []metav1.APIResource) []metav1.APIResource {\n\tseen := make(map[string]struct{}, len(resources))\n\ti := 0\n\tfor _, k := range resources {\n\t\tif _, ok := seen[k.Name]; ok {\n\t\t\tcontinue\n\t\t}\n\t\tseen[k.Name] = struct{}{}\n\t\tresources[i] = k\n\n\t\ti++\n\t}\n\treturn resources[:i]\n}" ]
[ "0.7373579", "0.67436945", "0.66443646", "0.63406587", "0.6209258", "0.605762", "0.6039649", "0.5944665", "0.5886248", "0.5750245", "0.56788707", "0.5572973", "0.55383927", "0.552214", "0.5503361", "0.55031127", "0.5499262", "0.5477553", "0.5433216", "0.53902084", "0.535577", "0.52868783", "0.5218482", "0.5204087", "0.51941985", "0.5169211", "0.51380396", "0.51134145", "0.5106708", "0.5090117", "0.50662315", "0.5052936", "0.5041121", "0.50354165", "0.5007837", "0.5007368", "0.4999884", "0.49854428", "0.49852842", "0.49726605", "0.49726474", "0.49437082", "0.4938736", "0.49363455", "0.49355167", "0.49321294", "0.49285063", "0.49049896", "0.4887764", "0.48821828", "0.48817712", "0.48803166", "0.4875778", "0.487134", "0.48707744", "0.4862691", "0.48601204", "0.48229778", "0.48090866", "0.48044574", "0.4804071", "0.4803175", "0.48019236", "0.47819176", "0.4778518", "0.4776585", "0.4771438", "0.47703734", "0.47482908", "0.4734014", "0.47294167", "0.47283155", "0.47263876", "0.47078046", "0.47041133", "0.4689539", "0.4682257", "0.4681459", "0.46722534", "0.4671265", "0.46697423", "0.4659851", "0.4644862", "0.46361467", "0.46358857", "0.46320394", "0.46299523", "0.46273217", "0.46271622", "0.46256393", "0.46070084", "0.46066692", "0.46012464", "0.4599022", "0.4597884", "0.45911103", "0.45907822", "0.4579066", "0.4575569", "0.4569986" ]
0.7210953
1
getFields looks up p.Fields from leaf and structures them into treeFields. TODO(pwittrock): simplify this function
func (p TreeWriter) getFields(leaf *yaml.RNode) (treeFields, error) { fieldsByName := map[string]*treeField{} // index nested and non-nested fields for i := range p.Fields { f := p.Fields[i] seq, err := leaf.Pipe(&f) if err != nil { return nil, err } if seq == nil { continue } if fieldsByName[f.Name] == nil { fieldsByName[f.Name] = &treeField{name: f.Name} } // non-nested field -- add directly to the treeFields list if f.SubName == "" { // non-nested field -- only 1 element val, err := yaml.String(seq.Content()[0], yaml.Trim, yaml.Flow) if err != nil { return nil, err } fieldsByName[f.Name].value = val continue } // nested-field -- create a parent elem, and index by the 'match' value if fieldsByName[f.Name].subFieldByMatch == nil { fieldsByName[f.Name].subFieldByMatch = map[string]treeFields{} } index := fieldsByName[f.Name].subFieldByMatch for j := range seq.Content() { elem := seq.Content()[j] matches := f.Matches[elem] str, err := yaml.String(elem, yaml.Trim, yaml.Flow) if err != nil { return nil, err } // map the field by the name of the element // index the subfields by the matching element so we can put all the fields for the // same element under the same branch matchKey := strings.Join(matches, "/") index[matchKey] = append(index[matchKey], &treeField{name: f.SubName, value: str}) } } // iterate over collection of all queried fields in the Resource for _, field := range fieldsByName { // iterate over collection of elements under the field -- indexed by element name for match, subFields := range field.subFieldByMatch { // create a new element for this collection of fields // note: we will convert name to an index later, but keep the match for sorting elem := &treeField{name: match} field.matchingElementsAndFields = append(field.matchingElementsAndFields, elem) // iterate over collection of queried fields for the element for i := range subFields { // add to the list of fields for this element elem.matchingElementsAndFields = append(elem.matchingElementsAndFields, subFields[i]) } } // clear this cached data field.subFieldByMatch = nil } // put the fields in a list so they are ordered fieldList := treeFields{} for _, v := range fieldsByName { fieldList = append(fieldList, v) } // sort the fields sort.Sort(fieldList) for i := range fieldList { field := fieldList[i] // sort the elements under this field sort.Sort(field.matchingElementsAndFields) for i := range field.matchingElementsAndFields { element := field.matchingElementsAndFields[i] // sort the elements under a list field by their name sort.Sort(element.matchingElementsAndFields) // set the name of the element to its index element.name = fmt.Sprintf("%d", i) } } return fieldList, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func getParentTreeFields(treePath string) (treeNames, treePaths []string) {\n\tif treePath == \"\" {\n\t\treturn treeNames, treePaths\n\t}\n\n\ttreeNames = strings.Split(treePath, \"/\")\n\ttreePaths = make([]string, len(treeNames))\n\tfor i := range treeNames {\n\t\ttreePaths[i] = strings.Join(treeNames[:i+1], \"/\")\n\t}\n\treturn treeNames, treePaths\n}", "func GetFields(reader *os.File) []Field {\n\tvar curbuf []byte\n\tdbfhead := GetDbfHead(reader)\n\n\toff := dbfhead.Headerlen - 32 - 264\n\tif off < 1 {\n\t\tpanic(\"invalid header\")\n\t}\n\tfieldlist := make([]Field, off/32)\n\tbuf := make([]byte, off)\n\t_, err := reader.ReadAt(buf, 32)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfor i, val := range fieldlist {\n\t\ta := i * 32\n\t\tcurbuf = buf[a:]\n\t\tval.Name = copyUpToNull(curbuf[0:11])\n\t\tval.Fieldtype = fmt.Sprintf(\"%s\", curbuf[11:12])\n\t\tval.FieldDataaddress = curbuf[12:16]\n\t\tval.FieldLen = Changebytetoint(curbuf[16:17])\n\t\tval.DecimalCount = curbuf[17:18]\n\t\tval.Workareaid = curbuf[20:21]\n\t\tfieldlist[i] = val\n\t}\n\treturn fieldlist\n}", "func (self *mergeState) getFields() []string {\n\tfields := make([]string, len(self.fields))\n\tfor f, i := range self.fields {\n\t\tfields[i] = f\n\t}\n\treturn fields\n}", "func (n *Node) Fields() []uintptr {\n\t// we store the offsets for the fields in type properties\n\ttprops := kit.Types.Properties(n.Type(), true) // true = makeNew\n\tpnm := \"__FieldOffs\"\n\tif foff, ok := tprops[pnm]; ok {\n\t\treturn foff.([]uintptr)\n\t}\n\tfoff := make([]uintptr, 0)\n\tkitype := KiType()\n\tFlatFieldsValueFunc(n.This, func(stru interface{}, typ reflect.Type, field reflect.StructField, fieldVal reflect.Value) bool {\n\t\tif fieldVal.Kind() == reflect.Struct && kit.EmbeddedTypeImplements(field.Type, kitype) {\n\t\t\tfoff = append(foff, field.Offset)\n\t\t}\n\t\treturn true\n\t})\n\ttprops[pnm] = foff\n\treturn foff\n}", "func (_struct *Struct) Fields() (Fields, error) {\n\tvar goFields Fields\n\tstructType := _struct.StructType()\n\tif structType == nil {\n\t\treturn nil, fmt.Errorf(\"no struct type in %#+v\", _struct)\n\t}\n\tfor idx, field := range structType.Fields.List {\n\t\ttyp, err := _struct.toType(field.Type)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"unable to lookup type '%s': %w\", field.Type, err)\n\t\t}\n\t\tgoFields = append(goFields, &Field{\n\t\t\tField: *field,\n\t\t\tStruct: _struct,\n\t\t\tNames: field.Names,\n\t\t\tIndex: uint(idx),\n\t\t\tTypeValue: typ,\n\t\t})\n\t}\n\treturn goFields, nil\n}", "func getNodeFields() []string {\n\trt := reflect.TypeOf((*tailcfg.Node)(nil)).Elem()\n\tret := make([]string, rt.NumField())\n\tfor i := 0; i < rt.NumField(); i++ {\n\t\tret[i] = rt.Field(i).Name\n\t}\n\treturn ret\n}", "func TraverseFields(v interface{}, processField ProcessValue) error {\n\treturn TraverseValueFields(reflect.ValueOf(v), processField)\n}", "func (*GetField) Children() []sql.Expression {\n\treturn nil\n}", "func fixFields(n, parent *node, depth int) {\n\tn.parent = parent\n\tn.depth = depth\n\tfor _, c := range n.children {\n\t\tfixFields(c, n, depth+1)\n\t}\n}", "func (n Node) AllFields() []interface{} {\n\tlabels := make([]interface{}, len(n.Labels))\n\tfor i, label := range n.Labels {\n\t\tlabels[i] = label\n\t}\n\treturn []interface{}{n.NodeIdentity, labels, n.Properties}\n}", "func (pce *ppdCacheEntry) getFields() (cdd.PrinterDescriptionSection, string, string, lib.DuplexVendorMap) {\n\tpce.mutex.Lock()\n\tdefer pce.mutex.Unlock()\n\treturn pce.description, pce.manufacturer, pce.model, pce.duplexMap\n}", "func typeFields(t reflect.Type) []field {\n\t// Anonymous fields to explore at the current level and the next.\n\tcurrent := []field{}\n\tnext := []field{{typ: t}}\n\n\t// Count of queued names for current level and the next.\n\tcount := map[reflect.Type]int{}\n\tnextCount := map[reflect.Type]int{}\n\n\t// Types already visited at an earlier level.\n\tvisited := map[reflect.Type]bool{}\n\n\t// Fields found.\n\tvar fields []field\n\n\tfor len(next) > 0 {\n\t\tcurrent, next = next, current[:0]\n\t\tcount, nextCount = nextCount, map[reflect.Type]int{}\n\n\t\tfor _, f := range current {\n\t\t\tif visited[f.typ] {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tvisited[f.typ] = true\n\n\t\t\t// Scan f.typ for fields to include.\n\t\t\tfor i := 0; i < f.typ.NumField(); i++ {\n\t\t\t\tsf := f.typ.Field(i)\n\t\t\t\tisUnexported := sf.PkgPath != \"\"\n\t\t\t\tif sf.Anonymous {\n\t\t\t\t\tt := sf.Type\n\t\t\t\t\tif isUnexported && t.Kind() != reflect.Struct {\n\t\t\t\t\t\t// Ignore embedded fields of unexported non-struct types.\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\t// Do not ignore embedded fields of unexported struct types\n\t\t\t\t\t// since they may have exported fields.\n\t\t\t\t} else if isUnexported {\n\t\t\t\t\t// Ignore unexported non-embedded fields.\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tindex := make([]int, len(f.index)+1)\n\t\t\t\tcopy(index, f.index)\n\t\t\t\tindex[len(f.index)] = i\n\n\t\t\t\tft := sf.Type\n\n\t\t\t\t// Record found field and index sequence.\n\t\t\t\tif !sf.Anonymous || ft.Kind() != reflect.Struct {\n\t\t\t\t\tfields = append(fields, field{\n\t\t\t\t\t\tname: sf.Name,\n\t\t\t\t\t\tindex: index,\n\t\t\t\t\t\ttyp: ft,\n\t\t\t\t\t})\n\t\t\t\t\tif count[f.typ] > 1 {\n\t\t\t\t\t\t// If there were multiple instances, add a second,\n\t\t\t\t\t\t// so that the annihilation code will see a duplicate.\n\t\t\t\t\t\t// It only cares about the distinction between 1 or 2,\n\t\t\t\t\t\t// so don't bother generating any more copies.\n\t\t\t\t\t\tfields = append(fields, fields[len(fields)-1])\n\t\t\t\t\t}\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\t// Record new anonymous struct to explore in next round.\n\t\t\t\tnextCount[ft]++\n\t\t\t\tif nextCount[ft] == 1 {\n\t\t\t\t\tnext = append(next, field{name: ft.Name(), index: index, typ: ft})\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\tsort.Sort(byIndex(fields))\n\n\treturn fields\n}", "func (s Strategy) GetFieldsForSubDocument(model string, foreignfield string) []map[string]interface{} {\n\tvar fields []map[string]interface{}\n\n\tfor _, f := range s.Map.Entities[model].Fields { // search foreign field in []map[string]interface{}\n\t\tif f[\"foreign\"] == foreignfield {\n\t\t\tfi := f[\"fields\"].([]interface{})\n\t\t\t// Convert the []interface into []map[string]interface{}\n\t\t\tfields = make([]map[string]interface{}, len(fi))\n\t\t\tfor i := range fields {\n\t\t\t\tfields[i] = fi[i].(map[string]interface{})\n\t\t\t}\n\t\t\treturn fields\n\t\t}\n\t}\n\treturn fields\n}", "func (o *NotificationAllOf) GetFields() map[string]interface{} {\n\tif o == nil || o.Fields == nil {\n\t\tvar ret map[string]interface{}\n\t\treturn ret\n\t}\n\treturn *o.Fields\n}", "func (jn *SemiJoin) GetFields(ctx context.Context, vcursor VCursor, bindVars map[string]*querypb.BindVariable) (*sqltypes.Result, error) {\n\treturn jn.Left.GetFields(ctx, vcursor, bindVars)\n}", "func TopLevelFields(paths []string) []string {\n\tseen := make(map[string]struct{}, len(paths))\n\tout := make([]string, 0, len(paths))\n\tfor _, path := range paths {\n\t\tparts := strings.SplitN(path, \".\", 2)\n\t\tif _, ok := seen[parts[0]]; ok {\n\t\t\tcontinue\n\t\t}\n\t\tseen[parts[0]] = struct{}{}\n\t\tout = append(out, parts[0])\n\t}\n\treturn out\n}", "func GetFields(q graphql.Querier, query *querybuilder.Query) ([]Field, error) {\n\ttyp, ok := schema.GetQueryType()\n\tif !ok {\n\t\treturn nil, errors.New(\"No QueryType present in schema\")\n\t}\n\n\tfor _, node := range query.List() {\n\t\tif node.ConcreteType == \"\" {\n\t\t\tfield, ok := typ.GetField(node.Name)\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"Missing field %q from type %q\", node.Name, typ.Name)\n\t\t\t}\n\n\t\t\ttyp, ok = schema.GetType(field.GetTypeName())\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"Missing type %q\", field.GetTypeName())\n\t\t\t}\n\t\t} else {\n\t\t\ttyp, ok = schema.GetType(node.ConcreteType)\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"Missing type %q\", node.ConcreteType)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn typ.Fields, nil\n}", "func (del *Delete) GetFields(context.Context, VCursor, map[string]*querypb.BindVariable) (*sqltypes.Result, error) {\n\treturn nil, fmt.Errorf(\"BUG: unreachable code for %q\", del.Query)\n}", "func (g *generator) structFields(t reflect.Type) []field {\n\tvar fields []field\n\tfor i := 0; i < t.NumField(); i++ {\n\t\tf := t.Field(i)\n\t\tif g.ignoreField(t, f) {\n\t\t\tcontinue\n\t\t}\n\t\tname, _ := parseTag(g.fieldTagKey, f.Tag)\n\t\tif name == \"\" {\n\t\t\tname = f.Name\n\t\t}\n\t\tfields = append(fields, field{\n\t\t\tName: name,\n\t\t\tType: f.Type,\n\t\t\tZero: zeroValue(f.Type),\n\t\t})\n\t}\n\treturn fields\n}", "func (e TypeReflectCacheEntry) Fields() map[string]*FieldCacheEntry {\n\treturn e.structFields\n}", "func BottomLevelFields(paths []string) []string {\n\tseen := make(map[string]struct{}, len(paths))\n\tfor _, path := range paths {\n\t\tprefix := path\n\t\tif i := strings.LastIndex(prefix, \".\"); i >= 0 {\n\t\t\tprefix = prefix[:i]\n\t\t}\n\t\tif _, ok := seen[prefix]; ok {\n\t\t\tdelete(seen, prefix)\n\t\t}\n\t\tseen[path] = struct{}{}\n\t}\n\tout := make([]string, 0, len(seen))\n\tfor k := range seen {\n\t\tout = append(out, k)\n\t}\n\treturn out\n}", "func getComparableFields() []string {\n\tfields := []string{}\n\n\tfor _, fieldName := range getContainerFields() {\n\t\t// Skip some fields\n\t\tif unicode.IsLower((rune)(fieldName[0])) {\n\t\t\tcontinue\n\t\t}\n\n\t\tskip := false\n\t\tfor _, f := range compareSkipFields {\n\t\t\tif f == fieldName {\n\t\t\t\tskip = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif !skip {\n\t\t\tfields = append(fields, fieldName)\n\t\t}\n\t}\n\n\treturn fields\n}", "func GetFields(c context.Context) Fields {\n\tret, _ := c.Value(fieldsKey).(Fields)\n\tif ret == nil {\n\t\treturn fields(nil)\n\t}\n\treturn ret\n}", "func IterFields(t *Type) (*Field, Iter)", "func (e Matcher_MatcherTreeValidationError) Field() string { return e.field }", "func (s *Search) GetFields() []string {\n\tq := s.buildQ()\n\tparams := map[string]string{\n\t\t\"query\": q,\n\t\t\"type\": \"File\",\n\t\t\"format\": \"application/solr+json\",\n\t\t\"fields\": \"*\",\n\t\t\"limit\": \"1\",\n\t}\n\n\tbody, err := s.performSearch(params)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn nil\n\t}\n\n\t// Parse response body as JSON\n\tvar result fieldResTop\n\tjson.Unmarshal(body, &result)\n\n\t// If no result was found\n\tif len(result.Res.Docs) != 1 {\n\t\treturn nil\n\t}\n\n\tvar fields []string\n\tfor key := range result.Res.Docs[0] {\n\t\tfields = append(fields, key)\n\t}\n\treturn fields\n}", "func StructFields(t reflect.Type) interface{} {\n\treturn structInfoForType(t).fields\n}", "func (fm *FieldModelOrder) GetFields(fbeValue *Order, fbeStructSize int) {\n fbeCurrentSize := 4 + 4\n\n if (fbeCurrentSize + fm.Id.FBESize()) <= fbeStructSize {\n fbeValue.Id, _ = fm.Id.Get()\n } else {\n fbeValue.Id = 0\n }\n fbeCurrentSize += fm.Id.FBESize()\n\n if (fbeCurrentSize + fm.Symbol.FBESize()) <= fbeStructSize {\n fbeValue.Symbol, _ = fm.Symbol.Get()\n } else {\n fbeValue.Symbol = \"\"\n }\n fbeCurrentSize += fm.Symbol.FBESize()\n\n if (fbeCurrentSize + fm.Side.FBESize()) <= fbeStructSize {\n _ = fm.Side.GetValue(&fbeValue.Side)\n } else {\n fbeValue.Side = *NewOrderSide()\n }\n fbeCurrentSize += fm.Side.FBESize()\n\n if (fbeCurrentSize + fm.Type.FBESize()) <= fbeStructSize {\n _ = fm.Type.GetValue(&fbeValue.Type)\n } else {\n fbeValue.Type = *NewOrderType()\n }\n fbeCurrentSize += fm.Type.FBESize()\n\n if (fbeCurrentSize + fm.Price.FBESize()) <= fbeStructSize {\n fbeValue.Price, _ = fm.Price.GetDefault(float64(0.0))\n } else {\n fbeValue.Price = float64(0.0)\n }\n fbeCurrentSize += fm.Price.FBESize()\n\n if (fbeCurrentSize + fm.Volume.FBESize()) <= fbeStructSize {\n fbeValue.Volume, _ = fm.Volume.GetDefault(float64(0.0))\n } else {\n fbeValue.Volume = float64(0.0)\n }\n fbeCurrentSize += fm.Volume.FBESize()\n}", "func (o *PublicViewInfo) GetFields() []PublicField {\n\tif o == nil {\n\t\tvar ret []PublicField\n\t\treturn ret\n\t}\n\n\treturn o.Fields\n}", "func loadFields(fieldBucket *bbolt.Bucket) (fields []field.Meta) {\n\tcursor := fieldBucket.Cursor()\n\tfor k, v := cursor.First(); k != nil; k, v = cursor.Next() {\n\t\tfields = append(fields, field.Meta{\n\t\t\tName: field.Name(k),\n\t\t\tID: field.ID(v[0]),\n\t\t\tType: field.Type(v[1]),\n\t\t})\n\t}\n\treturn\n}", "func _fields(args ...interface{}) *ast.FieldList {\n\tlist := []*ast.Field{}\n\tnames := []*ast.Ident{}\n\tlasti := interface{}(nil)\n\tmaybePop := func() {\n\t\tif len(names) > 0 {\n\t\t\tvar last ast.Expr\n\t\t\tif lastte_, ok := lasti.(string); ok {\n\t\t\t\tlast = _x(lastte_)\n\t\t\t} else {\n\t\t\t\tlast = lasti.(ast.Expr)\n\t\t\t}\n\t\t\tlist = append(list, &ast.Field{\n\t\t\t\tNames: names,\n\t\t\t\tType: last,\n\t\t\t})\n\t\t\tnames = []*ast.Ident{}\n\t\t}\n\t}\n\tfor i := 0; i < len(args); i++ {\n\t\tname, ok := args[i].(*ast.Ident)\n\t\tif !ok {\n\t\t\tname = _i(args[i].(string))\n\t\t}\n\t\tte_ := args[i+1]\n\t\ti += 1\n\t\t// NOTE: This comparison could be improved, to say, deep equality,\n\t\t// but is that the behavior we want?\n\t\tif lasti == te_ {\n\t\t\tnames = append(names, name)\n\t\t\tcontinue\n\t\t} else {\n\t\t\tmaybePop()\n\t\t\tnames = append(names, name)\n\t\t\tlasti = te_\n\t\t}\n\t}\n\tmaybePop()\n\treturn &ast.FieldList{\n\t\tList: list,\n\t}\n}", "func ReflectFieldsFq(\n\tt reflect.Type,\n\ttypeMap TypeMap,\n\texclude ExcludeFieldTag,\n) graphql.Fields {\n\tif t.Kind() != reflect.Struct {\n\t\tpanic(fmt.Sprintf(`ReflectFieldsFq can only work on struct types.\n\t\t\tReceived instead %s`, t.Kind()))\n\t}\n\tfields := make(graphql.Fields)\n\tfor i := 0; i < t.NumField(); i++ {\n\t\tf := t.Field(i)\n\t\tif includeField(f, exclude) {\n\t\t\tname := GqlName(GetFieldFirstTag(f, \"json\"))\n\t\t\tfields[string(name)] = ReflectFieldFq(name, f.Type, typeMap, exclude)\n\t\t}\n\t}\n\treturn fields\n}", "func (b *Builder) QueryFields(source reflect.Value, parent reflect.Value) (graphql.Fields, error) {\n\tresult := make(graphql.Fields, 0)\n\tif source.IsValid() && source.IsZero() {\n\t\tsource = reflect.New(source.Type())\n\t}\n\tnodes := b.buildObject(source, parent)\n\tfor _, node := range nodes {\n\t\tif node.skip {\n\t\t\tcontinue\n\t\t}\n\t\tif !node.source.CanSet() {\n\t\t\tcontinue\n\t\t}\n\t\tif node.inputOnly {\n\t\t\tcontinue\n\t\t}\n\t\tname := node.alias\n\t\tif name == \"\" {\n\t\t\tname = strcase.ToLowerCamel(node.name)\n\t\t}\n\t\tvar gType graphql.Type\n\t\tif node.isRelay {\n\t\t\tgType = b.buildConnection(node.source, parent)\n\t\t} else {\n\t\t\tgType = b.mapOutput(node.source, parent)\n\t\t}\n\t\tif gType == nil {\n\t\t\tcontinue\n\t\t}\n\t\tif node.required {\n\t\t\tgType = graphql.NewNonNull(gType)\n\t\t}\n\n\t\tfield := &graphql.Field{\n\t\t\tName: name,\n\t\t\tType: gType,\n\t\t\tDescription: node.description,\n\t\t\tResolve: node.resolver,\n\t\t\tArgs: node.resolverArgs,\n\t\t}\n\t\tresult[name] = field\n\t}\n\treturn result, nil\n}", "func (c MethodsCollection) FieldsGet() pFieldsGet {\n\treturn pFieldsGet{\n\t\tMethod: c.MustGet(\"FieldsGet\"),\n\t}\n}", "func (m *SubCategoryTemplate) GetFieldDeserializers()(map[string]func(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(error)) {\n res := m.FilePlanDescriptorTemplate.GetFieldDeserializers()\n return res\n}", "func Fields(msg proto.Message, processors ...FieldProcessor) Results {\n\tif len(processors) == 0 {\n\t\treturn nil\n\t}\n\t// 32 is a guess at how deep the Path could get.\n\t//\n\t// There's really no way to know ahead of time (since proto messages could\n\t// have a recursive structure, allowing the expression of trees, etc.)\n\treturn fieldsImpl(make(reflectutil.Path, 0, 32), msg.ProtoReflect(), lookupProcBundles(processors...))\n}", "func (r FieldRule) getFieldInfo(expr ast.Expr) fieldInfos {\n\tswitch t := expr.(type) {\n\tcase *ast.Ident:\n\t\tif t.Obj == nil || t.Obj.Decl == nil {\n\t\t\treturn nil\n\t\t}\n\n\t\treturn r.getFieldInfoFromDecl(t.Obj.Decl)\n\tdefault:\n\t\tpepperlint.Log(\"TODO: getInternalTypeSpec %T\", t)\n\t}\n\n\treturn nil\n}", "func (entry *Entry) Fields() Fields {\n\treturn entry.fields\n}", "func (LevelOfDangerous) Fields() []ent.Field {\r\n\treturn []ent.Field{\r\n\t\tfield.String(\"name\"),\r\n\t}\r\n}", "func (s *Set) GetFields(vcursor VCursor, bindVars map[string]*querypb.BindVariable) (*sqltypes.Result, error) {\n\tpanic(\"implement me\")\n}", "func (f *Field) Fields() []*Field {\n\treturn getFields(f.value, f.defaultTag)\n}", "func (d *Day16) getFieldOrder(order []field, fields []field, tickets []ticket) []field {\n\tif len(fields) == 0 && d.validOrderUntilNow(order, tickets) {\n\t\treturn order\n\t}\n\tfor i, f := range fields {\n\t\tnewOrder := make([]field, len(order), len(order)+1)\n\t\tcopy(newOrder, order)\n\t\tnewOrder = append(newOrder, f)\n\t\tif d.validOrderUntilNow(newOrder, tickets) {\n\t\t\tfieldsLeft := RemoveFieldAtIndex(fields, i)\n\t\t\tnewOrder = d.getFieldOrder(newOrder, fieldsLeft, tickets)\n\t\t\tif newOrder != nil {\n\t\t\t\treturn newOrder\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (m *ParentLabelDetails) GetFieldDeserializers()(map[string]func(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(error)) {\n res := make(map[string]func(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(error))\n res[\"color\"] = func (n i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode) error {\n val, err := n.GetStringValue()\n if err != nil {\n return err\n }\n if val != nil {\n m.SetColor(val)\n }\n return nil\n }\n res[\"description\"] = func (n i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode) error {\n val, err := n.GetStringValue()\n if err != nil {\n return err\n }\n if val != nil {\n m.SetDescription(val)\n }\n return nil\n }\n res[\"id\"] = func (n i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode) error {\n val, err := n.GetStringValue()\n if err != nil {\n return err\n }\n if val != nil {\n m.SetId(val)\n }\n return nil\n }\n res[\"isActive\"] = func (n i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode) error {\n val, err := n.GetBoolValue()\n if err != nil {\n return err\n }\n if val != nil {\n m.SetIsActive(val)\n }\n return nil\n }\n res[\"name\"] = func (n i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode) error {\n val, err := n.GetStringValue()\n if err != nil {\n return err\n }\n if val != nil {\n m.SetName(val)\n }\n return nil\n }\n res[\"@odata.type\"] = func (n i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode) error {\n val, err := n.GetStringValue()\n if err != nil {\n return err\n }\n if val != nil {\n m.SetOdataType(val)\n }\n return nil\n }\n res[\"parent\"] = func (n i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode) error {\n val, err := n.GetObjectValue(CreateParentLabelDetailsFromDiscriminatorValue)\n if err != nil {\n return err\n }\n if val != nil {\n m.SetParent(val.(ParentLabelDetailsable))\n }\n return nil\n }\n res[\"sensitivity\"] = func (n i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode) error {\n val, err := n.GetInt32Value()\n if err != nil {\n return err\n }\n if val != nil {\n m.SetSensitivity(val)\n }\n return nil\n }\n res[\"tooltip\"] = func (n i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode) error {\n val, err := n.GetStringValue()\n if err != nil {\n return err\n }\n if val != nil {\n m.SetTooltip(val)\n }\n return nil\n }\n return res\n}", "func (m *AuditLogRoot) GetFieldDeserializers()(map[string]func(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(error)) {\n res := m.Entity.GetFieldDeserializers()\n res[\"directoryAudits\"] = i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.SetCollectionOfObjectValues(CreateDirectoryAuditFromDiscriminatorValue , m.SetDirectoryAudits)\n res[\"provisioning\"] = i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.SetCollectionOfObjectValues(CreateProvisioningObjectSummaryFromDiscriminatorValue , m.SetProvisioning)\n res[\"signIns\"] = i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.SetCollectionOfObjectValues(CreateSignInFromDiscriminatorValue , m.SetSignIns)\n return res\n}", "func getAllFields(o interface{}) map[string]interface{} {\n\tt := reflect.TypeOf(o)\n\tif t.Kind() == reflect.Ptr || t.Kind() == reflect.Interface {\n\t\tt = reflect.ValueOf(o).Elem().Type()\n\t\to = reflect.ValueOf(o).Elem().Interface()\n\t}\n\n\tchanges := make(map[string]interface{})\n\n\tfor i := 0; i < t.NumField(); i++ {\n\t\tfield := t.Field(i)\n\t\tvalue, exists := field.Tag.Lookup(changeTag)\n\t\tif exists {\n\t\t\tif value == \"inner\" {\n\n\t\t\t\tchanges[field.Name] = getAllFields(reflect.ValueOf(o).FieldByName(field.Name).Interface())\n\t\t\t} else {\n\t\t\t\tjsonFieldName, exists := field.Tag.Lookup(\"json\")\n\n\t\t\t\tif !exists {\n\t\t\t\t\tjsonFieldName = field.Name\n\t\t\t\t} else {\n\t\t\t\t\tcommaIndex := strings.Index(jsonFieldName, \",\")\n\t\t\t\t\tif commaIndex != -1 {\n\t\t\t\t\t\tjsonFieldName = jsonFieldName[:commaIndex]\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tchangeField := &ChangeField{\n\t\t\t\t\tValue: getValue(o, field.Name),\n\t\t\t\t\tField: jsonFieldName,\n\t\t\t\t}\n\t\t\t\tchanges[field.Name] = changeField\n\t\t\t}\n\t\t}\n\t}\n\treturn changes\n}", "func StructFields(v interface{}) (vType reflect.Type, vFields []*TField) {\n\tvar (\n\t\tfield reflect.StructField\n\t)\n\tvType = reflect.Indirect(reflect.ValueOf(v)).Type()\n\tnumFields := vType.NumField()\n\tvFields = make([]*TField, 0, numFields)\n\tfor i := 0; i < numFields; i++ {\n\t\tfield = vType.Field(i)\n\t\tfieldInfo := &TField{\n\t\t\tFname: field.Name,\n\t\t\tFtype: field.Type.String(),\n\t\t\tFkind: field.Type.Kind(),\n\t\t\tFtags: field.Tag,\n\t\t}\n\t\tif field.PkgPath == \"\" {\n\t\t\tfieldInfo.Fexported = true\n\t\t}\n\t\tvFields = append(vFields, fieldInfo)\n\t}\n\treturn\n}", "func fields(spec *ast.TypeSpec) []*ast.Field {\n\ts := make([]*ast.Field, 0)\n\tif structType, ok := spec.Type.(*ast.StructType); ok {\n\t\tfor _, field := range structType.Fields.List {\n\t\t\tif keyname(field) != \"\" {\n\t\t\t\ts = append(s, field)\n\t\t\t}\n\t\t}\n\t}\n\treturn s\n}", "func (upd *Update) GetFields(vcursor VCursor, bindVars map[string]*querypb.BindVariable) (*sqltypes.Result, error) {\n\treturn nil, fmt.Errorf(\"BUG: unreachable code for %q\", upd.Query)\n}", "func GetFieldMap(obj interface{}) (ret map[string]string) {\n\tval := reflect.ValueOf(obj).Elem()\n\tret = make(map[string]string)\n\tfor i := 0; i < val.NumField(); i++ {\n\t\ttypeField := val.Type().Field(i)\n\t\tkey := strings.ToLower(typeField.Name)\n\t\tif typeField.PkgPath != \"\" {\n\t\t\t// Private method\n\t\t\tcontinue\n\t\t} else {\n\t\t\tret[key] = typeField.Name\n\t\t}\n\t}\n\treturn\n}", "func (e *Extractor) fields(s reflect.Value) []field {\n\tfields := make([]field, 0, s.NumField())\n\n\tfor i := 0; i < s.NumField(); i++ {\n\t\tif isIgnored(s.Type().Field(i).Name, e.ignoredFields) {\n\t\t\tcontinue\n\t\t}\n\n\t\tif s.Type().Field(i).Anonymous {\n\t\t\tif e.useEmbeddedStructs {\n\t\t\t\tfields = append(fields, e.fields(s.Field(i))...)\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\ttag := s.Type().Field(i).Tag\n\t\tname := s.Type().Field(i).Name\n\t\tvalue := s.Field(i)\n\t\tfields = append(fields, field{value, name, tag})\n\t}\n\n\treturn fields\n}", "func (e Matcher_MatcherTree_MatchMapValidationError) Field() string { return e.field }", "func (l *Lock) GetFields(vcursor VCursor, bindVars map[string]*querypb.BindVariable) (*sqltypes.Result, error) {\n\treturn nil, vterrors.New(vtrpc.Code_UNIMPLEMENTED, \"not implements in lock primitive\")\n}", "func (fm *FinalModelStructBytes) GetFields(fbeValue *StructBytes) (int, error) {\n var err error = nil\n fbeCurrentOffset := 0\n fbeCurrentSize := 0\n fbeFieldSize := 0\n\n fm.F1.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F1, fbeFieldSize, err = fm.F1.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F2.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F2, fbeFieldSize, err = fm.F2.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F3.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F3, fbeFieldSize, err = fm.F3.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n return fbeCurrentSize, err\n}", "func (i PullAll) Fields() []interface{} {\n\treturn nil\n}", "func (s *SoExtFollowingWrap) getModifiedFields(oriTable *SoExtFollowing, curTable *SoExtFollowing) (map[string]bool, bool, error) {\n\tif oriTable == nil {\n\t\treturn nil, false, errors.New(\"table info is nil, can't get modified fields\")\n\t}\n\thasWatcher := false\n\tfields := make(map[string]bool)\n\n\tif !reflect.DeepEqual(oriTable.FollowingCreatedOrder, curTable.FollowingCreatedOrder) {\n\t\tfields[\"FollowingCreatedOrder\"] = true\n\t\thasWatcher = hasWatcher || s.watcherFlag.HasFollowingCreatedOrderWatcher\n\t}\n\n\thasWatcher = hasWatcher || s.watcherFlag.WholeWatcher\n\treturn fields, hasWatcher, nil\n}", "func (e Department) EntFields() ent.Fields { return ent_Department_fields }", "func (Project) Fields() []ent.Field {\n\treturn nil\n}", "func (f Fields) Fields() map[string]interface{} {\n\treturn f\n}", "func (m *Store) GetFieldDeserializers()(map[string]func(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(error)) {\n res := m.Entity.GetFieldDeserializers()\n res[\"defaultLanguageTag\"] = i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.SetStringValue(m.SetDefaultLanguageTag)\n res[\"groups\"] = i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.SetCollectionOfObjectValues(CreateGroupFromDiscriminatorValue , m.SetGroups)\n res[\"languageTags\"] = i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.SetCollectionOfPrimitiveValues(\"string\" , m.SetLanguageTags)\n res[\"sets\"] = i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.SetCollectionOfObjectValues(CreateSetFromDiscriminatorValue , m.SetSets)\n return res\n}", "func expandFields(compiled *lang.CompiledExpr, define *lang.DefineExpr) lang.DefineFieldsExpr {\n\tvar fields lang.DefineFieldsExpr\n\tfor _, field := range define.Fields {\n\t\tif isEmbeddedField(field) {\n\t\t\tembedded := expandFields(compiled, compiled.LookupDefine(string(field.Type)))\n\t\t\tfields = append(fields, embedded...)\n\t\t} else {\n\t\t\tfields = append(fields, field)\n\t\t}\n\t}\n\treturn fields\n}", "func (w *RootWalker) GetFromRefsField(fieldName string, p interface{}, finder func(i int, ref skyobject.Reference) bool) error {\n\t_, e := w.getFromRefsField(fieldName, p, finder)\n\treturn e\n}", "func fields(t reflect.Type) map[string]interface{} {\n\tfieldCache.RLock()\n\tfs := fieldCache.m[t]\n\tfieldCache.RUnlock()\n\n\t//Cached version exists\n\tif fs != nil {\n\t\treturn fs\n\t}\n\t//This is to prevent multiple goroutines computing the same thing\n\tfieldCache.Lock()\n\tvar sy *sync.WaitGroup\n\tif sy, ok := fieldCache.create[t]; ok {\n\t\tfieldCache.Unlock()\n\t\tsy.Wait()\n\t\treturn fields(t)\n\t}\n\tsy = &sync.WaitGroup{}\n\tfieldCache.create[t] = sy\n\tsy.Add(1)\n\tfieldCache.Unlock()\n\n\tfs = compileStruct(t)\n\n\tfieldCache.Lock()\n\tfieldCache.m[t] = fs\n\tfieldCache.Unlock()\n\tsy.Done()\n\treturn fs\n}", "func (op *metadataLookup) field(parentFunc *stmt.CallExpr, expr stmt.Expr) {\n\tif op.err != nil {\n\t\treturn\n\t}\n\tswitch e := expr.(type) {\n\tcase *stmt.SelectItem:\n\t\top.field(nil, e.Expr)\n\tcase *stmt.CallExpr:\n\t\tif e.FuncType == function.Quantile {\n\t\t\top.planHistogramFields(e)\n\t\t\treturn\n\t\t}\n\t\tfor _, param := range e.Params {\n\t\t\top.field(e, param)\n\t\t}\n\tcase *stmt.ParenExpr:\n\t\top.field(nil, e.Expr)\n\tcase *stmt.BinaryExpr:\n\t\top.field(nil, e.Left)\n\t\top.field(nil, e.Right)\n\tcase *stmt.FieldExpr:\n\t\tqueryStmt := op.executeCtx.Query\n\t\tfieldMeta, err := op.metadata.GetField(queryStmt.Namespace, queryStmt.MetricName, field.Name(e.Name))\n\t\tif err != nil {\n\t\t\top.err = err\n\t\t\treturn\n\t\t}\n\n\t\top.planField(parentFunc, fieldMeta)\n\t}\n}", "func fileFields(path string) (base string, root string, ext string, dirname string) {\n\tbase = lastInPath(path)\n\tbaseNoLeadingPeriods, nPeriods := trimLeading(base, \".\")\n\ttmp := strings.Split(baseNoLeadingPeriods, \".\")\n\tif len(tmp) == 1 {\n\t\t// no file extension\n\t\troot = tmp[0]\n\t\text = \"\"\n\t} else {\n\t\troot = strings.Join(tmp[:len(tmp)-1], \".\")\n\t\text = \".\" + tmp[len(tmp)-1]\n\t}\n\t// add back any leading periods that were trimmed from base\n\troot = strings.Repeat(\".\", nPeriods) + root\n\tdirname = strings.TrimSuffix(path, fmt.Sprintf(\"/%v\", base))\n\treturn base, root, ext, dirname\n}", "func (s *scanner) structFields(structType reflect.Type) ([]*r.Field, error) {\n\tfmPtr := reflect.New(structType).Interface()\n\tm, err := r.StructFieldMap(fmPtr, true)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not decode struct of type %T: %v\", fmPtr, err)\n\t}\n\n\tstructFields := make([]*r.Field, len(s.stmt.fieldNames))\n\tfor i, fieldName := range s.stmt.fieldNames {\n\t\tfield, ok := m[strings.ToLower(fieldName)]\n\t\tif !ok { // the field doesn't have a destination\n\t\t\tstructFields[i] = nil\n\t\t} else {\n\t\t\tstructFields[i] = &field\n\t\t}\n\t}\n\treturn structFields, nil\n}", "func (r *Route) Fields() RouteInfo {\n\tr.mu.RLock()\n\tdefer r.mu.RUnlock()\n\treturn r.fieldsLocked()\n}", "func (i *Index) openFields() error {\n\tf, err := os.Open(i.path)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"opening directory\")\n\t}\n\tdefer f.Close()\n\n\tfis, err := f.Readdir(0)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"reading directory\")\n\t}\n\n\tfor _, fi := range fis {\n\t\tif !fi.IsDir() {\n\t\t\tcontinue\n\t\t}\n\n\t\tfld, err := i.newField(i.fieldPath(filepath.Base(fi.Name())), filepath.Base(fi.Name()))\n\t\tif err != nil {\n\t\t\treturn ErrName\n\t\t}\n\t\tif err := fld.Open(); err != nil {\n\t\t\treturn fmt.Errorf(\"open field: name=%s, err=%s\", fld.Name(), err)\n\t\t}\n\t\ti.fields[fld.Name()] = fld\n\t}\n\treturn nil\n}", "func (s *Struct) structFields() []reflect.StructField {\n\tt := s.value.Type()\n\n\tvar f []reflect.StructField\n\n\tfor i := 0; i < t.NumField(); i++ {\n\t\tfield := t.Field(i)\n\t\t// we can't access the value of unexported fields\n\t\tif field.PkgPath != \"\" {\n\t\t\tcontinue\n\t\t}\n\n\t\t// don't check if it's omitted\n\t\tif tag := field.Tag.Get(s.TagName); tag == \"-\" {\n\t\t\tcontinue\n\t\t}\n\n\t\tf = append(f, field)\n\t}\n\n\treturn f\n}", "func TypeFields(t *Type) (fields []*Field) {\n\n\tif t == nil {\n\t\treturn\n\t}\n\n\tfor _, spec := range t.Decl.Specs {\n\n\t\ttypeSpec := spec.(*ast.TypeSpec)\n\n\t\t// struct type\n\t\tif str, ok := typeSpec.Type.(*ast.StructType); ok {\n\n\t\t\tfor _, f := range str.Fields.List {\n\t\t\t\tfields = append(fields, &Field{\n\t\t\t\t\tField: f,\n\t\t\t\t\tType: t,\n\t\t\t\t})\n\t\t\t}\n\n\t\t\treturn\n\t\t}\n\n\t\t// interface type methods\n\t\tif str, ok := typeSpec.Type.(*ast.InterfaceType); ok {\n\t\t\tfor _, field := range str.Methods.List {\n\t\t\t\tif ident, ok := field.Type.(*ast.Ident); ok && ident.Obj != nil {\n\t\t\t\t\tfield.Names = []*ast.Ident{ident}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tfor _, f := range str.Methods.List {\n\t\t\t\tfields = append(fields, &Field{\n\t\t\t\t\tField: f,\n\t\t\t\t\tType: t,\n\t\t\t\t})\n\t\t\t}\n\n\t\t\treturn\n\t\t}\n\t}\n\n\treturn\n}", "func (svc record) readableFields(m *types.Module) []string {\n\tff := make([]string, 0)\n\n\t_ = m.Fields.Walk(func(f *types.ModuleField) error {\n\t\tif svc.ac.CanReadRecordValue(svc.ctx, f) {\n\t\t\tff = append(ff, f.Name)\n\t\t}\n\n\t\treturn nil\n\t})\n\n\treturn ff\n}", "func eachSubField(i interface{}, fn func(reflect.Value, string, []string) error, crumbs ...string) error {\n\tt := reflect.ValueOf(i)\n\tif t.Kind() != reflect.Ptr || t.Elem().Kind() != reflect.Struct {\n\t\treturn errors.New(\"eachSubField can only be called on a pointer-to-struct\")\n\t}\n\t// Sanity check. Should be true if it is a pointer-to-struct\n\tif !t.Elem().CanSet() {\n\t\treturn errors.New(\"eachSubField can only be called on a settable struct of structs\")\n\t}\n\n\tt = t.Elem()\n\tnf := t.NumField()\n\tfor i := 0; i < nf; i++ {\n\t\tfield := t.Field(i)\n\t\tsf := t.Type().Field(i)\n\t\tif sf.Tag.Get(\"flag\") == \"false\" {\n\t\t\tcontinue\n\t\t}\n\n\t\tif field.Kind() == reflect.Struct && field.CanSet() {\n\t\t\teachSubField(field.Addr().Interface(), fn, append(crumbs, sf.Name)...)\n\t\t} else if field.CanSet() {\n\t\t\tif err := fn(t, sf.Name, crumbs); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (t *Type) FieldSlice() []*Field", "func (c *ProjectsGetAncestryCall) Fields(s ...googleapi.Field) *ProjectsGetAncestryCall {\n\tc.urlParams_.Set(\"fields\", googleapi.CombineFields(s))\n\treturn c\n}", "func GetFields(m interface{}) (ns []string) {\n\ttyp := GetType(m)\n\t// Only structs are supported so return an empty result if the passed object\n\t// isn't a struct\n\tif typ.Kind() != reflect.Struct {\n\t\tfmt.Printf(\"%v type can't have attributes inspected\\n\", typ.Kind())\n\t\treturn\n\t}\n\t// loop through the struct's fields\n\tfor i := 0; i < typ.NumField(); i++ {\n\t\tf := typ.Field(i)\n\t\tif f.Anonymous {\n\t\t\tfkind := f.Type.Kind()\n\t\t\tif fkind == reflect.Struct || fkind == reflect.Ptr {\n\t\t\t\tfns := GetFields(reflect.New(f.Type).Interface())\n\t\t\t\tfor _, fn := range fns {\n\t\t\t\t\tif String(fn).IsInArray(ns) {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tns = append(ns, fn)\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tif String(f.Name).IsInArray(ns) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tns = append(ns, f.Name)\n\t\t}\n\t}\n\treturn ns\n}", "func GetRootFields() graphql.Fields {\n\treturn graphql.Fields{\n\t\t\"program\": GetProgramQuery(),\n\t\t\"programs\": GetProgramsQuery(),\n\t\t\"production\": GetProductionQuery(),\n\t\t\"productions\": GetProductionsQuery(),\n\t}\n}", "func (p *Parser) ReadFields(data interface{}) {\n\tconfig.ReadData(p, data)\n}", "func (n *nodeHeader) prefixFields() (*uint16, []byte) {\n\tswitch n.typ {\n\tcase typLeaf:\n\t\t// Leaves have no prefix\n\t\treturn nil, nil\n\tcase typNode4:\n\t\tn4 := n.node4()\n\t\treturn &n4.prefixLen, n4.prefix[:]\n\n\tcase typNode16:\n\t\tn16 := n.node16()\n\t\treturn &n16.prefixLen, n16.prefix[:]\n\n\tcase typNode48:\n\t\tn48 := n.node48()\n\t\treturn &n48.prefixLen, n48.prefix[:]\n\n\tcase typNode256:\n\t\tn256 := n.node256()\n\t\treturn &n256.prefixLen, n256.prefix[:]\n\t}\n\tpanic(\"invalid type\")\n}", "func (e FailError) Fields() vals.StructMap { return failFields{e} }", "func decorateFieldGetter(t *TopLevelType) *TopLevelType {\n\tnewT := t.DeepCopy().(*TopLevelType)\n\tnewT.OriginalType = t.Type\n\n\tswitch tt := newT.Type.(type) {\n\tcase *ModeledType:\n\t\tif tt.IsPtr || tt.Interface != nil || tt.IsBuiltin {\n\t\t\treturn t\n\t\t}\n\t\ttt.IsPtr = true\n\tcase *ArrayType:\n\t\treturn t\n\tcase *MapType:\n\t\treturn t\n\tdefault:\n\t\tpanic(fmt.Sprintf(\"unsupported type: %T\", newT.Type))\n\t}\n\n\treturn newT\n}", "func (fm *FinalModelStructOptional) GetFields(fbeValue *StructOptional) (int, error) {\n var err error = nil\n fbeCurrentOffset := 0\n fbeCurrentSize := 0\n fbeFieldSize := 0\n\n fm.FinalModelStructSimple.SetFBEOffset(fbeCurrentOffset)\n if fbeFieldSize, err = fm.FinalModelStructSimple.GetFields(fbeValue.StructSimple); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F100.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F100, fbeFieldSize, err = fm.F100.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F101.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F101, fbeFieldSize, err = fm.F101.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F102.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F102, fbeFieldSize, err = fm.F102.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F103.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F103, fbeFieldSize, err = fm.F103.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F104.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F104, fbeFieldSize, err = fm.F104.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F105.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F105, fbeFieldSize, err = fm.F105.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F106.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F106, fbeFieldSize, err = fm.F106.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F107.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F107, fbeFieldSize, err = fm.F107.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F108.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F108, fbeFieldSize, err = fm.F108.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F109.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F109, fbeFieldSize, err = fm.F109.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F110.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F110, fbeFieldSize, err = fm.F110.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F111.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F111, fbeFieldSize, err = fm.F111.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F112.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F112, fbeFieldSize, err = fm.F112.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F113.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F113, fbeFieldSize, err = fm.F113.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F114.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F114, fbeFieldSize, err = fm.F114.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F115.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F115, fbeFieldSize, err = fm.F115.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F116.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F116, fbeFieldSize, err = fm.F116.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F117.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F117, fbeFieldSize, err = fm.F117.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F118.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F118, fbeFieldSize, err = fm.F118.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F119.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F119, fbeFieldSize, err = fm.F119.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F120.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F120, fbeFieldSize, err = fm.F120.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F121.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F121, fbeFieldSize, err = fm.F121.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F122.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F122, fbeFieldSize, err = fm.F122.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F123.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F123, fbeFieldSize, err = fm.F123.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F124.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F124, fbeFieldSize, err = fm.F124.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F125.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F125, fbeFieldSize, err = fm.F125.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F126.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F126, fbeFieldSize, err = fm.F126.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F127.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F127, fbeFieldSize, err = fm.F127.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F128.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F128, fbeFieldSize, err = fm.F128.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F129.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F129, fbeFieldSize, err = fm.F129.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F130.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F130, fbeFieldSize, err = fm.F130.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F131.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F131, fbeFieldSize, err = fm.F131.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F132.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F132, fbeFieldSize, err = fm.F132.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F133.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F133, fbeFieldSize, err = fm.F133.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F134.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F134, fbeFieldSize, err = fm.F134.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F135.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F135, fbeFieldSize, err = fm.F135.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F136.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F136, fbeFieldSize, err = fm.F136.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F137.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F137, fbeFieldSize, err = fm.F137.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F138.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F138, fbeFieldSize, err = fm.F138.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F139.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F139, fbeFieldSize, err = fm.F139.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F140.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F140, fbeFieldSize, err = fm.F140.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F141.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F141, fbeFieldSize, err = fm.F141.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F142.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F142, fbeFieldSize, err = fm.F142.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F143.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F143, fbeFieldSize, err = fm.F143.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F144.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F144, fbeFieldSize, err = fm.F144.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F145.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F145, fbeFieldSize, err = fm.F145.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F146.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F146, fbeFieldSize, err = fm.F146.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F147.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F147, fbeFieldSize, err = fm.F147.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F148.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F148, fbeFieldSize, err = fm.F148.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F149.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F149, fbeFieldSize, err = fm.F149.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F150.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F150, fbeFieldSize, err = fm.F150.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F151.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F151, fbeFieldSize, err = fm.F151.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F152.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F152, fbeFieldSize, err = fm.F152.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F153.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F153, fbeFieldSize, err = fm.F153.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F154.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F154, fbeFieldSize, err = fm.F154.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F155.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F155, fbeFieldSize, err = fm.F155.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F156.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F156, fbeFieldSize, err = fm.F156.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F157.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F157, fbeFieldSize, err = fm.F157.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F158.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F158, fbeFieldSize, err = fm.F158.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F159.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F159, fbeFieldSize, err = fm.F159.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F160.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F160, fbeFieldSize, err = fm.F160.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F161.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F161, fbeFieldSize, err = fm.F161.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F162.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F162, fbeFieldSize, err = fm.F162.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F163.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F163, fbeFieldSize, err = fm.F163.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F164.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F164, fbeFieldSize, err = fm.F164.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n fm.F165.SetFBEOffset(fbeCurrentOffset)\n if fbeValue.F165, fbeFieldSize, err = fm.F165.Get(); err != nil {\n return fbeCurrentSize, err\n }\n fbeCurrentOffset += fbeFieldSize\n fbeCurrentSize += fbeFieldSize\n\n return fbeCurrentSize, err\n}", "func (pf *PathFilter) Fields(contentPath string) map[string]string {\n\tout := make(map[string]string)\n\n\tmatch := pf.re.FindStringSubmatch(contentPath)\n\tnames := pf.re.SubexpNames()\n\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tfmt.Println(\"path filter fields is crashing the app\")\n\t\t\tspew.Dump(contentPath, pf.filterPath, pf.re.String(), match, names)\n\t\t\tpanic(\"i'm done\")\n\t\t}\n\t}()\n\n\tfor i, name := range names {\n\t\tif i != 0 && name != \"\" {\n\t\t\tout[name] = match[i]\n\t\t}\n\t}\n\n\treturn out\n}", "func (m *LevelMutation) Fields() []string {\n\tfields := make([]string, 0, 1)\n\tif m._LevelName != nil {\n\t\tfields = append(fields, level.FieldLevelName)\n\t}\n\treturn fields\n}", "func (m *PolicyRule) GetFieldDeserializers()(map[string]func(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(error)) {\n res := m.Entity.GetFieldDeserializers()\n res[\"name\"] = func (n i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode) error {\n val, err := n.GetStringValue()\n if err != nil {\n return err\n }\n if val != nil {\n m.SetName(val)\n }\n return nil\n }\n return res\n}", "func (o *NotificationAllOf) GetFieldsOk() (*map[string]interface{}, bool) {\n\tif o == nil || o.Fields == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Fields, true\n}", "func (s UserSet) FieldsGet(args models.FieldsGetArgs) map[string]*models.FieldInfo {\n\tres := s.Collection().Call(\"FieldsGet\", args)\n\tresTyped, _ := res.(map[string]*models.FieldInfo)\n\treturn resTyped\n}", "func getField(n string, i interface{}) (interface{}, bool) {\n\tre := reflect.ValueOf(i).Elem()\n\tif re.Kind() == reflect.Struct {\n\t\tf := re.FieldByName(n)\n\n\t\tif f.IsValid() {\n\t\t\treturn f.Interface(), true\n\t\t}\n\t}\n\n\treturn nil, false\n}", "func (r *ReflectMapper) FieldMap(v reflect.Value) map[string]reflect.Value {\n\tv = reflect.Indirect(v)\n\n\tret := map[string]reflect.Value{}\n\ttm := r.mapper.TypeMap(v.Type())\n\tfor tagName, fi := range tm.Names {\n\t\t//fmt.Println(tagName,fi.Parent.Zero.Kind(),fi.Parent.Field.Anonymous)\n\t\tif (fi.Parent.Zero.Kind() == reflect.Struct || (fi.Zero.Kind() == reflect.Ptr && fi.Zero.Type().Elem().Kind() == reflect.Struct)) && !fi.Parent.Field.Anonymous {\n\t\t\tcontinue\n\t\t}\n\t\tret[tagName] = reflectx.FieldByIndexes(v, fi.Index)\n\t}\n\n\treturn ret\n}", "func (d Document) Fields() []string {\n\treturn d.data.fields\n}", "func (Menu) Fields() []ent.Field {\n\treturn nil\n}", "func (m *Reports) GetFieldDeserializers()(map[string]func(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(error)) {\n res := m.Entity.GetFieldDeserializers()\n return res\n}", "func isLteField(fl FieldLevel) bool {\n\tfield := fl.Field()\n\tkind := field.Kind()\n\n\tcurrentField, currentKind, ok := fl.GetStructFieldOK()\n\tif !ok || currentKind != kind {\n\t\treturn false\n\t}\n\n\tswitch kind {\n\n\tcase reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:\n\n\t\treturn field.Int() <= currentField.Int()\n\n\tcase reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr:\n\n\t\treturn field.Uint() <= currentField.Uint()\n\n\tcase reflect.Float32, reflect.Float64:\n\n\t\treturn field.Float() <= currentField.Float()\n\n\tcase reflect.Struct:\n\n\t\tfieldType := field.Type()\n\n\t\tif fieldType.ConvertibleTo(timeType) && currentField.Type().ConvertibleTo(timeType) {\n\n\t\t\tt := currentField.Convert(timeType).Interface().(time.Time)\n\t\t\tfieldTime := field.Convert(timeType).Interface().(time.Time)\n\n\t\t\treturn fieldTime.Before(t) || fieldTime.Equal(t)\n\t\t}\n\n\t\t// Not Same underlying type i.e. struct and time\n\t\tif fieldType != currentField.Type() {\n\t\t\treturn false\n\t\t}\n\t}\n\n\t// default reflect.String\n\treturn len(field.String()) <= len(currentField.String())\n}", "func (m *Planner) GetFieldDeserializers()(map[string]func(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(error)) {\n res := m.Entity.GetFieldDeserializers()\n res[\"buckets\"] = i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.SetCollectionOfObjectValues(CreatePlannerBucketFromDiscriminatorValue , m.SetBuckets)\n res[\"plans\"] = i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.SetCollectionOfObjectValues(CreatePlannerPlanFromDiscriminatorValue , m.SetPlans)\n res[\"tasks\"] = i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.SetCollectionOfObjectValues(CreatePlannerTaskFromDiscriminatorValue , m.SetTasks)\n return res\n}", "func (m *AdminReportSettings) GetFieldDeserializers()(map[string]func(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(error)) {\n res := m.Entity.GetFieldDeserializers()\n res[\"displayConcealedNames\"] = func (n i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode) error {\n val, err := n.GetBoolValue()\n if err != nil {\n return err\n }\n if val != nil {\n m.SetDisplayConcealedNames(val)\n }\n return nil\n }\n return res\n}", "func (m *Descriptor) GetMessageFields() []*FieldDescriptor { return m.Fields }", "func (fp *slicingFieldParser) getTaggedFields(str *[]byte) (map[string]string, error) {\n\tfields := make(map[string]string, 20)\n\tfor {\n\t\tnextTag, err := fp.nextField(str, equalsBytes, true) // Expect at least a value after the tag\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"parsing next tag: %w\", err)\n\t\t}\n\n\t\tnextValue, err := fp.nextField(str, spaceBytes, false) // We cannot expect any more fields as this may be the last\n\t\tif err != nil && err != io.EOF {\n\t\t\treturn nil, fmt.Errorf(\"parsing next tagged value: %w\", err)\n\t\t}\n\n\t\tfields[nextTag] = nextValue\n\n\t\tif err == io.EOF { // No more fields in stream\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn fields, nil\n}", "func (m *TargetManager) GetFieldDeserializers()(map[string]func(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(error)) {\n res := m.SubjectSet.GetFieldDeserializers()\n res[\"managerLevel\"] = i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.SetInt32Value(m.SetManagerLevel)\n return res\n}", "func (m *MacroEvaluator) GetFields() []Field {\n\tfields := make([]Field, len(m.FieldValues))\n\ti := 0\n\tfor key := range m.FieldValues {\n\t\tfields[i] = key\n\t\ti++\n\t}\n\treturn fields\n}", "func (sh *Hook) Fields() map[string]interface{} {\n\treturn sh.fields\n}", "func (u ConfluenceServerUserV1) GetFields() logrus.Fields {\n\treturn logrus.Fields{\n\t\t\"Host\": u.Host,\n\t\t\"Port\": u.Port,\n\t\t\"Description\": u.Description,\n\t\t\"Auth\": u.Auth.GetFields(),\n\t}\n}", "func (c MethodsCollection) FieldGet() pFieldGet {\n\treturn pFieldGet{\n\t\tMethod: c.MustGet(\"FieldGet\"),\n\t}\n}" ]
[ "0.61212295", "0.5837798", "0.5814727", "0.57181865", "0.56932044", "0.5691752", "0.56457984", "0.56295365", "0.5619157", "0.56047153", "0.55527204", "0.5524664", "0.5486085", "0.5475357", "0.5453594", "0.5439258", "0.5431331", "0.54223394", "0.54207104", "0.54119086", "0.53684807", "0.53610134", "0.5333137", "0.5312929", "0.5310961", "0.53053457", "0.53016186", "0.52949023", "0.52445155", "0.5238158", "0.52348644", "0.52335066", "0.5232413", "0.52235883", "0.52160823", "0.52118987", "0.5199386", "0.5197722", "0.5192848", "0.5186224", "0.5185207", "0.51843816", "0.5157361", "0.51572603", "0.5156397", "0.5136306", "0.5119913", "0.51093525", "0.5094425", "0.50698143", "0.5068588", "0.5066323", "0.50593615", "0.5057461", "0.5053655", "0.5051799", "0.5049374", "0.5044848", "0.50429153", "0.50394005", "0.5037606", "0.50358254", "0.5031969", "0.50217474", "0.50141734", "0.5011943", "0.5011099", "0.5008504", "0.49839044", "0.49817932", "0.4980802", "0.49779838", "0.49710244", "0.49579832", "0.4951412", "0.4941188", "0.4941126", "0.4922502", "0.49216717", "0.49209705", "0.49150345", "0.49127144", "0.49033496", "0.49026254", "0.48967338", "0.48945376", "0.48940632", "0.48938534", "0.4884627", "0.48804134", "0.48762256", "0.48722547", "0.4867748", "0.48670414", "0.48644316", "0.48634696", "0.48633027", "0.4856444", "0.48509735", "0.48507622" ]
0.79517883
0
NewCopyRecipeToMyRecipesWithChangesParams creates a new CopyRecipeToMyRecipesWithChangesParams object with the default values initialized.
func NewCopyRecipeToMyRecipesWithChangesParams() *CopyRecipeToMyRecipesWithChangesParams { var () return &CopyRecipeToMyRecipesWithChangesParams{ timeout: cr.DefaultTimeout, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewCopyRecipeToMyRecipesWithChangesParamsWithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\t\tHTTPClient: client,\n\t}\n}", "func NewCopyRecipeToMyRecipesWithChangesParamsWithTimeout(timeout time.Duration) *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithContext(ctx context.Context) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithTimeout(timeout time.Duration) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func NewCopyRecipeToMyRecipesWithChangesParamsWithContext(ctx context.Context) *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.JSONBody != nil {\n\t\tif err := r.SetBodyParam(o.JSONBody); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param recipeId\n\tif err := r.SetPathParam(\"recipeId\", o.RecipeID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithJSONBody(jSONBody *models.Recipe) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetJSONBody(jSONBody)\n\treturn o\n}", "func NewGetRecipesParams() GetRecipesParams {\n\n\tvar (\n\t\t// initialize parameters with default values\n\n\t\tingredient1Default = string(\"\")\n\t\tingredient2Default = string(\"\")\n\t\tingredient3Default = string(\"\")\n\t\tseasonDefault = string(\"\")\n\t)\n\n\treturn GetRecipesParams{\n\t\tIngredient1: &ingredient1Default,\n\n\t\tIngredient2: &ingredient2Default,\n\n\t\tIngredient3: &ingredient3Default,\n\n\t\tSeason: &seasonDefault,\n\t}\n}", "func (s *TemplateService) NewCopyTemplateParams(id string) *CopyTemplateParams {\n\tp := &CopyTemplateParams{}\n\tp.p = make(map[string]interface{})\n\tp.p[\"id\"] = id\n\treturn p\n}", "func NewCopyToArgs(nodeID NodeID, targetNodeID NodeID) *CopyToArgs {\n\targs := new(CopyToArgs)\n\targs.NodeID = nodeID\n\targs.TargetNodeID = targetNodeID\n\treturn args\n}", "func (f *FilterParams) Clone() *FilterParams {\n\tclone := &FilterParams{\n\t\tFilters: []*model.FilterSet{},\n\t}\n\tfor _, filters := range f.Filters {\n\t\tclone.Filters = append(clone.Filters, filters.Clone())\n\t}\n\tfor _, highlights := range f.Highlights {\n\t\tclone.Highlights = append(clone.Highlights, highlights.Clone())\n\t}\n\tclone.Invert = f.Invert\n\tclone.Variables = append(clone.Variables, f.Variables...)\n\tclone.Size = f.Size\n\tclone.DataMode = f.DataMode\n\treturn clone\n}", "func NewTransferParams(toWalletableId int32, toWalletableType string, fromWalletableId int32, fromWalletableType string, amount int32, date string, companyId int32, ) *TransferParams {\n\tthis := TransferParams{}\n\tthis.ToWalletableId = toWalletableId\n\tthis.ToWalletableType = toWalletableType\n\tthis.FromWalletableId = fromWalletableId\n\tthis.FromWalletableType = fromWalletableType\n\tthis.Amount = amount\n\tthis.Date = date\n\tthis.CompanyId = companyId\n\treturn &this\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithRecipeID(recipeID string) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetRecipeID(recipeID)\n\treturn o\n}", "func linkedConstructInputsCopyTo(ctx *pulumi.Context, inputs map[string]interface{}, args interface{}) error", "func NewParams(opts []copts.Opt) *Params {\r\n\tparams := &Params{}\r\n\tcopts.Apply(params, opts)\r\n\treturn params\r\n}", "func NewParams(opts []copts.Opt) *Params {\n\tparams := &Params{}\n\tcopts.Apply(params, opts)\n\treturn params\n}", "func NewCopy(ctx context.Context, cfg CopyConfig, reg *config_vars.Registry, client *http.Client, serverURL, workdir, userName, userEmail string, dep child.Child) (*gitilesParent, error) {\n\tif err := cfg.Validate(); err != nil {\n\t\treturn nil, skerr.Wrap(err)\n\t}\n\tgetContentsAtRev := func(ctx context.Context, rev *revision.Revision) (map[string]string, error) {\n\t\tfs, err := dep.VFS(ctx, rev)\n\t\tif err != nil {\n\t\t\treturn nil, skerr.Wrap(err)\n\t\t}\n\t\trv := map[string]string{}\n\t\tfor _, cp := range cfg.Copies {\n\t\t\tif err := vfs.Walk(ctx, fs, cp.SrcRelPath, func(fp string, info os.FileInfo, err error) error {\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn skerr.Wrap(err)\n\t\t\t\t}\n\t\t\t\tif info.IsDir() {\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\t\tcontents, err := vfs.ReadFile(ctx, fs, fp)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn skerr.Wrap(err)\n\t\t\t\t}\n\t\t\t\tif !strings.HasPrefix(fp, cp.SrcRelPath) {\n\t\t\t\t\treturn skerr.Fmt(\"Path %q does not have expected prefix %q\", fp, cp.SrcRelPath)\n\t\t\t\t}\n\t\t\t\tparentPath := path.Join(cp.DstRelPath, strings.TrimPrefix(fp, cp.SrcRelPath))\n\t\t\t\trv[parentPath] = string(contents)\n\t\t\t\treturn nil\n\t\t\t}); err != nil {\n\t\t\t\treturn nil, skerr.Wrap(err)\n\t\t\t}\n\t\t}\n\t\treturn rv, nil\n\t}\n\tgetChangesHelper := gitilesFileGetChangesForRollFunc(cfg.DependencyConfig)\n\tgetChangesForRoll := func(ctx context.Context, repo *gitiles_common.GitilesRepo, baseCommit string, from, to *revision.Revision, rolling []*revision.Revision) (map[string]string, error) {\n\t\tchanges, err := getChangesHelper(ctx, repo, baseCommit, from, to, rolling)\n\t\tif err != nil {\n\t\t\treturn nil, skerr.Wrap(err)\n\t\t}\n\t\tbefore, err := getContentsAtRev(ctx, from)\n\t\tif err != nil {\n\t\t\treturn nil, skerr.Wrap(err)\n\t\t}\n\t\tafter, err := getContentsAtRev(ctx, to)\n\t\tif err != nil {\n\t\t\treturn nil, skerr.Wrap(err)\n\t\t}\n\t\tfilenames := util.StringSet{}\n\t\tfor f := range before {\n\t\t\tfilenames[f] = true\n\t\t}\n\t\tfor f := range after {\n\t\t\tfilenames[f] = true\n\t\t}\n\t\tfor f := range filenames {\n\t\t\tif before[f] != after[f] {\n\t\t\t\tchanges[f] = after[f]\n\t\t\t}\n\t\t}\n\t\treturn changes, nil\n\t}\n\treturn newGitiles(ctx, cfg.GitilesConfig, reg, client, serverURL, getChangesForRoll)\n}", "func (in *NewPendingTxNotificationParams) DeepCopy() *NewPendingTxNotificationParams {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(NewPendingTxNotificationParams)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func NewCopyTemplateRequest(server string, id string, body CopyTemplateJSONRequestBody) (*http.Request, error) {\n\tvar bodyReader io.Reader\n\tbuf, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbodyReader = bytes.NewReader(buf)\n\treturn NewCopyTemplateRequestWithBody(server, id, \"application/json\", bodyReader)\n}", "func (params *headerParams) Copy() Params {\n\tdup := NewParams()\n\tfor _, key := range params.Keys() {\n\t\tif val, ok := params.Get(key); ok {\n\t\t\tdup.Add(key, val)\n\t\t}\n\t}\n\n\treturn dup\n}", "func Copy(scope *Scope, input tf.Output, optional ...CopyAttr) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{}\n\tfor _, a := range optional {\n\t\ta(attrs)\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"Copy\",\n\t\tInput: []tf.Input{\n\t\t\tinput,\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func copyParams(r *http.Request) map[string]string {\n\tparamMap := make(map[string]string)\n\n\tfor k, v := range r.Form {\n\t\tparamMap[k] = v[0]\n\t}\n\n\treturn paramMap\n}", "func (in *NewPendingTxBodyNotificationParams) DeepCopy() *NewPendingTxBodyNotificationParams {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(NewPendingTxBodyNotificationParams)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func NewQueryChangesParams() *QueryChangesParams {\n\treturn &QueryChangesParams{\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func NewGetPublicsRecipeParams() *GetPublicsRecipeParams {\n\n\treturn &GetPublicsRecipeParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func NewNarrowSearchRecipeParams() *NarrowSearchRecipeParams {\n\tvar (\n\t\tsortbyDefault = string(\"name\")\n\t\tsortdirDefault = string(\"desc\")\n\t)\n\treturn &NarrowSearchRecipeParams{\n\t\tSortby: &sortbyDefault,\n\t\tSortdir: &sortdirDefault,\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (m *Message) Copy() *Message {\n\t// Create a new message\n\tnewMessage := &Message{}\n\n\t// Copy stuff from the old message\n\t*newMessage = *m\n\n\t// Copy any IRcv3 tags\n\tnewMessage.Tags = m.Tags.Copy()\n\n\t// Copy the Prefix\n\tnewMessage.Prefix = m.Prefix.Copy()\n\n\t// Copy the Params slice\n\tnewMessage.Params = append(make([]string, 0, len(m.Params)), m.Params...)\n\n\t// Similar to parsing, if Params is empty, set it to nil\n\tif len(newMessage.Params) == 0 {\n\t\tnewMessage.Params = nil\n\t}\n\n\treturn newMessage\n}", "func NewChange(templateItem *ResourceItem, platformItem *ResourceItem, comparison map[string]*jsonPatch) *Change {\n\tc := &Change{\n\t\tKind: templateItem.Kind,\n\t\tName: templateItem.Name,\n\t\tPatches: []*jsonPatch{},\n\t\tCurrentState: platformItem.YamlConfig(),\n\t\tDesiredState: templateItem.YamlConfig(),\n\t}\n\n\tfor path, patch := range comparison {\n\t\tif patch.Op != \"noop\" {\n\t\t\tcli.DebugMsg(\"add path\", path)\n\t\t\tpatch.Path = path\n\t\t\tc.addPatch(patch)\n\t\t}\n\t}\n\n\tif len(c.Patches) > 0 {\n\t\tc.Action = \"Update\"\n\t} else {\n\t\tc.Action = \"Noop\"\n\t}\n\n\treturn c\n}", "func (in *HelmRequest) DeepCopy() *HelmRequest {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(HelmRequest)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func ParamChanges(r *rand.Rand) []simtypes.ParamChange {\n\treturn []simtypes.ParamChange{\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.KeyMaxContractSize),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"\\\"%d\\\"\", GenMaxContractSize(r))\n\t\t\t},\n\t\t),\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.KeyMaxContractGas),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"\\\"%d\\\"\", GenMaxContractGas(r))\n\t\t\t},\n\t\t),\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.KeyMaxContractMsgSize),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"\\\"%d\\\"\", GenMaxContractMsgSize(r))\n\t\t\t},\n\t\t),\n\t}\n}", "func (p *DiscoveryProtocol) copyNewDiscoveryRequest(request *api.DiscoveryRequest) *api.DiscoveryRequest {\n\treq := &api.DiscoveryRequest{DiscoveryMsgData: NewDiscoveryMsgData(uuid.Must(uuid.NewV4(), nil).String(), true, p.p2pHost),\n\t\tMessage: api.DiscoveryMessage_DiscoveryReq}\n\treq.DiscoveryMsgData.InitNodeID = request.DiscoveryMsgData.InitNodeID\n\treq.DiscoveryMsgData.TTL = request.DiscoveryMsgData.TTL\n\treq.DiscoveryMsgData.Expiry = request.DiscoveryMsgData.Expiry\n\treq.DiscoveryMsgData.InitHash = request.DiscoveryMsgData.InitHash\n\tlog.Println(\"COPYING: \", req.DiscoveryMsgData.InitHash)\n\n\tkey := p.p2pHost.Peerstore().PrivKey(p.p2pHost.ID())\n\treq.DiscoveryMsgData.MessageData.Sign = signProtoMsg(req, key)\n\treturn req\n}", "func (in *CredentialProviderRequest) DeepCopy() *CredentialProviderRequest {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(CredentialProviderRequest)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func NewGetReceiptsParams() *GetReceiptsParams {\n\tvar ()\n\treturn &GetReceiptsParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (s settings) deepCopy() settings {\n\t// vmodule is a slice and would be shared, so we have copy it.\n\tfilter := make([]modulePat, len(s.vmodule.filter))\n\tfor i := range s.vmodule.filter {\n\t\tfilter[i] = s.vmodule.filter[i]\n\t}\n\ts.vmodule.filter = filter\n\n\tif s.logger != nil {\n\t\tlogger := *s.logger\n\t\ts.logger = &logger\n\t}\n\n\treturn s\n}", "func GetParamsForPopularRecipe(hfrt PopularRecipeType) (types.CoinInputList, types.ItemInputList, types.EntriesList, types.WeightedOutputsList, int64) {\n\tswitch hfrt {\n\tcase Rcp5xWoodcoinTo1xChaircoin: // 5 x woodcoin -> 1 x chair coin recipe\n\t\treturn types.GenCoinInputList(\"wood\", 5),\n\t\t\ttypes.ItemInputList{},\n\t\t\ttypes.GenCoinOnlyEntry(\"chair\"),\n\t\t\ttypes.GenOneOutput(1),\n\t\t\t0\n\tcase Rcp5BlockDelayed5xWoodcoinTo1xChaircoin: // 5 x woodcoin -> 1 x chair coin recipe, 5 block delayed\n\t\treturn types.GenCoinInputList(\"wood\", 5),\n\t\t\ttypes.ItemInputList{},\n\t\t\ttypes.GenCoinOnlyEntry(\"chair\"),\n\t\t\ttypes.GenOneOutput(1),\n\t\t\t5\n\tcase Rcp5xWoodcoinTo1xRaichuItemBuy:\n\t\treturn types.GenCoinInputList(\"wood\", 5),\n\t\t\ttypes.ItemInputList{},\n\t\t\ttypes.GenItemOnlyEntry(\"Raichu\"),\n\t\t\ttypes.GenOneOutput(1),\n\t\t\t0\n\tcase RcpRaichuNameUpgrade:\n\t\treturn types.CoinInputList{},\n\t\t\ttypes.GenItemInputList(\"Raichu\"),\n\t\t\ttypes.GenEntriesFirstItemNameUpgrade(\"RaichuV2\"),\n\t\t\ttypes.GenOneOutput(1),\n\t\t\t0\n\tcase RcpRaichuNameUpgradeWithCatalyst:\n\t\treturn types.CoinInputList{},\n\t\t\ttypes.GenItemInputList(\"RaichuTC\", \"catalyst\"),\n\t\t\ttypes.GenEntriesFirstItemNameUpgrade(\"RaichuTCV2\"),\n\t\t\ttypes.GenOneOutput(1),\n\t\t\t0\n\tcase Rcp2BlockDelayedKnifeUpgrade:\n\t\treturn types.CoinInputList{},\n\t\t\ttypes.GenItemInputList(\"Knife\"),\n\t\t\ttypes.GenEntriesFirstItemNameUpgrade(\"KnifeV2\"),\n\t\t\ttypes.GenOneOutput(1),\n\t\t\t2\n\tcase Rcp2BlockDelayedKnifeMerge:\n\t\treturn types.CoinInputList{},\n\t\t\ttypes.GenItemInputList(\"Knife\", \"Knife\"),\n\t\t\ttypes.GenItemOnlyEntry(\"KnifeMRG\"),\n\t\t\ttypes.GenOneOutput(1),\n\t\t\t2\n\tcase Rcp2BlockDelayedKnifeBuyer:\n\t\treturn types.GenCoinInputList(\"wood\", 5),\n\t\t\ttypes.ItemInputList{},\n\t\t\ttypes.GenItemOnlyEntry(\"Knife\"),\n\t\t\ttypes.GenOneOutput(1),\n\t\t\t2\n\tdefault: // 5 x woodcoin -> 1 x chair coin recipe, no delay\n\t\treturn types.GenCoinInputList(\"wood\", 5),\n\t\t\ttypes.ItemInputList{},\n\t\t\ttypes.GenEntries(\"chair\", \"Raichu\"),\n\t\t\ttypes.GenOneOutput(1),\n\t\t\t0\n\t}\n}", "func NewParams(createWhoisPrice string, updateWhoisPrice string, deleteWhoisPrice string) Params {\n\treturn Params{\n\t\tCreateWhoisPrice: createWhoisPrice,\n\t\tUpdateWhoisPrice: updateWhoisPrice,\n\t\tDeleteWhoisPrice: deleteWhoisPrice,\n\t}\n}", "func NewPostReconciliationParams() *PostReconciliationParams {\n\tvar ()\n\treturn &PostReconciliationParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func NewParams(pricePerByte abi.TokenAmount, paymentInterval uint64, paymentIntervalIncrease uint64) (Params, error) {\n\treturn Params{\n\t\tPricePerByte: pricePerByte,\n\t\tPaymentInterval: paymentInterval,\n\t\tPaymentIntervalIncrease: paymentIntervalIncrease,\n\t}, nil\n}", "func NewCalcOperationParams() CalcOperationParams {\n\n\treturn CalcOperationParams{}\n}", "func NewChatNewParams() *ChatNewParams {\n\tvar (\n\t\tchannelIDDefault = float64(1)\n\t)\n\treturn &ChatNewParams{\n\t\tChannelID: &channelIDDefault,\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (m *ItemItemsDriveItemItemRequestBuilder) Copy()(*ItemItemsItemCopyRequestBuilder) {\n return NewItemItemsItemCopyRequestBuilderInternal(m.BaseRequestBuilder.PathParameters, m.BaseRequestBuilder.RequestAdapter)\n}", "func NewMsgUpdateRecipe(recipeName, cookbookID, id, description string,\n\tcoinInputs types.CoinInputList,\n\titemInputs types.ItemInputList,\n\tentries types.EntriesList,\n\toutputs types.WeightedOutputsList,\n\tsender sdk.AccAddress) MsgUpdateRecipe {\n\treturn MsgUpdateRecipe{\n\t\tName: recipeName,\n\t\tID: id,\n\t\tCookbookID: cookbookID,\n\t\tDescription: description,\n\t\tCoinInputs: coinInputs,\n\t\tItemInputs: itemInputs,\n\t\tEntries: entries,\n\t\tOutputs: outputs,\n\t\tBlockInterval: 0,\n\t\tSender: sender,\n\t}\n}", "func (inputs ConstructInputs) CopyTo(args interface{}) error {\n\treturn linkedConstructInputsCopyTo(inputs.ctx, inputs.inputs, args)\n}", "func callArgsCopyTo(ctx *Context, source map[string]interface{}, args interface{}) (Resource, error) {\n\t// Use the same implementation as construct.\n\tif err := constructInputsCopyTo(ctx, source, args); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Retrieve the `__self__` arg.\n\tself, err := callArgsSelf(ctx, source)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn self, nil\n}", "func NewConfigurationBackupModifyParams() *ConfigurationBackupModifyParams {\n\treturn &ConfigurationBackupModifyParams{\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (o *Entry) Copy(s Entry) {\n\to.Interval = s.Interval\n\to.Threshold = s.Threshold\n\to.Action = s.Action\n}", "func NewCopyToDefaultContentLocationPostRequestBody()(*CopyToDefaultContentLocationPostRequestBody) {\n m := &CopyToDefaultContentLocationPostRequestBody{\n }\n m.SetAdditionalData(make(map[string]interface{}));\n return m\n}", "func (c *RBController) NewRecipe(w http.ResponseWriter, r *http.Request) (err error) {\n\t// build data with anonymous struct\n\tdata := struct {\n\t\t*Recipe\n\t\tNewRecipe bool\n\t}{\n\t\tnew(Recipe),\n\t\ttrue,\n\t}\n\n\t// pass data to render\n\tc.HTML(w, http.StatusOK, \"recipes/edit\", data)\n\treturn nil\n}", "func (o *QueryChangesParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (sh *Hook) Copy() *Hook {\n\t// another map copy...\n\tf := make(map[string]interface{}, len(sh.fields))\n\tfor k, v := range sh.fields {\n\t\tf[k] = v\n\t}\n\n\treturn &Hook{\n\t\tCapturer: sh.Capturer,\n\t\tfields: f,\n\t\tminLevel: sh.minLevel,\n\t\ttraceEnabled: sh.traceEnabled,\n\t\ttraceSkipFrames: sh.traceSkipFrames,\n\t\ttraceContextLines: sh.traceContextLines,\n\t\ttraceAppPrefixes: sh.traceAppPrefixes,\n\t}\n}", "func NewParams() *Params {\n\treturn new(Params)\n}", "func NewChangeRequest() *ChangeRequest {\n\tattr := &ChangeRequest{}\n\treturn attr\n}", "func ParamChanges(r *rand.Rand) []simulation.ParamChange {\n\t// Note: params are encoded to JSON before being stored in the param store. These param changes\n\t// update the raw values in the store so values need to be JSON. This is why values that are represented\n\t// as strings in JSON (such as time.Duration) have the escaped quotes.\n\t// TODO should we encode the values properly with ModuleCdc.MustMarshalJSON()?\n\treturn []simulation.ParamChange{\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.KeyBidDuration),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"%d\", GenBidDuration(r))\n\t\t\t},\n\t\t),\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.KeyMaxAuctionDuration),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"%d\", GenMaxAuctionDuration(r))\n\t\t\t},\n\t\t),\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.KeyIncrementCollateral),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"%d\", GenIncrementCollateral(r))\n\t\t\t},\n\t\t),\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.KeyIncrementDebt),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"%d\", GenIncrementDebt(r))\n\t\t\t},\n\t\t),\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.KeyIncrementSurplus),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"%d\", GenIncrementSurplus(r))\n\t\t\t},\n\t\t),\n\t}\n}", "func (o *Entry) Copy(s Entry) {\n o.Description = s.Description\n o.Type = s.Type\n o.SourceZones = s.SourceZones\n o.DestinationZone = s.DestinationZone\n o.ToInterface = s.ToInterface\n o.Service = s.Service\n o.SourceAddresses = s.SourceAddresses\n o.DestinationAddresses = s.DestinationAddresses\n o.SatType = s.SatType\n o.SatAddressType = s.SatAddressType\n o.SatTranslatedAddresses = s.SatTranslatedAddresses\n o.SatInterface = s.SatInterface\n o.SatIpAddress = s.SatIpAddress\n o.SatFallbackType = s.SatFallbackType\n o.SatFallbackTranslatedAddresses = s.SatFallbackTranslatedAddresses\n o.SatFallbackInterface = s.SatFallbackInterface\n o.SatFallbackIpType = s.SatFallbackIpType\n o.SatFallbackIpAddress = s.SatFallbackIpAddress\n o.SatStaticTranslatedAddress = s.SatStaticTranslatedAddress\n o.SatStaticBiDirectional = s.SatStaticBiDirectional\n o.DatAddress = s.DatAddress\n o.DatPort = s.DatPort\n o.Disabled = s.Disabled\n o.Targets = s.Targets\n o.NegateTarget = s.NegateTarget\n o.Tags = s.Tags\n o.DatType = s.DatType\n o.DatDynamicDistribution = s.DatDynamicDistribution\n}", "func NewParams() *Params {\n\tp := Params{}\n\tp.names = []string{}\n\tp.values = map[string]interface{}{}\n\n\treturn &p\n}", "func (o *GetIngredientVersionRevisionParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func (pd *ParameterDefinition) DeepCopy() *ParameterDefinition {\n\tvar p2 ParameterDefinition\n\tp2 = *pd\n\tp2.ApplyTo = make([]string, len(pd.ApplyTo))\n\tcopy(p2.ApplyTo, pd.ApplyTo)\n\treturn &p2\n}", "func makeRequest(changes []store.Change) *store.ChangeRequest {\n\treq := new(store.ChangeRequest)\n\treq.RequestEntity = uint64(config.Id())\n\treq.RequestNode = config.Id()\n\treq.RequestId = store.AllocateRequestId()\n\treq.Changeset = changes\n\n\treturn req\n}", "func (v EditValidator) Copy() StakeMsg {\n\tv1 := v\n\tdesc := *v.Description\n\tv1.Description = &desc\n\treturn v1\n}", "func (in *FilterParameters) DeepCopy() *FilterParameters {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(FilterParameters)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (b *ClusterNodesBuilder) Copy(object *ClusterNodes) *ClusterNodesBuilder {\n\tif object == nil {\n\t\treturn b\n\t}\n\tb.compute = object.compute\n\tb.infra = object.infra\n\tb.master = object.master\n\tb.total = object.total\n\treturn b\n}", "func (p *Predicate) copy() *Predicate {\n\tparams := make([]Variable, len(p.parameters))\n\tfor i := range params {\n\t\tparams[i] = Variable{p.parameters[i].name, p.parameters[i].kind}\n\t}\n\treturn &Predicate{p.name, params, p.private}\n}", "func (b *AccessReviewRequestBuilder) Copy(object *AccessReviewRequest) *AccessReviewRequestBuilder {\n\tif object == nil {\n\t\treturn b\n\t}\n\tb.bitmap_ = object.bitmap_\n\tb.accountUsername = object.accountUsername\n\tb.action = object.action\n\tb.clusterID = object.clusterID\n\tb.clusterUUID = object.clusterUUID\n\tb.organizationID = object.organizationID\n\tb.resourceType = object.resourceType\n\tb.subscriptionID = object.subscriptionID\n\treturn b\n}", "func (s *Step) Copy() *Step {\n\tif s == nil {\n\t\treturn nil\n\t}\n\t// Unfortunately, because we don't know the type of StepData.Data, we\n\t// can't deep copy it.\n\tvar data []*StepData\n\tif s.Data != nil {\n\t\tdata = append(make([]*StepData, 0, len(s.Data)), s.Data...)\n\t}\n\treturn &Step{\n\t\tProperties: s.Properties.Copy(),\n\t\tData: data,\n\t\tStarted: s.Started,\n\t\tFinished: s.Finished,\n\t\tResult: s.Result,\n\t\tErrors: util.CopyStringSlice(s.Errors),\n\t}\n}", "func makeParamsWithVersion(action, version string) map[string]string {\n\tparams := make(map[string]string)\n\tparams[\"Action\"] = action\n\tparams[\"Version\"] = version\n\treturn params\n}", "func (r *Helm) Copy() *Helm {\n\treturn &Helm{\n\t\tID: r.ID,\n\t\t//ProjectName: r.ProjectName,\n\t\tType: r.Type,\n\t\tName: r.Name,\n\t\tAddress: r.Address,\n\t\tUsername: r.Username,\n\t\tPrefix: r.Prefix,\n\t}\n}", "func (s *Spec) Clone() *Spec {\n\tres := &Spec{Target: make(map[string]string)}\n\tfor k, v := range s.Target {\n\t\tres.Target[k] = v\n\t}\n\tfor _, app := range s.Apps {\n\t\tres.Apps = append(res.Apps, app.Clone())\n\t}\n\treturn res\n}", "func (in *V1SystemDeploymentParameters) DeepCopyToModel() *models.V1SystemDeploymentParameters {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(models.V1SystemDeploymentParameters)\n\tin.DeepCopyToModelInto(out)\n\treturn out\n}", "func (cs *CollectionSpec) Copy() *CollectionSpec {\n\tif cs == nil {\n\t\treturn nil\n\t}\n\n\tcpy := CollectionSpec{\n\t\tMaps: make(map[string]*MapSpec, len(cs.Maps)),\n\t\tPrograms: make(map[string]*ProgramSpec, len(cs.Programs)),\n\t}\n\n\tfor name, spec := range cs.Maps {\n\t\tcpy.Maps[name] = spec.Copy()\n\t}\n\n\tfor name, spec := range cs.Programs {\n\t\tcpy.Programs[name] = spec.Copy()\n\t}\n\n\treturn &cpy\n}", "func (s *VPCService) NewCreateVPCOfferingParams(displaytext string, name string, supportedservices []string) *CreateVPCOfferingParams {\n\tp := &CreateVPCOfferingParams{}\n\tp.p = make(map[string]interface{})\n\tp.p[\"displaytext\"] = displaytext\n\tp.p[\"name\"] = name\n\tp.p[\"supportedservices\"] = supportedservices\n\treturn p\n}", "func ParamChanges(r *rand.Rand) []simulation.ParamChange {\n\treturn []simulation.ParamChange{\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.ParamStoreKeyVotePeriod),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"\\\"%d\\\"\", GenVotePeriod(r))\n\t\t\t},\n\t\t),\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.ParamStoreKeyVoteThreshold),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"\\\"%s\\\"\", GenVoteThreshold(r))\n\t\t\t},\n\t\t),\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.ParamStoreKeyRewardBand),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"\\\"%s\\\"\", GenRewardBand(r))\n\t\t\t},\n\t\t),\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.ParamStoreKeyRewardDistributionWindow),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"\\\"%d\\\"\", GenRewardDistributionWindow(r))\n\t\t\t},\n\t\t),\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.ParamStoreKeySlashFraction),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"\\\"%s\\\"\", GenSlashFraction(r))\n\t\t\t},\n\t\t),\n\t\tsimulation.NewSimParamChange(types.ModuleName, string(types.ParamStoreKeySlashWindow),\n\t\t\tfunc(r *rand.Rand) string {\n\t\t\t\treturn fmt.Sprintf(\"\\\"%d\\\"\", GenSlashWindow(r))\n\t\t\t},\n\t\t),\n\t}\n}", "func NewModifiedResources() *ModifiedResources {\n\treturn &ModifiedResources{\n\t\titems: make(map[string]struct{}),\n\t\tauthors: make(map[string]struct{}),\n\t}\n}", "func (input *BeegoInput) ResetParams() {\n\tinput.pnames = input.pnames[:0]\n\tinput.pvalues = input.pvalues[:0]\n}", "func deepCopy(copy, orig interface{}) error {\n\tvar buf bytes.Buffer\n\tenc := gob.NewEncoder(&buf)\n\tdec := gob.NewDecoder(&buf)\n\terr := enc.Encode(orig)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn dec.Decode(copy)\n}", "func NewParams() *Parameters {\n\treturn &Parameters{\n\t\tTokenType: \"transit\",\n\t\tTLSMode: 1,\n\t\tLeaderOnly: true,\n\t\tConnectTimeout: defaultConnectTimeout,\n\t\tReadTimeout: defaultReadTimeout,\n\t\tRetryCount: defaultRetries,\n\t\tRequireType: \"master\",\n\t}\n}", "func copyComponentsPtr(s Servable, req *http.Request) {\n\tctx := context.WithValue(req.Context(), componentsKey, s.ServerField().Components)\n\t*req = *req.WithContext(ctx)\n}", "func (o *QueryChangesParams) WithDefaults() *QueryChangesParams {\n\to.SetDefaults()\n\treturn o\n}", "func (r *Resources) Copy(other *Resources) {\n\tr.CPU = other.CPU\n\tr.DISK = other.DISK\n\tr.MEMORY = other.MEMORY\n\tr.GPU = other.GPU\n}", "func (m *IntentsDeviceManagementIntentItemRequestBuilder) CreateCopy()(*IntentsItemCreateCopyRequestBuilder) {\n return NewIntentsItemCreateCopyRequestBuilderInternal(m.BaseRequestBuilder.PathParameters, m.BaseRequestBuilder.RequestAdapter)\n}", "func NewParams(tokenCourse, subscriptionPrice, VPNGBPrice,\n\tstorageGBPrice, baseVPNGb, baseStorageGb uint32, courseSigners []sdk.AccAddress) Params {\n\treturn Params{\n\t\tTokenCourse: tokenCourse,\n\t\tSubscriptionPrice: subscriptionPrice,\n\t\tVPNGBPrice: VPNGBPrice,\n\t\tStorageGBPrice: storageGBPrice,\n\t\tBaseVPNGb: baseVPNGb,\n\t\tBaseStorageGb: baseStorageGb,\n\t\tCourseChangeSigners: courseSigners[:],\n\t}\n}", "func DefaultParams() Params {\n\treturn Params{\n\t\tMintDenom: sdk.DefaultBondDenom,\n\t\tGenesisEpochProvisions: sdk.NewDec(5000000),\n\t\tEpochIdentifier: \"week\", // 1 week\n\t\tReductionPeriodInEpochs: 156, // 3 years\n\t\tReductionFactor: sdk.NewDecWithPrec(5, 1), // 0.5\n\t\tDistributionProportions: DistributionProportions{\n\t\t\tStaking: sdk.NewDecWithPrec(4, 1), // 0.4\n\t\t\tPoolIncentives: sdk.NewDecWithPrec(3, 1), // 0.3\n\t\t\tDeveloperRewards: sdk.NewDecWithPrec(2, 1), // 0.2\n\t\t\tCommunityPool: sdk.NewDecWithPrec(1, 1), // 0.1\n\t\t},\n\t\tWeightedDeveloperRewardsReceivers: []WeightedAddress{},\n\t\tMintingRewardsDistributionStartEpoch: 0,\n\t}\n}", "func (m *ItemMailFoldersItemMessagesMessageItemRequestBuilder) Copy()(*ItemMailFoldersItemMessagesItemCopyRequestBuilder) {\n return NewItemMailFoldersItemMessagesItemCopyRequestBuilderInternal(m.BaseRequestBuilder.PathParameters, m.BaseRequestBuilder.RequestAdapter)\n}", "func (v CreateValidator) Copy() StakeMsg {\n\tv1 := v\n\tdesc := *v.Description\n\tv1.Description = &desc\n\treturn v1\n}", "func (in *BindingVolumeParams) DeepCopy() *BindingVolumeParams {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(BindingVolumeParams)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (c *Change) Reset() *Change {\n\tc.params = make(map[string]interface{})\n\treturn c\n}", "func (t *TaskDriverRun) Copy() *TaskDriverRun {\n\tif t == nil {\n\t\treturn nil\n\t}\n\tvar steps map[string]*Step\n\tif t.Steps != nil {\n\t\tsteps = make(map[string]*Step, len(t.Steps))\n\t\tfor k, v := range t.Steps {\n\t\t\tsteps[k] = v.Copy()\n\t\t}\n\t}\n\treturn &TaskDriverRun{\n\t\tTaskId: t.TaskId,\n\t\tProperties: t.Properties.Copy(),\n\t\tSteps: steps,\n\t}\n}", "func NewParams(communityTax sdk.Dec, withdrawAddrEnabled bool) Params {\n\treturn Params{\n\t\tCommunityTax: communityTax,\n\t\tWithdrawAddrEnabled: withdrawAddrEnabled,\n\t}\n}", "func (original *Place) Copy() *Place {\n\treturn &Place{original.latitude, original.longitude, original.Name}\n}", "func (o *Opts) Copy() *Opts {\n\tnewOpts := *o\n\treturn &newOpts\n}", "func NewMakeSnapshotArgs(layerID LayerID) *MakeSnapshotArgs {\n\targs := new(MakeSnapshotArgs)\n\targs.LayerID = layerID\n\treturn args\n}", "func copyAttrs(attrs map[string]interface{}) map[string]interface{} {\n\tduplicate := make(map[string]interface{}, len(attrs))\n\tfor k, v := range attrs {\n\t\tduplicate[k] = v\n\t}\n\treturn duplicate\n}", "func (cc *CopyCommand) Init(args []string, options OptionMapType) error {\n\treturn cc.command.Init(args, options, cc)\n}", "func (o *ProjectWebhook) Clone() datamodel.Model {\n\tc := new(ProjectWebhook)\n\tc.FromMap(o.ToMap())\n\treturn c\n}", "func (in DefaultParams) DeepCopy() DefaultParams {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(DefaultParams)\n\tin.DeepCopyInto(out)\n\treturn *out\n}", "func DefaultParams() Params {\n\treturn Params{\n\t\tBudgets: []Budget{},\n\t\tEpochBlocks: DefaultEpochBlocks,\n\t}\n}", "func (this *L0JobContext) Copy(request string) JobContext {\n\treturn &L0JobContext{\n\t\trequest: request,\n\t\tjobID: this.jobID,\n\t\tlogic: this.logic,\n\t\tloadBalancerLogic: this.loadBalancerLogic,\n\t\tserviceLogic: this.serviceLogic,\n\t\tenvironmentLogic: this.environmentLogic,\n\t\tmutex: this.mutex,\n\t}\n}", "func (j *Job) CloneWithEmptyArgs() (r *Job) {\n\tcp := *j\n\tr = &cp\n\tcp.Args = []string{}\n\n\tcp.Out = make([]string, len(j.Out))\n\tcopy(cp.Out, j.Out)\n\tcp.Env = make([]string, len(j.Env))\n\tcopy(cp.Env, j.Env)\n\tcp.Finishaddr = make([]string, len(j.Finishaddr))\n\tcopy(cp.Finishaddr, j.Finishaddr)\n\tStampJob(r)\n\treturn\n}", "func init() {\n\tRootCmd.AddCommand(cloneCmd)\n\n\tcloneCmd.Flags().StringVarP(&BackrestPVCSize, \"pgbackrest-pvc-size\", \"\", \"\",\n\t\t`The size of the PVC capacity for the pgBackRest repository. Overrides the value set in the storage class. This is ignored if the storage type of \"local\" is not used. Must follow the standard Kubernetes format, e.g. \"10.1Gi\"`)\n\tcloneCmd.Flags().StringVarP(&BackrestStorageSource, \"pgbackrest-storage-source\", \"\", \"\",\n\t\t\"The data source for the clone when both \\\"local\\\" and \\\"s3\\\" are enabled in the \"+\n\t\t\t\"source cluster. Either \\\"local\\\", \\\"s3\\\" or both, comma separated. (default \\\"local\\\")\")\n\tcloneCmd.Flags().BoolVar(&MetricsFlag, \"enable-metrics\", false, `If sets, enables metrics collection on the newly cloned cluster`)\n\tcloneCmd.Flags().StringVarP(&PVCSize, \"pvc-size\", \"\", \"\",\n\t\t`The size of the PVC capacity for primary and replica PostgreSQL instances. Overrides the value set in the storage class. Must follow the standard Kubernetes format, e.g. \"10.1Gi\"`)\n}", "func (in *PeeringRequest) DeepCopy() *PeeringRequest {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(PeeringRequest)\n\tin.DeepCopyInto(out)\n\treturn out\n}" ]
[ "0.7398709", "0.7115698", "0.6506534", "0.5838536", "0.575506", "0.55052036", "0.54349566", "0.54136103", "0.5332016", "0.5116452", "0.50422174", "0.48769543", "0.48056495", "0.47824895", "0.47173387", "0.470763", "0.4704235", "0.46622336", "0.46519724", "0.45844418", "0.4572863", "0.4571199", "0.4553177", "0.45291772", "0.45016047", "0.44923818", "0.4451014", "0.4430593", "0.44229752", "0.44178048", "0.43726182", "0.43614635", "0.4337656", "0.43368137", "0.43014342", "0.42683175", "0.4241892", "0.42098635", "0.42066357", "0.41897637", "0.41891125", "0.4187208", "0.41769338", "0.4171703", "0.41682282", "0.41510686", "0.41416737", "0.41267204", "0.41054416", "0.41051856", "0.41006398", "0.40940455", "0.40934187", "0.40875527", "0.40847808", "0.4083641", "0.40746537", "0.4072627", "0.40681276", "0.4067471", "0.40664482", "0.40620583", "0.40568003", "0.4045866", "0.404472", "0.40443504", "0.40404072", "0.4032808", "0.40295076", "0.40213028", "0.40159816", "0.40108806", "0.40082905", "0.40041715", "0.39942926", "0.3992565", "0.39907628", "0.3989566", "0.3989199", "0.3987946", "0.39862043", "0.3977043", "0.39768663", "0.39752552", "0.3973195", "0.39713553", "0.39666712", "0.3960216", "0.39556953", "0.3952889", "0.3950046", "0.39497396", "0.39335942", "0.39302617", "0.39253962", "0.39246148", "0.39233983", "0.39184326", "0.39172593", "0.39164925" ]
0.83219415
0
NewCopyRecipeToMyRecipesWithChangesParamsWithTimeout creates a new CopyRecipeToMyRecipesWithChangesParams object with the default values initialized, and the ability to set a timeout on a request
func NewCopyRecipeToMyRecipesWithChangesParamsWithTimeout(timeout time.Duration) *CopyRecipeToMyRecipesWithChangesParams { var () return &CopyRecipeToMyRecipesWithChangesParams{ timeout: timeout, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *CopyRecipeToMyRecipesWithChangesParams) WithTimeout(timeout time.Duration) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewCopyRecipeToMyRecipesWithChangesParams() *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func NewCopyRecipeToMyRecipesWithChangesParamsWithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\t\tHTTPClient: client,\n\t}\n}", "func (o *GetPublicsRecipeParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewNarrowSearchRecipeParamsWithTimeout(timeout time.Duration) *NarrowSearchRecipeParams {\n\tvar (\n\t\tsortbyDefault = string(\"name\")\n\t\tsortdirDefault = string(\"desc\")\n\t)\n\treturn &NarrowSearchRecipeParams{\n\t\tSortby: &sortbyDefault,\n\t\tSortdir: &sortdirDefault,\n\n\t\ttimeout: timeout,\n\t}\n}", "func NewQueryChangesParamsWithTimeout(timeout time.Duration) *QueryChangesParams {\n\treturn &QueryChangesParams{\n\t\ttimeout: timeout,\n\t}\n}", "func (o *QueryChangesParams) WithTimeout(timeout time.Duration) *QueryChangesParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *QueryChangesParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewPostReconciliationParamsWithTimeout(timeout time.Duration) *PostReconciliationParams {\n\tvar ()\n\treturn &PostReconciliationParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func NewGetPublicsRecipeParamsWithTimeout(timeout time.Duration) *GetPublicsRecipeParams {\n\n\treturn &GetPublicsRecipeParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func NewCreateWidgetParamsWithTimeout(timeout time.Duration) *CreateWidgetParams {\n\tvar (\n\t\tacceptDefault = string(\"application/json\")\n\t\tcontentTypeDefault = string(\"application/json\")\n\t)\n\treturn &CreateWidgetParams{\n\t\tAccept: &acceptDefault,\n\t\tContentType: &contentTypeDefault,\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *ChatNewParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewConfigurationBackupModifyParamsWithTimeout(timeout time.Duration) *ConfigurationBackupModifyParams {\n\treturn &ConfigurationBackupModifyParams{\n\t\ttimeout: timeout,\n\t}\n}", "func NewContainerUpdateParamsWithTimeout(timeout time.Duration) *ContainerUpdateParams {\n\tvar ()\n\treturn &ContainerUpdateParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *CreateGitWebhookUsingPOSTParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetReceiptsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewGetReceiptsParamsWithTimeout(timeout time.Duration) *GetReceiptsParams {\n\tvar ()\n\treturn &GetReceiptsParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *NarrowSearchRecipeParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.JSONBody != nil {\n\t\tif err := r.SetBodyParam(o.JSONBody); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param recipeId\n\tif err := r.SetPathParam(\"recipeId\", o.RecipeID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func NewEditParamsWithTimeout(timeout time.Duration) *EditParams {\n\treturn &EditParams{\n\t\ttimeout: timeout,\n\t}\n}", "func (o *PetCreateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ActionDeploymentRequestUsingPOST2Params) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewUpdateBuildPropertiesParamsWithTimeout(timeout time.Duration) *UpdateBuildPropertiesParams {\n\tvar ()\n\treturn &UpdateBuildPropertiesParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func NewSizeParamsWithTimeout(timeout time.Duration) *SizeParams {\n\tvar ()\n\treturn &SizeParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func NewFreezeParamsWithTimeout(timeout time.Duration) *FreezeParams {\n\treturn &FreezeParams{\n\t\ttimeout: timeout,\n\t}\n}", "func NewPostApplyManifestParamsWithTimeout(timeout time.Duration) *PostApplyManifestParams {\n\tvar (\n\t\tdisableUpdatePostDefault = bool(false)\n\t\tdisableUpdatePreDefault = bool(false)\n\t\tdryRunDefault = bool(false)\n\t\tenableChartCleanupDefault = bool(false)\n\t\ttillerHostDefault = string(\"None\")\n\t\ttimeoutDefault = int64(3600)\n\t)\n\treturn &PostApplyManifestParams{\n\t\tDisableUpdatePost: &disableUpdatePostDefault,\n\t\tDisableUpdatePre: &disableUpdatePreDefault,\n\t\tDryRun: &dryRunDefault,\n\t\tEnableChartCleanup: &enableChartCleanupDefault,\n\t\tTillerHost: &tillerHostDefault,\n\t\tTimeout: &timeoutDefault,\n\n\t\trequestTimeout: timeout,\n\t}\n}", "func NewContainerRenameParamsWithTimeout(timeout time.Duration) *ContainerRenameParams {\n\tvar ()\n\treturn &ContainerRenameParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *CreateRunbookRunCreateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetBuildPropertiesParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewTimeout(parameters ...wparams.ParamStorer) Error {\n\treturn newGenericError(nil, DefaultTimeout, wparams.NewParamStorer(parameters...))\n}", "func (o *GetIngredientVersionRevisionParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *FileInfoCreateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *RevertProductSnapshotRequestUsingPOSTParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewSwarmUpdateParamsWithTimeout(timeout time.Duration) *SwarmUpdateParams {\n\tvar (\n\t\trotateManagerTokenDefault = bool(false)\n\t\trotateManagerUnlockKeyDefault = bool(false)\n\t\trotateWorkerTokenDefault = bool(false)\n\t)\n\treturn &SwarmUpdateParams{\n\t\tRotateManagerToken: &rotateManagerTokenDefault,\n\t\tRotateManagerUnlockKey: &rotateManagerUnlockKeyDefault,\n\t\tRotateWorkerToken: &rotateWorkerTokenDefault,\n\n\t\ttimeout: timeout,\n\t}\n}", "func (c *OrganizationsEnvironmentsApisRevisionsDebugsessionsCreateCall) Timeout(timeout int64) *OrganizationsEnvironmentsApisRevisionsDebugsessionsCreateCall {\n\tc.urlParams_.Set(\"timeout\", fmt.Sprint(timeout))\n\treturn c\n}", "func (o *GetContentSourceUsingGETParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *AddRepositoryParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func WithTimeout(timeoutType fab.TimeoutType, timeout time.Duration) RequestOption {\n\treturn func(ctx context.Client, o *requestOptions) error {\n\t\tif o.Timeouts == nil {\n\t\t\to.Timeouts = make(map[fab.TimeoutType]time.Duration)\n\t\t}\n\t\to.Timeouts[timeoutType] = timeout\n\t\treturn nil\n\t}\n}", "func NewUpdateWidgetParamsWithTimeout(timeout time.Duration) *UpdateWidgetParams {\n\tvar (\n\t\tacceptDefault = string(\"application/json\")\n\t\tcontentTypeDefault = string(\"application/json\")\n\t)\n\treturn &UpdateWidgetParams{\n\t\tAccept: &acceptDefault,\n\t\tContentType: &contentTypeDefault,\n\n\t\ttimeout: timeout,\n\t}\n}", "func NewInventoryStocktakingSearchParamsWithTimeout(timeout time.Duration) *InventoryStocktakingSearchParams {\n\tvar (\n\t\tcountDefault = int64(1000)\n\t\tfromDefault = int64(0)\n\t)\n\treturn &InventoryStocktakingSearchParams{\n\t\tCount: &countDefault,\n\t\tFrom: &fromDefault,\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *CloudTargetCreateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewGetRequestDetailsParamsWithTimeout(timeout time.Duration) *GetRequestDetailsParams {\n\tvar ()\n\treturn &GetRequestDetailsParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *CreateLifecycleParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *StorageServiceOwnershipGetParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateCrossConnectParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CapacityPoolGetParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ServiceInstanceGetParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewGetBuildPropertiesParamsWithTimeout(timeout time.Duration) *GetBuildPropertiesParams {\n\tvar ()\n\treturn &GetBuildPropertiesParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *NarrowSearchRecipeParams) WithTimeout(timeout time.Duration) *NarrowSearchRecipeParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *GetRequestDetailsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetWorkItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewUpdateCredentialParamsWithTimeout(timeout time.Duration) *UpdateCredentialParams {\n\tvar ()\n\treturn &UpdateCredentialParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *RegenerateDeployKeyParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *EditParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateCartUsingPOSTParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ChatNewParams) WithTimeout(timeout time.Duration) *ChatNewParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *ConfigurationBackupGetParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ConfigurationBackupModifyParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *BackupsCreateStatusParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PostReconciliationParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewConvertParamsWithTimeout(timeout time.Duration) *ConvertParams {\n\tvar ()\n\treturn &ConvertParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func NewDeleteKeyPairsParamsWithTimeout(timeout time.Duration) *DeleteKeyPairsParams {\n\tvar ()\n\treturn &DeleteKeyPairsParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *IntegrationsManualHTTPSCreateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewAddItemParamsWithTimeout(timeout time.Duration) *AddItemParams {\n\tvar ()\n\treturn &AddItemParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func NewChatNewParamsWithTimeout(timeout time.Duration) *ChatNewParams {\n\tvar (\n\t\tchannelIDDefault = float64(1)\n\t)\n\treturn &ChatNewParams{\n\t\tChannelID: &channelIDDefault,\n\n\t\ttimeout: timeout,\n\t}\n}", "func NewBarParamsWithTimeout(timeout time.Duration) *BarParams {\n\tvar ()\n\treturn &BarParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *CreateWidgetParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetRedbeamsFlowLogsProgressByResourceCrnParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewPatchZoneParamsWithTimeout(timeout time.Duration) *PatchZoneParams {\n\tvar ()\n\treturn &PatchZoneParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *UpdateBuildPropertiesParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetRemotesupportConnectemcParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *EditParams) WithTimeout(timeout time.Duration) *EditParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func NewRegenerateDeployKeyParamsWithTimeout(timeout time.Duration) *RegenerateDeployKeyParams {\n\treturn &RegenerateDeployKeyParams{\n\t\ttimeout: timeout,\n\t}\n}", "func (o *GetPrivateToggleDepositAddressCreationParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreatePolicyResetItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewRevertProductSnapshotRequestUsingPOSTParamsWithTimeout(timeout time.Duration) *RevertProductSnapshotRequestUsingPOSTParams {\n\tvar ()\n\treturn &RevertProductSnapshotRequestUsingPOSTParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *SyncStatusUsingGETParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithContext(ctx context.Context) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (o *PutParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewPatchAddonParamsWithTimeout(timeout time.Duration) *PatchAddonParams {\n\tvar ()\n\treturn &PatchAddonParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *CreateInstantPaymentParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetTerraformConfigurationSourcesUsingGET1Params) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *InventoryStocktakingSearchParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetRepository15Params) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ConvertParams) WithTimeout(timeout time.Duration) *ConvertParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *PostPartsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ApplicationComponentSnapshotCollectionGetParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *DeleteKeyPairsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateBlueprintInWorkspaceInternalParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func TimeoutSeconds(timeout int) Option {\n\treturn func(opts workerOpts) workerOpts {\n\t\topts.jobTimeoutSeconds = timeout\n\t\treturn opts\n\t}\n}", "func (o *GetDeploymentByIDV3UsingGETParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func WithTimeout(duration time.Duration) ReconcilerOption {\n\treturn func(r *Reconciler) {\n\t\tr.timeout = duration\n\t}\n}", "func NewUpdateTableMetadataParamsWithTimeout(timeout time.Duration) *UpdateTableMetadataParams {\n\tvar ()\n\treturn &UpdateTableMetadataParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func NewFileInfoCreateParamsWithTimeout(timeout time.Duration) *FileInfoCreateParams {\n\treturn &FileInfoCreateParams{\n\t\ttimeout: timeout,\n\t}\n}", "func (o *GetPaymentsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *BudgetAddParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetContentSourcesUsingGETParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreatePolicyParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}" ]
[ "0.7334397", "0.6605116", "0.63352853", "0.5800872", "0.56409484", "0.5607914", "0.54550415", "0.5356853", "0.5310841", "0.53100616", "0.5263849", "0.5247955", "0.52475536", "0.5226695", "0.52168256", "0.5173177", "0.5161624", "0.5160964", "0.5154121", "0.5137564", "0.5129438", "0.509061", "0.5075947", "0.5044079", "0.5043573", "0.50402194", "0.50364774", "0.5015112", "0.50103754", "0.49984336", "0.4994417", "0.49911186", "0.4985919", "0.49803686", "0.4977538", "0.49705553", "0.49669132", "0.4957732", "0.49575713", "0.4945712", "0.49426445", "0.49413744", "0.49387884", "0.49382064", "0.49378327", "0.49368423", "0.49167964", "0.49158952", "0.49145314", "0.49061963", "0.49027383", "0.49021363", "0.48972574", "0.48949277", "0.48921365", "0.48889342", "0.48874995", "0.48872876", "0.48839194", "0.48815405", "0.4879294", "0.4875162", "0.4873673", "0.48722887", "0.48593533", "0.48591775", "0.48553243", "0.48535115", "0.48518384", "0.48517892", "0.48504266", "0.48492536", "0.48456073", "0.4843525", "0.4841754", "0.48108134", "0.4804985", "0.4801075", "0.4800462", "0.4796924", "0.47938764", "0.47933123", "0.47928", "0.4790575", "0.47857007", "0.47820508", "0.47798878", "0.4778629", "0.4777421", "0.47754964", "0.47671118", "0.4759266", "0.47591543", "0.4759048", "0.4757747", "0.47555506", "0.47548163", "0.47480404", "0.4747348", "0.4744736" ]
0.7710058
0
NewCopyRecipeToMyRecipesWithChangesParamsWithContext creates a new CopyRecipeToMyRecipesWithChangesParams object with the default values initialized, and the ability to set a context for a request
func NewCopyRecipeToMyRecipesWithChangesParamsWithContext(ctx context.Context) *CopyRecipeToMyRecipesWithChangesParams { var () return &CopyRecipeToMyRecipesWithChangesParams{ Context: ctx, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewCopyRecipeToMyRecipesWithChangesParams() *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithContext(ctx context.Context) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func NewCopyRecipeToMyRecipesWithChangesParamsWithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\t\tHTTPClient: client,\n\t}\n}", "func NewCopyRecipeToMyRecipesWithChangesParamsWithTimeout(timeout time.Duration) *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithRecipeID(recipeID string) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetRecipeID(recipeID)\n\treturn o\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithTimeout(timeout time.Duration) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithJSONBody(jSONBody *models.Recipe) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetJSONBody(jSONBody)\n\treturn o\n}", "func NewQueryChangesParamsWithContext(ctx context.Context) *QueryChangesParams {\n\treturn &QueryChangesParams{\n\t\tContext: ctx,\n\t}\n}", "func (o *QueryChangesParams) WithContext(ctx context.Context) *QueryChangesParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (in *GitContext) DeepCopy() *GitContext {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(GitContext)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.JSONBody != nil {\n\t\tif err := r.SetBodyParam(o.JSONBody); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param recipeId\n\tif err := r.SetPathParam(\"recipeId\", o.RecipeID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (c *SMS) GenerateChangeSetWithContext(ctx aws.Context, input *GenerateChangeSetInput, opts ...request.Option) (*GenerateChangeSetOutput, error) {\n\treq, out := c.GenerateChangeSetRequest(input)\n\treq.SetContext(ctx)\n\treq.ApplyOptions(opts...)\n\treturn out, req.Send()\n}", "func NewPostV1ChangesChangeIDIdentitiesParamsWithContext(ctx context.Context) *PostV1ChangesChangeIDIdentitiesParams {\n\tvar ()\n\treturn &PostV1ChangesChangeIDIdentitiesParams{\n\n\t\tContext: ctx,\n\t}\n}", "func NewConfigurationBackupModifyParamsWithContext(ctx context.Context) *ConfigurationBackupModifyParams {\n\treturn &ConfigurationBackupModifyParams{\n\t\tContext: ctx,\n\t}\n}", "func NewPutClientConfigV2NamespaceChangesParamsWithContext(ctx context.Context) *PutClientConfigV2NamespaceChangesParams {\n\tvar ()\n\treturn &PutClientConfigV2NamespaceChangesParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (o *CommitOptions) WithChanges(value []string) *CommitOptions {\n\to.Changes = value\n\treturn o\n}", "func NewGetRecentFoodsParamsWithContext(ctx context.Context) *GetRecentFoodsParams {\n\n\treturn &GetRecentFoodsParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (o *GetReceiptsParams) WithContext(ctx context.Context) *GetReceiptsParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) SetRecipeID(recipeID string) {\n\to.RecipeID = recipeID\n}", "func (o *PatchV1ChangesEventsChangeEventIDParams) WithContext(ctx context.Context) *PatchV1ChangesEventsChangeEventIDParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (c *Client) CopyWithContext(ctx context.Context, src FileID, dst string) (FileID, error) {\n\tfileID, err := c.iclient.Copy(ctx, &pb.CopyRequest{SrcId: src[:], Dst: dst})\n\tif e, ok := err.(twirp.Error); ok && e.Code() == twirp.NotFound {\n\t\treturn FileID{}, ErrNotFound\n\t}\n\tif err != nil {\n\t\treturn FileID{}, err\n\t}\n\ts, err := UnmarshalFileID(fileID.Sum)\n\tif err != nil {\n\t\treturn FileID{}, err\n\t}\n\treturn s, nil\n}", "func NewGetReceiptsParamsWithContext(ctx context.Context) *GetReceiptsParams {\n\tvar ()\n\treturn &GetReceiptsParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (o *QueryChangesParams) WithFilter(filter *string) *QueryChangesParams {\n\to.SetFilter(filter)\n\treturn o\n}", "func NewConversationsSendToConversationParamsWithContext(ctx context.Context) *ConversationsSendToConversationParams {\n\tvar ()\n\treturn &ConversationsSendToConversationParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (in *HelmRequest) DeepCopy() *HelmRequest {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(HelmRequest)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (o *GetIngredientVersionRevisionParams) WithContext(ctx context.Context) *GetIngredientVersionRevisionParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func NewPatchV1ChangesEventsChangeEventIDParamsWithContext(ctx context.Context) *PatchV1ChangesEventsChangeEventIDParams {\n\tvar ()\n\treturn &PatchV1ChangesEventsChangeEventIDParams{\n\n\t\tContext: ctx,\n\t}\n}", "func NewCopyToArgs(nodeID NodeID, targetNodeID NodeID) *CopyToArgs {\n\targs := new(CopyToArgs)\n\targs.NodeID = nodeID\n\targs.TargetNodeID = targetNodeID\n\treturn args\n}", "func (o *ConfigurationBackupModifyParams) WithContext(ctx context.Context) *ConfigurationBackupModifyParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (in *HelmRequestSpec) DeepCopy() *HelmRequestSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(HelmRequestSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (s *TemplateService) NewCopyTemplateParams(id string) *CopyTemplateParams {\n\tp := &CopyTemplateParams{}\n\tp.p = make(map[string]interface{})\n\tp.p[\"id\"] = id\n\treturn p\n}", "func (p *PullRequestEvent) GetChanges() *EditChange {\n\tif p == nil {\n\t\treturn nil\n\t}\n\treturn p.Changes\n}", "func (m *ChargeRequestAdditionalRecipient) ContextValidate(ctx context.Context, formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := m.contextValidateAmountMoney(ctx, formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (in *ChoiceSpec) DeepCopy() *ChoiceSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(ChoiceSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (p *PullRequestReviewCommentEvent) GetChanges() *EditChange {\n\tif p == nil {\n\t\treturn nil\n\t}\n\treturn p.Changes\n}", "func NewNotifyStateChangedParamsWithContext(ctx context.Context) *NotifyStateChangedParams {\n\tvar ()\n\treturn &NotifyStateChangedParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (a *ManagementApiService) GetChanges(ctx _context.Context) apiGetChangesRequest {\n\treturn apiGetChangesRequest{\n\t\tapiService: a,\n\t\tctx: ctx,\n\t}\n}", "func (in *NewPendingTxNotificationParams) DeepCopy() *NewPendingTxNotificationParams {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(NewPendingTxNotificationParams)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func NewPostReconciliationParamsWithContext(ctx context.Context) *PostReconciliationParams {\n\tvar ()\n\treturn &PostReconciliationParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (o *PostV1ChangesChangeIDIdentitiesParams) WithContext(ctx context.Context) *PostV1ChangesChangeIDIdentitiesParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func NewQueryChangesParamsWithHTTPClient(client *http.Client) *QueryChangesParams {\n\treturn &QueryChangesParams{\n\t\tHTTPClient: client,\n\t}\n}", "func NewQueryChangesParams() *QueryChangesParams {\n\treturn &QueryChangesParams{\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func NewChatNewParamsWithContext(ctx context.Context) *ChatNewParams {\n\tvar (\n\t\tchannelIdDefault = float64(1)\n\t)\n\treturn &ChatNewParams{\n\t\tChannelID: &channelIdDefault,\n\n\t\tContext: ctx,\n\t}\n}", "func (project *ProjectV1) GetConfigDiffWithContext(ctx context.Context, getConfigDiffOptions *GetConfigDiffOptions) (result *ProjectConfigDiff, response *core.DetailedResponse, err error) {\n\terr = core.ValidateNotNil(getConfigDiffOptions, \"getConfigDiffOptions cannot be nil\")\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.ValidateStruct(getConfigDiffOptions, \"getConfigDiffOptions\")\n\tif err != nil {\n\t\treturn\n\t}\n\n\tpathParamsMap := map[string]string{\n\t\t\"project_id\": *getConfigDiffOptions.ProjectID,\n\t\t\"id\": *getConfigDiffOptions.ID,\n\t}\n\n\tbuilder := core.NewRequestBuilder(core.GET)\n\tbuilder = builder.WithContext(ctx)\n\tbuilder.EnableGzipCompression = project.GetEnableGzipCompression()\n\t_, err = builder.ResolveRequestURL(project.Service.Options.URL, `/v1/projects/{project_id}/configs/{id}/diff`, pathParamsMap)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tfor headerName, headerValue := range getConfigDiffOptions.Headers {\n\t\tbuilder.AddHeader(headerName, headerValue)\n\t}\n\n\tsdkHeaders := common.GetSdkHeaders(\"project\", \"V1\", \"GetConfigDiff\")\n\tfor headerName, headerValue := range sdkHeaders {\n\t\tbuilder.AddHeader(headerName, headerValue)\n\t}\n\tbuilder.AddHeader(\"Accept\", \"application/json\")\n\n\trequest, err := builder.Build()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tvar rawResponse map[string]json.RawMessage\n\tresponse, err = project.Service.Request(request, &rawResponse)\n\tif err != nil {\n\t\treturn\n\t}\n\tif rawResponse != nil {\n\t\terr = core.UnmarshalModel(rawResponse, \"\", &result, UnmarshalProjectConfigDiff)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tresponse.Result = result\n\t}\n\n\treturn\n}", "func NewRevertProductSnapshotRequestUsingPOSTParamsWithContext(ctx context.Context) *RevertProductSnapshotRequestUsingPOSTParams {\n\tvar ()\n\treturn &RevertProductSnapshotRequestUsingPOSTParams{\n\n\t\tContext: ctx,\n\t}\n}", "func NewGetRecipesParams() GetRecipesParams {\n\n\tvar (\n\t\t// initialize parameters with default values\n\n\t\tingredient1Default = string(\"\")\n\t\tingredient2Default = string(\"\")\n\t\tingredient3Default = string(\"\")\n\t\tseasonDefault = string(\"\")\n\t)\n\n\treturn GetRecipesParams{\n\t\tIngredient1: &ingredient1Default,\n\n\t\tIngredient2: &ingredient2Default,\n\n\t\tIngredient3: &ingredient3Default,\n\n\t\tSeason: &seasonDefault,\n\t}\n}", "func NewMonitorCheckGetScreenshotsParamsWithContext(ctx context.Context) *MonitorCheckGetScreenshotsParams {\n\tvar ()\n\treturn &MonitorCheckGetScreenshotsParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (in *BuildContext) DeepCopy() *BuildContext {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(BuildContext)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func NewPostV1IncidentsIncidentIDRelatedChangeEventsParamsWithContext(ctx context.Context) *PostV1IncidentsIncidentIDRelatedChangeEventsParams {\n\tvar ()\n\treturn &PostV1IncidentsIncidentIDRelatedChangeEventsParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (ctx *Context) Copy() *Context {\n\tvar pathParams Params\n\tif len(ctx.PathParams) > 0 {\n\t\tpathParams = append(pathParams, ctx.PathParams...)\n\t}\n\treturn &Context{\n\t\tresponseWriter2: nil,\n\t\tResponseWriter: nil,\n\t\tRequest: ctx.Request,\n\t\tPathParams: pathParams,\n\t\tqueryParams: ctx.queryParams,\n\t\tValidator: ctx.Validator,\n\t\tfetchClientIPFromHeader: ctx.fetchClientIPFromHeader,\n\t\thandlers: nil,\n\t\thandlerIndex: __abortHandlerIndex,\n\t\tkvs: ctx.kvs,\n\t}\n}", "func NewGetLolCatalogV1ItemsParamsWithContext(ctx context.Context) *GetLolCatalogV1ItemsParams {\n\tvar ()\n\treturn &GetLolCatalogV1ItemsParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (o *PostV1IncidentsIncidentIDRelatedChangeEventsParams) WithContext(ctx context.Context) *PostV1IncidentsIncidentIDRelatedChangeEventsParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func NewPostAPIV10PeerReviewsParamsWithContext(ctx context.Context) *PostAPIV10PeerReviewsParams {\n\tvar ()\n\treturn &PostAPIV10PeerReviewsParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (in *CredentialProviderRequest) DeepCopy() *CredentialProviderRequest {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(CredentialProviderRequest)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (in *CloudEventContext) DeepCopy() *CloudEventContext {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(CloudEventContext)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func Copy(scope *Scope, input tf.Output, optional ...CopyAttr) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{}\n\tfor _, a := range optional {\n\t\ta(attrs)\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"Copy\",\n\t\tInput: []tf.Input{\n\t\t\tinput,\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func (in *Choice) DeepCopy() *Choice {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Choice)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func NewRewardCommentsParamsWithContext(ctx context.Context) *RewardCommentsParams {\n\tvar ()\n\treturn &RewardCommentsParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (in *GitCreds) DeepCopy() *GitCreds {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(GitCreds)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (in *GitSpec) DeepCopy() *GitSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(GitSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func NewGetIngredientVersionRevisionParamsWithContext(ctx context.Context) *GetIngredientVersionRevisionParams {\n\treturn &GetIngredientVersionRevisionParams{\n\t\tContext: ctx,\n\t}\n}", "func (m *StepData) GetContextParameterFilters() StepFilters {\n\tvar filters StepFilters\n\tcontextFilters := []string{}\n\tfor _, secret := range m.Spec.Inputs.Secrets {\n\t\tcontextFilters = append(contextFilters, secret.Name)\n\t}\n\n\tif len(m.Spec.Inputs.Resources) > 0 {\n\t\tfor _, res := range m.Spec.Inputs.Resources {\n\t\t\tif res.Type == \"stash\" {\n\t\t\t\tcontextFilters = append(contextFilters, \"stashContent\")\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\tif len(m.Spec.Containers) > 0 {\n\t\tparameterKeys := []string{\"containerCommand\", \"containerShell\", \"dockerEnvVars\", \"dockerImage\", \"dockerName\", \"dockerOptions\", \"dockerPullImage\", \"dockerVolumeBind\", \"dockerWorkspace\", \"dockerRegistryUrl\", \"dockerRegistryCredentialsId\"}\n\t\tfor _, container := range m.Spec.Containers {\n\t\t\tfor _, condition := range container.Conditions {\n\t\t\t\tfor _, dependentParam := range condition.Params {\n\t\t\t\t\tparameterKeys = append(parameterKeys, dependentParam.Value)\n\t\t\t\t\tparameterKeys = append(parameterKeys, dependentParam.Name)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\t// ToDo: append dependentParam.Value & dependentParam.Name only according to correct parameter scope and not generally\n\t\tcontextFilters = append(contextFilters, parameterKeys...)\n\t}\n\tif len(m.Spec.Sidecars) > 0 {\n\t\t//ToDo: support fallback for \"dockerName\" configuration property -> via aliasing?\n\t\tcontextFilters = append(contextFilters, []string{\"containerName\", \"containerPortMappings\", \"dockerName\", \"sidecarEnvVars\", \"sidecarImage\", \"sidecarName\", \"sidecarOptions\", \"sidecarPullImage\", \"sidecarReadyCommand\", \"sidecarVolumeBind\", \"sidecarWorkspace\"}...)\n\t\t//ToDo: add condition param.Value and param.Name to filter as for Containers\n\t}\n\n\tcontextFilters = addVaultContextParametersFilter(m, contextFilters)\n\n\tif len(contextFilters) > 0 {\n\t\tfilters.All = append(filters.All, contextFilters...)\n\t\tfilters.General = append(filters.General, contextFilters...)\n\t\tfilters.Steps = append(filters.Steps, contextFilters...)\n\t\tfilters.Stages = append(filters.Stages, contextFilters...)\n\t\tfilters.Parameters = append(filters.Parameters, contextFilters...)\n\t\tfilters.Env = append(filters.Env, contextFilters...)\n\n\t}\n\treturn filters\n}", "func (c *SpanContext) CopyFrom(ctx *SpanContext) {\n\tc.traceID = ctx.traceID\n\tc.spanID = ctx.spanID\n\tc.parentID = ctx.parentID\n\tc.samplingState = ctx.samplingState\n\tif l := len(ctx.baggage); l > 0 {\n\t\tc.baggage = make(map[string]string, l)\n\t\tfor k, v := range ctx.baggage {\n\t\t\tc.baggage[k] = v\n\t\t}\n\t} else {\n\t\tc.baggage = nil\n\t}\n}", "func (c *Krds) ModifySecurityGroupWithContext(ctx aws.Context, input *map[string]interface{}, opts ...request.Option) (*map[string]interface{}, error) {\n\treq, out := c.ModifySecurityGroupRequest(input)\n\treq.SetContext(ctx)\n\treq.ApplyOptions(opts...)\n\treturn out, req.Send()\n}", "func (in *BackupResourceHookSpec) DeepCopy() *BackupResourceHookSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(BackupResourceHookSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (o *CommitOptions) GetChanges() []string {\n\tif o.Changes == nil {\n\t\tvar z []string\n\t\treturn z\n\t}\n\treturn o.Changes\n}", "func (in *NewPendingTxBodyNotificationParams) DeepCopy() *NewPendingTxBodyNotificationParams {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(NewPendingTxBodyNotificationParams)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func NewChangeSet(ctx *pulumi.Context,\n\tname string, args *ChangeSetArgs, opts ...pulumi.ResourceOption) (*ChangeSet, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.ChangeSetName == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'ChangeSetName'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource ChangeSet\n\terr := ctx.RegisterResource(\"alicloud:ros/changeSet:ChangeSet\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (c *Krds) CloneSecurityGroupWithContext(ctx aws.Context, input *map[string]interface{}, opts ...request.Option) (*map[string]interface{}, error) {\n\treq, out := c.CloneSecurityGroupRequest(input)\n\treq.SetContext(ctx)\n\treq.ApplyOptions(opts...)\n\treturn out, req.Send()\n}", "func (in *ContinuousParameterRangeSpecification) DeepCopy() *ContinuousParameterRangeSpecification {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(ContinuousParameterRangeSpecification)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (o *GetPublicsRecipeParams) WithContext(ctx context.Context) *GetPublicsRecipeParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func NewUpdateBuildPropertiesParamsWithContext(ctx context.Context) *UpdateBuildPropertiesParams {\n\tvar ()\n\treturn &UpdateBuildPropertiesParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (o *GetRecentFoodsParams) WithContext(ctx context.Context) *GetRecentFoodsParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (om *Sdk) ApplyChanges(args []string) error {\n\tlogWriter := commands.NewLogWriter(os.Stdout)\n\tcmd := commands.NewApplyChanges(om.api, om.api, logWriter, om.logger, 10)\n\treturn cmd.Execute(args)\n}", "func (in *RestoreResourceHookSpec) DeepCopy() *RestoreResourceHookSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(RestoreResourceHookSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (mr *MockPullRequestClientMockRecorder) GetPullRequestChanges(org, repo, number interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"GetPullRequestChanges\", reflect.TypeOf((*MockPullRequestClient)(nil).GetPullRequestChanges), org, repo, number)\n}", "func NewNarrowSearchRecipeParamsWithContext(ctx context.Context) *NarrowSearchRecipeParams {\n\tvar (\n\t\tsortbyDefault = string(\"name\")\n\t\tsortdirDefault = string(\"desc\")\n\t)\n\treturn &NarrowSearchRecipeParams{\n\t\tSortby: &sortbyDefault,\n\t\tSortdir: &sortdirDefault,\n\n\t\tContext: ctx,\n\t}\n}", "func (m *InstallVmtoolsParams) ContextValidate(ctx context.Context, formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := m.contextValidateData(ctx, formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.contextValidateWhere(ctx, formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func NewExtrasSavedFiltersListParamsWithContext(ctx context.Context) *ExtrasSavedFiltersListParams {\n\treturn &ExtrasSavedFiltersListParams{\n\t\tContext: ctx,\n\t}\n}", "func NewGetServiceDetailsParamsWithContext(ctx context.Context) *GetServiceDetailsParams {\n\tvar ()\n\treturn &GetServiceDetailsParams{\n\n\t\tContext: ctx,\n\t}\n}", "func RevisionContext(ns, svc, cfg, rev string) context.Context {\n\tkey := types.NamespacedName{Namespace: ns, Name: rev}\n\tif ctx, ok := contextCache.Get(key); ok {\n\t\treturn ctx.(context.Context)\n\t}\n\n\tctx := augmentWithRevision(context.Background(), ns, svc, cfg, rev)\n\tcontextCache.Add(key, ctx)\n\n\treturn ctx\n}", "func CopyWithContext(ctx context.Context, dst *Writer, src Stream) error {\n\tif err := src.Open(); err != nil {\n\t\treturn err\n\t}\n\tvar err error\n\tfor ctx.Err() == nil {\n\t\tvar pair Pair\n\t\tpair, err = src.Read()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif pair.Key == nil {\n\t\t\tbreak\n\t\t}\n\t\terr = dst.Write(pair)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn ctx.Err()\n}", "func (m *MockchangeSetAPI) WaitUntilChangeSetCreateCompleteWithContext(arg0 aws.Context, arg1 *cloudformation.DescribeChangeSetInput, arg2 ...request.WaiterOption) error {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"WaitUntilChangeSetCreateCompleteWithContext\", varargs...)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (in *S3Context) DeepCopy() *S3Context {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(S3Context)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (o *GetContactsParams) WithModifiedSince(modifiedSince *strfmt.DateTime) *GetContactsParams {\n\to.SetModifiedSince(modifiedSince)\n\treturn o\n}", "func NewGetUserAuditLogsConnectionParamsWithContext(ctx context.Context) *GetUserAuditLogsConnectionParams {\n\treturn &GetUserAuditLogsConnectionParams{\n\t\tContext: ctx,\n\t}\n}", "func (m *Mockapi) WaitUntilChangeSetCreateCompleteWithContext(arg0 aws.Context, arg1 *cloudformation.DescribeChangeSetInput, arg2 ...request.WaiterOption) error {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"WaitUntilChangeSetCreateCompleteWithContext\", varargs...)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (in *CrdSpec) DeepCopy() *CrdSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(CrdSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (in *CommonSpec) DeepCopy() *CommonSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(CommonSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (in *CommonSpec) DeepCopy() *CommonSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(CommonSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (m *MockRDSAPI) CopyDBClusterSnapshotWithContext(arg0 aws.Context, arg1 *rds.CopyDBClusterSnapshotInput, arg2 ...request.Option) (*rds.CopyDBClusterSnapshotOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"CopyDBClusterSnapshotWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.CopyDBClusterSnapshotOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (m *MockS3API) UploadPartCopyWithContext(arg0 context.Context, arg1 *s3.UploadPartCopyInput, arg2 ...request.Option) (*s3.UploadPartCopyOutput, error) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"UploadPartCopyWithContext\", varargs...)\n\tret0, _ := ret[0].(*s3.UploadPartCopyOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func NewCustomerGatewayUpdateOwnershipParamsWithContext(ctx context.Context) *CustomerGatewayUpdateOwnershipParams {\n\tvar ()\n\treturn &CustomerGatewayUpdateOwnershipParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (o *WorkflowCatalogServiceRequest) GetSelectionCriteriaInputs() []ServiceitemSelectionCriteriaInput {\n\tif o == nil {\n\t\tvar ret []ServiceitemSelectionCriteriaInput\n\t\treturn ret\n\t}\n\treturn o.SelectionCriteriaInputs\n}", "func (o *QueryChangesParams) WithDefaults() *QueryChangesParams {\n\to.SetDefaults()\n\treturn o\n}", "func NewMakeSnapshotArgs(layerID LayerID) *MakeSnapshotArgs {\n\targs := new(MakeSnapshotArgs)\n\targs.LayerID = layerID\n\treturn args\n}", "func (m Manager) GetChanges(context.Context, *pb.GetChangesRequest) (*pb.GetChangesResponse, error) {\n\tchanges, err := m.repo.GetAll()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &pb.GetChangesResponse{Changes: changes}, nil\n}", "func (l *LabelEvent) GetChanges() *EditChange {\n\tif l == nil {\n\t\treturn nil\n\t}\n\treturn l.Changes\n}" ]
[ "0.7866671", "0.78459364", "0.7305147", "0.65556616", "0.59407496", "0.5868117", "0.5521566", "0.50582105", "0.48954266", "0.4673309", "0.44319955", "0.43159604", "0.4264343", "0.42311433", "0.40826327", "0.40715033", "0.4071279", "0.40607655", "0.39594117", "0.38937762", "0.3864576", "0.3842568", "0.38322037", "0.38188782", "0.3813321", "0.3774964", "0.37609944", "0.37572223", "0.37567598", "0.3752687", "0.37335092", "0.3704852", "0.36798385", "0.36753094", "0.3674576", "0.36708575", "0.36669868", "0.366307", "0.36578894", "0.36366335", "0.3626938", "0.36021295", "0.35986683", "0.35965535", "0.35858685", "0.35779923", "0.3573573", "0.3568671", "0.35369998", "0.35282803", "0.35276163", "0.35245803", "0.35160434", "0.3500678", "0.3498514", "0.34974495", "0.34967914", "0.34946638", "0.3493289", "0.34830326", "0.3457799", "0.34573483", "0.34558386", "0.3453062", "0.3450547", "0.34477168", "0.34440368", "0.34405082", "0.34389812", "0.34374878", "0.3437466", "0.3436546", "0.34287992", "0.34238464", "0.34231818", "0.34225973", "0.3419285", "0.3417254", "0.34152198", "0.34142503", "0.3410739", "0.3409042", "0.34075728", "0.3401796", "0.33976027", "0.3390453", "0.3389154", "0.33848533", "0.33821052", "0.337943", "0.33792663", "0.33792663", "0.3376983", "0.3372704", "0.33662507", "0.33657944", "0.3365439", "0.33646902", "0.33643103", "0.33628964" ]
0.83000755
0
NewCopyRecipeToMyRecipesWithChangesParamsWithHTTPClient creates a new CopyRecipeToMyRecipesWithChangesParams object with the default values initialized, and the ability to set a custom HTTPClient for a request
func NewCopyRecipeToMyRecipesWithChangesParamsWithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams { var () return &CopyRecipeToMyRecipesWithChangesParams{ HTTPClient: client, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *CopyRecipeToMyRecipesWithChangesParams) WithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetPublicsRecipeParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *QueryChangesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ChatNewParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *RevertProductSnapshotRequestUsingPOSTParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostReconciliationParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateGitWebhookUsingPOSTParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetIngredientVersionRevisionParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *BudgetAddParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetReceiptsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ConfigurationBackupModifyParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostPartsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func NewPostReconciliationParamsWithHTTPClient(client *http.Client) *PostReconciliationParams {\n\tvar ()\n\treturn &PostReconciliationParams{\n\t\tHTTPClient: client,\n\t}\n}", "func (o *QueryChangesParams) WithHTTPClient(client *http.Client) *QueryChangesParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (o *AddBranchParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *AddItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *NarrowSearchRecipeParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ApplicationComponentSnapshotCollectionGetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetRequestDetailsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateCrossConnectParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateWidgetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *EditParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetWorkItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PetCreateParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ActionDeploymentRequestUsingPOST2Params) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *AddRepositoryParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostReconciliationParams) WithHTTPClient(client *http.Client) *PostReconciliationParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func WithHTTPClient(httpClient *http.Client) ClientOption {\n\treturn func(c *client) error {\n\t\tif httpClient == nil {\n\t\t\treturn errors.InvalidParameterError{Parameter: \"httpClient\", Reason: \"cannot be empty\"}\n\t\t}\n\n\t\tc.requester.Client = httpClient\n\t\treturn nil\n\t}\n}", "func (o *CreateCartUsingPOSTParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostAPIV3MachinesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreatePackageRepositoryDeltaUploadParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateVolumeBackupParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreatePolicyResetItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PutParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostMeArticlesDraftsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetRedbeamsFlowLogsProgressByResourceCrnParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ContainerUpdateParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ConfigurationBackupGetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *FileInfoCreateParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *InventoryStocktakingSearchParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetBackupLocationsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateRunbookParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *RepoGetAllCommitsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetRecentFoodsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PcloudPvminstancesSnapshotsRestorePostParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateInstantPaymentParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *LogRequestDownloadParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateRunbookRunCreateParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *IntegrationsManualHTTPSCreateParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PatchReferenceEntityRecordsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *RewardCommentsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateStockReceiptParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func NewConvertParamsWithHTTPClient(client *http.Client) *ConvertParams {\n\tvar ()\n\treturn &ConvertParams{\n\t\tHTTPClient: client,\n\t}\n}", "func (o *EstimateCoinBuyParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ImagePushParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *AddOrUpdateNodePoolConfigItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func NewAddItemParamsWithHTTPClient(client *http.Client) *AddItemParams {\n\tvar ()\n\treturn &AddItemParams{\n\t\tHTTPClient: client,\n\t}\n}", "func (o *GetPaymentsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func NewHTTPClient(tc *trace.Client, orig *http.Client) *HTTPClient {\n\tif orig == nil {\n\t\torig = http.DefaultClient\n\t}\n\trt := orig.Transport\n\tif rt == nil {\n\t\trt = http.DefaultTransport\n\t}\n\tclient := http.Client{\n\t\tTransport: &tracerTransport{base: rt},\n\t\tCheckRedirect: orig.CheckRedirect,\n\t\tJar: orig.Jar,\n\t\tTimeout: orig.Timeout,\n\t}\n\treturn &HTTPClient{\n\t\tClient: client,\n\t\ttc: tc,\n\t}\n}", "func (o *PostMenuItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ConvertParams) WithHTTPClient(client *http.Client) *ConvertParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (o *CreateLifecycleParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *BackupsCreateStatusParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *WaitListParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CapacityPoolGetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PollersPostParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateCardPaymentSourceParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func NewBarParamsWithHTTPClient(client *http.Client) *BarParams {\n\tvar ()\n\treturn &BarParams{\n\t\tHTTPClient: client,\n\t}\n}", "func (o *SharedCatalogSharedCatalogRepositoryV1SavePostParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CustomerGatewayUpdateOwnershipParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetDistroXOperationProgressByResourceCrnParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *LedgerVoucherPutParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateListParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *BucketsCollectionGetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func NewHTTPClient(options ...Opt) *HTTP {\n\tc := &HTTP{\n\t\tHTTPClient: &http.Client{},\n\t}\n\n\tfor _, option := range options {\n\t\toption(c)\n\t}\n\n\tif c.latestManifestURLFmt == \"\" {\n\t\tc.latestManifestURLFmt = defaultLatestManifestURLFmt\n\t}\n\n\tif c.manifestURLFmt == \"\" {\n\t\tc.manifestURLFmt = defaultManifestURLFmt\n\t}\n\n\treturn c\n}", "func (o *GetCreationTasksParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateBuildPropertiesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PayAllInvoicesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PayAllInvoicesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetContentSourceUsingGETParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *FreezeParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *NearestUsingGET1Params) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostDocumentMergeParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CloudTargetCreateParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GiftMessageCartRepositoryV1SavePostParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ConvertParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetZippedParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CheckTransactionCreditLimitParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateCredentialParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostAPIV10PeerReviewsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostV1ChangesChangeIDIdentitiesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateRepoNotificationParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CompanyCreditCreditHistoryManagementV1UpdatePutParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetGCParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *SavePreferencesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetV1IntegrationsAwsCloudtrailBatchesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ChatNewParams) WithHTTPClient(client *http.Client) *ChatNewParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (o *PatchAddonParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}" ]
[ "0.7614183", "0.7191112", "0.60642815", "0.6035904", "0.6035271", "0.601885", "0.596128", "0.5958554", "0.5913607", "0.5900584", "0.58948666", "0.5868178", "0.5853816", "0.5807214", "0.5771283", "0.5759591", "0.5750537", "0.57373995", "0.5735678", "0.57353026", "0.5731023", "0.5728886", "0.5717779", "0.571097", "0.5705245", "0.5704444", "0.57024497", "0.57021934", "0.5697534", "0.56789833", "0.5672361", "0.5668083", "0.56656694", "0.56633604", "0.56613314", "0.5660155", "0.5655626", "0.56448334", "0.56381404", "0.5637204", "0.5633252", "0.56267506", "0.56253225", "0.5620838", "0.56169116", "0.5612991", "0.56084293", "0.5603547", "0.56028503", "0.55996263", "0.5599019", "0.5594215", "0.5593145", "0.559071", "0.55899197", "0.55895495", "0.5574314", "0.55686575", "0.556664", "0.5557696", "0.55560964", "0.55516654", "0.5550406", "0.55494535", "0.5547117", "0.5546348", "0.55420315", "0.5541407", "0.55401", "0.5532894", "0.5531079", "0.5526014", "0.5524486", "0.5524455", "0.552166", "0.5516429", "0.55163825", "0.5516117", "0.5515396", "0.55149424", "0.55149424", "0.5514349", "0.5513633", "0.5509976", "0.550896", "0.5507775", "0.5507527", "0.55066335", "0.55062", "0.55042285", "0.5503505", "0.5503007", "0.5501636", "0.55005294", "0.5499335", "0.5496578", "0.5489046", "0.54874665", "0.54872245", "0.5487055" ]
0.746175
1
WithTimeout adds the timeout to the copy recipe to my recipes with changes params
func (o *CopyRecipeToMyRecipesWithChangesParams) WithTimeout(timeout time.Duration) *CopyRecipeToMyRecipesWithChangesParams { o.SetTimeout(timeout) return o }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *CopyRecipeToMyRecipesWithChangesParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetPublicsRecipeParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewCopyRecipeToMyRecipesWithChangesParamsWithTimeout(timeout time.Duration) *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (b *taskBuilder) timeout(timeout time.Duration) {\n\tb.Spec.ExecutionTimeout = timeout\n\tb.Spec.IoTimeout = timeout // With kitchen, step logs don't count toward IoTimeout.\n}", "func (o *GetIngredientVersionRevisionParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *NarrowSearchRecipeParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *NarrowSearchRecipeParams) WithTimeout(timeout time.Duration) *NarrowSearchRecipeParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func WithTimeout(duration time.Duration) ReconcilerOption {\n\treturn func(r *Reconciler) {\n\t\tr.timeout = duration\n\t}\n}", "func WithTimeout(t time.Duration) OptFunc {\n\treturn func(d *Downloader) {\n\t\td.timeout = t\n\t}\n}", "func (o *GetPublicsRecipeParams) WithTimeout(timeout time.Duration) *GetPublicsRecipeParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func WithTimeout(timeout time.Duration) Option {\n\treturn func(opts *Opts) error {\n\t\topts.Timeout = timeout\n\t\treturn nil\n\t}\n}", "func (o *GetNutritionForSingleParsedPlainTextIngredientParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func WithTimeout(timeout time.Duration) BuilderOptionFunc {\n\treturn func(b *Builder) error {\n\t\tb.timeout = timeout\n\t\treturn nil\n\t}\n}", "func (o *GetReceiptsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func WithTimeout(timeout time.Duration) Option {\n\treturn func(o *options) {\n\t\to.timeout = timeout\n\t}\n}", "func WithTimeout(timeout time.Duration) Option {\n\treturn func(o *options) {\n\t\to.timeout = timeout\n\t}\n}", "func (o *AddRepositoryParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *AddItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (pool *ComplexPool) SetTimeout(timeout time.Duration) {\n\tlogger.Debugf(\"prox (%p): setting timeout: %v\", pool, timeout)\n\tpool.timeout = timeout\n}", "func WithTimeout(timeout time.Duration) Option {\n\treturn func(opts *Options) {\n\t\topts.Timeout = timeout\n\t}\n}", "func NewNarrowSearchRecipeParamsWithTimeout(timeout time.Duration) *NarrowSearchRecipeParams {\n\tvar (\n\t\tsortbyDefault = string(\"name\")\n\t\tsortdirDefault = string(\"desc\")\n\t)\n\treturn &NarrowSearchRecipeParams{\n\t\tSortby: &sortbyDefault,\n\t\tSortdir: &sortdirDefault,\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *CreateGitWebhookUsingPOSTParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func WithTimeout(duration time.Duration) Option {\n\treturn wrappedOption{otlpconfig.WithTimeout(duration)}\n}", "func WithTimeout(timeout time.Duration) ClientOption {\n\treturn withTimeout{timeout}\n}", "func (o *RevertProductSnapshotRequestUsingPOSTParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *BudgetAddParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateRunbookRunCreateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *AddOrUpdateNodePoolConfigItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func WithTimeout(duration time.Duration) Option {\n\treturn wrappedOption{oconf.WithTimeout(duration)}\n}", "func cloneWithTimeout(httpClient *http.Client, t time.Duration) (*http.Client, error) {\n\tif httpClient == nil {\n\t\treturn nil, fmt.Errorf(\"nil HTTP client\")\n\t} else if httpClient.Transport == nil {\n\t\treturn nil, fmt.Errorf(\"nil HTTP client transport\")\n\t}\n\n\tif t.Nanoseconds() < 0 {\n\t\treturn httpClient, nil\n\t}\n\n\ttr, ok := httpClient.Transport.(*http.Transport)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"unexpected HTTP transport: %T\", httpClient.Transport)\n\t}\n\n\t// copy all public fields, to avoid copying transient state and locks\n\tntr := &http.Transport{\n\t\tProxy: tr.Proxy,\n\t\tDialContext: tr.DialContext,\n\t\tDial: tr.Dial,\n\t\tDialTLS: tr.DialTLS,\n\t\tTLSClientConfig: tr.TLSClientConfig,\n\t\tTLSHandshakeTimeout: tr.TLSHandshakeTimeout,\n\t\tDisableKeepAlives: tr.DisableKeepAlives,\n\t\tDisableCompression: tr.DisableCompression,\n\t\tMaxIdleConns: tr.MaxIdleConns,\n\t\tMaxIdleConnsPerHost: tr.MaxIdleConnsPerHost,\n\t\tMaxConnsPerHost: tr.MaxConnsPerHost,\n\t\tIdleConnTimeout: tr.IdleConnTimeout,\n\t\tResponseHeaderTimeout: tr.ResponseHeaderTimeout,\n\t\tExpectContinueTimeout: tr.ExpectContinueTimeout,\n\t\tTLSNextProto: tr.TLSNextProto,\n\t\tProxyConnectHeader: tr.ProxyConnectHeader,\n\t\tMaxResponseHeaderBytes: tr.MaxResponseHeaderBytes,\n\t}\n\n\t// apply timeout\n\tntr.DialContext = (&net.Dialer{\n\t\tTimeout: t,\n\t\tKeepAlive: 30 * time.Second,\n\t}).DialContext\n\n\t// clone http client with new transport\n\tnc := *httpClient\n\tnc.Transport = ntr\n\treturn &nc, nil\n}", "func WithTimeout(t time.Duration) Option {\n\treturn func(c *Client) { c.httpClient.Timeout = t }\n}", "func (o *CreateCartUsingPOSTParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *FileInfoCreateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func WithTimeout(timeout time.Duration) Option {\n\treturn func(c *Client) {\n\t\tc.client.Timeout = timeout\n\t}\n}", "func WithTimeout(timeout time.Duration) Option {\n\treturn func(c *Client) {\n\t\tc.client.Timeout = timeout\n\t}\n}", "func WithTimeout(timeout time.Duration) ClientOption {\n\treturn optionFunc(func(c *Client) {\n\t\tc.WithTimeout(timeout)\n\t})\n}", "func (o *EditParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func WithTimeout(timeout time.Duration) Opt {\n\treturn func(c *Client) error {\n\t\tc.client.Timeout = timeout\n\t\treturn nil\n\t}\n}", "func (o *PutMenuItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *EditParams) WithTimeout(timeout time.Duration) *EditParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *GetActionTemplateLogoVersionParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (c *OrganizationsEnvironmentsApisRevisionsDebugsessionsCreateCall) Timeout(timeout int64) *OrganizationsEnvironmentsApisRevisionsDebugsessionsCreateCall {\n\tc.urlParams_.Set(\"timeout\", fmt.Sprint(timeout))\n\treturn c\n}", "func (c *Connection) CopyToRemoteWithRetry(hostname, path string, sleep, timeout time.Duration) error {\n\tctx, cancel := context.WithTimeout(context.Background(), timeout)\n\tdefer cancel()\n\tch := make(chan error)\n\tvar mostRecentCopyToRemoteWithRetry error\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn\n\t\t\tdefault:\n\t\t\t\tch <- c.CopyToRemote(hostname, path)\n\t\t\t\ttime.Sleep(sleep)\n\t\t\t}\n\t\t}\n\t}()\n\tfor {\n\t\tselect {\n\t\tcase result := <-ch:\n\t\t\tmostRecentCopyToRemoteWithRetry = result\n\t\t\tif mostRecentCopyToRemoteWithRetry == nil {\n\t\t\t\treturn nil\n\t\t\t}\n\t\tcase <-ctx.Done():\n\t\t\treturn errors.Errorf(\"CopyToRemoteWithRetry timed out: %s\\n\", mostRecentCopyToRemoteWithRetry)\n\t\t}\n\t}\n}", "func (o *AddItemParams) WithTimeout(timeout time.Duration) *AddItemParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *PetCreateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *RegenerateDeployKeyParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetPrivateToggleDepositAddressCreationParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func WithTimeout(ctx context.Context, time time.Duration) (ret context.Context) {\n\tret = context.WithValue(ctx, liverpc.KeyTimeout, time)\n\treturn\n}", "func (o *AddBranchParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func Timeout(t time.Duration) ClientOpt {\n\treturn func(c *Client) {\n\t\tc.HTTPClient.Timeout = t\n\t}\n}", "func (c *Closer) AddTimeout(close func(ctx context.Context) error, timeout time.Duration) {\n\tc.closers = append(c.closers, &timeoutCloser{close: close, timeout: timeout})\n}", "func (o *GetRepository15Params) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func AddTimeout() {}", "func (x Go) Timeout(timeout time.Duration) Go {\n\tx.timeout = timeout\n\treturn x\n}", "func WithTimeout(timeout time.Duration) configF {\n\treturn func(c *config) *config {\n\t\tc.defaultTimeout = timeout\n\t\treturn c\n\t}\n}", "func (o *GetZippedParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetBuildPropertiesParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetDistroXOperationProgressByResourceCrnParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func WithTimeout(timeout time.Duration) Option {\n\treturn func(opts *VDRI) {\n\t\topts.client.Timeout = timeout\n\t}\n}", "func (o *CreatePolicyResetItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func WithTimeout(timeout int) CreateDBOpFn {\n\treturn func(op *CreateDBOp) {\n\t\tif 0 == timeout {\n\t\t\treturn\n\t\t}\n\t\top.timeout = &timeout\n\t\top.set = true\n\t}\n}", "func (ini *Init) AddWithTimeout(timeout time.Duration, f func(ctx context.Context) (any, error)) *Init {\n\treturn ini.Add(func(ctx context.Context) (any, error) {\n\t\treturn ini.withTimeout(ctx, timeout, f)\n\t})\n}", "func (o *CreateBlueprintInWorkspaceInternalParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func Timeout(timeout int64) Option {\n\treturn func(opts *options) {\n\t\topts.timeout = time.Duration(timeout) * time.Second\n\t}\n}", "func WithTimeout(t time.Duration) apiOption {\n\treturn func(m *Management) {\n\t\tm.timeout = t\n\t}\n}", "func WithPoolTimeout(timeout time.Duration) OptsFunc {\n\treturn func(o *redis.Options) {\n\t\to.PoolTimeout = timeout\n\t}\n}", "func (o *PatchAddonParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateWidgetParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func Timeout(timeout time.Duration) OptionFunc {\n\treturn func(tc *TracedClient) error {\n\t\tif timeout <= 0 {\n\t\t\treturn errors.New(\"timeout must be positive\")\n\t\t}\n\t\ttc.cl.Timeout = timeout\n\t\treturn nil\n\t}\n}", "func (o *FreezeParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PostMenuItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (m *monitor) withTimeout(timeout time.Duration) *monitor {\n\tm.timeout = timeout\n\treturn m\n}", "func (o *CreateCrossConnectParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CloudNFSExportAddParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ConfigurationBackupModifyParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PatchLibrariesByIDContentByIDParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *UpdateRunbookParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PostReconciliationParams) WithTimeout(timeout time.Duration) *PostReconciliationParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *UpdateBuildPropertiesParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ListDCForSeedParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetInterceptionitemsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PostReconciliationParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetProductsByIDVariationAttributesByIDParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetExampleNewProjectDescriptionCompatibilityVersion1Params) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func Timeout(t time.Duration) Option {\n\treturn func(c *Config) Option {\n\t\tprevious := c.Timeout\n\t\tc.Timeout = t\n\t\treturn Timeout(previous)\n\t}\n}", "func (o *GetRestoreDatalakeStatusParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *InventoryStocktakingSearchParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (r *Search) Timeout(timeout string) *Search {\n\n\tr.req.Timeout = &timeout\n\n\treturn r\n}", "func (f Reindex) WithTimeout(v time.Duration) func(*ReindexRequest) {\n\treturn func(r *ReindexRequest) {\n\t\tr.Timeout = v\n\t}\n}", "func (o *PcloudPvminstancesSnapshotsRestorePostParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ActionDeploymentRequestUsingPOST2Params) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetClusterTemplateByNameInWorkspaceParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetItemByAppIDParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateCardPaymentSourceParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetLolCatalogV1ItemsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CustomerGatewayUpdateOwnershipParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetContentSourceUsingGETParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (c *CentralCacheTestImpl) SetWithTimeout(item Item, serialize bool, compress bool, ttl int32) error {\n\treturn nil\n}", "func Timeout(timeout time.Duration) Option {\n\treturn func(client *http.Client) {\n\t\tclient.Timeout = timeout\n\t}\n}", "func Timeout(timeout time.Duration) Option {\n\treturn func(client *http.Client) {\n\t\tclient.Timeout = timeout\n\t}\n}" ]
[ "0.6520962", "0.5858076", "0.57581943", "0.5501371", "0.52934045", "0.5210673", "0.52082425", "0.5182034", "0.5154618", "0.508878", "0.50811505", "0.5077336", "0.50460726", "0.50446784", "0.5041395", "0.5041395", "0.5026333", "0.50209403", "0.49826825", "0.49667513", "0.49350506", "0.49000707", "0.48991832", "0.4899181", "0.48789322", "0.4878208", "0.48661205", "0.4862972", "0.4859411", "0.48487413", "0.4819714", "0.48195753", "0.48054454", "0.47872126", "0.47872126", "0.47812507", "0.4779445", "0.47775674", "0.476841", "0.47561425", "0.47559643", "0.47523353", "0.47439048", "0.4737756", "0.4732429", "0.47218335", "0.47116056", "0.47065336", "0.46980414", "0.46864238", "0.46831948", "0.46825367", "0.46799856", "0.46797565", "0.4672737", "0.46667558", "0.46653873", "0.46607205", "0.46479845", "0.46294", "0.4628776", "0.46286863", "0.462782", "0.46273237", "0.46265063", "0.462428", "0.46225348", "0.46161023", "0.4609931", "0.46094197", "0.46013364", "0.4598539", "0.45956787", "0.45955542", "0.45949632", "0.4591241", "0.4590414", "0.45893514", "0.4586748", "0.45867077", "0.4583725", "0.4581197", "0.45792067", "0.45781863", "0.4574672", "0.45735958", "0.457203", "0.457064", "0.45685962", "0.45676365", "0.45647547", "0.45629534", "0.4559099", "0.4558703", "0.45582998", "0.4557191", "0.45559624", "0.4547357", "0.45460075", "0.45460075" ]
0.7105756
0
SetTimeout adds the timeout to the copy recipe to my recipes with changes params
func (o *CopyRecipeToMyRecipesWithChangesParams) SetTimeout(timeout time.Duration) { o.timeout = timeout }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *CopyRecipeToMyRecipesWithChangesParams) WithTimeout(timeout time.Duration) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *GetPublicsRecipeParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetIngredientVersionRevisionParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *NarrowSearchRecipeParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (pool *ComplexPool) SetTimeout(timeout time.Duration) {\n\tlogger.Debugf(\"prox (%p): setting timeout: %v\", pool, timeout)\n\tpool.timeout = timeout\n}", "func (o *GetReceiptsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetNutritionForSingleParsedPlainTextIngredientParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *EditParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateGitWebhookUsingPOSTParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetActionTemplateLogoVersionParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (b *taskBuilder) timeout(timeout time.Duration) {\n\tb.Spec.ExecutionTimeout = timeout\n\tb.Spec.IoTimeout = timeout // With kitchen, step logs don't count toward IoTimeout.\n}", "func (o *AddOrUpdateNodePoolConfigItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *AddItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *AddRepositoryParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *RevertProductSnapshotRequestUsingPOSTParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateCartUsingPOSTParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetWorkflowBuildTaskMetaMoidParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetBuildPropertiesParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *FileInfoCreateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *FreezeParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetRestoreDatalakeStatusParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *UpdateNetworkCellularGatewaySettingsSubnetPoolParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (p *WorkPool) SetTimeout(timeout time.Duration) { // 设置超时时间\n\tp.timeout = timeout\n}", "func (o *UpdateBuildPropertiesParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func SetTimeout(dur time.Duration) { note.Timeout = dur }", "func (o *PutMenuItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ConfigurationBackupModifyParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateWidgetParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *UpdateRunbookParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateRunbookRunCreateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetZippedParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PatchAddonParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetDistroXOperationProgressByResourceCrnParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateScriptParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PostMenuItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *RegenerateDeployKeyParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetPrivateToggleDepositAddressCreationParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreatePolicyResetItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *VectorThumbnailParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *SetUniverseBackupFlagParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetRedbeamsFlowLogsProgressByResourceCrnParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ToggleNetworkGeneratorsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ActionDeploymentRequestUsingPOST2Params) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CustomerGatewayUpdateOwnershipParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetExampleNewProjectDescriptionCompatibilityVersion1Params) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PcloudPvminstancesSnapshotsRestorePostParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateLifecycleParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateBlueprintInWorkspaceInternalParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PostReconciliationParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetRepository15Params) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetClusterTemplateByNameInWorkspaceParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreatePackageRepositoryDeltaUploadParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetProductsByIDVariationAttributesByIDParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *ListDCForSeedParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetInterceptionitemsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetLolCatalogV1ItemsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateVolumeBackupParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PostPartsParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *UpdateDmrClusterLinkParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *SkuPackPostParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateCrossConnectParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *AddBranchParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *InventoryStocktakingSearchParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetIconParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetWorkItemParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PostAPIV3MachinesParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetNetworkAppliancePortParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *BudgetAddParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func SetTimeout(timeout time.Duration) {\n\tclient.SetTimeout(timeout)\n}", "func (o *GetDatalakeDbConfigParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetContentSourceUsingGETParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *DevicesGetModuleComponentCommandHistoryParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CloudTargetCreateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func NewCopyRecipeToMyRecipesWithChangesParamsWithTimeout(timeout time.Duration) *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *GetaspecificPbxDeviceFirmwareBinaryParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetRemotesupportConnectemcParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *UpdateStockReceiptParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PetCreateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (m *RedirectPostRequestBody) SetTimeout(value *int32)() {\n m.timeout = value\n}", "func (o *AddVMParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *UpdateSubnetParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PutFlagSettingParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *SharedCatalogSharedCatalogRepositoryV1SavePostParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *UploadWorkflowTemplateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetBootstrapParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *RebuildIndexSetParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *UploadDeployFileParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetPrivateOrderstateParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *DeleteNetworksNetworkIDTiersTierIDImagesImageNameParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PatchSepainstantIDParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *PatchAssetDeviceConfigurationsMoidParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *UpdateAddonParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CreateCardPaymentSourceParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (xmlmc *XmlmcInstStruct) SetTimeout(timeout int) {\n\txmlmc.timeout = timeout\n}", "func (o *CatalogProductTierPriceManagementV1AddPostParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *UpdateTransactionCategorizationRuleParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *BackupsCreateStatusParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *GetBundleByKeyParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *CloudNFSExportAddParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}", "func (o *VirtualizationChoicesReadParams) SetTimeout(timeout time.Duration) {\n\to.timeout = timeout\n}" ]
[ "0.6950563", "0.6872008", "0.63673514", "0.63112706", "0.62673944", "0.6094319", "0.6092785", "0.6001111", "0.59669495", "0.59356844", "0.5930924", "0.59246594", "0.5920255", "0.58911455", "0.58869755", "0.58824813", "0.5843332", "0.58411634", "0.5800582", "0.5796857", "0.57948065", "0.57835805", "0.57832783", "0.5781934", "0.5781316", "0.57804054", "0.5777273", "0.57694006", "0.57664967", "0.5754906", "0.57374287", "0.57355785", "0.5734713", "0.5725351", "0.572487", "0.5718388", "0.5711212", "0.5707071", "0.5701812", "0.5701267", "0.5693274", "0.56899804", "0.56861067", "0.56817704", "0.56743205", "0.5674021", "0.5669791", "0.5663938", "0.56621635", "0.56573623", "0.56402415", "0.5637113", "0.5635821", "0.5630474", "0.5629576", "0.5623465", "0.56203043", "0.5619131", "0.56182027", "0.56177384", "0.561342", "0.5612413", "0.5611636", "0.5595247", "0.5594694", "0.5593178", "0.55898327", "0.5587626", "0.5586399", "0.5581521", "0.5581301", "0.55789953", "0.5576991", "0.5576102", "0.5575919", "0.5568991", "0.55657613", "0.5562352", "0.55609363", "0.5549682", "0.5545799", "0.5545378", "0.55443364", "0.5543263", "0.5542479", "0.55385107", "0.55372375", "0.5535468", "0.5535421", "0.55328566", "0.5531095", "0.55301094", "0.5523148", "0.55207884", "0.5513605", "0.55127084", "0.5511358", "0.55055845", "0.5505483", "0.55045646" ]
0.74348015
0
WithContext adds the context to the copy recipe to my recipes with changes params
func (o *CopyRecipeToMyRecipesWithChangesParams) WithContext(ctx context.Context) *CopyRecipeToMyRecipesWithChangesParams { o.SetContext(ctx) return o }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func CopyWithContext(ctx context.Context, dst *Writer, src Stream) error {\n\tif err := src.Open(); err != nil {\n\t\treturn err\n\t}\n\tvar err error\n\tfor ctx.Err() == nil {\n\t\tvar pair Pair\n\t\tpair, err = src.Read()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif pair.Key == nil {\n\t\t\tbreak\n\t\t}\n\t\terr = dst.Write(pair)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn ctx.Err()\n}", "func (ctx *Context) Copy() *Context {\n\tvar pathParams Params\n\tif len(ctx.PathParams) > 0 {\n\t\tpathParams = append(pathParams, ctx.PathParams...)\n\t}\n\treturn &Context{\n\t\tresponseWriter2: nil,\n\t\tResponseWriter: nil,\n\t\tRequest: ctx.Request,\n\t\tPathParams: pathParams,\n\t\tqueryParams: ctx.queryParams,\n\t\tValidator: ctx.Validator,\n\t\tfetchClientIPFromHeader: ctx.fetchClientIPFromHeader,\n\t\thandlers: nil,\n\t\thandlerIndex: __abortHandlerIndex,\n\t\tkvs: ctx.kvs,\n\t}\n}", "func (c *Context) Copy() *Context {\n\tret := *c\n\tret.init(&fasthttp.RequestCtx{})\n\tc.Request.CopyTo(&ret.Request)\n\tc.Response.CopyTo(&ret.Response)\n\tret.WSConn = c.WSConn\n\tret.data = c.data\n\treturn &ret\n}", "func (mr *MockRDSAPIMockRecorder) CopyDBSnapshotWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CopyDBSnapshotWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).CopyDBSnapshotWithContext), varargs...)\n}", "func (evalCtx *extendedEvalContext) copy() *extendedEvalContext {\n\tcpy := *evalCtx\n\tcpy.EvalContext = *evalCtx.EvalContext.Copy()\n\treturn &cpy\n}", "func (ctx context) clone() context {\n\treturn context{\n\t\tkeyvals: safeSlice(ctx.keyvals),\n\t}\n}", "func (tx *WriteTx) RunWithContext(ctx context.Context) error {\n\tif tx.err != nil {\n\t\treturn tx.err\n\t}\n\tinput, err := tx.input()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = retry(ctx, func() error {\n\t\tout, err := tx.db.client.TransactWriteItemsWithContext(ctx, input)\n\t\tif tx.cc != nil && out != nil {\n\t\t\tfor _, cc := range out.ConsumedCapacity {\n\t\t\t\taddConsumedCapacity(tx.cc, cc)\n\t\t\t}\n\t\t}\n\t\treturn err\n\t})\n\treturn err\n}", "func (c *Client) CopyWithContext(ctx context.Context, src FileID, dst string) (FileID, error) {\n\tfileID, err := c.iclient.Copy(ctx, &pb.CopyRequest{SrcId: src[:], Dst: dst})\n\tif e, ok := err.(twirp.Error); ok && e.Code() == twirp.NotFound {\n\t\treturn FileID{}, ErrNotFound\n\t}\n\tif err != nil {\n\t\treturn FileID{}, err\n\t}\n\ts, err := UnmarshalFileID(fileID.Sum)\n\tif err != nil {\n\t\treturn FileID{}, err\n\t}\n\treturn s, nil\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (this *L0JobContext) Copy(request string) JobContext {\n\treturn &L0JobContext{\n\t\trequest: request,\n\t\tjobID: this.jobID,\n\t\tlogic: this.logic,\n\t\tloadBalancerLogic: this.loadBalancerLogic,\n\t\tserviceLogic: this.serviceLogic,\n\t\tenvironmentLogic: this.environmentLogic,\n\t\tmutex: this.mutex,\n\t}\n}", "func MetaWithContext(ctx context.Context, newMeta map[string]interface{}) context.Context {\n\tprevMeta := MetaFromContext(ctx)\n\n\tif prevMeta == nil {\n\t\tprevMeta = make(map[string]interface{})\n\t}\n\n\tfor k, v := range newMeta {\n\t\tprevMeta[k] = v\n\t}\n\n\treturn context.WithValue(ctx, MetaCtxKey, prevMeta)\n}", "func (m *MockRDSAPI) CopyOptionGroupWithContext(arg0 aws.Context, arg1 *rds.CopyOptionGroupInput, arg2 ...request.Option) (*rds.CopyOptionGroupOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"CopyOptionGroupWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.CopyOptionGroupOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (mr *MockRDSAPIMockRecorder) CopyDBClusterSnapshotWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CopyDBClusterSnapshotWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).CopyDBClusterSnapshotWithContext), varargs...)\n}", "func (m *MockRDSAPI) CopyDBParameterGroupWithContext(arg0 aws.Context, arg1 *rds.CopyDBParameterGroupInput, arg2 ...request.Option) (*rds.CopyDBParameterGroupOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"CopyDBParameterGroupWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.CopyDBParameterGroupOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (m *MockRDSAPI) CopyDBSnapshotWithContext(arg0 aws.Context, arg1 *rds.CopyDBSnapshotInput, arg2 ...request.Option) (*rds.CopyDBSnapshotOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"CopyDBSnapshotWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.CopyDBSnapshotOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (mr *MockRDSAPIMockRecorder) CopyOptionGroupWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CopyOptionGroupWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).CopyOptionGroupWithContext), varargs...)\n}", "func (m *MockRDSAPI) CopyDBClusterSnapshotWithContext(arg0 aws.Context, arg1 *rds.CopyDBClusterSnapshotInput, arg2 ...request.Option) (*rds.CopyDBClusterSnapshotOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"CopyDBClusterSnapshotWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.CopyDBClusterSnapshotOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (_obj *DataService) CreateApplyWithContext(tarsCtx context.Context, wx_id string, club_id string, _opt ...map[string]string) (ret int32, err error) {\n\n\tvar length int32\n\tvar have bool\n\tvar ty byte\n\t_os := codec.NewBuffer()\n\terr = _os.Write_string(wx_id, 1)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\terr = _os.Write_string(club_id, 2)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tvar _status map[string]string\n\tvar _context map[string]string\n\tif len(_opt) == 1 {\n\t\t_context = _opt[0]\n\t} else if len(_opt) == 2 {\n\t\t_context = _opt[0]\n\t\t_status = _opt[1]\n\t}\n\t_resp := new(requestf.ResponsePacket)\n\n\terr = _obj.s.Tars_invoke(tarsCtx, 0, \"createApply\", _os.ToBytes(), _status, _context, _resp)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\t_is := codec.NewReader(tools.Int8ToByte(_resp.SBuffer))\n\terr = _is.Read_int32(&ret, 0, true)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tif len(_opt) == 1 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t} else if len(_opt) == 2 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t\tfor k := range _status {\n\t\t\tdelete(_status, k)\n\t\t}\n\t\tfor k, v := range _resp.Status {\n\t\t\t_status[k] = v\n\t\t}\n\n\t}\n\t_ = length\n\t_ = have\n\t_ = ty\n\treturn ret, nil\n}", "func (_obj *Apilangpack) AddServantWithContext(imp _impApilangpackWithContext, obj string) {\n\ttars.AddServantWithContext(_obj, imp, obj)\n}", "func (m *MockRDSAPI) ModifyDBClusterWithContext(arg0 aws.Context, arg1 *rds.ModifyDBClusterInput, arg2 ...request.Option) (*rds.ModifyDBClusterOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"ModifyDBClusterWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.ModifyDBClusterOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (mr *MockS3APIMockRecorder) UploadPartCopyWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"UploadPartCopyWithContext\", reflect.TypeOf((*MockS3API)(nil).UploadPartCopyWithContext), varargs...)\n}", "func (obj *ConfigWriter) AddServantWithContext(imp impConfigWriterWithContext, objStr string) {\n\ttars.AddServantWithContext(obj, imp, objStr)\n}", "func (_obj *DataService) AddServantWithContext(imp _impDataServiceWithContext, obj string) {\n\ttars.AddServantWithContext(_obj, imp, obj)\n}", "func (mr *MockRDSAPIMockRecorder) CopyDBParameterGroupWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CopyDBParameterGroupWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).CopyDBParameterGroupWithContext), varargs...)\n}", "func (mr *MockRDSAPIMockRecorder) ModifyDBSnapshotWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ModifyDBSnapshotWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).ModifyDBSnapshotWithContext), varargs...)\n}", "func CopyContext(hglrcSrc unsafe.Pointer, hglrcDst unsafe.Pointer, mask unsafe.Pointer) unsafe.Pointer {\n\tret, _, _ := syscall.Syscall(gpCopyContext, 3, uintptr(hglrcSrc), uintptr(hglrcDst), uintptr(mask))\n\treturn (unsafe.Pointer)(ret)\n}", "func CopyContext(dest context.Context, src context.Context) (context.Context, error) {\n\ttrace := GetTraceFromContext(src)\n\tspan := GetSpanFromContext(src)\n\tif trace == nil || span == nil {\n\t\treturn dest, ErrTraceNotFoundInContext\n\t}\n\tdest = PutTraceInContext(dest, trace)\n\tdest = PutSpanInContext(dest, span)\n\treturn dest, nil\n}", "func (mr *MockRDSAPIMockRecorder) ModifyDBSnapshotAttributeWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ModifyDBSnapshotAttributeWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).ModifyDBSnapshotAttributeWithContext), varargs...)\n}", "func (_obj *LacService) AddServantWithContext(imp _impLacServiceWithContext, obj string) {\n\ttars.AddServantWithContext(_obj, imp, obj)\n}", "func (mr *MockRDSAPIMockRecorder) ModifyDBInstanceWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ModifyDBInstanceWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).ModifyDBInstanceWithContext), varargs...)\n}", "func With(ctx context.Context, kvs ...interface{}) context.Context {\n\tl := fromCtx(ctx)\n\tl = l.With(kvs...)\n\treturn toCtx(ctx, l)\n}", "func (m *MockRDSAPI) CopyDBClusterParameterGroupWithContext(arg0 aws.Context, arg1 *rds.CopyDBClusterParameterGroupInput, arg2 ...request.Option) (*rds.CopyDBClusterParameterGroupOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"CopyDBClusterParameterGroupWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.CopyDBClusterParameterGroupOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (m *MockS3API) UploadPartCopyWithContext(arg0 context.Context, arg1 *s3.UploadPartCopyInput, arg2 ...request.Option) (*s3.UploadPartCopyOutput, error) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"UploadPartCopyWithContext\", varargs...)\n\tret0, _ := ret[0].(*s3.UploadPartCopyOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (r *Request) WithContext(ctx context.Context) *Request", "func (c *Client) PatchWithContext(ctx context.Context, url string, reqBody, resType interface{}) error {\n\treturn c.CallAPIWithContext(ctx, \"PATCH\", url, reqBody, resType, true)\n}", "func (mr *MockRDSAPIMockRecorder) ModifyDBClusterSnapshotAttributeWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ModifyDBClusterSnapshotAttributeWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).ModifyDBClusterSnapshotAttributeWithContext), varargs...)\n}", "func (_obj *Apichannels) Channels_editCreatorWithContext(tarsCtx context.Context, params *TLchannels_editCreator, _opt ...map[string]string) (ret Updates, err error) {\n\n\tvar length int32\n\tvar have bool\n\tvar ty byte\n\t_os := codec.NewBuffer()\n\terr = params.WriteBlock(_os, 1)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tvar _status map[string]string\n\tvar _context map[string]string\n\tif len(_opt) == 1 {\n\t\t_context = _opt[0]\n\t} else if len(_opt) == 2 {\n\t\t_context = _opt[0]\n\t\t_status = _opt[1]\n\t}\n\t_resp := new(requestf.ResponsePacket)\n\n\terr = _obj.s.Tars_invoke(tarsCtx, 0, \"channels_editCreator\", _os.ToBytes(), _status, _context, _resp)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\t_is := codec.NewReader(tools.Int8ToByte(_resp.SBuffer))\n\terr = ret.ReadBlock(_is, 0, true)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tif len(_opt) == 1 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t} else if len(_opt) == 2 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t\tfor k := range _status {\n\t\t\tdelete(_status, k)\n\t\t}\n\t\tfor k, v := range _resp.Status {\n\t\t\t_status[k] = v\n\t\t}\n\n\t}\n\t_ = length\n\t_ = have\n\t_ = ty\n\treturn ret, nil\n}", "func sendWithContext(ctx context.Context, httpClient *http.Client, url string, body io.Reader, opt *Options) (*http.Response, error) {\n\tv, _ := query.Values(opt)\n\n\t// fmt.Print(v.Encode()) will output: \"city=0&mr=1&pb=4&pro=0&yys=0\"\n\tAPIEndpoint := fmt.Sprintf(\"%s&%s\", url, v.Encode())\n\tfmt.Println(APIEndpoint)\n\t// Change NewRequest to NewRequestWithContext and pass context it\n\treq, err := http.NewRequestWithContext(ctx, http.MethodGet, APIEndpoint, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// http.DefaultClient\n\tres, err := httpClient.Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn res, nil\n}", "func (h *KubernetesHelper) KubectlApplyWithContext(stdin string, context string, arg ...string) (string, error) {\n\targs := append([]string{\"apply\"}, arg...)\n\treturn h.KubectlWithContext(stdin, context, args...)\n}", "func (c *Client) PostWithContext(ctx context.Context, url string, reqBody, resType interface{}) error {\n\treturn c.CallAPIWithContext(ctx, \"POST\", url, reqBody, resType, true)\n}", "func copyComponentsPtr(s Servable, req *http.Request) {\n\tctx := context.WithValue(req.Context(), componentsKey, s.ServerField().Components)\n\t*req = *req.WithContext(ctx)\n}", "func (ctx context) Clone() logging.Context {\n\treturn &context{\n\t\tcorrID: ctx.corrID,\n\t\textra: ctx.extra,\n\t}\n}", "func (_obj *Apipayments) AddServantWithContext(imp _impApipaymentsWithContext, obj string) {\n\ttars.AddServantWithContext(_obj, imp, obj)\n}", "func AddServantWithContext(v dispatch, f interface{}, obj string) {\n\taddServantCommon(v, f, obj, true)\n}", "func (req *UpsertRequest) Context(ctx context.Context) *UpsertRequest {\n\treq.impl = req.impl.Context(ctx)\n\n\treturn req\n}", "func CrtlfWithContext(ctx context.Context, format string, args ...interface{}) {\n\tif hub := sentry.GetHubFromContext(ctx); hub != nil {\n\t\tcreticaldeps(hub.CaptureMessage, 3, format, args...)\n\t\treturn\n\t}\n\n\tcreticaldeps(sentry.CaptureMessage, 3, format, args...)\n}", "func (o *GetPublicsRecipeParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func RevisionContext(ns, svc, cfg, rev string) context.Context {\n\tkey := types.NamespacedName{Namespace: ns, Name: rev}\n\tif ctx, ok := contextCache.Get(key); ok {\n\t\treturn ctx.(context.Context)\n\t}\n\n\tctx := augmentWithRevision(context.Background(), ns, svc, cfg, rev)\n\tcontextCache.Add(key, ctx)\n\n\treturn ctx\n}", "func (req *UpsertObjectRequest) Context(ctx context.Context) *UpsertObjectRequest {\n\treq.impl = req.impl.Context(ctx)\n\n\treturn req\n}", "func (c *Context) Clone() *Context {\n\treturn &Context{\n\t\tStore: c.Store,\n\t\tSupervisor: c.Supervisor,\n\t\tProvisioner: c.Provisioner,\n\t\tLogger: c.Logger,\n\t}\n}", "func withContext(borrower ContextBorrower, worker Worker) Worker {\n\n\treturn func(t *T, _ Context) {\n\n\t\tif t.Failed() {\n\t\t\treturn\n\t\t}\n\n\t\tctx, release, err := borrower.Borrow()\n\t\tif err != nil {\n\t\t\tt.Errorf(\"%s\", err)\n\t\t\tt.FailNow()\n\t\t}\n\n\t\tdefer release()\n\t\tworkerRunner(nil, worker, t, ctx)\n\t}\n}", "func (o *RevertProductSnapshotRequestUsingPOSTParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (b *Builder) WithContext(context interface{}) *Builder {\n\tb.context = context\n\treturn b\n}", "func (mr *MockRDSAPIMockRecorder) ModifyDBClusterWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ModifyDBClusterWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).ModifyDBClusterWithContext), varargs...)\n}", "func (ctx *Ciphertext) Copy(ctxCopy BfvElement) error {\n\n\tif !checkContext([]BfvElement{ctx, ctxCopy}) {\n\t\treturn errors.New(\"input ciphertext are not using the same bfvcontext\")\n\t}\n\n\tfor i := range ctxCopy.Value() {\n\t\tctxCopy.Value()[i].Copy(ctx.Value()[i])\n\t}\n\tctxCopy.SetIsNTT(ctx.IsNTT())\n\n\treturn nil\n}", "func (d *Data) WithContext(ctx context.Context) *Data {\n\tcop := new(Data)\n\t*cop = *d\n\tcop.context = ctx\n\treturn cop\n}", "func (_obj *Apichannels) Channels_editPhotoWithContext(tarsCtx context.Context, params *TLchannels_editPhoto, _opt ...map[string]string) (ret Updates, err error) {\n\n\tvar length int32\n\tvar have bool\n\tvar ty byte\n\t_os := codec.NewBuffer()\n\terr = params.WriteBlock(_os, 1)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tvar _status map[string]string\n\tvar _context map[string]string\n\tif len(_opt) == 1 {\n\t\t_context = _opt[0]\n\t} else if len(_opt) == 2 {\n\t\t_context = _opt[0]\n\t\t_status = _opt[1]\n\t}\n\t_resp := new(requestf.ResponsePacket)\n\n\terr = _obj.s.Tars_invoke(tarsCtx, 0, \"channels_editPhoto\", _os.ToBytes(), _status, _context, _resp)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\t_is := codec.NewReader(tools.Int8ToByte(_resp.SBuffer))\n\terr = ret.ReadBlock(_is, 0, true)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tif len(_opt) == 1 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t} else if len(_opt) == 2 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t\tfor k := range _status {\n\t\t\tdelete(_status, k)\n\t\t}\n\t\tfor k, v := range _resp.Status {\n\t\t\t_status[k] = v\n\t\t}\n\n\t}\n\t_ = length\n\t_ = have\n\t_ = ty\n\treturn ret, nil\n}", "func (blk *Block) DrawWithContext(d Drawable, ctx DrawContext) error {\n\tblocks, _, err := d.GeneratePageBlocks(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(blocks) != 1 {\n\t\treturn errors.New(\"too many output blocks\")\n\t}\n\n\tfor _, newBlock := range blocks {\n\t\tif err := blk.mergeBlocks(newBlock); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func RequestWithContext(ctx context.Context, req *http.Request) *http.Request {\n\turl := req.URL\n\treq.URL = nil\n\treqCopy := req.WithContext(ctx)\n\treqCopy.URL = url\n\treq.URL = url\n\treturn reqCopy\n}", "func (mr *MockRDSAPIMockRecorder) CopyDBClusterParameterGroupWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CopyDBClusterParameterGroupWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).CopyDBClusterParameterGroupWithContext), varargs...)\n}", "func (c *VariantsetsPatchCall) Context(ctx context.Context) *VariantsetsPatchCall {\n\tc.ctx_ = ctx\n\treturn c\n}", "func With(ctx context.Context, app *App) context.Context {\n\treturn context.WithValue(ctx, appContextKey, app)\n}", "func (s *SizedWaitGroup) AddWithContext(ctx context.Context) error {\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\tcase s.current <- struct{}{}:\n\t\tbreak\n\t}\n\ts.wg.Add(1)\n\treturn nil\n}", "func (c *Context) Clone() *Context {\n\treturn &Context{\n\t\tLogger: c.Logger,\n\t}\n}", "func (obj *ShopSys) AddServantWithContext(imp impShopSysWithContext, objStr string) {\n\ttars.AddServantWithContext(obj, imp, objStr)\n}", "func (obj *ShopSys) AddServantWithContext(imp impShopSysWithContext, objStr string) {\n\ttars.AddServantWithContext(obj, imp, objStr)\n}", "func (mr *MockRDSAPIMockRecorder) ModifyDBSubnetGroupWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ModifyDBSubnetGroupWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).ModifyDBSubnetGroupWithContext), varargs...)\n}", "func (m *MockRDSAPI) ModifyDBSnapshotWithContext(arg0 aws.Context, arg1 *rds.ModifyDBSnapshotInput, arg2 ...request.Option) (*rds.ModifyDBSnapshotOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"ModifyDBSnapshotWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.ModifyDBSnapshotOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (c MethodsCollection) WithContext() pWithContext {\n\treturn pWithContext{\n\t\tMethod: c.MustGet(\"WithContext\"),\n\t}\n}", "func (m *MockRDSAPI) ModifyDBSnapshotAttributeWithContext(arg0 aws.Context, arg1 *rds.ModifyDBSnapshotAttributeInput, arg2 ...request.Option) (*rds.ModifyDBSnapshotAttributeOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"ModifyDBSnapshotAttributeWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.ModifyDBSnapshotAttributeOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (c *VariantsPatchCall) Context(ctx context.Context) *VariantsPatchCall {\n\tc.ctx_ = ctx\n\treturn c\n}", "func FromContext(c echo.Context) *newrelic.Transaction {\n\treturn newrelic.FromContext(c.Request().Context())\n}", "func prepContext(r *http.Request) *http.Request {\n\tskip := getSkipFile(r)\n\tres := r.WithContext(\n\t\tcontext.WithValue(r.Context(), toSkip, skip+1),\n\t)\n\n\treturn res\n}", "func (c Client) UpdateWithContext(context context.Context, input *UpdateRecordingInput) (*UpdateRecordingResponse, error) {\n\top := client.Operation{\n\t\tMethod: http.MethodPost,\n\t\tURI: \"/Trunks/{trunkSid}/Recording\",\n\t\tContentType: client.URLEncoded,\n\t\tPathParams: map[string]string{\n\t\t\t\"trunkSid\": c.trunkSid,\n\t\t},\n\t}\n\n\tif input == nil {\n\t\tinput = &UpdateRecordingInput{}\n\t}\n\n\tresponse := &UpdateRecordingResponse{}\n\tif err := c.client.Send(context, op, input, response); err != nil {\n\t\treturn nil, err\n\t}\n\treturn response, nil\n}", "func (c *SpanContext) CopyFrom(ctx *SpanContext) {\n\tc.traceID = ctx.traceID\n\tc.spanID = ctx.spanID\n\tc.parentID = ctx.parentID\n\tc.samplingState = ctx.samplingState\n\tif l := len(ctx.baggage); l > 0 {\n\t\tc.baggage = make(map[string]string, l)\n\t\tfor k, v := range ctx.baggage {\n\t\t\tc.baggage[k] = v\n\t\t}\n\t} else {\n\t\tc.baggage = nil\n\t}\n}", "func (obj *ShopSys) FakerBuyWithContext(ctx context.Context, input Faker, _opt ...map[string]string) (output Faker, err error) {\n\tvar inputMarshal []byte\n\tinputMarshal, err = proto.Marshal(&input)\n\tif err != nil {\n\t\treturn output, err\n\t}\n\n\tvar _status map[string]string\n\tvar _context map[string]string\n\tif len(_opt) == 1 {\n\t\t_context = _opt[0]\n\t} else if len(_opt) == 2 {\n\t\t_context = _opt[0]\n\t\t_status = _opt[1]\n\t}\n\n\tresp := new(requestf.ResponsePacket)\n\n\terr = obj.s.Tars_invoke(ctx, 0, \"FakerBuy\", inputMarshal, _status, _context, resp)\n\tif err != nil {\n\t\treturn output, err\n\t}\n\tif err = proto.Unmarshal(tools.Int8ToByte(resp.SBuffer), &output); err != nil {\n\t\treturn output, err\n\t}\n\n\tif len(_opt) == 1 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t} else if len(_opt) == 2 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t\tfor k := range _status {\n\t\t\tdelete(_status, k)\n\t\t}\n\t\tfor k, v := range resp.Status {\n\t\t\t_status[k] = v\n\t\t}\n\t}\n\n\treturn output, nil\n}", "func (obj *ShopSys) FakerBuyWithContext(ctx context.Context, input Faker, _opt ...map[string]string) (output Faker, err error) {\n\tvar inputMarshal []byte\n\tinputMarshal, err = proto.Marshal(&input)\n\tif err != nil {\n\t\treturn output, err\n\t}\n\n\tvar _status map[string]string\n\tvar _context map[string]string\n\tif len(_opt) == 1 {\n\t\t_context = _opt[0]\n\t} else if len(_opt) == 2 {\n\t\t_context = _opt[0]\n\t\t_status = _opt[1]\n\t}\n\n\tresp := new(requestf.ResponsePacket)\n\n\terr = obj.s.Tars_invoke(ctx, 0, \"FakerBuy\", inputMarshal, _status, _context, resp)\n\tif err != nil {\n\t\treturn output, err\n\t}\n\tif err = proto.Unmarshal(tools.Int8ToByte(resp.SBuffer), &output); err != nil {\n\t\treturn output, err\n\t}\n\n\tif len(_opt) == 1 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t} else if len(_opt) == 2 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t\tfor k := range _status {\n\t\t\tdelete(_status, k)\n\t\t}\n\t\tfor k, v := range resp.Status {\n\t\t\t_status[k] = v\n\t\t}\n\t}\n\n\treturn output, nil\n}", "func (c *Client) httpRequestWithContext(ctx context.Context, r *Request) (*Response, error) {\n\treq, err := http.NewRequestWithContext(ctx, r.Method, r.URL.RequestURI(), r.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tc.modifyLock.RLock()\n\ttoken := c.token\n\n\tc.config.modifyLock.RLock()\n\tlimiter := c.config.Limiter\n\thttpClient := c.config.HttpClient\n\toutputCurlString := c.config.OutputCurlString\n\toutputPolicy := c.config.OutputPolicy\n\tdisableRedirects := c.config.DisableRedirects\n\n\t// add headers\n\tif c.headers != nil {\n\t\tfor header, vals := range c.headers {\n\t\t\tfor _, val := range vals {\n\t\t\t\treq.Header.Add(header, val)\n\t\t\t}\n\t\t}\n\t\t// explicitly set the namespace header to current client\n\t\tif ns := c.headers.Get(NamespaceHeaderName); ns != \"\" {\n\t\t\tr.Headers.Set(NamespaceHeaderName, ns)\n\t\t}\n\t}\n\n\tc.config.modifyLock.RUnlock()\n\tc.modifyLock.RUnlock()\n\n\t// OutputCurlString and OutputPolicy logic rely on the request type to be retryable.Request\n\tif outputCurlString {\n\t\treturn nil, fmt.Errorf(\"output-curl-string is not implemented for this request\")\n\t}\n\tif outputPolicy {\n\t\treturn nil, fmt.Errorf(\"output-policy is not implemented for this request\")\n\t}\n\n\treq.URL.User = r.URL.User\n\treq.URL.Scheme = r.URL.Scheme\n\treq.URL.Host = r.URL.Host\n\treq.Host = r.URL.Host\n\n\tif len(r.ClientToken) != 0 {\n\t\treq.Header.Set(AuthHeaderName, r.ClientToken)\n\t}\n\n\tif len(r.WrapTTL) != 0 {\n\t\treq.Header.Set(\"X-Vault-Wrap-TTL\", r.WrapTTL)\n\t}\n\n\tif len(r.MFAHeaderVals) != 0 {\n\t\tfor _, mfaHeaderVal := range r.MFAHeaderVals {\n\t\t\treq.Header.Add(\"X-Vault-MFA\", mfaHeaderVal)\n\t\t}\n\t}\n\n\tif r.PolicyOverride {\n\t\treq.Header.Set(\"X-Vault-Policy-Override\", \"true\")\n\t}\n\n\tif limiter != nil {\n\t\tlimiter.Wait(ctx)\n\t}\n\n\t// check the token before potentially erroring from the API\n\tif err := validateToken(token); err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar result *Response\n\n\tresp, err := httpClient.Do(req)\n\n\tif resp != nil {\n\t\tresult = &Response{Response: resp}\n\t}\n\n\tif err != nil {\n\t\tif strings.Contains(err.Error(), \"tls: oversized\") {\n\t\t\terr = errwrap.Wrapf(\"{{err}}\\n\\n\"+TLSErrorString, err)\n\t\t}\n\t\treturn result, err\n\t}\n\n\t// Check for a redirect, only allowing for a single redirect, if redirects aren't disabled\n\tif (resp.StatusCode == 301 || resp.StatusCode == 302 || resp.StatusCode == 307) && !disableRedirects {\n\t\t// Parse the updated location\n\t\trespLoc, err := resp.Location()\n\t\tif err != nil {\n\t\t\treturn result, fmt.Errorf(\"redirect failed: %s\", err)\n\t\t}\n\n\t\t// Ensure a protocol downgrade doesn't happen\n\t\tif req.URL.Scheme == \"https\" && respLoc.Scheme != \"https\" {\n\t\t\treturn result, fmt.Errorf(\"redirect would cause protocol downgrade\")\n\t\t}\n\n\t\t// Update the request\n\t\treq.URL = respLoc\n\n\t\t// Reset the request body if any\n\t\tif err := r.ResetJSONBody(); err != nil {\n\t\t\treturn result, fmt.Errorf(\"redirect failed: %s\", err)\n\t\t}\n\n\t\t// Retry the request\n\t\tresp, err = httpClient.Do(req)\n\t\tif err != nil {\n\t\t\treturn result, fmt.Errorf(\"redirect failed: %s\", err)\n\t\t}\n\t}\n\n\tif err := result.Error(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn result, nil\n}", "func (c *Env) Copy() *Env {\n\tcontext := *c\n\treturn &context\n}", "func (m *MockRDSAPI) ModifyDBClusterSnapshotAttributeWithContext(arg0 aws.Context, arg1 *rds.ModifyDBClusterSnapshotAttributeInput, arg2 ...request.Option) (*rds.ModifyDBClusterSnapshotAttributeOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"ModifyDBClusterSnapshotAttributeWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.ModifyDBClusterSnapshotAttributeOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (c *CreativesInsertCall) Context(ctx context.Context) *CreativesInsertCall {\n\tc.ctx_ = ctx\n\treturn c\n}", "func (mr *MockRDSAPIMockRecorder) ModifyEventSubscriptionWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ModifyEventSubscriptionWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).ModifyEventSubscriptionWithContext), varargs...)\n}", "func (c Client) FetchWithContext(context context.Context) (*FetchTrunkResponse, error) {\n\top := client.Operation{\n\t\tMethod: http.MethodGet,\n\t\tURI: \"/Trunks/{sid}\",\n\t\tPathParams: map[string]string{\n\t\t\t\"sid\": c.sid,\n\t\t},\n\t}\n\n\tresponse := &FetchTrunkResponse{}\n\tif err := c.client.Send(context, op, nil, response); err != nil {\n\t\treturn nil, err\n\t}\n\treturn response, nil\n}", "func (a *Application) PutContext(r *http.Request) *http.Request {\n\treturn page.PutContext(r, os.Args[1:])\n}", "func ContextTransfer(ctx, targetCtx context.Context) context.Context {\n\tif r := requestFromContext(ctx); r != nil {\n\t\treturn contextWithRequest(targetCtx, r)\n\t}\n\treturn targetCtx\n}", "func (c *CreativesPatchCall) Context(ctx context.Context) *CreativesPatchCall {\n\tc.ctx_ = ctx\n\treturn c\n}", "func withTodo(ctx context.Context, todo *types.Todo) context.Context {\n\treturn context.WithValue(ctx, todoKey, todo)\n}", "func (m *MockRDSAPI) ModifyDBInstanceWithContext(arg0 aws.Context, arg1 *rds.ModifyDBInstanceInput, arg2 ...request.Option) (*rds.ModifyDBInstanceOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"ModifyDBInstanceWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.ModifyDBInstanceOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (env *environment) copy() *environment {\n\tcpy := &environment{\n\t\tsigner: env.signer,\n\t\tstate: env.state.Copy(),\n\t\tancestors: env.ancestors.Clone(),\n\t\tfamily: env.family.Clone(),\n\t\ttcount: env.tcount,\n\t\tcoinbase: env.coinbase,\n\t\theader: model.CopyHeader(env.header),\n\t\treceipts: copyReceipts(env.receipts),\n\t}\n\tif env.gasPool != nil {\n\t\tgasPool := *env.gasPool\n\t\tcpy.gasPool = &gasPool\n\t}\n\t// The content of txs and uncles are immutable, unnecessary\n\t// to do the expensive deep copy for them.\n\tcpy.txs = make([]*model.Transaction, len(env.txs))\n\tcopy(cpy.txs, env.txs)\n\tcpy.uncles = make(map[common.Hash]*model.Header)\n\tfor hash, uncle := range env.uncles {\n\t\tcpy.uncles[hash] = uncle\n\t}\n\treturn cpy\n}", "func WithOriginalSrc(parent context.Context, src string) context.Context {\n\treturn context.WithValue(parent, originalSrcCtxKey, src)\n}", "func (c *OrganizationsSharedflowsRevisionsUpdateSharedFlowRevisionCall) Context(ctx context.Context) *OrganizationsSharedflowsRevisionsUpdateSharedFlowRevisionCall {\n\tc.ctx_ = ctx\n\treturn c\n}", "func (_obj *Apichannels) Channels_editLocationWithContext(tarsCtx context.Context, params *TLchannels_editLocation, _opt ...map[string]string) (ret Bool, err error) {\n\n\tvar length int32\n\tvar have bool\n\tvar ty byte\n\t_os := codec.NewBuffer()\n\terr = params.WriteBlock(_os, 1)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tvar _status map[string]string\n\tvar _context map[string]string\n\tif len(_opt) == 1 {\n\t\t_context = _opt[0]\n\t} else if len(_opt) == 2 {\n\t\t_context = _opt[0]\n\t\t_status = _opt[1]\n\t}\n\t_resp := new(requestf.ResponsePacket)\n\n\terr = _obj.s.Tars_invoke(tarsCtx, 0, \"channels_editLocation\", _os.ToBytes(), _status, _context, _resp)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\t_is := codec.NewReader(tools.Int8ToByte(_resp.SBuffer))\n\terr = ret.ReadBlock(_is, 0, true)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tif len(_opt) == 1 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t} else if len(_opt) == 2 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t\tfor k := range _status {\n\t\t\tdelete(_status, k)\n\t\t}\n\t\tfor k, v := range _resp.Status {\n\t\t\t_status[k] = v\n\t\t}\n\n\t}\n\t_ = length\n\t_ = have\n\t_ = ty\n\treturn ret, nil\n}", "func (o *GetReceiptsParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (_obj *Apichannels) Channels_editCreatorOneWayWithContext(tarsCtx context.Context, params *TLchannels_editCreator, _opt ...map[string]string) (ret Updates, err error) {\n\n\tvar length int32\n\tvar have bool\n\tvar ty byte\n\t_os := codec.NewBuffer()\n\terr = params.WriteBlock(_os, 1)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tvar _status map[string]string\n\tvar _context map[string]string\n\tif len(_opt) == 1 {\n\t\t_context = _opt[0]\n\t} else if len(_opt) == 2 {\n\t\t_context = _opt[0]\n\t\t_status = _opt[1]\n\t}\n\t_resp := new(requestf.ResponsePacket)\n\n\terr = _obj.s.Tars_invoke(tarsCtx, 1, \"channels_editCreator\", _os.ToBytes(), _status, _context, _resp)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tif len(_opt) == 1 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t} else if len(_opt) == 2 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t\tfor k := range _status {\n\t\t\tdelete(_status, k)\n\t\t}\n\t\tfor k, v := range _resp.Status {\n\t\t\t_status[k] = v\n\t\t}\n\n\t}\n\t_ = length\n\t_ = have\n\t_ = ty\n\treturn ret, nil\n}", "func (this *Context) Clone() *Context {\n\tvar clone = *this\n\tclone.Parent = this\n\treturn &clone\n}", "func (_obj *Apichannels) Channels_editAdminWithContext(tarsCtx context.Context, params *TLchannels_editAdmin, _opt ...map[string]string) (ret Updates, err error) {\n\n\tvar length int32\n\tvar have bool\n\tvar ty byte\n\t_os := codec.NewBuffer()\n\terr = params.WriteBlock(_os, 1)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tvar _status map[string]string\n\tvar _context map[string]string\n\tif len(_opt) == 1 {\n\t\t_context = _opt[0]\n\t} else if len(_opt) == 2 {\n\t\t_context = _opt[0]\n\t\t_status = _opt[1]\n\t}\n\t_resp := new(requestf.ResponsePacket)\n\n\terr = _obj.s.Tars_invoke(tarsCtx, 0, \"channels_editAdmin\", _os.ToBytes(), _status, _context, _resp)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\t_is := codec.NewReader(tools.Int8ToByte(_resp.SBuffer))\n\terr = ret.ReadBlock(_is, 0, true)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tif len(_opt) == 1 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t} else if len(_opt) == 2 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t\tfor k := range _status {\n\t\t\tdelete(_status, k)\n\t\t}\n\t\tfor k, v := range _resp.Status {\n\t\t\t_status[k] = v\n\t\t}\n\n\t}\n\t_ = length\n\t_ = have\n\t_ = ty\n\treturn ret, nil\n}", "func (o *GetPublicsRecipeParams) WithContext(ctx context.Context) *GetPublicsRecipeParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (_obj *DataService) DeleteApplyWithContext(tarsCtx context.Context, wx_id string, club_id string, affectRows *int32, _opt ...map[string]string) (ret int32, err error) {\n\n\tvar length int32\n\tvar have bool\n\tvar ty byte\n\t_os := codec.NewBuffer()\n\terr = _os.Write_string(wx_id, 1)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\terr = _os.Write_string(club_id, 2)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\terr = _os.Write_int32((*affectRows), 3)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tvar _status map[string]string\n\tvar _context map[string]string\n\tif len(_opt) == 1 {\n\t\t_context = _opt[0]\n\t} else if len(_opt) == 2 {\n\t\t_context = _opt[0]\n\t\t_status = _opt[1]\n\t}\n\t_resp := new(requestf.ResponsePacket)\n\n\terr = _obj.s.Tars_invoke(tarsCtx, 0, \"deleteApply\", _os.ToBytes(), _status, _context, _resp)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\t_is := codec.NewReader(tools.Int8ToByte(_resp.SBuffer))\n\terr = _is.Read_int32(&ret, 0, true)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\terr = _is.Read_int32(&(*affectRows), 3, true)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\n\tif len(_opt) == 1 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t} else if len(_opt) == 2 {\n\t\tfor k := range _context {\n\t\t\tdelete(_context, k)\n\t\t}\n\t\tfor k, v := range _resp.Context {\n\t\t\t_context[k] = v\n\t\t}\n\t\tfor k := range _status {\n\t\t\tdelete(_status, k)\n\t\t}\n\t\tfor k, v := range _resp.Status {\n\t\t\t_status[k] = v\n\t\t}\n\n\t}\n\t_ = length\n\t_ = have\n\t_ = ty\n\treturn ret, nil\n}", "func (c *LicenseAssignmentsInsertCall) Context(ctx context.Context) *LicenseAssignmentsInsertCall {\n\tc.ctx_ = ctx\n\treturn c\n}", "func (m pWithContext) Extend(fnct func(m.UserSet, string, interface{}) m.UserSet) pWithContext {\n\treturn pWithContext{\n\t\tMethod: m.Method.Extend(fnct),\n\t}\n}" ]
[ "0.6065512", "0.5775695", "0.5548469", "0.53784853", "0.5308844", "0.5259793", "0.52479076", "0.52151114", "0.52060354", "0.5190454", "0.5179951", "0.51040375", "0.50996315", "0.50958705", "0.5086578", "0.5032771", "0.49986193", "0.49502012", "0.49191132", "0.48936892", "0.4882121", "0.48802188", "0.48737976", "0.48696804", "0.48543766", "0.48386228", "0.4837272", "0.48248476", "0.4814945", "0.48115858", "0.48026013", "0.4786724", "0.478555", "0.47824612", "0.47735", "0.47720903", "0.47708943", "0.4766081", "0.4764084", "0.47599772", "0.47582117", "0.4743759", "0.47433895", "0.47381237", "0.47346005", "0.47307238", "0.47256112", "0.47211534", "0.47176212", "0.47154757", "0.47105044", "0.47046226", "0.46989408", "0.4682128", "0.4672137", "0.4667094", "0.4662599", "0.46576428", "0.46560192", "0.4643769", "0.46188334", "0.4607595", "0.46046776", "0.46032894", "0.4602275", "0.4602275", "0.4601795", "0.4600154", "0.45919228", "0.45850405", "0.45832887", "0.45827544", "0.45775145", "0.4573782", "0.4569334", "0.4561205", "0.4561205", "0.455889", "0.4551789", "0.45514125", "0.4550337", "0.45447358", "0.45441964", "0.45420322", "0.45377967", "0.453446", "0.4529983", "0.4524906", "0.45245603", "0.4523205", "0.45227644", "0.45223987", "0.45150778", "0.45064634", "0.45046923", "0.4499884", "0.4498217", "0.44924864", "0.44862428", "0.44781783" ]
0.58584803
1
SetContext adds the context to the copy recipe to my recipes with changes params
func (o *CopyRecipeToMyRecipesWithChangesParams) SetContext(ctx context.Context) { o.Context = ctx }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *RevertProductSnapshotRequestUsingPOSTParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetPublicsRecipeParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithContext(ctx context.Context) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (ctx *Context) Copy() *Context {\n\tvar pathParams Params\n\tif len(ctx.PathParams) > 0 {\n\t\tpathParams = append(pathParams, ctx.PathParams...)\n\t}\n\treturn &Context{\n\t\tresponseWriter2: nil,\n\t\tResponseWriter: nil,\n\t\tRequest: ctx.Request,\n\t\tPathParams: pathParams,\n\t\tqueryParams: ctx.queryParams,\n\t\tValidator: ctx.Validator,\n\t\tfetchClientIPFromHeader: ctx.fetchClientIPFromHeader,\n\t\thandlers: nil,\n\t\thandlerIndex: __abortHandlerIndex,\n\t\tkvs: ctx.kvs,\n\t}\n}", "func (c *Context) Copy() *Context {\n\tret := *c\n\tret.init(&fasthttp.RequestCtx{})\n\tc.Request.CopyTo(&ret.Request)\n\tc.Response.CopyTo(&ret.Response)\n\tret.WSConn = c.WSConn\n\tret.data = c.data\n\treturn &ret\n}", "func (o *GetReceiptsParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetIngredientVersionRevisionParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (rec *RawEventCreate) SetContext(m map[string]interface{}) *RawEventCreate {\n\trec.mutation.SetContext(m)\n\treturn rec\n}", "func (o *CreateGitWebhookUsingPOSTParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *PcloudPvminstancesSnapshotsRestorePostParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetRepository15Params) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *AddItemParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (f *Fastglue) SetContext(c interface{}) {\n\tf.context = c\n}", "func (o *ConfigurationBackupModifyParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetNutritionForSingleParsedPlainTextIngredientParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *EditParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *AddRepositoryParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *PostPartsParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CreateCartUsingPOSTParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *PatchAddonParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetV1IntegrationsAwsCloudtrailBatchesParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *FreezeParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CreateVolumeBackupParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *PostReconciliationParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *AddOrUpdateNodePoolConfigItemParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CreateScriptParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *SetUniverseBackupFlagParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (evalCtx *extendedEvalContext) copy() *extendedEvalContext {\n\tcpy := *evalCtx\n\tcpy.EvalContext = *evalCtx.EvalContext.Copy()\n\treturn &cpy\n}", "func (o *FileInfoCreateParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *UpdateSubnetParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CreatePackageRepositoryDeltaUploadParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *NarrowSearchRecipeParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *StartGatewayBundleUsingPOSTParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func SetContext(ctx *apm.Context, req *http.Request, resp *Response, body *apm.BodyCapturer) {\n\tctx.SetHTTPRequest(req)\n\tctx.SetHTTPRequestBody(body)\n\tctx.SetHTTPStatusCode(resp.StatusCode)\n\tctx.SetHTTPResponseHeaders(resp.Headers)\n}", "func (o *ChatNewParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *VectorThumbnailParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (c *fakeRedisConn) SetContext(v interface{}) {}", "func (o *UpdateStockReceiptParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *UpdateAddonParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func ContextTransfer(sourceCtx context.Context, targetCtx context.Context) context.Context {\n\ttoken := tokenFromContext(sourceCtx)\n\treturn context.WithValue(targetCtx, tokenKey, token)\n}", "func (o *GetLolCatalogV1ItemsParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (d *DialogueService) SetContext(context string) {\n\td.client.context = context\n}", "func (o *GetZippedParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *PostApplyManifestParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetRestoreDatalakeStatusParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *UpdateRunbookParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CreateIscsiLunSnapshotParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *ImportApplicationUsingPOSTParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *UpdateNetworkCellularGatewaySettingsSubnetPoolParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CreateWidgetParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *PatchStorageVirtualDriveExtensionsMoidParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *PostAPIV3MachinesParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *SharedCatalogSharedCatalogRepositoryV1SavePostParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *PostMeArticlesDraftsParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetaspecificPbxDeviceFirmwareBinaryParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetBundleByKeyParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *PostReplicationPoliciesParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *SearchKeywordChunkedParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CustomerGatewayUpdateOwnershipParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetInterceptionitemsParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (ctx context) clone() context {\n\treturn context{\n\t\tkeyvals: safeSlice(ctx.keyvals),\n\t}\n}", "func (o *APIServiceMissedBlocksParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetDeploymentPreview1Params) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *SkuPackPostParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetMarketsRegionIDHistoryParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetRackTopoesParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *AddVMParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (f *flags) SetContext(ctx *cli.Context) {\n\tf.ctx = ctx\n}", "func (o *PostDeviceRackParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetProductsByIDVariationAttributesByIDParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (this *L0JobContext) Copy(request string) JobContext {\n\treturn &L0JobContext{\n\t\trequest: request,\n\t\tjobID: this.jobID,\n\t\tlogic: this.logic,\n\t\tloadBalancerLogic: this.loadBalancerLogic,\n\t\tserviceLogic: this.serviceLogic,\n\t\tenvironmentLogic: this.environmentLogic,\n\t\tmutex: this.mutex,\n\t}\n}", "func (o *GetRacksParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetWorkItemParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *UpdateFolderParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetModerationRulesParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *BudgetAddParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *PostAPI24ProtectionGroupSnapshotsParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CreatePolicyResetItemParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *AllLookmlTestsParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CloudNFSExportAddParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func CopyContext(dest context.Context, src context.Context) (context.Context, error) {\n\ttrace := GetTraceFromContext(src)\n\tspan := GetSpanFromContext(src)\n\tif trace == nil || span == nil {\n\t\treturn dest, ErrTraceNotFoundInContext\n\t}\n\tdest = PutTraceInContext(dest, trace)\n\tdest = PutSpanInContext(dest, span)\n\treturn dest, nil\n}", "func (o *ExtractionListV1Params) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *SafeObjectListParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (c *VariantsetsPatchCall) Context(ctx context.Context) *VariantsetsPatchCall {\n\tc.ctx_ = ctx\n\treturn c\n}", "func (s *Steps) SetContext(ctx *context.BDDContext) {\n\ts.bddContext = ctx\n}", "func (o *PostDocumentMergeParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetContentSourceUsingGETParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *ExtrasSavedFiltersListParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CreatePublicIPAdressUsingPOSTParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func ContextClone(ctx *Context) (*Context, error) {\n\tother := newContext()\n\tother.ctx = C.secp256k1_context_clone(ctx.ctx)\n\treturn other, nil\n}", "func (o *GetBackupLocationsParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CreateRunbookRunCreateParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *UpdateFeaturesConfigurationParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetBlockLatestParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetSingleBeadSimulationsParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CatalogProductTierPriceManagementV1AddPostParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetAiProductRecommendationsByIDParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *CreateSwiftPasswordParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetPackageSearchActionOldSpacesParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (o *GetDatalakeDbConfigParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}" ]
[ "0.62970084", "0.61870456", "0.61496747", "0.6082606", "0.5998157", "0.59815085", "0.5933783", "0.5929907", "0.59284055", "0.58162916", "0.5791815", "0.57468003", "0.5741497", "0.57338977", "0.5702138", "0.56917036", "0.56914335", "0.566592", "0.5653271", "0.56424826", "0.5628637", "0.5615693", "0.56129295", "0.5605682", "0.55995166", "0.55878526", "0.55660164", "0.55564713", "0.55423534", "0.55383945", "0.552462", "0.551025", "0.5479709", "0.5478579", "0.54756117", "0.5464842", "0.54646087", "0.54636085", "0.5460248", "0.54600966", "0.5459484", "0.5454265", "0.5452136", "0.5451395", "0.54505026", "0.5444703", "0.5444363", "0.5442037", "0.5441142", "0.54408103", "0.54402757", "0.5437061", "0.5434735", "0.5434455", "0.54286224", "0.54174906", "0.54170144", "0.54036176", "0.5402292", "0.53992724", "0.5397063", "0.53927225", "0.5391971", "0.53908134", "0.53903806", "0.53863347", "0.53781486", "0.5375762", "0.5375378", "0.53743136", "0.53681445", "0.536756", "0.5367492", "0.5366743", "0.53664505", "0.5361862", "0.5357045", "0.5356894", "0.5356474", "0.53515327", "0.53509367", "0.53478944", "0.5344313", "0.5343958", "0.5342382", "0.53418183", "0.53399336", "0.5338999", "0.5337036", "0.5335565", "0.53329873", "0.5328937", "0.53266007", "0.5324556", "0.53237337", "0.53192794", "0.531852", "0.53153527", "0.53141755", "0.53128165" ]
0.67339253
0
WithHTTPClient adds the HTTPClient to the copy recipe to my recipes with changes params
func (o *CopyRecipeToMyRecipesWithChangesParams) WithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams { o.SetHTTPClient(client) return o }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *CopyRecipeToMyRecipesWithChangesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetPublicsRecipeParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *AddItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *RevertProductSnapshotRequestUsingPOSTParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func NewCopyRecipeToMyRecipesWithChangesParamsWithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\t\tHTTPClient: client,\n\t}\n}", "func (o *GetIngredientVersionRevisionParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetReceiptsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *BudgetAddParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func WithHTTPClient(httpClient *http.Client) ClientOption {\n\treturn func(c *Client) {\n\t\tc.sling.Client(httpClient)\n\t}\n}", "func (o *GetNutritionForSingleParsedPlainTextIngredientParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *NarrowSearchRecipeParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func WithHTTPClient(c *http.Client) func(*Client) {\n\treturn func(mr *Client) {\n\t\tmr.client = c\n\t}\n}", "func (o *CreateGitWebhookUsingPOSTParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateCartUsingPOSTParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *EditParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func withHTTPClient(target *http.Client) ClientOption {\n\treturn func(subject *client) {\n\t\tsubject.client = target\n\t}\n}", "func (o *PostReconciliationParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func WithHTTPClient(httpClient *http.Client) ClientOption {\n\treturn func(c *client) error {\n\t\tif httpClient == nil {\n\t\t\treturn errors.InvalidParameterError{Parameter: \"httpClient\", Reason: \"cannot be empty\"}\n\t\t}\n\n\t\tc.requester.Client = httpClient\n\t\treturn nil\n\t}\n}", "func (o *PostPartsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *FreezeParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ConfigurationBackupModifyParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func WithHTTPClient(client *http.Client) OptionFunc {\n\treturn func(c *Client) {\n\t\tc.client = client\n\t}\n}", "func WithHTTPClient(client *http.Client) func(c *Client) error {\n\treturn func(c *Client) error {\n\t\tif client == nil {\n\t\t\treturn errors.New(\"HTTP client is nil\")\n\t\t}\n\t\tc.client = client\n\t\treturn nil\n\t}\n}", "func WithHTTPClient(client *http.Client) Opt {\n\treturn func(c *Client) error {\n\t\tif client != nil {\n\t\t\tc.client = client\n\t\t}\n\t\treturn nil\n\t}\n}", "func (o *AddOrUpdateNodePoolConfigItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func WithHTTPClient(client *http.Client) Opt {\n\treturn func(c *Client) {\n\t\tc.httpClient = client\n\t}\n}", "func (o *AddBranchParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func WithHTTPClient(h *http.Client) Opts {\n\treturn func(r *retryable) {\n\t\tr.httpClient = h\n\t}\n}", "func (o *CreatePolicyResetItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func WithHTTPClient(httpclient *http.Client) ClientOption {\n\treturn func(client *Client) {\n\t\tclient.httpClient = httpclient\n\t}\n}", "func (o *PostMenuItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func SetHTTPClient(httpClient *http.Client) func(*Client) error {\n\treturn func(client *Client) error {\n\t\tclient.client = httpClient\n\n\t\treturn nil\n\t}\n}", "func WithHTTPClient(httpClient *http.Client) ClientOption {\n\treturn func(client *Client) {\n\t\tclient.httpClient = httpClient\n\t}\n}", "func (o *PatchAddonParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetPrivateToggleDepositAddressCreationParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *AddRepositoryParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *RewardCommentsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PutMenuItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func WithHTTPClient(c *http.Client) Option {\n\treturn func(args *Client) {\n\t\targs.httpClient = c\n\t}\n}", "func (o *PostMeArticlesDraftsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateStockReceiptParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *RemoveDropRequestParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateRunbookParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateAddonParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *InventoryStocktakingSearchParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetInterceptionitemsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func WithHTTPClient(httpClient *http.Client) ClientOption {\n\treturn func(c *Client) {\n\t\tc.httpClient = httpClient\n\t}\n}", "func (o *CustomerGatewayUpdateOwnershipParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ChatNewParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func AddClient(nbmaster string, httpClient *http.Client, jwt string) {\r\n fmt.Printf(\"\\nSending a PUT request to add client %s to policy %s...\\n\", testClientName, testPolicyName)\r\n\r\n client := map[string]interface{}{\r\n \"data\": map[string]interface{}{\r\n \"type\": \"client\",\r\n \"attributes\": map[string]string{\r\n \"hardware\": \"VMware\",\r\n \"hostName\": \"MEDIA_SERVER\",\r\n \"OS\": \"VMware\"}}}\r\n\r\n clientRequest, _ := json.Marshal(client)\r\n\r\n uri := \"https://\" + nbmaster + \":\" + port + \"/netbackup/\" + policiesUri + testPolicyName + \"/clients/\" + testClientName\r\n\r\n request, _ := http.NewRequest(http.MethodPut, uri, bytes.NewBuffer(clientRequest))\r\n request.Header.Add(\"Content-Type\", contentTypeV2);\r\n request.Header.Add(\"Authorization\", jwt);\r\n request.Header.Add(\"If-Match\", \"1\");\r\n request.Header.Add(\"X-NetBackup-Audit-Reason\", \"added client \" + testClientName + \" to policy \" + testPolicyName);\r\n\r\n response, err := httpClient.Do(request)\r\n\r\n if err != nil {\r\n fmt.Printf(\"The HTTP request failed with error: %s\\n\", err)\r\n panic(\"Unable to add client to policy.\\n\")\r\n } else {\r\n if response.StatusCode != 201 {\r\n printErrorResponse(response)\r\n } else {\r\n fmt.Printf(\"%s added to %s successfully.\\n\", testClientName, testPolicyName);\r\n responseDetails, _ := httputil.DumpResponse(response, true);\r\n fmt.Printf(string(responseDetails))\r\n }\r\n }\r\n}", "func WithHTTPClient(client *http.Client) ClientOption {\n\treturn func(c *Client) {\n\t\tc.httpClient = client\n\t}\n}", "func (o *ExtrasSavedFiltersListParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PatchReferenceEntityRecordsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func SetHTTPClient(newClient *http.Client) {\n\thttpClient = newClient\n}", "func (o *CreateWidgetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateListParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (cb *ClientBuilder) HTTPClient(httpClient *http.Client) *ClientBuilder {\n\tcb.client.httpClient = httpClient\n\treturn cb\n}", "func (cb *ClientBuilder) HTTPClient(httpClient *http.Client) *ClientBuilder {\n\tcb.client.httpClient = httpClient\n\treturn cb\n}", "func (o *GetWorkItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetZippedParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func WithHTTPClient(client *http.Client) Option {\n\treturn func(c *Client) error {\n\t\tif client == nil {\n\t\t\treturn errors.New(\"client cannot be nil\")\n\t\t}\n\n\t\tc.client = client\n\t\treturn nil\n\t}\n}", "func WithHTTPClient(client HTTPClient) Option {\n\treturn func(opts *Client) {\n\t\topts.httpClient = client\n\t}\n}", "func (o *ActionDeploymentRequestUsingPOST2Params) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostMeOvhAccountOvhAccountIDCreditOrderParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostApplyManifestParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *WaitListParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetItemByAppIDParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func WithHTTPClient(h HTTPClient) ClientOption {\n\treturn clientOptionFunc(func(c interface{}) {\n\t\tswitch c := c.(type) {\n\t\tcase *Client:\n\t\t\tc.httpClient = h\n\t\tdefault:\n\t\t\tpanic(\"unknown type\")\n\t\t}\n\t})\n}", "func (o *GetV1IntegrationsAwsCloudtrailBatchesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ChargeAddonInvoiceParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetShopItemListParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func WithHTTPClient(hClient *http.Client) clientOption {\n\treturn func(c *client) {\n\t\tc.httpClient = hClient\n\t}\n}", "func (o *AddItemParams) WithHTTPClient(client *http.Client) *AddItemParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (o *CloudNFSExportAddParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ListDCForSeedParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ToggleNetworkGeneratorsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CheckTransactionCreditLimitParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *EstimateCoinBuyParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateFlowParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PatchRetryEventUsingPATCHParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetContentSourceUsingGETParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *SetUniverseBackupFlagParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreatePackageRepositoryDeltaUploadParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdatePriceListParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (c *Client) WithHTTPClient(cl *http.Client) *Client {\n\tc.Client = cl\n\treturn c\n}", "func (o *PetCreateParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *TurnOnLightParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateRowParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetRequestDetailsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateSubnetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ImportApplicationUsingPOSTParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *SharedCatalogSharedCatalogRepositoryV1SavePostParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PatchLibrariesByIDContentByIDParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetLolCatalogV1ItemsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *RebuildIndexSetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateTransactionCategorizationRuleParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetRecentFoodsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PutParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *APIServiceMissedBlocksParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *NearestUsingGET1Params) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}" ]
[ "0.68199915", "0.6054078", "0.59095484", "0.58934873", "0.58423924", "0.5841846", "0.5841459", "0.58388424", "0.5804307", "0.5749258", "0.5740945", "0.57321674", "0.57095027", "0.5701237", "0.5682556", "0.56596583", "0.5657443", "0.5632348", "0.5626681", "0.5624085", "0.5623318", "0.56171", "0.56163394", "0.56127757", "0.55901235", "0.5561946", "0.5551378", "0.55486834", "0.55430585", "0.55350083", "0.5524862", "0.55246913", "0.5524687", "0.5519972", "0.55071783", "0.5504641", "0.54954094", "0.5492696", "0.54696715", "0.5468691", "0.5459408", "0.54502624", "0.5446549", "0.5440246", "0.54323316", "0.543021", "0.5429159", "0.5417493", "0.54173154", "0.54160476", "0.541047", "0.5407576", "0.540627", "0.5401972", "0.54001075", "0.53974944", "0.5395133", "0.5395133", "0.53944147", "0.53921086", "0.5381287", "0.53735566", "0.53717893", "0.536705", "0.53660005", "0.53634524", "0.5360766", "0.53599477", "0.5356654", "0.53521496", "0.53481114", "0.534749", "0.53426087", "0.53395337", "0.53374094", "0.5335667", "0.5329211", "0.53286874", "0.53286123", "0.5328211", "0.5326901", "0.53252643", "0.5324756", "0.5319993", "0.5316284", "0.53154224", "0.531322", "0.53076625", "0.5305749", "0.5304368", "0.53028965", "0.5300504", "0.5300174", "0.52993345", "0.5294867", "0.5294246", "0.5294246", "0.5293764", "0.52937335", "0.52927196" ]
0.7069977
0
SetHTTPClient adds the HTTPClient to the copy recipe to my recipes with changes params
func (o *CopyRecipeToMyRecipesWithChangesParams) SetHTTPClient(client *http.Client) { o.HTTPClient = client }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *GetPublicsRecipeParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *RevertProductSnapshotRequestUsingPOSTParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (o *GetIngredientVersionRevisionParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *AddItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func SetHTTPClient(newClient *http.Client) {\n\thttpClient = newClient\n}", "func (o *GetReceiptsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateGitWebhookUsingPOSTParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *NarrowSearchRecipeParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateCartUsingPOSTParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *EditParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetNutritionForSingleParsedPlainTextIngredientParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func SetHTTPClient(httpClient *http.Client) func(*Client) error {\n\treturn func(client *Client) error {\n\t\tclient.client = httpClient\n\n\t\treturn nil\n\t}\n}", "func (o *AddOrUpdateNodePoolConfigItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *FreezeParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ConfigurationBackupModifyParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostReconciliationParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostPartsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func SetHTTPClient(client *http.Client) {\n\thttpClient = client\n}", "func (o *CreatePolicyResetItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PatchAddonParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *BudgetAddParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *AddRepositoryParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostMenuItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *RewardCommentsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ToggleNetworkGeneratorsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateAddonParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *AddBranchParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *RebuildIndexSetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateStockReceiptParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetContentSourceUsingGETParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ChatNewParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateRunbookParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PutMenuItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateWidgetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PatchRetryEventUsingPATCHParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ActionDeploymentRequestUsingPOST2Params) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *SetUniverseBackupFlagParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *IntegrationsManualHTTPSCreateParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (c *Client) SetHTTPClient(client *http.Client) {\n\tc.client = client\n}", "func (c *Client) SetHTTPClient(client *http.Client) {\n\tc.client = client\n}", "func (o *UpdateNetworkSwitchAccessControlListsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *TurnOnLightParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreatePackageRepositoryDeltaUploadParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateNetworkCellularGatewaySettingsSubnetPoolParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *WaitListParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PatchReferenceEntityRecordsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetPrivateToggleDepositAddressCreationParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateFlowParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateFeaturesConfigurationParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostApplyManifestParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (sm *Manager) SetHTTPClient(c *http.Client) {\n\tsm.client = c\n}", "func (sm *Manager) SetHTTPClient(c *http.Client) {\n\tsm.client = c\n}", "func (o *GetRuleChainParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetWorkItemParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetZippedParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetTerraformConfigurationSourcesUsingGET1Params) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetInterceptionitemsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ImportApplicationUsingPOSTParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *SharedCatalogSharedCatalogRepositoryV1SavePostParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *NearestUsingGET1Params) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateTransactionCategorizationRuleParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateCredentialParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateDmrClusterLinkParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetRequestDetailsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CheckTransactionCreditLimitParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *RemoveDropRequestParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *InventoryStocktakingSearchParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateCrossConnectParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *CreateScriptParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ExtractionListV1Params) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (c *Client) SetHTTPClient(httpClient HTTPClient) {\n\tif httpClient == nil {\n\t\tc.client = http.DefaultClient\n\t} else {\n\t\tc.client = httpClient\n\t}\n}", "func (o *CustomerGatewayUpdateOwnershipParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *SaveTemplateParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (service *BaseService) SetHTTPClient(client *http.Client) {\n\tsetMinimumTLSVersion(client)\n\n\tif isRetryableClient(service.Client) {\n\t\t// If \"service\" is currently holding a retryable client,\n\t\t// then set \"client\" as the embedded client used for individual requests.\n\t\ttr := service.Client.Transport.(*retryablehttp.RoundTripper)\n\t\ttr.Client.HTTPClient = client\n\t} else {\n\t\t// Otherwise, just hang \"client\" directly off the base service.\n\t\tservice.Client = client\n\t}\n}", "func (o *DevicesGetModuleComponentCommandHistoryParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetContentSourcesUsingGETParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateWidgetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ListDCForSeedParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetSingleBeadSimulationsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostMeArticlesDraftsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ResolveBatchParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PostAsyncParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetLolCatalogV1ItemsParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ExtrasSavedFiltersListParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetRepository15Params) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *SearchKeywordChunkedParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ChargeAddonInvoiceParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateSubnetParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdatePriceListParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetRestoreDatalakeStatusParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func SetHTTPClient(client *http.Client) error {\n\tif client == nil {\n\t\treturn errHTTPClientInvalid\n\t}\n\tm.Lock()\n\t_HTTPClient = client\n\tm.Unlock()\n\treturn nil\n}", "func (o *PostIPAMSwitchesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *GetRedbeamsFlowLogsProgressByResourceCrnParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PutFlagSettingParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *QueryChangesParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ExtrasGraphsReadParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *ModifyProxyConfigInternalParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *PutCwfNetworkIDParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}", "func (o *UpdateAccountStateParams) SetHTTPClient(client *http.Client) {\n\to.HTTPClient = client\n}" ]
[ "0.7151808", "0.7038088", "0.6908971", "0.6900971", "0.68990266", "0.6891095", "0.6878877", "0.68639123", "0.68540007", "0.6851541", "0.68416125", "0.6839482", "0.68369865", "0.68172693", "0.67863685", "0.67825335", "0.6768729", "0.67661315", "0.67284036", "0.669015", "0.6666474", "0.66595423", "0.66272116", "0.66251993", "0.6624819", "0.66120535", "0.6600765", "0.6595484", "0.6585106", "0.65792525", "0.6576837", "0.65640575", "0.65532804", "0.65520924", "0.6549552", "0.6542287", "0.65351915", "0.65347975", "0.6531616", "0.6530088", "0.6530088", "0.6524064", "0.6521851", "0.6519381", "0.65123063", "0.6498727", "0.6497561", "0.649555", "0.64926445", "0.64852166", "0.6482254", "0.6476503", "0.6476503", "0.6475066", "0.6471768", "0.6470399", "0.64627075", "0.6460342", "0.645829", "0.64564514", "0.64564407", "0.64563817", "0.6451665", "0.64506894", "0.6445098", "0.6444671", "0.64438236", "0.6442815", "0.644216", "0.64374065", "0.64351547", "0.6435119", "0.6434535", "0.6434402", "0.64330906", "0.6431951", "0.64316833", "0.64298123", "0.64272606", "0.64260316", "0.6422784", "0.6417391", "0.6416443", "0.6415855", "0.6407109", "0.6406952", "0.64056545", "0.6403929", "0.6401559", "0.6400267", "0.6399428", "0.639881", "0.63949907", "0.6386722", "0.6385927", "0.6385854", "0.63853043", "0.6384726", "0.6383152", "0.6380801" ]
0.7771742
0
WithJSONBody adds the jSONBody to the copy recipe to my recipes with changes params
func (o *CopyRecipeToMyRecipesWithChangesParams) WithJSONBody(jSONBody *models.Recipe) *CopyRecipeToMyRecipesWithChangesParams { o.SetJSONBody(jSONBody) return o }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *CopyRecipeToMyRecipesWithChangesParams) SetJSONBody(jSONBody *models.Recipe) {\n\to.JSONBody = jSONBody\n}", "func (a *API) JSONBody(ctx *fasthttp.RequestCtx, model interface{}) {\n\tr := bytes.NewReader(ctx.PostBody())\n\tjson.NewDecoder(r).Decode(&model)\n}", "func (c *RBController) RecipeJSONAdvanced(w http.ResponseWriter, r *http.Request) (err error) {\n\tr.ParseForm()\n\tstrict, err := strconv.Atoi(r.PostFormValue(\"strict\"))\n\tname := r.PostFormValue(\"name\")\n\tcuisine, _ := strconv.Atoi(r.PostFormValue(\"cuisine\"))\n\tseason, _ := strconv.Atoi(r.PostFormValue(\"season\"))\n\tmealtype, _ := strconv.Atoi(r.PostFormValue(\"mealtype\"))\n\n\t// get all the recipes that match\n\tvar recipes *list.List\n\tif strict == 0 {\n\t\trecipes, err = c.GetRecipesLoose(name, cuisine, mealtype, season)\n\t} else {\n\t\trecipes, err = c.GetRecipesStrict(name, cuisine, mealtype, season)\n\t}\n\n\t// slice of jsons\n\tjsons := make([]string, recipes.Len())\n\n\tif err == nil {\n\t\tindex := 0\n\t\tfor e := recipes.Front(); e != nil; e = e.Next() {\n\t\t\trec := e.Value.(*Recipe)\n\t\t\tjsons[index] = rec.ToJSON()\n\t\t\tindex++\n\t\t}\n\t\trequest := strings.Join(jsons, \"\\n\")\n\t\tfmt.Fprintf(w, request)\n\t} else {\n\t\tfmt.Fprintf(w, \"%v\", err.Error())\n\t}\n\treturn\n}", "func (r *Request) SetJSONBody(val interface{}) error {\n\tbuf := bytes.NewBuffer(nil)\n\tenc := json.NewEncoder(buf)\n\tif err := enc.Encode(val); err != nil {\n\t\treturn err\n\t}\n\n\tr.Obj = val\n\tr.Body = buf\n\tr.BodySize = int64(buf.Len())\n\treturn nil\n}", "func (c *RBController) RecipeJSON(w http.ResponseWriter, r *http.Request) (err error) {\n\tvars := mux.Vars(r)\n\tid, _ := strconv.Atoi(vars[\"id\"])\n\trecipe, err := c.GetRecipe(id)\n\tif err == nil {\n\t\tc.JSON(w, http.StatusOK, recipe)\n\t} else if err == sql.ErrNoRows {\n\t\tc.RenderError(w, 404, \"Sorry, your page wasn't found\")\n\t\terr = nil\n\t}\n\treturn\n}", "func TestAddRecipe(t *testing.T) {\n\tpayload := fmt.Sprintf(`\n {\n \"mealtype\": \"Breakfast\",\n \"name\": \"Pancakes\",\n \"Ingredients\": [ \"150g all purpose flour\",\n \t\t\t\t \"150ml of milk\"],\n \"preparation\": \"Add all ingredients and mix. Put in Pan.\"\n}`)\n\n\tresponse, err := http.Post(baseURL+\"/recipes\", \"application/json\", strings.NewReader(payload))\n\tif err != nil {\n\t\tt.Fatalf(\"failed to get json, %s\", err)\n\t}\n\n\tcheckResponseCode(t, http.StatusOK, response.StatusCode)\n}", "func (o *PostMenuItemParams) SetJSONBody(jSONBody *models.MenuItem) {\n\to.JSONBody = jSONBody\n}", "func (o *PutMenuItemParams) SetJSONBody(jSONBody *models.MenuItem) {\n\to.JSONBody = jSONBody\n}", "func (api *FoodRecipeAPI) partialRecipeUpdate(w http.ResponseWriter, req *http.Request) {\n\tdefer DrainBody(req)\n\tctx := req.Context()\n\n\tvars := mux.Vars(req)\n\tid := vars[\"id\"]\n\tlogData := log.Data{\"id\": id}\n\n\tvar errorObjects []*models.ErrorObject\n\n\tpatchJSON, recipePatches, err := patch.Get(ctx, req.Body)\n\tif err != nil {\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// Validate patch request\n\tfor i, recipePatch := range *recipePatches {\n\t\tif err = recipePatch.Validate(nil); err != nil {\n\t\t\tif _, ok := err.(*validator.InvalidValidationError); ok {\n\t\t\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\t\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tfor _, err := range err.(validator.ValidationErrors) {\n\t\t\t\terrorObjects = append(errorObjects, models.HandleValidationErrors(strconv.Itoa(i), err.ActualTag(), err.StructField(), err.Value().(string), err.Param()))\n\t\t\t}\n\t\t}\n\t}\n\tif len(errorObjects) > 0 {\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// apply patch against recipe resource\n\tp, err := jsonpatch.DecodePatch(patchJSON)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: unable to decode patch\", err)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// find current recipe doc\n\tvar recipe models.Recipe\n\n\tcollection := api.MongoClient.Database(\"food-recipes\").Collection(\"recipes\")\n\tif err = collection.FindOne(ctx, bson.M{\"_id\": id}).Decode(&recipe); err != nil {\n\t\tif err == mongo.ErrNoDocuments {\n\t\t\tlog.Warn(ctx, \"patch recipe: failed to find recipe\", log.FormatErrors([]error{err}), logData)\n\t\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrRecipeNotFound.Error()})\n\t\t\tErrorResponse(ctx, w, http.StatusNotFound, &models.ErrorResponse{Errors: errorObjects})\n\t\t\treturn\n\t\t}\n\n\t\tlog.Error(ctx, \"patch recipe: failed to find recipe, bad connection?\", err)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\tb, err := json.Marshal(recipe)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: error returned from json marshal\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// apply patch to existing recipe\n\tmodified, err := p.Apply(b)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: unable to apply patch to recipe\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\terr = json.Unmarshal(modified, &recipe)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: unmarshal modified recipe into recipe struct\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// store new recipe\n\tif _, err = collection.ReplaceOne(ctx, bson.M{\"_id\": id}, recipe); err != nil {\n\t\tif err == mongo.ErrNoDocuments {\n\t\t\tlog.Error(ctx, \"update recipe: failed to update recipe, recipe deos not exists\", err, logData)\n\t\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrRecipeNotFound.Error()})\n\t\t\tErrorResponse(ctx, w, http.StatusNotFound, &models.ErrorResponse{Errors: errorObjects})\n\t\t\treturn\n\t\t}\n\n\t\tlog.Error(ctx, \"update recipe: failed to insert recipe\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\n\tlog.Info(ctx, \"update recipe: request successful\", logData)\n}", "func NewJSONBody(v interface{}) (contentType string, bodyReader io.Reader, err error) {\n\tb, err := json.Marshal(v)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn \"application/json;charset=utf-8\", bytes.NewReader(b), nil\n}", "func BindJSON(r *http.Request, target interface{}) error {\n\tdata, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = json.Unmarshal(data, target)\n\n\treturn err\n}", "func (c *apiClient) newRequestWithJSONBody(\n\tctx context.Context, method, resourcePath string,\n\tquery url.Values, body interface{}) (*http.Request, error) {\n\tdata, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tc.Logger.Debugf(\"httpx: request body length: %d bytes\", len(data))\n\tif c.LogBody {\n\t\tc.Logger.Debugf(\"httpx: request body: %s\", string(data))\n\t}\n\trequest, err := c.newRequest(\n\t\tctx, method, resourcePath, query, bytes.NewReader(data))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif body != nil {\n\t\trequest.Header.Set(\"Content-Type\", \"application/json\")\n\t}\n\treturn request, nil\n}", "func MutateRequestBody(v RequestBody) *RequestBodyMutator {\n\treturn &RequestBodyMutator{\n\t\ttarget: v.(*requestBody),\n\t\tproxy: v.Clone().(*requestBody),\n\t}\n}", "func (ctx *HijackRequest) JSONBody() gjson.Result {\n\treturn gjson.Parse(ctx.Body())\n}", "func BlendJSON(source string, dest map[string]interface{}) error {\n\tsourceMap, err := JsonToMSI(source)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn Blend(sourceMap, dest)\n}", "func (c *ThreeScaleClient) buildUpdateJSONReq(ep string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(\"PUT\", c.adminPortal.rawURL+ep, body)\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Authorization\", \"Basic \"+basicAuth(\"\", c.credential))\n\treturn req, err\n}", "func (r *Request) ResetJSONBody() error {\n\tif r.Body == nil {\n\t\treturn nil\n\t}\n\treturn r.SetJSONBody(r.Obj)\n}", "func (input *BeegoInput) CopyBody(MaxMemory int64) []byte {\n\tif input.Context.Request.Body == nil {\n\t\treturn []byte{}\n\t}\n\n\tvar requestbody []byte\n\tsafe := &io.LimitedReader{R: input.Context.Request.Body, N: MaxMemory}\n\tif input.Header(\"Content-Encoding\") == \"gzip\" {\n\t\treader, err := gzip.NewReader(safe)\n\t\tif err != nil {\n\t\t\treturn nil\n\t\t}\n\t\trequestbody, _ = ioutil.ReadAll(reader)\n\t} else {\n\t\trequestbody, _ = ioutil.ReadAll(safe)\n\t}\n\n\tinput.Context.Request.Body.Close()\n\tbf := bytes.NewBuffer(requestbody)\n\tinput.Context.Request.Body = http.MaxBytesReader(input.Context.ResponseWriter, ioutil.NopCloser(bf), MaxMemory)\n\tinput.RequestBody = requestbody\n\treturn requestbody\n}", "func (recipe *Recipe) FromJSON(r io.Reader) error {\n\tdecoder := json.NewDecoder(r)\n\treturn decoder.Decode(recipe)\n}", "func (obj *JSONObject) Copy() JSONObject {\n\ttmp := obj.Export()\n\ttmp2 := JSONObject{}\n\ttmp2.ImportRaw(tmp)\n\treturn tmp2\n}", "func (e *ChefEnvironment) UpdateFromJSON(jsonEnv map[string]interface{}) util.Gerror {\n\tif e.Name != jsonEnv[\"name\"].(string) {\n\t\terr := util.Errorf(\"Environment name %s and %s from JSON do not match\", e.Name, jsonEnv[\"name\"].(string))\n\t\treturn err\n\t} else if e.Name == \"_default\" {\n\t\terr := util.Errorf(\"The '_default' environment cannot be modified.\")\n\t\terr.SetStatus(http.StatusMethodNotAllowed)\n\t\treturn err\n\t}\n\n\t/* Validations */\n\tvalidElements := []string{\"name\", \"chef_type\", \"json_class\", \"description\", \"default_attributes\", \"override_attributes\", \"cookbook_versions\"}\nValidElem:\n\tfor k := range jsonEnv {\n\t\tfor _, i := range validElements {\n\t\t\tif k == i {\n\t\t\t\tcontinue ValidElem\n\t\t\t}\n\t\t}\n\t\terr := util.Errorf(\"Invalid key %s in request body\", k)\n\t\treturn err\n\t}\n\n\tvar verr util.Gerror\n\n\tattrs := []string{\"default_attributes\", \"override_attributes\"}\n\tfor _, a := range attrs {\n\t\tjsonEnv[a], verr = util.ValidateAttributes(a, jsonEnv[a])\n\t\tif verr != nil {\n\t\t\treturn verr\n\t\t}\n\t}\n\n\tjsonEnv[\"json_class\"], verr = util.ValidateAsFieldString(jsonEnv[\"json_class\"])\n\tif verr != nil {\n\t\tif verr.Error() == \"Field 'name' nil\" {\n\t\t\tjsonEnv[\"json_class\"] = e.JSONClass\n\t\t} else {\n\t\t\treturn verr\n\t\t}\n\t} else {\n\t\tif jsonEnv[\"json_class\"].(string) != \"Chef::Environment\" {\n\t\t\tverr = util.Errorf(\"Field 'json_class' invalid\")\n\t\t\treturn verr\n\t\t}\n\t}\n\n\tjsonEnv[\"chef_type\"], verr = util.ValidateAsFieldString(jsonEnv[\"chef_type\"])\n\tif verr != nil {\n\t\tif verr.Error() == \"Field 'name' nil\" {\n\t\t\tjsonEnv[\"chef_type\"] = e.ChefType\n\t\t} else {\n\t\t\treturn verr\n\t\t}\n\t} else {\n\t\tif jsonEnv[\"chef_type\"].(string) != \"environment\" {\n\t\t\tverr = util.Errorf(\"Field 'chef_type' invalid\")\n\t\t\treturn verr\n\t\t}\n\t}\n\n\tjsonEnv[\"cookbook_versions\"], verr = util.ValidateAttributes(\"cookbook_versions\", jsonEnv[\"cookbook_versions\"])\n\tif verr != nil {\n\t\treturn verr\n\t}\n\tfor k, v := range jsonEnv[\"cookbook_versions\"].(map[string]interface{}) {\n\t\tif !util.ValidateEnvName(k) || k == \"\" {\n\t\t\tmerr := util.Errorf(\"Cookbook name %s invalid\", k)\n\t\t\tmerr.SetStatus(http.StatusBadRequest)\n\t\t\treturn merr\n\t\t}\n\n\t\tif v == nil {\n\t\t\tverr = util.Errorf(\"Invalid version number\")\n\t\t\treturn verr\n\t\t}\n\t\t_, verr = util.ValidateAsConstraint(v)\n\t\tif verr != nil {\n\t\t\t/* try validating as a version */\n\t\t\tv, verr = util.ValidateAsVersion(v)\n\t\t\tif verr != nil {\n\t\t\t\treturn verr\n\t\t\t}\n\t\t}\n\t}\n\n\tjsonEnv[\"description\"], verr = util.ValidateAsString(jsonEnv[\"description\"])\n\tif verr != nil {\n\t\tif verr.Error() == \"Field 'name' missing\" {\n\t\t\tjsonEnv[\"description\"] = \"\"\n\t\t} else {\n\t\t\treturn verr\n\t\t}\n\t}\n\n\te.ChefType = jsonEnv[\"chef_type\"].(string)\n\te.JSONClass = jsonEnv[\"json_class\"].(string)\n\te.Description = jsonEnv[\"description\"].(string)\n\te.Default = jsonEnv[\"default_attributes\"].(map[string]interface{})\n\te.Override = jsonEnv[\"override_attributes\"].(map[string]interface{})\n\t/* clear out, then loop over the cookbook versions */\n\te.CookbookVersions = make(map[string]string, len(jsonEnv[\"cookbook_versions\"].(map[string]interface{})))\n\tfor c, v := range jsonEnv[\"cookbook_versions\"].(map[string]interface{}) {\n\t\te.CookbookVersions[c] = v.(string)\n\t}\n\n\treturn nil\n}", "func RequestJSONBody(w http.ResponseWriter, r *http.Request, code int, payload interface{}) error {\n\tbuf := &bytes.Buffer{}\n\tenc := json.NewEncoder(buf)\n\tenc.SetEscapeHTML(true)\n\tif err := enc.Encode(payload); err != nil {\n\t\tw.Header().Set(\"X-Content-Type-Options\", \"nosniff\")\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn err\n\t}\n\tw.WriteHeader(code)\n\t_, err := w.Write(buf.Bytes())\n\tif err != nil {\n\t\tw.Header().Set(\"X-Content-Type-Options\", \"nosniff\")\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (o *BudgetAddParams) SetBody(body *models.BudgetAddRequest) {\n\to.Body = body\n}", "func ConversionRequestBody(imageLocation, desiredFormat string) ConversionRequest {\n\treturn ConversionRequest{\n\t\tImageLocation: imageLocation,\n\t\tDesiredFormat: desiredFormat,\n\t}\n}", "func (c *ThreeScaleClient) buildPatchJSONReq(ep string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(\"PATCH\", c.adminPortal.rawURL+ep, body)\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Authorization\", \"Basic \"+basicAuth(\"\", c.credential))\n\treturn req, err\n}", "func copyBody(r interface{}) (io.ReadCloser, error) {\n\tvar originalBody io.ReadCloser\n\tswitch r := r.(type) {\n\tcase *http.Request:\n\t\toriginalBody = r.Body\n\tcase *http.Response:\n\t\toriginalBody = r.Body\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unsupported type (%T) for copyBody\", r)\n\t}\n\n\tcontent, err := ioutil.ReadAll(originalBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tnewBody := ioutil.NopCloser(bytes.NewReader(content))\n\tswitch r := r.(type) {\n\tcase *http.Request:\n\t\tr.Body = newBody\n\tcase *http.Response:\n\t\tr.Body = newBody\n\t}\n\n\treturn ioutil.NopCloser(bytes.NewReader(content)), nil\n}", "func makeRequestBody(t interface{}) io.Reader {\n\tdebug := debug.Debug(\"oktad:makeRequestBody\")\n\tvar b bytes.Buffer\n\tenc := json.NewEncoder(&b)\n\terr := enc.Encode(t)\n\tif err != nil {\n\t\tdebug(\"Error encoding json! %s\", err)\n\t}\n\treturn &b\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.JSONBody != nil {\n\t\tif err := r.SetBodyParam(o.JSONBody); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param recipeId\n\tif err := r.SetPathParam(\"recipeId\", o.RecipeID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (s *Nap) BodyJSON(bodyJSON interface{}) *Nap {\n\tif bodyJSON == nil {\n\t\treturn s\n\t}\n\treturn s.BodyProvider(jsonBodyProvider{payload: bodyJSON})\n}", "func ReadJSONBody(writer http.ResponseWriter, request *http.Request, obj interface{}) error {\n\tb, err := ioutil.ReadAll(request.Body)\n\tif err != nil {\n\t\thttp.Error(writer, err.Error(), http.StatusInternalServerError)\n\t\treturn err\n\t}\n\terr = json.Unmarshal(b, obj)\n\tif err != nil {\n\t\thttp.Error(writer, err.Error(), http.StatusBadRequest)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (g *Github) PatchJSON(url, content string, v interface{}) error {\n\tbytes, err := g.patchBytes(url, content)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err := json.Unmarshal(bytes, v); err != nil {\n\t\treturn fmt.Errorf(\"could not parse json for url %s: %v\", url, err)\n\t}\n\treturn nil\n}", "func PostJSONWithBody(url string, params interface{}) (*http.Response, error) {\n\treader := new(bytes.Reader)\n\tif params != nil {\n\t\traw, err := json.Marshal(params)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treader = bytes.NewReader(raw)\n\t}\n\n\treturn http.Post(url, \"application/json; charset=utf-8\", reader)\n}", "func (r *StandardResponse) SetJSONBody(body interface{}) error {\n\tbodyBytes, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed JSON conversion: %s\", err.Error())\n\t}\n\n\tr.SetBody(bodyBytes)\n\treturn nil\n}", "func BindJSON(r *http.Request, i interface{}) error {\n\tb := &binder.JSON{}\n\treturn b.Bind(r.Body, i)\n}", "func (r ApiPatchOAuth2ClientRequest) JsonPatch(jsonPatch []JsonPatch) ApiPatchOAuth2ClientRequest {\n\tr.jsonPatch = &jsonPatch\n\treturn r\n}", "func (o *PostMenuItemParams) WithJSONBody(jSONBody *models.MenuItem) *PostMenuItemParams {\n\to.SetJSONBody(jSONBody)\n\treturn o\n}", "func ReadJSONFromBody(ctx echo.Context, value interface{}) error {\n\tdata, err := ioutil.ReadAll(ctx.Request().Body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(data) > 0 {\n\t\terr = json.Unmarshal(data, value)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func createAppJson(dm util.DepManager, appDir, appName, appJson string) error {\n\n\tupdatedJson, err := getAndUpdateAppJson(dm, appName, appJson)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = ioutil.WriteFile(filepath.Join(appDir, fileFlogoJson), []byte(updatedJson), 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func CreateAndMergeJsonPatch(original, override crv1alpha1.JSONMap) (crv1alpha1.JSONMap, error) {\n\t// Merge json specs with StrategicMerge\n\tmergedPatch, err := strategicMergeJsonPatch(original, override)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Convert merged json to map[string]interface{}\n\tvar merged map[string]interface{}\n\terr = json.Unmarshal(mergedPatch, &merged)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn merged, err\n}", "func (r *Request) PatchJSON(path string, data interface{}) {\n\tb, err := json.Marshal(data)\n\tif err != nil {\n\t\tr.t.Fatalf(\"httptesting: PatchJSON:json.Marshal(%T): %v\", data, err)\n\t}\n\n\tr.Patch(path, \"application/json\", b)\n}", "func PatchJSONWithParams(url string, v interface{}) (newreq *PatchJSONRequest, err error) {\n\tbs, err := json.Marshal(v)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treq, err := http.NewRequest(\"PATCH\", url, bytes.NewBuffer(bs))\n\tif err != nil {\n\t\treturn\n\t}\n\n\tnewreq = &PatchJSONRequest{req: req}\n\n\treturn\n}", "func prepareJSONPayload(rawReq *GenericRequest) (*bytes.Buffer, error) {\n\t// When payload ready, convert it to Json format\n\tbReqData, err := json.Marshal(&rawReq)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// @TODO Debug print marshal body\n\tfmt.Println(\"RAW Marshal BODY \" + string(bReqData))\n\n\t// Write json object to buffer\n\tbuffer := bytes.NewBuffer(bReqData)\n\n\treturn buffer, nil\n}", "func BindJSON(ctx *fasthttp.RequestCtx, d interface{}) error {\n\treturn json.Unmarshal(ctx.PostBody(), d)\n}", "func mapBody(match HTTPMatch, contentType string, body []byte) ([]byte, error) {\n\tnewBody := []byte{}\n\n\tif !isSameCaseInsensitive(contentType, JSON) {\n\t\treturn newBody, nil\n\t}\n\n\tvar parsed interface{}\n\terr := json.Unmarshal(body, &parsed)\n\tif err != nil {\n\t\treturn newBody, err\n\t}\n\n\tredacted := redact(match, parsed, \"$\")\n\n\tnewBody, err = json.Marshal(redacted)\n\tif err != nil {\n\t\treturn []byte{}, err\n\t}\n\n\treturn newBody, nil\n}", "func CreateRequestBody(target interface{}) (io.Reader, error) {\n\tbodyBytes, err := json.Marshal(target)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to marshal into json: %v\", err)\n\t}\n\n\treturn bytes.NewReader(bodyBytes), nil\n}", "func (o *PutMenuItemParams) WithJSONBody(jSONBody *models.MenuItem) *PutMenuItemParams {\n\to.SetJSONBody(jSONBody)\n\treturn o\n}", "func Copy(dest interface{}, src interface{}) error {\n\tdata, err := json.Marshal(src)\n\tif err != nil {\n\t\treturn errors.ErrorMarshal.Newm(err.Error())\n\t}\n\n\terr = json.Unmarshal(data, dest)\n\tif err != nil {\n\t\treturn errors.ErrorUnmarshal.Newm(err.Error())\n\t}\n\n\treturn nil\n}", "func (o *SavePreferencesParams) SetJSONBody(jSONBody *models.UpdateUserPreferences) {\n\to.JSONBody = jSONBody\n}", "func (s *ValidateService) BodyJson(body interface{}) *ValidateService {\n\ts.bodyJson = body\n\treturn s\n}", "func Copy(dst, src interface{}) interface{} {\n\tb, err := json.Marshal(src)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\terr = json.Unmarshal(b, dst)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn dst\n}", "func ToJSONBody(t *testing.T, body interface{}) *bytes.Buffer {\n\tjsonParams, err := json.Marshal(body)\n\tassert.NoErr(t, err)\n\treturn bytes.NewBuffer(jsonParams)\n}", "func setJSONData(req *http.Request, data interface{}) error {\n\tif data == nil {\n\t\treturn nil\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tbody, err := json.Marshal(data)\n\tif err != nil {\n\t\treturn err\n\t}\n\treq.Body = ioutil.NopCloser(bytes.NewReader(body))\n\treturn nil\n}", "func parseNDJSONRequestBody(body io.ReadCloser, structs []interface{}) {\n\tbuf := new(bytes.Buffer)\n\t_, err := buf.ReadFrom(body)\n\tExpect(err).ToNot(HaveOccurred())\n\n\trequestPayload := strings.TrimSuffix(buf.String(), \"\\n\") // _bulk requests need to end in a newline\n\tjsonPayloads := strings.Split(requestPayload, \"\\n\")\n\tExpect(jsonPayloads).To(HaveLen(len(structs)))\n\n\tfor i, s := range structs {\n\t\terr = json.Unmarshal([]byte(jsonPayloads[i]), s)\n\t\tExpect(err).ToNot(HaveOccurred())\n\t}\n}", "func GetBodyJSON(r *http.Request, v interface{}) error {\n\terr := json.NewDecoder(r.Body).Decode(&v)\n\treturn err\n}", "func (c *ThreeScaleClient) buildPostJSONReq(ep string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(\"POST\", c.adminPortal.rawURL+ep, body)\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Authorization\", \"Basic \"+basicAuth(\"\", c.credential))\n\treturn req, err\n}", "func (r *Recipe) MarshalJSON() ([]byte, error) {\n\tresp := struct {\n\t\tID int `json:\"id\"`\n\t\tName string `json:\"name\"`\n\t\tDescription string `json:\"description\"`\n\t\tPicture string `json:\"picture\"`\n\t\tCategory string `json:\"category\"`\n\t}{r.ID, r.Name, r.Description, r.Picture, r.Category}\n\tb, err := json.Marshal(resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tingredientsJSON := make([]*IngredientResponse, len(r.RecipesIngredients))\n\tfor i, v := range r.RecipesIngredients {\n\t\tingredientsJSON[i] = &IngredientResponse{Ingredient: v.Ingredient, Amount: v.Amount, Unit: v.Unit}\n\t}\n\tingredientsRaw, err := json.Marshal(ingredientsJSON)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tjsonSteps := fmt.Sprintf(`,\"steps\": %s, \"ingredients\": %v }`, r.Steps, string(ingredientsRaw))\n\treturn append(b[:len(b)-1], []byte(jsonSteps)...), nil\n}", "func WriteJSONBody(writer http.ResponseWriter, obj interface{}) {\n\twriter.Header().Add(\"Content-Type\", \"application/json\")\n\tb, err := json.Marshal(obj)\n\tif err != nil {\n\t\thttp.Error(writer, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\twriter.Write(b)\n}", "func copyBody(c io.ReadCloser) (*bytes.Reader, error) {\n\tif c == nil {\n\t\treturn nil, nil\n\t}\n\tdefer c.Close()\n\n\tvar b []byte\n\tbuf := bytes.NewBuffer(b)\n\t_, err := io.Copy(buf, c)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tr := bytes.NewReader(buf.Bytes())\n\treturn r, nil\n}", "func SendJSON(url string, body interface{}) error {\n\tclient := &http.Client{}\n\n\tb, err := json.Marshal(body)\n\tif err != nil {\n\t\tlog.Error(\"failed marshal request\", \"err\", err)\n\t\treturn err\n\t}\n\tj := bytes.NewReader(b)\n\n\treq, err := http.NewRequest(\"POST\", url, j)\n\tif err != nil {\n\t\treturn err\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\n\trespBody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif resp.StatusCode != 200 {\n\t\treturn badResponse(resp.StatusCode, respBody)\n\t}\n\n\treturn nil\n}", "func (bs *Bindings) Copy() (*Bindings, error) {\r\n\tbytes, err := json.Marshal(bs)\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\r\n\tret := NewBindings()\r\n\terr = json.Unmarshal(bytes, &ret)\r\n\r\n\treturn &ret, nil\r\n}", "func Copy(src interface{}, dst interface{}) error {\n\tif err := validateCopy(src, dst); err != nil {\n\t\treturn err\n\t}\n\tbytes, err := json.Marshal(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn json.Unmarshal(bytes, dst)\n}", "func AddJsonBookWithID(id string, bk Book) (Book, error) {\n\n\t// validate form values\n\tif bk.Isbn == \"\" || bk.Title == \"\" || bk.Author == \"\" || bk.Price == \"\" {\n\t\treturn bk, errors.New(\"400. Bad request. All fields must be complete.\")\n\t}\n ID := id\n\t//doc := make(map[string]interface{})\n\t//doc[\"Isbn\"] = bk.Isbn\n\t//doc[\"Title\"] = bk.Title\n\t//doc[\"Author\"] = bk.Author\n\t//doc[\"Price\"] = bk.Price\n\n ctx := context.Background()\n\t _, err := config.Client.Collection(\"books\").Doc(ID).Set(ctx,\n\t\tmap[string]interface{}{\n\t\t\t\"Isbn\": bk.Isbn,\n\t\t\t\"Title\": bk.Title,\n\t\t\t\"Author\": bk.Author,\n\t\t\t\"Price\": bk.Price,\n\t\t})\n\n\tif err != nil {\n\t\tlog.Fatalf(\"Failed to add a new book: %w\", err)\n\t\t//fmt.Errorf(\"Failed to iterate the list of requests: %w\", err)\n\n\t}\n\n\treturn bk, nil\n}", "func addBook(w http.ResponseWriter, r *http.Request) {\n\tvar book Book\n\t_= json.NewDecoder(r.Body).Decode(&book)\n\n\tbooks = append(books, book)\n\n\tw.WriteHeader(http.StatusCreated)\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tjson.NewEncoder(w).Encode(book)\n\t\n}", "func (a BodyWithAddPropsJSONBody) MarshalJSON() ([]byte, error) {\n\tvar err error\n\tobject := make(map[string]json.RawMessage)\n\n\tobject[\"inner\"], err = json.Marshal(a.Inner)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error marshaling 'inner': %w\", err)\n\t}\n\n\tobject[\"name\"], err = json.Marshal(a.Name)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error marshaling 'name': %w\", err)\n\t}\n\n\tfor fieldName, field := range a.AdditionalProperties {\n\t\tobject[fieldName], err = json.Marshal(field)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"error marshaling '%s': %w\", fieldName, err)\n\t\t}\n\t}\n\treturn json.Marshal(object)\n}", "func (a BodyWithAddPropsJSONBody) MarshalJSON() ([]byte, error) {\n\tvar err error\n\tobject := make(map[string]json.RawMessage)\n\n\tobject[\"inner\"], err = json.Marshal(a.Inner)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error marshaling 'inner': %w\", err)\n\t}\n\n\tobject[\"name\"], err = json.Marshal(a.Name)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error marshaling 'name': %w\", err)\n\t}\n\n\tfor fieldName, field := range a.AdditionalProperties {\n\t\tobject[fieldName], err = json.Marshal(field)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"error marshaling '%s': %w\", fieldName, err)\n\t\t}\n\t}\n\treturn json.Marshal(object)\n}", "func resetBody(req *http.Request, body []byte) {\n\tif req == nil || req.Body == nil {\n\t\treturn\n\t}\n\n\treq.Body = ioutil.NopCloser(bytes.NewReader(body))\n\n\t// do not modify existing GetBody function\n\tif req.GetBody == nil {\n\t\treq.GetBody = func() (io.ReadCloser, error) {\n\t\t\treturn ioutil.NopCloser(bytes.NewReader(body)), nil\n\t\t}\n\t}\n}", "func ReadJSONBody(request *http.Request, value interface{}) error {\n\tcontentType := request.Header.Get(\"Content-Type\")\n\tif !strings.Contains(contentType, \"application/json\") {\n\t\treturn fmt.Errorf(\"Invalid media type provided: %s\", contentType)\n\t}\n\tdecoder := json.NewDecoder(request.Body)\n\tif err := decoder.Decode(value); err != nil {\n\t\treturn fmt.Errorf(\"Failed to decode request body: %s\", err)\n\t}\n\treturn nil\n}", "func Body(data ...interface{}) AdditionalAttribute {\n return func(rb *Builder) error {\n rb.SetBody(data...)\n return nil\n }\n}", "func CreateJSONPatch(port C.int64_t, sourceC *C.char, sourceLen C.int, targetC *C.char, targetLen C.int) {\n\tsource := C.GoStringN(sourceC, sourceLen)\n\ttarget := C.GoStringN(targetC, targetLen)\n\n\tgo createJSONPatch(int64(port), source, target)\n}", "func (a *BodyWithAddPropsJSONBody) UnmarshalJSON(b []byte) error {\n\tobject := make(map[string]json.RawMessage)\n\terr := json.Unmarshal(b, &object)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif raw, found := object[\"inner\"]; found {\n\t\terr = json.Unmarshal(raw, &a.Inner)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error reading 'inner': %w\", err)\n\t\t}\n\t\tdelete(object, \"inner\")\n\t}\n\n\tif raw, found := object[\"name\"]; found {\n\t\terr = json.Unmarshal(raw, &a.Name)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error reading 'name': %w\", err)\n\t\t}\n\t\tdelete(object, \"name\")\n\t}\n\n\tif len(object) != 0 {\n\t\ta.AdditionalProperties = make(map[string]interface{})\n\t\tfor fieldName, fieldBuf := range object {\n\t\t\tvar fieldVal interface{}\n\t\t\terr := json.Unmarshal(fieldBuf, &fieldVal)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"error unmarshaling field %s: %w\", fieldName, err)\n\t\t\t}\n\t\t\ta.AdditionalProperties[fieldName] = fieldVal\n\t\t}\n\t}\n\treturn nil\n}", "func (a *BodyWithAddPropsJSONBody) UnmarshalJSON(b []byte) error {\n\tobject := make(map[string]json.RawMessage)\n\terr := json.Unmarshal(b, &object)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif raw, found := object[\"inner\"]; found {\n\t\terr = json.Unmarshal(raw, &a.Inner)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error reading 'inner': %w\", err)\n\t\t}\n\t\tdelete(object, \"inner\")\n\t}\n\n\tif raw, found := object[\"name\"]; found {\n\t\terr = json.Unmarshal(raw, &a.Name)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error reading 'name': %w\", err)\n\t\t}\n\t\tdelete(object, \"name\")\n\t}\n\n\tif len(object) != 0 {\n\t\ta.AdditionalProperties = make(map[string]interface{})\n\t\tfor fieldName, fieldBuf := range object {\n\t\t\tvar fieldVal interface{}\n\t\t\terr := json.Unmarshal(fieldBuf, &fieldVal)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"error unmarshaling field %s: %w\", fieldName, err)\n\t\t\t}\n\t\t\ta.AdditionalProperties[fieldName] = fieldVal\n\t\t}\n\t}\n\treturn nil\n}", "func (HTTPOperation) SetRequestBody(time time.Time, inputType api.InputTypeEnum, location int, numberAvailable int, numberTotal *int, tags *string, vaccine *int) error {\n\tbody.Date = time\n\tbody.InputType = inputType\n\tbody.Location = location\n\tbody.NumberAvailable = numberAvailable\n\tbody.NumberTotal = numberTotal\n\tbody.Tags = tags\n\tbody.Vaccine = vaccine\n\treturn nil\n}", "func JSONAddTblProduct(c *gin.Context) {\n\tDb, err := config.DbConnect()\n\tif err != nil {\n\t\tpanic(\"Not Connect database\")\n\t}\n\n\tdata, e := ioutil.ReadAll(c.Request.Body)\n\tif e != nil {\n\t\tc.JSON(http.StatusBadRequest, e.Error())\n\t}\n\tvar tabelproduct entities.TabelProduct\n\terr2 := json.Unmarshal(data, &tabelproduct)\n\tif err2 != nil {\n\t \tfmt.Println(err2)\n\t}\n\t\n\timagepath := tabelproduct.ImagePath\n\ttitle := tabelproduct.Title\n\tdescription := tabelproduct.Description\n\tprice := strconv.Itoa(tabelproduct.Price)\n\tmQuery := `INSERT INTO tabelproduct SET imagepath ='` + imagepath + `', title='` + title + `', description='` + description + `',\n\t\t\t\tprice='` + price + `';`\n\tupdDB, err := Db.Query(mQuery)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer updDB.Close()\n\n\tDb.Close()\n\tc.JSON(http.StatusOK, \"Add record successfully\")\n}", "func (c *RBController) SaveRecipe(w http.ResponseWriter, r *http.Request) (err error) {\n\t// Get properties\n\tname := r.PostFormValue(`name`)\n\tcuisine, err := strconv.Atoi(r.PostFormValue(`cuisine`))\n\n\t// Get the mealtype and season encoded ints\n\tmealtype := EncodeMealtype(r.Form[`mealtype`])\n\tseason := EncodeSeason(r.Form[`season`])\n\n\t// get everything else\n\tdescription := r.PostFormValue(`description`)\n\tingredients := r.PostFormValue(`ingredients`)\n\tinstructions := r.PostFormValue(`instructions`)\n\n\t// TODO better error handling\n\tif err != nil {\n\t\tfmt.Println(\"[WARNING] Something went wrong in SaveRecipe\")\n\t\tc.RenderError(w, 500, \"Sorry, something went wrong.\")\n\t\treturn\n\t}\n\n\t// everything OK: build the recipe, and send it to the database\n\trecipe := Recipe{ID: 0, Name: name, Cuisine: cuisine, Mealtype: mealtype,\n\t\tSeason: season, Description: description, Ingredientlist: ingredients,\n\t\tInstructions: instructions}\n\n\t// if we don't have the id string, then this is a new request.\n\tvars := mux.Vars(r)\n\tidStr := vars[\"id\"]\n\tid := 0\n\n\tif idStr != \"\" {\n\t\tid, _ = strconv.Atoi(idStr)\n\t\trecipe.ID = id\n\t\terr = c.RecipeDB.UpdateRecipe(&recipe)\n\t} else {\n\t\tid, err = c.RecipeDB.NewRecipe(&recipe)\n\t}\n\n\tif err == nil {\n\t\thttp.Redirect(w, r, \"/recipes/\"+fmt.Sprintf(\"%v\", id)+\"/\", http.StatusFound)\n\t}\n\treturn\n}", "func (req *Request) JSONBody() (url.Values, error) {\n\toutput := url.Values{}\n\n\tif req.ContentType() != ContentTypeJSON {\n\t\treturn output, nil\n\t}\n\n\tvars := map[string]interface{}{}\n\tif err := json.NewDecoder(req.Request.Body).Decode(&vars); err != nil && err != io.EOF {\n\t\treturn nil, err\n\t}\n\n\tfor k, v := range vars {\n\t\toutput.Set(k, fmt.Sprintf(\"%v\", v))\n\t}\n\n\treturn output, nil\n}", "func (o *PostApplyManifestParams) SetRequestBody(requestBody PostApplyManifestBody) {\n\to.RequestBody = requestBody\n}", "func CreatePatch(in *ecr.Repository, target *v1alpha1.RepositoryParameters) (*v1alpha1.RepositoryParameters, error) {\n\tcurrentParams := &v1alpha1.RepositoryParameters{}\n\tLateInitializeRepository(currentParams, in)\n\n\tjsonPatch, err := awsclients.CreateJSONPatch(currentParams, target)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tpatch := &v1alpha1.RepositoryParameters{}\n\tif err := json.Unmarshal(jsonPatch, patch); err != nil {\n\t\treturn nil, err\n\t}\n\treturn patch, nil\n}", "func decodeJSONBody(r *http.Request, dst interface{}) error {\n\tif r.Header.Get(\"Content-Type\") != \"\" {\n\t\tvalue, _ := header.ParseValueAndParams(r.Header, \"Content-Type\")\n\t\tif value != \"application/json\" {\n\t\t\treturn ContentHeaderError\n\t\t}\n\t}\n\n\tdec := json.NewDecoder(r.Body)\n\tdec.DisallowUnknownFields()\n\n\terr := dec.Decode(&dst)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif dec.More() {\n\t\tmsg := \"Request body must only contain a single JSON object\"\n\t\treturn errors.New(msg)\n\t}\n\n\treturn nil\n}", "func (l *List) PrependJSON(json interface{}) (err error) {\n\titem := MakeZeroValue(l.valType)\n\n\tif err := item.Set(json); err != nil {\n\t\treturn err\n\t}\n\n\treturn l.Prepend(item)\n}", "func (api *Client) CallJSON(opts *Opts, request interface{}, response interface{}) (resp *http.Response, err error) {\n\tvar requestBody []byte\n\t// Marshal the request if given\n\tif request != nil {\n\t\topts = opts.Copy()\n\t\trequestBody, err = json.Marshal(request)\n\t\topts.ContentType = \"application/json\"\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\t// Set the body up as a JSON object if required\n\t\tif opts.Body == nil {\n\t\t\topts.Body = bytes.NewBuffer(requestBody)\n\t\t}\n\t}\n\terrChan := make(chan error, 1)\n\tisMultipart := opts.MultipartMetadataName != \"\" && opts.Body != nil && request != nil\n\tif isMultipart {\n\t\tbodyReader, bodyWriter := io.Pipe()\n\t\twriter := multipart.NewWriter(bodyWriter)\n\t\topts.ContentType = writer.FormDataContentType()\n\t\tin := opts.Body\n\t\topts.Body = bodyReader\n\t\tgo func() {\n\t\t\tdefer func() { _ = bodyWriter.Close() }()\n\t\t\tvar err error\n\n\t\t\t// Create the first part\n\t\t\terr = writer.WriteField(opts.MultipartMetadataName, string(requestBody))\n\t\t\tif err != nil {\n\t\t\t\terrChan <- err\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// Add the file part\n\t\t\tpart, err := writer.CreateFormFile(opts.MultipartContentName, opts.MultipartFileName)\n\t\t\tif err != nil {\n\t\t\t\terrChan <- err\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// Copy it in\n\t\t\tif _, err := io.Copy(part, in); err != nil {\n\t\t\t\terrChan <- err\n\t\t\t\treturn\n\t\t\t}\n\t\t\terrChan <- writer.Close()\n\t\t}()\n\t}\n\tresp, err = api.Call(opts)\n\tif err != nil {\n\t\treturn resp, err\n\t}\n\tif response == nil || opts.NoResponse {\n\t\treturn resp, nil\n\t}\n\tif isMultipart {\n\t\terr = <-errChan\n\t\tif err != nil {\n\t\t\treturn resp, err\n\t\t}\n\t}\n\terr = DecodeJSON(resp, response)\n\treturn resp, err\n}", "func (o *CustomerGatewayUpdateOwnershipParams) SetBody(body *models.V1UpdateOwnershipRequest) {\n\to.Body = body\n}", "func (b *OperationMutator) RequestBody(v RequestBody) *OperationMutator {\n\tb.proxy.requestBody = v\n\treturn b\n}", "func (ctx *SimpleContext) RequestJsonBody(typ interface{}) error {\n\tcontentType := ctx.request.Header.Get(\"Content-Type\")\n\tif strings.Contains(contentType, \"json\") {\n\t\tbody, err := ioutil.ReadAll(ctx.request.Body)\n\t\tif nil != err {\n\t\t\treturn err\n\t\t}\n\t\terr = json.Unmarshal(body, typ)\n\t\tif nil != err {\n\t\t\treturn err\n\t\t}\n\t\treturn ctx.validate(typ)\n\t}\n\treturn errors.New(\"request body content-type is not contains json \")\n}", "func (spec *MachineSpec) Copy() *MachineSpec {\n\tvar specCopy MachineSpec\n\tp, err := json.Marshal(spec)\n\tif err != nil {\n\t\tpanic(\"internal error copying a MachineSpec: \" + err.Error())\n\t}\n\terr = json.Unmarshal(p, &specCopy)\n\tif err != nil {\n\t\tpanic(\"internal error copying a MachineSpec: \" + err.Error())\n\t}\n\treturn &specCopy\n}", "func (c *Client) newRequestJSON(method, urlStr string, v interface{}) (*http.Request, error) {\n\tb, err := json.Marshal(v)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq, err := c.newRequest(method, urlStr, bytes.NewReader(b))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treturn req, nil\n}", "func RequestBody(bodyReader io.Reader, data interface{}) error {\n\tbody, _ := ioutil.ReadAll(bodyReader)\n\terr := json.Unmarshal(body, data)\n\treturn err\n}", "func (r ApiApiProjectsIdDeploymentsPostRequest) Body(body RequestsCreateProjectDeploymentRequest) ApiApiProjectsIdDeploymentsPostRequest {\n\tr.body = &body\n\treturn r\n}", "func (r Recipe) MarshalJSON() ([]byte, error) {\n\n\t// specify the naming strategy here\n\treturn marshalJSON(\"SetJSONname\", r)\n}", "func Test_jsonpatch_Add_WithParents(t *testing.T) {\n\tg := NewWithT(t)\n\n\tpatch1, _ := DecodePatch([]byte(`\n[\n {\"op\":\"add\", \"path\":\"/level1\", \"value\":{}},\n {\"op\":\"add\", \"path\":\"/level1/level2\", \"value\":{}},\n {\"op\":\"add\", \"path\":\"/level1/level2/test_key\", \"value\":\"qwe\"}\n]\n`))\n\n\torigDoc := []byte(`{\"bar\":\"foo\"}`)\n\n\texpectNewDoc := []byte(`{\"bar\":\"foo\", \"level1\":{\"level2\":{\"test_key\":\"qwe\"}}}`)\n\n\tnewDoc, err := patch1.Apply(origDoc)\n\tg.Expect(err).ShouldNot(HaveOccurred(), \"patch apply\")\n\tg.Expect(JSONEqual(newDoc, expectNewDoc)).Should(BeTrue(), \"%v is not equal to %v\", string(newDoc), string(expectNewDoc))\n}", "func PatchJSON(left []byte, patch map[string]interface{}) ([]byte, error) {\n\tvLeft, err := parseFragment(left)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn json.Marshal(mergeRecursive(vLeft, patch))\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func RequireJSON(w http.ResponseWriter, r *http.Request, data interface{}) bool {\n\terr := json.NewDecoder(r.Body).Decode(data)\n\tif err != nil {\n\t\thttp.Error(w, \"request body is not valid JSON: \"+err.Error(), 400)\n\t\treturn false\n\t}\n\treturn true\n}", "func TestUpdateCategoryWrongJSONSyntax(t *testing.T) {\n\t//initial length of []products\n\tinitialLen := len(Categories)\n\t//parameters passed to request body\n\trequestBody := `{{\"CategoryID\":\"bq4fasj7jhfi127rimlg\",\"CategoryName\":\"Name\",,,}}`\n\treq, err := http.NewRequest(\"PATCH\", \"/categories/bq4fasj7jhfi127rimlg\", bytes.NewBufferString(requestBody))\n\treq = mux.SetURLVars(req, map[string]string{\"id\": \"bq4fasj7jhfi127rimlg\"})\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\trr := httptest.NewRecorder()\n\thandler := http.HandlerFunc(UpdateCategory)\n\n\thandler.ServeHTTP(rr, req)\n\n\tassert.Equal(t, 400, rr.Code, \"Bad request response is expected\")\n\tassert.Equal(t, initialLen, len(Categories), \"Expected length to stay the same after updating product\")\n}", "func Test_jsonpatch_Add_WithParents(t *testing.T) {\n\tg := NewWithT(t)\n\n\tpatch1, _ := jsonpatch.DecodePatch([]byte(`\n[\n {\"op\":\"add\", \"path\":\"/level1\", \"value\":{}},\n {\"op\":\"add\", \"path\":\"/level1/level2\", \"value\":{}},\n {\"op\":\"add\", \"path\":\"/level1/level2/test_key\", \"value\":\"qwe\"}\n]\n`))\n\n\torigDoc := []byte(`{\"bar\":\"foo\"}`)\n\n\texpectNewDoc := []byte(`{\"bar\":\"foo\", \"level1\":{\"level2\":{\"test_key\":\"qwe\"}}}`)\n\n\tnewDoc, err := patch1.Apply(origDoc)\n\tg.Expect(err).ShouldNot(HaveOccurred(), \"patch apply\")\n\tg.Expect(jsonpatch.Equal(newDoc, expectNewDoc)).Should(BeTrue(), \"%v is not equal to %v\", string(newDoc), string(expectNewDoc))\n}", "func (req *Request) UpdateFromJson(data []byte) error {\n\treturn json.Unmarshal(data, req)\n}", "func PatchJSON(object map[string]interface{},\n\toperation []map[string]interface{}) (map[string]interface{}, error) {\n\t// convert operation into JSON format\n\tpatchJson, err := json.Marshal(operation)\n\tif err != nil {\n\t\tlog.Error(fmt.Errorf(\"unable to convert patch operation to JSON: %+v\", err))\n\t\treturn map[string]interface{}{}, ErrInvalidPatch\n\t}\n\t// decode JSON patch operation\n\tpatch, err := jsonpatch.DecodePatch(patchJson)\n\tif err != nil {\n\t\tlog.Error(fmt.Errorf(\"unable to parse Json Patch operation: %+v\", err))\n\t\treturn map[string]interface{}{}, ErrInvalidPatch\n\t}\n\n\t// convert metadata to json\n\tvar metaJson []byte\n\tmetaJson, err = json.Marshal(object)\n\tif err != nil {\n\t\tlog.Error(fmt.Errorf(\"unable to convert object to JSON: %+v\", err))\n\t\treturn map[string]interface{}{}, ErrInvalidJSON\n\t}\n\n\t// apply JSON patch operation\n\tmodified, err := patch.Apply(metaJson)\n\tif err != nil {\n\t\tlog.Error(fmt.Errorf(\"unable to apply JSON patch: %+v\", err))\n\t\treturn map[string]interface{}{}, ErrInvalidPatch\n\t}\n\n\tlog.Debug(fmt.Sprintf(\"successfully applied JSON patch to object: %s\", modified))\n\t// convert final JSON string back to interface\n\tvar meta map[string]interface{}\n\tif err := json.Unmarshal(modified, &meta); err != nil {\n\t\treturn meta, ErrInvalidJSON\n\t}\n\treturn meta, nil\n}", "func encodeBody(obj any) (io.Reader, error) {\n\tbuf := bytes.NewBuffer(nil)\n\tenc := json.NewEncoder(buf)\n\tif err := enc.Encode(obj); err != nil {\n\t\treturn nil, err\n\t}\n\treturn buf, nil\n}", "func (o *UpdateBuildPropertiesParams) SetBody(body *models.JSONPatchDocument) {\n\to.Body = body\n}", "func generateJSONPatches(jsonPatches []clusterv1.JSONPatch, variables map[string]apiextensionsv1.JSON) ([]byte, error) {\n\tres := []jsonPatchRFC6902{}\n\n\tfor _, jsonPatch := range jsonPatches {\n\t\tvar value *apiextensionsv1.JSON\n\t\tif jsonPatch.Op == \"add\" || jsonPatch.Op == \"replace\" {\n\t\t\tvar err error\n\t\t\tvalue, err = calculateValue(jsonPatch, variables)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\n\t\tres = append(res, jsonPatchRFC6902{\n\t\t\tOp: jsonPatch.Op,\n\t\t\tPath: jsonPatch.Path,\n\t\t\tValue: value,\n\t\t})\n\t}\n\n\t// Render JSON Patches.\n\tresJSON, err := json.Marshal(res)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to marshal JSON Patch %v\", jsonPatches)\n\t}\n\n\treturn resJSON, nil\n}", "func getSecretPatchBody(svc string, nv config.NameValue) types.Patch {\n\tph := getSuperSecretTemplate(svc)\n\tph.Data = map[string]string{\n\t\tnv.Name: fmt.Sprintf(`(( index (ds \"data\") \"%s\" ))`, nv.Name),\n\t}\n\tphb, _ := yaml.Marshal(ph)\n\tp1 := types.Patch{\n\t\tPatch: strings.Replace(string(phb), \": |\", \": |-\", -1), // when load again the |- will disappaer and the value will be on one line\n\t\tTarget: getSelector(\"SuperSecret\", svc),\n\t}\n\treturn p1\n}" ]
[ "0.63932765", "0.5612401", "0.5401041", "0.53624415", "0.528843", "0.51485735", "0.51017815", "0.5067448", "0.5022532", "0.49914533", "0.49796784", "0.4972427", "0.4907961", "0.48759767", "0.48699957", "0.48650312", "0.48458964", "0.48397836", "0.48314732", "0.48046854", "0.48015723", "0.47635266", "0.4754551", "0.47458997", "0.47440606", "0.47380224", "0.47324312", "0.47265822", "0.47112268", "0.4702513", "0.4698319", "0.4694028", "0.46796072", "0.46758914", "0.46708143", "0.46648023", "0.46631253", "0.46562946", "0.465385", "0.46012995", "0.4590407", "0.45823577", "0.45755932", "0.4552863", "0.4548016", "0.45391262", "0.4535911", "0.45325446", "0.4528459", "0.45183516", "0.45145702", "0.4492726", "0.44904932", "0.4472443", "0.44640148", "0.44596308", "0.44591406", "0.4450284", "0.44463122", "0.44326714", "0.44223806", "0.4421589", "0.44209316", "0.43994793", "0.43994793", "0.43994352", "0.43991417", "0.43988067", "0.43954444", "0.43829498", "0.43829498", "0.43812072", "0.43581843", "0.4351877", "0.4350641", "0.4347879", "0.4339248", "0.43331987", "0.43229958", "0.4322255", "0.43204024", "0.4317967", "0.43156585", "0.4313201", "0.43077868", "0.4297466", "0.42968974", "0.4296016", "0.42957562", "0.4290836", "0.4281028", "0.42785916", "0.4278322", "0.4275504", "0.4270869", "0.42697158", "0.42661643", "0.4265445", "0.4262461", "0.42621738" ]
0.7077129
0
SetJSONBody adds the jsonBody to the copy recipe to my recipes with changes params
func (o *CopyRecipeToMyRecipesWithChangesParams) SetJSONBody(jSONBody *models.Recipe) { o.JSONBody = jSONBody }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *CopyRecipeToMyRecipesWithChangesParams) WithJSONBody(jSONBody *models.Recipe) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetJSONBody(jSONBody)\n\treturn o\n}", "func (r *Request) SetJSONBody(val interface{}) error {\n\tbuf := bytes.NewBuffer(nil)\n\tenc := json.NewEncoder(buf)\n\tif err := enc.Encode(val); err != nil {\n\t\treturn err\n\t}\n\n\tr.Obj = val\n\tr.Body = buf\n\tr.BodySize = int64(buf.Len())\n\treturn nil\n}", "func (o *PostMenuItemParams) SetJSONBody(jSONBody *models.MenuItem) {\n\to.JSONBody = jSONBody\n}", "func (r *StandardResponse) SetJSONBody(body interface{}) error {\n\tbodyBytes, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed JSON conversion: %s\", err.Error())\n\t}\n\n\tr.SetBody(bodyBytes)\n\treturn nil\n}", "func (o *PutMenuItemParams) SetJSONBody(jSONBody *models.MenuItem) {\n\to.JSONBody = jSONBody\n}", "func (a *API) JSONBody(ctx *fasthttp.RequestCtx, model interface{}) {\n\tr := bytes.NewReader(ctx.PostBody())\n\tjson.NewDecoder(r).Decode(&model)\n}", "func (HTTPOperation) SetRequestBody(time time.Time, inputType api.InputTypeEnum, location int, numberAvailable int, numberTotal *int, tags *string, vaccine *int) error {\n\tbody.Date = time\n\tbody.InputType = inputType\n\tbody.Location = location\n\tbody.NumberAvailable = numberAvailable\n\tbody.NumberTotal = numberTotal\n\tbody.Tags = tags\n\tbody.Vaccine = vaccine\n\treturn nil\n}", "func (r *Request) ResetJSONBody() error {\n\tif r.Body == nil {\n\t\treturn nil\n\t}\n\treturn r.SetJSONBody(r.Obj)\n}", "func (o *PostApplyManifestParams) SetRequestBody(requestBody PostApplyManifestBody) {\n\to.RequestBody = requestBody\n}", "func (o *SavePreferencesParams) SetJSONBody(jSONBody *models.UpdateUserPreferences) {\n\to.JSONBody = jSONBody\n}", "func NewJSONBody(v interface{}) (contentType string, bodyReader io.Reader, err error) {\n\tb, err := json.Marshal(v)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn \"application/json;charset=utf-8\", bytes.NewReader(b), nil\n}", "func setJSONData(req *http.Request, data interface{}) error {\n\tif data == nil {\n\t\treturn nil\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tbody, err := json.Marshal(data)\n\tif err != nil {\n\t\treturn err\n\t}\n\treq.Body = ioutil.NopCloser(bytes.NewReader(body))\n\treturn nil\n}", "func (ctx *HijackRequest) SetBody(obj interface{}) *HijackRequest {\n\t// reset to empty\n\tctx.req.StringBody(\"\")\n\tctx.req.JSONBody(nil)\n\n\tswitch body := obj.(type) {\n\tcase []byte:\n\t\tbuf := bytes.NewBuffer(body)\n\t\tctx.req.Body(buf)\n\tcase string:\n\t\tctx.req.StringBody(body)\n\tdefault:\n\t\tctx.req.JSONBody(obj)\n\t}\n\treturn ctx\n}", "func (o *BudgetAddParams) SetBody(body *models.BudgetAddRequest) {\n\to.Body = body\n}", "func (o *UpdateBuildPropertiesParams) SetBody(body *models.JSONPatchDocument) {\n\to.Body = body\n}", "func (o *CustomerGatewayUpdateOwnershipParams) SetBody(body *models.V1UpdateOwnershipRequest) {\n\to.Body = body\n}", "func BindJSON(r *http.Request, target interface{}) error {\n\tdata, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = json.Unmarshal(data, target)\n\n\treturn err\n}", "func (r *Request) SetBody(reader io.Reader) {\n\tr.request.Body = ioutil.NopCloser(reader)\n}", "func MutateRequestBody(v RequestBody) *RequestBodyMutator {\n\treturn &RequestBodyMutator{\n\t\ttarget: v.(*requestBody),\n\t\tproxy: v.Clone().(*requestBody),\n\t}\n}", "func (o *CreateIscsiLunSnapshotParams) SetRequestBody(requestBody []*models.IscsiLunSnapshotCreationParams) {\n\to.RequestBody = requestBody\n}", "func (zr *ZRequest) SetBody(body interface{}) *ZRequest {\n\tif zr.ended {\n\t\treturn zr\n\t}\n\tzr.body = body\n\treturn zr\n}", "func (input *BeegoInput) CopyBody(MaxMemory int64) []byte {\n\tif input.Context.Request.Body == nil {\n\t\treturn []byte{}\n\t}\n\n\tvar requestbody []byte\n\tsafe := &io.LimitedReader{R: input.Context.Request.Body, N: MaxMemory}\n\tif input.Header(\"Content-Encoding\") == \"gzip\" {\n\t\treader, err := gzip.NewReader(safe)\n\t\tif err != nil {\n\t\t\treturn nil\n\t\t}\n\t\trequestbody, _ = ioutil.ReadAll(reader)\n\t} else {\n\t\trequestbody, _ = ioutil.ReadAll(safe)\n\t}\n\n\tinput.Context.Request.Body.Close()\n\tbf := bytes.NewBuffer(requestbody)\n\tinput.Context.Request.Body = http.MaxBytesReader(input.Context.ResponseWriter, ioutil.NopCloser(bf), MaxMemory)\n\tinput.RequestBody = requestbody\n\treturn requestbody\n}", "func ReadJSONBody(writer http.ResponseWriter, request *http.Request, obj interface{}) error {\n\tb, err := ioutil.ReadAll(request.Body)\n\tif err != nil {\n\t\thttp.Error(writer, err.Error(), http.StatusInternalServerError)\n\t\treturn err\n\t}\n\terr = json.Unmarshal(b, obj)\n\tif err != nil {\n\t\thttp.Error(writer, err.Error(), http.StatusBadRequest)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (req *Request) SetBody(body []byte) {\n\treq.Req.Body = ioutil.NopCloser(bytes.NewReader(body))\n}", "func makeRequestBody(t interface{}) io.Reader {\n\tdebug := debug.Debug(\"oktad:makeRequestBody\")\n\tvar b bytes.Buffer\n\tenc := json.NewEncoder(&b)\n\terr := enc.Encode(t)\n\tif err != nil {\n\t\tdebug(\"Error encoding json! %s\", err)\n\t}\n\treturn &b\n}", "func (ctx *HijackResponse) SetBody(obj interface{}) *HijackResponse {\n\tswitch body := obj.(type) {\n\tcase []byte:\n\t\tctx.payload.Body = body\n\tcase string:\n\t\tctx.payload.Body = []byte(body)\n\tdefault:\n\t\tctx.SetHeader(\"Content-Type\", \"application/json; charset=utf-8\")\n\t\tvar err error\n\t\tctx.payload.Body, err = json.Marshal(obj)\n\t\tutils.E(err)\n\t}\n\treturn ctx\n}", "func (c *ThreeScaleClient) buildUpdateJSONReq(ep string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(\"PUT\", c.adminPortal.rawURL+ep, body)\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Authorization\", \"Basic \"+basicAuth(\"\", c.credential))\n\treturn req, err\n}", "func (s *Nap) BodyJSON(bodyJSON interface{}) *Nap {\n\tif bodyJSON == nil {\n\t\treturn s\n\t}\n\treturn s.BodyProvider(jsonBodyProvider{payload: bodyJSON})\n}", "func (e *ChefEnvironment) UpdateFromJSON(jsonEnv map[string]interface{}) util.Gerror {\n\tif e.Name != jsonEnv[\"name\"].(string) {\n\t\terr := util.Errorf(\"Environment name %s and %s from JSON do not match\", e.Name, jsonEnv[\"name\"].(string))\n\t\treturn err\n\t} else if e.Name == \"_default\" {\n\t\terr := util.Errorf(\"The '_default' environment cannot be modified.\")\n\t\terr.SetStatus(http.StatusMethodNotAllowed)\n\t\treturn err\n\t}\n\n\t/* Validations */\n\tvalidElements := []string{\"name\", \"chef_type\", \"json_class\", \"description\", \"default_attributes\", \"override_attributes\", \"cookbook_versions\"}\nValidElem:\n\tfor k := range jsonEnv {\n\t\tfor _, i := range validElements {\n\t\t\tif k == i {\n\t\t\t\tcontinue ValidElem\n\t\t\t}\n\t\t}\n\t\terr := util.Errorf(\"Invalid key %s in request body\", k)\n\t\treturn err\n\t}\n\n\tvar verr util.Gerror\n\n\tattrs := []string{\"default_attributes\", \"override_attributes\"}\n\tfor _, a := range attrs {\n\t\tjsonEnv[a], verr = util.ValidateAttributes(a, jsonEnv[a])\n\t\tif verr != nil {\n\t\t\treturn verr\n\t\t}\n\t}\n\n\tjsonEnv[\"json_class\"], verr = util.ValidateAsFieldString(jsonEnv[\"json_class\"])\n\tif verr != nil {\n\t\tif verr.Error() == \"Field 'name' nil\" {\n\t\t\tjsonEnv[\"json_class\"] = e.JSONClass\n\t\t} else {\n\t\t\treturn verr\n\t\t}\n\t} else {\n\t\tif jsonEnv[\"json_class\"].(string) != \"Chef::Environment\" {\n\t\t\tverr = util.Errorf(\"Field 'json_class' invalid\")\n\t\t\treturn verr\n\t\t}\n\t}\n\n\tjsonEnv[\"chef_type\"], verr = util.ValidateAsFieldString(jsonEnv[\"chef_type\"])\n\tif verr != nil {\n\t\tif verr.Error() == \"Field 'name' nil\" {\n\t\t\tjsonEnv[\"chef_type\"] = e.ChefType\n\t\t} else {\n\t\t\treturn verr\n\t\t}\n\t} else {\n\t\tif jsonEnv[\"chef_type\"].(string) != \"environment\" {\n\t\t\tverr = util.Errorf(\"Field 'chef_type' invalid\")\n\t\t\treturn verr\n\t\t}\n\t}\n\n\tjsonEnv[\"cookbook_versions\"], verr = util.ValidateAttributes(\"cookbook_versions\", jsonEnv[\"cookbook_versions\"])\n\tif verr != nil {\n\t\treturn verr\n\t}\n\tfor k, v := range jsonEnv[\"cookbook_versions\"].(map[string]interface{}) {\n\t\tif !util.ValidateEnvName(k) || k == \"\" {\n\t\t\tmerr := util.Errorf(\"Cookbook name %s invalid\", k)\n\t\t\tmerr.SetStatus(http.StatusBadRequest)\n\t\t\treturn merr\n\t\t}\n\n\t\tif v == nil {\n\t\t\tverr = util.Errorf(\"Invalid version number\")\n\t\t\treturn verr\n\t\t}\n\t\t_, verr = util.ValidateAsConstraint(v)\n\t\tif verr != nil {\n\t\t\t/* try validating as a version */\n\t\t\tv, verr = util.ValidateAsVersion(v)\n\t\t\tif verr != nil {\n\t\t\t\treturn verr\n\t\t\t}\n\t\t}\n\t}\n\n\tjsonEnv[\"description\"], verr = util.ValidateAsString(jsonEnv[\"description\"])\n\tif verr != nil {\n\t\tif verr.Error() == \"Field 'name' missing\" {\n\t\t\tjsonEnv[\"description\"] = \"\"\n\t\t} else {\n\t\t\treturn verr\n\t\t}\n\t}\n\n\te.ChefType = jsonEnv[\"chef_type\"].(string)\n\te.JSONClass = jsonEnv[\"json_class\"].(string)\n\te.Description = jsonEnv[\"description\"].(string)\n\te.Default = jsonEnv[\"default_attributes\"].(map[string]interface{})\n\te.Override = jsonEnv[\"override_attributes\"].(map[string]interface{})\n\t/* clear out, then loop over the cookbook versions */\n\te.CookbookVersions = make(map[string]string, len(jsonEnv[\"cookbook_versions\"].(map[string]interface{})))\n\tfor c, v := range jsonEnv[\"cookbook_versions\"].(map[string]interface{}) {\n\t\te.CookbookVersions[c] = v.(string)\n\t}\n\n\treturn nil\n}", "func (o *CheckTransactionCreditLimitParams) SetBody(body *platformclientmodels.CreditRequest) {\n\to.Body = body\n}", "func (c *baseClient) SetBody(params interface{}) *baseClient {\n\tc.body = params\n\treturn c\n}", "func (c *RBController) RecipeJSON(w http.ResponseWriter, r *http.Request) (err error) {\n\tvars := mux.Vars(r)\n\tid, _ := strconv.Atoi(vars[\"id\"])\n\trecipe, err := c.GetRecipe(id)\n\tif err == nil {\n\t\tc.JSON(w, http.StatusOK, recipe)\n\t} else if err == sql.ErrNoRows {\n\t\tc.RenderError(w, 404, \"Sorry, your page wasn't found\")\n\t\terr = nil\n\t}\n\treturn\n}", "func RequestJSONBody(w http.ResponseWriter, r *http.Request, code int, payload interface{}) error {\n\tbuf := &bytes.Buffer{}\n\tenc := json.NewEncoder(buf)\n\tenc.SetEscapeHTML(true)\n\tif err := enc.Encode(payload); err != nil {\n\t\tw.Header().Set(\"X-Content-Type-Options\", \"nosniff\")\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn err\n\t}\n\tw.WriteHeader(code)\n\t_, err := w.Write(buf.Bytes())\n\tif err != nil {\n\t\tw.Header().Set(\"X-Content-Type-Options\", \"nosniff\")\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn err\n\t}\n\treturn nil\n}", "func TestAddRecipe(t *testing.T) {\n\tpayload := fmt.Sprintf(`\n {\n \"mealtype\": \"Breakfast\",\n \"name\": \"Pancakes\",\n \"Ingredients\": [ \"150g all purpose flour\",\n \t\t\t\t \"150ml of milk\"],\n \"preparation\": \"Add all ingredients and mix. Put in Pan.\"\n}`)\n\n\tresponse, err := http.Post(baseURL+\"/recipes\", \"application/json\", strings.NewReader(payload))\n\tif err != nil {\n\t\tt.Fatalf(\"failed to get json, %s\", err)\n\t}\n\n\tcheckResponseCode(t, http.StatusOK, response.StatusCode)\n}", "func (c *RBController) RecipeJSONAdvanced(w http.ResponseWriter, r *http.Request) (err error) {\n\tr.ParseForm()\n\tstrict, err := strconv.Atoi(r.PostFormValue(\"strict\"))\n\tname := r.PostFormValue(\"name\")\n\tcuisine, _ := strconv.Atoi(r.PostFormValue(\"cuisine\"))\n\tseason, _ := strconv.Atoi(r.PostFormValue(\"season\"))\n\tmealtype, _ := strconv.Atoi(r.PostFormValue(\"mealtype\"))\n\n\t// get all the recipes that match\n\tvar recipes *list.List\n\tif strict == 0 {\n\t\trecipes, err = c.GetRecipesLoose(name, cuisine, mealtype, season)\n\t} else {\n\t\trecipes, err = c.GetRecipesStrict(name, cuisine, mealtype, season)\n\t}\n\n\t// slice of jsons\n\tjsons := make([]string, recipes.Len())\n\n\tif err == nil {\n\t\tindex := 0\n\t\tfor e := recipes.Front(); e != nil; e = e.Next() {\n\t\t\trec := e.Value.(*Recipe)\n\t\t\tjsons[index] = rec.ToJSON()\n\t\t\tindex++\n\t\t}\n\t\trequest := strings.Join(jsons, \"\\n\")\n\t\tfmt.Fprintf(w, request)\n\t} else {\n\t\tfmt.Fprintf(w, \"%v\", err.Error())\n\t}\n\treturn\n}", "func (o *UpdateCredentialParams) SetBody(body dynatrace.Credentials) {\n\to.Body = body\n}", "func (o *CreateBlueprintInWorkspaceInternalParams) SetBody(body *model.BlueprintV4Request) {\n\to.Body = body\n}", "func CreateRequestBody(target interface{}) (io.Reader, error) {\n\tbodyBytes, err := json.Marshal(target)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to marshal into json: %v\", err)\n\t}\n\n\treturn bytes.NewReader(bodyBytes), nil\n}", "func (ctx *HijackRequest) JSONBody() gjson.Result {\n\treturn gjson.Parse(ctx.Body())\n}", "func WriteJSONBody(writer http.ResponseWriter, obj interface{}) {\n\twriter.Header().Add(\"Content-Type\", \"application/json\")\n\tb, err := json.Marshal(obj)\n\tif err != nil {\n\t\thttp.Error(writer, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\twriter.Write(b)\n}", "func (s *ValidateService) BodyJson(body interface{}) *ValidateService {\n\ts.bodyJson = body\n\treturn s\n}", "func (o *ServiceBindingBindingParams) SetBody(body *models.ServiceBindingRequest) {\n\to.Body = body\n}", "func (o *IndicatorCreateV1Params) SetBody(body *models.APIIndicatorCreateReqsV1) {\n\to.Body = body\n}", "func ReadJSONFromBody(ctx echo.Context, value interface{}) error {\n\tdata, err := ioutil.ReadAll(ctx.Request().Body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(data) > 0 {\n\t\terr = json.Unmarshal(data, value)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o *FetchIntegrationFormParams) SetBody(body map[string]string) {\n\to.Body = body\n}", "func mapBody(match HTTPMatch, contentType string, body []byte) ([]byte, error) {\n\tnewBody := []byte{}\n\n\tif !isSameCaseInsensitive(contentType, JSON) {\n\t\treturn newBody, nil\n\t}\n\n\tvar parsed interface{}\n\terr := json.Unmarshal(body, &parsed)\n\tif err != nil {\n\t\treturn newBody, err\n\t}\n\n\tredacted := redact(match, parsed, \"$\")\n\n\tnewBody, err = json.Marshal(redacted)\n\tif err != nil {\n\t\treturn []byte{}, err\n\t}\n\n\treturn newBody, nil\n}", "func (o *CreateRepoNotificationParams) SetBody(body *models.NotificationCreateRequest) {\n\to.Body = body\n}", "func (c *ThreeScaleClient) buildPatchJSONReq(ep string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(\"PATCH\", c.adminPortal.rawURL+ep, body)\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Authorization\", \"Basic \"+basicAuth(\"\", c.credential))\n\treturn req, err\n}", "func resetBody(req *http.Request, body []byte) {\n\tif req == nil || req.Body == nil {\n\t\treturn\n\t}\n\n\treq.Body = ioutil.NopCloser(bytes.NewReader(body))\n\n\t// do not modify existing GetBody function\n\tif req.GetBody == nil {\n\t\treq.GetBody = func() (io.ReadCloser, error) {\n\t\t\treturn ioutil.NopCloser(bytes.NewReader(body)), nil\n\t\t}\n\t}\n}", "func (options *DeleteNotificationChannelsOptions) SetRequestBody(requestBody []string) *DeleteNotificationChannelsOptions {\n\toptions.RequestBody = requestBody\n\treturn options\n}", "func RequestBody(bodyReader io.Reader, data interface{}) error {\n\tbody, _ := ioutil.ReadAll(bodyReader)\n\terr := json.Unmarshal(body, data)\n\treturn err\n}", "func prepareJSONPayload(rawReq *GenericRequest) (*bytes.Buffer, error) {\n\t// When payload ready, convert it to Json format\n\tbReqData, err := json.Marshal(&rawReq)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// @TODO Debug print marshal body\n\tfmt.Println(\"RAW Marshal BODY \" + string(bReqData))\n\n\t// Write json object to buffer\n\tbuffer := bytes.NewBuffer(bReqData)\n\n\treturn buffer, nil\n}", "func BindJSON(ctx *fasthttp.RequestCtx, d interface{}) error {\n\treturn json.Unmarshal(ctx.PostBody(), d)\n}", "func (r *Request) SetBody(b []byte) {\n\tr.body = b\n}", "func (o *PublicInteractiveLoginCredentialParams) SetBody(body *model.CredentialRequest) {\n\to.Body = body\n}", "func BindJSON(r *http.Request, i interface{}) error {\n\tb := &binder.JSON{}\n\treturn b.Bind(r.Body, i)\n}", "func (o *PostMenuItemParams) WithJSONBody(jSONBody *models.MenuItem) *PostMenuItemParams {\n\to.SetJSONBody(jSONBody)\n\treturn o\n}", "func (o *SearchWorkspacesParams) SetBody(body *models.RestSearchWorkspaceRequest) {\n\to.Body = body\n}", "func (o *GetUserRoleNextsParams) SetRequestBody(requestBody *models.GetUserRoleNextsRequestBody) {\n\to.RequestBody = requestBody\n}", "func (o *PatchAddonParams) SetBody(body *models.Addon) {\n\to.Body = body\n}", "func ReadJSONBody(request *http.Request, value interface{}) error {\n\tcontentType := request.Header.Get(\"Content-Type\")\n\tif !strings.Contains(contentType, \"application/json\") {\n\t\treturn fmt.Errorf(\"Invalid media type provided: %s\", contentType)\n\t}\n\tdecoder := json.NewDecoder(request.Body)\n\tif err := decoder.Decode(value); err != nil {\n\t\treturn fmt.Errorf(\"Failed to decode request body: %s\", err)\n\t}\n\treturn nil\n}", "func (o *PatchAssetDeviceConfigurationsMoidParams) SetBody(body *models.AssetDeviceConfiguration) {\n\to.Body = body\n}", "func (api *FoodRecipeAPI) partialRecipeUpdate(w http.ResponseWriter, req *http.Request) {\n\tdefer DrainBody(req)\n\tctx := req.Context()\n\n\tvars := mux.Vars(req)\n\tid := vars[\"id\"]\n\tlogData := log.Data{\"id\": id}\n\n\tvar errorObjects []*models.ErrorObject\n\n\tpatchJSON, recipePatches, err := patch.Get(ctx, req.Body)\n\tif err != nil {\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// Validate patch request\n\tfor i, recipePatch := range *recipePatches {\n\t\tif err = recipePatch.Validate(nil); err != nil {\n\t\t\tif _, ok := err.(*validator.InvalidValidationError); ok {\n\t\t\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\t\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tfor _, err := range err.(validator.ValidationErrors) {\n\t\t\t\terrorObjects = append(errorObjects, models.HandleValidationErrors(strconv.Itoa(i), err.ActualTag(), err.StructField(), err.Value().(string), err.Param()))\n\t\t\t}\n\t\t}\n\t}\n\tif len(errorObjects) > 0 {\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// apply patch against recipe resource\n\tp, err := jsonpatch.DecodePatch(patchJSON)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: unable to decode patch\", err)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// find current recipe doc\n\tvar recipe models.Recipe\n\n\tcollection := api.MongoClient.Database(\"food-recipes\").Collection(\"recipes\")\n\tif err = collection.FindOne(ctx, bson.M{\"_id\": id}).Decode(&recipe); err != nil {\n\t\tif err == mongo.ErrNoDocuments {\n\t\t\tlog.Warn(ctx, \"patch recipe: failed to find recipe\", log.FormatErrors([]error{err}), logData)\n\t\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrRecipeNotFound.Error()})\n\t\t\tErrorResponse(ctx, w, http.StatusNotFound, &models.ErrorResponse{Errors: errorObjects})\n\t\t\treturn\n\t\t}\n\n\t\tlog.Error(ctx, \"patch recipe: failed to find recipe, bad connection?\", err)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\tb, err := json.Marshal(recipe)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: error returned from json marshal\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// apply patch to existing recipe\n\tmodified, err := p.Apply(b)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: unable to apply patch to recipe\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\terr = json.Unmarshal(modified, &recipe)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: unmarshal modified recipe into recipe struct\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// store new recipe\n\tif _, err = collection.ReplaceOne(ctx, bson.M{\"_id\": id}, recipe); err != nil {\n\t\tif err == mongo.ErrNoDocuments {\n\t\t\tlog.Error(ctx, \"update recipe: failed to update recipe, recipe deos not exists\", err, logData)\n\t\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrRecipeNotFound.Error()})\n\t\t\tErrorResponse(ctx, w, http.StatusNotFound, &models.ErrorResponse{Errors: errorObjects})\n\t\t\treturn\n\t\t}\n\n\t\tlog.Error(ctx, \"update recipe: failed to insert recipe\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\n\tlog.Info(ctx, \"update recipe: request successful\", logData)\n}", "func (o *SwarmUpdateParams) SetBody(body *models.SwarmSpec) {\n\to.Body = body\n}", "func (m *ItemItemsItemWorkbookFunctionsRandBetweenPostRequestBody) SetAdditionalData(value map[string]any)() {\n err := m.GetBackingStore().Set(\"additionalData\", value)\n if err != nil {\n panic(err)\n }\n}", "func (r *Request) SetBody(body io.Reader) *Request {\n\tr.Body = body\n\treturn r\n}", "func (b *httpRequestWriter) setBody(body io.Reader) error {\n\trc, ok := body.(io.ReadCloser)\n\tif !ok && body != nil {\n\t\trc = ioutil.NopCloser(body)\n\t}\n\tb.Body = rc\n\tif body != nil {\n\t\tswitch v := body.(type) {\n\t\tcase *bytes.Buffer:\n\t\t\tb.ContentLength = int64(v.Len())\n\t\t\tbuf := v.Bytes()\n\t\t\tb.GetBody = func() (io.ReadCloser, error) {\n\t\t\t\tr := bytes.NewReader(buf)\n\t\t\t\treturn ioutil.NopCloser(r), nil\n\t\t\t}\n\t\tcase *bytes.Reader:\n\t\t\tb.ContentLength = int64(v.Len())\n\t\t\tsnapshot := *v\n\t\t\tb.GetBody = func() (io.ReadCloser, error) {\n\t\t\t\tr := snapshot\n\t\t\t\treturn ioutil.NopCloser(&r), nil\n\t\t\t}\n\t\tcase *strings.Reader:\n\t\t\tb.ContentLength = int64(v.Len())\n\t\t\tsnapshot := *v\n\t\t\tb.GetBody = func() (io.ReadCloser, error) {\n\t\t\t\tr := snapshot\n\t\t\t\treturn ioutil.NopCloser(&r), nil\n\t\t\t}\n\t\tdefault:\n\t\t\t// This is where we'd set it to -1 (at least\n\t\t\t// if body != NoBody) to mean unknown, but\n\t\t\t// that broke people during the Go 1.8 testing\n\t\t\t// period. People depend on it being 0 I\n\t\t\t// guess. Maybe retry later. See Issue 18117.\n\t\t}\n\t\t// For client requests, Request.ContentLength of 0\n\t\t// means either actually 0, or unknown. The only way\n\t\t// to explicitly say that the ContentLength is zero is\n\t\t// to set the Body to nil. But turns out too much code\n\t\t// depends on NewRequest returning a non-nil Body,\n\t\t// so we use a well-known ReadCloser variable instead\n\t\t// and have the http package also treat that sentinel\n\t\t// variable to mean explicitly zero.\n\t\tif b.GetBody != nil && b.ContentLength == 0 {\n\t\t\tb.Body = http.NoBody\n\t\t\tb.GetBody = func() (io.ReadCloser, error) { return http.NoBody, nil }\n\t\t}\n\t}\n\treturn nil\n}", "func (o *ValidateCreateServiceRequestNamingParams) SetBody(body *dynatrace.RequestNaming) {\n\to.Body = body\n}", "func ConversionRequestBody(imageLocation, desiredFormat string) ConversionRequest {\n\treturn ConversionRequest{\n\t\tImageLocation: imageLocation,\n\t\tDesiredFormat: desiredFormat,\n\t}\n}", "func ToJSONBody(t *testing.T, body interface{}) *bytes.Buffer {\n\tjsonParams, err := json.Marshal(body)\n\tassert.NoErr(t, err)\n\treturn bytes.NewBuffer(jsonParams)\n}", "func BlendJSON(source string, dest map[string]interface{}) error {\n\tsourceMap, err := JsonToMSI(source)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn Blend(sourceMap, dest)\n}", "func (r *Request) PatchJSON(path string, data interface{}) {\n\tb, err := json.Marshal(data)\n\tif err != nil {\n\t\tr.t.Fatalf(\"httptesting: PatchJSON:json.Marshal(%T): %v\", data, err)\n\t}\n\n\tr.Patch(path, \"application/json\", b)\n}", "func (self *Response) SetBody(responseBody interface{}) error {\n\tswitch responseBody.(type) {\n\tcase func(string, http.Request) string:\n\t\t//\n\t\t// Mock Response should have a Body Generator function registered?\n\t\t//\n\n\t\tbodyGenerator := responseBody.(func(string, http.Request) string)\n\t\tself.bodyGenerator = &bodyGenerator\n\n\tcase string:\n\t\t//\n\t\t// Mock Response should have a static body registered?\n\t\t//\n\n\t\tself.body = responseBody.(string)\n\n\tdefault:\n\t\treturn fmt.Errorf(\"unsupported body type. SetBody() can receive either a static body string or a Body Generator Callback with the signature 'func(string, http.Request) string', for a more detailed description kindly check Go Mock Yourself documentation\")\n\t}\n\n\treturn nil\n}", "func (_options *SendNotificationsOptions) SetBody(body *NotificationCreate) *SendNotificationsOptions {\n\t_options.Body = body\n\treturn _options\n}", "func (o *GetRackTopoesParams) SetRequestBody(requestBody *models.GetRackTopoesRequestBody) {\n\to.RequestBody = requestBody\n}", "func AccessRequestBody(jsonToken string) string {\n\tvals := url.Values{}\n\tvals.Set(\"client_id\", config.C.Enterprise[\"APIKey\"])\n\tvals.Set(\"client_secret\", config.C.Enterprise[\"ClientSecret\"])\n\tvals.Set(\"jwt_token\", jsonToken)\n\treturn vals.Encode()\n}", "func (c *ThreeScaleClient) buildPostJSONReq(ep string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(\"POST\", c.adminPortal.rawURL+ep, body)\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Authorization\", \"Basic \"+basicAuth(\"\", c.credential))\n\treturn req, err\n}", "func (r ApiPatchOAuth2ClientRequest) JsonPatch(jsonPatch []JsonPatch) ApiPatchOAuth2ClientRequest {\n\tr.jsonPatch = &jsonPatch\n\treturn r\n}", "func (o *UpdateStockReceiptParams) SetBody(body *models.StockReceipt) {\n\to.Body = body\n}", "func (b *OperationMutator) RequestBody(v RequestBody) *OperationMutator {\n\tb.proxy.requestBody = v\n\treturn b\n}", "func (o *PutMenuItemParams) WithJSONBody(jSONBody *models.MenuItem) *PutMenuItemParams {\n\to.SetJSONBody(jSONBody)\n\treturn o\n}", "func GetBodyJSON(r *http.Request, v interface{}) error {\n\terr := json.NewDecoder(r.Body).Decode(&v)\n\treturn err\n}", "func (b binder) setFromJSONBody() HTTPError {\n\tif b.req.ContentLength == 0 {\n\t\treturn nil\n\t}\n\t// Always parse the form; it's safe and idempotent.\n\tif err := b.req.ParseForm(); err != nil {\n\t\treturn NewHTTPError(http.StatusInternalServerError, fmt.Sprintf(\"failed to parse form: %s\", err.Error()))\n\t}\n\tctype := b.req.Header.Get(\"Content-Type\")\n\tswitch {\n\tcase strings.HasPrefix(ctype, \"application/json\"):\n\t\tbody, err := b.requestBody()\n\t\tif err != nil {\n\t\t\treturn NewHTTPError(http.StatusBadRequest, err.Error())\n\t\t}\n\t\treturn b.decodeJSON(body)\n\tcase ctype == \"application/x-www-form-urlencoded\":\n\t\t// Handled by ParseForm.\n\t\treturn nil\n\tdefault:\n\t\treturn NewHTTPError(http.StatusUnsupportedMediaType, \"\")\n\t}\n}", "func writeJsonBody(w http.ResponseWriter, data interface{}) {\n\tpayload, err := json.Marshal(data)\n\tif err != nil {\n\t\tlog.Println(\"fail to encode data to the JSON format, error:\", err)\n\t\tupdateHeader(w, err)\n\t\treturn\n\t}\n\n\tupdateHeader(w, err)\n\tw.Write(payload)\n}", "func copyBody(r interface{}) (io.ReadCloser, error) {\n\tvar originalBody io.ReadCloser\n\tswitch r := r.(type) {\n\tcase *http.Request:\n\t\toriginalBody = r.Body\n\tcase *http.Response:\n\t\toriginalBody = r.Body\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unsupported type (%T) for copyBody\", r)\n\t}\n\n\tcontent, err := ioutil.ReadAll(originalBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tnewBody := ioutil.NopCloser(bytes.NewReader(content))\n\tswitch r := r.(type) {\n\tcase *http.Request:\n\t\tr.Body = newBody\n\tcase *http.Response:\n\t\tr.Body = newBody\n\t}\n\n\treturn ioutil.NopCloser(bytes.NewReader(content)), nil\n}", "func (o *BatchUpsertCatalogObjectsParams) SetBody(body *models.BatchUpsertCatalogObjectsRequest) {\n\to.Body = body\n}", "func (ctx *SimpleContext) RequestJsonBody(typ interface{}) error {\n\tcontentType := ctx.request.Header.Get(\"Content-Type\")\n\tif strings.Contains(contentType, \"json\") {\n\t\tbody, err := ioutil.ReadAll(ctx.request.Body)\n\t\tif nil != err {\n\t\t\treturn err\n\t\t}\n\t\terr = json.Unmarshal(body, typ)\n\t\tif nil != err {\n\t\t\treturn err\n\t\t}\n\t\treturn ctx.validate(typ)\n\t}\n\treturn errors.New(\"request body content-type is not contains json \")\n}", "func createAppJson(dm util.DepManager, appDir, appName, appJson string) error {\n\n\tupdatedJson, err := getAndUpdateAppJson(dm, appName, appJson)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = ioutil.WriteFile(filepath.Join(appDir, fileFlogoJson), []byte(updatedJson), 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *PcloudPvminstancesSnapshotsRestorePostParams) SetBody(body *models.SnapshotRestore) {\n\to.Body = body\n}", "func (o *UserRolesActionV1Params) SetBody(body *models.DomainActionUserRolesRequest) {\n\to.Body = body\n}", "func (c *apiClient) newRequestWithJSONBody(\n\tctx context.Context, method, resourcePath string,\n\tquery url.Values, body interface{}) (*http.Request, error) {\n\tdata, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tc.Logger.Debugf(\"httpx: request body length: %d bytes\", len(data))\n\tif c.LogBody {\n\t\tc.Logger.Debugf(\"httpx: request body: %s\", string(data))\n\t}\n\trequest, err := c.newRequest(\n\t\tctx, method, resourcePath, query, bytes.NewReader(data))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif body != nil {\n\t\trequest.Header.Set(\"Content-Type\", \"application/json\")\n\t}\n\treturn request, nil\n}", "func (r *Request) setBodyObject(object interface{}) error {\n\tvar err error\n\tr.body, err = r.session.codec.Marshal(object, nil)\n\treturn err\n}", "func (b *BitcoinClient) createBody(rpcBody *RPCBody) (*bytes.Buffer, error) {\n\tbodyJSON, err := json.Marshal(rpcBody)\n\tif err != nil {\n\t\tlog.Println(ErrCreatingBody)\n\t\treturn nil, ErrCreatingBody\n\t}\n\n\treturn bytes.NewBuffer(bodyJSON), nil\n}", "func (o *PcloudV1CloudinstancesCosimagesPostParams) SetBody(body *models.CreateCosImageImportJob) {\n\to.Body = body\n}", "func (recipe *Recipe) FromJSON(r io.Reader) error {\n\tdecoder := json.NewDecoder(r)\n\treturn decoder.Decode(recipe)\n}", "func Body(data ...interface{}) AdditionalAttribute {\n return func(rb *Builder) error {\n rb.SetBody(data...)\n return nil\n }\n}", "func (g *Github) PatchJSON(url, content string, v interface{}) error {\n\tbytes, err := g.patchBytes(url, content)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err := json.Unmarshal(bytes, v); err != nil {\n\t\treturn fmt.Errorf(\"could not parse json for url %s: %v\", url, err)\n\t}\n\treturn nil\n}", "func update(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\tvar receipt Receipt\n\tvar err error\n\tfmt.Println(\"starting write\")\n\n\tif len(args[0]) == 0 {\n\t\treturn shim.Error(\"wrong arguments\")\n\t}\n\n\treceiptJson := []byte(args[0])\n\terr = json.Unmarshal(receiptJson, &receipt)\n\tif err != nil {\n\t\tfmt.Println(\"json is wrong,json is: \" + args[0])\n\t\treturn shim.Error(err.Error())\n\t}\n\n\t//check if new owner exists\n\towner, err := getOwner(stub, receipt.OwnerRelation.Id)\n\tif err != nil {\n\t\tfmt.Println(\"Failed to find owner - \" + owner.Id)\n\t\treturn shim.Error(err.Error())\n\t}\n\n\terr = stub.PutState(receipt.Id, receiptJson) //write the variable into the ledger\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\tfmt.Println(\"- end write\")\n\treturn shim.Success(nil)\n}", "func (o *DeleteRolesParams) SetBody(body *models.OpenpitrixDeleteRolesRequest) {\n\to.Body = body\n}", "func (o *DeleteKeyPairsParams) SetBody(body *models.OpenpitrixDeleteKeyPairsRequest) {\n\to.Body = body\n}" ]
[ "0.68018734", "0.65308696", "0.5895473", "0.58617973", "0.5805046", "0.57633257", "0.56163126", "0.55379707", "0.544437", "0.543002", "0.5392192", "0.53760797", "0.53396225", "0.5332443", "0.52658796", "0.5227039", "0.5201621", "0.51650983", "0.5164823", "0.5163957", "0.5132194", "0.5121809", "0.5109449", "0.5087797", "0.50841653", "0.50432533", "0.5040159", "0.5030128", "0.50232303", "0.5020545", "0.5010065", "0.5005683", "0.49898025", "0.49835432", "0.49659893", "0.49582815", "0.49573895", "0.49548724", "0.4941528", "0.49231824", "0.4918244", "0.4915779", "0.4915135", "0.49105522", "0.49072474", "0.49065325", "0.49024886", "0.48932314", "0.48789856", "0.48660737", "0.48330647", "0.4831856", "0.48259124", "0.48226297", "0.48135003", "0.48120487", "0.48021063", "0.47971144", "0.47920674", "0.47910425", "0.47888118", "0.47825927", "0.477518", "0.47706112", "0.47649541", "0.47539806", "0.47539166", "0.4737956", "0.4737685", "0.473124", "0.47171292", "0.47159696", "0.47062302", "0.47021443", "0.4687725", "0.4686038", "0.46844363", "0.4682353", "0.46776888", "0.46755472", "0.46685338", "0.4667529", "0.46671632", "0.46608132", "0.46535465", "0.46525648", "0.4643804", "0.4640378", "0.46394086", "0.46360043", "0.46241742", "0.46134353", "0.4611435", "0.46046832", "0.45937136", "0.4591276", "0.45818275", "0.45809174", "0.45768395", "0.45724404" ]
0.7069762
0
WithRecipeID adds the recipeID to the copy recipe to my recipes with changes params
func (o *CopyRecipeToMyRecipesWithChangesParams) WithRecipeID(recipeID string) *CopyRecipeToMyRecipesWithChangesParams { o.SetRecipeID(recipeID) return o }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *CopyRecipeToMyRecipesWithChangesParams) SetRecipeID(recipeID string) {\n\to.RecipeID = recipeID\n}", "func (o *RecipeAdditive) SetRecipe(ctx context.Context, exec boil.ContextExecutor, insert bool, related *Recipe) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(ctx, exec, boil.Infer()); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE \\\"recipe_additive\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, []string{\"recipe_id\"}),\n\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 2, recipeAdditivePrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.ID, o.ID}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, updateQuery)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tif _, err = exec.ExecContext(ctx, updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.RecipeID = related.ID\n\tif o.R == nil {\n\t\to.R = &recipeAdditiveR{\n\t\t\tRecipe: related,\n\t\t}\n\t} else {\n\t\to.R.Recipe = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &recipeR{\n\t\t\tRecipeAdditives: RecipeAdditiveSlice{o},\n\t\t}\n\t} else {\n\t\trelated.R.RecipeAdditives = append(related.R.RecipeAdditives, o)\n\t}\n\n\treturn nil\n}", "func (r *RecipeInfo) newRecipe(id int) error {\n\t// Get recipe from the database\n\trecipe, err := new(models.Recipe).GetByID(database, queries, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Setup Recipe info\n\tr.ID = recipe.ID\n\tr.TotalSteps = len(recipe.Steps)\n\tr.CurrentStep = recipe.Steps[0]\n\tr.PrevStep = nil\n\tr.NextStep = nil\n\tr.JobIDs = make([]int64, 0)\n\n\tif r.TotalSteps > 1 {\n\t\tr.NextStep = recipe.Steps[1]\n\t}\n\n\tr.recipe = recipe\n\tdone, err := r.initStep(0)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif done {\n\t\tlog.Warn(\"Just setup a newRecipe that is already done\")\n\t}\n\treturn err\n}", "func AddRecipe(recipe *Recipe) *Recipe {\n\tuuid, _ := uuid.NewRandom()\n\trecipe.ID = uuid.String()\n\trecipe.CreatedAt = time.Now().UTC().String()\n\trecipe.UpdatedAt = time.Now().UTC().String()\n\n\trecipeList = append(recipeList, recipe)\n\treturn recipe\n}", "func (o *RecipeLipid) SetRecipe(ctx context.Context, exec boil.ContextExecutor, insert bool, related *Recipe) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(ctx, exec, boil.Infer()); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE \\\"recipe_lipid\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, []string{\"recipe_id\"}),\n\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 2, recipeLipidPrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.ID, o.ID}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, updateQuery)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tif _, err = exec.ExecContext(ctx, updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.RecipeID = related.ID\n\tif o.R == nil {\n\t\to.R = &recipeLipidR{\n\t\t\tRecipe: related,\n\t\t}\n\t} else {\n\t\to.R.Recipe = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &recipeR{\n\t\t\tRecipeLipids: RecipeLipidSlice{o},\n\t\t}\n\t} else {\n\t\trelated.R.RecipeLipids = append(related.R.RecipeLipids, o)\n\t}\n\n\treturn nil\n}", "func (ds CBDataStore) RecipeUpdate(modelRecipe model.Recipe, id, userID string) error {\n\t_, err := ds.bucket.Replace(modelRecipe.ID, modelRecipe, 0, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithContext(ctx context.Context) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (ds CBDataStore) RecipeUpdate(modelRecipe model.Recipe, id, userID string) error {\n\t_, err := ds.bucket.Replace(modelRecipe.Id, modelRecipe, 0, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *RBController) SaveRecipe(w http.ResponseWriter, r *http.Request) (err error) {\n\t// Get properties\n\tname := r.PostFormValue(`name`)\n\tcuisine, err := strconv.Atoi(r.PostFormValue(`cuisine`))\n\n\t// Get the mealtype and season encoded ints\n\tmealtype := EncodeMealtype(r.Form[`mealtype`])\n\tseason := EncodeSeason(r.Form[`season`])\n\n\t// get everything else\n\tdescription := r.PostFormValue(`description`)\n\tingredients := r.PostFormValue(`ingredients`)\n\tinstructions := r.PostFormValue(`instructions`)\n\n\t// TODO better error handling\n\tif err != nil {\n\t\tfmt.Println(\"[WARNING] Something went wrong in SaveRecipe\")\n\t\tc.RenderError(w, 500, \"Sorry, something went wrong.\")\n\t\treturn\n\t}\n\n\t// everything OK: build the recipe, and send it to the database\n\trecipe := Recipe{ID: 0, Name: name, Cuisine: cuisine, Mealtype: mealtype,\n\t\tSeason: season, Description: description, Ingredientlist: ingredients,\n\t\tInstructions: instructions}\n\n\t// if we don't have the id string, then this is a new request.\n\tvars := mux.Vars(r)\n\tidStr := vars[\"id\"]\n\tid := 0\n\n\tif idStr != \"\" {\n\t\tid, _ = strconv.Atoi(idStr)\n\t\trecipe.ID = id\n\t\terr = c.RecipeDB.UpdateRecipe(&recipe)\n\t} else {\n\t\tid, err = c.RecipeDB.NewRecipe(&recipe)\n\t}\n\n\tif err == nil {\n\t\thttp.Redirect(w, r, \"/recipes/\"+fmt.Sprintf(\"%v\", id)+\"/\", http.StatusFound)\n\t}\n\treturn\n}", "func (ds CBDataStore) Recipe(id, userID string) (*model.Recipe, error) {\n\tr := recipe{}\n\t_, err := ds.bucket.Get(id, &r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif r.UserId != userID {\n\t\treturn nil, fmt.Errorf(\"\")\n\t}\n\n\treturn r.Recipe, nil\n}", "func (ds CBDataStore) Recipe(id, userID string) (*model.Recipe, error) {\n\tr := recipe{}\n\t_, err := ds.bucket.Get(id, &r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif r.UserID != userID {\n\t\treturn nil, fmt.Errorf(\"\")\n\t}\n\n\treturn r.Recipe, nil\n}", "func AddRecipe(db *database.DB, recipe *Recipe) {\n\tdb.Create(recipe)\n}", "func (m *Storage) AddRecipe(r adding.Recipe) error {\n\n\tnewR := Recipe{\n\t\tID: len(m.recipes) + 1,\n\t\tCreated: time.Now(),\n\t\tMealType: r.MealType,\n\t\tName: r.Name,\n\t\tIngredients: r.Ingredients,\n\t\tPreparation: r.Preparation,\n\t}\n\tm.recipes = append(m.recipes, newR)\n\n\treturn nil\n}", "func (c *RBController) Recipe(w http.ResponseWriter, r *http.Request) (err error) {\n\tvars := mux.Vars(r)\n\tid, _ := strconv.Atoi(vars[\"id\"])\n\trecipe, err := c.GetRecipe(id)\n\tif err == nil {\n\t\tc.HTML(w, http.StatusOK, \"recipes/recipe\", recipe)\n\t} else if err == sql.ErrNoRows {\n\t\t// this means that the recipe wasn't found, so we should return a 404 error\n\t\tc.RenderError(w, 404, \"Sorry, your page wasn't found\")\n\t\terr = nil\n\t}\n\treturn\n}", "func UpdateRecipe(c *gin.Context) {\n\tvar recipe model.Recipe\n\tif err := c.ShouldBind(&recipe); err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"illegal format for recipe: \" + err.Error()})\n\t\treturn\n\t}\n\n\tid, err := strconv.ParseUint(c.Param(\"id\"), 10, 64)\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"invalid id-format: \" + c.Param(\"id\")})\n\t\treturn\n\t}\n\trecipe.ID = uint(id)\n\n\tif recipe.ID == 0 {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"cannot update recipe without id\"})\n\t\treturn\n\t}\n\n\tif recipe.Title == \"\" || recipe.Resources == nil || len(recipe.Resources) == 0 {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"illegal content for recipe\"})\n\t\treturn\n\t}\n\n\terr = recipe.Update()\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": err.Error()})\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, gin.H{\"data\": recipe.ID})\n}", "func (db *DB) SaveRecipe(ctx context.Context, dr *domain.Recipe) error {\n\tvar r Recipe\n\terr := r.FromDomain(dr)\n\tif err != nil {\n\t\treturn err\n\t}\n\tnow := time.Now()\n\tr.CretedAt, r.ModifiedAt = &now, &now\n\n\tvar sb strings.Builder\n\t// t := template.Must(template.New(\"save.tmpl\").Funcs(fm).ParseFiles(\"../../../templates/dgraph/save.tmpl\"))\n\tt := template.Must(template.New(\"save.tmpl\").Funcs(fm).ParseFiles(\"/templates/dgraph/save.tmpl\"))\n\terr = t.Execute(&sb, dr)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treq := &api.Request{CommitNow: true}\n\treq.Vars = map[string]string{\"$xid\": dr.ExternalID}\n\treq.Query = sb.String()\n\n\tmutations := make([]*api.Mutation, 0, len(dr.Ingredients)*2+len(dr.Tags)*2+1)\n\n\t// keep any food and tag\n\tfor i, di := range dr.Ingredients {\n\t\tvar i0, i1 Ingredient\n\t\terr := i0.FromDomain(di)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = i1.FromDomain(di)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tid := fmt.Sprintf(\"_:i%d\", i)\n\t\ti0.ID, i1.ID = id, id\n\t\tr.Ingredients[i] = &Ingredient{ID: id} // only id, empty fields\n\n\t\t// food stem found\n\t\ti0.Food.ID = fmt.Sprintf(\"uid(f%d)\", i)\n\t\tji0, err := json.Marshal(i0)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tmu0 := &api.Mutation{\n\t\t\tSetJson: ji0,\n\t\t\tCond: fmt.Sprintf(\"@if(eq(len(r), 0) AND eq(len(f%d), 1))\", i),\n\t\t}\n\n\t\t// food stem not found\n\t\ti1.Food.ID = fmt.Sprintf(\"_:f%d\", i)\n\t\tji1, err := json.Marshal(i1)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tmu1 := &api.Mutation{\n\t\t\tSetJson: ji1,\n\t\t\tCond: fmt.Sprintf(\"@if(eq(len(r), 0) AND eq(len(f%d), 0))\", i),\n\t\t}\n\n\t\tmutations = append(mutations, mu0, mu1)\n\t}\n\n\t// using nquads to be able to directly link tag to recipe\n\tfor i := range dr.Tags {\n\t\t// tag name found\n\t\tnq := &api.NQuad{\n\t\t\tSubject: \"_:recipe\",\n\t\t\tPredicate: \"tags\",\n\t\t\tObjectId: fmt.Sprintf(\"uid(t%d)\", i),\n\t\t}\n\n\t\tmu0 := &api.Mutation{\n\t\t\tSet: []*api.NQuad{nq},\n\t\t\tCond: fmt.Sprintf(\"@if(eq(len(r), 0) AND eq(len(t%d), 1))\", i),\n\t\t}\n\n\t\t// tag name not found\n\t\tvar t Tag\n\t\terr := t.FromDomain(dr.Tags[i])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\ttag := fmt.Sprintf(\"_:tag%d)\", i)\n\t\tnq0 := &api.NQuad{\n\t\t\tSubject: \"_:recipe\",\n\t\t\tPredicate: \"tags\",\n\t\t\tObjectId: tag,\n\t\t}\n\t\tnq1 := &api.NQuad{\n\t\t\tSubject: tag,\n\t\t\tPredicate: \"tagName\",\n\t\t\tObjectValue: &api.Value{Val: &api.Value_StrVal{StrVal: t.TagName}},\n\t\t}\n\t\tnq2 := &api.NQuad{\n\t\t\tSubject: tag,\n\t\t\tPredicate: \"tagStem\",\n\t\t\tObjectValue: &api.Value{Val: &api.Value_StrVal{StrVal: t.TagStem}},\n\t\t}\n\t\tnq3 := &api.NQuad{\n\t\t\tSubject: tag,\n\t\t\tPredicate: \"dgraph.type\",\n\t\t\tObjectValue: &api.Value{Val: &api.Value_StrVal{StrVal: t.DType[0]}},\n\t\t}\n\n\t\tmu1 := &api.Mutation{\n\t\t\tSet: []*api.NQuad{nq0, nq1, nq2, nq3},\n\t\t\tCond: fmt.Sprintf(\"@if(eq(len(r), 0) AND eq(len(t%d), 0))\", i),\n\t\t}\n\n\t\tmutations = append(mutations, mu0, mu1)\n\t}\n\n\tr.ID = \"_:recipe\"\n\tr.Tags = nil // don't overwrite tags\n\tjr, err := json.Marshal(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\tmu := &api.Mutation{\n\t\tSetJson: jr,\n\t\tCond: \"@if(eq(len(r), 0))\",\n\t}\n\tmutations = append(mutations, mu)\n\n\treq.Mutations = mutations\n\n\tres, err := db.Dgraph.NewTxn().Do(ctx, req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif ruid, created := res.Uids[\"recipe\"]; created {\n\t\tdr.ID = ruid\n\t} else {\n\t\treturn errors.ErrDuplicateID{ID: dr.ExternalID}\n\t}\n\n\treturn nil\n}", "func AddRecipeLipidHook(hookPoint boil.HookPoint, recipeLipidHook RecipeLipidHook) {\n\tswitch hookPoint {\n\tcase boil.BeforeInsertHook:\n\t\trecipeLipidBeforeInsertHooks = append(recipeLipidBeforeInsertHooks, recipeLipidHook)\n\tcase boil.BeforeUpdateHook:\n\t\trecipeLipidBeforeUpdateHooks = append(recipeLipidBeforeUpdateHooks, recipeLipidHook)\n\tcase boil.BeforeDeleteHook:\n\t\trecipeLipidBeforeDeleteHooks = append(recipeLipidBeforeDeleteHooks, recipeLipidHook)\n\tcase boil.BeforeUpsertHook:\n\t\trecipeLipidBeforeUpsertHooks = append(recipeLipidBeforeUpsertHooks, recipeLipidHook)\n\tcase boil.AfterInsertHook:\n\t\trecipeLipidAfterInsertHooks = append(recipeLipidAfterInsertHooks, recipeLipidHook)\n\tcase boil.AfterSelectHook:\n\t\trecipeLipidAfterSelectHooks = append(recipeLipidAfterSelectHooks, recipeLipidHook)\n\tcase boil.AfterUpdateHook:\n\t\trecipeLipidAfterUpdateHooks = append(recipeLipidAfterUpdateHooks, recipeLipidHook)\n\tcase boil.AfterDeleteHook:\n\t\trecipeLipidAfterDeleteHooks = append(recipeLipidAfterDeleteHooks, recipeLipidHook)\n\tcase boil.AfterUpsertHook:\n\t\trecipeLipidAfterUpsertHooks = append(recipeLipidAfterUpsertHooks, recipeLipidHook)\n\t}\n}", "func (h mainUC) Add(req requests.RecipeRequest) (interface{}, error) {\n\t/** validate category */\n\t_, err := h.catRepo.Find(req.CategoryID)\n\tif err != nil {\n\t\treturn nil, errors.New(\"selected category not found\")\n\t}\n\n\trecipe := models.NewRecipe(req.Name, h.slugHelp.Make(\"recipes\", req.Name),\n\t\treq.CategoryID)\n\n\ttx := h.recipeDtlRepo.DBConn().Begin()\n\terr = h.recipeRepo.Save(recipe, tx)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn nil, errors.New(\"error while saving recipe\")\n\t}\n\n\tvar details []*models.RecipeDetail\n\tfor _, item := range req.Ingredients {\n\t\t_, err := h.ingRepo.Find(item.IngredientID)\n\t\tif err != nil {\n\t\t\ttx.Rollback()\n\t\t\treturn nil, errors.New(\"selected ingredient not found\")\n\t\t}\n\t\tdtl := models.NewRecipeDetail(recipe.ID, item.IngredientID, item.Notes)\n\t\terr = h.recipeDtlRepo.Save(dtl, tx)\n\t\tif err != nil {\n\t\t\treturn nil, errors.New(\"error while saving recipe item\")\n\t\t}\n\t\tdetails = append(details, dtl)\n\t}\n\n\ttx.Commit()\n\n\treturn h.createResponse(recipe, details)\n}", "func (api *FoodRecipeAPI) partialRecipeUpdate(w http.ResponseWriter, req *http.Request) {\n\tdefer DrainBody(req)\n\tctx := req.Context()\n\n\tvars := mux.Vars(req)\n\tid := vars[\"id\"]\n\tlogData := log.Data{\"id\": id}\n\n\tvar errorObjects []*models.ErrorObject\n\n\tpatchJSON, recipePatches, err := patch.Get(ctx, req.Body)\n\tif err != nil {\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// Validate patch request\n\tfor i, recipePatch := range *recipePatches {\n\t\tif err = recipePatch.Validate(nil); err != nil {\n\t\t\tif _, ok := err.(*validator.InvalidValidationError); ok {\n\t\t\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\t\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tfor _, err := range err.(validator.ValidationErrors) {\n\t\t\t\terrorObjects = append(errorObjects, models.HandleValidationErrors(strconv.Itoa(i), err.ActualTag(), err.StructField(), err.Value().(string), err.Param()))\n\t\t\t}\n\t\t}\n\t}\n\tif len(errorObjects) > 0 {\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// apply patch against recipe resource\n\tp, err := jsonpatch.DecodePatch(patchJSON)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: unable to decode patch\", err)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// find current recipe doc\n\tvar recipe models.Recipe\n\n\tcollection := api.MongoClient.Database(\"food-recipes\").Collection(\"recipes\")\n\tif err = collection.FindOne(ctx, bson.M{\"_id\": id}).Decode(&recipe); err != nil {\n\t\tif err == mongo.ErrNoDocuments {\n\t\t\tlog.Warn(ctx, \"patch recipe: failed to find recipe\", log.FormatErrors([]error{err}), logData)\n\t\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrRecipeNotFound.Error()})\n\t\t\tErrorResponse(ctx, w, http.StatusNotFound, &models.ErrorResponse{Errors: errorObjects})\n\t\t\treturn\n\t\t}\n\n\t\tlog.Error(ctx, \"patch recipe: failed to find recipe, bad connection?\", err)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\tb, err := json.Marshal(recipe)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: error returned from json marshal\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// apply patch to existing recipe\n\tmodified, err := p.Apply(b)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: unable to apply patch to recipe\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\terr = json.Unmarshal(modified, &recipe)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: unmarshal modified recipe into recipe struct\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// store new recipe\n\tif _, err = collection.ReplaceOne(ctx, bson.M{\"_id\": id}, recipe); err != nil {\n\t\tif err == mongo.ErrNoDocuments {\n\t\t\tlog.Error(ctx, \"update recipe: failed to update recipe, recipe deos not exists\", err, logData)\n\t\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrRecipeNotFound.Error()})\n\t\t\tErrorResponse(ctx, w, http.StatusNotFound, &models.ErrorResponse{Errors: errorObjects})\n\t\t\treturn\n\t\t}\n\n\t\tlog.Error(ctx, \"update recipe: failed to insert recipe\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\n\tlog.Info(ctx, \"update recipe: request successful\", logData)\n}", "func UpdateRecipe(id string, recipe *Recipe) (*Recipe, error) {\n\tindex, err := findRecipe(id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trecipeList[index] = recipe\n\treturn recipe, nil\n}", "func (db RecipeDB) addRecipe(name, version string, success bool) error {\n\tversionNum, err := convertVersion(version)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdb[name] = Recipe{Name: name, Version: versionNum, InstallTime: time.Now().Unix(), Success: success}\n\n\tvar recipelist []Recipe\n\tfor _, recipe := range db {\n\t\trecipelist = append(recipelist, recipe)\n\t}\n\tdbBytes, err := json.Marshal(recipelist)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdbDir := getDbDir()\n\tif err := os.MkdirAll(dbDir, 0755); err != nil {\n\t\treturn err\n\t}\n\n\tf, err := ioutil.TempFile(dbDir, dbFileName+\"_*\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := f.Write(dbBytes); err != nil {\n\t\tf.Close()\n\t\treturn err\n\t}\n\tif err := f.Close(); err != nil {\n\t\treturn err\n\t}\n\n\treturn os.Rename(f.Name(), filepath.Join(dbDir, dbFileName))\n}", "func (r *Recipe) UpdateRecipe(db *sql.DB) (res sql.Result, err error) {\n\tfmt.Printf(\"Entro aqui malditasea\")\n\tres, err = db.Exec(\"UPDATE recipes SET name=$1, preptime=$2, difficulty=$3, vegetarian=$4 WHERE id=$5\",\n\t\tr.Name, r.PrepTime, r.Difficulty, r.Vegetarian, r.ID)\n\treturn res, err\n}", "func PostRecipeDetails(ctx context.Context, newRecipe *models.NewRecipe) error {\n\tdb, err := mysql.Connect(ctx, constant.DBs.RecipeRolodex)\n\tif err != nil {\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\treturn constant.Errors.DbConnectionFailure\n\t}\n\n\ttx, err := db.BeginTx(ctx, nil)\n\tif err != nil {\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\treturn constant.Errors.InternalServer\n\t}\n\n\t// Add recipe details to Recipes table\n\terr = InsertRecipeDetails(ctx, tx, *newRecipe.Season, *newRecipe.Title, *newRecipe.Author, *newRecipe.Link)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\treturn err\n\t}\n\n\t// Add ingredients to Ingredients table if they don't already exist\n\tif len(newRecipe.IngredientList) != 0 {\n\t\terr = InsertIngredients(ctx, tx, newRecipe.IngredientList)\n\t\tif err != nil {\n\t\t\ttx.Rollback()\n\t\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\t\treturn err\n\t\t}\n\t}\n\n\terr = InsertLink(ctx, tx, *newRecipe.Title, newRecipe.IngredientList)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\treturn err\n\t}\n\n\terr = tx.Commit()\n\tif err != nil {\n\t\ttx.Rollback()\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\treturn constant.Errors.DbInsertFailure\n\t}\n\n\treturn nil\n}", "func (r *Recipe) CreateRecipe(db *sql.DB) error {\n\terr := db.QueryRow(\n\t\t\"INSERT INTO recipes(id, name, preptime, difficulty, vegetarian) VALUES($1, $2, $3, $4, $5) RETURNING id\",\n\t\tr.ID, r.Name, r.PrepTime, r.Difficulty, r.Vegetarian).Scan(&r.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithTimeout(timeout time.Duration) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func CreateRecipe(c *gin.Context) {\n\tvar recipe model.Recipe\n\tif err := c.ShouldBind(&recipe); err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"illegal format for recipe: \" + err.Error()})\n\t\treturn\n\t}\n\n\tif recipe.ID != 0 {\n\t\tc.JSON(http.StatusConflict, gin.H{\"message\": \"cannot create recipe with specific id\"})\n\t\treturn\n\t}\n\n\tif recipe.Title == \"\" || recipe.Resources == nil || len(recipe.Resources) == 0 {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"illegal content for recipe\"})\n\t\treturn\n\t}\n\n\tvar duplicate model.Recipe\n\tmodel.DB.Where(\"title = ?\", recipe.Title).First(&duplicate)\n\tif duplicate.Title != \"\" {\n\t\tc.JSON(http.StatusConflict, gin.H{\"message\": \"recipe with title \" + duplicate.Title + \" already exists\"})\n\t\treturn\n\t}\n\n\trecipe.Create()\n\tc.JSON(http.StatusCreated, gin.H{\"data\": recipe.ID})\n}", "func (ds CBDataStore) RecipeDelete(id, userID string) error {\n\tfrag, err := ds.bucket.LookupIn(id).Get(\"user_id\").Execute()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar recipeUserID string\n\terr = frag.Content(\"user_id\", &recipeUserID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif recipeUserID != userID {\n\t\treturn fmt.Errorf(\"recipe does not exist\")\n\t}\n\n\t_, err = ds.bucket.Remove(id, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (ds CBDataStore) RecipeDelete(id, userID string) error {\n\tfrag, err := ds.bucket.LookupIn(id).Get(\"user_id\").Execute()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar recipeUserID string\n\terr = frag.Content(\"user_id\", &recipeUserID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif recipeUserID != userID {\n\t\treturn fmt.Errorf(\"recipe does not exist\")\n\t}\n\n\t_, err = ds.bucket.Remove(id, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func NewCopyRecipeToMyRecipesWithChangesParams() *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (c *RBController) NewRecipe(w http.ResponseWriter, r *http.Request) (err error) {\n\t// build data with anonymous struct\n\tdata := struct {\n\t\t*Recipe\n\t\tNewRecipe bool\n\t}{\n\t\tnew(Recipe),\n\t\ttrue,\n\t}\n\n\t// pass data to render\n\tc.HTML(w, http.StatusOK, \"recipes/edit\", data)\n\treturn nil\n}", "func (m *Storage) DeleteRecipe(ID int) {\n\tfor i, recipe := range m.recipes {\n\t\tif ID == recipe.ID {\n\t\t\tm.recipes = append(m.recipes[:i], m.recipes[i+1:]...)\n\t\t\tbreak\n\t\t}\n\t}\n}", "func (ds CBDataStore) RecipeCreate(modelRecipe model.Recipe, userID string) (*model.Recipe, error) {\n\tkey := fmt.Sprintf(\"recipe::%s::%s\", userID, modelRecipe.Name)\n\tnewR := new(recipe)\n\tnewR.Recipe = &modelRecipe\n\tnewR.Type = \"recipe\"\n\n\t_, err := ds.bucket.Insert(key, newR, 0)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn newR.Recipe, nil\n}", "func (ds CBDataStore) RecipeCreate(modelRecipe model.Recipe, userID string) (*model.Recipe, error) {\n\tkey := fmt.Sprintf(\"recipe::%s::%s\", userID, modelRecipe.Name)\n\tnewR := new(recipe)\n\tnewR.Recipe = &modelRecipe\n\tnewR.Type = \"recipe\"\n\n\t_, err := ds.bucket.Insert(key, newR, 0)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn newR.Recipe, nil\n}", "func adminRecipe(w http.ResponseWriter, r *http.Request, param string) {\n\n tmpl, _ := template.ParseFiles(\"admin_header.html\", \"admin_recipe.html\", \"admin_footer.html\")\n\n // Open database\n db := getDBConnection()\n defer db.Close()\n r.ParseForm()\n \n recipe_id, err := strconv.Atoi(r.Form.Get(\"recipe_selection\"))\n if err != nil {\n recipe_id = -1 \n } \n\n if (param == \"add_drink\") {\n // returned form is receipe_name=<drink name entered>\n if len(r.Form.Get(\"recipe_add\")) <= 1 {\n http.Redirect(w, r, \"/admin/recipe/\", http.StatusSeeOther)\n return\n }\n \n // Get glass selection\n glass_type_id, err := strconv.Atoi(r.Form.Get(\"glass_selection\"))\n if err != nil {\n http.Redirect(w, r, \"/admin/recipe/\", http.StatusSeeOther)\n return\n } \n\n _, err = db.Exec(\"insert into recipe (name, glass_type_id) values (?, ?)\", r.Form.Get(\"recipe_add\"), glass_type_id)\n if err != nil {\n panic(fmt.Sprintf(\"Failed to update db: %v\", err))\n }\n \n // get inserted id\n row := db.QueryRow(\"select max(id) from recipe\")\n err = row.Scan(&recipe_id)\n if err != nil {\n http.Redirect(w, r, \"/admin/recipe/\", http.StatusSeeOther)\n return\n }\n \n // http.Redirect(w, r, \"/admin/recipe/\", http.StatusSeeOther)\n // return\n }\n \n if (param == \"add_ingrediant\") {\n // returned form is wanting to add an ingrediant to a drink\n// NSERT INTO recipe_ingredient (recipe_id, ingredient_id, seq, qty) SELECT r.id, i.id, 4, 1 FROM recipe r, ingredient i WHERE r.name = 'Gin and tonic (lemon lime)' AND i.name = 'Lemon'; \n\n ingredient_id, err := strconv.Atoi(r.Form.Get(\"ingrediant_selection\"))\n if err != nil {\n http.Redirect(w, r, \"/admin/recipe/\", http.StatusSeeOther)\n }\n \n ingredient_id_remove, err := strconv.Atoi(r.Form.Get(\"remove_ingr\"))\n if err != nil {\n ingredient_id_remove = -1\n }\n ingredient_qty, err := strconv.Atoi(r.Form.Get(\"ingrediant_qty\"))\n if err != nil {\n ingredient_qty = -1\n }\n \n // Default to a quantity of 1 if nothing entered or invalid entry\n if ingredient_qty <= 0 {\n ingredient_qty = 1\n }\n \n if ingredient_id_remove > 0 {\n _, err := db.Exec(\"delete from recipe_ingredient where recipe_id=? and ingredient_id=? \", recipe_id, ingredient_id_remove)\n if err != nil {\n panic(fmt.Sprintf(\"Failed to update db: %v\", err))\n }\n } else {\n\n // get next seq number\n var seq_num int\n row := db.QueryRow(\"select max(seq)+1 from recipe_ingredient where recipe_id=?\", recipe_id)\n err = row.Scan(&seq_num)\n if err != nil {\n seq_num = 1\n }\n \n \n _, err = db.Exec(\"insert into recipe_ingredient (recipe_id, ingredient_id, seq, qty) values (?, ?, ?, ?)\", recipe_id, ingredient_id, seq_num, ingredient_qty)\n if err != nil {\n panic(fmt.Sprintf(\"Failed to update db (add ingrediant): %v\", err))\n }\n }\n // http.Redirect(w, r, \"/admin/recipe/\", http.StatusSeeOther)\n // return\n }\n \n var adminR AdminRecipe \n \n \n if (recipe_id > 0) {\n adminR.RecipieSelected = true\n } else\n {\n adminR.RecipieSelected = false\n }\n \n // Get a list of all drinks for list box\n rows, err := db.Query(\"select r.id, r.name, r.glass_type_id from recipe r order by r.name\")\n if err != nil {\n panic(fmt.Sprintf(\"%v\", err))\n }\n defer rows.Close()\n glass_type_id := -1\n var tmp_glass_type_id int\n for rows.Next() {\n var recipe Recipe\n rows.Scan(&recipe.Id, &recipe.Name, &tmp_glass_type_id)\n if recipe_id == recipe.Id {\n recipe.Selected = true\n glass_type_id = tmp_glass_type_id\n } else {\n recipe.Selected = false\n }\n adminR.Recipes = append(adminR.Recipes, recipe)\n }\n rows.Close()\n \n // Get a list of glass types for the glass selection listbox\n rows, err = db.Query(\"select g.id, g.name from glass_type g order by g.name\")\n if err != nil {\n panic(fmt.Sprintf(\"%v\", err))\n }\n defer rows.Close()\n\n for rows.Next() {\n var glass GlassType\n rows.Scan(&glass.Id, &glass.Name)\n if glass.Id == glass_type_id {\n glass.Selected = true\n } else {\n glass.Selected = false\n }\n adminR.GlassTypes = append(adminR.GlassTypes, glass)\n }\n rows.Close() \n \n // Get a list of all ingrediants for the \"add\" list box\n rows, err = db.Query(\"select i.id, i.name from ingredient i order by i.name\")\n if err != nil {\n panic(fmt.Sprintf(\"%v\", err))\n }\n defer rows.Close()\n\n for rows.Next() {\n var recipeIngr AdminRecipeIngr\n rows.Scan(&recipeIngr.Id, &recipeIngr.Name)\n adminR.AllIngredients = append(adminR.AllIngredients, recipeIngr)\n }\n rows.Close()\n \n // Get a list of all ingrediants in the currently selected drink\n adminR.RecipieId = recipe_id\n \n sqlstr := \n ` select\n i.id, \n i.name,\n ri.qty * dt.unit_size,\n case when ri.qty = 1 then dt.unit_name else dt.unit_plural end as uom\n from recipe_ingredient ri\n inner join ingredient i on ri.ingredient_id = i.id\n inner join dispenser_type dt on dt.id = i.dispenser_type_id\n where ri.recipe_id = ?\n order by ri.seq`\n \n rows, err = db.Query(sqlstr, recipe_id)\n if err != nil {\n panic(fmt.Sprintf(\"%v\", err))\n }\n defer rows.Close()\n\n for rows.Next() {\n var recipeIngr AdminRecipeIngr\n rows.Scan(&recipeIngr.Id, &recipeIngr.Name, &recipeIngr.Qty, &recipeIngr.UoM)\n adminR.RecIngredients = append(adminR.RecIngredients, recipeIngr)\n }\n rows.Close()\n \n var adminHead AdminHeader\n adminHead.AllowMaint = AllowMaint\n \n tmpl.ExecuteTemplate(w, \"admin_header\", adminHead)\n tmpl.ExecuteTemplate(w, \"admin_recipe\", adminR)\n tmpl.ExecuteTemplate(w, \"admin_footer\", nil)\n return\n}", "func TestAddRecipe(t *testing.T) {\n\tpayload := fmt.Sprintf(`\n {\n \"mealtype\": \"Breakfast\",\n \"name\": \"Pancakes\",\n \"Ingredients\": [ \"150g all purpose flour\",\n \t\t\t\t \"150ml of milk\"],\n \"preparation\": \"Add all ingredients and mix. Put in Pan.\"\n}`)\n\n\tresponse, err := http.Post(baseURL+\"/recipes\", \"application/json\", strings.NewReader(payload))\n\tif err != nil {\n\t\tt.Fatalf(\"failed to get json, %s\", err)\n\t}\n\n\tcheckResponseCode(t, http.StatusOK, response.StatusCode)\n}", "func (m *Storage) GetRecipe(id int) (listing.Recipe, error) {\n\tvar recipe listing.Recipe\n\n\tfor i := range m.recipes {\n\n\t\tif m.recipes[i].ID == id {\n\t\t\trecipe.ID = m.recipes[i].ID\n\t\t\trecipe.MealType = m.recipes[i].MealType\n\t\t\trecipe.Name = m.recipes[i].Name\n\t\t\trecipe.Ingredients = m.recipes[i].Ingredients\n\t\t\trecipe.Preparation = m.recipes[i].Preparation\n\t\t\trecipe.Created = m.recipes[i].Created\n\n\t\t\treturn recipe, nil\n\t\t}\n\t}\n\n\treturn recipe, listing.ErrNotFound\n}", "func (o *RecipeLipid) Recipe(mods ...qm.QueryMod) recipeQuery {\n\tqueryMods := []qm.QueryMod{\n\t\tqm.Where(\"\\\"id\\\" = ?\", o.RecipeID),\n\t}\n\n\tqueryMods = append(queryMods, mods...)\n\n\tquery := Recipes(queryMods...)\n\tqueries.SetFrom(query.Query, \"\\\"recipe\\\"\")\n\n\treturn query\n}", "func GetRecipe(id int) (recipe Recipe) {\n\tvar httpClient = &http.Client{}\n\tr, _ := httpClient.Get(fmt.Sprintf(\"http://recipes:8080/recipes/%d\", id))\n\tdefer r.Body.Close()\n\tjson.NewDecoder(r.Body).Decode(&recipe)\n\treturn\n}", "func (db *DB) Copy(ctx context.Context, targetID, sourceID string, options ...Options) (targetRev string, err error) {\n\topts, err := mergeOptions(options...)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif copier, ok := db.driverDB.(driver.Copier); ok {\n\t\ttargetRev, err = copier.Copy(ctx, targetID, sourceID, opts)\n\t\tif errors.StatusCode(err) != StatusNotImplemented {\n\t\t\treturn targetRev, err\n\t\t}\n\t}\n\trow, err := db.Get(ctx, sourceID, opts)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tvar doc map[string]interface{}\n\tif err = row.ScanDoc(&doc); err != nil {\n\t\treturn \"\", err\n\t}\n\tdelete(doc, \"_rev\")\n\tdoc[\"_id\"] = targetID\n\treturn db.Put(ctx, targetID, doc)\n}", "func DeleteRecipe(c *gin.Context) {\n\tvar recipe model.Recipe\n\tid, err := strconv.ParseUint(c.Param(\"id\"), 10, 64)\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"invalid id-format: \" + c.Param(\"id\")})\n\t\treturn\n\t}\n\trecipe.ID = uint(id)\n\terr = recipe.Delete()\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": err.Error()})\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, gin.H{\"message\": \"recipe was deleted succesfully\"})\n}", "func UpdateRecipe(db *database.DB, recipe *Recipe) {\n\tdb.Save(recipe)\n}", "func withRestaurantID(id int) restaurantOption {\n\treturn func(m *RestaurantMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Restaurant\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Restaurant, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Restaurant.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func withRestaurantID(id int) restaurantOption {\n\treturn func(m *RestaurantMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Restaurant\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Restaurant, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Restaurant.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func copyResourceToHookResource(src Resource, uuid string) (dst HookResource) {\n\tdst.ID = uuid\n\tdst.Name = src.Name\n\tdst.Address = src.Address\n\n\treturn dst\n}", "func (o *RecipeAdditive) Recipe(mods ...qm.QueryMod) recipeQuery {\n\tqueryMods := []qm.QueryMod{\n\t\tqm.Where(\"\\\"id\\\" = ?\", o.RecipeID),\n\t}\n\n\tqueryMods = append(queryMods, mods...)\n\n\tquery := Recipes(queryMods...)\n\tqueries.SetFrom(query.Query, \"\\\"recipe\\\"\")\n\n\treturn query\n}", "func steps2recipe(steps []Step) (R *Recipe, err error) {\n\t//initialize output\n\tR = &Recipe{}\n\n\t//copy steps into recipe\n\tR.Steps = make([]Step, len(steps))\n\tcopy(R.Steps, steps)\n\n\t//initialize nutrition map\n\tR.Nutrition = make(map[string]units.Quantity)\n\n\t//find ingrediants and fill in\n\tfor i := 0; i < len(R.Steps) && err == nil; i++ {\n\t\ts := R.Steps[i]\n\t\tif s.IsIngrediant() {\n\n\t\t\ting := s.(*Ingrediant)\n\n\t\t\t//look for closest/slightly random food\n\t\t\tmeasurement, data, nutrition, err := searchForFood(ing.Name, ing.Measurement)\n\t\t\tif err != nil {\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\t//add nutrition to totals\n\t\t\tfor k, v := range nutrition {\n\t\t\t\t_, ok := R.Nutrition[k]\n\t\t\t\tif ok {\n\t\t\t\t\tq := R.Nutrition[k]\n\t\t\t\t\tR.Nutrition[k] = units.Quantity{Unit: v.Unit, Amount: q.Amount + v.Amount, Type: v.Type}\n\t\t\t\t} else {\n\t\t\t\t\tR.Nutrition[k] = v\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t//add price to total if possible\n\t\t\t_, ok := data[\"price\"]\n\t\t\tif ok {\n\t\t\t\tx, _ := strconv.ParseFloat(data[\"price\"], 64)\n\t\t\t\tR.Price += x\n\t\t\t}\n\n\t\t\t//set measurement\n\t\t\ting.Measurement = measurement\n\t\t\ting.Data = data\n\n\t\t\t//replace old step\n\t\t\tR.Steps[i] = ing\n\t\t}\n\t}\n\tif err != nil {\n\t\treturn\n\t}\n\n\t//build a title\n\tR.Title = randomTitle(R)\n\n\t//return recipe\n\treturn\n}", "func AddCopy(ctx context.Context, config *config.Config, mgr manager.Manager) error {\n\tctx = ctxlog.NewContextWithRecorder(ctx, \"copy-reconciler\", mgr.GetEventRecorderFor(\"copy-recorder\"))\n\tlog := ctxlog.ExtractLogger(ctx)\n\tr := NewCopyReconciler(ctx, config, mgr, credsgen.NewInMemoryGenerator(log), controllerutil.SetControllerReference)\n\n\tc, err := controller.New(\"copy-controller\", mgr, controller.Options{\n\t\tReconciler: r,\n\t\tMaxConcurrentReconciles: config.MaxQuarksSecretWorkers,\n\t})\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Adding copy controller to manager failed.\")\n\t}\n\n\tnsPred := newNSPredicate(ctx, mgr.GetClient(), config.MonitoredID)\n\n\t// Watch for changes to the copied status of QuarksSecrets\n\tp := predicate.Funcs{\n\t\tCreateFunc: func(e event.CreateEvent) bool { return false },\n\t\tDeleteFunc: func(e event.DeleteEvent) bool { return false },\n\t\tGenericFunc: func(e event.GenericEvent) bool { return false },\n\t\tUpdateFunc: func(e event.UpdateEvent) bool {\n\t\t\tn := e.ObjectNew.(*qsv1a1.QuarksSecret)\n\n\t\t\tif n.Status.Copied != nil {\n\t\t\t\tctxlog.Debugf(ctx, \"Skipping QuarksSecret '%s', if copy status '%v' is true\", n.Name, *n.Status.Copied)\n\t\t\t\treturn !(*n.Status.Copied)\n\t\t\t}\n\n\t\t\treturn true\n\t\t},\n\t}\n\terr = c.Watch(&source.Kind{Type: &qsv1a1.QuarksSecret{}}, &handler.EnqueueRequestForObject{}, nsPred, p)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"Watching quarks secrets failed in copy controller.\")\n\t}\n\n\t// Watch for changes to user created secrets\n\tp = predicate.Funcs{\n\t\tCreateFunc: func(e event.CreateEvent) bool { return false },\n\t\tDeleteFunc: func(e event.DeleteEvent) bool { return false },\n\t\tGenericFunc: func(e event.GenericEvent) bool { return false },\n\t\tUpdateFunc: func(e event.UpdateEvent) bool {\n\t\t\tn := e.ObjectNew.(*corev1.Secret)\n\t\t\to := e.ObjectOld.(*corev1.Secret)\n\n\t\t\tshouldProcessReconcile := isUserCreatedSecret(n)\n\t\t\tif reflect.DeepEqual(n.Data, o.Data) && reflect.DeepEqual(n.Labels, o.Labels) &&\n\t\t\t\treflect.DeepEqual(n.Annotations, o.Annotations) {\n\t\t\t\treturn false\n\t\t\t}\n\n\t\t\treturn shouldProcessReconcile\n\t\t},\n\t}\n\terr = c.Watch(&source.Kind{Type: &corev1.Secret{}}, handler.EnqueueRequestsFromMapFunc(\n\t\tfunc(a crc.Object) []reconcile.Request {\n\t\t\tsecret := a.(*corev1.Secret)\n\n\t\t\tif skip.Reconciles(ctx, mgr.GetClient(), secret) {\n\t\t\t\treturn []reconcile.Request{}\n\t\t\t}\n\n\t\t\treconciles, err := listQuarksSecretsReconciles(ctx, mgr.GetClient(), secret, secret.Namespace)\n\t\t\tif err != nil {\n\t\t\t\tctxlog.Errorf(ctx, \"Failed to calculate reconciles for secret '%s/%s': %v\", secret.Namespace, secret.Name, err)\n\t\t\t}\n\t\t\tif len(reconciles) > 0 {\n\t\t\t\treturn reconciles\n\t\t\t}\n\n\t\t\treturn reconciles\n\t\t}), nsPred, p)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"Watching user defined secrets failed in copy controller.\")\n\t}\n\n\treturn nil\n}", "func (db *DB) UpdateRecipe(ctx context.Context, dr *domain.Recipe) (string, error) {\n\tvar r Recipe\n\terr := r.FromDomain(dr)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tnow := time.Now()\n\tr.ModifiedAt = &now\n\n\tvar sb strings.Builder\n\t// t := template.Must(template.New(\"update.tmpl\").Funcs(fm).ParseFiles(\"../../../templates/dgraph/update.tmpl\"))\n\tt := template.Must(template.New(\"update.tmpl\").Funcs(fm).ParseFiles(\"/templates/dgraph/update.tmpl\"))\n\terr = t.Execute(&sb, dr)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treq := &api.Request{CommitNow: true}\n\treq.Vars = map[string]string{\"$xid\": dr.ExternalID}\n\treq.Query = sb.String()\n\n\tmutations := make([]*api.Mutation, 0, len(dr.Ingredients)*2+len(dr.Tags)*2+2)\n\n\t// remove old edges\n\trdel := map[string]interface{}{\n\t\t\"uid\": \"uid(r)\",\n\t\t\"ingredients\": map[string]interface{}{\"uid\": \"uid(i)\"},\n\t\t\"steps\": map[string]interface{}{\"uid\": \"uid(s)\"},\n\t\t\"tags\": map[string]interface{}{\"uid\": \"uid(t)\"},\n\t}\n\tjdel, err := json.Marshal(rdel)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tmdel := &api.Mutation{\n\t\tDeleteJson: jdel,\n\t\tCond: \"@if(eq(len(r), 1))\",\n\t}\n\tmutations = append(mutations, mdel)\n\n\t// keep any food and tag\n\tfor i, di := range dr.Ingredients {\n\t\tvar i0, i1 Ingredient\n\t\terr := i0.FromDomain(di)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\terr = i1.FromDomain(di)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tid := fmt.Sprintf(\"_:i%d\", i)\n\t\ti0.ID, i1.ID = id, id\n\t\tr.Ingredients[i] = &Ingredient{ID: id} // only id, empty fields\n\n\t\t// food stem found\n\t\ti0.Food.ID = fmt.Sprintf(\"uid(f%d)\", i)\n\t\tji0, err := json.Marshal(i0)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tmu0 := &api.Mutation{\n\t\t\tSetJson: ji0,\n\t\t\tCond: fmt.Sprintf(\"@if(eq(len(r), 1) AND eq(len(f%d), 1))\", i),\n\t\t}\n\n\t\t// food stem not found\n\t\ti1.Food.ID = fmt.Sprintf(\"_:f%d\", i)\n\t\tji1, err := json.Marshal(i1)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tmu1 := &api.Mutation{\n\t\t\tSetJson: ji1,\n\t\t\tCond: fmt.Sprintf(\"@if(eq(len(r), 1) AND eq(len(f%d), 0))\", i),\n\t\t}\n\n\t\tmutations = append(mutations, mu0, mu1)\n\t}\n\n\t// using nquads to be able to directly link tag to recipe\n\tfor i := range dr.Tags {\n\t\t// tag name found\n\t\tnq := &api.NQuad{\n\t\t\tSubject: \"uid(r)\",\n\t\t\tPredicate: \"tags\",\n\t\t\tObjectId: fmt.Sprintf(\"uid(t%d)\", i),\n\t\t}\n\n\t\tmu0 := &api.Mutation{\n\t\t\tSet: []*api.NQuad{nq},\n\t\t\tCond: fmt.Sprintf(\"@if(eq(len(r), 1) AND eq(len(t%d), 1))\", i),\n\t\t}\n\n\t\t// tag name not found\n\t\tvar t Tag\n\t\terr := t.FromDomain(dr.Tags[i])\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\ttag := fmt.Sprintf(\"_:tag%d)\", i)\n\t\tnq0 := &api.NQuad{\n\t\t\tSubject: \"uid(r)\",\n\t\t\tPredicate: \"tags\",\n\t\t\tObjectId: tag,\n\t\t}\n\t\tnq1 := &api.NQuad{\n\t\t\tSubject: tag,\n\t\t\tPredicate: \"tagName\",\n\t\t\tObjectValue: &api.Value{Val: &api.Value_StrVal{StrVal: t.TagName}},\n\t\t}\n\t\tnq2 := &api.NQuad{\n\t\t\tSubject: tag,\n\t\t\tPredicate: \"tagStem\",\n\t\t\tObjectValue: &api.Value{Val: &api.Value_StrVal{StrVal: t.TagStem}},\n\t\t}\n\t\tnq3 := &api.NQuad{\n\t\t\tSubject: tag,\n\t\t\tPredicate: \"dgraph.type\",\n\t\t\tObjectValue: &api.Value{Val: &api.Value_StrVal{StrVal: t.DType[0]}},\n\t\t}\n\n\t\tmu1 := &api.Mutation{\n\t\t\tSet: []*api.NQuad{nq0, nq1, nq2, nq3},\n\t\t\tCond: fmt.Sprintf(\"@if(eq(len(r), 1) AND eq(len(t%d), 0))\", i),\n\t\t}\n\n\t\tmutations = append(mutations, mu0, mu1)\n\t}\n\n\tr.ID = \"uid(r)\"\n\tr.Tags = nil // don't overwrite tags\n\tjr, err := json.Marshal(r)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tmu := &api.Mutation{\n\t\tSetJson: jr,\n\t\tCond: \"@if(eq(len(r), 1))\",\n\t}\n\tmutations = append(mutations, mu)\n\n\treq.Mutations = mutations\n\n\tres, err := db.Dgraph.NewTxn().Do(ctx, req)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// catch updated recipe ID (it's stored in the response Json field)\n\tvar resj struct {\n\t\tRecipeUID []struct {\n\t\t\tUID string `json:\"uid\"`\n\t\t} `json:\"recipeUID\"`\n\t}\n\n\terr = json.Unmarshal(res.Json, &resj)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif len(resj.RecipeUID) == 0 {\n\t\treturn \"\", nil\n\t}\n\n\treturn resj.RecipeUID[0].UID, nil\n}", "func withCarRepairrecordID(id int) carrepairrecordOption {\n\treturn func(m *CarRepairrecordMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *CarRepairrecord\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*CarRepairrecord, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().CarRepairrecord.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func withCarID(id int) carOption {\n\treturn func(m *CarMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Car\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Car, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Car.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func withCarID(id int) carOption {\n\treturn func(m *CarMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Car\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Car, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Car.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func (recipe *Recipe) ValidateAddRecipe(ctx context.Context) error {\n\tvar missingFields []string\n\tvar invalidFields []string\n\n\t//recipe.ID generated by API if ID not given so never missing (generates a V4 UUID)\n\n\tif recipe.Alias == \"\" {\n\t\tmissingFields = append(missingFields, \"alias\")\n\t}\n\tif recipe.Format == \"\" {\n\t\tmissingFields = append(missingFields, \"format\")\n\t} else {\n\t\tif !validFormats[recipe.Format] {\n\t\t\tinvalidFields = append(invalidFields, \"format is not valid\")\n\t\t}\n\t}\n\n\tif recipe.InputFiles != nil && len(recipe.InputFiles) > 0 {\n\t\tfor i, file := range recipe.InputFiles {\n\t\t\tif file.Description == \"\" {\n\t\t\t\tmissingFields = append(missingFields, \"input-files[\"+strconv.Itoa(i)+\"].description\")\n\t\t\t}\n\t\t}\n\t} else {\n\t\tmissingFields = append(missingFields, \"input-files\")\n\t}\n\n\tif recipe.OutputInstances != nil && len(recipe.OutputInstances) > 0 {\n\t\tfor i, instance := range recipe.OutputInstances {\n\t\t\tinstanceMissingFields, instanceInvalidFields := instance.validateInstance(ctx)\n\t\t\tif len(instanceMissingFields) > 0 {\n\t\t\t\tfor mIndex, mField := range instanceMissingFields {\n\t\t\t\t\tinstanceMissingFields[mIndex] = \"output-instances[\" + strconv.Itoa(i) + \"].\" + mField\n\t\t\t\t}\n\t\t\t}\n\t\t\tif len(instanceInvalidFields) > 0 {\n\t\t\t\tfor iIndex, iField := range instanceInvalidFields {\n\t\t\t\t\tinstanceInvalidFields[iIndex] = \"output-instances[\" + strconv.Itoa(i) + \"].\" + iField\n\t\t\t\t}\n\t\t\t}\n\t\t\tmissingFields = append(missingFields, instanceMissingFields...)\n\t\t\tinvalidFields = append(invalidFields, instanceInvalidFields...)\n\t\t}\n\t} else {\n\t\tmissingFields = append(missingFields, \"output-instances\")\n\t}\n\n\tif missingFields != nil {\n\t\treturn fmt.Errorf(\"missing mandatory fields: %v\", missingFields)\n\t}\n\n\tif invalidFields != nil {\n\t\treturn fmt.Errorf(\"invalid fields: %v\", invalidFields)\n\t}\n\n\treturn nil\n\n}", "func withReviewID(id int) reviewOption {\n\treturn func(m *ReviewMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Review\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Review, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Review.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func (b *taskBuilder) recipeProp(key, value string) {\n\tif b.recipeProperties == nil {\n\t\tlog.Fatal(\"taskBuilder.recipeProp() cannot be called after taskBuilder.getRecipeProps()!\")\n\t}\n\tb.recipeProperties[key] = value\n}", "func (o *RecipeLipid) SetLipid(ctx context.Context, exec boil.ContextExecutor, insert bool, related *Lipid) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(ctx, exec, boil.Infer()); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE \\\"recipe_lipid\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, []string{\"lipid_id\"}),\n\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 2, recipeLipidPrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.ID, o.ID}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, updateQuery)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tif _, err = exec.ExecContext(ctx, updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.LipidID = related.ID\n\tif o.R == nil {\n\t\to.R = &recipeLipidR{\n\t\t\tLipid: related,\n\t\t}\n\t} else {\n\t\to.R.Lipid = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &lipidR{\n\t\t\tRecipeLipid: o,\n\t\t}\n\t} else {\n\t\trelated.R.RecipeLipid = o\n\t}\n\n\treturn nil\n}", "func (db *DB) DeleteRecipe(ctx context.Context, recipeID string) error {\n\tr, err := db.getRecipeByID(ctx, recipeID)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif r == nil {\n\t\treturn nil\n\t}\n\tr.Tags = nil\n\n\td := make([]interface{}, 0, len(r.Ingredients)+len(r.Steps)+1)\n\td = append(d, r)\n\tfor _, i := range r.Ingredients {\n\t\ti.Food = nil\n\t\td = append(d, *i)\n\t}\n\tfor _, s := range r.Steps {\n\t\td = append(d, *s)\n\t}\n\n\tpb, err := json.Marshal(d)\n\tif err != nil {\n\t\treturn err\n\t}\n\tmu := &api.Mutation{\n\t\tDeleteJson: pb,\n\t}\n\treq := &api.Request{CommitNow: true}\n\treq.Mutations = []*api.Mutation{mu}\n\n\t_, err = db.Dgraph.NewTxn().Do(ctx, req)\n\n\treturn err\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.JSONBody != nil {\n\t\tif err := r.SetBodyParam(o.JSONBody); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param recipeId\n\tif err := r.SetPathParam(\"recipeId\", o.RecipeID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (ctx *Context) SpecificFavRecipeHandler(w http.ResponseWriter, r *http.Request) {\n\tstate := &SessionState{}\n\t//grabbing the session ID from the sessionsstore\n\ts, err := sessions.GetSessionID(r, ctx.SessionKey)\n\tif err != nil {\n\t\thttp.Error(w, \"Error getting authenticated user\", http.StatusInternalServerError)\n\t\treturn\n\t}\n\t//getting the state and putting into state\n\terr = ctx.SessionStore.Get(s, state)\n\tif err != nil {\n\t\thttp.Error(w, \"Error getting session state\", http.StatusInternalServerError)\n\t\treturn\n\t}\n\t//get the url from the path\n\t_, recipeName := path.Split(r.URL.String())\n\tuser := state.User\n\tif err != nil {\n\t\thttp.Error(w, \"Error finding user\", http.StatusInternalServerError)\n\t\treturn\n\t}\n\tswitch r.Method {\n\tcase \"POST\":\n\t\t//Adds the recipe\n\t\terr := ctx.UserStore.AddToBook(user, recipeName)\n\t\tif err != nil {\n\t\t\thttp.Error(w, \"Error adding favorite recipe\", http.StatusInternalServerError)\n\t\t}\n\t\t//Recipe events adding recipe to book\n\t\tntfy := &notification.RecipesEvent{\n\t\t\tEventType: notification.NewBook,\n\t\t\tMessage: recipeName,\n\t\t}\n\t\tctx.Notifier.Notify(ntfy)\n\t\tw.Header().Add(\"Content-Type\", contentTypeTextUTF8)\n\t\tw.Write([]byte(\"Recipe has been added from the book\"))\n\tcase \"DELETE\":\n\t\terr := ctx.UserStore.DeleteFromBook(user, recipeName)\n\t\tif err != nil {\n\t\t\thttp.Error(w, \"Unable to remove that recipe\", http.StatusBadRequest)\n\t\t}\n\t\t//Recipe events adding recipe to book\n\t\tntfy := &notification.RecipesEvent{\n\t\t\tEventType: notification.RemoveBook,\n\t\t\tMessage: recipeName,\n\t\t}\n\t\tctx.Notifier.Notify(ntfy)\n\t\tw.Header().Add(\"Content-Type\", contentTypeTextUTF8)\n\t\tw.Write([]byte(\"Recipe has been removed from the book\"))\n\t}\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func NewCopyRecipeToMyRecipesWithChangesParamsWithHTTPClient(client *http.Client) *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\t\tHTTPClient: client,\n\t}\n}", "func (c *UsageController) Add(recipeID int64, userID int64) error {\n\tc.Usage = append(c.Usage, models.Usage{\n\t\tID: c.getNewID(),\n\t\tRecipeID: recipeID,\n\t\tDate: time.Now(),\n\t\tUserID: userID,\n\t})\n\n\treturn nil\n}", "func (cc *Chaincode) addJudgementID(stub shim.ChaincodeStubInterface, params []string) sc.Response {\n\t// Check Access\n\tcreatorOrg, creatorCertIssuer, err := getTxCreatorInfo(stub)\n\tif !authenticateCourt(creatorOrg, creatorCertIssuer) {\n\t\treturn shim.Error(\"{\\\"Error\\\":\\\"Access Denied!\\\",\\\"Payload\\\":{\\\"MSP\\\":\\\"\" + creatorOrg + \"\\\",\\\"CA\\\":\\\"\" + creatorCertIssuer + \"\\\"}}\")\n\t}\n\n\t// Check if sufficient Params passed\n\tif len(params) != 2 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting 2!\")\n\t}\n\n\t// Check if Params are non-empty\n\tfor a := 0; a < 2; a++ {\n\t\tif len(params[a]) <= 0 {\n\t\t\treturn shim.Error(\"Argument must be a non-empty string\")\n\t\t}\n\t}\n\n\t// Copy the Values from params[]\n\tID := params[0]\n\tJudgementID := params[1]\n\n\t// Check if ChargeSheet exists with Key => ID\n\tchargeSheetAsBytes, err := stub.GetState(ID)\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to get ChargeSheet Details!\")\n\t} else if chargeSheetAsBytes == nil {\n\t\treturn shim.Error(\"Error: ChargeSheet Does NOT Exist!\")\n\t}\n\n\t// Create Update struct var\n\tchargeSheetToUpdate := chargeSheet{}\n\terr = json.Unmarshal(chargeSheetAsBytes, &chargeSheetToUpdate) //unmarshal it aka JSON.parse()\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\t// Update ChargeSheet.InvestigationIDs to append => NewInvestigationID\n\tchargeSheetToUpdate.InvestigationIDs = append(chargeSheetToUpdate.InvestigationIDs, JudgementID)\n\n\t// Convert to JSON bytes\n\tchargeSheetJSONasBytes, err := json.Marshal(chargeSheetToUpdate)\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\t// Put State of newly generated ChargeSheet with Key => ID\n\terr = stub.PutState(ID, chargeSheetJSONasBytes)\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\tfor a := 0; a < len(chargeSheetToUpdate.FIRIDs); a++ {\n\t\t// Add Judgement ID to The FIR with FIRID\n\t\targs := util.ToChaincodeArgs(\"addJudgementToFIR\", chargeSheetToUpdate.FIRIDs[a], JudgementID)\n\t\tresponse := stub.InvokeChaincode(\"fir_cc\", args, \"mainchannel\")\n\t\tif response.Status != shim.OK {\n\t\t\treturn shim.Error(response.Message)\n\t\t}\n\t}\n\n\t// Returned on successful execution of the function\n\treturn shim.Success(chargeSheetJSONasBytes)\n}", "func (o *GetIngredientVersionRevisionParams) SetIngredientID(ingredientID strfmt.UUID) {\n\to.IngredientID = ingredientID\n}", "func (r *Recipe) DeleteRecipe(db *sql.DB) (res sql.Result, err error) {\n\tres, err = db.Exec(\"DELETE FROM recipes WHERE id=$1\", r.ID)\n\treturn res, err\n}", "func withPledgeID(id int) pledgeOption {\n\treturn func(m *PledgeMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Pledge\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Pledge, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Pledge.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func AddJsonBookWithID(id string, bk Book) (Book, error) {\n\n\t// validate form values\n\tif bk.Isbn == \"\" || bk.Title == \"\" || bk.Author == \"\" || bk.Price == \"\" {\n\t\treturn bk, errors.New(\"400. Bad request. All fields must be complete.\")\n\t}\n ID := id\n\t//doc := make(map[string]interface{})\n\t//doc[\"Isbn\"] = bk.Isbn\n\t//doc[\"Title\"] = bk.Title\n\t//doc[\"Author\"] = bk.Author\n\t//doc[\"Price\"] = bk.Price\n\n ctx := context.Background()\n\t _, err := config.Client.Collection(\"books\").Doc(ID).Set(ctx,\n\t\tmap[string]interface{}{\n\t\t\t\"Isbn\": bk.Isbn,\n\t\t\t\"Title\": bk.Title,\n\t\t\t\"Author\": bk.Author,\n\t\t\t\"Price\": bk.Price,\n\t\t})\n\n\tif err != nil {\n\t\tlog.Fatalf(\"Failed to add a new book: %w\", err)\n\t\t//fmt.Errorf(\"Failed to iterate the list of requests: %w\", err)\n\n\t}\n\n\treturn bk, nil\n}", "func (handler *RecipesHandler) NewRecipeHandler(c *gin.Context) {\n\tvar recipe models.Recipe\n\terr := c.ShouldBindJSON(&recipe)\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"error\": err.Error()})\n\t\treturn\n\t}\n\trecipe.ID = primitive.NewObjectID()\n\trecipe.PublishedAt = time.Now()\n\t_, err = handler.collection.InsertOne(handler.ctx, recipe)\n\tif err != nil {\n\t\tc.JSON(http.StatusInternalServerError, gin.H{\"error\": err.Error()})\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, recipe)\n}", "func GetRecipeByID(db *database.DB, id int) (Recipe, error) {\n\tvar recipe Recipe\n\terr := db.First(&recipe, id).Error\n\treturn recipe, err\n}", "func NewCopyRecipeToMyRecipesWithChangesParamsWithTimeout(timeout time.Duration) *CopyRecipeToMyRecipesWithChangesParams {\n\tvar ()\n\treturn &CopyRecipeToMyRecipesWithChangesParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func withDrugID(id int) drugOption {\n\treturn func(m *DrugMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Drug\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Drug, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Drug.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func (db *DB) GetRecipeByID(ctx context.Context, id string) (*domain.Recipe, error) {\n\tr, err := db.getRecipeByID(ctx, id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif r == nil {\n\t\treturn nil, nil\n\t}\n\n\treturn r.ToDomain(), nil\n}", "func AddRecipeAdditiveHook(hookPoint boil.HookPoint, recipeAdditiveHook RecipeAdditiveHook) {\n\tswitch hookPoint {\n\tcase boil.BeforeInsertHook:\n\t\trecipeAdditiveBeforeInsertHooks = append(recipeAdditiveBeforeInsertHooks, recipeAdditiveHook)\n\tcase boil.BeforeUpdateHook:\n\t\trecipeAdditiveBeforeUpdateHooks = append(recipeAdditiveBeforeUpdateHooks, recipeAdditiveHook)\n\tcase boil.BeforeDeleteHook:\n\t\trecipeAdditiveBeforeDeleteHooks = append(recipeAdditiveBeforeDeleteHooks, recipeAdditiveHook)\n\tcase boil.BeforeUpsertHook:\n\t\trecipeAdditiveBeforeUpsertHooks = append(recipeAdditiveBeforeUpsertHooks, recipeAdditiveHook)\n\tcase boil.AfterInsertHook:\n\t\trecipeAdditiveAfterInsertHooks = append(recipeAdditiveAfterInsertHooks, recipeAdditiveHook)\n\tcase boil.AfterSelectHook:\n\t\trecipeAdditiveAfterSelectHooks = append(recipeAdditiveAfterSelectHooks, recipeAdditiveHook)\n\tcase boil.AfterUpdateHook:\n\t\trecipeAdditiveAfterUpdateHooks = append(recipeAdditiveAfterUpdateHooks, recipeAdditiveHook)\n\tcase boil.AfterDeleteHook:\n\t\trecipeAdditiveAfterDeleteHooks = append(recipeAdditiveAfterDeleteHooks, recipeAdditiveHook)\n\tcase boil.AfterUpsertHook:\n\t\trecipeAdditiveAfterUpsertHooks = append(recipeAdditiveAfterUpsertHooks, recipeAdditiveHook)\n\t}\n}", "func CopyID(id ID) (ret ID) {\n\tfor i := 0; i < IDBytes; i++ {\n\t\tret[i] = id[i]\n\t}\n\treturn\n}", "func (recipeLipidL) LoadRecipe(ctx context.Context, e boil.ContextExecutor, singular bool, maybeRecipeLipid interface{}, mods queries.Applicator) error {\n\tvar slice []*RecipeLipid\n\tvar object *RecipeLipid\n\n\tif singular {\n\t\tobject = maybeRecipeLipid.(*RecipeLipid)\n\t} else {\n\t\tslice = *maybeRecipeLipid.(*[]*RecipeLipid)\n\t}\n\n\targs := make([]interface{}, 0, 1)\n\tif singular {\n\t\tif object.R == nil {\n\t\t\tobject.R = &recipeLipidR{}\n\t\t}\n\t\targs = append(args, object.RecipeID)\n\n\t} else {\n\tOuter:\n\t\tfor _, obj := range slice {\n\t\t\tif obj.R == nil {\n\t\t\t\tobj.R = &recipeLipidR{}\n\t\t\t}\n\n\t\t\tfor _, a := range args {\n\t\t\t\tif a == obj.RecipeID {\n\t\t\t\t\tcontinue Outer\n\t\t\t\t}\n\t\t\t}\n\n\t\t\targs = append(args, obj.RecipeID)\n\n\t\t}\n\t}\n\n\tif len(args) == 0 {\n\t\treturn nil\n\t}\n\n\tquery := NewQuery(\n\t\tqm.From(`recipe`),\n\t\tqm.WhereIn(`recipe.id in ?`, args...),\n\t)\n\tif mods != nil {\n\t\tmods.Apply(query)\n\t}\n\n\tresults, err := query.QueryContext(ctx, e)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to eager load Recipe\")\n\t}\n\n\tvar resultSlice []*Recipe\n\tif err = queries.Bind(results, &resultSlice); err != nil {\n\t\treturn errors.Wrap(err, \"failed to bind eager loaded slice Recipe\")\n\t}\n\n\tif err = results.Close(); err != nil {\n\t\treturn errors.Wrap(err, \"failed to close results of eager load for recipe\")\n\t}\n\tif err = results.Err(); err != nil {\n\t\treturn errors.Wrap(err, \"error occurred during iteration of eager loaded relations for recipe\")\n\t}\n\n\tif len(recipeLipidAfterSelectHooks) != 0 {\n\t\tfor _, obj := range resultSlice {\n\t\t\tif err := obj.doAfterSelectHooks(ctx, e); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(resultSlice) == 0 {\n\t\treturn nil\n\t}\n\n\tif singular {\n\t\tforeign := resultSlice[0]\n\t\tobject.R.Recipe = foreign\n\t\treturn nil\n\t}\n\n\tfor _, local := range slice {\n\t\tfor _, foreign := range resultSlice {\n\t\t\tif local.RecipeID == foreign.ID {\n\t\t\t\tlocal.R.Recipe = foreign\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func getRecipe(w http.ResponseWriter, r *http.Request) {\n\tparams := mux.Vars(r)\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\n\trecipe, err := allrecipes.GetRecipe(params[\"id\"])\n\tif err != nil {\n\t\tif strings.Contains(err.Error(), \"GetRecipeInfo\") {\n\t\t\thttp.Error(w, err.Error(), http.StatusNotFound)\n\t\t\treturn\n\t\t}\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tb, err := json.Marshal(recipe)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\t// fmt.Println(\"str \", string(b))\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.Write(b)\n}", "func deleteRecipes(recipeData map[string]RecipeDetails, attributeCounts AttributeCounts, attributeVals AttributeVals, crockPot []IngredientDetails) map[string]RecipeDetails {\n\n\t// AMBEROSIA\n\tif countIngName(\"Collected Dust\", crockPot) < 1 {\n\t\tdelete(recipeData, \"AMBEROSIA\")\n\t}\n\n\t// ASPARAGUS SOUP\n\tif countIngName(\"Asparagus\", crockPot) < 1 || attributeVals.vegVal-1 < .5 || attributeCounts.meatCount != 0 || attributeCounts.inedibleCount != 0 {\n\t\tdelete(recipeData, \"ASPARAGUS SOUP\")\n\t}\n\n\t// Bacon and Eggs\n\tif attributeVals.meatVal <= 1 || attributeVals.eggVal <= 1 || attributeVals.vegVal != 0 {\n\t\tdelete(recipeData, \"BACON AND EGGS\")\n\t}\n\n\t// BANANA POP\n\tif countIngName(\"Banana\", crockPot) < 1 || countIngName(\"Twigs\", crockPot) < 1 || countIngName(\"Ice\", crockPot) < 1 || attributeCounts.meatCount != 0 || attributeCounts.fishCount != 0 {\n\t\tdelete(recipeData, \"BANANA POP\")\n\t}\n\n\t// Barnacle Linguine\n\tif countIngName(\"Barnacles\", crockPot) != 2 || attributeVals.vegVal < 2 {\n\t\tdelete(recipeData, \"BARNACLE LINGUINE\")\n\t}\n\n\t// Barnacle Nigiri\n\tif countIngName(\"Barnacles\", crockPot) < 1 || countIngName(\"Kelp Fronds\", crockPot) < 1 || attributeCounts.eggCount < 1 {\n\t\tdelete(recipeData, \"BARNACLE NIGIRI\")\n\t}\n\n\t// BARNACLE PITA\n\t// attributeVals.fishval over 1 can cause stuffed fish head but idk the percentage\n\tif countIngName(\"Barnacles\", crockPot) < 1 || attributeVals.vegVal < 0.5 {\n\t\tdelete(recipeData, \"BARNACLE PITA\")\n\t}\n\n\t// BEEFY GREENS\n\tif countIngName(\"Leafy Meat\", crockPot) < 1 || attributeVals.vegVal < 3 {\n\t\tdelete(recipeData, \"BEEFY GREENS\")\n\t}\n\n\t// butter muffin\n\tif countIngName(\"Butterfly Wings\", crockPot) < 1 || attributeCounts.vegeCount < 1 || stringInSlice(\"Mandrake\", crockPot) || attributeCounts.meatCount != 0 {\n\t\tdelete(recipeData, \"BUTTER MUFFIN\")\n\t}\n\n\t// CALIFORNIA ROLL\n\tif countIngName(\"Kelp Fronds\", crockPot) < 2 || attributeVals.fishVal < 1 {\n\t\tdelete(recipeData, \"CALIFORNIA ROLL\")\n\t}\n\n\t// ceviche\n\tif countIngName(\"Ice\", crockPot) < 2 || attributeVals.fishVal < 2 || attributeCounts.eggCount != 0 || attributeCounts.inedibleCount != 0 {\n\t\tdelete(recipeData, \"CEVICHE\")\n\t}\n\n\t// CREAMY POTATO PURÉE**\n\t// need to look at potato or roasted potato\n\tif countIngName(\"Potato\", crockPot) < 2 || countIngName(\"Garlic\", crockPot) < 1 || stringInSlice(\"Twigs\", crockPot) || attributeCounts.meatCount != 0 {\n\t\tdelete(recipeData, \"CREAMY POTATO PURÉE\")\n\t}\n\n\t// dragonpie\n\t// need to look at dragon fruit or prepared dragon fruit\n\tif countIngName(\"Dragon Fruit\", crockPot) < 1 || stringInSlice(\"Mandrake\", crockPot) || attributeCounts.meatCount != 0 {\n\t\tdelete(recipeData, \"DRAGONPIE\")\n\t}\n\n\t// FANCY SPIRALLED TUBERS\n\t// need to look at potato or roasted potato\n\tif countIngName(\"Potato\", crockPot) < 1 || countIngName(\"Twigs\", crockPot) < 1 || attributeCounts.inedibleCount-1 > 1 || attributeCounts.meatCount != 0 {\n\t\tdelete(recipeData, \"FANCY SPIRALLED TUBERS\")\n\t}\n\n\t// fish tacos\n\t// twig value - 1 twig 50% chance of fish sticks\n\t// need to look at corn or popcorn\n\tif attributeVals.fishVal < 0.5 || (countIngName(\"Corn\", crockPot) < 1 || countIngName(\"Popcorn\", crockPot) < 1) {\n\t\tdelete(recipeData, \"FISH TACOS\")\n\t}\n\n\t// fishsticks\n\tif attributeVals.fishVal < 0.25 || countIngName(\"Twigs\", crockPot) != 1 || stringInSlice(\"Moleworm\", crockPot) {\n\t\tdelete(recipeData, \"FISHSTICKS\")\n\t}\n\n\t// fist full of jam\n\tif attributeVals.fruitVal < 0.5 || attributeVals.meatVal != 0 || attributeVals.vegVal != 0 || attributeVals.inedVal != 0 || stringInSlice(\"Dragon Fruit\", crockPot) {\n\t\tdelete(recipeData, \"FIST FULL OF JAM\")\n\t}\n\n\t// flower salad\n\tif countIngName(\"Cactus Flower\", crockPot) < 1 || attributeVals.vegVal-.5 < 1.5 || attributeVals.meatVal != 0 || attributeVals.fruitVal != 0 || attributeVals.eggVal != 0 || attributeVals.sweetVal != 0 || stringInSlice(\"Twigs\", crockPot) {\n\t\tdelete(recipeData, \"FLOWER SALAD\")\n\t}\n\n\t// froggle bunwich\n\t// need to look at frog legs or cooked frog legs\n\t// makes kabob if only one stick\n\tif countIngName(\"Frog Legs\", crockPot) < 1 || attributeCounts.vegeCount < 1 || attributeVals.eggVal != 0 || attributeVals.sweetVal != 0 || stringInSlice(\"Mandrake\", crockPot) {\n\t\tdelete(recipeData, \"FROGGLE BUNWICH\")\n\t}\n\n\t// fruit medley\n\t// twigs is safest anything else 50% of fist full of jam\n\tif attributeVals.fruitVal < 3 || attributeVals.meatVal != 0 || attributeVals.vegVal != 0 || stringInSlice(\"Dragon Fruit\", crockPot) {\n\t\tdelete(recipeData, \"FRUIT MEDLEY\")\n\t}\n\n\t// guacamole\n\t// need to look at cactus flesh or stone fruit\n\tif countIngName(\"Moleworm\", crockPot) < 1 || (countIngName(\"Cactus Flesh\", crockPot) < 1 || countIngName(\"Ripe Stone Fruit\", crockPot) < 1) || attributeVals.fruitVal != 0 {\n\t\tdelete(recipeData, \"GUACAMOLE\")\n\t}\n\n\t// Honey Ham\n\tif countIngName(\"Honey\", crockPot) < 1 || attributeVals.meatVal <= 1.5 || stringInSlice(\"Twigs\", crockPot) || stringInSlice(\"Moleworm\", crockPot) || stringInSlice(\"Mandrake\", crockPot) || stringInSlice(\"Tallbird Egg\", crockPot) {\n\t\tdelete(recipeData, \"HONEY HAM\")\n\t}\n\n\t// Honey Nuggets\n\tif countIngName(\"Honey\", crockPot) < 1 || attributeVals.meatVal > 1.5 || attributeVals.meatVal == 0 || attributeVals.inedVal != 0 {\n\t\tdelete(recipeData, \"HONEY NUGGETS\")\n\t}\n\n\t// ice cream\n\tif countIngName(\"Ice\", crockPot) < 1 || attributeCounts.dairyCount < 1 || attributeCounts.sweetenerCount < 1 || attributeVals.meatVal != 0 || attributeVals.vegVal != 0 || attributeVals.eggVal != 0 || stringInSlice(\"Twigs\", crockPot) {\n\t\tdelete(recipeData, \"ICE CREAM\")\n\t}\n\n\t// jelly salad\n\t// look for cooked version\n\tif countIngName(\"Leafy Meat\", crockPot) < 2 || attributeVals.sweetVal < 2 {\n\t\tdelete(recipeData, \"JELLY SALAD\")\n\t}\n\n\t// jellybeans\n\tif countIngName(\"Royal Jelly\", crockPot) < 1 || attributeVals.monVal != 0 || attributeVals.inedVal != 0 {\n\t\tdelete(recipeData, \"JELLYBEANS\")\n\t}\n\n\t// kabobs\n\tif attributeCounts.meatCount < 1 || countIngName(\"Twigs\", crockPot) != 1 || stringInSlice(\"Moleworm\", crockPot) || stringInSlice(\"Mandrake\", crockPot) || attributeVals.fishVal != 0 {\n\t\tdelete(recipeData, \"KABOBS\")\n\t}\n\n\t// leafy meatloaf\n\tif countIngName(\"Leafy Meat\", crockPot) < 2 {\n\t\tdelete(recipeData, \"LEAFY MEATLOAF\")\n\t}\n\n\t// LOBSTER BISQUE\n\tif countIngName(\"Wobster\", crockPot) < 1 || countIngName(\"Ice\", crockPot) < 1 {\n\t\tdelete(recipeData, \"LOBSTER BISQUE\")\n\t}\n\n\t// mandrake soup\n\tif countIngName(\"Mandrake\", crockPot) < 1 {\n\t\tdelete(recipeData, \"MANDRAKE SOUP\")\n\t}\n\n\t// Meatballs\n\tif attributeVals.meatVal >= 3 || attributeVals.meatVal == 0 || stringInSlice(\"Twigs\", crockPot) {\n\t\tdelete(recipeData, \"MEATBALLS\")\n\t}\n\n\t// meaty stew\n\tif attributeVals.meatVal < 3 || stringInSlice(\"Twigs\", crockPot) || stringInSlice(\"Moleworm\", crockPot) || stringInSlice(\"Honey\", crockPot) || stringInSlice(\"Mandrake\", crockPot) || stringInSlice(\"Tallbird Egg\", crockPot) {\n\t\tdelete(recipeData, \"MEATY STEW\")\n\t}\n\n\t// melonsicle\n\tif countIngName(\"Watermelon\", crockPot) < 1 || countIngName(\"Ice\", crockPot) < 1 || countIngName(\"Twigs\", crockPot) < 1 || attributeVals.meatVal != 0 || attributeVals.vegVal != 0 || attributeVals.eggVal != 0 {\n\t\tdelete(recipeData, \"MELONSICLE\")\n\t}\n\n\t// MILKMADE HAT\n\tif countIngName(\"Nostrils\", crockPot) < 1 || countIngName(\"Kelp Fronds\", crockPot) < 1 || attributeCounts.dairyCount < 1 {\n\t\tdelete(recipeData, \"MILKMADE HAT\")\n\t}\n\n\t// monster lasagna\n\tif attributeCounts.monsterCount < 2 || stringInSlice(\"Twigs\", crockPot) {\n\t\tdelete(recipeData, \"MONSTER LASAGNA\")\n\t}\n\n\t// Mushy Cake\n\tif countIngName(\"Moon Shroom\", crockPot) != 1 || countIngName(\"Red Cap\", crockPot) != 1 || countIngName(\"Blue Cap\", crockPot) != 1 || countIngName(\"Green Cap\", crockPot) != 1 {\n\t\tdelete(recipeData, \"MUSHY CAKE\")\n\t}\n\n\t// Pierogi\n\tif attributeCounts.meatCount < 1 || attributeCounts.eggCount < 1 || attributeCounts.vegeCount < 1 || stringInSlice(\"Twigs\", crockPot) || stringInSlice(\"Mandrake\", crockPot) {\n\t\tdelete(recipeData, \"PIEROGI\")\n\t}\n\n\t//powdercake\n\tif countIngName(\"Corn\", crockPot) < 1 || countIngName(\"Honey\", crockPot) < 1 || countIngName(\"Twigs\", crockPot) < 1 {\n\t\tdelete(recipeData, \"POWDERCAKE\")\n\t}\n\n\t//Pumkin cookie\n\t// 3 honey or comb 50% chance of making taffy\n\tif countIngName(\"Pumpkin\", crockPot) < 1 || countIngName(\"Honey\", crockPot) <= 1 {\n\t\tdelete(recipeData, \"PUMPKIN COOKIE\")\n\t}\n\n\t// RATATOUILLE\n\tif attributeCounts.vegeCount < 1 || stringInSlice(\"Twigs\", crockPot) || stringInSlice(\"Mandrake\", crockPot) || stringInSlice(\"Butterfly Wings\", crockPot) || stringInSlice(\"Dragon Fruit\", crockPot) {\n\t\tdelete(recipeData, \"RATATOUILLE\")\n\t}\n\n\t// SALSA FRESCA\n\t// look for cooked versions\n\tif countIngName(\"Toma Root\", crockPot) < 1 || attributeCounts.meatCount != 0 || attributeCounts.inedibleCount != 0 || attributeCounts.eggCount != 0 {\n\t\tdelete(recipeData, \"SALSA FRESCA\")\n\t}\n\n\t// SEAFOOD GUMBO\n\tif countIngName(\"Eel\", crockPot) < 1 || attributeVals.fishVal <= 2 {\n\t\tdelete(recipeData, \"SEAFOOD GUMBO\")\n\t}\n\n\t// Soothing tea\n\t// look for honey or comb\n\tif countIngName(\"Forget-Me-Lots\", crockPot) < 1 || countIngName(\"Honey\", crockPot) < 1 || countIngName(\"Ice\", crockPot) < 1 || attributeCounts.monsterCount != 0 || attributeCounts.meatCount != 0 || attributeCounts.fishCount != 0 || attributeCounts.eggCount != 0 || attributeCounts.inedibleCount != 0 || attributeCounts.dairyCount != 0 {\n\t\tdelete(recipeData, \"SOOTHING TEA\")\n\t}\n\n\t// Spicy Chili\n\tif attributeCounts.meatCount != 2 || attributeCounts.vegeCount != 2 || attributeVals.meatVal < 1.5 || attributeVals.vegVal < 1.5 {\n\t\tdelete(recipeData, \"SPICY CHILI\")\n\t}\n\n\t// STUFFED EGGPLANT\n\tif countIngName(\"Eggplant\", crockPot) < 1 || attributeCounts.vegeCount < 1 {\n\t\tdelete(recipeData, \"STUFFED EGGPLANT\")\n\t}\n\n\t// stuffed fish heads\n\tif countIngName(\"Barnacles\", crockPot) < 1 || attributeVals.fishVal-0.5 < 1 {\n\t\tdelete(recipeData, \"STUFFED FISH HEADS\")\n\t}\n\n\t// STUFFED PEPPER POPPERS\n\tif countIngName(\"Pepper\", crockPot) < 1 || attributeVals.meatVal > 1.5 || attributeCounts.meatCount == 0 || stringInSlice(\"Twigs\", crockPot) {\n\t\tdelete(recipeData, \"STUFFED PEPPER POPPERS\")\n\t}\n\n\t// SURF 'N' TURF\n\tif attributeVals.meatVal < 2.5 || attributeVals.fishVal < 1.5 || stringInSlice(\"Ice\", crockPot) {\n\t\tdelete(recipeData, \"SURF 'N' TURF\")\n\t}\n\n\t// Taffy\n\tif attributeVals.sweetVal < 3 || attributeCounts.meatCount != 0 {\n\t\tdelete(recipeData, \"TAFFY\")\n\t}\n\n\t// trail mix\n\t// required berries must be uncooked\n\tif countIngName(\"Roasted Birchnut\", crockPot) < 1 || countIngName(\"Berries\", crockPot) < 1 || countIngName(\"Roasted\", crockPot) > 3 || attributeCounts.fruitCount-1 < 1 || attributeCounts.meatCount != 0 || attributeCounts.fishCount != 0 || attributeCounts.eggCount != 0 || attributeCounts.vegeCount != 0 || attributeCounts.dairyCount != 0 {\n\t\tdelete(recipeData, \"TRAIL MIX\")\n\t}\n\n\t// TURKEY DINNER\n\t// checkingg twice for veg or fruit\n\tif countIngName(\"Drumstick\", crockPot) < 2 || attributeVals.meatVal-1 < 0.25 || (attributeVals.fruitVal < .5 || attributeVals.vegVal < .5) {\n\t\tdelete(recipeData, \"TURKEY DINNER\")\n\t}\n\n\t// unagi\n\tif countIngName(\"Eel\", crockPot) < 1 || (countIngName(\"Lichen\", crockPot) < 1 || countIngName(\"Kelp Fronds\", crockPot) < 1) {\n\t\tdelete(recipeData, \"UNAGI\")\n\t}\n\n\t// VEGETABLE STINGER\n\tif (countIngName(\"Toma Root\", crockPot) < 1 || countIngName(\"Asparagus\", crockPot) < 1) || countIngName(\"Ice\", crockPot) < 1 || attributeVals.vegVal-1 < 1.5 {\n\t\tdelete(recipeData, \"VEGETABLE STINGER\")\n\t}\n\n\t// Veggie Burger\n\tif countIngName(\"Leafy Meat\", crockPot) < 1 || countIngName(\"Onion\", crockPot) < 1 || attributeVals.vegVal-1 < 1 {\n\t\tdelete(recipeData, \"VEGGIE BURGER\")\n\t}\n\n\t// waffles\n\tif countIngName(\"Butter\", crockPot) < 1 || countIngName(\"Berries\", crockPot) < 1 || attributeCounts.eggCount < 1 {\n\t\tdelete(recipeData, \"WAFFLES\")\n\t}\n\n\t// wet goop\n\t// if everything is false\n\n\t// WOBSTER DINNER\n\tif countIngName(\"Wobster\", crockPot) < 1 || countIngName(\"Butter\", crockPot) < 1 || attributeCounts.meatCount != 0 || attributeCounts.fishCount != 0 || stringInSlice(\"Twigs\", crockPot) {\n\t\tdelete(recipeData, \"WOBSTER DINNER\")\n\t}\n\treturn recipeData\n}", "func withBookcourseID(id int) bookcourseOption {\n\treturn func(m *BookcourseMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Bookcourse\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Bookcourse, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Bookcourse.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func CreateRecipe(ctx context.Context, db *gorm.DB, input model.NewRecipeInput) (*model.Recipe, error) {\n\tuser, err := IsAuthenticated(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif user == nil {\n\t\treturn nil, auth.ErrUnauthenticated\n\t}\n\ttx := db.Begin()\n\tnewRecipe, err := model.NewRecipe(&input, user.ID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttx = tx.Create(newRecipe)\n\tif err := tx.Error; err != nil {\n\t\ttx.Rollback()\n\t\treturn nil, err\n\t}\n\tbuilt, err := model.BuildRecipe(newRecipe)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttx = tx.Commit()\n\tif err := tx.Error; err != nil {\n\t\treturn nil, err\n\t}\n\treturn built, nil\n}", "func withCarregisterID(id int) carregisterOption {\n\treturn func(m *CarregisterMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Carregister\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Carregister, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Carregister.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func withStreetID(id int) streetOption {\n\treturn func(m *StreetMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Street\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Street, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Street.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func (m *Storage) AddSampleRecipes() {\n\n\tfor _, recipe := range SampleMeals {\n\t\tm.recipes = append(m.recipes, recipe)\n\t}\n\n}", "func GetRecipeDetails(ctx context.Context, recipeID int64) (*models.Recipe, error) {\n\tdb, err := mysql.Connect(ctx, constant.DBs.RecipeRolodex)\n\tif err != nil {\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\treturn nil, constant.Errors.DbConnectionFailure\n\t}\n\n\tquery := `SELECT season, title, author, link FROM ` + constant.RR.Recipes +\n\t\t` WHERE id=?`\n\n\tvar details models.Recipe\n\terr = db.QueryRowContext(ctx, query, recipeID).Scan(\n\t\t&details.Season,\n\t\t&details.Title,\n\t\t&details.Author,\n\t\t&details.Link,\n\t)\n\tif err != nil {\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\tif err == sql.ErrNoRows {\n\t\t\treturn nil, constant.Errors.NoRecipeIDFound\n\t\t}\n\t\treturn nil, constant.Errors.DbQueryFailure\n\t}\n\n\tlogger.Info.Printf(\"%s Successfully retrieved recipe details for recipe with ID %v\", logger.GetCallInfo(), recipeID)\n\treturn &details, nil\n}", "func (c *cookRun) runRecipe(ctx context.Context, env environ.Env) *build.BuildRunResult {\n\tresult := &build.BuildRunResult{\n\t\tRecipe: &build.BuildRunResult_Recipe{\n\t\t\tName: c.RecipeName,\n\t\t},\n\t}\n\n\tfail := func(err error) *build.BuildRunResult {\n\t\tif err == nil {\n\t\t\tpanic(\"do not call fail with nil err\")\n\t\t}\n\t\tif result.InfraFailure != nil {\n\t\t\tpanic(\"bug! forgot to return the result on previous error\")\n\t\t}\n\t\tresult.InfraFailure = infraFailure(err)\n\t\treturn result\n\t}\n\n\t// The ready-to-run recipe must already be present on the file system in\n\t// checkout-dir.\n\trecipesPath, err := exec.LookPath(filepath.Join(c.CheckoutDir, \"recipes\"))\n\tif err != nil {\n\t\treturn fail(errors.Annotate(err, \"could not find bundled recipes\").Err())\n\t}\n\t// LookPath can return an absolute OR relative path. Use Abs to make sure.\n\trecipesPath, err = filepath.Abs(recipesPath)\n\tif err != nil {\n\t\treturn fail(errors.Annotate(err, \"could not convert bundled recipes to abspath\").Err())\n\t}\n\tc.engine.cmdPrefix = []string{recipesPath}\n\n\t// Setup our working directory. This is cwd for the recipe itself.\n\t// Previously this was unnecessarily configurable; Now we hard-code it to\n\t// \"$CWD/k\", which is the shortest path we can make. This is important to\n\t// allow tasks on Windows to have as many characters as possible; otherwise\n\t// they run into MAX_PATH issues.\n\tc.engine.workDir, err = prepareRecipeRunWorkDir(\"k\")\n\tif err != nil {\n\t\treturn fail(errors.Annotate(err, \"failed to prepare workdir\").Err())\n\t}\n\n\t// Tell the recipe to write the result protobuf message to a file and read\n\t// it below.\n\tc.engine.outputResultJSONFile = filepath.Join(c.TempDir, \"recipe-result.json\")\n\n\t// Run the recipe in the appropriate auth context by exporting it into the\n\t// environ of the recipe engine.\n\texported, err := lucictx.ExportInto(c.recipeAuth.Export(ctx, env), c.TempDir)\n\tif err != nil {\n\t\treturn fail(errors.Annotate(err, \"failed to export LUCI_CONTEXT\").Err())\n\t}\n\tdefer exported.Close()\n\texported.SetInEnviron(env)\n\n\tresult.AnnotationUrl = c.AnnotationURL.String()\n\tif err = c.runWithLogdogButler(ctx, env, result); err != nil {\n\t\treturn fail(errors.Annotate(err, \"failed to run recipe\").Err())\n\t}\n\tsetAnnotationText(result.Annotations)\n\n\t// Now read the recipe result file.\n\trecipeResultFile, err := os.Open(c.engine.outputResultJSONFile)\n\tif err != nil {\n\t\t// The recipe result file must exist and be readable.\n\t\t// If it is not, it is a fatal error.\n\t\treturn fail(errors.Annotate(err,\n\t\t\t\"could not read recipe result file at %q\", c.engine.outputResultJSONFile).Err())\n\t}\n\tdefer recipeResultFile.Close()\n\n\tif c.RecipeResultByteLimit > 0 {\n\t\tst, err := recipeResultFile.Stat()\n\t\tif err != nil {\n\t\t\treturn fail(errors.Annotate(err,\n\t\t\t\t\"could not stat recipe result file at %q\", c.engine.outputResultJSONFile).Err())\n\t\t}\n\n\t\tif sz := st.Size(); sz > int64(c.RecipeResultByteLimit) {\n\t\t\treturn fail(errors.Reason(\"recipe result file is %d bytes which is more than %d\",\n\t\t\t\tsz, c.RecipeResultByteLimit).Err())\n\t\t}\n\t}\n\n\tresult.RecipeResult = &recipe_engine.Result{}\n\terr = (&jsonpb.Unmarshaler{\n\t\tAllowUnknownFields: true,\n\t}).Unmarshal(recipeResultFile, result.RecipeResult)\n\tif err != nil {\n\t\treturn fail(errors.Annotate(err, \"could not parse recipe result\").Err())\n\t}\n\n\t// TODO(nodir): remove result.\n\tif result.RecipeResult.GetFailure() != nil && result.RecipeResult.GetFailure().GetFailure() == nil {\n\t\t// The recipe run has failed and the failure type is not step failure.\n\t\tresult.InfraFailure = &build.InfraFailure{\n\t\t\tText: fmt.Sprintf(\"recipe infra failure: %s\", result.RecipeResult.GetFailure().HumanReason),\n\t\t\tType: build.InfraFailure_RECIPE_INFRA_FAILURE,\n\t\t}\n\t\treturn result\n\t}\n\n\treturn result\n}", "func CopySecretWithName(corev1Input clientcorev1.CoreV1Interface, srcNS, srcSecretName, tgtNS, tgtSecretName, svcAccount string) (*corev1.Secret, error) {\n\ttgtNamespaceSvcAcct := corev1Input.ServiceAccounts(tgtNS)\n\tsrcSecrets := corev1Input.Secrets(srcNS)\n\ttgtNamespaceSecrets := corev1Input.Secrets(tgtNS)\n\n\t// First try to find the secret we're supposed to copy\n\tsrcSecret, err := srcSecrets.Get(context.Background(), srcSecretName, metav1.GetOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// check for nil source secret\n\tif srcSecret == nil {\n\t\treturn nil, errors.New(\"error copying secret; there is no error but secret is nil\")\n\t}\n\n\t// Found the secret, so now make a copy in our new namespace\n\tnewSecret, err := tgtNamespaceSecrets.Create(\n\t\tcontext.Background(),\n\t\t&corev1.Secret{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tName: tgtSecretName,\n\t\t\t},\n\t\t\tData: srcSecret.Data,\n\t\t\tType: srcSecret.Type,\n\t\t},\n\t\tmetav1.CreateOptions{})\n\n\t// If the secret already exists then that's ok - may have already been created\n\tif err != nil && !apierrs.IsAlreadyExists(err) {\n\t\treturn nil, fmt.Errorf(\"error copying the Secret: %s\", err)\n\t}\n\n\ttgtSvcAccount, err := tgtNamespaceSvcAcct.Get(context.Background(), svcAccount, metav1.GetOptions{})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error getting service account %s: %w\", svcAccount, err)\n\t}\n\n\tfor _, secret := range tgtSvcAccount.ImagePullSecrets {\n\t\tif secret.Name == tgtSecretName {\n\t\t\treturn newSecret, nil\n\t\t}\n\t}\n\t// Prevent overwriting existing imagePullSecrets\n\tpatch := `[{\"op\":\"add\",\"path\":\"/imagePullSecrets/-\",\"value\":{\"name\":\"` + tgtSecretName + `\"}}]`\n\tif len(tgtSvcAccount.ImagePullSecrets) == 0 {\n\t\tpatch = `[{\"op\":\"add\",\"path\":\"/imagePullSecrets\",\"value\":[{\"name\":\"` + tgtSecretName + `\"}]}]`\n\t}\n\t_, err = tgtNamespaceSvcAcct.Patch(context.Background(), svcAccount, types.JSONPatchType,\n\t\t[]byte(patch), metav1.PatchOptions{})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"patch failed on NS/SA (%s/%s): %w\",\n\t\t\ttgtNS, svcAccount, err)\n\t}\n\treturn newSecret, nil\n}", "func (o *RecipeAdditive) SetAdditive(ctx context.Context, exec boil.ContextExecutor, insert bool, related *Additive) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(ctx, exec, boil.Infer()); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE \\\"recipe_additive\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, []string{\"additive_id\"}),\n\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 2, recipeAdditivePrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.ID, o.ID}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, updateQuery)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tif _, err = exec.ExecContext(ctx, updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.AdditiveID = related.ID\n\tif o.R == nil {\n\t\to.R = &recipeAdditiveR{\n\t\t\tAdditive: related,\n\t\t}\n\t} else {\n\t\to.R.Additive = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &additiveR{\n\t\t\tRecipeAdditive: o,\n\t\t}\n\t} else {\n\t\trelated.R.RecipeAdditive = o\n\t}\n\n\treturn nil\n}", "func (cc *Chaincode) addInvestigationID(stub shim.ChaincodeStubInterface, params []string) sc.Response {\n\t// Check Access\n\tcreatorOrg, creatorCertIssuer, err := getTxCreatorInfo(stub)\n\tif !authenticatePolice(creatorOrg, creatorCertIssuer) {\n\t\treturn shim.Error(\"{\\\"Error\\\":\\\"Access Denied!\\\",\\\"Payload\\\":{\\\"MSP\\\":\\\"\" + creatorOrg + \"\\\",\\\"CA\\\":\\\"\" + creatorCertIssuer + \"\\\"}}\")\n\t}\n\n\t// Check if sufficient Params passed\n\tif len(params) != 2 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting 2!\")\n\t}\n\n\t// Check if Params are non-empty\n\tfor a := 0; a < 2; a++ {\n\t\tif len(params[a]) <= 0 {\n\t\t\treturn shim.Error(\"Argument must be a non-empty string\")\n\t\t}\n\t}\n\n\t// Copy the Values from params[]\n\tID := params[0]\n\tNewInvestigationID := params[1]\n\n\t// Check if ChargeSheet exists with Key => ID\n\tchargeSheetAsBytes, err := stub.GetState(ID)\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to get ChargeSheet Details!\")\n\t} else if chargeSheetAsBytes == nil {\n\t\treturn shim.Error(\"Error: ChargeSheet Does NOT Exist!\")\n\t}\n\n\t// Create Update struct var\n\tchargeSheetToUpdate := chargeSheet{}\n\terr = json.Unmarshal(chargeSheetAsBytes, &chargeSheetToUpdate) //unmarshal it aka JSON.parse()\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\t// Check if Judgement is Complete or NOT\n\tif chargeSheetToUpdate.Complete {\n\t\treturn shim.Error(\"Error: ChargeSheet is Complete & Locked!\")\n\t}\n\n\t// Update ChargeSheet.InvestigationIDs to append => NewInvestigationID\n\tchargeSheetToUpdate.InvestigationIDs = append(chargeSheetToUpdate.InvestigationIDs, NewInvestigationID)\n\n\t// Convert to JSON bytes\n\tchargeSheetJSONasBytes, err := json.Marshal(chargeSheetToUpdate)\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\t// Put State of newly generated ChargeSheet with Key => ID\n\terr = stub.PutState(ID, chargeSheetJSONasBytes)\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\t// Returned on successful execution of the function\n\treturn shim.Success(chargeSheetJSONasBytes)\n}", "func (c *RBController) RecipeJSON(w http.ResponseWriter, r *http.Request) (err error) {\n\tvars := mux.Vars(r)\n\tid, _ := strconv.Atoi(vars[\"id\"])\n\trecipe, err := c.GetRecipe(id)\n\tif err == nil {\n\t\tc.JSON(w, http.StatusOK, recipe)\n\t} else if err == sql.ErrNoRows {\n\t\tc.RenderError(w, 404, \"Sorry, your page wasn't found\")\n\t\terr = nil\n\t}\n\treturn\n}", "func (c *Client) Recipe(artisanSlug, recipeSlug string) (*response.Recipe, error) {\n\tvar data *d3.Recipe\n\n\tep := endpointRecipe(c.region, artisanSlug, recipeSlug)\n\n\tq, err := c.get(ep, &data)\n\n\tif nil != err {\n\t\treturn nil, err\n\t}\n\n\treturn &response.Recipe{\n\t\tData: data,\n\t\tEndpoint: ep,\n\t\tQuota: q,\n\t\tRegion: c.region,\n\t}, nil\n}", "func (f ReservationFilter) WithID(id schema.ID) ReservationFilter {\n\treturn append(f, bson.E{Key: \"_id\", Value: id})\n}", "func UpdateWithID(c *server.Context) error {\n\tvar (\n\t\terr error\n\t\tw *ware.Ware\n\t\treq struct {\n\t\t\tID uint32 `json:\"id\" validate:\"required\"`\n\t\t\tName string `json:\"name\"`\n\t\t\tDesc string `json:\"desc\" validate:\"max=50\"`\n\t\t\tParentCategoryID uint32 `json:\"parent_category_id\"`\n\t\t\tCategoryID uint32 `json:\"category_id\"`\n\t\t\tTotalSale uint32 `json:\"total_sale\"`\n\t\t\tAvatar string `json:\"avatar\"`\n\t\t\tImage string `json:\"image\"`\n\t\t\tDetailPic string `json:\"detail_pic\"`\n\t\t\tInventory uint32 `json:\"inventory\"`\n\t\t}\n\t)\n\n\tisAdmin := c.Request().Context().Value(\"user\").(jwtgo.MapClaims)[util.IsAdmin].(bool)\n\tif !isAdmin {\n\t\tlogger.Error(\"You don't have access\")\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrToken, nil)\n\t}\n\n\terr = c.JSONBody(&req)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInvalidParam, nil)\n\t}\n\n\terr = c.Validate(req)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInvalidParam, nil)\n\t}\n\n\tconn, err := mysql.Pool.Get()\n\tdefer mysql.Pool.Release(conn)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrMysql, nil)\n\t}\n\n\tw, err = ware.Service.GetByID(conn, req.ID)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrMysql, nil)\n\t}\n\n\tif len(req.Avatar) > 0 {\n\t\terr = util.UpdatePic(req.Avatar, w.Avatar)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInternalServerError, nil)\n\t\t}\n\t}\n\tif len(req.Image) > 0 {\n\t\terr = util.UpdatePic(req.Image, w.Image)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInternalServerError, nil)\n\t\t}\n\t}\n\tif len(req.DetailPic) > 0 {\n\t\terr = util.UpdatePic(req.DetailPic, w.DetailPic)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInternalServerError, nil)\n\t\t}\n\t}\n\n\tw.Name = req.Name\n\tw.CategoryID = req.CategoryID\n\tw.Desc = req.Desc\n\tw.ParentCategoryID = req.ParentCategoryID\n\tw.TotalSale = req.TotalSale\n\tw.Inventory = req.Inventory\n\n\terr = ware.Service.UpdateWare(conn, w)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrMysql, nil)\n\t}\n\n\treturn core.WriteStatusAndDataJSON(c, constants.ErrSucceed, nil)\n}", "func (r *Replicator) ReplicateObjectTo(sourceObj interface{}, target *v1.Namespace) error {\n\tsource := sourceObj.(*v1.Secret)\n\ttargetLocation := fmt.Sprintf(\"%s/%s\", target.Name, source.Name)\n\n\tlogger := log.\n\t\tWithField(\"kind\", r.Kind).\n\t\tWithField(\"source\", common.MustGetKey(source)).\n\t\tWithField(\"target\", targetLocation)\n\n\ttargetResourceType := source.Type\n\ttargetResource, exists, err := r.Store.GetByKey(targetLocation)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"Could not get %s from cache!\", targetLocation)\n\t}\n\tlogger.Infof(\"Checking if %s exists? %v\", targetLocation, exists)\n\n\tvar resourceCopy *v1.Secret\n\tif exists {\n\t\ttargetObject := targetResource.(*v1.Secret)\n\t\ttargetVersion, ok := targetObject.Annotations[common.ReplicatedFromVersionAnnotation]\n\t\tsourceVersion := source.ResourceVersion\n\n\t\tif ok && targetVersion == sourceVersion {\n\t\t\tlogger.Debugf(\"Secret %s is already up-to-date\", common.MustGetKey(targetObject))\n\t\t\treturn nil\n\t\t}\n\n\t\ttargetResourceType = targetObject.Type\n\t\tresourceCopy = targetObject.DeepCopy()\n\t} else {\n\t\tresourceCopy = new(v1.Secret)\n\t}\n\n\tkeepOwnerReferences, ok := source.Annotations[common.KeepOwnerReferences]\n\tif ok && keepOwnerReferences == \"true\" {\n\t\tresourceCopy.OwnerReferences = source.OwnerReferences\n\t}\n\n\tif resourceCopy.Data == nil {\n\t\tresourceCopy.Data = make(map[string][]byte)\n\t}\n\tif resourceCopy.Annotations == nil {\n\t\tresourceCopy.Annotations = make(map[string]string)\n\t}\n\n\treplicatedKeys := r.extractReplicatedKeys(source, targetLocation, resourceCopy)\n\n\tsort.Strings(replicatedKeys)\n\n\tlabelsCopy := make(map[string]string)\n\n\tstripLabels, ok := source.Annotations[common.StripLabels]\n\tif !ok && stripLabels != \"true\" {\n\t\tif source.Labels != nil {\n\t\t\tfor key, value := range source.Labels {\n\t\t\t\tlabelsCopy[key] = value\n\t\t\t}\n\t\t}\n\t}\n\n\tresourceCopy.Name = source.Name\n\tresourceCopy.Labels = labelsCopy\n\tresourceCopy.Type = targetResourceType\n\tresourceCopy.Annotations[common.ReplicatedAtAnnotation] = time.Now().Format(time.RFC3339)\n\tresourceCopy.Annotations[common.ReplicatedFromVersionAnnotation] = source.ResourceVersion\n\tresourceCopy.Annotations[common.ReplicatedKeysAnnotation] = strings.Join(replicatedKeys, \",\")\n\n\tvar obj interface{}\n\tif exists {\n\t\tlogger.Debugf(\"Updating existing secret %s/%s\", target.Name, resourceCopy.Name)\n\t\tobj, err = r.Client.CoreV1().Secrets(target.Name).Update(context.TODO(), resourceCopy, metav1.UpdateOptions{})\n\t} else {\n\t\tlogger.Debugf(\"Creating a new secret secret %s/%s\", target.Name, resourceCopy.Name)\n\t\tobj, err = r.Client.CoreV1().Secrets(target.Name).Create(context.TODO(), resourceCopy, metav1.CreateOptions{})\n\t}\n\tif err != nil {\n\t\terr = errors.Wrapf(err, \"Failed to update secret %s/%s\", target.Name, resourceCopy.Name)\n\t} else if err = r.Store.Update(obj); err != nil {\n\t\terr = errors.Wrapf(err, \"Failed to update cache for %s/%s\", target.Name, resourceCopy)\n\t}\n\n\treturn err\n}", "func Get(recipe *models.RecipeModel, id string) (err error) {\n\tif err = Config.DB.Where(\"id = ?\", id).First(recipe).Error; err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithJSONBody(jSONBody *models.Recipe) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetJSONBody(jSONBody)\n\treturn o\n}", "func CreateFromID(id int32, isDrop bool) (Item, bool) {\n\tnewItem := Item{}\n\n\tif _, ok := nx.Items[id]; !ok {\n\t\treturn Item{}, false\n\t}\n\n\tnxInfo := nx.Items[id]\n\n\tnewItem.UUID = uuid.Must(uuid.NewRandom())\n\tnewItem.InvID = byte(id / 1e6)\n\tnewItem.ItemID = id\n\tnewItem.Accuracy = nxInfo.Accuracy\n\tnewItem.Avoid = nxInfo.Evasion\n\n\tnewItem.Matk = nxInfo.MagicAttack\n\tnewItem.Mdef = nxInfo.MagicDefence\n\tnewItem.Watk = nxInfo.WeaponAttack\n\tnewItem.Wdef = nxInfo.WeaponDefence\n\n\tnewItem.Str = nxInfo.Str\n\tnewItem.Dex = nxInfo.Dex\n\tnewItem.Int = nxInfo.Int\n\tnewItem.Luk = nxInfo.Luk\n\n\tnewItem.ReqLevel = nxInfo.ReqLevel\n\tnewItem.UpgradeSlots = nxInfo.Upgrades\n\n\tnewItem.Amount = 1\n\n\tlog.Println(\"Finish create item from ID function\", newItem)\n\n\treturn newItem, true\n}", "func (j JID) Copy() JID {\n\treturn j\n}", "func FindRecipeAdditive(ctx context.Context, exec boil.ContextExecutor, iD int, selectCols ...string) (*RecipeAdditive, error) {\n\trecipeAdditiveObj := &RecipeAdditive{}\n\n\tsel := \"*\"\n\tif len(selectCols) > 0 {\n\t\tsel = strings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, selectCols), \",\")\n\t}\n\tquery := fmt.Sprintf(\n\t\t\"select %s from \\\"recipe_additive\\\" where \\\"id\\\"=$1\", sel,\n\t)\n\n\tq := queries.Raw(query, iD)\n\n\terr := q.Bind(ctx, exec, recipeAdditiveObj)\n\tif err != nil {\n\t\tif errors.Cause(err) == sql.ErrNoRows {\n\t\t\treturn nil, sql.ErrNoRows\n\t\t}\n\t\treturn nil, errors.Wrap(err, \"models: unable to select from recipe_additive\")\n\t}\n\n\treturn recipeAdditiveObj, nil\n}", "func withRoadID(id int) roadOption {\n\treturn func(m *RoadMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Road\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Road, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Road.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func (pu *PendingloanbindingUpdate) AddFileimportID(i int) *PendingloanbindingUpdate {\n\tpu.mutation.AddFileimportID(i)\n\treturn pu\n}", "func (rc *RentalCreate) SetCarID(u uuid.UUID) *RentalCreate {\n\trc.mutation.SetCarID(u)\n\treturn rc\n}", "func (c *Client) Copy(src FileID, dst string) (FileID, error) {\n\treturn c.CopyWithContext(context.Background(), src, dst)\n}" ]
[ "0.615391", "0.599917", "0.59268034", "0.5921865", "0.59162635", "0.55177057", "0.5504684", "0.55040866", "0.54498166", "0.5376762", "0.5370218", "0.5241743", "0.52196133", "0.51959723", "0.51737744", "0.51557696", "0.5140836", "0.5042319", "0.5009742", "0.49966857", "0.49725023", "0.49621832", "0.49527502", "0.49525237", "0.49435815", "0.4940204", "0.493994", "0.493994", "0.49144047", "0.49084982", "0.49063814", "0.48571196", "0.48571196", "0.47555086", "0.47480747", "0.47447953", "0.473472", "0.46827728", "0.46278554", "0.4619091", "0.46102396", "0.45997623", "0.45997623", "0.45892903", "0.45871252", "0.45403004", "0.45359597", "0.45301524", "0.45233616", "0.4519417", "0.4519417", "0.45142537", "0.45008183", "0.44819453", "0.44651037", "0.44573596", "0.44488597", "0.44405115", "0.44359934", "0.44336587", "0.44329673", "0.44311476", "0.44293344", "0.4409243", "0.4406762", "0.43721727", "0.43517116", "0.4342994", "0.433717", "0.4331834", "0.43238336", "0.4318565", "0.42889178", "0.42812046", "0.42610234", "0.42521074", "0.42468873", "0.42467934", "0.42451903", "0.4240111", "0.42392808", "0.42325723", "0.4223583", "0.42232475", "0.4211341", "0.4208364", "0.4196575", "0.41962636", "0.41959286", "0.41788617", "0.41719985", "0.4167242", "0.41605538", "0.41575953", "0.4155871", "0.41540992", "0.41510785", "0.41467118", "0.4137722", "0.41362268" ]
0.6474311
0
SetRecipeID adds the recipeId to the copy recipe to my recipes with changes params
func (o *CopyRecipeToMyRecipesWithChangesParams) SetRecipeID(recipeID string) { o.RecipeID = recipeID }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *RecipeLipid) SetRecipe(ctx context.Context, exec boil.ContextExecutor, insert bool, related *Recipe) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(ctx, exec, boil.Infer()); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE \\\"recipe_lipid\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, []string{\"recipe_id\"}),\n\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 2, recipeLipidPrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.ID, o.ID}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, updateQuery)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tif _, err = exec.ExecContext(ctx, updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.RecipeID = related.ID\n\tif o.R == nil {\n\t\to.R = &recipeLipidR{\n\t\t\tRecipe: related,\n\t\t}\n\t} else {\n\t\to.R.Recipe = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &recipeR{\n\t\t\tRecipeLipids: RecipeLipidSlice{o},\n\t\t}\n\t} else {\n\t\trelated.R.RecipeLipids = append(related.R.RecipeLipids, o)\n\t}\n\n\treturn nil\n}", "func (o *RecipeAdditive) SetRecipe(ctx context.Context, exec boil.ContextExecutor, insert bool, related *Recipe) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(ctx, exec, boil.Infer()); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE \\\"recipe_additive\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, []string{\"recipe_id\"}),\n\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 2, recipeAdditivePrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.ID, o.ID}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, updateQuery)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tif _, err = exec.ExecContext(ctx, updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.RecipeID = related.ID\n\tif o.R == nil {\n\t\to.R = &recipeAdditiveR{\n\t\t\tRecipe: related,\n\t\t}\n\t} else {\n\t\to.R.Recipe = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &recipeR{\n\t\t\tRecipeAdditives: RecipeAdditiveSlice{o},\n\t\t}\n\t} else {\n\t\trelated.R.RecipeAdditives = append(related.R.RecipeAdditives, o)\n\t}\n\n\treturn nil\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithRecipeID(recipeID string) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetRecipeID(recipeID)\n\treturn o\n}", "func (r *RecipeInfo) newRecipe(id int) error {\n\t// Get recipe from the database\n\trecipe, err := new(models.Recipe).GetByID(database, queries, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Setup Recipe info\n\tr.ID = recipe.ID\n\tr.TotalSteps = len(recipe.Steps)\n\tr.CurrentStep = recipe.Steps[0]\n\tr.PrevStep = nil\n\tr.NextStep = nil\n\tr.JobIDs = make([]int64, 0)\n\n\tif r.TotalSteps > 1 {\n\t\tr.NextStep = recipe.Steps[1]\n\t}\n\n\tr.recipe = recipe\n\tdone, err := r.initStep(0)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif done {\n\t\tlog.Warn(\"Just setup a newRecipe that is already done\")\n\t}\n\treturn err\n}", "func (c *RBController) SaveRecipe(w http.ResponseWriter, r *http.Request) (err error) {\n\t// Get properties\n\tname := r.PostFormValue(`name`)\n\tcuisine, err := strconv.Atoi(r.PostFormValue(`cuisine`))\n\n\t// Get the mealtype and season encoded ints\n\tmealtype := EncodeMealtype(r.Form[`mealtype`])\n\tseason := EncodeSeason(r.Form[`season`])\n\n\t// get everything else\n\tdescription := r.PostFormValue(`description`)\n\tingredients := r.PostFormValue(`ingredients`)\n\tinstructions := r.PostFormValue(`instructions`)\n\n\t// TODO better error handling\n\tif err != nil {\n\t\tfmt.Println(\"[WARNING] Something went wrong in SaveRecipe\")\n\t\tc.RenderError(w, 500, \"Sorry, something went wrong.\")\n\t\treturn\n\t}\n\n\t// everything OK: build the recipe, and send it to the database\n\trecipe := Recipe{ID: 0, Name: name, Cuisine: cuisine, Mealtype: mealtype,\n\t\tSeason: season, Description: description, Ingredientlist: ingredients,\n\t\tInstructions: instructions}\n\n\t// if we don't have the id string, then this is a new request.\n\tvars := mux.Vars(r)\n\tidStr := vars[\"id\"]\n\tid := 0\n\n\tif idStr != \"\" {\n\t\tid, _ = strconv.Atoi(idStr)\n\t\trecipe.ID = id\n\t\terr = c.RecipeDB.UpdateRecipe(&recipe)\n\t} else {\n\t\tid, err = c.RecipeDB.NewRecipe(&recipe)\n\t}\n\n\tif err == nil {\n\t\thttp.Redirect(w, r, \"/recipes/\"+fmt.Sprintf(\"%v\", id)+\"/\", http.StatusFound)\n\t}\n\treturn\n}", "func (o *RecipeLipid) SetLipid(ctx context.Context, exec boil.ContextExecutor, insert bool, related *Lipid) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(ctx, exec, boil.Infer()); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE \\\"recipe_lipid\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, []string{\"lipid_id\"}),\n\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 2, recipeLipidPrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.ID, o.ID}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, updateQuery)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tif _, err = exec.ExecContext(ctx, updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.LipidID = related.ID\n\tif o.R == nil {\n\t\to.R = &recipeLipidR{\n\t\t\tLipid: related,\n\t\t}\n\t} else {\n\t\to.R.Lipid = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &lipidR{\n\t\t\tRecipeLipid: o,\n\t\t}\n\t} else {\n\t\trelated.R.RecipeLipid = o\n\t}\n\n\treturn nil\n}", "func AddRecipe(recipe *Recipe) *Recipe {\n\tuuid, _ := uuid.NewRandom()\n\trecipe.ID = uuid.String()\n\trecipe.CreatedAt = time.Now().UTC().String()\n\trecipe.UpdatedAt = time.Now().UTC().String()\n\n\trecipeList = append(recipeList, recipe)\n\treturn recipe\n}", "func UpdateRecipe(c *gin.Context) {\n\tvar recipe model.Recipe\n\tif err := c.ShouldBind(&recipe); err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"illegal format for recipe: \" + err.Error()})\n\t\treturn\n\t}\n\n\tid, err := strconv.ParseUint(c.Param(\"id\"), 10, 64)\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"invalid id-format: \" + c.Param(\"id\")})\n\t\treturn\n\t}\n\trecipe.ID = uint(id)\n\n\tif recipe.ID == 0 {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"cannot update recipe without id\"})\n\t\treturn\n\t}\n\n\tif recipe.Title == \"\" || recipe.Resources == nil || len(recipe.Resources) == 0 {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"illegal content for recipe\"})\n\t\treturn\n\t}\n\n\terr = recipe.Update()\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": err.Error()})\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, gin.H{\"data\": recipe.ID})\n}", "func (c *RBController) Recipe(w http.ResponseWriter, r *http.Request) (err error) {\n\tvars := mux.Vars(r)\n\tid, _ := strconv.Atoi(vars[\"id\"])\n\trecipe, err := c.GetRecipe(id)\n\tif err == nil {\n\t\tc.HTML(w, http.StatusOK, \"recipes/recipe\", recipe)\n\t} else if err == sql.ErrNoRows {\n\t\t// this means that the recipe wasn't found, so we should return a 404 error\n\t\tc.RenderError(w, 404, \"Sorry, your page wasn't found\")\n\t\terr = nil\n\t}\n\treturn\n}", "func (ds CBDataStore) RecipeUpdate(modelRecipe model.Recipe, id, userID string) error {\n\t_, err := ds.bucket.Replace(modelRecipe.ID, modelRecipe, 0, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (ds CBDataStore) RecipeUpdate(modelRecipe model.Recipe, id, userID string) error {\n\t_, err := ds.bucket.Replace(modelRecipe.Id, modelRecipe, 0, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (workItemType *WorkItemType) SetID(id string) error {\n workItemType.ID = bson.ObjectIdHex(id)\n return nil\n}", "func (o *GetIngredientVersionRevisionParams) SetIngredientID(ingredientID strfmt.UUID) {\n\to.IngredientID = ingredientID\n}", "func (o *GetRecipeInformation200ResponseExtendedIngredientsInner) SetId(v int32) {\n\to.Id = v\n}", "func (ds CBDataStore) Recipe(id, userID string) (*model.Recipe, error) {\n\tr := recipe{}\n\t_, err := ds.bucket.Get(id, &r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif r.UserId != userID {\n\t\treturn nil, fmt.Errorf(\"\")\n\t}\n\n\treturn r.Recipe, nil\n}", "func (ds CBDataStore) Recipe(id, userID string) (*model.Recipe, error) {\n\tr := recipe{}\n\t_, err := ds.bucket.Get(id, &r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif r.UserID != userID {\n\t\treturn nil, fmt.Errorf(\"\")\n\t}\n\n\treturn r.Recipe, nil\n}", "func UpdateRecipe(id string, recipe *Recipe) (*Recipe, error) {\n\tindex, err := findRecipe(id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trecipeList[index] = recipe\n\treturn recipe, nil\n}", "func (tx *Transaction) SetID() error {\n\tvar encoded bytes.Buffer\n\tvar hash [32]byte\n\n\tenc := json.NewEncoder(&encoded)\n\tif err := enc.Encode(tx); err != nil {\n\t\treturn err\n\t}\n\n\thash = sha256.Sum256(encoded.Bytes())\n\ttx.id = hash[:]\n\treturn nil\n}", "func (rc *RentalCreate) SetCarID(u uuid.UUID) *RentalCreate {\n\trc.mutation.SetCarID(u)\n\treturn rc\n}", "func (db *DB) SaveRecipe(ctx context.Context, dr *domain.Recipe) error {\n\tvar r Recipe\n\terr := r.FromDomain(dr)\n\tif err != nil {\n\t\treturn err\n\t}\n\tnow := time.Now()\n\tr.CretedAt, r.ModifiedAt = &now, &now\n\n\tvar sb strings.Builder\n\t// t := template.Must(template.New(\"save.tmpl\").Funcs(fm).ParseFiles(\"../../../templates/dgraph/save.tmpl\"))\n\tt := template.Must(template.New(\"save.tmpl\").Funcs(fm).ParseFiles(\"/templates/dgraph/save.tmpl\"))\n\terr = t.Execute(&sb, dr)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treq := &api.Request{CommitNow: true}\n\treq.Vars = map[string]string{\"$xid\": dr.ExternalID}\n\treq.Query = sb.String()\n\n\tmutations := make([]*api.Mutation, 0, len(dr.Ingredients)*2+len(dr.Tags)*2+1)\n\n\t// keep any food and tag\n\tfor i, di := range dr.Ingredients {\n\t\tvar i0, i1 Ingredient\n\t\terr := i0.FromDomain(di)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = i1.FromDomain(di)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tid := fmt.Sprintf(\"_:i%d\", i)\n\t\ti0.ID, i1.ID = id, id\n\t\tr.Ingredients[i] = &Ingredient{ID: id} // only id, empty fields\n\n\t\t// food stem found\n\t\ti0.Food.ID = fmt.Sprintf(\"uid(f%d)\", i)\n\t\tji0, err := json.Marshal(i0)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tmu0 := &api.Mutation{\n\t\t\tSetJson: ji0,\n\t\t\tCond: fmt.Sprintf(\"@if(eq(len(r), 0) AND eq(len(f%d), 1))\", i),\n\t\t}\n\n\t\t// food stem not found\n\t\ti1.Food.ID = fmt.Sprintf(\"_:f%d\", i)\n\t\tji1, err := json.Marshal(i1)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tmu1 := &api.Mutation{\n\t\t\tSetJson: ji1,\n\t\t\tCond: fmt.Sprintf(\"@if(eq(len(r), 0) AND eq(len(f%d), 0))\", i),\n\t\t}\n\n\t\tmutations = append(mutations, mu0, mu1)\n\t}\n\n\t// using nquads to be able to directly link tag to recipe\n\tfor i := range dr.Tags {\n\t\t// tag name found\n\t\tnq := &api.NQuad{\n\t\t\tSubject: \"_:recipe\",\n\t\t\tPredicate: \"tags\",\n\t\t\tObjectId: fmt.Sprintf(\"uid(t%d)\", i),\n\t\t}\n\n\t\tmu0 := &api.Mutation{\n\t\t\tSet: []*api.NQuad{nq},\n\t\t\tCond: fmt.Sprintf(\"@if(eq(len(r), 0) AND eq(len(t%d), 1))\", i),\n\t\t}\n\n\t\t// tag name not found\n\t\tvar t Tag\n\t\terr := t.FromDomain(dr.Tags[i])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\ttag := fmt.Sprintf(\"_:tag%d)\", i)\n\t\tnq0 := &api.NQuad{\n\t\t\tSubject: \"_:recipe\",\n\t\t\tPredicate: \"tags\",\n\t\t\tObjectId: tag,\n\t\t}\n\t\tnq1 := &api.NQuad{\n\t\t\tSubject: tag,\n\t\t\tPredicate: \"tagName\",\n\t\t\tObjectValue: &api.Value{Val: &api.Value_StrVal{StrVal: t.TagName}},\n\t\t}\n\t\tnq2 := &api.NQuad{\n\t\t\tSubject: tag,\n\t\t\tPredicate: \"tagStem\",\n\t\t\tObjectValue: &api.Value{Val: &api.Value_StrVal{StrVal: t.TagStem}},\n\t\t}\n\t\tnq3 := &api.NQuad{\n\t\t\tSubject: tag,\n\t\t\tPredicate: \"dgraph.type\",\n\t\t\tObjectValue: &api.Value{Val: &api.Value_StrVal{StrVal: t.DType[0]}},\n\t\t}\n\n\t\tmu1 := &api.Mutation{\n\t\t\tSet: []*api.NQuad{nq0, nq1, nq2, nq3},\n\t\t\tCond: fmt.Sprintf(\"@if(eq(len(r), 0) AND eq(len(t%d), 0))\", i),\n\t\t}\n\n\t\tmutations = append(mutations, mu0, mu1)\n\t}\n\n\tr.ID = \"_:recipe\"\n\tr.Tags = nil // don't overwrite tags\n\tjr, err := json.Marshal(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\tmu := &api.Mutation{\n\t\tSetJson: jr,\n\t\tCond: \"@if(eq(len(r), 0))\",\n\t}\n\tmutations = append(mutations, mu)\n\n\treq.Mutations = mutations\n\n\tres, err := db.Dgraph.NewTxn().Do(ctx, req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif ruid, created := res.Uids[\"recipe\"]; created {\n\t\tdr.ID = ruid\n\t} else {\n\t\treturn errors.ErrDuplicateID{ID: dr.ExternalID}\n\t}\n\n\treturn nil\n}", "func (tx *Transaction) SetID() {\n\tvar encoded bytes.Buffer\n\tvar hash [32]byte\n\tencoder := gob.NewEncoder(&encoded)\n\terr := encoder.Encode(tx)\n\tHandle(err)\n\thash = sha256.Sum256(encoded.Bytes())\n\ttx.ID = hash[:]\n}", "func AddRecipeLipidHook(hookPoint boil.HookPoint, recipeLipidHook RecipeLipidHook) {\n\tswitch hookPoint {\n\tcase boil.BeforeInsertHook:\n\t\trecipeLipidBeforeInsertHooks = append(recipeLipidBeforeInsertHooks, recipeLipidHook)\n\tcase boil.BeforeUpdateHook:\n\t\trecipeLipidBeforeUpdateHooks = append(recipeLipidBeforeUpdateHooks, recipeLipidHook)\n\tcase boil.BeforeDeleteHook:\n\t\trecipeLipidBeforeDeleteHooks = append(recipeLipidBeforeDeleteHooks, recipeLipidHook)\n\tcase boil.BeforeUpsertHook:\n\t\trecipeLipidBeforeUpsertHooks = append(recipeLipidBeforeUpsertHooks, recipeLipidHook)\n\tcase boil.AfterInsertHook:\n\t\trecipeLipidAfterInsertHooks = append(recipeLipidAfterInsertHooks, recipeLipidHook)\n\tcase boil.AfterSelectHook:\n\t\trecipeLipidAfterSelectHooks = append(recipeLipidAfterSelectHooks, recipeLipidHook)\n\tcase boil.AfterUpdateHook:\n\t\trecipeLipidAfterUpdateHooks = append(recipeLipidAfterUpdateHooks, recipeLipidHook)\n\tcase boil.AfterDeleteHook:\n\t\trecipeLipidAfterDeleteHooks = append(recipeLipidAfterDeleteHooks, recipeLipidHook)\n\tcase boil.AfterUpsertHook:\n\t\trecipeLipidAfterUpsertHooks = append(recipeLipidAfterUpsertHooks, recipeLipidHook)\n\t}\n}", "func (t *Transaction) SetID() {\n\t// Create some data as a buffer and a hash variable\n\tvar d bytes.Buffer\n\tvar h [32]byte\n\n\t// Create a new encoder, passing the data to it\n\tvar e = gob.NewEncoder(&d)\n\n\t// Encode the transaction, handling any errors\n\terr := e.Encode(t)\n\tHandleError(err)\n\n\t// Create a hash with the datas bytes and assign to the transaction\n\th = sha256.Sum256(d.Bytes())\n\tt.ID = h[:]\n\n}", "func (r *Recipe) UpdateRecipe(db *sql.DB) (res sql.Result, err error) {\n\tfmt.Printf(\"Entro aqui malditasea\")\n\tres, err = db.Exec(\"UPDATE recipes SET name=$1, preptime=$2, difficulty=$3, vegetarian=$4 WHERE id=$5\",\n\t\tr.Name, r.PrepTime, r.Difficulty, r.Vegetarian, r.ID)\n\treturn res, err\n}", "func SetID(clientID string) {\n\tfmt.Println(\"Setting clientID\")\n\ttID = clientID\n}", "func (m *Storage) DeleteRecipe(ID int) {\n\tfor i, recipe := range m.recipes {\n\t\tif ID == recipe.ID {\n\t\t\tm.recipes = append(m.recipes[:i], m.recipes[i+1:]...)\n\t\t\tbreak\n\t\t}\n\t}\n}", "func (tx *Transaction) SetID() {\n\tvar encoded bytes.Buffer\n\tvar hash [32]byte\n\n\tencode := gob.NewEncoder(&encoded)\n\terr := encode.Encode(tx)\n\tif err != nil {\n\t\tlog.Panic()\n\t}\n\n\thash = sha256.Sum256(encoded.Bytes())\n\ttx.ID = hash[:]\n}", "func (m *PaymentTerm) SetId(value *i561e97a8befe7661a44c8f54600992b4207a3a0cf6770e5559949bc276de2e22.UUID)() {\n err := m.GetBackingStore().Set(\"id\", value)\n if err != nil {\n panic(err)\n }\n}", "func SetID(ctx context.Context, requestID string) context.Context {\n\treturn context.WithValue(ctx, requestIDKey, requestID)\n}", "func (ds CBDataStore) RecipeDelete(id, userID string) error {\n\tfrag, err := ds.bucket.LookupIn(id).Get(\"user_id\").Execute()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar recipeUserID string\n\terr = frag.Content(\"user_id\", &recipeUserID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif recipeUserID != userID {\n\t\treturn fmt.Errorf(\"recipe does not exist\")\n\t}\n\n\t_, err = ds.bucket.Remove(id, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (ds CBDataStore) RecipeDelete(id, userID string) error {\n\tfrag, err := ds.bucket.LookupIn(id).Get(\"user_id\").Execute()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar recipeUserID string\n\terr = frag.Content(\"user_id\", &recipeUserID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif recipeUserID != userID {\n\t\treturn fmt.Errorf(\"recipe does not exist\")\n\t}\n\n\t_, err = ds.bucket.Remove(id, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func DeleteRecipe(c *gin.Context) {\n\tvar recipe model.Recipe\n\tid, err := strconv.ParseUint(c.Param(\"id\"), 10, 64)\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"invalid id-format: \" + c.Param(\"id\")})\n\t\treturn\n\t}\n\trecipe.ID = uint(id)\n\terr = recipe.Delete()\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": err.Error()})\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, gin.H{\"message\": \"recipe was deleted succesfully\"})\n}", "func (m *ParentLabelDetails) SetId(value *string)() {\n err := m.GetBackingStore().Set(\"id\", value)\n if err != nil {\n panic(err)\n }\n}", "func (rmd ResourceMetaData) SetID(formatter resourceids.Id) {\n\trmd.ResourceData.SetId(formatter.ID())\n}", "func (m *Storage) AddRecipe(r adding.Recipe) error {\n\n\tnewR := Recipe{\n\t\tID: len(m.recipes) + 1,\n\t\tCreated: time.Now(),\n\t\tMealType: r.MealType,\n\t\tName: r.Name,\n\t\tIngredients: r.Ingredients,\n\t\tPreparation: r.Preparation,\n\t}\n\tm.recipes = append(m.recipes, newR)\n\n\treturn nil\n}", "func PostRecipeDetails(ctx context.Context, newRecipe *models.NewRecipe) error {\n\tdb, err := mysql.Connect(ctx, constant.DBs.RecipeRolodex)\n\tif err != nil {\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\treturn constant.Errors.DbConnectionFailure\n\t}\n\n\ttx, err := db.BeginTx(ctx, nil)\n\tif err != nil {\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\treturn constant.Errors.InternalServer\n\t}\n\n\t// Add recipe details to Recipes table\n\terr = InsertRecipeDetails(ctx, tx, *newRecipe.Season, *newRecipe.Title, *newRecipe.Author, *newRecipe.Link)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\treturn err\n\t}\n\n\t// Add ingredients to Ingredients table if they don't already exist\n\tif len(newRecipe.IngredientList) != 0 {\n\t\terr = InsertIngredients(ctx, tx, newRecipe.IngredientList)\n\t\tif err != nil {\n\t\t\ttx.Rollback()\n\t\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\t\treturn err\n\t\t}\n\t}\n\n\terr = InsertLink(ctx, tx, *newRecipe.Title, newRecipe.IngredientList)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\treturn err\n\t}\n\n\terr = tx.Commit()\n\tif err != nil {\n\t\ttx.Rollback()\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\treturn constant.Errors.DbInsertFailure\n\t}\n\n\treturn nil\n}", "func (o *PostMenuItemParams) SetRecipeGroupID(recipeGroupID int64) {\n\to.RecipeGroupID = recipeGroupID\n}", "func SetHookId(etid int64, hook_id int64) error {\n\tvar dummy string\n\tif err := db.QueryRow(\"UPDATE event_tasks SET hook_id=$1 WHERE id=$2 \"+\n\t\t\"RETURNING id\", hook_id, etid).Scan(&dummy); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (p *Process) CmdSetID(pac teoapi.Packet) (err error) {\n\tdata := pac.RemoveTrailingZero(pac.Data())\n\trequest := cdb.KeyValue{Cmd: pac.Cmd()}\n\tif err = request.UnmarshalText(data); err != nil {\n\t\treturn\n\t} else if err = p.tcdb.SetID(request.Key, request.Value); err != nil {\n\t\treturn\n\t}\n\t// Return only Value for text requests and all fields for json\n\tresponce := request\n\tresponce.Value = nil\n\tif !request.RequestInJSON {\n\t\t_, err = p.tcdb.con.SendAnswer(pac, pac.Cmd(), responce.Value)\n\t} else if retdata, err := responce.MarshalText(); err == nil {\n\t\t_, err = p.tcdb.con.SendAnswer(pac, pac.Cmd(), retdata)\n\t}\n\treturn\n}", "func TestAddRecipe(t *testing.T) {\n\tpayload := fmt.Sprintf(`\n {\n \"mealtype\": \"Breakfast\",\n \"name\": \"Pancakes\",\n \"Ingredients\": [ \"150g all purpose flour\",\n \t\t\t\t \"150ml of milk\"],\n \"preparation\": \"Add all ingredients and mix. Put in Pan.\"\n}`)\n\n\tresponse, err := http.Post(baseURL+\"/recipes\", \"application/json\", strings.NewReader(payload))\n\tif err != nil {\n\t\tt.Fatalf(\"failed to get json, %s\", err)\n\t}\n\n\tcheckResponseCode(t, http.StatusOK, response.StatusCode)\n}", "func AddRecipe(db *database.DB, recipe *Recipe) {\n\tdb.Create(recipe)\n}", "func (tcdb *Teocdb) SetID(key string, value []byte) (err error) {\n\tnextID, err := strconv.Atoi(string(value))\n\tif err != nil {\n\t\treturn\n\t}\n\treturn tcdb.session.Query(`UPDATE ids SET next_id = ? WHERE id_name = ?`,\n\t\tnextID, key).Exec()\n}", "func adminRecipe(w http.ResponseWriter, r *http.Request, param string) {\n\n tmpl, _ := template.ParseFiles(\"admin_header.html\", \"admin_recipe.html\", \"admin_footer.html\")\n\n // Open database\n db := getDBConnection()\n defer db.Close()\n r.ParseForm()\n \n recipe_id, err := strconv.Atoi(r.Form.Get(\"recipe_selection\"))\n if err != nil {\n recipe_id = -1 \n } \n\n if (param == \"add_drink\") {\n // returned form is receipe_name=<drink name entered>\n if len(r.Form.Get(\"recipe_add\")) <= 1 {\n http.Redirect(w, r, \"/admin/recipe/\", http.StatusSeeOther)\n return\n }\n \n // Get glass selection\n glass_type_id, err := strconv.Atoi(r.Form.Get(\"glass_selection\"))\n if err != nil {\n http.Redirect(w, r, \"/admin/recipe/\", http.StatusSeeOther)\n return\n } \n\n _, err = db.Exec(\"insert into recipe (name, glass_type_id) values (?, ?)\", r.Form.Get(\"recipe_add\"), glass_type_id)\n if err != nil {\n panic(fmt.Sprintf(\"Failed to update db: %v\", err))\n }\n \n // get inserted id\n row := db.QueryRow(\"select max(id) from recipe\")\n err = row.Scan(&recipe_id)\n if err != nil {\n http.Redirect(w, r, \"/admin/recipe/\", http.StatusSeeOther)\n return\n }\n \n // http.Redirect(w, r, \"/admin/recipe/\", http.StatusSeeOther)\n // return\n }\n \n if (param == \"add_ingrediant\") {\n // returned form is wanting to add an ingrediant to a drink\n// NSERT INTO recipe_ingredient (recipe_id, ingredient_id, seq, qty) SELECT r.id, i.id, 4, 1 FROM recipe r, ingredient i WHERE r.name = 'Gin and tonic (lemon lime)' AND i.name = 'Lemon'; \n\n ingredient_id, err := strconv.Atoi(r.Form.Get(\"ingrediant_selection\"))\n if err != nil {\n http.Redirect(w, r, \"/admin/recipe/\", http.StatusSeeOther)\n }\n \n ingredient_id_remove, err := strconv.Atoi(r.Form.Get(\"remove_ingr\"))\n if err != nil {\n ingredient_id_remove = -1\n }\n ingredient_qty, err := strconv.Atoi(r.Form.Get(\"ingrediant_qty\"))\n if err != nil {\n ingredient_qty = -1\n }\n \n // Default to a quantity of 1 if nothing entered or invalid entry\n if ingredient_qty <= 0 {\n ingredient_qty = 1\n }\n \n if ingredient_id_remove > 0 {\n _, err := db.Exec(\"delete from recipe_ingredient where recipe_id=? and ingredient_id=? \", recipe_id, ingredient_id_remove)\n if err != nil {\n panic(fmt.Sprintf(\"Failed to update db: %v\", err))\n }\n } else {\n\n // get next seq number\n var seq_num int\n row := db.QueryRow(\"select max(seq)+1 from recipe_ingredient where recipe_id=?\", recipe_id)\n err = row.Scan(&seq_num)\n if err != nil {\n seq_num = 1\n }\n \n \n _, err = db.Exec(\"insert into recipe_ingredient (recipe_id, ingredient_id, seq, qty) values (?, ?, ?, ?)\", recipe_id, ingredient_id, seq_num, ingredient_qty)\n if err != nil {\n panic(fmt.Sprintf(\"Failed to update db (add ingrediant): %v\", err))\n }\n }\n // http.Redirect(w, r, \"/admin/recipe/\", http.StatusSeeOther)\n // return\n }\n \n var adminR AdminRecipe \n \n \n if (recipe_id > 0) {\n adminR.RecipieSelected = true\n } else\n {\n adminR.RecipieSelected = false\n }\n \n // Get a list of all drinks for list box\n rows, err := db.Query(\"select r.id, r.name, r.glass_type_id from recipe r order by r.name\")\n if err != nil {\n panic(fmt.Sprintf(\"%v\", err))\n }\n defer rows.Close()\n glass_type_id := -1\n var tmp_glass_type_id int\n for rows.Next() {\n var recipe Recipe\n rows.Scan(&recipe.Id, &recipe.Name, &tmp_glass_type_id)\n if recipe_id == recipe.Id {\n recipe.Selected = true\n glass_type_id = tmp_glass_type_id\n } else {\n recipe.Selected = false\n }\n adminR.Recipes = append(adminR.Recipes, recipe)\n }\n rows.Close()\n \n // Get a list of glass types for the glass selection listbox\n rows, err = db.Query(\"select g.id, g.name from glass_type g order by g.name\")\n if err != nil {\n panic(fmt.Sprintf(\"%v\", err))\n }\n defer rows.Close()\n\n for rows.Next() {\n var glass GlassType\n rows.Scan(&glass.Id, &glass.Name)\n if glass.Id == glass_type_id {\n glass.Selected = true\n } else {\n glass.Selected = false\n }\n adminR.GlassTypes = append(adminR.GlassTypes, glass)\n }\n rows.Close() \n \n // Get a list of all ingrediants for the \"add\" list box\n rows, err = db.Query(\"select i.id, i.name from ingredient i order by i.name\")\n if err != nil {\n panic(fmt.Sprintf(\"%v\", err))\n }\n defer rows.Close()\n\n for rows.Next() {\n var recipeIngr AdminRecipeIngr\n rows.Scan(&recipeIngr.Id, &recipeIngr.Name)\n adminR.AllIngredients = append(adminR.AllIngredients, recipeIngr)\n }\n rows.Close()\n \n // Get a list of all ingrediants in the currently selected drink\n adminR.RecipieId = recipe_id\n \n sqlstr := \n ` select\n i.id, \n i.name,\n ri.qty * dt.unit_size,\n case when ri.qty = 1 then dt.unit_name else dt.unit_plural end as uom\n from recipe_ingredient ri\n inner join ingredient i on ri.ingredient_id = i.id\n inner join dispenser_type dt on dt.id = i.dispenser_type_id\n where ri.recipe_id = ?\n order by ri.seq`\n \n rows, err = db.Query(sqlstr, recipe_id)\n if err != nil {\n panic(fmt.Sprintf(\"%v\", err))\n }\n defer rows.Close()\n\n for rows.Next() {\n var recipeIngr AdminRecipeIngr\n rows.Scan(&recipeIngr.Id, &recipeIngr.Name, &recipeIngr.Qty, &recipeIngr.UoM)\n adminR.RecIngredients = append(adminR.RecIngredients, recipeIngr)\n }\n rows.Close()\n \n var adminHead AdminHeader\n adminHead.AllowMaint = AllowMaint\n \n tmpl.ExecuteTemplate(w, \"admin_header\", adminHead)\n tmpl.ExecuteTemplate(w, \"admin_recipe\", adminR)\n tmpl.ExecuteTemplate(w, \"admin_footer\", nil)\n return\n}", "func (r *Recipe) CreateRecipe(db *sql.DB) error {\n\terr := db.QueryRow(\n\t\t\"INSERT INTO recipes(id, name, preptime, difficulty, vegetarian) VALUES($1, $2, $3, $4, $5) RETURNING id\",\n\t\tr.ID, r.Name, r.PrepTime, r.Difficulty, r.Vegetarian).Scan(&r.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (iteration *Iteration) SetID(id string) error {\n\titeration.ID = bson.ObjectIdHex(id)\n\treturn nil\n}", "func (c *RBController) NewRecipe(w http.ResponseWriter, r *http.Request) (err error) {\n\t// build data with anonymous struct\n\tdata := struct {\n\t\t*Recipe\n\t\tNewRecipe bool\n\t}{\n\t\tnew(Recipe),\n\t\ttrue,\n\t}\n\n\t// pass data to render\n\tc.HTML(w, http.StatusOK, \"recipes/edit\", data)\n\treturn nil\n}", "func (mdl *Model) SetID(id interface{}) {\n\tmdl.id = id\n}", "func (w *Wire) SetID(id uint32) {\n\tw.id = id\n}", "func (m CrossOrderCancelReplaceRequest) SetRegistID(v string) {\n\tm.Set(field.NewRegistID(v))\n}", "func GetRecipe(id int) (recipe Recipe) {\n\tvar httpClient = &http.Client{}\n\tr, _ := httpClient.Get(fmt.Sprintf(\"http://recipes:8080/recipes/%d\", id))\n\tdefer r.Body.Close()\n\tjson.NewDecoder(r.Body).Decode(&recipe)\n\treturn\n}", "func (r *ProductRow) SetID(id int32) { r.Data.ID = &id }", "func (r *CampaignRow) SetID(id string) { r.Data.ID = id }", "func (r *CampaignRow) SetID(id string) { r.Data.ID = id }", "func (rb *ResourceBuilder) SetContainerID(val string) {\n\tif rb.config.ContainerID.Enabled {\n\t\trb.res.Attributes().PutStr(\"container.id\", val)\n\t}\n}", "func (_options *CreateSecretVersionActionOptions) SetID(id string) *CreateSecretVersionActionOptions {\n\t_options.ID = core.StringPtr(id)\n\treturn _options\n}", "func (api *FoodRecipeAPI) partialRecipeUpdate(w http.ResponseWriter, req *http.Request) {\n\tdefer DrainBody(req)\n\tctx := req.Context()\n\n\tvars := mux.Vars(req)\n\tid := vars[\"id\"]\n\tlogData := log.Data{\"id\": id}\n\n\tvar errorObjects []*models.ErrorObject\n\n\tpatchJSON, recipePatches, err := patch.Get(ctx, req.Body)\n\tif err != nil {\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// Validate patch request\n\tfor i, recipePatch := range *recipePatches {\n\t\tif err = recipePatch.Validate(nil); err != nil {\n\t\t\tif _, ok := err.(*validator.InvalidValidationError); ok {\n\t\t\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\t\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tfor _, err := range err.(validator.ValidationErrors) {\n\t\t\t\terrorObjects = append(errorObjects, models.HandleValidationErrors(strconv.Itoa(i), err.ActualTag(), err.StructField(), err.Value().(string), err.Param()))\n\t\t\t}\n\t\t}\n\t}\n\tif len(errorObjects) > 0 {\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// apply patch against recipe resource\n\tp, err := jsonpatch.DecodePatch(patchJSON)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: unable to decode patch\", err)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// find current recipe doc\n\tvar recipe models.Recipe\n\n\tcollection := api.MongoClient.Database(\"food-recipes\").Collection(\"recipes\")\n\tif err = collection.FindOne(ctx, bson.M{\"_id\": id}).Decode(&recipe); err != nil {\n\t\tif err == mongo.ErrNoDocuments {\n\t\t\tlog.Warn(ctx, \"patch recipe: failed to find recipe\", log.FormatErrors([]error{err}), logData)\n\t\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrRecipeNotFound.Error()})\n\t\t\tErrorResponse(ctx, w, http.StatusNotFound, &models.ErrorResponse{Errors: errorObjects})\n\t\t\treturn\n\t\t}\n\n\t\tlog.Error(ctx, \"patch recipe: failed to find recipe, bad connection?\", err)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\tb, err := json.Marshal(recipe)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: error returned from json marshal\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// apply patch to existing recipe\n\tmodified, err := p.Apply(b)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: unable to apply patch to recipe\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\terr = json.Unmarshal(modified, &recipe)\n\tif err != nil {\n\t\tlog.Error(ctx, \"patch recipe: unmarshal modified recipe into recipe struct\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: err.Error()})\n\t\tErrorResponse(ctx, w, http.StatusBadRequest, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\t// store new recipe\n\tif _, err = collection.ReplaceOne(ctx, bson.M{\"_id\": id}, recipe); err != nil {\n\t\tif err == mongo.ErrNoDocuments {\n\t\t\tlog.Error(ctx, \"update recipe: failed to update recipe, recipe deos not exists\", err, logData)\n\t\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrRecipeNotFound.Error()})\n\t\t\tErrorResponse(ctx, w, http.StatusNotFound, &models.ErrorResponse{Errors: errorObjects})\n\t\t\treturn\n\t\t}\n\n\t\tlog.Error(ctx, \"update recipe: failed to insert recipe\", err, logData)\n\t\terrorObjects = append(errorObjects, &models.ErrorObject{Error: errs.ErrInternalServer.Error()})\n\t\tErrorResponse(ctx, w, http.StatusInternalServerError, &models.ErrorResponse{Errors: errorObjects})\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\n\tlog.Info(ctx, \"update recipe: request successful\", logData)\n}", "func (m *Storage) GetRecipe(id int) (listing.Recipe, error) {\n\tvar recipe listing.Recipe\n\n\tfor i := range m.recipes {\n\n\t\tif m.recipes[i].ID == id {\n\t\t\trecipe.ID = m.recipes[i].ID\n\t\t\trecipe.MealType = m.recipes[i].MealType\n\t\t\trecipe.Name = m.recipes[i].Name\n\t\t\trecipe.Ingredients = m.recipes[i].Ingredients\n\t\t\trecipe.Preparation = m.recipes[i].Preparation\n\t\t\trecipe.Created = m.recipes[i].Created\n\n\t\t\treturn recipe, nil\n\t\t}\n\t}\n\n\treturn recipe, listing.ErrNotFound\n}", "func (r resourceFactory) setStateID(resourceLocalData *schema.ResourceData, payload map[string]interface{}) error {\n\tresourceSchema, err := r.openAPIResource.getResourceSchema()\n\tif err != nil {\n\t\treturn err\n\t}\n\tidentifierProperty, err := resourceSchema.getResourceIdentifier()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif payload[identifierProperty] == nil {\n\t\treturn fmt.Errorf(\"response object returned from the API is missing mandatory identifier property '%s'\", identifierProperty)\n\t}\n\n\tswitch payload[identifierProperty].(type) {\n\tcase int:\n\t\tresourceLocalData.SetId(strconv.Itoa(payload[identifierProperty].(int)))\n\tcase float64:\n\t\tresourceLocalData.SetId(strconv.Itoa(int(payload[identifierProperty].(float64))))\n\tdefault:\n\t\tresourceLocalData.SetId(payload[identifierProperty].(string))\n\t}\n\treturn nil\n}", "func (wc *WalletCreate) SetID(u uuid.UUID) *WalletCreate {\n\twc.mutation.SetID(u)\n\treturn wc\n}", "func (this *TriggerAction) SetId(id int64) {\n this.id = id\n}", "func (lc *LoanbindingCreate) SetID(i int) *LoanbindingCreate {\n\tlc.mutation.SetID(i)\n\treturn lc\n}", "func (lc *LicenseCreate) SetID(i int) *LicenseCreate {\n\tlc.mutation.SetID(i)\n\treturn lc\n}", "func UpdateRecipe(db *database.DB, recipe *Recipe) {\n\tdb.Save(recipe)\n}", "func CreateRecipe(c *gin.Context) {\n\tvar recipe model.Recipe\n\tif err := c.ShouldBind(&recipe); err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"illegal format for recipe: \" + err.Error()})\n\t\treturn\n\t}\n\n\tif recipe.ID != 0 {\n\t\tc.JSON(http.StatusConflict, gin.H{\"message\": \"cannot create recipe with specific id\"})\n\t\treturn\n\t}\n\n\tif recipe.Title == \"\" || recipe.Resources == nil || len(recipe.Resources) == 0 {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"illegal content for recipe\"})\n\t\treturn\n\t}\n\n\tvar duplicate model.Recipe\n\tmodel.DB.Where(\"title = ?\", recipe.Title).First(&duplicate)\n\tif duplicate.Title != \"\" {\n\t\tc.JSON(http.StatusConflict, gin.H{\"message\": \"recipe with title \" + duplicate.Title + \" already exists\"})\n\t\treturn\n\t}\n\n\trecipe.Create()\n\tc.JSON(http.StatusCreated, gin.H{\"data\": recipe.ID})\n}", "func (packet *ChoicePacket) SetID(id int64) {\n\tpacket.ID = id\n}", "func (o *AnalyzeRecipeInstructions200ResponseParsedInstructionsInnerStepsInnerIngredientsInner) SetId(v float32) {\n\to.Id = v\n}", "func CopyID(id ID) (ret ID) {\n\tfor i := 0; i < IDBytes; i++ {\n\t\tret[i] = id[i]\n\t}\n\treturn\n}", "func (au *AntenatalinformationUpdate) SetRisksID(id int) *AntenatalinformationUpdate {\n\tau.mutation.SetRisksID(id)\n\treturn au\n}", "func (packet *UpdateResourcePacket) SetID(id int64) {\n\tpacket.ID = id\n}", "func (b *taskBuilder) recipeProp(key, value string) {\n\tif b.recipeProperties == nil {\n\t\tlog.Fatal(\"taskBuilder.recipeProp() cannot be called after taskBuilder.getRecipeProps()!\")\n\t}\n\tb.recipeProperties[key] = value\n}", "func (ctx *Context) SpecificFavRecipeHandler(w http.ResponseWriter, r *http.Request) {\n\tstate := &SessionState{}\n\t//grabbing the session ID from the sessionsstore\n\ts, err := sessions.GetSessionID(r, ctx.SessionKey)\n\tif err != nil {\n\t\thttp.Error(w, \"Error getting authenticated user\", http.StatusInternalServerError)\n\t\treturn\n\t}\n\t//getting the state and putting into state\n\terr = ctx.SessionStore.Get(s, state)\n\tif err != nil {\n\t\thttp.Error(w, \"Error getting session state\", http.StatusInternalServerError)\n\t\treturn\n\t}\n\t//get the url from the path\n\t_, recipeName := path.Split(r.URL.String())\n\tuser := state.User\n\tif err != nil {\n\t\thttp.Error(w, \"Error finding user\", http.StatusInternalServerError)\n\t\treturn\n\t}\n\tswitch r.Method {\n\tcase \"POST\":\n\t\t//Adds the recipe\n\t\terr := ctx.UserStore.AddToBook(user, recipeName)\n\t\tif err != nil {\n\t\t\thttp.Error(w, \"Error adding favorite recipe\", http.StatusInternalServerError)\n\t\t}\n\t\t//Recipe events adding recipe to book\n\t\tntfy := &notification.RecipesEvent{\n\t\t\tEventType: notification.NewBook,\n\t\t\tMessage: recipeName,\n\t\t}\n\t\tctx.Notifier.Notify(ntfy)\n\t\tw.Header().Add(\"Content-Type\", contentTypeTextUTF8)\n\t\tw.Write([]byte(\"Recipe has been added from the book\"))\n\tcase \"DELETE\":\n\t\terr := ctx.UserStore.DeleteFromBook(user, recipeName)\n\t\tif err != nil {\n\t\t\thttp.Error(w, \"Unable to remove that recipe\", http.StatusBadRequest)\n\t\t}\n\t\t//Recipe events adding recipe to book\n\t\tntfy := &notification.RecipesEvent{\n\t\t\tEventType: notification.RemoveBook,\n\t\t\tMessage: recipeName,\n\t\t}\n\t\tctx.Notifier.Notify(ntfy)\n\t\tw.Header().Add(\"Content-Type\", contentTypeTextUTF8)\n\t\tw.Write([]byte(\"Recipe has been removed from the book\"))\n\t}\n}", "func (o *RecipeAdditive) SetAdditive(ctx context.Context, exec boil.ContextExecutor, insert bool, related *Additive) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(ctx, exec, boil.Infer()); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE \\\"recipe_additive\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, []string{\"additive_id\"}),\n\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 2, recipeAdditivePrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.ID, o.ID}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, updateQuery)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tif _, err = exec.ExecContext(ctx, updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.AdditiveID = related.ID\n\tif o.R == nil {\n\t\to.R = &recipeAdditiveR{\n\t\t\tAdditive: related,\n\t\t}\n\t} else {\n\t\to.R.Additive = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &additiveR{\n\t\t\tRecipeAdditive: o,\n\t\t}\n\t} else {\n\t\trelated.R.RecipeAdditive = o\n\t}\n\n\treturn nil\n}", "func (r *PackageRow) SetID(id string) { r.Data.ID = id }", "func (this *SumBounds) SetID(propID core.PropId) {\n\tthis.id = propID\n}", "func (bc *BulkCreate) SetID(i int) *BulkCreate {\n\tbc.mutation.SetID(i)\n\treturn bc\n}", "func (m *ChatMessageAttachment) SetId(value *string)() {\n err := m.GetBackingStore().Set(\"id\", value)\n if err != nil {\n panic(err)\n }\n}", "func (testEntityRelated_EntityInfo) SetId(object interface{}, id uint64) {\n\tobject.(*TestEntityRelated).Id = id\n}", "func (_options *CreateSecretActionOptions) SetID(id string) *CreateSecretActionOptions {\n\t_options.ID = core.StringPtr(id)\n\treturn _options\n}", "func (packet *PlayerCraftingPacket) SetID(id int64) {\n\tpacket.ID = id\n}", "func (mySource *Source) SetID(val string) {\n\tmySource.IDvar = val\n}", "func (acc *AccessControlCreate) SetID(i int64) *AccessControlCreate {\n\tacc.mutation.SetID(i)\n\treturn acc\n}", "func (m *Manifest) SetID(mid ManifestID) {\n\tm.Source = mid.Source\n\tm.Flavor = mid.Flavor\n}", "func (nc *NodeCreate) SetID(i int64) *NodeCreate {\n\tnc.mutation.SetID(i)\n\treturn nc\n}", "func (o *RecipeLipid) Recipe(mods ...qm.QueryMod) recipeQuery {\n\tqueryMods := []qm.QueryMod{\n\t\tqm.Where(\"\\\"id\\\" = ?\", o.RecipeID),\n\t}\n\n\tqueryMods = append(queryMods, mods...)\n\n\tquery := Recipes(queryMods...)\n\tqueries.SetFrom(query.Query, \"\\\"recipe\\\"\")\n\n\treturn query\n}", "func (ds CBDataStore) RecipeCreate(modelRecipe model.Recipe, userID string) (*model.Recipe, error) {\n\tkey := fmt.Sprintf(\"recipe::%s::%s\", userID, modelRecipe.Name)\n\tnewR := new(recipe)\n\tnewR.Recipe = &modelRecipe\n\tnewR.Type = \"recipe\"\n\n\t_, err := ds.bucket.Insert(key, newR, 0)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn newR.Recipe, nil\n}", "func (ds CBDataStore) RecipeCreate(modelRecipe model.Recipe, userID string) (*model.Recipe, error) {\n\tkey := fmt.Sprintf(\"recipe::%s::%s\", userID, modelRecipe.Name)\n\tnewR := new(recipe)\n\tnewR.Recipe = &modelRecipe\n\tnewR.Type = \"recipe\"\n\n\t_, err := ds.bucket.Insert(key, newR, 0)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn newR.Recipe, nil\n}", "func (o *GetContentSourceUsingGETParams) SetID(id strfmt.UUID) {\n\to.ID = id\n}", "func (o *CopyRecipeToMyRecipesWithChangesParams) WithContext(ctx context.Context) *CopyRecipeToMyRecipesWithChangesParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (m *TermStoreRequestBuilder) SetsById(id string)(*i619b375874405f8a48a2830bcc3f479ceef5c227f0ff2205d5561454de690fef.SetItemRequestBuilder) {\n urlTplParams := make(map[string]string)\n for idx, item := range m.pathParameters {\n urlTplParams[idx] = item\n }\n if id != \"\" {\n urlTplParams[\"set%2Did\"] = id\n }\n return i619b375874405f8a48a2830bcc3f479ceef5c227f0ff2205d5561454de690fef.NewSetItemRequestBuilderInternal(urlTplParams, m.requestAdapter);\n}", "func (auo *AntenatalinformationUpdateOne) SetRisksID(id int) *AntenatalinformationUpdateOne {\n\tauo.mutation.SetRisksID(id)\n\treturn auo\n}", "func (m *DiscoveredSensitiveType) SetId(value *i561e97a8befe7661a44c8f54600992b4207a3a0cf6770e5559949bc276de2e22.UUID)() {\n err := m.GetBackingStore().Set(\"id\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o *GetaspecificPbxDeviceFirmwareBinaryParams) SetID(id string) {\n\to.ID = id\n}", "func (dc *DatasourceCreate) SetID(i int64) *DatasourceCreate {\n\tdc.mutation.SetID(i)\n\treturn dc\n}", "func (o *InventoryStocktakingSearchParams) SetID(id *string) {\n\to.ID = id\n}", "func (piuo *ProviderIDUpdateOne) SetParticpantID(id string) *ProviderIDUpdateOne {\n\tpiuo.mutation.SetParticpantID(id)\n\treturn piuo\n}", "func (r *Response) SetID(id string) { r.id = id }", "func (piu *ProviderIDUpdate) SetParticpantID(id string) *ProviderIDUpdate {\n\tpiu.mutation.SetParticpantID(id)\n\treturn piu\n}", "func (db *DB) DeleteRecipe(ctx context.Context, recipeID string) error {\n\tr, err := db.getRecipeByID(ctx, recipeID)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif r == nil {\n\t\treturn nil\n\t}\n\tr.Tags = nil\n\n\td := make([]interface{}, 0, len(r.Ingredients)+len(r.Steps)+1)\n\td = append(d, r)\n\tfor _, i := range r.Ingredients {\n\t\ti.Food = nil\n\t\td = append(d, *i)\n\t}\n\tfor _, s := range r.Steps {\n\t\td = append(d, *s)\n\t}\n\n\tpb, err := json.Marshal(d)\n\tif err != nil {\n\t\treturn err\n\t}\n\tmu := &api.Mutation{\n\t\tDeleteJson: pb,\n\t}\n\treq := &api.Request{CommitNow: true}\n\treq.Mutations = []*api.Mutation{mu}\n\n\t_, err = db.Dgraph.NewTxn().Do(ctx, req)\n\n\treturn err\n}", "func (p *PointsSgMutator) SetId(val uint64) bool { //nolint:dupl false positive\n\tif val != p.Id {\n\t\tp.mutations = append(p.mutations, A.X{`=`, 0, val})\n\t\tp.Id = val\n\t\treturn true\n\t}\n\treturn false\n}", "func (db RecipeDB) addRecipe(name, version string, success bool) error {\n\tversionNum, err := convertVersion(version)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdb[name] = Recipe{Name: name, Version: versionNum, InstallTime: time.Now().Unix(), Success: success}\n\n\tvar recipelist []Recipe\n\tfor _, recipe := range db {\n\t\trecipelist = append(recipelist, recipe)\n\t}\n\tdbBytes, err := json.Marshal(recipelist)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdbDir := getDbDir()\n\tif err := os.MkdirAll(dbDir, 0755); err != nil {\n\t\treturn err\n\t}\n\n\tf, err := ioutil.TempFile(dbDir, dbFileName+\"_*\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := f.Write(dbBytes); err != nil {\n\t\tf.Close()\n\t\treturn err\n\t}\n\tif err := f.Close(); err != nil {\n\t\treturn err\n\t}\n\n\treturn os.Rename(f.Name(), filepath.Join(dbDir, dbFileName))\n}" ]
[ "0.67791796", "0.6419247", "0.62198806", "0.58240676", "0.56545", "0.5435384", "0.5370785", "0.5254237", "0.5230423", "0.5207603", "0.5194222", "0.51442355", "0.51157427", "0.5088593", "0.5044387", "0.5031753", "0.49801058", "0.49731523", "0.49663797", "0.49481156", "0.49331784", "0.49308527", "0.491125", "0.487885", "0.4864659", "0.4848365", "0.48433378", "0.48412558", "0.48392588", "0.48297298", "0.48297298", "0.47817263", "0.47816685", "0.47714114", "0.47407225", "0.4712468", "0.47036576", "0.47010627", "0.46955335", "0.46867767", "0.46786436", "0.46774265", "0.46666515", "0.46635014", "0.46531823", "0.46499777", "0.4623568", "0.46186838", "0.4615443", "0.4602691", "0.45989504", "0.45931166", "0.45931166", "0.45811978", "0.45742273", "0.4573504", "0.45591924", "0.45488322", "0.45457405", "0.45453125", "0.45441926", "0.45378065", "0.45377752", "0.4530994", "0.4526386", "0.4523697", "0.45227695", "0.45201817", "0.45034388", "0.4500001", "0.44942972", "0.4491293", "0.44884658", "0.44879636", "0.4482806", "0.44814473", "0.44782272", "0.44774994", "0.44745252", "0.44736406", "0.446729", "0.4464323", "0.446013", "0.4457974", "0.44524783", "0.44524783", "0.44518435", "0.44483855", "0.44482112", "0.44427657", "0.4440092", "0.4438243", "0.44347268", "0.44276258", "0.44210243", "0.44202903", "0.44169107", "0.44131094", "0.44114703", "0.44078633" ]
0.6869463
0
WriteToRequest writes these params to a swagger request
func (o *CopyRecipeToMyRecipesWithChangesParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error { if err := r.SetTimeout(o.timeout); err != nil { return err } var res []error if o.JSONBody != nil { if err := r.SetBodyParam(o.JSONBody); err != nil { return err } } // path param recipeId if err := r.SetPathParam("recipeId", o.RecipeID); err != nil { return err } if len(res) > 0 { return errors.CompositeValidationError(res...) } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *FileInfoCreateParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.ByteOffset != nil {\n\n\t\t// query param byte_offset\n\t\tvar qrByteOffset int64\n\n\t\tif o.ByteOffset != nil {\n\t\t\tqrByteOffset = *o.ByteOffset\n\t\t}\n\t\tqByteOffset := swag.FormatInt64(qrByteOffset)\n\t\tif qByteOffset != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"byte_offset\", qByteOffset); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\tif o.Info != nil {\n\t\tif err := r.SetBodyParam(o.Info); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.Overwrite != nil {\n\n\t\t// query param overwrite\n\t\tvar qrOverwrite bool\n\n\t\tif o.Overwrite != nil {\n\t\t\tqrOverwrite = *o.Overwrite\n\t\t}\n\t\tqOverwrite := swag.FormatBool(qrOverwrite)\n\t\tif qOverwrite != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"overwrite\", qOverwrite); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\t// path param path\n\tif err := r.SetPathParam(\"path\", o.Path); err != nil {\n\t\treturn err\n\t}\n\n\tif o.ReturnRecords != nil {\n\n\t\t// query param return_records\n\t\tvar qrReturnRecords bool\n\n\t\tif o.ReturnRecords != nil {\n\t\t\tqrReturnRecords = *o.ReturnRecords\n\t\t}\n\t\tqReturnRecords := swag.FormatBool(qrReturnRecords)\n\t\tif qReturnRecords != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"return_records\", qReturnRecords); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.StreamName != nil {\n\n\t\t// query param stream_name\n\t\tvar qrStreamName string\n\n\t\tif o.StreamName != nil {\n\t\t\tqrStreamName = *o.StreamName\n\t\t}\n\t\tqStreamName := qrStreamName\n\t\tif qStreamName != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"stream_name\", qStreamName); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\t// path param volume.uuid\n\tif err := r.SetPathParam(\"volume.uuid\", o.VolumeUUID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// header param Device-Id\n\tif err := r.SetHeaderParam(\"Device-Id\", o.DeviceID); err != nil {\n\t\treturn err\n\t}\n\n\tif o.DeviceOS != nil {\n\n\t\t// header param Device-OS\n\t\tif err := r.SetHeaderParam(\"Device-OS\", *o.DeviceOS); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\n\t// path param fiscalDocumentNumber\n\tif err := r.SetPathParam(\"fiscalDocumentNumber\", swag.FormatUint64(o.FiscalDocumentNumber)); err != nil {\n\t\treturn err\n\t}\n\n\t// path param fiscalDriveNumber\n\tif err := r.SetPathParam(\"fiscalDriveNumber\", swag.FormatUint64(o.FiscalDriveNumber)); err != nil {\n\t\treturn err\n\t}\n\n\t// query param fiscalSign\n\tqrFiscalSign := o.FiscalSign\n\tqFiscalSign := swag.FormatUint64(qrFiscalSign)\n\tif qFiscalSign != \"\" {\n\t\tif err := r.SetQueryParam(\"fiscalSign\", qFiscalSign); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.SendToEmail != nil {\n\n\t\t// query param sendToEmail\n\t\tvar qrSendToEmail string\n\t\tif o.SendToEmail != nil {\n\t\t\tqrSendToEmail = *o.SendToEmail\n\t\t}\n\t\tqSendToEmail := qrSendToEmail\n\t\tif qSendToEmail != \"\" {\n\t\t\tif err := r.SetQueryParam(\"sendToEmail\", qSendToEmail); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *StartV1Params) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Environment != nil {\n\n\t\t// query param environment\n\t\tvar qrEnvironment string\n\t\tif o.Environment != nil {\n\t\t\tqrEnvironment = *o.Environment\n\t\t}\n\t\tqEnvironment := qrEnvironment\n\t\tif qEnvironment != \"\" {\n\t\t\tif err := r.SetQueryParam(\"environment\", qEnvironment); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *UpdateAutoTagParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Body != nil {\n\t\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param id\n\tif err := r.SetPathParam(\"id\", o.ID.String()); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetIntrospectionParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// header param Authorization\n\tif err := r.SetHeaderParam(\"Authorization\", o.Authorization); err != nil {\n\t\treturn err\n\t}\n\n\tif o.ResponseAsJwt != nil {\n\n\t\t// query param response_as_jwt\n\t\tvar qrResponseAsJwt bool\n\t\tif o.ResponseAsJwt != nil {\n\t\t\tqrResponseAsJwt = *o.ResponseAsJwt\n\t\t}\n\t\tqResponseAsJwt := swag.FormatBool(qrResponseAsJwt)\n\t\tif qResponseAsJwt != \"\" {\n\t\t\tif err := r.SetQueryParam(\"response_as_jwt\", qResponseAsJwt); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// query param token\n\tqrToken := o.Token\n\tqToken := qrToken\n\tif qToken != \"\" {\n\t\tif err := r.SetQueryParam(\"token\", qToken); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.TokenTypeHint != nil {\n\n\t\t// query param token_type_hint\n\t\tvar qrTokenTypeHint string\n\t\tif o.TokenTypeHint != nil {\n\t\t\tqrTokenTypeHint = *o.TokenTypeHint\n\t\t}\n\t\tqTokenTypeHint := qrTokenTypeHint\n\t\tif qTokenTypeHint != \"\" {\n\t\t\tif err := r.SetQueryParam(\"token_type_hint\", qTokenTypeHint); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *PostContextsAddPhpParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// query param name\n\tqrName := o.Name\n\tqName := qrName\n\tif qName != \"\" {\n\n\t\tif err := r.SetQueryParam(\"name\", qName); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.Private != nil {\n\n\t\t// query param private\n\t\tvar qrPrivate int64\n\n\t\tif o.Private != nil {\n\t\t\tqrPrivate = *o.Private\n\t\t}\n\t\tqPrivate := swag.FormatInt64(qrPrivate)\n\t\tif qPrivate != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"private\", qPrivate); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetInstancesDocsParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// header param Authorization\n\tif err := r.SetHeaderParam(\"Authorization\", o.Authorization); err != nil {\n\t\treturn err\n\t}\n\n\tif o.OperationID != nil {\n\n\t\t// query param operationId\n\t\tvar qrOperationID string\n\t\tif o.OperationID != nil {\n\t\t\tqrOperationID = *o.OperationID\n\t\t}\n\t\tqOperationID := qrOperationID\n\t\tif qOperationID != \"\" {\n\t\t\tif err := r.SetQueryParam(\"operationId\", qOperationID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Version != nil {\n\n\t\t// query param version\n\t\tvar qrVersion string\n\t\tif o.Version != nil {\n\t\t\tqrVersion = *o.Version\n\t\t}\n\t\tqVersion := qrVersion\n\t\tif qVersion != \"\" {\n\t\t\tif err := r.SetQueryParam(\"version\", qVersion); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *CloudTargetCreateParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.CheckOnly != nil {\n\n\t\t// query param check_only\n\t\tvar qrCheckOnly bool\n\n\t\tif o.CheckOnly != nil {\n\t\t\tqrCheckOnly = *o.CheckOnly\n\t\t}\n\t\tqCheckOnly := swag.FormatBool(qrCheckOnly)\n\t\tif qCheckOnly != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"check_only\", qCheckOnly); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.IgnoreWarnings != nil {\n\n\t\t// query param ignore_warnings\n\t\tvar qrIgnoreWarnings bool\n\n\t\tif o.IgnoreWarnings != nil {\n\t\t\tqrIgnoreWarnings = *o.IgnoreWarnings\n\t\t}\n\t\tqIgnoreWarnings := swag.FormatBool(qrIgnoreWarnings)\n\t\tif qIgnoreWarnings != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"ignore_warnings\", qIgnoreWarnings); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\tif o.Info != nil {\n\t\tif err := r.SetBodyParam(o.Info); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.ReturnRecords != nil {\n\n\t\t// query param return_records\n\t\tvar qrReturnRecords bool\n\n\t\tif o.ReturnRecords != nil {\n\t\t\tqrReturnRecords = *o.ReturnRecords\n\t\t}\n\t\tqReturnRecords := swag.FormatBool(qrReturnRecords)\n\t\tif qReturnRecords != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"return_records\", qReturnRecords); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.ReturnTimeout != nil {\n\n\t\t// query param return_timeout\n\t\tvar qrReturnTimeout int64\n\n\t\tif o.ReturnTimeout != nil {\n\t\t\tqrReturnTimeout = *o.ReturnTimeout\n\t\t}\n\t\tqReturnTimeout := swag.FormatInt64(qrReturnTimeout)\n\t\tif qReturnTimeout != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"return_timeout\", qReturnTimeout); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *SayParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\treturn err\n\t}\n\n\tif o.Sinkid != nil {\n\n\t\t// query param sinkid\n\t\tvar qrSinkid string\n\t\tif o.Sinkid != nil {\n\t\t\tqrSinkid = *o.Sinkid\n\t\t}\n\t\tqSinkid := qrSinkid\n\t\tif qSinkid != \"\" {\n\t\t\tif err := r.SetQueryParam(\"sinkid\", qSinkid); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Voiceid != nil {\n\n\t\t// query param voiceid\n\t\tvar qrVoiceid string\n\t\tif o.Voiceid != nil {\n\t\t\tqrVoiceid = *o.Voiceid\n\t\t}\n\t\tqVoiceid := qrVoiceid\n\t\tif qVoiceid != \"\" {\n\t\t\tif err := r.SetQueryParam(\"voiceid\", qVoiceid); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *HandleGetAboutUsingGETParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// header param Authorization\n\tif err := r.SetHeaderParam(\"Authorization\", o.Authorization); err != nil {\n\t\treturn err\n\t}\n\n\t// query param apiVersion\n\tqrAPIVersion := o.APIVersion\n\tqAPIVersion := qrAPIVersion\n\n\tif err := r.SetQueryParam(\"apiVersion\", qAPIVersion); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetFileSystemParametersInternalParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.AccountID != nil {\n\n\t\t// query param accountId\n\t\tvar qrAccountID string\n\t\tif o.AccountID != nil {\n\t\t\tqrAccountID = *o.AccountID\n\t\t}\n\t\tqAccountID := qrAccountID\n\t\tif qAccountID != \"\" {\n\t\t\tif err := r.SetQueryParam(\"accountId\", qAccountID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.AccountName != nil {\n\n\t\t// query param accountName\n\t\tvar qrAccountName string\n\t\tif o.AccountName != nil {\n\t\t\tqrAccountName = *o.AccountName\n\t\t}\n\t\tqAccountName := qrAccountName\n\t\tif qAccountName != \"\" {\n\t\t\tif err := r.SetQueryParam(\"accountName\", qAccountName); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.AttachedCluster != nil {\n\n\t\t// query param attachedCluster\n\t\tvar qrAttachedCluster bool\n\t\tif o.AttachedCluster != nil {\n\t\t\tqrAttachedCluster = *o.AttachedCluster\n\t\t}\n\t\tqAttachedCluster := swag.FormatBool(qrAttachedCluster)\n\t\tif qAttachedCluster != \"\" {\n\t\t\tif err := r.SetQueryParam(\"attachedCluster\", qAttachedCluster); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// query param blueprintName\n\tqrBlueprintName := o.BlueprintName\n\tqBlueprintName := qrBlueprintName\n\tif qBlueprintName != \"\" {\n\t\tif err := r.SetQueryParam(\"blueprintName\", qBlueprintName); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// query param clusterName\n\tqrClusterName := o.ClusterName\n\tqClusterName := qrClusterName\n\tif qClusterName != \"\" {\n\t\tif err := r.SetQueryParam(\"clusterName\", qClusterName); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// query param fileSystemType\n\tqrFileSystemType := o.FileSystemType\n\tqFileSystemType := qrFileSystemType\n\tif qFileSystemType != \"\" {\n\t\tif err := r.SetQueryParam(\"fileSystemType\", qFileSystemType); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.Secure != nil {\n\n\t\t// query param secure\n\t\tvar qrSecure bool\n\t\tif o.Secure != nil {\n\t\t\tqrSecure = *o.Secure\n\t\t}\n\t\tqSecure := swag.FormatBool(qrSecure)\n\t\tif qSecure != \"\" {\n\t\t\tif err := r.SetQueryParam(\"secure\", qSecure); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// query param storageName\n\tqrStorageName := o.StorageName\n\tqStorageName := qrStorageName\n\tif qStorageName != \"\" {\n\t\tif err := r.SetQueryParam(\"storageName\", qStorageName); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param workspaceId\n\tif err := r.SetPathParam(\"workspaceId\", swag.FormatInt64(o.WorkspaceID)); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ServeBuildFieldShortParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param btLocator\n\tif err := r.SetPathParam(\"btLocator\", o.BtLocator); err != nil {\n\t\treturn err\n\t}\n\n\t// path param buildLocator\n\tif err := r.SetPathParam(\"buildLocator\", o.BuildLocator); err != nil {\n\t\treturn err\n\t}\n\n\t// path param field\n\tif err := r.SetPathParam(\"field\", o.Field); err != nil {\n\t\treturn err\n\t}\n\n\t// path param projectLocator\n\tif err := r.SetPathParam(\"projectLocator\", o.ProjectLocator); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetRequestDetailsParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param extra\n\tif err := r.SetPathParam(\"extra\", o.Extra); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetWorkItemParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.DollarExpand != nil {\n\n\t\t// query param $expand\n\t\tvar qrNrDollarExpand string\n\t\tif o.DollarExpand != nil {\n\t\t\tqrNrDollarExpand = *o.DollarExpand\n\t\t}\n\t\tqNrDollarExpand := qrNrDollarExpand\n\t\tif qNrDollarExpand != \"\" {\n\t\t\tif err := r.SetQueryParam(\"$expand\", qNrDollarExpand); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// query param api-version\n\tqrAPIVersion := o.APIVersion\n\tqAPIVersion := qrAPIVersion\n\tif qAPIVersion != \"\" {\n\t\tif err := r.SetQueryParam(\"api-version\", qAPIVersion); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.AsOf != nil {\n\n\t\t// query param asOf\n\t\tvar qrAsOf strfmt.DateTime\n\t\tif o.AsOf != nil {\n\t\t\tqrAsOf = *o.AsOf\n\t\t}\n\t\tqAsOf := qrAsOf.String()\n\t\tif qAsOf != \"\" {\n\t\t\tif err := r.SetQueryParam(\"asOf\", qAsOf); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Fields != nil {\n\n\t\t// query param fields\n\t\tvar qrFields string\n\t\tif o.Fields != nil {\n\t\t\tqrFields = *o.Fields\n\t\t}\n\t\tqFields := qrFields\n\t\tif qFields != \"\" {\n\t\t\tif err := r.SetQueryParam(\"fields\", qFields); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// path param id\n\tif err := r.SetPathParam(\"id\", swag.FormatInt32(o.ID)); err != nil {\n\t\treturn err\n\t}\n\n\t// path param organization\n\tif err := r.SetPathParam(\"organization\", o.Organization); err != nil {\n\t\treturn err\n\t}\n\n\t// path param project\n\tif err := r.SetPathParam(\"project\", o.Project); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *IntegrationsManualHTTPSCreateParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif err := r.SetBodyParam(o.Data); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ValidateParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\tif o.Body != nil {\n\t\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *PostSecdefSearchParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif err := r.SetBodyParam(o.Symbol); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *UserShowV1Params) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param id\n\tif err := r.SetPathParam(\"id\", swag.FormatInt64(o.ID)); err != nil {\n\t\treturn err\n\t}\n\n\t// path param name\n\tif err := r.SetPathParam(\"name\", o.Name); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetLogsParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Fields != nil {\n\n\t\t// query param fields\n\t\tvar qrFields string\n\t\tif o.Fields != nil {\n\t\t\tqrFields = *o.Fields\n\t\t}\n\t\tqFields := qrFields\n\t\tif qFields != \"\" {\n\t\t\tif err := r.SetQueryParam(\"fields\", qFields); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.From != nil {\n\n\t\t// query param from\n\t\tvar qrFrom string\n\t\tif o.From != nil {\n\t\t\tqrFrom = *o.From\n\t\t}\n\t\tqFrom := qrFrom\n\t\tif qFrom != \"\" {\n\t\t\tif err := r.SetQueryParam(\"from\", qFrom); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.IncludeFields != nil {\n\n\t\t// query param include_fields\n\t\tvar qrIncludeFields bool\n\t\tif o.IncludeFields != nil {\n\t\t\tqrIncludeFields = *o.IncludeFields\n\t\t}\n\t\tqIncludeFields := swag.FormatBool(qrIncludeFields)\n\t\tif qIncludeFields != \"\" {\n\t\t\tif err := r.SetQueryParam(\"include_fields\", qIncludeFields); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.IncludeTotals != nil {\n\n\t\t// query param include_totals\n\t\tvar qrIncludeTotals bool\n\t\tif o.IncludeTotals != nil {\n\t\t\tqrIncludeTotals = *o.IncludeTotals\n\t\t}\n\t\tqIncludeTotals := swag.FormatBool(qrIncludeTotals)\n\t\tif qIncludeTotals != \"\" {\n\t\t\tif err := r.SetQueryParam(\"include_totals\", qIncludeTotals); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Page != nil {\n\n\t\t// query param page\n\t\tvar qrPage int64\n\t\tif o.Page != nil {\n\t\t\tqrPage = *o.Page\n\t\t}\n\t\tqPage := swag.FormatInt64(qrPage)\n\t\tif qPage != \"\" {\n\t\t\tif err := r.SetQueryParam(\"page\", qPage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.PerPage != nil {\n\n\t\t// query param per_page\n\t\tvar qrPerPage int64\n\t\tif o.PerPage != nil {\n\t\t\tqrPerPage = *o.PerPage\n\t\t}\n\t\tqPerPage := swag.FormatInt64(qrPerPage)\n\t\tif qPerPage != \"\" {\n\t\t\tif err := r.SetQueryParam(\"per_page\", qPerPage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Q != nil {\n\n\t\t// query param q\n\t\tvar qrQ string\n\t\tif o.Q != nil {\n\t\t\tqrQ = *o.Q\n\t\t}\n\t\tqQ := qrQ\n\t\tif qQ != \"\" {\n\t\t\tif err := r.SetQueryParam(\"q\", qQ); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Sort != nil {\n\n\t\t// query param sort\n\t\tvar qrSort string\n\t\tif o.Sort != nil {\n\t\t\tqrSort = *o.Sort\n\t\t}\n\t\tqSort := qrSort\n\t\tif qSort != \"\" {\n\t\t\tif err := r.SetQueryParam(\"sort\", qSort); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Take != nil {\n\n\t\t// query param take\n\t\tvar qrTake int64\n\t\tif o.Take != nil {\n\t\t\tqrTake = *o.Take\n\t\t}\n\t\tqTake := swag.FormatInt64(qrTake)\n\t\tif qTake != \"\" {\n\t\t\tif err := r.SetQueryParam(\"take\", qTake); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *PostGetOneParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param id\n\tif err := r.SetPathParam(\"id\", swag.FormatInt64(o.ID)); err != nil {\n\t\treturn err\n\t}\n\n\tvaluesRelated := o.Related\n\n\tjoinedRelated := swag.JoinByFormat(valuesRelated, \"\")\n\t// query array param related\n\tif err := r.SetQueryParam(\"related\", joinedRelated...); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *BarParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ConfigGetParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Option != nil {\n\n\t\t// binding items for option\n\t\tjoinedOption := o.bindParamOption(reg)\n\n\t\t// query array param option\n\t\tif err := r.SetQueryParam(\"option\", joinedOption...); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.ProjectID != nil {\n\n\t\t// query param project_id\n\t\tvar qrProjectID int64\n\n\t\tif o.ProjectID != nil {\n\t\t\tqrProjectID = *o.ProjectID\n\t\t}\n\t\tqProjectID := swag.FormatInt64(qrProjectID)\n\t\tif qProjectID != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"project_id\", qProjectID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.UserID != nil {\n\n\t\t// query param user_id\n\t\tvar qrUserID int64\n\n\t\tif o.UserID != nil {\n\t\t\tqrUserID = *o.UserID\n\t\t}\n\t\tqUserID := swag.FormatInt64(qrUserID)\n\t\tif qUserID != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"user_id\", qUserID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetSsoParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// query param code\n\tqrCode := o.Code\n\tqCode := qrCode\n\tif qCode != \"\" {\n\t\tif err := r.SetQueryParam(\"code\", qCode); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// query param resource_id\n\tqrResourceID := o.ResourceID\n\tqResourceID := qrResourceID\n\tif qResourceID != \"\" {\n\t\tif err := r.SetQueryParam(\"resource_id\", qResourceID); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *AllLookmlTestsParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.FileID != nil {\n\n\t\t// query param file_id\n\t\tvar qrFileID string\n\t\tif o.FileID != nil {\n\t\t\tqrFileID = *o.FileID\n\t\t}\n\t\tqFileID := qrFileID\n\t\tif qFileID != \"\" {\n\t\t\tif err := r.SetQueryParam(\"file_id\", qFileID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// path param project_id\n\tif err := r.SetPathParam(\"project_id\", o.ProjectID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *APIServiceHaltsParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Height != nil {\n\n\t\t// query param height\n\t\tvar qrHeight string\n\t\tif o.Height != nil {\n\t\t\tqrHeight = *o.Height\n\t\t}\n\t\tqHeight := qrHeight\n\t\tif qHeight != \"\" {\n\t\t\tif err := r.SetQueryParam(\"height\", qHeight); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetUsersParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Connection != nil {\n\n\t\t// query param connection\n\t\tvar qrConnection string\n\t\tif o.Connection != nil {\n\t\t\tqrConnection = *o.Connection\n\t\t}\n\t\tqConnection := qrConnection\n\t\tif qConnection != \"\" {\n\t\t\tif err := r.SetQueryParam(\"connection\", qConnection); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Fields != nil {\n\n\t\t// query param fields\n\t\tvar qrFields string\n\t\tif o.Fields != nil {\n\t\t\tqrFields = *o.Fields\n\t\t}\n\t\tqFields := qrFields\n\t\tif qFields != \"\" {\n\t\t\tif err := r.SetQueryParam(\"fields\", qFields); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.IncludeFields != nil {\n\n\t\t// query param include_fields\n\t\tvar qrIncludeFields bool\n\t\tif o.IncludeFields != nil {\n\t\t\tqrIncludeFields = *o.IncludeFields\n\t\t}\n\t\tqIncludeFields := swag.FormatBool(qrIncludeFields)\n\t\tif qIncludeFields != \"\" {\n\t\t\tif err := r.SetQueryParam(\"include_fields\", qIncludeFields); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.IncludeTotals != nil {\n\n\t\t// query param include_totals\n\t\tvar qrIncludeTotals bool\n\t\tif o.IncludeTotals != nil {\n\t\t\tqrIncludeTotals = *o.IncludeTotals\n\t\t}\n\t\tqIncludeTotals := swag.FormatBool(qrIncludeTotals)\n\t\tif qIncludeTotals != \"\" {\n\t\t\tif err := r.SetQueryParam(\"include_totals\", qIncludeTotals); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Page != nil {\n\n\t\t// query param page\n\t\tvar qrPage int64\n\t\tif o.Page != nil {\n\t\t\tqrPage = *o.Page\n\t\t}\n\t\tqPage := swag.FormatInt64(qrPage)\n\t\tif qPage != \"\" {\n\t\t\tif err := r.SetQueryParam(\"page\", qPage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.PerPage != nil {\n\n\t\t// query param per_page\n\t\tvar qrPerPage int64\n\t\tif o.PerPage != nil {\n\t\t\tqrPerPage = *o.PerPage\n\t\t}\n\t\tqPerPage := swag.FormatInt64(qrPerPage)\n\t\tif qPerPage != \"\" {\n\t\t\tif err := r.SetQueryParam(\"per_page\", qPerPage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Q != nil {\n\n\t\t// query param q\n\t\tvar qrQ string\n\t\tif o.Q != nil {\n\t\t\tqrQ = *o.Q\n\t\t}\n\t\tqQ := qrQ\n\t\tif qQ != \"\" {\n\t\t\tif err := r.SetQueryParam(\"q\", qQ); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.SearchEngine != nil {\n\n\t\t// query param search_engine\n\t\tvar qrSearchEngine string\n\t\tif o.SearchEngine != nil {\n\t\t\tqrSearchEngine = *o.SearchEngine\n\t\t}\n\t\tqSearchEngine := qrSearchEngine\n\t\tif qSearchEngine != \"\" {\n\t\t\tif err := r.SetQueryParam(\"search_engine\", qSearchEngine); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Sort != nil {\n\n\t\t// query param sort\n\t\tvar qrSort string\n\t\tif o.Sort != nil {\n\t\t\tqrSort = *o.Sort\n\t\t}\n\t\tqSort := qrSort\n\t\tif qSort != \"\" {\n\t\t\tif err := r.SetQueryParam(\"sort\", qSort); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetBlockGeneratorResultParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param id\n\tif err := r.SetPathParam(\"id\", o.ID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *CreateRuntimeMapParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.FileUpload != nil {\n\n\t\tif o.FileUpload != nil {\n\n\t\t\t// form file param file_upload\n\t\t\tif err := r.SetFileParam(\"file_upload\", o.FileUpload); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetPlatformsParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Extended != nil {\n\n\t\t// query param extended\n\t\tvar qrExtended bool\n\t\tif o.Extended != nil {\n\t\t\tqrExtended = *o.Extended\n\t\t}\n\t\tqExtended := swag.FormatBool(qrExtended)\n\t\tif qExtended != \"\" {\n\t\t\tif err := r.SetQueryParam(\"extended\", qExtended); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetUserUsageParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Cloud != nil {\n\n\t\t// query param cloud\n\t\tvar qrCloud string\n\t\tif o.Cloud != nil {\n\t\t\tqrCloud = *o.Cloud\n\t\t}\n\t\tqCloud := qrCloud\n\t\tif qCloud != \"\" {\n\t\t\tif err := r.SetQueryParam(\"cloud\", qCloud); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Filterenddate != nil {\n\n\t\t// query param filterenddate\n\t\tvar qrFilterenddate int64\n\t\tif o.Filterenddate != nil {\n\t\t\tqrFilterenddate = *o.Filterenddate\n\t\t}\n\t\tqFilterenddate := swag.FormatInt64(qrFilterenddate)\n\t\tif qFilterenddate != \"\" {\n\t\t\tif err := r.SetQueryParam(\"filterenddate\", qFilterenddate); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Since != nil {\n\n\t\t// query param since\n\t\tvar qrSince int64\n\t\tif o.Since != nil {\n\t\t\tqrSince = *o.Since\n\t\t}\n\t\tqSince := swag.FormatInt64(qrSince)\n\t\tif qSince != \"\" {\n\t\t\tif err := r.SetQueryParam(\"since\", qSince); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Zone != nil {\n\n\t\t// query param zone\n\t\tvar qrZone string\n\t\tif o.Zone != nil {\n\t\t\tqrZone = *o.Zone\n\t\t}\n\t\tqZone := qrZone\n\t\tif qZone != \"\" {\n\t\t\tif err := r.SetQueryParam(\"zone\", qZone); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetOrderParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// header param Authorization\n\tif err := r.SetHeaderParam(\"Authorization\", o.Authorization); err != nil {\n\t\treturn err\n\t}\n\n\t// path param id\n\tif err := r.SetPathParam(\"id\", o.ID); err != nil {\n\t\treturn err\n\t}\n\n\tif o.MerchantID != nil {\n\n\t\t// query param merchantId\n\t\tvar qrMerchantID int64\n\t\tif o.MerchantID != nil {\n\t\t\tqrMerchantID = *o.MerchantID\n\t\t}\n\t\tqMerchantID := swag.FormatInt64(qrMerchantID)\n\t\tif qMerchantID != \"\" {\n\t\t\tif err := r.SetQueryParam(\"merchantId\", qMerchantID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetPropertyDescriptorParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param id\n\tif err := r.SetPathParam(\"id\", o.ID); err != nil {\n\t\treturn err\n\t}\n\n\t// query param propertyName\n\tqrPropertyName := o.PropertyName\n\tqPropertyName := qrPropertyName\n\tif qPropertyName != \"\" {\n\n\t\tif err := r.SetQueryParam(\"propertyName\", qPropertyName); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetCurrentGenerationParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *CreateGitWebhookUsingPOSTParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// header param Authorization\n\tif err := r.SetHeaderParam(\"Authorization\", o.Authorization); err != nil {\n\t\treturn err\n\t}\n\n\t// query param apiVersion\n\tqrAPIVersion := o.APIVersion\n\tqAPIVersion := qrAPIVersion\n\n\tif err := r.SetQueryParam(\"apiVersion\", qAPIVersion); err != nil {\n\t\treturn err\n\t}\n\tif err := r.SetBodyParam(o.GitWebhookSpec); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ViewsGetByIDParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param identifier\n\tif err := r.SetPathParam(\"identifier\", o.Identifier); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *UpdateDeviceParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// header param Authorization\n\tif err := r.SetHeaderParam(\"Authorization\", o.Authorization); err != nil {\n\t\treturn err\n\t}\n\n\t// path param deviceId\n\tif err := r.SetPathParam(\"deviceId\", o.DeviceID); err != nil {\n\t\treturn err\n\t}\n\n\tif o.UpdateDeviceRequest != nil {\n\t\tif err := r.SetBodyParam(o.UpdateDeviceRequest); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *SaveTemplateParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// query param api-version\n\tqrAPIVersion := o.APIVersion\n\tqAPIVersion := qrAPIVersion\n\tif qAPIVersion != \"\" {\n\t\tif err := r.SetQueryParam(\"api-version\", qAPIVersion); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.Body != nil {\n\t\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param organization\n\tif err := r.SetPathParam(\"organization\", o.Organization); err != nil {\n\t\treturn err\n\t}\n\n\t// path param project\n\tif err := r.SetPathParam(\"project\", o.Project); err != nil {\n\t\treturn err\n\t}\n\n\t// path param templateId\n\tif err := r.SetPathParam(\"templateId\", o.TemplateID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ConvertParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param from.currency_code\n\tif err := r.SetPathParam(\"from.currency_code\", o.FromCurrencyCode); err != nil {\n\t\treturn err\n\t}\n\n\tif o.FromNanos != nil {\n\n\t\t// query param from.nanos\n\t\tvar qrFromNanos int32\n\t\tif o.FromNanos != nil {\n\t\t\tqrFromNanos = *o.FromNanos\n\t\t}\n\t\tqFromNanos := swag.FormatInt32(qrFromNanos)\n\t\tif qFromNanos != \"\" {\n\t\t\tif err := r.SetQueryParam(\"from.nanos\", qFromNanos); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.FromUnits != nil {\n\n\t\t// query param from.units\n\t\tvar qrFromUnits string\n\t\tif o.FromUnits != nil {\n\t\t\tqrFromUnits = *o.FromUnits\n\t\t}\n\t\tqFromUnits := qrFromUnits\n\t\tif qFromUnits != \"\" {\n\t\t\tif err := r.SetQueryParam(\"from.units\", qFromUnits); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// path param to_code\n\tif err := r.SetPathParam(\"to_code\", o.ToCode); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *SystemEventsParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Filters != nil {\n\n\t\t// query param filters\n\t\tvar qrFilters string\n\t\tif o.Filters != nil {\n\t\t\tqrFilters = *o.Filters\n\t\t}\n\t\tqFilters := qrFilters\n\t\tif qFilters != \"\" {\n\t\t\tif err := r.SetQueryParam(\"filters\", qFilters); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Since != nil {\n\n\t\t// query param since\n\t\tvar qrSince string\n\t\tif o.Since != nil {\n\t\t\tqrSince = *o.Since\n\t\t}\n\t\tqSince := qrSince\n\t\tif qSince != \"\" {\n\t\t\tif err := r.SetQueryParam(\"since\", qSince); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Until != nil {\n\n\t\t// query param until\n\t\tvar qrUntil string\n\t\tif o.Until != nil {\n\t\t\tqrUntil = *o.Until\n\t\t}\n\t\tqUntil := qrUntil\n\t\tif qUntil != \"\" {\n\t\t\tif err := r.SetQueryParam(\"until\", qUntil); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetBundleByKeyParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Audit != nil {\n\n\t\t// query param audit\n\t\tvar qrAudit string\n\n\t\tif o.Audit != nil {\n\t\t\tqrAudit = *o.Audit\n\t\t}\n\t\tqAudit := qrAudit\n\t\tif qAudit != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"audit\", qAudit); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\t// query param externalKey\n\tqrExternalKey := o.ExternalKey\n\tqExternalKey := qrExternalKey\n\tif qExternalKey != \"\" {\n\n\t\tif err := r.SetQueryParam(\"externalKey\", qExternalKey); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.IncludedDeleted != nil {\n\n\t\t// query param includedDeleted\n\t\tvar qrIncludedDeleted bool\n\n\t\tif o.IncludedDeleted != nil {\n\t\t\tqrIncludedDeleted = *o.IncludedDeleted\n\t\t}\n\t\tqIncludedDeleted := swag.FormatBool(qrIncludedDeleted)\n\t\tif qIncludedDeleted != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"includedDeleted\", qIncludedDeleted); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\t// header param WithProfilingInfo\n\tif o.WithProfilingInfo != nil && len(*o.WithProfilingInfo) > 0 {\n\t\tif err := r.SetHeaderParam(\"X-Killbill-Profiling-Req\", *o.WithProfilingInfo); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// header param withStackTrace\n\tif o.WithStackTrace != nil && *o.WithStackTrace {\n\t\tif err := r.SetQueryParam(\"withStackTrace\", \"true\"); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *SwarmUpdateParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Body != nil {\n\t\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.RotateManagerToken != nil {\n\n\t\t// query param rotateManagerToken\n\t\tvar qrRotateManagerToken bool\n\t\tif o.RotateManagerToken != nil {\n\t\t\tqrRotateManagerToken = *o.RotateManagerToken\n\t\t}\n\t\tqRotateManagerToken := swag.FormatBool(qrRotateManagerToken)\n\t\tif qRotateManagerToken != \"\" {\n\t\t\tif err := r.SetQueryParam(\"rotateManagerToken\", qRotateManagerToken); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.RotateManagerUnlockKey != nil {\n\n\t\t// query param rotateManagerUnlockKey\n\t\tvar qrRotateManagerUnlockKey bool\n\t\tif o.RotateManagerUnlockKey != nil {\n\t\t\tqrRotateManagerUnlockKey = *o.RotateManagerUnlockKey\n\t\t}\n\t\tqRotateManagerUnlockKey := swag.FormatBool(qrRotateManagerUnlockKey)\n\t\tif qRotateManagerUnlockKey != \"\" {\n\t\t\tif err := r.SetQueryParam(\"rotateManagerUnlockKey\", qRotateManagerUnlockKey); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.RotateWorkerToken != nil {\n\n\t\t// query param rotateWorkerToken\n\t\tvar qrRotateWorkerToken bool\n\t\tif o.RotateWorkerToken != nil {\n\t\t\tqrRotateWorkerToken = *o.RotateWorkerToken\n\t\t}\n\t\tqRotateWorkerToken := swag.FormatBool(qrRotateWorkerToken)\n\t\tif qRotateWorkerToken != \"\" {\n\t\t\tif err := r.SetQueryParam(\"rotateWorkerToken\", qRotateWorkerToken); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// query param version\n\tqrVersion := o.Version\n\tqVersion := swag.FormatInt64(qrVersion)\n\tif qVersion != \"\" {\n\t\tif err := r.SetQueryParam(\"version\", qVersion); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ServiceInstanceGetParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.XBrokerAPIOriginatingIdentity != nil {\n\n\t\t// header param X-Broker-API-Originating-Identity\n\t\tif err := r.SetHeaderParam(\"X-Broker-API-Originating-Identity\", *o.XBrokerAPIOriginatingIdentity); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// header param X-Broker-API-Version\n\tif err := r.SetHeaderParam(\"X-Broker-API-Version\", o.XBrokerAPIVersion); err != nil {\n\t\treturn err\n\t}\n\n\t// path param instance_id\n\tif err := r.SetPathParam(\"instance_id\", o.InstanceID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ShowPackageParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param media_type\n\tif err := r.SetPathParam(\"media_type\", o.MediaType); err != nil {\n\t\treturn err\n\t}\n\n\t// path param namespace\n\tif err := r.SetPathParam(\"namespace\", o.Namespace); err != nil {\n\t\treturn err\n\t}\n\n\t// path param package\n\tif err := r.SetPathParam(\"package\", o.Package); err != nil {\n\t\treturn err\n\t}\n\n\t// path param release\n\tif err := r.SetPathParam(\"release\", o.Release); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *SizeParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif err := r.SetBodyParam(o.Parameters); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetOutagesParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// query param count\n\tqrCount := o.Count\n\tqCount := swag.FormatFloat64(qrCount)\n\tif qCount != \"\" {\n\n\t\tif err := r.SetQueryParam(\"count\", qCount); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.DeviceID != nil {\n\n\t\t// query param deviceId\n\t\tvar qrDeviceID string\n\n\t\tif o.DeviceID != nil {\n\t\t\tqrDeviceID = *o.DeviceID\n\t\t}\n\t\tqDeviceID := qrDeviceID\n\t\tif qDeviceID != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"deviceId\", qDeviceID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.InProgress != nil {\n\n\t\t// query param inProgress\n\t\tvar qrInProgress bool\n\n\t\tif o.InProgress != nil {\n\t\t\tqrInProgress = *o.InProgress\n\t\t}\n\t\tqInProgress := swag.FormatBool(qrInProgress)\n\t\tif qInProgress != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"inProgress\", qInProgress); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\t// query param page\n\tqrPage := o.Page\n\tqPage := swag.FormatFloat64(qrPage)\n\tif qPage != \"\" {\n\n\t\tif err := r.SetQueryParam(\"page\", qPage); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.Period != nil {\n\n\t\t// query param period\n\t\tvar qrPeriod float64\n\n\t\tif o.Period != nil {\n\t\t\tqrPeriod = *o.Period\n\t\t}\n\t\tqPeriod := swag.FormatFloat64(qrPeriod)\n\t\tif qPeriod != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"period\", qPeriod); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Query != nil {\n\n\t\t// query param query\n\t\tvar qrQuery string\n\n\t\tif o.Query != nil {\n\t\t\tqrQuery = *o.Query\n\t\t}\n\t\tqQuery := qrQuery\n\t\tif qQuery != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"query\", qQuery); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Type != nil {\n\n\t\t// query param type\n\t\tvar qrType string\n\n\t\tif o.Type != nil {\n\t\t\tqrType = *o.Type\n\t\t}\n\t\tqType := qrType\n\t\tif qType != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"type\", qType); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *TerminateParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param extractorId\n\tif err := r.SetPathParam(\"extractorId\", o.ExtractorID); err != nil {\n\t\treturn err\n\t}\n\n\t// path param inputId\n\tif err := r.SetPathParam(\"inputId\", o.InputID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ServeFieldParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param field\n\tif err := r.SetPathParam(\"field\", o.Field); err != nil {\n\t\treturn err\n\t}\n\n\t// path param vcsRootLocator\n\tif err := r.SetPathParam(\"vcsRootLocator\", o.VcsRootLocator); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *PostV1DevicesParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// form param device_identifier\n\tfrDeviceIdentifier := o.DeviceIdentifier\n\tfDeviceIdentifier := frDeviceIdentifier\n\tif fDeviceIdentifier != \"\" {\n\t\tif err := r.SetFormParam(\"device_identifier\", fDeviceIdentifier); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// form param kind\n\tfrKind := o.Kind\n\tfKind := frKind\n\tif fKind != \"\" {\n\t\tif err := r.SetFormParam(\"kind\", fKind); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// form param name\n\tfrName := o.Name\n\tfName := frName\n\tif fName != \"\" {\n\t\tif err := r.SetFormParam(\"name\", fName); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.NotificationIdentifier != nil {\n\n\t\t// form param notification_identifier\n\t\tvar frNotificationIdentifier string\n\t\tif o.NotificationIdentifier != nil {\n\t\t\tfrNotificationIdentifier = *o.NotificationIdentifier\n\t\t}\n\t\tfNotificationIdentifier := frNotificationIdentifier\n\t\tif fNotificationIdentifier != \"\" {\n\t\t\tif err := r.SetFormParam(\"notification_identifier\", fNotificationIdentifier); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.SubscribeNotification != nil {\n\n\t\t// form param subscribe_notification\n\t\tvar frSubscribeNotification bool\n\t\tif o.SubscribeNotification != nil {\n\t\t\tfrSubscribeNotification = *o.SubscribeNotification\n\t\t}\n\t\tfSubscribeNotification := swag.FormatBool(frSubscribeNotification)\n\t\tif fSubscribeNotification != \"\" {\n\t\t\tif err := r.SetFormParam(\"subscribe_notification\", fSubscribeNotification); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *QueryFirewallFieldsParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Limit != nil {\n\n\t\t// query param limit\n\t\tvar qrLimit int64\n\n\t\tif o.Limit != nil {\n\t\t\tqrLimit = *o.Limit\n\t\t}\n\t\tqLimit := swag.FormatInt64(qrLimit)\n\t\tif qLimit != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"limit\", qLimit); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Offset != nil {\n\n\t\t// query param offset\n\t\tvar qrOffset string\n\n\t\tif o.Offset != nil {\n\t\t\tqrOffset = *o.Offset\n\t\t}\n\t\tqOffset := qrOffset\n\t\tif qOffset != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"offset\", qOffset); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.PlatformID != nil {\n\n\t\t// query param platform_id\n\t\tvar qrPlatformID string\n\n\t\tif o.PlatformID != nil {\n\t\t\tqrPlatformID = *o.PlatformID\n\t\t}\n\t\tqPlatformID := qrPlatformID\n\t\tif qPlatformID != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"platform_id\", qPlatformID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetCatalogXMLParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.AccountID != nil {\n\n\t\t// query param accountId\n\t\tvar qrAccountID strfmt.UUID\n\n\t\tif o.AccountID != nil {\n\t\t\tqrAccountID = *o.AccountID\n\t\t}\n\t\tqAccountID := qrAccountID.String()\n\t\tif qAccountID != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"accountId\", qAccountID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.RequestedDate != nil {\n\n\t\t// query param requestedDate\n\t\tvar qrRequestedDate strfmt.DateTime\n\n\t\tif o.RequestedDate != nil {\n\t\t\tqrRequestedDate = *o.RequestedDate\n\t\t}\n\t\tqRequestedDate := qrRequestedDate.String()\n\t\tif qRequestedDate != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"requestedDate\", qRequestedDate); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\t// header param WithProfilingInfo\n\tif o.WithProfilingInfo != nil && len(*o.WithProfilingInfo) > 0 {\n\t\tif err := r.SetHeaderParam(\"X-Killbill-Profiling-Req\", *o.WithProfilingInfo); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// header param withStackTrace\n\tif o.WithStackTrace != nil && *o.WithStackTrace {\n\t\tif err := r.SetQueryParam(\"withStackTrace\", \"true\"); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetClockParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// header param X-Killbill-ApiKey\n\tif err := r.SetHeaderParam(\"X-Killbill-ApiKey\", o.XKillbillAPIKey); err != nil {\n\t\treturn err\n\t}\n\n\t// header param X-Killbill-ApiSecret\n\tif err := r.SetHeaderParam(\"X-Killbill-ApiSecret\", o.XKillbillAPISecret); err != nil {\n\t\treturn err\n\t}\n\n\t// header param WithProfilingInfo\n\tif o.WithProfilingInfo != nil && len(*o.WithProfilingInfo) > 0 {\n\t\tif err := r.SetHeaderParam(\"X-Killbill-Profiling-Req\", *o.WithProfilingInfo); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// header param withStackTrace\n\tif o.WithStackTrace != nil && *o.WithStackTrace {\n\t\tif err := r.SetQueryParam(\"withStackTrace\", \"true\"); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *AdminCreateJusticeUserParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param namespace\n\tif err := r.SetPathParam(\"namespace\", o.Namespace); err != nil {\n\t\treturn err\n\t}\n\n\t// path param targetNamespace\n\tif err := r.SetPathParam(\"targetNamespace\", o.TargetNamespace); err != nil {\n\t\treturn err\n\t}\n\n\t// path param userId\n\tif err := r.SetPathParam(\"userId\", o.UserID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *UpdateWidgetParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Accept != nil {\n\n\t\t// header param Accept\n\t\tif err := r.SetHeaderParam(\"Accept\", *o.Accept); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\n\tif o.ContentType != nil {\n\n\t\t// header param Content-Type\n\t\tif err := r.SetHeaderParam(\"Content-Type\", *o.ContentType); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\n\t// path param uuid\n\tif err := r.SetPathParam(\"uuid\", o.UUID.String()); err != nil {\n\t\treturn err\n\t}\n\n\tif err := r.SetBodyParam(o.WidgetBody); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *TestEndpointParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetLogsParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.XRequestID != nil {\n\n\t\t// header param X-Request-Id\n\t\tif err := r.SetHeaderParam(\"X-Request-Id\", *o.XRequestID); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\n\tif o.Page != nil {\n\n\t\t// query param page\n\t\tvar qrPage int64\n\t\tif o.Page != nil {\n\t\t\tqrPage = *o.Page\n\t\t}\n\t\tqPage := swag.FormatInt64(qrPage)\n\t\tif qPage != \"\" {\n\t\t\tif err := r.SetQueryParam(\"page\", qPage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.PageSize != nil {\n\n\t\t// query param page_size\n\t\tvar qrPageSize int64\n\t\tif o.PageSize != nil {\n\t\t\tqrPageSize = *o.PageSize\n\t\t}\n\t\tqPageSize := swag.FormatInt64(qrPageSize)\n\t\tif qPageSize != \"\" {\n\t\t\tif err := r.SetQueryParam(\"page_size\", qPageSize); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// path param project_name\n\tif err := r.SetPathParam(\"project_name\", o.ProjectName); err != nil {\n\t\treturn err\n\t}\n\n\tif o.Q != nil {\n\n\t\t// query param q\n\t\tvar qrQ string\n\t\tif o.Q != nil {\n\t\t\tqrQ = *o.Q\n\t\t}\n\t\tqQ := qrQ\n\t\tif qQ != \"\" {\n\t\t\tif err := r.SetQueryParam(\"q\", qQ); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ListSourceFileOfProjectVersionParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Fields != nil {\n\n\t\t// query param fields\n\t\tvar qrFields string\n\t\tif o.Fields != nil {\n\t\t\tqrFields = *o.Fields\n\t\t}\n\t\tqFields := qrFields\n\t\tif qFields != \"\" {\n\t\t\tif err := r.SetQueryParam(\"fields\", qFields); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// path param parentId\n\tif err := r.SetPathParam(\"parentId\", swag.FormatInt64(o.ParentID)); err != nil {\n\t\treturn err\n\t}\n\n\tif o.Q != nil {\n\n\t\t// query param q\n\t\tvar qrQ string\n\t\tif o.Q != nil {\n\t\t\tqrQ = *o.Q\n\t\t}\n\t\tqQ := qrQ\n\t\tif qQ != \"\" {\n\t\t\tif err := r.SetQueryParam(\"q\", qQ); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetDrgParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param drgId\n\tif err := r.SetPathParam(\"drgId\", o.DrgID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *UpdateFlowParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\tif o.Body != nil {\n\t\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param bucketId\n\tif err := r.SetPathParam(\"bucketId\", o.BucketID); err != nil {\n\t\treturn err\n\t}\n\n\t// path param flowId\n\tif err := r.SetPathParam(\"flowId\", o.FlowID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *CreateWidgetParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Accept != nil {\n\n\t\t// header param Accept\n\t\tif err := r.SetHeaderParam(\"Accept\", *o.Accept); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\n\tif o.ContentType != nil {\n\n\t\t// header param Content-Type\n\t\tif err := r.SetHeaderParam(\"Content-Type\", *o.ContentType); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\n\tif err := r.SetBodyParam(o.WidgetBody); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetBodyResourceByDatePeriodParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param date\n\tif err := r.SetPathParam(\"date\", o.Date.String()); err != nil {\n\t\treturn err\n\t}\n\n\t// path param period\n\tif err := r.SetPathParam(\"period\", o.Period); err != nil {\n\t\treturn err\n\t}\n\n\t// path param resource-path\n\tif err := r.SetPathParam(\"resource-path\", o.ResourcePath); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetAboutUserParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tvaluesSelect := o.Select\n\n\tjoinedSelect := swag.JoinByFormat(valuesSelect, \"csv\")\n\t// query array param select\n\tif err := r.SetQueryParam(\"select\", joinedSelect...); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ExtractionListV1Params) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// query param id\n\tqrID := o.ID\n\tqID := qrID\n\tif qID != \"\" {\n\n\t\tif err := r.SetQueryParam(\"id\", qID); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.Limit != nil {\n\n\t\t// query param limit\n\t\tvar qrLimit int64\n\n\t\tif o.Limit != nil {\n\t\t\tqrLimit = *o.Limit\n\t\t}\n\t\tqLimit := swag.FormatInt64(qrLimit)\n\t\tif qLimit != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"limit\", qLimit); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Offset != nil {\n\n\t\t// query param offset\n\t\tvar qrOffset string\n\n\t\tif o.Offset != nil {\n\t\t\tqrOffset = *o.Offset\n\t\t}\n\t\tqOffset := qrOffset\n\t\tif qOffset != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"offset\", qOffset); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetAuditEventsParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Page != nil {\n\n\t\t// query param page\n\t\tvar qrPage int32\n\t\tif o.Page != nil {\n\t\t\tqrPage = *o.Page\n\t\t}\n\t\tqPage := swag.FormatInt32(qrPage)\n\t\tif qPage != \"\" {\n\t\t\tif err := r.SetQueryParam(\"page\", qPage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// query param resourceCrn\n\tqrResourceCrn := o.ResourceCrn\n\tqResourceCrn := qrResourceCrn\n\tif qResourceCrn != \"\" {\n\t\tif err := r.SetQueryParam(\"resourceCrn\", qResourceCrn); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.Size != nil {\n\n\t\t// query param size\n\t\tvar qrSize int32\n\t\tif o.Size != nil {\n\t\t\tqrSize = *o.Size\n\t\t}\n\t\tqSize := swag.FormatInt32(qrSize)\n\t\tif qSize != \"\" {\n\t\t\tif err := r.SetQueryParam(\"size\", qSize); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *PcloudSystempoolsGetParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param cloud_instance_id\n\tif err := r.SetPathParam(\"cloud_instance_id\", o.CloudInstanceID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *WaitListParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param address\n\tif err := r.SetPathParam(\"address\", o.Address); err != nil {\n\t\treturn err\n\t}\n\n\tif o.Height != nil {\n\n\t\t// query param height\n\t\tvar qrHeight uint64\n\n\t\tif o.Height != nil {\n\t\t\tqrHeight = *o.Height\n\t\t}\n\t\tqHeight := swag.FormatUint64(qrHeight)\n\t\tif qHeight != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"height\", qHeight); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.PublicKey != nil {\n\n\t\t// query param public_key\n\t\tvar qrPublicKey string\n\n\t\tif o.PublicKey != nil {\n\t\t\tqrPublicKey = *o.PublicKey\n\t\t}\n\t\tqPublicKey := qrPublicKey\n\t\tif qPublicKey != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"public_key\", qPublicKey); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *BudgetAddParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\tif o.Body != nil {\n\t\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetGCParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param gc_id\n\tif err := r.SetPathParam(\"gc_id\", swag.FormatInt64(o.GcID)); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *PartialUpdateAppParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tr.SetTimeout(o.timeout)\n\tvar res []error\n\n\t// path param app_id\n\tif err := r.SetPathParam(\"app_id\", swag.FormatInt64(o.AppID)); err != nil {\n\t\treturn err\n\t}\n\n\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\treturn err\n\t}\n\n\t// path param team_id\n\tif err := r.SetPathParam(\"team_id\", swag.FormatInt64(o.TeamID)); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *StartPacketCaptureParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param id\n\tif err := r.SetPathParam(\"id\", o.ID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *TaskSchemasIDGetParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param identifier\n\tif err := r.SetPathParam(\"identifier\", o.Identifier); err != nil {\n\t\treturn err\n\t}\n\n\tif o.ResolveRef != nil {\n\n\t\t// query param resolveRef\n\t\tvar qrResolveRef bool\n\t\tif o.ResolveRef != nil {\n\t\t\tqrResolveRef = *o.ResolveRef\n\t\t}\n\t\tqResolveRef := swag.FormatBool(qrResolveRef)\n\t\tif qResolveRef != \"\" {\n\t\t\tif err := r.SetQueryParam(\"resolveRef\", qResolveRef); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *UploadTaskFileParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Description != nil {\n\n\t\t// form param description\n\t\tvar frDescription string\n\t\tif o.Description != nil {\n\t\t\tfrDescription = *o.Description\n\t\t}\n\t\tfDescription := frDescription\n\t\tif fDescription != \"\" {\n\t\t\tif err := r.SetFormParam(\"description\", fDescription); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.File != nil {\n\n\t\tif o.File != nil {\n\t\t\t// form file param file\n\t\t\tif err := r.SetFileParam(\"file\", o.File); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\t// path param id\n\tif err := r.SetPathParam(\"id\", swag.FormatInt64(o.ID)); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *PetCreateParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\tif o.Body != nil {\n\t\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *QueryChangesParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Filter != nil {\n\n\t\t// query param filter\n\t\tvar qrFilter string\n\n\t\tif o.Filter != nil {\n\t\t\tqrFilter = *o.Filter\n\t\t}\n\t\tqFilter := qrFilter\n\t\tif qFilter != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"filter\", qFilter); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Limit != nil {\n\n\t\t// query param limit\n\t\tvar qrLimit int64\n\n\t\tif o.Limit != nil {\n\t\t\tqrLimit = *o.Limit\n\t\t}\n\t\tqLimit := swag.FormatInt64(qrLimit)\n\t\tif qLimit != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"limit\", qLimit); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Offset != nil {\n\n\t\t// query param offset\n\t\tvar qrOffset int64\n\n\t\tif o.Offset != nil {\n\t\t\tqrOffset = *o.Offset\n\t\t}\n\t\tqOffset := swag.FormatInt64(qrOffset)\n\t\tif qOffset != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"offset\", qOffset); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Sort != nil {\n\n\t\t// query param sort\n\t\tvar qrSort string\n\n\t\tif o.Sort != nil {\n\t\t\tqrSort = *o.Sort\n\t\t}\n\t\tqSort := qrSort\n\t\tif qSort != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"sort\", qSort); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetInstrumentsInstrumentOrderBookParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.AcceptDatetimeFormat != nil {\n\n\t\t// header param Accept-Datetime-Format\n\t\tif err := r.SetHeaderParam(\"Accept-Datetime-Format\", *o.AcceptDatetimeFormat); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\n\t// header param Authorization\n\tif err := r.SetHeaderParam(\"Authorization\", o.Authorization); err != nil {\n\t\treturn err\n\t}\n\n\t// path param instrument\n\tif err := r.SetPathParam(\"instrument\", o.Instrument); err != nil {\n\t\treturn err\n\t}\n\n\tif o.Time != nil {\n\n\t\t// query param time\n\t\tvar qrTime string\n\t\tif o.Time != nil {\n\t\t\tqrTime = *o.Time\n\t\t}\n\t\tqTime := qrTime\n\t\tif qTime != \"\" {\n\t\t\tif err := r.SetQueryParam(\"time\", qTime); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetScopeConfigurationParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param scope_id\n\tif err := r.SetPathParam(\"scope_id\", o.ScopeID); err != nil {\n\t\treturn err\n\t}\n\n\t// path param site_id\n\tif err := r.SetPathParam(\"site_id\", o.SiteID); err != nil {\n\t\treturn err\n\t}\n\n\t// path param stack_id\n\tif err := r.SetPathParam(\"stack_id\", o.StackID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *UpdateEventParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Body != nil {\n\t\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param eventId\n\tif err := r.SetPathParam(\"eventId\", o.EventID); err != nil {\n\t\treturn err\n\t}\n\n\t// path param koronaAccountId\n\tif err := r.SetPathParam(\"koronaAccountId\", o.KoronaAccountID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetV1FunctionalitiesParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Impacted != nil {\n\n\t\t// query param impacted\n\t\tvar qrImpacted string\n\n\t\tif o.Impacted != nil {\n\t\t\tqrImpacted = *o.Impacted\n\t\t}\n\t\tqImpacted := qrImpacted\n\t\tif qImpacted != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"impacted\", qImpacted); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Labels != nil {\n\n\t\t// query param labels\n\t\tvar qrLabels string\n\n\t\tif o.Labels != nil {\n\t\t\tqrLabels = *o.Labels\n\t\t}\n\t\tqLabels := qrLabels\n\t\tif qLabels != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"labels\", qLabels); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Lite != nil {\n\n\t\t// query param lite\n\t\tvar qrLite bool\n\n\t\tif o.Lite != nil {\n\t\t\tqrLite = *o.Lite\n\t\t}\n\t\tqLite := swag.FormatBool(qrLite)\n\t\tif qLite != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"lite\", qLite); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Name != nil {\n\n\t\t// query param name\n\t\tvar qrName string\n\n\t\tif o.Name != nil {\n\t\t\tqrName = *o.Name\n\t\t}\n\t\tqName := qrName\n\t\tif qName != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"name\", qName); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Owner != nil {\n\n\t\t// query param owner\n\t\tvar qrOwner string\n\n\t\tif o.Owner != nil {\n\t\t\tqrOwner = *o.Owner\n\t\t}\n\t\tqOwner := qrOwner\n\t\tif qOwner != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"owner\", qOwner); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Page != nil {\n\n\t\t// query param page\n\t\tvar qrPage int32\n\n\t\tif o.Page != nil {\n\t\t\tqrPage = *o.Page\n\t\t}\n\t\tqPage := swag.FormatInt32(qrPage)\n\t\tif qPage != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"page\", qPage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.PerPage != nil {\n\n\t\t// query param per_page\n\t\tvar qrPerPage int32\n\n\t\tif o.PerPage != nil {\n\t\t\tqrPerPage = *o.PerPage\n\t\t}\n\t\tqPerPage := swag.FormatInt32(qrPerPage)\n\t\tif qPerPage != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"per_page\", qPerPage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Query != nil {\n\n\t\t// query param query\n\t\tvar qrQuery string\n\n\t\tif o.Query != nil {\n\t\t\tqrQuery = *o.Query\n\t\t}\n\t\tqQuery := qrQuery\n\t\tif qQuery != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"query\", qQuery); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ContainerUpdateParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param id\n\tif err := r.SetPathParam(\"id\", o.ID); err != nil {\n\t\treturn err\n\t}\n\n\tif err := r.SetBodyParam(o.Update); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetPointsByQueryParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.DollarSkip != nil {\n\n\t\t// query param $skip\n\t\tvar qrNrDollarSkip int32\n\t\tif o.DollarSkip != nil {\n\t\t\tqrNrDollarSkip = *o.DollarSkip\n\t\t}\n\t\tqNrDollarSkip := swag.FormatInt32(qrNrDollarSkip)\n\t\tif qNrDollarSkip != \"\" {\n\t\t\tif err := r.SetQueryParam(\"$skip\", qNrDollarSkip); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.DollarTop != nil {\n\n\t\t// query param $top\n\t\tvar qrNrDollarTop int32\n\t\tif o.DollarTop != nil {\n\t\t\tqrNrDollarTop = *o.DollarTop\n\t\t}\n\t\tqNrDollarTop := swag.FormatInt32(qrNrDollarTop)\n\t\tif qNrDollarTop != \"\" {\n\t\t\tif err := r.SetQueryParam(\"$top\", qNrDollarTop); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// query param api-version\n\tqrAPIVersion := o.APIVersion\n\tqAPIVersion := qrAPIVersion\n\tif qAPIVersion != \"\" {\n\t\tif err := r.SetQueryParam(\"api-version\", qAPIVersion); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.Body != nil {\n\t\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param organization\n\tif err := r.SetPathParam(\"organization\", o.Organization); err != nil {\n\t\treturn err\n\t}\n\n\t// path param project\n\tif err := r.SetPathParam(\"project\", o.Project); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *SyncStatusUsingGETParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param namespaceSelfLinkId\n\tif err := r.SetPathParam(\"namespaceSelfLinkId\", o.NamespaceSelfLinkID); err != nil {\n\t\treturn err\n\t}\n\n\t// path param requestId\n\tif err := r.SetPathParam(\"requestId\", o.RequestID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ResolveBatchParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Account != nil {\n\n\t\t// query param account\n\t\tvar qrAccount string\n\t\tif o.Account != nil {\n\t\t\tqrAccount = *o.Account\n\t\t}\n\t\tqAccount := qrAccount\n\t\tif qAccount != \"\" {\n\t\t\tif err := r.SetQueryParam(\"account\", qAccount); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Environment != nil {\n\n\t\t// query param environment\n\t\tvar qrEnvironment string\n\t\tif o.Environment != nil {\n\t\t\tqrEnvironment = *o.Environment\n\t\t}\n\t\tqEnvironment := qrEnvironment\n\t\tif qEnvironment != \"\" {\n\t\t\tif err := r.SetQueryParam(\"environment\", qEnvironment); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.From != nil {\n\n\t\t// query param from\n\t\tvar qrFrom string\n\t\tif o.From != nil {\n\t\t\tqrFrom = *o.From\n\t\t}\n\t\tqFrom := qrFrom\n\t\tif qFrom != \"\" {\n\t\t\tif err := r.SetQueryParam(\"from\", qFrom); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Page != nil {\n\n\t\t// query param page\n\t\tvar qrPage int32\n\t\tif o.Page != nil {\n\t\t\tqrPage = *o.Page\n\t\t}\n\t\tqPage := swag.FormatInt32(qrPage)\n\t\tif qPage != \"\" {\n\t\t\tif err := r.SetQueryParam(\"page\", qPage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Region != nil {\n\n\t\t// query param region\n\t\tvar qrRegion string\n\t\tif o.Region != nil {\n\t\t\tqrRegion = *o.Region\n\t\t}\n\t\tqRegion := qrRegion\n\t\tif qRegion != \"\" {\n\t\t\tif err := r.SetQueryParam(\"region\", qRegion); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.To != nil {\n\n\t\t// query param to\n\t\tvar qrTo string\n\t\tif o.To != nil {\n\t\t\tqrTo = *o.To\n\t\t}\n\t\tqTo := qrTo\n\t\tif qTo != \"\" {\n\t\t\tif err := r.SetQueryParam(\"to\", qTo); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetAccountParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tr.SetTimeout(o.timeout)\n\tvar res []error\n\n\tif o.Authorization != nil {\n\n\t\t// header param Authorization\n\t\tif err := r.SetHeaderParam(\"Authorization\", *o.Authorization); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\n\tif o.Country != nil {\n\n\t\t// query param country\n\t\tvar qrCountry string\n\t\tif o.Country != nil {\n\t\t\tqrCountry = *o.Country\n\t\t}\n\t\tqCountry := qrCountry\n\t\tif qCountry != \"\" {\n\t\t\tif err := r.SetQueryParam(\"country\", qCountry); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Email != nil {\n\n\t\t// query param email\n\t\tvar qrEmail string\n\t\tif o.Email != nil {\n\t\t\tqrEmail = *o.Email\n\t\t}\n\t\tqEmail := qrEmail\n\t\tif qEmail != \"\" {\n\t\t\tif err := r.SetQueryParam(\"email\", qEmail); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tvaluesFields := o.Fields\n\n\tjoinedFields := swag.JoinByFormat(valuesFields, \"csv\")\n\t// query array param fields\n\tif err := r.SetQueryParam(\"fields\", joinedFields...); err != nil {\n\t\treturn err\n\t}\n\n\tif o.PersonID != nil {\n\n\t\t// query param person_id\n\t\tvar qrPersonID string\n\t\tif o.PersonID != nil {\n\t\t\tqrPersonID = *o.PersonID\n\t\t}\n\t\tqPersonID := qrPersonID\n\t\tif qPersonID != \"\" {\n\t\t\tif err := r.SetQueryParam(\"person_id\", qPersonID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetDeviceHealthParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// header param Authorization\n\tif err := r.SetHeaderParam(\"Authorization\", o.Authorization); err != nil {\n\t\treturn err\n\t}\n\n\t// path param deviceId\n\tif err := r.SetPathParam(\"deviceId\", o.DeviceID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *UpdatePatientParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param ID\n\tif err := r.SetPathParam(\"ID\", o.ID); err != nil {\n\t\treturn err\n\t}\n\n\tif o.Patient != nil {\n\t\tif err := r.SetBodyParam(o.Patient); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *UpdateCustomIDPParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.CustomIDP != nil {\n\t\tif err := r.SetBodyParam(o.CustomIDP); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param aid\n\tif err := r.SetPathParam(\"aid\", o.Aid); err != nil {\n\t\treturn err\n\t}\n\n\t// path param iid\n\tif err := r.SetPathParam(\"iid\", o.Iid); err != nil {\n\t\treturn err\n\t}\n\n\t// path param tid\n\tif err := r.SetPathParam(\"tid\", o.Tid); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetSeriesIDFilterParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.AcceptLanguage != nil {\n\n\t\t// header param Accept-Language\n\t\tif err := r.SetHeaderParam(\"Accept-Language\", *o.AcceptLanguage); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\n\t// path param id\n\tif err := r.SetPathParam(\"id\", swag.FormatInt64(o.ID)); err != nil {\n\t\treturn err\n\t}\n\n\t// query param keys\n\tqrKeys := o.Keys\n\tqKeys := qrKeys\n\tif qKeys != \"\" {\n\t\tif err := r.SetQueryParam(\"keys\", qKeys); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *CreateBlueprintInWorkspaceInternalParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.AccountID != nil {\n\n\t\t// query param accountId\n\t\tvar qrAccountID string\n\t\tif o.AccountID != nil {\n\t\t\tqrAccountID = *o.AccountID\n\t\t}\n\t\tqAccountID := qrAccountID\n\t\tif qAccountID != \"\" {\n\t\t\tif err := r.SetQueryParam(\"accountId\", qAccountID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Body != nil {\n\t\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param workspaceId\n\tif err := r.SetPathParam(\"workspaceId\", swag.FormatInt64(o.WorkspaceID)); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *OrgGetParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param org\n\tif err := r.SetPathParam(\"org\", o.Org); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *ExtrasGraphsReadParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param id\n\tif err := r.SetPathParam(\"id\", swag.FormatInt64(o.ID)); err != nil {\n\t\treturn err\n\t}\n\n\tif o.Name != nil {\n\n\t\t// query param name\n\t\tvar qrName string\n\t\tif o.Name != nil {\n\t\t\tqrName = *o.Name\n\t\t}\n\t\tqName := qrName\n\t\tif qName != \"\" {\n\t\t\tif err := r.SetQueryParam(\"name\", qName); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Type != nil {\n\n\t\t// query param type\n\t\tvar qrType string\n\t\tif o.Type != nil {\n\t\t\tqrType = *o.Type\n\t\t}\n\t\tqType := qrType\n\t\tif qType != \"\" {\n\t\t\tif err := r.SetQueryParam(\"type\", qType); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetVersioningPolicyParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Description != nil {\n\n\t\t// query param Description\n\t\tvar qrDescription string\n\t\tif o.Description != nil {\n\t\t\tqrDescription = *o.Description\n\t\t}\n\t\tqDescription := qrDescription\n\t\tif qDescription != \"\" {\n\t\t\tif err := r.SetQueryParam(\"Description\", qDescription); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.IgnoreFilesGreaterThan != nil {\n\n\t\t// query param IgnoreFilesGreaterThan\n\t\tvar qrIgnoreFilesGreaterThan string\n\t\tif o.IgnoreFilesGreaterThan != nil {\n\t\t\tqrIgnoreFilesGreaterThan = *o.IgnoreFilesGreaterThan\n\t\t}\n\t\tqIgnoreFilesGreaterThan := qrIgnoreFilesGreaterThan\n\t\tif qIgnoreFilesGreaterThan != \"\" {\n\t\t\tif err := r.SetQueryParam(\"IgnoreFilesGreaterThan\", qIgnoreFilesGreaterThan); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.MaxSizePerFile != nil {\n\n\t\t// query param MaxSizePerFile\n\t\tvar qrMaxSizePerFile string\n\t\tif o.MaxSizePerFile != nil {\n\t\t\tqrMaxSizePerFile = *o.MaxSizePerFile\n\t\t}\n\t\tqMaxSizePerFile := qrMaxSizePerFile\n\t\tif qMaxSizePerFile != \"\" {\n\t\t\tif err := r.SetQueryParam(\"MaxSizePerFile\", qMaxSizePerFile); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.MaxTotalSize != nil {\n\n\t\t// query param MaxTotalSize\n\t\tvar qrMaxTotalSize string\n\t\tif o.MaxTotalSize != nil {\n\t\t\tqrMaxTotalSize = *o.MaxTotalSize\n\t\t}\n\t\tqMaxTotalSize := qrMaxTotalSize\n\t\tif qMaxTotalSize != \"\" {\n\t\t\tif err := r.SetQueryParam(\"MaxTotalSize\", qMaxTotalSize); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Name != nil {\n\n\t\t// query param Name\n\t\tvar qrName string\n\t\tif o.Name != nil {\n\t\t\tqrName = *o.Name\n\t\t}\n\t\tqName := qrName\n\t\tif qName != \"\" {\n\t\t\tif err := r.SetQueryParam(\"Name\", qName); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// path param Uuid\n\tif err := r.SetPathParam(\"Uuid\", o.UUID); err != nil {\n\t\treturn err\n\t}\n\n\tif o.VersionsDataSourceBucket != nil {\n\n\t\t// query param VersionsDataSourceBucket\n\t\tvar qrVersionsDataSourceBucket string\n\t\tif o.VersionsDataSourceBucket != nil {\n\t\t\tqrVersionsDataSourceBucket = *o.VersionsDataSourceBucket\n\t\t}\n\t\tqVersionsDataSourceBucket := qrVersionsDataSourceBucket\n\t\tif qVersionsDataSourceBucket != \"\" {\n\t\t\tif err := r.SetQueryParam(\"VersionsDataSourceBucket\", qVersionsDataSourceBucket); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.VersionsDataSourceName != nil {\n\n\t\t// query param VersionsDataSourceName\n\t\tvar qrVersionsDataSourceName string\n\t\tif o.VersionsDataSourceName != nil {\n\t\t\tqrVersionsDataSourceName = *o.VersionsDataSourceName\n\t\t}\n\t\tqVersionsDataSourceName := qrVersionsDataSourceName\n\t\tif qVersionsDataSourceName != \"\" {\n\t\t\tif err := r.SetQueryParam(\"VersionsDataSourceName\", qVersionsDataSourceName); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetBuildPropertiesParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// query param api-version\n\tqrAPIVersion := o.APIVersion\n\tqAPIVersion := qrAPIVersion\n\tif qAPIVersion != \"\" {\n\t\tif err := r.SetQueryParam(\"api-version\", qAPIVersion); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param buildId\n\tif err := r.SetPathParam(\"buildId\", swag.FormatInt32(o.BuildID)); err != nil {\n\t\treturn err\n\t}\n\n\tif o.Filter != nil {\n\n\t\t// query param filter\n\t\tvar qrFilter string\n\t\tif o.Filter != nil {\n\t\t\tqrFilter = *o.Filter\n\t\t}\n\t\tqFilter := qrFilter\n\t\tif qFilter != \"\" {\n\t\t\tif err := r.SetQueryParam(\"filter\", qFilter); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// path param organization\n\tif err := r.SetPathParam(\"organization\", o.Organization); err != nil {\n\t\treturn err\n\t}\n\n\t// path param project\n\tif err := r.SetPathParam(\"project\", o.Project); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *AdminGetBannedDevicesV4Params) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param namespace\n\tif err := r.SetPathParam(\"namespace\", o.Namespace); err != nil {\n\t\treturn err\n\t}\n\n\tif o.DeviceType != nil {\n\n\t\t// query param deviceType\n\t\tvar qrDeviceType string\n\t\tif o.DeviceType != nil {\n\t\t\tqrDeviceType = *o.DeviceType\n\t\t}\n\t\tqDeviceType := qrDeviceType\n\t\tif qDeviceType != \"\" {\n\t\t\tif err := r.SetQueryParam(\"deviceType\", qDeviceType); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.EndDate != nil {\n\n\t\t// query param endDate\n\t\tvar qrEndDate string\n\t\tif o.EndDate != nil {\n\t\t\tqrEndDate = *o.EndDate\n\t\t}\n\t\tqEndDate := qrEndDate\n\t\tif qEndDate != \"\" {\n\t\t\tif err := r.SetQueryParam(\"endDate\", qEndDate); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Limit != nil {\n\n\t\t// query param limit\n\t\tvar qrLimit int64\n\t\tif o.Limit != nil {\n\t\t\tqrLimit = *o.Limit\n\t\t}\n\t\tqLimit := swag.FormatInt64(qrLimit)\n\t\tif qLimit != \"\" {\n\t\t\tif err := r.SetQueryParam(\"limit\", qLimit); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Offset != nil {\n\n\t\t// query param offset\n\t\tvar qrOffset int64\n\t\tif o.Offset != nil {\n\t\t\tqrOffset = *o.Offset\n\t\t}\n\t\tqOffset := swag.FormatInt64(qrOffset)\n\t\tif qOffset != \"\" {\n\t\t\tif err := r.SetQueryParam(\"offset\", qOffset); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.StartDate != nil {\n\n\t\t// query param startDate\n\t\tvar qrStartDate string\n\t\tif o.StartDate != nil {\n\t\t\tqrStartDate = *o.StartDate\n\t\t}\n\t\tqStartDate := qrStartDate\n\t\tif qStartDate != \"\" {\n\t\t\tif err := r.SetQueryParam(\"startDate\", qStartDate); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// setting the default header value\n\tif err := r.SetHeaderParam(\"User-Agent\", utils.UserAgentGen()); err != nil {\n\t\treturn err\n\t}\n\n\tif err := r.SetHeaderParam(\"X-Amzn-Trace-Id\", utils.AmazonTraceIDGen()); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\n\treturn nil\n}", "func (o *BikePointGetAllParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *DecryptParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif err := r.SetBodyParam(o.Parameters); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *DeleteRequestsRequestNameParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// query param api-version\n\tqrAPIVersion := o.APIVersion\n\tqAPIVersion := qrAPIVersion\n\tif qAPIVersion != \"\" {\n\t\tif err := r.SetQueryParam(\"api-version\", qAPIVersion); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param organization\n\tif err := r.SetPathParam(\"organization\", o.Organization); err != nil {\n\t\treturn err\n\t}\n\n\t// query param requestName\n\tqrRequestName := o.RequestName\n\tqRequestName := qrRequestName\n\tif qRequestName != \"\" {\n\t\tif err := r.SetQueryParam(\"requestName\", qRequestName); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.Synchronous != nil {\n\n\t\t// query param synchronous\n\t\tvar qrSynchronous bool\n\t\tif o.Synchronous != nil {\n\t\t\tqrSynchronous = *o.Synchronous\n\t\t}\n\t\tqSynchronous := swag.FormatBool(qrSynchronous)\n\t\tif qSynchronous != \"\" {\n\t\t\tif err := r.SetQueryParam(\"synchronous\", qSynchronous); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *SearchAbsoluteParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Decorate != nil {\n\n\t\t// query param decorate\n\t\tvar qrDecorate bool\n\t\tif o.Decorate != nil {\n\t\t\tqrDecorate = *o.Decorate\n\t\t}\n\t\tqDecorate := swag.FormatBool(qrDecorate)\n\t\tif qDecorate != \"\" {\n\t\t\tif err := r.SetQueryParam(\"decorate\", qDecorate); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Fields != nil {\n\n\t\t// query param fields\n\t\tvar qrFields string\n\t\tif o.Fields != nil {\n\t\t\tqrFields = *o.Fields\n\t\t}\n\t\tqFields := qrFields\n\t\tif qFields != \"\" {\n\t\t\tif err := r.SetQueryParam(\"fields\", qFields); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Filter != nil {\n\n\t\t// query param filter\n\t\tvar qrFilter string\n\t\tif o.Filter != nil {\n\t\t\tqrFilter = *o.Filter\n\t\t}\n\t\tqFilter := qrFilter\n\t\tif qFilter != \"\" {\n\t\t\tif err := r.SetQueryParam(\"filter\", qFilter); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// query param from\n\tqrFrom := o.From\n\tqFrom := qrFrom\n\tif qFrom != \"\" {\n\t\tif err := r.SetQueryParam(\"from\", qFrom); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.Limit != nil {\n\n\t\t// query param limit\n\t\tvar qrLimit int64\n\t\tif o.Limit != nil {\n\t\t\tqrLimit = *o.Limit\n\t\t}\n\t\tqLimit := swag.FormatInt64(qrLimit)\n\t\tif qLimit != \"\" {\n\t\t\tif err := r.SetQueryParam(\"limit\", qLimit); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Offset != nil {\n\n\t\t// query param offset\n\t\tvar qrOffset int64\n\t\tif o.Offset != nil {\n\t\t\tqrOffset = *o.Offset\n\t\t}\n\t\tqOffset := swag.FormatInt64(qrOffset)\n\t\tif qOffset != \"\" {\n\t\t\tif err := r.SetQueryParam(\"offset\", qOffset); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// query param query\n\tqrQuery := o.Query\n\tqQuery := qrQuery\n\tif qQuery != \"\" {\n\t\tif err := r.SetQueryParam(\"query\", qQuery); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif o.Sort != nil {\n\n\t\t// query param sort\n\t\tvar qrSort string\n\t\tif o.Sort != nil {\n\t\t\tqrSort = *o.Sort\n\t\t}\n\t\tqSort := qrSort\n\t\tif qSort != \"\" {\n\t\t\tif err := r.SetQueryParam(\"sort\", qSort); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// query param to\n\tqrTo := o.To\n\tqTo := qrTo\n\tif qTo != \"\" {\n\t\tif err := r.SetQueryParam(\"to\", qTo); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GetCountersParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.ClusterNodeID != nil {\n\n\t\t// query param clusterNodeId\n\t\tvar qrClusterNodeID string\n\n\t\tif o.ClusterNodeID != nil {\n\t\t\tqrClusterNodeID = *o.ClusterNodeID\n\t\t}\n\t\tqClusterNodeID := qrClusterNodeID\n\t\tif qClusterNodeID != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"clusterNodeId\", qClusterNodeID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Nodewise != nil {\n\n\t\t// query param nodewise\n\t\tvar qrNodewise bool\n\n\t\tif o.Nodewise != nil {\n\t\t\tqrNodewise = *o.Nodewise\n\t\t}\n\t\tqNodewise := swag.FormatBool(qrNodewise)\n\t\tif qNodewise != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"nodewise\", qNodewise); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *MetroclusterInterconnectGetParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param adapter\n\tif err := r.SetPathParam(\"adapter\", o.Adapter); err != nil {\n\t\treturn err\n\t}\n\n\tif o.Fields != nil {\n\n\t\t// binding items for fields\n\t\tjoinedFields := o.bindParamFields(reg)\n\n\t\t// query array param fields\n\t\tif err := r.SetQueryParam(\"fields\", joinedFields...); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param node.uuid\n\tif err := r.SetPathParam(\"node.uuid\", o.NodeUUID); err != nil {\n\t\treturn err\n\t}\n\n\t// path param partner_type\n\tif err := r.SetPathParam(\"partner_type\", o.PartnerType); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *PutParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.Item != nil {\n\t\tif err := r.SetBodyParam(o.Item); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// path param itemId\n\tif err := r.SetPathParam(\"itemId\", o.ItemID); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *CreateAccessPolicyParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\tif o.Body != nil {\n\t\tif err := r.SetBodyParam(o.Body); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *DeleteDataSourceParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.APIKey != nil {\n\n\t\t// query param ApiKey\n\t\tvar qrAPIKey string\n\n\t\tif o.APIKey != nil {\n\t\t\tqrAPIKey = *o.APIKey\n\t\t}\n\t\tqAPIKey := qrAPIKey\n\t\tif qAPIKey != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"ApiKey\", qAPIKey); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.APISecret != nil {\n\n\t\t// query param ApiSecret\n\t\tvar qrAPISecret string\n\n\t\tif o.APISecret != nil {\n\t\t\tqrAPISecret = *o.APISecret\n\t\t}\n\t\tqAPISecret := qrAPISecret\n\t\tif qAPISecret != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"ApiSecret\", qAPISecret); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.CreationDate != nil {\n\n\t\t// query param CreationDate\n\t\tvar qrCreationDate int32\n\n\t\tif o.CreationDate != nil {\n\t\t\tqrCreationDate = *o.CreationDate\n\t\t}\n\t\tqCreationDate := swag.FormatInt32(qrCreationDate)\n\t\tif qCreationDate != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"CreationDate\", qCreationDate); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Disabled != nil {\n\n\t\t// query param Disabled\n\t\tvar qrDisabled bool\n\n\t\tif o.Disabled != nil {\n\t\t\tqrDisabled = *o.Disabled\n\t\t}\n\t\tqDisabled := swag.FormatBool(qrDisabled)\n\t\tif qDisabled != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"Disabled\", qDisabled); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.EncryptionKey != nil {\n\n\t\t// query param EncryptionKey\n\t\tvar qrEncryptionKey string\n\n\t\tif o.EncryptionKey != nil {\n\t\t\tqrEncryptionKey = *o.EncryptionKey\n\t\t}\n\t\tqEncryptionKey := qrEncryptionKey\n\t\tif qEncryptionKey != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"EncryptionKey\", qEncryptionKey); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.EncryptionMode != nil {\n\n\t\t// query param EncryptionMode\n\t\tvar qrEncryptionMode string\n\n\t\tif o.EncryptionMode != nil {\n\t\t\tqrEncryptionMode = *o.EncryptionMode\n\t\t}\n\t\tqEncryptionMode := qrEncryptionMode\n\t\tif qEncryptionMode != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"EncryptionMode\", qEncryptionMode); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.FlatStorage != nil {\n\n\t\t// query param FlatStorage\n\t\tvar qrFlatStorage bool\n\n\t\tif o.FlatStorage != nil {\n\t\t\tqrFlatStorage = *o.FlatStorage\n\t\t}\n\t\tqFlatStorage := swag.FormatBool(qrFlatStorage)\n\t\tif qFlatStorage != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"FlatStorage\", qFlatStorage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.LastSynchronizationDate != nil {\n\n\t\t// query param LastSynchronizationDate\n\t\tvar qrLastSynchronizationDate int32\n\n\t\tif o.LastSynchronizationDate != nil {\n\t\t\tqrLastSynchronizationDate = *o.LastSynchronizationDate\n\t\t}\n\t\tqLastSynchronizationDate := swag.FormatInt32(qrLastSynchronizationDate)\n\t\tif qLastSynchronizationDate != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"LastSynchronizationDate\", qLastSynchronizationDate); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\t// path param Name\n\tif err := r.SetPathParam(\"Name\", o.Name); err != nil {\n\t\treturn err\n\t}\n\n\tif o.ObjectsBaseFolder != nil {\n\n\t\t// query param ObjectsBaseFolder\n\t\tvar qrObjectsBaseFolder string\n\n\t\tif o.ObjectsBaseFolder != nil {\n\t\t\tqrObjectsBaseFolder = *o.ObjectsBaseFolder\n\t\t}\n\t\tqObjectsBaseFolder := qrObjectsBaseFolder\n\t\tif qObjectsBaseFolder != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"ObjectsBaseFolder\", qObjectsBaseFolder); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.ObjectsBucket != nil {\n\n\t\t// query param ObjectsBucket\n\t\tvar qrObjectsBucket string\n\n\t\tif o.ObjectsBucket != nil {\n\t\t\tqrObjectsBucket = *o.ObjectsBucket\n\t\t}\n\t\tqObjectsBucket := qrObjectsBucket\n\t\tif qObjectsBucket != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"ObjectsBucket\", qObjectsBucket); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.ObjectsHost != nil {\n\n\t\t// query param ObjectsHost\n\t\tvar qrObjectsHost string\n\n\t\tif o.ObjectsHost != nil {\n\t\t\tqrObjectsHost = *o.ObjectsHost\n\t\t}\n\t\tqObjectsHost := qrObjectsHost\n\t\tif qObjectsHost != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"ObjectsHost\", qObjectsHost); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.ObjectsPort != nil {\n\n\t\t// query param ObjectsPort\n\t\tvar qrObjectsPort int32\n\n\t\tif o.ObjectsPort != nil {\n\t\t\tqrObjectsPort = *o.ObjectsPort\n\t\t}\n\t\tqObjectsPort := swag.FormatInt32(qrObjectsPort)\n\t\tif qObjectsPort != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"ObjectsPort\", qObjectsPort); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.ObjectsSecure != nil {\n\n\t\t// query param ObjectsSecure\n\t\tvar qrObjectsSecure bool\n\n\t\tif o.ObjectsSecure != nil {\n\t\t\tqrObjectsSecure = *o.ObjectsSecure\n\t\t}\n\t\tqObjectsSecure := swag.FormatBool(qrObjectsSecure)\n\t\tif qObjectsSecure != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"ObjectsSecure\", qObjectsSecure); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.ObjectsServiceName != nil {\n\n\t\t// query param ObjectsServiceName\n\t\tvar qrObjectsServiceName string\n\n\t\tif o.ObjectsServiceName != nil {\n\t\t\tqrObjectsServiceName = *o.ObjectsServiceName\n\t\t}\n\t\tqObjectsServiceName := qrObjectsServiceName\n\t\tif qObjectsServiceName != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"ObjectsServiceName\", qObjectsServiceName); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.PeerAddress != nil {\n\n\t\t// query param PeerAddress\n\t\tvar qrPeerAddress string\n\n\t\tif o.PeerAddress != nil {\n\t\t\tqrPeerAddress = *o.PeerAddress\n\t\t}\n\t\tqPeerAddress := qrPeerAddress\n\t\tif qPeerAddress != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"PeerAddress\", qPeerAddress); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.SkipSyncOnRestart != nil {\n\n\t\t// query param SkipSyncOnRestart\n\t\tvar qrSkipSyncOnRestart bool\n\n\t\tif o.SkipSyncOnRestart != nil {\n\t\t\tqrSkipSyncOnRestart = *o.SkipSyncOnRestart\n\t\t}\n\t\tqSkipSyncOnRestart := swag.FormatBool(qrSkipSyncOnRestart)\n\t\tif qSkipSyncOnRestart != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"SkipSyncOnRestart\", qSkipSyncOnRestart); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.StorageType != nil {\n\n\t\t// query param StorageType\n\t\tvar qrStorageType string\n\n\t\tif o.StorageType != nil {\n\t\t\tqrStorageType = *o.StorageType\n\t\t}\n\t\tqStorageType := qrStorageType\n\t\tif qStorageType != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"StorageType\", qStorageType); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.VersioningPolicyName != nil {\n\n\t\t// query param VersioningPolicyName\n\t\tvar qrVersioningPolicyName string\n\n\t\tif o.VersioningPolicyName != nil {\n\t\t\tqrVersioningPolicyName = *o.VersioningPolicyName\n\t\t}\n\t\tqVersioningPolicyName := qrVersioningPolicyName\n\t\tif qVersioningPolicyName != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"VersioningPolicyName\", qVersioningPolicyName); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif o.Watch != nil {\n\n\t\t// query param Watch\n\t\tvar qrWatch bool\n\n\t\tif o.Watch != nil {\n\t\t\tqrWatch = *o.Watch\n\t\t}\n\t\tqWatch := swag.FormatBool(qrWatch)\n\t\tif qWatch != \"\" {\n\n\t\t\tif err := r.SetQueryParam(\"Watch\", qWatch); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}" ]
[ "0.7198161", "0.714435", "0.70471495", "0.7021836", "0.69967365", "0.69959503", "0.6979433", "0.6979074", "0.69695425", "0.6966308", "0.69242847", "0.6908102", "0.69045216", "0.6871055", "0.68575305", "0.68564737", "0.6851862", "0.6845359", "0.6844677", "0.684266", "0.68300045", "0.68283993", "0.68213093", "0.68209827", "0.681858", "0.6810088", "0.67938024", "0.6792597", "0.67860335", "0.6781293", "0.6778168", "0.67739195", "0.676121", "0.676101", "0.6760405", "0.675646", "0.67500865", "0.67439634", "0.6743771", "0.6742206", "0.67405975", "0.67344606", "0.67331755", "0.67328155", "0.67320985", "0.67255586", "0.6724229", "0.67159885", "0.67127234", "0.67094815", "0.67085487", "0.6705413", "0.67020816", "0.6698303", "0.66938454", "0.6692077", "0.66849154", "0.6677396", "0.66709644", "0.6670931", "0.6670394", "0.6666765", "0.6666114", "0.6665216", "0.6662671", "0.66568", "0.6653157", "0.6646967", "0.6645966", "0.6642767", "0.6640608", "0.6638263", "0.6634605", "0.66345316", "0.6625767", "0.66217625", "0.6619463", "0.66181296", "0.66144806", "0.6606646", "0.6605777", "0.66039085", "0.659942", "0.6598786", "0.65961313", "0.65951705", "0.6591678", "0.6586235", "0.65826946", "0.658246", "0.6574939", "0.6572858", "0.6569902", "0.6568091", "0.65670824", "0.65661305", "0.6565009", "0.6561903", "0.65615994", "0.6560924", "0.6557316" ]
0.0
-1
To reduce request to GKE metadata server, the base token source is reused across syncers. Note: Initialization is deferred because there are possible to use serviceAccountSecretRef with no available default token source.
func initializedBaseTokenSource() (oauth2.TokenSource, error) { baseTokenSourceOnce.Do(func() { baseTokenSource, baseTokenSourceErr = google.DefaultTokenSource(context.Background(), cloudPlatformScope) }) return baseTokenSource, baseTokenSourceErr }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func tokenSource(ctx context.Context) (oauth2.TokenSource, error) {\n\tok, err := credsFile.Exists()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar bootstrapToken *oauth2.Token\n\tif !ok {\n\t\ttok, err := authenticate()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tbootstrapToken = tok\n\t}\n\treturn newCachedTokenFile(ctx, bootstrapToken, credsFile.Path())\n}", "func initializeTokenRequester(centralCfg config.CentralConfig) {\n\tagent.tokenRequester = apicauth.NewPlatformTokenGetter(\n\t\tcentralCfg.GetAuthConfig().GetPrivateKey(),\n\t\tcentralCfg.GetAuthConfig().GetPublicKey(),\n\t\tcentralCfg.GetAuthConfig().GetKeyPassword(),\n\t\tcentralCfg.GetAuthConfig().GetTokenURL(),\n\t\tcentralCfg.GetAuthConfig().GetAudience(),\n\t\tcentralCfg.GetAuthConfig().GetClientID(),\n\t\tcentralCfg.GetAuthConfig().GetTimeout(),\n\t)\n}", "func newTokenSource() *tokenReplacer {\n\t// nil token will cause a refresh\n\ttok, _ := readToken()\n\treturn &tokenReplacer{tok, oauthConfig.TokenSource(context.Background(), tok), &tokenPrompt{}}\n}", "func (c *Config) tokenSource(ctx context.Context, scheme string) (oauth2.TokenSource, error) {\n\tif c.WorkforcePoolUserProject != \"\" {\n\t\tvalid := validateWorkforceAudience(c.Audience)\n\t\tif !valid {\n\t\t\treturn nil, fmt.Errorf(\"oauth2/google: workforce_pool_user_project should not be set for non-workforce pool credentials\")\n\t\t}\n\t}\n\n\tts := tokenSource{\n\t\tctx: ctx,\n\t\tconf: c,\n\t}\n\tif c.ServiceAccountImpersonationURL == \"\" {\n\t\treturn oauth2.ReuseTokenSource(nil, ts), nil\n\t}\n\tscopes := c.Scopes\n\tts.conf.Scopes = []string{\"https://www.googleapis.com/auth/cloud-platform\"}\n\timp := ImpersonateTokenSource{\n\t\tCtx: ctx,\n\t\tURL: c.ServiceAccountImpersonationURL,\n\t\tScopes: scopes,\n\t\tTs: oauth2.ReuseTokenSource(nil, ts),\n\t\tTokenLifetimeSeconds: c.ServiceAccountImpersonationLifetimeSeconds,\n\t}\n\treturn oauth2.ReuseTokenSource(nil, imp), nil\n}", "func tokenSource(ctx context.Context, accessToken, credentials string) (oauth2.TokenSource, error) {\n\t// Try access token first\n\tif accessToken != \"\" {\n\t\tlog.Printf(\"[INFO] authenticating via access_token\")\n\n\t\tcontents, _, err := pathorcontents.Read(accessToken)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to load access token: %w\", err)\n\t\t}\n\n\t\treturn oauth2.StaticTokenSource(&oauth2.Token{\n\t\t\tAccessToken: contents,\n\t\t}), nil\n\t}\n\n\t// Then credentials\n\tif credentials != \"\" {\n\t\tlog.Printf(\"[INFO] authenticating via credentials\")\n\n\t\tcontents, _, err := pathorcontents.Read(credentials)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to load credentials: %w\", err)\n\t\t}\n\n\t\tcreds, err := google.CredentialsFromJSON(ctx, []byte(contents), cloudPlatformScope)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to parse credentials: %w\", err)\n\t\t}\n\n\t\treturn creds.TokenSource, nil\n\t}\n\n\t// Fallback to default credentials\n\tlog.Printf(\"[INFO] authenticating via default credentials\")\n\tsource, err := google.DefaultTokenSource(ctx, cloudPlatformScope)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get default credentials: %w\", err)\n\t}\n\treturn source, nil\n}", "func (w *GCPAuthWrapper) SetTokenSource(permissionCode string) error {\n\tvar err error\n\n\tctx := context.Background()\n\n\tw.OauthToken, err = w.Config.Exchange(ctx, permissionCode)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *Config) TokenSource(ctx context.Context, t *Token) TokenSource {\n\ttkr := &tokenRefresher{\n\t\tctx: ctx,\n\t\tconf: c,\n\t}\n\tif t != nil {\n\t\ttkr.refreshToken = t.RefreshToken\n\t}\n\treturn &reuseTokenSource{\n\t\tt: t,\n\t\tnew: tkr,\n\t}\n}", "func Init(ctx context.Context, local bool) (oauth2.TokenSource, error) {\n\treturn auth_steps.Init(ctx, local, auth.ScopeGerrit, auth.ScopeUserinfoEmail)\n}", "func (b *BungieConfig) TokenSource(t *oauth2.Token) oauth2.TokenSource {\n\treturn newTokenSource(t, b.cfg.ClientID)\n}", "func StorageTokenSource(ctx context.Context, c *Config, t *oauth2.Token) oauth2.TokenSource {\n\tif t == nil || !t.Valid() {\n\t\tif tok, err := c.Storage.GetToken(); err == nil {\n\t\t\tt = tok\n\t\t}\n\t}\n\tts := c.Config.TokenSource(ctx, t)\n\treturn &storageTokenSource{c, ts}\n}", "func (c *Credentials) TokenSource(ctx context.Context) (oauth2.TokenSource, error) {\n\tswitch c.Type {\n\tcase CredentialsTypeADC:\n\t\treturn initializedBaseTokenSource()\n\tcase CredentialsTypeServiceAccountJSON:\n\t\tcred, err := google.CredentialsFromJSON(ctx, c.ServiceAccountJSON, cloudPlatformScope)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn cred.TokenSource, nil\n\tcase CredentialsTypeImpersonation:\n\t\tbaseTS, err := initializedBaseTokenSource()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tts, err := impersonate.CredentialsTokenSource(ctx, impersonate.CredentialsConfig{\n\t\t\tTargetPrincipal: c.ImpersonateConfig.TargetServiceAccount,\n\t\t\tDelegates: c.ImpersonateConfig.Delegates,\n\t\t\tScopes: []string{cloudPlatformScope},\n\t\t},\n\t\t\toption.WithTokenSource(baseTS),\n\t\t)\n\t\treturn ts, err\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"credentials type unknown: %v\", c.Type)\n\t}\n}", "func (ts tokenSource) Token() (*oauth2.Token, error) {\n\tconf := ts.conf\n\n\tcredSource, err := conf.parse(ts.ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsubjectToken, err := credSource.subjectToken()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tstsRequest := stsTokenExchangeRequest{\n\t\tGrantType: \"urn:ietf:params:oauth:grant-type:token-exchange\",\n\t\tAudience: conf.Audience,\n\t\tScope: conf.Scopes,\n\t\tRequestedTokenType: \"urn:ietf:params:oauth:token-type:access_token\",\n\t\tSubjectToken: subjectToken,\n\t\tSubjectTokenType: conf.SubjectTokenType,\n\t}\n\theader := make(http.Header)\n\theader.Add(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\tclientAuth := clientAuthentication{\n\t\tAuthStyle: oauth2.AuthStyleInHeader,\n\t\tClientID: conf.ClientID,\n\t\tClientSecret: conf.ClientSecret,\n\t}\n\tvar options map[string]interface{}\n\t// Do not pass workforce_pool_user_project when client authentication is used.\n\t// The client ID is sufficient for determining the user project.\n\tif conf.WorkforcePoolUserProject != \"\" && conf.ClientID == \"\" {\n\t\toptions = map[string]interface{}{\n\t\t\t\"userProject\": conf.WorkforcePoolUserProject,\n\t\t}\n\t}\n\tstsResp, err := exchangeToken(ts.ctx, conf.TokenURL, &stsRequest, clientAuth, header, options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\taccessToken := &oauth2.Token{\n\t\tAccessToken: stsResp.AccessToken,\n\t\tTokenType: stsResp.TokenType,\n\t}\n\tif stsResp.ExpiresIn < 0 {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: got invalid expiry from security token service\")\n\t} else if stsResp.ExpiresIn >= 0 {\n\t\taccessToken.Expiry = now().Add(time.Duration(stsResp.ExpiresIn) * time.Second)\n\t}\n\n\tif stsResp.RefreshToken != \"\" {\n\t\taccessToken.RefreshToken = stsResp.RefreshToken\n\t}\n\treturn accessToken, nil\n}", "func attemptInit(ctx context.Context, account string, scopes []string, audience string) (TokenProvider, error) {\n\t// This mutex is used to avoid hitting GKE metadata server concurrently if\n\t// we have a stampede of goroutines. It doesn't actually protect any shared\n\t// state in the current process.\n\tglobalGCELock.Lock()\n\tdefer globalGCELock.Unlock()\n\n\tif account == \"\" {\n\t\taccount = \"default\"\n\t}\n\n\t// Grab an email associated with the account. This must not be failing on\n\t// a healthy VM if the account is present. If it does, the metadata server is\n\t// broken.\n\temail, err := metadataClient.Email(account)\n\tif err != nil {\n\t\t// Note: we purposefully delay this check only after the first call to\n\t\t// the metadata fails because metadata.OnGCE was observed to often report\n\t\t// \"false\" when running on GKE due to gke-metadata-server being slow. Our\n\t\t// metadataClient has (much) higher timeouts that the client used by\n\t\t// metadata.OnGCE, and it handles slow gke-metadata-server better. So if we\n\t\t// end up here and metadata.OnGCE also says \"false\", then we are not on GCE\n\t\t// with high probability. The downside is that it may take up to 15 sec to\n\t\t// detect this (or whatever ResponseHeaderTimeout in metadataClient is).\n\t\tif !metadata.OnGCE() {\n\t\t\treturn nil, ErrBadCredentials\n\t\t}\n\t\tif _, yep := err.(metadata.NotDefinedError); yep {\n\t\t\treturn nil, ErrInsufficientAccess\n\t\t}\n\t\treturn nil, transient.Tag.Apply(err)\n\t}\n\n\t// Ensure the account has requested scopes. Assume 'cloud-platform' scope\n\t// covers all possible scopes. This is important when using GKE Workload\n\t// Identities: the metadata server always reports only 'cloud-platform' scope\n\t// there. Its presence should be enough to cover all scopes used in practice.\n\t// The exception is non-cloud scopes (like gerritcodereview or G Suite). To\n\t// use such scopes, one will have to use impersonation through Cloud IAM APIs,\n\t// which *are* covered by cloud-platform (see ActAsServiceAccount in auth.go).\n\tif audience == \"\" {\n\t\tavailableScopes, err := metadataClient.Scopes(account)\n\t\tif err != nil {\n\t\t\treturn nil, transient.Tag.Apply(err)\n\t\t}\n\t\tavailableSet := stringset.NewFromSlice(availableScopes...)\n\t\tif !availableSet.Has(\"https://www.googleapis.com/auth/cloud-platform\") {\n\t\t\tfor _, requested := range scopes {\n\t\t\t\tif !availableSet.Has(requested) {\n\t\t\t\t\tlogging.Warningf(ctx, \"GCE service account %q doesn't have required scope %q (all scopes: %q)\", account, requested, availableScopes)\n\t\t\t\t\treturn nil, ErrInsufficientAccess\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn &gceTokenProvider{\n\t\taccount: account,\n\t\temail: email,\n\t\taudience: audience,\n\t\tcacheKey: CacheKey{\n\t\t\tKey: fmt.Sprintf(\"gce/%s\", account),\n\t\t\tScopes: scopes,\n\t\t},\n\t}, nil\n}", "func (c *Config) TokenSource() oauth2.TokenSource {\n\tpk, err := jwt.ParseRSAPrivateKeyFromPEM(c.PrivateKey)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn oauth2.ReuseTokenSource(nil, appSource{\n\t\tappID: c.AppID,\n\t\texpire: c.Expire,\n\t\tpk: pk,\n\t})\n}", "func newTokenInjector(meta *metadata.Client, acc string) func(*http.Request) error {\n\tif acc == \"\" {\n\t\tacc = \"default\"\n\t}\n\tacc = url.PathEscape(acc)\n\treturn func(req *http.Request) error {\n\t\taud := fmt.Sprintf(\"%s://%s\", req.URL.Scheme, req.URL.Host)\n\t\taud = url.QueryEscape(aud)\n\t\t// TODO(smut): Cache the token and reuse if not yet expired.\n\t\t// Currently the only user of this package only makes one\n\t\t// request per boot so caching isn't too important yet.\n\t\ttok, err := meta.Get(fmt.Sprintf(tokMetadata, acc, aud))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treq.Header.Set(vmtoken.Header, tok)\n\t\treturn nil\n\t}\n}", "func appEngineTokenSource(ctx context.Context, scope ...string) oauth2.TokenSource {\n\tlogOnce.Do(func() {\n\t\tlog.Print(\"google: AppEngineTokenSource is deprecated on App Engine standard second generation runtimes (>= Go 1.11) and App Engine flexible. Please use DefaultTokenSource or ComputeTokenSource.\")\n\t})\n\treturn ComputeTokenSource(\"\")\n}", "func OIDCFederatedTokenSource(tokenConfig *OIDCFederatedTokenConfig) (oauth2.TokenSource, error) {\n\n\tif &tokenConfig.SourceTokenSource == nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: Source OIDC Token cannot be nil\")\n\t}\n\n\tif tokenConfig.Scope == \"\" {\n\t\ttokenConfig.Scope = GCP_OIDC_CLOUD_PLATFORM_SCOPE\n\t}\n\treturn &oidcFederatedTokenSource{\n\t\trefreshMutex: &sync.Mutex{},\n\t\trootSource: tokenConfig.SourceTokenSource,\n\t\tscope: tokenConfig.Scope,\n\t\ttargetResource: tokenConfig.TargetResource,\n\t\ttargetServiceAccount: tokenConfig.TargetServiceAccount,\n\t\tuseIAMToken: tokenConfig.UseIAMToken,\n\t}, nil\n}", "func (c *Config) TokenSource(ctx context.Context) (oauth2.TokenSource, error) {\n\treturn c.tokenSource(ctx, \"https\")\n}", "func Token(c *gin.Context) {\n\n\tclientStorage, err := models.NewClientStorage(ginject.Deps(c))\n\tif err != nil {\n\t\tpanic(err)\n\t\treturn\n\t}\n\n\tmanager := models.NewManager(ginject.Deps(c))\n\n\tserver.InitServer(manager)\n\tserver.SetAllowedGrantType(oauth2.AuthorizationCode, oauth2.Refreshing)\n\tserver.SetAllowedResponseType(oauth2.Token)\n\n\tserver.SetClientInfoHandler(func(r *http.Request) (clientID, clientSecret string, err error) {\n\n\t\tcid := r.FormValue(\"client_id\")\n\n\t\tif cid == \"\" {\n\n\t\t}\n\n\t\tinfo, err := clientStorage.GetByID(cid)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\n\t\tclientID = info.GetID()\n\t\tclientSecret = info.GetSecret()\n\t\treturn\n\t})\n\n\tserver.HandleTokenRequest(c)\n}", "func New(name string, configPath string, userName string) (tokenauth.Source, error) {\n\tif configPath == \"\" {\n\t\tconfigPath = k8s.DefaultKubeConfigPath\n\t}\n\tk8sConfig, err := cfg.LoadFromFile(configPath)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"Failed to load k8s config from file %v. Make sure it is there or change\"+\n\t\t\t\" permissions.\", configPath)\n\t}\n\n\tinfo, ok := k8sConfig.AuthInfos[userName]\n\tif !ok {\n\t\treturn nil, errors.Errorf(\"Failed to find user %s inside k8s config AuthInfo from file %v\", userName, configPath)\n\t}\n\n\t// Currently supported:\n\t// - token\n\t// - OIDC\n\t// - Google compute platform via Oauth2\n\tif info.AuthProvider != nil {\n\t\tswitch info.AuthProvider.Name {\n\t\tcase \"oidc\":\n\t\t\tcache, err := k8s.NewCacheFromUser(configPath, userName)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrap(err, \"Failed to get OIDC configuration from user. \")\n\t\t\t}\n\t\t\ts, _, err := oidcauth.NewWithCache(name, cache, nil)\n\t\t\treturn s, err\n\t\tcase \"gcp\":\n\t\t\tc, err := oauth2auth.NewConfigFromMap(info.AuthProvider.Config)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrap(err, \"Failed to create OAuth2 config from map.\")\n\t\t\t}\n\t\t\treturn oauth2auth.NewGCP(name, userName, configPath, c)\n\t\tdefault:\n\t\t\t// TODO(bplotka): Add support for more of them if needed.\n\t\t\treturn nil, errors.Errorf(\"Not supported k8s Auth provider %v\", info.AuthProvider.Name)\n\t\t}\n\t}\n\n\tif info.Token != \"\" {\n\t\treturn directauth.New(name, info.Token), nil\n\t}\n\n\treturn nil, errors.Errorf(\"Not found supported auth source from k8s config %+v\", info)\n}", "func RecoverToken(currentTarget *types.Target) (*oauth2.Token, error) {\n\tvar storage []keyring.BackendType\n\tstorageConfig := viper.GetString(\"targets.\" + currentTarget.Label + \".token-storage\")\n\tstorage = append(storage, keyring.BackendType(storageConfig))\n\tring, err := keyring.Open(keyring.Config{\n\t\t// Configuration for keychain\n\t\tAllowedBackends: storage,\n\t\tServiceName: \"gsh\",\n\n\t\t// Configuration for encrypted file\n\t\tFileDir: \"~/.gsh/\" + currentTarget.Label,\n\t\tFilePasswordFunc: terminalPrompt,\n\n\t\t// Configuration for KWallet\n\t\tKWalletAppID: \"gsh\",\n\t\tKWalletFolder: currentTarget.Label,\n\n\t\t// Configuration for pass (https://www.passwordstore.org/)\n\t\tPassDir: \"~/.gsh/\" + currentTarget.Label,\n\n\t\t// Configuration for Secret Service (https://secretstorage.readthedocs.io/en/latest/)\n\t\tLibSecretCollectionName: \"gsh\",\n\t})\n\tif err != nil {\n\t\tfmt.Printf(\"Client error opening token-storage: (%s)\\n\", err.Error())\n\t\treturn nil, err\n\t}\n\n\ttokenKeyItem, err := ring.Get(currentTarget.Label)\n\tif err != nil {\n\t\tfmt.Printf(\"Client error reading token storage: (%s)\\n\", err.Error())\n\t\treturn nil, err\n\t}\n\n\ttoken := new(oauth2.Token)\n\tif err := json.Unmarshal(tokenKeyItem.Data, &token); err != nil {\n\t\tfmt.Printf(\"Client error unmarshalling stored token: (%s)\\n\", err.Error())\n\t\treturn nil, err\n\t}\n\n\t// Setting custom HTTP client with timeouts\n\tvar netTransport = &http.Transport{\n\t\tDial: (&net.Dialer{\n\t\t\tTimeout: 10 * time.Second,\n\t\t}).Dial,\n\t\tTLSHandshakeTimeout: 10 * time.Second,\n\t}\n\tvar netClient = &http.Client{\n\t\tTimeout: 10 * time.Second,\n\t\tTransport: netTransport,\n\t}\n\n\t// Making discovery GSH request\n\tresp, err := netClient.Get(currentTarget.Endpoint + \"/status/config\")\n\tif err != nil {\n\t\tfmt.Printf(\"GSH API is down: %s (%s)\\n\", currentTarget.Endpoint, err.Error())\n\t\tos.Exit(1)\n\t}\n\tdefer resp.Body.Close()\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Printf(\"GSH API body response error: %s\\n\", err.Error())\n\t\tos.Exit(1)\n\t}\n\tif resp.StatusCode != http.StatusOK {\n\t\tfmt.Printf(\"GSH API status response error: %v\\n\", resp.StatusCode)\n\t\tos.Exit(1)\n\t}\n\ttype ConfigResponse struct {\n\t\tBaseURL string `json:\"oidc_base_url\"`\n\t\tRealm string `json:\"oidc_realm\"`\n\t\tAudience string `json:\"oidc_audience\"`\n\t\tIssuer string `json:\"oidc_issuer\"`\n\t\tCerts string `json:\"oidc_certs\"`\n\t}\n\tconfigResponse := new(ConfigResponse)\n\tif err := json.Unmarshal(body, &configResponse); err != nil {\n\t\tfmt.Printf(\"GSH API body unmarshal error: %s\\n\", err.Error())\n\t\tos.Exit(1)\n\t}\n\n\tctx := context.Background()\n\toauth2provider, err := oidc.NewProvider(ctx, configResponse.Issuer)\n\tif err != nil {\n\t\tfmt.Printf(\"GSH client setting OIDC provider error: %s\\n\", err.Error())\n\t\tos.Exit(1)\n\t}\n\n\t// Configure an OpenID Connect aware OAuth2 client.\n\toauth2config := &oauth2.Config{\n\t\tClientID: configResponse.Audience,\n\t\tEndpoint: oauth2provider.Endpoint(),\n\t}\n\ttokenRefreshed, err := oauth2config.TokenSource(ctx, token).Token()\n\tif err != nil {\n\t\tfmt.Printf(\"GSH client renew token error: %s\\n\", err.Error())\n\t\tos.Exit(1)\n\t}\n\n\treturn tokenRefreshed, nil\n}", "func init() {\n\toauthConfig, err := adal.NewOAuthConfig(azure.PublicCloud.ActiveDirectoryEndpoint, config.TenantID)\n\tif err != nil {\n\t\tlog.Fatalf(\"Failed to get OAuth config: %v\\n\", err)\n\t}\n\ttoken, err = adal.NewServicePrincipalToken(\n\t\t*oauthConfig,\n\t\tconfig.ServicePrincipalID,\n\t\tconfig.ServicePrincipalSecret,\n\t\tazure.PublicCloud.ResourceManagerEndpoint)\n\tif err != nil {\n\t\tlog.Fatalf(\"faled to get token: %v\\n\", err)\n\t}\n}", "func CredentialsTokenSource(creds *google.Credentials) TokenSource {\n\tif creds == nil {\n\t\treturn nil\n\t}\n\treturn TokenSource(creds.TokenSource)\n}", "func Init(paramClientID, paramClientSecret, paramAuth0URL, paramAudience string) {\n\tf := logrus.Fields{\n\t\t\"functionName\": \"token.Init\",\n\t\t\"auth0URL\": paramAuth0URL,\n\t\t\"audience\": paramAudience,\n\t}\n\tlog.WithFields(f).Debug(\"token init running...\")\n\n\tclientID = paramClientID\n\tclientSecret = paramClientSecret\n\taudience = paramAudience\n\toauthTokenURL = paramAuth0URL\n\n\tif expiry.Year() == 1 {\n\t\texpiry = time.Now()\n\t}\n\n\tgo retrieveToken() //nolint\n}", "func (gsp *GoogleServiceProvider) TokenSource(c context.Context, scopes ...string) (oauth2.TokenSource, error) {\n\tcbts := contextBoundTokenSource{\n\t\tContext: c,\n\t\tcache: gsp.Cache,\n\t\tcacheKey: accessTokenKeyForScopes(scopes),\n\t\tmakeTokenSource: func(c context.Context) (oauth2.TokenSource, error) {\n\t\t\treturn google.DefaultTokenSource(c, scopes...)\n\t\t},\n\t}\n\treturn &cbts, nil\n}", "func appEngineTokenSource(ctx context.Context, scope ...string) oauth2.TokenSource {\n\tscopes := append([]string{}, scope...)\n\tsort.Strings(scopes)\n\treturn &gaeTokenSource{\n\t\tctx: ctx,\n\t\tscopes: scopes,\n\t\tkey: strings.Join(scopes, \" \"),\n\t}\n}", "func sourceFromDefault(ctx context.Context, targetAudience string, tokenURL string) (*oidcTokenSource, error) {\n\tcredentials, err := google.FindDefaultCredentials(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif len(credentials.JSON) == 0 {\n\t\treturn nil, ErrComputeEngineNotSupported\n\t}\n\tconfig, err := google.JWTConfigFromJSON(credentials.JSON)\n\tif err != nil {\n\t\t// friendly error message that we found user credentials\n\t\tif strings.Contains(err.Error(), \"authorized_user\") {\n\t\t\treturn nil, ErrUserCredentialsNotSupported\n\t\t}\n\t\treturn nil, err\n\t}\n\tprivateKey, err := parseKey(config.PrivateKey)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &oidcTokenSource{config.Email, config.PrivateKeyID, privateKey, targetAudience, tokenURL}, nil\n}", "func TokenSource(ctx context.Context, config *oauth2.Config, state string, authHandler AuthorizationHandler) oauth2.TokenSource {\n\treturn TokenSourceWithPKCE(ctx, config, state, authHandler, nil)\n}", "func getClient(config *oauth2.Config) *http.Client {\n // The file token.json stores the user's access and refresh tokens, and is\n // created automatically when the authorization flow completes for the first\n // time.\n tokFile := \"token.json\"\n tok, err := tokenFromFile(tokFile)\n if err != nil {\n tok = getTokenFromWeb(config)\n saveToken(tokFile, tok)\n }\n return config.Client(context.Background(), tok)\n}", "func (s *Store) Init(ctx context.Context, metadataRaw secretstores.Metadata) error {\n\tmetadata, err := s.parseSecretManagerMetadata(metadataRaw)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tclient, err := s.getClient(ctx, metadata)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to setup secretmanager client: %s\", err)\n\t}\n\n\ts.client = client\n\ts.ProjectID = metadata.ProjectID\n\n\treturn nil\n}", "func (c *contextBoundTokenSource) Token() (*oauth2.Token, error) {\n\tnow := clock.Now(c)\n\n\t// Get the current token value. We do this without locking around the token\n\t// element.\n\tif tokIface, ok := c.cache.Get(c, c.cacheKey); ok {\n\t\ttok := tokIface.(*oauth2.Token)\n\t\tif !c.closeToExpRandomized(now, tok.Expiry, accessTokenExpirationRandomization) {\n\t\t\treturn tok, nil\n\t\t}\n\t}\n\n\t// Either the token is expired, or we are selected randomly as a refresh case.\n\t// Get a new TokenSource to refresh the token with.\n\tts, err := c.makeTokenSource(c)\n\tif err != nil {\n\t\treturn nil, errors.Annotate(err, \"failed to create new TokenSource\").Err()\n\t}\n\n\t// While refreshing, we lock around the cache key via GetOrCreate in case\n\t// multiple requests are either selected or have expired.\n\ttokIface, err := c.cache.GetOrCreate(c, c.cacheKey, func() (interface{}, time.Duration, error) {\n\t\ttok, err := ts.Token()\n\t\tif err != nil {\n\t\t\treturn nil, 0, err\n\t\t}\n\n\t\texpiryDelta := tok.Expiry.Sub(now)\n\t\tswitch {\n\t\tcase expiryDelta <= 0:\n\t\t\treturn nil, 0, errors.Reason(\"retrieved expired access token (%s < %s)\", tok.Expiry, now).Err()\n\t\tcase expiryDelta > accessTokenMinExpiration:\n\t\t\t// Subtract some time from the token's expiry so we don't use it immediately\n\t\t\t// before it actually expires.\n\t\t\ttok.Expiry = tok.Expiry.Add(-accessTokenMinExpiration)\n\t\t\texpiryDelta -= accessTokenMinExpiration\n\t\t}\n\t\treturn tok, expiryDelta, nil\n\t})\n\tif err != nil {\n\t\treturn nil, errors.Annotate(err, \"failed to mint new access token\").Err()\n\t}\n\treturn tokIface.(*oauth2.Token), nil\n}", "func (its ImpersonateTokenSource) Token() (*oauth2.Token, error) {\n\tlifetimeString := \"3600s\"\n\tif its.TokenLifetimeSeconds != 0 {\n\t\tlifetimeString = fmt.Sprintf(\"%ds\", its.TokenLifetimeSeconds)\n\t}\n\treqBody := generateAccessTokenReq{\n\t\tLifetime: lifetimeString,\n\t\tScope: its.Scopes,\n\t\tDelegates: its.Delegates,\n\t}\n\tb, err := json.Marshal(reqBody)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: unable to marshal request: %v\", err)\n\t}\n\tclient := oauth2.NewClient(its.Ctx, its.Ts)\n\treq, err := http.NewRequest(\"POST\", its.URL, bytes.NewReader(b))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: unable to create impersonation request: %v\", err)\n\t}\n\treq = req.WithContext(its.Ctx)\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: unable to generate access token: %v\", err)\n\t}\n\tdefer resp.Body.Close()\n\tbody, err := ioutil.ReadAll(io.LimitReader(resp.Body, 1<<20))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: unable to read body: %v\", err)\n\t}\n\tif c := resp.StatusCode; c < 200 || c > 299 {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: status code %d: %s\", c, body)\n\t}\n\n\tvar accessTokenResp impersonateTokenResponse\n\tif err := json.Unmarshal(body, &accessTokenResp); err != nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: unable to parse response: %v\", err)\n\t}\n\texpiry, err := time.Parse(time.RFC3339, accessTokenResp.ExpireTime)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: unable to parse expiry: %v\", err)\n\t}\n\treturn &oauth2.Token{\n\t\tAccessToken: accessTokenResp.AccessToken,\n\t\tExpiry: expiry,\n\t\tTokenType: \"Bearer\",\n\t}, nil\n}", "func TokenSource(aud string) oauth2.TokenSource {\n\tidSrc := idTokenSrc{aud: aud}\n\tinitialToken := &oauth2.Token{}\n\treturn oauth2.ReuseTokenSource(initialToken, idSrc)\n}", "func grpcGetServiceConfigWithOauthToken(sa *string, ctx context.Context) (*serviceconfig.Service, error) {\n\t//scope := sm.CloudPlatformScope\n\tserverAddr := \"servicemanagement.googleapis.com\"\n\tserverAddrWithPort := \"servicemanagement.googleapis.com:443\"\n\tgetServiceConfigProtoReq := smp.GetServiceConfigRequest{\n\t\tServiceName: serviceName,\n\t}\n\tpool, _ := x509.SystemCertPool()\n\t// error handling omitted\n\tcreds := credentials.NewClientTLSFromCert(pool, serverAddrWithPort)\n\tcreds.OverrideServerName(serverAddr)\n\t//perRPC, _ := grpcOauth.NewServiceAccountFromFile(*sa, scope)\n\tbearerToken := \"Bearer ya29.c.ElrSBZKqpjJDEyFjqpfWF1s62FplR8at1Lvt2NDxFKShwNzJr6x2T0YK6ycldNv_ZlA4aNxBjL1jmZdBmjvf6733o8G9sCsxDWHWNgy9Wewz7Fz_Jo7bSaz0psc\"\n\n\t//md := metadata.Pairs(\"Authorization\", bearerToken)\n\t//cos := grpc.HeaderCallOption{\n\t//\tHeaderAddr: &md,\n\t//}\n\n\tperRPC := customJwt{\n\t\ttoken: bearerToken,\n\t}\n\n\tconn, _ := grpc.Dial(\n\t\tserverAddrWithPort,\n\t\tgrpc.WithPerRPCCredentials(&perRPC),\n\t\tgrpc.WithTransportCredentials(creds),\n\t)\n\t//grpc.Header()\n\tgscp, err := smp.NewServiceManagerClient(conn).GetServiceConfig(ctx, &getServiceConfigProtoReq)\n\tif err != nil {\n\t\tfmt.Println(\"Error while making grpc call: \", err)\n\t}\n\tfmt.Println(\"grpc call get name : \", gscp.GetName())\n\treturn gscp, err\n}", "func newReloadingTokenSource(getToken func() []byte) *reloadingTokenSource {\n\treturn &reloadingTokenSource{\n\t\tgetToken: getToken,\n\t}\n}", "func (o *AzureCLICredentialOptions) init() {\n\tif o.tokenProvider == nil {\n\t\to.tokenProvider = defaultTokenProvider()\n\t}\n}", "func (src *gcloudTokenSource) Token() (*oauth2.Token, error) {\n\tcfg, err := GcloudConfig()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn cfg.oauthToken(), nil\n}", "func (dts downscopingTokenSource) Token() (*oauth2.Token, error) {\n\n\tdownscopedOptions := struct {\n\t\tBoundary accessBoundary `json:\"accessBoundary\"`\n\t}{\n\t\tBoundary: accessBoundary{\n\t\t\tAccessBoundaryRules: dts.config.Rules,\n\t\t},\n\t}\n\n\ttok, err := dts.config.RootSource.Token()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"downscope: unable to obtain root token: %v\", err)\n\t}\n\n\tb, err := json.Marshal(downscopedOptions)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"downscope: unable to marshal AccessBoundary payload %v\", err)\n\t}\n\n\tform := url.Values{}\n\tform.Add(\"grant_type\", \"urn:ietf:params:oauth:grant-type:token-exchange\")\n\tform.Add(\"subject_token_type\", \"urn:ietf:params:oauth:token-type:access_token\")\n\tform.Add(\"requested_token_type\", \"urn:ietf:params:oauth:token-type:access_token\")\n\tform.Add(\"subject_token\", tok.AccessToken)\n\tform.Add(\"options\", string(b))\n\n\tmyClient := oauth2.NewClient(dts.ctx, nil)\n\tresp, err := myClient.PostForm(identityBindingEndpoint, form)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to generate POST Request %v\", err)\n\t}\n\tdefer resp.Body.Close()\n\trespBody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"downscope: unable to read response body: %v\", err)\n\t}\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn nil, fmt.Errorf(\"downscope: unable to exchange token; %v. Server responded: %s\", resp.StatusCode, respBody)\n\t}\n\n\tvar tresp downscopedTokenResponse\n\n\terr = json.Unmarshal(respBody, &tresp)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"downscope: unable to unmarshal response body: %v\", err)\n\t}\n\n\t// an exchanged token that is derived from a service account (2LO) has an expired_in value\n\t// a token derived from a users token (3LO) does not.\n\t// The following code uses the time remaining on rootToken for a user as the value for the\n\t// derived token's lifetime\n\tvar expiryTime time.Time\n\tif tresp.ExpiresIn > 0 {\n\t\texpiryTime = time.Now().Add(time.Duration(tresp.ExpiresIn) * time.Second)\n\t} else {\n\t\texpiryTime = tok.Expiry\n\t}\n\n\tnewToken := &oauth2.Token{\n\t\tAccessToken: tresp.AccessToken,\n\t\tTokenType: tresp.TokenType,\n\t\tExpiry: expiryTime,\n\t}\n\treturn newToken, nil\n}", "func TokenSource(ctx context.Context, scopes ...string) oauth2.TokenSource {\n\treturn TokenSourceEnv(ctx, envPrivateKey, scopes...)\n}", "func getClient(ctx context.Context, config *oauth2.Config) *http.Client {\n cacheFile, err := tokenCacheFile()\n if err != nil {\n log.Fatalf(\"Unable to get path to cached credential file. %v\", err)\n }\n tok, err := tokenFromFile(cacheFile)\n if err != nil {\n tok = getTokenFromWeb(config)\n saveToken(cacheFile, tok)\n }\n return config.Client(ctx, tok)\n}", "func getClient(ctx context.Context, config *oauth2.Config) *http.Client {\n cacheFile, err := tokenCacheFile()\n if err != nil {\n log.Fatalf(\"Unable to get path to cached credential file. %v\", err)\n }\n tok, err := tokenFromFile(cacheFile)\n if err != nil {\n tok = getTokenFromWeb(config)\n saveToken(cacheFile, tok)\n }\n return config.Client(ctx, tok)\n}", "func getClient(ctx context.Context, config *oauth2.Config) *http.Client {\n cacheFile, err := tokenCacheFile()\n if err != nil {\n log.Fatalf(\"Unable to get path to cached credential file. %v\", err)\n }\n tok, err := tokenFromFile(cacheFile)\n if err != nil {\n tok = getTokenFromWeb(config)\n saveToken(cacheFile, tok)\n }\n return config.Client(ctx, tok)\n}", "func (a *auth) GetTokenSource(ctx context.Context, ref reference.Named) (oauth2.TokenSource, error) {\n\tsource, err := google.DefaultTokenSource(ctx, gcrScope)\n\tif nil != err {\n\t\terr = fmt.Errorf(\"failed to get Google Auth token source: %s\", err)\n\t}\n\n\treturn source, err\n\n}", "func newTokenSource(ctx context.Context, settings *Settings) (*internal.TokenSource, error) {\n\tvar ts internal.TokenSource\n\tvar err error\n\tif settings == nil {\n\t\tts, err = DefaultTokenSource(ctx, DefaultScope)\n\t} else if settings.APIKey != \"\" {\n\t\treturn nil, nil\n\t} else if settings.Scope != \"\" {\n\t\tts, err = OAuthJSONTokenSource(ctx, settings)\n\t} else {\n\t\tts, err = JWTTokenSource(ctx, settings)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &ts, err\n}", "func getClient(ctx context.Context, config *oauth2.Config, tokenFile string) (*http.Client, error) {\n\tloadedTok, err := tokenFromFile(tokenFile)\n\tif err != nil {\n\t\tlog.Printf(\"Could not load token from '%s'\\n\", tokenFile)\n\t\treturn nil, err\n\t}\n\t// force refresh\n\ttokenSource := config.TokenSource(ctx, loadedTok)\n\ttok, err := tokenSource.Token()\n\tif err != nil {\n\t\tlog.Printf(\"Could not refresh token from '%s'\\n\", tokenFile)\n\t\treturn nil, err\n\t}\n\treturn config.Client(ctx, tok), nil\n}", "func InitPoolToken(poolTokenName string) token.Token {\n\treturn token.Token{\n\t\tDescription: poolTokenName,\n\t\tSymbol: poolTokenName,\n\t\tOriginalSymbol: poolTokenName,\n\t\tWholeName: poolTokenName,\n\t\tOriginalTotalSupply: sdk.NewDec(0),\n\t\tOwner: supply.NewModuleAddress(ModuleName),\n\t\tType: GenerateTokenType,\n\t\tMintable: true,\n\t}\n}", "func GCPTokenSource(ts gcp.TokenSource) broker.Option {\n\treturn optfunc(gcpTokenSourceKey{}, ts)\n}", "func (p *provider) Init(ctx servicehub.Context) error {\n\tp.accessKeyValidator = &accessKeyValidator{\n\t\tTokenService: p.TokenService,\n\t\tcollection: AccessItemCollection{},\n\t}\n\tctx.AddTask(p.InitAKItemTask)\n\tctx.AddTask(p.SyncAKItemTask)\n\treturn nil\n}", "func SetToken(token string) {\n\tDefaultClient.SetToken(token)\n}", "func (s *reloadingTokenSource) Token() (*oauth2.Token, error) {\n\treturn &oauth2.Token{\n\t\tAccessToken: string(s.getToken()),\n\t}, nil\n}", "func InitTokenGenerator(env conf.EnvironmentKey) (*TokenGenerator, error) {\n\tconfig, err := conf.ReadTokenConfig(env)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttokenGenerator, err := parseTokenConfig(config)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn tokenGenerator, nil\n}", "func (c *client) newClientToken(token string) *gitea.Client {\n\tclient := gitea.NewClient(c.URL, token)\n\tif c.SkipVerify {\n\t\thttpClient := &http.Client{}\n\t\thttpClient.Transport = &http.Transport{\n\t\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t\t}\n\t\tclient.SetHTTPClient(httpClient)\n\t}\n\treturn client\n}", "func (ts *azureTokenSource) Token() (*azureToken, error) {\n\tts.lock.Lock()\n\tdefer ts.lock.Unlock()\n\n\tvar err error\n\ttoken := ts.cache.getToken(azureTokenKey)\n\n\tif token != nil && !token.token.IsExpired() {\n\t\treturn token, nil\n\t}\n\n\t// retrieve from config if no cache\n\tif token == nil {\n\t\ttokenFromCfg, err := ts.retrieveTokenFromCfg()\n\n\t\tif err == nil {\n\t\t\ttoken = tokenFromCfg\n\t\t}\n\t}\n\n\tif token != nil {\n\t\t// cache and return if the token is as good\n\t\t// avoids frequent persistor calls\n\t\tif !token.token.IsExpired() {\n\t\t\tts.cache.setToken(azureTokenKey, token)\n\t\t\treturn token, nil\n\t\t}\n\n\t\tklog.V(4).Info(\"Refreshing token.\")\n\t\ttokenFromRefresh, err := ts.Refresh(token)\n\t\tswitch {\n\t\tcase err == nil:\n\t\t\ttoken = tokenFromRefresh\n\t\tcase autorest.IsTokenRefreshError(err):\n\t\t\tklog.V(4).Infof(\"Failed to refresh expired token, proceed to auth: %v\", err)\n\t\t\t// reset token to nil so that the token source will be used to acquire new\n\t\t\ttoken = nil\n\t\tdefault:\n\t\t\treturn nil, fmt.Errorf(\"unexpected error when refreshing token: %v\", err)\n\t\t}\n\t}\n\n\tif token == nil {\n\t\ttokenFromSource, err := ts.source.Token()\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed acquiring new token: %v\", err)\n\t\t}\n\t\ttoken = tokenFromSource\n\t}\n\n\t// sanity check\n\tif token == nil {\n\t\treturn nil, fmt.Errorf(\"unable to acquire token\")\n\t}\n\n\t// corner condition, newly got token is valid but expired\n\tif token.token.IsExpired() {\n\t\treturn nil, fmt.Errorf(\"newly acquired token is expired\")\n\t}\n\n\terr = ts.storeTokenInCfg(token)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"storing the refreshed token in configuration: %v\", err)\n\t}\n\tts.cache.setToken(azureTokenKey, token)\n\n\treturn token, nil\n}", "func init() {\n\ttoken = nep17.Token{\n\t\tName: \"Awesome NEO Token\",\n\t\tSymbol: \"ANT\",\n\t\tDecimals: decimals,\n\t\tOwner: owner,\n\t\tTotalSupply: 11000000 * multiplier,\n\t\tCirculationKey: \"TokenCirculation\",\n\t}\n\tctx = storage.GetContext()\n}", "func newClient(token string) *github.Client {\n\tts := oauth2.StaticTokenSource(\n\t\t&oauth2.Token{AccessToken: token},\n\t)\n\ttc := oauth2.NewClient(context.Background(), ts)\n\n\treturn github.NewClient(tc)\n}", "func (g *gcs) Init(ctx context.Context) (err error) {\n\tg.context = context.Background()\n\n\tgcsClient, err := storage.NewClient(g.context, option.WithCredentialsFile(g.credentialsJSON))\n\tif err != nil {\n\t\treturn\n\t}\n\n\tg.bucket = gcsClient.Bucket(g.bucketName)\n\tg.client = gcsClient\n\n\treturn\n}", "func TestGetToken(t *testing.T) {\n\tmc := MockClient{t: t}\n\tmc.DoFunc = validDo\n\tmc.GetFunc = validGet\n\tconfig := ClientConfig{\n\t\tScopes: []string{\"thing\"},\n\t\tOktaDomain: \"mockta.local\",\n\t\tHTTPClient: &mc,\n\t}\n\n\tclient, err := NewClient(config)\n\n\tif err != nil {\n\t\tt.Fatalf(\"Failed: %s\", err)\n\t}\n\n\t// Test surge of requests these should all use the same key\n\tresult := testConcurrency(client, 0, 100, t)\n\tif len(result) > 1 {\n\t\tt.Fatalf(\"Concurrency Test 1 Failed: got %d, want 1\\n\", len(result))\n\t}\n\n\t// Test renewals\n\tresult = testConcurrency(client, 1000, 10, t)\n\tif len(result) != 10 {\n\t\tt.Fatalf(\"Concurrency Test 2 Failed: got %d, want 10\\n\", len(result))\n\t}\n}", "func Provider() terraform.ResourceProvider {\n\treturn &schema.Provider{\n\t\tSchema: map[string]*schema.Schema{\n\t\t\t\"credentials\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tOptional: true,\n\t\t\t\tDefaultFunc: schema.MultiEnvDefaultFunc([]string{\n\t\t\t\t\t\"GOOGLE_CREDENTIALS\",\n\t\t\t\t\t\"GOOGLE_CLOUD_KEYFILE_JSON\",\n\t\t\t\t\t\"GCLOUD_KEYFILE_JSON\",\n\t\t\t\t}, nil),\n\t\t\t\tValidateFunc: validateCredentials,\n\t\t\t},\n\n\t\t\t\"access_token\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tOptional: true,\n\t\t\t\tDefaultFunc: schema.MultiEnvDefaultFunc([]string{\n\t\t\t\t\t\"GOOGLE_OAUTH_ACCESS_TOKEN\",\n\t\t\t\t}, nil),\n\t\t\t\tConflictsWith: []string{\"credentials\"},\n\t\t\t},\n\n\t\t\t\"project\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tOptional: true,\n\t\t\t\tDefaultFunc: schema.MultiEnvDefaultFunc([]string{\n\t\t\t\t\t\"GOOGLE_PROJECT\",\n\t\t\t\t\t\"GOOGLE_CLOUD_PROJECT\",\n\t\t\t\t\t\"GCLOUD_PROJECT\",\n\t\t\t\t\t\"CLOUDSDK_CORE_PROJECT\",\n\t\t\t\t}, nil),\n\t\t\t},\n\n\t\t\t\"region\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tOptional: true,\n\t\t\t\tDefaultFunc: schema.MultiEnvDefaultFunc([]string{\n\t\t\t\t\t\"GOOGLE_REGION\",\n\t\t\t\t\t\"GCLOUD_REGION\",\n\t\t\t\t\t\"CLOUDSDK_COMPUTE_REGION\",\n\t\t\t\t}, nil),\n\t\t\t},\n\n\t\t\t\"zone\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tOptional: true,\n\t\t\t\tDefaultFunc: schema.MultiEnvDefaultFunc([]string{\n\t\t\t\t\t\"GOOGLE_ZONE\",\n\t\t\t\t\t\"GCLOUD_ZONE\",\n\t\t\t\t\t\"CLOUDSDK_COMPUTE_ZONE\",\n\t\t\t\t}, nil),\n\t\t\t},\n\t\t\t\"scopes\": {\n\t\t\t\tType: schema.TypeList,\n\t\t\t\tOptional: true,\n\t\t\t\tElem: &schema.Schema{Type: schema.TypeString},\n\t\t\t},\n\t\t},\n\n\t\tDataSourcesMap: map[string]*schema.Resource{\n\t\t\t\"google_active_folder\": dataSourceGoogleActiveFolder(),\n\t\t\t\"google_billing_account\": dataSourceGoogleBillingAccount(),\n\t\t\t\"google_dns_managed_zone\": dataSourceDnsManagedZone(),\n\t\t\t\"google_client_config\": dataSourceGoogleClientConfig(),\n\t\t\t\"google_cloudfunctions_function\": dataSourceGoogleCloudFunctionsFunction(),\n\t\t\t\"google_compute_address\": dataSourceGoogleComputeAddress(),\n\t\t\t\"google_compute_backend_service\": dataSourceGoogleComputeBackendService(),\n\t\t\t\"google_compute_default_service_account\": dataSourceGoogleComputeDefaultServiceAccount(),\n\t\t\t\"google_compute_forwarding_rule\": dataSourceGoogleComputeForwardingRule(),\n\t\t\t\"google_compute_image\": dataSourceGoogleComputeImage(),\n\t\t\t\"google_compute_instance\": dataSourceGoogleComputeInstance(),\n\t\t\t\"google_compute_global_address\": dataSourceGoogleComputeGlobalAddress(),\n\t\t\t\"google_compute_instance_group\": dataSourceGoogleComputeInstanceGroup(),\n\t\t\t\"google_compute_lb_ip_ranges\": dataSourceGoogleComputeLbIpRanges(),\n\t\t\t\"google_compute_network\": dataSourceGoogleComputeNetwork(),\n\t\t\t\"google_compute_regions\": dataSourceGoogleComputeRegions(),\n\t\t\t\"google_compute_region_instance_group\": dataSourceGoogleComputeRegionInstanceGroup(),\n\t\t\t\"google_compute_subnetwork\": dataSourceGoogleComputeSubnetwork(),\n\t\t\t\"google_compute_zones\": dataSourceGoogleComputeZones(),\n\t\t\t\"google_compute_vpn_gateway\": dataSourceGoogleComputeVpnGateway(),\n\t\t\t\"google_compute_ssl_policy\": dataSourceGoogleComputeSslPolicy(),\n\t\t\t\"google_container_cluster\": dataSourceGoogleContainerCluster(),\n\t\t\t\"google_container_engine_versions\": dataSourceGoogleContainerEngineVersions(),\n\t\t\t\"google_container_registry_repository\": dataSourceGoogleContainerRepo(),\n\t\t\t\"google_container_registry_image\": dataSourceGoogleContainerImage(),\n\t\t\t\"google_iam_policy\": dataSourceGoogleIamPolicy(),\n\t\t\t\"google_iam_role\": dataSourceGoogleIamRole(),\n\t\t\t\"google_kms_secret\": dataSourceGoogleKmsSecret(),\n\t\t\t\"google_kms_key_ring\": dataSourceGoogleKmsKeyRing(),\n\t\t\t\"google_kms_crypto_key\": dataSourceGoogleKmsCryptoKey(),\n\t\t\t\"google_folder\": dataSourceGoogleFolder(),\n\t\t\t\"google_netblock_ip_ranges\": dataSourceGoogleNetblockIpRanges(),\n\t\t\t\"google_organization\": dataSourceGoogleOrganization(),\n\t\t\t\"google_project\": dataSourceGoogleProject(),\n\t\t\t\"google_project_services\": dataSourceGoogleProjectServices(),\n\t\t\t\"google_service_account\": dataSourceGoogleServiceAccount(),\n\t\t\t\"google_service_account_key\": dataSourceGoogleServiceAccountKey(),\n\t\t\t\"google_storage_bucket_object\": dataSourceGoogleStorageBucketObject(),\n\t\t\t\"google_storage_object_signed_url\": dataSourceGoogleSignedUrl(),\n\t\t\t\"google_storage_project_service_account\": dataSourceGoogleStorageProjectServiceAccount(),\n\t\t\t\"google_storage_transfer_project_service_account\": dataSourceGoogleStorageTransferProjectServiceAccount(),\n\t\t},\n\n\t\tResourcesMap: ResourceMap(),\n\n\t\tConfigureFunc: providerConfigure,\n\t}\n}", "func (gts *GithubTokenSource) Token() (*oauth2.Token, error) {\n\treturn &oauth2.Token{AccessToken: gts.PersonalToken}, nil\n}", "func preConfigureCallback(vars resource.PropertyMap, c tfshim.ResourceConfig) error {\n\n\tenvName := stringValue(vars, \"environment\", []string{\"ARM_ENVIRONMENT\"})\n\tif envName == \"\" {\n\t\tenvName = \"public\"\n\t}\n\n\tenv, err := environments.FromName(envName)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to read Azure environment \\\"%s\\\": %v\", envName, err)\n\t}\n\n\tauthConfig := auth.Credentials{\n\t\tEnvironment: *env,\n\t\tEnableAuthenticatingUsingClientSecret: true,\n\t\tEnableAuthenticatingUsingAzureCLI: true,\n\t\tTenantID: stringValue(vars, \"tenantId\", []string{\"ARM_TENANT_ID\"}),\n\t\tClientID: stringValue(vars, \"clientId\", []string{\"ARM_CLIENT_ID\"}),\n\t\tClientSecret: stringValue(vars, \"clientSecret\", []string{\"ARM_CLIENT_SECRET\"}),\n\n\t\tEnableAuthenticatingUsingClientCertificate: true,\n\t\t// We don't handle ClientCertData yet, which is the actual base-64 encoded cert in config\n\t\tClientCertificatePassword: stringValue(vars, \"clientCertificatePassword\", []string{\"ARM_CLIENT_CERTIFICATE_PASSWORD\"}),\n\t\tClientCertificatePath: stringValue(vars, \"clientCertificatePath\", []string{\"ARM_CLIENT_CERTIFICATE_PATH\"}),\n\n\t\tEnableAuthenticatingUsingManagedIdentity: boolValue(vars, \"msiEndpoint\", []string{\"ARM_USE_MSI\"}),\n\t\tCustomManagedIdentityEndpoint: stringValue(vars, \"msiEndpoint\", []string{\"ARM_MSI_ENDPOINT\"}),\n\n\t\t// The configuration below would enable OIDC auth which we haven't tested and documented yet.\n\t\t//FederatedAssertion: idToken,\n\t\t//IDTokenRequestURL: d.Get(\"oidc_request_url\").(string),\n\t\t//IDTokenRequestToken: d.Get(\"oidc_request_token\").(string),\n\t\t//EnableClientFederatedAuth: d.Get(\"use_oidc\").(bool),\n\t\t//EnableGitHubOIDCAuth: d.Get(\"use_oidc\").(bool),\n\t}\n\n\t_, err = auth.NewAuthorizerFromCredentials(context.Background(), authConfig, env.MicrosoftGraph)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to load Azure credentials.\\n\"+\n\t\t\t\"Details: %v\\n\\n\"+\n\t\t\t\"\\tPlease make sure you have signed in via 'az login' or configured another authentication method.\\n\\n\"+\n\t\t\t\"\\tSee https://www.pulumi.com/registry/packages/azuread/installation-configuration/ for more information.\", err)\n\t}\n\n\treturn nil\n}", "func (s *storageTokenSource) Token() (*oauth2.Token, error) {\n\tif token, err := s.Config.Storage.GetToken(); err == nil && token.Valid() {\n\t\treturn token, err\n\t}\n\ttoken, err := s.TokenSource.Token()\n\tif err != nil {\n\t\treturn token, err\n\t}\n\tif err := s.Config.Storage.SetToken(token); err != nil {\n\t\treturn nil, err\n\t}\n\treturn token, nil\n}", "func initClientContext(c *GSSAPIContext, service string, inputToken []byte) error {\n\tc.ServiceName = service\n\n\tvar _inputToken *gssapi.Buffer\n\tvar err error\n\tif inputToken == nil {\n\t\t_inputToken = c.GSS_C_NO_BUFFER\n\t} else {\n\t\t_inputToken, err = c.MakeBufferBytes(inputToken)\n\t\tdefer _inputToken.Release()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tpreparedName := prepareServiceName(c)\n\tdefer preparedName.Release()\n\n\tcontextId, _, token, outputRetFlags, _, err := c.InitSecContext(\n\t\tnil,\n\t\tc.contextId,\n\t\tpreparedName,\n\t\tc.GSS_MECH_KRB5,\n\t\tc.reqFlags,\n\t\t0,\n\t\tc.GSS_C_NO_CHANNEL_BINDINGS,\n\t\t_inputToken)\n\tdefer token.Release()\n\n\tc.token = token.Bytes()\n\tc.contextId = contextId\n\tc.availFlags = outputRetFlags\n\treturn nil\n}", "func (idx *Unique) Init() error {\n\ttokenManager, err := jwt.New(map[string]interface{}{\n\t\t\"secret\": idx.cs3conf.JWTSecret,\n\t})\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tidx.tokenManager = tokenManager\n\n\tclient, err := pool.GetStorageProviderServiceClient(idx.cs3conf.ProviderAddr)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tidx.storageProvider = client\n\n\tctx := context.Background()\n\ttk, err := idx.authenticate(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\tctx = metadata.AppendToOutgoingContext(ctx, revactx.TokenHeader, tk)\n\n\tif err := idx.makeDirIfNotExists(ctx, idx.indexBaseDir); err != nil {\n\t\treturn err\n\t}\n\n\tif err := idx.makeDirIfNotExists(ctx, idx.indexRootDir); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func getOauthClient(token string) (*github.Client, context.Context) {\n\toauthToken := &oauth2.Token{\n\t\tAccessToken: token,\n\t}\n\tctx := context.Background()\n\tts := oauth2.StaticTokenSource(oauthToken)\n\ttc := oauth2.NewClient(ctx, ts)\n\tghClient := github.NewClient(tc)\n\treturn ghClient, ctx\n}", "func initConfig() {\n\ttoken := rootCmd.Flag(\"token\").Value.String()\n\tgiturl := rootCmd.Flag(\"giturl\").Value.String()\n\tGitClient = gitlab.NewClient(nil, token)\n\tGitClient.SetBaseURL(giturl + \"/api/v4/\")\n}", "func Initialize(config *viper.Viper) (tokenConfig *Config, err error) {\n\ttokenConfig = &Config{PlatformName: config.GetString(\"PlatformName\")}\n\n\tbytes, err := getKey(config, \"Public\")\n\tif err != nil {\n\t\treturn\n\t}\n\ttokenConfig.PublicKey, err = crypto.ParseRSAPublicKeyFromPEM(bytes)\n\tif err != nil {\n\t\treturn\n\t}\n\tbytes, err = getKey(config, \"Private\")\n\tif err != nil {\n\t\treturn\n\t}\n\ttokenConfig.PrivateKey, err = crypto.ParseRSAPrivateKeyFromPEM(bytes)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn\n}", "func fetchToken(settings *Settings, taskSettings *TaskSettings) *oauth2.Token {\n\ttoken, err := LookupCache(settings)\n\ttokenExpired := isTokenExpired(token)\n\tif token == nil || tokenExpired {\n\t\tif taskSettings.AuthType == \"sso\" {\n\t\t\ttoken, err = SSOFetch(taskSettings.SsoCli, settings.Email, settings.Scope)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(err)\n\t\t\t\treturn nil\n\t\t\t}\n\t\t} else {\n\t\t\tfetchSettings := settings\n\t\t\tif tokenExpired && taskSettings.Refresh {\n\t\t\t\t// If creds cannot be retrieved here, which is unexpected, we will ignore\n\t\t\t\t// the error and let FetchToken return a standardized error message\n\t\t\t\t// in the subsequent step.\n\t\t\t\tcreds, _ := FindJSONCredentials(context.Background(), settings)\n\t\t\t\trefreshTokenJSON := BuildRefreshTokenJSON(token.RefreshToken, creds)\n\t\t\t\tif refreshTokenJSON != \"\" {\n\t\t\t\t\trefreshSettings := *settings // Make a shallow copy\n\t\t\t\t\trefreshSettings.CredentialsJSON = refreshTokenJSON\n\t\t\t\t\tfetchSettings = &refreshSettings\n\t\t\t\t}\n\t\t\t}\n\t\t\ttoken, err = FetchToken(context.Background(), fetchSettings)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(err)\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t\tif settings.ServiceAccount != \"\" {\n\t\t\ttoken, err = GenerateServiceAccountAccessToken(token.AccessToken, settings.ServiceAccount, settings.Scope)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(err)\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t\tif settings.Sts {\n\t\t\ttoken, err = StsExchange(token.AccessToken, EncodeClaims(settings))\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(err)\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t\terr = InsertCache(settings, token)\n\t\tif err != nil {\n\t\t\tfmt.Println(err)\n\t\t\treturn nil\n\t\t}\n\t}\n\treturn token\n}", "func (t *TokenSource) Token() (*oauth2.Token, error) {\n\ttoken, err := t.TokenGenerator.GetWithSTS(context.Background(), t.ClusterID)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error generating token: %w\", err)\n\t}\n\treturn &oauth2.Token{\n\t\tAccessToken: token.Token,\n\t\tExpiry: token.Expiration.Add(-t.Leeway),\n\t}, nil\n}", "func init() {\n RootCmd.AddCommand(RefreshTokenCmd)\n}", "func ClientConfigurer(d *schema.ResourceData) (interface{}, error) {\n\t// req, err := http.NewRequest(\"POST\", d.Get(\"auth_server\").(string), nil)\n\t// if err != nil {\n\t// \treturn nil, err\n\t// }\n\t// req.SetBasicAuth(\"SpotifyAuthProxy\", d.Get(\"api_key\").(string))\n\t// resp, err := http.DefaultClient.Do(req)\n\t// if err != nil {\n\t// \treturn nil, err\n\t// }\n\t// defer resp.Body.Close()\n\t// body, err := ioutil.ReadAll(resp.Body)\n\t// if err != nil {\n\t// \treturn nil, err\n\t// }\n\t// if resp.StatusCode != http.StatusOK {\n\t// \treturn nil, fmt.Errorf(\"%s\", string(body))\n\t// }\n\n\t// tokenData := struct {\n\t// \tAccessToken string `json:\"access_token\"`\n\t// \tRefreshToken string `json:\"refresh_token\"`\n\t// \tExpiresIn int `json:\"expires_in\"`\n\t// \tTokenType string `json:\"token_type\"`\n\t// }{}\n\n\t// if err := json.Unmarshal(body, &tokenData); err != nil {\n\t// \treturn nil, err\n\t// }\n\n\t// token := &oauth2.Token{\n\t// \tAccessToken: tokenData.AccessToken,\n\t// \tRefreshToken: tokenData.RefreshToken,\n\t// \tTokenType: tokenData.TokenType,\n\t// \tExpiry: time.Now().Add(time.Duration(tokenData.ExpiresIn) * time.Second),\n\t// }\n\n\t// cnf := &oauth2.Config{\n\t// \t// ClientID: d.Get(\"client_id\").(string),\n\t// \tClientID: \"SpotifyAuthProxy\",\n\t// \tClientSecret: d.Get(\"api_key\").(string),\n\t// \tEndpoint: oauth2.Endpoint{\n\t// \t\tTokenURL: d.Get(\"auth_server\").(string),\n\t// \t\tAuthStyle: oauth2.AuthStyleInHeader,\n\t// \t},\n\t// }\n\n\ttransport := &transport{\n\t\tAPIKey: d.Get(\"api_key\").(string),\n\t\tServer: d.Get(\"auth_server\").(string),\n\t}\n\ttransport.getToken()\n\n\tclient := spotify.NewClient(&http.Client{\n\t\tTransport: transport,\n\t})\n\treturn &client, nil\n}", "func getArmClient(authCfg *authentication.Config, tfVersion string, skipProviderRegistration bool) (*ArmClient, error) {\n\tenv, err := authentication.LoadEnvironmentFromUrl(authCfg.CustomResourceManagerEndpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// client declarations:\n\tclient := ArmClient{\n\t\tclientId: authCfg.ClientID,\n\t\ttenantId: authCfg.TenantID,\n\t\tsubscriptionId: authCfg.SubscriptionID,\n\t\tterraformVersion: tfVersion,\n\t\tenvironment: *env,\n\t\tusingServicePrincipal: authCfg.AuthenticatedAsAServicePrincipal,\n\t\tskipProviderRegistration: skipProviderRegistration,\n\t}\n\n\toauth, err := authCfg.BuildOAuthConfig(env.ActiveDirectoryEndpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsender := sender.BuildSender(\"AzureStack\")\n\n\t// Resource Manager endpoints\n\tendpoint := env.ResourceManagerEndpoint\n\n\t// Instead of the same endpoint use token audience to get the correct token.\n\tauth, err := authCfg.GetAuthorizationToken(sender, oauth, env.TokenAudience)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Graph Endpoints\n\tgraphEndpoint := env.GraphEndpoint\n\tgraphAuth, err := authCfg.GetAuthorizationToken(sender, oauth, graphEndpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tclient.registerAuthentication(graphEndpoint, client.tenantId, graphAuth, sender)\n\tclient.registerComputeClients(endpoint, client.subscriptionId, auth)\n\tclient.registerDNSClients(endpoint, client.subscriptionId, auth)\n\tclient.registerNetworkingClients(endpoint, client.subscriptionId, auth)\n\tclient.registerResourcesClients(endpoint, client.subscriptionId, auth)\n\tclient.registerStorageClients(endpoint, client.subscriptionId, auth)\n\n\treturn &client, nil\n}", "func mustProviderClient() *gophercloud.ProviderClient {\n\topt, err := openstack.AuthOptionsFromEnv()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tprovider, err := openstack.AuthenticatedClient(opt)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tprovider.ReauthFunc = func() error {\n\t\topt, err := openstack.AuthOptionsFromEnv()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tnewprov, err := openstack.AuthenticatedClient(opt)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tprovider.CopyTokenFrom(newprov)\n\t\treturn nil\n\t}\n\treturn provider\n}", "func (gr *Reconciler) Init() {\n\tkm := k8s.NewRsrcManager().WithName(\"basek8s\").WithClient(gr.Manager.GetClient()).WithScheme(gr.Manager.GetScheme())\n\tgr.RsrcMgr.Add(k8s.Type, km)\n\tapp.AddToScheme(&AddToSchemes)\n\tAddToSchemes.AddToScheme(gr.Manager.GetScheme())\n}", "func (t *Handler) Init() error {\n\tlog.Info(\"TeamHandler.Init\")\n\tvar err error\n\tt.clientset, err = authorization.CreateClientSet()\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\tpanic(err.Error())\n\t}\n\tt.edgenetClientset, err = authorization.CreateEdgeNetClientSet()\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\tpanic(err.Error())\n\t}\n\tt.resourceQuota = &corev1.ResourceQuota{}\n\tt.resourceQuota.Name = \"team-quota\"\n\tt.resourceQuota.Spec = corev1.ResourceQuotaSpec{\n\t\tHard: map[corev1.ResourceName]resource.Quantity{\n\t\t\t\"cpu\": resource.MustParse(\"5m\"),\n\t\t\t\"memory\": resource.MustParse(\"1Mi\"),\n\t\t\t\"requests.storage\": resource.MustParse(\"1Mi\"),\n\t\t\t\"pods\": resource.Quantity{Format: \"0\"},\n\t\t\t\"count/persistentvolumeclaims\": resource.Quantity{Format: \"0\"},\n\t\t\t\"count/services\": resource.Quantity{Format: \"0\"},\n\t\t\t\"count/configmaps\": resource.Quantity{Format: \"0\"},\n\t\t\t\"count/replicationcontrollers\": resource.Quantity{Format: \"0\"},\n\t\t\t\"count/deployments.apps\": resource.Quantity{Format: \"0\"},\n\t\t\t\"count/deployments.extensions\": resource.Quantity{Format: \"0\"},\n\t\t\t\"count/replicasets.apps\": resource.Quantity{Format: \"0\"},\n\t\t\t\"count/replicasets.extensions\": resource.Quantity{Format: \"0\"},\n\t\t\t\"count/statefulsets.apps\": resource.Quantity{Format: \"0\"},\n\t\t\t\"count/statefulsets.extensions\": resource.Quantity{Format: \"0\"},\n\t\t\t\"count/jobs.batch\": resource.Quantity{Format: \"0\"},\n\t\t\t\"count/cronjobs.batch\": resource.Quantity{Format: \"0\"},\n\t\t},\n\t}\n\treturn err\n}", "func newJWTBase(ctx context.Context, cfg Config) (string, error) {\n\tserviceAccount, project, tokenSource, err := getServiceAccountInfo(ctx, cfg)\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"unable to get service account from environment\")\n\t}\n\n\tpayload, err := json.Marshal(map[string]interface{}{\n\t\t\"aud\": \"vault/\" + cfg.Role,\n\t\t\"sub\": serviceAccount,\n\t\t\"exp\": time.Now().UTC().Add(5 * time.Minute).Unix(),\n\t})\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"unable to encode JWT payload\")\n\t}\n\n\thc := getHTTPClient(ctx, cfg)\n\t// reuse base transport and timeout but sprinkle on the token source for IAM access\n\thcIAM := &http.Client{\n\t\tTimeout: hc.Timeout,\n\t\tTransport: &oauth2.Transport{\n\t\t\tSource: tokenSource,\n\t\t\tBase: hc.Transport,\n\t\t},\n\t}\n\tiamClient, err := iam.New(hcIAM)\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"unable to init IAM client\")\n\t}\n\n\tif cfg.IAMAddress != \"\" {\n\t\tiamClient.BasePath = cfg.IAMAddress\n\t}\n\n\tresp, err := iamClient.Projects.ServiceAccounts.SignJwt(\n\t\tfmt.Sprintf(\"projects/%s/serviceAccounts/%s\",\n\t\t\tproject, serviceAccount),\n\t\t&iam.SignJwtRequest{Payload: string(payload)}).Context(ctx).Do()\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"unable to sign JWT\")\n\t}\n\treturn resp.SignedJwt, nil\n}", "func newV1TokenAuthenticator(serverURL string, clientCert, clientKey, ca []byte, cacheTime time.Duration, implicitAuds authenticator.Audiences, metrics AuthenticatorMetrics) (authenticator.Token, error) {\n\ttempfile, err := ioutil.TempFile(\"\", \"\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tp := tempfile.Name()\n\tdefer os.Remove(p)\n\tconfig := v1.Config{\n\t\tClusters: []v1.NamedCluster{\n\t\t\t{\n\t\t\t\tCluster: v1.Cluster{Server: serverURL, CertificateAuthorityData: ca},\n\t\t\t},\n\t\t},\n\t\tAuthInfos: []v1.NamedAuthInfo{\n\t\t\t{\n\t\t\t\tAuthInfo: v1.AuthInfo{ClientCertificateData: clientCert, ClientKeyData: clientKey},\n\t\t\t},\n\t\t},\n\t}\n\tif err := json.NewEncoder(tempfile).Encode(config); err != nil {\n\t\treturn nil, err\n\t}\n\n\tclientConfig, err := webhookutil.LoadKubeconfig(p, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tc, err := tokenReviewInterfaceFromConfig(clientConfig, \"v1\", testRetryBackoff)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tauthn, err := newWithBackoff(c, testRetryBackoff, implicitAuds, 10*time.Second, metrics)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn cache.New(authn, false, cacheTime, cacheTime), nil\n}", "func (t *tokenSource) Token() (*oauth2.Token, error) {\n\treturn t.token, nil\n}", "func (t *tokenSource) Token() (*oauth2.Token, error) {\n\treturn t.token, nil\n}", "func (ap *Provider) updateToken(client clients.AzureClient, token *v3.Token) error {\n\t// For the new flow via Microsoft Graph, the caching and updating of the token to the Microsoft Graph API\n\t// is handled separately via the SDK client cache.\n\tcfg, err := ap.GetAzureConfigK8s()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !IsConfigDeprecated(cfg) {\n\t\treturn nil\n\t}\n\n\tcurrent, err := client.MarshalTokenJSON()\n\tif err != nil {\n\t\treturn errors.New(\"failed to unmarshal token\")\n\t}\n\n\tsecret, err := ap.tokenMGR.GetSecret(token.UserID, token.AuthProvider, []*v3.Token{token})\n\tif err != nil {\n\t\tif apierrors.IsNotFound(err) {\n\t\t\t// providerToken doesn't exist as a secret, update on token.\n\t\t\tif current, ok := token.ProviderInfo[\"access_token\"]; ok && current != current {\n\t\t\t\ttoken.ProviderInfo[\"access_token\"] = current\n\t\t\t}\n\t\t\treturn nil\n\t\t}\n\t\treturn err\n\t}\n\n\tif current == secret {\n\t\treturn nil\n\t}\n\n\treturn ap.tokenMGR.UpdateSecret(token.UserID, token.AuthProvider, current)\n}", "func (s *StoreNotifyingSource) Token() (*oauth2.Token, error) {\n\tt, err := s.src.Token()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = s.store.Save(t)\n\treturn t, err\n}", "func getClient(config *oauth2.Config) *http.Client {\r\n\t// The file token.json stores the user's access and refresh tokens, and is\r\n\t// created automatically when the authorization flow completes for the first\r\n\t// time.\r\n\ttokFile := path.Join(secretDir, \"token.json\")\r\n\ttok, err := tokenFromFile(tokFile)\r\n\tif err != nil {\r\n\t\ttok = getTokenFromWeb(config)\r\n\t\t// saveToken(tokFile, tok)\r\n\t}\r\n\treturn config.Client(context.Background(), tok)\r\n}", "func (c *Client) Token(ctx context.Context, opts *tokenOptions) (*logical.Response, error) {\n\t// Marshal a request body only if there are any user-specified GitHub App\n\t// token constraints.\n\tvar body io.ReadWriter\n\tif opts != nil {\n\t\tbody = new(bytes.Buffer)\n\t\tif err := json.NewEncoder(body).Encode(opts); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Build the token request.\n\treq, err := http.NewRequestWithContext(ctx, http.MethodPost, c.url.String(), body)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"%w: %v\", errUnableToBuildAccessTokenReq, err)\n\t}\n\n\treq.Header.Set(\"User-Agent\", projectName)\n\n\tif body != nil {\n\t\treq.Header.Set(\"Content-Type\", \"application/json\")\n\t}\n\n\t// Perform the request, re-using the shared transport.\n\tres, err := c.transport.RoundTrip(req)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"%w: RoundTrip error: %v\", errUnableToCreateAccessToken, err)\n\t}\n\n\tdefer res.Body.Close()\n\n\tif statusCode(res.StatusCode).Unsuccessful() {\n\t\tbodyBytes, err := ioutil.ReadAll(res.Body)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"%w: %s: error reading error response body: %v\",\n\t\t\t\terrUnableToCreateAccessToken, res.Status, err)\n\t\t}\n\n\t\tbodyErr := fmt.Errorf(\"%w: %v\", errBody, string(bodyBytes))\n\n\t\treturn nil, fmt.Errorf(\"%w: %s: %v\", errUnableToCreateAccessToken,\n\t\t\tres.Status, bodyErr)\n\t}\n\n\tvar resData map[string]interface{}\n\tif err := json.NewDecoder(res.Body).Decode(&resData); err != nil {\n\t\treturn nil, fmt.Errorf(\"%w: %v\", errUnableToDecodeAccessTokenRes, err)\n\t}\n\n\ttokenRes := &logical.Response{Data: resData}\n\n\t// As per the issue request in https://git.io/JUhRk, return a Vault \"lease\"\n\t// aligned to the GitHub token's `expires_at` field.\n\tif expiresAt, ok := resData[\"expires_at\"]; ok {\n\t\tif expiresAtStr, ok := expiresAt.(string); ok {\n\t\t\tif expiresAtTime, err := time.Parse(time.RFC3339, expiresAtStr); err == nil {\n\t\t\t\ttokenRes.Secret = &logical.Secret{\n\t\t\t\t\tInternalData: map[string]interface{}{\"secret_type\": backendSecretType},\n\t\t\t\t\tLeaseOptions: logical.LeaseOptions{\n\t\t\t\t\t\tTTL: time.Until(expiresAtTime),\n\t\t\t\t\t},\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn tokenRes, nil\n}", "func (m *Manager) RefreshAccessToken(tgr *oauth2.TokenGenerateRequest) (accessToken oauth2.TokenInfo, err error) {\n\tcli, err := m.GetClient(tgr.ClientID)\n\tif err != nil {\n\t\treturn\n\t} else if tgr.ClientSecret != cli.GetSecret() {\n\t\terr = errors.ErrInvalidClient\n\t\treturn\n\t}\n\tti, err := m.LoadRefreshToken(tgr.Refresh)\n\tif err != nil {\n\t\treturn\n\t} else if ti.GetClientID() != tgr.ClientID {\n\t\terr = errors.ErrInvalidRefreshToken\n\t\treturn\n\t}\n\toldAccess := ti.GetAccess()\n\t_, ierr := m.injector.Invoke(func(stor oauth2.TokenStore, gen oauth2.AccessGenerate) {\n\t\ttd := &oauth2.GenerateBasic{\n\t\t\tClient: cli,\n\t\t\tUserID: ti.GetUserID(),\n\t\t\tCreateAt: time.Now(),\n\t\t}\n\t\tisGenRefresh := false\n\t\tif rcfg, ok := m.gtcfg[oauth2.Refreshing]; ok {\n\t\t\tisGenRefresh = rcfg.IsGenerateRefresh\n\t\t}\n\t\ttv, rv, terr := gen.Token(td, isGenRefresh)\n\t\tif terr != nil {\n\t\t\terr = terr\n\t\t\treturn\n\t\t}\n\t\tti.SetAccess(tv)\n\t\tti.SetAccessCreateAt(td.CreateAt)\n\t\tif scope := tgr.Scope; scope != \"\" {\n\t\t\tti.SetScope(scope)\n\t\t}\n\t\tif rv != \"\" {\n\t\t\tti.SetRefresh(rv)\n\t\t}\n\t\tif verr := stor.Create(ti); verr != nil {\n\t\t\terr = verr\n\t\t\treturn\n\t\t}\n\t\t// remove the old access token\n\t\tif verr := stor.RemoveByAccess(oldAccess); verr != nil {\n\t\t\terr = verr\n\t\t\treturn\n\t\t}\n\t\taccessToken = ti\n\t})\n\tif ierr != nil && err == nil {\n\t\terr = ierr\n\t}\n\treturn\n}", "func defConfig() Config {\n\treturn Config{\n\t\tAPI: api{\n\t\t\tServiceName: \"token-svc\",\n\t\t\tMetricsPort: \"4001\",\n\t\t\tPort: \"4000\",\n\t\t\tShutdownTimeoutSecs: 120,\n\t\t\tIdleTimeOutSecs: 90,\n\t\t\tWriteTimeOutSecs: 30,\n\t\t\tReadTimeOutSecs: 5,\n\t\t\tTimeoutSecs: 30,\n\t\t\tAllowedHeaders: []string{\"X-Requested-With\", \"X-Request-ID\", \"jaeger-debug-id\", \"Content-Type\", \"Authorization\"},\n\t\t\tAllowedOrigins: []string{\"*\"},\n\t\t\tAllowedMethods: []string{\"GET\", \"HEAD\", \"POST\", \"PUT\", \"OPTIONS\", \"DELETE\"},\n\t\t\tOpenEndPoints: []string{\"/login\", \"/health/ping\", \"/register\"},\n\t\t},\n\t\tLogger: logger{\n\t\t\tLevel: \"debug\",\n\t\t\tEncoding: \"json\",\n\t\t\tOutputPaths: []string{\"stdout\", \"/tmp/logs/tokensvc.logs\"},\n\t\t\tErrorOutputPaths: []string{\"stderr\"},\n\t\t},\n\t\tDB: db{\n\t\t\tUser: \"postgres\",\n\t\t\tPass: \"postgres\",\n\t\t\tHost: \"postgres\",\n\t\t\tPort: \"5432\",\n\t\t\tName: \"postgres\",\n\t\t\tTimeout: \"30\",\n\t\t},\n\t\tToken: token{\n\t\t\tAccessTokenLifeSpanMins: 30, // half hour\n\t\t\tRefreshTokenLifeSpanMins: 10080, // 1 week\n\t\t\tFailedLoginAttemptCacheLifeSpanMins: 30,\n\t\t\tFailedLoginAttemptsMax: 5,\n\t\t\tAuthPrivateKeyPath: \"/tmp/certs/app.rsa\", // TODO: Let's read these in from Vault\n\t\t\tAuthPublicKeyPath: \"/tmp/certs/app.rsa.pub\",\n\t\t\tIssuer: \"homerow.tech\",\n\t\t\tAccessCacheKeyID: \"token-access-user\",\n\t\t\tRefreshCacheKeyID: \"token-refresh-user\",\n\t\t\tFailedLoginCacheKeyID: \"failed-login-user\",\n\t\t},\n\t\tCookie: cookie{\n\t\t\tLifeSpanDays: 7,\n\t\t\tHashKey: \"something-that-is-32-byte-secret\",\n\t\t\tBlockKey: \"something-else-16-24-or-32secret\",\n\t\t\tName: \"homerow.tech\",\n\t\t\tDomain: \"dev.homerow.tech\",\n\t\t\tKeyUserID: \"id\",\n\t\t\tKeyEmail: \"email\",\n\t\t\tKeyJWTAccessID: \"jti-access\",\n\t\t\tKeyJWTRefreshID: \"jti-refresh\",\n\t\t},\n\t\tCache: cache{\n\t\t\tHost: \"redis\",\n\t\t\tPort: \"6379\",\n\t\t\tUserAccountLockedLifeSpanMins: 60,\n\t\t\tUserAccountLockedKeyID: \"account-locked-user\",\n\t\t},\n\t}\n}", "func (i *InternalTokenHelper) populateTokenPath() {\n\ti.tokenPath = filepath.Join(i.homeDir, \".vault-token\")\n}", "func (K *KWAPI) refreshToken(username string, auth *KWAuth) (*KWAuth, error) {\n\tif auth == nil {\n\t\treturn nil, fmt.Errorf(\"No refresh token found for %s.\", username)\n\t}\n\tpath := fmt.Sprintf(\"https://%s/oauth/token\", K.Server)\n\n\treq, err := http.NewRequest(http.MethodPost, path, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\thttp_header := make(http.Header)\n\thttp_header.Set(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\tif K.AgentString == NONE {\n\t\tK.AgentString = \"SnugLib/1.0\"\n\t}\n\thttp_header.Set(\"User-Agent\", K.AgentString)\n\n\treq.Header = http_header\n\n\tclient_id := K.ApplicationID\n\n\tpostform := &url.Values{\n\t\t\"client_id\": {client_id},\n\t\t\"client_secret\": {K.secrets.decrypt(K.secrets.client_secret_key)},\n\t\t\"grant_type\": {\"refresh_token\"},\n\t\t\"refresh_token\": {auth.RefreshToken},\n\t}\n\n\tif K.Snoop {\n\t\tStdout(\"\\n[kiteworks]: %s\\n--> ACTION: \\\"POST\\\" PATH: \\\"%s\\\"\", username, path)\n\t\tfor k, v := range *postform {\n\t\t\tif k == \"grant_type\" || k == \"RedirectURI\" || k == \"scope\" {\n\t\t\t\tStdout(\"\\\\-> POST PARAM: %s VALUE: %s\", k, v)\n\t\t\t} else {\n\t\t\t\tStdout(\"\\\\-> POST PARAM: %s VALUE: [HIDDEN]\", k)\n\t\t\t}\n\t\t}\n\t}\n\n\treq.Body = ioutil.NopCloser(bytes.NewReader([]byte(postform.Encode())))\n\n\tclient := K.Session(username).NewClient()\n\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := K.decodeJSON(resp, &auth); err != nil {\n\t\treturn nil, err\n\t}\n\n\tauth.Expires = auth.Expires + time.Now().Unix()\n\treturn auth, nil\n}", "func GToken() string {\n\treturn viper.GetString(\"google-safile\")\n}", "func (t *TokenSource) Token() (*oauth2.Token, error) {\n\treturn &oauth2.Token{\n\t\tAccessToken: t.AccessToken,\n\t}, nil\n}", "func (s *notifyRefreshTokenSource) Token() (*oauth2.Token, error) {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\tif s.t.Valid() {\n\t\treturn s.t, nil\n\t}\n\tt, err := s.new.Token()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ts.t = t\n\treturn t, s.f(t)\n}", "func (g *GitHub) getToken(code, state string) error {\n\tif code == \"\" || state == \"\" {\n\t\tlog.ErrorWithFields(\"code or state is nil\", log.Fields{\"code\": code, \"state\": state})\n\t\treturn fmt.Errorf(\"code or state is nil\")\n\t}\n\tlog.InfoWithFields(\"cyclone receives auth code\", log.Fields{\"request code\": code})\n\n\t// Get a object to request token.\n\tconf, err := g.getConf()\n\tif err != nil {\n\t\tlog.Warnf(\"Unable to get the conf according coderepository\")\n\t\treturn err\n\t}\n\n\t// To communication with githubo or other vcs to get token.\n\tvar tok *oauth2.Token\n\ttok, err = conf.Exchange(oauth2.NoContext, code) // Post a token request and receive toeken.\n\tif err != nil {\n\t\tlog.Error(err)\n\t\treturn err\n\t}\n\n\tif !tok.Valid() {\n\t\tlog.Fatalf(\"Token invalid. Got: %#v\", tok)\n\t\treturn err\n\t}\n\tlog.Info(\"get the token successfully!\")\n\n\t// Create service in database (but not ready to be used yet).\n\tvcstoken := api.VscToken{\n\t\tUserID: state,\n\t\tVsc: \"github\",\n\t\tVsctoken: *tok,\n\t}\n\n\tds := store.NewStore()\n\tdefer ds.Close()\n\n\t_, err = ds.FindtokenByUserID(state, \"github\")\n\tif err != nil {\n\t\terr = ds.NewTokenDocument(&vcstoken)\n\t\tif err != nil {\n\t\t\tlog.ErrorWithFields(\"NewTokenDocument\", log.Fields{\"user_id\": state,\n\t\t\t\t\"token\": tok, \"error\": err})\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\terr = ds.UpdateToken(&vcstoken)\n\t\tif err != nil {\n\t\t\tlog.ErrorWithFields(\"UpdateToken\", log.Fields{\"user_id\": state,\n\t\t\t\t\"token\": tok, \"error\": err})\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func ReuseTokenSource(t *Token, src TokenSource) TokenSource {\n\t// Don't wrap a reuseTokenSource in itself. That would work,\n\t// but cause an unnecessary number of mutex operations.\n\t// Just build the equivalent one.\n\tif rt, ok := src.(*reuseTokenSource); ok {\n\t\tif t == nil {\n\t\t\t// Just use it directly.\n\t\t\treturn rt\n\t\t}\n\t\tsrc = rt.new\n\t}\n\treturn &reuseTokenSource{\n\t\tt: t,\n\t\tnew: src,\n\t}\n}", "func (h *handler) initFromActionsEnv(ctx context.Context) {\n\ttoken := h.GetInput(\"repo_token\")\n\tif token == \"\" {\n\t\th.Fatalf(\"Empty repo_token\")\n\t}\n\th.Client = github.NewClient(oauth2.NewClient(ctx, oauth2.StaticTokenSource(\n\t\t&oauth2.Token{AccessToken: token},\n\t)))\n}", "func Initialize(ctx context.Context, global *Global) (err error) {\n\tlog.SetFlags(0)\n\tglobal.ctx = ctx\n\n\tvar instanceDeployment InstanceDeployment\n\tvar storageClient *storage.Client\n\n\tinitID := fmt.Sprintf(\"%v\", uuid.New())\n\terr = ffo.ReadUnmarshalYAML(solution.PathToFunctionCode+solution.SettingsFileName, &instanceDeployment)\n\tif err != nil {\n\t\tlog.Println(glo.Entry{\n\t\t\tSeverity: \"CRITICAL\",\n\t\t\tMessage: \"init_failed\",\n\t\t\tDescription: fmt.Sprintf(\"ReadUnmarshalYAML %s %v\", solution.SettingsFileName, err),\n\t\t\tInitID: initID,\n\t\t})\n\t\treturn err\n\t}\n\n\tglobal.environment = instanceDeployment.Core.EnvironmentName\n\tglobal.instanceName = instanceDeployment.Core.InstanceName\n\tglobal.microserviceName = instanceDeployment.Core.ServiceName\n\n\tlog.Println(glo.Entry{\n\t\tMicroserviceName: global.microserviceName,\n\t\tInstanceName: global.instanceName,\n\t\tEnvironment: global.environment,\n\t\tSeverity: \"NOTICE\",\n\t\tMessage: \"coldstart\",\n\t\tInitID: initID,\n\t})\n\n\tglobal.assetsCollectionID = instanceDeployment.Core.SolutionSettings.Hosting.FireStore.CollectionIDs.Assets\n\tglobal.ownerLabelKeyName = instanceDeployment.Core.SolutionSettings.Monitoring.LabelKeyNames.Owner\n\tglobal.retryTimeOutSeconds = instanceDeployment.Settings.Service.GCF.RetryTimeOutSeconds\n\tglobal.violationResolverLabelKeyName = instanceDeployment.Core.SolutionSettings.Monitoring.LabelKeyNames.ViolationResolver\n\tprojectID := instanceDeployment.Core.SolutionSettings.Hosting.ProjectID\n\n\tstorageClient, err = storage.NewClient(ctx)\n\tif err != nil {\n\t\tlog.Println(glo.Entry{\n\t\t\tMicroserviceName: global.microserviceName,\n\t\t\tInstanceName: global.instanceName,\n\t\t\tEnvironment: global.environment,\n\t\t\tSeverity: \"CRITICAL\",\n\t\t\tMessage: \"init_failed\",\n\t\t\tDescription: fmt.Sprintf(\"storage.NewClient(ctx) %v\", err),\n\t\t\tInitID: initID,\n\t\t})\n\t\treturn err\n\t}\n\t// bucketHandle must be evaluated after storateClient init\n\tglobal.bucketHandle = storageClient.Bucket(instanceDeployment.Core.SolutionSettings.Hosting.GCS.Buckets.AssetsJSONFile.Name)\n\n\tglobal.cloudresourcemanagerService, err = cloudresourcemanager.NewService(ctx)\n\tif err != nil {\n\t\tlog.Println(glo.Entry{\n\t\t\tMicroserviceName: global.microserviceName,\n\t\t\tInstanceName: global.instanceName,\n\t\t\tEnvironment: global.environment,\n\t\t\tSeverity: \"CRITICAL\",\n\t\t\tMessage: \"init_failed\",\n\t\t\tDescription: fmt.Sprintf(\"cloudresourcemanager.NewService(ctx) %v\", err),\n\t\t\tInitID: initID,\n\t\t})\n\t\treturn err\n\t}\n\tglobal.cloudresourcemanagerServiceV2, err = cloudresourcemanagerv2.NewService(ctx)\n\tif err != nil {\n\t\tlog.Println(glo.Entry{\n\t\t\tMicroserviceName: global.microserviceName,\n\t\t\tInstanceName: global.instanceName,\n\t\t\tEnvironment: global.environment,\n\t\t\tSeverity: \"CRITICAL\",\n\t\t\tMessage: \"init_failed\",\n\t\t\tDescription: fmt.Sprintf(\"cloudresourcemanagerv2.NewService(ctx) %v\", err),\n\t\t\tInitID: initID,\n\t\t})\n\t\treturn err\n\t}\n\tglobal.firestoreClient, err = firestore.NewClient(ctx, projectID)\n\tif err != nil {\n\t\tlog.Println(glo.Entry{\n\t\t\tMicroserviceName: global.microserviceName,\n\t\t\tInstanceName: global.instanceName,\n\t\t\tEnvironment: global.environment,\n\t\t\tSeverity: \"CRITICAL\",\n\t\t\tMessage: \"init_failed\",\n\t\t\tDescription: fmt.Sprintf(\"firestore.NewClient(ctx, projectID) %v\", err),\n\t\t\tInitID: initID,\n\t\t})\n\t\treturn err\n\t}\n\treturn nil\n}", "func init() {\n\n\tflag.StringVar(&Token, \"t\", \"\", \"Bot Token\")\n}", "func Initialize(client kubernetes.Interface, namespace, serviceAccount string, imagePullSecrets []string) error {\n\tvar kc authn.Keychain\n\tkcOpts := &k8schain.Options{\n\t\tNamespace: namespace,\n\t\tServiceAccountName: serviceAccount,\n\t\tImagePullSecrets: imagePullSecrets,\n\t}\n\n\tkc, err := k8schain.New(context.Background(), client, *kcOpts)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to initialize registry keychain\")\n\t}\n\n\tauthn.DefaultKeychain = kc\n\treturn nil\n}", "func (c *Easee) refresh() {\n\tfor range time.Tick(5 * time.Minute) {\n\t\tif _, err := c.Client.Transport.(*oauth2.Transport).Source.Token(); err != nil {\n\t\t\tc.log.ERROR.Println(\"token refresh:\", err)\n\t\t}\n\t}\n}", "func (c *Client) Token(token string) *Client {\n\tc.api.Set(\"Authorization\", \"Discogs token=\"+token)\n\treturn c\n}", "func (w TokenFileWriter) CreateMgmtTokenForConsulSecretsEngine(rootToken string) (map[string]interface{},\n\ttokencreatable.RevokeFunc, error) {\n\tconsulSecretsEngineOpsPolicyDocument := `\n# allow to configure the access information for Consul\npath \"` + secretsengine.ConsulSecretEngineMountPoint + `/config/access\" {\n capabilities = [\"create\", \"update\"]\n}\n\n# allow to create, update, read, list, or delete the Consul role definition\npath \"` + secretsengine.ConsulSecretEngineMountPoint + `/roles/*\" {\n capabilities = [\"create\", \"read\", \"update\", \"delete\", \"list\"]\n}\n`\n\n\tif err := w.secretClient.InstallPolicy(rootToken,\n\t\tconsulSecretsEngineOpsPolicyName,\n\t\tconsulSecretsEngineOpsPolicyDocument); err != nil {\n\t\treturn nil, nil, fmt.Errorf(\"failed to install Consul secrets engine operations policy: %v\", err)\n\t}\n\n\t// setup new token's properties\n\ttokenParams := make(map[string]interface{})\n\ttokenParams[\"type\"] = \"service\"\n\t// Vault prefixes \"token\" in front of display_name\n\ttokenParams[\"display_name\"] = \"for Consul ACL bootstrap\"\n\ttokenParams[\"no_parent\"] = true\n\ttokenParams[\"period\"] = \"1h\"\n\ttokenParams[\"policies\"] = []string{consulSecretsEngineOpsPolicyName}\n\ttokenParams[\"meta\"] = map[string]interface{}{\n\t\t\"description\": \"Consul secrets engine management token\",\n\t}\n\tresponse, err := w.secretClient.CreateToken(rootToken, tokenParams)\n\tif err != nil {\n\t\treturn nil, nil, fmt.Errorf(\"failed to create token for Consul secrets engine operations: %v\", err)\n\t}\n\n\treturn response, nil, nil\n}", "func (c *TokensCommand) Initialize(app *kingpin.Application, config *servicecfg.Config) {\n\tc.config = config\n\n\ttokens := app.Command(\"tokens\", \"List or revoke invitation tokens\")\n\n\tformats := []string{teleport.Text, teleport.JSON, teleport.YAML}\n\n\t// tctl tokens add ...\"\n\tc.tokenAdd = tokens.Command(\"add\", \"Create a invitation token.\")\n\tc.tokenAdd.Flag(\"type\", \"Type(s) of token to add, e.g. --type=node,app,db,proxy,etc\").Required().StringVar(&c.tokenType)\n\tc.tokenAdd.Flag(\"value\", \"Override the default random generated token with a specified value\").StringVar(&c.value)\n\tc.tokenAdd.Flag(\"labels\", \"Set token labels, e.g. env=prod,region=us-west\").StringVar(&c.labels)\n\tc.tokenAdd.Flag(\"ttl\", fmt.Sprintf(\"Set expiration time for token, default is %v minutes\",\n\t\tint(defaults.ProvisioningTokenTTL/time.Minute))).\n\t\tDefault(fmt.Sprintf(\"%v\", defaults.ProvisioningTokenTTL)).\n\t\tDurationVar(&c.ttl)\n\tc.tokenAdd.Flag(\"app-name\", \"Name of the application to add\").Default(\"example-app\").StringVar(&c.appName)\n\tc.tokenAdd.Flag(\"app-uri\", \"URI of the application to add\").Default(\"http://localhost:8080\").StringVar(&c.appURI)\n\tc.tokenAdd.Flag(\"db-name\", \"Name of the database to add\").StringVar(&c.dbName)\n\tc.tokenAdd.Flag(\"db-protocol\", fmt.Sprintf(\"Database protocol to use. Supported are: %v\", defaults.DatabaseProtocols)).StringVar(&c.dbProtocol)\n\tc.tokenAdd.Flag(\"db-uri\", \"Address the database is reachable at\").StringVar(&c.dbURI)\n\tc.tokenAdd.Flag(\"format\", \"Output format, 'text', 'json', or 'yaml'\").EnumVar(&c.format, formats...)\n\n\t// \"tctl tokens rm ...\"\n\tc.tokenDel = tokens.Command(\"rm\", \"Delete/revoke an invitation token.\").Alias(\"del\")\n\tc.tokenDel.Arg(\"token\", \"Token to delete\").StringVar(&c.value)\n\n\t// \"tctl tokens ls\"\n\tc.tokenList = tokens.Command(\"ls\", \"List node and user invitation tokens.\")\n\tc.tokenList.Flag(\"format\", \"Output format, 'text', 'json' or 'yaml'\").EnumVar(&c.format, formats...)\n\n\tif c.stdout == nil {\n\t\tc.stdout = os.Stdout\n\t}\n}", "func GenerateK8sApiFromToken(host string, caPath string, token string) (*K8sApi, error) {\n\tconfig := &rest.Config{\n\t\tHost: host,\n\t\tBearerToken: token,\n\t\tTLSClientConfig: rest.TLSClientConfig{\n\t\t\tCAFile: caPath,\n\t\t},\n\t}\n\tclientSet, dif, err := GetClientSet(config)\n\tif err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tapi := &K8sApi{\n\t\t\tClientSet: clientSet,\n\t\t\tDif: dif,\n\t\t}\n\t\tif rbacClient, err := rbac.NewForConfig(config); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tapi.RbacClient = rbacClient\n\t\t\tif v1beta1Client, err := v1beta1.NewForConfig(config); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t} else {\n\t\t\t\tapi.V1beta1Client = v1beta1Client\n\t\t\t\treturn api, nil\n\t\t\t}\n\t\t}\n\t}\n}" ]
[ "0.6248717", "0.59793735", "0.5819379", "0.5771171", "0.5746096", "0.57363856", "0.5698077", "0.56682885", "0.56210387", "0.560057", "0.55441684", "0.55334824", "0.5487145", "0.5487065", "0.545568", "0.5444413", "0.540357", "0.536407", "0.5363666", "0.5356209", "0.5340606", "0.53262603", "0.52807415", "0.52508456", "0.5244249", "0.5240652", "0.52141625", "0.51998115", "0.5182218", "0.5173219", "0.51634437", "0.516319", "0.5157955", "0.5155673", "0.51468414", "0.5136258", "0.5109961", "0.51060593", "0.5092437", "0.5090331", "0.5090331", "0.5090331", "0.5088441", "0.506864", "0.50550437", "0.50426227", "0.50342405", "0.5033609", "0.5029531", "0.5014899", "0.50067204", "0.50059974", "0.5000696", "0.49995935", "0.49966198", "0.49901855", "0.4989285", "0.4982872", "0.49827713", "0.49794254", "0.497637", "0.4967891", "0.4965149", "0.4929004", "0.49240944", "0.4920136", "0.4915101", "0.491126", "0.49044043", "0.49040434", "0.48862103", "0.4886109", "0.48533753", "0.4844106", "0.4841585", "0.4827553", "0.48109105", "0.48109105", "0.4792825", "0.47921476", "0.47789103", "0.47735044", "0.47693276", "0.47601682", "0.47548503", "0.47488436", "0.47413442", "0.47169647", "0.47135827", "0.4713057", "0.47113636", "0.4700624", "0.4699288", "0.46950927", "0.46909702", "0.4690951", "0.46855864", "0.46796447", "0.46794888", "0.46785042" ]
0.7220189
0
TokenSource create oauth2.TokenSource for Credentials. Note: We can specify scopes needed for spannerautoscaler but it does increase maintenance cost. We should already use least privileged Google Service Accounts so it use cloudPlatformScope.
func (c *Credentials) TokenSource(ctx context.Context) (oauth2.TokenSource, error) { switch c.Type { case CredentialsTypeADC: return initializedBaseTokenSource() case CredentialsTypeServiceAccountJSON: cred, err := google.CredentialsFromJSON(ctx, c.ServiceAccountJSON, cloudPlatformScope) if err != nil { return nil, err } return cred.TokenSource, nil case CredentialsTypeImpersonation: baseTS, err := initializedBaseTokenSource() if err != nil { return nil, err } ts, err := impersonate.CredentialsTokenSource(ctx, impersonate.CredentialsConfig{ TargetPrincipal: c.ImpersonateConfig.TargetServiceAccount, Delegates: c.ImpersonateConfig.Delegates, Scopes: []string{cloudPlatformScope}, }, option.WithTokenSource(baseTS), ) return ts, err default: return nil, fmt.Errorf("credentials type unknown: %v", c.Type) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func tokenSource(ctx context.Context, accessToken, credentials string) (oauth2.TokenSource, error) {\n\t// Try access token first\n\tif accessToken != \"\" {\n\t\tlog.Printf(\"[INFO] authenticating via access_token\")\n\n\t\tcontents, _, err := pathorcontents.Read(accessToken)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to load access token: %w\", err)\n\t\t}\n\n\t\treturn oauth2.StaticTokenSource(&oauth2.Token{\n\t\t\tAccessToken: contents,\n\t\t}), nil\n\t}\n\n\t// Then credentials\n\tif credentials != \"\" {\n\t\tlog.Printf(\"[INFO] authenticating via credentials\")\n\n\t\tcontents, _, err := pathorcontents.Read(credentials)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to load credentials: %w\", err)\n\t\t}\n\n\t\tcreds, err := google.CredentialsFromJSON(ctx, []byte(contents), cloudPlatformScope)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to parse credentials: %w\", err)\n\t\t}\n\n\t\treturn creds.TokenSource, nil\n\t}\n\n\t// Fallback to default credentials\n\tlog.Printf(\"[INFO] authenticating via default credentials\")\n\tsource, err := google.DefaultTokenSource(ctx, cloudPlatformScope)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get default credentials: %w\", err)\n\t}\n\treturn source, nil\n}", "func CredentialsTokenSource(creds *google.Credentials) TokenSource {\n\tif creds == nil {\n\t\treturn nil\n\t}\n\treturn TokenSource(creds.TokenSource)\n}", "func tokenSource(ctx context.Context) (oauth2.TokenSource, error) {\n\tok, err := credsFile.Exists()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar bootstrapToken *oauth2.Token\n\tif !ok {\n\t\ttok, err := authenticate()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tbootstrapToken = tok\n\t}\n\treturn newCachedTokenFile(ctx, bootstrapToken, credsFile.Path())\n}", "func StorageTokenSource(ctx context.Context, c *Config, t *oauth2.Token) oauth2.TokenSource {\n\tif t == nil || !t.Valid() {\n\t\tif tok, err := c.Storage.GetToken(); err == nil {\n\t\t\tt = tok\n\t\t}\n\t}\n\tts := c.Config.TokenSource(ctx, t)\n\treturn &storageTokenSource{c, ts}\n}", "func NewTokenSource(ctx context.Context, conf DownscopingConfig) (oauth2.TokenSource, error) {\n\tif conf.RootSource == nil {\n\t\treturn nil, fmt.Errorf(\"downscope: rootSource cannot be nil\")\n\t}\n\tif len(conf.Rules) == 0 {\n\t\treturn nil, fmt.Errorf(\"downscope: length of AccessBoundaryRules must be at least 1\")\n\t}\n\tif len(conf.Rules) > 10 {\n\t\treturn nil, fmt.Errorf(\"downscope: length of AccessBoundaryRules may not be greater than 10\")\n\t}\n\tfor _, val := range conf.Rules {\n\t\tif val.AvailableResource == \"\" {\n\t\t\treturn nil, fmt.Errorf(\"downscope: all rules must have a nonempty AvailableResource: %+v\", val)\n\t\t}\n\t\tif len(val.AvailablePermissions) == 0 {\n\t\t\treturn nil, fmt.Errorf(\"downscope: all rules must provide at least one permission: %+v\", val)\n\t\t}\n\t}\n\treturn downscopingTokenSource{ctx: ctx, config: conf}, nil\n}", "func TokenSource(ctx context.Context, config *oauth2.Config, state string, authHandler AuthorizationHandler) oauth2.TokenSource {\n\treturn TokenSourceWithPKCE(ctx, config, state, authHandler, nil)\n}", "func (gsp *GoogleServiceProvider) TokenSource(c context.Context, scopes ...string) (oauth2.TokenSource, error) {\n\tcbts := contextBoundTokenSource{\n\t\tContext: c,\n\t\tcache: gsp.Cache,\n\t\tcacheKey: accessTokenKeyForScopes(scopes),\n\t\tmakeTokenSource: func(c context.Context) (oauth2.TokenSource, error) {\n\t\t\treturn google.DefaultTokenSource(c, scopes...)\n\t\t},\n\t}\n\treturn &cbts, nil\n}", "func OIDCFederatedTokenSource(tokenConfig *OIDCFederatedTokenConfig) (oauth2.TokenSource, error) {\n\n\tif &tokenConfig.SourceTokenSource == nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: Source OIDC Token cannot be nil\")\n\t}\n\n\tif tokenConfig.Scope == \"\" {\n\t\ttokenConfig.Scope = GCP_OIDC_CLOUD_PLATFORM_SCOPE\n\t}\n\treturn &oidcFederatedTokenSource{\n\t\trefreshMutex: &sync.Mutex{},\n\t\trootSource: tokenConfig.SourceTokenSource,\n\t\tscope: tokenConfig.Scope,\n\t\ttargetResource: tokenConfig.TargetResource,\n\t\ttargetServiceAccount: tokenConfig.TargetServiceAccount,\n\t\tuseIAMToken: tokenConfig.UseIAMToken,\n\t}, nil\n}", "func TokenSource(ctx context.Context, scopes ...string) oauth2.TokenSource {\n\treturn TokenSourceEnv(ctx, envPrivateKey, scopes...)\n}", "func TokenSource(aud string) oauth2.TokenSource {\n\tidSrc := idTokenSrc{aud: aud}\n\tinitialToken := &oauth2.Token{}\n\treturn oauth2.ReuseTokenSource(initialToken, idSrc)\n}", "func NewTokenSource(name string, path string) oauth2.TokenSource {\n\treturn &tokenSource{\n\t\tname: name,\n\t\tpath: path,\n\t}\n}", "func newTokenSource(ctx context.Context, settings *Settings) (*internal.TokenSource, error) {\n\tvar ts internal.TokenSource\n\tvar err error\n\tif settings == nil {\n\t\tts, err = DefaultTokenSource(ctx, DefaultScope)\n\t} else if settings.APIKey != \"\" {\n\t\treturn nil, nil\n\t} else if settings.Scope != \"\" {\n\t\tts, err = OAuthJSONTokenSource(ctx, settings)\n\t} else {\n\t\tts, err = JWTTokenSource(ctx, settings)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &ts, err\n}", "func (c *Config) TokenSource(ctx context.Context) (oauth2.TokenSource, error) {\n\treturn c.tokenSource(ctx, \"https\")\n}", "func (c *Config) TokenSource() oauth2.TokenSource {\n\tpk, err := jwt.ParseRSAPrivateKeyFromPEM(c.PrivateKey)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn oauth2.ReuseTokenSource(nil, appSource{\n\t\tappID: c.AppID,\n\t\texpire: c.Expire,\n\t\tpk: pk,\n\t})\n}", "func (c *Config) tokenSource(ctx context.Context, scheme string) (oauth2.TokenSource, error) {\n\tif c.WorkforcePoolUserProject != \"\" {\n\t\tvalid := validateWorkforceAudience(c.Audience)\n\t\tif !valid {\n\t\t\treturn nil, fmt.Errorf(\"oauth2/google: workforce_pool_user_project should not be set for non-workforce pool credentials\")\n\t\t}\n\t}\n\n\tts := tokenSource{\n\t\tctx: ctx,\n\t\tconf: c,\n\t}\n\tif c.ServiceAccountImpersonationURL == \"\" {\n\t\treturn oauth2.ReuseTokenSource(nil, ts), nil\n\t}\n\tscopes := c.Scopes\n\tts.conf.Scopes = []string{\"https://www.googleapis.com/auth/cloud-platform\"}\n\timp := ImpersonateTokenSource{\n\t\tCtx: ctx,\n\t\tURL: c.ServiceAccountImpersonationURL,\n\t\tScopes: scopes,\n\t\tTs: oauth2.ReuseTokenSource(nil, ts),\n\t\tTokenLifetimeSeconds: c.ServiceAccountImpersonationLifetimeSeconds,\n\t}\n\treturn oauth2.ReuseTokenSource(nil, imp), nil\n}", "func newTokenSource() *tokenReplacer {\n\t// nil token will cause a refresh\n\ttok, _ := readToken()\n\treturn &tokenReplacer{tok, oauthConfig.TokenSource(context.Background(), tok), &tokenPrompt{}}\n}", "func NewIAMTokenSource(ctx context.Context, cfg IAMConfig) (oauth2.TokenSource, error) {\n\tvar (\n\t\terr error\n\t\ttknSrc oauth2.TokenSource\n\t)\n\tif cfg.JSON != nil {\n\t\tcreds, err := google.CredentialsFromJSON(ctx, cfg.JSON, iam.CloudPlatformScope)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ttknSrc = creds.TokenSource\n\t} else {\n\t\ttknSrc, err = defaultTokenSource(ctx, iam.CloudPlatformScope)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsvc, err := iam.New(oauth2.NewClient(ctx, tknSrc))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif cfg.IAMAddress != \"\" {\n\t\tsvc.BasePath = cfg.IAMAddress\n\t}\n\n\tsrc := &iamTokenSource{\n\t\tcfg: cfg,\n\t\tsvc: svc,\n\t}\n\n\ttkn, err := src.Token()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"unable to create initial token\")\n\t}\n\n\treturn oauth2.ReuseTokenSource(tkn, src), nil\n}", "func (w *GCPAuthWrapper) SetTokenSource(permissionCode string) error {\n\tvar err error\n\n\tctx := context.Background()\n\n\tw.OauthToken, err = w.Config.Exchange(ctx, permissionCode)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func appEngineTokenSource(ctx context.Context, scope ...string) oauth2.TokenSource {\n\tscopes := append([]string{}, scope...)\n\tsort.Strings(scopes)\n\treturn &gaeTokenSource{\n\t\tctx: ctx,\n\t\tscopes: scopes,\n\t\tkey: strings.Join(scopes, \" \"),\n\t}\n}", "func New(name string, configPath string, userName string) (tokenauth.Source, error) {\n\tif configPath == \"\" {\n\t\tconfigPath = k8s.DefaultKubeConfigPath\n\t}\n\tk8sConfig, err := cfg.LoadFromFile(configPath)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"Failed to load k8s config from file %v. Make sure it is there or change\"+\n\t\t\t\" permissions.\", configPath)\n\t}\n\n\tinfo, ok := k8sConfig.AuthInfos[userName]\n\tif !ok {\n\t\treturn nil, errors.Errorf(\"Failed to find user %s inside k8s config AuthInfo from file %v\", userName, configPath)\n\t}\n\n\t// Currently supported:\n\t// - token\n\t// - OIDC\n\t// - Google compute platform via Oauth2\n\tif info.AuthProvider != nil {\n\t\tswitch info.AuthProvider.Name {\n\t\tcase \"oidc\":\n\t\t\tcache, err := k8s.NewCacheFromUser(configPath, userName)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrap(err, \"Failed to get OIDC configuration from user. \")\n\t\t\t}\n\t\t\ts, _, err := oidcauth.NewWithCache(name, cache, nil)\n\t\t\treturn s, err\n\t\tcase \"gcp\":\n\t\t\tc, err := oauth2auth.NewConfigFromMap(info.AuthProvider.Config)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrap(err, \"Failed to create OAuth2 config from map.\")\n\t\t\t}\n\t\t\treturn oauth2auth.NewGCP(name, userName, configPath, c)\n\t\tdefault:\n\t\t\t// TODO(bplotka): Add support for more of them if needed.\n\t\t\treturn nil, errors.Errorf(\"Not supported k8s Auth provider %v\", info.AuthProvider.Name)\n\t\t}\n\t}\n\n\tif info.Token != \"\" {\n\t\treturn directauth.New(name, info.Token), nil\n\t}\n\n\treturn nil, errors.Errorf(\"Not found supported auth source from k8s config %+v\", info)\n}", "func (c *Config) TokenSource(ctx context.Context, t *Token) TokenSource {\n\ttkr := &tokenRefresher{\n\t\tctx: ctx,\n\t\tconf: c,\n\t}\n\tif t != nil {\n\t\ttkr.refreshToken = t.RefreshToken\n\t}\n\treturn &reuseTokenSource{\n\t\tt: t,\n\t\tnew: tkr,\n\t}\n}", "func initializedBaseTokenSource() (oauth2.TokenSource, error) {\n\tbaseTokenSourceOnce.Do(func() {\n\t\tbaseTokenSource, baseTokenSourceErr = google.DefaultTokenSource(context.Background(), cloudPlatformScope)\n\t})\n\treturn baseTokenSource, baseTokenSourceErr\n}", "func (ts tokenSource) Token() (*oauth2.Token, error) {\n\tconf := ts.conf\n\n\tcredSource, err := conf.parse(ts.ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsubjectToken, err := credSource.subjectToken()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tstsRequest := stsTokenExchangeRequest{\n\t\tGrantType: \"urn:ietf:params:oauth:grant-type:token-exchange\",\n\t\tAudience: conf.Audience,\n\t\tScope: conf.Scopes,\n\t\tRequestedTokenType: \"urn:ietf:params:oauth:token-type:access_token\",\n\t\tSubjectToken: subjectToken,\n\t\tSubjectTokenType: conf.SubjectTokenType,\n\t}\n\theader := make(http.Header)\n\theader.Add(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\tclientAuth := clientAuthentication{\n\t\tAuthStyle: oauth2.AuthStyleInHeader,\n\t\tClientID: conf.ClientID,\n\t\tClientSecret: conf.ClientSecret,\n\t}\n\tvar options map[string]interface{}\n\t// Do not pass workforce_pool_user_project when client authentication is used.\n\t// The client ID is sufficient for determining the user project.\n\tif conf.WorkforcePoolUserProject != \"\" && conf.ClientID == \"\" {\n\t\toptions = map[string]interface{}{\n\t\t\t\"userProject\": conf.WorkforcePoolUserProject,\n\t\t}\n\t}\n\tstsResp, err := exchangeToken(ts.ctx, conf.TokenURL, &stsRequest, clientAuth, header, options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\taccessToken := &oauth2.Token{\n\t\tAccessToken: stsResp.AccessToken,\n\t\tTokenType: stsResp.TokenType,\n\t}\n\tif stsResp.ExpiresIn < 0 {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: got invalid expiry from security token service\")\n\t} else if stsResp.ExpiresIn >= 0 {\n\t\taccessToken.Expiry = now().Add(time.Duration(stsResp.ExpiresIn) * time.Second)\n\t}\n\n\tif stsResp.RefreshToken != \"\" {\n\t\taccessToken.RefreshToken = stsResp.RefreshToken\n\t}\n\treturn accessToken, nil\n}", "func (a *auth) GetTokenSource(ctx context.Context, ref reference.Named) (oauth2.TokenSource, error) {\n\tsource, err := google.DefaultTokenSource(ctx, gcrScope)\n\tif nil != err {\n\t\terr = fmt.Errorf(\"failed to get Google Auth token source: %s\", err)\n\t}\n\n\treturn source, err\n\n}", "func (b *BungieConfig) TokenSource(t *oauth2.Token) oauth2.TokenSource {\n\treturn newTokenSource(t, b.cfg.ClientID)\n}", "func IDTokenSource(ctx context.Context, config IDTokenConfig, opts ...option.ClientOption) (oauth2.TokenSource, error) {\n\tif config.Audience == \"\" {\n\t\treturn nil, fmt.Errorf(\"impersonate: an audience must be provided\")\n\t}\n\tif config.TargetPrincipal == \"\" {\n\t\treturn nil, fmt.Errorf(\"impersonate: a target service account must be provided\")\n\t}\n\n\tclientOpts := append(defaultClientOptions(), opts...)\n\tclient, _, err := htransport.NewClient(ctx, clientOpts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tits := impersonatedIDTokenSource{\n\t\tclient: client,\n\t\ttargetPrincipal: config.TargetPrincipal,\n\t\taudience: config.Audience,\n\t\tincludeEmail: config.IncludeEmail,\n\t}\n\tfor _, v := range config.Delegates {\n\t\tits.delegates = append(its.delegates, formatIAMServiceAccountName(v))\n\t}\n\treturn oauth2.ReuseTokenSource(nil, its), nil\n}", "func WithTokenSource(ctx context.Context, ts oauth2.TokenSource) context.Context {\n\treturn context.WithValue(ctx, contextKey, ts)\n}", "func JWTAccessTokenSource(ctx context.Context, config *gcpjwt.IAMConfig, audience string) (oauth2.TokenSource, error) {\n\tctx = gcpjwt.NewIAMContext(ctx, config)\n\tts := &jwtAccessTokenSource{\n\t\tctx: ctx,\n\t\taudience: audience,\n\t\tjwtConfig: config,\n\t}\n\ttok, err := ts.Token()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn oauth2.ReuseTokenSource(tok, ts), nil\n}", "func (h *WebFlowHelper) GetTokenSource(ctx context.Context) (oauth2.TokenSource, error) {\n\tauthURL := h.config.AuthCodeURL(\"state-token\", oauth2.AccessTypeOffline)\n\n\t// TODO(jlewi): How to open it automatically?\n\tfmt.Printf(\"Go to the following link in your browser then type the \"+\n\t\t\"authorization code: \\n%v\\n\", authURL)\n\n\tvar authCode string\n\tif _, err := fmt.Scan(&authCode); err != nil {\n\t\treturn nil, errors.Wrapf(err, \"Unable to read authorization code\")\n\t}\n\n\ttok, err := h.config.Exchange(context.TODO(), authCode)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"Unable to retrieve token from web: %v\")\n\t}\n\n\treturn h.config.TokenSource(ctx, tok), nil\n}", "func (dts downscopingTokenSource) Token() (*oauth2.Token, error) {\n\n\tdownscopedOptions := struct {\n\t\tBoundary accessBoundary `json:\"accessBoundary\"`\n\t}{\n\t\tBoundary: accessBoundary{\n\t\t\tAccessBoundaryRules: dts.config.Rules,\n\t\t},\n\t}\n\n\ttok, err := dts.config.RootSource.Token()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"downscope: unable to obtain root token: %v\", err)\n\t}\n\n\tb, err := json.Marshal(downscopedOptions)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"downscope: unable to marshal AccessBoundary payload %v\", err)\n\t}\n\n\tform := url.Values{}\n\tform.Add(\"grant_type\", \"urn:ietf:params:oauth:grant-type:token-exchange\")\n\tform.Add(\"subject_token_type\", \"urn:ietf:params:oauth:token-type:access_token\")\n\tform.Add(\"requested_token_type\", \"urn:ietf:params:oauth:token-type:access_token\")\n\tform.Add(\"subject_token\", tok.AccessToken)\n\tform.Add(\"options\", string(b))\n\n\tmyClient := oauth2.NewClient(dts.ctx, nil)\n\tresp, err := myClient.PostForm(identityBindingEndpoint, form)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to generate POST Request %v\", err)\n\t}\n\tdefer resp.Body.Close()\n\trespBody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"downscope: unable to read response body: %v\", err)\n\t}\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn nil, fmt.Errorf(\"downscope: unable to exchange token; %v. Server responded: %s\", resp.StatusCode, respBody)\n\t}\n\n\tvar tresp downscopedTokenResponse\n\n\terr = json.Unmarshal(respBody, &tresp)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"downscope: unable to unmarshal response body: %v\", err)\n\t}\n\n\t// an exchanged token that is derived from a service account (2LO) has an expired_in value\n\t// a token derived from a users token (3LO) does not.\n\t// The following code uses the time remaining on rootToken for a user as the value for the\n\t// derived token's lifetime\n\tvar expiryTime time.Time\n\tif tresp.ExpiresIn > 0 {\n\t\texpiryTime = time.Now().Add(time.Duration(tresp.ExpiresIn) * time.Second)\n\t} else {\n\t\texpiryTime = tok.Expiry\n\t}\n\n\tnewToken := &oauth2.Token{\n\t\tAccessToken: tresp.AccessToken,\n\t\tTokenType: tresp.TokenType,\n\t\tExpiry: expiryTime,\n\t}\n\treturn newToken, nil\n}", "func sourceFromDefault(ctx context.Context, targetAudience string, tokenURL string) (*oidcTokenSource, error) {\n\tcredentials, err := google.FindDefaultCredentials(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif len(credentials.JSON) == 0 {\n\t\treturn nil, ErrComputeEngineNotSupported\n\t}\n\tconfig, err := google.JWTConfigFromJSON(credentials.JSON)\n\tif err != nil {\n\t\t// friendly error message that we found user credentials\n\t\tif strings.Contains(err.Error(), \"authorized_user\") {\n\t\t\treturn nil, ErrUserCredentialsNotSupported\n\t\t}\n\t\treturn nil, err\n\t}\n\tprivateKey, err := parseKey(config.PrivateKey)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &oidcTokenSource{config.Email, config.PrivateKeyID, privateKey, targetAudience, tokenURL}, nil\n}", "func GCPTokenSource(ts gcp.TokenSource) broker.Option {\n\treturn optfunc(gcpTokenSourceKey{}, ts)\n}", "func (its ImpersonateTokenSource) Token() (*oauth2.Token, error) {\n\tlifetimeString := \"3600s\"\n\tif its.TokenLifetimeSeconds != 0 {\n\t\tlifetimeString = fmt.Sprintf(\"%ds\", its.TokenLifetimeSeconds)\n\t}\n\treqBody := generateAccessTokenReq{\n\t\tLifetime: lifetimeString,\n\t\tScope: its.Scopes,\n\t\tDelegates: its.Delegates,\n\t}\n\tb, err := json.Marshal(reqBody)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: unable to marshal request: %v\", err)\n\t}\n\tclient := oauth2.NewClient(its.Ctx, its.Ts)\n\treq, err := http.NewRequest(\"POST\", its.URL, bytes.NewReader(b))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: unable to create impersonation request: %v\", err)\n\t}\n\treq = req.WithContext(its.Ctx)\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: unable to generate access token: %v\", err)\n\t}\n\tdefer resp.Body.Close()\n\tbody, err := ioutil.ReadAll(io.LimitReader(resp.Body, 1<<20))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: unable to read body: %v\", err)\n\t}\n\tif c := resp.StatusCode; c < 200 || c > 299 {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: status code %d: %s\", c, body)\n\t}\n\n\tvar accessTokenResp impersonateTokenResponse\n\tif err := json.Unmarshal(body, &accessTokenResp); err != nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: unable to parse response: %v\", err)\n\t}\n\texpiry, err := time.Parse(time.RFC3339, accessTokenResp.ExpireTime)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"oauth2/google: unable to parse expiry: %v\", err)\n\t}\n\treturn &oauth2.Token{\n\t\tAccessToken: accessTokenResp.AccessToken,\n\t\tExpiry: expiry,\n\t\tTokenType: \"Bearer\",\n\t}, nil\n}", "func newProxyTokenSource(\n\tctx context.Context,\n\tendpoint string,\n\treuseTokenFromUrl bool,\n) (ts oauth2.TokenSource, err error) {\n\tu, err := url.Parse(endpoint)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"newProxyTokenSource cannot parse endpoint %s: %w\", endpoint, err)\n\t\treturn\n\t}\n\n\tclient := &http.Client{}\n\tif u.Scheme == \"unix\" {\n\t\tclient.Transport = &http.Transport{\n\t\t\tDialContext: func(ctx context.Context, _, _ string) (net.Conn, error) {\n\t\t\t\tdialer := net.Dialer{}\n\t\t\t\treturn dialer.DialContext(ctx, u.Scheme, u.Path)\n\t\t\t},\n\t\t}\n\t\tendpoint = \"http://unix?\" + u.RawQuery\n\t}\n\n\tts = proxyTokenSource{\n\t\tctx: ctx,\n\t\tendpoint: endpoint,\n\t\tclient: client,\n\t}\n\tif reuseTokenFromUrl {\n\t\treturn oauth2.ReuseTokenSource(nil, ts), nil\n\t}\n\treturn ts, nil\n}", "func TokenSourceEnv(ctx context.Context, envVar string, scopes ...string) oauth2.TokenSource {\n\tif impKey := os.Getenv(envImpersonate); impKey == \"true\" {\n\t\treturn impersonatedTokenSource(ctx, scopes)\n\t}\n\tkey := os.Getenv(envVar)\n\tif key == \"\" { // Try for application default credentials.\n\t\tts, err := google.DefaultTokenSource(ctx, scopes...)\n\t\tif err != nil {\n\t\t\tlog.Println(\"No 'Application Default Credentials' found.\")\n\t\t\treturn nil\n\t\t}\n\t\treturn ts\n\t}\n\tconf, err := jwtConfigFromFile(key, scopes)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn conf.TokenSource(ctx)\n}", "func NewContextIAMTokenSource(ctx context.Context, cfg IAMConfig) (ContextTokenSource, error) {\n\tsrc := &iamTokenSource{cfg: cfg}\n\n\ttkn, err := src.ContextToken(ctx)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"unable to create initial token\")\n\t}\n\n\treturn &reuseTokenSource{t: tkn, new: src}, nil\n}", "func appEngineTokenSource(ctx context.Context, scope ...string) oauth2.TokenSource {\n\tlogOnce.Do(func() {\n\t\tlog.Print(\"google: AppEngineTokenSource is deprecated on App Engine standard second generation runtimes (>= Go 1.11) and App Engine flexible. Please use DefaultTokenSource or ComputeTokenSource.\")\n\t})\n\treturn ComputeTokenSource(\"\")\n}", "func NewStkSource() (StkSource, error) {\n\tsecret := make([]byte, 32)\n\tif _, err := rand.Read(secret); err != nil {\n\t\treturn nil, err\n\t}\n\tkey, err := deriveKey(secret)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tc, err := aes.NewCipher(key)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\taead, err := cipher.NewGCMWithNonceSize(c, stkNonceSize)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &stkSource{aead: aead}, nil\n}", "func (src *gcloudTokenSource) Token() (*oauth2.Token, error) {\n\tcfg, err := GcloudConfig()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn cfg.oauthToken(), nil\n}", "func TokenSourceWithPKCE(ctx context.Context, config *oauth2.Config, state string, authHandler AuthorizationHandler, pkce *PKCEParams) oauth2.TokenSource {\n\treturn oauth2.ReuseTokenSource(nil, authHandlerSource{config: config, ctx: ctx, authHandler: authHandler, state: state, pkce: pkce})\n}", "func NewSourceCodeToken(ctx *pulumi.Context,\n\tname string, args *SourceCodeTokenArgs, opts ...pulumi.ResourceOption) (*SourceCodeToken, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.Token == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Token'\")\n\t}\n\tif args.Type == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Type'\")\n\t}\n\tif args.Token != nil {\n\t\targs.Token = pulumi.ToSecret(args.Token).(pulumi.StringInput)\n\t}\n\tif args.TokenSecret != nil {\n\t\targs.TokenSecret = pulumi.ToSecret(args.TokenSecret).(pulumi.StringPtrInput)\n\t}\n\tsecrets := pulumi.AdditionalSecretOutputs([]string{\n\t\t\"token\",\n\t\t\"tokenSecret\",\n\t})\n\topts = append(opts, secrets)\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource SourceCodeToken\n\terr := ctx.RegisterResource(\"azure:appservice/sourceCodeToken:SourceCodeToken\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewSourceFromDefault(ctx context.Context, targetAudience string) (oauth2.TokenSource, error) {\n\n\treturn newSourceFromDefaultURL(ctx, targetAudience, googleTokenURL)\n}", "func (d *Disco) CredentialsSource() auth.CredentialsSource {\n\tif d.credsSrc == nil {\n\t\t// We'll return an empty one just to save the caller from having to\n\t\t// protect against the nil case, since this interface already allows\n\t\t// for the possibility of there being no credentials at all.\n\t\treturn auth.StaticCredentialsSource(nil)\n\t}\n\treturn d.credsSrc\n}", "func Provider() terraform.ResourceProvider {\n\treturn &schema.Provider{\n\t\tSchema: map[string]*schema.Schema{\n\t\t\t\"credentials\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tOptional: true,\n\t\t\t\tDefaultFunc: schema.MultiEnvDefaultFunc([]string{\n\t\t\t\t\t\"GOOGLE_CREDENTIALS\",\n\t\t\t\t\t\"GOOGLE_CLOUD_KEYFILE_JSON\",\n\t\t\t\t\t\"GCLOUD_KEYFILE_JSON\",\n\t\t\t\t}, nil),\n\t\t\t\tValidateFunc: validateCredentials,\n\t\t\t},\n\n\t\t\t\"access_token\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tOptional: true,\n\t\t\t\tDefaultFunc: schema.MultiEnvDefaultFunc([]string{\n\t\t\t\t\t\"GOOGLE_OAUTH_ACCESS_TOKEN\",\n\t\t\t\t}, nil),\n\t\t\t\tConflictsWith: []string{\"credentials\"},\n\t\t\t},\n\n\t\t\t\"project\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tOptional: true,\n\t\t\t\tDefaultFunc: schema.MultiEnvDefaultFunc([]string{\n\t\t\t\t\t\"GOOGLE_PROJECT\",\n\t\t\t\t\t\"GOOGLE_CLOUD_PROJECT\",\n\t\t\t\t\t\"GCLOUD_PROJECT\",\n\t\t\t\t\t\"CLOUDSDK_CORE_PROJECT\",\n\t\t\t\t}, nil),\n\t\t\t},\n\n\t\t\t\"region\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tOptional: true,\n\t\t\t\tDefaultFunc: schema.MultiEnvDefaultFunc([]string{\n\t\t\t\t\t\"GOOGLE_REGION\",\n\t\t\t\t\t\"GCLOUD_REGION\",\n\t\t\t\t\t\"CLOUDSDK_COMPUTE_REGION\",\n\t\t\t\t}, nil),\n\t\t\t},\n\n\t\t\t\"zone\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tOptional: true,\n\t\t\t\tDefaultFunc: schema.MultiEnvDefaultFunc([]string{\n\t\t\t\t\t\"GOOGLE_ZONE\",\n\t\t\t\t\t\"GCLOUD_ZONE\",\n\t\t\t\t\t\"CLOUDSDK_COMPUTE_ZONE\",\n\t\t\t\t}, nil),\n\t\t\t},\n\t\t\t\"scopes\": {\n\t\t\t\tType: schema.TypeList,\n\t\t\t\tOptional: true,\n\t\t\t\tElem: &schema.Schema{Type: schema.TypeString},\n\t\t\t},\n\t\t},\n\n\t\tDataSourcesMap: map[string]*schema.Resource{\n\t\t\t\"google_active_folder\": dataSourceGoogleActiveFolder(),\n\t\t\t\"google_billing_account\": dataSourceGoogleBillingAccount(),\n\t\t\t\"google_dns_managed_zone\": dataSourceDnsManagedZone(),\n\t\t\t\"google_client_config\": dataSourceGoogleClientConfig(),\n\t\t\t\"google_cloudfunctions_function\": dataSourceGoogleCloudFunctionsFunction(),\n\t\t\t\"google_compute_address\": dataSourceGoogleComputeAddress(),\n\t\t\t\"google_compute_backend_service\": dataSourceGoogleComputeBackendService(),\n\t\t\t\"google_compute_default_service_account\": dataSourceGoogleComputeDefaultServiceAccount(),\n\t\t\t\"google_compute_forwarding_rule\": dataSourceGoogleComputeForwardingRule(),\n\t\t\t\"google_compute_image\": dataSourceGoogleComputeImage(),\n\t\t\t\"google_compute_instance\": dataSourceGoogleComputeInstance(),\n\t\t\t\"google_compute_global_address\": dataSourceGoogleComputeGlobalAddress(),\n\t\t\t\"google_compute_instance_group\": dataSourceGoogleComputeInstanceGroup(),\n\t\t\t\"google_compute_lb_ip_ranges\": dataSourceGoogleComputeLbIpRanges(),\n\t\t\t\"google_compute_network\": dataSourceGoogleComputeNetwork(),\n\t\t\t\"google_compute_regions\": dataSourceGoogleComputeRegions(),\n\t\t\t\"google_compute_region_instance_group\": dataSourceGoogleComputeRegionInstanceGroup(),\n\t\t\t\"google_compute_subnetwork\": dataSourceGoogleComputeSubnetwork(),\n\t\t\t\"google_compute_zones\": dataSourceGoogleComputeZones(),\n\t\t\t\"google_compute_vpn_gateway\": dataSourceGoogleComputeVpnGateway(),\n\t\t\t\"google_compute_ssl_policy\": dataSourceGoogleComputeSslPolicy(),\n\t\t\t\"google_container_cluster\": dataSourceGoogleContainerCluster(),\n\t\t\t\"google_container_engine_versions\": dataSourceGoogleContainerEngineVersions(),\n\t\t\t\"google_container_registry_repository\": dataSourceGoogleContainerRepo(),\n\t\t\t\"google_container_registry_image\": dataSourceGoogleContainerImage(),\n\t\t\t\"google_iam_policy\": dataSourceGoogleIamPolicy(),\n\t\t\t\"google_iam_role\": dataSourceGoogleIamRole(),\n\t\t\t\"google_kms_secret\": dataSourceGoogleKmsSecret(),\n\t\t\t\"google_kms_key_ring\": dataSourceGoogleKmsKeyRing(),\n\t\t\t\"google_kms_crypto_key\": dataSourceGoogleKmsCryptoKey(),\n\t\t\t\"google_folder\": dataSourceGoogleFolder(),\n\t\t\t\"google_netblock_ip_ranges\": dataSourceGoogleNetblockIpRanges(),\n\t\t\t\"google_organization\": dataSourceGoogleOrganization(),\n\t\t\t\"google_project\": dataSourceGoogleProject(),\n\t\t\t\"google_project_services\": dataSourceGoogleProjectServices(),\n\t\t\t\"google_service_account\": dataSourceGoogleServiceAccount(),\n\t\t\t\"google_service_account_key\": dataSourceGoogleServiceAccountKey(),\n\t\t\t\"google_storage_bucket_object\": dataSourceGoogleStorageBucketObject(),\n\t\t\t\"google_storage_object_signed_url\": dataSourceGoogleSignedUrl(),\n\t\t\t\"google_storage_project_service_account\": dataSourceGoogleStorageProjectServiceAccount(),\n\t\t\t\"google_storage_transfer_project_service_account\": dataSourceGoogleStorageTransferProjectServiceAccount(),\n\t\t},\n\n\t\tResourcesMap: ResourceMap(),\n\n\t\tConfigureFunc: providerConfigure,\n\t}\n}", "func (t *TokenSource) Token() (*oauth2.Token, error) {\n\ttoken, err := t.TokenGenerator.GetWithSTS(context.Background(), t.ClusterID)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error generating token: %w\", err)\n\t}\n\treturn &oauth2.Token{\n\t\tAccessToken: token.Token,\n\t\tExpiry: token.Expiration.Add(-t.Leeway),\n\t}, nil\n}", "func NewClientCredentialsSource(name string, path string) ClientCredentialsSource {\n\treturn &clientCredentialsSource{\n\t\tname: name,\n\t\tpath: path,\n\t}\n}", "func Init(ctx context.Context, local bool) (oauth2.TokenSource, error) {\n\treturn auth_steps.Init(ctx, local, auth.ScopeGerrit, auth.ScopeUserinfoEmail)\n}", "func (c *contextBoundTokenSource) Token() (*oauth2.Token, error) {\n\tnow := clock.Now(c)\n\n\t// Get the current token value. We do this without locking around the token\n\t// element.\n\tif tokIface, ok := c.cache.Get(c, c.cacheKey); ok {\n\t\ttok := tokIface.(*oauth2.Token)\n\t\tif !c.closeToExpRandomized(now, tok.Expiry, accessTokenExpirationRandomization) {\n\t\t\treturn tok, nil\n\t\t}\n\t}\n\n\t// Either the token is expired, or we are selected randomly as a refresh case.\n\t// Get a new TokenSource to refresh the token with.\n\tts, err := c.makeTokenSource(c)\n\tif err != nil {\n\t\treturn nil, errors.Annotate(err, \"failed to create new TokenSource\").Err()\n\t}\n\n\t// While refreshing, we lock around the cache key via GetOrCreate in case\n\t// multiple requests are either selected or have expired.\n\ttokIface, err := c.cache.GetOrCreate(c, c.cacheKey, func() (interface{}, time.Duration, error) {\n\t\ttok, err := ts.Token()\n\t\tif err != nil {\n\t\t\treturn nil, 0, err\n\t\t}\n\n\t\texpiryDelta := tok.Expiry.Sub(now)\n\t\tswitch {\n\t\tcase expiryDelta <= 0:\n\t\t\treturn nil, 0, errors.Reason(\"retrieved expired access token (%s < %s)\", tok.Expiry, now).Err()\n\t\tcase expiryDelta > accessTokenMinExpiration:\n\t\t\t// Subtract some time from the token's expiry so we don't use it immediately\n\t\t\t// before it actually expires.\n\t\t\ttok.Expiry = tok.Expiry.Add(-accessTokenMinExpiration)\n\t\t\texpiryDelta -= accessTokenMinExpiration\n\t\t}\n\t\treturn tok, expiryDelta, nil\n\t})\n\tif err != nil {\n\t\treturn nil, errors.Annotate(err, \"failed to mint new access token\").Err()\n\t}\n\treturn tokIface.(*oauth2.Token), nil\n}", "func JWTAccessTokenSourceFromJSON(jsonKey []byte, audience string) (internal.TokenSource, error) {\n\tcfg, err := JWTConfigFromJSON(jsonKey, []string{}, \"\")\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"google: could not parse JSON key: %v\", err)\n\t}\n\tpk, err := parseKey(cfg.PrivateKey)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"google: could not parse key: %v\", err)\n\t}\n\tts := &jwtAccessTokenSource{\n\t\temail: cfg.Email,\n\t\taudience: audience,\n\t\tpk: pk,\n\t\tpkID: cfg.PrivateKeyID,\n\t}\n\ttok, err := ts.Token()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn internal.ReuseTokenSource(tok, ts), nil\n}", "func (ac *AuthorizationCodeTokenSource) Token() (*oauth2.Token, error) {\n\t// Generate a random code verifier string\n\tverifierBytes := make([]byte, 32)\n\tif _, err := rand.Read(verifierBytes); err != nil {\n\t\treturn nil, err\n\t}\n\n\tverifier := base64.RawURLEncoding.EncodeToString(verifierBytes)\n\n\t// Generate a code challenge. Only the challenge is sent when requesting a\n\t// code which allows us to keep it secret for now.\n\tshaBytes := sha256.Sum256([]byte(verifier))\n\tchallenge := base64.RawURLEncoding.EncodeToString(shaBytes[:])\n\n\t// Generate a URL with the challenge to have the user log in.\n\turl := fmt.Sprintf(\"%s?response_type=code&code_challenge=%s&code_challenge_method=S256&client_id=%s&redirect_uri=http://localhost:8484/&scope=%s\", ac.AuthorizeURL, challenge, ac.ClientID, strings.Join(ac.Scopes, `%20`))\n\n\tif len(*ac.EndpointParams) > 0 {\n\t\turl += \"&\" + ac.EndpointParams.Encode()\n\t}\n\n\t// Run server before opening the user's browser so we are ready for any redirect.\n\tcodeChan := make(chan string)\n\thandler := authHandler{\n\t\tc: codeChan,\n\t}\n\n\ts := &http.Server{\n\t\tAddr: \":8484\",\n\t\tHandler: handler,\n\t\tReadTimeout: 5 * time.Second,\n\t\tWriteTimeout: 5 * time.Second,\n\t\tMaxHeaderBytes: 1024,\n\t}\n\n\tgo func() {\n\t\t// Run in a goroutine until the server is closed or we get an error.\n\t\tif err := s.ListenAndServe(); err != http.ErrServerClosed {\n\t\t\tlog.Fatal().Err(err).Msg(\"Server exited unexpectedly\")\n\t\t}\n\t}()\n\n\t// Open auth URL in browser, print for manual use in case open fails.\n\tfmt.Println(\"Open your browser to log in using the URL:\")\n\tfmt.Println(url)\n\topen(url)\n\n\t// Get code from handler, exchange it for a token, and then return it. This\n\t// channel read blocks until one code becomes available.\n\t// There is currently no timeout.\n\tcode := <-codeChan\n\ts.Shutdown(context.Background())\n\n\tpayload := fmt.Sprintf(\"grant_type=authorization_code&client_id=%s&code_verifier=%s&code=%s&redirect_uri=http://localhost:8484/\", ac.ClientID, verifier, code)\n\n\treturn requestToken(ac.TokenURL, payload)\n}", "func NewSource(ctx *pulumi.Context,\n\tname string, args *SourceArgs, opts ...pulumi.ResourceOption) (*Source, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.OrganizationId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'OrganizationId'\")\n\t}\n\treplaceOnChanges := pulumi.ReplaceOnChanges([]string{\n\t\t\"organizationId\",\n\t})\n\topts = append(opts, replaceOnChanges)\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Source\n\terr := ctx.RegisterResource(\"google-native:securitycenter/v1:Source\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (t *TokenSource) Token() (*oauth2.Token, error) {\n\treturn &oauth2.Token{\n\t\tAccessToken: t.AccessToken,\n\t}, nil\n}", "func NewServicePrincipalTokenFromCredentials(c map[string]string, scope string) (*adal.ServicePrincipalToken, error) {\n\toauthConfig, err := adal.NewOAuthConfig(azure.PublicCloud.ActiveDirectoryEndpoint, c[\"AZURE_TENANT_ID\"])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn adal.NewServicePrincipalToken(*oauthConfig, c[\"AZURE_CLIENT_ID\"], c[\"AZURE_CLIENT_SECRET\"], scope)\n}", "func NewConfigWithToken(token string) oauth2.TokenSource {\n\treturn oauth2.StaticTokenSource(\n\t\t&oauth2.Token{AccessToken: token},\n\t)\n}", "func New(path string, mask int, d debug.Debugger) (*Source, error) {\n\tif _, err := os.Stat(path); os.IsNotExist(err) {\n\t\treturn nil, ErrFileNotFound\n\t}\n\n\tts := &Source{\n\t\tpath: path,\n\t\ttoken: new(oauth2.Token),\n\t\tmask: mask,\n\t\tDebugger: d,\n\t}\n\tts.readToken()\n\n\treturn ts, nil\n}", "func NewSource(opts ...Option) (*Source, error) {\n\ts := Source{\n\t\tprefix: \"nsscache\",\n\t\tmountPath: \"secret\",\n\t}\n\n\tfor _, opt := range opts {\n\t\topt(&s)\n\t}\n\n\tif s.client == nil {\n\t\tcl, err := api.NewClient(nil)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ts.client = cl\n\t}\n\n\treturn &s, nil\n}", "func (t *TokenSource) Token() (*oauth2.Token, error) {\n\ttoken := &oauth2.Token{\n\t\tAccessToken: t.AccessToken,\n\t}\n\treturn token, nil\n}", "func (t *TokenSource) Token() (*oauth2.Token, error) {\n\ttoken := &oauth2.Token{\n\t\tAccessToken: t.AccessToken,\n\t}\n\treturn token, nil\n}", "func NewHTTPClient(transport http.RoundTripper, ts TokenSource) (*HTTPClient, error) {\n\tif ts == nil {\n\t\treturn nil, errors.New(\"gcp: no credentials available\")\n\t}\n\treturn &HTTPClient{\n\t\tClient: http.Client{\n\t\t\tTransport: &oauth2.Transport{\n\t\t\t\tBase: transport,\n\t\t\t\tSource: ts,\n\t\t\t},\n\t\t},\n\t}, nil\n}", "func NewSecretsSource(prefix, delimiter string) (source *SecretsSource) {\n\treturn &SecretsSource{\n\t\tkoanf: koanf.New(constDelimiter),\n\t\tprefix: prefix,\n\t\tdelimiter: delimiter,\n\t}\n}", "func NewApplicationDefault(ctx context.Context, scope ...string) (credentials.PerRPCCredentials, error) {\n\tcreds, err := google.FindDefaultCredentials(ctx, scope...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// If JSON is nil, the authentication is provided by the environment and not\n\t// with a credentials file, e.g. when code is running on Google Cloud\n\t// Platform. Use the returned token source.\n\tif creds.JSON == nil {\n\t\treturn TokenSource{creds.TokenSource}, nil\n\t}\n\n\t// If auth is provided by env variable or creds file, the behavior will be\n\t// different based on whether scope is set. Because the returned\n\t// creds.TokenSource does oauth with jwt by default, and it requires scope.\n\t// We can only use it if scope is not empty, otherwise it will fail with\n\t// missing scope error.\n\t//\n\t// If scope is set, use it, it should just work.\n\t//\n\t// If scope is not set, we try to use jwt directly without oauth (this only\n\t// works if it's a service account).\n\n\tif len(scope) != 0 {\n\t\treturn TokenSource{creds.TokenSource}, nil\n\t}\n\n\t// Try to convert JSON to a jwt config without setting the optional scope\n\t// parameter to check if it's a service account (the function errors if it's\n\t// not). This is necessary because the returned config doesn't show the type\n\t// of the account.\n\tif _, err := google.JWTConfigFromJSON(creds.JSON); err != nil {\n\t\t// If this fails, it's not a service account, return the original\n\t\t// TokenSource from above.\n\t\treturn TokenSource{creds.TokenSource}, nil\n\t}\n\n\t// If it's a service account, create a JWT only access with the key.\n\treturn NewJWTAccessFromKey(creds.JSON)\n}", "func (s *reuseTokenSource) ContextToken(ctx context.Context) (*oauth2.Token, error) {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\tif s.t.Valid() {\n\t\treturn s.t, nil\n\t}\n\tt, err := s.new.ContextToken(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ts.t = t\n\treturn t, nil\n}", "func CreateTargetClientFromDynamicServiceAccountToken(ctx context.Context, targetClient kubernetes.Interface, name string) (kubernetes.Interface, error) {\n\treturn createTargetClientFromServiceAccount(ctx, targetClient, name, labelsE2ETestDynamicServiceAccountTokenAccess, func(serviceAccount *corev1.ServiceAccount) (string, error) {\n\t\ttokenRequest := &authenticationv1.TokenRequest{\n\t\t\tSpec: authenticationv1.TokenRequestSpec{\n\t\t\t\tAudiences: []string{v1beta1constants.GardenerAudience},\n\t\t\t\tExpirationSeconds: pointer.Int64(3600),\n\t\t\t},\n\t\t}\n\n\t\tif err := targetClient.Client().SubResource(\"token\").Create(ctx, serviceAccount, tokenRequest); err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\treturn tokenRequest.Status.Token, nil\n\t})\n}", "func NewWithCredentialsSource(credsSrc auth.CredentialsSource) *Disco {\n\treturn &Disco{\n\t\thostCache: make(map[svchost.Hostname]*Host),\n\t\tcredsSrc: credsSrc,\n\t\tTransport: httpTransport,\n\t}\n}", "func (gts *GithubTokenSource) Token() (*oauth2.Token, error) {\n\treturn &oauth2.Token{AccessToken: gts.PersonalToken}, nil\n}", "func AAZSTokenSource(log *util.Logger, tox vag.TokenExchanger, azsConfig string, q url.Values) (vag.TokenSource, vag.TokenSource, error) {\n\ttoken, err := tox.Exchange(q)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\ttrs := tox.TokenSource(token)\n\tazs := aazsproxy.New(log)\n\n\tmts := vag.MetaTokenSource(func() (*vag.Token, error) {\n\t\t// get TRS token from refreshing TRS token source\n\t\titoken, err := trs.TokenEx()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// exchange TRS id_token for AAZS token\n\t\tatoken, err := azs.Exchange(azsConfig, itoken.IDToken)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn atoken, err\n\n\t\t// produce tokens from refresh MBB token source\n\t}, azs.TokenSource)\n\n\treturn mts, trs, nil\n}", "func ReuseTokenSource(t *Token, src TokenSource) TokenSource {\n\t// Don't wrap a reuseTokenSource in itself. That would work,\n\t// but cause an unnecessary number of mutex operations.\n\t// Just build the equivalent one.\n\tif rt, ok := src.(*reuseTokenSource); ok {\n\t\tif t == nil {\n\t\t\t// Just use it directly.\n\t\t\treturn rt\n\t\t}\n\t\tsrc = rt.new\n\t}\n\treturn &reuseTokenSource{\n\t\tt: t,\n\t\tnew: src,\n\t}\n}", "func (p *tokenSource) Token() (*oauth2.Token, error) {\n\ttoken, err := readFileContent(path.Join(p.path, fmt.Sprintf(\"%s-token-secret\", p.name)))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttokenType, err := readFileContent(path.Join(p.path, fmt.Sprintf(\"%s-token-type\", p.name)))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// parse the token claims to get expiry time\n\tclaims, err := jwt.ParseClaims(token)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &oauth2.Token{\n\t\tAccessToken: token,\n\t\tTokenType: tokenType,\n\t\tExpiry: time.Unix(int64(claims.Exp), 0),\n\t}, nil\n}", "func NewOauthTokenCredentials(oauthToken string) Credentials {\n\treturn &oauthTokenCredentials{oauthToken: oauthToken}\n}", "func NewSSOTokenProvider(client CreateTokenAPIClient, cachedTokenFilepath string, optFns ...func(o *SSOTokenProviderOptions)) *SSOTokenProvider {\n\toptions := SSOTokenProviderOptions{\n\t\tClient: client,\n\t\tCachedTokenFilepath: cachedTokenFilepath,\n\t}\n\tfor _, fn := range optFns {\n\t\tfn(&options)\n\t}\n\n\tprovider := &SSOTokenProvider{\n\t\toptions: options,\n\t}\n\n\treturn provider\n}", "func NewSSOTokenProvider(client CreateTokenAPIClient, cachedTokenFilepath string, optFns ...func(o *SSOTokenProviderOptions)) *SSOTokenProvider {\n\toptions := SSOTokenProviderOptions{\n\t\tClient: client,\n\t\tCachedTokenFilepath: cachedTokenFilepath,\n\t}\n\tfor _, fn := range optFns {\n\t\tfn(&options)\n\t}\n\n\tprovider := &SSOTokenProvider{\n\t\toptions: options,\n\t}\n\n\treturn provider\n}", "func (s *storageTokenSource) Token() (*oauth2.Token, error) {\n\tif token, err := s.Config.Storage.GetToken(); err == nil && token.Valid() {\n\t\treturn token, err\n\t}\n\ttoken, err := s.TokenSource.Token()\n\tif err != nil {\n\t\treturn token, err\n\t}\n\tif err := s.Config.Storage.SetToken(token); err != nil {\n\t\treturn nil, err\n\t}\n\treturn token, nil\n}", "func NewTokenProvider() sarama.AccessTokenProvider {\n\tspt, err := getServicePrincipalToken(\n\t\tos.Getenv(\"AAD_TENANT_ID\"),\n\t\tos.Getenv(\"AAD_APPLICATION_ID\"),\n\t\tos.Getenv(\"AAD_APPLICATION_SECRET\"),\n\t\tos.Getenv(\"AAD_AUDIENCE\"))\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\treturn &TokenProvider{\n\t\tservicePrincipalToken : spt,\n\t}\n}", "func New(authToken string) (*datastore, error) {\n\tctx := context.Background()\n\n\tts := oauth2.StaticTokenSource(\n\t\t&oauth2.Token{AccessToken: authToken},\n\t)\n\ttc := oauth2.NewClient(ctx, ts)\n\tif tc == nil {\n\t\treturn nil, errors.New(\"Access Token Invalid\")\n\t}\n\n\tclient := github.NewClient(tc)\n\tif client == nil {\n\t\treturn nil, errors.New(\"Error creating Github client\")\n\t}\n\n\treturn &datastore{\n\t\tContext: ctx,\n\t\tClient: client,\n\t\tService: client.Git,\n\t}, nil\n}", "func (ap *oauth2ClientCredentialsAuthPlugin) requestToken() (*oauth2Token, error) {\n\tbody := url.Values{\"grant_type\": []string{\"client_credentials\"}}\n\tif len(*ap.Scopes) > 0 {\n\t\tbody[\"scope\"] = []string{strings.Join(*ap.Scopes, \" \")}\n\t}\n\n\tr, err := http.NewRequest(\"POST\", ap.TokenURL, strings.NewReader(body.Encode()))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr.Header.Set(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\tr.SetBasicAuth(ap.ClientID, ap.ClientSecret)\n\n\tclient := defaultRoundTripperClient(&tls.Config{InsecureSkipVerify: ap.tlsSkipVerify}, 10)\n\tresponse, err := client.Do(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbodyRaw, err := ioutil.ReadAll(response.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif response.StatusCode != 200 {\n\t\treturn nil, fmt.Errorf(\"error in response from OAuth2 token endpoint: %v\", string(bodyRaw))\n\t}\n\n\tvar tokenResponse tokenEndpointResponse\n\terr = json.Unmarshal(bodyRaw, &tokenResponse)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif strings.ToLower(tokenResponse.TokenType) != \"bearer\" {\n\t\treturn nil, errors.New(\"unknown token type returned from token endpoint\")\n\t}\n\n\treturn &oauth2Token{\n\t\tToken: strings.TrimSpace(tokenResponse.AccessToken),\n\t\tExpiresAt: time.Now().Add(time.Duration(tokenResponse.ExpiresIn) * time.Second),\n\t}, nil\n}", "func (t *tokenSource) Token() (*oauth2.Token, error) {\n\treturn t.token, nil\n}", "func (t *tokenSource) Token() (*oauth2.Token, error) {\n\treturn t.token, nil\n}", "func NewChromeOSTokenSourceForAccount(ctx context.Context,\n\ttconn *chrome.TestConn, scopes []string, email string) oauth2.TokenSource {\n\treturn &chromeOSTokenSource{\n\t\tctx: ctx,\n\t\ttconn: tconn,\n\t\temail: email,\n\t\tscopes: scopes,\n\t}\n}", "func newClient(token string) *github.Client {\n\tts := oauth2.StaticTokenSource(\n\t\t&oauth2.Token{AccessToken: token},\n\t)\n\ttc := oauth2.NewClient(context.Background(), ts)\n\n\treturn github.NewClient(tc)\n}", "func newReloadingTokenSource(getToken func() []byte) *reloadingTokenSource {\n\treturn &reloadingTokenSource{\n\t\tgetToken: getToken,\n\t}\n}", "func (ts *apiTokenSource) Token() (*oauth2.Token, error) {\n\tts.m.Lock()\n\tdefer ts.m.Unlock()\n\tif ts.token.Valid() {\n\t\treturn ts.token, nil\n\t}\n\n\tresp, err := ts.requestToken()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := ts.treatResponseBody(resp.Body); err != nil {\n\t\treturn nil, err\n\t}\n\treturn ts.token, nil\n}", "func (c *ChainedTokenCredential) GetToken(ctx context.Context, opts policy.TokenRequestOptions) (*azcore.AccessToken, error) {\n\tif c.successfulCredential != nil && !c.retrySources {\n\t\treturn c.successfulCredential.GetToken(ctx, opts)\n\t}\n\n\tvar errs []error\n\tfor _, cred := range c.sources {\n\t\ttoken, err := cred.GetToken(ctx, opts)\n\t\tif err == nil {\n\t\t\tlog.Writef(EventAuthentication, \"Azure Identity => %s authenticated with %s\", c.name, extractCredentialName(cred))\n\t\t\tc.successfulCredential = cred\n\t\t\treturn token, nil\n\t\t}\n\t\terrs = append(errs, err)\n\t\tif _, ok := err.(credentialUnavailableError); !ok {\n\t\t\tres := getResponseFromError(err)\n\t\t\tmsg := createChainedErrorMessage(errs)\n\t\t\treturn nil, newAuthenticationFailedError(c.name, msg, res)\n\t\t}\n\t}\n\t// if we get here, all credentials returned credentialUnavailableError\n\tmsg := createChainedErrorMessage(errs)\n\terr := newCredentialUnavailableError(c.name, msg)\n\tlog.Write(EventAuthentication, \"Azure Identity => ERROR: \"+err.Error())\n\treturn nil, err\n}", "func createOAuthClient(token string) *http.Client {\n\tctx := context.Background()\n\tsrc := oauth2.StaticTokenSource(&oauth2.Token{\n\t\tAccessToken: token,\n\t})\n\n\treturn oauth2.NewClient(ctx, src)\n}", "func NewClientFromBearerToken(req *http.Request) (Client, http.Header, error) {\n\tauthorization := req.Header.Get(\"Authorization\")\n\n\tfields := strings.Split(authorization, \" \")\n\tif len(fields) != 2 || fields[0] != \"Bearer\" {\n\t\treturn nil, nil, errMissingOrInvalidToken\n\t}\n\n\ttoken := oauth2.Token{\n\t\tTokenType: fields[0],\n\t\tAccessToken: fields[1],\n\t}\n\tclient, err := storage.NewClient(req.Context(), option.WithTokenSource(oauth2.StaticTokenSource(&token)))\n\tif err != nil {\n\t\treturn nil, nil, fmt.Errorf(\"creating client with token source: %v\", err)\n\t}\n\n\treturn GCSClient{client}, map[string][]string{\n\t\t\"Authorization\": []string{authorization},\n\t}, nil\n}", "func (d *Disco) SetCredentialsSource(src auth.CredentialsSource) {\n\td.credsSrc = src\n}", "func (settings FileSettings) ServicePrincipalTokenFromClientCredentialsWithResource(resource string) (*adal.ServicePrincipalToken, error) {\n\tif _, ok := settings.Values[ClientSecret]; !ok {\n\t\treturn nil, errors.New(\"missing client secret\")\n\t}\n\tconfig, err := adal.NewOAuthConfig(settings.getAADEndpoint(), settings.Values[TenantID])\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn adal.NewServicePrincipalToken(*config, settings.Values[ClientID], settings.Values[ClientSecret], resource)\n}", "func (c *Client) NewSdkToken(ctx context.Context, id, referrer string) (*SdkToken, error) {\n\tt := &SdkToken{\n\t\tApplicantID: id,\n\t\tReferrer: referrer,\n\t}\n\tjsonStr, err := json.Marshal(t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := c.newRequest(\"POST\", \"/sdk_token\", bytes.NewBuffer(jsonStr))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar resp SdkToken\n\tif _, err := c.do(ctx, req, &resp); err != nil {\n\t\treturn nil, err\n\t}\n\tt.Token = resp.Token\n\treturn t, err\n}", "func Credentials(ctx context.Context, scopes ...string) *google.Credentials {\n\treturn CredentialsEnv(ctx, envPrivateKey, scopes...)\n}", "func (ts *Source) Token() (*oauth2.Token, error) {\n\tif !ts.token.Valid() {\n\t\tts.Log(ts.mask, \"[TKN] token is not valid, it has probably expired\")\n\t\tif err := ts.refreshToken(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif err := ts.saveToken(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn ts.token, nil\n}", "func NewTokenCredentialAdapter(credential azcore.TokenCredential, scopes []string) autorest.Authorizer {\n\ttkPolicy := runtime.NewBearerTokenPolicy(credential, scopes, nil)\n\treturn &policyAdapter{\n\t\tpl: runtime.NewPipeline(\"azidext\", \"v0.4.0\", runtime.PipelineOptions{\n\t\t\tPerRetry: []policy.Policy{tkPolicy, nullPolicy{}},\n\t\t}, nil),\n\t}\n}", "func NewServiceAccountToken(fp string, isKeyFile bool) (*ServiceAccountToken, error) {\n\trv := &ServiceAccountToken{\n\t\tfilename: fp,\n\t}\n\t// Set the update function whether the provided file contains a cached token\n\t// or a service account keyfile.\n\trv.updateFn = rv.readTokenFromFile\n\tif isKeyFile {\n\t\tvar err error\n\t\trv.tokenSrc, err = auth.NewJWTServiceAccountTokenSource(context.TODO(), \"#bogus\", fp, \"\", \"\", compute.CloudPlatformScope, auth.ScopeUserinfoEmail)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\trv.updateFn = rv.tokenSrc.Token\n\t}\n\treturn rv, rv.Update()\n}", "func (fts ErroringTokenSource) Token() (*oauth2.Token, error) {\n\treturn nil, errors.New(\"intentional error\")\n}", "func New(ctx context.Context, credentials, token string, opts ...ClientOption) (*Client, error) {\r\n\r\n\tclient := &Client{\r\n\t\tscope: \"https://www.googleapis.com/auth/spreadsheets.readonly\",\r\n\t}\r\n\r\n\tfor _, opt := range opts {\r\n\t\tclient = opt(client)\r\n\t}\r\n\r\n\treturn new(ctx, credentials, token, client)\r\n}", "func NewSourceProvider(config Config) *SourceProvider {\n\tconn := NewConnection(config)\n\treturn &SourceProvider{\n\t\tConfig: config,\n\t\tConnection: conn,\n\t\tTaskFlow: make(chan provider.Task),\n\t\tQuitChan: make(chan bool),\n\t}\n}", "func Token(c *gin.Context) {\n\n\tclientStorage, err := models.NewClientStorage(ginject.Deps(c))\n\tif err != nil {\n\t\tpanic(err)\n\t\treturn\n\t}\n\n\tmanager := models.NewManager(ginject.Deps(c))\n\n\tserver.InitServer(manager)\n\tserver.SetAllowedGrantType(oauth2.AuthorizationCode, oauth2.Refreshing)\n\tserver.SetAllowedResponseType(oauth2.Token)\n\n\tserver.SetClientInfoHandler(func(r *http.Request) (clientID, clientSecret string, err error) {\n\n\t\tcid := r.FormValue(\"client_id\")\n\n\t\tif cid == \"\" {\n\n\t\t}\n\n\t\tinfo, err := clientStorage.GetByID(cid)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\n\t\tclientID = info.GetID()\n\t\tclientSecret = info.GetSecret()\n\t\treturn\n\t})\n\n\tserver.HandleTokenRequest(c)\n}", "func GetSourceCodeToken(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *SourceCodeTokenState, opts ...pulumi.ResourceOption) (*SourceCodeToken, error) {\n\tvar resource SourceCodeToken\n\terr := ctx.ReadResource(\"azure:appservice/sourceCodeToken:SourceCodeToken\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (c Config) NewClient(tok oauth2.Token) *Client {\n\tts := tokenSource{\n\t\ttoken: tok,\n\t\tconfig: c,\n\t}\n\t_ = ts\n\tb, _ := url.Parse(c.BaseURL)\n\treturn &Client{\n\t\tTokenSource: ts,\n\t\tClient: http.Client{\n\t\t\tTransport: &oauth2.Transport{\n\t\t\t\tBase: &Transport{BaseURL: b},\n\t\t\t\tSource: ts,\n\t\t\t},\n\t\t},\n\t}\n}", "func getGoogleTokenFromFile(file string) (*oauth2.Token, error) {\n\n\t// os.Open opens the credential file.\n\t// creates File object\n\tf, err := os.Open(file)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// initialize Token object\n\tt := &oauth2.Token{}\n\n\t// json.NewDecoder returns a new decoder that reads from r.\n\t// the decoder introduces its own buddering and may read data\n\t// from r beyond the JSON values requested.\n\td := json.NewDecoder(f)\n\n\t// Decode reads the next JSON-encoded value from its input\n\t// and stores it in the value pointed to by v.\n\terr = d.Decode(t)\n\tdefer f.Close()\n\treturn t, err\n}", "func newSource(opts ...SourceOption) *sourcesv1alpha1.HTTPSource {\n\tsrc := &sourcesv1alpha1.HTTPSource{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tNamespace: tNs,\n\t\t\tName: tName,\n\t\t\tUID: tUID,\n\t\t},\n\t\tSpec: sourcesv1alpha1.HTTPSourceSpec{\n\t\t\tSource: tSource,\n\t\t},\n\t}\n\n\tsrc.Status.InitializeConditions()\n\n\tfor _, opt := range opts {\n\t\topt(src)\n\t}\n\n\treturn src\n}", "func (ts *azureTokenSource) Token() (*azureToken, error) {\n\tts.lock.Lock()\n\tdefer ts.lock.Unlock()\n\n\tvar err error\n\ttoken := ts.cache.getToken(azureTokenKey)\n\n\tif token != nil && !token.token.IsExpired() {\n\t\treturn token, nil\n\t}\n\n\t// retrieve from config if no cache\n\tif token == nil {\n\t\ttokenFromCfg, err := ts.retrieveTokenFromCfg()\n\n\t\tif err == nil {\n\t\t\ttoken = tokenFromCfg\n\t\t}\n\t}\n\n\tif token != nil {\n\t\t// cache and return if the token is as good\n\t\t// avoids frequent persistor calls\n\t\tif !token.token.IsExpired() {\n\t\t\tts.cache.setToken(azureTokenKey, token)\n\t\t\treturn token, nil\n\t\t}\n\n\t\tklog.V(4).Info(\"Refreshing token.\")\n\t\ttokenFromRefresh, err := ts.Refresh(token)\n\t\tswitch {\n\t\tcase err == nil:\n\t\t\ttoken = tokenFromRefresh\n\t\tcase autorest.IsTokenRefreshError(err):\n\t\t\tklog.V(4).Infof(\"Failed to refresh expired token, proceed to auth: %v\", err)\n\t\t\t// reset token to nil so that the token source will be used to acquire new\n\t\t\ttoken = nil\n\t\tdefault:\n\t\t\treturn nil, fmt.Errorf(\"unexpected error when refreshing token: %v\", err)\n\t\t}\n\t}\n\n\tif token == nil {\n\t\ttokenFromSource, err := ts.source.Token()\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed acquiring new token: %v\", err)\n\t\t}\n\t\ttoken = tokenFromSource\n\t}\n\n\t// sanity check\n\tif token == nil {\n\t\treturn nil, fmt.Errorf(\"unable to acquire token\")\n\t}\n\n\t// corner condition, newly got token is valid but expired\n\tif token.token.IsExpired() {\n\t\treturn nil, fmt.Errorf(\"newly acquired token is expired\")\n\t}\n\n\terr = ts.storeTokenInCfg(token)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"storing the refreshed token in configuration: %v\", err)\n\t}\n\tts.cache.setToken(azureTokenKey, token)\n\n\treturn token, nil\n}" ]
[ "0.74979717", "0.7401055", "0.70478684", "0.70462215", "0.7018204", "0.7013672", "0.6994854", "0.69563544", "0.6934276", "0.68950164", "0.6852529", "0.6838926", "0.6566733", "0.6493974", "0.64755976", "0.64075357", "0.63987464", "0.63816744", "0.6311301", "0.6255845", "0.6236213", "0.62148416", "0.6172333", "0.615314", "0.611329", "0.6101506", "0.6006135", "0.5999053", "0.5910152", "0.58661383", "0.5844087", "0.5822731", "0.5811801", "0.57878125", "0.577386", "0.5749013", "0.5741592", "0.56212234", "0.5512078", "0.54932404", "0.5488484", "0.5486315", "0.54288495", "0.54228663", "0.53945506", "0.5391912", "0.53631216", "0.53590333", "0.5354607", "0.5327732", "0.5293619", "0.5288975", "0.5262917", "0.5249633", "0.52204764", "0.52184", "0.5209877", "0.5209877", "0.51496667", "0.51468015", "0.51454616", "0.51165134", "0.5115445", "0.51007855", "0.509807", "0.506307", "0.50311613", "0.50306505", "0.5020333", "0.49871963", "0.49871963", "0.49863034", "0.49862713", "0.49767163", "0.49757373", "0.4974377", "0.4974377", "0.49693817", "0.4949857", "0.49476585", "0.49344978", "0.49335757", "0.49291018", "0.49157667", "0.49052602", "0.48751166", "0.48312938", "0.48265323", "0.4821572", "0.48153564", "0.4794772", "0.47920808", "0.47867814", "0.47670457", "0.47658384", "0.47526658", "0.47372767", "0.47291675", "0.47164917", "0.47124854" ]
0.7532123
0
New returns a new Syncer.
func New( ctx context.Context, ctrlClient ctrlclient.Client, namespacedName types.NamespacedName, projectID string, instanceID string, credentials *Credentials, recorder record.EventRecorder, opts ...Option, ) (Syncer, error) { ts, err := credentials.TokenSource(ctx) if err != nil { return nil, err } log := logging.FromContext(ctx) spannerClient, err := spanner.NewClient( ctx, projectID, spanner.WithTokenSource(ts), spanner.WithLog(log), ) if err != nil { return nil, err } metricsClient, err := metrics.NewClient( ctx, projectID, metrics.WithTokenSource(ts), ) if err != nil { return nil, err } s := &syncer{ projectID: projectID, instanceID: instanceID, credentials: credentials, ctrlClient: ctrlClient, spannerClient: spannerClient, metricsClient: metricsClient, namespacedName: namespacedName, interval: time.Minute, stopCh: make(chan struct{}), log: zapr.NewLogger(zap.NewNop()), clock: utilclock.RealClock{}, recorder: recorder, } for _, opt := range opts { opt(s) } return s, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewSyncer(cfg *Config) *Syncer {\n\tsyncer := new(Syncer)\n\tsyncer.cfg = cfg\n\tsyncer.meta = NewLocalMeta(cfg.Meta)\n\tsyncer.closed.Set(false)\n\tsyncer.lastCount.Set(0)\n\tsyncer.count.Set(0)\n\tsyncer.insertCount.Set(0)\n\tsyncer.updateCount.Set(0)\n\tsyncer.deleteCount.Set(0)\n\tsyncer.done = make(chan struct{})\n\tsyncer.jobs = newJobChans(cfg.WorkerCount)\n\tsyncer.tables = make(map[string]*table)\n\tsyncer.ctx, syncer.cancel = context.WithCancel(context.Background())\n\tsyncer.reMap = make(map[string]*regexp.Regexp)\n\treturn syncer\n}", "func NewSyncer() *Syncer {\n\tkubeconfig := filepath.Join(\n\t\tos.Getenv(\"HOME\"), \".kube\", \"config\",\n\t)\n\tconfig, err := clientcmd.BuildConfigFromFlags(\"\", kubeconfig)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\n\tclient, err := kubernetes.NewForConfig(config)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\n\tdsn := *dbUser + \":\" + *dbPass + \"@\" + *dbHost + \"/\" + *dbName + \"?charset=utf8\"\n\tdb, err := sql.Open(\"mysql\", dsn)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\n\treturn &Syncer{\n\t\tclient: *client,\n\t\tdata: make(map[string]*v1.ResourceQuota),\n\t\tdb: db,\n\t}\n}", "func New(baseRuntime *base.Runtime, owner, ownee client.Object, fn controllerutil.MutateFn) Syncer {\n\treturn &k8sSyncer{\n\t\tRuntime: baseRuntime,\n\t\tlog: baseRuntime.Log.WithName(\"syncer\"),\n\t\towner: owner,\n\t\townee: ownee,\n\t\tuserMutateFn: fn,\n\t}\n}", "func NewSyncer(url url.URL, archs map[string]bool, storage Storage) *Syncer {\n\treturn &Syncer{url, archs, storage}\n}", "func New(dst, src string) *Sync {\n\treturn &Sync{\n\t\tVerbose: true,\n\t\tDst: dst,\n\t\tSrc: src,\n\t}\n}", "func NewSyncer(dir string) *Syncer {\n\treturn &Syncer{\n\t\tdir: dir,\n\t\tsyncedFiles: make(map[string]bool),\n\t}\n}", "func New(ag agent.Agent, validators []types.Address, storage storage.Storage) (*MerkleSyncer, error) {\n\tctx, cancel := context.WithCancel(context.Background())\n\n\treturn &MerkleSyncer{\n\t\twrapperC: make(chan *pb.MerkleWrapper, wrapperCNumber),\n\t\tagent: ag,\n\t\tvalidators: validators,\n\t\tstorage: storage,\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t}, nil\n}", "func NewSyncer(uploadDriver uploader.Driver) *Syncer {\n\treturn &Syncer{\n\t\tuploader: uploadDriver,\n\n\t\ttobeUploadedObjects: make([]uploader.Object, 0),\n\t\ttobeDeletedObjects: make([]uploader.Object, 0),\n\t}\n}", "func NewSyncer(ds dtypes.MetadataDS, sm *stmgr.StateManager, exchange exchange.Client, syncMgrCtor SyncManagerCtor, connmgr connmgr.ConnManager, self peer.ID, beacon beacon.Schedule, verifier ffiwrapper.Verifier) (*Syncer, error) {\n\tgen, err := sm.ChainStore().GetGenesis()\n\tif err != nil {\n\t\treturn nil, xerrors.Errorf(\"getting genesis block: %w\", err)\n\t}\n\n\tgent, err := types.NewTipSet([]*types.BlockHeader{gen})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcp, err := loadCheckpoint(ds)\n\tif err != nil {\n\t\treturn nil, xerrors.Errorf(\"error loading mpool config: %w\", err)\n\t}\n\n\ts := &Syncer{\n\t\tds: ds,\n\t\tcheckpt: cp,\n\t\tbeacon: beacon,\n\t\tbad: NewBadBlockCache(),\n\t\tGenesis: gent,\n\t\tExchange: exchange,\n\t\tstore: sm.ChainStore(),\n\t\tsm: sm,\n\t\tself: self,\n\t\treceiptTracker: newBlockReceiptTracker(),\n\t\tconnmgr: connmgr,\n\t\tverifier: verifier,\n\n\t\tincoming: pubsub.New(50),\n\t}\n\n\tif build.InsecurePoStValidation {\n\t\tlog.Warn(\"*********************************************************************************************\")\n\t\tlog.Warn(\" [INSECURE-POST-VALIDATION] Insecure test validation is enabled. If you see this outside of a test, it is a severe bug! \")\n\t\tlog.Warn(\"*********************************************************************************************\")\n\t}\n\n\ts.syncmgr = syncMgrCtor(s.Sync)\n\treturn s, nil\n}", "func NewSync() (Sync, error) {\n\tlogStore, err := datastore.NewEventLogStore()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"Unable to Sync events, cannot connect to log store\")\n\t}\n\tlinkedCustomerStore, err := datastore.NewLinkedCustomerStore()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"Unable to Sync events, cannot connect to linked customer store\")\n\t}\n\treturn &sync{\n\t\tstop: make(chan bool),\n\t\tsubscriber: messaging.NewRedisStreamSubscriber(),\n\t\tlogStore: logStore,\n\t\tlinkedCustomerStore: linkedCustomerStore,\n\t}, nil\n}", "func NewSync(mountID mount.ID, m mount.Mount, opts SyncOpts) (*Sync, error) {\n\tif err := opts.Valid(); err != nil {\n\t\treturn nil, err\n\t}\n\n\ts := &Sync{\n\t\topts: opts,\n\t\tmountID: mountID,\n\t\tm: m,\n\t}\n\n\tif opts.Log != nil {\n\t\ts.log = opts.Log.New(\"sync\")\n\t} else {\n\t\ts.log = machine.DefaultLogger.New(\"sync\")\n\t}\n\n\t// Create directory structure if it doesn't exist.\n\tif err := os.MkdirAll(filepath.Join(s.opts.WorkDir, \"data\"), 0755); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Fetch or load remote index.\n\tvar err error\n\tif s.ridx, err = s.loadIdx(RemoteIndexName, s.fetchRemoteIdx); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Create or load local index.\n\tif s.lidx, err = s.loadIdx(LocalIndexName, s.fetchLocalIdx); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Update local index if needed.\n\tif err := s.updateLocal(); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Create FS event consumer queue.\n\ts.a = NewAnteroom()\n\n\t// Create file system notification object.\n\ts.n, err = opts.NotifyBuilder.Build(&notify.BuildOpts{\n\t\tMountID: mountID,\n\t\tMount: m,\n\t\tCache: s.a,\n\t\tCacheDir: filepath.Join(s.opts.WorkDir, \"data\"),\n\t\tRemoteIdx: s.ridx,\n\t\tLocalIdx: s.lidx,\n\t})\n\tif err != nil {\n\t\ts.a.Close()\n\t\treturn nil, err\n\t}\n\n\t// Create file synchronization object.\n\ts.s, err = opts.SyncBuilder.Build(&BuildOpts{\n\t\tRemoteIdx: s.ridx,\n\t\tLocalIdx: s.lidx,\n\t})\n\tif err != nil {\n\t\ts.n.Close()\n\t\ts.a.Close()\n\t\treturn nil, err\n\t}\n\n\treturn s, nil\n}", "func New(option Option) (*Synchronizer, error) {\n\treturn newWithNow(option, time.Now)\n}", "func NewSynchronization()(*Synchronization) {\n m := &Synchronization{\n Entity: *NewEntity(),\n }\n return m\n}", "func New(cfg *Config) (*SyncManager, error) {\n\tsm := SyncManager{\n\t\tcfg: *cfg,\n\t\ttxMemPool: make(map[common.Uint256]struct{}),\n\t\trequestedTxns: make(map[common.Uint256]struct{}),\n\t\trequestedBlocks: make(map[common.Uint256]struct{}),\n\t\tpeerStates: make(map[*peer.Peer]*peerSyncState),\n\t\tmsgChan: make(chan interface{}, cfg.MaxPeers*3),\n\t\tquit: make(chan struct{}),\n\t}\n\n\treturn &sm, nil\n}", "func NewSync(c *cli.Context) Sync {\n\treturn Sync{\n\t\tsrc: c.Args().Get(0),\n\t\tdst: c.Args().Get(1),\n\t\top: c.Command.Name,\n\t\tfullCommand: commandFromContext(c),\n\n\t\t// flags\n\t\tdelete: c.Bool(\"delete\"),\n\t\tsizeOnly: c.Bool(\"size-only\"),\n\t\texitOnError: c.Bool(\"exit-on-error\"),\n\n\t\t// flags\n\t\tfollowSymlinks: !c.Bool(\"no-follow-symlinks\"),\n\t\tstorageClass: storage.StorageClass(c.String(\"storage-class\")),\n\t\traw: c.Bool(\"raw\"),\n\t\t// region settings\n\t\tsrcRegion: c.String(\"source-region\"),\n\t\tdstRegion: c.String(\"destination-region\"),\n\t\tstorageOpts: NewStorageOpts(c),\n\t}\n}", "func NewSyncer(config *Config, outputCollection OutputCollection, logger *logrus.Entry, metricsHandle *sqmetrics.SquareMetrics) (*Syncer, error) {\n\t// Pre-parse poll interval\n\tpollInterval := time.Duration(0)\n\tif config.PollInterval != \"\" {\n\t\tvar err error\n\t\tpollInterval, err = time.ParseDuration(config.PollInterval)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Couldn't parse Poll Interval '%s': %v\\n\", config.PollInterval, err)\n\t\t}\n\t\tlogger.Infof(\"Poll interval is %s\", config.PollInterval)\n\t}\n\n\tsyncer := Syncer{\n\t\tconfig: config,\n\t\tclients: map[string]syncerEntry{},\n\t\toldClients: map[string]syncerEntry{},\n\t\tlogger: logger,\n\t\tmetricsHandle: metricsHandle,\n\t\tpollInterval: pollInterval,\n\t\toutputCollection: outputCollection,\n\t}\n\n\tserverURL, err := url.Parse(\"https://\" + config.Server)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed parsing server: %s\", config.Server)\n\t}\n\tsyncer.server = serverURL\n\n\t// Add callback for last success gauge\n\tmetricsHandle.AddGauge(\"seconds_since_last_success\", func() int64 {\n\t\tsince, _ := syncer.timeSinceLastSuccess()\n\t\treturn int64(since / time.Second)\n\t})\n\n\tsyncer.updateMostRecentError(nilError)\n\treturn &syncer, nil\n}", "func New(conf config.Config, orderer gomel.Orderer, log zerolog.Logger, setup bool) (gomel.Syncer, error) {\n\ts := &syncer{}\n\n\t// init fetch\n\tvar netserv network.Server\n\tvar err error\n\tnetserv, s.subservices, err = getNetServ(conf.FetchNetType, conf.Pid, conf.FetchAddresses, s.subservices, conf.Timeout, log)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tserv, ftrigger := fetch.NewServer(conf, orderer, netserv, log.With().Int(lg.Service, lg.FetchService).Logger())\n\ts.servers = append(s.servers, serv)\n\ts.fetch = ftrigger\n\t// init gossip\n\tnetserv, s.subservices, err = getNetServ(conf.GossipNetType, conf.Pid, conf.GossipAddresses, s.subservices, conf.Timeout, log)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tserv, gtrigger := gossip.NewServer(conf, orderer, netserv, log.With().Int(lg.Service, lg.GossipService).Logger())\n\ts.servers = append(s.servers, serv)\n\ts.gossip = gtrigger\n\tif setup {\n\t\t// init rmc\n\t\tnetserv, s.subservices, err = getNetServ(conf.RMCNetType, conf.Pid, conf.RMCAddresses, s.subservices, conf.Timeout, log)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tserv, s.mcast = rmc.NewServer(conf, orderer, netserv, log.With().Int(lg.Service, lg.RMCService).Logger())\n\t\ts.servers = append(s.servers, serv)\n\t} else {\n\t\t// init mcast\n\t\tnetserv, s.subservices, err = getNetServ(conf.MCastNetType, conf.Pid, conf.MCastAddresses, s.subservices, conf.Timeout, log)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tserv, s.mcast = multicast.NewServer(conf, orderer, netserv, log.With().Int(lg.Service, lg.MCService).Logger())\n\t\ts.servers = append(s.servers, serv)\n\t}\n\treturn s, nil\n}", "func New(config *Config) *SyncManager {\n\tsm := SyncManager{\n\t\tpeerNotifier: config.PeerNotifier,\n\t\tchain: config.Chain,\n\t\tchainParams: config.ChainParams,\n\t\ttxMemPool: config.TxMemPool,\n\t\tblockMemPool: config.BlockMemPool,\n\t\trejectedTxns: make(map[common.Uint256]struct{}),\n\t\trequestedTxns: make(map[common.Uint256]struct{}),\n\t\trequestedBlocks: make(map[common.Uint256]struct{}),\n\t\trequestedConfirmedBlocks: make(map[common.Uint256]struct{}),\n\t\tpeerStates: make(map[*peer.Peer]*peerSyncState),\n\t\tmsgChan: make(chan interface{}, config.MaxPeers*3),\n\t\tquit: make(chan struct{}),\n\t}\n\n\tevents.Subscribe(sm.handleBlockchainEvents)\n\n\treturn &sm\n}", "func New() Awaiter {\n\ta := new(awaiter)\n\ta.cancel = make(chan interface{})\n\ta.wg = new(sync.WaitGroup)\n\ta.locker = new(sync.Mutex)\n\treturn a\n}", "func (p *PersistentSyncMap) New() {\n\tp.storage = make(map[interface{}]interface{}, 100)\n}", "func New() *differ {\n\treturn &differ{}\n}", "func New(kubeconfig *rest.Config, opa opa_client.Data, ns types.ResourceType) *GenericSync {\n\topaPrefix := ns.Resource\n\tif ns.Namespace != \"\" {\n\t\topaPrefix = fmt.Sprintf(\"%s/%s\", ns.Resource, ns.Namespace)\n\t}\n\n\treturn &GenericSync{\n\t\tkubeconfig: kubeconfig,\n\t\tns: ns,\n\t\topa: opa.Prefix(opaPrefix),\n\t}\n}", "func newSyncCommand() cli.Command {\n\treturn new(syncCommand).getCommand()\n}", "func New(name string) *Locker {\n\treturn &Locker{\n\t\tRedis: redis.New(name),\n\t}\n}", "func New(dev *model.Dev) (*Syncthing, error) {\n\tfullPath := getInstallPath()\n\tremotePort, err := model.GetAvailablePort(dev.Interface)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tremoteGUIPort, err := model.GetAvailablePort(dev.Interface)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tguiPort, err := model.GetAvailablePort(dev.Interface)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlistenPort, err := model.GetAvailablePort(dev.Interface)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tpwd := uuid.New().String()\n\thash, err := bcrypt.GenerateFromPassword([]byte(pwd), 10)\n\tif err != nil {\n\t\toktetoLog.Infof(\"couldn't hash the password %s\", err)\n\t\thash = []byte(\"\")\n\t}\n\n\tcompression := \"metadata\"\n\tif dev.Sync.Compression {\n\t\tcompression = \"always\"\n\t}\n\ts := &Syncthing{\n\t\tAPIKey: \"cnd\",\n\t\tGUIPassword: pwd,\n\t\tGUIPasswordHash: string(hash),\n\t\tbinPath: fullPath,\n\t\tClient: NewAPIClient(),\n\t\tFileWatcherDelay: DefaultFileWatcherDelay,\n\t\tGUIAddress: net.JoinHostPort(dev.Interface, strconv.Itoa(guiPort)),\n\t\tHome: config.GetAppHome(dev.Namespace, dev.Name),\n\t\tLogPath: GetLogFile(dev.Namespace, dev.Name),\n\t\tListenAddress: net.JoinHostPort(dev.Interface, strconv.Itoa(listenPort)),\n\t\tRemoteAddress: fmt.Sprintf(\"tcp://%s:%d\", dev.Interface, remotePort),\n\t\tRemoteDeviceID: DefaultRemoteDeviceID,\n\t\tRemoteGUIAddress: net.JoinHostPort(dev.Interface, strconv.Itoa(remoteGUIPort)),\n\t\tLocalGUIPort: guiPort,\n\t\tLocalPort: listenPort,\n\t\tRemoteGUIPort: remoteGUIPort,\n\t\tRemotePort: remotePort,\n\t\tType: \"sendonly\",\n\t\tIgnoreDelete: true,\n\t\tVerbose: dev.Sync.Verbose,\n\t\tFolders: []*Folder{},\n\t\tRescanInterval: strconv.Itoa(dev.Sync.RescanInterval),\n\t\tCompression: compression,\n\t\ttimeout: time.Duration(dev.Timeout.Default),\n\t}\n\tindex := 1\n\tfor _, sync := range dev.Sync.Folders {\n\t\tresult, err := dev.IsSubPathFolder(sync.LocalPath)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif !result {\n\t\t\ts.Folders = append(\n\t\t\t\ts.Folders,\n\t\t\t\t&Folder{\n\t\t\t\t\tName: strconv.Itoa(index),\n\t\t\t\t\tLocalPath: sync.LocalPath,\n\t\t\t\t\tRemotePath: sync.RemotePath,\n\t\t\t\t},\n\t\t\t)\n\t\t\tindex++\n\t\t}\n\t}\n\n\treturn s, nil\n}", "func New(com bool) SchTask {\n\treturn SchTask{\n\t\tbin: taskerFile,\n\t\tprefix: \"go-wintask-\",\n\t\tcompatibility: com,\n\t}\n}", "func New(localNodes ipld.NodeGetter, blockStore coreiface.BlockAPI, opts ...func(cfg *Config)) (*Dsync, error) {\n\tcfg := &Config{\n\t\tPushPreCheck: DefaultDagPrecheck,\n\t\tPushFinalCheck: DefaultDagFinalCheck,\n\t}\n\n\tfor _, opt := range opts {\n\t\topt(cfg)\n\t}\n\n\tif err := cfg.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tds := &Dsync{\n\t\tlng: localNodes,\n\t\tbapi: blockStore,\n\n\t\trequireAllBlocks: cfg.RequireAllBlocks,\n\t\tallowRemoves: cfg.AllowRemoves,\n\n\t\tpreCheck: cfg.PushPreCheck,\n\t\tfinalCheck: cfg.PushFinalCheck,\n\t\tonCompleteHook: cfg.PushComplete,\n\t\tgetDagInfoCheck: cfg.GetDagInfoCheck,\n\t\topenBlockStreamCheck: cfg.OpenBlockStreamCheck,\n\t\tremoveCheck: cfg.RemoveCheck,\n\n\t\tsessionPool: map[string]*session{},\n\t\tsessionCancels: map[string]context.CancelFunc{},\n\t\tsessionTTLDur: time.Hour * 5,\n\t}\n\n\tif cfg.PinAPI != nil {\n\t\tds.pin = cfg.PinAPI\n\t}\n\tif cfg.InfoStore != nil {\n\t\tds.infoStore = cfg.InfoStore\n\t}\n\n\tif cfg.HTTPRemoteAddress != \"\" {\n\t\tm := http.NewServeMux()\n\t\tm.Handle(\"/dsync\", HTTPRemoteHandler(ds))\n\n\t\tds.httpServer = &http.Server{\n\t\t\tAddr: cfg.HTTPRemoteAddress,\n\t\t\tHandler: m,\n\t\t}\n\t}\n\n\tif cfg.Libp2pHost != nil {\n\t\tlog.Debug(\"dsync: adding p2p handler\")\n\t\tds.p2pHandler = newp2pHandler(ds, cfg.Libp2pHost)\n\t}\n\n\treturn ds, nil\n}", "func New(dependencies Dependencies) {\n\twriter = dependencies.Writer\n\treader = dependencies.Reader\n\thost = dependencies.Host\n\tnotifierService = dependencies.NotifierService\n}", "func newSyncable(sqlConfig *sqlConfig, databases map[string]syncable.Database) (syncable.Syncable, error) {\n\tdatabase := databases[sqlConfig.sqlDB]\n\tif database == nil {\n\t\treturn &syncable.ZeroSyncable{}, fmt.Errorf(\"Database %s is not setup\", sqlConfig.sqlDB)\n\t}\n\tif database.Type() != \"sql\" {\n\t\treturn &syncable.ZeroSyncable{}, fmt.Errorf(\"Database %s is not a sql database\", sqlConfig.sqlDB)\n\t}\n\tsqlDB := database.(*DB)\n\n\treturn &Syncable{config: sqlConfig, database: sqlDB}, nil\n}", "func New(cb Done, transport http.RoundTripper) *Manager {\n\treturn &Manager{\n\t\tkeys: sets.NewString(),\n\t\tcb: cb,\n\t\ttransport: transport,\n\t}\n}", "func New() Lock {\n\treturn &lock{sem: make(chan int, 1)}\n}", "func New() target.Locker {\n\treturn &Noop{}\n}", "func New() Manager {\n\treturn Manager{\n\t\tState: make(map[string]string),\n\t\tClientHolder: make(map[string]utils.Set),\n\t\tClientQueue: make(map[string]utils.Queue),\n\t}\n}", "func New() lock.Locker {\n\tc := make(chan struct{}, 1)\n\tc <- struct{}{}\n\treturn &trivial {\n\t\tc: c,\n\t}\n}", "func New() *nauth {\n\tsingleton.mutex.Lock()\n\tdefer singleton.mutex.Unlock()\n\n\tif singleton.generatedKeysC == nil && PrecalculatedKeysNum > 0 {\n\t\tsingleton.generatedKeysC = make(chan keyPair, PrecalculatedKeysNum)\n\t\tgo singleton.precalculateKeys()\n\t}\n\treturn &singleton\n}", "func New() GoroutineLock {\n\treturn GoroutineLock(getGoroutineID())\n}", "func newRelaySync(shared sharedSync) *relaySync {\n\treturn &relaySync{\n\t\tdone: false,\n\t\tshared: shared,\n\t\tfirstSetupDone: false,\n\t\tfirstSetupWait: make(chan struct{}),\n\t\tsetupCompleteHeld: false,\n\t\tlegCompleteHeld: false,\n\t}\n}", "func New(cb ReceiveMessageCallback, syncCB ReceiveSyncMessageCallback) *Worker {\n\tid := nextWorkerId()\n\n\tcbWrapper := &callbacks{\n\t\tcb: cb,\n\t\tsyncCB: syncCB,\n\t}\n\tcallbacksMapLocker.Lock()\n\tcallbacksMap[id] = cbWrapper\n\tcallbacksMapLocker.Unlock()\n\n\tinitV8Once.Do(func() {\n\t\tC.v8_init()\n\t})\n\n\tworker := &Worker{}\n\tworker.cWorker = C.worker_new(C.int(id))\n\truntime.SetFinalizer(worker, func(final_worker *Worker) {\n\t\tC.worker_dispose(final_worker.cWorker)\n\t\tcallbacksMapLocker.Lock()\n\t\tdelete(callbacksMap, id)\n\t\tcallbacksMapLocker.Unlock()\n\t})\n\treturn worker\n}", "func New(apiKey string) *Clockwork {\n\treturn &Clockwork{\n\t\tapiKey: apiKey,\n\t}\n}", "func New(u *url.URL) (godilock.DLocker, error) {\n\thost := u.Host\n\tif err := zklock.Connect(strings.Split(host, \",\"), 20*time.Second); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &LockManager{\n\t\tprefix: u.Path,\n\t\tsession: make(map[string]*zklock.Dlocker),\n\t}, nil\n}", "func New(sync *contract.Sync) Service {\n\treturn &service{\n\t\tSync: sync,\n\t\tsource: &dbResource{Resource: sync.Source},\n\t\tdest: &dbResource{Resource: sync.Dest},\n\t}\n}", "func New() (*T) {\n\n\tme := T{\n\t\tcount: 0,\n\t\tdatum: \"\",\n\t}\n\n\treturn &me\n}", "func New(cfg clients.Config) (*Client, error) {\n\tif err := cfg.Validate(); err != nil {\n\t\treturn nil, errors.Trace(err)\n\t}\n\n\tproducerCfg := sarama.NewConfig()\n\tproducerCfg.Producer.Return.Successes = true\n\n\tproducer, err := sarama.NewSyncProducer(\n\t\tcfg.Brokers,\n\t\tproducerCfg,\n\t)\n\tif err != nil {\n\t\treturn nil, errors.Trace(err)\n\t}\n\n\treturn &Client{\n\t\tproducer: producer,\n\t\tConfig: cfg,\n\t}, nil\n}", "func New(\n\tctx context.Context,\n\tlogger logr.Logger,\n\tclock clock.Clock,\n\tc client.Client,\n\tnamespace string,\n\tidentity string,\n\trotation Config,\n) (\n\tInterface,\n\terror,\n) {\n\tm := &manager{\n\t\tstore: make(secretStore),\n\t\tclock: clock,\n\t\tlogger: logger.WithValues(\"namespace\", namespace),\n\t\tclient: c,\n\t\tnamespace: namespace,\n\t\tidentity: identity,\n\t\tlastRotationInitiationTimes: make(nameToUnixTime),\n\t}\n\n\tif err := m.initialize(ctx, rotation); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn m, nil\n}", "func newSharedSync(channel *Channel) sharedSync {\n\tshared := sharedSync{\n\t\ttransportCtx: channel.ctx,\n\t\ttransportLock: channel.transportLock,\n\t\tlegComplete: new(sync.WaitGroup),\n\t\trunSetup: new(sync.WaitGroup),\n\t\tsetupComplete: new(sync.WaitGroup),\n\t\trunLeg: new(sync.WaitGroup),\n\t}\n\n\tshared.runSetup.Add(1)\n\treturn shared\n}", "func New() (interface{}, error) {\n\treturn Backend(), nil\n}", "func (m *Manager) Create(vdiskID string) (tlog.SlaveSyncer, error) {\n\tm.mux.Lock()\n\tdefer m.mux.Unlock()\n\n\t// check if it already exist\n\tss, ok := m.syncers[vdiskID]\n\tif ok {\n\t\treturn ss, nil\n\t}\n\n\t// creates if not exist\n\tss, err := newSlaveSyncer(m.ctx, m.configSource, vdiskID, m.privKey, m)\n\tif err != nil {\n\t\tlog.Errorf(\"slavesync mgr: failed to create syncer for vdisk: %v, err: %v\", vdiskID, err)\n\t\treturn nil, err\n\t}\n\tm.syncers[vdiskID] = ss\n\treturn ss, nil\n}", "func New(name string, client *clientv3.Client) *Ring {\n\tpkgMu.Lock()\n\tdefer pkgMu.Unlock()\n\treturn &Ring{\n\t\tName: name,\n\t\tclient: client,\n\t\tkv: clientv3.NewKV(client),\n\t\tbackendID: getBackendID(),\n\t\tleaseTimeout: 120, // 120 seconds\n\t}\n}", "func New(ReqSize int, BufSzie int) *Async {\n\tas := Async{\n\t\tquit: make(chan bool),\n\t\ttaskChan: make(chan *task, ReqSize),\n\t\tbufSize: BufSzie,\n\t\twait: &sync.WaitGroup{},\n\t\tdone: make(chan bool),\n\t}\n\n\tgo as.watcher()\n\treturn &as\n}", "func New(config Config) zapcore.WriteSyncer {\n\treturn &gelf{Config: config}\n}", "func New(v interface{}, config Config) (*twerk, error) {\n\n\tcallableFunc, err := callable.New(v)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = isValid(config)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttwrkr := &twerk{\n\t\tcallable: callableFunc,\n\n\t\tconfig: config,\n\n\t\tjobListener: make(chan jobInstruction, config.Max),\n\n\t\tliveWorkersNum: newAtomicNumber(0),\n\t\tcurrentlyWorkingNum: newAtomicNumber(0),\n\n\t\tbroadcastDie: make(chan bool),\n\n\t\torchestrator: &defaultOrchestrator{config: &config},\n\n\t\tstopTheWorldLock: &sync.RWMutex{},\n\t}\n\n\ttwrkr.startInBackground()\n\n\treturn twrkr, nil\n}", "func New(st *store.Store) Scheduler {\n\treturn Scheduler{\n\t\tstore: st,\n\t\treadyCh: make(chan struct{}),\n\t\tupdateCh: make(chan struct{}),\n\t\tprevT: time.Unix(0, 0),\n\t\tstartTime: time.Unix(0, 0),\n\t}\n}", "func New() *Updater {\n\treturn &Updater{}\n}", "func New() (*Serf, error) {\n\tserf := &Serf{\n\t\tshutdownCh: make(chan struct{}),\n\t}\n\n\treturn serf, nil\n}", "func NewSync(opts ...Option) Sync {\n\toptions := Options{}\n\tfor _, o := range opts {\n\t\to(&options)\n\t}\n\n\treturn &memorySync{\n\t\toptions: options,\n\t\tlocks: make(map[string]*memoryLock),\n\t}\n}", "func New(dir string, conn *websocket.Conn) *Watcher {\n\treturn &Watcher{\n\t\tConn: conn,\n\t\tDir: dir,\n\t}\n}", "func New() Closer {\n\treturn &closer{ch: make(chan struct{})}\n}", "func New(cfg *config.Config, rootPath string) (*Manager, error) {\n\tself := peer.NewSelf(cfg, rootPath)\n\tm := &Manager{\n\t\tself: self,\n\t}\n\tp := self.ToPeer()\n\trecent, err := p.ReadRecent()\n\tif log.If(err) {\n\t\tlog.Println(err)\n\t}\n\tm.recent, err = thread.NewList(self, recent)\n\tm.peers, err = p.ReadPeers()\n\tlog.If(err)\n\tif len(m.peers) == 0 {\n\t\tm.peers = self.LoadInitPeers()\n\t}\n\tm.tags, err = p.ReadTags()\n\tlog.If(err)\n\tm.spams, err = p.ReadSpams()\n\tlog.If(err)\n\tm.readThreads = self.ReadThreads()\n\tif err := self.StartListen(m.accept); log.If(err) {\n\t\treturn nil, err\n\t}\n\tgo func() {\n\t\tfor {\n\t\t\tm.cron()\n\t\t\ttime.Sleep(time.Hour)\n\t\t}\n\t}()\n\treturn m, nil\n}", "func New(id string, quit qu.C, uuid uint64) (w *Worker, conn net.Conn) {\n\t// log.L.SetLevel(\"trace\", true)\n\tsc := stdconn.New(os.Stdin, os.Stdout, quit)\n\t\n\treturn NewWithConnAndSemaphore(id, sc, quit, uuid), sc\n}", "func New(t *testing.T) Tester {\n\treturn Tester{t}\n}", "func NewSyncerFromFile(config *Config, clientConfig ClientConfig, bundle string, logger *logrus.Entry, metricsHandle *sqmetrics.SquareMetrics) (*Syncer, error) {\n\tsyncer := Syncer{\n\t\tconfig: config,\n\t\tclients: map[string]syncerEntry{},\n\t\toldClients: map[string]syncerEntry{},\n\t\tlogger: logger,\n\t\tmetricsHandle: metricsHandle,\n\t\tdisableClientReloading: true,\n\t}\n\n\tclient, err := NewBackupBundleClient(bundle, logger)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\toutput, err := OutputDirCollection{Config: config}.NewOutput(clientConfig, logger)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsyncer.clients[clientConfig.DirName] = syncerEntry{\n\t\tclient,\n\t\tclientConfig,\n\t\toutput,\n\t\tmap[string]secretState{},\n\t}\n\n\tsyncer.updateMostRecentError(nilError)\n\n\treturn &syncer, nil\n}", "func (c *Creator) New() (filestorage.FileStorage, error) {\n\tfs := New(c.apiKey, c.secret)\n\treturn fs, fs.SetBucket(c.defaultBucket)\n}", "func New() *Storage {\n\treturn &Storage{\n\t\tstat: &statApp{},\n\t}\n}", "func New(t T) *Tester {\n\ttt := &Tester{\n\t\tt: t,\n\n\t\tclients: make(map[string]*client),\n\n\t\tcodecs: make(map[string]goka.Codec),\n\t\ttopicQueues: make(map[string]*queue),\n\t\tstorages: make(map[string]storage.Storage),\n\t}\n\ttt.tmgr = NewMockTopicManager(tt, 1, 1)\n\ttt.producer = newProducerMock(tt.handleEmit)\n\n\treturn tt\n}", "func New(cfg *config.Config, a aws.Client, g google.Client) SyncGSuite {\n\treturn &syncGSuite{\n\t\taws: a,\n\t\tgoogle: g,\n\t\tcfg: cfg,\n\t\tusers: make(map[string]*aws.User),\n\t}\n}", "func NewRateLimitSyncer(registry *ratelimit.Registry, serviceLister externalServiceLister) *RateLimitSyncer {\n\tr := &RateLimitSyncer{\n\t\tregistry: registry,\n\t\tserviceLister: serviceLister,\n\t\tlimit: 500,\n\t}\n\treturn r\n}", "func (correlation) New() string {\n\treturn utils.NewUUID()\n}", "func New(sess *session.Session, config *client.Config) *TaskRouter {\n\treturn &TaskRouter{\n\t\tV1: v1.New(sess, config),\n\t}\n}", "func NewSyncthing() *Syncthing {\n\treturn &Syncthing{}\n}", "func New(url string, username string, password string) TransmissionClient {\n\tapiclient := NewClient(url, username, password)\n\ttc := TransmissionClient{apiclient: apiclient}\n\treturn tc\n}", "func New() *Serial {\n\treturn &Serial{}\n}", "func New(ctx context.Context, s *Session, health healthcheck.Handler) Controller {\n\ttemp := controller{\n\t\tcmd: make(chan int),\n\t\tsong: make(chan string),\n\t\tplaylist: make(chan string),\n\t\tready: make(chan struct{}),\n\t\tcurrentToken: make(chan oauth2.Token),\n\t\thealth: health,\n\t}\n\n\tgo run(ctx, s, temp)\n\treturn temp\n}", "func New() (interfaces.AsyncWork, error) {\n\treturn NewWithContext(context.Background())\n}", "func New(logger logrus.FieldLogger, conf Config) (*Collector, error) {\n\tproducer, err := sarama.NewSyncProducer(conf.Brokers, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Collector{\n\t\tProducer: producer,\n\t\tConfig: conf,\n\t\tlogger: logger,\n\t}, nil\n}", "func NewSyncMachine(\n\tlogger log.StandardLogger,\n\tchannel net.BroadcastChannel,\n\tblockCounter chain.BlockCounter,\n\tinitialState SyncState,\n) *SyncMachine {\n\treturn &SyncMachine{\n\t\tlogger: logger,\n\t\tchannel: channel,\n\t\tblockCounter: blockCounter,\n\t\tinitialState: initialState,\n\t}\n}", "func New(r *httprouter.Router) Transport {\n\treturn &transport{r}\n}", "func New(tracker *bestblock.Tracker, lag int64, config Config) (*Manager, error) {\n\tm := &Manager{\n\t\tconfig: config,\n\t\tlag: lag,\n\t\ttracker: tracker,\n\t\tclosed: false,\n\t\tstopChan: make(chan bool),\n\t}\n\n\tlog.Info(\"setting up redis connection\")\n\tm.redis = redis.NewClient(&redis.Options{\n\t\tAddr: config.RedisServer,\n\t\tPassword: config.RedisPassword,\n\t\tDB: 0,\n\t\tReadTimeout: time.Second * 1,\n\t})\n\n\terr := m.redis.Ping().Err()\n\tif err != nil {\n\t\tlog.Error(err)\n\t\treturn nil, err\n\t}\n\n\tlog.Info(\"connected to redis successfully\")\n\n\tgo m.watchNewBlocks()\n\n\treturn m, nil\n}", "func New(st Storage) *Server {\n\tsrv := &Server{\n\t\tst: st,\n\t}\n\tsrv.setupRouter()\n\treturn srv\n}", "func New(drv Driver) FileSystem {\n\treturn FileSystem{drv}\n}", "func New() Clock {\n\treturn &clock{}\n}", "func New(cfg *config.Config) *Core {\n\treturn &Core{\n\t\tcfg: cfg,\n\t\tlockWatchCh: make(chan int),\n\t\torder: types.NewOrder(),\n\t}\n}", "func NewSynchronizationJob()(*SynchronizationJob) {\n m := &SynchronizationJob{\n Entity: *NewEntity(),\n }\n return m\n}", "func New(client plugins.Client, cryptor txcrypto.Cryptor) (*AppchainMonitor, error) {\n\tmeta, err := client.GetOutMeta()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"get out interchainCounter from broker contract :%w\", err)\n\t}\n\n\tctx, cancel := context.WithCancel(context.Background())\n\treturn &AppchainMonitor{\n\t\tclient: client,\n\t\tinterchainCounter: meta,\n\t\tcryptor: cryptor,\n\t\trecvCh: make(chan *pb.IBTP, 1024),\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t}, nil\n}", "func New(sync *contract.Sync, dao dao.Service, mutex *shared.Mutex, jobService jobs.Service, historyService history.Service) Service {\n\treturn newService(sync, dao, mutex, jobService, historyService)\n}", "func New(\n\tclient client.Client,\n\ttaskQueue string,\n\toptions Options,\n) Worker {\n\treturn internal.NewWorker(client, taskQueue, options)\n}", "func New(ctx context.Context, servsHandler services.Handler) Queue {\n\tqueue := &senderWorkQueue{\n\t\tmainCtx: ctx,\n\t\twakeUp: make(chan int),\n\t\tqueue: map[string]*openapi.Event{},\n\t\tservsHandler: servsHandler,\n\t}\n\n\tgo queue.work()\n\n\treturn queue\n}", "func New(qw QueueWriter) *Updater {\n\treturn &Updater{qw, sync.Map{}}\n}", "func newWatcher(loader *Loader, uri string, interval time.Duration, onStop func()) *watcher {\n\treturn &watcher{\n\t\tstate: isCreated,\n\t\tupdatedAt: 0,\n\t\tloader: loader,\n\t\turi: uri,\n\t\tupdates: make(chan Update, 1),\n\t\tinterval: interval,\n\t\tonStop: onStop,\n\t}\n}", "func New(m map[string]interface{}) (share.Manager, error) {\n\tc, err := parseConfig(m)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tc.init()\n\n\treturn NewShareManager(c)\n}", "func (f *Factory) Sync(syncFn SyncFunc) *Factory {\n\tf.sync = syncFn\n\treturn f\n}", "func New() *Worker {\n\treturn &Worker{\n\t\tQueue: make(chan Match),\n\t\tContent: content.New(),\n\t\tdone: make(chan bool),\n\t}\n}", "func NewBroadcaster() *Broadcaster {\n\treturn &Broadcaster{\n\t\tsync.Mutex{},\n\t\tmake(chan message, 1),\n\t}\n}", "func New() iface.Backend {\n\treturn &Backend{\n\t\tBackend: common.NewBackend(new(config.Config)),\n\t\tgroups: make(map[string][]string),\n\t\ttasks: make(map[string][]byte),\n\t}\n}", "func NewSyncTransport(token, endpoint string) *SyncTransport {\n\treturn &SyncTransport{\n\t\tbaseTransport{\n\t\t\tToken: token,\n\t\t\tEndpoint: endpoint,\n\t\t\tRetryAttempts: DefaultRetryAttempts,\n\t\t\tPrintPayloadOnError: true,\n\t\t\tItemsPerMinute: 0,\n\t\t\tperMinCounter: 0,\n\t\t\tstartTime: time.Now(),\n\t\t},\n\t}\n}", "func New(orderStream chan orders.Order) *Worker {\n\treturn &Worker{\n\t\torderStream: orderStream,\n\t\tresultStream: make(chan anydb.Record),\n\t}\n}", "func New(pushURL, owner string) {\n\tSave(pushURL, config.Tart{\n\t\tName: pushURL,\n\t\tPushURL: pushURL,\n\t\tIsRunning: false,\n\t\tOwners: []string{owner},\n\t\tPID: -1,\n\t\tRestartDelaySecs: 30,\n\t\tRestartOnStop: false,\n\t\tLogStdout: true,\n\t})\n}", "func newStorage() *storage {\n\treturn &storage{\n\t\tsto: make(map[uint16]mqtt.Message),\n\t\tmux: new(sync.RWMutex),\n\t}\n}", "func New(f internalinterfaces.SharedInformerFactory) Interface {\n\treturn &version{f}\n}", "func New(f internalinterfaces.SharedInformerFactory) Interface {\n\treturn &version{f}\n}", "func New(f internalinterfaces.SharedInformerFactory) Interface {\n\treturn &version{f}\n}" ]
[ "0.7282004", "0.721814", "0.7176084", "0.70723194", "0.7026891", "0.69710535", "0.68922335", "0.6689", "0.6485574", "0.6469128", "0.634105", "0.62894493", "0.6284482", "0.62109864", "0.6156181", "0.6142157", "0.59772795", "0.59637696", "0.5897962", "0.5894021", "0.58785456", "0.58574104", "0.57843095", "0.5766043", "0.57290614", "0.56936884", "0.56178874", "0.56099695", "0.5601514", "0.5597902", "0.5592118", "0.55882746", "0.5573639", "0.55584633", "0.5555459", "0.55412793", "0.55354494", "0.54884464", "0.54803324", "0.54671186", "0.54651064", "0.5463053", "0.54399955", "0.541742", "0.54169685", "0.54112893", "0.539796", "0.53977054", "0.53941846", "0.5388436", "0.5386041", "0.5381723", "0.53615457", "0.53610915", "0.5358668", "0.5357931", "0.5340808", "0.5339524", "0.5334168", "0.5332895", "0.5312567", "0.53098106", "0.52899915", "0.5288315", "0.52823323", "0.52752066", "0.52733713", "0.52700496", "0.5265257", "0.52637136", "0.5259203", "0.52467674", "0.5236427", "0.5232166", "0.52296585", "0.522877", "0.52278125", "0.52261764", "0.52234274", "0.5215607", "0.52146494", "0.5212882", "0.5209605", "0.5209307", "0.52041733", "0.5203845", "0.51961267", "0.5195047", "0.5190949", "0.5190064", "0.51894206", "0.5179941", "0.5169141", "0.51689804", "0.5168031", "0.51664364", "0.51662076", "0.5164786", "0.5164786", "0.5164786" ]
0.67866945
7
UpdateTarget updates target and returns wether did update or not.
func (s *syncer) UpdateTarget(projectID, instanceID string, credentials *Credentials) bool { updated := false if s.projectID != projectID { updated = true s.projectID = projectID } if s.instanceID != instanceID { updated = true s.instanceID = instanceID } // TODO: Consider deepCopy if !reflect.DeepEqual(s.credentials, credentials) { updated = true s.credentials = credentials } return updated }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (mr *ClientMockRecorder) UpdateTarget(arg0, arg1, arg2 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"UpdateTarget\", reflect.TypeOf((*Client)(nil).UpdateTarget), arg0, arg1, arg2)\n}", "func (m *Client) UpdateTarget(arg0 context.Context, arg1 int64, arg2 zendesk.Target) (zendesk.Target, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"UpdateTarget\", arg0, arg1, arg2)\n\tret0, _ := ret[0].(zendesk.Target)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func NeedsUpdate() (bool, error) {\n\ttarget := findTarget()\n\tif target == nil {\n\t\treturn false, nil\n\t}\n\texists, err := target.isSetup()\n\treturn !exists, err\n}", "func (o *RequestTarget) HasTarget() bool {\n\tif o != nil && o.Target != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (t *Target) Update(newTarget *Target) {\n\tmutableMutex.Lock()\n defer mutableMutex.Unlock()\n\tt.Protocol = newTarget.Protocol\n\tt.Dest = newTarget.Dest\n\tt.TCPTLS = newTarget.TCPTLS\n\tt.HTTPMethod = newTarget.HTTPMethod\n\tt.HTTPStatusList = newTarget.HTTPStatusList\n\tt.Regexp = newTarget.Regexp\n\tt.ResSize = newTarget.ResSize\n\tt.Retry = newTarget.Retry\n\tt.RetryWait = newTarget.RetryWait\n\tt.Timeout = newTarget.Timeout\n\tt.TLSSkipVerify = newTarget.TLSSkipVerify\n}", "func (c *FakeAWSSNSTargets) Update(ctx context.Context, aWSSNSTarget *v1alpha1.AWSSNSTarget, opts v1.UpdateOptions) (result *v1alpha1.AWSSNSTarget, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewUpdateAction(awssnstargetsResource, c.ns, aWSSNSTarget), &v1alpha1.AWSSNSTarget{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1alpha1.AWSSNSTarget), err\n}", "func (o *SLOOverallStatuses) HasTarget() bool {\n\treturn o != nil && o.Target != nil\n}", "func (a *Agent) updateTarget() error {\n\tif a.episodes%a.UpdateTargetEpisodes == 0 {\n\t\tlog.Infof(\"updating target model - current steps %v target update %v\", a.steps, a.updateTargetSteps)\n\t\terr := a.Policy.(*model.Sequential).CloneLearnablesTo(a.TargetPolicy.(*model.Sequential))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (u *Updater) Update() (bool, error) {\n\tisPatched := false\n\tfor _, updater := range u.updaters {\n\t\tif !u.dryRun {\n\t\t\terr := updater.buildNewAndOldObject()\n\t\t\tif err != nil {\n\t\t\t\treturn false, errors.Annotatef(err, \"build components:\")\n\t\t\t}\n\t\t}\n\t\tisUpdated, err := updater.add(isPatched)\n\t\tisPatched = isPatched || u.isPatched || isUpdated\n\t\tif err != nil {\n\t\t\treturn false, errors.Annotatef(err, \"add/patch components:\")\n\t\t}\n\t\tif !u.dryRun {\n\t\t\terr = updater.remove()\n\t\t\tif err != nil {\n\t\t\t\treturn false, errors.Annotatef(err, \"remove components:\")\n\t\t\t}\n\t\t}\n\t}\n\treturn isPatched, nil\n}", "func UpdateTargetHandler(w http.ResponseWriter, r *http.Request) {\n\tenv := envFromRequest(r)\n\n\tip, _, err := net.SplitHostPort(r.RemoteAddr)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tif err := r.ParseForm(); err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\tmac, scriptName, environment, params := parsePostForm(r.PostForm)\n\tif mac == \"\" || scriptName == \"\" {\n\t\thttp.Error(w, \"MAC address and target must not be empty\", http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tserver := server.New(mac, ip, \"\")\n\tinputErr, err := polling.UpdateTarget(\n\t\tenv.Logger, env.ServerStates, env.Templates, env.EventLog, env.BaseURL, server,\n\t\tscriptName, environment, params)\n\n\tif err != nil {\n\t\tif inputErr {\n\t\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\t} else {\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t}\n\t\treturn\n\t}\n\thttp.Redirect(w, r, \"/\", http.StatusFound)\n}", "func (c *FakeCloudwatchEventTargets) Update(cloudwatchEventTarget *v1alpha1.CloudwatchEventTarget) (result *v1alpha1.CloudwatchEventTarget, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewUpdateAction(cloudwatcheventtargetsResource, c.ns, cloudwatchEventTarget), &v1alpha1.CloudwatchEventTarget{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1alpha1.CloudwatchEventTarget), err\n}", "func (a Actions) Update() bool {\n\tif len(a) != 1 {\n\t\treturn false\n\t}\n\n\treturn a[0] == ActionUpdate\n}", "func targetChanged(s1, s2 *core.BuildState, t1, t2 *core.BuildTarget) bool {\n\th1 := build.RuleHash(s1, t1, true, false)\n\th2 := build.RuleHash(s2, t2, true, false)\n\tif !bytes.Equal(h1, h2) {\n\t\treturn true\n\t}\n\th1, err1 := sourceHash(s1, t1)\n\th2, err2 := sourceHash(s2, t2)\n\treturn !bytes.Equal(h1, h2) || err1 != nil || err2 != nil\n}", "func (o *SearchSLOResponseDataAttributesFacets) HasTarget() bool {\n\tif o != nil && o.Target != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (s *TargetCRUD) Update(arg ...crud.Arg) (crud.Arg, error) {\n\tevent := eventFromArg(arg[0])\n\ttarget := targetFromStuct(event)\n\toldTarget, ok := event.OldObj.(*state.Target)\n\tif !ok {\n\t\tpanic(\"unexpected type, expected *state.Target\")\n\t}\n\tprint.DeletePrintln(\"deleting target\", *oldTarget.Target.Target,\n\t\t\"from upstream\", *oldTarget.Upstream.ID)\n\tprint.CreatePrintln(\"creating target\", *target.Target.Target,\n\t\t\"on upstream\", *target.Upstream.ID)\n\treturn target, nil\n}", "func (c *CoWIdler) Updated() bool {\n\treturn c.newStatus != nil\n}", "func (m *ParcelMock) DefaultTargetFinished() bool {\n\t// if expectation series were set then invocations count should be equal to expectations count\n\tif len(m.DefaultTargetMock.expectationSeries) > 0 {\n\t\treturn atomic.LoadUint64(&m.DefaultTargetCounter) == uint64(len(m.DefaultTargetMock.expectationSeries))\n\t}\n\n\t// if main expectation was set then invocations count should be greater than zero\n\tif m.DefaultTargetMock.mainExpectation != nil {\n\t\treturn atomic.LoadUint64(&m.DefaultTargetCounter) > 0\n\t}\n\n\t// if func was set then invocations count should be greater than zero\n\tif m.DefaultTargetFunc != nil {\n\t\treturn atomic.LoadUint64(&m.DefaultTargetCounter) > 0\n\t}\n\n\treturn true\n}", "func (p *Privmsg) IsValidTarget() bool {\n\treturn len(p.Target) > 0\n}", "func (p *Proxy) UpdateTargets(groups map[string][]*target.Target) error {\n\tp.targetsLock.Lock()\n\tdefer p.targetsLock.Unlock()\n\n\tnewTargets := map[uint64]*target.ScrapeStatus{}\n\tfor _, ts := range groups {\n\t\tfor _, t := range ts {\n\t\t\told := p.targets[t.Hash]\n\t\t\tif old == nil {\n\t\t\t\told = target.NewScrapeStatus(t.Series)\n\t\t\t}\n\t\t\tnewTargets[t.Hash] = old\n\t\t}\n\t}\n\tp.targets = newTargets\n\treturn nil\n}", "func (m *Mounter) HasTarget(targetPath string) (string, bool) {\n\tm.Lock()\n\tdefer m.Unlock()\n\n\tfor k, v := range m.mounts {\n\t\tfor _, p := range v.Mountpoint {\n\t\t\tif p.Path == targetPath {\n\t\t\t\treturn k, true\n\t\t\t}\n\t\t}\n\t}\n\treturn \"\", false\n}", "func (e *Explore) UpdateTargets(targets map[string][]*discovery.SDTargets) {\n\te.targetsLock.Lock()\n\tdefer e.targetsLock.Unlock()\n\n\tfor job, ts := range targets {\n\t\tall := map[uint64]*exploringTarget{}\n\t\tfor _, t := range ts {\n\t\t\thash := t.ShardTarget.Hash\n\n\t\t\tif e.targets[job] != nil && e.targets[job][hash] != nil {\n\t\t\t\tall[hash] = e.targets[job][hash]\n\t\t\t} else {\n\t\t\t\tall[hash] = &exploringTarget{\n\t\t\t\t\tjob: job,\n\t\t\t\t\trt: target.NewScrapeStatus(0),\n\t\t\t\t\ttarget: t.ShardTarget,\n\t\t\t\t}\n\t\t\t\te.needExplore <- all[hash]\n\t\t\t}\n\t\t}\n\t\te.targets[job] = all\n\t}\n}", "func (b Block) CheckTarget(target Target) bool {\n\tblockHash := b.ID()\n\treturn bytes.Compare(target[:], blockHash[:]) >= 0\n}", "func checkBuildTarget(path string, update bool) bool {\n\tif f, err := os.Stat(path); err == nil {\n\t\tif update && !f.IsDir() {\n\t\t\tsylog.Fatalf(\"Only sandbox updating is supported.\")\n\t\t}\n\t\tif !update && !force {\n\t\t\treader := bufio.NewReader(os.Stdin)\n\t\t\tfmt.Print(\"Build target already exists. Do you want to overwrite? [N/y] \")\n\t\t\tinput, err := reader.ReadString('\\n')\n\t\t\tif err != nil {\n\t\t\t\tsylog.Fatalf(\"Error parsing input: %s\", err)\n\t\t\t}\n\t\t\tif val := strings.Compare(strings.ToLower(input), \"y\\n\"); val == 0 {\n\t\t\t\tforce = true\n\t\t\t} else {\n\t\t\t\tsylog.Errorf(\"Stopping build.\")\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t}\n\treturn true\n}", "func (u *Updater) HasUpdate() (bool, error) {\n\treturn u.pkg.CanUpdate()\n}", "func (ctl *Ctl) CanUpdate() (bool, error) {\n\treturn true, nil\n}", "func (ctl *Ctl) CanUpdate() (bool, error) {\n\treturn true, nil\n}", "func VerifyTargetGroup() bool {\n\treturn defaultVerifyTargetGroup\n}", "func (me TxsdSystemCategory) IsTarget() bool { return me.String() == \"target\" }", "func (o *RequestStatusMetadata) HasTargets() bool {\n\tif o != nil && o.Targets != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (n *Sub) IsUpdate(w *model.Watcher) bool {\n\treturn n.w.Mtime.Unix() != w.Mtime.Unix()\n}", "func (o *RequestTarget) GetTargetOk() (*ResourceReference, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Target, true\n}", "func (p MetadataChangedPredicate) Update(e event.UpdateEvent) bool {\n\tif e.ObjectOld == nil {\n\t\treturn false\n\t}\n\n\tif e.ObjectNew == nil {\n\t\treturn false\n\t}\n\n\tmetaChanged := !reflect.DeepEqual(e.ObjectOld.GetLabels(), e.ObjectNew.GetLabels()) ||\n\t\t!reflect.DeepEqual(e.ObjectOld.GetAnnotations(), e.ObjectNew.GetAnnotations()) ||\n\t\t!reflect.DeepEqual(e.ObjectOld.GetFinalizers(), e.ObjectNew.GetFinalizers())\n\n\treturn metaChanged\n}", "func (c *rolloutContext) shouldVerifyTargetGroup(svc *corev1.Service) bool {\n\tif !defaults.VerifyTargetGroup() {\n\t\t// feature is disabled\n\t\treturn false\n\t}\n\tdesiredPodHash := c.newRS.Labels[v1alpha1.DefaultRolloutUniqueLabelKey]\n\tif c.rollout.Spec.Strategy.BlueGreen != nil {\n\t\tif c.rollout.Status.StableRS == desiredPodHash {\n\t\t\t// for blue-green, we only verify targets right after switching active service. So if\n\t\t\t// we are fully promoted, then there is no need to verify targets.\n\t\t\t// NOTE: this is the opposite of canary, where we only verify targets if stable == desired\n\t\t\treturn false\n\t\t}\n\t\tsvcPodHash := svc.Spec.Selector[v1alpha1.DefaultRolloutUniqueLabelKey]\n\t\tif svcPodHash != desiredPodHash {\n\t\t\t// we have not yet switched service selector\n\t\t\treturn false\n\t\t}\n\t\tif c.rollout.Status.BlueGreen.PostPromotionAnalysisRunStatus != nil {\n\t\t\t// we already started post-promotion analysis, so verification already occurred\n\t\t\treturn false\n\t\t}\n\t\treturn true\n\t} else if c.rollout.Spec.Strategy.Canary != nil {\n\t\tif c.rollout.Spec.Strategy.Canary.TrafficRouting == nil || c.rollout.Spec.Strategy.Canary.TrafficRouting.ALB == nil {\n\t\t\t// not ALB canary, so no need to verify targets\n\t\t\treturn false\n\t\t}\n\t\tif c.rollout.Status.StableRS != desiredPodHash {\n\t\t\t// for canary, we only verify targets right after switching stable service, which happens\n\t\t\t// after the update. So if stable != desired, we are still in the middle of an update\n\t\t\t// and there is no need to verify targets.\n\t\t\t// NOTE: this is the opposite of blue-green, where we only verify targets if stable != active\n\t\t\treturn false\n\t\t}\n\t\treturn true\n\t}\n\t// should not get here\n\treturn false\n}", "func ShouldUpdate(rollout Rollout, nodeID storj.NodeID) bool {\n\treturn isRolloutCandidate(nodeID, rollout)\n}", "func (t *Target) GetAlive() (bool) {\n\tmutableMutex.Lock()\n defer mutableMutex.Unlock()\n\treturn t.alive\n}", "func (SourceChangePredicate) Update(e event.UpdateEvent) bool {\n\tif e.MetaOld == nil || e.MetaNew == nil {\n\t\t// ignore objects without metadata\n\t\treturn false\n\t}\n\tif e.MetaNew.GetGeneration() != e.MetaOld.GetGeneration() {\n\t\t// reconcile on spec changes\n\t\treturn true\n\t}\n\n\t// handle force sync\n\tif val, ok := e.MetaNew.GetAnnotations()[sourcev1.SyncAtAnnotation]; ok {\n\t\tif valOld, okOld := e.MetaOld.GetAnnotations()[sourcev1.SyncAtAnnotation]; okOld {\n\t\t\tif val != valOld {\n\t\t\t\treturn true\n\t\t\t}\n\t\t} else {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (c *CLIUpdater) IsUpdatable() (bool, error) {\n\texecutablePath, err := osext.Executable()\n\tif err != nil {\n\t\treturn false, err\n\t}\n\terr = unix.Access(executablePath, unix.W_OK)\n\tif err == nil {\n\t\treturn true, nil\n\t}\n\t// File is not writable, check if the current user is owner.\n\ts := &unix.Stat_t{}\n\terr = unix.Stat(executablePath, s)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tif int(s.Uid) != os.Getuid() {\n\t\treturn false, nil\n\t}\n\treturn true, nil\n}", "func (c *rolloutContext) areTargetsVerified() bool {\n\treturn c.targetsVerified == nil || *c.targetsVerified\n}", "func (op *UpdateTensorboardOperation) Done() bool {\n\treturn op.lro.Done()\n}", "func DoesTargetMatchPods(spec accessClient.TrafficTargetSpec, srcPod *corev1.Pod, dstPod *corev1.Pod) bool {\n\treturn cli.DoesTargetRefDstPod(spec, dstPod) && cli.DoesTargetRefSrcPod(spec, srcPod)\n}", "func (r *Elasticsearch) Update() bool {\n\treturn true\n}", "func (op *UpdateSnapshotOperation) Done() bool {\n\treturn op.lro.Done()\n}", "func TestApplyAddTargetTwice(t *testing.T) {\n\trepo, _, err := testutils.EmptyRepo(\"docker.com/notary\")\n\trequire.NoError(t, err)\n\t_, err = repo.InitTargets(data.CanonicalTargetsRole)\n\trequire.NoError(t, err)\n\thash := sha256.Sum256([]byte{})\n\tf := &data.FileMeta{\n\t\tLength: 1,\n\t\tHashes: map[string][]byte{\n\t\t\t\"sha256\": hash[:],\n\t\t},\n\t}\n\tfjson, err := json.Marshal(f)\n\trequire.NoError(t, err)\n\n\tcl := changelist.NewMemChangelist()\n\trequire.NoError(t, cl.Add(&changelist.TUFChange{\n\t\tActn: changelist.ActionCreate,\n\t\tRole: changelist.ScopeTargets,\n\t\tChangeType: \"target\",\n\t\tChangePath: \"latest\",\n\t\tData: fjson,\n\t}))\n\trequire.NoError(t, cl.Add(&changelist.TUFChange{\n\t\tActn: changelist.ActionCreate,\n\t\tRole: changelist.ScopeTargets,\n\t\tChangeType: \"target\",\n\t\tChangePath: \"latest\",\n\t\tData: fjson,\n\t}))\n\n\trequire.NoError(t, applyChangelist(repo, nil, cl))\n\trequire.Len(t, repo.Targets[\"targets\"].Signed.Targets, 1)\n\trequire.NotEmpty(t, repo.Targets[\"targets\"].Signed.Targets[\"latest\"])\n\n\trequire.NoError(t, applyTargetsChange(repo, nil, &changelist.TUFChange{\n\t\tActn: changelist.ActionCreate,\n\t\tRole: changelist.ScopeTargets,\n\t\tChangeType: \"target\",\n\t\tChangePath: \"latest\",\n\t\tData: fjson,\n\t}))\n\trequire.Len(t, repo.Targets[\"targets\"].Signed.Targets, 1)\n\trequire.NotEmpty(t, repo.Targets[\"targets\"].Signed.Targets[\"latest\"])\n}", "func (LabelsAndGenerationPredicate) Update(e event.UpdateEvent) bool {\n\n\tif e.MetaOld == nil {\n\t\tlog.Error(nil, \"Update event has no old metadata\", \"event\", e)\n\t\treturn false\n\t}\n\tif e.ObjectOld == nil {\n\t\tlog.Error(nil, \"Update event has no old runtime object to update\", \"event\", e)\n\t\treturn false\n\t}\n\tif e.ObjectNew == nil {\n\t\tlog.Error(nil, \"Update event has no new runtime object for update\", \"event\", e)\n\t\treturn false\n\t}\n\tif e.MetaNew == nil {\n\t\tlog.Error(nil, \"Update event has no new metadata\", \"event\", e)\n\t\treturn false\n\t}\n\n\t// reconcile if the labels have changed\n\treturn !reflect.DeepEqual(e.MetaOld.GetLabels(), e.MetaNew.GetLabels()) ||\n\t\te.MetaNew.GetGeneration() != e.MetaOld.GetGeneration()\n}", "func (a *HyperflexApiService) UpdateHyperflexTarget(ctx context.Context, moid string) ApiUpdateHyperflexTargetRequest {\n\treturn ApiUpdateHyperflexTargetRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tmoid: moid,\n\t}\n}", "func (p UserSignupChangedPredicate) Update(e runtimeevent.UpdateEvent) bool {\n\tif !checkMetaObjects(changedLog, e) {\n\t\treturn false\n\t}\n\treturn e.ObjectNew.GetGeneration() != e.ObjectOld.GetGeneration() ||\n\t\tp.annotationChanged(e, toolchainv1alpha1.UserSignupUserEmailAnnotationKey) ||\n\t\tp.labelChanged(e, toolchainv1alpha1.UserSignupUserEmailHashLabelKey)\n}", "func (qp QueryParsed) IsUpdate() bool {\n\treturn qp.Structure.GetKind() == lib.QueryKindCreate ||\n\t\tqp.Structure.GetKind() == lib.QueryKindDrop ||\n\t\tqp.Structure.GetKind() == lib.QueryKindDelete ||\n\t\tqp.Structure.GetKind() == lib.QueryKindInsert ||\n\t\tqp.Structure.GetKind() == lib.QueryKindUpdate\n}", "func (x *fastReflection_MsgUpdateParams) IsValid() bool {\n\treturn x != nil\n}", "func (o *Ga4ghChemotherapy) HasUpdated() bool {\n\tif o != nil && o.Updated != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (t *Texture) SetNeedsUpdate(value bool) *Texture {\n\tt.p.Set(\"needsUpdate\", value)\n\treturn t\n}", "func (p OwnerInNamespacePredicate) Update(e event.UpdateEvent) bool {\n\treturn p.ownerInNamespace(e.MetaNew.GetOwnerReferences())\n}", "func (o *Ga4ghTumourboard) HasUpdated() bool {\n\tif o != nil && o.Updated != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (_ *updateDebugPredicate) Update(evt event.UpdateEvent) bool {\n\tif os.Getenv(\"DEBUG_UPDATE\") == \"true\" {\n\t\tobj := fmt.Sprintf(\"%s/%s\", evt.ObjectNew.GetNamespace(), evt.ObjectNew.GetName())\n\t\tdiff, err := client.MergeFrom(evt.ObjectOld).Data(evt.ObjectNew)\n\t\tif err != nil {\n\t\t\tupdateDebugLog.Info(\"error generating diff\", \"err\", err, \"obj\", obj)\n\t\t} else {\n\t\t\tupdateDebugLog.Info(\"Update diff\", \"diff\", string(diff), \"obj\", obj)\n\t\t}\n\t}\n\treturn true\n}", "func (c *Candy) Update(g *Game) bool {\n\treturn c.state.Update(g, c)\n}", "func (c *ClusterStateImpl) instanceUpdateCompleted(asgName, instanceId string) bool {\n\treturn c.getInstanceState(asgName, instanceId) == updateCompleted\n}", "func (t *Target) ValidateUpdate(tx *pop.Connection) (*validate.Errors, error) {\n\treturn validate.NewErrors(), nil\n}", "func (s *Member) SetTargetName(v string) *Member {\n\ts.TargetName = &v\n\treturn s\n}", "func (t1 *Target) Equal(t2 *Target) bool {\n\treturn reflect.DeepEqual(t1.Target, t2.Target)\n}", "func (op *UpdateReplicationOperation) Done() bool {\n\treturn op.lro.Done()\n}", "func (s *txAllocState) Updated() bool {\n\treturn s.meta.Updated() || s.data.Updated()\n}", "func (kew *KeyspaceEventWatcher) TargetIsBeingResharded(ctx context.Context, target *querypb.Target) bool {\n\tif target.TabletType != topodatapb.TabletType_PRIMARY {\n\t\treturn false\n\t}\n\tks := kew.getKeyspaceStatus(ctx, target.Keyspace)\n\tif ks == nil {\n\t\treturn false\n\t}\n\treturn ks.beingResharded(target.Shard)\n}", "func (b *BrBuilder) CanUpdate() bool {\n\tfpath := filepath.Join(b.BuildPath, config.LatestBuildFile)\n\tif st, _ := os.Stat(fpath); st != nil && !st.IsDir() {\n\t\treturn true\n\t}\n\tlog.Trace(\"[Branch] Access build path %s failed.\", fpath)\n\treturn false\n}", "func (s *UpdateStatus) CheckUpdate(localVerStr string, remoteVerStr string, remoteAddr string) error {\n\ts.Lock()\n\tdefer s.Unlock()\n\n\tif !s.seedSet.Has(remoteAddr) {\n\t\treturn nil\n\t}\n\n\tlocalVersion, err := gover.NewVersion(localVerStr)\n\tif err != nil {\n\t\treturn err\n\t}\n\tremoteVersion, err := gover.NewVersion(remoteVerStr)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif remoteVersion.GreaterThan(localVersion) {\n\t\tif s.versionStatus == noUpdate {\n\t\t\ts.versionStatus = hasUpdate\n\t\t}\n\n\t\tmaxVersion, err := gover.NewVersion(s.maxVerSeen)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif remoteVersion.GreaterThan(maxVersion) {\n\t\t\ts.maxVerSeen = remoteVerStr\n\t\t}\n\t}\n\tif remoteVersion.Segments()[0] > localVersion.Segments()[0] {\n\t\ts.versionStatus = hasMUpdate\n\t}\n\tif s.versionStatus != noUpdate {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"module\": logModule,\n\t\t\t\"Current version\": localVerStr,\n\t\t\t\"Newer version\": remoteVerStr,\n\t\t\t\"seed\": remoteAddr,\n\t\t}).Warn(\"Please update your bytomd via https://github.com/Bytom/bytom/releases/ or http://bytom.io/wallet/\")\n\t\ts.notified = true\n\t}\n\treturn nil\n}", "func (fs *FrameworkState) Update(ctx context.Context, b []byte) (stateChanged bool, err error) {\n\t// Operates in a locked environment.\n\n\tevent, err := ParseOperatorSubscribe(b)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"framework state update: %s\", err)\n\t}\n\n\tchanged := false\n\tswitch *event.Type {\n\tcase mesos_v1_master.Event_SUBSCRIBED:\n\t\tchanged = parseSubscribeEvent(fs, event)\n\tcase mesos_v1_master.Event_TASK_ADDED:\n\t\tchanged = parseTaskAddEvent(ctx, fs, event)\n\tcase mesos_v1_master.Event_TASK_UPDATED:\n\t\tchanged = parseTaskUpdateEvent(fs, event)\n\tcase mesos_v1_master.Event_AGENT_ADDED:\n\t\t// An agent may be re-added even if the agent was never removed, for\n\t\t// example if mesos agent restarts then it re-adds itself when coming\n\t\t// back.\n\t\tchanged = parseAgentAddEvent(fs, event)\n\tcase mesos_v1_master.Event_AGENT_REMOVED:\n\t\tchanged = parseAgentRemoveEvent(fs, event)\n\tcase mesos_v1_master.Event_UNKNOWN:\n\t\tlogger.Debug(\"mesos unknown mesos event\")\n\tdefault:\n\t\tlogger.WithField(\"type\", *event.Type).Debug(\"mesos unrecognized mesos event\")\n\t}\n\n\treturn changed, nil\n}", "func (DaemonSetStatusChangedPredicate) Update(e event.UpdateEvent) bool {\n\tnewDS := e.ObjectNew.DeepCopyObject().(*appsv1.DaemonSet)\n\toldDS := e.ObjectOld.DeepCopyObject().(*appsv1.DaemonSet)\n\tplog.V(2).Info(\"newDS\", \"nUNS:=\", newDS.Status.UpdatedNumberScheduled, \"oUNS:=\", oldDS.Status.UpdatedNumberScheduled, \"nDNS:=\", newDS.Status.DesiredNumberScheduled, \"nNR:=\", newDS.Status.NumberReady, \"nNA:=\", newDS.Status.NumberAvailable)\n\tif newDS.Status.UpdatedNumberScheduled >= oldDS.Status.UpdatedNumberScheduled {\n\t\tif (newDS.Status.UpdatedNumberScheduled == newDS.Status.NumberReady) &&\n\t\t\t(newDS.Status.UpdatedNumberScheduled == newDS.Status.NumberAvailable) {\n\t\t\treturn true\n\t\t}\n\t}\n\tif e.MetaOld == nil {\n\t\tplog.Error(nil, \"Update event has no old metadata\", \"event\", e)\n\t\treturn false\n\t}\n\tif e.ObjectOld == nil {\n\t\tplog.Error(nil, \"Update event has no old runtime object to update\", \"event\", e)\n\t\treturn false\n\t}\n\tif e.ObjectNew == nil {\n\t\tplog.Error(nil, \"Update event has no new runtime object for update\", \"event\", e)\n\t\treturn false\n\t}\n\tif e.MetaNew == nil {\n\t\tplog.Error(nil, \"Update event has no new metadata\", \"event\", e)\n\t\treturn false\n\t}\n\tif e.MetaNew.GetGeneration() == e.MetaOld.GetGeneration() {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func HasUpdate(dir string) error {\n\th, err := hasUpdate(federalRevenueURL, dir)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error getting updated at: %w\", err)\n\t}\n\tif !h {\n\t\tos.Exit(1)\n\t}\n\treturn nil\n}", "func (x *fastReflection_MsgUpdateParamsResponse) IsValid() bool {\n\treturn x != nil\n}", "func (op *UpdateActiveDirectoryOperation) Done() bool {\n\treturn op.lro.Done()\n}", "func (op *UpdateVolumeOperation) Done() bool {\n\treturn op.lro.Done()\n}", "func (op *UpdateJobOperation) Done() bool {\n\treturn op.lro.Done()\n}", "func IsUpdate() bool {\n\tdataDir, err := FindDirectory(DataDirectory)\n\tif err != nil {\n\t\treturn false\n\t}\n\tif _, err := os.Stat(filepath.Join(dataDir, \"containerservice.yaml\")); err == nil {\n\t\treturn true\n\t}\n\treturn false\n}", "func (OnlyUpdateWhenGenerationNotChanged) Update(e event.UpdateEvent) bool {\n\tif e.MetaOld == nil {\n\t\tlog.Error(nil, \"Update event has no old metadata\", \"event\", e)\n\t\treturn false\n\t}\n\tif e.ObjectOld == nil {\n\t\tlog.Error(nil, \"Update event has no old runtime object to update\", \"event\", e)\n\t\treturn false\n\t}\n\tif e.ObjectNew == nil {\n\t\tlog.Error(nil, \"Update event has no new runtime object for update\", \"event\", e)\n\t\treturn false\n\t}\n\tif e.MetaNew == nil {\n\t\tlog.Error(nil, \"Update event has no new metadata\", \"event\", e)\n\t\treturn false\n\t}\n\treturn e.MetaNew.GetGeneration() == e.MetaOld.GetGeneration()\n}", "func (tc *Configs) Target(name string) (*Target, bool) {\n\tfilePrefix, target := splitTarget(name)\n\tfor _, tf := range tc.Files {\n\t\tif filePrefix != \"\" && tf.Basename() != filePrefix {\n\t\t\tcontinue\n\t\t}\n\t\ttarget, ok := tf.Targets[target]\n\t\tif ok {\n\t\t\treturn target, ok\n\t\t}\n\t}\n\treturn nil, false\n}", "func (a *AdminApiService) ModifyTarget(ctx _context.Context, id string, target Target) (Target, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPatch\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue Target\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/admin/target/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(parameterToString(id, \"\")) , -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &target\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 401 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 403 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 409 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 500 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (t *Target) EqualTo(other Target) bool {\n\treturn t.Host == other.Host &&\n\t\tt.Port == other.Port\n}", "func (o *SLOOverallStatuses) GetTargetOk() (*float64, bool) {\n\tif o == nil || o.Target == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Target, true\n}", "func (p OwnersOwnerInNamespacePredicate) Update(e event.UpdateEvent) bool {\n\treturn p.ownersOwnerInNamespace(e.MetaNew.GetOwnerReferences())\n}", "func FilterTarget(opt Options, store *core.StoreInfo, filters []Filter) bool {\n\tstoreID := fmt.Sprintf(\"store%d\", store.GetId())\n\tfor _, filter := range filters {\n\t\tif filter.FilterTarget(opt, store) {\n\t\t\tlog.Debugf(\"[filter %T] filters store %v from target\", filter, store)\n\t\t\tfilterCounter.WithLabelValues(\"filter-target\", storeID, filter.Type()).Inc()\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (s *RepositorySyncDefinition) SetTarget(v string) *RepositorySyncDefinition {\n\ts.Target = &v\n\treturn s\n}", "func (m *StorageMock) MinimockUpdateDone() bool {\n\tfor _, e := range m.UpdateMock.expectations {\n\t\tif mm_atomic.LoadUint64(&e.Counter) < 1 {\n\t\t\treturn false\n\t\t}\n\t}\n\n\t// if default expectation was set then invocations count should be greater than zero\n\tif m.UpdateMock.defaultExpectation != nil && mm_atomic.LoadUint64(&m.afterUpdateCounter) < 1 {\n\t\treturn false\n\t}\n\t// if func was set then invocations count should be greater than zero\n\tif m.funcUpdate != nil && mm_atomic.LoadUint64(&m.afterUpdateCounter) < 1 {\n\t\treturn false\n\t}\n\treturn true\n}", "func (pa *PodAutoscaler) Target() (float64, bool) {\n\treturn pa.annotationFloat64(autoscaling.TargetAnnotation)\n}", "func (agent *Agent) update_target_model() {\n\tagent.target_model = agent.model\n}", "func AutoUpdate(pjPath string, out output.Outputer) (updated bool, resultVersion string) {\n\tif versionInfo, _ := projectfile.ParseVersionInfo(pjPath); versionInfo != nil {\n\t\treturn false, \"\"\n\t}\n\n\t// Check for an update, but timeout after one second.\n\tlogging.Debug(\"Checking for updates.\")\n\tupdate := Updater{\n\t\tCurrentVersion: constants.Version,\n\t\tAPIURL: constants.APIUpdateURL,\n\t\tCmdName: constants.CommandName,\n\t}\n\tseconds := 1\n\tif secondsOverride := os.Getenv(constants.AutoUpdateTimeoutEnvVarName); secondsOverride != \"\" {\n\t\toverride, err := strconv.Atoi(secondsOverride)\n\t\tif err == nil {\n\t\t\tseconds = override\n\t\t}\n\t}\n\tinfo, err := timeout(update.Info, time.Duration(seconds)*time.Second)\n\tif err != nil {\n\t\tif err.Error() != \"timeout\" {\n\t\t\tlogging.Error(\"Unable to automatically check for updates: %s\", err)\n\t\t} else {\n\t\t\tlogging.Debug(\"Automatically checking for updates timed out\")\n\t\t}\n\t\treturn false, \"\"\n\t} else if info == nil {\n\t\tlogging.Debug(\"No update available.\")\n\t\treturn false, \"\"\n\t}\n\n\t// Self-update.\n\tlogging.Debug(\"Self-updating.\")\n\terr = update.Run(out, true)\n\tif err != nil {\n\t\tlog := logging.Error\n\t\tif os.IsPermission(errs.InnerError(err)) {\n\t\t\tout.Error(locale.T(\"auto_update_permission_err\"))\n\t\t}\n\t\tif errors.As(err, new(*osutils.AlreadyLockedError)) {\n\t\t\tlog = logging.Debug\n\t\t}\n\t\tlog(\"Unable to self update: %s\", err)\n\t\treturn false, \"\"\n\t}\n\n\treturn true, info.Version\n}", "func (c *Client) NeedsUpdate(ctx context.Context) (bool, error) {\n\tmeta, err := c.GetMetadata()\n\tif err != nil {\n\t\treturn true, nil\n\t}\n\n\t// No need to update if it's been within a day since the last update.\n\tif c.clock.Now().Before(meta.DownloadedAt.Add(updateInterval)) {\n\t\treturn false, nil\n\t}\n\n\tif err = c.populateOCIArtifact(); err != nil {\n\t\treturn false, xerrors.Errorf(\"OPA bundle error: %w\", err)\n\t}\n\n\tdigest, err := c.artifact.Digest(ctx)\n\tif err != nil {\n\t\treturn false, xerrors.Errorf(\"digest error: %w\", err)\n\t}\n\n\tif meta.Digest != digest {\n\t\treturn true, nil\n\t}\n\n\t// Update DownloadedAt with the current time.\n\t// Otherwise, if there are no updates in the remote registry,\n\t// the digest will be fetched every time even after this.\n\tif err = c.updateMetadata(meta.Digest, time.Now()); err != nil {\n\t\treturn false, xerrors.Errorf(\"unable to update the policy metadata: %w\", err)\n\t}\n\n\treturn false, nil\n}", "func (s *ResourceSyncAttempt) SetTarget(v string) *ResourceSyncAttempt {\n\ts.Target = &v\n\treturn s\n}", "func (s *peerRESTServer) TargetExistsHandler(w http.ResponseWriter, r *http.Request) {\n\tctx := newContext(r, w, \"TargetExists\")\n\tif !s.IsValid(w, r) {\n\t\ts.writeErrorResponse(w, errors.New(\"Invalid request\"))\n\t\treturn\n\t}\n\n\tvars := mux.Vars(r)\n\tbucketName := vars[peerRESTBucket]\n\tif bucketName == \"\" {\n\t\ts.writeErrorResponse(w, errors.New(\"Bucket name is missing\"))\n\t\treturn\n\t}\n\tvar targetID event.TargetID\n\tif r.ContentLength <= 0 {\n\t\ts.writeErrorResponse(w, errInvalidArgument)\n\t\treturn\n\t}\n\n\terr := gob.NewDecoder(r.Body).Decode(&targetID)\n\tif err != nil {\n\t\ts.writeErrorResponse(w, err)\n\t\treturn\n\t}\n\n\tvar targetExists remoteTargetExistsResp\n\ttargetExists.Exists = globalNotificationSys.RemoteTargetExist(bucketName, targetID)\n\n\tdefer w.(http.Flusher).Flush()\n\tlogger.LogIf(ctx, gob.NewEncoder(w).Encode(&targetExists))\n}", "func (s *DataStore) UpdateBackupTargetStatus(backupTarget *longhorn.BackupTarget) (*longhorn.BackupTarget, error) {\n\tobj, err := s.lhClient.LonghornV1beta2().BackupTargets(s.namespace).UpdateStatus(context.TODO(), backupTarget, metav1.UpdateOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tverifyUpdate(backupTarget.Name, obj, func(name string) (runtime.Object, error) {\n\t\treturn s.GetBackupTargetRO(name)\n\t})\n\treturn obj, nil\n}", "func (o *RequestStatusMetadata) GetTargetsOk() (*[]RequestTarget, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Targets, true\n}", "func (i *InvalidateDeadline) SetTarget(t time.Time) {\n\ti.Active = true\n\ti.Target = t\n}", "func (c *FakeAWSSNSTargets) UpdateStatus(ctx context.Context, aWSSNSTarget *v1alpha1.AWSSNSTarget, opts v1.UpdateOptions) (*v1alpha1.AWSSNSTarget, error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewUpdateSubresourceAction(awssnstargetsResource, \"status\", c.ns, aWSSNSTarget), &v1alpha1.AWSSNSTarget{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1alpha1.AWSSNSTarget), err\n}", "func (l Lambda) IsUpdatable() bool {\n\treturn l.updatable\n}", "func hasUpdateArg(args []string) bool {\n\tfor _, arg := range args {\n\t\tif ArgWorkflowUpdate == arg {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func vnodeNeedsUpdate(desired *appmeshv1beta1.VirtualNode, target *aws.VirtualNode) bool {\n\tif desired.Spec.ServiceDiscovery != nil &&\n\t\tdesired.Spec.ServiceDiscovery.Dns != nil {\n\t\t// If Service discovery is desired, verify target is equal\n\t\tif desired.Spec.ServiceDiscovery.Dns.HostName != target.HostName() {\n\t\t\treturn true\n\t\t}\n\t} else {\n\t\t// If no desired Service Discovery, verify target is not set\n\t\tif target.HostName() != \"\" {\n\t\t\treturn true\n\t\t}\n\t}\n\n\tif desired.Spec.Listeners != nil {\n\t\tdesiredSet := set.NewSet()\n\t\tfor i := range desired.Spec.Listeners {\n\t\t\tdesiredSet.Add(desired.Spec.Listeners[i])\n\t\t}\n\t\tcurrSet := target.ListenersSet()\n\t\tif !desiredSet.Equal(currSet) {\n\t\t\treturn true\n\t\t}\n\t} else {\n\t\t// If the spec doesn't have any listeners, make sure target is not set\n\t\tif len(target.Listeners()) != 0 {\n\t\t\treturn true\n\t\t}\n\t}\n\n\t// This needs to be updated since AppMesh VN name isn't the same as k8s VN name.\n\tif desired.Spec.Backends != nil {\n\t\tdesiredSet := set.NewSet()\n\t\tfor i := range desired.Spec.Backends {\n\t\t\tdesiredSet.Add(desired.Spec.Backends[i])\n\t\t}\n\t\tcurrSet := target.BackendsSet()\n\t\tif !desiredSet.Equal(currSet) {\n\t\t\treturn true\n\t\t}\n\t} else {\n\t\t// If the spec doesn't have any backends, make sure target is not set\n\t\tif len(target.Backends()) != 0 {\n\t\t\treturn true\n\t\t}\n\t}\n\n\tif vnodeLoggingNeedsUpdate(desired, target) {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (clt *SMServiceClient) IsDemandTarget(dm Demand, idlist []uint64) bool {\n\tdmid := dm.TargetId\n\tfor _, id := range idlist {\n\t\tif id == dmid {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func SatisfiesTargetValue(targetValue int64, minChange int64, utxos []*common.UTXO) bool {\n\ttotalValue := int64(0)\n\tfor _, utxo := range utxos {\n\t\ttotalValue += utxo.Value\n\t}\n\n\treturn (totalValue == targetValue || totalValue >= targetValue+minChange)\n}", "func (o *SLOOverallStatuses) SetTarget(v float64) {\n\to.Target = &v\n}", "func (req *DronaRequest) IsDnUpdate() bool {\n\treq.Lock()\n\tdefer req.Unlock()\n\treturn !req.processed\n}", "func (s *TestExecutionSummary) SetTarget(v *TestExecutionTarget) *TestExecutionSummary {\n\ts.Target = v\n\treturn s\n}", "func IsUpdateNeeded(filepath string, updateInterval time.Duration) (bool, error) {\n\tinfo, err := os.Stat(filepath)\n\tif os.IsNotExist(err) {\n\t\treturn true, ResetUpdateTime(filepath)\n\t} else if err != nil {\n\t\treturn false, err\n\t}\n\n\ttimeSinceMod := time.Since(info.ModTime())\n\n\tif timeSinceMod >= updateInterval {\n\t\treturn true, nil\n\t}\n\n\treturn false, nil\n}", "func (p *ClientUpdatePoller) Done() bool {\n\treturn p.pt.Done()\n}" ]
[ "0.6614034", "0.64920354", "0.6105699", "0.60342765", "0.6030683", "0.59310377", "0.5721925", "0.56897306", "0.56685436", "0.5625591", "0.5597833", "0.55948585", "0.55275893", "0.5479221", "0.5437466", "0.5431092", "0.5352646", "0.53513056", "0.5247706", "0.52319986", "0.5218859", "0.5126008", "0.5123713", "0.5102578", "0.50980705", "0.50980705", "0.5077669", "0.5068585", "0.504985", "0.5047485", "0.5029949", "0.50288457", "0.49995494", "0.49883005", "0.49707618", "0.4969256", "0.49562407", "0.49236235", "0.49181753", "0.4887698", "0.48850834", "0.48804533", "0.48715234", "0.48534343", "0.48514205", "0.48442078", "0.48311225", "0.48306647", "0.48262447", "0.48238328", "0.48171017", "0.479466", "0.47933492", "0.478729", "0.47868702", "0.47794142", "0.47716722", "0.4770499", "0.4768803", "0.4763252", "0.47622558", "0.4748089", "0.47319576", "0.4730715", "0.47262576", "0.47218654", "0.47106868", "0.4709026", "0.47065383", "0.47061", "0.47012338", "0.46982333", "0.46900722", "0.4689173", "0.4662192", "0.4657857", "0.46552193", "0.46472454", "0.4637955", "0.46331123", "0.46259266", "0.4620435", "0.46172634", "0.46153527", "0.46128142", "0.46062654", "0.46053547", "0.46003383", "0.45993", "0.45987377", "0.45937958", "0.45933825", "0.4590953", "0.45874238", "0.45827252", "0.45759052", "0.4574269", "0.45631516", "0.45609513", "0.45585743" ]
0.73240125
0
NewPoller returns an instance of Poller but do not start polling goroutine
func NewPoller(url string, interval time.Duration, out chan PollMsg, shutdown chan *sync.WaitGroup) *Poller { return &Poller{ URL: url, Interval: interval, Out: out, Shutdown: shutdown, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewPoller() *Poller {\n\treturn &Poller{}\n}", "func NewPoller() *Poller {\n\treturn &Poller{\n\t\tinterval: DefaultPollingInterval,\n\t}\n}", "func NewPoller(getFunc GetFunc, period time.Duration, store Store) *Poller {\n\treturn &Poller{\n\t\tgetFunc: getFunc,\n\t\tperiod: period,\n\t\tstore: store,\n\t}\n}", "func NewPoller(poll PollerFunc, interval time.Duration) *Poller {\n\treturn &Poller{\n\t\tChannel: make(chan interface{}),\n\t\tPoll: poll,\n\t\tWaitInterval: interval,\n\t\tisStopped: false,\n\t\tisFinished: false,\n\t\tgroup: &sync.WaitGroup{},\n\t\tstopMutex: &sync.Mutex{},\n\t}\n}", "func NewPoller() (*Epoll, error) {\n\treturn NewPollerWithBuffer(128)\n}", "func New() (*Poller, error) {\n\tfd, err := unix.EpollCreate1(unix.EPOLL_CLOEXEC)\n\tif err != nil {\n\t\terr = os.NewSyscallError(\"epoll_create1\", err)\n\t\tlog.WithError(err).Error(\"failed to create an epoll instance\")\n\t\treturn nil, err\n\t}\n\n\tp := &Poller{\n\t\tepfd: fd,\n\t}\n\n\treturn p, nil\n}", "func New() *PollerRegistry {\n\treturn &PollerRegistry{\n\t\tRegistry: make(map[string]*poller.Poller),\n\t\tToDB: make(chan interface{}),\n\t\tUpdateStatus: make(chan string),\n\t}\n}", "func New(p *Page) (Poller, error) {\n\tid := \"\"\n\tl := log.With().Str(\"func\", \"poller.New\").Logger()\n\n\tif p.ID != nil && len(*p.ID) > 0 {\n\t\tid = *p.ID\n\t} else {\n\t\tid = randomdata.SillyName()\n\t\tl.Info().Msg(\"generating random name...\")\n\t}\n\tl = l.With().Str(\"id\", id).Logger()\n\n\t// -- Validation\n\tmethod, err := parseHTTPMethod(p.Method)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tparsedURL, err := parseURL(p.URL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// -- Set ups\n\trandomFrequency, ticks, offset := parsePollOptions(id, p.PollOptions)\n\tif randomFrequency {\n\t\tticks = nextRandomTick(ticks-offset, ticks+offset)\n\t}\n\n\trandUA, userAgents := parseUserAgentOptions(id, p.UserAgentOptions)\n\n\theaders := http.Header{}\n\tif p.Headers == nil {\n\t\tl.Warn().Msg(\"no headers provided\")\n\t} else {\n\t\tfor headerKey, headerVal := range p.Headers {\n\t\t\theaders[headerKey] = []string{headerVal}\n\t\t}\n\t\tswitch hlen := len(p.Headers); {\n\t\tcase hlen == 0:\n\t\t\tl.Warn().Msg(\"no headers provided\")\n\t\tcase hlen < 3:\n\t\t\tl.Warn().Msg(\"few headers provided\")\n\t\t}\n\t}\n\n\thttpClient := &http.Client{\n\t\tTimeout: time.Duration(defaultHTTPClientTimeout) * time.Second,\n\t\tCheckRedirect: func(req *http.Request, via []*http.Request) error {\n\t\t\treturn http.ErrUseLastResponse\n\t\t},\n\t}\n\tif p.FollowRedirect {\n\t\thttpClient.CheckRedirect = nil\n\t}\n\n\trequest, err := http.NewRequestWithContext(context.Background(), method, parsedURL.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\trequest.Header = headers\n\n\t// -- Complete and return\n\treturn &pagePoller{\n\t\tid: id,\n\t\thttpClient: httpClient,\n\t\trequest: request,\n\t\tuserAgents: userAgents,\n\t\tticks: ticks,\n\t\trandTick: randomFrequency,\n\t\toffsetRange: offset,\n\t\tlastUAIndex: -1,\n\t\trandUa: randUA,\n\t}, nil\n}", "func New[T any](pl exported.Pipeline, resp *http.Response) (*Poller[T], error) {\n\tif resp == nil {\n\t\tlog.Write(log.EventLRO, \"Resuming Core-Fake-Poller poller.\")\n\t\treturn &Poller[T]{pl: pl}, nil\n\t}\n\n\tlog.Write(log.EventLRO, \"Using Core-Fake-Poller poller.\")\n\tfakeStatus := resp.Header.Get(shared.HeaderFakePollerStatus)\n\tif fakeStatus == \"\" {\n\t\treturn nil, errors.New(\"response is missing Fake-Poller-Status header\")\n\t}\n\n\tctxVal := resp.Request.Context().Value(shared.CtxAPINameKey{})\n\tif ctxVal == nil {\n\t\treturn nil, errors.New(\"missing value for CtxAPINameKey\")\n\t}\n\n\tapiName, ok := ctxVal.(string)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"expected string for CtxAPINameKey, the type was %T\", ctxVal)\n\t}\n\n\tqp := \"\"\n\tif resp.Request.URL.RawQuery != \"\" {\n\t\tqp = \"?\" + resp.Request.URL.RawQuery\n\t}\n\n\tp := &Poller[T]{\n\t\tpl: pl,\n\t\tresp: resp,\n\t\tAPIName: apiName,\n\t\t// NOTE: any changes to this path format MUST be reflected in SanitizePollerPath()\n\t\tFakeURL: fmt.Sprintf(\"%s://%s%s%s%s\", resp.Request.URL.Scheme, resp.Request.URL.Host, resp.Request.URL.Path, lroStatusURLSuffix, qp),\n\t\tFakeStatus: fakeStatus,\n\t}\n\treturn p, nil\n}", "func New(c *Config) (Poller, error) {\n\tcfg := c.withDefaults()\n\n\tkq, err := KqueueCreate(&KqueueConfig{\n\t\tOnWaitError: cfg.OnWaitError,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn poller{kq}, nil\n}", "func (p *StreamingEndpointsCreatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func NewPollerJob(resourceUrl string) PollerJob{\n\treturn PollerJob{\n\t\tresourceUrl:resourceUrl,\n\t\tsuspend: make(chan bool),\n\t\tresume: make(chan bool),\n\t}\n}", "func New(cache *ipcache.RecentIPCache) ConnectionPoller {\n\tc := &connectionPoller{\n\t\tfinder: &ssFinder{},\n\t\trecentIPCache: cache,\n\t\tconnectionPool: make(map[connection.Connection]struct{}),\n\t}\n\tc.GetConnections()\n\treturn c\n}", "func (p *Poller) Start() {\n\tgo func(url string, interval time.Duration, out chan<- PollMsg, shutdown <-chan *sync.WaitGroup) {\n\t\twg := &sync.WaitGroup{}\n\t\tdefer func() {\n\t\t\twg.Done()\n\t\t}()\n\n\t\tticker := time.NewTicker(interval)\n\t\tmsg := PollMsg{}\n\t\tresp, err := http.Get(url)\n\t\tif err != nil {\n\t\t\tmsg.Error = append(msg.Error, err)\n\t\t}\n\n\t\tdata, err2 := ioutil.ReadAll(resp.Body)\n\t\tif err2 != nil {\n\t\t\tmsg.Error = append(msg.Error, err2)\n\t\t}\n\n\t\tmsg.Payload = data\n\t\tout <- msg\n\n\t\tresp.Body.Close()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase wg = <-shutdown:\n\t\t\t\treturn\n\t\t\tcase <-ticker.C:\n\t\t\t\tmsg := PollMsg{}\n\t\t\t\tresp, err := http.Get(url)\n\t\t\t\tif err != nil {\n\t\t\t\t\tmsg.Error = append(msg.Error, err)\n\t\t\t\t}\n\n\t\t\t\tdata, err2 := ioutil.ReadAll(resp.Body)\n\t\t\t\tif err2 != nil {\n\t\t\t\t\tmsg.Error = append(msg.Error, err2)\n\t\t\t\t}\n\n\t\t\t\tmsg.Payload = data\n\t\t\t\tout <- msg\n\n\t\t\t\tresp.Body.Close()\n\t\t\t}\n\t\t}\n\t}(p.URL, p.Interval, p.Out, p.Shutdown)\n}", "func (f *Input) startPoller(ctx context.Context) {\n\tf.wg.Add(1)\n\tgo func() {\n\t\tdefer f.wg.Done()\n\t\tglobTicker := time.NewTicker(f.PollInterval)\n\t\tdefer globTicker.Stop()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn\n\t\t\tcase <-globTicker.C:\n\t\t\t}\n\n\t\t\tf.poll(ctx)\n\t\t}\n\t}()\n}", "func NewLogPollerWrapper(t mockConstructorTestingTNewLogPollerWrapper) *LogPollerWrapper {\n\tmock := &LogPollerWrapper{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func newManualPollingPolicy(\n\tconfigFetcher configProvider,\n\tstore *configStore,\n\tlogger Logger) *manualPollingPolicy {\n\n\treturn &manualPollingPolicy{configRefresher: configRefresher{configFetcher: configFetcher, store: store, logger: logger}}\n}", "func (h *Handler) UpdatePoller(c echo.Context) error {\n\tid := c.Param(\"id\")\n\trequest := &PollerRequest{}\n\tvar err error\n\tif err = c.Bind(request); err != nil {\n\t\treturn err\n\t}\n\n\tdb := h.DB.Clone()\n\tdefer db.Close()\n\n\tnewInterval, err := time.ParseDuration(request.PollInterval)\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\tif err := db.DB(\"oxylus\").C(\"pollers\").Update(\n\t\tbson.M{\"uuid\": id},\n\t\tbson.M{\"$set\": bson.M{\n\t\t\t\"ispolling\": request.IsPolling,\n\t\t\t\"action\": request.Action,\n\t\t\t\"pollinterval\": newInterval}}); err != nil {\n\t\treturn err\n\t}\n\n\tif request.IsPolling {\n\t\tvar p poller.Poller\n\t\tif err := db.DB(\"oxylus\").C(\"pollers\").Find(bson.M{\"uuid\": id}).One(&p); err != nil {\n\t\t\treturn err\n\t\t}\n\t\th.PollerRegistry.Add(id, &p)\n\t} else {\n\t\th.PollerRegistry.Remove(id)\n\t}\n\n\treturn c.NoContent(http.StatusOK)\n}", "func (p *Poller[T]) Poll(ctx context.Context) (*http.Response, error) {\n\tctx = context.WithValue(ctx, shared.CtxAPINameKey{}, p.APIName)\n\terr := pollers.PollHelper(ctx, p.FakeURL, p.pl, func(resp *http.Response) (string, error) {\n\t\tif !poller.StatusCodeValid(resp) {\n\t\t\tp.resp = resp\n\t\t\treturn \"\", exported.NewResponseError(resp)\n\t\t}\n\t\tfakeStatus := resp.Header.Get(shared.HeaderFakePollerStatus)\n\t\tif fakeStatus == \"\" {\n\t\t\treturn \"\", errors.New(\"response is missing Fake-Poller-Status header\")\n\t\t}\n\t\tp.resp = resp\n\t\tp.FakeStatus = fakeStatus\n\t\treturn p.FakeStatus, nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn p.resp, nil\n}", "func (p *NotebookWorkspacesClientStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *StreamingEndpointsStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *Poller) Start() {\n\tif p.IsStopped() {\n\t\tpanic(\"cannot restart poller after it has been stopped\")\n\t}\n\n\tp.group.Add(1)\n\tgo p.runPolling()\n}", "func Poller(in <-chan *Resource, out chan<- *Resource, status chan<- State) {\n\tfor r := range in {\n\t\ts := r.Poll()\n\t\tstatus <- State{r.url, s}\n\t\tout <- r\n\t}\n}", "func (p *PrivateEndpointConnectionsCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *Poller) Poll(ctx context.Context, pf PollFunc) (interface{}, error) {\n\tif p.timeout > 0 {\n\t\tctxWithTimeout, cancel := context.WithTimeout(ctx, p.timeout)\n\t\tdefer cancel()\n\t\tctx = ctxWithTimeout\n\t}\n\n\tticker := time.NewTicker(p.interval)\n\tdefer ticker.Stop()\n\n\tfor {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\tdone, res, err := pf(ctx)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tif !done {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\treturn res, nil\n\n\t\tcase <-ctx.Done():\n\t\t\treturn nil, ctx.Err()\n\t\t}\n\t}\n}", "func (p *PrivateEndpointConnectionsClientCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *LiveEventsCreatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func NewPollee(url string, httpMethod string, maxErrorCount int, successStatus string, responseChannel chan<- *PollResponse) (*Pollee, error) {\n\tvar netTransport = &http.Transport{\n\t\tDial: (&net.Dialer{\n\t\t\tTimeout: DefaultDialTimeout,\n\t\t}).Dial,\n\t\tTLSHandshakeTimeout: DefaultTLSHandshakeTimeout,\n\t}\n\tvar netClient = &http.Client{\n\t\tTimeout: DefaultConnectionTimeout,\n\t\tTransport: netTransport,\n\t}\n\treq, err := http.NewRequest(httpMethod, url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Pollee{\n\t\turl: url,\n\t\tErrorCount: 0,\n\t\tMaxErrorCount: maxErrorCount,\n\t\tSuccessStatus: successStatus,\n\t\tclient: netClient,\n\t\trequest: req,\n\t\tpollResponseChannel: responseChannel,\n\t}, nil\n}", "func (p *ServicesCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *ServicesCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *SharedPrivateLinkResourcesCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (g *GardenerAPI) Poll(ctx context.Context,\n\ttoRunnable func(o *storage.ObjectAttrs) Runnable, maxWorkers int, period time.Duration) {\n\t// Poll no faster than period.\n\tticker := time.NewTicker(period)\n\tthrottle := NewWSTokenSource(maxWorkers)\n\tfor {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\tlog.Println(\"Poller context done\")\n\t\t\treturn\n\t\tdefault:\n\t\t\terr := g.pollAndRun(ctx, toRunnable, throttle)\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(err)\n\t\t\t}\n\t\t}\n\n\t\t<-ticker.C // Wait for next tick, to avoid fast spinning on errors.\n\t}\n}", "func (p *NotebookWorkspacesStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *ServicesStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *rpioPoller) poll() {\npollLoop:\n\tfor {\n\t\tselect {\n\t\tcase <-p.ticker.C:\n\t\t\t// Read pins and handle edge detection\n\t\t\tfor pin, registration := range p.registeredPins {\n\t\t\t\tif pin.EdgeDetected() {\n\t\t\t\t\tgo registration.callback(registration.edge)\n\t\t\t\t}\n\t\t\t}\n\t\tcase newRegistration := <-p.newPin:\n\t\t\t// Add pin registration to pins to poll\n\t\t\tp.registeredPins[newRegistration.pin] = newRegistration\n\t\tcase registrationToRemove := <-p.removePin:\n\t\t\t// Remove pin registration from pins to poll\n\t\t\tdelete(p.registeredPins, registrationToRemove)\n\t\tcase newPollFreq := <-p.newPollFreq:\n\t\t\t// Update the ticker polling frequency\n\t\t\tp.ticker.Reset(newPollFreq)\n\t\tcase <-p.stop:\n\t\t\tbreak pollLoop\n\t\t}\n\t}\n}", "func (p *NotebookWorkspacesClientCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *StreamingEndpointsStopPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *AgentPoolsCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *SQLResourcesCreateUpdateSQLStoredProcedurePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (s *Store) startPoller() {\n\tfor {\n\t\tselect {\n\t\tcase req := <-s.getReqQueue:\n\t\t\treq.respCh <- s.performGetOperation(req.key)\n\n\t\tcase req := <-s.modifyReqQueue:\n\t\t\terr := s.performModifyOperation(req)\n\t\t\treq.respCh <- err\n\n\t\t\ts.fanOutSubscriptions(req)\n\n\t\tcase sub := <-s.subscribeQueue:\n\t\t\ts.registerSubscription(sub)\n\t\t}\n\t}\n}", "func (TelegramBotApp *TelegramBotApp) setupPolling() (tgbotapi.UpdatesChannel, error) {\n\tTelegramBotApp.bot.RemoveWebhook()\n\tupdateConfig := tgbotapi.NewUpdate(0)\n\tupdateConfig.Timeout = 5\n\tfmt.Println(\"[+] Pooling method selected\")\n\treturn TelegramBotApp.bot.GetUpdatesChan(updateConfig)\n}", "func (p *SQLResourcesClientCreateUpdateSQLStoredProcedurePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *LiveOutputsCreatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func NewPollingListener(token string) (PollingListener, MessageSender) {\n\treturn PollingListener{\n\t\ts: make(chan interface{}),\n\t\tToken: token,\n\t\tr: MessageSender{token: token},\n\t}, MessageSender{token: token}\n}", "func (p *GremlinResourcesCreateUpdateGremlinGraphPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *ClientCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (h *Handler) AddPoller(c echo.Context) error {\n\tid := c.Param(\"id\")\n\trequest := &PollerRequest{}\n\tvar err error\n\tif err = c.Bind(request); err != nil {\n\t\treturn err\n\t}\n\tp := &poller.Poller{}\n\tp.UUID = uuid.NewV4().String()\n\tp.Action = request.Action\n\tp.Driver = &particleio.ParticleIO{\n\t\tUUID: p.UUID,\n\t\tDeviceID: request.DeviceID,\n\t\tAccessToken: request.AccessToken,\n\t}\n\tp.PollInterval, err = time.ParseDuration(request.PollInterval)\n\tif err != nil {\n\t\treturn err\n\t}\n\tp.IsPolling = request.IsPolling\n\tp.User = id\n\n\tdb := h.DB.Clone()\n\tdefer db.Close()\n\n\tif err := db.DB(\"oxylus\").C(\"pollers\").Insert(&p); err != nil {\n\t\treturn err\n\t}\n\t// if ispolling then send the poller to the registry\n\t// turn this into a channel\n\th.PollerRegistry.Add(p.UUID, p)\n\treturn c.NoContent(http.StatusCreated)\n}", "func NewPollDriver(log gopi.Logger) (*PollDriver, error) {\n\tvar err error\n\n\tlog.Debug(\"<linux.Poll>Open\")\n\n\tthis := new(PollDriver)\n\tif this.handle, err = syscall.EpollCreate1(syscall.EPOLL_CLOEXEC); err != nil {\n\t\treturn nil, err\n\t}\n\tthis.events = make(map[int]uint32)\n\tthis.log = log\n\n\t// success\n\treturn this, nil\n}", "func (p *LiveEventsStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *VaultsCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func NewPoll() (record *Poll) {\n\treturn new(Poll)\n}", "func (p *LiveEventsResetPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *DeploymentsStartJFRPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *CertificatesCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *GremlinResourcesClientCreateUpdateGremlinGraphPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *restorePoller[T]) Poll(ctx context.Context) (*http.Response, error) {\n\terr := pollHelper(ctx, p.AsyncURL, p.pl, func(resp *http.Response) (string, error) {\n\t\tif !poller.StatusCodeValid(resp) {\n\t\t\tp.resp = resp\n\t\t\treturn \"\", runtime.NewResponseError(resp)\n\t\t}\n\t\tstate, err := poller.GetStatus(resp)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t} else if state == \"\" {\n\t\t\treturn \"\", errors.New(\"the response did not contain a status\")\n\t\t}\n\t\tp.resp = resp\n\t\tp.CurState = state\n\t\treturn p.CurState, nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn p.resp, nil\n}", "func NewPollerWithBuffer(count int) (*Epoll, error) {\n\tfd, err := unix.EpollCreate1(0)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Epoll{\n\t\tfd: fd,\n\t\tlock: &sync.RWMutex{},\n\t\tconns: make(map[int]net.Conn),\n\t\tconnbuf: make([]net.Conn, count, count),\n\t\tevents: make([]unix.EpollEvent, count, count),\n\t}, nil\n}", "func (p *ServicesStopPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *StreamingEndpointsScalePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *LiveEventsStopPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func NewRestorePoller[T any](pl runtime.Pipeline, resp *http.Response, finalState runtime.FinalStateVia) (*restorePoller[T], error) {\n\tif resp == nil {\n\t\treturn &restorePoller[T]{pl: pl}, nil\n\t}\n\t//log.Write(log.EventLRO, \"Using Azure-AsyncOperation poller.\")\n\tasyncURL := resp.Header.Get(\"Azure-AsyncOperation\")\n\tif asyncURL == \"\" {\n\t\treturn nil, errors.New(\"response is missing Azure-AsyncOperation header\")\n\t}\n\tif !poller.IsValidURL(asyncURL) {\n\t\treturn nil, fmt.Errorf(\"invalid polling URL %s\", asyncURL)\n\t}\n\t// check for provisioning state. if the operation is a RELO\n\t// and terminates synchronously this will prevent extra polling.\n\t// it's ok if there's no provisioning state.\n\tstate, _ := poller.GetProvisioningState(resp)\n\tif state == \"\" {\n\t\tstate = poller.StatusInProgress\n\t}\n\tp := &restorePoller[T]{\n\t\tpl: pl,\n\t\tresp: resp,\n\t\tAsyncURL: asyncURL,\n\t\tLocURL: resp.Header.Get(\"Location\"),\n\t\tOrigURL: resp.Request.URL.String(),\n\t\tMethod: resp.Request.Method,\n\t\tFinalState: finalState,\n\t\tCurState: state,\n\t}\n\treturn p, nil\n}", "func (p *DeploymentsStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *NotebookWorkspacesCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *Poller) runPolling() {\n\tfor !p.IsStopped() {\n\t\tif res := p.Poll(); res != nil {\n\t\t\tp.sendResult(res)\n\t\t}\n\t\ttime.Sleep(p.WaitInterval)\n\t}\n\tp.group.Done()\n\tp.isFinished = true\n}", "func (p *EventSubscriptionsCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *Pollee) Poll() {\n\tlog.Infof(\"ErrorCount %d, MaxErrorCount %d\", p.ErrorCount, p.MaxErrorCount)\n\tif p.ErrorCount >= p.MaxErrorCount {\n\t\tp.sendPollResponse(\"\", ErrorMaxCountExceeded)\n\t\treturn\n\t}\n\tp.ErrorCount++\n\tlog.Debugf(\"Polling %s ....\", p.url)\n\tresp, err := p.client.Do(p.request)\n\tif err != nil {\n\t\tlog.Debugf(\"Failed polling %s with %v\", p.url, err)\n\t\tp.sendPollResponse(\"\", err)\n\t}\n\trespStatus := \"\"\n\tif resp != nil {\n\t\tdefer resp.Body.Close()\n\t\trespStatus = resp.Status\n\t\tlog.Debugf(\"Response statue %s\", respStatus)\n\t}\n\tif respStatus == p.SuccessStatus {\n\t\tlog.Debugf(\"Expected %s. Got %s\", p.SuccessStatus, respStatus)\n\t\tp.ErrorCount = 0\n\t}\n\tp.sendPollResponse(respStatus, nil)\n}", "func (p *DeploymentsClientWhatIfPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *DeploymentsStopPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (r *rPIO) Poll() {\n\tif r.polling {\n\t\t// Only poll once\n\t\treturn\n\t}\n\n\t// Start polling\n\tgo r.poller.poll()\n\n\tr.polling = true\n}", "func NewPoll(address common.Address, backend bind.ContractBackend) (*Poll, error) {\n\tcontract, err := bindPoll(address, backend, backend, backend)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Poll{PollCaller: PollCaller{contract: contract}, PollTransactor: PollTransactor{contract: contract}, PollFilterer: PollFilterer{contract: contract}}, nil\n}", "func (p *StreamingEndpointsUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *DeploymentsClientCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func startPollWorker(site *ElectionSite) *pollWorker {\n\n\tworker := &pollWorker{site: site,\n\t\tballot: nil,\n\t\tkillch: make(chan bool, 1), // make sure sender won't block\n\t\tlistench: make(chan *Ballot, 1)} // make sure sender won't block\n\n\tgo worker.listen()\n\n\treturn worker\n}", "func NewLogPoller(orm *ORM, ec Client, lggr logger.Logger, pollPeriod time.Duration,\n\tfinalityDepth int64, backfillBatchSize int64, rpcBatchSize int64, keepBlocksDepth int64) *logPoller {\n\n\treturn &logPoller{\n\t\tec: ec,\n\t\torm: orm,\n\t\tlggr: lggr.Named(\"LogPoller\"),\n\t\treplayStart: make(chan int64),\n\t\treplayComplete: make(chan error),\n\t\tpollPeriod: pollPeriod,\n\t\tfinalityDepth: finalityDepth,\n\t\tbackfillBatchSize: backfillBatchSize,\n\t\trpcBatchSize: rpcBatchSize,\n\t\tkeepBlocksDepth: keepBlocksDepth,\n\t\tfilters: make(map[string]Filter),\n\t\tfilterDirty: true, // Always build Filter on first call to cache an empty filter if nothing registered yet.\n\t}\n}", "func (p *LiveEventsAllocatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *ManagedClustersStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *ManagedClustersRunCommandPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (ex *ExchangeClient) LongPoll(stopPollCh, stopAllCh chan int, errCh chan error) {\n\tif ex.pi == nil {\n\t\tpanic(\"No pi in ex\")\n\t}\n\tex.Info(\"Starting LongPoll|msgCode=POLLING\")\n\tdefer Utils.RecoverCrash(ex.logger) // catch all panic. RecoverCrash logs information needed for debugging.\n\tex.wg.Add(1)\n\tdefer ex.wg.Done()\n\n\tdefer func() {\n\t\tex.Info(\"Stopping LongPoll...\")\n\t\tex.cancel()\n\t}()\n\n\tvar err error\n\treqTimeout := ex.pi.ResponseTimeout\n\treqTimeout += uint64(float64(reqTimeout) * 0.1) // add 10% so we don't step on the HeartbeatInterval inside the ping\n\n\tif err != nil {\n\t}\n\tex.transport = &http.Transport{\n\t\tTLSClientConfig: &tls.Config{\n\t\t\tInsecureSkipVerify: false,\n\t\t\tRootCAs: globals.config.RootCerts(),\n\t\t},\n\t\tResponseHeaderTimeout: time.Duration(reqTimeout) * time.Millisecond,\n\t}\n\n\t// check for the proxy setting. Useful for mitmproxy testing\n\tproxy := os.Getenv(\"PINGER_PROXY\")\n\tif proxy != \"\" {\n\t\tproxyUrl, err := url.Parse(proxy)\n\t\tif err != nil {\n\t\t\tex.sendError(errCh, err)\n\t\t\treturn\n\t\t}\n\t\tex.transport.Proxy = http.ProxyURL(proxyUrl)\n\t}\n\n\tex.httpClient = &http.Client{\n\t\tTransport: ex.transport,\n\t}\n\tuseCookieJar := false\n\tif useCookieJar {\n\t\tcookieJar, err := cookiejar.New(nil)\n\t\tif err != nil {\n\t\t\tex.sendError(errCh, err)\n\t\t\treturn\n\t\t}\n\t\tex.httpClient.Jar = cookieJar\n\t}\n\tredactedUrl := strings.Split(ex.pi.MailServerUrl, \"?\")[0]\n\n\tex.Info(\"New HTTP Client with timeout %s %s<redacted>\", ex.transport.ResponseHeaderTimeout, redactedUrl)\n\tsleepTime := 0\n\ttooFastResponse := (time.Duration(ex.pi.ResponseTimeout) * time.Millisecond) / 4\n\tex.Debug(\"TooFast timeout set to %s\", tooFastResponse)\n\tvar responseCh chan *http.Response\n\tvar responseErrCh chan error\n\tfor {\n\t\tif sleepTime > 0 {\n\t\t\ts := time.Duration(sleepTime) * time.Second\n\t\t\tex.Info(\"Sleeping %s before retry\", s)\n\t\t\ttime.Sleep(s)\n\t\t}\n\t\tif responseErrCh != nil {\n\t\t\tclose(responseErrCh)\n\t\t}\n\t\tresponseErrCh = make(chan error)\n\t\tif responseCh != nil {\n\t\t\tclose(responseCh)\n\t\t}\n\t\tresponseCh = make(chan *http.Response)\n\n\t\ttimeSent := time.Now()\n\t\tex.wg.Add(1)\n\t\tex.cancelled = false\n\t\tgo ex.doRequestResponse(responseCh, responseErrCh)\n\t\tselect {\n\t\tcase err = <-responseErrCh:\n\t\t\tif err == NoSuchHostError || err == UnknownCertificateAuthority {\n\t\t\t\terrCh <- LongPollReRegister\n\t\t\t} else {\n\t\t\t\tex.sendError(errCh, err)\n\t\t\t}\n\t\t\treturn\n\n\t\tcase response := <-responseCh:\n\t\t\tif response == retryResponse {\n\t\t\t\tex.Debug(\"Retry-response from response reader.\")\n\t\t\t\tsleepTime = ex.exponentialBackoff(sleepTime)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// the response body tends to be pretty short (and we've capped it anyway). Let's just read it all.\n\t\t\tresponseBody, err := ioutil.ReadAll(response.Body)\n\t\t\tif err != nil {\n\t\t\t\tresponse.Body.Close() // attempt to close. Ignore any errors.\n\t\t\t\tex.sendError(errCh, err)\n\t\t\t\treturn\n\t\t\t}\n\t\t\terr = response.Body.Close()\n\t\t\tif err != nil {\n\t\t\t\tex.sendError(errCh, err)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tswitch {\n\t\t\tcase response.StatusCode != 200:\n\t\t\t\tswitch {\n\t\t\t\tcase response.StatusCode == 401:\n\t\t\t\t\t// ask the client to re-register, since nothing we could do would fix this\n\t\t\t\t\tex.Info(\"401 response. Telling client to re-register|msgCode=EAS_AUTH_ERR_REREGISTER\")\n\t\t\t\t\terrCh <- LongPollReRegister\n\t\t\t\t\treturn\n\n\t\t\t\tdefault:\n\t\t\t\t\t// just retry\n\t\t\t\t\tsleepTime = ex.exponentialBackoff(sleepTime)\n\t\t\t\t\tex.Info(\"Response Status %s. Back to polling\", response.Status)\n\t\t\t\t}\n\t\t\t\t//EAS Ping\n\t\t\tcase ex.pi.ASIsSyncRequest == false && (ex.pi.NoChangeReply != nil && bytes.Compare(responseBody, ex.pi.NoChangeReply) == 0):\n\t\t\t\t// go back to polling\n\t\t\t\tif time.Since(timeSent) <= tooFastResponse {\n\t\t\t\t\tex.Warning(\"Ping: NoChangeReply was too fast. Doing backoff. This usually indicates that the client is still connected to the exchange server.\")\n\t\t\t\t\tsleepTime = ex.exponentialBackoff(sleepTime)\n\t\t\t\t} else {\n\t\t\t\t\tex.Info(\"Ping: NoChangeReply after %s. Back to polling\", time.Since(timeSent))\n\t\t\t\t\tsleepTime = 0 // good reply. Reset any exponential backoff stuff.\n\t\t\t\t}\n\t\t\t\t// EAS Ping\n\t\t\tcase ex.pi.ASIsSyncRequest == false && (ex.pi.ExpectedReply == nil || bytes.Compare(responseBody, ex.pi.ExpectedReply) == 0):\n\t\t\t\t// there's new mail!\n\t\t\t\tif ex.pi.ExpectedReply != nil {\n\t\t\t\t\tex.Debug(\"Ping: Reply matched ExpectedReply\")\n\t\t\t\t}\n\t\t\t\tex.Debug(\"Ping: Got mail. Setting LongPollNewMail|msgCode=EAS_NEW_EMAIL\")\n\t\t\t\terrCh <- LongPollNewMail\n\t\t\t\treturn\n\t\t\t\t// EAS Sync\n\t\t\tcase ex.pi.ASIsSyncRequest == true && len(responseBody) == 0:\n\t\t\t\t// go back to polling\n\t\t\t\tif time.Since(timeSent) <= tooFastResponse {\n\t\t\t\t\tex.Warning(\"Sync: NoChangeReply after %s was too fast. Doing backoff. This usually indicates that the client is still connected to the exchange server.\", time.Since(timeSent))\n\t\t\t\t\tsleepTime = ex.exponentialBackoff(sleepTime)\n\t\t\t\t} else {\n\t\t\t\t\tex.Info(\"Sync: NoChangeReply after %s. Back to polling\", time.Since(timeSent))\n\t\t\t\t\tsleepTime = 0 // good reply. Reset any exponential backoff stuff.\n\t\t\t\t}\n\n\t\t\tcase ex.pi.ASIsSyncRequest == true && len(responseBody) > 0:\n\t\t\t\t// there's new mail!\n\t\t\t\tif ex.pi.ExpectedReply != nil {\n\t\t\t\t\tex.Debug(\"Sync: Reply matched ExpectedReply\")\n\t\t\t\t}\n\t\t\t\tex.Debug(\"Sync: Got mail. Setting LongPollNewMail|msgCode=EAS_NEW_EMAIL\")\n\t\t\t\terrCh <- LongPollNewMail\n\t\t\t\treturn\n\t\t\tdefault:\n\t\t\t\tex.Warning(\"Unhandled response. Just keep polling: Headers:%+v Body:%s\", response.Header, base64.StdEncoding.EncodeToString(responseBody))\n\t\t\t\tsleepTime = ex.exponentialBackoff(sleepTime)\n\t\t\t}\n\n\t\tcase <-stopPollCh: // parent will close this, at which point this will trigger.\n\t\t\tex.Debug(\"Was told to stop. Stopping\")\n\t\t\treturn\n\n\t\tcase <-stopAllCh: // parent will close this, at which point this will trigger.\n\t\t\tex.Debug(\"Was told to stop (allStop). Stopping\")\n\t\t\treturn\n\t\t}\n\t}\n}", "func (p *GremlinResourcesCreateUpdateGremlinDatabasePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *DeploymentsRestartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *SQLResourcesCreateUpdateSQLUserDefinedFunctionPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func newOOMv2Poller(publisher shim.Publisher) (oomPoller, error) {\n\treturn &watcherV2{\n\t\titemCh: make(chan itemV2),\n\t\tpublisher: publisher,\n\t}, nil\n}", "func (p *PrivateEndpointConnectionsUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func createPollingTracker(resp *http.Response) (pollingTracker, error) {\n\tvar pt pollingTracker\n\tswitch strings.ToUpper(resp.Request.Method) {\n\tcase http.MethodDelete:\n\t\tpt = &pollingTrackerDelete{pollingTrackerBase: pollingTrackerBase{resp: resp}}\n\tcase http.MethodPatch:\n\t\tpt = &pollingTrackerPatch{pollingTrackerBase: pollingTrackerBase{resp: resp}}\n\tcase http.MethodPost:\n\t\tpt = &pollingTrackerPost{pollingTrackerBase: pollingTrackerBase{resp: resp}}\n\tcase http.MethodPut:\n\t\tpt = &pollingTrackerPut{pollingTrackerBase: pollingTrackerBase{resp: resp}}\n\tdefault:\n\t\treturn nil, autorest.NewError(\"azure\", \"createPollingTracker\", \"unsupported HTTP method %s\", resp.Request.Method)\n\t}\n\tif err := pt.initializeState(); err != nil {\n\t\treturn pt, err\n\t}\n\t// this initializes the polling header values, we do this during creation in case the\n\t// initial response send us invalid values; this way the API call will return a non-nil\n\t// error (not doing this means the error shows up in Future.Done)\n\treturn pt, pt.updatePollingMethod()\n}", "func (p *ClientCreateOrUpdateByIDPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (pt *pollingTrackerBase) initPollingMethod() error {\n\tif ao, err := getURLFromAsyncOpHeader(pt.resp); err != nil {\n\t\treturn err\n\t} else if ao != \"\" {\n\t\tpt.URI = ao\n\t\tpt.Pm = PollingAsyncOperation\n\t\treturn nil\n\t}\n\tif lh, err := getURLFromLocationHeader(pt.resp); err != nil {\n\t\treturn err\n\t} else if lh != \"\" {\n\t\tpt.URI = lh\n\t\tpt.Pm = PollingLocation\n\t\treturn nil\n\t}\n\t// it's ok if we didn't find a polling header, this will be handled elsewhere\n\treturn nil\n}", "func (p *ManagedClustersStopPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *ServicesUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (c *ManagedServiceClient) Poll() *ManagedServicePollRequest {\n\treturn &ManagedServicePollRequest{\n\t\trequest: c.Get(),\n\t}\n}", "func NewPoll(creator, question string, answerOptions, settings []string) (*Poll, *ErrorMessage) {\n\tp := Poll{\n\t\tID: model.NewId(),\n\t\tCreatedAt: model.GetMillis(),\n\t\tCreator: creator,\n\t\tQuestion: question,\n\t}\n\tfor _, answerOption := range answerOptions {\n\t\tif errMsg := p.AddAnswerOption(answerOption); errMsg != nil {\n\t\t\treturn nil, errMsg\n\t\t}\n\t}\n\tfor _, s := range settings {\n\t\tswitch s {\n\t\tcase \"anonymous\":\n\t\t\tp.Settings.Anonymous = true\n\t\tcase \"progress\":\n\t\t\tp.Settings.Progress = true\n\t\tcase \"public-add-option\":\n\t\t\tp.Settings.PublicAddOption = true\n\t\tdefault:\n\t\t\treturn nil, &ErrorMessage{\n\t\t\t\tMessage: &i18n.Message{\n\t\t\t\t\tID: \"poll.newPoll.unrecognizedSetting\",\n\t\t\t\t\tOther: \"Unrecognized poll setting: {{.Setting}}\",\n\t\t\t\t},\n\t\t\t\tData: map[string]interface{}{\n\t\t\t\t\t\"Setting\": s,\n\t\t\t\t},\n\t\t\t}\n\t\t}\n\t}\n\treturn &p, nil\n}", "func (p *ClientUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *SQLResourcesCreateUpdateSQLContainerPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *SQLResourcesClientCreateUpdateSQLUserDefinedFunctionPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *GremlinResourcesClientCreateUpdateGremlinDatabasePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (p *ClientMoveResourcesPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func New() *Limiter {\n\tl := &Limiter{make(chan int, 1), make(chan struct{}, 1)}\n\tl.waiter <- struct{}{}\n\treturn l\n}", "func (p *BindingsCreateOrUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func NewFilePoller(ctx context.Context, f FileChannel, pollTimeout time.Duration) *FilePoller {\n\treturn &FilePoller{File: f, ctx: ctx, pollTimeout: pollTimeout}\n}", "func (p *EventSubscriptionsUpdatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func NewWatcher(\n\tlister ListerService,\n\tinterval time.Duration,\n\topts ...func(*WatcherConfig),\n) *Watcher {\n\tcfg := WatcherConfig{\n\t\tLogger: zero.Logger(),\n\t}\n\tfor _, opt := range opts {\n\t\topt(&cfg)\n\t}\n\tw := &Watcher{\n\t\tlister: lister,\n\t\tstreamer: stream.NewPoller(\n\t\t\tfunc() (zero.Interface, error) {\n\t\t\t\treturn lister.ListGoRepos()\n\t\t\t},\n\t\t\tinterval,\n\t\t),\n\t\tlog: cfg.Logger,\n\t}\n\tgo w.run()\n\treturn w\n}" ]
[ "0.8291853", "0.8011239", "0.78253216", "0.77730536", "0.7642704", "0.73517257", "0.7298127", "0.71799994", "0.707358", "0.6794879", "0.65679336", "0.6537219", "0.6506827", "0.6469193", "0.63708735", "0.621865", "0.62098026", "0.6207545", "0.6190991", "0.61326003", "0.6106173", "0.6079252", "0.60733724", "0.6059773", "0.6042892", "0.6042766", "0.60419714", "0.6025855", "0.6009047", "0.6009047", "0.6002667", "0.5978744", "0.5977028", "0.5959582", "0.5953964", "0.5944533", "0.59442073", "0.5942539", "0.5939524", "0.5931469", "0.5906757", "0.59019935", "0.58947146", "0.5867558", "0.58640593", "0.586134", "0.58608955", "0.5847723", "0.58465594", "0.58276117", "0.582623", "0.58135194", "0.57920927", "0.57914793", "0.57881016", "0.5783758", "0.57800055", "0.57748544", "0.57666874", "0.57556653", "0.57535815", "0.5751446", "0.57470864", "0.57372606", "0.57298666", "0.5715409", "0.5707901", "0.5703567", "0.5699213", "0.56950444", "0.56850535", "0.567939", "0.5660578", "0.5645768", "0.56422925", "0.56348073", "0.56335354", "0.5626659", "0.5626133", "0.56245416", "0.561528", "0.5609632", "0.5604238", "0.5583702", "0.557905", "0.55725956", "0.55683523", "0.55671334", "0.5566826", "0.5556223", "0.55478287", "0.5539472", "0.5533329", "0.5524122", "0.5523599", "0.55223167", "0.55202186", "0.55094457", "0.5503603", "0.5502593" ]
0.77754176
3
Start spins up a goroutine that continously polls given API endpoint at interval Poller.Interval
func (p *Poller) Start() { go func(url string, interval time.Duration, out chan<- PollMsg, shutdown <-chan *sync.WaitGroup) { wg := &sync.WaitGroup{} defer func() { wg.Done() }() ticker := time.NewTicker(interval) msg := PollMsg{} resp, err := http.Get(url) if err != nil { msg.Error = append(msg.Error, err) } data, err2 := ioutil.ReadAll(resp.Body) if err2 != nil { msg.Error = append(msg.Error, err2) } msg.Payload = data out <- msg resp.Body.Close() for { select { case wg = <-shutdown: return case <-ticker.C: msg := PollMsg{} resp, err := http.Get(url) if err != nil { msg.Error = append(msg.Error, err) } data, err2 := ioutil.ReadAll(resp.Body) if err2 != nil { msg.Error = append(msg.Error, err2) } msg.Payload = data out <- msg resp.Body.Close() } } }(p.URL, p.Interval, p.Out, p.Shutdown) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *AutoScaler) Run() {\n\tticker := s.clock.NewTicker(s.pollPeriod)\n\ts.readyCh <- struct{}{} // For testing.\n\n\t// Don't wait for ticker and execute pollAPIServer() for the first time.\n\ts.pollAPIServer()\n\n\tfor {\n\t\tselect {\n\t\tcase <-ticker.C():\n\t\t\ts.pollAPIServer()\n\t\tcase <-s.stopCh:\n\t\t\treturn\n\t\t}\n\t}\n}", "func (p *Poller) Start() {\n\tif p.IsStopped() {\n\t\tpanic(\"cannot restart poller after it has been stopped\")\n\t}\n\n\tp.group.Add(1)\n\tgo p.runPolling()\n}", "func (s *Service) Start(ctx context.Context) <-chan error {\n\ttick := make(chan error, 1)\n\n\tif s.Interval <= 0 {\n\t\ttick <- errors.New(\"cannot run poll service for non-positive interval\")\n\t\tclose(tick)\n\t\treturn tick\n\t}\n\n\tgo func(c context.Context, t chan<- error, interval time.Duration) {\n\t\t// update once when the service starts.\n\t\tt <- nil\n\n\t\tticker := time.NewTicker(interval)\n\t\tdefer ticker.Stop()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ticker.C:\n\t\t\t\tt <- nil\n\n\t\t\tcase <-ctx.Done():\n\t\t\t\tt <- ctx.Err()\n\t\t\t\tclose(t)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}(ctx, tick, time.Duration(s.Interval))\n\n\treturn tick\n}", "func (p *APIPingProbe) Start() error {\n\tif p.Active {\n\t\treturn nil\n\t}\n\n\tt := time.Duration(p.Config.Frequency * float64(time.Second))\n\twriteTicker := time.NewTicker(t)\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-p.StopChannel:\n\t\t\t\treturn\n\t\t\tcase <-writeTicker.C:\n\t\t\t\tgo p.PingAPI()\n\t\t\t}\n\t\t}\n\t}()\n\n\tp.Active = true\n\treturn nil\n}", "func (w *IndexPoller) Start() {\n\tw.channel <- w.run()\n\tticker := time.NewTicker(w.pollRate)\n\tfor {\n\t\tselect {\n\t\tcase <-w.controlChannel:\n\t\t\tdefer close(w.channel)\n\t\t\tticker.Stop()\n\t\t\treturn\n\t\tcase <-ticker.C:\n\t\t\tw.channel <- w.run()\n\t\t}\n\t}\n}", "func (m *Microservice) StartOperationPolling() {\n\tinterval := m.Config.viper.GetString(\"agent.operations.pollRate\")\n\n\tzap.S().Infof(\"Adding operation polling task with interval: %s\", interval)\n\t_, err := m.Scheduler.cronjob.AddFunc(interval, func() {\n\t\tm.CheckForNewConfiguration()\n\t})\n\n\tif err != nil {\n\t\tzap.S().Errorf(\"Could not create polling task with interval [%s]. %s\", interval, err)\n\t}\n}", "func (f *Input) startPoller(ctx context.Context) {\n\tf.wg.Add(1)\n\tgo func() {\n\t\tdefer f.wg.Done()\n\t\tglobTicker := time.NewTicker(f.PollInterval)\n\t\tdefer globTicker.Stop()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn\n\t\t\tcase <-globTicker.C:\n\t\t\t}\n\n\t\t\tf.poll(ctx)\n\t\t}\n\t}()\n}", "func (p *StreamingEndpointsStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (e *EndpointsManager) Run() {\n\tticker := time.NewTicker(time.Second * 10)\n\tdefer ticker.Stop()\n\tfor {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\te.watchAliveEndpoint()\n\t\tcase <-e.exit:\n\t\t\tclose(e.closed)\n\t\t\tcommon.Logger.Info(\"service done!!!\")\n\t\t\treturn\n\t\t}\n\t}\n}", "func (d *Daemon) Start(interval int, data chan *structs.Data) {\n\t// immediately fetch the first set of data and send it over the data channel\n\td.fetchData()\n\td.buildData()\n\n\tselect {\n\tcase data <- d.Data:\n\t\tlogger.Trace(\"Sending initial results on the 'data' channel\")\n\tdefault:\n\t\tlogger.Trace(\"Could not send initial results on the 'data' channel\")\n\t}\n\n\t// fetch new data every interval\n\tduration := time.Duration(interval) * time.Second\n\tfor _ = range time.Tick(duration) {\n\t\td.resetData()\n\t\td.fetchData()\n\t\td.buildData()\n\n\t\t// send the result over the data channel\n\t\tselect {\n\t\tcase data <- d.Data:\n\t\t\tlogger.Trace(\"Sending results on the 'data' channel\")\n\t\tdefault:\n\t\t\tlogger.Trace(\"Could not send results on the 'data' channel\")\n\t\t}\n\t}\n}", "func (s *Store) startPoller() {\n\tfor {\n\t\tselect {\n\t\tcase req := <-s.getReqQueue:\n\t\t\treq.respCh <- s.performGetOperation(req.key)\n\n\t\tcase req := <-s.modifyReqQueue:\n\t\t\terr := s.performModifyOperation(req)\n\t\t\treq.respCh <- err\n\n\t\t\ts.fanOutSubscriptions(req)\n\n\t\tcase sub := <-s.subscribeQueue:\n\t\t\ts.registerSubscription(sub)\n\t\t}\n\t}\n}", "func StartPing(interval time.Duration){\n\tlog.Println(\"Ping Time\")\n\tpinging := true\n\tfor pinging {\n\t\tpinging = false\n\t\tlog.Print(\"Pinging set to \" + strconv.FormatBool(pinging))\n\n\t\tif shouldIPing() {\n\t\t\tpinging = true\n\t\t\tlog.Print(\"Pinging set to \" + strconv.FormatBool(pinging))\n\n\t\t\tbullyImpl.SetIsCoordinatorAlive(false)\n\t\t\tlog.Print(bullyImpl.IsCoordinatorAlive())\n\t\t\tbullyImpl.GetMoi().Ping(bullyImpl.GetCoordinator())\n\n\t\t\ttimer := time.NewTimer(interval)\n\t\t\tselect {\n\t\t\tcase <- endTimer:\n\t\t\t\tlog.Print(\"Pinging was ended\")\n\t\t\tcase <- timer.C:\n\t\t\t\tif shouldIPing() && !bullyImpl.IsCoordinatorAlive() {\n\t\t\t\t\tpinging = false\n\t\t\t\t\tlog.Print(\"Pinging set to \" + strconv.FormatBool(pinging))\n\t\t\t\t\tlog.Print(\"Coordinator is not alive, launching a new Election\")\n\t\t\t\t\tgo func(){ electionChannel <- struct{}{} }()\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func (b *Bootstrapper) Start() {\n\tb.ctx, b.cancel = context.WithCancel(b.ctx)\n\tb.ticker = time.NewTicker(b.config.Period)\n\n\tgo func() {\n\t\tdefer b.ticker.Stop()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-b.ctx.Done():\n\t\t\t\treturn\n\t\t\tcase <-b.ticker.C:\n\t\t\t\tb.checkConnectivity()\n\t\t\t}\n\t\t}\n\t}()\n}", "func (p *Prober) Start() {\n\t// Get static copy of the config object\n\tcfg := p.config.Copy()\n\n\tfor _, svc := range cfg.Monitor.Services {\n\t\t// Create new Probe Bot and start it\n\t\tif svc.Interval == 0 {\n\t\t\tsvc.Interval = cfg.Monitor.Interval\n\t\t}\n\n\t\tgo NewProbeBot(\n\t\t\tp.eb,\n\t\t\tsvc,\n\t\t\tp.status.Update,\n\t\t).Start()\n\t}\n}", "func (p *Poller) runPolling() {\n\tfor !p.IsStopped() {\n\t\tif res := p.Poll(); res != nil {\n\t\t\tp.sendResult(res)\n\t\t}\n\t\ttime.Sleep(p.WaitInterval)\n\t}\n\tp.group.Done()\n\tp.isFinished = true\n}", "func (p *LiveEventsStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func Start(host string, username string, password string, activeSeconds int, inactiveSeconds int, actionLevel string) {\n\tif !communication.TestConnectivity(host) {\n\t\tlog.Panic(\"The URL \" + host + \" is not the API Endpoint\")\n\t}\n\tvar login = authenticate.HandleLogin(username, password, host)\n\ttrigger.Run(login.Value, host, activeSeconds, inactiveSeconds, actionLevel)\n}", "func (m *Module) Start(initStartup bool) error {\n\tinterval := m.settings.FetchInterval\n\tif interval == 0 || interval < minFetchInterval {\n\t\tinterval = minFetchInterval\n\t}\n\n\t// Get some readings at start\n\tgo m.getReadings()\n\n\tm.ticker = time.NewTicker(time.Second * time.Duration(interval))\n\tgo func() {\n\t\tfor range m.ticker.C {\n\t\t\tm.getReadings()\n\t\t}\n\t}()\n\n\treturn nil\n}", "func (p *ServicesStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (r *MyMonitor) StartMonitoring(urlinfo st.URLInfo) {\n\n\tgo func() {\n\t\tticker := time.NewTicker(time.Duration(urlinfo.Frequency) * time.Second) // trigers at rate of Frequency\n\t\trequestStatus := make(chan string) // result of http request will come on requestStatus channel\n\t\tdataBase := database.GetDatabase()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase idStop := <-r.MonitorStp: // if deactivated Monitoring\n\t\t\t\tif idStop == urlinfo.ID { // check if this url need to be Deactivate\n\t\t\t\t\treturn // stop monitering\n\t\t\t\t}\n\t\t\tcase <-ticker.C: // at Frequency time\n\t\t\t\tfmt.Printf(\"Request to %s\\t\", urlinfo.URL)\n\t\t\t\tgo Request(urlinfo.URL, urlinfo.CrawlTimeout, requestStatus)\n\t\t\tcase st := <-requestStatus: // if status for request comes\n\t\t\t\tfmt.Println(\"Status: \", st)\n\t\t\t\t// if not 200 OK increase Failure count and update it to Database\n\t\t\t\t// if Failure count reaches to failure threshold then mark url inactive , stop crawling and update status into database\n\t\t\t\tif st != \"200 OK\" {\n\t\t\t\t\turlinfo.FailureCount++\n\t\t\t\t\tdataBase.UpdateColumnInDatabase(urlinfo.ID, \"failure_count\", urlinfo.FailureCount)\n\t\t\t\t\tif urlinfo.FailureCount == urlinfo.FailureThreshold {\n\t\t\t\t\t\tdataBase.UpdateColumnInDatabase(urlinfo.ID, \"status\", \"inactive\")\n\t\t\t\t\t\tdataBase.UpdateColumnInDatabase(urlinfo.ID, \"crawling\", false)\n\t\t\t\t\t\treturn //Stop Monitering\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n}", "func (r *PeriodicGoroutine) Start() {\n\tdefer close(r.finished)\n\nloop:\n\tfor {\n\t\tif shutdown, err := runPeriodicHandler(r.ctx, r.handler, r.operation); shutdown {\n\t\t\tbreak\n\t\t} else if h, ok := r.handler.(ErrorHandler); ok && err != nil {\n\t\t\th.HandleError(err)\n\t\t}\n\n\t\tselect {\n\t\tcase <-r.clock.After(r.interval):\n\t\tcase <-r.ctx.Done():\n\t\t\tbreak loop\n\t\t}\n\t}\n\n\tif h, ok := r.handler.(Finalizer); ok {\n\t\th.OnShutdown()\n\t}\n}", "func (bt *Esbeat) Polling(name string, host *url.URL, fetchData FuncFetchData) error {\n\tlogp.Info(\"esbeat-%s-%s is running\", name, host.String())\n\n\thttp := helper.NewHTTP(bt.config)\n\tticker := time.NewTicker(bt.config.Period)\n\n\tfor {\n\t\tselect {\n\t\tcase <-bt.done:\n\t\t\tlogp.Info(\"esbeat-%s-%s is stopping\", name, host.String())\n\t\t\treturn nil\n\t\tcase <-ticker.C:\n\t\t}\n\n\t\tbody, err := fetchData(http, host)\n\t\tif err != nil {\n\t\t\tlogp.Err(\"Error reading cluster node: %v\", err)\n\t\t} else {\n\t\t\tevent := common.MapStr{\n\t\t\t\t\"@timestamp\": common.Time(time.Now()),\n\t\t\t\t\"type\": name,\n\t\t\t\t\"url\": host.String(),\n\t\t\t\tname: body,\n\t\t\t}\n\t\t\tbt.client.PublishEvent(event)\n\t\t}\n\t}\n}", "func Start(h Handler) {\n for {\n log.Println(\"Polling ..!!\")\n resp, err := SqsService.Receive()\n\n if err != nil {\n log.Println(err)\n continue\n }\n\n if len(resp.Messages) > 0 {\n process(h, resp.Messages)\n }\n }\n}", "func startPeriodicHealthCheck(m *MicroService, interval time.Duration, d *net.Dialer) {\r\n\tticker := time.NewTicker(interval)\r\n\tdefer ticker.Stop()\r\n\tfor t := range ticker.C {\r\n\t\tlogVerbose(\"Checking health of Service:\", m.Route, \" ---tick:\", t)\r\n\t\thandleHealthCheck(m, d)\r\n\t}\r\n}", "func (p *Poller) Run() {\n\tgo util.Forever(func() {\n\t\te, err := p.getFunc()\n\t\tif err != nil {\n\t\t\tglog.Errorf(\"failed to list: %v\", err)\n\t\t\treturn\n\t\t}\n\t\tp.sync(e)\n\t}, p.period)\n}", "func (p *NotebookWorkspacesClientStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (c *TempController) Poll(w http.ResponseWriter, req *http.Request) {\n\tvar poll PollRequest\n\n\tbody, err := ioutil.ReadAll(io.LimitReader(req.Body, 500000))\n\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tif err := req.Body.Close(); err != nil {\n\t\tpanic(err)\n\t}\n\n\tif err := json.Unmarshal(body, &poll); err != nil {\n\t\trespondMessage(http.StatusNotAcceptable, w, err.Error())\n\t}\n\n\tswitch poll.Action {\n\tcase \"START\":\n\t\t//validate interval > 0\n\t\tif poll.IntervalMilliseconds <= 0 {\n\t\t\trespondMessage(http.StatusBadRequest, w, fmt.Sprintf(\"Expected intervalMilliseconds field > 0. Got %v\", poll.IntervalMilliseconds))\n\t\t\treturn\n\t\t}\n\n\t\t//call poller.start\n\t\terr = c.poller.Start(poll.IntervalMilliseconds)\n\n\t\tif err != nil {\n\t\t\trespondMessage(http.StatusInternalServerError, w, fmt.Sprintf(\"Error starting temperature poller. Error %v\", err))\n\t\t\treturn\n\t\t}\n\n\t\trespondMessage(http.StatusOK, w, fmt.Sprintf(\"Started polling temperature with interval %v\", poll.IntervalMilliseconds))\n\tcase \"STOP\":\n\t\t//call poller.stop\n\t\terr = c.poller.Stop()\n\n\t\tif err != nil {\n\t\t\trespondMessage(http.StatusInternalServerError, w, fmt.Sprintf(\"Error stopping temperature poller. Error %v\", err))\n\t\t\treturn\n\t\t}\n\n\t\trespondMessage(http.StatusOK, w, \"Stopped polling temperature.\")\n\tdefault:\n\t\trespondMessage(http.StatusBadRequest, w, fmt.Sprintf(\"Invalid 'action' value. Expected 'START' or 'STOP'. Got %v\", poll.Action))\n\t}\n}", "func (c *ProxyMarketClient) StartWatcher(updatedTime time.Duration) {\n\tgo func() {\n\t\tfor {\n\t\t\tres, err := c.GetProxyListAllByNewest()\n\t\t\tif err != nil {\n\t\t\t\ttime.Sleep(updatedTime * time.Millisecond)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tc.Proxies = res.List.Data\n\t\t\ttime.Sleep(updatedTime * time.Millisecond)\n\t\t}\n\t}()\n}", "func (sb *StatusBeater) Start(stopChan chan struct{}, publish func(event beat.Event)) {\n\tgo func() {\n\t\tsb.Beat(ServiceStarted, \"Service started\", publish)\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-sb.IntervalFunc():\n\t\t\t\tsb.Beat(ServiceRunning, \"Service is Running\", publish)\n\t\t\tcase <-stopChan:\n\t\t\t\tsb.Beat(ServiceStopped, \"Service is Stopped\", publish)\n\t\t\t\tsb.doneChan <- struct{}{}\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n}", "func (api *API) Start() {\n\t// timeout before forcing shutdown\n\twait := time.Second * 3\n\n\t// this is from the mux documentation\n\tsrv := &http.Server{\n\t\tAddr: api.IPPort,\n\t\t// Good practice to set timeouts to avoid Slowloris attacks.\n\t\tWriteTimeout: time.Second * 15,\n\t\tReadTimeout: time.Second * 15,\n\t\tIdleTimeout: time.Second * 60,\n\t\t//TOFIX ! load from config\n\t\tHandler: handlers.CORS(\n\t\t\thandlers.AllowedOrigins([]string{\"*\"}),\n\t\t\thandlers.AllowedMethods([]string{\"GET\", \"POST\", \"PUT\", \"DELETE\", \"OPTIONS\"}),\n\t\t\thandlers.AllowedHeaders([]string{\"X-Requested-With\", \"Content-Type\", \"X-Session-Token\"}),\n\t\t)(api.Router), // Pass our instance of gorilla/mux in.\n\t}\n\n\tgo func() {\n\t\tlog.Infof(\"API Listenning : %s, version : %s\", api.IPPort, api.Version)\n\t\tlog.Infof(\"API Endpoint : %s\", api.IPPort+api.Prefix)\n\t\tif err := srv.ListenAndServe(); err != nil {\n\t\t\tlog.Println(err)\n\t\t}\n\t}()\n\t// Block until we receive our signal.\n\t<-c\n\n\t// Create a deadline to wait for.\n\tctx, cancel := context.WithTimeout(context.Background(), wait)\n\tdefer cancel()\n\t// Doesn't block if no connections, but will otherwise wait\n\t// until the timeout deadline.\n\tsrv.Shutdown(ctx)\n\t// Optionally, you could run srv.Shutdown in a goroutine and block on\n\t// <-ctx.Done() if your application should wait for other services\n\t// to finalize based on context cancellation.\n\tlog.Println(\"shutting down\")\n\tos.Exit(0)\n}", "func (c *Client) RunDispatch() {\n for {\n time.Sleep(pollInterval)\n c.SendRequests()\n c.RecvResponses()\n }\n}", "func PollArtiMetricsRestEndpoint(artDetails *jfauth.ServiceDetails, intervalSecs int) {\n\tjflog.Info(fmt.Sprintf(\"Polling api/v1/metrics REST end point\"))\n\turl := \"api/v1/metrics\"\n\tfor {\n\t\tresp, err := getHttpResp(artDetails, url)\n\t\tif err != nil {\n\t\t\tfmt.Printf(\"GET HTTP failed for url : %s, resp = %s\\n\", url, resp)\n\t\t\tjflog.Error(fmt.Sprintf(\"GET HTTP failed for url : %s, resp = %s\", url, resp))\n\t\t}\n\t\ttime.Sleep(time.Duration(intervalSecs) * time.Second)\n\t}\n}", "func (p *pollerAutoScaler) Start() {\n\tlogger := p.logger.Sugar()\n\tp.wg.Add(1)\n\tgo func() {\n\t\tdefer p.wg.Done()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-p.ctx.Done():\n\t\t\t\treturn\n\t\t\tcase <-time.After(p.cooldownTime):\n\t\t\t\tcurrentResource := autoscaler.ResourceUnit(p.sem.GetLimit())\n\t\t\t\tcurrentUsages, err := p.pollerUsageEstimator.Estimate()\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogger.Warnw(\"poller autoscaler skip due to estimator error\", \"error\", err)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tproposedResource := p.recommender.Recommend(currentResource, currentUsages)\n\t\t\t\tlogger.Debugw(\"poller autoscaler recommendation\",\n\t\t\t\t\t\"currentUsage\", currentUsages,\n\t\t\t\t\t\"current\", uint64(currentResource),\n\t\t\t\t\t\"recommend\", uint64(proposedResource),\n\t\t\t\t\t\"isDryRun\", p.isDryRun)\n\t\t\t\tif !p.isDryRun {\n\t\t\t\t\tp.sem.SetLimit(int(proposedResource))\n\t\t\t\t}\n\t\t\t\tp.pollerUsageEstimator.Reset()\n\n\t\t\t\t// hooks\n\t\t\t\tfor i := range p.onAutoScale {\n\t\t\t\t\tp.onAutoScale[i]()\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\treturn\n}", "func (t Ticker) Start() {\n\t// Create new Kite ticker instance\n\tticker = kiteticker.New(t.APIKey, t.APIAccesToken)\n\tsubscriptions = t.Subscrptions\n\n\t// Assign callbacks\n\tticker.OnError(onError)\n\tticker.OnClose(onClose)\n\tticker.OnConnect(onConnect)\n\tticker.OnReconnect(onReconnect)\n\tticker.OnNoReconnect(onNoReconnect)\n\tticker.OnTick(onTick)\n\tticker.OnOrderUpdate(onOrderUpdate)\n\n\t// Start the connection\n\tticker.Serve()\n}", "func (pool *SubPool) Start() {\n\tfor {\n\t\tselect {\n\t\tcase sub := <-pool.Register:\n\t\t\tlog.Printf(\"Registering subscriber %v to pool\", sub.ID)\n\t\t\tpool.Subscribers = append(pool.Subscribers, sub)\n\t\t\tlog.Printf(\"Successfully registered subscriber %v to pool\", sub.ID)\n\t\tcase message := <-pool.Broadcast:\n\t\t\tlog.Printf(\"Broadcasting message to all subscribers...\")\n\t\t\tfor _, sub := range pool.Subscribers {\n\t\t\t\tif err := sub.connection.WriteJSON(message); err != nil {\n\t\t\t\t\tlog.Printf(\"Error writing message %v to subscriber %v\", message, sub.ID)\n\t\t\t\t}\n\t\t\t}\n\t\tcase <-pool.StopChan:\n\t\t\tfor _, sub := range pool.Subscribers {\n\t\t\t\tlog.Printf(\"shutting down connection for subscriber %v...\", sub.ID)\n\t\t\t\tsub.connection.Close()\n\t\t\t}\n\t\t}\n\t}\n}", "func (mon Monitor) Start(ip string, port int, cloudFunctionsPattern string, chanAnalyzer chan []CloudService) {\n\n\tmon.cloudFunctionsPattern = cloudFunctionsPattern\n\n\tfor {\n\t\t//mon.lookup = *dist.NewLookupProxy(ip, port)\n\n\t\tmon.refreshCloudServices(ip, port)\n\n\t\t//err := mon.lookup.Close()\n\t\t//if err != nil {\n\t\t//\tlib.PrintlnError(\"Error at closing lookup. Error:\", err)\n\t\t//}\n\n\t\tfor i := range mon.cloudServices {\n\t\t\tmon.cloudServices[i].RefreshPrice()\n\t\t\tmon.cloudServices[i].RefreshStatus()\n\t\t}\n\n\t\tif len(mon.cloudServices) > 0 {\n\t\t\tchanAnalyzer <- mon.cloudServices\n\t\t}\n\n\t\ttime.Sleep(5 * time.Second)\n\t}\n}", "func TrackStreams() {\n // Sleep until the time is a multiple of the refresh period\n now := time.Now()\n wakeUpTime := now.Truncate(config.Timing.Period).Add(config.Timing.Period)\n fmt.Print(\"Waiting...\")\n time.Sleep(wakeUpTime.Sub(now))\n fmt.Println(\"Go\")\n\n // Start periodic updates\n ticker := time.NewTicker(config.Timing.Period)\n Update() // Update immediately, since ticker waits for next interval\n for {\n <-ticker.C\n Update()\n }\n}", "func Start(meter metric.Meter, interval time.Duration) error {\n\tr := &runtime{\n\t\tmeter: meter,\n\t\tinterval: interval,\n\t}\n\treturn r.register()\n}", "func (p *Proxy) Start() {\n\tp.stream.EventKv(\"proxy.starting\", kvs{\"public_key\": p.publicKey, \"speed\": strconv.Itoa(p.speed)})\n\n\tp.Fetch()\n\tfor {\n\t\tselect {\n\t\tcase <-p.stopCh:\n\t\t\tclose(p.doneCh)\n\t\t\treturn\n\t\tcase <-p.ticker.C:\n\t\t\tp.Fetch()\n\t\t}\n\t}\n}", "func (r *ManagedServicePollRequest) Interval(value time.Duration) *ManagedServicePollRequest {\n\tr.interval = value\n\treturn r\n}", "func (pr *PeriodicReader) start(ctx context.Context) {\n\tdefer pr.wait.Done()\n\tticker := time.NewTicker(pr.interval)\n\tfor {\n\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn\n\t\tcase <-ticker.C:\n\t\t\tif err := pr.collectWithTimeout(ctx, pr.exporter.ExportMetrics); err != nil {\n\t\t\t\totel.Handle(err)\n\t\t\t}\n\t\t}\n\t}\n}", "func (a *Aggregator) Run() {\n\tfor {\n\t\tnow := time.Now()\n\t\twait := now.Add(a.pollingInterval).\n\t\t\tTruncate(a.pollingInterval).\n\t\t\tSub(now)\n\t\ttime.Sleep(wait)\n\n\t\tts := time.Now().Truncate(a.pollingInterval)\n\t\tcounts := a.counter.Reset()\n\n\t\ta.mu.Lock()\n\t\ta.data = a.data.Next()\n\t\ta.data.Value = Rate{\n\t\t\tTimestamp: ts.Unix(),\n\t\t\tCounts: counts,\n\t\t}\n\t\ta.mu.Unlock()\n\t}\n}", "func (dispatcher *Dispatcher) Start() {\n\tfor {\n\t\tselect {\n\t\tcase statusUpdate := <-dispatcher.statusChan:\n\t\t\tif !dispatcher.shouldPublish(statusUpdate) {\n\t\t\t\tlog.Debugf(\"suppressing: %+v\", statusUpdate)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tpingResult := statusUpdate.Status.LatestResult\n\t\t\tvar error string\n\t\t\tif pingResult.Error != nil {\n\t\t\t\terror = pingResult.Error.Error()\n\t\t\t}\n\t\t\tstatus := alerter.PingerStatus{\n\t\t\t\tOK: pingResult.Status == ping.StatusOK,\n\t\t\t\tError: error,\n\t\t\t\tOutputURL: outputURL(dispatcher.advertisedBaseURL, statusUpdate.Name),\n\t\t\t}\n\n\t\t\tupdate := alerter.PingerUpdate{\n\t\t\t\tName: statusUpdate.Name,\n\t\t\t\tStatus: status,\n\t\t\t\tConsecutive: statusUpdate.Status.Consecutive,\n\t\t\t\tLatestOK: statusUpdate.Status.LatestOK,\n\t\t\t\tLatestNOK: statusUpdate.Status.LatestNOK,\n\t\t\t}\n\n\t\t\tlog.Debugf(\"dispatching %+v\", statusUpdate)\n\t\t\tdispatcher.dispatch(update)\n\t\t}\n\t}\n\n}", "func (this *BaseUnit) Start(ch chan<- model.Metric) {\n\tthis.RLock()\n\tfetchInterval := this.fetchInterval\n\tthis.RUnlock()\n\n\tlog.Debugf(\"Starting fetch for unit %v...\", this)\n\n\tticker := time.NewTicker(fetchInterval)\n\tdefer ticker.Stop()\n\n\tc := reflect.ValueOf(this.unit)\n\tmethodFetch := c.MethodByName(\"Fetch\")\n\tmethodFetch.Call([]reflect.Value{reflect.ValueOf(ch)})\n\n\tfor {\n\t\tselect {\n\t\tcase <-this.fetchStop:\n\t\t\treturn\n\t\tcase <-ticker.C:\n\t\t\tmethodFetch.Call([]reflect.Value{reflect.ValueOf(ch)})\n\t\t}\n\t}\n}", "func PollItself() {\n\tfor {\n\t\tresp, err := http.Get(\"http://localhost:1234/\")\n\t\tif err != nil {\n\t\t\tlogger.Sugar().Errorf(\"HTTP request failed: %w\", err)\n\t\t} else {\n\t\t\tbody, err := ioutil.ReadAll(resp.Body)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Sugar().Errorf(\"Failed to read response: %w\", err)\n\t\t\t} else {\n\t\t\t\tlogger.Sugar().Debugf(\"Response: %s\", string(body))\n\t\t\t\t_ = resp.Body.Close()\n\t\t\t}\n\t\t}\n\t\ttime.Sleep(time.Second * time.Duration(rand.Intn(3)))\n\t}\n}", "func (t *TestMonitor) Start(_ time.Duration) (chan statedb.PricePoint, chan error) {\n\n\tpriceChan := make(chan PricePoint)\n\n\tt.tick = time.NewTicker(t.interval)\n\tgo pitcher(t.tick.C, t.points, priceChan)\n\n\tt.active = true\n\n\treturn priceChan, nil\n}", "func (t *Timer) Start() {\n\tif !t.running {\n\t\tt.running = true\n\t\tgo func() {\n\t\t\ttime.Sleep(t.interval)\n\t\t\tt.elapsedCallback()\n\t\t\tt.running = false\n\t\t}()\n\t}\n}", "func (api *API) Start(block bool) error {\n\n\tif !api.hasHandlers {\n\t\tlog.Debug(\"Watchtower HTTP API skipped.\")\n\t\treturn nil\n\t}\n\n\tif api.Token == \"\" {\n\t\tlog.Fatal(tokenMissingMsg)\n\t}\n\n\tif block {\n\t\trunHTTPServer()\n\t} else {\n\t\tgo func() {\n\t\t\trunHTTPServer()\n\t\t}()\n\t}\n\treturn nil\n}", "func (n *Notifier) Start() error {\n\tnotificationForChannels := time.NewTicker(time.Second * 60).C\n\tnotificationForTimeTable := time.NewTicker(time.Second * 60).C\n\tfor {\n\t\tselect {\n\t\tcase <-notificationForChannels:\n\t\t\tn.NotifyChannels()\n\t\tcase <-notificationForTimeTable:\n\t\t\tn.NotifyIndividuals()\n\t\t}\n\t}\n}", "func (l *Log) start(ch chan<- *Event) {\n\tdefer close(ch)\n\n\tl.Log.Debug(\"enter\")\n\tdefer l.Log.Debug(\"exit\")\n\n\tvar start = l.StartTime.UnixNano() / int64(time.Millisecond)\n\tvar nextToken *string\n\tvar err error\n\n\tfor {\n\t\tl.Log.WithField(\"start\", start).Debug(\"request\")\n\t\tnextToken, start, err = l.fetch(nextToken, start, ch)\n\n\t\tif err != nil {\n\t\t\tl.err = fmt.Errorf(\"log %q: %s\", l.GroupName, err)\n\t\t\tbreak\n\t\t}\n\n\t\tif nextToken == nil && l.Follow {\n\t\t\ttime.Sleep(l.PollInterval)\n\t\t\tl.Log.WithField(\"start\", start).Debug(\"poll\")\n\t\t\tcontinue\n\t\t}\n\n\t\tif nextToken == nil {\n\t\t\tbreak\n\t\t}\n\t}\n}", "func (ex *ExchangeClient) LongPoll(stopPollCh, stopAllCh chan int, errCh chan error) {\n\tif ex.pi == nil {\n\t\tpanic(\"No pi in ex\")\n\t}\n\tex.Info(\"Starting LongPoll|msgCode=POLLING\")\n\tdefer Utils.RecoverCrash(ex.logger) // catch all panic. RecoverCrash logs information needed for debugging.\n\tex.wg.Add(1)\n\tdefer ex.wg.Done()\n\n\tdefer func() {\n\t\tex.Info(\"Stopping LongPoll...\")\n\t\tex.cancel()\n\t}()\n\n\tvar err error\n\treqTimeout := ex.pi.ResponseTimeout\n\treqTimeout += uint64(float64(reqTimeout) * 0.1) // add 10% so we don't step on the HeartbeatInterval inside the ping\n\n\tif err != nil {\n\t}\n\tex.transport = &http.Transport{\n\t\tTLSClientConfig: &tls.Config{\n\t\t\tInsecureSkipVerify: false,\n\t\t\tRootCAs: globals.config.RootCerts(),\n\t\t},\n\t\tResponseHeaderTimeout: time.Duration(reqTimeout) * time.Millisecond,\n\t}\n\n\t// check for the proxy setting. Useful for mitmproxy testing\n\tproxy := os.Getenv(\"PINGER_PROXY\")\n\tif proxy != \"\" {\n\t\tproxyUrl, err := url.Parse(proxy)\n\t\tif err != nil {\n\t\t\tex.sendError(errCh, err)\n\t\t\treturn\n\t\t}\n\t\tex.transport.Proxy = http.ProxyURL(proxyUrl)\n\t}\n\n\tex.httpClient = &http.Client{\n\t\tTransport: ex.transport,\n\t}\n\tuseCookieJar := false\n\tif useCookieJar {\n\t\tcookieJar, err := cookiejar.New(nil)\n\t\tif err != nil {\n\t\t\tex.sendError(errCh, err)\n\t\t\treturn\n\t\t}\n\t\tex.httpClient.Jar = cookieJar\n\t}\n\tredactedUrl := strings.Split(ex.pi.MailServerUrl, \"?\")[0]\n\n\tex.Info(\"New HTTP Client with timeout %s %s<redacted>\", ex.transport.ResponseHeaderTimeout, redactedUrl)\n\tsleepTime := 0\n\ttooFastResponse := (time.Duration(ex.pi.ResponseTimeout) * time.Millisecond) / 4\n\tex.Debug(\"TooFast timeout set to %s\", tooFastResponse)\n\tvar responseCh chan *http.Response\n\tvar responseErrCh chan error\n\tfor {\n\t\tif sleepTime > 0 {\n\t\t\ts := time.Duration(sleepTime) * time.Second\n\t\t\tex.Info(\"Sleeping %s before retry\", s)\n\t\t\ttime.Sleep(s)\n\t\t}\n\t\tif responseErrCh != nil {\n\t\t\tclose(responseErrCh)\n\t\t}\n\t\tresponseErrCh = make(chan error)\n\t\tif responseCh != nil {\n\t\t\tclose(responseCh)\n\t\t}\n\t\tresponseCh = make(chan *http.Response)\n\n\t\ttimeSent := time.Now()\n\t\tex.wg.Add(1)\n\t\tex.cancelled = false\n\t\tgo ex.doRequestResponse(responseCh, responseErrCh)\n\t\tselect {\n\t\tcase err = <-responseErrCh:\n\t\t\tif err == NoSuchHostError || err == UnknownCertificateAuthority {\n\t\t\t\terrCh <- LongPollReRegister\n\t\t\t} else {\n\t\t\t\tex.sendError(errCh, err)\n\t\t\t}\n\t\t\treturn\n\n\t\tcase response := <-responseCh:\n\t\t\tif response == retryResponse {\n\t\t\t\tex.Debug(\"Retry-response from response reader.\")\n\t\t\t\tsleepTime = ex.exponentialBackoff(sleepTime)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// the response body tends to be pretty short (and we've capped it anyway). Let's just read it all.\n\t\t\tresponseBody, err := ioutil.ReadAll(response.Body)\n\t\t\tif err != nil {\n\t\t\t\tresponse.Body.Close() // attempt to close. Ignore any errors.\n\t\t\t\tex.sendError(errCh, err)\n\t\t\t\treturn\n\t\t\t}\n\t\t\terr = response.Body.Close()\n\t\t\tif err != nil {\n\t\t\t\tex.sendError(errCh, err)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tswitch {\n\t\t\tcase response.StatusCode != 200:\n\t\t\t\tswitch {\n\t\t\t\tcase response.StatusCode == 401:\n\t\t\t\t\t// ask the client to re-register, since nothing we could do would fix this\n\t\t\t\t\tex.Info(\"401 response. Telling client to re-register|msgCode=EAS_AUTH_ERR_REREGISTER\")\n\t\t\t\t\terrCh <- LongPollReRegister\n\t\t\t\t\treturn\n\n\t\t\t\tdefault:\n\t\t\t\t\t// just retry\n\t\t\t\t\tsleepTime = ex.exponentialBackoff(sleepTime)\n\t\t\t\t\tex.Info(\"Response Status %s. Back to polling\", response.Status)\n\t\t\t\t}\n\t\t\t\t//EAS Ping\n\t\t\tcase ex.pi.ASIsSyncRequest == false && (ex.pi.NoChangeReply != nil && bytes.Compare(responseBody, ex.pi.NoChangeReply) == 0):\n\t\t\t\t// go back to polling\n\t\t\t\tif time.Since(timeSent) <= tooFastResponse {\n\t\t\t\t\tex.Warning(\"Ping: NoChangeReply was too fast. Doing backoff. This usually indicates that the client is still connected to the exchange server.\")\n\t\t\t\t\tsleepTime = ex.exponentialBackoff(sleepTime)\n\t\t\t\t} else {\n\t\t\t\t\tex.Info(\"Ping: NoChangeReply after %s. Back to polling\", time.Since(timeSent))\n\t\t\t\t\tsleepTime = 0 // good reply. Reset any exponential backoff stuff.\n\t\t\t\t}\n\t\t\t\t// EAS Ping\n\t\t\tcase ex.pi.ASIsSyncRequest == false && (ex.pi.ExpectedReply == nil || bytes.Compare(responseBody, ex.pi.ExpectedReply) == 0):\n\t\t\t\t// there's new mail!\n\t\t\t\tif ex.pi.ExpectedReply != nil {\n\t\t\t\t\tex.Debug(\"Ping: Reply matched ExpectedReply\")\n\t\t\t\t}\n\t\t\t\tex.Debug(\"Ping: Got mail. Setting LongPollNewMail|msgCode=EAS_NEW_EMAIL\")\n\t\t\t\terrCh <- LongPollNewMail\n\t\t\t\treturn\n\t\t\t\t// EAS Sync\n\t\t\tcase ex.pi.ASIsSyncRequest == true && len(responseBody) == 0:\n\t\t\t\t// go back to polling\n\t\t\t\tif time.Since(timeSent) <= tooFastResponse {\n\t\t\t\t\tex.Warning(\"Sync: NoChangeReply after %s was too fast. Doing backoff. This usually indicates that the client is still connected to the exchange server.\", time.Since(timeSent))\n\t\t\t\t\tsleepTime = ex.exponentialBackoff(sleepTime)\n\t\t\t\t} else {\n\t\t\t\t\tex.Info(\"Sync: NoChangeReply after %s. Back to polling\", time.Since(timeSent))\n\t\t\t\t\tsleepTime = 0 // good reply. Reset any exponential backoff stuff.\n\t\t\t\t}\n\n\t\t\tcase ex.pi.ASIsSyncRequest == true && len(responseBody) > 0:\n\t\t\t\t// there's new mail!\n\t\t\t\tif ex.pi.ExpectedReply != nil {\n\t\t\t\t\tex.Debug(\"Sync: Reply matched ExpectedReply\")\n\t\t\t\t}\n\t\t\t\tex.Debug(\"Sync: Got mail. Setting LongPollNewMail|msgCode=EAS_NEW_EMAIL\")\n\t\t\t\terrCh <- LongPollNewMail\n\t\t\t\treturn\n\t\t\tdefault:\n\t\t\t\tex.Warning(\"Unhandled response. Just keep polling: Headers:%+v Body:%s\", response.Header, base64.StdEncoding.EncodeToString(responseBody))\n\t\t\t\tsleepTime = ex.exponentialBackoff(sleepTime)\n\t\t\t}\n\n\t\tcase <-stopPollCh: // parent will close this, at which point this will trigger.\n\t\t\tex.Debug(\"Was told to stop. Stopping\")\n\t\t\treturn\n\n\t\tcase <-stopAllCh: // parent will close this, at which point this will trigger.\n\t\t\tex.Debug(\"Was told to stop (allStop). Stopping\")\n\t\t\treturn\n\t\t}\n\t}\n}", "func (t *RisingMonitor) Start(_ time.Duration) (chan PricePoint, chan error) {\n\n\tpriceChan := make(chan PricePoint)\n\tt.tick = time.NewTicker(t.interval)\n\tgo risingPitcher(t.tick.C, t.start, priceChan)\n\tt.active = true\n\treturn priceChan, nil\n}", "func (c *watchAggregator) Start() {\n\tif c.autoWatch {\n\t\tc.startAutoWatch(true)\n\t} else if c.passiveClient != nil {\n\t\tc.startAutoWatch(false)\n\t}\n}", "func Start(ctx context.Context, restConfig *rest.Config) (*Monitor, error) {\n\tm := NewMonitorWithInterval(time.Second)\n\tclient, err := kubernetes.NewForConfig(restConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tconfigClient, err := configclientset.NewForConfig(restConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := StartKubeAPIMonitoringWithNewConnections(ctx, m, restConfig, 5*time.Second); err != nil {\n\t\treturn nil, err\n\t}\n\tif err := StartOpenShiftAPIMonitoringWithNewConnections(ctx, m, restConfig, 5*time.Second); err != nil {\n\t\treturn nil, err\n\t}\n\tif err := StartOAuthAPIMonitoringWithNewConnections(ctx, m, restConfig, 5*time.Second); err != nil {\n\t\treturn nil, err\n\t}\n\tif err := StartKubeAPIMonitoringWithConnectionReuse(ctx, m, restConfig, 5*time.Second); err != nil {\n\t\treturn nil, err\n\t}\n\tif err := StartOpenShiftAPIMonitoringWithConnectionReuse(ctx, m, restConfig, 5*time.Second); err != nil {\n\t\treturn nil, err\n\t}\n\tif err := StartOAuthAPIMonitoringWithConnectionReuse(ctx, m, restConfig, 5*time.Second); err != nil {\n\t\treturn nil, err\n\t}\n\tstartPodMonitoring(ctx, m, client)\n\tstartNodeMonitoring(ctx, m, client)\n\tstartEventMonitoring(ctx, m, client)\n\n\t// add interval creation at the same point where we add the monitors\n\tstartClusterOperatorMonitoring(ctx, m, configClient)\n\tm.intervalCreationFns = append(\n\t\tm.intervalCreationFns,\n\t\tintervalcreation.IntervalsFromEvents_OperatorAvailable,\n\t\tintervalcreation.IntervalsFromEvents_OperatorProgressing,\n\t\tintervalcreation.IntervalsFromEvents_OperatorDegraded,\n\t\tintervalcreation.IntervalsFromEvents_E2ETests,\n\t\tintervalcreation.IntervalsFromEvents_NodeChanges,\n\t)\n\n\tm.StartSampling(ctx)\n\treturn m, nil\n}", "func (h *LinkerdInfo) Start() {\n\tticker := time.NewTicker(linkerdInfoInterval)\n\tfor {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\th.handleCertsInfo()\n\t\tcase <-h.stopCh:\n\t\t\treturn\n\t\t}\n\t}\n}", "func (a *Agent) start() {\n\ta.initAPI()\n\tnb := 0\n\tfor {\n\t\ta.updateStreams()\n\t\tnb++\n\t\tif nb == 10 {\n\t\t\tlog.Printf(\"Sent %d logs and %d metrics on the last %d seconds\\n\", a.nbLogs, a.nbMetrics, nb*conf.period)\n\t\t\tnb = 0\n\t\t\ta.nbLogs = 0\n\t\t\ta.nbMetrics = 0\n\t\t}\n\t\ttime.Sleep(time.Duration(conf.period) * time.Second)\n\t}\n}", "func (p *NotebookWorkspacesStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (w *Worker) Start() {\n\tfor {\n\t\tselect {\n\t\tcase <-w.stopCh:\n\t\t\treturn\n\n\t\tcase services := <-w.serviceCh:\n\t\t\tlog.Info(\"Got services\")\n\n\t\t\tvhosts := make([]VirtualHost, 0)\n\n\t\t\tfor name := range services {\n\t\t\t\tvhost := VirtualHost{\n\t\t\t\t\tName: name,\n\t\t\t\t\tDomains: []string{\n\t\t\t\t\t\tfmt.Sprintf(\"%s.service.%s\", name, w.consulDomain),\n\t\t\t\t\t},\n\t\t\t\t\tRoutes: []Route{\n\t\t\t\t\t\tRoute{\n\t\t\t\t\t\t\tCluster: name,\n\t\t\t\t\t\t\tPrefix: \"/\",\n\t\t\t\t\t\t\tTimeoutMS: 3 * time.Minute,\n\t\t\t\t\t\t\tRetryPolicy: &RetryPolicy{\n\t\t\t\t\t\t\t\tRetryOn: \"5xx,connect-failure\",\n\t\t\t\t\t\t\t\tNumRetries: 1,\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t}\n\n\t\t\t\tif _, ok := services[\"api-users\"]; ok && name == \"api\" {\n\t\t\t\t\tvhost.Routes = append([]Route{{Cluster: \"api-users\", Prefix: \"/users\", RetryPolicy: vhost.Routes[0].RetryPolicy}}, vhost.Routes...)\n\t\t\t\t\tvhost.Routes = append([]Route{{Cluster: \"api-users\", Prefix: \"/oauth\", RetryPolicy: vhost.Routes[0].RetryPolicy}}, vhost.Routes...)\n\t\t\t\t\tvhost.Routes = append([]Route{{Cluster: \"api-users\", Prefix: \"/me\", RetryPolicy: vhost.Routes[0].RetryPolicy}}, vhost.Routes...)\n\t\t\t\t\tvhost.Routes = append([]Route{{Cluster: \"api-users\", Prefix: \"/emails\", RetryPolicy: vhost.Routes[0].RetryPolicy}}, vhost.Routes...)\n\t\t\t\t}\n\n\t\t\t\tvhosts = append(vhosts, vhost)\n\t\t\t}\n\n\t\t\tw.response = Response{VirtualHosts: vhosts}\n\t\t}\n\t}\n}", "func (p *DeploymentsStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (s *keyvisualService) run() {\n\t// TODO: make the ticker consistent with heartbeat interval\n\tticker := time.NewTicker(time.Minute)\n\tdefer ticker.Stop()\n\tfor {\n\t\tselect {\n\t\tcase <-s.ctx.Done():\n\t\t\treturn\n\t\tcase <-ticker.C:\n\t\t\tcluster := s.svr.GetRaftCluster()\n\t\t\tif cluster == nil || !serverapi.IsServiceAllowed(s.svr, defaultRegisterAPIGroupInfo) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\ts.scanRegions(cluster)\n\t\t\t// TODO: implements the stats\n\t\t}\n\t}\n}", "func (t *pollTrigger) Start(ctx context.Context) (<-chan bool, error) {\n\ttrigger := make(chan bool)\n\n\tticker := time.NewTicker(t.Interval)\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ticker.C:\n\t\t\t\ttrigger <- true\n\t\t\tcase <-ctx.Done():\n\t\t\t\tticker.Stop()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn trigger, nil\n}", "func (a *API) Run() error {\n\treturn a.e.Start(a.addr)\n}", "func (a *Agent) Run() error {\n\tuserCredentials := fmt.Sprintf(\"%s:%s\", a.config.User, a.config.Password)\n\tuserCredentials = base64.StdEncoding.EncodeToString([]byte(userCredentials))\n\theader := a.buildTransportHeaderMap()\n\theader.Set(\"Authorization\", \"Basic \"+userCredentials)\n\n\tconn, err := transport.Connect(a.backendSelector.Select(), a.config.TLS, header)\n\tif err != nil {\n\t\treturn err\n\t}\n\ta.conn = conn\n\n\tif _, _, err := a.createListenSockets(); err != nil {\n\t\treturn err\n\t}\n\n\t// These are in separate goroutines so that they can, theoretically, be executing\n\t// concurrently.\n\tgo a.sendPump(conn)\n\tgo a.receivePump(conn)\n\n\t// Send an immediate keepalive once we've connected.\n\tif err := a.sendKeepalive(); err != nil {\n\t\tlogger.Error(err)\n\t}\n\n\tgo func() {\n\t\tkeepaliveTicker := time.NewTicker(time.Duration(a.config.KeepaliveInterval) * time.Second)\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-keepaliveTicker.C:\n\t\t\t\tif err := a.sendKeepalive(); err != nil {\n\t\t\t\t\tlogger.WithError(err).Error(\"failed sending keepalive\")\n\t\t\t\t}\n\t\t\tcase <-a.stopping:\n\t\t\t\treturn\n\t\t\t}\n\n\t\t}\n\t}()\n\n\t// Prepare the HTTP API server\n\ta.api = newServer(a)\n\n\t// Start the HTTP API server\n\tgo func() {\n\t\tlogger.Info(\"starting api on address: \", a.api.Addr)\n\n\t\tif err := a.api.ListenAndServe(); err != http.ErrServerClosed {\n\t\t\tlogger.Fatal(err)\n\t\t}\n\t}()\n\n\t// Allow Stop() to block until the HTTP server shuts down.\n\ta.wg.Add(1)\n\tgo func() {\n\t\t// NOTE: This does not guarantee a clean shutdown of the HTTP API.\n\t\t// This is _only_ for the purpose of making Stop() a blocking call.\n\t\t// The goroutine running the HTTP Server has to return before Stop()\n\t\t// can return, so we use this to signal that goroutine to shutdown.\n\t\t<-a.stopping\n\t\tlogger.Info(\"api shutting down\")\n\n\t\tctx, cancel := context.WithTimeout(context.Background(), 1*time.Second)\n\t\tdefer cancel()\n\n\t\tif err := a.api.Shutdown(ctx); err != nil {\n\t\t\tlogger.Error(err)\n\t\t}\n\t\ta.wg.Done()\n\t}()\n\n\treturn nil\n}", "func (i *Interval) Start() error {\n\tif !i.latch.CanStart() {\n\t\treturn exception.New(ErrCannotStart)\n\t}\n\n\ti.latch.Starting()\n\tgo func() {\n\t\ti.latch.Started()\n\n\t\tif i.delay > 0 {\n\t\t\ttime.Sleep(i.delay)\n\t\t}\n\n\t\ttick := time.Tick(i.interval)\n\t\tvar err error\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-tick:\n\t\t\t\terr = i.action()\n\t\t\t\tif err != nil && i.errors != nil {\n\t\t\t\t\ti.errors <- err\n\t\t\t\t}\n\t\t\tcase <-i.latch.NotifyStopping():\n\t\t\t\ti.latch.Stopped()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\t<-i.latch.NotifyStarted()\n\treturn nil\n}", "func (p *StreamingEndpointsCreatePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (g *Engine) Start() error {\n\tudpListeners := make([]*net.UDPConn, len(g.addrs))[0:0]\n\tswitch g.network {\n\tcase \"tcp\", \"tcp4\", \"tcp6\":\n\t\tfor i := range g.addrs {\n\t\t\tln, err := newPoller(g, true, i)\n\t\t\tif err != nil {\n\t\t\t\tfor j := 0; j < i; j++ {\n\t\t\t\t\tg.listeners[j].stop()\n\t\t\t\t}\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tg.addrs[i] = ln.listener.Addr().String()\n\t\t\tg.listeners = append(g.listeners, ln)\n\t\t}\n\tcase \"udp\", \"udp4\", \"udp6\":\n\t\tfor i, addrStr := range g.addrs {\n\t\t\taddr, err := net.ResolveUDPAddr(g.network, addrStr)\n\t\t\tif err != nil {\n\t\t\t\tfor j := 0; j < i; j++ {\n\t\t\t\t\tudpListeners[j].Close()\n\t\t\t\t}\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tln, err := g.listenUDP(\"udp\", addr)\n\t\t\tif err != nil {\n\t\t\t\tfor j := 0; j < i; j++ {\n\t\t\t\t\tudpListeners[j].Close()\n\t\t\t\t}\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tg.addrs[i] = ln.LocalAddr().String()\n\t\t\tudpListeners = append(udpListeners, ln)\n\t\t}\n\t}\n\n\tfor i := 0; i < g.pollerNum; i++ {\n\t\tp, err := newPoller(g, false, i)\n\t\tif err != nil {\n\t\t\tfor j := 0; j < len(g.listeners); j++ {\n\t\t\t\tg.listeners[j].stop()\n\t\t\t}\n\n\t\t\tfor j := 0; j < i; j++ {\n\t\t\t\tg.pollers[j].stop()\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\t\tg.pollers[i] = p\n\t}\n\n\tfor i := 0; i < g.pollerNum; i++ {\n\t\tg.Add(1)\n\t\tgo g.pollers[i].start()\n\t}\n\tfor _, l := range g.listeners {\n\t\tg.Add(1)\n\t\tgo l.start()\n\t}\n\n\tfor _, ln := range udpListeners {\n\t\t_, err := g.AddConn(ln)\n\t\tif err != nil {\n\t\t\tfor j := 0; j < len(g.listeners); j++ {\n\t\t\t\tg.listeners[j].stop()\n\t\t\t}\n\n\t\t\tfor j := 0; j < len(g.pollers); j++ {\n\t\t\t\tg.pollers[j].stop()\n\t\t\t}\n\n\t\t\tfor j := 0; j < len(udpListeners); j++ {\n\t\t\t\tudpListeners[j].Close()\n\t\t\t}\n\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// g.Timer.Start()\n\n\tif len(g.addrs) == 0 {\n\t\tlogging.Info(\"NBIO[%v] start\", g.Name)\n\t} else {\n\t\tlogging.Info(\"NBIO[%v] start listen on: [\\\"%v@%v\\\"]\", g.Name, g.network, strings.Join(g.addrs, `\", \"`))\n\t}\n\treturn nil\n}", "func (am *AppManager) Start() {\n\tlogger.Printf(\"Starting Goroutine to refresh applications data every %d minute(s)\\n\", am.appUpdateInterval)\n\t//get the data as soon as possible\n\tgo am.refreshAppData()\n\tticker := time.NewTicker(time.Duration(int64(am.appUpdateInterval)) * time.Minute)\n\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ticker.C:\n\t\t\t\tgo am.refreshAppData()\n\n\t\t\tcase tempAppInfo := <-am.updateChannel:\n\t\t\t\tlogger.Printf(\"App Update....received %d app details\", len(tempAppInfo))\n\t\t\t\tam.appData = tempAppInfo\n\n\t\t\tcase rr := <-am.readChannel:\n\t\t\t\tad := am.getAppData(rr.appGUID)\n\t\t\t\trr.responseChan <- ad\n\n\t\t\tcase <-am.closeChannel:\n\t\t\t\tlogger.Print(\"quit \\r\\n\")\n\t\t\t\tticker.Stop()\n\t\t\t}\n\t\t}\n\t}()\n}", "func (p *ManagedClustersStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func StartFetcher(f Fetcher, timer time.Duration) (<-chan *darksky.Forecast, chan string) {\n\tcontrolChannel := make(chan string)\n\tforecastChannel := make(chan *darksky.Forecast)\n\n\tticker := time.NewTicker(timer)\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ticker.C:\n\t\t\t\tforecast, err := f.Fetch()\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Println(\"[ERROR] unable to get forecast\", err)\n\t\t\t\t}\n\t\t\t\tforecastChannel <- forecast\n\t\t\tcase <-controlChannel:\n\t\t\t\tclose(forecastChannel)\n\t\t\t\tticker.Stop()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn forecastChannel, controlChannel\n}", "func (c *Client) Start(ctx context.Context) error {\n\tif c.observer == nil || c.observer.listener == nil {\n\t\tlevel.Warn(c.logger).Log(xlog.MessageKey(), \"No listener was setup to receive updates.\")\n\t\treturn nil\n\t}\n\tif c.observer.ticker == nil {\n\t\tlevel.Error(c.logger).Log(xlog.MessageKey(), \"Observer ticker is nil\")\n\t\treturn ErrUndefinedIntervalTicker\n\t}\n\n\tif !atomic.CompareAndSwapInt32(&c.observer.state, stopped, transitioning) {\n\t\tlevel.Error(c.logger).Log(xlog.MessageKey(), \"Start called when a listener was not in stopped state\", \"err\", ErrListenerNotStopped)\n\t\treturn ErrListenerNotStopped\n\t}\n\n\tc.observer.ticker.Reset(c.observer.pullInterval)\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-c.observer.shutdown:\n\t\t\t\treturn\n\t\t\tcase <-c.observer.ticker.C:\n\t\t\t\toutcome := SuccessOutcome\n\t\t\t\tctx := c.setLogger(context.Background(), c.logger)\n\t\t\t\titems, err := c.GetItems(ctx, \"\")\n\t\t\t\tif err == nil {\n\t\t\t\t\tc.observer.listener.Update(items)\n\t\t\t\t} else {\n\t\t\t\t\toutcome = FailureOutcome\n\t\t\t\t\tlevel.Error(c.logger).Log(xlog.MessageKey(), \"Failed to get items for listeners\", xlog.ErrorKey(), err)\n\t\t\t\t}\n\t\t\t\tc.observer.measures.Polls.With(prometheus.Labels{\n\t\t\t\t\tOutcomeLabel: outcome}).Add(1)\n\t\t\t}\n\t\t}\n\t}()\n\n\tatomic.SwapInt32(&c.observer.state, running)\n\treturn nil\n}", "func (w *WatchManager) run() {\n\tw.pollUpdatesInWasp() // initial pull from WASP\n\trunning := true\n\tfor running {\n\t\tselect {\n\t\tcase <-time.After(1 * time.Minute):\n\t\t\tw.pollUpdatesInWasp()\n\n\t\tcase <-w.stopChannel:\n\t\t\trunning = false\n\t\t}\n\n\t\ttime.Sleep(1 * time.Second)\n\t}\n}", "func (TelegramBotApp *TelegramBotApp) setupPolling() (tgbotapi.UpdatesChannel, error) {\n\tTelegramBotApp.bot.RemoveWebhook()\n\tupdateConfig := tgbotapi.NewUpdate(0)\n\tupdateConfig.Timeout = 5\n\tfmt.Println(\"[+] Pooling method selected\")\n\treturn TelegramBotApp.bot.GetUpdatesChan(updateConfig)\n}", "func PollLocal(ConfigObject *common.Config, aggregateFile string, forever bool) {\n\tduration := ConfigObject.JsonConfig.LocalMetrics.Scrape_interval\n\tserviceType := ConfigObject.JsonConfig.LocalMetrics.Type\n\t// Setup storage dir\n\tcommon.SetupStorage(ConfigObject, serviceType, aggregateFile)\n\tfor {\n\t\tif len(ConfigObject.JsonConfig.LocalMetrics.Urls) <= 0 {\n\t\t\tfmt.Println(\"-------------- Nothing to monitor for Local Url --------------\")\n\t\t\tfmt.Println(\"-------------- Stopping the thread --------------\")\n\t\t\tbreak\n\t\t}\n\t\tfor i := 0; i < len(ConfigObject.JsonConfig.LocalMetrics.Urls); i++ {\n\t\t\tappID := ConfigObject.JsonConfig.LocalMetrics.Urls[i].Name\n\t\t\turl := ConfigObject.JsonConfig.LocalMetrics.Urls[i].Url\n\t\t\tresultTmpDir := strings.Replace(appID, \"/\", \"_\", -1)\n\t\t\tConfigObject.TW.Add(1)\n\t\t\tgo common.ScrapeMetrics(ConfigObject, url, serviceType, resultTmpDir, ConfigObject, appID, hrc)\n\t\t}\n\t\tConfigObject.TW.Wait()\n\t\tConfigObject.TW.Add(1)\n\t\tgo common.AggregateData(ConfigObject, aggregateFile, serviceType)\n\t\tConfigObject.TW.Wait()\n\t\t<-time.After(time.Duration(duration) * time.Second)\n\t}\n}", "func (n *Nozzle) Start() {\n\trx := n.s.Stream(context.Background(), n.buildBatchReq())\n\n\tgo n.timerProcessor()\n\tgo n.timerEmitter()\n\tgo n.envelopeReader(rx)\n\n\tn.log.Info(\"starting workers\", logger.Count(2*runtime.NumCPU()))\n\tfor i := 0; i < 2*runtime.NumCPU(); i++ {\n\t\tgo n.pointWriter()\n\t}\n\n\tgo n.pointBatcher()\n}", "func (r *AutoRoller) Start(ctx context.Context, tickFrequency time.Duration) {\n\tsklog.Infof(\"Starting autoroller.\")\n\tlv := metrics2.NewLiveness(\"last_successful_autoroll_tick\", map[string]string{\"roller\": r.roller})\n\tcleanup.Repeat(tickFrequency, func(_ context.Context) {\n\t\t// Explicitly ignore the passed-in context; this allows us to\n\t\t// continue running even if the context is canceled, which helps\n\t\t// to prevent errors due to interrupted syncs, etc.\n\t\tctx := context.Background()\n\t\tif err := r.Tick(ctx); err != nil {\n\t\t\t// Hack: we frequently get failures from GoB which trigger error-rate alerts.\n\t\t\t// These alerts are noise and sometimes hide real failures. If the error is\n\t\t\t// due to a sync failure, log it as a warning instead of an error. We'll rely\n\t\t\t// on the liveness alert in the case where we see persistent sync failures.\n\t\t\tif isSyncError(err) {\n\t\t\t\tsklog.Warningf(\"Failed to run autoroll: %s\", err)\n\t\t\t} else {\n\t\t\t\tsklog.Errorf(\"Failed to run autoroll: %s\", err)\n\t\t\t}\n\t\t} else {\n\t\t\tlv.Reset()\n\t\t}\n\t}, nil)\n\n\t// Update the current reviewers in a loop.\n\tlvReviewers := metrics2.NewLiveness(\"last_successful_reviewers_retrieval\", map[string]string{\"roller\": r.roller})\n\tcleanup.Repeat(30*time.Minute, func(ctx context.Context) {\n\t\temails := GetReviewers(r.client, r.cfg.RollerName, r.cfg.Reviewer, r.cfg.ReviewerBackup)\n\t\tr.emailsMtx.Lock()\n\t\tdefer r.emailsMtx.Unlock()\n\t\tr.emails = emails\n\n\t\tconfigCopies := replaceReviewersPlaceholder(r.cfg.Notifiers, emails)\n\t\tif err := r.notifier.ReloadConfigs(ctx, configCopies); err != nil {\n\t\t\tsklog.Errorf(\"Failed to reload configs: %s\", err)\n\t\t\treturn\n\t\t}\n\t\tlvReviewers.Reset()\n\t}, nil)\n\n\t// Handle requests for manual rolls.\n\tif r.cfg.SupportsManualRolls {\n\t\tlvManualRolls := metrics2.NewLiveness(\"last_successful_manual_roll_check\", map[string]string{\"roller\": r.roller})\n\t\tcleanup.Repeat(time.Minute, func(_ context.Context) {\n\t\t\t// Explicitly ignore the passed-in context; this allows\n\t\t\t// us to continue handling manual rolls even if the\n\t\t\t// context is canceled, which helps to prevent errors\n\t\t\t// due to interrupted syncs, etc.\n\t\t\tctx := context.Background()\n\t\t\tif err := r.handleManualRolls(ctx); err != nil {\n\t\t\t\tsklog.Error(err)\n\t\t\t} else {\n\t\t\t\tlvManualRolls.Reset()\n\t\t\t}\n\t\t}, nil)\n\t}\n}", "func (s *Refresh) StartBackgroundRefresh() {\n\tgo s.FindServerStateAdded()\n}", "func (h *HookbotTrigger) Start() error {\n\tfinishCh := make(chan struct{})\n\tmsgCh, errCh := listen.RetryingWatch(h.Endpoint, http.Header{}, finishCh)\n\tgo h.errorHandler(errCh)\n\tgo h.msgHandler(msgCh)\n\treturn nil\n}", "func (m *Merge) Start(ctx context.Context, interval time.Duration) {\n\tticker := time.NewTicker(interval)\n\tdefer ticker.Stop()\n\tfor {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\tm.Main(ctx)\n\t\tcase <-m.ctx.Done():\n\t\t\treturn\n\t\t}\n\t}\n}", "func StartPollingHandler(w http.ResponseWriter, r *http.Request) {\n\tenv := envFromRequest(r)\n\n\tscript := polling.GenStartScript(env.Logger, env.BaseURL)\n\n\tw.Write([]byte(script))\n}", "func (rl *RateLimit) startRequests() {\n\tfor {\n\t\t// start a user request every 1 second\n\t\tgo func() {\n\t\t\tsize := rl.requestSize()\n\t\t\terr := make(chan error)\n\t\t\tgo rl.VerifyQuota(size, err)\n\t\t\tif e := <-err; e != nil {\n\t\t\t\t// respond to the user about quota exhaustion\n\t\t\t\tlog.Println(e)\n\t\t\t} else {\n\t\t\t\tlog.Println(\"request OK to be sent downstream\")\n\t\t\t}\n\t\t}()\n\t\ttime.Sleep(500 * time.Millisecond)\n\t}\n}", "func (s *JRPCServer) Start() {\n\thttpCall := s.httpSrv.On(s.endpointURL)\n\thttpCall.Forever()\n\thttpCall.handlerFunc = func(w http.ResponseWriter, req *http.Request) error {\n\t\t// ctx := context.Background()\n\t\ts.guard.Lock()\n\t\tdefer s.guard.Unlock()\n\t\tjReq := btcjson.Request{}\n\t\tbuf, err := ioutil.ReadAll(req.Body)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unable to decode jRPC request: %v\", err)\n\t\t}\n\t\terr = req.Body.Close()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treq.Body = ioutil.NopCloser(bytes.NewBuffer(buf))\n\t\tmustUnmarshal(buf, &jReq)\n\t\tcall, err := s.findCall(jReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// put unmarshalled JRPC request into a context\n\t\tctx := context.WithValue(req.Context(), jRPCRequestKey, jReq)\n\t\treturn call.execute(w, req.WithContext(ctx))\n\t}\n\ts.httpSrv.Start()\n}", "func runTick() {\n ticker := time.NewTicker(time.Duration(*requestBurst * 1000000 / *requestRate) * time.Microsecond)\n for range ticker.C {\n //create \"burst\" go routines that each make 1 get request\n for i := 0; i < *requestBurst; i++ {\n\t // generate 1 request\n\t go generateRequest()\n }\n }\n}", "func (p *DeploymentsStartJFRPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func pollerActor(logger log.Logger, status chan<- []*ec2.VpnConnection, svc ec2iface.EC2API, interval *time.Duration) actor.Actor {\n\n\tcancel := make(chan struct{})\n\tinput := &ec2.DescribeVpnConnectionsInput{}\n\tticker := time.NewTicker(*interval)\n\n\treturn actor.NewActor(\n\t\tfunc() error {\n\n\t\t\tfor {\n\n\t\t\t\tresult, err := svc.DescribeVpnConnections(input)\n\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tstatus <- result.VpnConnections\n\t\t\t\t_ = level.Debug(logger).Log(\"msg\", \"Sent updated VPN telemetry data to next stage\")\n\n\t\t\t\tselect {\n\n\t\t\t\tcase <-ticker.C:\n\t\t\t\t\t_ = level.Debug(logger).Log(\"msg\", \"Waking up\")\n\t\t\t\t\tcontinue\n\n\t\t\t\tcase <-cancel:\n\t\t\t\t\t_ = level.Info(logger).Log(\"cancelled\", \"Asked to terminate\")\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\t}\n\t\t},\n\t\tfunc(err error) {\n\t\t\t_ = level.Info(logger).Log(\"interrupted\", fmt.Sprintf(\"interrupted with %v\", err))\n\t\t\tclose(cancel)\n\t\t},\n\t)\n\n}", "func (mc *EndpointsWatcher) StartWatcher(quitCh chan struct{}, wg *sync.WaitGroup) {\n\tdefer wg.Done()\n\tfor {\n\t\twatcher := cache.NewListWatchFromClient(mc.clientset.CoreV1().RESTClient(), mc.resourceStr, v1.NamespaceAll, fields.Everything())\n\t\tretryWatcher, err := watchClient.NewRetryWatcher(mc.lastRV, watcher)\n\t\tif err != nil {\n\t\t\tlog.WithError(err).Fatal(\"Could not start watcher for k8s resource: \" + mc.resourceStr)\n\t\t}\n\n\t\tresCh := retryWatcher.ResultChan()\n\t\trunWatcher := true\n\t\tfor runWatcher {\n\t\t\tselect {\n\t\t\tcase <-quitCh:\n\t\t\t\treturn\n\t\t\tcase c := <-resCh:\n\t\t\t\ts, ok := c.Object.(*metav1.Status)\n\t\t\t\tif ok && s.Status == metav1.StatusFailure {\n\t\t\t\t\tif s.Reason == metav1.StatusReasonGone {\n\t\t\t\t\t\tlog.WithField(\"resource\", mc.resourceStr).Info(\"Requested resource version too old, no longer stored in K8S API\")\n\t\t\t\t\t\trunWatcher = false\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t\t// Ignore and let the retry watcher retry.\n\t\t\t\t\tlog.WithField(\"resource\", mc.resourceStr).WithField(\"object\", c.Object).Info(\"Failed to read from k8s watcher\")\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\t// Update the lastRV, so that if the watcher restarts, it starts at the correct resource version.\n\t\t\t\to, ok := c.Object.(*v1.Endpoints)\n\t\t\t\tif !ok {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tmc.lastRV = o.ObjectMeta.ResourceVersion\n\n\t\t\t\tpb, err := protoutils.EndpointsToProto(o)\n\t\t\t\tif err != nil {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tr := &storepb.K8SResource{\n\t\t\t\t\tResource: &storepb.K8SResource_Endpoints{\n\t\t\t\t\t\tEndpoints: pb,\n\t\t\t\t\t},\n\t\t\t\t}\n\n\t\t\t\tmsg := &K8sResourceMessage{\n\t\t\t\t\tObject: r,\n\t\t\t\t\tObjectType: mc.resourceStr,\n\t\t\t\t\tEventType: c.Type,\n\t\t\t\t}\n\t\t\t\tmc.updateCh <- msg\n\t\t\t}\n\t\t}\n\n\t\tlog.WithField(\"resource\", mc.resourceStr).Info(\"K8s watcher channel closed. Retrying\")\n\n\t\t// Wait 5 minutes before retrying, however if stop is called, just return.\n\t\tselect {\n\t\tcase <-quitCh:\n\t\t\treturn\n\t\tcase <-time.After(5 * time.Minute):\n\t\t\tcontinue\n\t\t}\n\t}\n}", "func (iw *InstanceWorker) APILocalFeedMonitoring() {\n\tfmt.Printf(\"starting local feed API monitoring for %s\\n\", iw.Instance.Domain)\n\tvar timeFrame time.Duration\n\tu := url.URL{\n\t\tScheme: \"https\",\n\t\tHost: iw.Instance.Domain,\n\t\tPath: \"/api/v1/timelines/public\",\n\t\tRawQuery: \"local=true&limit=50\",\n\t}\n\tfor {\n\t\tresp, err := http.Get(u.String())\n\t\tif err == nil {\n\t\t\tbody, err := ioutil.ReadAll(resp.Body)\n\t\t\tresp.Body.Close()\n\t\t\tif err == nil {\n\t\t\t\tvar statuses []mastodon.Status\n\t\t\t\tif err = json.Unmarshal(body, &statuses); err == nil {\n\t\t\t\t\tif len(statuses) > 10 { // don't lose time with sleeping instances…\n\t\t\t\t\t\toldest := time.Now()\n\t\t\t\t\t\tmost_recent := time.Now()\n\t\t\t\t\t\tfor _, status := range statuses {\n\t\t\t\t\t\t\tiw.SaveIfUnknown(status.Account)\n\t\t\t\t\t\t\tif status.CreatedAt.After(most_recent) {\n\t\t\t\t\t\t\t\tmost_recent = status.CreatedAt\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tif status.CreatedAt.Before(oldest) {\n\t\t\t\t\t\t\t\toldest = status.CreatedAt\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t\ttimeFrame = most_recent.Sub(oldest)\n\t\t\t\t\t} else {\n\t\t\t\t\t\ttimeFrame = 17 * time.Hour\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\ttimeFrame = time.Duration(float64(timeFrame) * 0.75)\n\t\tif timeFrame > (12 * time.Hour) {\n\t\t\ttimeFrame = 12 * time.Hour\n\t\t}\n\t\ttime.Sleep(timeFrame)\n\t}\n\t/*\n\t\t- fetch json from instance's URL\n\t\t- partially unmarshal json to retrieve toots' id and accounts\n\t\t- lookup worker's toots map\n\t\t- if toot not found, add it to worker's map and launch 'SaveIfUnknown' func\n\t\t- cleanup vars (json, etc.)\n\t*/\n}", "func (s *StopWatch) Start() {\n if !s.running {\n s.start = time.Now()\n s.running = true\n }\n}", "func (p *Probe) Start(ctx context.Context, dataChan chan *metrics.EventMetrics) {\n\tdefer p.wait()\n\n\tp.updateTargetsAndStartProbes(ctx, dataChan)\n\n\t// Do more frequent listing of targets until we get a non-zero list of\n\t// targets.\n\tinitialRefreshInterval := p.opts.Interval\n\t// Don't wait too long if p.opts.Interval is large.\n\tif initialRefreshInterval > time.Second {\n\t\tinitialRefreshInterval = time.Second\n\t}\n\n\tfor {\n\t\tif ctxDone(ctx) {\n\t\t\treturn\n\t\t}\n\t\tif len(p.targets) != 0 {\n\t\t\tbreak\n\t\t}\n\t\tp.updateTargetsAndStartProbes(ctx, dataChan)\n\t\ttime.Sleep(initialRefreshInterval)\n\t}\n\n\ttargetsUpdateTicker := time.NewTicker(p.targetsUpdateInterval)\n\tdefer targetsUpdateTicker.Stop()\n\n\tfor {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn\n\t\tcase <-targetsUpdateTicker.C:\n\t\t\tp.updateOauthToken()\n\t\t\tp.updateTargetsAndStartProbes(ctx, dataChan)\n\t\t}\n\t}\n}", "func (e *Epazote) Start(sk Scheduler, debug bool) {\n\tif debug {\n\t\te.debug = true\n\t}\n\n\tfor k, v := range e.Services {\n\t\t// Set service name\n\t\tv.Name = k\n\n\t\t// Status\n\t\tif v.Expect.Status < 1 {\n\t\t\tv.Expect.Status = 200\n\t\t}\n\n\t\t// rxBody\n\t\tif v.Expect.Body != \"\" {\n\t\t\tre := regexp.MustCompile(v.Expect.Body)\n\t\t\tv.Expect.body = re\n\t\t}\n\n\t\t// retry\n\t\tif v.RetryInterval == 0 {\n\t\t\tv.RetryInterval = 500\n\t\t}\n\t\tif v.RetryLimit == 0 {\n\t\t\tv.RetryLimit = 3\n\t\t}\n\n\t\tif v.Test.Test != \"\" {\n\t\t\tv.Test.Test = strings.TrimSpace(v.Test.Test)\n\t\t}\n\n\t\tif e.debug {\n\t\t\tif v.URL != \"\" {\n\t\t\t\tlog.Printf(Green(\"Adding service: %s URL: %s\"), v.Name, v.URL)\n\t\t\t} else {\n\t\t\t\tlog.Printf(Green(\"Adding service: %s Test: %s\"), v.Name, v.Test.Test)\n\t\t\t}\n\t\t}\n\n\t\t// schedule the service\n\t\tsk.AddScheduler(k, GetInterval(60, v.Every), e.Supervice(v))\n\t}\n\n\t// initialize map in case of no services\n\tif e.Services == nil {\n\t\te.Services = make(map[string]*Service)\n\t}\n\n\tif len(e.Config.Scan.Paths) > 0 {\n\t\tfor _, v := range e.Config.Scan.Paths {\n\t\t\tsk.AddScheduler(v, GetInterval(300, e.Config.Scan.Every), e.Scan(v))\n\t\t\t// schedule the scan but also scan at the beginning\n\t\t\tgo e.search(v, false)\n\t\t}\n\t}\n\n\tlog.Printf(\"Epazote %c on %d services, scan paths: %s [pid: %d]\",\n\t\tIcon(herb),\n\t\tlen(e.Services),\n\t\tstrings.Join(e.Config.Scan.Paths, \",\"),\n\t\tos.Getpid())\n}", "func (s *Service) Boot() {\n\n\t// run tick goroutine\n\ttickChan := make(chan bool)\n\ts.logger.LogDebug(\"booting loop for interval %ds\", int(s.fetchInterval.Seconds()))\n\tgo intervalTick(int(s.fetchInterval.Seconds()), s.jitterSec, tickChan)\n\tgo s.notificationSentTimestampOperator()\n\n\t// run infinite loop\n\tfor {\n\t\t// wait until we reached another interval tick\n\t\tselect {\n\t\tcase <-tickChan:\n\t\t}\n\t\terr := s.mainLoop()\n\n\t\tif err != nil {\n\t\t\ts.logger.LogError(err, \"mainLoop failed\")\n\t\t}\n\t}\n\n}", "func (wss *WssClientContext) Start() {\n\tfor {\n\n\t\tu := url.URL{\n\t\t\tScheme: \"ws\",\n\t\t\tOpaque: \"\",\n\t\t\tUser: &url.Userinfo{},\n\t\t\tHost: wssConfig.Server + \":\" + strconv.Itoa(wssConfig.Port),\n\t\t\tPath: \"\",\n\t\t\tRawPath: \"\",\n\t\t\tForceQuery: false,\n\t\t\tRawQuery: \"\",\n\t\t\tFragment: \"\",\n\t\t\tRawFragment: \"\",\n\t\t}\n\t\tc, _, err := websocket.DefaultDialer.Dial(u.String(), nil)\n\t\tif err != nil {\n\t\t\tlogger.Errorf(\"Can't access server \")\n\t\t\tcontinue\n\t\t}\n\t\tdefer c.Close()\n\t\t_, msg, err := c.ReadMessage()\n\t\tif err != nil {\n\t\t\tlogger.Errorf(\"Can't read message\")\n\t\t\tcontinue\n\t\t}\n\t\tmsgString := string(msg)\n\t\tlogger.Infof(\"Recieve message: \" + msgString)\n\n\t\ttime.Sleep(time.Duration(3) * time.Second)\n\t}\n}", "func (a *api) Run() {\n\tport := strconv.FormatInt(int64(a.Config.HttpPort), 10)\n\ta.logger.Info(\"Running API on \", port)\n\tgraceful.Run(\":\"+port, 100*time.Millisecond, a.ge)\n}", "func StartGettingWeather() {\n\t// get some initial data from start\n\t// mainWeatherGetter()\n\n\tfor i := range time.Tick(time.Second * time.Duration(delay)) {\n\t\t_ = i\n\t\tloopCounter++\n\t\tfmt.Println(time.Now().Format(time.RFC850), \" counter: \", loopCounter)\n\t\tmainWeatherGetter()\n\t}\n}", "func (c Consumer) Start() {\n\tfor {\n\t\tlog.Print(\"Consumer: Request initiated from consumer: \", c.id)\n\t\tc.RequestWidgets()\n\t}\n}", "func (p *CassandraClustersClientStartPoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (site *Site) Run(interval time.Duration) {\n\tloadpointChan := make(chan Updater)\n\tgo site.loopLoadpoints(loadpointChan)\n\n\tticker := time.NewTicker(interval)\n\tsite.update(<-loadpointChan) // start immediately\n\n\tfor {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\tsite.update(<-loadpointChan)\n\t\tcase lp := <-site.lpUpdateChan:\n\t\t\tsite.update(lp)\n\t\t}\n\t}\n}", "func startCrawling(start string) {\n\tcheckIndexPresence()\n\n\tvar wg sync.WaitGroup\n\tnoOfWorkers := 10\n\n\t// Send first url to the channel\n\tgo func(s string) {\n\t\tqueue <- s\n\t}(start)\n\n\t// Create worker pool with noOfWorkers workers\n\twg.Add(noOfWorkers)\n\tfor i := 1; i <= noOfWorkers; i++ {\n\t\tgo worker(&wg, i)\n\t}\n\twg.Wait()\n}", "func (tb *TelemetryBuffer) Start(intervalms time.Duration) {\n\tdefer tb.close()\n\tif !tb.fdExists && tb.connected {\n\t\tif intervalms < DefaultInterval {\n\t\t\tintervalms = DefaultInterval\n\t\t}\n\n\t\tinterval := time.NewTicker(intervalms).C\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-interval:\n\t\t\t\t// Send payload to host and clear cache when sent successfully\n\t\t\t\t// To-do : if we hit max slice size in payload, write to disk and process the logs on disk on future sends\n\t\t\t\tif err := tb.sendToHost(); err == nil {\n\t\t\t\t\ttb.payload.reset()\n\t\t\t\t}\n\t\t\tcase report := <-tb.data:\n\t\t\t\ttb.payload.push(report)\n\t\t\tcase <-tb.cancel:\n\t\t\t\tgoto EXIT\n\t\t\t}\n\t\t}\n\t} else {\n\t\t<-tb.cancel\n\t}\n\nEXIT:\n}", "func (socket *Socket) StartPolling() {\r\n\tsocket.stopPollChan = make(chan struct{})\r\n\tsocket.wg.Add(1)\r\n\tgo func() {\r\n\t\tvar events [maxEpollEvents]syscall.EpollEvent\r\n\t\tdefer socket.wg.Done()\r\n\r\n\t\tfor {\r\n\t\t\tselect {\r\n\t\t\tcase <-socket.stopPollChan:\r\n\t\t\t\treturn\r\n\t\t\tdefault:\r\n\t\t\t\tnum, err := syscall.EpollWait(socket.epfd, events[:], -1)\r\n\t\t\t\tif err != nil {\r\n\t\t\t\t\tsocket.ErrChan <- fmt.Errorf(\"epollWait: %v\", err)\r\n\t\t\t\t\treturn\r\n\t\t\t\t}\r\n\r\n\t\t\t\tfor ev := 0; ev < num; ev++ {\r\n\t\t\t\t\tif events[0].Fd == socket.wakeEvent.Fd {\r\n\t\t\t\t\t\tcontinue\r\n\t\t\t\t\t}\r\n\t\t\t\t\terr = socket.handleEvent(&events[0])\r\n\t\t\t\t\tif err != nil {\r\n\t\t\t\t\t\tsocket.ErrChan <- fmt.Errorf(\"handleEvent: %v\", err)\r\n\t\t\t\t\t}\r\n\t\t\t\t}\r\n\t\t\t}\r\n\t\t}\r\n\t}()\r\n}", "func (c *Client) Start(ctx context.Context, productMap exchange.ProductMap, exchangeDoneCh chan<- struct{}) error {\n\tc.productMap = productMap[c.exchangeName]\n\terr := c.GetPairs()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = c.API.Connect(c.GetURL())\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = c.API.SendSubscribeRequest(c.FormatSubscribeRequest())\n\tif err != nil {\n\t\treturn err\n\t}\n\tgo c.StartTickerListener(ctx, exchangeDoneCh)\n\treturn nil\n}" ]
[ "0.6498689", "0.64851135", "0.64312106", "0.6397514", "0.63825715", "0.6344594", "0.6055021", "0.59994274", "0.59954834", "0.5936349", "0.5913254", "0.59047526", "0.5847434", "0.5748641", "0.56952643", "0.5693969", "0.5684528", "0.56824774", "0.5681835", "0.5674555", "0.5671568", "0.5653857", "0.56536263", "0.5650007", "0.5602763", "0.559277", "0.5588531", "0.55848986", "0.55549675", "0.55397743", "0.553308", "0.55329716", "0.5525544", "0.5498398", "0.54924744", "0.5475592", "0.5465721", "0.5439053", "0.54322684", "0.54248863", "0.54209423", "0.5408447", "0.5400997", "0.54007477", "0.53985304", "0.5397891", "0.53893906", "0.5381034", "0.53725773", "0.5365288", "0.5362025", "0.5343434", "0.5342817", "0.53358585", "0.5322", "0.53097546", "0.52966934", "0.52890736", "0.5286533", "0.5282733", "0.5276533", "0.5275376", "0.5262321", "0.52545995", "0.5249601", "0.5247767", "0.5239548", "0.5238626", "0.52288795", "0.52130985", "0.5212979", "0.5203716", "0.5203139", "0.5196553", "0.51955765", "0.5190943", "0.5188744", "0.5186378", "0.51773864", "0.5175815", "0.51754016", "0.51691586", "0.5166455", "0.5164783", "0.51634413", "0.5158356", "0.51582134", "0.5153815", "0.51483804", "0.51371086", "0.5132352", "0.5131929", "0.5130965", "0.51303935", "0.51285285", "0.5121558", "0.51165813", "0.5112623", "0.5110724", "0.51062757" ]
0.7321983
0
Stop sends a shutdown signal to the polling goroutine to return
func (p *Poller) Stop() { wg := &sync.WaitGroup{} wg.Add(1) p.Shutdown <- wg wg.Wait() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (rcsw *RemoteClusterServiceWatcher) Stop(cleanupState bool) {\n\trcsw.probeEventsSink.send(&ClusterNotRegistered{\n\t\tclusterName: rcsw.clusterName,\n\t})\n\tclose(rcsw.stopper)\n\tif cleanupState {\n\t\trcsw.eventsQueue.Add(&ClusterUnregistered{})\n\t}\n\trcsw.eventsQueue.ShutDown()\n}", "func (w *UnbondingWatcher) Stop() {\n\tclose(w.quit)\n}", "func (m *Mainloop) Stop() {\n\tgo func() { m.termchan <- 1 }()\n\treturn\n}", "func (p *Poller) Stop() {\n\tp.stopMutex.Lock()\n\tdefer p.stopMutex.Unlock()\n\n\tp.isStopped = true\n\tclose(p.Channel)\n}", "func (m *WebsocketRoutineManager) Stop() error {\n\tif m == nil {\n\t\treturn fmt.Errorf(\"websocket routine manager %w\", ErrNilSubsystem)\n\t}\n\n\tm.mu.Lock()\n\tif atomic.LoadInt32(&m.state) == stoppedState {\n\t\tm.mu.Unlock()\n\t\treturn fmt.Errorf(\"websocket routine manager %w\", ErrSubSystemNotStarted)\n\t}\n\tatomic.StoreInt32(&m.state, stoppedState)\n\tm.mu.Unlock()\n\n\tclose(m.shutdown)\n\tm.wg.Wait()\n\n\treturn nil\n}", "func Stop() {\n\tquit <- true\n\tfor id, socket := range sockets {\n\t\tfmt.Printf(\"Closing socket %s\\n\", id)\n\t\tsocket.Close()\n\t\tpoller.RemoveBySocket(socket)\n\t}\n\tzctx.Term()\n}", "func (w *Watcher) Stop() {\n\tw.StopChannel <- true\n}", "func Stop() {\n\tstopRunning <- true\n\n}", "func (hb *heartbeat) stop() {\n\tselect {\n\tcase hb.stopChan <- struct{}{}:\n\tdefault:\n\t}\n}", "func (brw *blockRetrievalWorker) Shutdown() {\n\tselect {\n\tcase <-brw.stopCh:\n\tdefault:\n\t\tclose(brw.stopCh)\n\t}\n}", "func (w *WatchManager) Stop() {\n\tlog.Println(\"Stopping Watcher...\")\n\tclose(w.stopChannel)\n\tw.RunWaitGroup.Done()\n}", "func Stop() {\n\t// /bin/dbus-send --system --dest=org.ganesha.nfsd --type=method_call /org/ganesha/nfsd/admin org.ganesha.nfsd.admin.shutdown\n}", "func (w *Worker) Stop() {\n\tgo func() {\n\t\tw.QuitChan <- true\n\t}()\n}", "func (w *Worker) Stop() {\n\tgo func() {\n\t\tw.QuitChan <- true\n\t}()\n}", "func (w *Worker) Stop() {\n\tgo func() {\n\t\tw.QuitChan <- true\n\t}()\n}", "func (w *Worker) Stop() {\n\tgo func() {\n\t\tw.QuitChan <- true\n\t}()\n}", "func (w *Worker) Stop() {\n\tgo func() {\n\t\tw.QuitChan <- true\n\t}()\n}", "func (w *Worker) Stop() {\n\tgo func() {\n\t\tw.QuitChan <- true\n\t}()\n}", "func (d *Driver) Stop() error {\n\tif err := d.verifyRootPermissions(); err != nil {\n\t\treturn err\n\t}\n\n\ts, err := d.GetState()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif s != state.Stopped {\n\t\terr := d.sendSignal(syscall.SIGTERM)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"hyperkit sigterm failed\")\n\t\t}\n\t\t// wait 120s for graceful shutdown\n\t\tfor i := 0; i < 60; i++ {\n\t\t\ttime.Sleep(2 * time.Second)\n\t\t\ts, _ := d.GetState()\n\t\t\tlog.Debugf(\"VM state: %s\", s)\n\t\t\tif s == state.Stopped {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t\treturn errors.New(\"VM Failed to gracefully shutdown, try the kill command\")\n\t}\n\treturn nil\n}", "func (bt *BackTest) Stop() {\n\tclose(bt.shutdown)\n}", "func (u *utxoNursery) Stop() error {\n\tif !atomic.CompareAndSwapUint32(&u.stopped, 0, 1) {\n\t\treturn nil\n\t}\n\n\tutxnLog.Infof(\"UTXO nursery shutting down\")\n\n\tclose(u.quit)\n\tu.wg.Wait()\n\n\treturn nil\n}", "func (w *IndexPoller) Stop() {\n\tdefer close(w.controlChannel)\n\tw.controlChannel <- true\n}", "func Stop() {\n\tExitChannel <- true\n}", "func (w *Worker) Stop() {\n go func() {\n w.QuitChan <- true\n }()\n}", "func (s *SlaveHealthChecker) Stop() {\n\tclose(s.stop)\n}", "func (w *Watcher) Stop() {\n\tw.stopChan <- struct{}{}\n}", "func (w Worker) Stop() {\n\tgo func() {\n\t\tw.QuitChan <- true\n\t}()\n}", "func (p *ProbeHandler) Stop() {\n\tp.quit <- struct{}{}\n}", "func (management *Management) Stop() {\n\tlog.Info(management.logPrefix, \"Shutdown\")\n\tmanagement.closesOnce.Do(func() {\n\t\tclose(management.shutdownStarted)\n\t})\n\n\tmanagement.shutdownWaiter.Wait()\n\n\tlog.Info(management.logPrefix, \"Shutdown finished\")\n}", "func (lp *LongPoll) Shutdown() {\n\tif lp.cancel != nil {\n\t\tlp.cancel()\n\t}\n}", "func Stop() {\n\tclose(shutdownChannel)\n\tshutdownWaitGroup.Wait()\n\tlog4go.Info(\"Console shutdown complete\")\n}", "func (pool *SubPool) Stop() {\n\tpool.StopChan <- struct{}{}\n}", "func (pw *PurgeWorker) Stop() {\n\tpw.StopChan <- true\n}", "func (w *Worker) Stop() {\n\tgo func() {\n\t\tw.Quit <- true\n\t}()\n}", "func (b *BTCVanity) Stop() {\n\tb.stop <- true\n}", "func (b *BTCVanity) Stop() {\n\tb.stop <- true\n}", "func (w *Worker) Shutdown() {\n\tw.Stopped = true\n\tclose(w.stopChan)\n}", "func (w *worker) stop() {\n\tw.quitChan <- true\n}", "func (s *Streamer) Stop() error {\n\ts.mu.Lock()\n\tif s.state != stateRunning {\n\t\ts.mu.Unlock()\n\t\treturn ErrNotRunning\n\t}\n\ts.state = stateStopping\n\ts.mu.Unlock()\n\n\ts.fsNotify.Close() // trigger stop chain: fsNotify -> (sendChangeEvents,logNotifyErrors) -> eventsRouter\n\ts.threads.Wait()\n\n\ts.mu.Lock()\n\ts.state = stateStopped\n\ts.mu.Unlock()\n\n\treturn nil\n}", "func (c *Crawler) Stop() {\t\n\tch := make(chan bool)\n\tc.stopChan <- ch\n\n\t// Wait until it has stopped\n\t<- ch\n}", "func (sp *StreamPool) Stop() {\n\t//sw.quitCh <- true\n}", "func (h *handler) Stop(ctx context.Context) {\n\th.closeOnce.Do(func() {\n\t\th.startClosingTime = h.clock.Time()\n\n\t\t// Must hold the locks here to ensure there's no race condition in where\n\t\t// we check the value of [h.closing] after the call to [Signal].\n\t\th.syncMessageQueue.Shutdown()\n\t\th.asyncMessageQueue.Shutdown()\n\t\tclose(h.closingChan)\n\n\t\t// TODO: switch this to use a [context.Context] with a cancel function.\n\t\t//\n\t\t// Don't process any more bootstrap messages. If a dispatcher is\n\t\t// processing a bootstrap message, stop. We do this because if we\n\t\t// didn't, and the engine was in the middle of executing state\n\t\t// transitions during bootstrapping, we wouldn't be able to grab\n\t\t// [h.ctx.Lock] until the engine finished executing state transitions,\n\t\t// which may take a long time. As a result, the router would time out on\n\t\t// shutting down this chain.\n\t\tstate := h.ctx.State.Get()\n\t\tbootstrapper, ok := h.engineManager.Get(state.Type).Get(snow.Bootstrapping)\n\t\tif !ok {\n\t\t\th.ctx.Log.Error(\"bootstrapping engine doesn't exists\",\n\t\t\t\tzap.Stringer(\"type\", state.Type),\n\t\t\t)\n\t\t\treturn\n\t\t}\n\t\tbootstrapper.Halt(ctx)\n\t})\n}", "func (n *SQSNotify) Stop() {\n\tn.running = false\n\t_ = n.flushDeleteQueue0()\n}", "func (bt *Heartbeat) Stop() {\n\tclose(bt.done)\n}", "func (is *idleSweep) Stop() {\n\tif !is.started {\n\t\treturn\n\t}\n\n\tis.started = false\n\tis.ch.log.Info(\"Stopping idle connections poller.\")\n\tclose(is.stopCh)\n}", "func (sh *SignalHandler) Stop() {\n\tsh.lock.Lock()\n\tdefer sh.lock.Unlock()\n\n\tif sh.isRunning == true {\n\t\tsh.isRunning = false\n\t}\n}", "func (p *pollerAutoScaler) Stop() {\n\tp.cancel()\n\tp.wg.Wait()\n}", "func Stopping() error {\n\treturn SdNotify(\"STOPPING=1\")\n}", "func (s *T) Stop() {\n\tclose(s.stopCh)\n\ts.wg.Wait()\n}", "func (w *Worker) Stop() {\n\tgo func() {\n\t\tw.quit <- true\n\t}()\n}", "func (p *PollingListener) Shutdown() {\n\tclose(p.s)\n}", "func handleStop(s *Server, cmd interface{}, closeChan <-chan struct{}) (interface{}, error) {\n\tselect {\n\tcase s.requestProcessShutdown <- struct{}{}:\n\tdefault:\n\t}\n\treturn \"kaspad stopping.\", nil\n}", "func (s *Service) Stop() {\n\tclose(s.stopChan)\n}", "func (w *Worker) Stop() {\n\tw.QuitChan <- true\n\t// fmt.Fprintln(os.Stderr, \"Worker got stop call\")\n}", "func (p *statusUpdate) Stop() {\n\tfor _, client := range p.eventClients {\n\t\tclient.Stop()\n\t}\n\tlog.Info(\"Task status updater stopped\")\n\tfor _, listener := range p.listeners {\n\t\tlistener.Stop()\n\t}\n\tp.applier.drainAndShutdown()\n}", "func (z *Zipkin) Stop() {\n\tctx, cancel := context.WithTimeout(context.Background(), defaultShutdownTimeout)\n\n\tdefer z.waitGroup.Wait()\n\tdefer cancel()\n\n\tz.server.Shutdown(ctx) //nolint:errcheck // Ignore the returned error as we cannot do anything about it anyway\n}", "func (w *Worker) Stop() {\n\tw.quit <- true\n}", "func (f *framework) stop() {\n\tclose(f.epochChan)\n}", "func (w Worker) Stop() {\n\tgo func() {\n\t\tw.quit <- true\n\t}()\n}", "func (w Worker) Stop() {\n\tgo func() {\n\t\tw.quit <- true\n\t}()\n}", "func (w Worker) Stop() {\n\tgo func() {\n\t\tw.quit <- true\n\t}()\n}", "func (w Worker) Stop() {\n\tgo func() {\n\t\tw.quit <- true\n\t}()\n}", "func (w Worker) Stop() {\n go func() {\n w.quit <- true\n }()\n}", "func (s *Server) Stop(ctx context.Context) {\n\ts.shutdownFuncsM.Lock()\n\tdefer s.shutdownFuncsM.Unlock()\n\ts.shutdownOnce.Do(func() {\n\t\tclose(s.shuttingDown)\n\t\t// Shut down the HTTP server in parallel to calling any custom shutdown functions\n\t\twg := sync.WaitGroup{}\n\t\twg.Add(1)\n\t\tgo func() {\n\t\t\tdefer wg.Done()\n\t\t\tif err := s.srv.Shutdown(ctx); err != nil {\n\t\t\t\tslog.Debug(ctx, \"Graceful shutdown failed; forcibly closing connections 👢\")\n\t\t\t\tif err := s.srv.Close(); err != nil {\n\t\t\t\t\tslog.Critical(ctx, \"Forceful shutdown failed, exiting 😱: %v\", err)\n\t\t\t\t\tpanic(err) // Something is super hosed here\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\t\tfor _, f := range s.shutdownFuncs {\n\t\t\tf := f // capture range variable\n\t\t\twg.Add(1)\n\t\t\tgo func() {\n\t\t\t\tdefer wg.Done()\n\t\t\t\tf(ctx)\n\t\t\t}()\n\t\t}\n\t\twg.Wait()\n\t})\n}", "func (m *MaintenanceScheduler) Stop() {\n\tm.schan <- true\n}", "func (s *SimpleServer) Stop() {\n\tif s != nil {\n\t\ts.shutdownReq <- true\n\t}\n}", "func (c *Concentrator) Stop() {\n\tclose(c.exit)\n\tc.exitWG.Wait()\n}", "func (w *Worker) Stop() {\n\tw.os.Do(func() {\n\t\tw.l.Info(\"astikit: stopping worker...\")\n\t\tw.cancel()\n\t\tw.wg.Done()\n\t})\n}", "func (eis *eventSocket) stop() error {\n\teis.log.Info(\"closing Chain IPC\")\n\terrs := wrappers.Errs{}\n\terrs.Add(eis.unregisterFn(), eis.socket.Close())\n\treturn errs.Err\n}", "func (k *Kinsumer) Stop() {\n\tk.stoprequest <- true\n\tk.mainWG.Wait()\n}", "func (api *API) Stop() error {\n\n\t// context: wait for 3 seconds\n\tctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)\n\tdefer cancel()\n\n\terr := api.srv.Shutdown(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (r *ring) Stop() {\n\tif !atomic.CompareAndSwapInt32(\n\t\t&r.status,\n\t\tcommon.DaemonStatusStarted,\n\t\tcommon.DaemonStatusStopped,\n\t) {\n\t\treturn\n\t}\n\n\tr.peerProvider.Stop()\n\tr.value.Store(emptyHashring())\n\n\tr.subscribers.Lock()\n\tdefer r.subscribers.Unlock()\n\tr.subscribers.keys = make(map[string]chan<- *ChangedEvent)\n\tclose(r.shutdownCh)\n\n\tif success := common.AwaitWaitGroup(&r.shutdownWG, time.Minute); !success {\n\t\tr.logger.Warn(\"service resolver timed out on shutdown.\")\n\t}\n}", "func (f5 *BigIP) Stop() error {\n\tf5.shutdown <- true\n\n\treturn nil\n}", "func (s *server) Stop() error {\n\t// Make sure this only happens once.\n\tif atomic.AddInt32(&s.shutdown, 1) != 1 {\n\t\tlogging.CPrint(logging.INFO, \"server is already in the process of shutting down\", logging.LogFormat{})\n\t\treturn nil\n\t}\n\n\ts.syncManager.Stop()\n\n\t// Signal the remaining goroutines to quit.\n\tclose(s.quit)\n\n\ts.wg.Done()\n\n\treturn nil\n}", "func (dw *DirWatcher) Stop() {\n\tdw.qrun <- false\n}", "func (w *Worker) Stop() {\n\tw.wQuit <- true\n}", "func (tkr *RandTicker) Stop() {\n\tclose(tkr.done)\n}", "func (hmr *receiver) Shutdown(ctx context.Context) error {\n\tclose(hmr.done)\n\treturn hmr.closeScrapers(ctx)\n}", "func (t *Ticker) Stop() {\n\tselect {\n\tcase t.stop <- struct{}{}:\n\tdefault:\n\t}\n}", "func (s *SignalMonitor) Stop() {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\n\tif s.isOn {\n\t\ts.isOn = false\n\t\ts.offc <- struct{}{}\n\t}\n}", "func (d *loadReporterDaemonImpl) Stop() {\n\tif !atomic.CompareAndSwapInt32(&d.started, 1, 0) {\n\t\treturn\n\t}\n\n\tclose(d.shutdownCh)\n\tif success := AwaitWaitGroup(&d.shutdownWG, time.Minute); !success {\n\t\td.logger.Warn(\"Timed out waiting to shutdown load reporter.\")\n\t}\n\n\td.logger.Info(\"Load reporter stopped.\")\n}", "func (m *ManagerImpl) Stop() error {\n\tif !m.lifecycle.BeginShutdown() {\n\t\treturn ErrNotRunning\n\t}\n\tm.statusTracker.NotifySSEShutdownExpected()\n\tm.withRefreshTokenLock(func() {\n\t\tif m.nextRefresh != nil {\n\t\t\tm.nextRefresh.Stop()\n\t\t}\n\t})\n\tm.StopWorkers()\n\tm.sseClient.StopStreaming()\n\tm.lifecycle.AwaitShutdownComplete()\n\treturn nil\n}", "func (c *ZKCluster) Stop() {\n\tif c.checkerdone != nil {\n\t\tclose(c.checkerdone)\n\t\tclose(c.updates)\n\t\tc.checkerdone = nil\n\t}\n}", "func (r *RoverDriver) Stop() {\n r.commands <- stop\n}", "func (a API) Stop(ctx context.Context) error {\n\treturn a.srv.Shutdown(ctx)\n}", "func (h *Harness) Stop() {\n\tfin := make(chan struct{})\n\tgo func() {\n\t\tdefer close(fin)\n\t\tif err := h.Client.Close(); err != nil {\n\t\t\th.t.Fatal(err)\n\t\t}\n\n\t\tif err := h.cmd.Process.Kill(); err != nil {\n\t\t\th.t.Fatal(err)\n\t\t}\n\n\t\tif err := os.Remove(h.configPath); err != nil {\n\t\t\th.t.Fatal(err)\n\t\t}\n\t}()\n\tselect {\n\tcase <-fin:\n\tcase <-time.After(h.StopTimeout):\n\t}\n}", "func (n *DcrdNotifier) Stop() error {\n\t// Already shutting down?\n\tif atomic.AddInt32(&n.stopped, 1) != 1 {\n\t\treturn nil\n\t}\n\n\t// Shutdown the rpc client, this gracefully disconnects from dcrd, and\n\t// cleans up all related resources.\n\tn.chainConn.Shutdown()\n\n\tclose(n.quit)\n\tn.wg.Wait()\n\n\tn.chainUpdates.Stop()\n\n\t// Notify all pending clients of our shutdown by closing the related\n\t// notification channels.\n\tfor _, epochClient := range n.blockEpochClients {\n\t\tclose(epochClient.cancelChan)\n\t\tepochClient.wg.Wait()\n\n\t\tclose(epochClient.epochChan)\n\t}\n\tn.txNotifier.TearDown()\n\n\treturn nil\n}", "func (e *Engine) Stop() {\n\tif atomic.CompareAndSwapInt32(&e.stopping, 0, 1) {\n\t\te.wg.Wait()\n\t\te.running = 0\n\t\te.stopping = 0\n\t}\n}", "func (lt *Logtailer) Stop() {\n\tclose(lt.shutdown)\n}", "func (srv *Server) StopNotify() <-chan struct{} {\n return srv.stopc\n}", "func (b *Backend) Stop() error {\n\tb.eventSock.Close()\n\tb.commandSock.Close()\n\n\tb.eventSockCancel()\n\tb.commandSockCancel()\n\n\treturn nil\n}", "func (t *Ticker) Stop() {\n\tt.Done <- struct{}{}\n}", "func (n *Node) stop() (err error) {\n\tif err = n.stateCheck(nodeRunning, nodeHealthChecking); err != nil {\n\t\treturn\n\t}\n\tn.setState(nodeShuttingDown)\n\tn.stopChan <- true\n\tn.expireTicker.Stop()\n\tclose(n.stopChan)\n\tlogDebug(\"[Node]\", \"(%v) shutting down.\", n)\n\tn.shutdown()\n\treturn\n}", "func (r *runtime) Stop() {\n\tr.logger.Info(\"stopping broker server...\")\n\tdefer r.cancel()\n\n\tr.Shutdown()\n\n\tif r.httpServer != nil {\n\t\tr.logger.Info(\"stopping http server...\")\n\t\tif err := r.httpServer.Close(r.ctx); err != nil {\n\t\t\tr.logger.Error(\"shutdown http server error\", logger.Error(err))\n\t\t} else {\n\t\t\tr.logger.Info(\"stopped http server successfully\")\n\t\t}\n\t}\n\n\t// close registry, deregister broker node from active list\n\tif r.registry != nil {\n\t\tr.logger.Info(\"closing discovery-registry...\")\n\t\tif err := r.registry.Deregister(r.node); err != nil {\n\t\t\tr.logger.Error(\"unregister broker node error\", logger.Error(err))\n\t\t}\n\t\tif err := r.registry.Close(); err != nil {\n\t\t\tr.logger.Error(\"unregister broker node error\", logger.Error(err))\n\t\t} else {\n\t\t\tr.logger.Info(\"closed discovery-registry successfully\")\n\t\t}\n\t}\n\n\tif r.master != nil {\n\t\tr.logger.Info(\"stopping master...\")\n\t\tr.master.Stop()\n\t}\n\n\tif r.stateMachineFactory != nil {\n\t\tr.stateMachineFactory.Stop()\n\t}\n\n\tif r.repo != nil {\n\t\tr.logger.Info(\"closing state repo...\")\n\t\tif err := r.repo.Close(); err != nil {\n\t\t\tr.logger.Error(\"close state repo error, when broker stop\", logger.Error(err))\n\t\t} else {\n\t\t\tr.logger.Info(\"closed state repo successfully\")\n\t\t}\n\t}\n\tif r.stateMgr != nil {\n\t\tr.stateMgr.Close()\n\t}\n\tif r.srv.channelManager != nil {\n\t\tr.logger.Info(\"closing write channel manager...\")\n\t\tr.srv.channelManager.Close()\n\t\tr.logger.Info(\"closed write channel successfully\")\n\t}\n\n\tif r.factory.connectionMgr != nil {\n\t\tif err := r.factory.connectionMgr.Close(); err != nil {\n\t\t\tr.logger.Error(\"close connection manager error, when broker stop\", logger.Error(err))\n\t\t} else {\n\t\t\tr.logger.Info(\"closed connection manager successfully\")\n\t\t}\n\t}\n\tr.logger.Info(\"close connections successfully\")\n\n\t// finally, shutdown rpc server\n\tif r.grpcServer != nil {\n\t\tr.logger.Info(\"stopping grpc server...\")\n\t\tr.grpcServer.Stop()\n\t\tr.logger.Info(\"stopped grpc server successfully\")\n\t}\n\n\tr.state = server.Terminated\n\tr.logger.Info(\"stopped broker server successfully\")\n}", "func (f *HTTPFeeder) Stop(stopChan chan bool) {\n\tif f.IsRunning {\n\t\tf.Server.Shutdown(context.TODO())\n\t}\n\tclose(stopChan)\n}", "func (w *wsEvents) Stop() error { return w.ws.Stop() }", "func (t *TCPTest) Stop() {\n\tt.exit<-struct{}{}\n}", "func (wsServer *WsServer) Stop() {\n\tif atomic.AddInt32(&wsServer.shutdown, 1) != 1 {\n\t\tLogger.log.Info(\"RPC server is already in the process of shutting down\")\n\t}\n\tLogger.log.Info(\"RPC server shutting down\")\n\tif wsServer.started != 0 {\n\t\twsServer.server.Close()\n\t}\n\tfor _, listen := range wsServer.config.HttpListenters {\n\t\tlisten.Close()\n\t}\n\tLogger.log.Warn(\"RPC server shutdown complete\")\n\twsServer.started = 0\n\twsServer.shutdown = 1\n}", "func (a *appsec) stop() {\n\ta.unregisterWAF()\n\ta.limiter.Stop()\n}", "func (w *worker) stop() {\n\tselect {\n\tcase w.stopCh <- struct{}{}:\n\tdefault: // Non-blocking.\n\t}\n}" ]
[ "0.7100459", "0.70205927", "0.69549215", "0.69389325", "0.69152033", "0.6909807", "0.6899172", "0.68864995", "0.6879378", "0.6826131", "0.6764922", "0.6750816", "0.67412126", "0.67412126", "0.67412126", "0.67412126", "0.67412126", "0.67412126", "0.6737912", "0.672382", "0.6704135", "0.6684673", "0.6676277", "0.6670724", "0.66676867", "0.6665554", "0.6663922", "0.66551757", "0.6623341", "0.66232324", "0.66222405", "0.66117454", "0.66031784", "0.66025573", "0.65995336", "0.65995336", "0.6594916", "0.6589645", "0.6582771", "0.6579329", "0.65666395", "0.65598315", "0.6559553", "0.6556075", "0.65534836", "0.65476596", "0.65435827", "0.65416014", "0.653483", "0.6531247", "0.6527875", "0.651898", "0.65171874", "0.6515303", "0.65085745", "0.65021634", "0.64956844", "0.64866644", "0.6480522", "0.6480522", "0.6480522", "0.6480522", "0.64761156", "0.64678836", "0.6464896", "0.6456774", "0.64556086", "0.6453572", "0.64515907", "0.64478856", "0.6447824", "0.6443389", "0.64420176", "0.6441434", "0.64399296", "0.6438904", "0.6436266", "0.643566", "0.64352477", "0.64349556", "0.642367", "0.64182574", "0.64153856", "0.64113283", "0.64104265", "0.64090365", "0.64053553", "0.64036644", "0.6401902", "0.6395193", "0.6395024", "0.6394285", "0.6393274", "0.63908595", "0.6384889", "0.6380325", "0.6372703", "0.6371762", "0.6371052", "0.6369634" ]
0.78706795
0
GetAndDelete retrieves a PendingFileshare from the repository and then deletes it.
func (r *inMemoryFileshareRepository) GetAndDelete(key string) (PendingFileshare, bool) { r.Lock() defer r.Unlock() fileshare, ok := r.pendingFileshares[key] if !ok { return PendingFileshare{}, false } delete(r.pendingFileshares, key) return fileshare, true }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (am *AutogitManager) Delete(\n\tctx context.Context, dstTLF *libkbfs.TlfHandle, dstDir string,\n\trepo, branchName string) (doneCh <-chan struct{}, err error) {\n\tam.log.CDebugf(ctx, \"Autogit delete request for %s/%s:%s\",\n\t\tdstTLF.GetCanonicalPath(), dstDir, repo, branchName)\n\tdefer func() {\n\t\tam.deferLog.CDebugf(ctx, \"Delete request processed: %+v\", err)\n\t}()\n\n\treq := deleteReq{\n\t\tdstTLF, dstDir, repo, branchName, make(chan struct{}),\n\t}\n\n\tselect {\n\tcase am.deleteQueue.In() <- req:\n\tcase <-ctx.Done():\n\t\treturn nil, ctx.Err()\n\t}\n\treturn req.doneCh, nil\n}", "func (s *SharemeService) Delete(c *gae.Context, session *Session, key string) (ret Share) {\n\tstat := s.Stat(c, key)\n\tif stat.IsError() {\n\t\treturn stat\n\t}\n\n\terr := stat.Delete(c)\n\tif err != nil {\n\t\treturn *noShare(err)\n\t}\n\n\tc.HTTPContext.ReturnStatus = http.StatusOK\n\tsession.Delete(fmt.Sprintf(\"%s%s\", KeySessionPrefix, key))\n\tret = Share{Key: ret.Key, Expires: 0, Size: 0}\n\treturn\n}", "func (s *shares) Delete(shareID int) error {\n\t_, err := s.c.baseRequest(http.MethodDelete, routes.shares, nil, strconv.Itoa(shareID))\n\treturn err\n}", "func (vrfs *VRFShare) Delete(ctx context.Context) error {\n\treturn vrfs.GetEntityMetadata().GetStore().Delete(ctx, vrfs)\n}", "func RelationshipPendingDelete(w http.ResponseWriter, r *http.Request) {\n\taccessToken := auth.ParseApiKey(r, accessTokenKey, true)\n\tif !checkToken(accessToken, w) {\n\t\treturn\n\t}\n\n\tk := r.URL.Query().Get(\"key\")\n\tif k == \"\" {\n\t\tapi.Api.BuildMissingParameter(w)\n\t\treturn\n\t}\n\n\tout, err := managers.RelationshipPendingManagerDelete(k)\n\tif err != nil {\n\t\tlogger.Error(err.Error())\n\t\tapi.Api.BuildErrorResponse(http.StatusInternalServerError, \"failed to delete rel\", w)\n\t\treturn\n\t}\n\n\tapi.Api.BuildJsonResponse(true, \"rel deleted\", out, w)\n}", "func (fs *Stow) Delete(ctx context.Context, id string) (string, interface{}, error) {\n\tdg, close, err := dgraph.NewClient(ctx)\n\tif err != nil {\n\t\treturn \"\", nil, err\n\t}\n\tdefer close()\n\n\ttx := dg.NewTxn()\n\tdefer dgraph.Discard(ctx, tx)\n\n\tfile, err := FindFileImpl(ctx, tx, id)\n\tif err != nil {\n\t\treturn \"\", nil, err\n\t}\n\n\tif file == nil {\n\t\treturn \"\", nil, fmt.Errorf(\"file not found: %s\", id)\n\t}\n\n\tlocation, err := stow.Dial(fs.kind, fs.config)\n\tif err != nil {\n\t\tlog.Errorf(\"stow.Dial fail: %v\", err)\n\t\treturn \"\", nil, err\n\t}\n\tdefer location.Close()\n\n\tcontainer, err := location.Container(fs.bucket)\n\tif err != nil {\n\t\tlog.Errorf(\"stow.GetContainer fail: %v\", err)\n\t\treturn \"\", nil, err\n\t}\n\n\terr = container.RemoveItem(file.Path)\n\tif err != nil {\n\t\tlog.Errorf(\"stow.Container.RemoveItem fail: %v\", err)\n\t\treturn id, nil, err\n\t}\n\n\tresp2, err := dgraph.DeleteNode(ctx, tx, file.ID)\n\tif err != nil {\n\t\treturn file.ID, nil, err\n\t}\n\n\treturn file.ID, resp2, nil\n}", "func (t *SimpleChaincode) delete(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\tfmt.Println(\"- start delete transfer\")\n\n\ttype transferDeleteTransientInput struct {\n\t\tName string `json:\"name\"`\n\t}\n\n\tif len(args) != 0 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Private transfer name must be passed in transient map.\")\n\t}\n\n\ttransMap, err := stub.GetTransient()\n\tif err != nil {\n\t\treturn shim.Error(\"Error getting transient: \" + err.Error())\n\t}\n\n\tif _, ok := transMap[\"transfer_delete\"]; !ok {\n\t\treturn shim.Error(\"transfer_delete must be a key in the transient map\")\n\t}\n\n\tif len(transMap[\"transfer_delete\"]) == 0 {\n\t\treturn shim.Error(\"transfer_delete value in the transient map must be a non-empty JSON string\")\n\t}\n\n\tvar transferDeleteInput transferDeleteTransientInput\n\terr = json.Unmarshal(transMap[\"transfer_delete\"], &transferDeleteInput)\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to decode JSON of: \" + string(transMap[\"transfer_delete\"]))\n\t}\n\n\tif len(transferDeleteInput.Name) == 0 {\n\t\treturn shim.Error(\"name field must be a non-empty string\")\n\t}\n\n\t// to maintain the authorization~name index, we need to read the transfer first and get its authorization\n\tvalAsbytes, err := stub.GetPrivateData(\"collectionFileTransfer\", transferDeleteInput.Name) //get the marble from chaincode state\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to get state for \" + transferDeleteInput.Name)\n\t} else if valAsbytes == nil {\n\t\treturn shim.Error(\"Transfer does not exist: \" + transferDeleteInput.Name)\n\t}\n\n\tvar transferToDelete fileTransfer\n\terr = json.Unmarshal([]byte(valAsbytes), &transferToDelete)\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to decode JSON of: \" + string(valAsbytes))\n\t}\n\n\t// delete the transfer from state\n\terr = stub.DelPrivateData(\"collectionFileTransfer\", transferDeleteInput.Name)\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to delete state:\" + err.Error())\n\t}\n\n\t// Also delete the transfer from the authorization~name index\n\tindexName := \"authorization~name\"\n\tauthorizationNameIndexKey, err := stub.CreateCompositeKey(indexName, []string{transferToDelete.Authorization, transferToDelete.Name})\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\terr = stub.DelPrivateData(\"collectionFileTransfer\", authorizationNameIndexKey)\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to delete state:\" + err.Error())\n\t}\n\n\t// Finally, delete private details of transfer\n\terr = stub.DelPrivateData(\"collectionFileTransferPrivateDetails\", transferDeleteInput.Name)\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\treturn shim.Success(nil)\n}", "func (g *GistFile) Delete(id string) (*http.Response, error) {\n\turll := fmt.Sprintf(\"/gists/%s\", id)\n\treq, err := http.NewRequest(http.MethodDelete, urll, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp, err := auth.Session.Client.Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func (a *GoogleAuth) GetAndDelete(state string) bool {\n\ta.mu.Lock()\n\tdefer a.mu.Unlock()\n\n\t_, ok := a.states[state]\n\tif ok {\n\t\tdelete(a.states, state)\n\t\treturn true\n\t}\n\treturn false\n}", "func (s *gcBlobTaskStore) Delete(ctx context.Context, b *models.GCBlobTask) error {\n\tdefer metrics.InstrumentQuery(\"gc_blob_task_delete\")()\n\n\tq := \"DELETE FROM gc_blob_review_queue WHERE digest = decode($1, 'hex')\"\n\tdgst, err := NewDigest(b.Digest)\n\tif err != nil {\n\t\treturn err\n\t}\n\tres, err := s.db.ExecContext(ctx, q, dgst)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"deleting GC blob task: %w\", err)\n\t}\n\tcount, err := res.RowsAffected()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"deleting GC blob task: %w\", err)\n\t}\n\tif count == 0 {\n\t\treturn fmt.Errorf(\"GC blob task not found\")\n\t}\n\n\treturn nil\n}", "func (h Handler) makeDelete(v Injection) http.HandlerFunc {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tparams := mux.Vars(r)\n\t\tmtx.Lock()\n\t\t_, deleted := v.Share[params[\"id\"]]\n\t\tdelete(v.Share, params[\"id\"])\n\t\tmtx.Unlock()\n\t\tjson.NewEncoder(w).Encode(deleted)\n\t})\n}", "func (is *ObjectStorage) DeleteBlob(repo string, digest godigest.Digest) error {\n\tvar lockLatency time.Time\n\n\tif err := digest.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tblobPath := is.BlobPath(repo, digest)\n\n\tis.Lock(&lockLatency)\n\tdefer is.Unlock(&lockLatency)\n\n\t_, err := is.store.Stat(context.Background(), blobPath)\n\tif err != nil {\n\t\tis.log.Error().Err(err).Str(\"blob\", blobPath).Msg(\"failed to stat blob\")\n\n\t\treturn zerr.ErrBlobNotFound\n\t}\n\n\t// first check if this blob is not currently in use\n\tif ok, _ := common.IsBlobReferenced(is, repo, digest, is.log); ok {\n\t\treturn zerr.ErrBlobReferenced\n\t}\n\n\tif fmt.Sprintf(\"%v\", is.cache) != fmt.Sprintf(\"%v\", nil) {\n\t\tdstRecord, err := is.cache.GetBlob(digest)\n\t\tif err != nil && !errors.Is(err, zerr.ErrCacheMiss) {\n\t\t\tis.log.Error().Err(err).Str(\"blobPath\", dstRecord).Msg(\"dedupe: unable to lookup blob record\")\n\n\t\t\treturn err\n\t\t}\n\n\t\t// remove cache entry and move blob contents to the next candidate if there is any\n\t\tif ok := is.cache.HasBlob(digest, blobPath); ok {\n\t\t\tif err := is.cache.DeleteBlob(digest, blobPath); err != nil {\n\t\t\t\tis.log.Error().Err(err).Str(\"digest\", digest.String()).Str(\"blobPath\", blobPath).\n\t\t\t\t\tMsg(\"unable to remove blob path from cache\")\n\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t\t// if the deleted blob is one with content\n\t\tif dstRecord == blobPath {\n\t\t\t// get next candidate\n\t\t\tdstRecord, err := is.cache.GetBlob(digest)\n\t\t\tif err != nil && !errors.Is(err, zerr.ErrCacheMiss) {\n\t\t\t\tis.log.Error().Err(err).Str(\"blobPath\", dstRecord).Msg(\"dedupe: unable to lookup blob record\")\n\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// if we have a new candidate move the blob content to it\n\t\t\tif dstRecord != \"\" {\n\t\t\t\tif err := is.store.Move(context.Background(), blobPath, dstRecord); err != nil {\n\t\t\t\t\tis.log.Error().Err(err).Str(\"blobPath\", blobPath).Msg(\"unable to remove blob path\")\n\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t}\n\n\tif err := is.store.Delete(context.Background(), blobPath); err != nil {\n\t\tis.log.Error().Err(err).Str(\"blobPath\", blobPath).Msg(\"unable to remove blob path\")\n\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (gc *GalleryContext) DeleteAndConfirm() uiauto.Action {\n\tdeleteButtonFinder := nodewith.Role(role.Button).Name(\"Delete\").Ancestor(RootFinder)\n\tconfirmButtonFinder := nodewith.Role(role.Button).Name(\"Delete\").Ancestor(DialogFinder)\n\treturn uiauto.Combine(\"remove current opened media file\",\n\t\tgc.ui.WithTimeout(30*time.Second).WithInterval(1*time.Second).LeftClickUntil(\n\t\t\tdeleteButtonFinder, gc.ui.WithTimeout(3*time.Second).WaitUntilExists(confirmButtonFinder)),\n\t\tgc.ui.LeftClick(confirmButtonFinder),\n\t)\n}", "func (m *Manager) Get(id string) *Transfer {\n\tm.mu.RLock()\n\tdefer m.mu.RUnlock()\n\n\treturn m.transfers[id]\n}", "func TestDelete(t *testing.T) {\n\tth.SetupHTTP()\n\tdefer th.TeardownHTTP()\n\n\tMockDeleteResponse(t)\n\tres := sharetypes.Delete(client.ServiceClient(), \"shareTypeID\")\n\tth.AssertNoErr(t, res.Err)\n}", "func (ss *redisStore) GetAndDelete(key string) (*storage.Secret, error) {\n\tbb, err := ss.rdb.GetDel(ctx, key).Bytes()\n\tif err == redis.Nil {\n\t\treturn nil, storage.ErrNoRecord\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar s storage.Secret\n\terr = json.Unmarshal(bb, &s)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &s, nil\n}", "func (s RPCService) Delete(ctx context.Context, req *IdRequest) (*Void, error) {\n\treturn &Void{}, s.storage.Delete(req.Id)\n}", "func (r *Remoter) Get(waitCtx context.Context, remote string, local string) error {\n\tclt, err := sftp.NewClient(r.clt)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_ = os.Remove(local)\n\trf, err := clt.Open(remote)\n\tif err != nil {\n\t\t_ = clt.Close()\n\t\treturn err\n\t}\n\twf, err := os.Create(local)\n\tif err != nil {\n\t\t_ = clt.Close()\n\t\t_ = rf.Close()\n\t\treturn err\n\t}\n\tnoWait, waitGrp := r.wait(waitCtx)\n\t_, err = io.Copy(wf, rf)\n\tclose(noWait)\n\twaitGrp.Wait()\n\t_ = wf.Close()\n\t_ = rf.Close()\n\t_ = clt.Close()\n\treturn err\n}", "func (s *SharemeService) Get(c *gae.Context, key string) interface{} { // Return either Share in case of Error or Binary.\n\tst := s.Stat(c, key)\n\tif st.IsError() {\n\t\treturn noShare(fmt.Errorf(st.Error))\n\t}\n\n\t/* retrieve blob */\n\trc, err := s.storageService.Get(c, st.StorageKey)\n\tif err != nil {\n\t\tc.HTTPContext.ReturnStatus = http.StatusNotFound\n\t\t//TODO: delete key from datastore and session.\n\t\treturn noShare(err)\n\t}\n\t/* generate return object */\n\treturn &BlobBinary{Reader: rc, mimeType: st.MimeType}\n}", "func (rc *RequiredCapability) Delete() (error, error, int) {\n\tauthorized, err := rc.isTenantAuthorized()\n\tif !authorized {\n\t\treturn errors.New(\"not authorized on this tenant\"), nil, http.StatusForbidden\n\t} else if err != nil {\n\t\treturn nil, fmt.Errorf(\"checking authorization for existing DS ID: %s\" + err.Error()), http.StatusInternalServerError\n\t}\n\t_, cdnName, _, err := dbhelpers.GetDSNameAndCDNFromID(rc.ReqInfo.Tx.Tx, *rc.DeliveryServiceID)\n\tif err != nil {\n\t\treturn nil, err, http.StatusInternalServerError\n\t}\n\tuserErr, sysErr, errCode := dbhelpers.CheckIfCurrentUserCanModifyCDN(rc.ReqInfo.Tx.Tx, string(cdnName), rc.ReqInfo.User.UserName)\n\tif userErr != nil || sysErr != nil {\n\t\treturn userErr, sysErr, errCode\n\t}\n\treturn api.GenericDelete(rc)\n}", "func (l *Locker) LoadAndDelete(key Flags) (interface{}, bool) {\n\treturn l.data.LoadAndDelete(key)\n}", "func (storage *B2Storage) DeleteFile(threadIndex int, filePath string) (err error) {\n\n if strings.HasSuffix(filePath, \".fsl\") {\n filePath = filePath[:len(filePath) - len(\".fsl\")]\n entries, err := storage.clients[threadIndex].ListFileNames(filePath, true, true)\n if err != nil {\n return err\n }\n\n toBeDeleted := false\n\n for _, entry := range entries {\n if entry.FileName != filePath || (!toBeDeleted && entry.Action != \"hide\" ) {\n continue\n }\n\n toBeDeleted = true\n\n err = storage.clients[threadIndex].DeleteFile(filePath, entry.FileID)\n if err != nil {\n return err\n }\n }\n\n return nil\n\n } else {\n entries, err := storage.clients[threadIndex].ListFileNames(filePath, true, false)\n if err != nil {\n return err\n }\n\n if len(entries) == 0 {\n return nil\n }\n return storage.clients[threadIndex].DeleteFile(filePath, entries[0].FileID)\n }\n}", "func (api *bucketAPI) SyncDelete(obj *objstore.Bucket) error {\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, writeErr = apicl.ObjstoreV1().Bucket().Delete(context.Background(), &obj.ObjectMeta)\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleBucketEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\t}\n\n\treturn writeErr\n}", "func (repo Repository) Delete(fullPath string) error {\n\tfilePath := path.Join(repo.StorageDir, fullPath)\n\n\treturn os.Remove(filePath)\n}", "func NewDeleteSmbShareDefault(code int) *DeleteSmbShareDefault {\n\treturn &DeleteSmbShareDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (svc *SSHKeysService) Delete(ctx context.Context, prj, id string) (*http.Response, error) {\n\tpath := deleteSSHKeyPath(prj, id)\n\treturn svc.client.resourceDelete(ctx, path)\n}", "func (s *Store) Delete(id packet.ID) {\n\t// acquire mutex\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\n\t// delete future\n\tdelete(s.store, id)\n}", "func (client *GCSBlobstore) Delete(dest string) error {\n\tif client.readOnly() {\n\t\treturn ErrInvalidROWriteOperation\n\t}\n\n\terr := client.getObjectHandle(client.authenticatedGCS, dest).Delete(context.Background())\n\tif err == storage.ErrObjectNotExist {\n\t\treturn nil\n\t}\n\treturn err\n}", "func (pg *PGClient) DeletePendingTransfer(id int64) error {\n\ttx, err := pg.DB.Beginx()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer func() {\n\t\tswitch err {\n\t\tcase nil:\n\t\t\terr = tx.Commit()\n\t\tdefault:\n\t\t\ttx.Rollback()\n\t\t}\n\t}()\n\n\tcheck := &model.PendingTransfer{}\n\tif err = tx.Get(check, qPendingTransferBy + `id = $1`, id); err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\treturn ErrPendingTransferNotFound\n\t\t}\n\t\treturn err\n\t}\n\n\tresult, err := tx.Exec(qDeletePendingTransfer, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\trows, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif rows != 1 {\n\t\treturn ErrPendingTransferNotFound\n\t}\n\treturn err\n}", "func (api *snapshotrestoreAPI) SyncDelete(obj *cluster.SnapshotRestore) error {\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, writeErr = apicl.ClusterV1().SnapshotRestore().Delete(context.Background(), &obj.ObjectMeta)\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleSnapshotRestoreEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\t}\n\n\treturn writeErr\n}", "func (api *dscprofileAPI) SyncDelete(obj *cluster.DSCProfile) error {\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, writeErr = apicl.ClusterV1().DSCProfile().Delete(context.Background(), &obj.ObjectMeta)\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleDSCProfileEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\t}\n\n\treturn writeErr\n}", "func (c *Client) Delete(d core.Digest) error {\n\t_, err := httputil.Delete(fmt.Sprintf(\"http://%s/blobs/%s\", c.addr, d))\n\treturn err\n}", "func (s *GDrive) Delete(ctx context.Context, token string, filename string) (err error) {\n\tmetadata, _ := s.findID(fmt.Sprintf(\"%s.metadata\", filename), token)\n\t_ = s.service.Files.Delete(metadata).Do()\n\n\tvar fileID string\n\tfileID, err = s.findID(filename, token)\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = s.service.Files.Delete(fileID).Context(ctx).Do()\n\treturn\n}", "func (f FileRepo) Delete(context context.Context, id string) (string, error) {\n\topts := options.FindOneAndDelete().SetProjection(bson.D{{\"_id\", 1}})\n\tobjID, err := primitive.ObjectIDFromHex(id)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tquery := bson.M{\n\t\t\"_id\": objID,\n\t}\n\tvar delFile model.File\n\terr = f.collection.FindOneAndDelete(context, query, opts).Decode(&delFile)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn delFile.ID.Hex(), nil\n}", "func (api *distributedservicecardAPI) SyncDelete(obj *cluster.DistributedServiceCard) error {\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, writeErr = apicl.ClusterV1().DistributedServiceCard().Delete(context.Background(), &obj.ObjectMeta)\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleDistributedServiceCardEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\t}\n\n\treturn writeErr\n}", "func ExampleFileSharesClient_Delete() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclient, err := armstorage.NewFileSharesClient(\"{subscription-id}\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\t_, err = client.Delete(ctx,\n\t\t\"res4079\",\n\t\t\"sto4506\",\n\t\t\"share9689\",\n\t\t&armstorage.FileSharesClientDeleteOptions{XMSSnapshot: nil,\n\t\t\tInclude: nil,\n\t\t})\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n}", "func (e *etcdStore) GetAndDelete(ctx context.Context, codec store.Codec) error {\n\tvar (\n\t\tkey = codec.Key()\n\t\tresp *clientv3.DeleteResponse\n\t\terr error\n\t)\n\tif key == \"\" {\n\t\treturn errKeyIsBlank\n\t}\n\tif resp, err = e.cli.Delete(ctx, key, clientv3.WithPrevKV()); err != nil {\n\t\treturn err\n\t}\n\tif len(resp.PrevKvs) == 0 {\n\t\treturn store.ErrKVNotExists\n\t}\n\tcodec.SetVersion(0)\n\treturn codec.Decode(string(resp.PrevKvs[0].Value))\n}", "func (s *State) DeletePending(peerID core.PeerID, h core.InfoHash) {\n\tif s.get(h, peerID).status != _pending {\n\t\treturn\n\t}\n\ts.delete(h, peerID)\n\ts.log(\"hash\", h, \"peer\", peerID).Infof(\n\t\t\"Deleted pending conn, capacity now at %d\", s.capacity(h))\n}", "func Delete(client *gophercloud.ServiceClient, id string, bearer map[string]string) (r volumes.DeleteResult) {\n\t_, r.Err = client.Delete(deleteURL(client, id), &gophercloud.RequestOpts{\n\t\tMoreHeaders: bearer,\n\t})\n\treturn\n}", "func (c *BinaryFileClient) Delete() *BinaryFileDelete {\n\tmutation := newBinaryFileMutation(c.config, OpDelete)\n\treturn &BinaryFileDelete{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (a *UtilsApiService) DeleteFeatureTrackUsingDelete(ctx context.Context, featureTrackId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\ta.client = NewAPIClient(&Configuration{\n\t\tBasePath: ctx.Value(\"BasePath\").(string),\n\t\tDefaultHeader: make(map[string]string),\n\t\tUserAgent: \"Swagger-Codegen/1.0.0/go\",\n\t})\n\tlocalVarPath := a.client.cfg.BasePath + \"/nucleus/v1/feature_track/{feature_track_id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"feature_track_id\"+\"}\", fmt.Sprintf(\"%v\", featureTrackId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"*/*\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func UnshareFile(w http.ResponseWriter, r *http.Request) {\n\t//log\n\tnow, userIP := globalPkg.SetLogObj(r)\n\tlogobj := logpkg.LogStruct{\"_\", now, userIP, \"macAdress\", \"UnshareFile\", \"file\", \"_\", \"_\", \"_\", 0}\n\tvar requestObj RetrieveBody\n\tdecoder := json.NewDecoder(r.Body)\n\tdecoder.DisallowUnknownFields()\n\terr := decoder.Decode(&requestObj)\n\tif err != nil {\n\t\tglobalPkg.SendError(w, \"please enter your correct request\")\n\t\tglobalPkg.WriteLog(logobj, \"please enter your correct request\", \"failed\")\n\t\treturn\n\t}\n\ttime.Sleep(time.Millisecond * 10)\n\taccountObj := account.GetAccountByAccountPubicKey(requestObj.Publickey)\n\tif accountObj.AccountPublicKey != requestObj.Publickey {\n\t\tglobalPkg.SendError(w, \"error in public key\")\n\t\tglobalPkg.WriteLog(logobj, \"error in public key\", \"failed\")\n\t\treturn\n\t}\n\tif accountObj.AccountPassword != requestObj.Password {\n\t\tglobalPkg.SendError(w, \"error in password\")\n\t\tglobalPkg.WriteLog(logobj, \"error in password\", \"failed\")\n\t\treturn\n\t}\n\t// Signture string\n\tvalidSig := false\n\tpk1 := account.FindpkByAddress(accountObj.AccountPublicKey).Publickey\n\tif pk1 != \"\" {\n\t\tpublickey1 := cryptogrpghy.ParsePEMtoRSApublicKey(pk1)\n\t\tsignatureData := requestObj.Publickey + requestObj.Password + requestObj.FileID\n\t\tvalidSig = cryptogrpghy.VerifyPKCS1v15(requestObj.Signture, signatureData, *publickey1)\n\t} else {\n\t\tvalidSig = false\n\t}\n\t// validSig = true\n\tif !validSig {\n\t\tglobalPkg.SendError(w, \"you are not allowed to delete unshare file\")\n\t\tglobalPkg.WriteLog(logobj, \"you are not allowed to delete unshare file\", \"failed\")\n\t\treturn\n\t}\n\tfound := false\n\tsharefile := filestorage.FindSharedfileByAccountIndex(accountObj.AccountIndex)\n\tif len(sharefile.OwnerSharefile) != 0 {\n\t\tfor sharefileindex, sharefileObj := range sharefile.OwnerSharefile {\n\t\t\tfileindex := containsfileidindex(sharefileObj.Fileid, requestObj.FileID)\n\t\t\tif fileindex != -1 {\n\t\t\t\tfound = true\n\t\t\t\tsharefileObj.Fileid = append(sharefileObj.Fileid[:fileindex], sharefileObj.Fileid[fileindex+1:]...)\n\t\t\t\tsharefile.OwnerSharefile = append(sharefile.OwnerSharefile[:sharefileindex], sharefile.OwnerSharefile[sharefileindex+1:]...)\n\t\t\t\t// fmt.Println(\"============== file ids :\", len(sharefileObj.Fileid), \"============\", len(sharefile.OwnerSharefile))\n\t\t\t\t// delete from permission list\n\t\t\t\taccountOwnerObj := account.GetAccountByAccountPubicKey(sharefileObj.OwnerPublicKey)\n\t\t\t\tFilelistOwner := accountOwnerObj.Filelist\n\t\t\t\tvar indexpk int = -1\n\t\t\t\tvar indexfile int = -1\n\t\t\t\tfor j, fileOwnerObj := range FilelistOwner {\n\t\t\t\t\tif fileOwnerObj.Fileid == requestObj.FileID {\n\t\t\t\t\t\tif len(fileOwnerObj.PermissionList) != 0 {\n\t\t\t\t\t\t\tfor k, pkpermission := range fileOwnerObj.PermissionList {\n\t\t\t\t\t\t\t\tif pkpermission == requestObj.Publickey {\n\t\t\t\t\t\t\t\t\tindexpk = k\n\t\t\t\t\t\t\t\t\tindexfile = j\n\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tif indexpk != -1 {\n\t\t\t\t\taccountOwnerObj.Filelist[indexfile].PermissionList = append(accountOwnerObj.Filelist[indexfile].PermissionList[:indexpk], accountOwnerObj.Filelist[indexfile].PermissionList[indexpk+1:]...)\n\t\t\t\t\tbroadcastTcp.BoardcastingTCP(accountOwnerObj, \"updateaccountFilelist\", \"file\")\n\t\t\t\t\t// accountOwnerObj.Filelist = FilelistOwner\n\t\t\t\t}\n\t\t\t\t//\n\t\t\t\tif len(sharefileObj.Fileid) != 0 && len(sharefile.OwnerSharefile) >= 1 {\n\t\t\t\t\tsharefile.OwnerSharefile = append(sharefile.OwnerSharefile, sharefileObj)\n\t\t\t\t} else if len(sharefileObj.Fileid) != 0 && len(sharefile.OwnerSharefile) == 0 {\n\t\t\t\t\tsharefile.OwnerSharefile = append(sharefile.OwnerSharefile, sharefileObj)\n\t\t\t\t}\n\t\t\t\tbroadcastTcp.BoardcastingTCP(sharefile, \"updatesharefile\", \"file\")\n\n\t\t\t\tif len(sharefile.OwnerSharefile) == 0 {\n\t\t\t\t\tbroadcastTcp.BoardcastingTCP(sharefile, \"deleteaccountindex\", \"file\")\n\t\t\t\t}\n\t\t\t\tglobalPkg.SendResponseMessage(w, \"you unshare file successfully\")\n\t\t\t\tglobalPkg.WriteLog(logobj, \"you unshare file successfully\", \"success\")\n\t\t\t\treturn\n\n\t\t\t}\n\t\t}\n\t}\n\n\tif !found {\n\t\tglobalPkg.SendError(w, \"you not take share file\")\n\t\tglobalPkg.WriteLog(logobj, \"you not take share file\", \"failed\")\n\t\treturn\n\t}\n\n}", "func GetShare(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *ShareState, opts ...pulumi.ResourceOption) (*Share, error) {\n\tvar resource Share\n\terr := ctx.ReadResource(\"google-native:file/v1beta1:Share\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (api *objectAPI) SyncDelete(obj *objstore.Object) error {\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, writeErr = apicl.ObjstoreV1().Object().Delete(context.Background(), &obj.ObjectMeta)\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleObjectEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\t}\n\n\treturn writeErr\n}", "func PostDeleteTrack(w http.ResponseWriter, r *http.Request) {\n\n\treqData := map[string]string{}\n\n\t// Parse JSON Data\n\tdec := json.NewDecoder(r.Body)\n\terr := dec.Decode(&reqData)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tt := reqData[\"track_id\"]\n\n\tcontext.tq.remove(t)\n\n\tw.WriteHeader(204)\n\tw.Write([]byte(`{\"status\":\"deleted\", \"track\":\"` + t + `\"}`))\n}", "func (m *manifestService) Delete(ctx context.Context, dgst digest.Digest) error {\n\tcontext.GetLogger(ctx).Debugf(\"(*manifestService).Delete\")\n\treturn m.manifests.Delete(withRepository(ctx, m.repo), dgst)\n}", "func GetPendingTask(uid int64, shared bool) (*Task, error) {\n\t//declarations\n\trows, err := db.Query(\"SELECT tasks.id, users.token FROM tasks \"+\n\t\t\"INNER JOIN group_tasks ON tasks.gid = group_tasks.id \"+\n\t\t\"INNER JOIN users ON group_tasks.uid = users.id \"+\n\t\t\"WHERE group_tasks.uid = $1 AND tasks.status = $2\", uid, Pending)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer rows.Close()\n\n\t// fetch task\n\tfor rows.Next() {\n\t\tvar tid, user_token string\n\t\tif err := rows.Scan(&tid, &user_token); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\ttask, err := GetTask(tid, user_token)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn task, nil\n\t}\n\n\tif shared {\n\t\t//declarations\n\t\trows, err := db.Query(\"SELECT tasks.id, users.token FROM tasks \"+\n\t\t\t\"INNER JOIN group_tasks ON tasks.gid = group_tasks.id \"+\n\t\t\t\"INNER JOIN users ON group_tasks.uid = users.id \"+\n\t\t\t\"WHERE tasks.status = $1\", Pending)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// fetch task\n\t\tdefer rows.Close()\n\t\tfor rows.Next() {\n\t\t\tvar tid, user_token string\n\t\t\tif err := rows.Scan(&tid, &user_token); err != nil {\n\t\t\t\treturn nil, nil\n\t\t\t}\n\n\t\t\ttask, err := GetTask(tid, user_token)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\treturn task, nil\n\t\t}\n\t}\n\n\treturn nil, nil\n}", "func (cache *diskBlockCacheWrapped) Delete(ctx context.Context,\n\tblockIDs []kbfsblock.ID) (numRemoved int, sizeRemoved int64, err error) {\n\t// This is a write operation but we are only reading the pointers to the\n\t// caches. So we use a read lock.\n\tcache.mtx.RLock()\n\tdefer cache.mtx.RUnlock()\n\tnumRemoved, sizeRemoved, err = cache.workingSetCache.Delete(ctx, blockIDs)\n\tif cache.syncCache == nil || err != nil {\n\t\treturn numRemoved, sizeRemoved, err\n\t}\n\tsyncNumRemoved, syncSizeRemoved, err :=\n\t\tcache.syncCache.Delete(ctx, blockIDs)\n\treturn numRemoved + syncNumRemoved, sizeRemoved + syncSizeRemoved, err\n}", "func (c *BinaryFileClient) DeleteOne(bf *BinaryFile) *BinaryFileDeleteOne {\n\treturn c.DeleteOneID(bf.ID)\n}", "func (httpfs *FS) Delete(ctx context.Context, path string) (err error) {\n\treturn qfs.ErrReadOnly\n}", "func (p *Peer) Delete(ctx context.Context, dgst digest.Digest) error {\n\tc, err := digestconv.DigestToCid(dgst)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert digest\")\n\t}\n\n\treturn p.dserv.Remove(ctx, c)\n}", "func (p *GetService) Get(request string, reply *string) error {\n\tvar sdfsfilename string\n\tvar localfilename string\n\tvar IP string\n\tIP = strings.Split(request, \" \")[0]\n\tsdfsfilename = strings.Split(request, \" \")[1]\n\tlocalfilename = strings.Split(request, \" \")[2]\n\ttime.Sleep(time.Second * 1)\n\tputfile(sdfsfilename, localfilename, IP)\n\treturn nil\n}", "func GetDatastoreFileshare(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *DatastoreFileshareState, opts ...pulumi.ResourceOption) (*DatastoreFileshare, error) {\n\tvar resource DatastoreFileshare\n\terr := ctx.ReadResource(\"azure:machinelearning/datastoreFileshare:DatastoreFileshare\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (r *repository) Delete(id uint) error {\n\tif err := r.db.Where(\"id = ?\", id).Delete(&models.Upload{}).Error; err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (l *RemoteProvider) DeleteSchedule(req *http.Request, scheduleID string) ([]byte, error) {\n\tif !l.Capabilities.IsSupported(PersistSchedules) {\n\t\tlogrus.Error(\"operation not available\")\n\t\treturn nil, ErrInvalidCapability(\"PersistSchedules\", l.ProviderName)\n\t}\n\n\tep, _ := l.Capabilities.GetEndpointForFeature(PersistSchedules)\n\n\tlogrus.Infof(\"attempting to fetch schedule from cloud for id: %s\", scheduleID)\n\n\tremoteProviderURL, _ := url.Parse(fmt.Sprintf(\"%s%s/%s\", l.RemoteProviderURL, ep, scheduleID))\n\tlogrus.Debugf(\"constructed schedule url: %s\", remoteProviderURL.String())\n\tcReq, _ := http.NewRequest(http.MethodDelete, remoteProviderURL.String(), nil)\n\n\ttokenString, err := l.GetToken(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := l.DoRequest(cReq, tokenString)\n\tif err != nil {\n\t\tlogrus.Errorf(\"unable to delete schedules: %v\", err)\n\t\treturn nil, ErrDelete(err, \"Perf Schedule :\"+scheduleID, resp.StatusCode)\n\t}\n\tdefer func() {\n\t\t_ = resp.Body.Close()\n\t}()\n\tbdr, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn nil, ErrDataRead(err, \"Perf Schedule :\"+scheduleID)\n\t}\n\n\tif resp.StatusCode == http.StatusOK {\n\t\tlogrus.Infof(\"schedule successfully retrieved from remote provider\")\n\t\treturn bdr, nil\n\t}\n\treturn nil, ErrDelete(err, fmt.Sprint(bdr), resp.StatusCode)\n}", "func (fs *Ipfs) Delete(path string) error {\n\t// Remoe file if on disk and unpinn\n\tif fname, err := fs.makeFilename(path); err == nil {\n\t\tos.Remove(fname)\n\t}\n\n\tipath := ipath.New(path)\n\treturn fs.coreAPI.Pin().Rm(context.Background(), ipath)\n}", "func (self *CassandraMetaStore) GetPending(v *meta.RequestVars) (*meta.Object, error) {\n\treturn self.findOid(v.Oid, true)\n}", "func (r *RecordingService) Delete(ctx context.Context, sid string) error {\n\treturn r.client.DeleteResource(ctx, recordingsPathPart, sid)\n}", "func (f FileURL) Delete(ctx context.Context) (*FileDeleteResponse, error) {\n\treturn f.fileClient.Delete(ctx, nil)\n}", "func (s *server) Delete(_ context.Context, request *pb.DeleteRequest) (*pb.DeleteResponse, error) {\n\tfmt.Printf(\"deleting %s\\n\", path.Join(s.localReplicaPath, request.FullPath))\n\terr := os.Remove(path.Join(s.localReplicaPath, request.FullPath))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to delete at %s: %w\", request.FullPath, err)\n\t}\n\treturn &pb.DeleteResponse{}, nil\n}", "func (f *Fs) delete(ctx context.Context, isFile bool, id string, remote string, hardDelete bool) (err error) {\n\tif hardDelete {\n\t\topts := rest.Opts{\n\t\t\tMethod: \"DELETE\",\n\t\t\tRootURL: id,\n\t\t\tNoResponse: true,\n\t\t}\n\t\treturn f.pacer.Call(func() (bool, error) {\n\t\t\tresp, err := f.srv.Call(ctx, &opts)\n\t\t\treturn shouldRetry(ctx, resp, err)\n\t\t})\n\t}\n\t// Move file/dir to deleted files if not hard delete\n\tleaf := path.Base(remote)\n\tif isFile {\n\t\t_, err = f.moveFile(ctx, id, leaf, f.opt.DeletedID)\n\t} else {\n\t\terr = f.moveDir(ctx, id, leaf, f.opt.DeletedID)\n\t}\n\treturn err\n}", "func (dtm *DfgetTaskManager) Delete(ctx context.Context, clientID, taskID string) error {\n\tkey, err := generateKey(clientID, taskID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn dtm.dfgetTaskStore.Delete(key)\n}", "func (p *SharedPrivateLinkResourcesDeletePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (s *shares) Get(shareID string) (Share, error) {\n\tres, err := s.c.baseRequest(http.MethodGet, routes.shares, nil, shareID)\n\tif err != nil {\n\t\treturn Share{}, err\n\t}\n\tvar r sharesListResponse\n\tres.JSON(&r)\n\treturn r.Ocs.Data[0], nil\n}", "func (dq *deleteQueue) deleteFromPending() {\n\tdq.Lock()\n\tdefer dq.Unlock()\n\n\tn := len(dq.entries)\n\tif n > awsBatchSizeLimit {\n\t\tn = awsBatchSizeLimit\n\t}\n\tfails, err := dq.deleteBatch(dq.entries[:n])\n\tif err != nil {\n\t\tdq.svc.Logger(\"Error deleting batch: %s\", err)\n\t\treturn\n\t}\n\n\tdq.entries = dq.entries[n:]\n\n\tif len(fails) > 0 {\n\t\tdq.entries = append(dq.entries, fails...)\n\t}\n}", "func (obj *MessengerCloudFsFileInfoList) Delete() {\n\tif obj == nil {\n\t\treturn\n\t}\n\truntime.SetFinalizer(obj, nil)\n\tobj.delete()\n}", "func Get() (int, error) {\n\tstorage, err := New(SharedDir)\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\n\tif err := storage.Lock(); err != nil {\n\t\treturn -1, err\n\t}\n\tdefer storage.Unlock()\n\ts, err := storage.Acquire()\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\treturn s, nil\n}", "func (sfs *SafeFileSet) Delete(metahash [32]byte) {\n\tsfs.filesSet.Delete(metahash)\n}", "func (s *Server) Delete(ctx context.Context, req *pb.DeleteRequest) (*pb.Empty, error) {\n\tp, ok := peer.FromContext(ctx)\n\tif !ok {\n\t\treturn nil, errors.New(\"failed to extract peer details from context\")\n\t}\n\ts.logger.Info(\"delete request received\", zap.Stringer(\"addr\", p.Addr), zap.String(\"key\", req.Key))\n\n\tif err := s.store.Delete(req.GetKey(), time.Now().UTC().UnixNano()); err != nil {\n\t\treturn &pb.Empty{}, fmt.Errorf(\"failed to process delete: %w\", err)\n\t}\n\treturn &pb.Empty{}, nil\n}", "func (g *gcs) Delete(ctx context.Context, remotePath string) (err error) {\n\tif err = g.bucket.Object(remotePath).Delete(g.context); err != nil {\n\t\treturn err\n\t}\n\n\treturn\n}", "func (b *Blob) Delete() error {\n\tif ref := b.DecRefCount(); ref == 0 {\n\t\tutils.Remove(b.FilePath())\n\t\tutils.Remove(b.RefCountPath())\n\t}\n\treturn nil\n}", "func (s *scheduledWorkQueue) Forget(obj interface{}) {\n\ts.workLock.Lock()\n\tdefer s.workLock.Unlock()\n\n\tif cancel, ok := s.work[obj]; ok {\n\t\tcancel()\n\t\tdelete(s.work, obj)\n\t}\n}", "func (a *UtilsApiService) DeleteNotificationUsingDelete(ctx context.Context, notificationId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\ta.client = NewAPIClient(&Configuration{\n\t\tBasePath: ctx.Value(\"BasePath\").(string),\n\t\tDefaultHeader: make(map[string]string),\n\t\tUserAgent: \"Swagger-Codegen/1.0.0/go\",\n\t})\n\tlocalVarPath := a.client.cfg.BasePath + \"/nucleus/v1/notification/{notification_id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"notification_id\"+\"}\", fmt.Sprintf(\"%v\", notificationId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"*/*\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (s TensePresIndResource) Delete(id string, r api2go.Request) (api2go.Responder, error) {\n\terr := s.TensePresIndStorage.Delete(id)\n\n\tif err != nil {\n\t\treturn &Response{Code: http.StatusNotFound}, api2go.NewHTTPError(err, err.Error(), http.StatusNotFound)\n\t}\n\n\treturn &Response{Code: http.StatusNoContent}, err\n}", "func (o *SwiftObject) Delete(metadata map[string]string) error {\n\tif _, err := o.newFile(\"ts\", 0); err != nil {\n\t\treturn err\n\t} else {\n\t\tdefer o.Close()\n\t\treturn o.Commit(metadata)\n\t}\n}", "func delete(resource string, id string) ([]byte, error) {\n\thttpParams := &HTTPParams{\n\t\tResource: resource + \"/\" + id,\n\t\tVerb: \"DELETE\",\n\t}\n\treturn processRequest(httpParams)\n}", "func (r *MonitorGTPResource) Delete(id string) error {\n\tif err := r.c.ModQuery(\"DELETE\", BasePath+MonitorGTPEndpoint+\"/\"+id, nil); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *ResourcesHandler) Delete(event.DeleteEvent, workqueue.RateLimitingInterface) {}", "func (s *CarShareStorage) Delete(id string, context api2go.APIContexter) error {\n\t_, exists := s.carShares[id]\n\tif !exists {\n\t\treturn storage.ErrNotFound\n\t}\n\tdelete(s.carShares, id)\n\n\treturn nil\n}", "func (s *s3ManifestService) Delete(ctx context.Context, dgst godigest.Digest) error {\n\treturn fmt.Errorf(\"unimplemented\")\n}", "func (p *StoragesDeletePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func (c *PluginContext) SyncDelete(key ContextKey) {\n\tc.Mx.Lock()\n\tdefer c.Mx.Unlock()\n\tc.Delete(key)\n}", "func (obj *MessengerFileCipher) Delete() {\n\tif obj == nil {\n\t\treturn\n\t}\n\truntime.SetFinalizer(obj, nil)\n\tobj.delete()\n}", "func (h *HubService) GetDeleteToken(repo string, auth Authenticator) (*TokenAuth, error) {\n\tpath := fmt.Sprintf(\"v1/repositories/%s/\", repo)\n\treq, err := h.newRequest(\"DELETE\", path, auth)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn h.do(req)\n}", "func (a *UtilsApiService) DeleteNotificationSettingUsingDelete(ctx context.Context, notificationSettingId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\ta.client = NewAPIClient(&Configuration{\n\t\tBasePath: ctx.Value(\"BasePath\").(string),\n\t\tDefaultHeader: make(map[string]string),\n\t\tUserAgent: \"Swagger-Codegen/1.0.0/go\",\n\t})\n\tlocalVarPath := a.client.cfg.BasePath + \"/nucleus/v1/notification_setting/{notification_setting_id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"notification_setting_id\"+\"}\", fmt.Sprintf(\"%v\", notificationSettingId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"*/*\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func Delete(c *gophercloud.ServiceClient, networkID string) os.DeleteResult {\n\treturn os.Delete(c, networkID)\n}", "func (c *UnsavedPostAttachmentClient) Delete() *UnsavedPostAttachmentDelete {\n\tmutation := newUnsavedPostAttachmentMutation(c.config, OpDelete)\n\treturn &UnsavedPostAttachmentDelete{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (f File) Delete() File {\n\t//DPrintf(\"delete %v from versions %v\", f.Path, f)\n\tif f.Deleted {\n\t\tDPrintf(\"WARNING: deleting an already deleted file\")\n\t}\n\tf.Deleted = true\n\treturn f\n}", "func (_BaseAccessWallet *BaseAccessWalletTransactor) ClearPending(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _BaseAccessWallet.contract.Transact(opts, \"clearPending\")\n}", "func (m *DisksClientMock) Get(ctx context.Context, resourceGroupName string, diskName string) (result compute.Disk, rerr *retry.Error) {\n\treturn compute.Disk{}, nil\n}", "func (t *Token) Delete() error {\n\treturn t.deleteFile()\n}", "func (w *wireMap) AtomicDelete(wire *GRPCWire) error {\n\tw.mu.Lock()\n\tdefer w.mu.Unlock()\n\tdelete(w.wires, linkKey{\n\t\tnamespace: wire.LocalPodNetNS,\n\t\tlinkUID: wire.UID,\n\t})\n\n\tdelete(w.handles, wire.LocalNodeIfaceID)\n\n\treturn nil\n}", "func Delete(instructionData reflect.Value, finished chan bool) int {\n\tfmt.Println(\"FIBER INFO: Deleting File ...\")\n\n\tpath, err := variable.GetValue(instructionData, \"PathVarName\", \"PathIsVar\", \"Path\")\n\tif err != nil {\n\t\tfinished <- true\n\t\treturn -1\n\t}\n\n\tos.Remove(path.(string))\n\tfinished <- true\n\treturn -1\n}", "func (p *DeleteService) ForwardDelete(request string, reply *string) error {\n\t_, index := Getip()\n\tmemid := election()\n\tif _, ok := fileList[request]; ok {\n\t\tnodelist := fileList[request].NodeList\n\t\tfor _, i := range nodelist {\n\t\t\tif i == index {\n\t\t\t\tos.Remove(request)\n\t\t\t} else {\n\t\t\t\tclient, err := rpc.Dial(\"tcp\", straddr[i]+\":4000\")\n\t\t\t\tif (err != nil) {\n\t\t\t\t\tfmt.Println(err)\n\t\t\t\t}\n\t\t\t\terr = client.Call(\"DeleteService.Delete\", request, &reply)\n\t\t\t\tif err != nil {\n\t\t\t\t\tfmt.Println(err)\n\t\t\t\t}\n\t\t\t\t*reply = \"file delete!\"\n\t\t\t\tdefer client.Close()\n\t\t\t}\n\t\t}\n\t\tfor _, index1 := range memid {\n\t\t\tif index1 == index {\n\t\t\t\tdelete(fileList, request)\n\t\t\t} else {\n\t\t\t\tvar temp = make(map[string]pair)\n\t\t\t\ttemp[request] = pair{\n\t\t\t\t\tTimeStamp: 0,\n\t\t\t\t\tNodeList: nil,\n\t\t\t\t}\n\t\t\t\tsendlist(temp, straddr[index1])\n\t\t\t}\n\t\t}\n\t} else {\n\t\t*reply = \"no such file!\"\n\t}\n\treturn nil\n}", "func (c *ServerConn) Delete(path string) error {\n\t_, _, err := c.cmd(StatusRequestedFileActionOK, \"DELE %s\", path)\n\treturn err\n}", "func (pc pidController) delete() {\n\tpid := pc.pidProvider.provide()\n\tfilePID, err := pc.get()\n\tif err != nil {\n\t\toktetoLog.Infof(\"unable to delete PID file at %s: %s\", pc.pidFilePath, err)\n\t}\n\n\tif err := pc.watcher.Close(); err != nil {\n\t\toktetoLog.Infof(\"could not close watcher: %w\", err)\n\t}\n\n\tif strconv.Itoa(pid) != filePID {\n\t\toktetoLog.Infof(\"okteto process with PID '%s' has the ownership of the file.\", filePID)\n\t\toktetoLog.Info(\"skipping deletion of PID file\")\n\t\treturn\n\t}\n\terr = pc.filesystem.Remove(pc.pidFilePath)\n\tif err != nil && !os.IsNotExist(err) {\n\t\toktetoLog.Infof(\"unable to delete PID file at %s: %s\", pc.pidFilePath, err)\n\t}\n}", "func (t *FakeObjectTracker) Delete(gvr schema.GroupVersionResource, ns, name string) error {\n\treturn nil\n}", "func TestDeleteFile(t *testing.T) {\n\tfmt.Println(\"Get File info test\")\n\tcc := getClientConnection()\n\n\t// when done the connection will be close.\n\tdefer cc.Close()\n\n\t// Create a new client service...\n\tc := filepb.NewFileServiceClient(cc)\n\n\trqst := &filepb.DeleteFileRequest{\n\t\tPath: \"C:\\\\Temp\\\\toto.bmp\",\n\t}\n\n\trsp, err := c.DeleteFile(context.Background(), rqst)\n\tif err != nil {\n\t\tlog.Fatalf(\"error while testing get file info: %v\", err)\n\t}\n\n\tlog.Println(\"Delete file succeed:\", rsp.Result)\n}", "func ExampleFileSharesClient_Get() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclient, err := armstorage.NewFileSharesClient(\"{subscription-id}\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := client.Get(ctx,\n\t\t\"res9871\",\n\t\t\"sto6217\",\n\t\t\"share1634\",\n\t\t&armstorage.FileSharesClientGetOptions{Expand: to.Ptr(\"stats\"),\n\t\t\tXMSSnapshot: nil,\n\t\t})\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// TODO: use response item\n\t_ = res\n}", "func (_BaseContentSpace *BaseContentSpaceTransactor) ClearPending(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _BaseContentSpace.contract.Transact(opts, \"clearPending\")\n}" ]
[ "0.50653577", "0.5034428", "0.4935411", "0.4918671", "0.47782397", "0.47302368", "0.46873987", "0.46159875", "0.45806068", "0.45560926", "0.4527142", "0.45016083", "0.44963208", "0.44932404", "0.44923154", "0.44845128", "0.4465482", "0.44443616", "0.44268724", "0.442413", "0.44082674", "0.44048682", "0.44005018", "0.43910846", "0.43831557", "0.43810475", "0.43648708", "0.4353177", "0.43451512", "0.43398616", "0.4319744", "0.43115503", "0.43091625", "0.43084115", "0.4307879", "0.43002096", "0.42881882", "0.42863333", "0.4267286", "0.425731", "0.42411047", "0.42373547", "0.41976142", "0.4197086", "0.41945192", "0.41881", "0.4184357", "0.41768727", "0.41714287", "0.4167408", "0.41647607", "0.41629493", "0.41627523", "0.41506726", "0.41478115", "0.41429427", "0.41390234", "0.41254988", "0.41241863", "0.411722", "0.41035706", "0.40868276", "0.40857258", "0.4085581", "0.4082384", "0.40784743", "0.40677637", "0.4067659", "0.40650195", "0.40628585", "0.40624565", "0.40601888", "0.40564308", "0.40552875", "0.40537018", "0.40471813", "0.4042135", "0.4038989", "0.40293342", "0.40176928", "0.40136307", "0.40126547", "0.40122333", "0.4012112", "0.40117556", "0.4008499", "0.40054828", "0.4002601", "0.3997073", "0.39961907", "0.39922374", "0.39905113", "0.39863193", "0.3984429", "0.39837417", "0.3983345", "0.39819688", "0.39810938", "0.39750004", "0.39738235" ]
0.758004
0
Set adds a new PendingFileshare in the repository. If the key already exists, it returns false, otherwise it returns true.
func (r *inMemoryFileshareRepository) Set(key string, fd PendingFileshare) bool { r.Lock() defer r.Unlock() if _, ok := r.pendingFileshares[key]; ok { return false // Key already exists. } if r.pendingFileshares == nil { r.pendingFileshares = make(map[string]PendingFileshare) } r.pendingFileshares[key] = fd return true }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (f *Filter) Set(key string) (bool, error) {\n\tcmd := \"s \" + f.Name + \" \" + f.getKey(key)\n\tresp, err := f.Conn.SendAndReceive(cmd)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tif resp == \"Yes\" || resp == \"No\" {\n\t\treturn resp == \"Yes\", nil\n\t}\n\treturn false, errInvalidResponse(resp)\n}", "func (r *inMemoryFileshareRepository) GetAndDelete(key string) (PendingFileshare, bool) {\n\tr.Lock()\n\tdefer r.Unlock()\n\n\tfileshare, ok := r.pendingFileshares[key]\n\tif !ok {\n\t\treturn PendingFileshare{}, false\n\t}\n\n\tdelete(r.pendingFileshares, key)\n\treturn fileshare, true\n}", "func (l *localStore) Set(key string, b []byte) error {\n\t_, ok := l.m[key]\n\tl.m[key] = b\n\tif ok {\n\t\treturn ErrKeyExist\n\t}\n\treturn nil\n}", "func (s *SyncStorage) SetIfNotExists(ns string, key string, data interface{}) (bool, error) {\n\treturn s.getDbBackend(ns).SetNX(getNsPrefix(ns)+key, data, 0)\n}", "func (kvs *FS) Set(key string, value []byte) error {\n\treturn lockedfile.Write(kvs.filename(key), bytes.NewReader(value), 0600)\n}", "func (c *fileStorageClient) Set(key string, value []byte) error {\n\treturn c.Batch(SetOperation(key, value))\n}", "func (fs *FSCache) Set(key string, content []byte) error {\n\treturn ioutil.WriteFile(\n\t\tpath.Join(fs.Root, key),\n\t\tcontent,\n\t\t0600,\n\t)\n}", "func (f *NaiveMap) Set(key string, value interface{}) bool {\n\tf.lock.Lock()\n\tdefer f.lock.Unlock()\n\n\tif f.capacity <= f.count {\n\t\tlog.Print(\"At maximum capacity!\")\n\t\treturn false\n\t}\n\n\tif key == \"\" {\n\t\tlog.Print(\"Invalid input.\")\n\t\treturn false\n\t}\n\t// Check if they key has been set, if so, put value with key.\n\tfor i, _ := range f.keys {\n\t\tif f.keys[i] == key {\n\t\t\tf.values[i] = value\n\t\t\treturn true\n\t\t}\n\t}\n\n\t// Make sure there's an an empty spot, then add.\n\tfor i, _ := range f.keys {\n\t\tif f.keys[i] == \"\" {\n\t\t\tf.keys[i] = key\n\t\t\tf.values[i] = value\n\t\t\tf.count++\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (c *Cache) Set(key string, msg *CacheItem) bool {\n\tc.mu.Lock()\n\tc.backend[key] = msg\n\tc.mu.Unlock()\n\n\treturn true\n}", "func (storage *PaymentChannelStorage) PutIfAbsent(key *PaymentChannelKey, state *PaymentChannelData) (ok bool, err error) {\n\treturn storage.delegate.PutIfAbsent(key, state)\n}", "func (s *SharemeService) Add(c *gae.Context, session *Session, key string) (stat Share) {\n\tstat = s.Stat(c, key)\n\tif stat.IsError() {\n\t\treturn\n\t}\n\tsession.Set(fmt.Sprintf(\"%s%s\", KeySessionPrefix, key), stat.Name)\n\treturn\n}", "func (s *SyncStorage) SetIfNotExistsAndPublish(ns string, channelsAndEvents []string, key string, data interface{}) (bool, error) {\n\tnsPrefix := getNsPrefix(ns)\n\tif len(channelsAndEvents) == 0 {\n\t\treturn s.getDbBackend(ns).SetNX(nsPrefix+key, data, 0)\n\t}\n\tif err := s.checkChannelsAndEvents(\"SetIfNotExistsAndPublish\", channelsAndEvents); err != nil {\n\t\treturn false, err\n\t}\n\tchannelsAndEventsPrepared := s.prepareChannelsAndEvents(nsPrefix, channelsAndEvents)\n\treturn s.getDbBackend(ns).SetNXPub(channelsAndEventsPrepared, nsPrefix+key, data)\n}", "func (m *mSignatureKeyHolderMockEquals) Set(f func(p SignatureKeyHolder) (r bool)) *SignatureKeyHolderMock {\n\tm.mainExpectation = nil\n\tm.expectationSeries = nil\n\n\tm.mock.EqualsFunc = f\n\treturn m.mock\n}", "func (storage *remoteStorage) Set(key *data.ID, value []byte) error {\n\tnode := storage.node\n\n\turl := fmt.Sprintf(\"http://%s/storage/%s\", node.TCPAddr().String(), key.String())\n\n\t// Base64 encoding, RFC 4648.\n\t// str := base64.StdEncoding.EncodeToString(value)\n\treq, err := http.NewRequest(http.MethodPut, url, bytes.NewReader(value))\n\tif req.Body != nil {\n\t\tdefer req.Body.Close()\n\t}\n\tif err != nil {\n\t\tnode.disconnect(err)\n\t\treturn err\n\t}\n\tres, err := http.DefaultClient.Do(req)\n\tif err != nil {\n\t\tnode.disconnect(err)\n\t\treturn err\n\t}\n\tif res.Body != nil {\n\t\tdefer res.Body.Close()\n\t}\n\tif res.StatusCode < 200 || res.StatusCode > 299 {\n\t\terr := fmt.Errorf(\"HTTP storage Put %s -> %d %s\", url, res.StatusCode, res.Status)\n\t\tnode.disconnect(err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (ak *AlertKey) Set(key string) {\n\t_ = ak.Akdb.Put([]byte(key), []byte(\"true\"), nil)\n}", "func (s *CacheServer) Set(ctx context.Context, item *pb.CacheItem) (*pb.Success, error) {\n\texpire, _ := time.Parse(time.RFC3339, item.ExpireTime)\n\terr := cache.Set(item.Key, item.Value, expire)\n\tif err != nil {\n\t\treturn &pb.Success{\n\t\t\tSuccess: false,\n\t\t}, err\n\t}\n\tlog.Printf(\"set item: %v, %v, %v\", item.Key, item.Value, expire)\n\treturn &pb.Success{\n\t\tSuccess: true,\n\t}, nil\n}", "func (s *InMemoryHashStorage) Set(key int64, value string) {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\n\ts.hashes[key] = value\n}", "func (c *BlockCache) Set(key string) {\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\tkey = strings.ToLower(key)\n\tc.m[key] = true\n}", "func (s *Store) Put(key string, value []byte) bool {\n\n\tif _, ok := s.objects.Load(key); !ok {\n\t\ts.objects.Store(key, value)\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (fc *fileCache) Set(key, value string, ttl int) {\n\tfc.cache.Set(key, &cacheObject{\n\t\tValue: value,\n\t\tTimestamp: time.Now().Unix(),\n\t\tTTL: ttl,\n\t})\n\tfc.dirty = true\n}", "func (us *urlStatus) Set(key string, val bool) {\n\tus.mu.Lock()\n\t// Lock so only one goroutine at a time can access the map c.v.\n\tdefer us.mu.Unlock()\n\tus.v[key] = val\n}", "func (t *SBF) Add(hashes []uint64) bool {\n // Check if the key is in SBF already\n t.mutex.RLock()\n for _, pf := range t.plainFilters {\n if pf.Has(hashes) {\n t.mutex.RUnlock()\n // Has the key already.\n return true\n }\n }\n t.mutex.RUnlock()\n\n // Add the key to SBF\n t.mutex.Lock()\n defer t.mutex.Unlock()\n pf := t.plainFilters[len(t.plainFilters)-1]\n\n if t.keys == t.capacity {\n // SBF is full. Expand it by attaching another plainFilter\n pf := plainFilter.NewPlainFilter(scale_size*pf.Capacity, r*pf.Probability)\n t.plainFilters = append(t.plainFilters, pf)\n atomic.AddUint64(&t.capacity, pf.Capacity)\n }\n\n // In most cases added is false. Since we checked the key is not in the filter in the\n // top half of this function. But there is a tiny chance there is a context switch happens\n // between the RWLock and we could add the same key twice. So double check added here.\n added := pf.Add(hashes)\n if !added {\n atomic.AddUint64(&t.keys, 1)\n }\n \n return added\n}", "func (c *SimpleCache) Set(k string, v interface{}, ttl int64) bool {\n\tc.Lock.Lock()\n\tdata, present := c.getData(k, ttl)\n\tif present == true {\n\t\tif ttl == -1 {\n\t\t\tdata.ExpireAt = time.Now().Unix() + c.TTL\n\t\t} else {\n\t\t\tdata.ExpireAt = time.Now().Unix() + ttl\n\t\t}\n\t\tc.Queue.update(data)\n\t\tc.Data[k] = data\n\t} else {\n\t\tvar ttx int64\n\t\tif ttl == -1 {\n\t\t\tttx = time.Now().Unix() + c.TTL\n\t\t} else {\n\t\t\tttx = time.Now().Unix() + ttl\n\t\t}\n\t\tnewData := newQueueItem(k, v, ttx)\n\t\tc.Queue.push(newData)\n\t\tc.Data[k] = newData\n\t}\n\tc.Lock.Unlock()\n\treturn true\n}", "func (ps *peerSet) Put(addr net.UDPAddr) bool {\n\ts := addr.String()\n\n\tif _, ok := ps.values[s]; ok {\n\t\treturn false\n\t}\n\n\tps.values[s] = struct{}{}\n\n\tr := &ring.Ring{\n\t\tValue: addr,\n\t}\n\n\tif ps.ring == nil {\n\t\tps.ring = r\n\t} else {\n\t\tps.ring.Link(r)\n\t}\n\n\treturn true\n}", "func (b *Bcache) Set(key, val string, expiredTimestamp int64) {\n\tb.peer.Set(key, val, expiredTimestamp)\n}", "func (b *SharingKeys) Add(name string, key *[32]byte) (*SharingKey, error) {\n\tif name == \"\" {\n\t\treturn nil, ErrSharingKeyNameInvalid\n\t}\n\tn := []byte(name)\n\tif v := b.b.Get(n); v != nil {\n\t\treturn nil, ErrSharingKeyExist\n\t}\n\tif err := b.b.Put([]byte(name), key[:]); err != nil {\n\t\treturn nil, err\n\t}\n\ts := &SharingKey{\n\t\tb: b,\n\t\tname: n,\n\t\tsecret: key[:],\n\t}\n\treturn s, nil\n}", "func (c *Cache) Set(key string, value []byte) (changed bool) {\n\tif len(value) != ValueSize {\n\t\t// check done to enable future optimization with fixed sized records.\n\t\tpanic(\"value with bad length provided\")\n\t}\n\n\t// do as much work as possible outside a transaction\n\tk := []byte(key)\n\tchanged = true\n\n\terr := c.db.Update(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket(bkt)\n\t\tcv := b.Get(k)\n\t\tif bytes.Equal(value, cv) {\n\t\t\tchanged = false\n\t\t}\n\t\treturn b.Put(k, value)\n\t})\n\tif err != nil {\n\t\tchanged = true\n\t\tif c.err == nil {\n\t\t\tc.err = err\n\t\t}\n\t}\n\treturn\n}", "func (this HashString) Set(val string) <-chan bool {\n\treturn BoolCommand(this.parent, this.args(\"hset\", val)...)\n}", "func (l *Locker) Set(key Flags, value interface{}) {\n\tl.data.Store(key, value)\n}", "func (s *FileSet) Add(file string) {\n\ts.files[file] = true\n}", "func (c *Cache) IsSet(key string) bool {\n\t_, ok := c.candidateState[key]\n\n\treturn ok\n}", "func (obj *OrderedObject) Set(key string, value interface{}) (exists bool) {\n\tif idx, ok := obj.keys[key]; ok {\n\t\tobj.Object[idx].Value = value\n\t\treturn false\n\t} else {\n\t\tobj.keys[key] = len(obj.Object)\n\t\tobj.Object = append(obj.Object, json.Member{Key: key, Value: value})\n\t\treturn true\n\t}\n}", "func (s Set) Add(key string) {\n\ts[key] = true\n}", "func (mmHasPendings *mClientMockHasPendings) Set(f func(ctx context.Context, object insolar.Reference) (b1 bool, err error)) *ClientMock {\n\tif mmHasPendings.defaultExpectation != nil {\n\t\tmmHasPendings.mock.t.Fatalf(\"Default expectation is already set for the Client.HasPendings method\")\n\t}\n\n\tif len(mmHasPendings.expectations) > 0 {\n\t\tmmHasPendings.mock.t.Fatalf(\"Some expectations are already set for the Client.HasPendings method\")\n\t}\n\n\tmmHasPendings.mock.funcHasPendings = f\n\treturn mmHasPendings.mock\n}", "func (c *AdapterMemory) SetIfNotExist(ctx context.Context, key interface{}, value interface{}, duration time.Duration) (bool, error) {\n\tisContained, err := c.Contains(ctx, key)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tif !isContained {\n\t\tif _, err = c.doSetWithLockCheck(ctx, key, value, duration); err != nil {\n\t\t\treturn false, err\n\t\t}\n\t\treturn true, nil\n\t}\n\treturn false, nil\n}", "func (t *T) Set(key, value string) error {\n\tif _, err := t.file.WriteString(fmt.Sprintf(\"\\n%s,%s\", key, value)); err != nil {\n\t\treturn err\n\t}\n\tt.links[key] = value\n\treturn nil\n}", "func (o *Object) Set(key StringOrSymbol, val, receiver Value) Boolean {\n\tif p, ok := o.fields[key]; ok {\n\t\treturn p.Set()(val)\n\t}\n\treturn False\n}", "func (l *LocalStore) Set(ctx context.Context, key, data string) error {\n\tl.lock.Lock()\n\tl.store[key] = data\n\tl.lock.Unlock()\n\n\treturn nil\n}", "func (h Hash) Add(in string) bool {\n\tif _, ok := h[in]; ok {\n\t\t// already in the set\n\t\treturn false\n\t}\n\t// not in the set\n\th[in] = struct{}{}\n\treturn true\n}", "func (this HashFloat) Set(val float64) <-chan bool {\n\treturn BoolCommand(this.parent, this.args(\"hset\", ftoa(val))...)\n}", "func (s *server) Set(key string, value *string) error {\n\tres := make(chan error)\n\ts.ops <- func() {\n\t\ts.store[key] = value\n\t\tres <- nil\n\t}\n\treturn <-res\n}", "func (s *Storage) Set(key []byte, value []byte) error {\n\treturn s.SetWithTTL(key, value, 0)\n}", "func (mmAdd *mListRepositoryMockAdd) Set(f func(ip *net.IPNet) (b1 bool, err error)) *ListRepositoryMock {\n\tif mmAdd.defaultExpectation != nil {\n\t\tmmAdd.mock.t.Fatalf(\"Default expectation is already set for the ListRepository.Add method\")\n\t}\n\n\tif len(mmAdd.expectations) > 0 {\n\t\tmmAdd.mock.t.Fatalf(\"Some expectations are already set for the ListRepository.Add method\")\n\t}\n\n\tmmAdd.mock.funcAdd = f\n\treturn mmAdd.mock\n}", "func (sc *LoanMarketShareContract) Put(ctx contractapi.TransactionContextInterface, shareID int, titleholder string, amount float64, repayments float64, statutes string, rating float64, status string)\t(err error) {\n\n\tif shareID == 0 {\n\t\terr = errors.New(\"Loan Rating ID can not be empty\")\n\t\treturn\n\t}\n\t\n\tobj := new(LoanMarketShareObj)\n\tobj.ShareID = shareID\n\tobj.TitleHolder = titleholder\n\tobj.Amount = amount\n\tobj.Repayments = repayments\n\tobj.Statutes = statutes\n\tobj.Rating = rating\n\tobj.Status = status\n\t\n\n\tif obj.Created, err = GetTimestamp(ctx); err != nil {\n\t\treturn\n\t}\n\n\tif obj.Createdby, err = GetCallerID(ctx); err != nil {\n\t\treturn\n\t}\n\n\tkey := strconv.Itoa(shareID)\n\tobjBytes, _ := json.Marshal(obj)\t\n\terr = ctx.GetStub().PutState(key, []byte(objBytes))\n return \n}", "func (m *BoltMeta) Set(key, value []byte) error {\n\terr := m.DS.Update(func(tx *bolt.Tx) error {\n\t\tbucket, err := tx.CreateBucketIfNotExists(m.MetadataBucket)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = bucket.Put(key, value)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n\n\treturn err\n}", "func (session KeyValueSession) Set(key DatabaseKey, value interface{}) bool {\n\tbuffer := convertData(value)\n\tsession.store.tree.Set(key, buffer)\n\n\treturn true\n}", "func (m *Module) SetKeyIfNotExists(ctx context.Context, key, value string, t time.Duration) (bool, error) {\n\tm.lock.Lock()\n\tdefer m.lock.Unlock()\n\n\treturn m.client.SetNX(ctx, key, value, t).Result()\n}", "func (sfs *SafeFileSet) AddUnique(metahash [32]byte) (isNew bool) {\n\t_, notNew := sfs.filesSet.LoadOrStore(metahash, true)\n\treturn !notNew\n}", "func (s *SyncStorage) SetIf(ns string, key string, oldData, newData interface{}) (bool, error) {\n\treturn s.getDbBackend(ns).SetIE(getNsPrefix(ns)+key, oldData, newData)\n}", "func (v *Viper) Set(key string, value interface{}) bool {\n\treturn true\n}", "func (c *Cache) Add(key string, value []byte) bool {\n\tkh := c.hash64([]byte(key))\n\n\tc.mutex.Lock()\n\t_, addFail := c.chunkMap[kh]\n\n\tif addFail == false {\n\t\t// cache miss: we can write to the index of nextChunk\n\t\tchunk := c.nextChunk\n\t\t// this index will be changed -> update in-memory struct\n\t\toldMeta := c.metaMap[chunk]\n\t\tnewMeta := MetaEntry{Key: kh, Len: uint32(len(value)), Checksum: c.hash32(value)}\n\n\t\tc.replaceMeta(oldMeta.Key, newMeta, false)\n\n\t\tc.seekToData(chunk)\n\t\tbinary.Write(c.fh, binary.LittleEndian, value)\n\n\t\tc.nextChunk++\n\t\tif c.nextChunk >= c.chunkCount {\n\t\t\tc.nextChunk = 0 // overflowed -> next chunk shall be 0\n\t\t}\n\t}\n\tc.mutex.Unlock()\n\treturn !addFail\n}", "func (p *Pair) Set(fkey, fvalue string) bool {\n\tvar (\n\t\tidx int = -1\n\t)\n\n\tcomponents := strings.SplitN(p.Value, \",\", len(p.Format))\n\n\tif len(p.Format) == 0 {\n\t\treturn false\n\t}\n\tif len(components) != len(p.Format) {\n\t\treturn false\n\t}\n\n\tfor i, formatKey := range p.Format {\n\t\tif formatKey == fkey {\n\t\t\tidx = i\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif idx == -1 {\n\t\treturn false\n\t}\n\n\tif idx >= len(components) {\n\t\treturn false\n\t}\n\n\tcomponents[idx] = fvalue\n\n\tp.Value = strings.Join(components, \",\")\n\n\treturn true\n}", "func (s *Flaky) Set(k, v []byte) error {\n\tif s.fail() {\n\t\treturn s.err\n\t}\n\treturn s.Txn.Set(k, v)\n}", "func (s *StorageBase) Set(ctx context.Context, sessionId string, key string, value interface{}, ttl time.Duration) error {\n\treturn ErrorDisabled\n}", "func (m MockKeyring) Set(item keyring.Item) error {\n\treturn nil\n}", "func (bs *cachedFlatFsDataStore) Put(key string, value []byte) (err error) {\n\n\tbs.Lock()\n\tdefer bs.Unlock()\n\tif err = bs.dataStore.Put(key, value); err != nil {\n\t\tbs.cacheSet[key] = struct{}{}\n\t}\n\n\treturn err\n}", "func (s *TXPoolServer) setPendingTx(tx *tx.Transaction,\n\tsender tc.SenderType, txResultCh chan *tc.TxResult) bool {\n\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\tif ok := s.allPendingTxs[tx.Hash()]; ok != nil {\n\t\tlog.Debugf(\"setPendingTx: transaction %x already in the verifying process\",\n\t\t\ttx.Hash())\n\t\treturn false\n\t}\n\n\tpt := &serverPendingTx{\n\t\ttx: tx,\n\t\tsender: sender,\n\t\tch: txResultCh,\n\t}\n\n\ts.allPendingTxs[tx.Hash()] = pt\n\treturn true\n}", "func (mi *MasterIndex) addPending(id restic.ID, tpe restic.BlobType) bool {\n\n\tmi.idxMutex.Lock()\n\tdefer mi.idxMutex.Unlock()\n\n\t// Check if blob is pending or in index\n\tif mi.pendingBlobs.Has(restic.BlobHandle{ID: id, Type: tpe}) {\n\t\treturn false\n\t}\n\n\tfor _, idx := range mi.idx {\n\t\tif idx.Has(id, tpe) {\n\t\t\treturn false\n\t\t}\n\t}\n\n\t// really not known -> insert\n\tmi.pendingBlobs.Insert(restic.BlobHandle{ID: id, Type: tpe})\n\treturn true\n}", "func (s UserSet) SetShare(value bool) {\n\ts.RecordCollection.Set(models.NewFieldName(\"Share\", \"share\"), value)\n}", "func (m *mCryptographyServiceMockVerify) Set(f func(p crypto.PublicKey, p1 insolar.Signature, p2 []byte) (r bool)) *CryptographyServiceMock {\n\tm.mainExpectation = nil\n\tm.expectationSeries = nil\n\n\tm.mock.VerifyFunc = f\n\treturn m.mock\n}", "func (c *Conn) Set(key, value []byte) error {\n\tresponse := c.client.Cmd(cmdSet, key, value)\n\tif !isOK(response) {\n\t\treturn errx.Errorf(\"set command failed\")\n\t}\n\treturn nil\n}", "func (mmIssuePendingExecution *mDelegationTokenFactoryMockIssuePendingExecution) Set(f func(msg mm_insolar.Message, pulse mm_insolar.PulseNumber) (d1 mm_insolar.DelegationToken, err error)) *DelegationTokenFactoryMock {\n\tif mmIssuePendingExecution.defaultExpectation != nil {\n\t\tmmIssuePendingExecution.mock.t.Fatalf(\"Default expectation is already set for the DelegationTokenFactory.IssuePendingExecution method\")\n\t}\n\n\tif len(mmIssuePendingExecution.expectations) > 0 {\n\t\tmmIssuePendingExecution.mock.t.Fatalf(\"Some expectations are already set for the DelegationTokenFactory.IssuePendingExecution method\")\n\t}\n\n\tmmIssuePendingExecution.mock.funcIssuePendingExecution = f\n\treturn mmIssuePendingExecution.mock\n}", "func (https *FS) Has(ctx context.Context, path string) (bool, error) {\n\treturn false, nil\n}", "func (fr *FileRing) Contains(f File) bool {\n\tfr.lock.RLock()\n\tdefer fr.lock.RUnlock()\n\n\th := f.Id()\n\tif m, ok := fr.hashMap[h]; ok {\n\t\tfound, ok := m.versionMap[f.Sequence]\n\t\tif ok {\n\t\t\t// We have assumed till now that there are no hash cnoflicts. What if there is one?\n\t\t\t// Panic for now. Fix later :P\n\t\t\tif found.File != f.File {\n\t\t\t\tlog.Fatal(\"Hash conflict!\")\n\t\t\t}\n\t\t\tif found.AutoTombstone == false && f.AutoTombstone == true {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (p *peerContactsSet) put(peerContact string) bool {\n\tif len(peerContact) < 6 {\n\t\treturn false\n\t}\n\tif ok := p.set[peerContact]; ok {\n\t\treturn false\n\t}\n\tp.set[peerContact] = true\n\tr := &ring.Ring{Value: peerContact}\n\tif p.ring == nil {\n\t\tp.ring = r\n\t} else {\n\t\tp.ring.Link(r)\n\t}\n\treturn true\n}", "func (r *Connection) Set(key string, value []byte) error {\n\treturn r.cache.Set([]byte(key), value, r.expiry)\n}", "func (s *metadataSupplier) Set(key string, value string) {\n\ts.metadata.Set(key, value)\n}", "func (txn *Txn) Set(key, val []byte) error {\n\treturn txn.SetEntry(NewEntry(key, val))\n}", "func (mmEquals *mDigestHolderMockEquals) Set(f func(other DigestHolder) (b1 bool)) *DigestHolderMock {\n\tif mmEquals.defaultExpectation != nil {\n\t\tmmEquals.mock.t.Fatalf(\"Default expectation is already set for the DigestHolder.Equals method\")\n\t}\n\n\tif len(mmEquals.expectations) > 0 {\n\t\tmmEquals.mock.t.Fatalf(\"Some expectations are already set for the DigestHolder.Equals method\")\n\t}\n\n\tmmEquals.mock.funcEquals = f\n\treturn mmEquals.mock\n}", "func (lmem *lockedMemRepo) Put(name string, key types.KeyInfo) error {\n\tif err := lmem.checkToken(); err != nil {\n\t\treturn err\n\t}\n\tlmem.Lock()\n\tdefer lmem.Unlock()\n\n\t_, isThere := lmem.mem.keystore[name]\n\tif isThere {\n\t\treturn xerrors.Errorf(\"putting key '%s': %w\", name, types.ErrKeyExists)\n\t}\n\n\tlmem.mem.keystore[name] = key\n\treturn nil\n}", "func (c *concurrentStorage) add(u url.URL) (bool) {\n\tc.Lock()\n\tdefer c.Unlock()\n\tif _, ok := c.urls[u]; ok{\n\t\treturn false\n\t}\n\tc.urls[u] = true\n\tc.urlsSize++\n\treturn true\n}", "func (ts *TweetSearch) MarkTweetAsAlreadyTweeted(projectName string) (bool, error) {\n\tstorageConn := ts.Storage.Get()\n\tdefer storageConn.Close()\n\n\t// Generate score in format YYYYMMDDHHiiss\n\tnow := time.Now()\n\tscore := now.Format(\"20060102150405\")\n\n\tres, err := storageConn.MarkRepositoryAsTweeted(projectName, score)\n\tif err != nil || !res {\n\t\tlog.Printf(\"Adding project %s to tweeted list: ❌ s%s (%v)\\n\", projectName, err, res)\n\t}\n\n\treturn res, err\n}", "func (tsm *ThreadSafeMap) SetIfNotExists(key string, val interface{}) bool {\n\ttsm.Lock()\n\tdefer tsm.Unlock()\n\n\tif _, ok := tsm.items[key]; !ok {\n\t\ttsm.items[key] = val\n\t\treturn true\n\t}\n\treturn false\n}", "func (s *devStorage) setIfNot(imei string, dr devReq) bool {\n\ts.mux.Lock()\n\tdefer s.mux.Unlock()\n\tif _, ok := s.storage[imei]; ok {\n\t\treturn false\n\t}\n\ts.storage[imei] = dr\n\treturn true\n}", "func (a *Adapter) Set(key uint64, response []byte, expiration time.Time) {\n\ta.mutex.Lock()\n\tdefer a.mutex.Unlock()\n\n\tif _, ok := a.store[key]; ok {\n\t\t// Known key, overwrite previous item.\n\t\ta.store[key] = response\n\t\treturn\n\t}\n\n\t// New key, make sure we have the capacity.\n\tif len(a.store) == a.capacity {\n\t\ta.evict()\n\t}\n\n\ta.store[key] = response\n}", "func (b *BadgerDB) Set(key Key, value []byte) error {\n\tfullKey := append(key.Scope().Bytes(), key.ID()...)\n\n\terr := b.backend.Update(func(txn *badger.Txn) error {\n\t\treturn txn.Set(fullKey, value)\n\t})\n\n\treturn err\n}", "func (f *Flag) Set() { atomic.CompareAndSwapUint32((*uint32)(unsafe.Pointer(f)), 0, 1) }", "func (s *SyncStorage) Set(ns string, pairs ...interface{}) error {\n\tif len(pairs) == 0 {\n\t\treturn nil\n\t}\n\n\tkeyAndData, err := s.setNamespaceToKeys(getNsPrefix(ns), pairs...)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn s.getDbBackend(ns).MSet(keyAndData...)\n}", "func (c *FileSystemCache) Set(data []byte, expire time.Duration, key ...string) error {\n\tif len(key) < 1 {\n\t\treturn fmt.Errorf(\"no key specified\")\n\t}\n\tfolder := c.keyPath(key[:len(key)-1]...)\n\t_, err := os.Stat(folder)\n\tif os.IsNotExist(err) {\n\t\tos.MkdirAll(folder, 0700)\n\t}\n\tif expire > 0 {\n\t\texpireTS := time.Now().Add(expire).UnixNano()\n\t\terr := ioutil.WriteFile(c.expirePath(key...), []byte(strconv.FormatInt(expireTS, 10)), 0600)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn ioutil.WriteFile(c.keyPath(key...), data, 0600)\n}", "func (c *DetailsDB) Set(key, value []byte) error {\n\terr := c.db.Update(func(tx *bolt.Tx) error {\n\t\tbucket, err := tx.CreateBucketIfNotExists(c.bucket)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = bucket.Put(key, value)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n\n\treturn err\n}", "func (s Set) Add(st string) {\n\tif _, ok := s[st]; !ok {\n\t\ts[st] = true\n\t}\n}", "func (s *Set) Add(str string) bool {\n\tif s.Exist(str) {\n\t\treturn false\n\t}\n\ts.m[str] = struct{}{}\n\treturn false\n}", "func (l *ShardedNoTS) Set(tenantID, key string, val interface{}) error {\n\t_, ok := l.cache[tenantID]\n\tif !ok {\n\t\tl.cache[tenantID] = l.constructor()\n\t\tl.itemCount[tenantID] = 0\n\t}\n\n\tl.itemCount[tenantID]++\n\treturn l.cache[tenantID].Set(key, val)\n}", "func (s *Storage) Set(key string, value interface{}) {\n\ts.mu.Lock()\n\ts.data[key] = value\n\tctx, cancel := context.WithTimeout(context.Background(), s.timeout)\n\ts.cancelFuncs[key] = cancel\n\ts.mu.Unlock()\n\tgo func(ctx context.Context) {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\ts.mu.Lock()\n\t\t\tdelete(s.data, key)\n\t\t\tdelete(s.cancelFuncs, key)\n\t\t\ts.mu.Unlock()\n\t\t}\n\t}(ctx)\n}", "func (ss Set) Put(k string) {\n\tss[k] = true\n}", "func (self *Store) Set(k string) []byte {\n\tself.mu.Lock()\n\tself.mu.Unlock()\n\treturn nil\n}", "func (_m *DataStoreTxn) Set(key []byte, value []byte) error {\n\tret := _m.Called(key, value)\n\n\tvar r0 error\n\tif rf, ok := ret.Get(0).(func([]byte, []byte) error); ok {\n\t\tr0 = rf(key, value)\n\t} else {\n\t\tr0 = ret.Error(0)\n\t}\n\n\treturn r0\n}", "func (h *hashDisk) Set(value []byte, fileIndex, fileOffset uint32) error {\n\tif bytes.Equal(value, h.emptyValue) {\n\t\treturn ErrInvalidKey\n\t}\n\tif h.totalEntries >= h.MaxSize {\n\t\treturn ErrNoSpace\n\t}\n\tnewEntry := true\n\t// Compute hash\n\tslot := hyperloglog.MurmurBytes(value) % h.entries\n\toffset := slot * h.entrySize\n\tfor { // Try to find an empty slot\n\t\tslotValue := h.m[offset : offset+keySize]\n\t\tif bytes.Equal(slotValue, value) {\n\t\t\t// Found same key, override. We could just return instead but it was found in\n\t\t\t// benchmarks that it hardly change anything at all so it's better to be able to override\n\t\t\tnewEntry = false\n\t\t\tbreak\n\t\t}\n\t\tif bytes.Equal(slotValue, h.emptyValue) {\n\t\t\t// Found empty slot\n\t\t\tbreak\n\t\t}\n\t\tslot = (slot + 1) % h.entries\n\t\toffset = slot * h.entrySize\n\t}\n\t// Insert\n\tindexes := make([]byte, 4+4)\n\tencoding.PutUint32(indexes[0:4], fileIndex)\n\tencoding.PutUint32(indexes[4:8], fileOffset)\n\tcopy(h.m[offset:offset+keySize], value)\n\tcopy(h.m[offset+keySize:offset+keySize+8], indexes)\n\tif newEntry {\n\t\th.totalEntries++\n\t}\n\treturn nil\n}", "func (f File) Set(groupName string, group HostGroup) {\n\tf[groupName] = group\n}", "func (that *StrAnyMap) SetIfNotExist(key string, value interface{}) bool {\n\tif !that.Contains(key) {\n\t\tthat.doSetWithLockCheck(key, value)\n\t\treturn true\n\t}\n\treturn false\n}", "func (o *PtNaturalBcs) Set(key string, nod *Node, fcn fun.Func, extra string) (setisok bool) {\n\td := nod.GetDof(key)\n\tif d == nil { // handle LBB nodes\n\t\treturn\n\t}\n\tif idx, ok := o.Eq2idx[d.Eq]; ok {\n\t\to.Bcs[idx].Key = \"f\" + key\n\t\to.Bcs[idx].Eq = d.Eq\n\t\to.Bcs[idx].X = nod.Vert.C\n\t\to.Bcs[idx].Fcn = fcn\n\t\to.Bcs[idx].Extra = extra\n\t} else {\n\t\to.Eq2idx[d.Eq] = len(o.Bcs)\n\t\to.Bcs = append(o.Bcs, &PtNaturalBc{\"f\" + key, d.Eq, nod.Vert.C, fcn, extra})\n\t}\n\treturn true\n}", "func (c *cache) Set(id types.UID, status *PodStatus, err error, timestamp time.Time) (updated bool) {\n\tc.lock.Lock()\n\tdefer c.lock.Unlock()\n\n\tif utilfeature.DefaultFeatureGate.Enabled(features.EventedPLEG) {\n\t\t// Set the value in the cache only if it's not present already\n\t\t// or the timestamp in the cache is older than the current update timestamp\n\t\tif cachedVal, ok := c.pods[id]; ok && cachedVal.modified.After(timestamp) {\n\t\t\treturn false\n\t\t}\n\t}\n\n\tc.pods[id] = &data{status: status, err: err, modified: timestamp}\n\tc.notify(id, timestamp)\n\treturn true\n}", "func (o *FileversionFileversion) HasPendingFileRef() bool {\n\tif o != nil && o.PendingFileRef != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (q *UniqueQueue) Enqueue(key string, val interface{}) bool {\n\tq.mu.Lock()\n\tdefer q.mu.Unlock()\n\n\tif q.doneChanClosed {\n\t\treturn false\n\t}\n\n\t// Key is already in the queue. Update the set's entry and\n\t// return without modifying its place in the queue.\n\tif entry, ok := q.queuedSet[key]; ok {\n\t\tentry.val = val\n\t\treturn true\n\t}\n\n\tif q.full() {\n\t\treturn false\n\t}\n\n\tq.queue[q.tail] = entry{\n\t\tkey: key,\n\t\tval: val,\n\t}\n\tq.queuedSet[key] = &q.queue[q.tail]\n\tq.tail = q.inc(q.tail)\n\n\tq.trySchedule()\n\n\treturn true\n}", "func (userdata *User) ShareFile(filename string, recipient string) (\r\n\taccessToken uuid.UUID, err error) {\r\n\tvar shareInvite ShareInvite\r\n\tvar fileKeyMeta FileKeyMeta\r\n\t//IMPLEMENT SHAREDUSERS AND USERTOKENS in FILEKEYSTRUCT\r\n\r\n\t//check is file exists in users file space\r\n\tupdatedUser, err := GetLatestUser(userdata.UUID_, userdata.PasswordHash, userdata.HMACKey, userdata.EncKey)\r\n\tif err != nil {\r\n\t\treturn uuid.Nil, errors.New(\"Failed to retrieve latest user info.\")\r\n\t}\r\n\r\n\tvar fk FileKey\r\n\tfileKey, fileFound := updatedUser.Filespace[filename]\r\n\r\n\tif !fileFound {\r\n\t\treturn uuid.Nil, errors.New(\"File does not exist in caller's personal filespace.\")\r\n\t}\r\n\r\n\t//what if this person is sharing a shared file? check access token!\r\n\t_, own := updatedUser.FilesOwned[filename]\r\n\tif !own { //this is just in case a revocation has happened and the fileKey data has changed\r\n\t\tat := userdata.AccessTokens[filename]\r\n\t\tfk, err = updatedUser.RetrieveAccessToken(at)\r\n\t\tif err != nil {\r\n\t\t\treturn uuid.Nil, errors.New(\"Failed to retrieve access token.\")\r\n\t\t}\r\n\t} else { //just in case fileKey data has changed from a recent session\r\n\t\tcurrFileKey, _ := userlib.DatastoreGet(fileKey.KeyId)\r\n\t\tlen_data := len(currFileKey) - userlib.HashSizeBytes\r\n\r\n\t\tif len_data < 0 || len_data > len(currFileKey) || len(currFileKey[:len_data]) < userlib.HashSizeBytes {\r\n\t\t\t//automatically return error, file has been changed\r\n\t\t\treturn uuid.Nil, errors.New(\"FileKey data length has changed.\")\r\n\t\t}\r\n\t\t//verify integrity of both fileKey struct and the file itself\r\n\t\tcomputedMac, _ := userlib.HMACEval(fileKey.HMAC_key, currFileKey[:len_data])\r\n\t\tif !userlib.HMACEqual(computedMac, currFileKey[len_data:]) {\r\n\t\t\treturn uuid.Nil, errors.New(\"File key struct has been tampered with in Datastore.\")\r\n\t\t}\r\n\t\t//decrypt + depad fileKey from DS to current fileKey var (overwrite)\r\n\t\tdecrypt := userlib.SymDec(fileKey.Enc_key, currFileKey[:len_data])\r\n\t\tdecrypt = PKCS(decrypt, \"remove\")\r\n\r\n\t\terr = json.Unmarshal(decrypt, &fk)\r\n\t\tif err != nil {\r\n\t\t\treturn uuid.Nil, errors.New(\"Error demarshaling.\")\r\n\t\t}\r\n\t}\r\n\r\n\t//check if recipient exists\r\n\tpubKey, userFound := userlib.KeystoreGet(recipient + \"public_key\")\r\n\tif !userFound {\r\n\t\treturn uuid.Nil, errors.New(\"Recepient not found.\")\r\n\t}\r\n\r\n\t//populate Shareinvite and FileKeyMeta struct\r\n\tfileKeyMeta.DSid = fk.KeyId\r\n\tfileKeyMeta.HMACkey = fk.HMAC_key\r\n\tfileKeyMeta.ENCkey = fk.Enc_key\r\n\r\n\tfkm_json, _ := json.Marshal(fileKeyMeta)\r\n\t//encrypt FileKeyMeta using RSA\r\n\tfileKeyMeta_enc, _ := userlib.PKEEnc(pubKey, fkm_json) //dont need to pad?\r\n\r\n\t//Marshal the fileKeyMeta info\r\n\tshareInvite.RSAFileKey = fileKeyMeta_enc\r\n\t//msg for signature is the RSA encrypted, MARSHALED FileMetaKey struct\r\n\tshareInvite.Signature, _ = userlib.DSSign(userdata.PrivSignKey, shareInvite.RSAFileKey)\r\n\tshareInvite.Sender = updatedUser.Username\r\n\tshareInvite_json, _ := json.Marshal(shareInvite)\r\n\r\n\taccessToken = uuid.New() //generate random accessToken\r\n\tuserlib.DatastoreSet(accessToken, shareInvite_json)\r\n\r\n\t//update SharedUsers and UserTokens fields in fileKey\r\n\t//if they are a DIRECT sharer (meaning theyre one level below owner) generate key\r\n\t//and put recipient in their list\r\n\t//else, put them under the direct sharer from their lineage\r\n\tfk.UserTokens[recipient] = accessToken //update user tokens for possible revoking later\r\n\tlist, direct_user := fk.SharedUsers[updatedUser.Username] //should return true if user is a key\r\n\tif own {\r\n\t\tvar emptyList []string\r\n\t\tfk.SharedUsers[recipient] = emptyList\r\n\t} else if direct_user {\r\n\t\tlist = append(list, recipient) //add recipient to direct sharer's list\r\n\t\tfk.SharedUsers[recipient] = list\r\n\t} else { //indirect user case, iterate over map of shared users\r\n\t\tvar originalSharer string\r\n\t\t//provides each key and respective value\r\n\t\tfor directSharer, listShared := range fk.SharedUsers {\r\n\t\t\tfor _, indirectSharer := range listShared {\r\n\t\t\t\tif updatedUser.Username == indirectSharer {\r\n\t\t\t\t\toriginalSharer = directSharer\r\n\t\t\t\t\tbreak\r\n\t\t\t\t}\r\n\t\t\t}\r\n\t\t\tif originalSharer != \"\" {\r\n\t\t\t\tbreak //break out of second for loop if you've found the original sharer\r\n\t\t\t}\r\n\t\t}\r\n\t\tif originalSharer == \"\" {\r\n\t\t\treturn uuid.Nil, errors.New(\"User is not owner but could not find who shared file with them.\")\r\n\t\t}\r\n\t\tlist = append(list, recipient)\r\n\t\tfk.SharedUsers[originalSharer] = list //add the recipient of this indirect user to list\r\n\t}\r\n\t//Now lets send the updated fileKey to the Datastore\r\n\tfileKey_json, _ := json.Marshal(fk)\r\n\tfk_IV := userlib.RandomBytes(userlib.AESBlockSizeBytes)\r\n\tfileKey_enc := userlib.SymEnc(fk.Enc_key, fk_IV, PKCS(fileKey_json, \"add\"))\r\n\t//Add HMACs for both file key and file elem struct (runtime corresponds to size of appended file, nothing else)\r\n\tfk_hmac, _ := userlib.HMACEval(fk.HMAC_key, fileKey_enc)\r\n\tfileKey_enc = append(fileKey_enc, fk_hmac...)\r\n\tuserlib.DatastoreSet(fileKey.KeyId, fileKey_enc)\r\n\r\n\treturn accessToken, nil\r\n}", "func (s BlobSet) Has(h BlobHandle) bool {\n\t_, ok := s[h]\n\treturn ok\n}", "func (self *SafeMap) Set(k interface{}, v interface{}) bool {\n\tself.lock.Lock()\n\tdefer self.lock.Unlock()\n\n\tif val, ok := self.sm[k]; !ok {\n\t\tself.sm[k] = v\n\t} else if val != v {\n\t\tself.sm[k] = v\n\t} else {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (b *Bloom) AddExisted(value string) bool {\n\texisted := true\n\tfor i := 0; i < len(b.keys); i++ {\n\t\thash := SipHash24([]byte(value), b.keys[i]) % b.size\n\t\tif !b.get(hash) {\n\t\t\texisted = false\n\t\t}\n\t\tb.set(hash)\n\t}\n\treturn existed\n}", "func ShareFiles(w http.ResponseWriter, r *http.Request) {\n\t//log\n\tnow, userIP := globalPkg.SetLogObj(r)\n\tlogobj := logpkg.LogStruct{\"_\", now, userIP, \"macAdress\", \"ShareFiles\", \"file\", \"_\", \"_\", \"_\", 0}\n\tvar ShareFiledataObj ShareFiledata\n\tdecoder := json.NewDecoder(r.Body)\n\tdecoder.DisallowUnknownFields()\n\terr := decoder.Decode(&ShareFiledataObj)\n\tif err != nil {\n\t\tglobalPkg.SendError(w, \"please enter your correct request\")\n\t\tglobalPkg.WriteLog(logobj, \"please enter your correct request\", \"failed\")\n\t\treturn\n\t}\n\ttime.Sleep(time.Millisecond * 10) // for handle unknown issue\n\taccountObj := account.GetAccountByAccountPubicKey(ShareFiledataObj.Publickey)\n\tif accountObj.AccountPublicKey != ShareFiledataObj.Publickey {\n\t\tglobalPkg.SendError(w, \"error in public key\")\n\t\tglobalPkg.WriteLog(logobj, \"error in public key\", \"failed\")\n\t\treturn\n\t}\n\tif accountObj.AccountPassword != ShareFiledataObj.Password {\n\t\tglobalPkg.SendError(w, \"error in password\")\n\t\tglobalPkg.WriteLog(logobj, \"error in password\", \"failed\")\n\t\treturn\n\t}\n\t// check user own this file id\n\tfiles := accountObj.Filelist\n\tfound := false\n\tfor _, fileObj := range files {\n\t\tif fileObj.Fileid == ShareFiledataObj.FileID {\n\t\t\tfound = true\n\t\t}\n\t}\n\tif !found {\n\t\tglobalPkg.SendError(w, \"You don't have this file\")\n\t\tglobalPkg.WriteLog(logobj, \"You don't have this file\", \"failed\")\n\t\treturn\n\t}\n\t// check pk already exist in blockchain\n\taccountList := accountdb.GetAllAccounts()\n\tfor _, pk := range ShareFiledataObj.PermissionPkList {\n\t\tif !containspk(accountList, pk) {\n\t\t\tglobalPkg.SendError(w, \"this public key is not associated with any account\")\n\t\t\tglobalPkg.WriteLog(logobj, \"You don't have this file\", \"failed\")\n\t\t\treturn\n\t\t}\n\t}\n\t// Signture string\n\tvalidSig := false\n\tpk1 := account.FindpkByAddress(accountObj.AccountPublicKey).Publickey\n\tif pk1 != \"\" {\n\t\tpublickey1 := cryptogrpghy.ParsePEMtoRSApublicKey(pk1)\n\t\tstrpermissionlist := strings.Join(ShareFiledataObj.PermissionPkList, \"\")\n\t\tfmt.Println(\"strpermissionlist : \", strpermissionlist)\n\t\tsignatureData := strpermissionlist + ShareFiledataObj.FileID + ShareFiledataObj.Publickey\n\t\tvalidSig = cryptogrpghy.VerifyPKCS1v15(ShareFiledataObj.Signture, signatureData, *publickey1)\n\t} else {\n\t\tvalidSig = false\n\t}\n\t// validSig = true\n\tif !validSig {\n\t\tglobalPkg.SendError(w, \"you are not allowed to share file\")\n\t\tglobalPkg.WriteLog(logobj, \"you are not allowed to share file\", \"failed\")\n\t\treturn\n\t}\n\t//\n\n\tfilelistOwner := accountObj.Filelist\n\t// add account index see file , ownerpk , fileid\n\t//append share file id , ownerpk to account index want to share file to you\n\tfor _, pk := range ShareFiledataObj.PermissionPkList {\n\t\tvar sharedfileObj filestorage.SharedFile\n\t\tvar ownerfileObj filestorage.OwnersharedFile\n\t\tvar ownerfileObj2 filestorage.OwnersharedFile\n\t\tvar foundOwnerpk bool\n\t\taccountind := account.GetAccountByAccountPubicKey(pk)\n\t\tsharedfileObj.AccountIndex = accountind.AccountIndex\n\t\townedsharefile := filestorage.FindSharedfileByAccountIndex(sharedfileObj.AccountIndex)\n\t\tif pk != ShareFiledataObj.Publickey { //same owner share to himself\n\t\t\tif len(ownedsharefile.OwnerSharefile) != 0 {\n\t\t\t\tfor _, ownedsharefileObj := range ownedsharefile.OwnerSharefile {\n\n\t\t\t\t\tif ownedsharefileObj.OwnerPublicKey == ShareFiledataObj.Publickey {\n\t\t\t\t\t\tfoundOwnerpk = true\n\t\t\t\t\t\tif !containsfileid(ownedsharefileObj.Fileid, ShareFiledataObj.FileID) {\n\t\t\t\t\t\t\townedsharefileObj.Fileid = append(ownedsharefileObj.Fileid, ShareFiledataObj.FileID)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\tsharedfileObj.OwnerSharefile = append(sharedfileObj.OwnerSharefile, ownedsharefileObj)\n\t\t\t\t}\n\t\t\t\tif !foundOwnerpk {\n\t\t\t\t\townerfileObj2.OwnerPublicKey = ShareFiledataObj.Publickey\n\t\t\t\t\townerfileObj2.Fileid = append(ownerfileObj2.Fileid, ShareFiledataObj.FileID)\n\t\t\t\t\tsharedfileObj.OwnerSharefile = append(sharedfileObj.OwnerSharefile, ownerfileObj2)\n\t\t\t\t}\n\n\t\t\t} else {\n\t\t\t\townerfileObj.OwnerPublicKey = ShareFiledataObj.Publickey\n\t\t\t\townerfileObj.Fileid = append(ownerfileObj.Fileid, ShareFiledataObj.FileID)\n\t\t\t\tsharedfileObj.OwnerSharefile = append(sharedfileObj.OwnerSharefile, ownerfileObj)\n\t\t\t}\n\t\t\tbroadcastTcp.BoardcastingTCP(sharedfileObj, \"sharefile\", \"file\")\n\t\t\t//append permisssionlist to account owner filelist\n\t\t\tfor m := range filelistOwner {\n\t\t\t\tif filelistOwner[m].Fileid == ShareFiledataObj.FileID {\n\t\t\t\t\tif !containsfileid(filelistOwner[m].PermissionList, pk) {\n\t\t\t\t\t\tfilelistOwner[m].PermissionList = append(filelistOwner[m].PermissionList, pk)\n\t\t\t\t\t}\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\taccountObj.Filelist = filelistOwner\n\tbroadcastTcp.BoardcastingTCP(accountObj, \"updateaccountFilelist\", \"file\")\n\n\tglobalPkg.SendResponseMessage(w, \"you shared file successfully\")\n\tglobalPkg.WriteLog(logobj, \"you shared file successfully\", \"success\")\n}", "func (s *OrderedRequestBodies) Set(key string, val *RequestBody) bool {\n\treturn s.data.Set(key, val)\n}" ]
[ "0.56819904", "0.5656244", "0.56193894", "0.5489106", "0.547759", "0.5415303", "0.5358624", "0.5329123", "0.5298075", "0.5251422", "0.5236577", "0.518625", "0.51822776", "0.51691777", "0.51528174", "0.5129096", "0.5112244", "0.5102831", "0.5098065", "0.5075395", "0.5075193", "0.5069218", "0.50587875", "0.50115156", "0.50058746", "0.49935445", "0.49846864", "0.4969345", "0.4963284", "0.4958479", "0.49434528", "0.49427998", "0.49327493", "0.49325326", "0.49287853", "0.4912423", "0.49069577", "0.48915246", "0.48909193", "0.48884398", "0.48815522", "0.48809186", "0.48790106", "0.4873914", "0.48673648", "0.48570365", "0.48534843", "0.48378074", "0.48351842", "0.48254824", "0.48254266", "0.48142982", "0.4812445", "0.48069984", "0.48056346", "0.48041052", "0.47874373", "0.47744697", "0.4773371", "0.47695345", "0.4761128", "0.47582212", "0.47570246", "0.47510725", "0.47437623", "0.47336155", "0.47332895", "0.47312617", "0.47295195", "0.47254533", "0.47162998", "0.4701438", "0.46996418", "0.4693732", "0.46876216", "0.46843454", "0.46841013", "0.46839523", "0.4683913", "0.46813342", "0.46796092", "0.4678921", "0.46763113", "0.4664546", "0.46639284", "0.46620402", "0.46354783", "0.46354", "0.46339247", "0.46329728", "0.463194", "0.4631577", "0.4631392", "0.4629464", "0.46242055", "0.4623186", "0.46225983", "0.4618754", "0.46157706", "0.4614538" ]
0.85217196
0
Deprecated: use Default instead.
func New(t LoggerReporter, baseURL string) *Expect { return WithConfig(Config{ BaseURL: baseURL, Reporter: NewAssertReporter(t), Printers: []Printer{ NewCompactPrinter(t), }, }) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (in *ManagedCluster) Default() {\n}", "func (f Base) Default() interface{} {\n\treturn f.defaultValue\n}", "func (t TestDescription) Default() TestDescription {\n\treturn t.newLabel(\"DEFAULT\")\n}", "func Default() *API {\n\tapi := New()\n\tapi.Use(\n\t\ttimeout.Default(\"12s\"),\n\t\tlogging.Default(\"TONIC\"))\n\n\t// adding default encoders..\n\tapi.Encoder(\n\t\tencoding.JSON,\n\t\tencoding.XML,\n\t\tencoding.YAML,\n\t\tencoding.FORM)\n\n\treturn api\n}", "func (i *Internationalization) Default() language.Tag { return i.fallback }", "func (*Namespace) Default(defaultv any, givenv ...any) (any, error) {\n\t// given is variadic because the following construct will not pass a piped\n\t// argument when the key is missing: {{ index . \"key\" | default \"foo\" }}\n\t// The Go template will complain that we got 1 argument when we expected 2.\n\n\tif len(givenv) == 0 {\n\t\treturn defaultv, nil\n\t}\n\tif len(givenv) != 1 {\n\t\treturn nil, fmt.Errorf(\"wrong number of args for default: want 2 got %d\", len(givenv)+1)\n\t}\n\n\tg := reflect.ValueOf(givenv[0])\n\tif !g.IsValid() {\n\t\treturn defaultv, nil\n\t}\n\n\tset := false\n\n\tswitch g.Kind() {\n\tcase reflect.Bool:\n\t\tset = true\n\tcase reflect.String, reflect.Array, reflect.Slice, reflect.Map:\n\t\tset = g.Len() != 0\n\tcase reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:\n\t\tset = g.Int() != 0\n\tcase reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr:\n\t\tset = g.Uint() != 0\n\tcase reflect.Float32, reflect.Float64:\n\t\tset = g.Float() != 0\n\tcase reflect.Complex64, reflect.Complex128:\n\t\tset = g.Complex() != 0\n\tcase reflect.Struct:\n\t\tswitch actual := givenv[0].(type) {\n\t\tcase time.Time:\n\t\t\tset = !actual.IsZero()\n\t\tdefault:\n\t\t\tset = true\n\t\t}\n\tdefault:\n\t\tset = !g.IsNil()\n\t}\n\n\tif set {\n\t\treturn givenv[0], nil\n\t}\n\n\treturn defaultv, nil\n}", "func Default(s string) string {\n\treturn s\n}", "func (r *KeystoneAPI) Default() {\n\tkeystoneapilog.Info(\"default\", \"name\", r.Name)\n\n\tr.Spec.Default()\n}", "func (p *param) Default() []byte {\n\treturn []byte(`{\"descr\":\"Normal network L0 server\",\"prefix\":[\"tg001\"]}`)\n}", "func Default(u []*url.URL, s Handler, verbose bool) *GGet {\n\treturn &GGet{URLS: u, Strategy: s, r: &httpRetriever{}, Verbose: verbose}\n}", "func (t Term) Default(args ...interface{}) Term {\n\treturn constructMethodTerm(t, \"Default\", p.Term_DEFAULT, args, map[string]interface{}{})\n}", "func Default(ctx context.Context, t *mold.Transformer, v reflect.Value, param string) error {\n\tif !v.IsZero() {\n\t\treturn nil\n\t}\n\n\tswitch v.Interface().(type) {\n\tcase string:\n\t\tv.SetString(param)\n\tcase int, int8, int16, int32, int64:\n\t\tvalue, err := strconv.Atoi(param)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tv.SetInt(int64(value))\n\n\tcase uint, uint8, uint16, uint32, uint64:\n\t\tvalue, err := strconv.Atoi(param)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tv.SetUint(uint64(value))\n\tcase float32, float64:\n\t\tvalue, err := strconv.ParseFloat(param, 64)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tv.SetFloat(value)\n\tcase bool:\n\t\tvalue, err := strconv.ParseBool(param)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tv.SetBool(value)\n\tcase time.Duration:\n\t\td, err := time.ParseDuration(param)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tv.SetInt(int64(d))\n\t}\n\treturn nil\n}", "func (r *OpenStackCluster) Default() {\n\tif r.Spec.IdentityRef != nil && r.Spec.IdentityRef.Kind == \"\" {\n\t\tr.Spec.IdentityRef.Kind = defaultIdentityRefKind\n\t}\n}", "func (ts *TestSrv) Default() {\n\tts.Title = kTsTitle\n\tts.GameServer.Default()\n\tts.AcctServer.Http = 8000\n\tts.AcctServer.Https = 8080\n\tts.AcctServer.Rpc = 5001\n\tts.Game.Default(\"game\")\n\tts.Acct.Default(\"acct\")\n}", "func (r *Friend) Default() {\n\tfriendlog.Info(\"default\", \"name\", r.Name)\n\n\t// TODO(user): fill in your defaulting logic.\n}", "func newDefault(compilers Compilers) *Context {\n\tc := &Context{compilers, options{}}\n\t_ = defaultOptions(c)\n\treturn c\n}", "func (r *Unit) Default() {\n\tunitlog.Info(\"default\", \"name\", r.Name)\n\n\t// TODO(user): fill in your defaulting logic.\n\t// 这里可以加入一些Unit 结构体对象初始化之前的一些默认的逻辑,比如给一些字段填充默认值\n\n\t// default replicas set to 1\n\tunitlog.Info(\"default\", \"name\", r.Name)\n\n\tif r.Spec.Replicas == nil {\n\t\tdefaultReplicas := int32(1)\n\t\tr.Spec.Replicas = &defaultReplicas\n\t}\n\n\t// add default selector label\n\tlabelMap := make(map[string]string, 1)\n\tlabelMap[\"app\"] = r.Name\n\tr.Spec.Selector = &metav1.LabelSelector{\n\t\tMatchLabels: labelMap,\n\t}\n\n\t// add default template label\n\tr.Spec.Template.Labels = labelMap\n\n\tr.Status.LastUpdateTime = metav1.Now()\n\n\t// 当然,还可以根据需求加一些适合在初始化时做的逻辑,例如为pod注入sidecar\n\t// ...\n\n}", "func (s *defaultServicesrvc) Default(ctx context.Context, p *defaultservice.DefaultPayload) (err error) {\n\ts.logger.Print(\"defaultService.default\")\n\treturn\n}", "func (cluster *Cluster) Default() {\n\tklog.Info(\"default\", \"name\", cluster.Name)\n\tcluster.addFinalizer()\n\tcluster.defaultKubernetesVersion()\n\tcluster.defaultControlPlaneReplicas()\n\tcluster.defaultVPN()\n\tcluster.defaultUninitializedCertificatesLabel()\n\tcluster.defaultNetworking()\n}", "func (e Environment) Default(name string, a ...interface{}) {\n\te[fmt.Sprintf(\"%s.default\", name)] = fmt.Sprint(a...)\n}", "func (obj *Response) Default() {\n\tobj.Program = \"Beq\"\n\tobj.Version = \"0.01\"\n}", "func (c *ConfHolder) Default() []byte {\n\treturn nil\n}", "func DefaultUsage() {\n\tfmt.Println(\"Usage\")\n\tPrintDefaults()\n}", "func (machine *VirtualMachine) Default() {\n\tmachine.defaultImpl()\n\tvar temp any = machine\n\tif runtimeDefaulter, ok := temp.(genruntime.Defaulter); ok {\n\t\truntimeDefaulter.CustomDefault()\n\t}\n}", "func (client *BaseClient) Default(str string, defaultStr string) string {\n\tif str == \"\" {\n\t\treturn defaultStr\n\t}\n\treturn str\n}", "func Default() *Root {\n\tdef := Root{\n\t\tAPI: API{\n\t\t\tListenAddress: \"/ip6/::1/tcp/1234/http\",\n\t\t\tTimeout: Duration(30 * time.Second),\n\t\t},\n\t\tLibp2p: Libp2p{\n\t\t\tListenAddresses: []string{\n\t\t\t\t\"/ip4/0.0.0.0/tcp/0\",\n\t\t\t\t\"/ip6/::/tcp/0\",\n\t\t\t},\n\t\t},\n\t}\n\treturn &def\n}", "func Default(name, instance string) {\n\tglobalsMu.Lock()\n\tdefer globalsMu.Unlock()\n\tif _, found := globals[name]; found {\n\t\tpanic(\"config.Default: default \" + name + \" has same name as a global\")\n\t}\n\tif _, found := globals[instance]; !found {\n\t\tif _, found = defaults[instance]; !found {\n\t\t\tpanic(\"config.Default: instance \" + instance + \" does not exist\")\n\t\t}\n\t}\n\tdefaults[name] = instance\n}", "func (database *SqlDatabase) Default() {\n\tdatabase.defaultImpl()\n\tvar temp any = database\n\tif runtimeDefaulter, ok := temp.(genruntime.Defaulter); ok {\n\t\truntimeDefaulter.CustomDefault()\n\t}\n}", "func (store *ConfigurationStore) Default() {\n\tstore.defaultImpl()\n\tvar temp any = store\n\tif runtimeDefaulter, ok := temp.(genruntime.Defaulter); ok {\n\t\truntimeDefaulter.CustomDefault()\n\t}\n}", "func Default(defaultValue interface{}) MethodParamOption {\r\n\treturn func(p *MethodParam) {\r\n\t\tif defaultValue != nil {\r\n\t\t\tp.defaultValue = fmt.Sprint(defaultValue)\r\n\t\t}\r\n\t}\r\n}", "func Default() Logger {\n\tprotect.Do(func() {\n\t\t// If SetDefault was used before Default() was called, we need to\n\t\t// detect that here.\n\t\tif def == nil {\n\t\t\tdef = New(DefaultOptions)\n\t\t}\n\t})\n\n\treturn def\n}", "func Default() Logger {\n\tprotect.Do(func() {\n\t\t// If SetDefault was used before Default() was called, we need to\n\t\t// detect that here.\n\t\tif def == nil {\n\t\t\tdef = New(DefaultOptions)\n\t\t}\n\t})\n\n\treturn def\n}", "func (c *KubeadmConfig) Default() {\n\tDefaultKubeadmConfigSpec(&c.Spec)\n}", "func (*Colors) Default(value string) string {\n\treturn Color(\"default\", value)\n}", "func (credential *FederatedIdentityCredential) Default() {\n\tcredential.defaultImpl()\n\tvar temp any = credential\n\tif runtimeDefaulter, ok := temp.(genruntime.Defaulter); ok {\n\t\truntimeDefaulter.CustomDefault()\n\t}\n}", "func Default() []byte {\n\tb, err := buildDefault()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn b\n}", "func configDefault(config ...Config) Config {\n\t// Return default config if nothing provided\n\tif len(config) < 1 {\n\t\treturn ConfigDefault\n\t}\n\n\t// Override default config\n\tcfg := config[0]\n\n\t// Set default values\n\tif int(cfg.Duration.Seconds()) > 0 {\n\t\tlog.Warn(\"[LIMITER] Duration is deprecated, please use Expiration\")\n\t\tcfg.Expiration = cfg.Duration\n\t}\n\tif cfg.Key != nil {\n\t\tlog.Warn(\"[LIMITER] Key is deprecated, please us KeyGenerator\")\n\t\tcfg.KeyGenerator = cfg.Key\n\t}\n\tif cfg.Store != nil {\n\t\tlog.Warn(\"[LIMITER] Store is deprecated, please use Storage\")\n\t\tcfg.Storage = cfg.Store\n\t}\n\tif cfg.Next == nil {\n\t\tcfg.Next = ConfigDefault.Next\n\t}\n\tif cfg.Max <= 0 {\n\t\tcfg.Max = ConfigDefault.Max\n\t}\n\tif int(cfg.Expiration.Seconds()) <= 0 {\n\t\tcfg.Expiration = ConfigDefault.Expiration\n\t}\n\tif cfg.KeyGenerator == nil {\n\t\tcfg.KeyGenerator = ConfigDefault.KeyGenerator\n\t}\n\tif cfg.LimitReached == nil {\n\t\tcfg.LimitReached = ConfigDefault.LimitReached\n\t}\n\tif cfg.LimiterMiddleware == nil {\n\t\tcfg.LimiterMiddleware = ConfigDefault.LimiterMiddleware\n\t}\n\treturn cfg\n}", "func (d Data) Default(key string, def interface{}) interface{} {\n\tval, ok := d[key]\n\tif ok {\n\t\treturn val\n\t}\n\n\treturn def\n}", "func (ruleset *DnsForwardingRuleset) Default() {\n\truleset.defaultImpl()\n\tvar temp any = ruleset\n\tif runtimeDefaulter, ok := temp.(genruntime.Defaulter); ok {\n\t\truntimeDefaulter.CustomDefault()\n\t}\n}", "func (subscription *NamespacesTopicsSubscription) Default() {\n\tsubscription.defaultImpl()\n\tvar temp any = subscription\n\tif runtimeDefaulter, ok := temp.(genruntime.Defaulter); ok {\n\t\truntimeDefaulter.CustomDefault()\n\t}\n}", "func Default() *Baa {\n\treturn Instance(defaultAppName)\n}", "func (config *DeleteConfig) Default() {\r\n\tconfig.CaseName = \"*\" // default: delete from all cases\r\n\tconfig.Endpoint = \"\" // default detele single endpoint\r\n\tconfig.EndpointList = nil // Used if need to delete more than one endpoint\r\n}", "func Default() *Engine {\n\tengine := New()\n\tengine.Use(Logger(), Recovery())\n\treturn engine\n}", "func (record *PrivateDnsZonesSRVRecord) Default() {\n\trecord.defaultImpl()\n\tvar temp any = record\n\tif runtimeDefaulter, ok := temp.(genruntime.Defaulter); ok {\n\t\truntimeDefaulter.CustomDefault()\n\t}\n}", "func Default() *Config {\n\treturn &defaultConfig\n}", "func Default(v Value) Value {\n\tswitch x := v.(type) {\n\tcase *Vertex:\n\t\treturn x.Default()\n\tcase *Disjunction:\n\t\treturn x.Default()\n\tdefault:\n\t\treturn v\n\t}\n}", "func (Pipe) Default(ctx *context.Context) error {\n\tids := ids.New(\"nfpms\")\n\tfor i := range ctx.Config.NFPMs {\n\t\tfpm := &ctx.Config.NFPMs[i]\n\t\tif fpm.ID == \"\" {\n\t\t\tfpm.ID = \"default\"\n\t\t}\n\t\tif fpm.Bindir == \"\" {\n\t\t\tfpm.Bindir = \"/usr/bin\"\n\t\t}\n\t\tif fpm.PackageName == \"\" {\n\t\t\tfpm.PackageName = ctx.Config.ProjectName\n\t\t}\n\t\tif fpm.FileNameTemplate == \"\" {\n\t\t\tfpm.FileNameTemplate = defaultNameTemplate\n\t\t}\n\t\tif fpm.Maintainer == \"\" {\n\t\t\tdeprecate.NoticeCustom(ctx, \"nfpms.maintainer\", \"`{{ .Property }}` should always be set, check {{ .URL }} for more info\")\n\t\t}\n\t\tids.Inc(fpm.ID)\n\t}\n\n\tdeprecation.Noticer = io.Discard\n\treturn ids.Validate()\n}", "func (rule *NamespacesEventhubsAuthorizationRule) Default() {\n\trule.defaultImpl()\n\tvar temp any = rule\n\tif runtimeDefaulter, ok := temp.(genruntime.Defaulter); ok {\n\t\truntimeDefaulter.CustomDefault()\n\t}\n}", "func (r *AccessToken) Default() {\n\taccesstokenlog.Info(\"default\", \"name\", r.Name)\n}", "func Default() *Floki {\n\tm := New()\n\n\tm.loadConfig()\n\n\tif Env == Dev {\n\t\tm.Use(Logger())\n\n\t\tif boolValue(m.config[\"EnableProfiling\"]) {\n\t\t\tRegisterProfiler(m)\n\t\t}\n\t}\n\n\tm.Use(Recovery())\n\n\treturn m\n}", "func (m *ParameterMutator) Default(v interface{}) *ParameterMutator {\n\tm.lock.Lock()\n\tdefer m.lock.Unlock()\n\tm.proxy.defaultValue = v\n\treturn m\n}", "func (r *Storage) Default() {\n\tstoragelog.Info(\"default\", \"name\", r.Name)\n\tif r.Spec.ConnectType == \"internal\" {\n\t\tif r.Spec.Image == \"\" {\n\t\t\tr.Spec.Image = \"docker.elastic.co/elasticsearch/elasticsearch:7.5.1\"\n\t\t}\n\t\tif r.Spec.Instances == 0 {\n\t\t\tr.Spec.Instances = 3\n\t\t}\n\t}\n}", "func Default() *Server {\n\tf := logrus.Fields{\n\t\tLogLatency: 0,\n\t\tLogServerHostname: \"\",\n\t\tLogRemoteAddr: \"\",\n\t\tLogRequestSize: 0,\n\t\tLogResponseSize: 0,\n\t}\n\tl := logrus.New()\n\tl.Formatter = &logrus.TextFormatter{DisableTimestamp: true}\n\th := New()\n\th.Use(Logger(l, f))\n\th.Use(Recovery())\n\treturn h\n}", "func replaceDefault(in string) string {\n\tif in == \"default\" {\n\t\treturn defaultToken\n\t}\n\treturn in\n}", "func (cfg *Config) Default() error {\n\tcfg.MetricTTL = DefaultMetricTTL\n\tcfg.MetricType = DefaultMetricType\n\treturn nil\n}", "func DefaultUsage() {\n\tfmt.Println(\"Usage\")\n\tgoflags.PrintDefaults()\n\tgoenv.PrintDefaults()\n\tPrintDefaults()\n}", "func DefaultUsage() {\n\tfmt.Println(\"Usage\")\n\tgoflags.PrintDefaults()\n\tgoenv.PrintDefaults()\n\tPrintDefaults()\n}", "func (cmd *CLI) DefaultUsage() {\n\tfmt.Fprintln(cmd.StdOutput(), cmd.UsageString())\n}", "func (r *MetadataBackupPolicy) Default() {\n\tlog.Info(\"default\", \"name\", r.Name)\n\n\tif r.Spec.EtcdEndpoint == \"\" {\n\t\tlog.Info(\"Initializing EtcdEndpoint\")\n\t\tr.Spec.EtcdEndpoint = \"https://127.0.0.1:2379\"\n\t}\n\n\tif r.Spec.EtcdCreds == \"\" {\n\t\tlog.Info(\"Initializing EtcdCreds\")\n\t\tr.Spec.EtcdCreds = \"etcd-creds\"\n\t}\n\n\tif r.Spec.RetainNumBackups == nil || *r.Spec.RetainNumBackups == 0 {\n\t\tlog.Info(\"Initializing RetainNumBackups\")\n\t\tr.Spec.RetainNumBackups = new(int64)\n\t\t*r.Spec.RetainNumBackups = 120\n\t}\n\n\tif r.Spec.Suspend == nil {\n\t\tlog.Info(\"Initializing 'Suspend'\")\n\t\t// Initialized to false.\n\t\tr.Spec.Suspend = new(bool)\n\t}\n}", "func Default(def string, value interface{}) string {\n\tif set, ok := template.IsTrue(value); ok && set {\n\t\treturn fmt.Sprint(value)\n\t}\n\treturn def\n}", "func Default() *Server {\n\treturn New(Logger(), Recovery())\n}", "func Default() *MalService {\n\treturn &MalService{\n\t\tConfig: config.DefaultConfig,\n\t}\n}", "func DefaultExporter() Exporter {\n\treturn Exporter{now: defaultNow}\n}", "func DefaultOptions() Options {\n\treturn Options{}\n}", "func (r *AssetRendering) Default() {\n\tr.Value = nil\n}", "func NewDefault() *sanitizer {\n\ts := New()\n\ts.StripHtml = false\n\ts.Elements = acceptableTagsList\n\ts.Attributes = acceptableAttributesList\n\ts.URISchemes = acceptableUriSchemes\n\ts.StrictMode()\n\treturn s\n}", "func Default() *Registry {\n\treturn defaultInstance\n}", "func withDefault(key, defaultValue string) (value string) {\n\tif value == \"\" {\n\t\tvalue = defaultValue\n\t}\n\treturn value\n}", "func (f *File) Default(defaultVal interface{}) *File {\n\tf.defaultVal = defaultVal\n\treturn f\n}", "func SetDefault(key string, value interface{}) {\n\t// do nothing\n}", "func (m *ItemsMutator) Default(v interface{}) *ItemsMutator {\n\tm.lock.Lock()\n\tdefer m.lock.Unlock()\n\tm.proxy.defaultValue = v\n\treturn m\n}", "func Default() *AAAConfig {\n\treturn defaultStbConfig\n}", "func DefaultOptions() Options {\r\n\treturn DefaultServer.Options()\r\n}", "func (r *BackupLocation) Default() {\n\tbackuplocationlog.Info(\"default\", \"name\", r.Name)\n\n\t// TODO(user): fill in your defaulting logic.\n}", "func defaultOptions() *options {\n\treturn &options{\n\t\tDialFunc: DefaultDialFunc,\n\t\tLogFunc: DefaultLogFunc,\n\t}\n}", "func (e *Common) Default() string {\n\tif e.Def == nil {\n\t\treturn \"\"\n\t}\n\tif e.Def.Choice != \"\" {\n\t\treturn e.Def.Choice\n\t} else if e.Def.Type != \"\" {\n\t\t// Type is still used by the default element in collation.\n\t\treturn e.Def.Type\n\t}\n\treturn \"\"\n}", "func (s *SubnetTest) Default(cant bool) {\n\ts.CantIsBootstrapped = cant\n\ts.CantBootstrapped = cant\n}", "func (spec *KeystoneAPISpec) Default() {\n\tif spec.ContainerImage == \"\" {\n\t\tspec.ContainerImage = keystoneAPIDefaults.ContainerImageURL\n\t}\n}", "func (self *ScriptDef) makeDefault() error {\n if def, err := self.toInterfaceVal(self.DefaultStr); err != nil {\n return err\n } else {\n self.Default = def\n }\n return nil\n}", "func DefaultURL() string {\n\treturn ValueOrDefaultURL(\"\")\n}", "func (*Prefix) Default() string { return defaultPrefix }", "func (t *Table) Default(val interface{}) *Table {\n\tv := convertToStr(val)\n\tt.columns[len(t.columns)-1].Default = &v\n\treturn t\n}", "func NewDefault(opts ...DefaultOption) *Default {\n\tc := &Default{\n\t\tminSleep: 10 * time.Millisecond,\n\t\tmaxSleep: 2 * time.Second,\n\t\tdecayConstant: 2,\n\t\tattackConstant: 1,\n\t}\n\tc.Update(opts...)\n\treturn c\n}", "func Default() *Venom {\n\tven := New()\n\tven.RegisterResolver(EnvironmentLevel, defaultEnvResolver)\n\treturn ven\n}", "func Default() helmify.Processor {\n\treturn &dft{}\n}", "func Default() *Config {\n\tconf := &Config{\n\t\tProtocol: \"tcp\",\n\t\tAddr: \"0.0.0.0:25565\",\n\t\tHosts: []HostConfig{\n\t\t\t{\n\t\t\t\tName: \"Server-1\",\n\t\t\t\tAddr: \"localhost:25580\",\n\t\t\t},\n\t\t},\n\t\tLogConfig: LogConfig{\n\t\t\tLogConnections: true,\n\t\t\tLogDisconnect: false,\n\t\t},\n\t\tHealthCheckTime: 5,\n\t\tUDPTimeout: 3000,\n\t\tSaveConfigOnClose: false,\n\t\tInterfaces: []string{},\n\t}\n\tconf.fillFlags()\n\treturn conf\n}", "func (setting *MongodbDatabaseCollectionThroughputSetting) Default() {\n\tsetting.defaultImpl()\n\tvar temp any = setting\n\tif runtimeDefaulter, ok := temp.(genruntime.Defaulter); ok {\n\t\truntimeDefaulter.CustomDefault()\n\t}\n}", "func Default() (e *Engine, ie *InternalEngine) {\n\n\t// setup database\n\tdb := database.DefaultSQL()\n\n\t// migrate cms db\n\tdb.SQL.MigrateSql()\n\n\t// check for rsa keys\n\tsecurity.CheckOrGenRSAKeysAndSecrets(db.SQL.Dbx)\n\n\t// setup log level\n\tswitch context.Config.EnvVars.LogLevel {\n\tcase log.LOG_LEVEL_CRITICAL:\n\t\tfallthrough\n\tcase log.LOG_LEVEL_ERROR:\n\t\tgin.SetMode(gin.ReleaseMode)\n\tcase log.LOG_LEVEL_WARNING:\n\t\tgin.SetMode(gin.TestMode)\n\tcase log.LOG_LEVEL_DEBUG:\n\t\tgin.SetMode(gin.DebugMode)\n\t}\n\tr := gin.Default()\n\tir := gin.Default()\n\n\t// setup repositories\n\trg := repository.DefaultRepositoriesGroup(db.SQL.Dbx)\n\n\t// setup services\n\tsg := service.DefaultServicesGroup(rg, db)\n\n\t// setup controllers\n\tcg := controller.DefaultControllerGroup(r, sg)\n\ticg := controller.DefaultInternalControllerGroup(ir, sg)\n\n\t// create engine\n\te = &Engine{\n\t\tGin: r,\n\t\tControllersGroup: cg,\n\t\tServicesGroup: sg,\n\t\tRepositoriesGroup: rg,\n\t\tDatabase: db,\n\t}\n\n\t// create engine\n\tie = &InternalEngine{\n\t\tGin: ir,\n\t\tInternalControllersGroup: icg,\n\t\tServicesGroup: sg,\n\t\tRepositoriesGroup: rg,\n\t\tDatabase: db,\n\t}\n\n\treturn e, ie\n}", "func (rule *NamespacesTopicsSubscriptionsRule) Default() {\n\trule.defaultImpl()\n\tvar temp any = rule\n\tif runtimeDefaulter, ok := temp.(genruntime.Defaulter); ok {\n\t\truntimeDefaulter.CustomDefault()\n\t}\n}", "func DefaultLocation() string {\n\treturn locationDefault\n}", "func DefaultLocation() string {\n\treturn locationDefault\n}", "func (xdc *XxxDemoCreate) defaults() {\n\tif _, ok := xdc.mutation.IsDel(); !ok {\n\t\tv := xxxdemo.DefaultIsDel\n\t\txdc.mutation.SetIsDel(v)\n\t}\n\tif _, ok := xdc.mutation.Memo(); !ok {\n\t\tv := xxxdemo.DefaultMemo\n\t\txdc.mutation.SetMemo(v)\n\t}\n\tif _, ok := xdc.mutation.Sort(); !ok {\n\t\tv := xxxdemo.DefaultSort\n\t\txdc.mutation.SetSort(v)\n\t}\n\tif _, ok := xdc.mutation.CreatedAt(); !ok {\n\t\tv := xxxdemo.DefaultCreatedAt()\n\t\txdc.mutation.SetCreatedAt(v)\n\t}\n\tif _, ok := xdc.mutation.UpdatedAt(); !ok {\n\t\tv := xxxdemo.DefaultUpdatedAt()\n\t\txdc.mutation.SetUpdatedAt(v)\n\t}\n\tif _, ok := xdc.mutation.Status(); !ok {\n\t\tv := xxxdemo.DefaultStatus\n\t\txdc.mutation.SetStatus(v)\n\t}\n\tif _, ok := xdc.mutation.ID(); !ok {\n\t\tv := xxxdemo.DefaultID()\n\t\txdc.mutation.SetID(v)\n\t}\n}", "func (in *KubeadmControlPlane) Default() {\n\tif in.Spec.Replicas == nil {\n\t\treplicas := int32(1)\n\t\tin.Spec.Replicas = &replicas\n\t}\n\n\tif in.Spec.InfrastructureTemplate.Namespace == \"\" {\n\t\tin.Spec.InfrastructureTemplate.Namespace = in.Namespace\n\t}\n}", "func Default() *Batis {\n\tif batis != nil {\n\t\tbatis.parseEnv()\n\t}\n\treturn batis\n}", "func (v *VerticaDB) Default() {\n\tverticadblog.Info(\"default\", \"name\", v.Name)\n\n\t// imagePullPolicy: if not set should default to Always if the tag in the image is latest,\n\t// otherwise it should be IfNotPresent (set in verticadb_types.go)\n\tif strings.HasSuffix(v.Spec.Image, \":latest\") {\n\t\tv.Spec.ImagePullPolicy = v1.PullAlways\n\t}\n}", "func (r *DaisyInstallation) Default() {\n\tdaisyinstallationlog.Info(\"default\", \"name\", r.Name)\n\n\t// TODO(user): fill in your defaulting logic.\n\t// To avoid call Default() multi times when requeue\n\tif r.Labels == nil {\n\t\tr.Labels = make(map[string]string)\n\t}\n\tr.Labels[\"daisy.com/daisy-webhook-default\"] = \"handled\"\n}", "func (c *Client) defaultify() {\n\tif c.apiBase == \"\" {\n\t\tc.apiBase = apiBase //global var\n\t}\n}", "func Default() *Module {\n\treturn New(\"BAT0\")\n}", "func (policy *ServersConnectionPolicy) Default() {\n\tpolicy.defaultImpl()\n\tvar temp any = policy\n\tif runtimeDefaulter, ok := temp.(genruntime.Defaulter); ok {\n\t\truntimeDefaulter.CustomDefault()\n\t}\n}", "func defaultConfig() *config {\n\treturn &config{}\n}", "func (f *Funcs) alterDefault(s string) string {\n\tswitch f.driver {\n\tcase \"postgres\":\n\t\tif m := postgresDefaultCastRE.FindStringSubmatch(s); m != nil {\n\t\t\treturn m[1]\n\t\t}\n\tcase \"mysql\":\n\t\tif v := strings.ToUpper(s); v == \"CURRENT_TIMESTAMP()\" {\n\t\t\treturn \"CURRENT_TIMESTAMP\"\n\t\t}\n\tcase \"sqlite3\":\n\t\tif !sqliteDefaultNeedsParenRE.MatchString(s) {\n\t\t\treturn \"(\" + s + \")\"\n\t\t}\n\t}\n\treturn s\n}" ]
[ "0.66580963", "0.6516262", "0.6497345", "0.6331331", "0.6288372", "0.62188077", "0.62092525", "0.61736196", "0.6171395", "0.6143951", "0.6123842", "0.61111015", "0.61039007", "0.6092376", "0.6084715", "0.6078167", "0.60673964", "0.6048669", "0.6029358", "0.60164595", "0.60038024", "0.59953755", "0.597434", "0.59709674", "0.59663206", "0.5946038", "0.5934835", "0.5929184", "0.59159213", "0.5902315", "0.58894867", "0.58894867", "0.5886033", "0.5884788", "0.5880246", "0.5872957", "0.5867326", "0.58497477", "0.5849611", "0.58487356", "0.5845156", "0.5844274", "0.5841071", "0.5830357", "0.5829621", "0.5822546", "0.5819843", "0.581359", "0.5810674", "0.58007187", "0.5798811", "0.5790314", "0.5783981", "0.578066", "0.5779741", "0.57787204", "0.57787204", "0.5777051", "0.57753944", "0.57733583", "0.5768848", "0.57615113", "0.57570475", "0.5754305", "0.57542235", "0.5753977", "0.5753471", "0.57520497", "0.5748512", "0.57485044", "0.57401174", "0.573812", "0.5732673", "0.5722398", "0.57104075", "0.57043946", "0.5701134", "0.5692832", "0.5683263", "0.56772333", "0.5673221", "0.5665281", "0.56626195", "0.56606185", "0.5656799", "0.56487244", "0.56436336", "0.56370956", "0.5634679", "0.5632713", "0.5632713", "0.5632513", "0.5610608", "0.55902207", "0.5576535", "0.5573339", "0.5562552", "0.5561035", "0.5559965", "0.55561894", "0.5556111" ]
0.0
-1
Env returns Environment associated with Expect instance. Tests can use it to store arbitrary data. Example: e := httpexpect.Default(t, " e.Env().Put("key", "value") value := e.Env().GetString("key")
func (e *Expect) Env() *Environment { return e.chain.env() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Env(env string) (value []byte, err error) {\n\tvalue, err = exec.Command(\"go\", \"env\", env).Output()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tvalue = bytes.TrimSpace(value)\n\n\tif len(value) == 0 {\n\t\terr = ErrEmptyEnv{env}\n\t}\n\n\treturn\n}", "func (envManager *TestEnvManager) GetEnv() TestEnv {\n\treturn envManager.testEnv\n}", "func (e *echo) Env(val string) *echo {\n\tvars := e.declareVars(val)\n\tfor k, v := range vars {\n\t\tif err := os.Setenv(k, v); err != nil {\n\t\t\te.shouldPanic(err.Error())\n\t\t} else {\n\t\t\tdelete(e.vars, k) // overwrite local var\n\t\t}\n\t}\n\treturn e\n}", "func Env() *Environment {\n\tonce.Do(func() {\n\t\te = &Environment{}\n\t})\n\treturn e\n}", "func (b *Executable) Env(arg, value string) *Executable {\n\tif b.Environment == nil {\n\t\tb.Environment = make(map[string]string)\n\t}\n\tb.Environment[arg] = value\n\treturn b\n}", "func (suite *Suite[Env]) Env() *Env {\n\tif suite.env == nil || !suite.isUpdateEnvCalledInThisTest {\n\t\tsuite.UpdateEnv(suite.defaultStackDef)\n\t}\n\treturn suite.env\n}", "func GetENV(experimentDetails *experimentTypes.ExperimentDetails) {\n\texperimentDetails.ExperimentName = Getenv(\"EXPERIMENT_NAME\", \"\")\n\texperimentDetails.AppNS = Getenv(\"APP_NS\", \"\")\n\texperimentDetails.TargetContainer = Getenv(\"APP_CONTAINER\", \"\")\n\texperimentDetails.TargetPods = Getenv(\"APP_POD\", \"\")\n\texperimentDetails.AppLabel = Getenv(\"APP_LABEL\", \"\")\n\texperimentDetails.ChaosDuration, _ = strconv.Atoi(Getenv(\"TOTAL_CHAOS_DURATION\", \"30\"))\n\texperimentDetails.ChaosNamespace = Getenv(\"CHAOS_NAMESPACE\", \"litmus\")\n\texperimentDetails.EngineName = Getenv(\"CHAOS_ENGINE\", \"\")\n\texperimentDetails.ChaosUID = clientTypes.UID(Getenv(\"CHAOS_UID\", \"\"))\n\texperimentDetails.ChaosPodName = Getenv(\"POD_NAME\", \"\")\n\texperimentDetails.ContainerRuntime = Getenv(\"CONTAINER_RUNTIME\", \"\")\n\texperimentDetails.NetworkInterface = Getenv(\"NETWORK_INTERFACE\", \"eth0\")\n\texperimentDetails.TargetIPs = Getenv(\"TARGET_IPs\", \"\")\n}", "func (cmd Cmd) Env(env *Env) Cmd {\n\tcmd.Environment = env\n\treturn cmd\n}", "func Env(env Environ) func(*Runner) error {\n\treturn func(r *Runner) error {\n\t\tif env == nil {\n\t\t\tenv, _ = EnvFromList(os.Environ())\n\t\t}\n\t\tr.Env = env\n\t\treturn nil\n\t}\n}", "func getTestEnv() *Env {\n\tdb := createMockDB()\n\n\toauthConf := &oauth2.Config{\n\t\tClientID: \"abcdef0123abcdef4567\",\n\t\tClientSecret: \"abcdef0123abcdef4567abcdef8901abcdef2345\",\n\t\tScopes: []string{\"user:email\"},\n\t\tEndpoint: githuboauth.Endpoint,\n\t}\n\n\tenv := &Env{\n\t\tdb: db,\n\t\tjwtSecretKey: \"keyForTesting\",\n\t\toauthConf: oauthConf,\n\t\toauthState: \"nonRandomStateString\",\n\t}\n\treturn env\n}", "func NewEnv(t *testing.T) *Env {\n\treturn &Env{t, make(chan struct{}), sync.Mutex{}, make([]string, 0)}\n}", "func (s RunSpec) GetEnv() (env []string) {\n\tif s.Environment == nil {\n\t\ts.Environment = make(map[string]string)\n\t}\n\n\ts.Environment[\"RANNA_HOSTDIR\"] = s.HostDir\n\n\tenv = make([]string, len(s.Environment))\n\ti := 0\n\tfor k, v := range s.Environment {\n\t\tenv[i] = fmt.Sprintf(`%s=%s`, k, v)\n\t\ti++\n\t}\n\n\treturn\n}", "func GetEnv() Environment {\n\treturn sharedEnv\n}", "func Env() EnvType {\n\treturn env.Load().(EnvType)\n}", "func (app *Application) GetEnv() string {\n return app.env\n}", "func GetEnv() string {\n\tenv := os.Getenv(\"ENV\")\n\tif len(env) == 0 {\n\t\treturn TestEnv\n\t}\n\n\treturn env\n}", "func (ci MrbCallInfo) Env() REnv {\n\treturn REnv{C.mrb_vm_ci_env(ci.p), nil}\n}", "func (p RProc) Env() REnv {\n\tif !p.HasEnv() {\n\t\treturn REnv{nil, p.mrb}\n\t}\n\treturn REnv{C._MRB_PROC_ENV(p.p), p.mrb}\n}", "func FakeEnv() Env {\n\treturn NewScope(nil)\n}", "func InjectEnv() env.Env {\n\twire.Build(\n\t\twire.Bind(new(env.Env), new(env.GoDotEnv)),\n\t\tenv.NewGoDotEnv,\n\t)\n\treturn env.GoDotEnv{}\n}", "func Env() string {\n\tr := Get(\"ENV\")\n\tif r != Dev && r != Test && r != Acc && r != Prod {\n\t\tr = Dev\n\t}\n\treturn r\n}", "func (c *Context) Env() map[string]string {\n\treturn c.env\n}", "func NewExpectWithEnv(name string, args []string, env []string, serverProcessConfigName string) (ep *ExpectProcess, err error) {\n\tep = &ExpectProcess{\n\t\tcfg: expectConfig{\n\t\t\tname: serverProcessConfigName,\n\t\t\tcmd: name,\n\t\t\targs: args,\n\t\t\tenv: env,\n\t\t},\n\t\treadCloseCh: make(chan struct{}),\n\t}\n\tep.cmd = commandFromConfig(ep.cfg)\n\n\tif ep.fpty, err = pty.Start(ep.cmd); err != nil {\n\t\treturn nil, err\n\t}\n\n\tep.wg.Add(2)\n\tgo ep.read()\n\tgo ep.waitSaveExitErr()\n\treturn ep, nil\n}", "func Env(env ...string) string {\n\tvar returnVar string\n\tif len(env) == 0 {\n\t\tlogger.Error(\" env number is zero\")\n\t\treturn \"\"\n\t}\n\tfor i := 0; i < len(env); i++ {\n\t\tif returnVar == \"\" {\n\t\t\treturnVar = os.Getenv(env[i])\n\t\t} else {\n\t\t\treturn returnVar\n\t\t}\n\t}\n\treturn returnVar\n}", "func Env(name string) string {\n\treturn os.Getenv(name)\n}", "func (p *Process) Env() map[string]string {\n\treturn static.CopyStringMap(p.Builder.Env)\n}", "func (cb *clientBase) GetEnv() string {\n\treturn cb.env\n}", "func (env *Environment) Get() interface{} {\n\treturn *env\n}", "func (c *LocalCmd) Env(env ...string) *LocalCmd {\n\tc.env = env\n\treturn c\n}", "func Env(key string) EnvVars {\n\treturn EnvVars{\n\t\tKey: key,\n\t}\n}", "func (o BuildSpecRuntimePtrOutput) Env() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v *BuildSpecRuntime) map[string]string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Env\n\t}).(pulumi.StringMapOutput)\n}", "func GetEnv(name string, defaultValue string) string {\n\tif strVal, ok := os.LookupEnv(name); ok && len(strVal) > 0 {\n\t\treturn strVal\n\t}\n\n\treturn defaultValue\n}", "func (h *Helpers) Env(s string) string {\n\ts = strings.ToUpper(s)\n\tif v, ok := os.LookupEnv(s); ok {\n\t\treturn v\n\t}\n\n\treturn \"\"\n}", "func NewEnv() Env {\n\tenv := Env{}\n\tenv.LoadEnv()\n\treturn env\n}", "func (dir EnvDir) Env(secrets map[string]string, _ tpl.SecretReader) (map[string]string, error) {\n\treturn dir, nil\n}", "func (t *TestContext) Environment() environment.Implementation {\n\treturn t.impl\n}", "func GetENV() string {\n\tif env == \"\" {\n\t\treturn Dev\n\t}\n\treturn env\n}", "func NewEnv() (*Env, error) {\n\tctx := context.Background()\n\tdomainID := fmt.Sprintf(\"domain %d\", rand.Int()) // nolint: gas\n\tdb, err := testdb.New(ctx)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"env: failed to open database: %v\", err)\n\t}\n\n\t// Map server\n\tmapEnv, err := maptest.NewMapEnv(ctx)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"env: failed to create trillian map server: %v\", err)\n\t}\n\n\ttlog := fake.NewTrillianLogClient()\n\n\t// Configure domain, which creates new map and log trees.\n\tdomainStorage, err := domain.NewStorage(db)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"env: failed to create domain storage: %v\", err)\n\t}\n\tadminSvr := adminserver.New(tlog, mapEnv.Map, mapEnv.Admin, mapEnv.Admin, domainStorage, vrfKeyGen)\n\tdomainPB, err := adminSvr.CreateDomain(ctx, &pb.CreateDomainRequest{\n\t\tDomainId: domainID,\n\t\tMinInterval: ptypes.DurationProto(1 * time.Second),\n\t\tMaxInterval: ptypes.DurationProto(5 * time.Second),\n\t})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"env: CreateDomain(): %v\", err)\n\t}\n\n\tmapID := domainPB.Map.TreeId\n\tlogID := domainPB.Log.TreeId\n\tmapPubKey, err := der.UnmarshalPublicKey(domainPB.Map.GetPublicKey().GetDer())\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"env: Failed to load signing keypair: %v\", err)\n\t}\n\tvrfPub, err := p256.NewVRFVerifierFromRawKey(domainPB.Vrf.GetDer())\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"env: Failed to load vrf pubkey: %v\", err)\n\t}\n\n\t// Common data structures.\n\tmutations, err := mutationstorage.New(db)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"env: Failed to create mutations object: %v\", err)\n\t}\n\tauth := authentication.NewFake()\n\tauthz := authorization.New()\n\n\tqueue := mutator.MutationQueue(mutations)\n\tserver := keyserver.New(tlog, mapEnv.Map, mapEnv.Admin, mapEnv.Admin,\n\t\tentry.New(), auth, authz, domainStorage, queue, mutations)\n\tgsvr := grpc.NewServer()\n\tpb.RegisterKeyTransparencyServer(gsvr, server)\n\n\t// Sequencer\n\tseq := sequencer.New(tlog, mapEnv.Map, entry.New(), domainStorage, mutations, queue)\n\t// Only sequence when explicitly asked with receiver.Flush()\n\td := &domaindef.Domain{\n\t\tDomainID: domainID,\n\t\tLogID: logID,\n\t\tMapID: mapID,\n\t}\n\treceiver := seq.NewReceiver(ctx, d, 60*time.Hour, 60*time.Hour)\n\treceiver.Flush(ctx)\n\n\taddr, lis, err := Listen()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tgo gsvr.Serve(lis)\n\n\t// Client\n\tcc, err := grpc.Dial(addr, grpc.WithInsecure())\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Dial(%v) = %v\", addr, err)\n\t}\n\tktClient := pb.NewKeyTransparencyClient(cc)\n\tclient := grpcc.New(ktClient, domainID, vrfPub, mapPubKey, coniks.Default, fake.NewFakeTrillianLogVerifier())\n\tclient.RetryCount = 0\n\n\treturn &Env{\n\t\tEnv: &integration.Env{\n\t\t\tClient: client,\n\t\t\tCli: ktClient,\n\t\t\tDomain: domainPB,\n\t\t\tReceiver: receiver,\n\t\t},\n\t\tmapEnv: mapEnv,\n\t\tgrpcServer: gsvr,\n\t\tgrpcCC: cc,\n\t\tdb: db,\n\t}, nil\n}", "func (p RProc) EnvGet(i int) Value {\n\tif !p.HasEnv() {\n\t\treturn nilValue\n\t}\n\treturn p.mrb.ProcCFuncEnvGet(i)\n}", "func (s *EnvironmentsServiceOp) Get(env string) (*Environment, *Response, error) {\n\tpath := path.Join(environmentsPath, env)\n\treq, e := s.client.NewRequest(\"GET\", path, nil)\n\tif e != nil {\n\t\treturn nil, nil, e\n\t}\n\treturnedEnv := Environment{}\n\tresp, e := s.client.Do(req, &returnedEnv)\n\tif e != nil {\n\t\treturn nil, resp, e\n\t}\n\treturn &returnedEnv, resp, e\n}", "func SetEnv(e EnvType) {\n\tif e == EnvTypeDev ||\n\t\te == EnvTypeProd ||\n\t\te == EnvTypeTest {\n\t\tenv.Store(e)\n\t}\n}", "func (o StorageClusterSpecStorkPtrOutput) Env() StorageClusterSpecStorkEnvArrayOutput {\n\treturn o.ApplyT(func(v *StorageClusterSpecStork) []StorageClusterSpecStorkEnv {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Env\n\t}).(StorageClusterSpecStorkEnvArrayOutput)\n}", "func (b *taskBuilder) env(key, value string) {\n\tif b.Spec.Environment == nil {\n\t\tb.Spec.Environment = map[string]string{}\n\t}\n\tb.Spec.Environment[key] = value\n}", "func (o BuildSpecRuntimeOutput) Env() pulumi.StringMapOutput {\n\treturn o.ApplyT(func(v BuildSpecRuntime) map[string]string { return v.Env }).(pulumi.StringMapOutput)\n}", "func (o StorageClusterSpecStorkOutput) Env() StorageClusterSpecStorkEnvArrayOutput {\n\treturn o.ApplyT(func(v StorageClusterSpecStork) []StorageClusterSpecStorkEnv { return v.Env }).(StorageClusterSpecStorkEnvArrayOutput)\n}", "func (lkp OsEnvVariableLookuper) GetEnv(envVarName string) string {\n\tif TestingEnvVariableLookup != nil {\n\t\treturn TestingEnvVariableLookup.GetEnv(envVarName)\n\t}\n\n\treturn GetEnv(envVarName)\n}", "func TestEnvironmentGet(t *testing.T) {\n\tport := make(chan int, 1)\n\tdefer createTestServer(port, t).Close()\n\taddr := <-port\n\tresp, err := http.Get(fmt.Sprintf(\"http://localhost:%d/env/get\", addr))\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tdefer resp.Body.Close()\n\tbodyContent, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tif string(bodyContent) != \"Testing env.set function\" {\n\t\tt.Fatalf(\"Wrong env.get value. Expected 'Testing env.set function' but got '%s'\", string(bodyContent))\n\t}\n}", "func Env() Vars {\n\tif _env == nil {\n\t\t_envLock.Lock()\n\t\tdefer _envLock.Unlock()\n\t\tif _env == nil {\n\t\t\t_env = NewVarsFromEnvironment()\n\t\t}\n\t}\n\treturn _env\n}", "func (v *VirtualEnvironment) GetEnv(inherit bool, projectDir string) (map[string]string, error) {\n\tenv := make(map[string]string)\n\tif v.getEnv == nil {\n\t\t// if runtime is not explicitly disabled, this is an error\n\t\tif os.Getenv(constants.DisableRuntime) != \"true\" {\n\t\t\treturn nil, locale.NewError(\n\t\t\t\t\"err_get_env_unactivated\", \"Trying to set up an environment in an un-activated environment. This should not happen. Please report this issue in our forum: %s\",\n\t\t\t\tconstants.ForumsURL,\n\t\t\t)\n\t\t}\n\t\tenv[\"PATH\"] = os.Getenv(\"PATH\")\n\t} else {\n\t\tvar err error\n\t\tenv, err = v.getEnv(inherit, projectDir)\n\t\tif err != nil {\n\t\t\treturn env, err\n\t\t}\n\t}\n\n\tif projectDir != \"\" {\n\t\tenv[constants.ActivatedStateEnvVarName] = projectDir\n\t\tenv[constants.ActivatedStateIDEnvVarName] = v.activationID\n\n\t\t// Get project from explicitly defined configuration file\n\t\tpj, fail := project.Parse(filepath.Join(projectDir, constants.ConfigFileName))\n\t\tif fail != nil {\n\t\t\treturn env, fail.ToError()\n\t\t}\n\t\tfor _, constant := range pj.Constants() {\n\t\t\tenv[constant.Name()] = constant.Value()\n\t\t}\n\t}\n\n\tif inherit {\n\t\treturn inheritEnv(env), nil\n\t}\n\n\treturn env, nil\n}", "func Setenv(t testing.TB, name, val string) {\n\tNew(t).Setenv(name, val)\n}", "func (gf *genericFramework) Env(key, value string) error {\n\tif gf.adam.Variables == nil {\n\t\tgf.adam.Variables = jsonutil.NewVariableMap(\"\", nil)\n\t}\n\tif _, ok := gf.adam.Variables.Get(key); ok {\n\t\treturn fmt.Errorf(\"%v has been defined\", key)\n\t}\n\tgf.adam.Variables.Set(key, jsonutil.NewStringVariable(key, value))\n\treturn nil\n}", "func Env() string {\n\treturn AppConfig.Env\n}", "func Env(key, defaultValue string) string {\n\tvalue, defined := os.LookupEnv(key)\n\tif !defined {\n\t\treturn defaultValue\n\t}\n\n\treturn value\n}", "func (e *Env) NewEnv() *Env {\n\treturn &Env{\n\t\tenv: make(map[string]interface{}),\n\t\tparent: e,\n\t\tbuiltin: e.builtin,\n\t\tglobal: e.global,\n\t\tfuncArg: make(map[string]interface{}),\n\t\t//importFunc: e.importFunc,\n\t\tfileInfo: e.fileInfo,\n\t}\n}", "func env() *awscdk.Environment {\n\treturn &awscdk.Environment{\n\t Account: aws.String(os.Getenv(\"CDK_DEFAULT_ACCOUNT\")),\n\t Region: aws.String(os.Getenv(\"CDK_DEFAULT_REGION\")),\n\t}\n}", "func (o StorageClusterSpecPtrOutput) Env() StorageClusterSpecEnvArrayOutput {\n\treturn o.ApplyT(func(v *StorageClusterSpec) []StorageClusterSpecEnv {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Env\n\t}).(StorageClusterSpecEnvArrayOutput)\n}", "func env() *awscdk.Environment {\n\treturn nil\n\n\t//---------------------------------------------------------------------------\n\t// return &awscdk.Environment{\n\t// Account: jsii.String(os.Getenv(\"CDK_DEFAULT_ACCOUNT\")),\n\t// Region: jsii.String(os.Getenv(\"CDK_DEFAULT_REGION\")),\n\t// }\n}", "func NewEnv(context *libcoap.Context) *Env {\n return &Env{\n context,\n nil,\n make(chan Event, 32),\n nil,\n }\n}", "func Env(varName string, defaultValue string) string {\n\tvarValue := os.Getenv(varName)\n\tif varValue == \"\" {\n\t\treturn defaultValue\n\t}\n\treturn varValue\n}", "func (this *Go) GetEnviron() map[string]string {\n\tenv := make(map[string]string)\n\tfor _, p := range os.Environ() {\n\t\tarr := strings.SplitN(p, \"=\", 2)\n\t\tenv[arr[0]] = arr[1]\n\t}\n\treturn env\n}", "func NewMockEnv(t *testing.T) *MockEnv {\n\troot := testlib.TempDirAbs(t)\n\n\tlocal := filepath.Join(root, \"local\")\n\tif err := os.Mkdir(local, os.ModePerm); err != nil {\n\t\tpanic(err)\n\t}\n\tremote := filepath.Join(root, \"remote\")\n\tif err := os.Mkdir(remote, os.ModePerm); err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn &MockEnv{\n\t\tlocalRoot: local,\n\t\tremoteRoot: remote,\n\n\t\tplumbing: DevPlumbing{\n\t\t\tPlumbing: gggit.NewPlumbing(),\n\n\t\t\tSilenceStderr: true,\n\t\t\tURLMap: make(map[string]string),\n\t\t},\n\n\t\tvars: env.Variables{\n\t\t\tHOME: local,\n\t\t\tPATH: \"\",\n\t\t\tGGROOT: local,\n\t\t},\n\t}\n}", "func (o StorageClusterSpecOutput) Env() StorageClusterSpecEnvArrayOutput {\n\treturn o.ApplyT(func(v StorageClusterSpec) []StorageClusterSpecEnv { return v.Env }).(StorageClusterSpecEnvArrayOutput)\n}", "func (m *MockenvDescriber) EXPECT() *MockenvDescriberMockRecorder {\n\treturn m.recorder\n}", "func (c *Cli) Getenv(name string) string {\n\tvar s string\n\te, err := c.EnvAttribute(name)\n\tif err != nil {\n\t\treturn s\n\t}\n\tswitch e.Type {\n\tcase \"bool\":\n\t\treturn fmt.Sprintf(\"%t\", e.BoolValue)\n\tcase \"int\":\n\t\treturn fmt.Sprintf(\"%d\", e.IntValue)\n\tcase \"int64\":\n\t\treturn fmt.Sprintf(\"%d\", e.Int64Value)\n\tcase \"uint\":\n\t\treturn fmt.Sprintf(\"%d\", e.UintValue)\n\tcase \"uint64\":\n\t\treturn fmt.Sprintf(\"%d\", e.Uint64Value)\n\tcase \"float64\":\n\t\treturn fmt.Sprintf(\"%f\", e.Float64Value)\n\tcase \"time.Duration\":\n\t\treturn fmt.Sprintf(\"%s\", e.DurationValue)\n\t}\n\treturn e.StringValue\n}", "func (o VirtualDatabaseSpecOutput) Env() VirtualDatabaseSpecEnvArrayOutput {\n\treturn o.ApplyT(func(v VirtualDatabaseSpec) []VirtualDatabaseSpecEnv { return v.Env }).(VirtualDatabaseSpecEnvArrayOutput)\n}", "func Getenv(dst interface{}, key string) error {\n\treturn DefaultReader.Getenv(dst, key)\n}", "func (o VirtualDatabaseSpecPtrOutput) Env() VirtualDatabaseSpecEnvArrayOutput {\n\treturn o.ApplyT(func(v *VirtualDatabaseSpec) []VirtualDatabaseSpecEnv {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Env\n\t}).(VirtualDatabaseSpecEnvArrayOutput)\n}", "func GetEnv(root *pb.Span, t *pb.TraceChunk) string {\n\tif v, ok := root.Meta[envKey]; ok {\n\t\treturn v\n\t}\n\tfor _, s := range t.Spans {\n\t\tif s.SpanID == root.SpanID {\n\t\t\tcontinue\n\t\t}\n\t\tif v, ok := s.Meta[envKey]; ok {\n\t\t\treturn v\n\t\t}\n\t}\n\treturn \"\"\n}", "func env(key string, defaultValue string) string {\n\tif value, exists := os.LookupEnv(key); exists {\n\t\treturn value\n\t}\n\treturn defaultValue\n}", "func (e EnvFile) Env(secrets map[string]string, sr tpl.SecretReader) (map[string]string, error) {\n\tenv, err := e.env.Env(secrets, sr)\n\tif err != nil {\n\t\treturn nil, ErrParsingTemplate(e.path, err)\n\t}\n\treturn env, nil\n}", "func GetEnv(key string) string {\n\treturn envHash[key]\n}", "func NewEnv(files []string) *Env {\n\tglobal := make(map[string]interface{})\n\tglobal[\"ENVIRON\"] = getEnvVars()\n\n\treturn &Env{\n\t\tenv: make(map[string]interface{}),\n\t\tparent: nil,\n\t\tbuiltin: newBuiltIn(files),\n\t\tglobal: global,\n\t\tfuncArg: make(map[string]interface{}),\n\t\t//importFunc: make(map[string]func(*Env) (reflect.Value, error)),\n\t\tfileInfo: &FileInfo{\n\t\t\tfiles: files,\n\t\t\tcurFileIndex: 0,\n\t\t\treadCloser: make(map[string]*io.ReadCloser),\n\t\t\tscanner: make(map[string]*bufio.Scanner),\n\t\t},\n\t}\n}", "func (c *Client) EnvGet(ctx context.Context) (map[string]string, error) {\n\tvar resp EnvGetResponse\n\tif err := c.client.Do(ctx, \"GET\", envURL, nil, &resp); err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp.Env, nil\n}", "func NewEnv() *Env {\n\tenv := new(Env)\n\t// env.EnvParams.DefaultEnv()\n\t// env.EnvParams.Initialize()\n\treturn env\n}", "func Mock() Env {\n\treturn mock.New()\n}", "func (p RProc) SetEnv(stackItems ...Value) {\n\tif len(stackItems) == 0 {\n\t\tC._mrb_create_env(p.mrb.p, p.p, 0, nil)\n\t\treturn\n\t}\n\n\tC._mrb_create_env(p.mrb.p, p.p, C.mrb_int(len(stackItems)), &(stackItems[0].v))\n}", "func env(key string, defaultValue string) string {\n\tvalue := os.Getenv(key)\n\n\tif len(value) > 0 {\n\t\treturn value\n\t}\n\treturn defaultValue\n\n}", "func (context *Context) GetEnv(target string) (string, bool) {\n\tv, ok := context.Env.Get(target)\n\tif !ok {\n\t\treturn context.Global.GEnv.Get(target)\n\t}\n\treturn v, ok\n}", "func NewEnv() *Env {\n\treturn &Env{\n\t\tos.Getenv(clusterRoot),\n\t\tos.Getenv(kubeConfig),\n\t}\n}", "func assertEnv(ctx context.Context, client client.Client, spec corev1.PodSpec, component, container, key, expectedValue string) error {\n\tvalue, err := getEnv(ctx, client, spec, component, container, key)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif value != nil && strings.ToLower(*value) != expectedValue {\n\t\treturn ErrIncompatibleCluster{\n\t\t\terr: fmt.Sprintf(\"%s=%s is not supported\", key, *value),\n\t\t\tcomponent: component,\n\t\t\tfix: fmt.Sprintf(\"remove the %s env var or set it to '%s'\", key, expectedValue),\n\t\t}\n\t}\n\n\treturn nil\n}", "func Env(prefix string) Provider {\n\treturn &env{prefix}\n}", "func (m *wasiSnapshotPreview1Impl) environGet(penviron pointer, penvironBuf pointer) (err wasiErrno) {\n\tfor _, s := range m.env {\n\t\tbuf := m.bytes(penvironBuf)\n\t\tcopy(buf, s)\n\t\tbuf[len(s)] = 0\n\n\t\tm.putUint32(uint32(penvironBuf), penviron)\n\t\tpenvironBuf, penviron = penvironBuf+pointer(len(s))+1, penviron+4\n\t}\n\treturn wasiErrnoSuccess\n}", "func (o VirtualDatabaseSpecBuildOutput) Env() VirtualDatabaseSpecBuildEnvArrayOutput {\n\treturn o.ApplyT(func(v VirtualDatabaseSpecBuild) []VirtualDatabaseSpecBuildEnv { return v.Env }).(VirtualDatabaseSpecBuildEnvArrayOutput)\n}", "func (c *CmdReal) GetEnv() []string {\n\treturn c.cmd.Env\n}", "func (c *Config) Env() map[string]interface{} {\r\n\tc.mu.Lock()\r\n\tdefer c.mu.Unlock()\r\n\r\n\treturn cloneMap(c.env)\r\n}", "func (envx *EnvBufferedExec) SetEnv(key string, value string) *EnvBufferedExec {\n\tenvx.env[key] = value\n\treturn envx\n}", "func (o VirtualDatabaseSpecBuildPtrOutput) Env() VirtualDatabaseSpecBuildEnvArrayOutput {\n\treturn o.ApplyT(func(v *VirtualDatabaseSpecBuild) []VirtualDatabaseSpecBuildEnv {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Env\n\t}).(VirtualDatabaseSpecBuildEnvArrayOutput)\n}", "func Environ() string {\n\treturn environ\n}", "func (c *Cli) Env(name string) string {\n\te, ok := c.env[name]\n\tif ok == false {\n\t\treturn fmt.Sprintf(\"%q not documented for environment\", name)\n\t}\n\treturn e.Usage\n}", "func Setenv(key, value string) error", "func GetEnv(fileName string) (*Env, error) {\n\tv := viper.New()\n\tv.AutomaticEnv()\n\tv.AddConfigPath(configPath)\n\tv.SetConfigType(configType)\n\tv.SetConfigName(fileName)\n\n\tif err := v.ReadInConfig(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tlogger := v.GetString(envKeyLogger)\n\tif logger == \"\" {\n\t\tlogger = defaultLoggerEnvironment\n\t}\n\n\tmySQLHost := v.GetString(envKeyMySQLHost)\n\tif mySQLHost == \"\" {\n\t\treturn nil, missingEnvError(envKeyMySQLHost)\n\t}\n\n\tmySQLPort := v.GetString(envKeyMySQLPort)\n\tif mySQLPort == \"\" {\n\t\treturn nil, missingEnvError(envKeyMySQLPort)\n\t}\n\n\tmySQLName := v.GetString(envKeyMySQLName)\n\tif mySQLName == \"\" {\n\t\treturn nil, missingEnvError(envKeyMySQLName)\n\t}\n\n\tmySQLPassword := v.GetString(envKeyMySQLPassword)\n\n\tmySQLUser := v.GetString(envKeyMySQLUser)\n\n\tredisHost := v.GetString(envKeyRedisHost)\n\tif redisHost == \"\" {\n\t\treturn nil, missingEnvError(envKeyRedisHost)\n\t}\n\n\tredisPort := v.GetInt(envKeyRedisPort)\n\tif redisPort == 0 {\n\t\treturn nil, missingEnvError(envKeyRedisPort)\n\t}\n\n\tredisPassword := v.GetString(envKeyRedisPassword)\n\n\tredisKeyPrefix := v.GetString(envKeyRedisKeyPrefix)\n\n\tport := v.GetString(envKeyPort)\n\tif port == \"\" {\n\t\tport = defaultValuePort\n\t}\n\n\tenv := &Env{\n\t\tlogger: logger,\n\t\tmySQLHost: mySQLHost,\n\t\tmySQLName: mySQLName,\n\t\tmySQLPassword: mySQLPassword,\n\t\tmySQLPort: mySQLPort,\n\t\tmySQLUser: mySQLUser,\n\t\tport: port,\n\t\tredisHost: redisHost,\n\t\tredisPassword: redisPassword,\n\t\tredisPort: redisPort,\n\t\tredisKeyPrefix: redisKeyPrefix,\n\t}\n\n\treturn env, nil\n}", "func (c *Action) SetEnv(k, v string) {\n\tfmt.Fprintf(c.w, setEnvFmt, k, escapeData(v))\n}", "func (s *Action) Env(c *cli.Context) error {\n\tctx := ctxutil.WithGlobalFlags(c)\n\tname := c.Args().First()\n\targs := c.Args().Tail()\n\tkeepCase := c.Bool(\"keep-case\")\n\n\tif len(args) == 0 {\n\t\treturn exit.Error(exit.Usage, nil, \"Missing subcommand to execute\")\n\t}\n\n\tif !s.Store.Exists(ctx, name) && !s.Store.IsDir(ctx, name) {\n\t\treturn exit.Error(exit.NotFound, nil, \"Secret %s not found\", name)\n\t}\n\n\tkeys := make([]string, 0, 1)\n\tif s.Store.IsDir(ctx, name) {\n\t\tdebug.Log(\"%q is a dir, adding it's entries\", name)\n\n\t\tl, err := s.Store.Tree(ctx)\n\t\tif err != nil {\n\t\t\treturn exit.Error(exit.List, err, \"failed to list store: %s\", err)\n\t\t}\n\n\t\tsubtree, err := l.FindFolder(name)\n\t\tif err != nil {\n\t\t\treturn exit.Error(exit.NotFound, nil, \"Entry %q not found\", name)\n\t\t}\n\n\t\tfor _, e := range subtree.List(tree.INF) {\n\t\t\tdebug.Log(\"found key: %s\", e)\n\t\t\tkeys = append(keys, e)\n\t\t}\n\t} else {\n\t\tkeys = append(keys, name)\n\t}\n\n\tenv := make([]string, 0, 1)\n\tfor _, key := range keys {\n\t\tdebug.Log(\"exporting to environment key: %s\", key)\n\t\tsec, err := s.Store.Get(ctx, key)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"failed to get entry for env prefix %q: %w\", name, err)\n\t\t}\n\t\tenvKey := path.Base(key)\n\t\tif !keepCase {\n\t\t\tenvKey = strings.ToUpper(envKey)\n\t\t}\n\t\tenv = append(env, fmt.Sprintf(\"%s=%s\", envKey, sec.Password()))\n\t}\n\n\tcmd := exec.CommandContext(ctx, args[0], args[1:]...)\n\tcmd.Env = append(os.Environ(), env...)\n\tcmd.Stdin = os.Stdin\n\tcmd.Stdout = stdout\n\tcmd.Stderr = os.Stderr\n\n\treturn cmd.Run()\n}", "func (t *Task) Env() map[string]string {\n\treturn t.opts.env\n}", "func (config *Configuration) GetEnv() string {\n\treturn string(config.env)\n}", "func TestEnviron(t *testing.T) {\n\ttests := map[string]string{\n\t\t\"KEY_0\": \"Abc\",\n\t\t\"KEY_1\": \"Def\",\n\t}\n\n\t// Set test data.\n\tos.Clearenv()\n\tfor key, value := range tests {\n\t\tif err := os.Setenv(key, value); err != nil {\n\t\t\tt.Error(err)\n\t\t}\n\t}\n\n\t// Test function.\n\tfor i, str := range Environ() {\n\t\ttmp := strings.Split(str, \"=\")\n\t\tkey, value := tmp[0], tmp[1]\n\t\tif v, ok := tests[key]; v != value || !ok {\n\t\t\tif !ok {\n\t\t\t\tt.Errorf(\"test %v. extra key`%v`\", i, key)\n\t\t\t} else {\n\t\t\t\tt.Errorf(\"test %v. expected `%v` but `%v`\", i, v, value)\n\t\t\t}\n\t\t}\n\t}\n}", "func (r *CheckedDaemonSet) assertEnv(ctx context.Context, client client.Client, container, key, expectedValue string) error {\n\tif err := assertEnv(ctx, client, r.Spec.Template.Spec, ComponentCalicoNode, container, key, expectedValue); err != nil {\n\t\treturn err\n\t}\n\tr.ignoreEnv(container, key)\n\treturn nil\n}", "func TestSetEnv(t *testing.T, key string, value string) func() {\n\tt.Helper()\n\toriginalValue := os.Getenv(key)\n\tos.Setenv(key, value)\n\treturn func() { os.Setenv(key, originalValue) }\n}", "func NewEnvironment(testKeys []string) Environment {\n\tvars := make(map[string]string)\n\tfor _, key := range testKeys {\n\t\tvars[key] = os.Getenv(key)\n\t}\n\treturn Environment{backup: vars}\n}", "func GetEnv(name string) string {\n\treturn os.Getenv(name)\n}" ]
[ "0.69212854", "0.66415465", "0.65178424", "0.6469878", "0.63406515", "0.63362306", "0.6283041", "0.62473047", "0.6217694", "0.6203327", "0.619105", "0.61533636", "0.6140795", "0.61385715", "0.6101667", "0.609115", "0.6076825", "0.603624", "0.60137403", "0.59723485", "0.5957127", "0.59388554", "0.5920079", "0.58856606", "0.587825", "0.5873134", "0.58566993", "0.5851227", "0.5830275", "0.58210164", "0.5800738", "0.57835776", "0.57793826", "0.57723325", "0.57713693", "0.57635313", "0.57602006", "0.5758193", "0.5753125", "0.57499605", "0.57354426", "0.5729966", "0.57240933", "0.57232356", "0.5722918", "0.5713317", "0.57131594", "0.5699462", "0.5691218", "0.5683483", "0.56799144", "0.5672831", "0.5668093", "0.5659458", "0.5657843", "0.56556475", "0.56528175", "0.5639351", "0.56380147", "0.5634281", "0.5625658", "0.56179756", "0.56152016", "0.56125575", "0.5610904", "0.56096786", "0.5609369", "0.5608131", "0.5607647", "0.56071424", "0.55994487", "0.5597153", "0.55965644", "0.5593531", "0.55898863", "0.5586755", "0.5581781", "0.5570247", "0.55677176", "0.5563317", "0.55616945", "0.5550284", "0.55305046", "0.55275595", "0.5526128", "0.552028", "0.55191404", "0.55163205", "0.5509128", "0.5508744", "0.55020726", "0.54898137", "0.54886657", "0.5485311", "0.5483277", "0.548059", "0.5478879", "0.54714453", "0.5470773", "0.54704654" ]
0.7403274
0
Request returns a new Request instance. Arguments are similar to NewRequest. After creating request, all builders attached to Expect instance are invoked. See Builder.
func (e *Expect) Request(method, path string, pathargs ...interface{}) *Request { opChain := e.chain.enter("Request(%q)", method) defer opChain.leave() req := newRequest(opChain, e.config, method, path, pathargs...) for _, builder := range e.builders { builder(req) } for _, matcher := range e.matchers { req.WithMatcher(matcher) } return req }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *Request) Request() *http.Request {\n\treq := httptest.NewRequest(r.Method, r.Target, strings.NewReader(r.Body))\n\treq.Header = r.Header\n\treturn req\n}", "func (f Factory) Request(method string, arguments interface{}) (request Request) {\n\trequest.Token = f.Token\n\trequest.Method = method\n\tif arguments != nil {\n\t\tvar err error\n\t\trequest.Values, err = query.Values(arguments)\n\t\tmust(err)\n\t}\n\treturn request\n}", "func NewRequest(requestName string, params rata.Params, header http.Header, query url.Values, body ...io.Reader) Request {\n\tif header == nil {\n\t\theader = http.Header{}\n\t}\n\theader.Set(\"Accept\", \"application/json\")\n\n\trequest := Request{\n\t\tRequestName: requestName,\n\t\tParams: params,\n\t\tHeader: header,\n\t\tQuery: query,\n\t}\n\n\tif len(body) == 1 {\n\t\trequest.Body = body[0]\n\t}\n\n\treturn request\n}", "func NewRequest() (r *Request) {\n\treturn &Request{}\n}", "func (r *Request) Request() (*http.Request, error) {\n\tvar req *http.Request\n\tvar err error\n\n\tif r.body != nil {\n\t\tbody, err := json.Marshal(r.body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tbuff := bytes.NewBuffer(body)\n\n\t\treq, err = http.NewRequest(r.method, r.url, buff)\n\t} else {\n\t\treq, err = http.NewRequest(r.method, r.url, nil)\n\t}\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tv, err := query.Values(r.query)\n\tif err == nil {\n\t\treq.URL.RawQuery = v.Encode()\n\t}\n\n\treturn req, nil\n}", "func NewRequest() *Request {\n\treturn defaul.NewRequest()\n}", "func NewRequest() *Request {\n\treturn &Request{}\n}", "func NewRequest() *Request {\n\treturn &Request{}\n}", "func NewRequest() *Request {\n\tr := &Request{}\n\treturn r\n}", "func NewRequest() *Request {\n\tr := &Request{}\n\treturn r\n}", "func NewRequest() *Request {\n\tr := &Request{}\n\treturn r\n}", "func (s APIv1) NewRequest(ctx context.Context, method, path string, data interface{}) (req *http.Request, err error) {\n\t// Resolve the URL reference from the path\n\tendpoint := s.endpoint.ResolveReference(&url.URL{Path: path})\n\n\tvar body io.ReadWriter\n\tif data != nil {\n\t\tbody = &bytes.Buffer{}\n\t\tif err = json.NewEncoder(body).Encode(data); err != nil {\n\t\t\treturn nil, fmt.Errorf(\"could not serialize request data: %s\", err)\n\t\t}\n\t} else {\n\t\tbody = nil\n\t}\n\n\t// Create the http request\n\tif req, err = http.NewRequestWithContext(ctx, method, endpoint.String(), body); err != nil {\n\t\treturn nil, fmt.Errorf(\"could not create request: %s\", err)\n\t}\n\n\t// Set the headers on the request\n\treq.Header.Add(\"User-Agent\", \"Whisper/1.0\")\n\treq.Header.Add(\"Accept\", \"application/json\")\n\treq.Header.Add(\"Accept-Language\", \"en-US,en\")\n\treq.Header.Add(\"Accept-Encoding\", \"gzip, deflate, br\")\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\n\treturn req, nil\n}", "func NewRequest() *Request {\n\tr := &Request{\n\t\tAggregations: make(map[string]types.Aggregations, 0),\n\t\tExt: make(map[string]json.RawMessage, 0),\n\t\tScriptFields: make(map[string]types.ScriptField, 0),\n\t}\n\treturn r\n}", "func NewRequest() *Request {\n\tr := &Request{\n\t\tAggregations: make(map[string]types.Aggregations, 0),\n\t\tExt: make(map[string]json.RawMessage, 0),\n\t\tScriptFields: make(map[string]types.ScriptField, 0),\n\t}\n\treturn r\n}", "func NewRequest(robot *Robot, message *Message, query []string) *Request {\n\treturn &Request{\n\t\tMessage: message,\n\t\tQuery: query,\n\t\trobot: robot,\n\t}\n}", "func NewRequest(bufid string, acts ...Action) Request {\n\tr := Request{BufferID: bufid}\n\tr.Add(acts...)\n\treturn r\n}", "func NewRequest() *HttpRequest {\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{},\n\t}\n\n\thR := HttpRequest{client: &http.Client{Transport: tr}}\n\thR.cache = cache.NewMemoryCache()\n\thR.params = Params{}\n\thR.parsers = []bodyParser{rawParser()}\n\thR.handlers = []ResponseStatusHandler{handler()}\n\thR.fetch = fetchXML()\n\treturn &hR\n}", "func NewRequest(t *testing.T, h http.Handler, target string) *Request {\n\treturn &Request{\n\t\tT: t,\n\t\tHandler: h,\n\t\tTarget: target,\n\t}\n}", "func NewRequest(addr string, body []byte) *Request {\n\treturn &Request{\n\t\tAddr: addr,\n\t\tHeader: defaultHeader(len(body)),\n\t\tBody: body,\n\t}\n}", "func NewRequest(params interface{}, atta map[string]interface{}) *DubboRequest {\n\tif atta == nil {\n\t\tatta = make(map[string]interface{})\n\t}\n\treturn &DubboRequest{\n\t\tParams: params,\n\t\tAttachments: atta,\n\t}\n}", "func NewRequest(id, replyTo string, object ObjectID, method string, args ...interface{}) (*Request, error) {\n\tinputs, err := newTuple(args...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Request{\n\t\tID: id,\n\t\tInputs: inputs,\n\t\tObject: object,\n\t\tReplyTo: replyTo,\n\t\tMethod: method,\n\t}, nil\n}", "func New() *Request {\n\treturn &Request{\n\t\tclient: &http.Client{Timeout: time.Second * 3},\n\t\tmethod: \"GET\",\n\t\theader: make(http.Header),\n\t}\n}", "func NewRequest(session *Session, path string) *Request {\n\trequest := new(Request)\n\trequest.UnderlyingRequest = api.NewRequest(session.underlyingSession, path)\n\trequest.session = session\n\treturn request\n}", "func NewRequest() *Request {\n\treturn &Request{\n\t\tCounter: 1,\n\t\tURLStruct: &url.URL{},\n\t\tHeader: make(http.Header),\n\t\tPathParams: make(map[string]string),\n\t}\n}", "func NewRequest(url string, branch string, author string, email string, date string, deps *[]Request) *Request {\n\treturn &Request{\n\t\turl,\n\t\tbranch,\n\t\tauthor,\n\t\temail,\n\t\tdate,\n\t\tdeps,\n\t}\n}", "func NewRequest(r *http.Request) *Request {\n\tvar request Request\n\trequest.ID = atomic.AddUint32(&requestID, 1)\n\trequest.Method = r.Method\n\trequest.Body = r.Body\n\trequest.BodyBuff = new(bytes.Buffer)\n\trequest.BodyBuff.ReadFrom(r.Body)\n\trequest.RemoteAddr = r.Header.Get(\"X-Forwarded-For\")\n\trequest.Header = r.Header\n\tif request.RemoteAddr == \"\" {\n\t\trequest.RemoteAddr = r.RemoteAddr\n\t}\n\trequest.UrlParams = mux.Vars(r)\n\trequest.QueryParams = r.URL.Query()\n\treturn &request\n}", "func (s *Nap) Request() (*http.Request, error) {\n\treqURL, err := url.Parse(s.rawURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = buildQueryParamUrl(reqURL, s.queryStructs, s.queryParams)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar body io.Reader\n\tif s.bodyProvider != nil {\n\t\tbody, err = s.bodyProvider.Body()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treq, err := http.NewRequestWithContext(s.Context(), s.method, reqURL.String(), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\taddHeaders(req, s.header)\n\treturn req, err\n}", "func NewRequest(m Manager) *Request {\n\treturn &Request{\n\t\tm: m,\n\t\trLocks: make(map[string]struct{}),\n\t\twLocks: make(map[string]struct{}),\n\t}\n}", "func NewRequest(path string, mode xrdfs.OpenMode, options xrdfs.OpenOptions) *Request {\n\treturn &Request{Mode: mode, Options: options, Path: path}\n}", "func (r *RecordedRequest) Request() *http.Request {\n\turl := r.URL\n\treturn &http.Request{\n\t\tMethod: r.Method,\n\t\tURL: &url,\n\t\tProto: r.Proto,\n\t\tProtoMajor: r.ProtoMajor,\n\t\tProtoMinor: r.ProtoMinor,\n\t\tHeader: r.Header,\n\t\tBody: ioutil.NopCloser(bytes.NewBuffer(r.Body)),\n\t\tContentLength: r.ContentLength,\n\t\tTransferEncoding: r.TransferEncoding,\n\t\tHost: r.Host,\n\t\tForm: r.Form,\n\t\tPostForm: r.PostForm,\n\t\tMultipartForm: r.MultipartForm,\n\t\tTrailer: r.Trailer,\n\t\tRemoteAddr: r.RemoteAddr,\n\t\tRequestURI: r.RequestURI,\n\t\tTLS: r.TLS,\n\t}\n}", "func NewRequest(session *Session, path string) *Request {\n\tr := new(Request)\n\tr.path = path\n\tr.session = session\n\tr.queryValues = make(url.Values)\n\treturn r\n}", "func newRequest(db *Database) *Request {\n\treq := &Request{\n\t\tdb: db,\n\t\tpath: \"/\",\n\t\tquery: url.Values{},\n\t\theader: http.Header{},\n\t}\n\treturn req\n}", "func NewRequest(t Type, body io.WriterTo) *Request {\n\treq := &Request{\n\t\tBody: &copyReader{WriterTo: body},\n\t\tProto: \"OFP/1.3\",\n\t\tProtoMajor: 1, ProtoMinor: 3,\n\t}\n\n\treq.Header.Version = uint8(req.ProtoMajor + req.ProtoMinor)\n\treq.Header.Type = t\n\n\treturn req\n}", "func (c *baseClient) Request() (*http.Request, error) {\n\treqURL, err := url.Parse(c.url)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif c.queryStruct != nil {\n\t\terr = addQueryStruct(reqURL, c.queryStruct)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tbody := &bytes.Buffer{}\n\tif c.body != nil {\n\t\tif err := json.NewEncoder(body).Encode(c.body); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(c.method, reqURL.String(), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Add headers to request\n\tfor k, vs := range c.header {\n\t\tfor _, v := range vs {\n\t\t\treq.Header.Add(k, v)\n\t\t}\n\t}\n\n\treturn req, nil\n}", "func (c *apiHttp) NewRequest(command string) RequestBuilder {\n\theaders := make(map[string]string)\n\tif c.Headers != nil {\n\t\tfor k := range c.Headers {\n\t\t\theaders[k] = c.Headers.Get(k)\n\t\t}\n\t}\n\n\treturn &requestBuilder{\n\t\tcommand: command,\n\t\tclient: c,\n\t\theaders: headers,\n\t}\n}", "func (i *instance) NewRequest(method, urlStr string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(method, urlStr, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Make a context for this request.\n\tc := &context{\n\t\treq: req,\n\t\tsession: newSessionID(),\n\t\tinstance: i,\n\t}\n\n\t// Associate this request.\n\trelease := appengine_internal.RegisterTestContext(req, c)\n\ti.relFuncs = append(i.relFuncs, release)\n\n\treturn req, nil\n}", "func NewRequest(method, urlStr string, body io.Reader, meta MetaMap) (req *Request, err error) {\n\tif meta == nil {\n\t\tmeta = make(MetaMap, 2)\n\t}\n\n\tif httpReq, err := http.NewRequest(method, urlStr, body); err == nil {\n\t\treq = &Request{\n\t\t\tRequest: httpReq,\n\t\t\tMeta: meta,\n\t\t}\n\t}\n\treturn\n}", "func (c *Client) Request() *Request {\n\treq := &Request{\n\t\tmake(map[string]string),\n\t\tmake(map[string]string),\n\t}\n\n\tfor _, modifier := range c.modifiers {\n\t\tmodifier(req)\n\t}\n\n\treturn req\n}", "func NewRequest(q string) *Request {\n\treq := &Request{\n\t\tq: q,\n\t\tHeader: make(map[string][]string),\n\t}\n\treturn req\n}", "func (g *Group) newRequest(operation *request.Operation, input interface{}, output interface{}, handlers *request.Handlers) *request.Request {\n return request.New(g.Client, operation, input, output, handlers)\n}", "func NewRequest(method, url string) *Request {\n\treturn &Request{method: method, url: url}\n}", "func NewRequest(url string) *Request {\n\treturn client.NewRequest(url)\n}", "func NewRequest(r *http.Request, ns string, body []byte) (*Request, error) {\n\th, err := json.Marshal(r.Header)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Request{\n\t\tID: uuid.NewV4(),\n\t\tMethod: r.Method,\n\t\tHeaders: h,\n\t\tPayload: body,\n\t\tHost: r.Host,\n\t\tScheme: r.URL.Scheme,\n\t\tPath: r.URL.Path,\n\t\tNamespace: ns,\n\t\tRetries: 0,\n\t}, nil\n}", "func NewRequest(req Requester) *Transaction {\n\treturn &Transaction{\n\t\ttp: Request,\n\t\tReq: req,\n\t}\n}", "func NewRequest(src, dest string) *Request {\n\treturn &Request{src: src, dest: dest}\n}", "func New(handler http.Handler) *Request {\n\tapiTest := &APITest{}\n\n\trequest := &Request{apiTest: apiTest}\n\tresponse := &Response{apiTest: apiTest}\n\tapiTest.request = request\n\tapiTest.response = response\n\tapiTest.handler = handler\n\n\treturn apiTest.request\n}", "func (c *Client) newRequest(t RequestType) *Request {\n\tc.seqID++\n\treturn &Request{\n\t\tAPIVersion: \"v1\",\n\t\tRequestType: t,\n\t\tTracerTime: time.Now().Unix(),\n\t\tRuntimeID: globalconfig.RuntimeID(),\n\t\tSeqID: c.seqID,\n\t\tDebug: c.debug,\n\t\tApplication: Application{\n\t\t\tServiceName: c.Service,\n\t\t\tEnv: c.Env,\n\t\t\tServiceVersion: c.Version,\n\t\t\tTracerVersion: version.Tag,\n\t\t\tLanguageName: \"go\",\n\t\t\tLanguageVersion: runtime.Version(),\n\t\t},\n\t\tHost: Host{\n\t\t\tHostname: hostname,\n\t\t\tContainerID: internal.ContainerID(),\n\t\t\tOS: getOSName(),\n\t\t\tOSVersion: getOSVersion(),\n\t\t},\n\t}\n}", "func newRequest(req *http.Request) *Request {\n\trequest := &Request{\n\t\tRequest: req,\n\t}\n\n\treturn request\n}", "func (c *Context) Request(body interface{}) (req *Request, err error) {\n\tdefer c.Req.Body.Close()\n\n\treq, err = RequestReader(c.Req.Body, body)\n\treturn\n}", "func (m *InMemManager) NewRequest() *Request {\n\treturn NewRequest(m)\n}", "func NewRequest(id string, method string, params interface{}) *Request {\n\tbuffer, err := json.Marshal(params)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn &Request{JsonRPC: \"2.0\", Id: id, Method: method, Params: buffer}\n}", "func NewRequest(methodName string, params []interface{}) (req Request, err error) {\n\treq = Request{\n\t\tMethodName: methodName,\n\t\tParams: make([]Value, 0, len(params)),\n\t}\n\n\tfor _, p := range params {\n\t\tv := NewValue(p)\n\t\tif v == nil {\n\t\t\treturn req, fmt.Errorf(\"NewRequest: cannot convert parameter %v\", p)\n\t\t}\n\n\t\treq.Params = append(req.Params, *v)\n\t}\n\n\treturn req, nil\n}", "func NewRequest(op OpCode, bdy proto.Message, authenticator auth.Authenticator, provider ProviderID) (*Request, error) {\n\tbodyBuf, err := proto.Marshal(bdy)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tauthtok, err := authenticator.NewRequestAuth()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr := &Request{\n\t\tHeader: wireHeader{\n\t\t\tversionMajor: versionMajorOne,\n\t\t\tversionMinor: versionMinorZero,\n\t\t\tflags: flagsZero,\n\t\t\tprovider: provider,\n\t\t\t// todo set session handles\n\t\t\tcontentType: contentTypeProtobuf,\n\t\t\tauthType: authtok.AuthType(),\n\t\t\tbodyLen: uint32(len(bodyBuf)),\n\t\t\tauthLen: uint16(authtok.Buffer().Len()),\n\t\t\topCode: op,\n\t\t\tStatus: StatusSuccess,\n\t\t},\n\t\tBody: RequestBody{\n\t\t\tbytes.NewBuffer(bodyBuf),\n\t\t},\n\t\tAuth: authtok,\n\t}\n\treturn r, nil\n}", "func New(env *solo.Solo) *RequestManager {\n\trequestManager := &RequestManager{env: env}\n\treturn requestManager\n}", "func newRequest(req *http.Request) *Request {\n\treturn &Request{\n\t\tRequest: req,\n\t}\n}", "func NewRequest(f map[string]string, r io.Reader, h map[string]string, d encoding.DecodeFunc) *Request {\n\treturn &Request{Fields: f, Raw: r, Headers: h, decode: d}\n}", "func NewRequest(method, toURL string, body io.Reader) (*Request, error) {\n\tif method == \"\" {\n\t\t// We document that \"\" means \"GET\" for Request.Method, and people have\n\t\t// relied on that from NewRequest, so keep that working.\n\t\t// We still enforce ValidMethod for non-empty methods.\n\t\tmethod = GET\n\t}\n\tif !ValidMethod(method) {\n\t\treturn nil, fmt.Errorf(\"github.com/badu//http: invalid method %q\", method)\n\t}\n\tu, err := url.Parse(toURL) // Just url.Parse (url is shadowed for godoc).\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\trc, ok := body.(io.ReadCloser)\n\tif !ok && body != nil {\n\t\trc = ioutil.NopCloser(body)\n\t}\n\t// The host's colon:port should be normalized. See Issue 14836.\n\tu.Host = removeEmptyPort(u.Host)\n\treq := &Request{\n\t\tMethod: method,\n\t\tURL: u,\n\t\tProto: HTTP1_1,\n\t\tProtoMajor: 1,\n\t\tProtoMinor: 1,\n\t\tHeader: make(hdr.Header),\n\t\tBody: rc,\n\t\tHost: u.Host,\n\t}\n\tif body != nil {\n\t\tswitch v := body.(type) {\n\t\tcase *bytes.Buffer:\n\t\t\treq.ContentLength = int64(v.Len())\n\t\t\tbuf := v.Bytes()\n\t\t\treq.GetBody = func() (io.ReadCloser, error) {\n\t\t\t\tr := bytes.NewReader(buf)\n\t\t\t\treturn ioutil.NopCloser(r), nil\n\t\t\t}\n\t\tcase *bytes.Reader:\n\t\t\treq.ContentLength = int64(v.Len())\n\t\t\tsnapshot := *v\n\t\t\treq.GetBody = func() (io.ReadCloser, error) {\n\t\t\t\tr := snapshot\n\t\t\t\treturn ioutil.NopCloser(&r), nil\n\t\t\t}\n\t\tcase *strings.Reader:\n\t\t\treq.ContentLength = int64(v.Len())\n\t\t\tsnapshot := *v\n\t\t\treq.GetBody = func() (io.ReadCloser, error) {\n\t\t\t\tr := snapshot\n\t\t\t\treturn ioutil.NopCloser(&r), nil\n\t\t\t}\n\t\tdefault:\n\t\t\t// This is where we'd set it to -1 (at least\n\t\t\t// if body != NoBody) to mean unknown, but\n\t\t\t// that broke people during the Go 1.8 testing\n\t\t\t// period. People depend on it being 0 I\n\t\t\t// guess. Maybe retry later. See Issue 18117.\n\t\t}\n\t\t// For client requests, Request.ContentLength of 0\n\t\t// means either actually 0, or unknown. The only way\n\t\t// to explicitly say that the ContentLength is zero is\n\t\t// to set the Body to nil. But turns out too much code\n\t\t// depends on NewRequest returning a non-nil Body,\n\t\t// so we use a well-known ReadCloser variable instead\n\t\t// and have the http package also treat that sentinel\n\t\t// variable to mean explicitly zero.\n\t\tif req.GetBody != nil && req.ContentLength == 0 {\n\t\t\treq.Body = NoBody\n\t\t\treq.GetBody = func() (io.ReadCloser, error) { return NoBody, nil }\n\t\t}\n\t}\n\n\treturn req, nil\n}", "func NewRequest(conn Conn, data []byte, offset int) Request {\n\treturn &request{conn: conn, data: data, offset: offset}\n}", "func NewRequest(name string) *Request {\n\tresult := Request{name: name}\n\tresult.pending = nil\n\tresult.approved = nil\n\treturn &result\n}", "func newRequest(t *testing.T, path string) *http.Request {\n\tif !strings.HasPrefix(path, \"/\") {\n\t\tpath = \"/\" + path\n\t}\n\tu := fmt.Sprintf(\"http://localhost:%d%s\", expectedPort, path)\n\treq, err := http.NewRequestWithContext(\n\t\tcontext.Background(),\n\t\thttp.MethodGet, u,\n\t\tnil,\n\t)\n\trequire.NoError(t, err)\n\treturn req\n}", "func Request(url, method string, args interface{}) (req *http.Request, err error) {\n\n\tmessage, err := rpc.EncodeClientRequest(method, args)\n\tif err != nil {\n\t\treturn\n\t}\n\treq, err = http.NewRequest(\"POST\", url, bytes.NewBuffer(message))\n\tif err != nil {\n\t\treturn\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treturn\n}", "func (r *Search) Request(req *Request) *Search {\n\tr.req = req\n\n\treturn r\n}", "func NewRequest(ctx context.Context) *Request {\n\tr := &Request{\n\t\tctx: ctx,\n\t\tquery: url.Values{},\n\t}\n\n\tr.url, _ = url.Parse(endpoint)\n\n\treturn r\n}", "func (c *Client) NewRequest(method, path string) *Request {\n\trequestUrl := path\n\tif !strings.HasPrefix(path, \"http\") {\n\t\trequestUrl = c.Config.ApiAddress + path\n\t}\n\n\tr := &Request{\n\t\tmethod: method,\n\t\turl: requestUrl,\n\t\tparams: make(map[string][]string),\n\t}\n\treturn r\n}", "func newRequest(r *http.Request, vars map[string]string) *Request {\n\trequest := &Request{r, vars, nil}\n\treturn request\n}", "func NewRequest(endpoint Endpoint, payload interface{}, config *models.Config) *Request {\n\treturn &Request{\n\t\tEndpoint: endpoint,\n\t\tPayload: payload,\n\t\tConfig: config,\n\t\tMethod: POST,\n\t\tContentType: JSONContentType,\n\t}\n}", "func (fn RequestMakerFunc) MakeRequest() Request {\n\treturn fn()\n}", "func (s *Stein) newRequest(method string, path string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(method, path, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\n\treturn req, nil\n}", "func (b *AuroraBackend) NewRequest(params *CallParams) (*http.Request, error) {\n\treq, err := http.NewRequest(params.Method, params.Path, params.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\theaders := params.Headers\n\tif headers == nil {\n\t\theaders = make(http.Header)\n\t}\n\n\tif params.Credentials != nil {\n\t\theaders.Add(\"X-Application-ID\", params.Credentials.AppID)\n\t\theaders.Add(\"X-Application-Token\", params.Credentials.AppToken)\n\t\theaders.Add(\"X-Device-ID\", params.Credentials.DeviceID)\n\t}\n\n\treq.Header = headers\n\treturn req, nil\n}", "func (enc *Enqueuer) Request() (*http.Request, error) {\n\tbodyBytes, err := enc.BodyBytes()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq, err := http.NewRequest(\n\t\t\"POST\",\n\t\tenc.RequestPath(),\n\t\tbytes.NewReader(bodyBytes),\n\t)\n\treq.Header.Add(\"Content-Length\", strconv.Itoa(len(bodyBytes)))\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treturn req, nil\n}", "func (r *Request) New() *Request {\n\theaders := make(http.Header)\n\tfor key, value := range r.header {\n\t\theaders[key] = value\n\t}\n\n\treturn &Request{\n\t\tclient: r.client,\n\t\tmethod: r.method,\n\t\turl: r.url,\n\t\theader: headers,\n\t\tquery: r.query,\n\t\tbody: r.body,\n\t\tSuccess: r.Success,\n\t\tFailure: r.Failure,\n\t}\n}", "func mockRequest() *request.Request {\n\tr := request.Request{}\n\thl := request.HandlerList{\n\t\tAfterEachFn: func(item request.HandlerListRunItem) bool { return true },\n\t}\n\n\tr.Handlers = request.Handlers{\n\t\tValidate: hl,\n\t\tBuild: hl,\n\t\tSend: hl,\n\t\tSign: hl,\n\t\tValidateResponse: hl,\n\t\tUnmarshal: hl,\n\t\tUnmarshalMeta: hl,\n\t\tUnmarshalError: hl,\n\t\tRetry: hl,\n\t\tAfterRetry: hl,\n\t\tComplete: hl,\n\t}\n\n\treturn &r\n}", "func NewRequest(method, urlStr string, body ...[]byte) (*Request, error) {\n\tif method == \"\" {\n\t\tmethod = \"GET\"\n\t}\n\tvar r io.Reader\n\tif body != nil {\n\t\tr = bytes.NewReader(body[0])\n\t}\n\n\treq, err := http.NewRequest(method, urlStr, r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Request{Req: req}, nil\n}", "func NewRequest(method, urlStr string, body io.Reader) (*Request, error) {\n\treq, err := http.NewRequest(method, urlStr, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Request{\n\t\t\treq, Retries, Timeout, backoff.NewExponentialBackOff()},\n\t\tnil\n}", "func (s *ApplicationsService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func NewRequest(bufConn io.Reader) (*Request, error) {\n\t// Read the version byte\n\theader := []byte{0, 0, 0}\n\tif _, err := io.ReadAtLeast(bufConn, header, 3); err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to get command version: %v\", err)\n\t}\n\n\t// ensure compatibility\n\tif header[0] != socks5Version {\n\t\treturn nil, fmt.Errorf(\"Unsupported command version: %v\", header[0])\n\t}\n\n\t// Read in the destination address\n\tdest, err := readAddrSpec(bufConn)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Request{\n\t\tVersion: socks5Version,\n\t\tCommand: header[1],\n\t\tDestAddr: dest,\n\t\tbufConn: bufConn,\n\t}, nil\n}", "func (c *Client) NewRequest(ctx context.Context, method, path string, params interface{}, data interface{}) *request.Request {\n\thttpReq, _ := http.NewRequest(method, path, nil)\n\thttpReq = httpReq.WithContext(ctx)\n\thttpReq.URL, _ = url.Parse(c.Info.Endpoint + path)\n\n\tr := request.New(httpReq, c.Info, c.Handlers.Copy(), params, data)\n\tr.HTTPClient = c.HTTPClient\n\treturn r\n}", "func (api *API) Request(method string, result interface{}) *resty.Request {\n\trequest := resty.New().R()\n\tif result != nil {\n\t\trequest.SetResult(result)\n\t}\n\n\treturn request\n}", "func NewRequest(parent context.Context, br *openrtb.BidRequest, endpoint string, encoder Encoder) (*Request, error) {\n\tif br == nil {\n\t\treturn nil, ErrNilBidRequest\n\t} else if len(endpoint) == 0 {\n\t\treturn nil, ErrEmptyURL\n\t}\n\n\tif encoder == nil {\n\t\tencoder = DefaultEncoder\n\t}\n\n\t// TODO(@garukun): Consider implementing another constructor that does not marshal bid request\n\t// object to JSON reader every single time.\n\tbuf := bytes.NewBuffer(nil)\n\tif err := encoder.EncodeToWriter(buf, br); err != nil {\n\t\treturn nil, err\n\t}\n\n\thr, err := http.NewRequest(http.MethodPost, endpoint, buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Request{\n\t\tctx: parent,\n\t\tbr: br,\n\t\thr: hr,\n\t}, nil\n}", "func (c *IoT) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\t// Run custom request initialization if present\n\tif initRequest != nil {\n\t\tinitRequest(req)\n\t}\n\n\treturn req\n}", "func NewRequest(host string) *Request {\n\trequest := &Request{host, url.Values{}, http.Header{}, BasicAuth{}}\n\treturn request\n}", "func (c *Client) NewRequest(method, relPath string, body, options, headers interface{}) (*http.Request, error) {\n\trel, err := url.Parse(relPath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Make the full url based on the relative path\n\tu := c.baseURL.ResolveReference(rel)\n\n\t// Add custom options\n\tif options != nil {\n\t\toptionsQuery, err := query.Values(options)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tfor k, values := range u.Query() {\n\t\t\tfor _, v := range values {\n\t\t\t\toptionsQuery.Add(k, v)\n\t\t\t}\n\t\t}\n\t\tu.RawQuery = optionsQuery.Encode()\n\t}\n\n\t// A bit of JSON ceremony\n\tvar js []byte = nil\n\tif body != nil {\n\t\tjs, err = json.Marshal(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, u.String(), bytes.NewBuffer(js))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"Accept\", \"application/json\")\n\treq.Header.Add(\"User-Agent\", UserAgent)\n\n\tc.makeSignature(req)\n\n\treturn req, nil\n}", "func (c *Client) NewRequest(method, path string, reqBody interface{}, needAuth bool) (*http.Request, error) {\n\tvar body []byte\n\tvar err error\n\n\tif reqBody != nil {\n\t\tbody, err = json.Marshal(reqBody)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\ttarget := fmt.Sprintf(\"%s%s\", c.APIEndPoint, path)\n\treq, err := http.NewRequest(method, target, bytes.NewReader(body))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Inject headers\n\tif body != nil {\n\t\treq.Header.Set(\"Content-Type\", \"application/json;charset=utf-8\")\n\t}\n\treq.Header.Set(\"Authorization\", fmt.Sprintf(\"sso-key %s:%s\", c.APIKey, c.APISecret))\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"User-Agent\", \"ExternalDNS/\"+externaldns.Version)\n\n\t// Send the request with requested timeout\n\tc.Client.Timeout = c.Timeout\n\n\treturn req, nil\n}", "func NewRequest(r *http.Request, cfg *UploadsConfig) (req *Request, err error) {\n\treq = &Request{\n\t\tRemoteAddr: fetchIP(r.RemoteAddr),\n\t\tProtocol: r.Proto,\n\t\tMethod: r.Method,\n\t\tURI: uri(r),\n\t\tHeader: r.Header,\n\t\tCookies: make(map[string]string),\n\t\tRawQuery: r.URL.RawQuery,\n\t\tAttributes: attributes.All(r),\n\t}\n\n\tfor _, c := range r.Cookies() {\n\t\tif v, err := url.QueryUnescape(c.Value); err == nil {\n\t\t\treq.Cookies[c.Name] = v\n\t\t}\n\t}\n\n\tswitch req.contentType() {\n\tcase contentNone:\n\t\treturn req, nil\n\n\tcase contentStream:\n\t\treq.body, err = ioutil.ReadAll(r.Body)\n\t\treturn req, err\n\n\tcase contentMultipart:\n\t\tif err = r.ParseMultipartForm(defaultMaxMemory); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treq.Uploads = parseUploads(r, cfg)\n\t\tfallthrough\n\tcase contentFormData:\n\t\tif err = r.ParseForm(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treq.body = parseData(r)\n\t}\n\n\treq.Parsed = true\n\treturn req, nil\n}", "func Request(ctx context.Context) *http.Request {\n\treturn ctx.Value(reqKey).(*http.Request)\n}", "func newRequest(ctx context.Context, msg interface{}) *request {\n\treturn &request{\n\t\tctx: ctx,\n\t\tmsg: msg,\n\t\tfailure: make(chan error, 1),\n\t\tresponse: make(chan *Delivery, 1),\n\t}\n}", "func newRequest(method, path string, body io.Reader) (*http.Request, error) {\n\tr := http.Request{\n\t\tMethod: method,\n\t\tURL: &url.URL{Path: path},\n\t\tProto: \"HTTP/1.1\",\n\t\tProtoMajor: 1,\n\t\tProtoMinor: 1,\n\t\tHeader: make(http.Header),\n\t}\n\n\tif body != nil {\n\t\tswitch b := body.(type) {\n\t\tcase *bytes.Buffer:\n\t\t\tr.Body = ioutil.NopCloser(body)\n\t\t\tr.ContentLength = int64(b.Len())\n\t\tcase *bytes.Reader:\n\t\t\tr.Body = ioutil.NopCloser(body)\n\t\t\tr.ContentLength = int64(b.Len())\n\t\tcase *strings.Reader:\n\t\t\tr.Body = ioutil.NopCloser(body)\n\t\t\tr.ContentLength = int64(b.Len())\n\t\tdefault:\n\t\t\tr.Body = ioutil.NopCloser(body)\n\t\t}\n\t}\n\n\treturn &r, nil\n}", "func NewRequest(commandName string, arguments []string) ([]byte, error) {\n\tvar err error\n\tvar q Request\n\tvar payload []byte\n\n\tq = Request{Command: commandName, Arguments: arguments}\n\n\tif payload, err = json.Marshal(q); err != nil {\n\t\tlog.Fatalln(\"Error on JSON Marshal: \", err)\n\t\treturn nil, err\n\t}\n\n\treturn payload, nil\n}", "func (r *Request) Build() (*Request, error) {\n\tif err := r.checkError(); err != nil {\n\t\treturn nil, err\n\t}\n\tif \"\" == r.httpVerb {\n\t\treturn nil, errors.New(\"No HttpVerb Provided\")\n\t}\n\tif \"\" == r.url {\n\t\treturn nil, errors.New(\"No URL Provided\")\n\t}\n\n\t// Check if this can be moved to WithHeaders method or using url.Values{}\n\tif r.requestParams != nil {\n\t\treqParamString := \"?\"\n\t\tfor k, v := range r.requestParams {\n\t\t\tif reqParamString != \"?\" {\n\t\t\t\treqParamString += \"&\"\n\t\t\t}\n\t\t\treqParamString = reqParamString + k + `=` + v\n\t\t}\n\t\tr.url = r.url + reqParamString\n\t}\n\n\t// TODO: Move to a Function\n\tif len(r.requestBody) == 0 && len(r.requestBodyParams) != 0 {\n\t\tr.WithRequestBody(r.requestBodyParams)\n\t}\n\n\treturn r, nil\n}", "func NewRequest(command string) (qr quantum.Request, err error) {\n\tr := Request{\n\t\tCommand: command,\n\t}\n\td, err := json.Marshal(r)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tqr = NewRequestJSON(d)\n\treturn\n}", "func (r *Forecast) Request(req *Request) *Forecast {\n\tr.req = req\n\n\treturn r\n}", "func NewRequest(id string, method string) *Request {\n\n\treq := &Request{\n\t\tJSONRPC: \"2.0\",\n\t\tID: id,\n\t\tMethod: method,\n\t}\n\treturn req\n}", "func (c *Client) NewRequest(url string) *Request {\n\tvar r Request\n\tr.params = make(map[string]string)\n\n\tif url[0] == '/' {\n\t\turl = url[1:]\n\t}\n\tr.url = url\n\tr.client = c\n\n\treturn &r\n}", "func (c *Client) Request(verb, spath string, ro *RequestOptions) (*http.Request, error) {\n\tlog.Printf(\"[INFO] request: %s %s\", verb, spath)\n\n\tif ro == nil {\n\t\tro = new(RequestOptions)\n\t}\n\n\tu := *c.URL\n\tu.Path = path.Join(c.URL.Path, spath)\n\n\tif c.Token != \"\" {\n\t\tif ro.Headers == nil {\n\t\t\tro.Headers = make(map[string]string)\n\t\t}\n\t\tro.Headers[\"Authorization\"] = fmt.Sprintf(\"Bearer %s\", c.Token)\n\t}\n\n\treturn c.rawRequest(verb, &u, ro)\n}", "func (c *Client) NewRequest(method, requestPath string) *Request {\n\tc.modifyLock.RLock()\n\taddr := c.addr\n\ttoken := c.token\n\tmfaCreds := c.mfaCreds\n\twrappingLookupFunc := c.wrappingLookupFunc\n\tpolicyOverride := c.policyOverride\n\tc.modifyLock.RUnlock()\n\n\thost := addr.Host\n\t// if SRV records exist (see https://tools.ietf.org/html/draft-andrews-http-srv-02), lookup the SRV\n\t// record and take the highest match; this is not designed for high-availability, just discovery\n\t// Internet Draft specifies that the SRV record is ignored if a port is given\n\tif addr.Port() == \"\" && c.config.SRVLookup {\n\t\t_, addrs, err := net.LookupSRV(\"http\", \"tcp\", addr.Hostname())\n\t\tif err == nil && len(addrs) > 0 {\n\t\t\thost = fmt.Sprintf(\"%s:%d\", addrs[0].Target, addrs[0].Port)\n\t\t}\n\t}\n\n\treq := &Request{\n\t\tMethod: method,\n\t\tURL: &url.URL{\n\t\t\tUser: addr.User,\n\t\t\tScheme: addr.Scheme,\n\t\t\tHost: host,\n\t\t\tPath: path.Join(addr.Path, requestPath),\n\t\t},\n\t\tHost: addr.Host,\n\t\tClientToken: token,\n\t\tParams: make(map[string][]string),\n\t}\n\n\tvar lookupPath string\n\tswitch {\n\tcase strings.HasPrefix(requestPath, \"/v1/\"):\n\t\tlookupPath = strings.TrimPrefix(requestPath, \"/v1/\")\n\tcase strings.HasPrefix(requestPath, \"v1/\"):\n\t\tlookupPath = strings.TrimPrefix(requestPath, \"v1/\")\n\tdefault:\n\t\tlookupPath = requestPath\n\t}\n\n\treq.MFAHeaderVals = mfaCreds\n\n\tif wrappingLookupFunc != nil {\n\t\treq.WrapTTL = wrappingLookupFunc(method, lookupPath)\n\t} else {\n\t\treq.WrapTTL = DefaultWrappingLookupFunc(method, lookupPath)\n\t}\n\n\treq.Headers = c.Headers()\n\treq.PolicyOverride = policyOverride\n\n\treturn req\n}", "func NewRequest(token string) *Request {\n\treqID := fmt.Sprintf(\"alaudacli-%d\", time.Now().Unix())\n\trestyReq := resty.R()\n\trestyReq.SetHeader(\"Content-Type\", \"application/json\")\n\trestyReq.SetHeader(\"Alauda-Request-ID\", reqID)\n\n\tif token != \"\" {\n\t\trestyReq.SetHeader(\"Authorization\", fmt.Sprintf(\"Token %s\", token))\n\t}\n\n\treturn &Request{\n\t\trestyReq,\n\t}\n}", "func (a *APITest) Request() *Request {\n\treturn a.request\n}", "func (rt *registryRoundTripper) NewRequest(method, path string,\n\tquery *url.Values, body interface{}) (*http.Request, error) {\n\n\treq, err := rt.DefaultRequestDoer.NewRequest(method, path, query, body)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif tok := rt.holder.Token(); tok != \"\" {\n\t\treq.Header.Set(\"Authorization\", \"Bearer \"+tok)\n\t}\n\n\treq.Header.Set(\"User-Agent\", \"Torus-Daemon/\"+rt.version)\n\treq.Header.Set(\"X-Registry-Version\", rt.apiVersion)\n\n\treturn req, nil\n}", "func NewRequestObject(node *tree.Node, res www.ResponseWriter, r *www.Request) *Request {\n\n\treturn &Request{\n\t\tParameters: \tparameters.New(),\n\t\tconfig:\t\t\tnode.Config,\n\t\tNode:\t\t\tnode,\n\t\tres:\t\t \tres,\n\t\tr: \t\t\t\tr,\n\t\tmethod: \t\tr.Method,\n\t}\n}", "func (c client) newRequest(ctx context.Context, method string, url string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequestWithContext(ctx, method, url, body)\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to build request %w\", err)\n\t}\n\n\treturn req, nil\n}" ]
[ "0.73795944", "0.71671206", "0.70723724", "0.70567113", "0.69686484", "0.6907132", "0.68797153", "0.68797153", "0.68705773", "0.68705773", "0.68705773", "0.6861712", "0.68570507", "0.68570507", "0.6850988", "0.68318594", "0.6826593", "0.6780901", "0.67803687", "0.6775812", "0.6749095", "0.674192", "0.6670509", "0.66667384", "0.66025054", "0.6588116", "0.65837353", "0.65777725", "0.65765405", "0.6571973", "0.65408385", "0.65356886", "0.6498663", "0.6494698", "0.6486694", "0.6482431", "0.64689076", "0.6465789", "0.6465072", "0.64327115", "0.6425263", "0.64133984", "0.6410421", "0.640301", "0.63916427", "0.638521", "0.63847613", "0.6367591", "0.6362813", "0.63603306", "0.63492787", "0.6333264", "0.63244534", "0.6320266", "0.6311177", "0.6309158", "0.6307617", "0.63064253", "0.62992394", "0.629578", "0.6287799", "0.6286711", "0.6286627", "0.6273797", "0.62627524", "0.6257802", "0.62523466", "0.62505513", "0.6243899", "0.62398016", "0.6223657", "0.62171316", "0.62170917", "0.6216524", "0.6216206", "0.62115234", "0.6207934", "0.6204767", "0.6204446", "0.62034774", "0.61938286", "0.6191767", "0.61770177", "0.61766094", "0.6176006", "0.61753464", "0.6168175", "0.61657584", "0.61638665", "0.616349", "0.6158406", "0.6157383", "0.6156242", "0.61517197", "0.61486554", "0.61471725", "0.6141996", "0.61372423", "0.61368334", "0.6136393" ]
0.80211514
0
OPTIONS is a shorthand for e.Request("OPTIONS", path, pathargs...).
func (e *Expect) OPTIONS(path string, pathargs ...interface{}) *Request { return e.Request(http.MethodOptions, path, pathargs...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Options(url string, data ...interface{}) (*ClientResponse, error) {\n\treturn DoRequest(\"OPTIONS\", url, data...)\n}", "func (g *Group) OPTIONS(path string, handler Handler, middleware ...Middleware) *Group {\n\treturn g.Add(http.MethodOptions, path, handler, middleware...)\n}", "func (r *Request) Options(path string, params ...url.Values) {\n\tcontentType := \"text/html\"\n\n\tif len(params) == 0 {\n\t\tr.Send(\"OPTIONS\", path, contentType)\n\t} else {\n\t\tr.Send(\"OPTIONS\", path, contentType, params[0])\n\t}\n}", "func Options(path string, fn http.HandlerFunc, c ...alice.Constructor) {\n\trecord(\"OPTIONS\", path)\n\n\tinfoMutex.Lock()\n\tr.OPTIONS(path, Handler(alice.New(c...).ThenFunc(fn)))\n\tinfoMutex.Unlock()\n}", "func (serv *Server) OPTIONS(url string, handlers ...Handler) {\n\tserv.Handle(\"OPTIONS\", url, handlers...)\n}", "func (r *Router) OPTIONS(url string, viewFn View) *Path {\n\treturn r.Path(fasthttp.MethodOptions, url, viewFn)\n}", "func (router *Router) OPTIONS(relativePath string, handler Handler, decorators ...Decorator) {\n\trouter.createRouter(http.MethodOptions, relativePath, handler, \"\", decorators...)\n}", "func (c *Client) Options(url string, headers, queryParams map[string][]string) (response *http.Response, err error) {\n\treturn c.makeRequest(url, http.MethodOptions, headers, queryParams, nil)\n}", "func (engine *Engine) OPTIONS(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn engine.handle(http.MethodOptions, relativePath, handlers)\n}", "func (e *Engine) OPTIONS(path string, handler Handler) {\n\te.registerRoute(http.MethodOptions, path, handler)\n}", "func (c *Client) Options(ctx context.Context, url string, data ...interface{}) (*Response, error) {\n\treturn c.DoRequest(ctx, http.MethodOptions, url, data...)\n}", "func (f *Fastglue) OPTIONS(path string, h FastRequestHandler) {\n\tf.Router.OPTIONS(path, f.handler(h))\n}", "func (r *Router) OPTIONS(path string, handle HandlerFunc, middleware ...MiddlewareFunc) {\n\tr.Handle(\"OPTIONS\", path, handle, middleware...)\n}", "func (group *RouterGroup) OPTIONS(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn group.handle(http.MethodOptions, relativePath, handlers)\n}", "func (F *Frisby) Options(url string) *Frisby {\n\tF.Method = \"OPTIONS\"\n\tF.Url = url\n\treturn F\n}", "func (group *RouterGroup) OPTIONS(path string, handlers ...HandlerFunc) {\n\tgroup.Handle(\"OPTIONS\", path, handlers)\n}", "func (r *bitroute) OPTIONS(path string, f func(router.Control)) {\n\tr.register(\"OPTIONS\", path, f)\n}", "func (r *Router) OPTIONS(path string, h HandlerFunc) {\n\tr.router.OPTIONS(path, r.handle(h, r.getValidationForPath(path, \"OPTIONS\")))\n}", "func Options(c *gin.Context) {\n\tif c.Request.Method != \"OPTIONS\" {\n\t\tc.Next()\n\t} else {\n\t\tc.Header(\"Access-Control-Allow-Origin\", \"*\")\n\t\tc.Header(\"Access-Control-Allow-Methods\", \"GET,POST,PUT,PATCH,DELETE,OPTIONS\")\n\t\tc.Header(\"Access-Control-Allow-Headers\", \"authorization,origin,content-type,accept\")\n\t\tc.Header(\"Allow\", \"HEAD,GET,POST,PUT,PATCH,DELETE,OPTIONS\")\n\t\tc.Header(\"Content-Type\", \"application/json\")\n\t\tc.AbortWithStatus(200)\n\t}\n}", "func (router *Router) Options(path string, handler http.Handler) {\n\trouter.Handle(\"OPTIONS\", path, handler)\n}", "func (rg *RouteGroup) OPTIONS(path string, handlers ...Handler) *Route {\n\treturn rg.add(\"OPTIONS\", path, handlers)\n}", "func (r *Router) OPTIONS(path string, handle Handle) {\n\tr.Handle(http.MethodOptions, path, handle)\n}", "func (s *Server) OPTIONS(path string, handle http.HandlerFunc) {\n\ts.router.OPTIONS(\"/*path\", s.wrapHandler(handle))\n}", "func (r *Mux) OPTIONS(path string, handler http.HandlerFunc) {\n\tr.HandleFunc(\"OPTIONS\", path, handler)\n}", "func (n *NS) Options(url string) *route {\n\treturn &route{context: n.c, method: \"OPTIONS\", url: n.prefix + url, bCtrl: n.bCtrl, aCtrl: n.aCtrl}\n}", "func (r *Request) Options(url string) (*Response, error) {\n\treturn r.Execute(MethodOptions, url)\n}", "func (rb *RequestBuilder) Options(url string) *Response {\n\treturn rb.DoRequest(http.MethodOptions, url, nil)\n}", "func (g *Group) OPTIONS(path string, h Handler, gases ...Gas) {\n\tg.Air.OPTIONS(g.Prefix+path, h, append(g.Gases, gases...)...)\n}", "func Options(path string, fn httprouter.Handle) {\n\tr.Router.OPTIONS(path, fn)\n}", "func Option(c *gin.Context) {\n\tif c.Request.Method != \"OPTIONS\" {\n\t\tc.Next()\n\t} else {\n\t\tc.Header(\"Access-Control-Allow-Origin\", \"*\")\n\t\tc.Header(\"Access-Control-Allow-Methods\", \"GET,POST,PUT,OPTIONS,PATCH,DELETE\")\n\t\tc.Header(\"Access-Control-Allow-Headers\", \"authorization,origin,content-type,accept\")\n\t\tc.Header(\"Allow\", \"HEAD,GET,POST,PUT,OPTIONS,PATCH,DELETE\")\n\t\tc.Header(\"Content-Cype\", \"application/json\")\n\t\tc.AbortWithStatus(200)\n\t}\n\n}", "func (this *Route) OPTIONS(handlers ...HTTPHandler) *Route {\n\treturn this.addHandler(\"options\", handlers...)\n}", "func (r *Router) OPTIONS(route string, handler interface{}) *Router {\n\tr.Handle(MethodOPTIONS, route, handler)\n\treturn r\n}", "func (g *RouterGroup) OPTIONS(url string, handler ...Handler) *RouterGroup {\n\tg.app.routeANY = false\n\tg.AppendReqAndResp(url, \"options\", handler)\n\treturn g\n}", "func OptionsHandler(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Add(\"Allow\", http.MethodGet)\n\tw.Header().Add(\"Allow\", http.MethodHead)\n\tw.Header().Add(\"Allow\", http.MethodPost)\n\tw.Header().Add(\"Allow\", http.MethodPut)\n\tw.Header().Add(\"Allow\", http.MethodDelete)\n\tw.Header().Add(\"Allow\", http.MethodOptions)\n\tw.Header().Set(\"Transfer-Encoding\", \"chunked\")\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tw.WriteHeader(http.StatusNoContent)\n}", "func Options(ctx context.Context, url string, options ...RequestOption) (*Response, error) {\n\tr, err := newRequest(ctx, http.MethodConnect, url, nil, options...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn doRequest(http.DefaultClient, r)\n}", "func OptionsBytes(url string, data ...interface{}) []byte {\n\treturn RequestBytes(\"OPTIONS\", url, data...)\n}", "func (r *Route) Options(h http.Handler) *Route {\n\tr.Add(\"OPTIONS\", h)\n\n\treturn r\n}", "func (app *App) OPTIONS(url string, handler ...Handler) *App {\n\tapp.routeANY = false\n\tapp.AppendReqAndResp(url, \"options\", handler)\n\treturn app\n}", "func (c *Client) Request(verb, spath string, ro *RequestOptions) (*http.Request, error) {\n\tlog.Printf(\"[INFO] request: %s %s\", verb, spath)\n\n\tif ro == nil {\n\t\tro = new(RequestOptions)\n\t}\n\n\tu := *c.URL\n\tu.Path = path.Join(c.URL.Path, spath)\n\n\tif c.Token != \"\" {\n\t\tif ro.Headers == nil {\n\t\t\tro.Headers = make(map[string]string)\n\t\t}\n\t\tro.Headers[\"Authorization\"] = fmt.Sprintf(\"Bearer %s\", c.Token)\n\t}\n\n\treturn c.rawRequest(verb, &u, ro)\n}", "func (r *Router) Options(pattern string, handlers ...interface{}) *Router {\n\tc := r.pattern(pattern)\n\tc.Method = \"OPTIONS\"\n\tclone := r.clone(c)\n\tclone.handle(handlers)\n\treturn clone\n}", "func (rg *RouteGroup) Options(path string, handlers ...Handler) *Route {\n\treturn rg.add(\"OPTIONS\", path, handlers)\n}", "func (r *Router) OPTIONS(pattern string, handler func(Context)) Route {\n\treturn r.Handle(\"OPTIONS\", pattern, wrap(handler))\n}", "func Options(c *gin.Context) {\r\n\tc.Writer.Header().Set(\"Access-Control-Allow-Methods\", \"DELETE,POST,PUT\")\r\n\tc.Writer.Header().Set(\"Access-Control-Allow-Headers\", \"Content-Type\")\r\n\tc.Next()\r\n}", "func (r *Router) Options(pattern string, handler http.Handler) Route {\n\treturn r.Handle(\"OPTIONS\", pattern, handler)\n}", "func (s *System) Options(path string) string, error {\n\treq, err := http.NewRequest(http.MethodOptions, \"https://\"+s.host+\"/bps/api/v2/core/\"+path, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t//add the header\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tres, err := s.HTTPClient.Do(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer res.Body.Close()\n\n\tif res.StatusCode != http.StatusNoContent {\n\t\treturn fmt.Errorf(\"unknown error, status code: %d\", res.StatusCode)\n\t}\n\n\treturn nil\n}", "func (router *Router) OptionsFunc(path string, handler http.HandlerFunc) {\n\trouter.Handle(\"OPTIONS\", path, handler)\n}", "func (mm *Middleman) Options(path string, middleware Middleware) error {\r\n\treturn mm.addMiddleware(path, http.MethodOptions, middleware)\r\n}", "func (APIResourceBase) Options(session *Session, url string, queries url.Values, body io.Reader) (APIStatus, interface{}) {\n\treturn FailSimple(http.StatusMethodNotAllowed), nil\n}", "func (tr *Transport) Options(url string, fn HandlerFunc, options ...HandlerOption) {\n\ttr.mux.Handler(net_http.MethodOptions, url, encapsulate(fn, tr.options, options))\n}", "func OptionsResponse(w http.ResponseWriter, r *http.Request, argv map[string]string) error {\n\tw.Header().Set(\"Allow\", \"GET,PUT,POST,DELETE,OPTIONS\")\n\n\treturn nil\n}", "func OptionsContent(url string, data ...interface{}) string {\n\treturn RequestContent(\"OPTIONS\", url, data...)\n}", "func (r *Router) Options(path string, handle HandleFunc) {\n\tr.register(path, http.MethodOptions, handle)\n}", "func DefaultOptions(\n server Server,\n transactor *Transactor,\n methods []*RouteControllerHelper,\n) responses.Data {\n var methodStrings []string\n seenMethods := map[RequestMethod]bool{\n MethodOPTIONS: true,\n }\n for _, rch := range(methods) {\n for _, method := range rch.AllMethods() {\n if _, ok := seenMethods[method]; !ok {\n methodStrings = append(methodStrings, method.String())\n seenMethods[method] = true\n }\n }\n }\n sort.Strings(methodStrings)\n methodStrings = append([]string{MethodOPTIONS.String()}, methodStrings...)\n err := transactor.SetHeader(\n \"Access-Control-Allow-Methods\",\n strings.Join(methodStrings, \", \"),\n )\n if err != nil {\n return transactor.Abort(\n http.StatusInternalServerError,\n neterr.DefaultOptionsHeaderSetError,\n neterr.CodedErrorFromError(0, err),\n )\n }\n\n return transactor.Respond(http.StatusOK)\n}", "func (r *Router) Options(path, title string, fn Handle) {\n\tr.addRoute(\"OPTIONS\", path, title, fn)\n}", "func (m *MockRouterTx) OPTIONS(path string, handler interface{}, options ...interface{}) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{path, handler}\n\tfor _, a := range options {\n\t\tvarargs = append(varargs, a)\n\t}\n\tm.ctrl.Call(m, \"OPTIONS\", varargs...)\n}", "func (r *Route) Options(h interface{}) *Route {\n\treturn r.Handle(toHandler(h), \"OPTIONS\")\n}", "func (m *MockRouter) OPTIONS(path string, handler interface{}, options ...interface{}) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{path, handler}\n\tfor _, a := range options {\n\t\tvarargs = append(varargs, a)\n\t}\n\tm.ctrl.Call(m, \"OPTIONS\", varargs...)\n}", "func (server *HTTPServer) handleOptionsMethod(nextHandler http.Handler) http.Handler {\n\treturn http.HandlerFunc(\n\t\tfunc(w http.ResponseWriter, r *http.Request) {\n\t\t\tif r.Method == \"OPTIONS\" {\n\t\t\t\tw.WriteHeader(http.StatusOK)\n\t\t\t} else {\n\t\t\t\tnextHandler.ServeHTTP(w, r)\n\t\t\t}\n\t\t})\n}", "func (server *HTTPServer) handleOptionsMethod(nextHandler http.Handler) http.Handler {\n\treturn http.HandlerFunc(\n\t\tfunc(w http.ResponseWriter, r *http.Request) {\n\t\t\tif r.Method == \"OPTIONS\" {\n\t\t\t\tw.WriteHeader(http.StatusOK)\n\t\t\t} else {\n\t\t\t\tnextHandler.ServeHTTP(w, r)\n\t\t\t}\n\t\t})\n}", "func Options(c *gin.Context) {\n\tOrigin := c.MustGet(\"CorsOrigin\").(string)\n\n\tc.Writer.Header().Set(\"Access-Control-Allow-Origin\", Origin)\n\tc.Writer.Header().Set(\"Access-Control-Allow-Methods\", \"GET,DELETE,POST,PUT\")\n\tc.Writer.Header().Set(\"Access-Control-Allow-Credentials\", \"true\")\n\tc.Writer.Header().Set(\"Access-Control-Allow-Headers\", \"Content-Type\")\n\tc.Next()\n}", "func (p *PatternServeMux) Options(pat string, h http.Handler) {\n\tp.Add(\"OPTIONS\", pat, h)\n}", "func (b *Baa) Options(pattern string, h ...HandlerFunc) RouteNode {\n\treturn b.Router().Add(\"OPTIONS\", pattern, h)\n}", "func RequestOptions(opts ...RequestOpt) Opt {\n\tif len(opts) == 0 {\n\t\treturn nil\n\t}\n\treturn func(c *Client) Opt {\n\t\told := append([]RequestOpt{}, c.requestOpts...)\n\t\tc.requestOpts = opts\n\t\treturn RequestOptions(old...)\n\t}\n}", "func (r *Router) Options(path string, handlers ...Handler) {\n\tr.addToRoutes(http.MethodOptions, path, handlers...)\n}", "func OptionsHandler(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Access-Control-Allow-Methods\", \"GET, POST, PUT\")\n\tw.Header().Set(\"Access-Control-Max-Age\", \"120\")\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tw.Header().Set(\"Access-Control-Allow-Headers\", \"content-type\")\n\tw.WriteHeader(200)\n}", "func (tr *Transport) OPTION(\n\turl string,\n\tfn Handler,\n\toptions ...HandlerOption,\n) {\n\ttr.mux.Handler(\n\t\tnet_http.MethodOptions,\n\t\turl,\n\t\tnewHandler(fn, append(tr.options, options...)...),\n\t)\n}", "func (m *Mux) Options(pattern string, handler HandlerFunc) {\n\tm.Handle(http.MethodOptions, pattern, handler)\n}", "func (mx *Mux) Options(pattern string, handlerFn http.HandlerFunc) {\n\tmx.handle(mOPTIONS, pattern, handlerFn)\n}", "func (c *ClientConn) Options(u *base.URL) (*base.Response, error) {\n\terr := c.checkState(map[clientConnState]struct{}{\n\t\tclientConnStateInitial: {},\n\t\tclientConnStatePrePlay: {},\n\t\tclientConnStatePreRecord: {},\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres, err := c.Do(&base.Request{\n\t\tMethod: base.Options,\n\t\tURL: u,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif res.StatusCode != base.StatusOK {\n\t\t// since this method is not implemented by every RTSP server,\n\t\t// return only if status code is not 404\n\t\tif res.StatusCode == base.StatusNotFound {\n\t\t\treturn res, nil\n\t\t}\n\t\treturn res, liberrors.ErrClientWrongStatusCode{Code: res.StatusCode, Message: res.StatusMessage}\n\t}\n\n\tc.getParameterSupported = func() bool {\n\t\tpub, ok := res.Header[\"Public\"]\n\t\tif !ok || len(pub) != 1 {\n\t\t\treturn false\n\t\t}\n\n\t\tfor _, m := range strings.Split(pub[0], \",\") {\n\t\t\tif base.Method(m) == base.GetParameter {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t\treturn false\n\t}()\n\n\treturn res, nil\n}", "func (m *Mux) OptionsFunc(path string, handler http.HandlerFunc) *Route {\n\treturn m.register(\"OPTIONS\", path, handler)\n}", "func Options(url string, opts ...RequestOption) (*Response, error) {\n\treturn DefaultSession.Options(url, opts...)\n}", "func SendOptions(ctx context.Context, response http.ResponseWriter,\n\trequest *http.Request, pipeline filter.Pipeline) {\n\tresponse.Header().Add(\"Allow\", \"OPTIONS, GET, HEAD\")\n\tresponse.WriteHeader(http.StatusNoContent)\n\tmedia := strings.NewReader(\"\")\n\terr := error(nil)\n\tif len(pipeline) > 0 {\n\t\t// use a filter pipeline\n\t\t_, err = filter.PipelineCopy(ctx, response, media, request, pipeline)\n\t} else {\n\t\t// unfiltered, simple copy\n\t\t_, err = io.Copy(response, media)\n\t}\n\tif err != nil {\n\t\tlog.Error().Msgf(\"SendOptions: %v\", err)\n\t}\n}", "func (mr *MockRouterMockRecorder) OPTIONS(path, handler interface{}, options ...interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\tvarargs := append([]interface{}{path, handler}, options...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"OPTIONS\", reflect.TypeOf((*MockRouter)(nil).OPTIONS), varargs...)\n}", "func (mr *MockRouterTxMockRecorder) OPTIONS(path, handler interface{}, options ...interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\tvarargs := append([]interface{}{path, handler}, options...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"OPTIONS\", reflect.TypeOf((*MockRouterTx)(nil).OPTIONS), varargs...)\n}", "func (r *Route) Options(handler http.Handler) *Route {\n\tr.handlers[http.MethodOptions] = handler\n\treturn r\n}", "func (s *Nap) Options(pathURL string) *Nap {\n\ts.method = MethodOptions\n\treturn s.Path(pathURL)\n}", "func (ctl Controller) Options(ctx *gin.Context) {\n\n\tMethodNotAllowedJSON(ctx)\n}", "func (c *BaseController) Options(w http.ResponseWriter, r *Request) {\n\thttp.Error(w, \"Method Not Allowed\", 405)\n}", "func TestOptionsFilter(t *testing.T) {\n\ttearDown()\n\tws := new(WebService)\n\tws.Route(ws.GET(\"/candy/{kind}\").To(dummy))\n\tws.Route(ws.DELETE(\"/candy/{kind}\").To(dummy))\n\tws.Route(ws.POST(\"/candies\").To(dummy))\n\tAdd(ws)\n\tFilter(OPTIONSFilter())\n\n\thttpRequest, _ := http.NewRequest(\"OPTIONS\", \"http://here.io/candy/gum\", nil)\n\thttpWriter := httptest.NewRecorder()\n\tDefaultContainer.dispatch(httpWriter, httpRequest)\n\tactual := httpWriter.Header().Get(HEADER_Allow)\n\tif \"GET,DELETE\" != actual {\n\t\tt.Fatal(\"expected: GET,DELETE but got:\" + actual)\n\t}\n\n\thttpRequest, _ = http.NewRequest(\"OPTIONS\", \"http://here.io/candies\", nil)\n\thttpWriter = httptest.NewRecorder()\n\tDefaultContainer.dispatch(httpWriter, httpRequest)\n\tactual = httpWriter.Header().Get(HEADER_Allow)\n\tif \"POST\" != actual {\n\t\tt.Fatal(\"expected: POST but got:\" + actual)\n\t}\n}", "func handleOptions(w http.ResponseWriter, r *http.Request) {\n\tfmt.Println(\"Handle options!\")\n\tw.Header().Set(\"Access-Control-Allow-Credentials\", \"false\")\n\tw.Header().Set(\"Access-Control-Allow-Headers\", \"Accept, Content-Type, Pragma, X-Requested-With\")\n\tw.Header().Set(\"Access-Control-Allow-Methods\", \"POST, GET, PUT, DELETE, OPTIONS\")\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tw.Header().Set(\"Access-Control-Max-Age\", \"86400\")\n\tw.WriteHeader(http.StatusOK)\n\treturn\n}", "func (o *OptionsFilter) Filter(req *restful.Request, resp *restful.Response, chain *restful.FilterChain) {\n\tif \"OPTIONS\" != req.Request.Method {\n\t\tchain.ProcessFilter(req, resp)\n\t\treturn\n\t}\n\n\tarchs := req.Request.Header.Get(restful.HEADER_AccessControlRequestHeaders)\n\t//methods := strings.Join(o.getContainer().ComputeAllowedMethods(req), \",\")\n\torigin := req.Request.Header.Get(restful.HEADER_Origin)\n\tif len(origin) == 0 {\n\t\torigin = \"*\"\n\t}\n\tlog.Printf(\"request >> origin:%s\\n\", origin)\n\tresp.AddHeader(restful.HEADER_Allow, \"*\")\n\tresp.AddHeader(restful.HEADER_AccessControlAllowOrigin, origin)\n\tresp.AddHeader(restful.HEADER_AccessControlAllowHeaders, archs)\n\tresp.AddHeader(restful.HEADER_AccessControlAllowMethods, \"*\")\n\tresp.AddHeader(restful.HEADER_AccessControlAllowCredentials, \"true\")\n}", "func AccessControlRequestMethod(value string) Option {\n\treturn setHeader(\"Access-Control-Request-Method\", value)\n}", "func (s *HTTPServer) preparedQuerySpecificOptions(resp http.ResponseWriter, req *http.Request) interface{} {\n\tpath := req.URL.Path\n\tswitch {\n\tcase strings.HasSuffix(path, \"/execute\"):\n\t\tresp.Header().Add(\"Allow\", strings.Join([]string{\"OPTIONS\", \"GET\"}, \",\"))\n\t\treturn resp\n\n\tcase strings.HasSuffix(path, \"/explain\"):\n\t\tresp.Header().Add(\"Allow\", strings.Join([]string{\"OPTIONS\", \"GET\"}, \",\"))\n\t\treturn resp\n\n\tdefault:\n\t\tresp.Header().Add(\"Allow\", strings.Join([]string{\"OPTIONS\", \"GET\", \"PUT\", \"DELETE\"}, \",\"))\n\t\treturn resp\n\t}\n}", "func (c *Client) HTTP(verb, spath string, ro *RequestOptions) (*http.Response, error) {\n\treq, err := c.Request(verb, spath, ro)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres, err := dispose(c.HTTPClient.Do(req))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn res, nil\n}", "func OptGet() Option {\n\treturn func(r *Request) error {\n\t\tr.Method = \"GET\"\n\t\treturn nil\n\t}\n}", "func (r *Request) Get(path string) *Request {\n\treturn r.method(\"GET\", path)\n}", "func (mx *Mux) Options(pattern string, handler interface{}) {\n\tmx.handle(OPTIONS, pattern, handler)\n}", "func OptionsUser(c *gin.Context) {\n\tc.Writer.Header().Set(\"Access-Control-Allow-Methods\", \"DELETE,POST, PUT\")\n\tc.Writer.Header().Set(\"Access-Control-Allow-Headers\", \"Content-Type\")\n\tc.Next()\n}", "func (ctx *AppContext) OptionsHandler(h http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(writer http.ResponseWriter, request *http.Request) {\n\t\tif request.Method == \"OPTIONS\" {\n\t\t\tfmt.Fprintf(writer, \"\")\n\t\t\treturn\n\t\t}\n\n\t\th.ServeHTTP(writer, request)\n\t})\n}", "func (e *Expect) GET(path string, pathargs ...interface{}) *Request {\n\treturn e.Request(http.MethodGet, path, pathargs...)\n}", "func (api *Client) Call(opts *Opts) (resp *http.Response, err error) {\n\tapi.mu.RLock()\n\tdefer api.mu.RUnlock()\n\tif opts == nil {\n\t\treturn nil, errors.New(\"call() called with nil opts\")\n\t}\n\turl := api.rootURL\n\tif opts.RootURL != \"\" {\n\t\turl = opts.RootURL\n\t}\n\tif url == \"\" {\n\t\treturn nil, errors.New(\"RootURL not set\")\n\t}\n\turl += opts.Path\n\tif opts.Parameters != nil && len(opts.Parameters) > 0 {\n\t\turl += \"?\" + opts.Parameters.Encode()\n\t}\n\treq, err := http.NewRequest(opts.Method, url, opts.Body)\n\tif err != nil {\n\t\treturn\n\t}\n\theaders := make(map[string]string)\n\t// Set default headers\n\tfor k, v := range api.headers {\n\t\theaders[k] = v\n\t}\n\tif opts.ContentType != \"\" {\n\t\theaders[\"Content-Type\"] = opts.ContentType\n\t}\n\tif opts.ContentLength != nil {\n\t\treq.ContentLength = *opts.ContentLength\n\t}\n\tif opts.ContentRange != \"\" {\n\t\theaders[\"Content-Range\"] = opts.ContentRange\n\t}\n\t// Set any extra headers\n\tif opts.ExtraHeaders != nil {\n\t\tfor k, v := range opts.ExtraHeaders {\n\t\t\theaders[k] = v\n\t\t}\n\t}\n\t// add any options to the headers\n\tfs.OpenOptionAddHeaders(opts.Options, headers)\n\t// Now set the headers\n\tfor k, v := range headers {\n\t\tif v != \"\" {\n\t\t\treq.Header.Add(k, v)\n\t\t}\n\t}\n\tif opts.UserName != \"\" || opts.Password != \"\" {\n\t\treq.SetBasicAuth(opts.UserName, opts.Password)\n\t}\n\tc := ClientWithHeaderReset(api.c, headers)\n\tapi.mu.RUnlock()\n\tresp, err = c.Do(req)\n\tapi.mu.RLock()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !opts.IgnoreStatus {\n\t\tif resp.StatusCode < 200 || resp.StatusCode > 299 {\n\t\t\treturn resp, api.errorHandler(resp)\n\t\t}\n\t}\n\tif opts.NoResponse {\n\t\treturn resp, resp.Body.Close()\n\t}\n\treturn resp, nil\n}", "func digestOptions()(reqOpts RequestOptions, outOpts OutputOptions, err error) {\n\treturn RequestOptions{\n\t\tMethod : \"GET\",\n\t\tURL : \"localhost:8080/test\",\n\t}, OutputOptions {\n\t\tOutputHTML : true,\n\t\tShowFullJSON : true,\n\t}, nil\n}", "func Method(method string, urlPath string) (Request, error) {\n\tif method != \"GET\" && method != \"POST\" && method != \"PUT\" && method != \"DELETE\" &&\n\t\tmethod != \"HEAD\" && method != \"OPTIONS\" && method != \"PATCH\" {\n\t\treturn nil, errors.New(\"method not supported\")\n\t}\n\treturn newRequest(method, urlPath, sessionWithoutCookies)\n}", "func (e *Expect) PATCH(path string, pathargs ...interface{}) *Request {\n\treturn e.Request(http.MethodPatch, path, pathargs...)\n}", "func HttpOptions(httpOpts *server.Options) Option {\n\treturn func(c *Service) {\n\t\tc.serverOptions = httpOpts\n\t}\n}", "func optionsHandler(ctx apiContext, w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Access-Control-Allow-Origin\", originAddress)\n\tw.Header().Set(\"Access-Control-Allow-Headers\", supportedHeaders)\n\tw.Header().Set(\"Access-Control-Allow-Credentials\", \"true\")\n\tw.Header().Set(\"Access-Control-Allow-Methods\", methodsToString(ctx.supportedMethods))\n\tw.WriteHeader(http.StatusOK)\n}", "func (srv *Server) handleOptions(res http.ResponseWriter, req *http.Request) {\n\tmethods := make(map[string]bool)\n\n\tnode := srv.getFSNode(req.URL.Path)\n\tif node != nil {\n\t\tmethods[http.MethodGet] = true\n\t\tmethods[http.MethodHead] = true\n\t}\n\n\tfor _, rute := range srv.routeDeletes {\n\t\t_, ok := rute.parse(req.URL.Path)\n\t\tif ok {\n\t\t\tmethods[http.MethodDelete] = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\tfor _, rute := range srv.routeGets {\n\t\t_, ok := rute.parse(req.URL.Path)\n\t\tif ok {\n\t\t\tmethods[http.MethodGet] = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\tfor _, rute := range srv.routePatches {\n\t\t_, ok := rute.parse(req.URL.Path)\n\t\tif ok {\n\t\t\tmethods[http.MethodPatch] = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\tfor _, rute := range srv.routePosts {\n\t\t_, ok := rute.parse(req.URL.Path)\n\t\tif ok {\n\t\t\tmethods[http.MethodPost] = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\tfor _, rute := range srv.routePuts {\n\t\t_, ok := rute.parse(req.URL.Path)\n\t\tif ok {\n\t\t\tmethods[http.MethodPut] = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif len(methods) == 0 {\n\t\tres.WriteHeader(http.StatusNotFound)\n\t\treturn\n\t}\n\n\tmethods[http.MethodOptions] = true\n\n\tvar x int\n\tallows := make([]string, len(methods))\n\tfor k, v := range methods {\n\t\tif v {\n\t\t\tallows[x] = k\n\t\t\tx++\n\t\t}\n\t}\n\n\tsort.Strings(allows)\n\n\tres.Header().Set(\"Allow\", strings.Join(allows, \", \"))\n\tres.WriteHeader(http.StatusOK)\n}", "func OptionsAnswer(w http.ResponseWriter) {\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tw.Header().Set(\"Access-Control-Allow-Headers\", \"*\")\n\tw.Header().Set(\"Access-Control-Allow-Methods\", \"*\")\n}", "func OptionsUser(c *gin.Context) {\n\tc.Writer.Header().Add(\"Access-Control-Allow-Origin\", \"*\")\n\tc.Writer.Header().Set(\"Access-Control-Allow-Methods\", \"GET, DELETE, POST, PATCH, PUT, OPTIONS\")\n\tc.Writer.Header().Set(\"Access-Control-Allow-Headers\", \"Content-Type\")\n\tc.Writer.Header().Set(\"Access-Control-Allow-Headers\", \"X-Requested-With\")\n\tc.Next()\n}", "func ReqOpts(hdr http.Header) *toclient.RequestOptions {\n\topts := toclient.NewRequestOptions()\n\topts.Header = hdr\n\treturn &opts\n}" ]
[ "0.76900136", "0.7574512", "0.7364902", "0.7281106", "0.7234489", "0.7181617", "0.71771985", "0.71593726", "0.71545005", "0.7152557", "0.7115014", "0.70317966", "0.7006474", "0.697124", "0.69393027", "0.69371873", "0.69233954", "0.69009876", "0.6857603", "0.6837377", "0.6830049", "0.68299896", "0.68084", "0.6745392", "0.6663974", "0.66606045", "0.6617182", "0.6592707", "0.65867925", "0.65356016", "0.6523823", "0.65135735", "0.6508351", "0.64645416", "0.6446378", "0.6445187", "0.64395785", "0.6437644", "0.6435132", "0.63730514", "0.62997985", "0.6299091", "0.6237802", "0.6229727", "0.61944234", "0.6187936", "0.61652786", "0.6124756", "0.6122601", "0.61160296", "0.61151344", "0.61073035", "0.6058957", "0.60516125", "0.6046014", "0.60429704", "0.60131854", "0.60056454", "0.60056454", "0.59670436", "0.59312755", "0.592525", "0.5909391", "0.59046376", "0.5877589", "0.5865045", "0.58533645", "0.58007663", "0.579405", "0.57740045", "0.5767243", "0.57410824", "0.5734779", "0.57119584", "0.57031745", "0.5680191", "0.56750613", "0.56686246", "0.56661975", "0.56444114", "0.56357545", "0.5629346", "0.5611519", "0.5597881", "0.55878127", "0.55825984", "0.5556712", "0.5546201", "0.5519237", "0.5512131", "0.5511339", "0.5502821", "0.5473568", "0.546432", "0.5430084", "0.54158187", "0.5407375", "0.5397371", "0.5378444", "0.53723645" ]
0.8413449
0
HEAD is a shorthand for e.Request("HEAD", path, pathargs...).
func (e *Expect) HEAD(path string, pathargs ...interface{}) *Request { return e.Request(http.MethodHead, path, pathargs...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *Request) Head(path string) *Request {\n\treturn r.method(\"HEAD\", path)\n}", "func (r *Request) Head(path string, params ...url.Values) {\n\tcontentType := \"text/html\"\n\n\tif len(params) == 0 {\n\t\tr.Send(\"HEAD\", path, contentType)\n\t} else {\n\t\tr.Send(\"HEAD\", path, contentType, params[0])\n\t}\n}", "func Head(url string, data ...interface{}) (*ClientResponse, error) {\n\treturn DoRequest(\"HEAD\", url, data...)\n}", "func Head(path string, fn http.HandlerFunc, c ...alice.Constructor) {\n\trecord(\"HEAD\", path)\n\n\tinfoMutex.Lock()\n\tr.HEAD(path, Handler(alice.New(c...).ThenFunc(fn)))\n\tinfoMutex.Unlock()\n}", "func (r *Router) HEAD(url string, viewFn View) *Path {\n\treturn r.Path(fasthttp.MethodHead, url, viewFn)\n}", "func (F *Frisby) Head(url string) *Frisby {\n\tF.Method = \"HEAD\"\n\tF.Url = url\n\treturn F\n}", "func (rb *RequestBuilder) Head(url string) *Response {\n\treturn rb.DoRequest(http.MethodHead, url, nil)\n}", "func (tr *Transport) HEAD(\n\turl string,\n\tfn Handler,\n\toptions ...HandlerOption,\n) {\n\ttr.mux.Handler(\n\t\tnet_http.MethodHead,\n\t\turl,\n\t\tnewHandler(fn, append(tr.options, options...)...),\n\t)\n}", "func (r *Request) Head(url string) (*Response, error) {\n\treturn r.Execute(MethodHead, url)\n}", "func NewHead(url string) *Request { return NewRequest(\"HEAD\", url) }", "func (cl *Client) Head(c context.Context, url string, opts ...RequestOption) (*Response, error) {\n\treq, err := cl.NewRequest(c, http.MethodHead, url, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn cl.Do(c, req)\n}", "func (r *Router) HEAD(path string, h HandlerFunc) {\n\tr.router.HEAD(path, r.handle(h, r.getValidationForPath(path, \"HEAD\")))\n}", "func (c *Client) Head(url string) (*http.Response, error) {\n\treq, err := http.NewRequest(http.MethodHead, url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn c.Do(req)\n}", "func (f *Fastglue) HEAD(path string, h FastRequestHandler) {\n\tf.Router.HEAD(path, f.handler(h))\n}", "func (c *Client) Head(ctx context.Context, url string, data ...interface{}) (*Response, error) {\n\treturn c.DoRequest(ctx, http.MethodHead, url, data...)\n}", "func (serv *Server) HEAD(url string, handlers ...Handler) {\n\tserv.Handle(\"HEAD\", url, handlers...)\n}", "func (a *API) Head(path string) (resp *http.Response, err error) {\n\tu, err := url.ParseRequestURI(a.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tu.Path = path\n\n\treturn a.Client.Head(u.String())\n}", "func (g *Group) HEAD(path string, handler Handler, middleware ...Middleware) *Group {\n\treturn g.Add(http.MethodHead, path, handler, middleware...)\n}", "func (c *Client) Head(url string) (*Response, error) {\n\treq, err := c.NewRequest(\"HEAD\", url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresponse, err := c.Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn response, nil\n}", "func (r *Router) HEAD(path string, handle Handle) {\n\tr.Handle(http.MethodHead, path, handle)\n}", "func (e *Engine) HEAD(path string, handler Handler) {\n\te.registerRoute(http.MethodHead, path, handler)\n}", "func (c *Client) Head(ctx context.Context, hosts []string, path string) (http.Header, int, error) {\n\tresp, err := c.executeRequest(ctx, http.MethodHead, hosts, path, nil)\n\tif err != nil {\n\t\treturn nil, 0, errors.WithStack(err)\n\t}\n\tdefer resp.Body.Close()\n\treturn resp.Header, resp.StatusCode, nil\n}", "func Head(url string) (resp *http.Response, err error) {\n\treturn DefaultClient.Head(url)\n}", "func (c *Client) Head(url string) (response *http.Response, err error) {\n\treturn http.Head(url)\n}", "func (r *Router) HEAD(path string, handle HandlerFunc, middleware ...MiddlewareFunc) {\n\tr.Handle(\"HEAD\", path, handle, middleware...)\n}", "func Head(ctx context.Context, url string, options ...RequestOption) (*Response, error) {\n\tr, err := newRequest(ctx, http.MethodGet, url, nil, options...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn doRequest(http.DefaultClient, r)\n}", "func (m *Middleware) HEAD(path string, handler Handler) {\n\tm.engine.HEAD(path, m.apply(handler))\n}", "func (r *Mux) HEAD(path string, handler http.HandlerFunc) {\n\tr.HandleFunc(\"HEAD\", path, handler)\n}", "func (engine *Engine) HEAD(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn engine.handle(http.MethodHead, relativePath, handlers)\n}", "func Head(path string, fn httprouter.Handle) {\n\tr.Router.HEAD(path, fn)\n}", "func (d *Doer) Head(url string) (*http.Response, error) {\n\treq, err := d.newRequest(http.MethodHead, url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn d.do(req, nil)\n}", "func (router *Router) HEAD(relativePath string, handler Handler, decorators ...Decorator) {\n\trouter.createRouter(http.MethodHead, relativePath, handler, \"\", decorators...)\n}", "func Head(url string) (*http.Response, error) {\n\treturn DefaultClient.Head(url)\n}", "func (router *Router) Head(path string, handler http.Handler) {\n\trouter.Handle(\"HEAD\", path, handler)\n}", "func (s *Nap) Head(pathURL string) *Nap {\n\ts.method = MethodHead\n\treturn s.Path(pathURL)\n}", "func (r *bitroute) HEAD(path string, f func(router.Control)) {\n\tr.register(\"HEAD\", path, f)\n}", "func Head(url string, opts ...RequestOption) (*Response, error) {\n\treturn DefaultSession.Head(url, opts...)\n}", "func (o *Object) head(ctx context.Context) error {\n\tif o.fs.opt.NoHead {\n\t\to.size = -1\n\t\to.modTime = timeUnset\n\t\to.contentType = fs.MimeType(ctx, o)\n\t\treturn nil\n\t}\n\turl := o.url()\n\treq, err := http.NewRequestWithContext(ctx, \"HEAD\", url, nil)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"stat failed: %w\", err)\n\t}\n\to.fs.addHeaders(req)\n\tres, err := o.fs.httpClient.Do(req)\n\tif err == nil && res.StatusCode == http.StatusNotFound {\n\t\treturn fs.ErrorObjectNotFound\n\t}\n\terr = statusError(res, err)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to stat: %w\", err)\n\t}\n\treturn o.decodeMetadata(ctx, res)\n}", "func (r *Route) Head(h http.Handler) *Route {\n\tr.Add(\"HEAD\", h)\n\n\treturn r\n}", "func (c *Client) Head(url string) (*http.Response, error) {\n\tb := c.breakerLookup(url)\n\tif b == nil {\n\t\treturn c.client.Head(url)\n\t}\n\n\tctx := getHeadCtx()\n\tdefer releaseHeadCtx(ctx)\n\n\tctx.Client = c.client\n\tctx.ErrorOnBadStatus = c.errOnBadStatus\n\tctx.URL = url\n\tif err := b.Call(ctx, breaker.WithTimeout(c.timeout)); err != nil {\n\t\treturn nil, err\n\t}\n\treturn ctx.Response, ctx.Error\n}", "func (s *HTTPSet) Head(url string) (*http.Response, error) {\n\turl, err := s.replaceHost(url)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn s.HTTPClient.Head(url)\n}", "func (group *RouterGroup) HEAD(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn group.handle(http.MethodHead, relativePath, handlers)\n}", "func (tr *Transport) Head(url string, fn HandlerFunc, options ...HandlerOption) {\n\ttr.mux.Handler(net_http.MethodHead, url, encapsulate(fn, tr.options, options))\n}", "func (node *Node) HEAD(functions ...interface{}) *Handler {\n\n\thandler := &Handler{}\n\n\tif len(functions) > 0 { handler.function = functions[0].(func (req web.RequestInterface) *web.ResponseStatus) }\n\n\tnode.addHandler(\"HEAD\", handler)\n\n\treturn handler\n}", "func (b *Builder) Head(url string) *Builder {\n\tb.Url = url\n\tb.Method = http.MethodHead\n\treturn b\n}", "func (group *RouterGroup) HEAD(path string, handlers ...HandlerFunc) {\n\tgroup.Handle(\"HEAD\", path, handlers)\n}", "func (app *App) HEAD(url string, handler ...Handler) *App {\n\tapp.routeANY = false\n\tapp.AppendReqAndResp(url, \"head\", handler)\n\treturn app\n}", "func (g *RouterGroup) HEAD(url string, handler ...Handler) *RouterGroup {\n\tg.app.routeANY = false\n\tg.AppendReqAndResp(url, \"head\", handler)\n\treturn g\n}", "func HeadObject(url string) (*http.Response, error) {\n\treq, err := http.NewRequest(\"HEAD\", url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := http.DefaultClient.Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func (rg *RouteGroup) Head(path string, handlers ...Handler) *Route {\n\treturn rg.add(\"HEAD\", path, handlers)\n}", "func (r *Router) Head(path string, handle HandleFunc) {\n\tr.register(path, http.MethodHead, handle)\n}", "func (r *Route) Head(h interface{}) *Route {\n\treturn r.Handle(toHandler(h), \"HEAD\")\n}", "func (mx *Mux) Head(pattern string, handlerFn http.HandlerFunc) {\n\tmx.handle(mHEAD, pattern, handlerFn)\n}", "func (rg *RouteGroup) HEAD(path string, handlers ...Handler) *Route {\n\treturn rg.add(\"HEAD\", path, handlers)\n}", "func (r *Router) HEAD(route string, handler interface{}) *Router {\n\tr.Handle(MethodHEAD, route, handler)\n\treturn r\n}", "func (r *Router) HEAD(pattern string, handler func(Context)) Route {\n\treturn r.Handle(\"HEAD\", pattern, wrap(handler))\n}", "func (router *Router) HeadFunc(path string, handler http.HandlerFunc) {\n\trouter.Handle(\"HEAD\", path, handler)\n}", "func (r *Router) Head(pattern string, handler http.Handler) Route {\n\treturn r.Handle(\"HEAD\", pattern, handler)\n}", "func HeadBytes(url string, data ...interface{}) []byte {\n\treturn RequestBytes(\"HEAD\", url, data...)\n}", "func (this *Route) HEAD(handlers ...HTTPHandler) *Route {\n\treturn this.addHandler(\"head\", handlers...)\n}", "func (r *Router) Head(pattern string, handlers ...interface{}) *Router {\n\tc := r.pattern(pattern)\n\tc.Method = \"HEAD\"\n\tclone := r.clone(c)\n\tclone.handle(handlers)\n\treturn clone\n}", "func (group *RouterGroup) HEAD(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn group.Handler(\"HEAD\", relativePath, handlers...)\n}", "func (m *Mux) Head(pattern string, handler HandlerFunc) {\n\tm.Handle(http.MethodHead, pattern, handler)\n}", "func Head(url string) *THttpClient {\r\n\treturn NewHttpClient(url).Head(\"\")\r\n}", "func doGetHead(w http.ResponseWriter, r *http.Request) {\n\ts, err := newServerCGI(w, r)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\treg := regexp.MustCompile(\"^(get|head|removed)/([0-9A-Za-z_]+)/?([-0-9A-Za-z/]*)$\")\n\tm := reg.FindStringSubmatch(s.path())\n\tif m == nil {\n\t\tlog.Println(\"illegal url\", s.path())\n\t\treturn\n\t}\n\tmethod, datfile, stamp := m[1], m[2], m[3]\n\tca := thread.NewCache(datfile)\n\tbegin, end, id := s.parseStamp(stamp, math.MaxInt32)\n\tvar recs thread.RecordMap\n\tif method == \"removed\" {\n\t\trecs = ca.LoadRemovedRecords()\n\t} else {\n\t\trecs = ca.LoadRecords()\n\t}\n\tfor _, r := range recs {\n\t\tif r.InRange(begin, end, id) {\n\t\t\tif method == \"get\" {\n\t\t\t\tif err := r.Load(); err != nil {\n\t\t\t\t\tlog.Println(err)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tfmt.Fprintln(s.wr, r.Recstr())\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tfmt.Fprintln(s.wr, strings.Replace(r.Idstr(), \"_\", \"<>\", -1))\n\t\t}\n\t}\n\tif method == \"get\" {\n\t\tthread.UpdatedRecord.Inform(datfile, id, begin, end)\n\t}\n}", "func (b *Baa) Head(pattern string, h ...HandlerFunc) RouteNode {\n\treturn b.Router().Add(\"HEAD\", pattern, h)\n}", "func HeadHandler(w http.ResponseWriter, r *http.Request) {\n\tFilesLock.RLock()\n\tval, ok := Files.Get(fileKey(r.URL))\n\tFilesLock.RUnlock()\n\n\tif !ok {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\treturn\n\t}\n\tf := val.(*File)\n\n\tw.Header().Set(\"Content-Type\", f.ContentType)\n\tw.Header().Set(\"Transfer-Encoding\", \"chunked\")\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tw.WriteHeader(http.StatusOK)\n}", "func (p *PatternServeMux) Head(pat string, h http.Handler) {\n\tp.Add(\"HEAD\", pat, h)\n}", "func head(httplink string, wg *sync.WaitGroup) {\n\twg.Add(1)\n\tdefer wg.Done()\n\toutput := Result{}\n\toutput.URL = httplink\n\ttimeout := time.Duration(time.Duration(timeoutSeconds) * time.Second)\n\n\tclient := http.Client{\n\t\tTimeout: timeout,\n\t\tCheckRedirect: redirectMiddleware,\n\t}\n\n\treq, err := http.NewRequest(\"HEAD\", httplink, nil)\n\tif err != nil {\n\t\terrorsProcessed++\n\t\toutput.Errors = append(output.Errors, fmt.Sprintf(\"%s\", err))\n\t\tresults = append(results, output)\n\t\treturn\n\t}\n\n\treq.Header.Set(\"User-Agent\", userAgent)\n\n\tres, err := client.Do(req)\n\tif err != nil {\n\t\terrorsProcessed++\n\t\tif res != nil {\n\t\t\tloc := res.Header.Get(\"Location\")\n\t\t\toutput.StatusCode = res.StatusCode\n\t\t\tif loc != \"\" {\n\t\t\t\tfull, err := absoluteURL(loc, httplink)\n\t\t\t\tif err == nil {\n\t\t\t\t\toutput.Redirect = full\n\t\t\t\t\tresults = append(results, output)\n\t\t\t\t\taddQueueLink(full, \"head\", httplink, 0, wg)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\toutput.Errors = append(output.Errors, fmt.Sprintf(\"%s\", err))\n\t\tresults = append(results, output)\n\t\treturn\n\t}\n\n\tdefer res.Body.Close()\n\n\t// some hosts block HEAD requests, so we do a standard GET instead\n\tif res.StatusCode == 404 || res.StatusCode == 405 {\n\t\tisOutbound := baseDomain != \"\" && getHost(httplink) != baseDomain\n\n\t\tif isOutbound {\n\t\t\tgetResponse(httplink, wg)\n\t\t\treturn\n\t\t}\n\t}\n\n\toutput.StatusCode = res.StatusCode\n\n\tif output.StatusCode != 200 {\n\t\terrorsProcessed++\n\t\toutput.Errors = append(output.Errors, fmt.Sprintf(\"returned status %d\", output.StatusCode))\n\t}\n\n\tresults = append(results, output)\n}", "func (r *Route) Head(handler http.Handler) *Route {\n\tr.handlers[http.MethodHead] = handler\n\treturn r\n}", "func (mx *Mux) Head(pattern string, handler interface{}) {\n\tmx.handle(HEAD, pattern, handler)\n}", "func HeadContent(url string, data ...interface{}) string {\n\treturn RequestContent(\"HEAD\", url, data...)\n}", "func (mm *Middleman) Head(path string, middleware Middleware) error {\r\n\treturn mm.addMiddleware(path, http.MethodHead, middleware)\r\n}", "func (client HTTPSuccessClient) Head404() (result autorest.Response, err error) {\n req, err := client.Head404Preparer()\n if err != nil {\n return result, autorest.NewErrorWithError(err, \"httpinfrastructuregroup.HTTPSuccessClient\", \"Head404\", nil , \"Failure preparing request\")\n }\n\n resp, err := client.Head404Sender(req)\n if err != nil {\n result.Response = resp\n return result, autorest.NewErrorWithError(err, \"httpinfrastructuregroup.HTTPSuccessClient\", \"Head404\", resp, \"Failure sending request\")\n }\n\n result, err = client.Head404Responder(resp)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"httpinfrastructuregroup.HTTPSuccessClient\", \"Head404\", resp, \"Failure responding to request\")\n }\n\n return\n}", "func (g *Group) HEAD(path string, h Handler, gases ...Gas) {\n\tg.Air.HEAD(g.Prefix+path, h, append(g.Gases, gases...)...)\n}", "func (req *Request) Head() *ResultSet {\n\treturn req.do(http.MethodHead)\n}", "func (c *BaseController) Head(w http.ResponseWriter, r *Request) {\n\thttp.Error(w, \"Method Not Allowed\", 405)\n}", "func (c *Controller) Head() {\n\thttp.Error(c.Ctx.ResponseWriter, \"Method Not Allowed\", http.StatusMethodNotAllowed)\n}", "func (m *Mux) HeadFunc(path string, handler http.HandlerFunc) *Route {\n\treturn m.register(\"HEAD\", path, handler)\n}", "func (client HTTPSuccessClient) Head404Sender(req *http.Request) (*http.Response, error) {\n return autorest.SendWithSender(client, req)\n}", "func (c *Client) Head(name string, opts *HeadOpts) FileIterator {\n\treturn c.HeadWithContext(context.Background(), name, opts)\n}", "func (mr *MockRouterMockRecorder) HEAD(path, handler interface{}, options ...interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\tvarargs := append([]interface{}{path, handler}, options...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"HEAD\", reflect.TypeOf((*MockRouter)(nil).HEAD), varargs...)\n}", "func (rb *RequestBuilder) AsyncHead(url string, f func(*Response)) {\n\tgo func() {\n\t\tf(rb.Head(url))\n\t}()\n}", "func (ctl Controller) Head(ctx *gin.Context) {\n\n\tMethodNotAllowedJSON(ctx)\n}", "func (s *GDrive) Head(ctx context.Context, token string, filename string) (contentLength uint64, err error) {\n\tvar fileID string\n\tfileID, err = s.findID(filename, token)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tvar fi *drive.File\n\tif fi, err = s.service.Files.Get(fileID).Context(ctx).Fields(\"size\").Do(); err != nil {\n\t\treturn\n\t}\n\n\tcontentLength = uint64(fi.Size)\n\n\treturn\n}", "func (mr *MockRouterTxMockRecorder) HEAD(path, handler interface{}, options ...interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\tvarargs := append([]interface{}{path, handler}, options...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"HEAD\", reflect.TypeOf((*MockRouterTx)(nil).HEAD), varargs...)\n}", "func (m *MockRouter) HEAD(path string, handler interface{}, options ...interface{}) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{path, handler}\n\tfor _, a := range options {\n\t\tvarargs = append(varargs, a)\n\t}\n\tm.ctrl.Call(m, \"HEAD\", varargs...)\n}", "func (c *Client) HeadVar(url string, data ...interface{}) *gvar.Var {\n\treturn c.RequestVar(\"HEAD\", url, data...)\n}", "func HeadVar(url string, data ...interface{}) *gvar.Var {\n\treturn RequestVar(\"HEAD\", url, data...)\n}", "func (client HTTPSuccessClient) Head204() (result autorest.Response, err error) {\n req, err := client.Head204Preparer()\n if err != nil {\n return result, autorest.NewErrorWithError(err, \"httpinfrastructuregroup.HTTPSuccessClient\", \"Head204\", nil , \"Failure preparing request\")\n }\n\n resp, err := client.Head204Sender(req)\n if err != nil {\n result.Response = resp\n return result, autorest.NewErrorWithError(err, \"httpinfrastructuregroup.HTTPSuccessClient\", \"Head204\", resp, \"Failure sending request\")\n }\n\n result, err = client.Head204Responder(resp)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"httpinfrastructuregroup.HTTPSuccessClient\", \"Head204\", resp, \"Failure responding to request\")\n }\n\n return\n}", "func (r *Route) HeadFunc(f http.HandlerFunc) *Route {\n\treturn r.Head(http.HandlerFunc(f))\n}", "func Ping(w http.ResponseWriter, r *http.Request) {\n\turl := r.FormValue(\"url\")\n\n\tfmt.Println(\"Ping\", url)\n\n\tresp, err := http.Head(url)\n\tif err != nil {\n\t\thttp.Error(w, \"Unable to connect\", 500)\n\t\treturn\n\t}\n\n\tw.WriteHeader(resp.StatusCode)\n}", "func (s *ProjectsService) GetHEAD(projectName string) (string, *Response, error) {\n\tu := fmt.Sprintf(\"projects/%s/HEAD\", url.QueryEscape(projectName))\n\treturn getStringResponseWithoutOptions(s.client, u)\n}", "func (m *MockRouterTx) HEAD(path string, handler interface{}, options ...interface{}) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{path, handler}\n\tfor _, a := range options {\n\t\tvarargs = append(varargs, a)\n\t}\n\tm.ctrl.Call(m, \"HEAD\", varargs...)\n}", "func (r *Router) HeadFunc(pattern string, fn http.HandlerFunc) Route {\n\treturn r.Head(pattern, http.HandlerFunc(fn))\n}", "func getFsEndpoint(ctx context.Context, client *http.Client, url string, opt *Options) (string, bool) {\n\t// If url ends with '/' it is already a proper url always assumed to be a directory.\n\tif url[len(url)-1] == '/' {\n\t\treturn url, false\n\t}\n\n\t// If url does not end with '/' we send a HEAD request to decide\n\t// if it is directory or file, and if directory appends the missing\n\t// '/', or if file returns the directory url to parent instead.\n\tcreateFileResult := func() (string, bool) {\n\t\tfs.Debugf(nil, \"If path is a directory you must add a trailing '/'\")\n\t\tparent, _ := path.Split(url)\n\t\treturn parent, true\n\t}\n\tcreateDirResult := func() (string, bool) {\n\t\tfs.Debugf(nil, \"To avoid the initial HEAD request add a trailing '/' to the path\")\n\t\treturn url + \"/\", false\n\t}\n\n\t// If HEAD requests are not allowed we just have to assume it is a file.\n\tif opt.NoHead {\n\t\tfs.Debugf(nil, \"Assuming path is a file as --http-no-head is set\")\n\t\treturn createFileResult()\n\t}\n\n\t// Use a client which doesn't follow redirects so the server\n\t// doesn't redirect http://host/dir to http://host/dir/\n\tnoRedir := *client\n\tnoRedir.CheckRedirect = func(req *http.Request, via []*http.Request) error {\n\t\treturn http.ErrUseLastResponse\n\t}\n\treq, err := http.NewRequestWithContext(ctx, \"HEAD\", url, nil)\n\tif err != nil {\n\t\tfs.Debugf(nil, \"Assuming path is a file as HEAD request could not be created: %v\", err)\n\t\treturn createFileResult()\n\t}\n\taddHeaders(req, opt)\n\tres, err := noRedir.Do(req)\n\n\tif err != nil {\n\t\tfs.Debugf(nil, \"Assuming path is a file as HEAD request could not be sent: %v\", err)\n\t\treturn createFileResult()\n\t}\n\tif res.StatusCode == http.StatusNotFound {\n\t\tfs.Debugf(nil, \"Assuming path is a directory as HEAD response is it does not exist as a file (%s)\", res.Status)\n\t\treturn createDirResult()\n\t}\n\tif res.StatusCode == http.StatusMovedPermanently ||\n\t\tres.StatusCode == http.StatusFound ||\n\t\tres.StatusCode == http.StatusSeeOther ||\n\t\tres.StatusCode == http.StatusTemporaryRedirect ||\n\t\tres.StatusCode == http.StatusPermanentRedirect {\n\t\tredir := res.Header.Get(\"Location\")\n\t\tif redir != \"\" {\n\t\t\tif redir[len(redir)-1] == '/' {\n\t\t\t\tfs.Debugf(nil, \"Assuming path is a directory as HEAD response is redirect (%s) to a path that ends with '/': %s\", res.Status, redir)\n\t\t\t\treturn createDirResult()\n\t\t\t}\n\t\t\tfs.Debugf(nil, \"Assuming path is a file as HEAD response is redirect (%s) to a path that does not end with '/': %s\", res.Status, redir)\n\t\t\treturn createFileResult()\n\t\t}\n\t\tfs.Debugf(nil, \"Assuming path is a file as HEAD response is redirect (%s) but no location header\", res.Status)\n\t\treturn createFileResult()\n\t}\n\tif res.StatusCode < 200 || res.StatusCode > 299 {\n\t\t// Example is 403 (http.StatusForbidden) for servers not allowing HEAD requests.\n\t\tfs.Debugf(nil, \"Assuming path is a file as HEAD response is an error (%s)\", res.Status)\n\t\treturn createFileResult()\n\t}\n\n\tfs.Debugf(nil, \"Assuming path is a file as HEAD response is success (%s)\", res.Status)\n\treturn createFileResult()\n}", "func printHeadTag(r *http.Response) {\n\tvar head string\n\tvar err error\n\tvar b = bytes.NewBuffer(nil)\n\tvar n int64\n\tvar openInd, closeInd int\n\t//if len(r.Body) == 0 {\n\t//\tfmt.Printf(\"getrequester.getHeadTag: no body\\n\")\n\t//\treturn\n\t//}\n\tif n, err = b.ReadFrom(r.Body); err != nil {\n\t\tfmt.Printf(\"getrequester.getHeadTag err=%v\\n\", err.Error())\n\t\thead = b.String()\n\t\tif head == \"\" {\n\t\t\tfmt.Printf(\"getrequester.getHeadTag: nothing was read from the body\\n\")\n\t\t} else {\n\t\t\tfmt.Printf(\"%v\\n\", head)\n\t\t}\n\t\treturn\n\t}\n\tif n == 0 {\n\t\tfmt.Printf(\"getrequester.getHeadTag: no body\\n\")\n\t\treturn\n\t}\n\tr.Body.Close()\n\topenInd = strings.Index(b.String(), \"<head>\")\n\tcloseInd = strings.Index(b.String(), \"</head>\")\n\tif openInd == -1 || closeInd == -1 {\n\t\tfmt.Printf(\"getrequester.getHeadTag no head tag (%v;%v)\", openInd, closeInd)\n\t\treturn\n\t}\n\thead = b.String()\n\thead = head[openInd : closeInd+6]\n\tfmt.Printf(\"%v\\n\", head)\n}", "func (client HTTPSuccessClient) Head404Responder(resp *http.Response) (result autorest.Response, err error) { \n err = autorest.Respond(\n resp,\n client.ByInspecting(),\n azure.WithErrorUnlessStatusCode(http.StatusOK,http.StatusNoContent,http.StatusNotFound),\n autorest.ByClosing())\n result.Response = resp\n return\n}", "func (ctx *Context) IsHead() bool {\r\n\treturn ctx.Is(\"HEAD\")\r\n}", "func (client HTTPSuccessClient) Head204Sender(req *http.Request) (*http.Response, error) {\n return autorest.SendWithSender(client, req)\n}" ]
[ "0.8339463", "0.796774", "0.79031473", "0.79007864", "0.7859185", "0.78095627", "0.7662336", "0.76048857", "0.75982046", "0.7593806", "0.7559749", "0.7555149", "0.75462085", "0.75420874", "0.7530385", "0.7517558", "0.7505431", "0.7505388", "0.746149", "0.7457079", "0.7431187", "0.74300724", "0.73983973", "0.73963344", "0.736028", "0.73429185", "0.7339016", "0.73286027", "0.73218334", "0.7321662", "0.73112243", "0.72904205", "0.7255", "0.723798", "0.7230453", "0.7223893", "0.7147914", "0.7133416", "0.70610267", "0.7056589", "0.7036722", "0.70081884", "0.6983289", "0.6974984", "0.691673", "0.6906471", "0.6840314", "0.6801512", "0.67911077", "0.67888016", "0.67760164", "0.67710376", "0.67649615", "0.67604434", "0.6702921", "0.6612381", "0.66112685", "0.6594372", "0.6563715", "0.6562698", "0.65618086", "0.6504534", "0.6500028", "0.6499687", "0.64650726", "0.6460968", "0.6457972", "0.6443077", "0.6431901", "0.6422665", "0.64080995", "0.6358015", "0.6314226", "0.630902", "0.6306513", "0.62763166", "0.6264894", "0.6201519", "0.6138083", "0.6049126", "0.6016663", "0.59580034", "0.586425", "0.58414173", "0.5830797", "0.58236074", "0.5818969", "0.5804552", "0.580145", "0.57315546", "0.5667795", "0.56630605", "0.565248", "0.56455916", "0.56309783", "0.56232595", "0.5570354", "0.55524206", "0.5551861", "0.5526659" ]
0.8708814
0
GET is a shorthand for e.Request("GET", path, pathargs...).
func (e *Expect) GET(path string, pathargs ...interface{}) *Request { return e.Request(http.MethodGet, path, pathargs...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *Request) Get(path string) *Request {\n\treturn r.method(\"GET\", path)\n}", "func GET(t *testing.T, path string) *httpexpect.Request {\n\tu, err := url.Parse(path)\n\trequire.Nil(t, err)\n\treturn newAPI(t).\n\t\tGET(u.Path).\n\t\tWithQueryString(u.RawQuery)\n}", "func (r *Request) Get(url string) *Request {\n\tr.method = http.MethodGet\n\tr.url = url\n\treturn r\n}", "func Get (url string, args map[string]string) (*http.Response, error) {\n\t// create a client\n\tclient, req, _ := GetHttpClient(url)\n\t// build the query\n\tif len(args) > 0 {\n\t\treq = buildQuery(req, args)\n\t}\n\t// execute the request\n\t//fmt.Println(req.URL.String())\n\treturn client.Do(req)\n}", "func (t *Tester) GET(path string, variables interface{}, pathargs ...interface{}) *Request {\n\trr := t.E.GET(path, pathargs...).WithHeaders(t.Headers).WithHeader(\"Cookie\", AccessTokenCookie).WithQueryObject(variables)\n\treturn &Request{rr}\n}", "func (r *Router) GET(url string, viewFn View) *Path {\n\treturn r.Path(fasthttp.MethodGet, url, viewFn)\n}", "func (r *Router) GET(path string, handler RequestHandler) {\n\tr.setPath(httpGET, path, handler)\n}", "func (c *baseClient) Get(path string) *baseClient {\n\tc.method = \"GET\"\n\treturn c.Path(path)\n}", "func (r *Request) Get(path string, params ...url.Values) {\n\tcontentType := \"text/html\"\n\n\tif len(params) == 0 {\n\t\tr.Send(\"GET\", path, contentType)\n\t} else {\n\t\tr.Send(\"GET\", path, contentType, params[0])\n\t}\n}", "func (client *Client) Get(path string) Response {\n\treturn client.processRequest(path, \"GET\")\n}", "func (c *Client) Get(path string) (f interface{}, err error) {\n\treturn c.do(\"GET\", path, nil)\n}", "func (c *Case) GET(p string) *RequestBuilder {\n\treturn &RequestBuilder{\n\t\tmethod: http.MethodGet,\n\t\tpath: p,\n\t\tcas: c,\n\t\tfail: c.fail,\n\t}\n}", "func (tr *Transport) GET(\n\turi string,\n\tfn Handler,\n\toptions ...HandlerOption,\n) {\n\ttr.mux.Handler(\n\t\tnet_http.MethodGet,\n\t\turi,\n\t\tnewHandler(fn, append(tr.options, options...)...),\n\t)\n}", "func (api *Api) Get(path string, endpoint http.HandlerFunc, queries ...string) {\n\tapi.Router.HandleFunc(path, endpoint).Methods(\"GET\").Queries(queries...)\n}", "func Get(path string, fn http.HandlerFunc, c ...alice.Constructor) {\n\tinfoMutex.Lock()\n\trecord(\"GET\", path)\n\tr.Get(path, alice.New(c...).ThenFunc(fn).(http.HandlerFunc))\n\tinfoMutex.Unlock()\n}", "func (a *APITest) Get(url string) *Request {\n\ta.request.method = http.MethodGet\n\ta.request.url = url\n\treturn a.request\n}", "func (c Client) get(path string, params url.Values, holder interface{}) error {\n\treturn c.request(\"GET\", path, params, &holder)\n}", "func Get(path string, fn http.HandlerFunc, c ...alice.Constructor) {\n\trecord(\"GET\", path)\n\n\tinfoMutex.Lock()\n\tr.GET(path, Handler(alice.New(c...).ThenFunc(fn)))\n\tinfoMutex.Unlock()\n}", "func (e *Engine) GET(path string, handler Handler) {\n\te.registerRoute(http.MethodGet, path, handler)\n}", "func Get(url string, r io.Reader, w io.Writer, clientGenerator func() *http.Client, reqTuner ...func(*http.Request)) error {\n\treturn Request(\"GET\", url, r, w, clientGenerator, reqTuner...)\n}", "func GetRequest(host, path string) (*http.Response, error) {\n\tfmt.Println(\"GET\", \"/\"+path)\n\tresp, err := http.Get(host + \"/\" + path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func NewGet(url string) *Request { return NewRequest(\"GET\", url) }", "func (c *Client) Get(route string, queryValues map[string]string) (*RawResponse, error) {\n return c.doRequest(\"GET\", route, queryValues, nil)\n}", "func (c *Client) Get(url string, headers, queryParams map[string][]string) (response *http.Response, err error) {\n\treturn c.makeRequest(url, http.MethodGet, headers, queryParams, nil)\n}", "func Get(opts ...Option) ([]byte, error) {\n\treturn request(\"GET\", opts...)\n}", "func (c clientType) Get(path string) (*resty.Response, error) {\n\treturn c.execute(http.MethodGet, path, nil)\n}", "func (serv *Server) GET(url string, handlers ...Handler) {\n\tserv.Handle(\"GET\", url, handlers...)\n}", "func (req *Req) Get(u string) ([]byte, error) {\n\treturn req.request(\"GET\", u)\n}", "func (client *Client) Get(action string, params url.Values, header http.Header) (*Response, error) {\r\n\treturn client.Request(\"GET\", action, params, header, nil)\r\n}", "func (r *Router) GET(path string, h HandlerFunc) {\n\tr.router.GET(path, r.handle(h, r.getValidationForPath(path, \"GET\")))\n}", "func (s *Server) Get(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\ts.Router.HandleFunc(path, f).Methods(\"GET\")\n}", "func (app *App) GET(url string, handler ...Handler) *App {\n\tapp.routeANY = false\n\tapp.AppendReqAndResp(url, \"get\", handler)\n\treturn app\n}", "func (f *Fastglue) GET(path string, h FastRequestHandler) {\n\tf.Router.GET(path, f.handler(h))\n}", "func Get(path string, f func(http.ResponseWriter, *http.Request)) *Route {\n\treturn NewRoute().Path(path).Method(HttpGet).HandlerFunc(f)\n}", "func (g *Group) GET(path string, handler Handler, middleware ...Middleware) *Group {\n\treturn g.Add(http.MethodGet, path, handler, middleware...)\n}", "func (router *Router) GET(relativePath string, handler Handler, decorators ...Decorator) {\n\trouter.createRouter(http.MethodGet, relativePath, handler, \"\", decorators...)\n}", "func (h *Handler) Get(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\tpath = configuration.Instance.Service.Path + path\n\tlog.Println(\"Adding '\" + path + \"' as GET path\")\n\th.Router.HandleFunc(path, f).Methods(\"GET\")\n}", "func (rb *RequestBuilder) Get(url string) *Response {\n\treturn rb.DoRequest(http.MethodGet, url, nil)\n}", "func Get(url string, data ...interface{}) (*ClientResponse, error) {\n\treturn DoRequest(\"GET\", url, data...)\n}", "func (s *Nap) Get(pathURL string) *Nap {\n\ts.method = MethodGet\n\treturn s.Path(pathURL)\n}", "func (g *RouterGroup) GET(url string, handler ...Handler) *RouterGroup {\n\tg.app.routeANY = false\n\tg.AppendReqAndResp(url, \"get\", handler)\n\treturn g\n}", "func (m *Middleware) GET(path string, handler Handler) {\n\tm.engine.GET(path, m.apply(handler))\n}", "func (s *Server) GET(path string, handle http.HandlerFunc) {\n\ts.router.GET(path, s.wrapHandler(handle))\n}", "func (e *Engine) GET(path string, handler HandlerFunc) {\n\n\troute := Route{\n\t\tmethod: http.MethodGet,\n\t\thandler: handler,\n\t\tparam: map[int]string{},\n\t\tpath: path,\n\t}\n\tsPath := strings.Split(path, \"/\")\n\tre := regexp.MustCompile(`\\{\\w*\\}`)\n\tparams := re.FindAllString(path, -1)\n\n\tfor _, p := range params {\n\t\troute.path = strings.Replace(path, p, `\\w*`, -1)\n\t}\n\t//Revisit this logic. I have done string manipulation, just to unblock myself.\n\tfor i, p := range strings.Split(route.path, \"/\") {\n\t\tif p == `\\w*` {\n\t\t\troute.param[i] = strings.Trim(strings.Trim(sPath[i], \"{\"), \"}\")\n\t\t}\n\t}\n\n\trouteTable[route.path] = &route\n}", "func (e *Engine) GET(path string, fn Controller) {\n\te.gin.GET(path, adapt(fn))\n}", "func (c *Client) Get(ctx context.Context, url string, data ...interface{}) (*Response, error) {\n\treturn c.DoRequest(ctx, http.MethodGet, url, data...)\n}", "func (F *Frisby) Get(url string) *Frisby {\n\tF.Method = \"GET\"\n\tF.Url = url\n\treturn F\n}", "func (r *Router) GET(path string, handle HandlerFunc, middleware ...MiddlewareFunc) {\n\tr.Handle(\"GET\", path, handle, middleware...)\n}", "func (app *App) Get(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\tapp.Router.HandleFunc(path, f).Methods(\"GET\")\n}", "func Get(h http.Handler) http.Handler {\n\treturn HTTP(h, GET)\n}", "func (session *Session) Get(path string, params Params) (response []byte, err error) {\n\turlStr := session.getUrl(path, params)\n\tlog.Println(urlStr)\n\tresponse, err = session.sendGetRequest(urlStr)\n\treturn\n\n\t//res, err = MakeResult(response)\n\t//return\n\n}", "func (a *App) Get(path string, f func(w http.ResponseWriter, r *http.Request)) {\r\n\ta.Router.HandleFunc(path, f).Methods(\"GET\")\r\n}", "func (engine *Engine) GET(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn engine.handle(http.MethodGet, relativePath, handlers)\n}", "func Get(url string) (resp *http.Response, err error) {\n\treturn do(\"GET\", url, nil)\n}", "func (a *App) Get(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\ta.Router.HandleFunc(path, f).Methods(\"GET\")\n}", "func (a *App) Get(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\ta.Router.HandleFunc(path, f).Methods(\"GET\")\n}", "func (a *App) Get(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\ta.Router.HandleFunc(path, f).Methods(\"GET\")\n}", "func (r *bitroute) GET(path string, f func(router.Control)) {\n\tr.register(\"GET\", path, f)\n}", "func (mg *Module) GET(path string, handler gin.HandlerFunc) gin.IRoutes {\n\treturn mg.App.Router.GET(mg.MountPath+path, handler)\n}", "func (r Requester) Get(filters string) Requester {\n\treq, err := http.NewRequest(http.MethodGet, r.url+filters, nil)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tr.httpRequest = req\n\treturn r\n}", "func (r *Router) GET(path string, handler Handle) {\n\tr.Handle(\"GET\", path, handler)\n}", "func (r *Router) GET(path string, handle Handle) {\n\tr.Handle(http.MethodGet, path, handle)\n}", "func (r *Router) Get(path, title string, fn Handle) {\n\tr.addRoute(\"GET\", path, title, fn)\n}", "func (c *Client) Request(verb, spath string, ro *RequestOptions) (*http.Request, error) {\n\tlog.Printf(\"[INFO] request: %s %s\", verb, spath)\n\n\tif ro == nil {\n\t\tro = new(RequestOptions)\n\t}\n\n\tu := *c.URL\n\tu.Path = path.Join(c.URL.Path, spath)\n\n\tif c.Token != \"\" {\n\t\tif ro.Headers == nil {\n\t\t\tro.Headers = make(map[string]string)\n\t\t}\n\t\tro.Headers[\"Authorization\"] = fmt.Sprintf(\"Bearer %s\", c.Token)\n\t}\n\n\treturn c.rawRequest(verb, &u, ro)\n}", "func (c *Client) Get(path string, out interface{}) error {\n\treturn c.Send(\"GET\", path, nil, out)\n}", "func getRequest(client *Client, path string, query string, response interface{}) error {\n\treq := client.resty.R().SetResult(&response)\n\tif query != \"\" {\n\t\treq.SetQueryParam(\"query\", query)\n\t}\n\tr, err := req.Get(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif r.IsError() {\n\t\treturn fmt.Errorf(\"%s: %s\", r.Status(), r.Body())\n\t}\n\n\treturn nil\n}", "func Get(url string, data ...interface{}) (*Response, error) {\n\tr := NewRequest()\n\treturn r.Get(url, data...)\n}", "func (rg *RouteGroup) GET(path string, handlers ...Handler) *Route {\n\treturn rg.add(\"GET\", path, handlers)\n}", "func (tr *Transport) Get(url string, fn HandlerFunc, options ...HandlerOption) {\n\ttr.mux.Handler(net_http.MethodGet, url, encapsulate(fn, tr.options, options))\n}", "func (client *Client) Get(c context.Context, uri string, params interface{}, res interface{}) (err error) {\n\treq, err := client.NewRequest(xhttp.MethodGet, uri, params)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn client.client.Do(c, req, res)\n}", "func (session *Session) Get(path string, params Params) (res Result, err error) {\n\turlStr := session.app.BaseEndPoint + session.getRequestUrl(path, params)\n\n\tvar response []byte\n\tresponse, err = session.SendGetRequest(urlStr)\n\n\tif err != nil {\n\t\treturn\n\t}\n\n\tres, err = MakeResult(response)\n\treturn\n}", "func (b *Builder) Get(url string) *Builder {\n\tb.Url = url\n\tb.Method = http.MethodGet\n\treturn b\n}", "func Get(route string, do interface{}) *handler {\n\treturn handlerByMethod(&route, do, \"GET\")\n}", "func (c *Client) Get() *Request {\n\treturn NewRequest(c.httpClient, c.base, \"GET\", c.version, c.authstring, c.userAgent)\n}", "func (c *Client) Get(endpoint string, resource ...RequestResource) error {\n\twrapper := newWrapper(\"get\", endpoint, resource...)\n\treturn c.do(&wrapper)\n}", "func Get(path string, handler http.Handler) Route {\n\treturn NewRoute(\"GET\", path, handler)\n}", "func (a *API) Get(path string) (resp *http.Response, err error) {\n\tu, err := url.ParseRequestURI(a.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tu.Path = path\n\n\treturn a.Client.Get(u.String())\n}", "func (fs *IPfsfs) Req(path string, arg string) (resp *http.Response, err error) {\n\tu := url.URL{}\n\tu.Scheme = \"http\"\n\tu.Host = ipfsHost\n\tu.Path = api + path\n\tif arg != \"\" {\n\t\tval := url.Values{}\n\t\tval.Set(\"arg\", arg)\n\t\tval.Set(\"encoding\", \"json\")\n\t\tu.RawQuery = val.Encode()\n\t}\n\tlogger.Debug(\"URL : %s\", u.String())\n\tresp, err = http.Get(u.String())\n\tif resp == nil {\n\t\treturn nil, err\n\t}\n\tif resp.StatusCode != 200 {\n\t\treturn resp, errors.New(resp.Status)\n\t}\n\tif err != nil {\n\t\treturn resp, err\n\t}\n\treturn resp, err\n}", "func (s *Server) Get(path string, fn http.HandlerFunc) {\n\thandler := s.handlers.ThenFunc(fn)\n\ts.router.Handler(http.MethodGet, path, handler)\n}", "func HTTPGet(params *HTTPRequest) (*http.Response, error) {\n\tparams.Method = \"GET\"\n\tprintHTTPRequest(params)\n\n\tu, err := url.Parse(params.URL)\n\tif err != nil {\n\t\tprintErrorLine(err)\n\t\treturn nil, err\n\t}\n\tif len(params.Query) > 0 {\n\t\tu.RawQuery = params.Query.Encode()\n\t}\n\n\tresp, err := http.Get(u.String())\n\tif err != nil {\n\t\tprintErrorLine(err)\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\tif err := printHTTPResponse(params.URL, resp); err != nil {\n\t\tprintErrorLine(err)\n\t\treturn nil, err\n\t}\n\n\treturn resp, nil\n}", "func (r *RouterGroup) GET(path string, fn Controller) {\n\tr.gin.GET(path, adapt(fn))\n}", "func (f *Client) Get(path string, params GraphQueryString) *GraphRequest {\n\treturn f.NewGraphRequest(Get, path, params)\n}", "func (this *Route) GET(handlers ...HTTPHandler) *Route {\n\treturn this.addHandler(\"get\", handlers...)\n}", "func (c *Client) get(endpoint string, queries map[string]string) (*http.Response, error) {\n\t// Assemble request\n\treq, err := c.buildRequest(\"GET\", endpoint, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Add query strings\n\tif queries != nil {\n\t\tencodeQuery(req.URL, queries)\n\t}\n\n\tclient := buildHTTPSClient(c.verifySSL)\n\treturn client.Do(req)\n}", "func (r *Router) Get(path string, handlers ...Handler) {\n\tr.addToRoutes(http.MethodGet, path, handlers...)\n}", "func (r *Router) GET(partPath string, handlers ...HandlerFunc) {\n\tpath := path.Join(r.basePath, partPath)\n\thandlers = r.combineHandlers(handlers)\n\tr.addRoute(http.MethodGet, path, handlers)\n}", "func Get(path string, handler func(http.ResponseWriter, *http.Request)) {\n\tdefaultRoutes.get(path, handler)\n}", "func (r *Mux) GET(path string, handler http.HandlerFunc) {\n\tr.HandleFunc(\"GET\", path, handler)\n}", "func (cl *Client) Get(c context.Context, url string, opts ...RequestOption) (*Response, error) {\n\treq, err := cl.NewRequest(c, http.MethodGet, url, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn cl.Do(c, req)\n}", "func Get(path string, fn httprouter.Handle) {\n\tr.Router.GET(path, fn)\n}", "func (t transporter) Get(path string) (*http.Response, error) {\n\tresp, err := t.client.Get(t.scheme + path)\n\treturn resp, err\n}", "func Get(options RequestOptions) error {\n\thost, path := uriToHostAndPath(options.Uri)\n\toptions.Headers[\"Host\"] = host\n\trequest := fmt.Sprintf(\"GET %s HTTP/1.0\", path)\n\tprotocol := fmt.Sprintf(\"%s\\r\\n%s\\r\\n\", request, options.Headers)\n\treturn send(host, protocol, options)\n}", "func Get(ctx context.Context, url string, options ...RequestOption) (*Response, error) {\n\tr, err := newRequest(ctx, http.MethodGet, url, nil, options...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn doRequest(http.DefaultClient, r)\n}", "func (c *Client) Get(path string, resource, options interface{}) error {\n\treturn c.CreateAndDo(\"GET\", path, nil, options, nil, resource)\n}", "func (web *WebServer) Get(path string, handler http.HandlerFunc) {\n\tweb.router.Get(path, handler)\n}", "func (router *Router) Get(path string, handler http.Handler) {\n\trouter.Handle(\"GET\", path, handler)\n}", "func (group *RouterGroup) GET(path string, handlers ...HandlerFunc) {\n\tgroup.Handle(\"GET\", path, handlers)\n}", "func (APIResourceBase) Get(session *Session, url string, queries url.Values, body io.Reader) (APIStatus, interface{}) {\n\treturn FailSimple(http.StatusMethodNotAllowed), nil\n}", "func (n *NS) Get(url string) *route {\n\treturn &route{context: n.c, method: \"GET\", url: n.prefix + url, bCtrl: n.bCtrl, aCtrl: n.aCtrl}\n}", "func (c *Client) Get(path string, params map[string]string, res interface{}, auth bool) (err error) {\n\tif c.debug {\n\t\tfmt.Println(\"GET:\", auth, path, params)\n\t}\n\tif auth {\n\t\tc.Sign(path, \"\")\n\t}\n\n\tvar r *httpclient.Response\n\tr, err = c.api.Get(c.URL()+path, params)\n\tif err != nil {\n\t\treturn\n\t}\n\tdefer r.Body.Close()\n\n\terr = handleResponse(r, res, auth)\n\treturn\n}" ]
[ "0.81648576", "0.7915022", "0.7594921", "0.75366944", "0.7499703", "0.74763393", "0.73408014", "0.73338115", "0.7321939", "0.7278976", "0.7253074", "0.72435415", "0.72385293", "0.72086805", "0.7207457", "0.7175362", "0.7173006", "0.7165603", "0.7134777", "0.71175677", "0.7106533", "0.70983386", "0.70947814", "0.707551", "0.70506865", "0.7050413", "0.7046737", "0.70369107", "0.7016641", "0.7014082", "0.7012025", "0.69967544", "0.6987188", "0.69832975", "0.6976085", "0.6962892", "0.695707", "0.69510317", "0.69262695", "0.69193053", "0.6912874", "0.6892119", "0.6876617", "0.685673", "0.6842335", "0.6841285", "0.68399113", "0.6839215", "0.68363", "0.6823853", "0.682335", "0.6812193", "0.6798875", "0.6776956", "0.67650557", "0.67650557", "0.67650557", "0.67631924", "0.674381", "0.67418844", "0.6739588", "0.6724777", "0.66839474", "0.6679168", "0.66697454", "0.6665771", "0.6659088", "0.6645765", "0.6636608", "0.66345", "0.662029", "0.6612085", "0.6607359", "0.659871", "0.6596327", "0.65916485", "0.6591584", "0.65750796", "0.65747577", "0.6558734", "0.65563095", "0.6549486", "0.65378225", "0.6530636", "0.6513504", "0.65060604", "0.6505991", "0.65055275", "0.65021706", "0.64961976", "0.6481562", "0.64687866", "0.6460501", "0.6459771", "0.6446179", "0.64323723", "0.64132744", "0.6410876", "0.64103156", "0.6399102" ]
0.8767445
0
POST is a shorthand for e.Request("POST", path, pathargs...).
func (e *Expect) POST(path string, pathargs ...interface{}) *Request { return e.Request(http.MethodPost, path, pathargs...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *Request) Post(path string) *Request {\n\treturn r.method(\"POST\", path)\n}", "func (r *Request) Post(path, contentType string, data ...interface{}) {\n\tr.Send(\"POST\", path, contentType, data...)\n}", "func (client *Client) Post(path, data string, mods ...func(*Req)) (Res, error) {\n\treq := client.NewReq(\"POST\", path, strings.NewReader(data), mods...)\n\treturn client.Do(req)\n}", "func (r *Router) POST(path string, handler RequestHandler) {\n\tr.setPath(httpPOST, path, handler)\n}", "func (r *Router) POST(url string, viewFn View) *Path {\n\treturn r.Path(fasthttp.MethodPost, url, viewFn)\n}", "func (c clientType) Post(path string, body interface{}) (resp *resty.Response, err error) {\n\treturn c.execute(http.MethodPost, path, body)\n}", "func (r *Request) Post(url string) *Request {\n\tr.method = http.MethodPost\n\tr.url = url\n\treturn r\n}", "func (s *Nap) Post(pathURL string) *Nap {\n\ts.method = MethodPost\n\treturn s.Path(pathURL)\n}", "func Post(path string, f func(http.ResponseWriter, *http.Request)) *Route {\n\treturn NewRoute().Path(path).Method(HttpPost).HandlerFunc(f)\n}", "func (e *Engine) POST(path string, handler Handler) {\n\te.registerRoute(http.MethodPost, path, handler)\n}", "func (c *baseClient) Post(path string) *baseClient {\n\tc.method = \"POST\"\n\treturn c.Path(path)\n}", "func (s *Server) Post(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\ts.Router.HandleFunc(path, f).Methods(\"POST\")\n}", "func (g *Group) POST(path string, handler Handler, middleware ...Middleware) *Group {\n\treturn g.Add(http.MethodPost, path, handler, middleware...)\n}", "func (h *Handler) Post(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\tpath = configuration.Instance.Service.Path + path\n\tlog.Println(\"Adding '\" + path + \"' as POST path\")\n\th.Router.HandleFunc(path, f).Methods(\"POST\")\n}", "func (a *ApiServer) Post(path string, f func(w http.ResponseWriter, r *http.Request)) *ApiServer{\n\ta.Router.HandleFunc(path, f).Methods(\"POST\")\n\treturn a\n}", "func (router *Router) POST(relativePath string, handler Handler, decorators ...Decorator) {\n\trouter.createRouter(http.MethodPost, relativePath, handler, \"\", decorators...)\n}", "func (app *App) Post(path string, endpoint http.HandlerFunc, queries ...string) {\r\n\tapp.Router.HandleFunc(path, endpoint).Methods(\"POST\").Queries(queries...)\r\n}", "func (a *APITest) Post(url string) *Request {\n\tr := a.request\n\tr.method = http.MethodPost\n\tr.url = url\n\treturn r\n}", "func (t *Tester) POST(path string, variables interface{}, pathargs ...interface{}) *Request {\n\trr := t.E.POST(path, pathargs...).WithHeaders(t.Headers).WithHeader(\"Cookie\", AccessTokenCookie).WithForm(variables)\n\treturn &Request{rr}\n}", "func (app *App) Post(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\tapp.Router.HandleFunc(path, f).Methods(\"POST\")\n}", "func (a *App) Post(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\ta.Router.HandleFunc(path, f).Methods(\"POST\")\n}", "func (a *App) Post(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\ta.Router.HandleFunc(path, f).Methods(\"POST\")\n}", "func Post(path string, fn http.HandlerFunc, c ...alice.Constructor) {\n\tinfoMutex.Lock()\n\trecord(\"POST\", path)\n\tr.Post(path, alice.New(c...).ThenFunc(fn).(http.HandlerFunc))\n\tinfoMutex.Unlock()\n}", "func (a *App) Post(path string, f func(w http.ResponseWriter, r *http.Request)) {\r\n\ta.Router.HandleFunc(path, f).Methods(\"POST\")\r\n}", "func (api *Api) Post(path string, endpoint http.HandlerFunc, queries ...string) {\n\tapi.Router.HandleFunc(path, endpoint).Methods(\"POST\").Queries(queries...)\n}", "func (c *Client) Post(path string, in, out interface{}) error {\n\treturn c.Send(\"POST\", path, in, out)\n}", "func (f *Fastglue) POST(path string, h FastRequestHandler) {\n\tf.Router.POST(path, f.handler(h))\n}", "func Post(path string, fn http.HandlerFunc, c ...alice.Constructor) {\n\trecord(\"POST\", path)\n\n\tinfoMutex.Lock()\n\tr.POST(path, Handler(alice.New(c...).ThenFunc(fn)))\n\tinfoMutex.Unlock()\n}", "func (rb *RequestBuilder) Post(url string, body interface{}) *Response {\n\treturn rb.DoRequest(http.MethodPost, url, body)\n}", "func (web *WebServer) Post(path string, handler http.HandlerFunc) {\n\tweb.router.Post(path, handler)\n}", "func (c Client) post(path string, params url.Values, holder interface{}) error {\n\treturn c.request(\"POST\", path, params, &holder)\n}", "func (engine *Engine) POST(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn engine.handle(http.MethodPost, relativePath, handlers)\n}", "func (s *Server) Post(path string, fn http.HandlerFunc) {\n\thandler := s.handlers.ThenFunc(fn)\n\ts.router.Handler(http.MethodPost, path, handler)\n}", "func (r *bitroute) POST(path string, f func(router.Control)) {\n\tr.register(\"POST\", path, f)\n}", "func (r *Router) POST(path string, handle HandlerFunc, middleware ...MiddlewareFunc) {\n\tr.Handle(\"POST\", path, handle, middleware...)\n}", "func Post(path string, fn httprouter.Handle) {\n\tr.Router.POST(path, fn)\n}", "func (r *Router) POST(path string, handler Handle) {\n\tr.Handle(\"POST\", path, handler)\n}", "func (rg *RouteGroup) POST(path string, handlers ...Handler) *Route {\n\treturn rg.add(\"POST\", path, handlers)\n}", "func (c *Client) Post(path string, body io.Reader) (f interface{}, err error) {\n\treturn c.do(\"POST\", path, body)\n}", "func (c *Case) POST(p string) *RequestBuilder {\n\treturn &RequestBuilder{\n\t\tmethod: http.MethodPost,\n\t\tpath: p,\n\t\tcas: c,\n\t\tfail: c.fail,\n\t}\n}", "func (r *Request) Post(url string) (*Response, error) {\n\treturn r.Execute(MethodPost, url)\n}", "func (m *Middleware) POST(path string, handler Handler) {\n\tm.engine.POST(path, m.apply(handler))\n}", "func (r *Router) POST(path string, h HandlerFunc) {\n\tr.router.POST(path, r.handle(h, r.getValidationForPath(path, \"POST\")))\n}", "func (s *Server) POST(path string, handle http.HandlerFunc) {\n\ts.router.POST(path, s.wrapHandler(handle))\n}", "func (e *Engine) POST(path string, fn Controller) {\n\te.gin.POST(path, adapt(fn))\n}", "func (client *Client) Post(action string, params url.Values, header http.Header) (*Response, error) {\r\n\treturn client.Request(\"POST\", action, params, header, nil)\r\n}", "func Post(path string, handler http.Handler) Route {\n\treturn NewRoute(\"POST\", path, handler)\n}", "func Post(path string, handler func(http.ResponseWriter, *http.Request)) {\n\tdefaultRoutes.post(path, handler)\n}", "func Post(path string, data string) {\n\tres, err := postThis(path, data)\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\tfmt.Printf(sep)\n\tfmt.Printf(\"POST result: %s\\n\", res)\n}", "func Post(url string, r io.Reader, w io.Writer, clientGenerator func() *http.Client, reqTuner ...func(*http.Request)) error {\n\treturn Request(\"POST\", url, r, w, clientGenerator, reqTuner...)\n}", "func (serv *Server) POST(url string, handlers ...Handler) {\n\tserv.Handle(\"POST\", url, handlers...)\n}", "func (router *Router) Post(path string, handler http.Handler) {\n\trouter.Handle(\"POST\", path, handler)\n}", "func Post(h http.Handler) http.Handler {\n\treturn HTTP(h, POST)\n}", "func (req *Req) Post(u string) ([]byte, error) {\n\treturn req.request(\"POST\", u)\n}", "func (rg *RouteGroup) Post(path string, handlers ...Handler) *Route {\n\treturn rg.add(\"POST\", path, handlers)\n}", "func (c *Client) Post(route string, queryValues map[string]string, body []byte) (*RawResponse, error) {\n return c.doRequest(\"POST\", route, queryValues, body)\n}", "func (group *RouterGroup) POST(path string, handlers ...HandlerFunc) {\n\tgroup.Handle(\"POST\", path, handlers)\n}", "func (r *Router) Post(path, title string, fn Handle) {\n\tr.addRoute(\"POST\", path, title, fn)\n}", "func (group *RouterGroup) POST(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn group.Handler(\"POST\", relativePath, handlers...)\n}", "func (r *Router) Post(path string, handler httpHandler) {\n\tr.POST(path, wrapHandler(handler))\n}", "func (s *Server) Post(path string, body interface{}) (*http.Response, error) {\n\turl, err := s.urlFor(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbodyBytes, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn http.Post(url, \"application/json\", bytes.NewReader(bodyBytes))\n}", "func (g *RouterGroup) POST(url string, handler ...Handler) *RouterGroup {\n\tg.app.routeANY = false\n\tg.AppendReqAndResp(url, \"post\", handler)\n\treturn g\n}", "func (r *Router) POST(path string, handle Handle) {\n\tr.Handle(http.MethodPost, path, handle)\n}", "func (session *Session) Post(path string, data Params) (res Result, err error) {\n\turlStr := session.app.BaseEndPoint + session.getRequestUrl(path, nil)\n\n\tvar response []byte\n\tresponse, err = session.SendPostRequest(urlStr, data)\n\n\tif err != nil {\n\t\treturn\n\t}\n\n\tres, err = MakeResult(response)\n\treturn\n}", "func (r *Router) Post(path string, handler http.Handler) {\n\tr.POST(path, wrapHandler(handler))\n}", "func (r *Router) Post(path string, handler http.Handler) {\n\tr.POST(path, wrapHandler(handler))\n}", "func (group *RouterGroup) POST(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn group.handle(http.MethodPost, relativePath, handlers)\n}", "func (f *Client) Post(path string, params GraphQueryString) *GraphRequest {\n\treturn f.NewGraphRequest(Post, path, params)\n}", "func (r *Route) Post(h http.Handler) *Route {\n\tr.Add(\"POST\", h)\n\n\treturn r\n}", "func (mm *Middleman) Post(path string, middleware Middleware) error {\r\n\treturn mm.addMiddleware(path, http.MethodPost, middleware)\r\n}", "func (r *RouterGroup) POST(path string, fn Controller) {\n\tr.gin.POST(path, adapt(fn))\n}", "func (n *NS) Post(url string) *route {\n\treturn &route{context: n.c, method: \"POST\", url: n.prefix + url, bCtrl: n.bCtrl, aCtrl: n.aCtrl}\n}", "func (r *Mux) POST(path string, handler http.HandlerFunc) {\n\tr.HandleFunc(\"POST\", path, handler)\n}", "func (c *Client) Post(ctx context.Context, url string, data ...interface{}) (*Response, error) {\n\treturn c.DoRequest(ctx, http.MethodPost, url, data...)\n}", "func Post(host, path string, params map[string]string, requiredCode int) error {\n\tresp, err := PostRequest(host, path, params)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = out(resp, requiredCode)\n\treturn err\n}", "func (self *Client) Post(dst interface{}, path string, data url.Values) error {\n\tvar addr *url.URL\n\tvar err error\n\tvar body *strings.Reader\n\n\tif addr, err = url.Parse(self.Prefix + strings.TrimLeft(path, \"/\")); err != nil {\n\t\treturn err\n\t}\n\n\tif data != nil {\n\t\tbody = strings.NewReader(data.Encode())\n\t}\n\n\treturn self.newRequest(dst, \"POST\", addr, body)\n}", "func Post(url string, data ...interface{}) (*ClientResponse, error) {\n\treturn DoRequest(\"POST\", url, data...)\n}", "func (a *API) Post(path string, bodyType string, body io.Reader) (resp *http.Response, err error) {\n\tu, err := url.ParseRequestURI(a.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tu.Path = path\n\n\treturn a.Client.Post(u.String(), bodyType, body)\n}", "func (c *Client) Post(url string, headers, queryParams map[string][]string, data interface{}) (response *http.Response, err error) {\n\treturn c.makeRequest(url, http.MethodPost, headers, queryParams, data)\n}", "func (this *Route) POST(handlers ...HTTPHandler) *Route {\n\treturn this.addHandler(\"post\", handlers...)\n}", "func PostRequest(host, path string, params map[string]string) (*http.Response, error) {\n\tall := []string{}\n\tvalues := url.Values{}\n\tfor k, v := range params {\n\t\tall = append(all, k+\"='\"+v+\"'\")\n\t\tvalues[k] = []string{v}\n\t}\n\tfmt.Println(\"POST\", \"/\"+path, \"\", strings.Join(all, \", \"))\n\treturn http.PostForm(host+\"/\"+path, values)\n\n}", "func (a *App) POST(pattern string, handler Handler) {\n\ta.add(\"POST\", pattern, handler)\n}", "func (b *Builder) Post(url string) *Builder {\n\tb.Url = url\n\tb.Method = http.MethodPost\n\treturn b\n}", "func (app *App) POST(url string, handler ...Handler) *App {\n\tapp.routeANY = false\n\tapp.AppendReqAndResp(url, \"post\", handler)\n\treturn app\n}", "func (F *Frisby) Post(url string) *Frisby {\n\tF.Method = \"POST\"\n\tF.Url = url\n\treturn F\n}", "func Post(route string, do interface{}) *handler {\n\treturn handlerByMethod(&route, do, \"POST\")\n}", "func (r *Router) Post(path string, handle HandleFunc) {\n\tr.register(path, http.MethodPost, handle)\n}", "func (r *Request) PostForm(path string, data interface{}) {\n\tr.Post(path, \"application/x-www-form-urlencoded\", data)\n}", "func (r *Route) Post(handler http.Handler) *Route {\n\tr.handlers[http.MethodPost] = handler\n\treturn r\n}", "func (r *Router) Post(path string, handle requestHandler, config ...RouteConfig) {\n\tr.routers[\"POST\"] = append(r.routers[\"POST\"], newRouteHandler(path, handle, config...))\n}", "func (c *Client) Post(path string, data, resource interface{}) error {\n\treturn c.CreateAndDo(\"POST\", path, data, nil, nil, resource)\n}", "func (r *Router) POST(partPath string, handlers ...HandlerFunc) {\n\tpath := path.Join(r.basePath, partPath)\n\thandlers = r.combineHandlers(handlers)\n\tr.addRoute(http.MethodPost, path, handlers)\n}", "func (group *Groups) Post(path string, handler http.HandlerFunc) {\n\tgroup.webserver.router.Post(path, group.handler(handler))\n}", "func (t transporter) Post(path string, body io.Reader) (*http.Response, error) {\n\tresp, err := t.client.Post(t.scheme+path, \"application/json\", body)\n\treturn resp, err\n}", "func (client *Client) Post(c context.Context, uri string, params interface{}, res interface{}) (err error) {\n\treq, err := client.NewRequest(xhttp.MethodPost, uri, params)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn client.client.Do(c, req, res)\n}", "func (c *Client) Post() *Request {\n\treturn NewRequest(c.httpClient, c.base, \"POST\", c.version, c.authstring, c.userAgent)\n}", "func (tr *Transport) POST(\n\turl string,\n\tfn Handler,\n\toptions ...HandlerOption,\n) {\n\ttr.mux.Handler(\n\t\tnet_http.MethodPost,\n\t\turl,\n\t\tnewHandler(fn, append(tr.options, options...)...),\n\t)\n}", "func (tr *Transport) Post(url string, fn HandlerFunc, options ...HandlerOption) {\n\ttr.mux.Handler(net_http.MethodPost, url, encapsulate(fn, tr.options, options))\n}", "func (session *Session) Post(path string, urlParams, data Params) (response []byte, err error) {\n\turlStr := session.getUrl(path, urlParams)\n\n\tresponse, err = session.sendPostRequest(urlStr, data)\n\n\treturn\n\n\t//res, err = MakeResult(response)\n\t//return\n}", "func (c *Client) Post(rawurl string, in, out interface{}) error {\n\treturn c.Do(rawurl, \"POST\", in, out)\n}" ]
[ "0.8183926", "0.7732567", "0.76200134", "0.75784963", "0.75539345", "0.7427361", "0.7357585", "0.7335185", "0.7305281", "0.730186", "0.72804624", "0.727684", "0.7255811", "0.7240516", "0.72129816", "0.72095233", "0.7202608", "0.71791095", "0.71427095", "0.7142554", "0.71369547", "0.71369547", "0.71364075", "0.7135508", "0.7134434", "0.70815575", "0.7078616", "0.70781827", "0.706358", "0.7050629", "0.7049153", "0.7034477", "0.7032914", "0.70170915", "0.70155114", "0.7010572", "0.69836676", "0.69719934", "0.69632995", "0.6962234", "0.69483835", "0.69481456", "0.69337386", "0.69334674", "0.69150245", "0.6905602", "0.68927115", "0.6889825", "0.6877413", "0.6845533", "0.6844327", "0.6833629", "0.68273944", "0.6815093", "0.6814455", "0.6811714", "0.6809695", "0.68058634", "0.68048674", "0.6804152", "0.67972267", "0.678047", "0.6774031", "0.6771766", "0.67559665", "0.67559665", "0.675559", "0.6754227", "0.67530715", "0.67492473", "0.67371905", "0.67354983", "0.67280424", "0.672664", "0.67194295", "0.66998845", "0.6691052", "0.6683097", "0.6674177", "0.66611505", "0.663799", "0.6637488", "0.66347784", "0.66326404", "0.6625959", "0.6622738", "0.66174537", "0.6592155", "0.65898854", "0.6585473", "0.6584076", "0.65831906", "0.65479666", "0.6531224", "0.65080494", "0.64952433", "0.6492289", "0.6477051", "0.6466147", "0.6463049" ]
0.85327715
0
PUT is a shorthand for e.Request("PUT", path, pathargs...).
func (e *Expect) PUT(path string, pathargs ...interface{}) *Request { return e.Request(http.MethodPut, path, pathargs...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *Request) Put(path string) *Request {\n\treturn r.method(\"PUT\", path)\n}", "func (r *Request) Put(path, contentType string, data ...interface{}) {\n\tr.Send(\"PUT\", path, contentType, data...)\n}", "func (client *Client) Put(path string) Response {\n\treturn client.processRequest(path, \"PUT\")\n}", "func (r *Router) PUT(path string, handler RequestHandler) {\n\tr.setPath(httpPUT, path, handler)\n}", "func (c clientType) Put(path string, body interface{}) (resp *resty.Response, err error) {\n\treturn c.execute(http.MethodPut, path, body)\n}", "func (api *Api) Put(path string, endpoint http.HandlerFunc, queries ...string) {\n\tapi.Router.HandleFunc(path, endpoint).Methods(\"PUT\").Queries(queries...)\n}", "func (r *Router) PUT(url string, viewFn View) *Path {\n\treturn r.Path(fasthttp.MethodPut, url, viewFn)\n}", "func (g *Group) PUT(path string, handler Handler, middleware ...Middleware) *Group {\n\treturn g.Add(http.MethodPut, path, handler, middleware...)\n}", "func (c Client) put(path string, params url.Values, holder interface{}) error {\n\treturn c.request(\"PUT\", path, params, &holder)\n}", "func (r Requester) Update(path string, payload interface{}) Requester {\n\tb, err := json.Marshal(payload)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tbody := bytes.NewReader(b)\n\tr.httpRequest, err = http.NewRequest(http.MethodPut, r.url, body)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treturn r\n}", "func Put(path string, fn http.HandlerFunc, c ...alice.Constructor) {\n\tinfoMutex.Lock()\n\trecord(\"PUT\", path)\n\tr.Put(path, alice.New(c...).ThenFunc(fn).(http.HandlerFunc))\n\tinfoMutex.Unlock()\n}", "func (r *Router) PUT(path string, h HandlerFunc) {\n\tr.router.PUT(path, r.handle(h, r.getValidationForPath(path, \"PUT\")))\n}", "func Put(path string, fn http.HandlerFunc, c ...alice.Constructor) {\n\trecord(\"PUT\", path)\n\n\tinfoMutex.Lock()\n\tr.PUT(path, Handler(alice.New(c...).ThenFunc(fn)))\n\tinfoMutex.Unlock()\n}", "func (a *ApiServer) Put(path string, f func(w http.ResponseWriter, r *http.Request)) *ApiServer{\n\ta.Router.HandleFunc(path, f).Methods(\"PUT\")\n\treturn a\n}", "func (e *Engine) PUT(path string, fn Controller) {\n\te.gin.PUT(path, adapt(fn))\n}", "func (s *Server) Put(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\ts.Router.HandleFunc(path, f).Methods(\"PUT\")\n}", "func Put(url string, r io.Reader, w io.Writer, clientGenerator func() *http.Client, reqTuner ...func(*http.Request)) error {\n\treturn Request(\"PUT\", url, r, w, clientGenerator, reqTuner...)\n}", "func (e *Engine) PUT(path string, handler Handler) {\n\te.registerRoute(http.MethodPut, path, handler)\n}", "func (router *Router) PUT(relativePath string, handler Handler, decorators ...Decorator) {\n\trouter.createRouter(http.MethodPut, relativePath, handler, \"\", decorators...)\n}", "func (r *Router) PUT(path string, handle HandlerFunc, middleware ...MiddlewareFunc) {\n\tr.Handle(\"PUT\", path, handle, middleware...)\n}", "func (r *Request) Put(url string) *Request {\n\tr.method = http.MethodPut\n\tr.url = url\n\treturn r\n}", "func (r *bitroute) PUT(path string, f func(router.Control)) {\n\tr.register(\"PUT\", path, f)\n}", "func (h *Handler) Put(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\tpath = configuration.Instance.Service.Path + path\n\tlog.Println(\"Adding '\" + path + \"' as PUT path\")\n\th.Router.HandleFunc(path, f).Methods(\"PUT\")\n}", "func (f *Fastglue) PUT(path string, h FastRequestHandler) {\n\tf.Router.PUT(path, f.handler(h))\n}", "func (r *RouterGroup) PUT(path string, fn Controller) {\n\tr.gin.PUT(path, adapt(fn))\n}", "func (app *App) Put(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\tapp.Router.HandleFunc(path, f).Methods(\"PUT\")\n}", "func Put(url string, data ...interface{}) (*ClientResponse, error) {\n\treturn DoRequest(\"PUT\", url, data...)\n}", "func (m *Middleware) PUT(path string, handler Handler) {\n\tm.engine.PUT(path, m.apply(handler))\n}", "func (a *App) Put(path string, f func(w http.ResponseWriter, r *http.Request)) {\r\n\ta.Router.HandleFunc(path, f).Methods(\"PUT\")\r\n}", "func Put(path string, f func(http.ResponseWriter, *http.Request)) *Route {\n\treturn NewRoute().Path(path).Method(HttpPut).HandlerFunc(f)\n}", "func (a *App) Put(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\ta.Router.HandleFunc(path, f).Methods(\"PUT\")\n}", "func (a *App) Put(path string, f func(w http.ResponseWriter, r *http.Request)) {\n\ta.Router.HandleFunc(path, f).Methods(\"PUT\")\n}", "func Put(path string, handler http.Handler) Route {\n\treturn NewRoute(\"PUT\", path, handler)\n}", "func (c *Client) Put(url string, headers, queryParams map[string][]string, data interface{}) (response *http.Response, err error) {\n\treturn c.makeRequest(url, http.MethodPut, headers, queryParams, data)\n}", "func (f *Client) Put(path string, params GraphQueryString) *GraphRequest {\n\treturn f.NewGraphRequest(Put, path, params)\n}", "func (c *Client) Put(path string, data, resource interface{}) error {\n\treturn c.CreateAndDo(\"PUT\", path, data, nil, nil, resource)\n}", "func (r *Router) PUT(path string, handler Handle) {\n\tr.Handle(\"PUT\", path, handler)\n}", "func (F *Frisby) Put(url string) *Frisby {\n\tF.Method = \"PUT\"\n\tF.Url = url\n\treturn F\n}", "func (r *Router) PUT(path string, handle Handle) {\n\tr.Handle(http.MethodPut, path, handle)\n}", "func NewPut(url string) *Request { return NewRequest(\"PUT\", url) }", "func (rg *RouteGroup) PUT(path string, handlers ...Handler) *Route {\n\treturn rg.add(\"PUT\", path, handlers)\n}", "func (client *RestClient) Put(path string, obj interface{}) error {\n\tres, err := client.R().SetBody(obj).Put(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif res.StatusCode() >= 400 {\n\t\treturn UnmarshalError(res)\n\t}\n\n\treturn nil\n}", "func (rb *RequestBuilder) Put(url string, body interface{}) *Response {\n\treturn rb.DoRequest(http.MethodPut, url, body)\n}", "func (engine *Engine) PUT(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn engine.handle(http.MethodPut, relativePath, handlers)\n}", "func Put(path string, fn httprouter.Handle) {\n\tr.Router.PUT(path, fn)\n}", "func (r *Router) Put(path, title string, fn Handle) {\n\tr.addRoute(\"PUT\", path, title, fn)\n}", "func (c *Client) Put() *Request {\n\treturn NewRequest(c.httpClient, c.base, \"PUT\", c.version, c.authstring, c.userAgent)\n}", "func (a *APITest) Put(url string) *Request {\n\tr := a.request\n\tr.method = http.MethodPut\n\tr.url = url\n\treturn r\n}", "func (r *Mux) PUT(path string, handler http.HandlerFunc) {\n\tr.HandleFunc(\"PUT\", path, handler)\n}", "func (APIResourceBase) Put(session *Session, url string, queries url.Values, body io.Reader) (APIStatus, interface{}) {\n\treturn FailSimple(http.StatusMethodNotAllowed), nil\n}", "func (app *App) PUT(url string, handler ...Handler) *App {\n\tapp.routeANY = false\n\tapp.AppendReqAndResp(url, \"put\", handler)\n\treturn app\n}", "func (c *Client) Put(rawurl string, in, out interface{}) error {\n\treturn c.Do(rawurl, \"PUT\", in, out)\n}", "func (self *Client) Put(dst interface{}, path string, data url.Values) error {\n\tvar addr *url.URL\n\tvar err error\n\tvar body *strings.Reader\n\n\tif addr, err = url.Parse(self.Prefix + strings.TrimLeft(path, \"/\")); err != nil {\n\t\treturn err\n\t}\n\n\tif data != nil {\n\t\tbody = strings.NewReader(data.Encode())\n\t}\n\n\treturn self.newRequest(dst, \"PUT\", addr, body)\n}", "func (r *Request) Put(url string) (*Response, error) {\n\treturn r.Execute(MethodPut, url)\n}", "func (router *Router) Put(path string, handler http.Handler) {\n\trouter.Handle(\"PUT\", path, handler)\n}", "func (c *Client) put(rawURL string, authenticate bool, expectedStatus int, in interface{}, out interface{}) error {\n\terr := c.do(rawURL, \"PUT\", authenticate, expectedStatus, in, out)\n\treturn errio.Error(err)\n}", "func (tr *Transport) PUT(\n\turl string,\n\tfn Handler,\n\toptions ...HandlerOption,\n) {\n\ttr.mux.Handler(\n\t\tnet_http.MethodPut,\n\t\turl,\n\t\tnewHandler(fn, append(tr.options, options...)...),\n\t)\n}", "func (group *RouterGroup) PUT(path string, handlers ...HandlerFunc) {\n\tgroup.Handle(\"PUT\", path, handlers)\n}", "func Put(url string, payload interface{}, headers map[string]string) (int, error) {\n\treturn doHttpCall(url, \"PUT\", payload, headers)\n}", "func (serv *Server) PUT(url string, handlers ...Handler) {\n\tserv.Handle(\"PUT\", url, handlers...)\n}", "func Put(c context.Context, path string, data []byte) error {\n\treturn FromContext(c).Put(path, data)\n}", "func (g *RouterGroup) PUT(url string, handler ...Handler) *RouterGroup {\n\tg.app.routeANY = false\n\tg.AppendReqAndResp(url, \"put\", handler)\n\treturn g\n}", "func (c *Client) Put(ctx context.Context, url string, data ...interface{}) (*Response, error) {\n\treturn c.DoRequest(ctx, http.MethodPut, url, data...)\n}", "func (c *TogglHttpClient) PutRequest(endpoint string, body interface{}) (*json.RawMessage, error) {\n\treturn request(c, \"PUT\", endpoint, body)\n}", "func (c *Client) Put(endpoint string, resource ...RequestResource) error {\n\twrapper := newWrapper(\"put\", endpoint, resource...)\n\treturn c.do(&wrapper)\n}", "func Put(route string, do interface{}) *handler {\n\treturn handlerByMethod(&route, do, \"PUT\")\n}", "func (node *Node) PUT(functions ...interface{}) *Handler {\n\n\thandler := &Handler{}\n\n\tif len(functions) > 0 { handler.function = functions[0].(func (req web.RequestInterface) *web.ResponseStatus) }\n\n\tnode.addHandler(\"PUT\", handler)\n\n\treturn handler\n}", "func (mm *Middleman) Put(path string, middleware Middleware) error {\r\n\treturn mm.addMiddleware(path, http.MethodPut, middleware)\r\n}", "func (s *Nap) Put(pathURL string) *Nap {\n\ts.method = MethodPut\n\treturn s.Path(pathURL)\n}", "func (a *App) PUT(pattern string, handler Handler) {\n\ta.add(\"PUT\", pattern, handler)\n}", "func (c *Client) Put(url string, reqBody, resType interface{}) error {\n\treturn c.CallAPI(\"PUT\", url, reqBody, resType, true)\n}", "func (cl *Client) Put(c context.Context, url string, opts ...RequestOption) (*Response, error) {\n\treq, err := cl.NewRequest(c, http.MethodPut, url, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn cl.Do(c, req)\n}", "func Put(url string, data ...interface{}) (*Response, error) {\n\tr := NewRequest()\n\treturn r.Put(url, data...)\n}", "func Put(url string, bodyType string, body io.Reader) (*http.Response, error) {\n\treq, err := http.NewRequest(\"PUT\", url, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Header.Set(\"Content-Type\", bodyType)\n\treturn http.DefaultClient.Do(req)\n}", "func (r *Router) Put(path string, handle requestHandler, config ...RouteConfig) {\n\tr.routers[\"PUT\"] = append(r.routers[\"PUT\"], newRouteHandler(path, handle, config...))\n}", "func (r *Router) Put(path string, handler httpHandler) {\n\tr.PUT(path, wrapHandler(handler))\n}", "func (em *entityManager) Put(ctx context.Context, entityPath string, body []byte, mw ...MiddlewareFunc) (*http.Response, error) {\n\tctx, span := em.startSpanFromContext(ctx, \"sb.EntityManger.Put\")\n\tdefer span.End()\n\n\treturn em.Execute(ctx, http.MethodPut, entityPath, bytes.NewReader(body), mw...)\n}", "func (r *Route) Put(h http.Handler) *Route {\n\tr.Add(\"PUT\", h)\n\n\treturn r\n}", "func (group *RouterGroup) PUT(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn group.handle(http.MethodPut, relativePath, handlers)\n}", "func Put(path string, data string) {\n\tres, err := putThis(path, data)\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\tfmt.Printf(sep)\n\tfmt.Printf(\"PUT result: %s\\n\", res)\n}", "func put(resource string, data []byte) ([]byte, error) {\n\thttpParams := &HTTPParams{\"PUT\", resource, \"application/json\", data}\n\treturn processRequest(httpParams)\n}", "func (n *NS) Put(url string) *route {\n\treturn &route{context: n.c, method: \"PUT\", url: n.prefix + url, bCtrl: n.bCtrl, aCtrl: n.aCtrl}\n}", "func (r *Router) Put(path string, handler http.Handler) {\n\tr.PUT(path, wrapHandler(handler))\n}", "func (r *Router) Put(path string, handler http.Handler) {\n\tr.PUT(path, wrapHandler(handler))\n}", "func (r *Router) Put(path string, handle HandleFunc) {\n\tr.register(path, http.MethodPut, handle)\n}", "func Put(h http.Handler) http.Handler {\n\treturn HTTP(h, PUT)\n}", "func (group *RouterGroup) PUT(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn group.Handler(\"PUT\", relativePath, handlers...)\n}", "func (c *Client) Put(url string, headers map[string][]string, body io.Reader) (client.Status, map[string][]string, io.ReadCloser, error) {\n\treturn c.Do(\"PUT\", url, headers, body)\n}", "func (c *Client) put(urlStr string, body []byte) (*http.Response, error) {\n\treturn c.doAlwaysFollowingRedirects(\"PUT\", urlStr, body)\n}", "func makePutRequest(link string, data *bytes.Buffer) string {\n client := &http.Client{}\n req, _ := http.NewRequest(\"PUT\", link, data)\n response, clientErr := client.Do(req)\n if clientErr != nil {\n fmt.Println(clientErr)\n }\n bodyBytes, _ := ioutil.ReadAll(response.Body)\n response.Body.Close()\n return string(bodyBytes)\n}", "func (g *Group) PUT(path string, h Handler, gases ...Gas) {\n\tg.Air.PUT(g.Prefix+path, h, append(g.Gases, gases...)...)\n}", "func (b *Builder) Put(url string) *Builder {\n\tb.Url = url\n\tb.Method = http.MethodPut\n\treturn b\n}", "func (s *WebService) Put(route string, handler interface{}) {\r\n\ts.addRoute(route, \"PUT\", handler)\r\n}", "func (r *Request) PutForm(path string, data interface{}) {\n\tr.Put(path, \"application/x-www-form-urlencoded\", data)\n}", "func Put(config *HTTPConfig) (*HTTPResult, error) {\n\treturn HandleRequest(\"PUT\", config)\n}", "func (r *Router) PUT(pattern string, handler func(Context)) Route {\n\treturn r.Handle(\"PUT\", pattern, wrap(handler))\n}", "func IsPUT(r *http.Request) bool {\n\treturn strings.Compare(r.Method, \"PUT\") == 0\n}", "func (rg *RouteGroup) Put(path string, handlers ...Handler) *Route {\n\treturn rg.add(\"PUT\", path, handlers)\n}", "func Put(uri string) error {\n\treq, err := http.NewRequest(\"PUT\", Host+uri, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tres, err := Client.Do(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer res.Body.Close()\n\n\tif res.StatusCode != 204 {\n\t\treturn fmt.Errorf(\"got %d\", res.StatusCode)\n\t}\n\n\treturn nil\n}", "func Put(ctx context.Context, url string, body Body, options ...RequestOption) (*Response, error) {\n\tr, err := newRequest(ctx, http.MethodPut, url, body, options...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr.Header.Set(\"Content-Type\", body.ContentType())\n\treturn doRequest(http.DefaultClient, r)\n}" ]
[ "0.82787406", "0.7994318", "0.7698192", "0.7613152", "0.7606588", "0.75494", "0.7526452", "0.7382648", "0.73742276", "0.7360344", "0.735589", "0.73023254", "0.7295277", "0.7255101", "0.7236425", "0.72218215", "0.7220903", "0.72073925", "0.7194703", "0.7173302", "0.71618134", "0.71506995", "0.7142924", "0.7114693", "0.71133375", "0.7099384", "0.70991063", "0.70741767", "0.70646983", "0.70572865", "0.70505035", "0.70505035", "0.6991252", "0.69755644", "0.69719696", "0.6964554", "0.69634783", "0.69632566", "0.6945773", "0.6931826", "0.69221693", "0.6910999", "0.6909102", "0.68365324", "0.6829249", "0.6812156", "0.6801669", "0.67985475", "0.67925054", "0.6789405", "0.6769526", "0.6763695", "0.6761947", "0.6726006", "0.67154133", "0.6713628", "0.67086995", "0.6698083", "0.6691394", "0.66877127", "0.6679534", "0.66663885", "0.6658176", "0.66508794", "0.6641034", "0.6633264", "0.6615081", "0.65903634", "0.65814847", "0.6580279", "0.65649104", "0.65515816", "0.6548919", "0.653104", "0.652635", "0.6522962", "0.65187925", "0.65114015", "0.6506272", "0.6496171", "0.6490028", "0.6464805", "0.64541876", "0.64541876", "0.64315695", "0.642794", "0.6427653", "0.6420117", "0.6412596", "0.6365728", "0.6363343", "0.63616556", "0.6352011", "0.6324144", "0.63208926", "0.63090724", "0.63003916", "0.6299348", "0.6290136", "0.6283265" ]
0.880231
0
PATCH is a shorthand for e.Request("PATCH", path, pathargs...).
func (e *Expect) PATCH(path string, pathargs ...interface{}) *Request { return e.Request(http.MethodPatch, path, pathargs...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *Request) Patch(path string) *Request {\n\treturn r.method(\"PATCH\", path)\n}", "func (r *Router) PATCH(path string, handler RequestHandler) {\n\tr.setPath(httpPATCH, path, handler)\n}", "func (r *Router) PATCH(url string, viewFn View) *Path {\n\treturn r.Path(fasthttp.MethodPatch, url, viewFn)\n}", "func (r *Request) Patch(path, contentType string, data ...interface{}) {\n\tr.Send(\"PATCH\", path, contentType, data...)\n}", "func (r *Router) PATCH(path string, h HandlerFunc) {\n\tr.router.PATCH(path, r.handle(h, r.getValidationForPath(path, \"PATCH\")))\n}", "func (r *Router) PATCH(path string, handle HandlerFunc, middleware ...MiddlewareFunc) {\n\tr.Handle(\"PATCH\", path, handle, middleware...)\n}", "func (r *Router) PATCH(path string, handle Handle) {\n\tr.Handle(\"PATCH\", path, handle)\n}", "func (r *bitroute) PATCH(path string, f func(router.Control)) {\n\tr.register(\"PATCH\", path, f)\n}", "func (e *Engine) PATCH(path string, fn Controller) {\n\te.gin.PATCH(path, adapt(fn))\n}", "func (e *Engine) PATCH(path string, handler Handler) {\n\te.registerRoute(http.MethodPatch, path, handler)\n}", "func (tr *Transport) PATCH(\n\turl string,\n\tfn Handler,\n\toptions ...HandlerOption,\n) {\n\ttr.mux.Handler(\n\t\tnet_http.MethodPatch,\n\t\turl,\n\t\tnewHandler(fn, append(tr.options, options...)...),\n\t)\n}", "func (m *Middleware) PATCH(path string, handler Handler) {\n\tm.engine.PATCH(path, m.apply(handler))\n}", "func (r *RouterGroup) PATCH(path string, fn Controller) {\n\tr.gin.PATCH(path, adapt(fn))\n}", "func (serv *Server) PATCH(url string, handlers ...Handler) {\n\tserv.Handle(\"PATCH\", url, handlers...)\n}", "func (rg *RouteGroup) PATCH(path string, handlers ...Handler) *Route {\n\treturn rg.add(\"PATCH\", path, handlers)\n}", "func Patch(path string, fn http.HandlerFunc, c ...alice.Constructor) {\n\trecord(\"PATCH\", path)\n\n\tinfoMutex.Lock()\n\tr.PATCH(path, Handler(alice.New(c...).ThenFunc(fn)))\n\tinfoMutex.Unlock()\n}", "func PATCH(c *httputil.Client, data DataMultipartWriter, v interface{}, url string) error {\n\treturn Do(c, \"PATCH\", data, v, url)\n}", "func (router *Router) PATCH(relativePath string, handler Handler, decorators ...Decorator) {\n\trouter.createRouter(http.MethodPatch, relativePath, handler, \"\", decorators...)\n}", "func (r *Mux) PATCH(path string, handler http.HandlerFunc) {\n\tr.HandleFunc(\"PATCH\", path, handler)\n}", "func (api *Api) Patch(path string, endpoint http.HandlerFunc, queries ...string) {\n\tapi.Router.HandleFunc(path, endpoint).Methods(\"PATCH\").Queries(queries...)\n}", "func Patch(path string, fn http.HandlerFunc, c ...alice.Constructor) {\n\tinfoMutex.Lock()\n\trecord(\"PATCH\", path)\n\tr.Patch(path, alice.New(c...).ThenFunc(fn).(http.HandlerFunc))\n\tinfoMutex.Unlock()\n}", "func (engine *Engine) PATCH(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn engine.handle(http.MethodPatch, relativePath, handlers)\n}", "func (group *RouterGroup) PATCH(path string, handlers ...HandlerFunc) {\n\tgroup.Handle(\"PATCH\", path, handlers)\n}", "func (r *Router) PATCH(path string, handle Handle) {\n\tr.Handle(http.MethodPatch, path, handle)\n}", "func (node *Node) PATCH(functions ...interface{}) *Handler {\n\n\thandler := &Handler{}\n\n\tif len(functions) > 0 { handler.function = functions[0].(func (req web.RequestInterface) *web.ResponseStatus) }\n\n\tnode.addHandler(\"PATCH\", handler)\n\n\treturn handler\n}", "func Patch(url string, data ...interface{}) (*ClientResponse, error) {\n\treturn DoRequest(\"PATCH\", url, data...)\n}", "func (c *Client) Patch(url string, headers, queryParams map[string][]string, data interface{}) (response *http.Response, err error) {\n\treturn c.makeRequest(url, http.MethodPatch, headers, queryParams, data)\n}", "func (r *Request) Patch(url string) *Request {\n\tr.method = http.MethodPatch\n\tr.url = url\n\treturn r\n}", "func (rb *RequestBuilder) Patch(url string, body interface{}) *Response {\n\treturn rb.DoRequest(http.MethodPatch, url, body)\n}", "func (c *Client) Patch(ctx context.Context, url string, data ...interface{}) (*Response, error) {\n\treturn c.DoRequest(ctx, http.MethodPatch, url, data...)\n}", "func (c *Client) patch(rawURL string, authenticate bool, expectedStatus int, in interface{}, out interface{}) error {\n\terr := c.do(rawURL, \"PATCH\", authenticate, expectedStatus, in, out)\n\treturn errio.Error(err)\n}", "func (group *RouterGroup) PATCH(relativePath string, handlers ...HandlerFunc) IRoutes {\n\treturn group.handle(http.MethodPatch, relativePath, handlers)\n}", "func (cl *Client) Patch(c context.Context, url string, opts ...RequestOption) (*Response, error) {\n\treq, err := cl.NewRequest(c, http.MethodPatch, url, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn cl.Do(c, req)\n}", "func (r *Router) PATCH(pattern string, handler func(Context)) Route {\n\treturn r.Handle(\"PATCH\", pattern, wrap(handler))\n}", "func (c *Client) Patch(url string, reqBody, resType interface{}) error {\n\treturn c.CallAPI(\"PATCH\", url, reqBody, resType, true)\n}", "func (a *APITest) Patch(url string) *Request {\n\ta.request.method = http.MethodPatch\n\ta.request.url = url\n\treturn a.request\n}", "func (c *Client) Patch(rawurl string, in, out interface{}) error {\n\treturn c.Do(rawurl, \"PATCH\", in, out)\n}", "func Patch(path string, handler http.Handler) Route {\n\treturn NewRoute(\"PATCH\", path, handler)\n}", "func (r *Router) PATCH(route string, handler interface{}) *Router {\n\tr.Handle(MethodPATCH, route, handler)\n\treturn r\n}", "func (F *Frisby) Patch(url string) *Frisby {\n\tF.Method = \"PATCH\"\n\tF.Url = url\n\treturn F\n}", "func (g *Group) PATCH(path string, h Handler, gases ...Gas) {\n\tg.Air.PATCH(g.Prefix+path, h, append(g.Gases, gases...)...)\n}", "func (r *Request) Patch(url string) (*Response, error) {\n\treturn r.Execute(MethodPatch, url)\n}", "func (this *Route) PATCH(handlers ...HTTPHandler) *Route {\n\treturn this.addHandler(\"patch\", handlers...)\n}", "func (rg *RouteGroup) Patch(path string, handlers ...Handler) *Route {\n\treturn rg.add(\"PATCH\", path, handlers)\n}", "func Patch(path string, fn httprouter.Handle) {\n\tr.Router.PATCH(path, fn)\n}", "func (c *Client) Patch(url string, headers map[string][]string, body io.Reader) (client.Status, map[string][]string, io.ReadCloser, error) {\n\treturn c.Do(\"PATCH\", url, headers, body)\n}", "func (APIResourceBase) Patch(session *Session, url string, queries url.Values, body io.Reader) (APIStatus, interface{}) {\n\treturn FailSimple(http.StatusMethodNotAllowed), nil\n}", "func Patch(ctx context.Context, url string, body Body, options ...RequestOption) (*Response, error) {\n\tr, err := newRequest(ctx, http.MethodPatch, url, body, options...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr.Header.Set(\"Content-Type\", body.ContentType())\n\treturn doRequest(http.DefaultClient, r)\n}", "func (r *Router) Patch(path string, handler httpHandler) {\n\tr.PATCH(path, wrapHandler(handler))\n}", "func (r Requester) Update(path string, payload interface{}) Requester {\n\tb, err := json.Marshal(payload)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tbody := bytes.NewReader(b)\n\tr.httpRequest, err = http.NewRequest(http.MethodPut, r.url, body)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treturn r\n}", "func Patch(url string, opts ...RequestOption) (*Response, error) {\n\treturn DefaultSession.Patch(url, opts...)\n}", "func (mm *Middleman) Patch(path string, middleware Middleware) error {\r\n\treturn mm.addMiddleware(path, http.MethodPatch, middleware)\r\n}", "func (s *Nap) Patch(pathURL string) *Nap {\n\ts.method = MethodPatch\n\treturn s.Path(pathURL)\n}", "func Patch(route string, do interface{}) *handler {\n\treturn handlerByMethod(&route, do, \"PATCH\")\n}", "func (r *Route) Patch(h http.Handler) *Route {\n\tr.Add(\"PATCH\", h)\n\n\treturn r\n}", "func (tr *Transport) Patch(url string, fn HandlerFunc, options ...HandlerOption) {\n\ttr.mux.Handler(net_http.MethodPatch, url, encapsulate(fn, tr.options, options))\n}", "func (a *Router) Patch(pattern string, hs ...func(*Context) error) *Router {\n\treturn a.Handle(http.MethodPatch, pattern, hs...)\n}", "func (f5 *f5LTM) patch(url string, payload interface{}, result interface{}) error {\n\treturn f5.restRequestPayload(\"PATCH\", url, payload, result)\n}", "func (mr *MockRouterMockRecorder) PATCH(path, handler interface{}, options ...interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\tvarargs := append([]interface{}{path, handler}, options...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"PATCH\", reflect.TypeOf((*MockRouter)(nil).PATCH), varargs...)\n}", "func (m *MockRouter) PATCH(path string, handler interface{}, options ...interface{}) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{path, handler}\n\tfor _, a := range options {\n\t\tvarargs = append(varargs, a)\n\t}\n\tm.ctrl.Call(m, \"PATCH\", varargs...)\n}", "func (r *Request) PatchForm(path string, data interface{}) {\n\tr.Patch(path, \"application/x-www-form-urlencoded\", data)\n}", "func (ctl Controller) Patch(ctx *gin.Context) {\n\n\tMethodNotAllowedJSON(ctx)\n}", "func (r *Router) Patch(path string, handle HandleFunc) {\n\tr.register(path, http.MethodPatch, handle)\n}", "func (mr *MockRouterTxMockRecorder) PATCH(path, handler interface{}, options ...interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\tvarargs := append([]interface{}{path, handler}, options...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"PATCH\", reflect.TypeOf((*MockRouterTx)(nil).PATCH), varargs...)\n}", "func (m *MockRouterTx) PATCH(path string, handler interface{}, options ...interface{}) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{path, handler}\n\tfor _, a := range options {\n\t\tvarargs = append(varargs, a)\n\t}\n\tm.ctrl.Call(m, \"PATCH\", varargs...)\n}", "func PatchMethod(e *echo.Group, newFn newInstanceFn, pFn patchFn) {\n\te.PATCH(\"/:id\", func(c echo.Context) error {\n\t\tvar id int\n\n\t\tif err := Param(\"id\").InPath().Int(c, &id); err != nil {\n\t\t\treturn BadRequest(err)\n\t\t}\n\t\tresource := newFn()\n\t\tif err := c.Bind(resource); err != nil {\n\t\t\treturn BadRequest(err)\n\t\t}\n\n\t\tif err := pFn(resource); err != nil {\n\t\t\treturn InternalServerError(err)\n\t\t}\n\n\t\treturn c.NoContent(http.StatusNoContent)\n\t})\n\n}", "func (w *Worker) Patch(c *http.Client, url string, data interface{}, bind interface{}) (int, error) {\n\tbs, err := json.Marshal(data)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treq, err := http.NewRequest(\"PATCH\", url, bytes.NewReader(bs))\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tres, err := c.Do(req)\n\tif err != nil {\n\t\tif res != nil {\n\t\t\tioutil.ReadAll(res.Body)\n\t\t\tres.Body.Close()\n\t\t}\n\t\treturn 0, err\n\t}\n\tdefer res.Body.Close()\n\terr = json.NewDecoder(res.Body).Decode(bind)\n\tif res.StatusCode == http.StatusNoContent || bind == nil {\n\t\treturn res.StatusCode, nil\n\t}\n\treturn res.StatusCode, err\n}", "func (b *Baa) Patch(pattern string, h ...HandlerFunc) RouteNode {\n\treturn b.Router().Add(\"PATCH\", pattern, h)\n}", "func (r *Router) Patch(pattern string, handler http.Handler) Route {\n\treturn r.Handle(\"PATCH\", pattern, handler)\n}", "func OptPatch() Option {\n\treturn RequestOption(webutil.OptPatch())\n}", "func (conn Connection) Patch(cmd string, content, result interface{}) (resp *http.Response, err error) {\n\treturn conn.Send(http.MethodPatch, cmd, content, result)\n}", "func (m *Mux) Patch(pattern string, handler HandlerFunc) {\n\tm.Handle(http.MethodPatch, pattern, handler)\n}", "func (avisess *AviSession) Patch(uri string, payload interface{}, patchOp string, response interface{}, options ...ApiOptionsParams) error {\n\tvar patchPayload = make(map[string]interface{})\n\tpatchPayload[patchOp] = payload\n\tglog.Infof(\" PATCH OP %v data %v\", patchOp, payload)\n\treturn avisess.restRequestInterfaceResponse(\"PATCH\", uri, patchPayload, response, options...)\n}", "func Patch(url string, requestBody []byte, requestHeaders map[string]string, authHeader string) (int, []byte, map[string][]string) {\n\tlogs.Alert(fmt.Sprintf(\"<------------------------- %s ------------------------->\\n\", \"start\"))\n\tlogs.Alert(fmt.Sprintf(\"[HTTP PATCH: %s]\", url))\n\n\tr, _ := http.NewRequest(\"PATCH\", url, bytes.NewBuffer(requestBody))\n\n\tif authHeader != \"\" {\n\t\tr.Header.Add(\"Authorization\", authHeader)\n\t}\n\n\tfor name, value := range requestHeaders {\n\t\tr.Header.Add(name, value)\n\t}\n\n\tclient := &http.Client{Transport: transport, Timeout: timeout}\n\tresp, err := client.Do(r)\n\n\tif err != nil {\n\t\tlogs.Alert(fmt.Sprintf(\"[Request Error --> %v]\", err))\n\t\treturn http.StatusInternalServerError, nil, nil\n\t}\n\n\tdefer resp.Body.Close()\n\tjsonBytes, _ := ioutil.ReadAll(resp.Body)\n\tlogs.Alert(fmt.Sprintf(\"[Status Code --> %d]\", resp.StatusCode))\n\tlogs.Alert(fmt.Sprintf(\"[Response Headers --> %v]\", resp.Header))\n\tlogs.Alert(fmt.Sprintf(\"[Response --> %s]\", string(jsonBytes)))\n\tlogs.Alert(fmt.Sprintf(\"<------------------------- %s ------------------------->\", \"end\"))\n\treturn resp.StatusCode, jsonBytes, resp.Header\n}", "func (conn Connection) Patch(cmd string, content, result interface{}) (effect *SideEffect, resp *http.Response, err error) {\n\treturn conn.Send(http.MethodPatch, cmd, content, result)\n}", "func (r *Route) Patch(h interface{}) *Route {\n\treturn r.Handle(toHandler(h), \"PATCH\")\n}", "func (r *Request) PatchJSON(path string, data interface{}) {\n\tb, err := json.Marshal(data)\n\tif err != nil {\n\t\tr.t.Fatalf(\"httptesting: PatchJSON:json.Marshal(%T): %v\", data, err)\n\t}\n\n\tr.Patch(path, \"application/json\", b)\n}", "func (r *Route) Patch(handler http.Handler) *Route {\n\tr.handlers[http.MethodPatch] = handler\n\treturn r\n}", "func (res *Resource) Patch(storage store.Update) {\n\tres.HandleFuncC(\n\t\tpat.Patch(patID),\n\t\tfunc(ctx context.Context, w http.ResponseWriter, r *http.Request) {\n\t\t\tres.patchHandler(ctx, w, r, storage)\n\t\t},\n\t)\n\n\tres.addRoute(patch, patID)\n}", "func (a *APITest) Patchf(format string, args ...interface{}) *Request {\n\treturn a.Patch(fmt.Sprintf(format, args...))\n}", "func (r *ExternalRequest) Update(ctx context.Context, reqObj *External) error {\n\treturn r.JSONRequest(ctx, \"PATCH\", \"\", reqObj, nil)\n}", "func TestPatch(t *testing.T) {\n\tconst case1Empty = \"/\"\n\tconst case2SetHeader = \"/set_header\"\n\tconst case3SendJSON = \"/send_json\"\n\tts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t// check method is PATCH before going to check other features\n\t\tif r.Method != PATCH {\n\t\t\tt.Errorf(\"Expected method %q; got %q\", PATCH, r.Method)\n\t\t}\n\t\tif r.Header == nil {\n\t\t\tt.Errorf(\"Expected non-nil request Header\")\n\t\t}\n\t\tswitch r.URL.Path {\n\t\tdefault:\n\t\t\tt.Errorf(\"No testing for this case yet : %q\", r.URL.Path)\n\t\tcase case1Empty:\n\t\t\tt.Logf(\"case %v \", case1Empty)\n\t\tcase case2SetHeader:\n\t\t\tt.Logf(\"case %v \", case2SetHeader)\n\t\t\tif r.Header.Get(\"API-Key\") != \"fookey\" {\n\t\t\t\tt.Errorf(\"Expected 'API-Key' == %q; got %q\", \"fookey\", r.Header.Get(\"API-Key\"))\n\t\t\t}\n\t\tcase case3SendJSON:\n\t\t\tt.Logf(\"case %v \", case3SendJSON)\n\t\t\tdefer r.Body.Close()\n\t\t\tbody, _ := ioutil.ReadAll(r.Body)\n\t\t\tif string(body) != `{\"query1\":\"test\",\"query2\":\"test\"}` {\n\t\t\t\tt.Error(`Expected Body with {\"query1\":\"test\",\"query2\":\"test\"}`, \"| but got\", string(body))\n\t\t\t}\n\t\t}\n\t}))\n\n\tdefer ts.Close()\n\n\tNew().Patch(ts.URL + case1Empty).\n\t\tEnd()\n\n\tNew().Patch(ts.URL+case2SetHeader).\n\t\tSetHeader(\"API-Key\", \"fookey\").\n\t\tEnd()\n\n\tNew().Patch(ts.URL + case3SendJSON).\n\t\tSendMapString(`{\"query1\":\"test\"}`).\n\t\tSendMapString(`{\"query2\":\"test\"}`).\n\t\tEnd()\n}", "func (p *PatternServeMux) Patch(pat string, h http.Handler) {\n\tp.Add(\"PATCH\", pat, h)\n}", "func (k *kubectlContext) Patch(args ...string) error {\n\tout, err := k.do(append([]string{\"patch\"}, args...)...)\n\tk.t.Log(string(out))\n\treturn err\n}", "func (c *BaseController) Patch(w http.ResponseWriter, r *Request) {\n\thttp.Error(w, \"Method Not Allowed\", 405)\n}", "func (r *FakeClient) Patch(\n\tctx context.Context, obj client.Object, patch client.Patch, opts ...client.PatchOption,\n) error {\n\t// TODO (covariance) implement me!\n\tpanic(\"not implemented\")\n}", "func (r *versionResolver) Patch(ctx context.Context, obj *restModel.APIVersion) (*restModel.APIPatch, error) {\n\tif !evergreen.IsPatchRequester(*obj.Requester) {\n\t\treturn nil, nil\n\t}\n\tapiPatch, err := data.FindPatchById(*obj.Id)\n\tif err != nil {\n\t\treturn nil, InternalServerError.Send(ctx, fmt.Sprintf(\"Couldn't find a patch with id '%s': %s\", *obj.Id, err.Error()))\n\t}\n\treturn apiPatch, nil\n}", "func (e *Expect) PUT(path string, pathargs ...interface{}) *Request {\n\treturn e.Request(http.MethodPut, path, pathargs...)\n}", "func (mx *Mux) Patch(pattern string, handlerFn http.HandlerFunc) {\n\tmx.handle(mPATCH, pattern, handlerFn)\n}", "func (c *Controller) Patch() {\n\thttp.Error(c.Ctx.ResponseWriter, \"Method Not Allowed\", http.StatusMethodNotAllowed)\n}", "func Patch(pattern string, handler func(Context)) {\n\tmux.HandleFunc(pattern, func(w http.ResponseWriter, r *http.Request) {\n\t\tif r.Method != \"PATCH\" {\n\t\t\thttp.NotFound(w, r)\n\t\t\treturn\n\t\t}\n\n\t\thandler(Context{\n\t\t\tResponse: w,\n\t\t\tRequest: r,\n\t\t})\n\t})\n}", "func (router *Router) Patch(path string, handler Handle) *Router {\n\trouter.Mux.PATCH(path, handleProxy(handler))\n\treturn router\n}", "func (fkw *FakeClientWrapper) Patch(ctx context.Context, obj runtime.Object,\n\tpatch k8sCl.Patch, opts ...k8sCl.PatchOption) error {\n\treturn fkw.client.Patch(ctx, obj, patch, opts...)\n}", "func (c *SubresourceClient) Patch(namespace string, name string, data []byte) (e error) {\n\tif c.Error != \"\" {\n\t\te = fmt.Errorf(c.Error)\n\t}\n\treturn\n}", "func PatchBytes(url string, data ...interface{}) []byte {\n\treturn RequestBytes(\"PATCH\", url, data...)\n}", "func (client *MockClient) Patch(context ctx.Context, object ctrlClient.Object, patch ctrlClient.Patch, options ...ctrlClient.PatchOption) error {\n\treturn fmt.Errorf(\"Not implemented\")\n}", "func (r *ExtensionRequest) Update(ctx context.Context, reqObj *Extension) error {\n\treturn r.JSONRequest(ctx, \"PATCH\", \"\", reqObj, nil)\n}", "func (a *App) Patch(w http.ResponseWriter, r *http.Request) {\n\tw.Write([]byte(\"patching in testing mode. Get ready to send multipart-form data\"))\n}", "func (*HttpCffAction) PatchPath(ref string) string {\n\treturn fmt.Sprintf(\"/api/objects/http/cff_action/%s\", ref)\n}", "func (client HTTPSuccessClient) Patch202(booleanValue *bool) (result autorest.Response, err error) {\n req, err := client.Patch202Preparer(booleanValue)\n if err != nil {\n return result, autorest.NewErrorWithError(err, \"httpinfrastructuregroup.HTTPSuccessClient\", \"Patch202\", nil , \"Failure preparing request\")\n }\n\n resp, err := client.Patch202Sender(req)\n if err != nil {\n result.Response = resp\n return result, autorest.NewErrorWithError(err, \"httpinfrastructuregroup.HTTPSuccessClient\", \"Patch202\", resp, \"Failure sending request\")\n }\n\n result, err = client.Patch202Responder(resp)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"httpinfrastructuregroup.HTTPSuccessClient\", \"Patch202\", resp, \"Failure responding to request\")\n }\n\n return\n}" ]
[ "0.8082939", "0.79485875", "0.7912536", "0.7824908", "0.763242", "0.76136535", "0.7507971", "0.7493006", "0.7487783", "0.74693745", "0.7466364", "0.74389553", "0.74365085", "0.7419845", "0.74184185", "0.7400768", "0.7369115", "0.7331139", "0.7320122", "0.73052347", "0.72978497", "0.7279466", "0.7256477", "0.7235181", "0.71369684", "0.71277636", "0.7123945", "0.7122819", "0.7110203", "0.70212674", "0.7000825", "0.6986983", "0.6974243", "0.6951236", "0.69376606", "0.6933343", "0.6923582", "0.69046843", "0.6896538", "0.6891865", "0.6875577", "0.6856909", "0.67946696", "0.6782048", "0.67761827", "0.6701242", "0.66883475", "0.6688292", "0.66098213", "0.6592049", "0.65908647", "0.65868753", "0.6577811", "0.6566974", "0.6542181", "0.65197635", "0.64779514", "0.6470891", "0.64128727", "0.64037466", "0.63904923", "0.6382211", "0.6359057", "0.6357", "0.63515407", "0.6339525", "0.629366", "0.62833375", "0.62475675", "0.62325436", "0.62140095", "0.6211714", "0.61983925", "0.6187708", "0.6164913", "0.6163976", "0.6156128", "0.6152", "0.6144881", "0.6138464", "0.6112108", "0.61105466", "0.609833", "0.6077597", "0.6069947", "0.60429096", "0.6040407", "0.6038182", "0.60255843", "0.60138416", "0.600529", "0.5986722", "0.5984741", "0.5978965", "0.5955159", "0.5943534", "0.5934052", "0.5923601", "0.5912267", "0.587179" ]
0.8773625
0